Meso-Scale Deployment of OpenFlow: 1. Clemson University, this project will deploy an OpenFlow testbed on the Clemson University campus and connect with wireless mesh access points and mobile terminals. This trial will conduct OpenFlow experimentation focused on OpenFlow enabled network operation solutions as a precursor to deployment into Clemson research and production networks. Toroki LS4810 switch 2. Georgia Tech, this project will deploy an OpenFlow testbed in the CS building on the Georgia Tech Campus. This project will conduct OpenFlow experimentation focused on OpenFlow enabled next generation network admission control system as a precursor to deployment into campus residential network. The testbed will replicate a campus deployment required to handle thousands of end hosts. Previously, the GTech network admission control system is a VLAN based approach, but there are several disadvantages for VLAN based access control. For example, the number of VLAN is limited, and the control overhead to set the private and public IPs for the authorization is big. GTech OpenFlow deployed OpenFlow switches for campus access control. All the hosts will go through OpenFlow switches to a centralized controller for authorization. If it is authorized, it can access the network. Otherwise, it will be quarantined. 3. Indiana University, this project will deploy an OpenFlow testbed between two campuses. This trial will conduct OpenFlow experimentation focused on OpenFlow enabled dynamic provisioning and distributed monitoring as a precursor to deployment into the Indiana University campus production network. It will also examine and propose ways for network operators to manage and support their networks using Openflow enabled devices. HP Procurve 5406/6600 Distributed monitoring: cSamp implementation (Coordinated Sampling) – a measurement framework for network-wide flow monitoring. Idea from CMU, IU wants to implement on OpenFlow. Current flow monitoring uses uniform packet sampling based on a sampling probability. And current routers record flow measurements completely independently of each other, thus leading to redundant flow measurements and inefficient use of router resources. So they argue that a centralized system that coordinates monitoring across different routers can significantly increase the flow monitoring capabilities of a network. Three improvements: Use flow sampling instead of uniform packet sampling; hash-based packet selection to eliminate duplicate measurements in the networks; framework for distributing responsibilities across routers to achieve network-wide monitoring goals while respecting router resource constraints. Dynamic provisioning Deploy an inter-Campus Networking, between IU and IUPUI, investigating how OpenFlow will interoperate with existing inter and intra campus provisioning tools such as Sherpa (used on NLR): GMPLS, tools such as the Internet2 ION service. IPTV – Worked on design with IU Campus Network Architecture group for using OpenFlow as a mechanism for fine-grained control of IPTV related traffic flows. OpenFlow would be used for Layer 2 switching decisions to avoid issues of multicast flooding and non-optimal topologies. Data capture – Collaboration with Minaxi Gupta (IU) on designs for using OpenFlow as a mechanism to capture cybersecurity data sets on campus networks. 4. Rutgers, this project will deploy OpenFlow switches between wireless research networks and Internet2 to enable researcher programmability and virtualization in the wired network connecting the Rutgers wireless assets to the GENI backbone. OpenFlow switches in the ORBIT indoor mesh testbed will also enhance performance and flexibility and provide support for grid virtualization. NEC IP 8800 5. UW-Madison, this project will deploy an OpenFlow testbed throughout the University of Wisconsin campus. This project will develop OpenFlow extensions focused on the creation of flexible, heterogeneous slices, the development of a generic load balancer and a captive portal for authentication. This work is a precursor to the deployment of OpenFlow into campus production networks. We have designed the OpenSAFE system and an accompanying language, ALARMS, for scalable and flexible network monitoring. OpenSAFE and ALARMS are based on OpenFlow. The system can perform a variety of network monitoring functions at line-rates. It also allows flexible interposition of middleboxes as well as fine-grained load balancing across a diverse collection of hardware. This system was demo-ed at GEC7, where it garnered a lot of interest from other GENI participants. HP OpenSAFE: Problem: How to route the traffic for network analysis in a robust, high performance manner that does not impact normal network traffic? They focus on network monitoring via inspection of copies of traffic at interesting points in the network. Current “copies of traffic”: 1. Span ports are typically directed into a single computer running some sort of IDS (Intrusion Detection System), 2. Span ports on network equipment is often extremely limited. OpenSAFE can direct spanned network traffic in arbitrary ways. OpenSAFE can handle any number of network inputs and manage the traffic in such a way that it can be used by many services while filtering packets at line rate. Design: The input is a connection from the span path at the chosen network aggregation point to an OpenFlow switch port. Some number of filters are in use, attached to various OpenFlow switch ports. Finally, output is directed into some number of sinks. 6. University of Washington This project will deploy a hybrid OpenFlow and RouteBrick testbed within the computer science and engineering department. This project will develop building blocks allowing researchers to investigate the placement of middlebox functionality in enterprise networks. This work is a precursor to the deployment of OpenFlow and RouteBricks into campus production networks. Our goal is to build a testbed and develop building blocks that would allow researchers to investigate the optimal placement of middlebox functionality in today's enterprise networks. We have begun the deployment of OpenFlow enabled HP Procurve switches in the computer science building here at University of Washington. The deployment is initially being used by a small number of graduate students in systems and networking for both research and day-to-day use. Eventually it will be phased in as the primary network for a larger set of users. After phase in, we will continue to operate a research network so that researchers can test their ideas before validating them on the production system. HP ProCurve 6600 7. Princeton University This project will deploy an OpenFlow testbed within the CS self-operated IT department. This trial will conduct OpenFlow experimentation focused on OpenFlow enabled management of experimental and production services as a precursor to deployment into their department's production network. HP ProCurve switches Do not have any documents! 8. Stanford University This project will deploy an OpenFlow, testbed throughout the Stanford University campus. This project will develop OpenFlow extensions focused on the creation of flexible, heterogeneous slices, the development of a generic load balancer and a captive portal for authentication. This work is a precursor to the deployment of OpenFlow into campus production networks. Eight OpenFlow switches (HP, NEC, Quanta, and Toroki) Developer of OpenFlow!!!!!! Including a) OpenFlow reference implementation, b) FlowVisor, c) SNAC Controller, and d) Debugging/monitoring tools. 9. Internet2 Openflow Internet2 will implement Stage 0 of the Internet2 OpenFlow test network by creating a five node OpenFlow test network using small enterprise switches (HP ProCurve 5406 switches proposed, pending evaluation). The initial design has 10GE WAN interfaces, but that is subject to negotiation of the MOU Internet2 has with GENI for wide-area connectivity. Stage 0 comprises the appropriate OpenFlow controller(s), interfaces to connect several enterprise OpenFlow campuses, and interfaces to connect other projects such as ProtoGENI and SPP. The Stage 0 network will allow members and non-members of Internet2 to connect to GENI OpenFlow services. Internet2 will create OpenFlow nodes in five cities: Atlanta, Chicago, Houston, New York, and Salt Lake City. Internet2 will also explore integrating the OpenFlow nodes with OpenFlow deployments on ProtoGENI nodes. In this second, complementary-to-ProtoGENI scenario, Internet2 will create OpenFlow nodes in five cities: Atlanta, Chicago, Los Angeles, New York, and Seattle. 10. National LambdaRail Openflow Deploy and operate OpenFlow-enabled HP Procurve 6600 switches at 5 NLR POPs and interconnect NLR's FrameNet network to the GENI OpenFlow-enabled backbone, permitting members and non-members of NLR to connect to GENI OpenFlow services. OpenFlow at LSU This project will deploy an OpenFlow testbed in Frey Building on the LSU campus. The purpose of OpenFlow deployment at LSU is to help both experiment traffic and production traffic to make better use of our existing high speed network research testbed. We have already established an up to 10Gbps high speed network research testbed CRON (Cyberinfrastructure of Reconfigurable Optical Network) at LSU. For the experiment traffic, CRON has already established a 10Gbps connection to other ProtoGENI sites on the GENI backbone through Internet2 ION service. According to the final plan of GENI, the national backbones will be built out with OpenFlow. As a ProtoGENI site at LSU, we are responsible to provide OpenFlow capabilities on our ProtoGENI testbed, so that in the future all the ProtoGENI sites will have significant commonality in the technologies employed, and significant overlap in the sets of campuses. Also CRON is connected to Louisiana Optical Network Initiative (LONI) over 10Gbps links. Inside LONI network, there will be several Louisiana colleges and universities, such as Southern University, New Orleans University, University of Louisiana at Lafayette, and so on, which will use the high speed computing resources of CRON at LSU. Of course, these universities will be allowed to use CRON testbed for their high speed network research experiments the same as other ProtoGENI sites. Moreover, there will be some high speed production traffics requirements between these universities, for instance, the remote visualization application. These high speed production traffics may want to use the high speed hardware/software in CRON testbed. In that case, OpenFlow will be used for dynamically mapping the production traffic to the specific high speed hardware/software they required. For example, if one university wants to use the hardware emulator to create some specific delay or rate limit, the controller in OpenFlow can simply forward the traffic to the hardware emulator which they already configured by the software inside CRON. The controller can distinguish different requirements through the TCP ports or the VLAN id in the flow table. Besides, instead of just use VLAN to distinguish all the different experiment traffics and production traffics, we will choose to use OpenFlow. One reason is that the number of VLAN is limited. Another reason is that it is more flexible for OpenFlow to choose all the infrastructures inside CRON than the fixed VLANs, and OpenFlow can easily choose infrastructures since all the infrastructures have the uniform private IP. h
© Copyright 2026 Paperzz