Gaining Control of Virtualized Server Environments By Jim Metzler, Ashton Metzler & Associates Distinguished Research Fellow and Co-Founder, Webtorials Editorial/Analyst Division Introduction The traditional data center network was based on an architecture and design tailored to satisfy the connectivity requirements of a relatively simple server infrastructure based on the following set of assumptions: Each server was dedicated to a single application and had a single IP address and a single MAC address. Nearly all of the traffic within the data center was client/server related. This type of traffic is often referred to as “north-south” traffic. The relatively small amount of traffic that went between servers was readily visible on physical switch ports. This type of traffic is often referred to as “east-west” traffic. The physical separation of servers satisfied most of the requirements for security and for the delineation of departmental resources. The static nature of the application environment lent itself to manual processes for server deployment or redeployment. VLANs tended to be used primarily for traffic isolation and to provide high availability between the access and distribution tiers. As a result, VLANs tended to be localized within application pods. There were clear lines of demarcation between the server and application management domain and the network management domain. Interestingly enough, the implementation of server virtualization has negated the validity of each of the assumptions listed above. As a result, the traditional approach to designing and managing data center networks is no longer valid. The goal of this white paper is to discuss what IT organizations must do to successfully meet the networking challenges associated with a highly dynamic, virtualized data center. In order to achieve this goal, three IT professionals were interviewed. One of the interviewees was the senior manager of global network design for a major health care organization. His organization doesn’t allow him to be explicitly named in a document such as this, so he will be referred to in this white paper as The Senior Gaining Control of Virtualized Server Environments June 2011 Page 1 Manager. The other two interviewees were Aaron Larkins, the CEO of Profitability.net and Charles Hollingsworth, the manager of IT Architectural Planning at Georgia State University. The Adoption of Server Virtualization There are numerous benefits that can be derived from server virtualization, including: Server Consolidation Once it has been virtualized, a single physical server can support multiple virtual machines (VMs) and as a result, applications that would normally require a dedicated server can now share a single physical server. This enables IT organizations to reduce the number of servers in their data centers which yields a significant savings in both CAPEX (i.e., costs of server hardware, SAN Host bus adapters, and Ethernet NICs) and OPEX; i.e., server management labor expense, plus facility costs such as power, cooling, and floor space. Flexible Server and Application Provisioning A production VM can be transferred without service interruption from a given physical server to a different physical server, either within the same data center or between different data centers. This capability enables workload management and optimization across an IT organization’s virtualized data center(s). This capability also helps IT organizations meet business objectives by: o Streamlining the provisioning of new applications. o Improving backup and restoration operations. o Enabling zero-downtime maintenance. o Enable Other Initiatives In order to more effectively manage and control user desktops, many IT organizations are implementing virtual desktops. Virtualized servers are a prerequisite for virtualized desktops because a virtual desktop implementation requires that a VM in a data center server hosts a complete user desktop including all its applications, configurations, and privileges. A recent market research report1 contained the results of a survey that asked IT professionals about their organization’s adoption of server virtualization and the impact of that adoption on networking and management. Throughout this white paper that report will be referred to as The Research Report. The Research Report presented the responses to a survey question in which the respondents were asked to indicate the percentage of their company’s data center servers that have either already been virtualized or that they expected would be virtualized within the next year. Their responses are shown in Table 1. 1 http://www.webtorials.com/content/2010/11/2010-cloud-networking-report.html Gaining Control of Virtualized Server Environments June 2011 Page 2 None 1% to 25% 26% to 50% 51% to 75% 76% to 100% Have already been virtualized 21.6% 33.0% 18.9% 15.1% 11.3% Expect to be virtualized within a year 12.4% 25.6% 21.9% 21.9% 18.2% Table 1: Deployment of Virtualized Servers2 One observation that can be drawn from the data in Table 1 is that over the next year the deployment of server virtualization will continue to increase. For example, over the next year the percentage of IT organizations that will have virtualized the majority of their data center servers is anticipated to almost double. Georgia State University is an example of an organization that has made only a moderate deployment of server virtualization as they have virtualized roughly one third of their servers. The company that The Senior Manager works for is an example of an organization that has been very aggressive with implementing server virtualization. That company has already virtualized roughly eighty percent of their five thousand servers and their goal is to get as close as they can to having one hundred percent of their servers virtualized. The Senior Manager stated that they will not virtualize all of their servers in the near term in part because in some instances porting the code from a physical server to a VM is too expensive and in part because of the compliance concerns they have relative to some of their healthcare applications. As previously noted, once a server has been virtualized it can support multiple VMs. One measure of the efficiency, and hence the cost savings of server virtualization, is the average number of VMs on a single physical server. This number is typically referred to as the consolidation ratio. In the current environment, organizations are typically implementing consolidation ratios between 5:1 and 10:1. The consolidation ratio is expected to grow rapidly as IT departments become more familiar with virtualization technology and as they deploy multi-core server processors that are optimized for virtualization. The company that The Senior Manager works for is an example of a company that is driving cost savings by implementing a high concentration ratio. According to The Senior Manager they currently average fifty VMs per physical server and that number is likely to increase in part due to the fact that they are aggressively implementing 10 Gbps ports. IT organizations, however, will only be able to appreciate these added cost savings if they redesign their data center networks and change their approach to management. The Research Report indicated that most IT organizations understood that and were making progress in those directions. For example, The Research Report presented the responses to a survey question in which the respondents were asked “Has your IT organization already redesigned, or within the next year will it redesign, its data center LAN in order to support cloud computing in general, and virtualized servers in particular?” Those responses are shown in Table 2. 2 http://www.webtorials.com/content/2009/12/cloud-computing-a-reality-check-guide-to-risk-mitigation.html Gaining Control of Virtualized Server Environments June 2011 Page 3 Already Have Cloud Computing in General Virtualized Servers in Particular No Plans 28.6% Will Within the Next Year 42.9% 50.5% 30.7% 18.8% 28.6% Table 2: Redesign of the Data Center LAN The data in Table 2 indicates the significant impact that server virtualization is having on the design of data center networks. As discussed in The Research Report, some of the primary design changes that IT organizations are making to their data center networks in order to better support server virtualization include implementing: A 2-Tier network topology to improve traffic flows among servers. Edge Virtual Bridging (EVB) to provide scalable I/O performance and full visibility of VM-VM traffic on the same physical server. The combination of multi-chassis LAGs and virtual LAN switching to improve performance and availability via support for active-active Layer 2 switching and load balancing across redundant source/destination paths. The Research Report also indicated that most IT organizations understand the importance of modifying how they manage in a virtualized environment. In particular, The Research Report presented the responses to a survey question in which the survey respondents were asked to indicate how important it is to their organization to get better over the next year at managing some key tasks related to server virtualization. They were given the following five-point scale: 1. 2. 3. 4. 5. Not at all important Slightly important Moderately important Very important Extremely important Included in Table 3 are the tasks and the percentage of the survey respondents who indicated that the task was either very or extremely important for their organization to get better at over the next year. Gaining Control of Virtualized Server Environments June 2011 Page 4 Server Virtualization Management Task Perform traditional management tasks such as troubleshooting and performance management on a per VM basis Keep track of VMs as they are moved between physical servers Dynamically move VMs, and all of the supporting management functionality, between physical servers Discover VMs Manage the traffic that goes between VMs on a single physical server Importance: Very or Extremely 49% 38% 37% 33% 31% Table 3: Importance of Managing Server Virtualization As shown in Table 3, the survey respondents indicated that getting better at each of the individual challenges associated with server virtualization is important to their organization. In addition, it is reasonable to look at the five challenges contained in Table 3 as being a single challenge - managing server virtualization. When looked at that way, getting better at server virtualization is extremely important to IT organizations. Managing at the Virtual Machine Level The Senior Manager commented that server virtualization is “great when it works” but that they struggle when it doesn’t. When there is a problem caused by server virtualization, the approach that they take is to get all of the appropriate personnel into a war room and “tear apart the problem”. He added that, “this can take hours and can affect business”. Hollingsworth concurred and said that, “When looked at overall, server virtualization is wonderful, but there are some legitimate issues.” He added that server virtualization increases complexity and “makes it difficult to know what is running where and when.” Below are some techniques and functionality that IT organizations can implement to respond to the challenges that are associated with server virtualization. Network Configuration at the Virtual Machine or Virtual Port Level In the traditional data center, physical servers were typically dedicated to a single application. Therefore, network policies could be implemented by configuring the physical port of the network access switch with the Access Control Lists (ACLs), Quality of Service (QoS) classifications, and the rate-limiting characteristics required for specific applications. With virtualization, multiple disparate applications can be running on VMs sharing a single physical server. IT organizations need the ability to manage network traffic using the same granular policies relating to security, QoS, and regulatory compliance for applications running on VMs as it did previously for applications running on dedicated physical servers. This can be accomplished if the network access switch is capable of mapping each VM to a separate, individually configurable virtual switch port. The network access switch management system can then maintain a profile for each VM, including the VM’s IP address and virtual MAC address, VLAN memberships, ACLs and QoS classification. Gaining Control of Virtualized Server Environments June 2011 Page 5 The benefits of virtual ports can be maximized in those situations in which the physical network access switch’s operating system includes support for switching all VM-to-VM traffic even when the source and destination VMs are residing on the same physical server. With this form of virtualized switching at the network edge, the physical access switch processes all VM traffic rather than having VM-to-VM traffic within the same physical server being switched by a hypervisor’s software based virtual switch (vSwitch). Virtual port profiles and edge virtualized switching are two subjects that are currently being addressed by the IEEE 802.1Qbg working group which is expected to deliver ratified standards within the next two years. The standards will include specifications for a Virtual Ethernet Port Aggregator (VEPA) in external access switches and a Virtual Station Interface (VSI) Protocol to associate and deassociate VM MAC Addresses with a virtual port profile. The way that VEPA will work is that it will offload networking tasks from the server’s vSwitch, which will in turn maximize the server’s available CPU resources to support higher consolidation ratios. Virtual port capability, together with VEPA, also allows existing traffic monitoring and troubleshooting tools to gain visibility into all the traffic flows for each VM. This visibility makes it possible for IT organizations to continue to manage the availability and performance of applications. Additional benefits of VEPA include high performance, low latency, hardware-based switching for all VM traffic and the ability to minimize the number of switching tiers in the data center by potentially eliminating the hypervisor and blade server switching tiers. In heterogeneous hypervisor server environments, VEPA also eliminates the need to support the multiple vSwitch architectures and management tools that result from using multiple vendors, further reducing complexity. The use of VEPA can also eliminate the organizational and economic stress that so many companies have experienced as both the network and the server organizations claimed or disclaimed responsibility or were forced to increase staffing to manage the multitude of vSwitches. From the point of view of the network, VEPA functionality can be implemented on most switches as a relatively simple firmware upgrade rather than requiring new ASICs or other hardware. This means that data center networks can deploy pre-standard VEPA without running the risk of having to replace their access switches when the standard is finalized. A number of vendors already support pre-standard VEPA in their switches. However, for VEPA to work as intended, hypervisor vendors will need to provide software support for the server side of VEPA. While this also appears to be a relatively easy modification, the hypervisor vendors seem somewhat reluctant to embrace a pre-standard switching solution that can bypass their software switches. The Senior Manager stated that they are undergoing a proof of concept trial relative to using an external switch to switch between VMs. He added that in order to reduce organizational stress that his organization is discussing the creation of a data center group that is comprised of personnel who have a background in networking, servers, applications, storage and load balancing. Hollingsworth discussed some of the organizational challenges associated with server virtualization. One example he gave is the coordination needed between the network engineering group and the platform engineering group to make sure that vLANs are properly exposed on network equipment, virtual switches and virtual servers. This requires a degree of coordination and synchronization of group priorities that was not previously required. For example, it is necessary to make sure the synchronization of vLANs is not configured just for the initial host the VM is placed on, but also for any host where the VM might be moved. Hollingsworth added that debugging issues are also much more challenging. For example, in a conventional network connection the network engineering group can deploy packet capture and analysis tools between the host and the nearest switch. That is not always possible within a virtualized server environment where communications between VMs may never leave a physical server. In these cases, the network engineering group is dependent on the platform engineering group either capturing and supplying the appropriate management data for analysis, or Gaining Control of Virtualized Server Environments June 2011 Page 6 moving the appropriate VM from one server to another so that the traffic flow is forced to route to the physical network where network engineering can capture and analyze the data. The complexity of these issues are leading both the platform engineering group and the network engineering group to realize that the virtual switch provides an unnecessary layer of abstraction that doesn’t help either engineering group. Dynamic Policy In addition to enabling server consolidation, another major benefit of server virtualization is derived from the relative ease with which a VM can be provisioned or migrated among physical servers. This enables the creation of a dynamic server environment in which the compute capacity devoted to an application can expand and contract in real time to meet fluctuations in user demand, and virtual servers can be easily relocated to balance load on physical server platforms or to ensure higher application availability. The Senior Manager stated that in order to support patch management and to ensure adequate application performance that they move VMs between servers in the same data center on a weekly basis. He stated that the majority of these moves are automated and that they would like to move VMs between servers in disparate data centers but that they are constrained both by the requirement to not cross a layer 3 network boundary as well as the difficulty associated with moving the associated storage. Larkins commented that they automatically move VMs between servers within their data center in response to a server failure. They also move VMs manually in order to improve performance. He added that he expects to more fully automate the movement of VMs in the near future as they support more applications built for cloud computing which include APIs that enable appropriate resources to be dynamically added or removed. When a VM is provisioned or migrated to a new physical server, changes in the configuration of the network are required to ensure appropriate policies are enforced for the application that is running in the VM. VMs can be created or moved in a matter of seconds or minutes but it can take hours or even days to implement the appropriate network configuration if manual processes are employed. The delays associated with reconfiguring the network can result in performance shortfalls, security vulnerabilities and gaps in regulatory compliance. Taking full advantage of the dynamic server environment made possible by server virtualization requires that the virtual port profile associated with a VM can be automatically created within the network and migrated to the appropriate access switch whenever a VM is migrated among servers. The dynamic creation and movement of port profiles can be achieved in those instances in which the switch vendor provides the right level of integration between the hypervisor management system and the switch element management system via APIs supported by both systems. This integration allows the hypervisor manager to automatically notify the switch management system when a VM is created or moved, so that the virtual port profile in the network is created or moved as required without human intervention. Fully automated configuration of virtual port profiles not only enables a highly dynamic server environment, but it also minimizes operational workloads, reduces the human errors that occur during configuration and preserves the traditional responsibilities of the server and network administrations. Deployment of multiple types of hypervisors within a single data center is becoming increasingly common. This means that switch element management systems need to support the APIs of each hypervisor management system employed within the data center. Gaining Control of Virtualized Server Environments June 2011 Page 7 Visibility The dynamic nature of virtualized server environments can make it difficult for IT organizations to have sufficient visibility to keep track of the status of each VM throughout its lifecycle. This lack of visibility can result in VM Sprawl whereby obsolete or low activity VMs proliferate and waste system, storage and network resources. This lack of visibility can also make it difficult for organizations to comply with the large and growing list of regulatory requirements. From a server management perspective, full visibility into a virtualized server environment requires system management tools that can extract detailed life cycle data from the various hypervisors’ management systems. This capability enables the management system to catalog each VM in terms of characteristics such as its time of deployment, the version of the OS it is running, its location, the ownership of the VM, the allocated resources, the workload history and the patch status. In a similar fashion, network managers also need a network management tool that can both gather information about, and report on the network services provided to VMs. In those instances in which the data center network supports virtual port profiles, much of the data required by network managers can be collected via enhanced capabilities within the switches’ element management system. VM location and inventory data can be stored in a network management database where it can be correlated with both historical and real-time network performance statistics. This provides a valuable aid in troubleshooting application performance issues and verifying the correctness of network configurations. VM-aware network element management systems allow network managers to have direct access to the information required to troubleshoot virtualized server networking issues without the time delays that are typically involved when network managers must coordinate manually with the appropriate server management personnel. Larkins stated that driven in part by their regulatory requirements that his clients want to know where their data is at all times. As a result, he needs detailed visibility into the creation and movement of VMs in order to provide that insight. He added that his customers expect that he will be monitoring the systems that he provides to them and that to do this, his company “Monitors as much as we can, collects as much data as we can, and trends that data as much as we can.” Gaining Control of Virtualized Server Environments June 2011 Page 8 Summary The traditional data center network was based on an architecture and design that was tailored to satisfy the connectivity requirements of a relatively simple server infrastructure based on a variety of assumptions. This included the assumption that each server was dedicated to a single application and had a single IP address and a single MAC address and that the static nature of the application environment lent itself to manual processes for server deployment or redeployment. While server virtualization delivers efficiency, flexibility and agility to application provisioning and management, it has negated the validity of most, if not all, of the assumptions that were used to design traditional data center networks. As a result, IT organizations must develop a new architecture and design for their data center networks. As part of their new approach to architecting and designing data center networks, IT organizations need to manage network traffic using the same granular policies relating to security, QoS, and regulatory compliance for applications running on VMs as it did previously for applications running on dedicated physical servers. This can be accomplished if the network access switch is capable of mapping each VM to a separate, individually configurable virtual switch port. This topic is the subject of standards currently being developed by the IEEE. In order for IT organizations to take full advantage of the dynamic server environment made possible by server virtualization requires that the virtual port profile associated with a VM can be automatically created within the network and migrated to the appropriate access switch whenever a VM is migrated among servers. The dynamic creation and movement of port profiles can be achieved in those instances where the switch vendor provides the right level of integration between the hypervisor management system and the switch element management system via APIs that are supported by both systems. Given the dynamic nature of virtualized server environments, in order for IT organizations to have full visibility into these environments requires a system management tool that can extract detailed life cycle data from the various hypervisors’ management systems. In a similar fashion, IT organizations also need a network management tool that can both gather information about, and report on the network services provided to VMs. In those instances in which the data center network supports virtual port profiles, much of the data that is required by network managers can be collected via enhanced capabilities within the switches’ element management system. Gaining Control of Virtualized Server Environments June 2011 Page 9 About the Author – Jim Metzler Jim has a broad background in the IT industry. This includes being a software engineer, an engineering manager for high-speed data services for a major network service provider, a product manager for network hardware, a network manager at two Fortune 500 companies, and the principal of a consulting organization. In addition, Jim has created software tools for designing customer networks for a major network service provider and directed and performed market research at a major industry analyst firm. Jim’s current interests include cloud networking and application delivery. About the Webtorials® Editorial/Analyst Division The Webtorials® Editorial/Analyst Division, a joint venture of industry veterans Steven Taylor and Jim Metzler, is devoted to performing in-depth analysis and research in focused areas such as Metro Ethernet and MPLS, as well as in areas that cross the traditional functional boundaries of IT, such as Unified Communications and Application Delivery. The Editorial/Analyst Division’s focus is on providing actionable insight through custom research with a forward looking viewpoint. Through reports that examine industry dynamics from both a demand and a supply perspective, the firm educates the marketplace both on emerging trends and the role that IT products, services and processes play in responding to those trends. ® For more information and for additional Webtorials Editorial/Analyst Division products, please contact Jim Metzler at [email protected] or Steven Taylor at [email protected]. Published by Webtorials Editorial/Analyst Division www.Webtorials.com Division Cofounders: Jim Metzler [email protected] Steven Taylor [email protected] Professional Opinions Disclaimer All information presented and opinions expressed in this publication represent the current opinions of the author(s) based on professional judgment and best available information at the time of the presentation. Consequently, the information is subject to change, and no liability for advice presented is assumed. Ultimate responsibility for choice of appropriate solutions remains with the reader. Copyright © 2011 Webtorials For editorial and sponsorship information, contact Jim Metzler or Steven Taylor. The Webtorials Editorial/Analyst Division is an analyst and consulting joint venture of Steven Taylor and Jim Metzler. Gaining Control of Virtualized Server Environments June 2011 Page 10
© Copyright 2026 Paperzz