US9270619B2 - Logical switch - Google Patents

Logical switch Download PDF

Info

Publication number
US9270619B2
US9270619B2 US13/925,551 US201313925551A US9270619B2 US 9270619 B2 US9270619 B2 US 9270619B2 US 201313925551 A US201313925551 A US 201313925551A US 9270619 B2 US9270619 B2 US 9270619B2
Authority
US
United States
Prior art keywords
virtual
party
switch
physical host
components
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US13/925,551
Other versions
US20140376560A1 (en
Inventor
Pradeep Senniappan
Karthikeyan Nenmeli Ravichandran
Natalia Valeryevna Varava
Gregory M. Cusanza
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Microsoft Technology Licensing LLC
Original Assignee
Microsoft Technology Licensing LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Microsoft Technology Licensing LLC filed Critical Microsoft Technology Licensing LLC
Priority to US13/925,551 priority Critical patent/US9270619B2/en
Assigned to MICROSOFT CORPORATION reassignment MICROSOFT CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SENNIAPPAN, PRADEEP, CUSANZA, GREGORY M., RAVICHANDRAN, KARTHIKEYAN NENMELI, VARAVA, NATALIA VALERYEVNA
Priority to PCT/US2014/043593 priority patent/WO2014209846A1/en
Priority to CN201480036045.2A priority patent/CN105409172B/en
Publication of US20140376560A1 publication Critical patent/US20140376560A1/en
Assigned to MICROSOFT TECHNOLOGY LICENSING, LLC reassignment MICROSOFT TECHNOLOGY LICENSING, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MICROSOFT CORPORATION
Assigned to MICROSOFT TECHNOLOGY LICENSING, LLC reassignment MICROSOFT TECHNOLOGY LICENSING, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MICROSOFT CORPORATION
Application granted granted Critical
Publication of US9270619B2 publication Critical patent/US9270619B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/12Discovery or management of network topologies
    • H04L41/122Discovery or management of network topologies of virtualised topologies, e.g. software-defined networks [SDN] or network function virtualisation [NFV]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/25Routing or path finding in a switch fabric
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/12Discovery or management of network topologies
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/70Virtual switches

Definitions

  • Computers and computing systems have affected nearly every aspect of modern living. Computers are generally involved in work, recreation, healthcare, transportation, entertainment, household management, etc.
  • computing system functionality can be enhanced by a computing systems ability to be interconnected to other computing systems via network connections.
  • Network connections may include, but are not limited to, connections via wired or wireless Ethernet, cellular connections, or even computer to computer connections through serial, parallel, USB, or other connections. The connections allow a computing system to access services at other computing systems and to quickly and efficiently receive application data from other computing system.
  • Some computing functionality is implemented in the context of distributed computing and virtualization as implemented in datacenters.
  • virtual computing hardware is implemented using actual physical hardware.
  • several virtual machines may be installed on a physical machine or set of physical machines interconnected with physical networking hardware.
  • the virtual machines may be interconnected using virtual network adapters and virtual switches.
  • the virtual switches make use of an actual physical switch to accomplish communication between different virtual machines that requires communication between different physical machines.
  • virtual networks in a datacenter typically have various types of physical server hardware, physical network interface cards (nics) on host physical machines, physical top of rack (TOR) switches and network services configurations in their environments. Administrators currently configure this physical hardware independently, one piece at a time. This can be error prone and time consuming as an administrator needs to manually maintain consistency in such configurations across all virtualization hosts and switches these hosts are connected to in order to enable a) basic connectivity functions for virtualized workloads; b) adequate availability of physical nics and switches hosts are connected to; and c) VM migration capabilities across hosts.
  • nics physical network interface cards
  • TOR top of rack
  • VNics virtual nics
  • VMs virtual machines
  • One embodiment illustrated herein includes a method that may be practiced in a distributed computing environment.
  • the method includes acts for configuring third party solutions to operate with virtual machines and virtual switches in a distributed network environment.
  • the method includes receiving information at a logical switch about third party solutions in a distributed network.
  • the method further includes receiving information at the logical switch about requirements for virtual components of the distributed network.
  • the method further includes the logical switch automatically configuring third party solutions in the distributed network to meet the requirements for the virtual components of the distributed network.
  • FIG. 1 illustrates a virtualization fabric
  • FIG. 2 illustrates a method of configuring third party solutions to operate with virtual machines in a distributed network environment.
  • a logical switch centralizes and simplifies virtual network provisioning, administration, and monitoring using network aggregation.
  • an administrator can interact with the logical switch to define what functionality should be supported in a virtual network.
  • the logical switch can be connected to third party solutions, such as physical hardware devices, software solutions or combinations thereof, such as switches, TOR switches, forwarding hardware and/or applications, monitoring hardware and/or applications, quality of service hardware and/or applications, firewall hardware and/or applications, security gateway hardware and/or applications functions, DOS attack function.
  • the logical switch can then automatically, configure network hardware to allow the network hardware to meet virtual system requirements or to enable additional functionality.
  • the logical switch could be used to configure solutions to create or extend forwarding functionality, monitoring functionality, firewall functionality, DOS attack remediation, IP allocation and management, performance acceleration, gateway functions, etc.
  • an administrator can indicate that a virtual network should support certain forwarding, bandwidth, filtering, etc. parameters.
  • the administrator can configure these parameters one time in a logical switch user interface.
  • the logical switch can then be responsible for ensuring that all third party solutions, such as physical network hardware and/or software solutions plugged into the logical switch are configured to support such parameters. This allows for easy deployment and migration of workloads across a fabric as there are assurances that the various portions of the fabric support the desired parameters by automatic configuration by the logical switch without an administrator needing to configure each hardware device individually.
  • the logical switch can automatically, without network administrator interaction, cause the solution to be brought back into compliance and/or alert an administrator if the solution cannot be brought back into compliance.
  • the logical switch can automatically configure the solutions to be compliant with the specified parameters without a network administrator needing to manually configure the solutions individually.
  • FIG. 1 illustrates a distributed network topology fabric 100 .
  • a set 102 of datacenters Within the distributed network topology fabric 100 is a set 102 of datacenters. While three datacenters are illustrated in the set 102 , it should be appreciated that virtually any number of datacenters may be implemented within a distributed network topology fabric. Further, the datacenters may be of similar or different types of datacenters. For example, in one embodiment, a single private datacenter under virtually complete control of an enterprise may be implemented. In another example, multiple private datacenters under virtually complete control of an enterprise may be implemented. In yet another example, a datacenter may be a public cloud datacenter under the control of a hosting entity, but which hosts tenants' virtual networks. In yet another alternative embodiment, an enterprise may use a combination of private cloud data center(s) and public cloud data center(s). Etc.
  • FIG. 1 illustrates, for example, a datacenter 102 - 1 .
  • a datacenter 102 - 1 is a combination of physical components and virtual components.
  • FIG. 1 illustrates a set 104 of physical host machines.
  • Each of the physical host machines includes a physical network interface card (nic), such as the nic 106 - 1 illustrated on the physical host machine 104 - 1 .
  • the nics are connected to a physical switch 108 that connects the physical host machines in the datacenter 102 - 1 .
  • the switch 108 can also connect the physical host machines to other datacenters or other hardware and/or devices external to the datacenter 102 - 1 .
  • FIG. 1 further illustrates various virtual components implemented at the datacenter 102 - 1 .
  • the physical host machine 104 - 1 hosts a set 110 of virtual machines.
  • virtual machine 110 - 1 also includes a virtual network interface card (vnic) 112 - 1 .
  • vnic virtual network interface card
  • Each of the virtual machines includes a vnic connected to the virtual switch 114 - 1 that facilitates communication to and from virtual machines.
  • FIG. 1 further illustrates a virtual switch 114 - 1 on the host machine 104 - 1 .
  • Each host machine includes one or more virtual switches to enable communication between virtual machines on the same host and to components off of a given host.
  • the virtual switch leverages the capabilities of the physical switch 108 to enable communication between physical components.
  • FIG. 1 further illustrates a virtual machine monitor (VMM) 116 .
  • the VMM 116 is a server that may be used to place and configure virtual components in the datacenter 102 - 1 .
  • the VMM 116 may be implemented using one or more of the physical host machines in the set 104 or on some other available hardware device. Additionally, the physical switches may support functionality, and such functionality needs to be enabled if desired for use by a virtual components. Similarly, for vnics to support certain functionality, such functionality needs to be enabled by the underlying physical nics.
  • configuration of physical hardware would be performed by an administrator interacting with the devices individually to configure devices in the data center 102 - 1 .
  • placement of virtual components such as virtual machines
  • migration of virtual components as physical hardware changes or failures in the network would ordinarily be performed by administrator interaction with the VMM 116 .
  • embodiments herein implement a logical switch 118 implemented on top of the virtual switch 114 as part of the VMM 116 that can automatically configure and monitor aspects for network functionality, such as connectivity and capabilities, in a virtualized datacenter.
  • a logical switch is a scalable, pluggable, and extensible virtual network services platform that that supports third party network services to plug into the key virtual machine (VM) and virtual networking operations.
  • the integration with third party solutions can be enabled on various levels.
  • the logical switch may be configured to interact with third party solutions (hardware and/or software) by using a plugin.
  • the plugin is a server side solution deployed at the VMM that allows the logical switch 118 to communicate directly with the third party solutions.
  • third parties provide plugins that can be installed on the VMM that allows for communication with application programming interfaces (APIs) at the solutions provided by the third parties.
  • APIs application programming interfaces
  • host level extensions may be implemented.
  • drivers can be installed at the hosts in the set 104 of hosts. This allows the hosts to interact with third party solutions.
  • third parties can provide interface extensions that can be plugged into the user interface portion of the VMM.
  • administrators can configure third party solutions using interface components provided by third party solution providers, but integrated into the native VMM configuration interface.
  • native hardware APIs may be used.
  • the hardware may support certain APIs and the VMM can be configured to communicate natively to the third party solutions.
  • a VMM may be configured to communicate natively to TOR switches, VPN and NAT gateways, load balancers, and the like.
  • the virtual components are limited in their capabilities by what the physical components can provide.
  • the virtual switch 114 - 1 can provide functionality that is enabled by the physical switch 108 .
  • the logical switch 118 can determine and configure physical component properties, virtual component requirements, and match virtual component requirements to physical component (or other third party solution) capabilities.
  • embodiments can implement network classification that is independent from a VM network, facilitated by the logical switch. This allows fine-tuning networking capabilities that VMs demand and at the same time allows an administrator to move VM workloads across the datacenter to match these demands to the capabilities of the underlying networking equipment.
  • Configuration and monitoring of aspects of network functions (e.g. connectivity and capability) in a virtualized datacenter can be done at a central place, i.e. at the logical switch.
  • the logical switch functions as a single virtual switch across all hosts plugged into this logical switch 118 .
  • Embodiments provide simplified extensibility.
  • a logical switch 118 is built on top of an extensible virtual switch 114 and allows combining value-add services enabled by various network services.
  • an administrator can provide configuration in the logical switch to enable forwarding functionality, monitoring functionality, firewall functionality, DOS attack remediation, IP allocation and management, performance acceleration, gateway functions, etc.
  • the logical switch coordinates network connectivity and policy management across entities plugged into it. It also enables servicing functions (e.g. version management, maintenance) across a distributed environment.
  • the logical switch 118 may use the native virtual switch 114 for packet forwarding, third party switch extension for packet monitoring and an IP address management service for IP allocation management functions.
  • Embodiments may implement network topology and connectivity discovery.
  • Virtual network connectivity discovery in enabled by the logical switch, in combination with the functionality of physical and/or virtual networking equipment.
  • Network connectivity is learned from a network topology that is either declared by an administrator or discovered by the logical switch querying various physical and virtual components in the network.
  • network topology could be published by third parties via logical switch integration.
  • the logical switch may include APIs that allows third party equipment to declare functionality that is enabled and/or being implemented.
  • a topology published by one party or entered into the VMM 116 can be shared across multiple parties managing the same portion of the datacenter. This further lowers the entry point for third parties to innovate on top of the VMM 116 virtualization enabled datacenters.
  • Embodiments may enable advanced compliance capabilities.
  • the VMM 116 actively monitors the configuration of the physical environment enabling the logical switch 118 to detect the physical environment's health, availability and utilization. This may include, for example, VLAN, PVLAN, and other settings on TOR physical switches to which hosts are connected, physical NIC and NIC teaming parameters, management and other host vNics, third party virtual switch extensions that are part of the logical switch and VM vNics attached to the logical switch.
  • the VMM 116 may additionally or alternatively support a) automatic (built-in and enabled by third parties) and manual remediation of incompliance for various conditions; b) evaluating the larger impact of incompliance via aggregation; c) avoiding placing workloads on incompatible virtual and physical equipment. In this way the logical switch can either attempt to remediate solutions that are out of compliance, and/or alter a network administrator of such non-compliance.
  • Embodiments enable division of responsibilities between capability and connectivity.
  • the logical switch matches network capabilities and connectivity requested by the virtual workloads to the underlying capabilities and connectivity enabled by the physical infrastructure. This abstracts workloads from the specifics of the fulfilling such requests by the physical environment.
  • the logical switch platform supports dynamic changes in both physical and virtual networks. It also allows tenants to define and manage their own networks. For example, a tenant can define L 3 (IP space) for their networks, the VMM and a third party plugins can jointly enable L 2 isolation for such networks.
  • L 3 IP space
  • the VMM and a third party plugins can jointly enable L 2 isolation for such networks.
  • Embodiments may enable effective datacenter utilization.
  • VM placement in a datacenter can take into account what logical switch can provide connectivity and capability demanded by the VM.
  • Embodiments can also find a healthy host that can meet VM requirements at VM deployment or migration time.
  • the logical switch allows configuring the physical environment in a more dynamic and optimal fashion where the environment can be provisioned on demand instead of being statically overprovisioned ahead of time for both network virtualization-based and even VLAN-based VM workloads.
  • the method 200 may be practiced in a distributed computing environment and includes acts for configuring third party solutions to operate with virtual machines in a distributed network environment.
  • the method includes receiving information at a logical switch about third party solutions in a distributed network (act 202 ).
  • the logical switch 118 may receive information about physical switches, TOR switches, routers, gateways, load balancers, etc.
  • the method 200 further includes receiving information at the logical switch about requirements for virtual components of the distributed network (act 204 ).
  • the logical switch 118 may receive information about functionality and/or performance desired from the virtual components.
  • the logical switch may include or be coupled to a user interface that allows an administrator to specify certain forwarding functionality, monitoring functionality, firewall functionality, DOS attack remediation, IP allocation and management, performance acceleration, gateway functions, etc.
  • the method 200 further includes the logical switch automatically configuring third party solutions in the distributed network to meet the requirements for the virtual components of the distributed network (act 206 ).
  • the logical switch can coordinates network connectivity and policy management across entities plugged into it to accomplish the specified functionality.
  • the logical switch can configure third-party solutions to achieve the desired functionality.
  • the method 200 may be practiced where receiving information at a logical switch about third party solutions in a distributed network is performed by the logical switch communicating with the third party solutions and the third party solutions providing the information about the third party solutions directly to the logical switch.
  • the method 200 may be practiced where receiving information at a logical switch about third party solutions in a distributed network is performed by the logical switch receiving configuration metadata from a network administrator.
  • the method 200 may be practiced where receiving information at the logical switch about requirements for virtual components of the distributed network comprises receiving information about requirements for virtual components of the distributed network in requirement metadata from a network administrator.
  • the method 200 may be practiced where the logical switch automatically configuring third party solutions in the distributed network to meet the requirements for the virtual components of the distributed network comprises the logical switch directly configuring third party solutions by communicating with the third party solutions using native third party solution protocols.
  • communicating with third party solutions using third party solution protocols may include the logical switch communicating using a plug-in at the logical switch that enables the plug-in to communicate directly with third party solutions in third party solution's native protocol.
  • communicating with third party solutions using third party solution protocols comprises the logical switch communicating directly with third party solutions using APIs exposed by the third party solutions.
  • the methods may be practiced by a computer system including one or more processors and computer readable media such as computer memory.
  • the computer memory may store computer executable instructions that when executed by one or more processors cause various functions to be performed, such as the acts recited in the embodiments.
  • Embodiments of the present invention may comprise or utilize a special purpose or general-purpose computer including computer hardware, as discussed in greater detail below.
  • Embodiments within the scope of the present invention also include physical and other computer-readable media for carrying or storing computer-executable instructions and/or data structures.
  • Such computer-readable media can be any available media that can be accessed by a general purpose or special purpose computer system.
  • Computer-readable media that store computer-executable instructions are physical storage media.
  • Computer-readable media that carry computer-executable instructions are transmission media.
  • embodiments of the invention can comprise at least two distinctly different kinds of computer-readable media: physical computer readable storage media and transmission computer readable media.
  • Physical computer readable storage media includes RAM, ROM, EEPROM, CD-ROM or other optical disk storage (such as CDs, DVDs, etc), magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer.
  • a “network” is defined as one or more data links that enable the transport of electronic data between computer systems and/or modules and/or other electronic devices.
  • a network or another communications connection can include a network and/or data links which can be used to carry or desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer. Combinations of the above are also included within the scope of computer-readable media.
  • program code means in the form of computer-executable instructions or data structures can be transferred automatically from transmission computer readable media to physical computer readable storage media (or vice versa).
  • program code means in the form of computer-executable instructions or data structures received over a network or data link can be buffered in RAM within a network interface module (e.g., a “NIC”), and then eventually transferred to computer system RAM and/or to less volatile computer readable physical storage media at a computer system.
  • NIC network interface module
  • computer readable physical storage media can be included in computer system components that also (or even primarily) utilize transmission media.
  • Computer-executable instructions comprise, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions.
  • the computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, or even source code.
  • the invention may be practiced in network computing environments with many types of computer system configurations, including, personal computers, desktop computers, laptop computers, message processors, hand-held devices, multi-processor systems, microprocessor-based or programmable consumer electronics, network PCs, minicomputers, mainframe computers, mobile telephones, PDAs, pagers, routers, switches, and the like.
  • the invention may also be practiced in distributed system environments where local and remote computer systems, which are linked (either by hardwired data links, wireless data links, or by a combination of hardwired and wireless data links) through a network, both perform tasks.
  • program modules may be located in both local and remote memory storage devices.
  • the functionally described herein can be performed, at least in part, by one or more hardware logic components.
  • illustrative types of hardware logic components include Field-programmable Gate Arrays (FPGAs), Program-specific Integrated Circuits (ASICs), Program-specific Standard Products (ASSPs), System-on-a-chip systems (SOCs), Complex Programmable Logic Devices (CPLDs), etc.

Abstract

Configuring third party solutions to operate with virtual machines and virtual switches in a distributed network environment. The method includes receiving information at a logical switch about third party solutions in a distributed network. The method further includes receiving information at the logical switch about requirements for virtual components of the distributed network. The method further includes the logical switch automatically configuring third party solutions in the distributed network to meet the requirements for the virtual components of the distributed network.

Description

BACKGROUND Background and Relevant Art
Computers and computing systems have affected nearly every aspect of modern living. Computers are generally involved in work, recreation, healthcare, transportation, entertainment, household management, etc.
Further, computing system functionality can be enhanced by a computing systems ability to be interconnected to other computing systems via network connections. Network connections may include, but are not limited to, connections via wired or wireless Ethernet, cellular connections, or even computer to computer connections through serial, parallel, USB, or other connections. The connections allow a computing system to access services at other computing systems and to quickly and efficiently receive application data from other computing system.
Some computing functionality is implemented in the context of distributed computing and virtualization as implemented in datacenters. In virtualization scenarios, virtual computing hardware is implemented using actual physical hardware. For example, several virtual machines may be installed on a physical machine or set of physical machines interconnected with physical networking hardware. Further, the virtual machines may be interconnected using virtual network adapters and virtual switches. However, in many cases the virtual switches make use of an actual physical switch to accomplish communication between different virtual machines that requires communication between different physical machines.
Thus, virtual networks in a datacenter typically have various types of physical server hardware, physical network interface cards (nics) on host physical machines, physical top of rack (TOR) switches and network services configurations in their environments. Administrators currently configure this physical hardware independently, one piece at a time. This can be error prone and time consuming as an administrator needs to manually maintain consistency in such configurations across all virtualization hosts and switches these hosts are connected to in order to enable a) basic connectivity functions for virtualized workloads; b) adequate availability of physical nics and switches hosts are connected to; and c) VM migration capabilities across hosts.
Additionally, tenants of a datacenter hosting a virtual network for the tenant today have to configure virtual nics (vNics) of their virtual machines (VMs) one by one.
The subject matter claimed herein is not limited to embodiments that solve any disadvantages or that operate only in environments such as those described above. Rather, this background is only provided to illustrate one exemplary technology area where some embodiments described herein may be practiced.
BRIEF SUMMARY
One embodiment illustrated herein includes a method that may be practiced in a distributed computing environment. The method includes acts for configuring third party solutions to operate with virtual machines and virtual switches in a distributed network environment. The method includes receiving information at a logical switch about third party solutions in a distributed network. The method further includes receiving information at the logical switch about requirements for virtual components of the distributed network. The method further includes the logical switch automatically configuring third party solutions in the distributed network to meet the requirements for the virtual components of the distributed network.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
Additional features and advantages will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by the practice of the teachings herein. Features and advantages of the invention may be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims. Features of the present invention will become more fully apparent from the following description and appended claims, or may be learned by the practice of the invention as set forth hereinafter.
BRIEF DESCRIPTION OF THE DRAWINGS
In order to describe the manner in which the above-recited and other advantages and features can be obtained, a more particular description of the subject matter briefly described above will be rendered by reference to specific embodiments which are illustrated in the appended drawings. Understanding that these drawings depict only typical embodiments and are not therefore to be considered to be limiting in scope, embodiments will be described and explained with additional specificity and detail through the use of the accompanying drawings in which:
FIG. 1 illustrates a virtualization fabric; and
FIG. 2 illustrates a method of configuring third party solutions to operate with virtual machines in a distributed network environment.
DETAILED DESCRIPTION
The described embodiments implement what is referred to herein as a logical switch. A logical switch centralizes and simplifies virtual network provisioning, administration, and monitoring using network aggregation. In particular, an administrator can interact with the logical switch to define what functionality should be supported in a virtual network. The logical switch can be connected to third party solutions, such as physical hardware devices, software solutions or combinations thereof, such as switches, TOR switches, forwarding hardware and/or applications, monitoring hardware and/or applications, quality of service hardware and/or applications, firewall hardware and/or applications, security gateway hardware and/or applications functions, DOS attack function. The logical switch can then automatically, configure network hardware to allow the network hardware to meet virtual system requirements or to enable additional functionality. For example, the logical switch could be used to configure solutions to create or extend forwarding functionality, monitoring functionality, firewall functionality, DOS attack remediation, IP allocation and management, performance acceleration, gateway functions, etc.
For example, an administrator can indicate that a virtual network should support certain forwarding, bandwidth, filtering, etc. parameters. The administrator can configure these parameters one time in a logical switch user interface. The logical switch can then be responsible for ensuring that all third party solutions, such as physical network hardware and/or software solutions plugged into the logical switch are configured to support such parameters. This allows for easy deployment and migration of workloads across a fabric as there are assurances that the various portions of the fabric support the desired parameters by automatic configuration by the logical switch without an administrator needing to configure each hardware device individually. Additionally, when a solution goes out of compliance with the parameters, the logical switch can automatically, without network administrator interaction, cause the solution to be brought back into compliance and/or alert an administrator if the solution cannot be brought back into compliance. Further, when new solutions are added to the fabric and connected to the logical switch, the logical switch can automatically configure the solutions to be compliant with the specified parameters without a network administrator needing to manually configure the solutions individually.
Referring now to FIG. 1, an example is illustrated. FIG. 1 illustrates a distributed network topology fabric 100. Within the distributed network topology fabric 100 is a set 102 of datacenters. While three datacenters are illustrated in the set 102, it should be appreciated that virtually any number of datacenters may be implemented within a distributed network topology fabric. Further, the datacenters may be of similar or different types of datacenters. For example, in one embodiment, a single private datacenter under virtually complete control of an enterprise may be implemented. In another example, multiple private datacenters under virtually complete control of an enterprise may be implemented. In yet another example, a datacenter may be a public cloud datacenter under the control of a hosting entity, but which hosts tenants' virtual networks. In yet another alternative embodiment, an enterprise may use a combination of private cloud data center(s) and public cloud data center(s). Etc.
FIG. 1 illustrates, for example, a datacenter 102-1. At the datacenter 102-1, is a combination of physical components and virtual components. In particular, FIG. 1 illustrates a set 104 of physical host machines. Each of the physical host machines includes a physical network interface card (nic), such as the nic 106-1 illustrated on the physical host machine 104-1. The nics are connected to a physical switch 108 that connects the physical host machines in the datacenter 102-1. The switch 108 can also connect the physical host machines to other datacenters or other hardware and/or devices external to the datacenter 102-1.
FIG. 1 further illustrates various virtual components implemented at the datacenter 102-1. For example, the physical host machine 104-1 hosts a set 110 of virtual machines. In the illustrated example, virtual machine 110-1 also includes a virtual network interface card (vnic) 112-1. Each of the virtual machines includes a vnic connected to the virtual switch 114-1 that facilitates communication to and from virtual machines.
FIG. 1 further illustrates a virtual switch 114-1 on the host machine 104-1. Each host machine includes one or more virtual switches to enable communication between virtual machines on the same host and to components off of a given host. The virtual switch leverages the capabilities of the physical switch 108 to enable communication between physical components.
FIG. 1 further illustrates a virtual machine monitor (VMM) 116. The VMM 116 is a server that may be used to place and configure virtual components in the datacenter 102-1. The VMM 116 may be implemented using one or more of the physical host machines in the set 104 or on some other available hardware device. Additionally, the physical switches may support functionality, and such functionality needs to be enabled if desired for use by a virtual components. Similarly, for vnics to support certain functionality, such functionality needs to be enabled by the underlying physical nics.
Ordinarily, configuration of physical hardware would be performed by an administrator interacting with the devices individually to configure devices in the data center 102-1. Further, placement of virtual components (such as virtual machines), and migration of virtual components as physical hardware changes or failures in the network would ordinarily be performed by administrator interaction with the VMM 116. However, embodiments herein implement a logical switch 118 implemented on top of the virtual switch 114 as part of the VMM 116 that can automatically configure and monitor aspects for network functionality, such as connectivity and capabilities, in a virtualized datacenter.
A logical switch is a scalable, pluggable, and extensible virtual network services platform that that supports third party network services to plug into the key virtual machine (VM) and virtual networking operations. The integration with third party solutions (including hardware devices and/or software solutions) can be enabled on various levels. For example, in some embodiments the logical switch may be configured to interact with third party solutions (hardware and/or software) by using a plugin. The plugin is a server side solution deployed at the VMM that allows the logical switch 118 to communicate directly with the third party solutions. In particular, third parties provide plugins that can be installed on the VMM that allows for communication with application programming interfaces (APIs) at the solutions provided by the third parties. The VMM invokes third party code provided by the plugin.
Alternatively, host level extensions may be implemented. In particular, drivers can be installed at the hosts in the set 104 of hosts. This allows the hosts to interact with third party solutions.
In yet another alternative embodiment, third parties can provide interface extensions that can be plugged into the user interface portion of the VMM. Thus, administrators can configure third party solutions using interface components provided by third party solution providers, but integrated into the native VMM configuration interface.
In yet another embodiments, native hardware APIs may be used. In particular, the hardware may support certain APIs and the VMM can be configured to communicate natively to the third party solutions. For example, a VMM may be configured to communicate natively to TOR switches, VPN and NAT gateways, load balancers, and the like.
The virtual components are limited in their capabilities by what the physical components can provide. For example, the virtual switch 114-1 can provide functionality that is enabled by the physical switch 108. The logical switch 118 can determine and configure physical component properties, virtual component requirements, and match virtual component requirements to physical component (or other third party solution) capabilities. In particular, embodiments can implement network classification that is independent from a VM network, facilitated by the logical switch. This allows fine-tuning networking capabilities that VMs demand and at the same time allows an administrator to move VM workloads across the datacenter to match these demands to the capabilities of the underlying networking equipment.
Configuration and monitoring of aspects of network functions (e.g. connectivity and capability) in a virtualized datacenter can be done at a central place, i.e. at the logical switch. The logical switch functions as a single virtual switch across all hosts plugged into this logical switch 118.
Embodiments provide simplified extensibility. In particular a logical switch 118 is built on top of an extensible virtual switch 114 and allows combining value-add services enabled by various network services. For example, an administrator can provide configuration in the logical switch to enable forwarding functionality, monitoring functionality, firewall functionality, DOS attack remediation, IP allocation and management, performance acceleration, gateway functions, etc. The logical switch coordinates network connectivity and policy management across entities plugged into it. It also enables servicing functions (e.g. version management, maintenance) across a distributed environment.
For example the logical switch 118 may use the native virtual switch 114 for packet forwarding, third party switch extension for packet monitoring and an IP address management service for IP allocation management functions.
Embodiments may implement network topology and connectivity discovery. Virtual network connectivity discovery in enabled by the logical switch, in combination with the functionality of physical and/or virtual networking equipment. Network connectivity is learned from a network topology that is either declared by an administrator or discovered by the logical switch querying various physical and virtual components in the network. Alternatively or additionally, network topology could be published by third parties via logical switch integration. In particular, the logical switch may include APIs that allows third party equipment to declare functionality that is enabled and/or being implemented. A topology published by one party or entered into the VMM 116 can be shared across multiple parties managing the same portion of the datacenter. This further lowers the entry point for third parties to innovate on top of the VMM 116 virtualization enabled datacenters.
Embodiments may enable advanced compliance capabilities. In particular, the VMM 116 actively monitors the configuration of the physical environment enabling the logical switch 118 to detect the physical environment's health, availability and utilization. This may include, for example, VLAN, PVLAN, and other settings on TOR physical switches to which hosts are connected, physical NIC and NIC teaming parameters, management and other host vNics, third party virtual switch extensions that are part of the logical switch and VM vNics attached to the logical switch.
In a condition when either the configuration of a networking entity deviates from the desired state, or the health of the networking entity that is a part of a logical switch is affected, the state of an affected entity is flagged as incompliant providing a relevant detailed description of the problem detected and an alert is sent to an administrator. The effect of such conditions on a larger set of operations enabled by the logical switch is calculated. Notably, a failure\incompliance in one subcomponent can affect the state of a larger system.
In addition to flagging incompliance and alerting the administrator, the VMM 116 may additionally or alternatively support a) automatic (built-in and enabled by third parties) and manual remediation of incompliance for various conditions; b) evaluating the larger impact of incompliance via aggregation; c) avoiding placing workloads on incompatible virtual and physical equipment. In this way the logical switch can either attempt to remediate solutions that are out of compliance, and/or alter a network administrator of such non-compliance.
Embodiments enable division of responsibilities between capability and connectivity. The logical switch matches network capabilities and connectivity requested by the virtual workloads to the underlying capabilities and connectivity enabled by the physical infrastructure. This abstracts workloads from the specifics of the fulfilling such requests by the physical environment.
The logical switch platform supports dynamic changes in both physical and virtual networks. It also allows tenants to define and manage their own networks. For example, a tenant can define L3 (IP space) for their networks, the VMM and a third party plugins can jointly enable L2 isolation for such networks.
Embodiments may enable effective datacenter utilization. VM placement in a datacenter can take into account what logical switch can provide connectivity and capability demanded by the VM. Embodiments can also find a healthy host that can meet VM requirements at VM deployment or migration time. The logical switch allows configuring the physical environment in a more dynamic and optimal fashion where the environment can be provisioned on demand instead of being statically overprovisioned ahead of time for both network virtualization-based and even VLAN-based VM workloads.
The following discussion now refers to a number of methods and method acts that may be performed. Although the method acts may be discussed in a certain order or illustrated in a flow chart as occurring in a particular order, no particular ordering is required unless specifically stated, or required because an act is dependent on another act being completed prior to the act being performed.
Referring now to FIG. 2, a method 200 is illustrated. The method 200 may be practiced in a distributed computing environment and includes acts for configuring third party solutions to operate with virtual machines in a distributed network environment. The method includes receiving information at a logical switch about third party solutions in a distributed network (act 202). For example, the logical switch 118 may receive information about physical switches, TOR switches, routers, gateways, load balancers, etc.
The method 200 further includes receiving information at the logical switch about requirements for virtual components of the distributed network (act 204). For example, the logical switch 118 may receive information about functionality and/or performance desired from the virtual components. For example, the logical switch may include or be coupled to a user interface that allows an administrator to specify certain forwarding functionality, monitoring functionality, firewall functionality, DOS attack remediation, IP allocation and management, performance acceleration, gateway functions, etc.
The method 200 further includes the logical switch automatically configuring third party solutions in the distributed network to meet the requirements for the virtual components of the distributed network (act 206). For example, the logical switch can coordinates network connectivity and policy management across entities plugged into it to accomplish the specified functionality. In particular, the logical switch can configure third-party solutions to achieve the desired functionality.
As noted above, the method 200 may be practiced where receiving information at a logical switch about third party solutions in a distributed network is performed by the logical switch communicating with the third party solutions and the third party solutions providing the information about the third party solutions directly to the logical switch. Alternatively or additionally, the method 200 may be practiced where receiving information at a logical switch about third party solutions in a distributed network is performed by the logical switch receiving configuration metadata from a network administrator.
With respect to receiving information about requirements for virtual components, the method 200 may be practiced where receiving information at the logical switch about requirements for virtual components of the distributed network comprises receiving information about requirements for virtual components of the distributed network in requirement metadata from a network administrator.
The method 200 may be practiced where the logical switch automatically configuring third party solutions in the distributed network to meet the requirements for the virtual components of the distributed network comprises the logical switch directly configuring third party solutions by communicating with the third party solutions using native third party solution protocols. For example, communicating with third party solutions using third party solution protocols may include the logical switch communicating using a plug-in at the logical switch that enables the plug-in to communicate directly with third party solutions in third party solution's native protocol. Alternatively or additionally, the method of claim 5, wherein communicating with third party solutions using third party solution protocols comprises the logical switch communicating directly with third party solutions using APIs exposed by the third party solutions.
Further, the methods may be practiced by a computer system including one or more processors and computer readable media such as computer memory. In particular, the computer memory may store computer executable instructions that when executed by one or more processors cause various functions to be performed, such as the acts recited in the embodiments.
Embodiments of the present invention may comprise or utilize a special purpose or general-purpose computer including computer hardware, as discussed in greater detail below. Embodiments within the scope of the present invention also include physical and other computer-readable media for carrying or storing computer-executable instructions and/or data structures. Such computer-readable media can be any available media that can be accessed by a general purpose or special purpose computer system. Computer-readable media that store computer-executable instructions are physical storage media. Computer-readable media that carry computer-executable instructions are transmission media. Thus, by way of example, and not limitation, embodiments of the invention can comprise at least two distinctly different kinds of computer-readable media: physical computer readable storage media and transmission computer readable media.
Physical computer readable storage media includes RAM, ROM, EEPROM, CD-ROM or other optical disk storage (such as CDs, DVDs, etc), magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer.
A “network” is defined as one or more data links that enable the transport of electronic data between computer systems and/or modules and/or other electronic devices. When information is transferred or provided over a network or another communications connection (either hardwired, wireless, or a combination of hardwired or wireless) to a computer, the computer properly views the connection as a transmission medium. Transmissions media can include a network and/or data links which can be used to carry or desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer. Combinations of the above are also included within the scope of computer-readable media.
Further, upon reaching various computer system components, program code means in the form of computer-executable instructions or data structures can be transferred automatically from transmission computer readable media to physical computer readable storage media (or vice versa). For example, computer-executable instructions or data structures received over a network or data link can be buffered in RAM within a network interface module (e.g., a “NIC”), and then eventually transferred to computer system RAM and/or to less volatile computer readable physical storage media at a computer system. Thus, computer readable physical storage media can be included in computer system components that also (or even primarily) utilize transmission media.
Computer-executable instructions comprise, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, or even source code. Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the described features or acts described above. Rather, the described features and acts are disclosed as example forms of implementing the claims.
Those skilled in the art will appreciate that the invention may be practiced in network computing environments with many types of computer system configurations, including, personal computers, desktop computers, laptop computers, message processors, hand-held devices, multi-processor systems, microprocessor-based or programmable consumer electronics, network PCs, minicomputers, mainframe computers, mobile telephones, PDAs, pagers, routers, switches, and the like. The invention may also be practiced in distributed system environments where local and remote computer systems, which are linked (either by hardwired data links, wireless data links, or by a combination of hardwired and wireless data links) through a network, both perform tasks. In a distributed system environment, program modules may be located in both local and remote memory storage devices.
Alternatively, or in addition, the functionally described herein can be performed, at least in part, by one or more hardware logic components. For example, and without limitation, illustrative types of hardware logic components that can be used include Field-programmable Gate Arrays (FPGAs), Program-specific Integrated Circuits (ASICs), Program-specific Standard Products (ASSPs), System-on-a-chip systems (SOCs), Complex Programmable Logic Devices (CPLDs), etc.
The present invention may be embodied in other specific forms without departing from its spirit or characteristics. The described embodiments are to be considered in all respects only as illustrative and not restrictive. The scope of the invention is, therefore, indicated by the appended claims rather than by the foregoing description. All changes which come within the meaning and range of equivalency of the claims are to be embraced within their scope.

Claims (21)

What is claimed is:
1. A computer-implemented method of configuring third party solutions to operate with virtual machines in a distributed network environment, the computer-implemented method being performed by one or more processors when executing computer executable instructions for implementing the method, and wherein the computer-implemented method comprises:
monitoring at a server configured as a virtual machine monitor a plurality of requirements for virtual components of physical host machines running at a datacenter,
each physical host machine being connected to a physical switch, and
each physical host machine running one or more virtual machines that are connected to one another at the physical host machine and to other host components through a virtual switch that connects to other virtual machines on the physical host machine or other components on other host components by the physical switch;
providing a single virtual switch across the physical host machines running at a datacenter, physical host machines being connected to the logical switch through one or more of the virtual switches at the virtual machines running at the physical hosts,
wherein the single virtual switch coordinates network connectivity and policy management across all physical host machines connected to the single virtual switch by performing the following:
receiving information at the single virtual switch about one or more third party solutions based on network services comprising both third party software and hardware device solutions across all said physical host machines connected to the single virtual switch;
receiving information at the single virtual switch about said requirements for virtual components of the physical host machines; and
based on the information received about the one or more third party solutions and the requirements for virtual components of the physical host machines, the single virtual switch automatically configuring at least one of the third party solutions to meet the requirements of the virtual components of at least one physical host machine at which one or more virtual machines are deployed on behalf of the at least one third party solution.
2. The computer-implemented method of claim 1, wherein the various levels of the configured function of the single virtual switch comprises at least one of:
a configuration to interact with the third party solutions by using a plugin;
configuring lost level extensions with drivers installed at the physical host machines;
configuring interface extensions that can be plugged into a user interface portion of the virtual machine monitor; and
configuring the virtual machine monitor with native hardware for one or more application program interfaces (APIs) that communicate natively to the third party solutions.
3. The computer-implemented method of claim 1, wherein receiving information at the single virtual switch about the third party solutions is performed by receiving at the single virtual switch configuration metadata from a network administrator.
4. The computer-implemented method of claim 1, wherein receiving information at the single virtual switch about the requirements for virtual components comprises receiving information about the requirements for virtual components in the form of metadata from a network administrator.
5. The computer-implemented method of claim 1 wherein the single virtual switch automatically configuring the third party solutions to meet the requirements of the virtual components comprises the single virtual switch directly configuring third party solutions by communicating with the third party solutions using native third party solution protocols.
6. The computer-implemented method of claim 5, wherein communicating with third party solutions using third party solution protocols comprises the single virtual switch communicating using a plug-in that enables the plug-in to communicate directly with third party solutions in the third party solution's native protocol.
7. The computer-implemented method of claim 5, wherein communicating with third party solutions using third party solution protocols comprises the single virtual switch communicating directly with third party solutions using application program interfaces (APIs) exposed by the third party solutions.
8. The computer-implemented method of claim 1, wherein the third party solution comprises a top of rack (TOR) switch.
9. In a distributed computing environment, a system for configuring third party solutions to operate with virtual machines in a distributed network environment, the system comprising:
one or more processors; and
one or more computer readable devices containing computer executable instructions that when executed by the one or more processors cause the system to perform the following computer-implemented method:
monitoring at a server configured as a virtual machine monitor a plurality of requirements for virtual components of physical host machines running at a datacenter,
each physical host machine being connected to a physical switch, and
each physical host machine running one or more virtual machines that are connected to one another at the physical host machine and to other host components through a virtual switch that connects to other virtual machines on the physical host machine or other components on other host components by the physical switch;
providing a single virtual switch across the physical host machines running at a datacenter, physical host machines being connected to the logical switch through one or more of the virtual switches at the virtual machines running at the physical hosts,
wherein the single virtual switch coordinates network connectivity and policy management across all physical host machines connected to the single virtual switch by performing the following:
receiving information at the single virtual switch about one or more third party solutions based on network services comprising both third party software and hardware device solutions across all said physical host machines connected to the single virtual switch;
receiving information at the single virtual switch about said requirements for virtual components of the physical host machines; and
based on the information received about the one or more third party solutions and the requirements for virtual components of the physical host machines, the single virtual switch automatically configuring at least one of the third party solutions to meet the requirements of the virtual components of at least one physical host machine at which one or more virtual machines are deployed on behalf of the at least one third party solution.
10. The system of claim 9, wherein the various levels of the configured function of the single virtual switch comprises at least one of:
a configuration to interact with the third party solutions by using a plugin;
configuring lost level extensions with drivers installed at the physical host machines;
configuring interface extensions that can be plugged into a user interface portion of the virtual machine monitor; and
configuring the virtual machine monitor with native hardware APIs that communicate natively to the third party solutions.
11. The system of claim 9, wherein receiving information at the single virtual switch about the third party solutions is performed by receiving at the single virtual switch configuration metadata from a network administrator.
12. The system of claim 9, wherein receiving information at the single virtual switch about the requirements for virtual components comprises receiving information about the requirements for virtual components in the form of metadata from a network administrator.
13. The system of claim 9, wherein the single virtual switch automatically configuring the third party solutions to meet the requirements of the virtual components comprises the single virtual switch directly configuring third party solutions by communicating with the third party solutions using native third party solution protocols.
14. The system of claim 9, wherein communicating with third party solutions using third party solution protocols comprises the single virtual switch communicating using a plug-in that enables the plug-in to communicate directly with third party solutions in the third party solution's native protocol.
15. The system of claim 9, wherein communicating with third party solutions using third party solution protocols comprises the single virtual switch communicating directly with third party solutions using application program interfaces (APIs) exposed by the third party solutions.
16. The system of claim 9, wherein the third party solution comprises a top of rack (TOR) switch.
17. A computing system comprising:
a server comprising a virtual machine monitor that monitors a plurality of requirements for virtual components of physical host machines running at a datacenter,
each physical host machine being connected to a physical switch, and
each physical host machine running one or more virtual machines that are connected to one another at the physical host machine and to other host components through a virtual switch that connects to other virtual machines on the physical host machine or other components on other host components by the physical switch;
a single virtual switch across the physical host machines running at the datacenter, physical host machines being connected to the logical switch through one or more of the virtual switches at the virtual machines running at the physical hosts,
wherein the single virtual switch coordinates network connectivity and policy management across all physical host machines connected to the single virtual switch by performing the following:
receiving information at the single virtual switch about one or more third party solutions based on network services comprising both third party software and hardware device solutions across all said physical host machines connected to the single virtual switch;
receiving information at the single virtual switch about said requirements for virtual components of the physical host machines; and
based on the information received about the one or more third party solutions and the requirements for virtual components of the physical host machines, the single virtual switch automatically configuring at least one of the third party solutions to meet the requirements of the virtual components of at least one physical host machine at which one or more virtual machines are deployed on behalf of the at least one third party solution.
18. The computing system of claim 17, wherein receiving information at the single virtual switch about the third party solutions is performed by receiving at the single virtual switch configuration metadata from a network administrator, and wherein receiving information at the single virtual switch about the requirements for virtual components comprises receiving information about the requirements for virtual components in the form of metadata from a network administrator.
19. The computing system of claim 17, wherein the single virtual switch automatically configuring the third party solutions to meet the requirements of the virtual components comprises the single virtual switch directly configuring third party solutions by communicating with the third party solutions using native third party solution protocols.
20. The computing system of claim 19, wherein communicating with third party solutions using third party solution protocols comprises at least one of the following:
the single virtual switch communicating using a plug-in that enables the plug-in to communicate directly with third party solutions in the third party solution's native protocol; and
the single virtual switch communicating directly with third party solutions using application program interfaces (APIs) exposed by the third party solutions.
21. A computer-implemented method performed by one or more processors when executing computer executable instructions for implementing the method, and wherein the computer-implemented method comprises:
receiving information at a logical switch about one or more third party solutions based on network services comprising both third party software and hardware device solutions across a plurality of physical host machines connected to the logical switch;
wherein each physical host machine runs one or more virtual machines that are connected to one another at the physical host machine and to other host components through a virtual switch that connects to other virtual machines on the physical host machine or other components on other host components by a physical switch;
receiving information at the logical switch about said requirements for virtual components of the physical host machines; and
based on the information received about the one or more third party solutions and the requirements for virtual components of the physical host machines, the logical switch automatically configuring at least one of the third party solutions to meet the requirements of the virtual components of at least one physical host machine at which one or more virtual machines are deployed on behalf of the at least one third party solution.
US13/925,551 2013-06-24 2013-06-24 Logical switch Active 2033-08-29 US9270619B2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US13/925,551 US9270619B2 (en) 2013-06-24 2013-06-24 Logical switch
PCT/US2014/043593 WO2014209846A1 (en) 2013-06-24 2014-06-23 Configuring third party solutions through a logical switch interface
CN201480036045.2A CN105409172B (en) 2013-06-24 2014-06-23 Logical switch

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US13/925,551 US9270619B2 (en) 2013-06-24 2013-06-24 Logical switch

Publications (2)

Publication Number Publication Date
US20140376560A1 US20140376560A1 (en) 2014-12-25
US9270619B2 true US9270619B2 (en) 2016-02-23

Family

ID=51210799

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/925,551 Active 2033-08-29 US9270619B2 (en) 2013-06-24 2013-06-24 Logical switch

Country Status (3)

Country Link
US (1) US9270619B2 (en)
CN (1) CN105409172B (en)
WO (1) WO2014209846A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170054686A1 (en) * 2015-08-18 2017-02-23 Konstantin Malkov Agentless Security of Virtual Machines using a Filtering Platform
US11436053B2 (en) * 2019-05-24 2022-09-06 Microsoft Technology Licensing, Llc Third-party hardware integration in virtual networks

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8516241B2 (en) 2011-07-12 2013-08-20 Cisco Technology, Inc. Zone-based firewall policy model for a virtualized data center
JP2015122680A (en) * 2013-12-25 2015-07-02 株式会社日立製作所 Logical network construction method, and network system
US10075377B1 (en) 2015-04-23 2018-09-11 Cisco Technology, Inc. Statistical collection in a network switch natively configured as a load balancer
US10142231B2 (en) * 2016-03-31 2018-11-27 Intel Corporation Technologies for network I/O access
CN107294752A (en) * 2016-04-01 2017-10-24 中兴通讯股份有限公司 Realize framework, the method and device of network function communication
US10341371B2 (en) 2016-08-31 2019-07-02 Nicira, Inc. Identifying and handling threats to data compute nodes in public cloud
US10484302B2 (en) 2016-08-27 2019-11-19 Nicira, Inc. Managed forwarding element executing in public cloud data compute node with different internal and external network addresses
US10567482B2 (en) 2017-08-24 2020-02-18 Nicira, Inc. Accessing endpoints in logical networks and public cloud service providers native networks using a single network interface and a single routing table
US10491516B2 (en) 2017-08-24 2019-11-26 Nicira, Inc. Packet communication between logical networks and public cloud service providers native networks using a single network interface and a single routing table
CN114584465A (en) 2017-08-27 2022-06-03 Nicira股份有限公司 Executing online services in a public cloud
US10601705B2 (en) 2017-12-04 2020-03-24 Nicira, Inc. Failover of centralized routers in public cloud logical networks
US10862753B2 (en) 2017-12-04 2020-12-08 Nicira, Inc. High availability for stateful services in public cloud logical networks
US11343229B2 (en) 2018-06-28 2022-05-24 Vmware, Inc. Managed forwarding element detecting invalid packet addresses
US10491466B1 (en) 2018-08-24 2019-11-26 Vmware, Inc. Intelligent use of peering in public cloud
US11196591B2 (en) 2018-08-24 2021-12-07 Vmware, Inc. Centralized overlay gateway in public cloud
US11374794B2 (en) 2018-08-24 2022-06-28 Vmware, Inc. Transitive routing in public cloud
US11064018B1 (en) * 2020-01-15 2021-07-13 Vmware, Inc. Incorporating software defined networking resource utilization in workload placement

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050120160A1 (en) * 2003-08-20 2005-06-02 Jerry Plouffe System and method for managing virtual servers
US20070067435A1 (en) 2003-10-08 2007-03-22 Landis John A Virtual data center that allocates and manages system resources across multiple nodes
EP2431883A2 (en) 2010-09-20 2012-03-21 Huawei Technologies Co., Ltd. Method, network management center, and a related device for configuring a network policy for a virtual port
WO2012109868A1 (en) 2011-08-01 2012-08-23 华为技术有限公司 Network policy configuration method, management device and network management centre device
US8346935B2 (en) 2010-01-15 2013-01-01 Joyent, Inc. Managing hardware resources by sending messages amongst servers in a data center
US8345692B2 (en) 2010-04-27 2013-01-01 Cisco Technology, Inc. Virtual switching overlay for cloud computing
US8387060B2 (en) 2009-10-01 2013-02-26 Dell Products L.P. Virtual machine resource allocation group policy based on workload profile, application utilization and resource utilization
US20130058229A1 (en) * 2010-07-06 2013-03-07 Martin Casado Method and apparatus for robust packet distribution among hierarchical managed switching elements
US20130070762A1 (en) 2011-09-20 2013-03-21 Robert Edward Adams System and methods for controlling network traffic through virtual switches
US20130329725A1 (en) * 2012-06-06 2013-12-12 Juniper Networks, Inc. Facilitating operation of one or more virtual networks
US8806005B2 (en) 2011-09-12 2014-08-12 Microsoft Corporation Cross-machine event log correlation

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050120160A1 (en) * 2003-08-20 2005-06-02 Jerry Plouffe System and method for managing virtual servers
US20070067435A1 (en) 2003-10-08 2007-03-22 Landis John A Virtual data center that allocates and manages system resources across multiple nodes
US8387060B2 (en) 2009-10-01 2013-02-26 Dell Products L.P. Virtual machine resource allocation group policy based on workload profile, application utilization and resource utilization
US8346935B2 (en) 2010-01-15 2013-01-01 Joyent, Inc. Managing hardware resources by sending messages amongst servers in a data center
US8345692B2 (en) 2010-04-27 2013-01-01 Cisco Technology, Inc. Virtual switching overlay for cloud computing
US20130058229A1 (en) * 2010-07-06 2013-03-07 Martin Casado Method and apparatus for robust packet distribution among hierarchical managed switching elements
EP2431883A2 (en) 2010-09-20 2012-03-21 Huawei Technologies Co., Ltd. Method, network management center, and a related device for configuring a network policy for a virtual port
WO2012109868A1 (en) 2011-08-01 2012-08-23 华为技术有限公司 Network policy configuration method, management device and network management centre device
US8806005B2 (en) 2011-09-12 2014-08-12 Microsoft Corporation Cross-machine event log correlation
US20130070762A1 (en) 2011-09-20 2013-03-21 Robert Edward Adams System and methods for controlling network traffic through virtual switches
US20130329725A1 (en) * 2012-06-06 2013-12-12 Juniper Networks, Inc. Facilitating operation of one or more virtual networks

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
"How to Configure Network Settings on a Host by Applying a Logical Switch in VMM," TechNet Library for System Center 2012, 2013 Microsoft Available at: http://technet.microsoft.com/en-us/library/jj628156(d=printer).aspx.
"How to Create a Logical Switch in VMM," TechNet Library for System Center 2012, 2013 Microsoft Available at: http://technet.microsoft.com/en-us/library/jj628154(d=printer).aspx.
"Impact of Virtualization on Cloud Networking", In White Paper of Arista Networks, Retrieved on: May 1, 2013, 7 pages.
"International Search Report and Written Opinion received for PCT Patent Application No. PCT/US2014/043593", Mailed Date: Nov. 3, 2014, 14 pages.
"Simplify Virtual Machine Management and Migration with Ethernet Fabrics in the Datacenter", In White Paper of Microsoft-Virtualization, Mar. 2011, 20 pages.
"Windows Server 2012 Hyper-V Component Architecture", Published on: Feb. 29, 2012, Available at: http://download.microsoft.com/download/8/6/D/86D659D0-127F-4723-9D46-5AF03F8A92F3/Windows%20Server%20E2%80%9C8%E2%80%9D%20Beta%20Hyper-V%20Component%20Architecture%20Poster.pdf.
Cain, et al., "Networking in VMM 2012 SP1-Logical Networks (Part I)", Published on: Feb. 14, 2013, Available at: http://web.archive.org/web/20130606053029/http://blogs.technet.com/b/scvmm/archive/2013/02/14/networking-in-vmm-2012-sp1-logical-networks-part-i-aspx.
Gautreau, Brian, "Dell Reference Architecture for Hyper-V, PowerEdge Blade Servers, Force10 Switches, and Compellent Storage Center", In Dell Technical White Paper, Nov. 2011, 30 pages.

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170054686A1 (en) * 2015-08-18 2017-02-23 Konstantin Malkov Agentless Security of Virtual Machines using a Filtering Platform
US11436053B2 (en) * 2019-05-24 2022-09-06 Microsoft Technology Licensing, Llc Third-party hardware integration in virtual networks

Also Published As

Publication number Publication date
CN105409172A (en) 2016-03-16
WO2014209846A1 (en) 2014-12-31
US20140376560A1 (en) 2014-12-25
CN105409172B (en) 2019-03-29

Similar Documents

Publication Publication Date Title
US9270619B2 (en) Logical switch
US11716309B1 (en) Allocating external IP addresses from isolated pools
US10708082B1 (en) Unified control plane for nested clusters in a virtualized computing infrastructure
US11507401B2 (en) Framework for networking and security services in virtual networks
US11171834B1 (en) Distributed virtualized computing infrastructure management
US10841226B2 (en) Configuring service load balancers with specified backend virtual networks
US11128494B2 (en) Distributed virtual gateway appliance
US11863352B2 (en) Hierarchical networking for nested container clusters
US11870642B2 (en) Network policy generation for continuous deployment
US11748132B2 (en) Apparatus and method for configuring and enabling virtual applications
US11700236B2 (en) Packet steering to a host-based firewall in virtualized environments
US20230079209A1 (en) Containerized routing protocol process for virtual private networks
US20230104368A1 (en) Role-based access control autogeneration in a cloud native software-defined network architecture
US11863378B2 (en) Automated provisioning of endpoint devices with management connectivity
US10469374B2 (en) Multiple provider framework for virtual switch data planes and data plane migration
WO2023003812A1 (en) Automated provisioning of endpoint devices with management connectivity
US11444836B1 (en) Multiple clusters managed by software-defined network (SDN) controller
US20240129161A1 (en) Network segmentation for container orchestration platforms
US20230106531A1 (en) Virtual network routers for cloud native software-defined network architectures
US20240078123A1 (en) Auto-scaling service mesh for virtual machines
US20230409369A1 (en) Metric groups for software-defined network architectures
CN117255019A (en) System, method, and storage medium for virtualizing computing infrastructure

Legal Events

Date Code Title Description
AS Assignment

Owner name: MICROSOFT CORPORATION, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SENNIAPPAN, PRADEEP;VARAVA, NATALIA VALERYEVNA;RAVICHANDRAN, KARTHIKEYAN NENMELI;AND OTHERS;SIGNING DATES FROM 20130610 TO 20130701;REEL/FRAME:030944/0237

AS Assignment

Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034747/0417

Effective date: 20141014

Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:039025/0454

Effective date: 20141014

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8