architecture and design of vmware nsx-t for workload ... · architecture and design of vmware nsx-t...

48
Architecture and Design of VMware NSX-T for Workload Domains 19 MAR 2019 VMware Validated Design 5.0.1 VMware NSX-T 2.4

Upload: others

Post on 24-May-2020

20 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Architecture and Designof VMware NSX-T forWorkload Domains19 MAR 2019VMware Validated Design 5.0.1VMware NSX-T 2.4

Page 2: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 2

You can find the most up-to-date technical documentation on the VMware website at:

https://docs.vmware.com/

If you have comments about this documentation, submit your feedback to

[email protected]

Copyright © 2018–2019 VMware, Inc. All rights reserved. Copyright and trademark information.

VMware, Inc.3401 Hillview Ave.Palo Alto, CA 94304www.vmware.com

Page 3: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Contents

About Architecture and Design of VMware NSX-T for Workload Domains 4

1 Applying the Guidance for NSX-T for Workload Domains 5

2 Architecture Overview 7

Physical Network Architecture 7

Virtual Infrastructure Architecture 9

3 Detailed Design 16

Physical Infrastructure Design 16

Virtual Infrastructure Design 21

VMware, Inc. 3

Page 4: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

About Architecture and Design ofVMware NSX-T for Workload Domains

Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about therequirements for software, tools, and external services to implement VMware NSX-T® in a shared edgeand compute cluster in an SDDC that is compliant with VMware Validated Design for Software-DefinedData Center.

PrerequisitesDeploy the management cluster according to VMware Validated Design for Software-Defined Data Centerat least in a single region. See the VMware Validated Design documentation page.

Intended AudienceThis design is intended for architects and administrators who want to deploy NSX-T in a virtualinfrastructure workload domain for tenant workloads.

Required VMware SoftwareIn addition to the VMware Validated Design for Software-Defined Data Center 5.0.1 deployment, you mustdownload NSX-T 2.4. You then deploy and configure NSX-T in the shared edge and compute clusteraccording to this guide. See VMware Validated Design Release Notes for more information aboutsupported product versions

VMware, Inc. 4

Page 5: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Applying the Guidance for NSX-T for Workload Domains 1The content in Architecture and Design of VMware NSX-T for Workload Domains replaces certain parts ofArchitecture and Design in VMware Validated Design for Software-Defined Data Center, also referred toas the Standard SDDC.

Before You Design the Virtual Infrastructure WorkloadDomain with NSX-TBefore you follow this documentation, you must deploy the components for the SDDC managementcluster according to VMware Validated Design for Software-Defined Data Center at least in a singleregion. See Architecture and Design, Planning and Preparation and Deployment for Region A in the VMware Validated Design documentation.

n ESXi

n Platform Services Controller pair and Management vCenter Server

n NSX for vSphere

n vRealize Lifecycle Manager

n vSphere Update Manager

n vRealize Operations Manager

n vRealize Log Insight

n vRealize Automation with embedded vRealize Orchestrator

n vRealize Business

VMware, Inc. 5

Page 6: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Designing a Virtual Infrastructure Workload Domain withNSX-TNext, follow the guidance to design a virtual infrastructure (VI) workload domain with NSX-T deployed inthis way:

n In general, use the guidelines about the VI workload domain and shared edge and compute cluster inthe following sections of Architecture and Design in VMware Validated Design for Software-DefinedData Center:

n Architecture Overview > Physical Infrastructure Architecture

n Architecture Overview > Virtual Infrastructure Architecture

n Detailed Design > Physical Infrastructure Design

n Detailed Design > Virtual Infrastructure Design

n For the sections that are available in both Architecture and Design of VMware NSX-T for WorkloadDomains and Architecture and Design, follow the design guidelines in Architecture and Design ofVMware NSX-T for Workload Domains.

First-Level Chapter Places to Use the Guidance for NSX-T

Architecture Overview n Physical Infrastructure Architecturen Workload Domain Architecturen Cluster Typesn Physical Network Architecture

n Network Transportn Infrastructure Network Architecturen Physical Network Interfaces

n Virtual Infrastructure Architecture

Detailed Design n Physical Infrastructure Designn Physical Design Fundamentalsn Physical Networking Designn Physical Storage Design

n Virtual Infrastructure Designn vCenter Server Design

n vCenter Server Deploymentn vCenter Server Networkingn vCenter Server Redundancyn vCenter Server Appliance Sizingn vSphere Cluster Designn vCenter Server Customizationn Use of TLS Certificates in vCenter Server

n Virtualization Network Designn NSX-T Design

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 6

Page 7: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Architecture Overview 2VMware Validated Design for NSX-T enables IT organizations that have deployed VMware ValidatedDesign for Software-Defined Data Center 5.0 to create a shared edge and compute cluster that usesNSX-T capabilities.

This chapter includes the following topics:

n Physical Network Architecture

n Virtual Infrastructure Architecture

Physical Network ArchitectureVMware Validated Designs can use most physical network architectures.

Network TransportYou can implement the physical layer switch fabric of an SDDC by offering Layer 2 or Layer 3 transportservices. For a scalable and vendor-neutral data center network, use a Layer 3 transport.

VMware Validated Design supports both Layer 2 and Layer 3 transports. To decide whether to use Layer2 or Layer 3, consider the following factors:

n NSX-T service routers establish Layer 3 routing adjacency with the first upstream Layer 3 device toprovide equal cost routing for workloads.

n The investment you have today in your current physical network infrastructure.

n The benefits and drawbacks for both layer 2 and layer 3 designs.

Benefits and Drawbacks of Layer 2 TransportA design using Layer 2 transport has these considerations:

n In a design that uses Layer 2 transport, top of rack switches and upstream Layer 3 devices, such ascore switches or routers, form a switched fabric.

n The upstream Layer 3 device terminates each VLAN and provides default gateway functionality.

n Uplinks from the top of rack switch to the upstream Layer 3 devices are 802.1Q trunks carrying allrequired VLANs.

Using a Layer 2 transport has the following benefits and drawbacks:

VMware, Inc. 7

Page 8: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 2‑1. Benefits and Drawbacks for Layer 2 Transport

Characteristic Description

Benefits n More design freedom.n You can span VLANs across racks.

Drawbacks n The size of such a deployment is limited because the fabricelements have to share a limited number of VLANs.

n You might have to rely on a specialized data centerswitching fabric product from a single vendor.

n Traffic between VLANs must traverse to upstream Layer 3device to be routed.

Figure 2‑1. Example Layer 2 Transport

ToR ToR

ESXiHost

Upstream L3 DeviceL3

25 GbE 25 GbE

L2

Upstream L3 Device

Benefits and Drawbacks of Layer 3 TransportA design using Layer 3 transport requires these considerations:

n Layer 2 connectivity is limited within the data center rack up to the top of rack switches.

n The top of rack switch terminates each VLAN and provides default gateway functionality. The top ofrack switch has a switch virtual interface (SVI) for each VLAN.

n Uplinks from the top of rack switch to the upstream layer are routed point-to-point links. You cannotuse VLAN trunking on the uplinks.

n A dynamic routing protocol, such as BGP, connects the top of rack switches and upstream switches.Each top of rack switch in the rack advertises a small set of prefixes, typically one per VLAN orsubnet. In turn, the top of rack switch calculates equal cost paths to the prefixes it receives from othertop of rack switches.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 8

Page 9: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 2‑2. Benefits and Drawbacks of Layer 3 Transport

Characteristic Description

Benefits n You can select from many Layer 3 capable switch productsfor the physical switching fabric.

n You can mix switches from different vendors because ofgeneral interoperability between their implementation ofBGP.

n This approach is typically more cost effective because ituses only the basic functionality of the physical switches.

Drawbacks n VLANs are restricted to a single rack. The restriction canaffect vSphere Fault Tolerance, and storage networks.

Figure 2‑2. Example Layer 3 Transport

ToR

L3

ESXiHost

UpstreamSwitch

UpstreamSwitch

WAN/MPLSInternet

WAN/MPLSInternet

ToR

L2

25 GbE25 GbE

Virtual Infrastructure ArchitectureThe virtual infrastructure is the foundation of an operational SDDC. It contains the software-definedinfrastructure, software-defined networking and software-defined storage.

In the virtual infrastructure layer, access to the underlying physical infrastructure is controlled andallocated to the management and compute workloads. The virtual infrastructure layer consists of thehypervisors on the physical hosts and the control of these hypervisors. The management components ofthe SDDC consist of elements in the virtual management layer itself.

Figure 2‑3. Virtual Infrastructure Layer in the SDDC

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 9

Page 10: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

ServiceManagement

Portfolio Management

OperationsManagement

CloudManagement

Layer

Service Catalog

Self-Service Portal

Orchestration

BusinessContinuity

Fault Tolerance and Disaster

Recovery

Backup & Restore

Hypervisor

Pools of Resources

Virtualization Control

VirtualInfrastructure

Layer

Compute

Storage

Network

PhysicalLayer

Security

Replication Compliance

Risk

Governance

Virtual Infrastructure OverviewThe SDDC virtual infrastructure consists of workload domains. The SDDC virtual infrastructure includes amanagement workload domain that contains the management cluster and a virtual infrastructure workloaddomain that contains the shared edge and compute cluster.

Management ClusterThe management cluster runs the virtual machines that manage the SDDC. These virtual machines hostvCenter Server, vSphere Update Manager, NSX Manager, and other management components. Allmanagement, monitoring, and infrastructure services are provisioned to a vSphere cluster which provideshigh availability for these critical services. Permissions on the management cluster limit access only toadministrators. This limitation protects the virtual machines that are running the management, monitoring,and infrastructure services from unauthorized access. The management cluster leverages software-defined networking capabilities in NSX for vSphere.

The management cluster architecture and design is covered in the VMware Validated Design forSoftware-Defined Data Center. The NSX-T validated design does not include the design of themanagement cluster.

Shared Edge and Compute ClusterThe shared edge and compute cluster runs the NSX-T edge virtual machines and all tenant workloads.The edge virtual machines are responsible for North-South routing between compute workloads and theexternal network. This is often referred to as the on-off ramp of the SDDC.

The hosts in this cluster provide services such as high availability to the NSX-T edge virtual machinesand tenant workloads.

Figure 2‑4. SDDC Logical Design

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 10

Page 11: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Network Virtualization ComponentsThe NSX-T platform consists of several components that are relevant to the network virtualization design.

NSX-T PlatformNSX-T creates a network virtualization layer, which is an abstraction between the physical and virtualnetworks. You create all virtual networks on top of this layer.

Several components are required to create this network virtualization layer:

n NSX-T Managers

n NSX-T Edge Nodes

n NSX-T Distributed Routers (DR)

n NSX-T Service Routers (SR)

n NSX-T Segments (Logical Switches)

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 11

Page 12: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

These components are distributed in different planes to create communication boundaries and provideisolation of workload data from system control messages.

Data plane Performs stateless forwarding or transformation of packets based on tablespopulated by the control plane, reports topology information to the controlplane, and maintains packet level statistics.

The following traffic runs in the data plane:

n Workload data

n N-VDS virtual switch, distributed routing, and the distributed firewall inNSX-T

The data is carried over designated transport networks in the physicalnetwork.

Control plane Contains messages for network virtualization control. You place the controlplane communication on secure physical networks (VLANs) that areisolated from the transport networks for the data plane.

The control plane computes the runtime state based on configuration fromthe management plane. Control plane propagates topology informationreported by the data plane elements, and pushes stateless configuration toforwarding engines.

Control plane in NSX-T has two parts:

n Central Control Plane (CCP). The CCP is implemented as a cluster ofvirtual machines called CCP nodes. The cluster form factor providesboth redundancy and scalability of resources.

The CCP is logically separated from all data plane traffic, that is, afailure in the control plane does not affect existing data planeoperations.

n Local Control Plane (LCP). The LCP runs on transport nodes. It is nearto the data plane it controls and is connected to the CCP. The LCP isresponsible for programming the forwarding entries of the data plane.

Management plane Provides a single API entry point to the system, persists user configuration,handles user queries, and performs operational tasks on all management,control, and data plane nodes in the system.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 12

Page 13: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

For NSX-T, all querying, modifying, and persisting user configuration is inthe management plane. Propagation of that configuration down to thecorrect subset of data plane elements is in the control plane. As a result,some data belongs to multiple planes. Each plane uses this data accordingto stage of existence. The management plane also queries recent statusand statistics from the control plane, and under certain conditions directlyfrom the data plane.

The management plane is the only source of truth for the logical systembecause it is the only entry point for user configuration. You make changesusing either a RESTful API or the NSX-T user interface.

For example, responding to a vSphere vMotion operation of a virtualmachine is responsibility of the control plane, but connecting the virtualmachine to the logical network is responsibility of the management plane.

Network Virtualization ServicesNetwork virtualization services include segments, gateways, firewalls, and other components of NSX-T.

Segments (LogicalSwitch)

Reproduces switching functionality, broadcast, unknown unicast, andmulticast (BUM) traffic in a virtual environment that is decoupled from theunderlying hardware.

Segments are similar to VLANs because they provide network connectionsto which you can attach virtual machines. The virtual machines can thencommunicate with each other over tunnels between ESXi hosts. EachSegment has a virtual network identifier (VNI), like a VLAN ID. UnlikeVLANs, VNIs scale beyond the limits of VLAN IDs.

Gateway (LogicalRouter)

Provides North-South connectivity so that workloads can access externalnetworks, and East-West connectivity between logical networks.

A Logical Router is a configured partition of a traditional network hardwarerouter. It replicates the functionality of the hardware, creating multiplerouting domains in a single router. Logical Routers perform a subset of thetasks that are handled by the physical router, and each can contain multiple

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 13

Page 14: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

routing instances and routing tables. Using logical routers can be aneffective way to maximize router use, because a set of logical routers withina single physical router can perform the operations previously performed byseveral pieces of equipment.

n Distributed router (DR)

A DR spans ESXi hosts whose virtual machines are connected to thisgateway, and edge nodes the gateway is bound to. Functionally, theDR is responsible for one-hop distributed routing between segmentsand gateways connected to this gateway.

n One or more (optional) service routers (SR).

An SR is responsible for delivering services that are not currentlyimplemented in a distributed fashion, such as stateful NAT.

A gateway always has a DR. A gateway has SRs when it is a Tier-0gateway, or when it is a Tier-1 gateway and has routing services configuredsuch as NAT or DHCP.

NSX-T Edge Node Provides routing services and connectivity to networks that are external tothe NSX-T domain through a Tier-0 gateway over BGP or static routing.

You must deploy an NSX-T Edge for stateful services at either the Tier-0 orTier-1 gateways.

NSX-T Edge Cluster Represents a collection of NSX-T Edge nodes that host multiple servicerouters in highly available configurations. At a minimum, deploy a singleTier-0 SR to provide external connectivity.

An NSX-T Edge cluster does not have a one-to-one relationship with avSphere cluster. A vSphere cluster can run multiple NSX-T Edge clusters.

Transport Node Participates in NSX-T overlay or NSX-T VLAN networking. If a nodecontains an NSX-T Virtual Distributed Switch (N-VDS) such as ESXi hostsand NSX-T Edge nodes, it can be a transport node.

If an ESXi host contains at least one N-VDS, it can be a transport node.

Transport Zone A transport zone can span one or more vSphere clusters. Transport zonesdictate which ESXi hosts and which virtual machines can participate in theuse of a particular network.

A transport zone defines a collection of ESXi hosts that can communicatewith each other across a physical network infrastructure. Thiscommunication happens over one or more interfaces defined as TunnelEndpoints (TEPs).

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 14

Page 15: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

When you create an ESXi host transport node and then add the node to atransport zone, NSX-T installs an N-VDS on the host. For each transportzone that the host belongs to, a separate N-VDS is installed. The N-VDS isused for attaching virtual machines to NSX-T Segments and for creatingNSX-T gateway uplinks and downlinks.

NSX-T Controller As a component of the control plane, the controllers control virtual networksand overlay transport tunnels.

For stability and reliability of data transport, NSX-T deploys the NSX-TController as a role in the Manager cluster which consists of three highlyavailable virtual appliances. They are responsible for the programmaticdeployment of virtual networks across the entire NSX-T architecture.

Logical Firewall Responsible for traffic handling in and out the network according to firewallrules.

A logical firewall offers multiple sets of configurable Layer 3 and Layer 2rules. Layer 2 firewall rules are processed before Layer 3 rules. You canconfigure an exclusion list to exclude segments, logical ports, or groupsfrom firewall enforcement.

The default rule, located at the bottom of the rule table, is a catch-all rule.The logical firewall enforces the default rule on packets that do not matchother rules. After the host preparation operation, the default rule is set tothe allow action. Change this default rule to a block action and enforceaccess control through a positive control model, that is, only traffic definedin a firewall rule can flow on the network.

Logical Load Balancer Provides high-availability service for applications and distributes thenetwork traffic load among multiple servers.

The load balancer accepts TCP, UDP, HTTP, or HTTPS requests on thevirtual IP address and determines which pool server to use.

Logical load balancer is supported only in an SR on the Tier-1 gateway.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 15

Page 16: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Detailed Design 3The NSX-T detailed design considers both physical and virtual infrastructure design. It includes numbereddesign decisions and the justification and implications of each decision.

This chapter includes the following topics:

n Physical Infrastructure Design

n Virtual Infrastructure Design

Physical Infrastructure DesignThe physical infrastructure design includes design decision details for the physical network.

Figure 3‑1. Physical Infrastructure Design

ServiceManagement

Portfolio Management

OperationsManagement

CloudManagement

Layer

Service Catalog

Self-Service Portal

Orchestration

BusinessContinuity

Fault Tolerance and Disaster

Recovery

Backup & Restore

Hypervisor

Pools of Resources

Virtualization Control

VirtualInfrastructure

Layer

Compute

Storage

Network

PhysicalLayer

Security

Replication Compliance

Risk

Governance

Physical Networking DesignDesign of the physical SDDC network includes defining the network topology for connecting the physicalswitches and the ESXi hosts, determining switch port settings for VLANs and link aggregation, anddesigning routing. You can use the VMware Validated Design guidance for design and deployment withmost enterprise-grade physical network architectures.

VMware, Inc. 16

Page 17: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Switch Types and Network ConnectivityFollow the best practices for physical switches, switch connectivity, VLANs and subnets, and access portsettings.

Top of Rack Physical Switches

When configuring top of rack (ToR) switches, consider the following best practices:

n Configure redundant physical switches to enhance availability.

n Configure switch ports that connect to ESXi hosts manually as trunk ports. Virtual switches arepassive devices and do not support trunking protocols, such as Dynamic Trunking Protocol (DTP).

n Modify the Spanning Tree Protocol (STP) on any port that is connected to an ESXi NIC to reduce thetime to transition ports over to the forwarding state, for example using the Trunk PortFast featurefound in a Cisco physical switch.

n Provide DHCP or DHCP Helper capabilities on all VLANs used by TEP VMkernel ports. This setupsimplifies the configuration by using DHCP to assign IP address based on the IP subnet in use.

n Configure jumbo frames on all switch ports, inter-switch link (ISL), and switched virtual interfaces(SVIs).

Top of Rack Connectivity and Network Settings

Each ESXi host is connected redundantly to the ToR switches SDDC network fabric by two 25 GbE ports.Configure the ToR switches to provide all necessary VLANs using an 802.1Q trunk. These redundantconnections use features in vSphere Distributed Switch and NSX-T to guarantee that no physicalinterface is overrun and available redundant paths are used.

Figure 3‑2. Host to ToR Connectivity

ToR ToR

ESXiHost

25 GbE 25 GbE

VLANs and Subnets

Each ESXi host uses VLANs and corresponding subnets.

Follow these guidelines:

n Use only /24 subnets to reduce confusion and mistakes when handling IPv4 subnetting.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 17

Page 18: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

n Use the IP address .254 as the (floating) interface with .252 and .253 for Virtual Router RedundancyProtocol (VRPP) or Hot Standby Routing Protocol (HSRP).

n Use the RFC1918 IPv4 address space for these subnets and allocate one octet by region andanother octet by function.

Access Port Network Settings

Configure additional network settings on the access ports that connect the ToR switches to thecorresponding servers.

Spanning Tree Protocol(STP)

Although this design does not use the Spanning Tree Protocol, switchesusually include STP configured by default. Designate the access ports astrunk PortFast.

Trunking Configure the VLANs as members of a 802.1Q trunk with the managementVLAN acting as the native VLAN.

MTU Set MTU for all VLANs and SVIs (Management, vMotion, VXLAN, andStorage) to jumbo frames for consistency purposes.

DHCP Helper Configure a DHCP helper (sometimes called a DHCP relay) on all TEPVLANs.

Physical Network Design DecisionsThe physical network design decisions determine the physical layout and use of VLANs. They alsoinclude decisions on jumbo frames and on other network-related requirements such as DNS and NTP.

Physical Network Design Decisions

Routing protocols NSX-T supports only the BGP routing protocol.

DHCP Helper Set the DHCP helper (relay) to point to a DHCP server by IPv4 address.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 18

Page 19: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 3‑1. Physical Network Design Decisions

Decision ID Design Decision Design Justification Design Implication

NSXT-PHY-NET-001

Implement the following physicalnetwork architecture:n One 25 GbE (10 GbE minimum)

port on each ToR switch forESXi host uplinks.

n No EtherChannel(LAG/LACP/vPC) configurationfor ESXi host uplinks

n Layer 3 device that supportsBGP.

n Guarantees availability during aswitch failure.

n Uses BGP as the only dynamicrouting protocol that is supported byNSX-T.

n Might limit the hardwarechoice.

n Requires dynamicrouting protocolconfiguration in thephysical network.

NSXT-PHY-NET-002

Use a physical network that isconfigured for BGP routingadjacency.

n Supports flexibility in network designfor routing multi-site and multi-tenancy workloads.

n Uses BGP as the only dynamicrouting protocol that is supported byNSX-T.

Requires BGP configurationin the physical network.

NSXT-PHY-NET-003

Use two ToR switches for each rack. Supports the use of two 10 GbE (25GbE recommended) links to each serverand provides redundancy and reducesthe overall design complexity.

Requires two ToR switchesper rack which can increasecosts.

NSXT-PHY-NET-004

Use VLANs to segment physicalnetwork functions.

n Supports physical networkconnectivity without requiring manyNICs.

n Isolates the different networkfunctions of the SDDC so that youcan have differentiated services andprioritized traffic as needed.

Requires uniformconfiguration andpresentation on all the trunksmade available to the ESXihosts.

Additional Design Decisions

Additional design decisions deal with static IP addresses, DNS records, and the required NTP timesource.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 19

Page 20: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 3‑2. IP Assignment, DNS, and NTP Design Decisions

Decision ID Design Decision Design JustificationDesignImplication

NSXT-PHY-NET-005

Assign static IP addresses to allmanagement components in theSDDC infrastructure except forNSX-T TEPs.

NSX-T TEPs are assigned byusing a DHCP server. Set thelease duration for the TEPDHCP scope to at least 7 days.

Ensures that interfaces such as management andstorage always have the same IP address. In thisway, you provide support for continuous managementof ESXi hosts using vCenter Server and forprovisioning IP storage by storage administrators.

NSX-T TEPs do not have an administrative endpoint.As a result, they can use DHCP for automatic IPaddress assignment. You are also unable to assigndirectly a static IP address to the VMkernel port of anNSX-T TEP. IP pools are an option but the NSX-Tadministrator must create them. If you must change orexpand the subnet, changing the DHCP scope issimpler than creating an IP pool and assigning it tothe ESXi hosts.

Requires accurateIP addressmanagement.

NSXT-PHY-NET-006

Create DNS records for allmanagement nodes to enableforward, reverse, short, andFQDN resolution.

Ensures consistent resolution of management nodesusing both IP address (reverse lookup) and nameresolution.

None.

NSXT-PHY-NET-007

Use an NTP time source for allmanagement nodes.

Maintains accurate and synchronized time betweenmanagement nodes.

None.

Jumbo Frames Design Decisions

IP storage throughput can benefit from the configuration of jumbo frames. Increasing the per-framepayload from 1500 bytes to the jumbo frame setting improves the efficiency of data transfer. You mustconfigure jumbo frames end-to-end. Select an MTU that matches the MTU of the physical switch ports.

According to the purpose of the workload, determine whether to configure jumbo frames on a virtualmachine. If the workload consistently transfers large amounts of network data, configure jumbo frames, ifpossible. In that case, confirm that both the virtual machine operating system and the virtual machineNICs support jumbo frames.

Using jumbo frames also improves the performance of vSphere vMotion.

Note The Geneve overlay requires an MTU value of 1600 bytes or greater.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 20

Page 21: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 3‑3. Jumbo Frames Design Decisions

Decision ID Design Decision Design Justification Design Implication

NSXT-PHY-NET-008

Configure the MTU size to at least 9000bytes (jumbo frames) on the physical switchports, vSphere Distributed Switches,vSphere Distributed Switch port groups, andN-VDS switches that support the followingtraffic types.n Geneve (overlay)n vSANn vMotionn NFSn vSphere Replication

Improves trafficthroughput.

To support Geneve,increase the MTU settingto a minimum of 1600bytes.

When adjusting the MTU packetsize, you must also configure theentire network path (VMkernelports, virtual switches, physicalswitches, and routers) to supportthe same MTU packet size.

Virtual Infrastructure DesignThe virtual infrastructure design includes the NSX-T components that make up the virtual infrastructurelayer.

Figure 3‑3. Virtual Infrastructure Layer in the SDDC

ServiceManagement

Portfolio Management

OperationsManagement

CloudManagement

Layer

Service Catalog

Self-Service Portal

Orchestration

BusinessContinuity

Fault Tolerance and Disaster

Recovery

Backup & Restore

Hypervisor

Pools of Resources

Virtualization Control

VirtualInfrastructure

Layer

Compute

Storage

Network

PhysicalLayer

Security

Replication Compliance

Risk

Governance

vSphere Cluster DesignThe cluster design must consider the workload that the cluster handles. Different cluster types in thisdesign have different characteristics.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 21

Page 22: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

vSphere Cluster Design Decision BackgroundWhen you design the cluster layout in vSphere, consider the following guidelines:

n Use fewer, larger ESXi hosts, or more, smaller ESXi hosts.

n A scale-up cluster has fewer, larger ESXi hosts.

n A scale-out cluster has more, smaller ESXi hosts.

n Compare the capital costs of purchasing fewer, larger ESXi hosts with the costs of purchasing more,smaller ESXi hosts. Costs vary between vendors and models.

n Evaluate the operational costs of managing a few ESXi hosts with the costs of managing more ESXihosts.

n Consider the purpose of the cluster.

n Consider the total number of ESXi hosts and cluster limits.

vSphere High Availability DesignVMware vSphere High Availability (vSphere HA) protects your virtual machines in case of ESXi hostfailure by restarting virtual machines on other hosts in the cluster when an ESXi host fails.

vSphere HA Design Basics

During configuration of the cluster, the ESXi hosts elect a master ESXi host. The master ESXi hostcommunicates with the vCenter Server system and monitors the virtual machines and secondary ESXihosts in the cluster.

The master ESXi host detects different types of failure:

n ESXi host failure, for example an unexpected power failure

n ESXi host network isolation or connectivity failure

n Loss of storage connectivity

n Problems with virtual machine OS availability

Table 3‑4. vSphere HA Design Decisions

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-VC-001 Use vSphere HA toprotect all clustersagainst failures.

vSphere HA supports a robust levelof protection for both ESXi host andvirtual machine availability.

You must provide sufficient resources onthe remaining hosts so that virtualmachines can be migrated to thosehosts in the event of a host outage.

NSXT-VI-VC-002 Set vSphere HA HostIsolation Response toPower Off.

vSAN requires that the HA IsolationResponse be set to Power Off and torestart VMs on available ESXi hosts.

VMs are powered off in case of a falsepositive and an ESXi host is declaredisolated incorrectly.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 22

Page 23: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

vSphere HA Admission Control Policy Configuration

The vSphere HA Admission Control Policy allows an administrator to configure how the clusterdetermines available resources. In a smaller vSphere HA cluster, a larger proportion of the clusterresources are reserved to accommodate ESXi host failures, based on the selected policy.

The following policies are available:

Host failures the clustertolerates

vSphere HA ensures that a specified number of ESXi hosts can fail andsufficient resources remain in the cluster to fail over all the virtual machinesfrom those ESXi hosts.

Percentage of clusterresources reserved

vSphere HA reserves a specified percentage of aggregate CPU andmemory resources for failover.

Specify Failover Hosts When an ESXi host fails, vSphere HA attempts to restart its virtualmachines on any of the specified failover ESXi hosts. If restart is notpossible, for example, the failover ESXi hosts have insufficient resources orhave failed as well, then vSphere HA attempts to restart the virtualmachines on other ESXi hosts in the cluster.

Shared Edge and Compute Cluster DesignTenant workloads run on the ESXi hosts in the shared edge and compute cluster. Because of the sharednature of the cluster, NSX-T Edge appliances also run in this cluster. To support these workloads, youmust determine the number of ESXi hosts and vSphere HA settings and several other characteristics ofthe shared edge and compute cluster.

Table 3‑5. Shared Edge and Compute Cluster Design Decisions

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-VC-003

Create a shared edge andcompute cluster thatcontains tenant workloadsand NSX-T Edgeappliances.

Limits the footprint of the design bysaving the use of a vSphere clusterspecifically for the NSX-T Edge nodes.

In a shared cluster, the VLANs andsubnets between the VMkernel portsfor ESXi host overlay and the overlayports of the edge appliances must beseparate.

NSXT-VI-VC-004

Configure admission controlfor a failure of one ESXihost and percentage-basedfailover capacity.

vSphere HA protects the tenantworkloads and NSX-T Edgeappliances in the event of an ESXihost failure. vSphere HA powers onthe virtual machines from the non-responding ESXi hosts on theremaining ESXi hosts.

Only a single ESXi host failure istolerated before a resource contentionoccurs.

NSXT-VI-VC-005

Create a shared edge andcompute cluster thatconsists of a minimum offour ESXi hosts.

Allocating four ESXi hosts provides afull redundancy within the cluster.

Four ESXi hosts is the smallest startingpoint for the shared edge and computecluster for redundancy andperformance as a result increasingcost.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 23

Page 24: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 3‑5. Shared Edge and Compute Cluster Design Decisions (Continued)

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-VC-006

Create a resource pool forthe two large sized edgevirtual machines with a CPUshare level of High, amemory share of normal,and a 32-GB memoryreservation.

The NSX-T Edge appliances control allnetwork traffic in and out of the SDDC.In a contention situation, theseappliances must receive all theresources required.

During contention, the NSX-Tcomponents receive more resourcesthan the other workloads. As a result,monitoring and capacity managementmust be a proactive activity.

The resource pool memory reservationmust be expanded if you plan to deploymore NSX-T Edge appliances.

NSXT-VI-VC-007

Create a resource pool forall tenant workloads with aCPU share value of Normaland a memory share valueof Normal.

Running virtual machines at the clusterlevel has a negative impact on all othervirtual machines during contention. Toavoid an impact on networkconnectivity, in a shared edge andcompute cluster, the NSX-T Edgeappliances must receive resourceswith priority to the other workloads.Setting the share values to Normalincreases the resource shares of theNSX-T Edge appliances in the cluster.

During contention, tenant workloadsmight have insufficient resources andhave poor performance. Proactivelyperform monitoring and capacitymanagement, add capacity or dedicatean edge cluster before contentionoccurs.

NSXT-VI-VC-008

Create a host profile for theshared edge and computecluster.

Using host profiles simplifies theconfiguration of ESXi hosts andensures that settings are uniformacross the cluster.

The host profile is only useful for initialcluster deployment and configuration.

After you add the ESXi hosts astransport nodes to the NSX-Tdeployment, the host profile is nolonger usable and you must removethem from the cluster.

Compute Cluster DesignAs the SDDC expands, you can add compute-only clusters. Tenant workloads run on the ESXi hosts inthe compute cluster instances. One Compute vCenter Server instance manages multiple computeclusters. The design determines vSphere HA settings for the compute cluster.

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-VC-009

Configure vSphere HA to usepercentage-based failovercapacity to ensure n+1availability.

Using explicit host failoverlimits the total availableresources in a cluster.

The resources of one ESXi host in thecluster is reserved which can causeprovisioning to fail if resources areexhausted.

Virtualization Network DesignDesign the virtualization network according to the business goals of your organization. Prevent alsounauthorized access, and provide timely access to business data.

This network virtualization design uses vSphere and NSX-T to implement virtual networking.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 24

Page 25: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Virtual Network Design GuidelinesThis VMware Validated Design follows high-level network design guidelines and networking bestpractices.

Design Goals

You can apply the following high-level design goals to your environment:

n Meet diverse needs. The network must meet the diverse needs of many different entities in anorganization. These entities include applications, services, storage, administrators, and users.

n Reduce costs. Server consolidation alone reduces network costs by reducing the number of requirednetwork ports and NICs, but you should determine a more efficient network design. For example,configuring two 25 GbE NICs with VLANs might be more cost effective than configuring a dozen 1-GbE NICs on separate physical networks.

n Boost performance. You can achieve performance improvements and decrease the time required toperform maintenance by providing sufficient bandwidth, which reduces contention and latency.

n Improve availability. You usually improve availability by providing network redundancy.

n Support security. You can support an acceptable level of security through controlled access whererequired and isolation where necessary.

n Improve infrastructure functionality. You can configure the network to support vSphere features suchas vSphere vMotion, vSphere High Availability, and vSphere Fault Tolerance.

Best Practices

Follow the networking best practices throughout your environment.

n Separate network services from one another for greater security and better performance.

n Use Network I/O Control and traffic shaping to guarantee bandwidth to critical virtual machines.During network contention, these critical virtual machines receive a higher percentage of thebandwidth.

n Separate network services on an NSX-T Virtual Distributed Switch (N-VDS) by attaching them tosegments with different VLAN IDs.

n Keep vSphere vMotion traffic on a separate network. When migration with vMotion occurs, thecontents of the memory of the guest operating system is transmitted over the network. You can placevSphere vMotion on a separate network by using a dedicated vSphere vMotion VLAN.

n When using pass-through devices with Linux kernel version 2.6.20 or an earlier guest OS, avoid MSIand MSI-X modes. These modes have significant performance impact.

n For best performance, use VMXNET3 virtual machine NICs.

n Ensure that physical network adapters connected to the same virtual switch are also connected to thesame physical network.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 25

Page 26: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Network Segmentation and VLANs

You separate different types of traffic for access security and to reduce contention and latency.

High latency on a network can impact performance. Some components are more sensitive to high latencythan others. For example, reducing latency is important on the IP storage and the vSphere FaultTolerance logging network, because latency on these networks can negatively affect the performance ofmultiple virtual machines.

According to the application or service, high latency on specific virtual machine networks can alsonegatively affect performance. Use information gathered from the current state analysis and frominterviews with key stakeholder and SMEs to determine which workloads and networks are especiallysensitive to high latency.

Virtual Networks

Determine the number of networks or VLANs that are required according to the type of traffic.

n vSphere operational traffic.

n Management

n Geneve (overlay)

n vMotion

n vSAN

n NFS Storage

n vSphere Replication

n Traffic that supports the services and applications of the organization.

Virtual SwitchesVirtual switches simplify the configuration process by providing single pane of glass view for performingvirtual network management tasks.

Virtual Switch Design Background

vSphere Distributed Switch and NSX-T Virtual Distributed Switch (N-VDS) provide several advantagesover vSphere Standard Switch.

Centralizedmanagement

n A distributed switch is created and centrally managed on a vCenterServer system. The switch configuration is consistent across ESXihosts.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 26

Page 27: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

n An N-VDS is created and centrally managed in NSX-T Manager. Theswitch configuration is consistent across ESXi and edge transportnodes.

Centralized management saves time and reduces mistakes and operationalcosts.

Additional features Some of the features of distributed switches can be useful to theapplications and services running in the organization’s infrastructure. Forexample, NetFlow and port mirroring provide monitoring andtroubleshooting capabilities to the virtual infrastructure.

Consider the following caveats for distributed switches:

n Distributed switches are manageable only when the vCenter Server instance is available. As a result,vCenter Server becomes a Tier-1 application.

n N-VDS instances are manageable only when the NSX-T Manager cluster is available. As a result, theNSX-T Manager cluster becomes a Tier-1 application.

Virtual Switch Design Decisions

The virtual switch design decisions determine the use and placement of specific switch types.

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-NET-001

Use N-VDS for the NSX-T based sharededge and compute cluster, and for additionalNSX-T based compute clusters.

You need N-VDS foroverlay traffic.

N-VDS is not compatible withvSphere host profiles.

Shared Edge and Compute Cluster Switches

The shared edge and compute cluster uses a single N-VDS with a certain configuration for handled traffictypes, NIC teaming, and MTU size.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 27

Page 28: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Figure 3‑4. Virtual Switch Design for ESXi Hosts in the Shared Edge and Compute Cluster

nic0 nic1

VLAN NFS

VLAN vMotion

VLAN vSAN

Sample ESXi Shared Edge and Compute Host

sfo01-w-nvds01

VLAN ESXi Management

VLAN Trunking Uplink01

VLAN Trunking Uplink02

VLAN Trunking Edge Overlay (TEP)

Table 3‑6. Virtual Switches for the Shared Edge and Compute Cluster

N-VDS Switch Name FunctionNumber ofPhysical NIC Ports Teaming Policy MTU

sfo01-w-nvds01 n ESXi Managementn vSphere vMotionn vSANn NFSn Geneve Overlay (TEP)n Uplink trunking (2) for the NSX-T

Edge instances

2 n Load balance source forESXi traffic

n Failover order for edgeVM traffic

9000

sfo01-w-uplink01 n Uplink to enable ECMP 1 Failover order 9000

sfo01-w-uplink02 n Uplink to enable ECMP 1 Failover order 9000

Table 3‑7. Virtual Switches in the Shared Edge and Compute Cluster by Physical NIC

N-VDS Switch vmnic Function

sfo01-w-nvds01 0 Uplink

sfo01-w-nvds01 1 Uplink

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 28

Page 29: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Figure 3‑5. Segment Configuration on an ESXi Host That Runs an NSX-T Edge Node

NSX-T Edge Virtual Machine

fp-eth0 (overlay)

sfo01-w-uplink01(VLAN)

fp-eth1 (uplink 1)

fp-eth2 (uplink 2)

eth0 (management)

ESXi Host

sfo01-w-nvds01

vmnic0

vmnic1

sfo01-w-nvds01(Overlay)

sfo01-w-uplink02(VLAN)

sfo01-w-nvds01-uplink01(VLAN Trunking)

sfo01-w-overlay(VLAN Trunking)

sfo01-w-nvds01-uplink02(VLAN Trunking)

sfo01-w-nvds01-management(VLAN 1641)

Table 3‑8. Segments in the Shared Edge and Compute Cluster

N-VDS Switch Segment Name

sfo01-w-nvds01 sfo01-w-nvds01-management

sfo01-w-nvds01 sfo01-w-nvds01-vmotion

sfo01-w-nvds01 sfo01-w-nvds01-vsan

sfo01-w-nvds01 sfo01-w-nvds01-overlay

sfo01-w-nvds01 sfo01-w-nvds01-uplink01

sfo01-w-nvds01 sfo01-w-nvds01-uplink02

sfo01-w-nvds01 sfo01-w-nvds01-nfs

sfo01-w02-uplink01 sfo01-w02-uplink01

sfo01-w02-uplink02 sfo01-w02-uplink02

Table 3‑9. VMkernel Adapters for the Shared Edge and Compute Cluster

N-VDS Switch Segment Name Enabled Services

sfo01-w-nvds01 sfo01-w-nvds01-management Management Traffic

sfo01-w-nvds01 sfo01-w-nvds01-vmotion vMotion Traffic

sfo01-w-nvds01 sfo01-w-nvds01-vsan vSAN

sfo01-w-nvds01 sfo01-w-nvds01-nfs --

Note

When the NSX-T Edge appliance is on an N-VDS, it must use a different VLAN ID and subnet from theESXi hosts overlay (TEP) VLAN ID and subnet.

ESXi host TEP VMkernel ports are automatically created when you configure an ESXi host as a transportnode.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 29

Page 30: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

NIC TeamingYou can use NIC teaming to increase the network bandwidth available in a network path, and to providethe redundancy that supports higher availability.

Benefits and Overview

NIC teaming helps avoid a single point of failure and provides options for load balancing of traffic. Toreduce further the risk of a single point of failure, build NIC teams by using ports from multiple NIC andmotherboard interfaces.

Create a single virtual switch with teamed NICs across separate physical switches.

NIC Teaming Design Background

For a predictable level of performance, use multiple network adapters in one of the followingconfigurations.

n An active-passive configuration that uses explicit failover when connected to two separate switches.

n An active-active configuration in which two or more physical NICs in the server are assigned theactive role.

This validated design uses a non-LAG active-active configuration using the route based on physical NICload algorithm for vSphere Distributed Switch and load balance source algorithm for N-VDS. By using thisconfiguration, network cards remain active instead of remaining idle until a failure occurs.

Table 3‑10. NIC Teaming and Policy

Design Quality Active-Active Active-Passive Comments

Availability ↑ ↑ Using teaming regardless of the option increases the availability of theenvironment.

Manageability o o Neither design option impacts manageability.

Performance ↑ o An active-active configuration can send traffic across either NIC, therebyincreasing the available bandwidth. This configuration provides a benefit ifthe NICs are being shared among traffic types and Network I/O Control isused.

Recoverability o o Neither design option impacts recoverability.

Security o o Neither design option impacts security.

Legend: ↑ = positive impact on quality; ↓ = negative impact on quality; o = no impact on quality.

Table 3‑11. NIC Teaming Design Decisions

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-NET-002 In the shared edge and computecluster, use the Load balancesource teaming policy on N-VDS.

NSX-T Virtual Distributed Switch(N-VDS)supports Load balance source and Failoverteaming policies. When you use the Loadbalance source policy, both physical NICs can beactive and carry traffic.

None.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 30

Page 31: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Geneve OverlayGeneve provides the overlay capability in NSX-T to create isolated, multi-tenant broadcast domainsacross data center fabrics, and enables customers to create elastic, logical networks that span physicalnetwork boundaries.

The first step in creating these logical networks is to isolate and pool the networking resources. By usingthe Geneve overlay, NSX-T isolates the network into a pool of capacity and separates the consumption ofthese services from the underlying physical infrastructure. This model is similar to the model vSphereuses to abstract compute capacity from the server hardware to create virtual pools of resources that canbe consumed as a service. You can then organize the pool of network capacity in logical networks thatare directly attached to specific applications.

Geneve is a tunneling mechanism which provides extensibility while still using the offload capabilities ofNICs for performance improvement.

Geneve works by creating Layer 2 logical networks that are encapsulated in UDP packets. A Segment IDin every frame identifies the Geneve logical networks without the need for VLAN tags. As a result, manyisolated Layer 2 networks can coexist on a common Layer 3 infrastructure using the same VLAN ID.

In the vSphere architecture, the encapsulation is performed between the virtual NIC of the guest VM andthe logical port on the virtual switch, making the Geneve overlay transparent to both the guest virtualmachines and the underlying Layer 3 network. The Tier-0 Gateway performs gateway services betweenoverlay and non-overlay hosts, for example, a physical server or the Internet router. The NSX-T Edgevirtual machine translates overlay segment IDs to VLAN IDs, so that non-overlay hosts can communicatewith virtual machines on an overlay network.

The edge cluster hosts all NSX-T Edge virtual machine instances that connect to the corporate networkfor secure and centralized network administration.

Table 3‑12. Geneve Overlay Design Decisions

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-NET-003

Use NSX-T to introduceoverlay networks forworkloads.

Simplifies the networkconfiguration by using centralizedvirtual network management.

n Requires additional compute andstorage resources to deploy NSX-T components.

n Might require more training inNSX-T.

NSXT-VI-NET-004

To provide virtualized networkcapabilities to workloads, useoverlay networks with NSX-TEdge virtual machines anddistributed routing.

Creates isolated, multi-tenantbroadcast domains across datacenter fabrics to deploy elastic,logical networks that spanphysical network boundaries.

Requires configuring transportnetworks with an MTU size of at least1600 bytes.

vMotion TCP/IP StackUse the vMotion TCP/IP stack to isolate traffic for vSphere vMotion and to assign a dedicated defaultgateway for vSphere vMotion traffic.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 31

Page 32: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

By using a separate TCP/IP stack, you can manage vSphere vMotion and cold migration traffic accordingto the topology of the network, and as required for your organization.

n Route the traffic for the migration of virtual machines by using a default gateway that is different fromthe gateway assigned to the default stack on the ESXi host.

n Assign a separate set of buffers and sockets.

n Avoid routing table conflicts that might otherwise appear when many features are using a commonTCP/IP stack.

n Isolate traffic to improve security.

Table 3‑13. vMotion TCP/IP Stack Design Decision

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-NET-005

Use the vMotionTCP/IP stack forvSphere vMotiontraffic.

By using the vMotion TCP/IP stack,vSphere vMotion traffic can beassigned a default gateway on its ownsubnet and can go over Layer 3networks.

The vMotion TCP/IP stack is notavailable in the VMkernel adaptercreation wizard of vSphere DistributedSwitch. You must create the VMkerneladapter directly on the ESXi host.

NSX DesignThis design implements software-defined networking by using VMware NSX-T. By using NSX-T,virtualization delivers for networking what it has already delivered for compute and storage.

In much the same way that server virtualization programmatically creates, takes snapshots of, deletes,and restores software-based virtual machines (VMs), NSX network virtualization programmaticallycreates, takes snapshots of, deletes, and restores software-based virtual networks. As a result, you followa simplified operational model for the underlying physical network.

NSX-T is a nondisruptive solution. You can deploy it on any IP network, including existing traditionalnetworking models and next-generation fabric architectures, regardless of the vendor.

When administrators provision workloads, network management is a time-consuming task. You spendmost time configuring individual components in the physical infrastructure and verifying that networkchanges do not affect other devices that are using the same physical network infrastructure.

The need to pre-provision and configure networks is a constraint to cloud deployments where speed,agility, and flexibility are critical requirements. Pre-provisioned physical networks enable fast creation ofvirtual networks and faster deployment times of workloads using the virtual network. If the physicalnetwork that you need is already available on the ESXi host to run a workload, pre-provisioning physicalnetworks works well. However, if the network is not available on an ESXi host, you must find an ESXi hostwith the available network and allocate capacity to run workloads in your environment.

Decouple virtual networks from their physical counterparts. In the virtualized environment, you mustrecreate all physical networking attributes that are required by the workloads. Because networkvirtualization supports the creation of virtual networks without modification of the physical networkinfrastructure, you can provision the workload networks faster.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 32

Page 33: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

NSX-T DesignNSX-T components are not dedicated to a specific vCenter Server instance or vSphere construct. Youcan share them across vSphere environments in the same physical location.

Although an NSX-T deployment is not associated with a vCenter Server instance, it supports only single-region deployments.

Table 3‑14. NSX-T Design Decisions

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-001

Deploy a three node NSX-T Managercluster to configure and manage theshared edge and compute cluster and alladditional compute clusters in a singleregion.

Provides support for software-definednetworking (SDN) capabilities, suchas load balancing, firewalls, andlogical switching.

You must install andconfigure NSX-TManager in a highlyavailable cluster.

NSXT-VI-SDN-002

In the management cluster, add theNSX-T Manager nodes to the distributedfirewall exclusion list of NSX for vSphere.

Ensures that the management andcontrol plane is still available if amistake in the configuration of thedistributed firewall of NSX for vSphereoccurs.

None.

NSX-T ComponentsThe following sections describe the components in the solution and how they are relevant to the networkvirtualization design.

NSX-T Manager

NSX-T Manager provides the graphical user interface (GUI) and the RESTful API for creating,configuring, and monitoring NSX-T components, such as segments and gateways.

NSX-T Manager implements the management and control plane for the NSX-T infrastructure. NSX-TManager provides an aggregated system view and is the centralized network management component ofNSX-T. It provides a method for monitoring and troubleshooting workloads attached to virtual networks. Itprovides configuration and orchestration of the following services:

n Logical networking components, such as logical switching and routing

n Networking and edge services

n Security services and distributed firewall

NSX-T Manager also provides a RESTful API endpoint to automate consumption. Because of thisarchitecture, you can automate all configuration and monitoring operations using any cloud managementplatform, security vendor platform, or automation framework.

The NSX-T Management Plane Agent (MPA) is an NSX-T Manager component that is available on eachESXi host. The MPA is in charge of persisting the desired state of the system and for communicating non-flow-controlling (NFC) messages such as configuration, statistics, status, and real-time data betweentransport nodes and the management plane.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 33

Page 34: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

NSX-T Manager also contains the NSX-T Controller component. NSX-T Controllers control the virtualnetworks and overlay transport tunnels. The controllers are responsible for the programmatic deploymentof virtual networks across the entire NSX-T architecture.

The Central Control Plane (CCP) is logically separated from all data plane traffic, that is, a failure in thecontrol plane does not affect existing data plane operations. The controller provides configuration to otherNSX-T Controller components such as the segments, gateways, and edge virtual machine configuration.

Table 3‑15. NSX-T Manager Design Decisions

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-003 Deploy a three node NSX-TManager cluster using thelarge-size appliance.

The large-size appliancesupports more than 64 ESXihosts. The small-sizeappliance is for proof ofconcept and the medium sizeonly supports up to 64 ESXihosts.

The large size requires moreresources in the managementcluster.

NSXT-VI-SDN-004 Create a virtual IP (VIP) forthe NSX-T Manager cluster.

Provides HA for the NSX-TManager UI and API.

The VIP provides HA only, itdoes not load balancerequests across the managercluster.

NSXT-VI-SDN-005 n Grant administratorsaccess to both the NSX-TManager UI and itsRESTful API endpoint.

n Restrict end-user accessto the RESTful APIendpoint configured forend-user provisioning,such as vRealizeAutomation or PivotalContainer Service (PKS).

Ensures that tenants or non-provider staff cannot modifyinfrastructure components.

End-users typically interactonly indirectly with NSX-Tfrom their provisioning portal.Administrators interact withNSX-T using its UI and API.

End users have access onlyto end-point components.

NSX-T Virtual Distributed Switch

An NSX-T Virtual Distributed Switch (N-VDS) runs on ESXi hosts and provides physical traffic forwarding.It transparently provides the underlying forwarding service that each segment relies on. To implementnetwork virtualization, a network controller must configure the ESXi host virtual switch with network flowtables that form the logical broadcast domains the tenant administrators define when they create andconfigure segments.

NSX-T implements each logical broadcast domain by tunneling VM-to-VM traffic and VM-to-gatewaytraffic using the Geneve tunnel encapsulation mechanism. The network controller has a global view of thedata center and ensures that the ESXi host virtual switch flow tables are updated as VMs are created,moved, or removed.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 34

Page 35: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 3‑16. NSX-T N-VDS Design Decision

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-006 Deploy an N-VDS instance to eachESXi host in the shared edge andcompute cluster.

ESXi hosts in the shared edge andcompute cluster provide tunnel endpointsfor Geneve overlay encapsulation.

None.

Logical Switching

NSX-T Segments create logically abstracted segments to which you can connect tenant workloads. Asingle Segment is mapped to a unique Geneve segment that is distributed across the ESXi hosts in atransport zone. The Segment supports line-rate switching in the ESXi host without the constraints ofVLAN sprawl or spanning tree issues.

Table 3‑17. NSX-T Logical Switching Design Decision

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-007 Deploy all workloads onNSX-T Segments (logicalswitches).

To take advantage of features such asdistributed routing, tenant workloadsmust be connected to NSX-TSegments.

You must perform all networkmonitoring in the NSX-T ManagerUI or vRealize Network Insight.

Gateways (Logical Routers)

NSX-T Gateways provide North-South connectivity so that workloads can access external networks, andEast-West connectivity between different logical networks.

A Logical Router is a configured partition of a traditional network hardware router. It replicates thefunctionality of the hardware, creating multiple routing domains in a single router. Logical routers performa subset of the tasks that are handled by the physical router, and each can contain multiple routinginstances and routing tables. Using logical routers can be an effective way to maximize router use,because a set of logical routers within a single physical router can perform the operations previouslyperformed by several pieces of equipment.

n Distributed router (DR)

A DR spans ESXi hosts whose virtual machines are connected to this Gateway, and edge nodes theGateway is bound to. Functionally, the DR is responsible for one-hop distributed routing betweensegments and Gateways connected to this Gateway.

n One or more (optional) service routers (SR).

An SR is responsible for delivering services that are not currently implemented in a distributedfashion, such as stateful NAT.

A Gateway always has a DR. A Gateway has SRs when it is a Tier-0 Gateway, or when it is a Tier-1Gateway and has routing services configured such as NAT or DHCP.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 35

Page 36: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Tunnel Endpoint

Tunnel endpoints enable ESXi hosts to participate in an NSX-T overlay. The NSX-T overlay deploys aLayer 2 network on top of an existing Layer 3 network fabric by encapsulating frames inside packets andtransferring the packets over an underlying transport network. The underlying transport network can beanother Layer 2 networks or it can cross Layer 3 boundaries. The Tunnel Endpoint (TEP) is theconnection point at which the encapsulation and decapsulation take place.

NSX-T Edges

NSX-T Edges provide routing services and connectivity to networks that are external to the NSX-Tdeployment. You use an NSX-T Edge for establishing external connectivity from the NSX-T domain byusing a Tier-0 Gateway using BGP or static routing. Additionally, you deploy an NSX-T Edge to supportnetwork address translation (NAT) services at either the Tier-0 or Tier-1 Gateway.

The NSX-T Edge connects isolated, stub networks to shared uplink networks by providing commongateway services such as NAT, and dynamic routing.

Logical Firewall

NSX-T uses handles traffic in and out the network according to firewall rules.

A logical firewall offers multiple sets of configurable Layer 3 and Layer 2 rules. Layer 2 firewall rules areprocessed before Layer 3 rules. You can configure an exclusion list to exclude segments, logical ports, orgroups from firewall enforcement.

The default rule, that is at the bottom of the rule table, is a catchall rule. The logical firewall enforces thedefault rule on packets that do not match other rules. After the host preparation operation, the default ruleis set to the allow action. Change this default rule to a block action and apply access control through apositive control model, that is, only traffic defined in a firewall rule can flow on the network.

Logical Load Balancer

The NSX-T logical load balancer offers high-availability service for applications and distributes thenetwork traffic load among multiple servers.

The load balancer accepts TCP, UDP, HTTP, or HTTPS requests on the virtual IP address anddetermines which pool server to use.

Logical load balancer is supported only on the Tier-1 Gateway.

NSX-T Network Requirements and SizingNSX-T requirements impact both physical and virtual networks.

Physical Network Requirements

Physical requirements determine the MTU size for networks that carry overlay traffic, dynamic routingsupport, time synchronization through an NTP server, and forward and reverse DNS resolution.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 36

Page 37: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Requirement Comments

Provide an MTU size of 1600 or greater on anynetwork that carries Geneve overlay traffic must.

Geneve packets cannot be fragmented. The MTU size must be largeenough to support extra encapsulation overhead.

This design uses an MTU size of 9000 for Geneve traffic. See Table 3‑3.

Enable dynamic routing support on the upstreamLayer 3 devices.

You use BGP on the upstream Layer 3 devices to establish routingadjacency with the Tier-0 SRs.

Provide an NTP server. The NSX-T Manager requires NTP settings that synchronize it with the restof the environment.

Establish forward and reverse DNS resolution forall management VMs.

The NSX-T Controllers do not require DNS entries.

NSX-T Component Specifications

When you size the resources for NSX-T components, consider the compute and storage requirements foreach component, and the number of nodes per component type.

Size of NSX Edge services gateways might be different according to tenant requirements. Consider alloptions in such a case.

Table 3‑18. Resource Specification of the NSX-T Components

Virtual Machine vCPU Memory (GB) Storage (GB)Quantity per NSX-TDeployment

NSX-T Manager 12 (Large) 48 (Large) 200 (Large) 3

NSX-T Edge virtualmachine

2 (Small) 4 (Small) 120 (Small) Numbers are differentaccording to the usecase. At least two edgedevices are required toenable ECMP routing.

4 (Medium) 8 (Medium) 120 (Medium)

8 (Large) 16 (Large) 120 (Large)

Table 3‑19. Design Decisions on Sizing the NSX-T Edge Virtual Machines

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-008 Use large-size NSX Edge virtualmachines.

The large-size appliance provides all theperformance characteristics if a failureoccurs.

None.

Network Virtualization Conceptual Design for NSX-TThis conceptual design for NSX-T provides the network virtualization design of the logical componentsthat handle the data to and from tenant workloads in the environment.

The network virtualization conceptual design includes a perimeter firewall, a provider logical router, andthe NSX-T Gateway. It also considers the external network, internal workload networks, and themanagement network.

Figure 3‑6. NSX-T Conceptual Overview

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 37

Page 38: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

VM VMVM VM VMVM

Internet MPLS

MgmtNetwork

External Network

Perimeter Firewall

Upstream Layer 3Devices

NSX-Т DistributedRouter (DR)

Internal Tenant Networks (Segments)

vNIC-Level Distributed Firewall

NSX-T Service Router(SR)

The conceptual design has the following components.

External Networks Connectivity to and from external networks is through the perimeter firewall.

Perimeter Firewall The firewall exists at the perimeter of the data center to filter Internet traffic.

Upstream Layer 3Devices

The upstream Layer 3 devices are behind the perimeter firewall and handleNorth-South traffic that is entering and leaving the NSX-T environment. Inmost cases, this layer consists of a pair of top of rack switches orredundant upstream Layer 3 devices such as core routers.

NSX-T Gateway (SR) The SR component of the NSX-T Tier-0 Gateway is responsible forestablishing eBGP peering with the Upstream Layer 3 devices and enablingNorth-South routing.

NSX-T Gateway (DR) The DR component of the NSX-T Gateway is responsible for East-Westrouting.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 38

Page 39: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Management Network The management network is a VLAN-backed network that supports allmanagement components such as NSX-T Manager and NSX-T Controllers.

Internal WorkloadNetworks

Internal workload networks are NSX-T Segments and provide connectivityfor the tenant workloads. Workloads are directly connected to thesenetworks. Internal workload networks are then connected to a DR.

Cluster Design for NSX-TThe NSX-T design uses management, and shared edge and compute clusters. You can add morecompute clusters for scale-out, or different workload types or SLAs.

The logical NSX-T design considers the vSphere clusters and defines the place where each NSXcomponent runs.

Figure 3‑7. NSX-T Cluster Design

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 39

Page 40: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Management Cluster

The management cluster contains all components for managing the SDDC. This cluster is a corecomponent of the VMware Validated Design for Software-Defined Data Center. For information about themanagement cluster design, see the Architecture and Design documentation in VMware Validated Designfor Software-Defined Data Center.

NSX-T Edge Node Cluster

The NSX-T Edge cluster is a logical grouping of NSX-T Edge virtual machines. These NSX-T Edge virtualmachines run in the vSphere shared edge and compute cluster and provide North-South routing for theworkloads in the compute clusters.

Shared Edge and Compute Cluster

In the shared edge and compute cluster, ESXi hosts are prepared for NSX-T. As a result, they can beconfigured as transport nodes and can participate in the overlay network. All tenant workloads, and NSX-T Edge virtual machines run in this cluster.

Table 3‑20. Cluster Design Decisions

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-009

For the compute stack, do notdedicate a vSphere edgecluster.

Simplifies configuration andminimizes the number of ESXihosts required for initialdeployment.

The NSX-T Edge virtual machines aredeployed in the shared edge and computecluster.

Because of the shared nature of thecluster, you must scale out the cluster ascompute workloads are added to avoid animpact on network performance.

NSXT-VI-SDN-010

Deploy at least two large-sizeNSX-T Edge virtual machinesin the shared edge andcompute cluster.

Creates NSX-T Edge cluster,and meets availability and scalerequirements.

When additional Edge VM's are added,the Resource Pool Memory Reservationmust be adjusted.

NSXT-VI-SDN-011

Apply vSphere DistributedResource Scheduler (vSphereDRS) VM-Host anti-affinityrules to NSX-T Managers.

Prevents managers fromrunning on the same ESXi hostand thereby risking their highavailability capability.

Requires at least four physical hosts toguarantee the three NSX-T Managerscontinue to run if an ESXi host failureoccurs.

Additional configuration is required to setup anti-affinity rules.

NSXT-VI-SDN-012

Apply vSphere DRS VM-Hostanti-affinity rules to the virtualmachines of the NSX-T Edgecluster.

Prevents the NSX-T Edge virtualmachines from running on thesame ESXi host andcompromising their highavailability.

Additional configuration is required to setup anti-affinity rules.

High Availability of NSX-T Components

The NSX-T Managers run on the management cluster. vSphere HA protects the NSX-T Managers byrestarting the NSX-T Manager virtual machine on a different ESXi host if a primary ESXi host failureoccurs.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 40

Page 41: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

The data plane remains active during outages in the management and control planes although theprovisioning and modification of virtual networks is impaired until those planes become available again.

The NSX-T Edge virtual machines are deployed on the shared edge and compute cluster. vSphere DRSanti-affinity rules prevent NSX-T Edge virtual machines that belong to the same NSX-T Edge cluster fromrunning on the same ESXi host.

NSX-T SRs for North-South routing are configured in equal-cost multi-path (ECMP) mode that supportsroute failover in seconds.

Replication Mode of SegmentsThe control plane decouples NSX-T from the physical network, and handles the broadcast, unknownunicast, and multicast (BUM) traffic in the segments (logical switches).

The following options are available for BUM replication on segments.

Table 3‑21. BUM Replication Mode of NSX-T Segments

BUM Replication Mode Description

Hierarchical Two-Tier In this mode, the ESXi host transport nodes are groupedaccording to their TEP IP subnet. One ESXi host in eachsubnet is responsible for replication to a ESXi host in anothersubnet. The receiving ESXi host replicates the traffic to theESXi hosts in its local subnet.

The source ESXi host transport node knows about the groupsbased on information it has received from the NSX-T controlcluster. The system can select an arbitrary ESXi host transportnode as the mediator for the source subnet if the remotemediator ESXi host node is available.

Head-End In this mode, the ESXi host transport node at the origin of theframe to be flooded on a segment sends a copy to every otherESXi host transport node that is connected to this segment.

Table 3‑22. Design Decisions on Segment Replication Mode

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-013 Use hierarchical two-tierreplication on all segments.

Hierarchical two-tier replication is more efficientby reducing the number of ESXi hosts thesource ESXi host must replicate traffic to.

None.

Transport Zone DesignTransport zones determine which hosts can participate in the use of a particular network. A transportzone identifies the type of traffic, VLAN or overlay, and the N-VDS name. You can configure one or moretransport zones. A transport zone does not represent a security boundary.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 41

Page 42: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 3‑23. Transport Zones Design Decisions

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-014

Create a single transportzone for all overlay traffic.

Ensures all Segments are available toall ESXi hosts and edge virtualmachines configured as TransportNodes.

None.

NSXT-VI-SDN-015

Create a VLAN transportzone for ESXi hostVMkernel ports.

Enables the migration of ESXi hostVMkernel ports to the N-VDS.

The N-VDS name must match theN-VDS name in the overlaytransport zone.

NSXT-VI-SDN-016

Create two transport zonesfor edge virtual machineuplinks.

Enables the edge virtual machines touse equal-cost multi-path routing(ECMP).

You must specify a VLAN range,that is use VLAN trunking, on thesegment used as the uplinks.

Network I/O Control Design for NSX-TWhen a Network I/O Control profile is attached to an N-VDS, during contention the switch allocatesavailable bandwidth according to the configured shares, limit, and reservation for each vSphere traffictype.

How Network I/O Control Works

Network I/O Control enforces the share value specified for the different traffic types only when there isnetwork contention. When contention occurs, Network I/O Control applies the share values set to eachtraffic type. As a result, less important traffic, as defined by the share percentage, is throttled, grantingaccess to more network resources to more important traffic types.

Network I/O Control also supports the reservation of bandwidth for system traffic according to the overallpercentage of available bandwidth.

Network I/O Control Heuristics

The following heuristics can help with design decisions.

Shares vs. Limits When you use bandwidth allocation, consider using shares instead of limits.Limits impose hard limits on the amount of bandwidth used by a traffic floweven when network bandwidth is available.

Limits on NetworkResource Pools

Consider imposing limits on a resource pool. For example, set a limit onvSphere vMotion traffic to avoid oversubscription at the physical networklevel when multiple vSphere vMotion data transfers are initiated on differentESXi hosts at the same time. By limiting the available bandwidth forvSphere vMotion at the ESXi host level, you can prevent performancedegradation for other traffic.

Network I/O Control Design Decisions

Based on the heuristics, this design has the following decisions.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 42

Page 43: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 3‑24. Network I/O Control Design Decisions

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-SDN-017

Create and attach a NetworkI/O Control Policy on all N-DVSswitches.

Increases resiliency and performance of thenetwork.

If configured incorrectly,Network I/O Controlmight impact networkperformance for criticaltraffic types.

NSXT-VI-SDN-018

Set the share value forvSphere vMotion traffic to Low(25).

During times of network contention, vSpherevMotion traffic is not as important as virtualmachine or storage traffic.

During times of networkcontention, vMotiontakes longer than usualto complete.

NSXT-VI-SDN-019

Set the share value forvSphere Replication traffic toLow (25).

During times of network contention, vSphereReplication traffic is not as important as virtualmachine or storage traffic.

During times of networkcontention, vSphereReplication takes longerand might violate thedefined SLA.

NSXT-VI-SDN-020

Set the share value for vSANtraffic to High (100).

During times of network contention, vSAN trafficneeds a guaranteed bandwidth to support virtualmachine performance.

None.

NSXT-VI-SDN-021

Set the share value formanagement traffic to Normal(50).

By keeping the default setting of Normal,management traffic is prioritized higher thanvSphere vMotion and vSphere Replication butlower than vSAN traffic. Management traffic isimportant because it ensures that the hosts canstill be managed during times of networkcontention.

None.

NSXT-VI-SDN-022

Set the share value for NFStraffic to Low (25).

Because NFS is used for secondary storage,such as backups and vRealize Log Insightarchives, it is not as important as vSAN traffic.By prioritizing it lower, vSAN is not impacted.

During times of networkcontention, backups areslower than usual.

NSXT-VI-SDN-023

Set the share value for backuptraffic to Low (25).

During times of network contention, the primaryfunctions of the SDDC must continue to haveaccess to network resources with priority overbackup traffic.

During times of networkcontention, backups areslower than usual.

NSXT-VI-SDN-024

Set the share value for virtualmachines to High (100).

Virtual machines are the most important asset inthe SDDC. Leaving the default setting of Highensures that they always have access to thenetwork resources they need.

None.

NSXT-VI-SDN-025

Set the share value forvSphere Fault Tolerance toLow (25).

This design does not use vSphere FaultTolerance. Fault tolerance traffic can be set thelowest priority.

None.

NSXT-VI-SDN-026

Set the share value for iSCSItraffic to Low (25).

This design does not use iSCSI. iSCSI trafficcan be set the lowest priority.

None.

Transport Node and Uplink Policy DesignA transport node can participate in an NSX-T overlay or NSX-T VLAN network.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 43

Page 44: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Several types of transport nodes are available in NSX-T.

ESXi Host TransportNodes

ESXi host transport nodes are ESXi hosts prepared and configured forNSX-T. N-VDS provides network services to the virtual machines runningon these ESXi hosts.

Edge Nodes NSX-T Edge nodes are service appliances that run network services thatcannot be distributed to the hypervisors. They are grouped in one orseveral NSX-T Edge clusters. Each cluster represents a pool of capacity.

Uplink profiles define policies for the links from ESXi hosts to NSX-T Segments or from NSX-T Edgevirtual machines to top of rack switches. By using uplink profiles, you can apply consistent configurationof capabilities for network adapters across multiple ESXi hosts or edge virtual machines. Uplink profilesare containers for the properties or capabilities for the network adapters.

Uplink profiles can use either load balance source or failover order teaming. If using load balance source,multiple uplinks can be active. If using failover order, only a single uplink can be active.

Table 3‑25. Design Decisions on Transport Nodes and Uplink Policy

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-027 Create an uplink profile withthe load balance sourceteaming policy with two activeuplinks for ESXi hosts.

For increased resiliency andperformance, supports theconcurrent use of bothphysical NICs on the ESXihosts that are configured astransport nodes.

You can use this policy onlywith ESXi hosts. Edge virtualmachines must use thefailover order teaming policy.

NSXT-VI-SDN-028 Create an uplink profile withthe failover order teamingpolicy with one active uplinkand no standby uplinks foredge virtual machine overlaytraffic.

Provides a profile thataccording to the requirementsfor Edge virtual machines.Edge virtual machinessupport uplink profiles onlywith a failover order teamingpolicy.

VLAN ID is required in theuplink profile. Hence, youmust create an uplink profilefor each VLAN used by theedge virtual machines.

n You create and managemore uplink profiles.

n The VLAN ID used mustbe different than theVLAN ID for ESXi hostoverlay traffic.

NSXT-VI-SDN-029 Create two uplink profiles withthe failover order teamingpolicy with one active uplinkand no standby uplinks foredge virtual machine uplinktraffic.

Enables ECMP because theеdge virtual machine canuplink to the physical networkover two different VLANs.

You create and manage moreuplink profiles.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 44

Page 45: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 3‑25. Design Decisions on Transport Nodes and Uplink Policy (Continued)

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-030 Create a Transport NodePolicy with the VLAN andOverlay Transport Zones, N-VDS settings, Physical NICsand Network mappings toallow NSX-T to migrate theVMKernel adapter to the N-VDS.

Allows the profile to beassigned directly to thevSphere cluster and ensuresconsistent configurationacross all ESXi HostTransport Nodes.

You must create all requiredTransport Zones andSegments before creating theTransport Node Policy.

NSXT-VI-SDN-031 Add as transport nodes allESXi hosts in the SharedEdge and Compute cluster byapplying the Transport NodePolicy to the vSphere clusterobject.

Enables the participation ofESXi hosts and the virtualmachines on them in NSX-Toverlay and VLAN networks.

ESXi hosts VMKerneladapters are migrated to theN-VDS. Ensure the VLAN ID'sfor the VMKernel Segmentsare correct to ensure hostcommunication is not lost.

NSXT-VI-SDN-032 Add as transport nodes alledge virtual machines.

Enables the participation ofedge virtual machines in theoverlay network and thedelivery of services, such asrouting, by these machines.

None.

NSXT-VI-SDN-033 Create an NSX-T Edgecluster with the defaultBidirectional ForwardingDetection (BFD) settingscontaining the edge transportnodes.

Satisfies the availabilityrequirements by default.

Edge clusters are required tocreate services such as NAT,routing to physical networks,and load balancing.

None.

Routing Design by Using NSX-TThe routing design considers different levels of routing in the environment, such as number and type ofNSX-T routers, dynamic routing protocol, and so on. At each level, you apply a set of principles fordesigning a scalable routing solution.

Routing can be defined in the following directions: North-South and East-West.

n North-South traffic is traffic leaving or entering the NSX-T domain, for example, a virtual machine onan overlay network communicating with an end-user device on the corporate network.

n East-West traffic is traffic that remains in the NSX-T domain, for example, two virtual machines on thesame or different segments communicating with each other.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 45

Page 46: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Table 3‑26. Design Decisions on Routing Using NSX-T

Decision ID Design Decision Design Justification Design Implications

NSXT-VI-SDN-034

Create two VLANs to enableECMP between the Tier-0Gateway and the Layer 3device (ToR or upstreamdevice).

The ToR switches orupstream Layer 3 deviceshave an SVI on one of the twoVLANS and each edge virtualmachine has an interface oneach VLAN.

Supports multiple equal-cost routes onthe Tier-0 Gateway and provides moreresiliency and better bandwidth use inthe network.

Extra VLANs are required.

NSXT-VI-SDN-035

Deploy an Active-ActiveTier-0 Gateway.

Supports ECMP North-South routingon all edge virtual machines in theNSX-T Edge cluster.

Active-Active Tier-0 Gatewayscannot provide services such asNAT. If you deploy a specificsolution that requires statefulservices on the Tier-0 Gateway,such as VMware Pivotal ContainerService, you must deploy anadditional Tier-0 Gateway in Active-Standby mode.

NSXT-VI-SDN-036

Use BGP as the dynamicrouting protocol.

Enables the dynamic routing by usingNSX-T. NSX-T supports only BGP .

In environments where BGP cannotbe used, you must configure andmanage static routes.

NSXT-VI-SDN-037

Configure BGP Keep AliveTimer to 4 and Hold DownTimer to 12 between the ToRswitches and the Tier-0Gateway.

Provides a balance between failuredetection between the ToR switchesand the Tier-0 Gateway andoverburdening the ToRs with keepalive traffic.

By using longer timers to detect if arouter is not responding, the dataabout such a router remains in therouting table longer. As a result, theactive router continues to sendtraffic to a router that is down.

NSXT-VI-SDN-038

Do not enable GracefulRestart between BGPneighbors.

Avoids loss of traffic. Graceful Restartmaintains the forwarding table which inturn will forward packets to a downneighbor even after the BGP timershave expired causing loss of traffic.

None.

NSXT-VI-SDN-039

Deploy a Tier-1 Gateway tothe NSX-T Edge cluster andconnect it to the Tier-0Gateway.

Creates a two-tier routing architecturethat supports load balancers and NAT.

Because the Tier-1 is alwaysActive/Standby, creation of servicessuch as load balancers or NAT ispossible.

A Tier-1 Gateway can only beconnected to a single Tier-0Gateway.

In scenarios where multiple Tier-0Gateways are required, you mustcreate multiple Tier-1 Gateways.

Virtual Network Design Example Using NSX-TDesign a setup of virtual networks where you determine the connection of virtual machines to Segmentsand the routing between the Tier-1 Gateway and Tier-0 Gateway, and then between the Tier-0 Gatewayand the physical network.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 46

Page 47: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Figure 3‑8. Virtual Network Example

Physical Layer 3 Devices

Tier-1 to Tier-0 Connection

Legend:

Segment 1 - 192.168.100.0/24

VLANs

Segment

Tier-0GatewayActive/Active

ECMP

NSX-TEdge Cluster

Edge VM 1 Edge VM 2

Tier-1Gateway

Linux Windows Linux

Segment 2 - 192.168.200.0/24

Linux Windows Linux

Monitoring NSX-TMonitor the operation of NSX-T for identifying failures in the network setup. Use the vRealize Log InsightContent Pack for VMware NSX-T to see the logs generated by the components of NSX-T in the userinterface of vRealize Log Insight.

vRealize Log Insight saves log queries and alerts, and you can use dashboards for efficient monitoring.

Table 3‑27. Design Decisions on Monitoring NSX-T

Decision ID Design Decision Design Justification Design Implication

NSXT-VI-SDN-040 Install the content pack for NSX-Tin vRealize Log Insight.

Provides granular monitoring of theNSX-T infrastructure.

Requires manuallyinstalling the content pack.

NSXT-VI-SDN-041 Configure each NSX-T componentto send log information over syslogto the vRealize Log Insight clusterVIP.

Ensures that all NSX-T componentslog files are available for monitoringand troubleshooting in vRealize LogInsight.

Requires manuallyconfiguring syslog on eachNSX-T component.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 47

Page 48: Architecture and Design of VMware NSX-T for Workload ... · Architecture and Design of VMware NSX-T for Workload Domains provides detailed information about the ... n NSX for vSphere

Use of SSL Certificates in NSX-TBy default, NSX-T Manager uses a self-signed Secure Sockets Layer (SSL) certificate. This certificate isnot trusted by end-user devices or Web browsers.

As a best practice, replace self-signed certificates with certificates that are signed by a third-party orenterprise Certificate Authority (CA).

Table 3‑28. Design Decisions on the SSL Certificate of NSX-T Manager

Design ID Design Decision Design Justification Design Implication

NSXT-VI-SDN-042 Replace the NSX-T Managerscertificate with a certificatethat is signed by a third-partyPublic Key Infrastructure.

Ensures that thecommunication betweenNSX-T administrators and theNSX-T Manager is encryptedby using a trusted certificate.

Replacing and managingcertificates is an operationaloverhead.

NSXT-VI-SDN-043 Replace the NSX-T ManagerCluster certificate with acertificate that is signed by athird-party Public KeyInfrastructure.

Ensures that thecommunication between thevirtual IP on the NSX-TManager Cluster andadministrators is encrypted byusing a trusted certificate.

Replacing and managingcertificates is an operationaloverhead.

Architecture and Design of VMware NSX-T for Workload Domains

VMware, Inc. 48