Nutanix TechNote-VMware VSphere Networking With Nutanix
Nutanix TechNote-VMware VSphere Networking With Nutanix
Nutanix TechNote-VMware VSphere Networking With Nutanix
Networking
Nutanix Best Practices
Copyright
Copyright 2017 Nutanix, Inc.
Nutanix, Inc.
1740 Technology Drive, Suite 150
San Jose, CA 95110
All rights reserved. This product is protected by U.S. and international copyright and intellectual
property laws.
Nutanix is a trademark of Nutanix, Inc. in the United States and/or other jurisdictions. All other
marks and names mentioned herein may be trademarks of their respective companies.
Copyright | 2
VMware vSphere Networking
Contents
1. Executive Summary................................................................................ 5
4. Discovery Protocols..............................................................................14
7. Conclusion............................................................................................. 31
Appendix......................................................................................................................... 32
References.......................................................................................................................... 32
3
VMware vSphere Networking
List of Figures................................................................................................................34
List of Tables................................................................................................................. 35
4
VMware vSphere Networking
1. Executive Summary
With the VMware vSphere hypervisor, you can dynamically configure, balance, or share logical
networking components across various traffic types. Therefore, its imperative that we take virtual
networking into consideration when architecting a network solution for Nutanix hyperconverged
appliances.
This best practices guide addresses the fundamental features of VMwares vSphere switching
technology and details the configuration elements you need to run vSphere on the Nutanix
Enterprise Cloud Platform. The guidance in this document can help you quickly and easily
develop an appropriate design strategy for deploying the Nutanix hyperconverged platform with
VMware vSphere 6.x as the hypervisor.
1. Executive Summary | 5
VMware vSphere Networking
2.4. AHV
Nutanix ships with AHV, a built-in enterprise-ready hypervisor based on a hardened version of
proven open source technology. AHV is managed with the Prism interface, a robust REST API,
and an interactive command-line interface called aCLI (Acropolis CLI). These tools combine to
eliminate the management complexity typically associated with open source environments and
allow out-of-the-box virtualization on Nutanixall without the licensing fees associated with other
hypervisors.
The figure below shows an overview of the Nutanix architecture, including the hypervisor of
your choice (AHV, ESXi, Hyper-V, or XenServer), user VMs, the Nutanix storage CVM, and its
local disk devices. Each CVM connects directly to the local storage controller and its associated
disks. Using local storage controllers on each host localizes access to data through the DSF,
thereby reducing storage I/O latency. Moreover, having a local storage controller on each
node ensures that storage performance as well as storage capacity increase linearly with node
addition. The DSF replicates writes synchronously to at least one other Nutanix node in the
system, distributing data throughout the cluster for resiliency and availability. Replication factor
2 creates two identical data copies in the cluster, and replication factor 3 creates three identical
data copies.
Local storage for each Nutanix node in the architecture appears to the hypervisor as one large
pool of shared storage. This allows the DSF to support all key virtualization features. Data
localization maintains performance and quality of service (QoS) on each host, minimizing the
effect noisy VMs have on their neighbors performance. This functionality allows for large, mixed-
workload clusters that are more efficient and more resilient to failure than traditional architectures
with standalone, shared, and dual-controller storage arrays.
When VMs move from one hypervisor to another, such as during live migration or a high
availability (HA) event, the now local CVM serves a newly migrated VMs data. While all write I/
O occurs locally, when the local CVM reads old data stored on the now remote CVM, the local
CVM forwards the I/O request to the remote CVM. The DSF detects that I/O is occurring from a
different node and migrates the data to the local node in the background, ensuring that all read I/
O is served locally as well.
The next figure shows how data follows the VM as it moves between hypervisor nodes.
The vDS has matured over the several version iterations of the vSphere product, providing new
functionality as well as changes to existing components and behaviors.
vSphere 5.x provided the following capabilities and enhancements:
Inter-VM traffic visibility via the Netflow protocol.
LLDP (link layer discovery protocol) support, to provide upstream switch visibility.
Enhanced link aggregation support, including a variety of hashing algorithms.
Single-root I/O virtualization (SR-IOV) support.
Support for 40 Gb network interface cards (NICs).
Network I/O Control (NIOC) version 2.
In vSphere 6.x, VMware has introduced improvements as well as new features and functionality,
including:
NIOC version 3.
Support for IGMP snooping for IPv4 packets and MLD snooping for IPv6 packets.
Multiple TCP/IP stacks for vMotion.
The following table concisely compares the vSwitch and the vDS.
Originating port ID
IP hash
Host or
vSwitch Standard No No Source MAC hash No No
vCenter
Explicit failover
order
Originating port ID
IP hash
Source MAC hash
Enterprise
vDS vCenter Yes Yes Yes Yes
Plus Explicit failover
order
Route based on
physical NIC load
4. Discovery Protocols
From vSphere 5.0 onward, the vDS has supported two discovery protocols: LLDP and CDP
(Cisco discovery protocol). Discovery protocols give vSphere administrators visibility into
connectivity from the virtual network to the physical network, which simplifies troubleshooting in
the event of cabling issues, MTU (maximum transmission unit) or packet fragmentation, or other
problems. The only disadvantage to using discovery protocols is a potential security issue, as
they make both the topology of the network and switch-specific information visible.
VMware vSphere offers three configuration options or attributes for LLDP, as presented in the
table below. Each option varies the information that the selected discovery protocol sends and
receives. CDP is either enabled or notit does not have additional options.
Attribute Description
In this mode, ESXi detects and displays information from the upstream switch,
Listen
but it does not advertise the vDS configuration to the upstream network device.
In this mode, ESXi advertises information about the vDS to the switch
Advertise administrator, but it does not detect or display information about the physical
switch.
This attribute listens and advertises information between the vDS and upstream
Both
switch provider.
Note: Nutanix recommends the "Both" option, which ensures that ESXi collects and
displays information from the physical switch and sends information about the vDS to
the physical switch.
The following information is visible in the vSphere client when enabling discovery protocol data:
The physical switch interface connected to the dvUplink.
MTU size (in other words, whether or not jumbo frames are enabled).
The switch management IP address.
The switch name, description, software version, and capabilities.
4. Discovery Protocols | 14
VMware vSphere Networking
The following table presents some general guidelines, but Nutanix recommends that all
customers carefully consider the advantages and disadvantages of discovery protocols for their
specific security and discovery requirements.
4. Discovery Protocols | 15
VMware vSphere Networking
Note: None of the network resource pools configure artificial limits or reservations,
so leave the host limit set to unlimited.
The sections below describe our rationale for setting the share values in this table.
Note: The calculations of minimum bandwidth per traffic type presented here
assume that the system is not using iSCSI, vSphere Replication, vSphere Storage
Area Network traffic, or Virtual SAN traffic. Also, we assume that NFS traffic remains
with the host for the default "NFS traffic" resource pool.
Management Traffic
Management traffic requires minimal bandwidth. A share value of 25 (low) and two 10 Gb
interfaces ensure a minimum of approximately 1.5 Gbps for management traffic, which exceeds
the minimum bandwidth requirement of 1 Gb.
vMotion Traffic
vMotion is a burst-type workload that uses no bandwidth until the distributed resource scheduler
(DRS) (a vCenter service that actively rebalances VMs across hosts) invokes it or a vSphere
administrator starts a vMotion or puts a host into maintenance mode. As such, it is unlikely to
have any significant ongoing impact on the network traffic. A share value of 50 (normal) and two
10 Gb interfaces provide a minimum of approximately 3 Gbps, which is sufficient to complete
vMotion in a timely manner without degrading VM or storage performance and well above the
minimum bandwidth requirement of 1 Gb.
VM Traffic
VM traffic is the reason we have datacenters in the first place, so this traffic is always important,
if not critical. As slow VM network connectivity can quickly impact end users and reduce
productivity, we must ensure that this traffic has a significant share of the available bandwidth
during periods of contention. With a share value of 100 (high) and two 10 Gb interfaces, VM
traffic receives a minimum of approximately 6 Gbps. This bandwidth is more than what is
required in most environments and ensures a good amount of headroom in case of unexpected
burst activity.
NFS Traffic
NFS traffic is always critical, as its essential to the Acropolis DSF and to CVM and VM
performance, so it must receive a significant share of the available bandwidth during periods
of contention. However, as NFS traffic is serviced locally, it does not impact the physical
network card unless the Nutanix CVM fails or is offline for maintenance. Thus, under normal
circumstances, no NFS traffic crosses the physical NICs, so the NIOC share value has no impact
on other traffic. As such, we have excluded it from our calculations.
In case of network contention, CVM maintenance, or failure, we can assign NFS traffic a share
value of 100 (high) as a safety measure, ensuring a minimum of 6 Gbps bandwidth.
To ensure optimal DSF performance in the unlikely circumstance wherein both 10 Gb adapters
are saturated, we assign a share value of 100 (high), guaranteeing each CVM a minimum of
6 Gbps bandwidth. This bandwidth is more than what is required in most environments and
ensures a good amount of headroom in case of unexpected burst activity.
iSCSI Traffic
Like NFS traffic, iSCSI traffic is not used by default; however, if you are using it, it may be critical
to your environment. As such, we have given this traffic type a share value of 100.
Note: NIOC does not cover in-guest iSCSI. If you are using in-guest iSCSI, we
recommend creating a dvPortGroup for in-guest iSCSI traffic and assigning it to a
custom network resource pool called "In-Guest iSCSI." Assign this pool a share value
of 100 (high).
Although setting artificial limits and reservations on various traffic types guarantees quality of
service (QoS) and SLAs, it prevents other workloads from using the total available bandwidth.
Reserving a certain amount of bandwidth strictly for one workload or purpose constrains the
networks ability to sustain an unreserved workloads short bursts of traffic.
In the course of testing and evaluating the functionality of NIOC v3, Nutanix has encountered
various abnormalities in its usability and behavior as well as generally higher CPU overhead.
Because of these results, we recommend that customers implement NIOC v2 instead of NIOC
v3.
To implement NIOC v2, provision a vDS with 5.5.0 selected as the Distributed Switch version,
following the proportional shares we described above for NIOC v2. The section below offers
detailed instructions for provisioning your vDS.
LBT is a very simple and effective solution that works very well in Nutanix deployments.
Notify Switches
The purpose of the notify switches policy setting is to enable or disable communication by ESXi
with the physical switch in the event of a failover. Choosing Yes for this policy setting means that
when a failover event routes a virtual Ethernet adapters traffic over a different physical Ethernet
adapter in the team, ESXi sends a notification to the physical switch to update the lookup tables.
This configuration ensures that failover occurs in a timely manner and with minimal interruption to
network connectivity.
Failback
For customers not using Enterprise Plus or the vDS with load-based teaming, failback can help
rebalance network traffic across the original NIC, which may improve network performance.
The only significant disadvantage of setting failback to Yes is that, in the unlikely event of
network instability (or "flapping"), having network traffic fail back to the original NIC may result in
intermittent or degraded network connectivity.
Nutanix recommends setting failback to Yes when using vSwitch and No when using vDS.
Failover Order
Using failover order allows the vSphere administrator to specify the order in which NICs fail
over by assigning a pNIC to one of three groups: active adapters, standby adapters, or unused
adapters. For example, if all active adapters lose connectivity, the system uses the highest
priority standby adapter, and so on. This feature is only necessary in a Nutanix environment
when performing multi-NIC vMotion.
When configuring multi-NIC vMotion, set the first dvPortGroup used for vMotion to have one
dvUplink active and the other standby. Configure the reverse for the second dvPortGroup used
for vMotion.
For more information, see Multiple-NIC vMotion in vSphere 5 (KB2007467).
5.4. Security
When configuring a vSwitch or vDS, there are three configurable options under security that can
be set to Accept or Reject: promiscuous mode, MAC address changes, and forged transmits.
In general, the most secure and appropriate setting for each of these options is Reject unless
you have a specific requirement for Accept. Two examples of situations in which you might
consider configuring Accept on forged transmits and MAC address changes are:
1. Microsoft load balancing in Unicast mode.
2. iSCSI deployments on select storage types.
The following table offers general guidelines, but Nutanix recommends that all customers
carefully consider their requirements for specific applications.
All Nutanix deployments use an internal-only vSwitch for the NFS communication between the
ESXi host and the Nutanix CVM. This vSwitch remains unmodified regardless of the virtual
networking configuration for ESXi management, VM traffic, vMotion, and so on.
The configurations for vSwitch and vDS solutions look very similarthe main difference is the
location of the control plane.
With the default vSwitch, there is a control plane on each host, and each operates
independently. This independence requires an administrator to configure, maintain, and
operate each individual vSwitch separately.
With the vDS, the control plane is located within vCenter. This centrality requires the system to
keep vCenter online to ensure that configuration and maintenance activities propagate to the
ESXi hosts that are part of the vDS instance.
Both options offer the advantage of reduced cabling and switching requirements (because they
do not require 1 GbE ports). Moreover, they represent a simple solution that only requires 802.1q
configured on the physical network. These options are suitable for environments where logical
separation of management and VM traffic is acceptable.
Option 1 offers several benefits; the most important benefit is that the control plane is located
locally on every ESXi host, so there is no dependency on vCenter.
A major disadvantage of the vSwitch lies in its load balancing capabilities. Although both physical
uplinks are active, traffic across these pNICs does not actively rebalance unless the VM has
gone through a power cycle or a failure has occurred in the networking stack. Furthermore, a
standard vSwitch cannot manage contention or prioritize network traffic across various traffic
classes (for example, storage, vMotion, VM, and so on).
Figure 5: Virtual Networking Option 1: Single vSwitch with Nutanix Internal-Only vSwitch
networks, larger frames assist with both greater throughput and reduced network overhead.
By default, the network moves large numbers of 1,500-byte frames; however, with larger,
9,000-byte frames, the network can process six times fewer packets, thus reducing the CPU
overhead. Solutions such as VXLAN require a payload greater than 1,500 bytes; thus, VMware
recommends using jumbo frames for VXLAN to avoid packet fragmentation.
When considering jumbo frames, be sure to confirm that all network devices support them end-
to-end and that you can enable this support on all switches globally. If you can confirm both of
these capabilities, consider using jumbo frames.
Note: You can choose to enable jumbo frames for only specific traffic types;
however, ensure it is configured end-to-end
Tip: Nutanix supports using a standard (not jumbo) MTU of 1,500. Performance
using this MTU is still excellent.
The figure above illustrates how jumbo frames and standard frames can coexist and
communicate in a Nutanix deployment.
Suited to jumbo frames (1) (MTU 9,000) Nutanix CVM internal traffic
vMotion / multi-NIC vMotion
Fault tolerance
iSCSI
NFS
VXLAN (2)
The following table shows the various traffic types in a Nutanix environment with VMware
vSphere, as well as the Nutanix-recommended MTU.
In summary, jumbo frames offer the following benefits:
Reduced number of interrupts for the switches to process.
Lower CPU utilization on the switches and on Nutanix CVMs.
Increased performance for some workloads, including the Nutanix CVM, vMotion, and fault
tolerance.
Performance with jumbo frames is either the same or better than without them.
Note: Enabling multicast snooping on a vDS with a Nutanix CVM attached affects
its ability to discover and add new nodes in the cluster (the Cluster Expand option in
Prism and the Nutanix CLI).
If you need to implement multicast snooping, your standard operational procedures should
include manual steps for adding additional nodes to a cluster. For more information, refer to
Nutanix KB 3493 and VMwares vSphere 6.0 documentation on Multicast Filtering Modes.
7. Conclusion
Virtual networking within a vSphere 6.x environment plays an important role in infrastructure
availability, scalability, performance, management, and security. With hyperconverged
technology, customers need to evaluate their networking requirements carefully against the
various configuration options, implications, and features within the VMware vSphere 6.x
hypervisor.
In most deployments, Nutanix recommends the vSphere Distributed Switch (vDS) coupled with
Network I/O Control (Version 2) and load-based teaming, as this combination provides simplicity,
ensures traffic prioritization in the event of contention, and reduces operational management
overhead.
With a strategically networked vSphere deployment on Nutanix, customers can be confident that
their networking configuration conforms to field-tested best practices.
For feedback or questions, please contact us using the Nutanix NEXT Community forums.
7. Conclusion | 31
VMware vSphere Networking
Appendix
References
1. VMware vSphere 6.0 Documentation
2. Nutanix Jumbo Frames
3. Performance Evaluation of NIOC in VMware vSphere 6.0
Appendix | 32
VMware vSphere Networking
About Nutanix
Nutanix makes infrastructure invisible, elevating IT to focus on the applications and services that
power their business. The Nutanix Enterprise Cloud Platform leverages web-scale engineering
and consumer-grade design to natively converge compute, virtualization, and storage into
a resilient, software-defined solution with rich machine intelligence. The result is predictable
performance, cloud-like infrastructure consumption, robust security, and seamless application
mobility for a broad range of enterprise applications. Learn more at www.nutanix.com or follow up
on Twitter @nutanix.
Appendix | 33
VMware vSphere Networking
List of Figures
Figure 1: Nutanix Enterprise Cloud Platform.....................................................................7
Figure 5: Virtual Networking Option 1: Single vSwitch with Nutanix Internal-Only vSwitch 25
34
VMware vSphere Networking
List of Tables
Table 1: Document Version History.................................................................................. 5
35