VSAN 2 Node Guide
VSAN 2 Node Guide
VSAN 2 Node Guide
Table Of Contents
1. Overview
1.1.Introduction
2. Support Statements
2.1.vSphere Versions
2.2.vSphere & vSAN
2.3.Hybrid and All-Flash Support
2.4.vSAN Witness Host
2.5.Supported on vSAN but not vSAN 2 Node Clusters
3. New Concepts in vSAN - 2 Node Clusters
3.1.The Witness Host
3.2.Read Locality in vSAN 2 Node Clusters
3.3.Witness Traffic Separation (WTS)
4. Requirements
4.1.VMware vCenter Server
4.2.A Witness Host
4.3.Networking and Latency Requirements
5. Configuration Minimums and Maximums
5.1.Virtual Machines Per Host
5.2.Witness Host
5.3.vSAN Storage Policies
5.4.Fault Domains
6. Design Considerations
6.1.Cluster Compute Resource Utilization
6.2.Network Design Considerations
6.3.The Role of vSAN Heartbeats
6.4.Bandwidth Calculation
6.5.Using the vSAN Witness Appliance as a vSAN Witness Host
6.6.Using a Physical Host as a vSAN Witness Host
6.7.vSAN Witness Host Networking Examples
6.8.vSAN Witness Appliance Sizing
6.9.vSAN Witness Host Versioning & Updating
7. Cluster Settings – vSphere HA
7.1.Cluster Settings – vSphere HA
7.2.Turn on vSphere HA
7.3.Host Monitoring
7.4.Admission Control
7.5.Host Hardware Monitoring – VM Component Protection
7.6.Datastore for Heartbeating
7.7.Virtual Machine Response for Host Isolation
7.8.Advanced Options
8. Cluster Settings - DRS
8.1.Cluster Settings - DRS
8.2.Partially Automated or Fully Automated DRS
9. Using a vSAN Witness Appliance
9.1.Setup Step 1: Deploy the vSAN Witness Appliance
9.2.Setup Step 2: vSAN Witness Appliance Management
9.3.Setup Step 3: Add Witness to vCenter Server
9.4.Setup Step 4: Config vSAN Witness Host Networking
10. Configuring 2 Node vSAN
10.1.Pre-Requisites
10.2.VMkernel Interfaces
10.3.VMkernel Interfaces - Witness Traffic
10.4.VMkernel Interfaces - vSAN Traffic - VSS
10.5.VMkernel Interfaces - vSAN Traffic - VDS
10.6.Creating a New 2 Node vSAN Cluster
10.7.Upgrading a older 2 Node vSAN Cluster
10.8.Converting a 2 Node Cluster with WTS to a 3 Node Cluster
11. Management and Maintenance
11.1.Management and Maintenance
11.2.Maintenance Mode Consideration
12. Failure Scenarios
12.1.Failure Scenarios
12.2.Individual Drive Failure
12.3.Host Failure / Network Partitions
12.4.VM Provisioning When a Host is Offline
12.5.Multiple Simultaneous Failures
12.6.Replacing a Failed vSAN Witness Host
12.7.Failure Scenario Matrices
13. Appendix A
13.1.Appendix A: Additional Resources
13.2.Location of the vSAN Witness Appliance OVA
14. Appendix B
14.1.Appendix B: Commands for vSAN 2 Node Clusters
1. Overview
vSAN 2 Node is a specific configuration typically implemented in environments
where a minimal configuration is required.
1. 1 Introduction
The vSAN 2 Node configuration was introduced in vSAN 6.1. A vSAN 2 Node
Cluster is a specific configuration implemented in environments where a
minimal configuration is a key requirement. This guide was developed to
provide additional insight and information for installation, configuration and
operation of a vSAN 2 Node Cluster infrastructure in conjunction with VMware
vSphere. This guide will explain how vSphere handles specific failure scenarios
and discuss various design considerations and operational procedures for 2
Node vSAN Releases including 6.1 through 6.7.
A vSAN Witness Host that provides quorum for the 2 Nodes can be hosted on
a third site via low bandwidth/high latency links, or on alternate infrastructure
in the same location.
With two Fault Domains available, virtual machines deployed on vSAN 2 Node
Clusters typically have mirrored data protection, with one copy of data on
Node 1, a second copy of data on Node 2, and the Witness component placed
on the vSAN Witness Host.
In the event of a node or device failure, a full copy of the virtual machine data
is still available on the alternate node. Because the alternate replica and
Witness component are still available, the virtual machine remains accessible
on the vSAN datastore.
In cases where the virtual machine was running on the failed node, vSphere HA
can handle this task if properly configured.
2. Support Statements
vSAN Stretched Cluster configurations require vSphere 6.0 Update 1 (U1) or
greater.
2. 1 vSphere Versions
vSAN 6.1 vSAN 6.2 vSAN 6.5 vSAN 6.6 vSAN 6.7
Feature Requireme Requireme Requireme Requireme Requireme
nts nts nts nts nts
vSAN 6.1 vSAN 6.2 vSAN 6.5 vSAN 6.6 vSAN 6.7
Feature Requireme Requireme Requireme Requireme Requireme
nts nts nts nts nts
v5 On- v5 On-
Encryption Disk Disk
format format
VMware vSAN 6.1 introduced several features including All-Flash and 2 Node
Cluster functionality. There are no limitations on the edition of vSphere used
for vSAN. However, for vSAN 2 Node functionality, vSphere Distributed
Resource Scheduler (DRS) is very desirable . DRS will provide initial placement
assistance, load balance the environment when there's an imbalance, and will
also automatically migrate virtual machines to their correct site in accordance
to VM/Host affinity rules. It can also help with migrating virtual machines to
their correct site when a node recovers after a failure. Otherwise the
administrator will have to manually carry out these tasks.
Both physical ESXi hosts and vSAN Witness Appliances are supported as a
vSAN Witness Host.
VMware provides a vSAN Witness Appliance for those customers who do not
wish to use a physical host for this role. The vSAN Witness Appliance must run
on an ESXi 5.5 or higher host. *If the vSAN Witness Appliance is used for a
vSAN 6.7 2 Node cluster, the ESXi 5.5 or higher host must also meet the CPU
requirements of vSphere 6.7.
This can include an ESXi Free licensed host, a vSphere licensed (ESXi) host, or a
host residing in OVH (formerly vCloud Air), a vCloud Air Network (VCAN)
partner, or any hosted ESXi installation.
10
In a vSAN 2 Node Clusters, there are only 3 Fault Domains. Two fault
domains are configured (typically referred to as the Preferred and
Secondary) and the Witness fault domain is implied. Standard vSAN
configurations can be comprised of up to 32 Fault Domains.
11
12
The vSAN Witness Host is a dedicated ESXi host, which may be a physical ESXi
host or vSAN Witness Appliance, whose purpose is to host the Witness
component of vSAN objects.
The vSAN Witness must have connectivity to both vSAN nodes to join the
cluster. In steady state operations, the master node is the host that resides in
the “Preferred site”; the backup node resides in the “Secondary site”. Unless
the vSAN Witness Host connects to both the master and the backup nodes, it
will not join the vSAN cluster.
The vSAN Witness Host must be managed by the same vCenter Server
managing the vSAN Cluster. There must be connectivity between vCenter
Server and the vSAN Witness Host in the same fashion as vCenter controlling
other vSphere hosts.
The vSAN Witness Host must also have connectivity between the to each vSAN
node.
In vSAN 6.5 VMware publicly introduced the feature known as Witness Traffic
Separation, where a separately tagged VMkernel interface may be used for
connectivity to the vSAN Witness Host, rather than requiring direct
connectivity to the vSAN Data Network. This capability can only be enabled
from the command line, and is supported when used with 2 Node Direct
Connect configurations.
13
In a vSAN 2 Node Clusters, 100% of reads, occur in the site the VM resides on.
This aligns with the behavior of vSAN Stretched Clusters. Read locality
overrides the NumberOfFailuresToTolerate=1 policy’s behavior to distribute
reads across the components.
Notice the blue triangle in the cache devices? That’s 30% of the cache device
being allocated as the write buffer. The other 70% of the cache device is green
, demonstrating the read cache. It is important to note that All-Flash vSAN
14
Writes go to the write buffer on both Nodes 1 and 2. This is always the case
because writes occur on both nodes simultaneously.
By default, reads are only serviced by the host that the VM is running on.
The image shows a typical read operation. The virtual machine is running on
Node 1 and all reads are serviced by the cache device of Node 1’s disk group.
By default, reads do not traverse to the other node. This behavior is default in 2
Node configurations, as they are mechanically similar to Stretched Cluster
configurations. This behavior is preferred when the latency between sites is at
the upper end of the supported boundary of 5ms round-trip-time (RTT).
15
Because the cache has not been warmed on the host the virtual machine is
now running on, reads will have to occur on the capacity drives as the cache is
warmed.
The image shows only part of the cache device as green, indicating that as
reads occur, they are cached in the read cache of the disk group.
The process of invoking a vMotion could be from various DRS events, such as
putting a host in maintenance mode or balancing workloads. The default
Stretched Cluster recommendation, is to keep virtual machines on one site or
the other, unless there is a failure event.
Read operations after a disk failure, are going to behave similarly to those of a
vMotion. A single disk in the disk group has failed in the image on the right.
Reads are going to come from Node 2, and the cache device on Node 2 is
going to start caching content from the virtual machine’s disk.
Since only a capacity device failed, and there are others still contributing to the
capacity, reads will also traverse the network, as data is rewritten to one of the
surviving capacity devices on Node 1, if there is sufficient capacity.
Once data has been reprotected on Node 1, the cache will have to rewarm on
Node 1 again.
16
Read operations after a disk group failure, are also going to behave like that of
a disk failure.
In configurations where the host with a failed disk group has an additional disk
group, rebuilds will occur on the surviving disk group provided there is
capacity.
In hosts with a single disk group, rebuilds will not occur, as the disk group is
unavailable.
What can be done to prevent the need for rewarming the read cache?
Forcing the cache to be read across Stretched Cluster sites is bad, because
additional read latency is introduced.
vSAN 2 Node configurations are typically in a single location, directly
17
When it is False (0), site locality is in effect and reads are only occurring on the
site the VM resides on.
Not only does this help in the event of a virtual machine moving across hosts,
which would require cache to be rewarmed, but it also allows reads to occur
across both mirrors, distributing the load more evenly across both hosts.
To check the status, run the following command on each ESXi host:
esxcfg-advcfg -g /VSAN/DOMOwnerForceWarmCache
18
Forcing reads to be read across both nodes in cases where the Number of
Failures to Tolerate policy is 1 can prevent having to rewarm the disk group
cache in cases of vMotions, host maintenance, or device failures.
Client Cache
VMware vSAN 6.2 introduced Client Cache, a mechanism that allocates 0.4%
of host memory, up to 1GB, as an additional read cache tier. Virtual machines
leverage the Client Cache of the host they are running on. Client Cache is not
associated with Stretched Cluster read locality, and runs independently.
19
Host 1
vmk0 - Tagged for Management Traffic
vmk1 - Tagged for Witness Traffic - This must* be done using esxcli
vsan network ip add -i vmk1 -T=witness
vmk2 - Tagged for vSAN Traffic
vmk3 - Tagged for vMotion Traffic
Host 2
vmk0 - Tagged for Management Traffic
vmk1 - Tagged for Witness Traffic - This must* be done using esxcli
vsan network ip add -i vmk1 -T=witness
20
*Enabling Witness Traffic is not available from the vSphere Web Client.
**Any VMkernel port, not used for vSAN Traffic, can be used for Witness
Traffic. In a more simplistic configuration, the Management VMkernel interface
(vmk0) could be tagged for Witness Traffic. The VMkernel port used, will be
required to have connectivity to the vSAN Traffic tagged interface on the vSAN
Witness Appliance.
***The vmk0 VMkernel Interface, which is used for Management traffic may
also be used for vSAN Traffic. In this situation, vSAN Traffic must be
unchecked from vmk1.
****The vmk1 VMkernel interface must not have an address that is on the same
subnet as vmk0. Because vSAN uses the default tcp/ip stack, in cases where
vmk0 and vmk1 are on the same subnet, traffic will use vmk0 rather than
vmk1. This is detailed in KB 2010877 . Vmk1 should be configured with an
address on a different subnet than vmk0.
The ability to connect 2 Nodes directly removes the requirement for a high
speed switch. This design can be significantly more cost effective when
deploying tens or hundreds of 2 Node clusters.
vSAN 2 Node Direct Connect was announced with vSAN 6.5, and is available
with vSAN 6.6, 6.5, and 6.2*. *6.2 using vSphere 6.0 Patch 3 or higher without
an RPQ.
21
4. Requirements
List of requirements for implementing vSAN 2 Node Clusters
22
4. 2 A Witness Host
In a vSAN 2 NodeCluster, the Witness components are only ever placed on the
vSAN Witness Host. Either a physical ESXi host, or the vSAN Witness Appliance
provided by VMware, can be used as the vSAN Witness Host.
If a vSAN Witness Appliance is used for the Witness host, it will not consume
any of the customer’s vSphere licenses. A physical ESXi host that is used as a
vSAN Witness Host will need to be licensed accordingly, as this can still be
used to provision virtual machines should a customer choose to do so.
It is important that the vSAN Witness Host is not added to the vSAN cluster.
The vSAN Witness Host is selected during the creation of a vSAN 2 Node
Cluster.
The vSAN Witness appliance will have a unique identifier in the vSphere UI to
assist with identifying that a host is in fact a vSAN Witness Appliance. It is
shown as a “blue” host, as highlighted below:
Note: This is only visible when the vSAN Witness Appliance is deployed. If a
physical host is used as a vSAN Witness Host, then it does not change its
appearance in the web client. A dedicated vSAN Witness Host is required for
23
Both Layer 2 (same subnet) and Layer 3 (routed) configurations are used in a
recommended vSAN 2 Node Cluster deployment.
vSAN traffic between nodes is multicast for versions 6.5 and previous, while
unicast is used for version 6.6 or later . Witness traffic between each node in a
2 Node cluster and the Witness Host is unicast .
This refers to the communication between vSAN hosts and the Witness
Host/Site.
The latency to the Witness Host is dependent on the number of objects in the
cluster.
Bandwidth between vSAN Nodes hosting VM objects and the Witness Host are
dependent on the number of objects residing on vSAN. It is important to size
data site to witness bandwidth appropriately for both availability and growth. A
standard rule of thumb is 2Mbps for every 1000 components on vSAN.
Because vSAN nodes have a maximum number of 9000 components per host,
24
Please refer to the Design Considerations section of this guide for further
details on how to determine bandwidth requirements.
Knowledge Base Article 2141733 details a situation where data nodes have
an MTU of 9000 (Jumbo Frames) and the vSAN Witness Host has an MTU of
1500. The vSAN Health Check looks for a uniform MTU size across all
VMkernel interfaces that are tagged for traffic related to vSAN, and reports
any inconsistencies. It is important to maintain a consistent MTU size across all
vSAN VMkernel interfaces on data nodes and the vSAN Witness Host in a
vSAN 2 Node cluster to prevent traffic fragmentation.
As KB 2141733 indicates, the corrective actions are either to reduce the MTU
from 9000 on the data node VMkernel interfaces, or increase the MTU value
on the vSAN Witness Host's VMkernel interface that is tagged for vSAN Traffic.
Either of these are acceptable corrective actions.
The placement of the vSAN Witness Host will likely be the deciding factor in
which configuration will be used. Network capability, control, and cost to/from
the vSAN Witness Host as well as overall performance characteristics on data
nodes are items to consider when making this design decision.
In situations where the vSAN Witness Host VMkernel interface tagged for
vSAN traffic is not configured to use Jumbo Frames (or cannot due to
underlying infrastructure), VMware recommends that all vSAN VMkernel
interfaces use the default MTU of 1500.
Multiple 2 Node vSAN deployments can send their witness traffic on the same
25
shared VLAN.
26
27
The maximum number of virtual machines per ESXi host is unaffected by the
vSAN 2 Node Cluster configuration. The maximum is the same as normal vSAN
deployments.
In the event of node failure the virtual machines on the failed node can be
restarted on the surviving node.
5. 2 Witness Host
In Pre-vSAN 6.6 2 Node Clusters FTT may not be greater than 1. In vSAN 6.6 or
higher 2 Node Clusters, PFTT may not be greater than 1. This is because 2
Node Clusters are comprised of 3 Fault Domains.
28
Affinity
Affinity rules are used when the PFTT rule value is 0. This rule has 2 values,
Preferred or Secondary. This determines which site an Affinity based vmdk
would reside on.
Other policy settings are not impacted by deploying vSAN in a 2 Node Cluster
configuration and can be used as per a non-stretched vSAN cluster.
5. 4 Fault Domains
The "Preferred" Fault Domain is selected in the vSphere UI, and assigns the
"vSAN Master" node role to the host in that Fault Domain. The alternate Fault
Domain assigns the "vSAN Backup" node role to the host in that Fault Domain.
These Fault Domains are typically named "Preferred" and "Secondary" but are
not required to be. One of these Fault Domains must be designed with the
"Preferred" setting. The vSAN Witness Host provides a 3rd implied fault
domain.
29
6. Design Considerations
The witness host must be capable of running the same version of ESXi as vSAN
data nodes.
30
VMware understands that some customers will want to run levels of resource
utilization higher than 50%. While it is possible to run at higher utilization in
each site, customers should understand that in the event of failure, not all
virtual machines will be restarted on the surviving node.
vSA FT S
Capacity Capacity Capacity
N Protec T/P FT F
Required in Required in Require
Versi tion FT M T
Preferred Site Secondary Site ment
on T T
Across Mir
6.1- N
Sites 1 ror 100% 100% 200%
6.6 A
Only ing
31
A vSAN 2 Node Cluster typically runs in a single site, with the vSAN Witness Host residing in an alternate location
or on an alternate host in the same site.
Witness Site - Contains vSAN Witness host - Could be in a different site, or the
small site as the 2 Node cluster.
Managemen
Layer 2 or 3
t Network to Layer 2 or 3 Layer 2 or 3
vCenter
Managemen
t Network to
Typically Layer 2 Typically Layer 2
Alternate
Host
32
No requirement
for VM Network.
Running VMs on
the vSAN Witness
Appliance is not
VM Network Typically Layer 2 Typically Layer 2
supported.
Running VMs on
a Physical
Witness Host is
supported.
Layer 2
Layer 2
Multicast: vSAN
vSAN Multicast: vSAN
6.1-6.5
Network 6.1-6.5
Unicast: vSAN
Unicast: vSAN 6.6+
6.6+
Witness
Traffic Using To the vSAN To the vSAN
To the Data Site:
WTS & vSAN Witness Site: Witness Site: Layer
Layer 3
Witness Layer 3 3
Offsite
Witness
Traffic Using To the vSAN To the vSAN To the vSAN
WTS & vSAN Witness: Layer 2 Witness: Layer 2 or Nodes: Layer 2 or
Witness or 3 3 3
Locally
33
Port Requirements
VMware vSAN requires these ports to be open, both inbound and outbound:
Po Prot Connectivity
rt ocol To/From
12
vSAN 34
Clustering 5, UDP vSAN Hosts
Service 23
45
vSAN 22
TCP vSAN Hosts
Transport 33
vSAN VASA
80 vSAN Hosts
Vendor TCP
80 and vCenter
Provider
vSAN Hosts
vSAN Unicast 12
and vSAN
Agent (to 32 UDP
Witness
Witness Host) 1
Appliance
34
TCP/IP Stacks
At this time, the vSAN traffic does not have its own dedicated TCP/IP stack.
Custom TCP/IP stacks are also not applicable for vSAN traffic.
ESXi hosts come with a default TCP/IP stack. As a result, hosts have a single default gateway. This default
gateway is associated with the Management VMkernel interface (typically vmk0).
vSAN networking uses the same TCP/IP stack as the Management VMkernel interface. If the vSAN Data Network
were to attempt to use a Layer 3 network, static routing would be need to be added for the VMkernel interfaces
tagged for vSAN Traffic.
Communication with the vSAN Witness via an alternate VMkernel interface can
be performed by using a separate VMkernel interface, or the Management
interface, as long as the interface used has traffic tagged as "Witness Traffic".
It is important to remember that vSAN uses the same TCP/IP stack as the
Management interface, and therefore if an alternate interface is used, static
routing must be in place for the vSAN node to properly communicate with the
vSAN Witness Host
The Witness Host on the Witness Site will require a static route added so that
requests to reach the 2 Node Cluster are routed out the WitnessPg VMkernel
interface
Static routes are added via the esxcli network ip route or esxcfg-route commands. Refer to the appropriate
vSphere Command Line Guide for more information.
35
either site 1 or site 2 needed to have static routes manually added before they
can successfully communicate to the witness, and the other data site. Any
replacement of the witness host will also require the static routes to be
updated to facilitate communication to the data sites.
In the illustration below, each vSAN Host's vmk1 VMkernel interface is tagged
with "witness" traffic. Each vSAN Host must have a static route configured for
vmk1 able to properly access vmk1 on the vSAN Witness Host, which is tagged
with "vsan" traffic. The vmk1 interface on the vSAN Witness Host must have a
static route configured to be able to properly communicate with vmk1 on each
vSAN Host. This is a supported configuration.
Sample configuration using Witness Traffic Separation using only the Management
VMkernel Interface on each ESXi Host
In the illustration below, each vSAN Host's vmk0 VMkernel interface is tagged
with both "Management" and "witness" traffic. The vSAN Witness Host has the
vmk0 VMkernel interface tagged with both "Management" and "vsan" traffic.
This is also a supported configuration.
36
In this illustration, the vmk0 and vmk1 VMkernel interfaces are on the same
network. Vmk1 is tagged for "vsan" traffic and vmk0 is tagged for
"Managment" traffic. Because vSAN uses the default TCP/IP stack, vSAN
traffic does not properly flow form vmk1, which is tagged for "vsan" traffic, but
rather from vmk0, which is NOT tagged for "vsan" traffic. This causes an error
with the vSAN Health Check indicating that the vSAN Witness Host does not
have proper tagging.
37
The issue is not unique to vSAN, but rather occurs when using any VMkernel
interfaces using the default TCP/IP stack. KB Article 2010877 addresses this
condition.
Though not represented here, this is also true for the vSAN Data network.
The vSAN Master node and the vSAN Backup node send heartbeats every
second. If communication is lost for 5 consecutive heartbeats (5 seconds)
between the master and the backup due to an issue with the backup node, the
master chooses a different ESXi host as a backup on the remote site. This is
repeated until all hosts on the remote site are checked. If there is a complete
site failure, the master selects a backup node from the “Preferred” site.
When a node rejoins an empty site after a complete site failure, either the
master (in the case of the node joining the primary site) or the backup (in the
case where the node is joining the secondary site) will migrate to that site.
38
6. 4 Bandwidth Calculation
Hosts designated as a vSAN Witness Host do not maintain any VM data, but
rather only component metadata, the requirements are much smaller than that
of the backend vSAN data network.
Virtual Machines on vSAN are comprised of multiple objects, which can potentially be split into multiple
components, depending on factors like policy and size. The number of components on vSAN have a direct
impact on the bandwidth requirement between the data sites and the witness.
The required bandwidth between the vSAN Witness Host and each node is
equal to ~1138 B x Number of Components /5s
The 1138 B value comes from operations that occur when the Preferred Site
goes offline, and the Secondary Site takes ownership of all of the components.
When the Preferred Node goes offline, the Secondary Node becomes the
Master Node. The vSAN Witness Host sends updates to the new Master,
followed by the new Master replying to the vSAN Witness Host as ownership is
updated.
In the event of a Preferred Node failure, the link must be large enough to allow
39
Workload 1
Approximately 25 VMs with the above configuration would require the vSAN
Witness Host to contain 75 components.
With the 10% buffer included, a rule of thumb can be stated that for every
1,000 components, 2 Mbps is appropriate.
Workload 2
40
This section will address using the vSAN Witness Appliance as a vSAN Witness
Host. The vSAN Witness Appliance is available in an OVA (Open Virtual
Appliance) format from VMware. The vSAN Witness Appliance does need to
reside on a physical ESXi host.
The vSAN Witness Appliance must on an ESXi 5.5 or greater VMware host.
The ESXi 5.5 or greater host must meet the minimum requirements for the
vSAN Witness Host for the version of vSAN 2 Node that the vSAN Witness
Appliance will support.
Networking must be in place that allows for the vSAN Witness Appliance
to properly communicate with the vSAN 2 Node Cluster.
On a vSphere environment backed with any supported storage (vmfs datastore, NFS datastore, vSAN
Cluster)
On vCloud Air/OVH backed by supported storage
Any vCloud Air Network partner hosted solution
On a vSphere Hypervisor (free) installation using any supported storage (vmfs datastore or NFS datastore)
41
The vSAN Witness Appliance is supported running on top of another non-Stretched vSAN cluster.
The vSAN Witness Appliance is supported on a Stretched Cluster vSAN for another vSAN 2 Node cluster.
vSAN 2-node cluster hosting witness for another vSAN 2-node cluster witness, and vice versa, is not
recommended and requires an RPQ.
CPU Requirements
The vSAN Witness Appliance is a virtual machine that has special vSphere
installation that is used to provide quorum/tiebreaker services for a 2 Node
vSAN Cluster. The underlying CPU architecture must be supported by the
vSphere installation inside the vSAN Witness Appliance.
Networking
The vSAN Witness Appliance contains two network adapters that are
connected to separate vSphere Standard Switches (VSS).
42
43
Because of this, promiscuous mode is not required when using a vSAN Witness
Appliance.
Since the vSAN Witness Appliance will be deployed on a physical ESXi host the
underlying physical ESXi host will need to have a minimum of one VM network
preconfigured. This VM network will need to reach both the management
network and the vSAN network shared by the ESXi hosts on the data sites. An
alternative option that might be simpler to implement is to have two
preconfigured VM networks on the underlying physical ESXi host, one for the
management network and one for the vSAN network. When the virtual ESXi
witness is deployed on this physical ESXi host, the network will need to be
attached/configured accordingly.
44
When using a physical ESXi host as a vSAN Witness Host, the VMkernel
interface that will be tagged for "vsan" traffic must have connectivity to the
vSAN Data Node VMkernel interace that is tagged for "witness" traffic. This
could be over Layer 3, or even over Layer 2 if desired.
If using a physical host as the vSAN Witness Host there are some requirements
to consider.
Requirement
s
45
Required
46
Management traffic for the data nodes is typically automatically routed to the
vCenter server at the central data center. Because they reside in the same
physical location, vSAN traffic networking between data nodes is consistent
with that of a traditional vSAN cluster.
These vSAN Nodes are still required to communicate with the vSAN Witness
Host residing in the central datacenter. Witness Traffic Separation, allows for a
separate VMkernel interface for "witness" traffic. This VMkernel must be able
to communicate with the vSAN Witness Host.
47
In cases were the vSAN Witness Appliance uses the WitnessPG (vmk1) to
communicate with vSAN Data Nodes over Layer 3, static routing will be
required to ensure proper connectivity.
The below illustration shows a vSAN Witness Appliance with the Managment
(vmk0) and WitnessPg (vmk1) VMkernel interfaces on different networks. This
is a typical configuration.
In the illustration above, the vSAN Witness Appliance in the central datacenter
has an Management (vmk0) IP address of 192.168.109.23. This VMkernel
interface will use the default gateway to communicate with vCenter Server.
The WitnessPG VMkernel interface (vmk1) has an IP address of
192.168.110.23.
48
Because vSAN based traffic (tagged as "vsan" or "witness") use the default
gateway, static routes must be used in this configuration. A static route on
Host 1 and Host 2 for their vmk1 VMkernel interfaces to properly connect to
the WitnessPg VMkernel interface on the vSAN Witness Appliance. The
routing command on Host 1 and Host 2 would look like this:
*Note the address of x.x.x.1 is the gateway in each of these networks for this
example. This may differ from your environment.
The below illustration shows a vSAN Witness Appliance with the Managment
(vmk0) and WitnessPg (vmk1) VMkernel interfaces on the same network. This
is a not a typical configuration, but is supported if configured properly.
49
Notice that the WitnessPg (vmk1) VMkernel interface is NOT tagged with
"vsan" traffic, but rather the Management (vmk0) VMkernel interface has both
"Management" and "vsan" traffic tagged, In cases were vmk0 and vmk1 would
happen to reside on the same network, a multi-homing issue will occur if the
WitnessPg (vmk1) continues to have "vsan" traffic tagged.
The correct configuration in this situation, would be untag "vsan" traffic from
the WitnessPg (vmk1) VMkernel interface, and tag the Management (vmk0)
interface instead. This is also a supported configuration.
*Some concerns have been raised in the past about exposing the vSAN Data
Network to the WAN in 2 Node vSAN Clusters. Witness Traffic Separation
mitigates these concerns because only vSAN Obect metadata traverses the
WAN to the vSAN Witness Host.
50
Other customers have asked "why?" A 3rd host running locally could
potentially be a lesser capable host that has enough resources to run a minimal
workload that includes the vSAN Witness Host role, possibly local backups, as
well as other resources such as networking based virtual machines.
In the below illustration, the 2 Node vSAN Cluster Hosts, Host 1 and Host 2, are
both connected to the same Management network as the physical host that is
running the vSAN Witness Appliance.
51
interface tagged for "witness" traffic on the 192.168.15.x network. The vSAN
Witness Appliance has the WitnessPg (vmk1) VMkernel interface tagged for
"vsan" traffic, also on the 192.168.15.x network. In this configuration, static
routing is not required because Layer 2 networking is use.
In the below illustration, the 2 Node vSAN Cluster Hosts, Host 1 and Host 2, are
both connected to the same Management network as the physical host that is
running the vSAN Witness Appliance.
52
Compute Requirements
Memory Requirements
Storage Requirements
Cache Device Size: Each vSAN Witness Appliance deployment option has a
cache device size of 10GB. This is sufficient for each for the maximum of
45,000 components. In a typical vSAN deployment, the cache device must be a
Flash/SSD device. Because the vSAN Witness Appliance has virtual disks, the
10GB cache device is configured as a virtual SSD. There is no requirement for
this device to reside on a physical flash/SSD device. Traditional spinning drives
are sufficient.
Capacity Device Sizing: First consider that a capacity device can support up to
21,000 components. Also consider that a vSAN 2 Node Cluster can support a
maximum of 27,000 components. Each Witness Component is 16MB, as a
result, the largest capacity device that can be used for storing of Witness
Components is approaching 350GB.
53
A vSAN Witness Appliance is provided with each release of vSAN. Upon initial
deployment of the vSAN Witness Appliance, it is required to be the same as the
version of vSAN. The vSphere host that the vSAN Witness Appliance runs on, is
not required to be the same version.
54
When upgrading the vSAN Cluster, upgrade the vSAN Witness Appliance in the
same fashion as upgrading vSphere . This keeps the versions aligned. Be
certain to ensure that the underlying hosts and the vCenter Server version
supports the version of vSAN being upgraded to.
55
56
vSphere HA Turn on
Advanced Settings:
das.usedefaultisolationaddress False
57
7. 2 Turn on vSphere HA
To turn on vSphere HA, select the cluster object in the vCenter inventory,
Manage, then vSphere HA. From here, vSphere HA can be turned on and off
via a check box.
7. 3 Host Monitoring
58
7. 4 Admission Control
59
60
What do Heartbeat Datastores do, and when does it come in to play? The
heartbeat datastore is used by a host which is isolated to inform the rest of the
cluster what its state is and what the state of the VMs is. When a host is
61
62
7. 8 Advanced Options
63
If a host is not receiving any heartbeats, it uses a fail safe mechanism to detect
if it is merely isolated from its HA master node or completely isolated from the
network. It does this by pinging the default gateway.
When using vSAN 2 Node clusters in the same location, there is no need to
have a separate das.isolationaddress for each of the hosts.
64
65
With vSphere DRS enabled on the cluster, the virtual machines can simply be
deployed to the cluster, and then the virtual machine is powered on, DRS will
move the virtual machines to either host based on utilization.
With fully automated mode, DRS will take care of the initial placement and on-
going load balancing of virtual machines. DRS should balance virtual machines
across the hosts.
66
67
The first step is to download and deploy the vSAN Witness Appliance, or
deploy it directly via a URL, as shown below. In this example it has been
downloaded:
68
69
70
71
At this point a decision needs to be made regarding the expected size of the 2
Node vSAN Cluster configuration. There are three options offered. If you
expect the number of VMs deployed on the vSAN 2 Node Cluster to be 10 or
fewer, select the Tiny configuration. If you expect to deploy more than 10 VMs,
but less than 500 VMs, then the Normal (default option) should be chosen. On
selecting a particular configuration, the resources consumed by the appliance
and displayed in the wizard (CPU, Memory and Disk):
72
Select a datastore for the vSAN Witness Appliance. This will be one of the
datastore available to the underlying physical host. You should consider when
the vSAN Witness Appliance is deployed as thick or thin, as thin VMs may grow
over time, so ensure there is enough capacity on the selected datastore.
73
Select a network for the Management Network and for the Witness Network.
74
75
At this point, the vSAN Witness Appliance is ready to be deployed. It will need
to be powered on manually via the vSphere web client UI later:
76
Once the vSAN Witness Appliance is deployed and powered on, select it in the
vSphere web client UI and begin the next steps in the configuration process.
Once the vSAN Witness Appliance has been deployed, select it in the vSphere
web client UI, open the console.
The console of the vSAN Witness Appliance should be access to add the
correct networking information, such as IP address and DNS, for the
management network.
On launching the console, unless you have a DHCP server on the management
network, it is very likely that the landing page of the DCUI will look something
similar to the following:
77
Use the <F2> key to customize the system. The root login and password will
need to be provided at this point. This is the root password that was added
during the OVA deployment earlier.
Select the Network Adapters view. There will be two network adapters, each
corresponding to the network adapters on the virtual machine. You should
note that the MAC address of the network adapters from the DCUI view match
the MAC address of the network adapters from the virtual machine view.
Because these match, there is no need to use promiscuous mode on the
network, as discussed earlier.
Select vmnic0, and if you wish to view further information, select the key <D>
to see more details.
78
Navigate to the IPv4 Configuration section. This will be using DHCP by default.
Select the static option as shown below and add the appropriate IP address,
subnet mask and default gateway for this vSAN Witness Appliance
Management Network.
79
The next step is to configure DNS. A primary DNS server should be added and
an optional alternate DNS server can also be added. The FQDN, fully qualified
domain name, of the host should also be added at this point.
80
When all the tests have passed, and the FQDN is resolvable, administrators can
move onto the next step of the configuration, which is adding the vSAN
Witness Appliance ESXi instance to the vCenter server.
81
Provide the appropriate credentials. In this example, the root user and
password.
82
83
84
The vSAN Witness Appliance also comes with its own license. You do not need
to consume vSphere licenses for the witness appliance:
85
86
Click Finish when ready to complete the addition of the Witness to the vCenter
server:
87
One final item of note is the appearance of the vSAN Witness Appliance ESXi
instance in the vCenter inventory. It has a light blue shading, to differentiate it
from standard ESXi hosts. It might be a little difficult to see in the screen shot
below, but should be clearly visible in your infrastructure. (Note: In vSAN 6.1
and 6. 2 deployments, the “No datastores have been configured” message is
because the nested ESXi host has no VMFS datastore. This can be ignored.)
88
One final recommendation is to verify that the settings of the vSAN Witness
Appliance matches the Tiny, Normal or Large configuration selected during
deployment. For example, the Normal deployment should have an 12GB HDD
for boot in vSAN 6.5 (8GB for vSAN 6.1/6.2), a 10GB Flash that will be
configured later on as a cache device and another 350 HDD that will also be
configured later on as a capacity device.
89
Once confirmed, you can proceed to the next step of configuring the vSAN
network for the vSAN Witness Appliance.
The next step is to configure the vSAN network correctly on the vSAN Witness
Appliance. When the Witness is selected, navigate to Configure > Networking >
Virtual switches as shown below.
90
The Witness has a portgroup pre-defined called witnessPg. Here the VMkernel
port to be used for vSAN traffic is visible. If there is no DHCP server on the
vSAN network (which is likely), then the VMkernel adapter will not have a valid
IP address. Select VMkernel adapters > vmk1 to view the properties of the
witnessPg. Validate that "vSAN" is an enabled service as depicted below.
91
92
Select the witnessPg and edit the properties by selecting the pencil icon.
If vSAN is not an enabled service, select the witnessPg portgroup, and then
select the option to edit it. Tag the VMkernel port for vSAN traffic, as shown
below:
Next, ensure the MTU is set to the same value as the vSAN Data Node hosts’
vSAN VMkernel interface.
93
In the IPV4 settings, a default IP address has been allocated. Modify it for the
vSAN traffic network.
Static routes are still required by the witnessPg VMkernel interface (vmk1) as in
vSAN 6.1 or 6.2. The "Override default gateway for this adapter" setting is not
supported for the witness VMkernel interface (vmk1).
Once the witnessPg VMkernel interface address has been configured, click OK.
94
95
10. 1 Pre-Requisites
Pre-requisites:
Ensure connectivity between the 2 Node cluster and the vSAN Witness
Host
If using Witness Traffic Separation - A VMkernel interface other than
the vSAN Network will be required
If not using Witness Traffic Separation - The vSAN Network will be
required to have connectivity to the vSAN Witness Host
Configure appropriate networking
For vCenter
vCenter must be able to communicate with the Management
interface on each vSAN Host
vCenter must be able to communicate with the Management
interface for the vSAN Witness Host
For vSAN Hosts
The vSAN Host Management interface must be able to
communicate with vCenter
vSAN Hosts must be able to communicate with each other
vSAN Hosts must be able to communicate with the vSAN Witness
Host vSAN Tagged interface
For vSAN Witness Host
The vSAN Witness Host Management interface must be able to
communicate with vCenter
The vSAN Witness Host vSAN Tagged interface must be able to
communicate with the vSAN Nodes
96
VMkernel Interfaces
VMkernel interfaces provide connectivity for different inter-node
communication between vSpehre hosts. IP-based storage protocols typically
have dedicated VMkernel interfaces on isolated networks. It is considered a
best practice and VMware recommendation to isolate storage traffic. vSAN
may only use VMkernel interfaces that are appropriate tagged for vSAN traffic
types.
Witness Traffic Separation was publicly introduced with vSAN 6.5, but is also
included in vSAN 6.2 as of vSphere 6.0 Update 3 (vCenter and vSphere
versions). Witness Traffic Separation removes the requirement for traffic to
and from the vSAN Witness Host to be available from the vSAN data network.
A "front-end facing" VMkernel interface may be used, allowing the "back-end"
of 2 Node vSAN to use VMkernel ports that are directly connected across
hosts. The back-end vSAN network can still be connected to a switch if desired.
VMkernel Interfaces
97
Witness Traffic Separation was publicly introduced with vSAN 6.5, but is also
included in vSAN 6.2 as of vSphere 6.0 Update 3 (vCenter and vSphere
versions). Witness Traffic Separation removes the requirement for traffic to
and from the vSAN Witness Host to be available from the vSAN data network.
A "front-end facing" VMkernel interface may be used, allowing the "back-end"
of 2 Node vSAN to use VMkernel ports that are directly connected across
hosts. The back-end vSAN network can still be connected to a switch if desired.
98
Data nodes can use a direct connection between nodes or may be connected
to a traditional switch, which could be part of an isolated data network.
Communication with the vSAN Witness Host is configured separately from the
back-end vSAN Data network.
99
3. Give the VMkernel interface a descriptive label and be certain to select the
appropriate VLAN if necessary.
100
101
*Note: the Default gateway may not be set. vSAN uses the same TCP/IP
stack as the Management interface and may not use an alternate gateway.
It is important to also ensure that this VMkernel interface is NOT on the
same TCP/IP network as the Management interface.
102
103
2. Use the vSphere CLI to connect to the vSAN Host and execute the
following command:
3. Execute the following single line PowerCLI script against the vSAN
Host:
104
To add a static route on the vSAN data nodes to the vSAN Witness Host's
"vSAN Tagged" VMkernel interface, perform one of the following methods:
1. Enable SSH on the vSAN Host, login, and execute the following
command:
2. Use the vSphere CLI to connect to the vSAN Host and execute the
following command:
3. Execute the following single line PowerCLI script against the vSAN
Host:
105
10. Confirm the Witness Tagged VMkernel interface can ping the vSAN
Witness Host vSAN Traffic tagged VMkernel interface:
1. Enable SSH on the vSAN Host, login, and execute the following
command:
2. Execute the following single line PowerCLI script against the vSAN
Host:
Once Witness Traffic Separation networking has been configured on the first
host, repeat the process for the second host.
*Important things to consider/remember:
The VMkernel interface for Witness Traffic may not be on the same
network as the Management interface (vmk0). If Witness Traffic must be
run on the same network as vmk0, simply skip steps 1-6 and tag vmk0 in
step 7 instead.
Witness Traffic from data nodes to the vSAN Witness Host contain no
106
virtual machine data, but rather vSAN metadata, such as vSAN component
placement, vSAN node ownership, etc. Tagging vmk0 for Witness Traffic is
fully supported.
1. Create a new VMkernel port for use as a vSAN Traffic interface on the first
host.
107
108
109
4. Select one or more adapters that will be used for vSAN Traffic and select
OK.
110
111
6. Give the VMkernel interface a descriptive label and be certain to select the
appropriate VLAN if necessary.
112
*Note: the Default gateway may not be set. vSAN uses the same TCP/IP
stack as the Management interface and may not use an alternate gateway.
It is important to also ensure that this VMkernel interface is NOT on the
same TCP/IP network as the Management interface.
113
114
Notice vmk2 has vSAN Traffic Tagged and vmk1 has Witness Traffic
Tagged.
115
1. Create a new VMkernel port for use as a vSAN Traffic interface on the first
host.
116
3. Select browse and choose the vSAN Port Group on the VDS and select OK.
117
4. Select Next.
118
5. Select the appropriate VLAN if necessary and check vSAN in the Available
services. This is "tagging" this VMkernel interface for vSAN Traffic.
119
*Note: the Default gateway may not be set. vSAN uses the same TCP/IP
stack as the Management interface and may not use an alternate gateway.
It is important to also ensure that this VMkernel interface is NOT on the
same TCP/IP network as the Management interface.
120
121
Notice vmk2 has vSAN Traffic Tagged and vmk1 has Witness Traffic
Tagged.
10. Repeat this process for the vSAN and vMotion Network for the second
host.
The following steps should be followed to install a new 2 Node vSAN Cluster.
122
To setup 2 Node vSAN, he Manage > vSAN > Services . Click Configure to
begin the vSAN wizard.
The initial wizard allows for choosing various options like enabling
Deduplication and Compression (All-Flash architectures only with Advanced or
greater licensing) or Encryption (Enterprise licensing required) for vSAN. Select
Two host vSAN Cluster and Next.
123
Select the services desired that are compatible with the hardware and licensing
options for the cluster.
124
In 2 Node vSAN it is important to get into the habit of using Allow Reduced
Redundancy. This is because any time there is an requirement to
create/remove/recreate a vSAN Disk Group on a 2 Node vSAN Cluster, the
Allow Reduced Redundancy option is required. Without this setting checked,
changes will likely not occur due to the cluster not being able to maintain
storage policy compliance.
Disks should be selected for their appropriate role in the vSAN cluster.
125
Click Next .
The Create fault domain wizard will allow hosts to be selected for either the
Preferred or Secondary fault domain. The default naming of these two fault
domains are Preferred and Secondary.
126
Select one host and choose >> to move it to the Secondary fault domain.
Click Next .
The vSAN Witness host detailed earlier must be selected to act as the vSAN
Witness for the 2 Node vSAN Cluster.
127
Click Next .
Just like physical vSAN hosts, the vSAN Witness needs a cache tier and a
capacity tier. * Note: The vSAN Witness does not actually require SSD backing
and may reside on a traditional mechanical drive.
128
Be certain to select the 10GB device as the cache device and the 15GB device
as a capacity device.
Select Next .
Review the vSAN 2 Node Cluster configuration for accuracy and click Finish.
129
Log in to the VAM I interface and select Update from the Navigator pane to
begin the upgrade process.
130
After the upgrade has completed, the VCSA will have to be rebooted for the
updates to be applied. It is important to remember that vSAN Health Check will
not be available until after hosts have been upgraded.
Upgrading each host is the next task to be completed. There are a few
considerations to remember when performing these steps.
131
ensure reads are always distributed across hosts, which will mitigate the
warming process required when a virtual machine moves to the alternate host.
With vSphere DRS in place, will ensure that virtual machines are moved to the
alternate host. If DRS is set to “fully automated” virtual machines will vMotion
to the other host automatically, while “partially automated” or “manual” will
require the virtualization admin to vMotion the virtual machines to the other
host manually.
After both vSAN Data Nodes have been upgraded, the vSAN Witness Host will
also need to be upgraded. The Health Check will show that the vSAN Witness
Host has not been upgraded.
Upgrading the vSAN Witness Host is done in the same way that any other ESXi
hosts are updated. It is important to remember that as the vSAN Witness Host
is upgraded, the witness components will no longer be available and objects
will be noncompliant. Objects will report that the Witness component is not
found.
132
After the upgrade is complete, the Witness component will return, and will
reside on the vSAN Witness Host.
The final step in the upgrade process, will be to upgrade the on-disk format. To use the new features, from time
to time the on-disk format must be upgraded to the required version. The Health Check will assume that the
vSphere version will prefer a native on-disk format for that version, and as a result, it will throw an error until the
format is upgraded.
133
To upgrade the on-disk format from an older version to a newer version, select
M anage > General under vSAN. Then click Upgrade under the On-disk Format
Version section.
The depending on the on-disk format version an upgrade will either perform a
rolling upgrade across the hosts in the cluster, or make a small metadata
update. In the event a rolling upgrade is required, each host’s disk groups will
be removed and recreated with the new on-disk format. The amount of time
required to complete the on-disk upgrade will vary based on the cluster
hardware configuration, how much data is on the cluster, and whatever over
134
disk operations are occurring on the cluster. The on-disk rolling upgrade
process does not interrupt virtual machine disk access and is performed
automatically across the hosts in the cluster.
The witness components residing on the vSAN Witness Host will be deleted
and recreated. This process is relatively quick given the size of witness objects.
Basic Workflow
The process of converting a 2 Node Cluster to a Cluster with 3 or more hosts is
135
as follows:
These steps will ensure availability of vSAN data and virtual machines when
migrating to a switch for vSAN traffic.
1. Migrate any virtual machines to the host that is in the Fault Domain that is
designated as "Preferred".
1. This Fault Domain is often named "Preferred" but could have been
given a different name during configuration.
2. The asterisk in the Fault Domains UI will indicate which host is
136
"Preferred"
2. When all workloads have been migrated to the preferred host, place the
other host in Maintenance Mode, choosing Ensure Accessibility
3. Disconnect the direct connected uplink(s) from the alternate host (non-
preferred) and connect it to an appropriate switch
1. This will connect the preferred host to the switch
4. Connect the non-preferred node to the switch, just as the preferred node
is connected
1. Confirm connectivity between the preferred and non-preferred nodes
2. On each node, from the ESXi shell, use vmkping to confirm
connectivity
1. vmkping -I vmX (vSAN tagged VMkernel interface) <IP Address of
alternate host vSAN tagged VMkernel>
2. vmkping -I vmX (vMotion tagged VMkernel interface) <IP Address
of alternate host vMotion tagged VMkernel>
5. When connectivity is confirmed, exit maintenance mode on the non-
preferred node
137
138
139
140
2. Disk Groups can be added using the Add Disk Group option in
Disk Management
141
Below is a narrated video that goes through the process for a 2 Node vSAN 6.7
Update 1 cluster:
142
Summary
The process to convert a 2 Node vSAN Cluster to a 3 (or more) Node vSAN
Cluster is relatively easy, but does require a particular steps to be taken.
143
144
In any situation where a 2 Node vSAN cluster has an inaccessible host or disk group, vSAN objects are at risk of
becoming inaccessible should another failure occur.
When a host fails, vSAN cannot rebuild data on another host to protect against another failure.
If a host must enter maintenance mode, vSAN cannot evacuate data from the host to maintain policy
compliance. While the host is in maintenance mode, data is exposed to a potential failure or inaccessibility
should an additional failure occur.
When placing a vSAN host in maintenance mode, the vSAN data migration options are:
o Full data migration – Not available for two-node vSAN clusters using the default storage policy, as policy
compliance requires two for data and one for the witness object
o Ensure accessibility – The preferred option for two-host or three-host vSAN clusters using the default
storage policy. Ensure accessibility guarantees the enough components of the vSAN object are available for
the object to remain available. Though still accessible, vSAN objects on two-host or three-host clusters are no
longer policy
compliant. When the host is no longer in maintenance mode, objects will be rebuilt to ensure policy
compliance. During this time however, vSAN objects are at risk because they will become inaccessible if
another failure occurs.
Any objects that have a non-standard single copy storage policy (FTT=0) will be moved to an available host in
the cluster. If there is not sufficient capacity on any alternate hosts in the cluster, the host will not enter
maintenance mode.
145
o No data migration – This is not a recommended option for vSAN clusters. vSAN objects that use the default
vSAN Storage Policy may continue to be accessible, but vSAN does not ensure their accessibility.
Any objects that have a non-standard single copy storage policy (FTT=0) will become inaccessible until the
host exits maintenance mode.
o vSAN Witness Appliance (recommended) – No virtual machine workloads may run here. The only purpose
of this appliance is to provide quorum for the 2 Node vSAN Cluster. Maintenance mode should be brief,
typically associated with updates and patching.
o Physical ESXi Host – While not typical, a physical host may be used as a vSAN Witness Host. This
configuration will support virtual machine workloads. Keep in mind that a vSAN Witness Host may not be a
member of any vSphere cluster, and as a result virtual machines will have to be manually moved to an
alternate host for them to continue to be available.
146
147
The virtual machine's virtual disk (vmdk) has one component placed in the
Preferred Host, one component placed in the Secondary Host, and a Witness
component in the Witness Site that houses the vSAN Witness Host.
148
The illustration shows a storage policy that protect vSAN Objects across sites.
This is the default protection policy used with vSAN 2 Node Clusters for
versions 6.1 through 6.7.
149
Fails?
vSAN will mark the component degraded and the VM will continue to
run.
Reads are immediately serviced from the alternate node
The component will be rebuilt within the same host if there is an
additional drive or disk group available.
If there are no additional drives or disk groups available in the host, the
component will only be rebuilt if the failed/isolated device comes back
online.
In cases where the component cannot be rebuilt, reads will continue
to be serviced from the alternate host.
150
Goes Offline?
vSAN will mark the component absent and the VM will continue to run.
151
152
In the event the Preferred Host fails or is partitioned, vSAN powers the virtual
machines running in that Host off. The reason for this, is because the virtual
machine's components are not accessible due to the loss of quorum. The vSAN
2 Node has now experienced a single Host failure. The loss of either Host in
addition to the vSAN Witness is two failures, will take the entire cluster offline.
153
An the Secondary Node will be elected as the HA master, which will validate
which virtual machines are to be powered on. Because quorum has been
formed between the vSAN Witness Host and the Secondary Host, virtual
machines on the Secondary Host will have access to their data, and can be
powered on.
154
In the event the Secondary Host fails or is partitioned, vSAN powers the virtual
machines running in that host off. The reason for this, is because the virtual
machine's components are not accessible due to the loss of quorum. The vSAN
2 Node Cluster has now experienced a single site failure. The loss of either
node in addition to the witness is two failures, will take the entire cluster
offline.
155
The Preferred Host, will validate which virtual machines are to be powered on.
Virtual machines which have been moved to the Preferred Host will now have
access to their data, and can be powered on.
156
Virtual machines running in both nods of a 2 Node Cluster are not impacted by
the vSAN Witness Host being partitioned. Virtual machines continue to run.
The vSAN 2 Node Cluster has now experienced a single failure. The loss of
either host in addition to the witness is two failures, will take the entire cluster
offline.
157
In the event the vSAN Witness Host has failed, the behavior is the same as if
the vSAN Witness Host has been partitioned from the cluster. Virtual machines
continue to run at both locations. Because the 2 Node vSAN Cluster has now
experienced a single site failure, it is important to either get the vSAN Witness
Host back online, or deploy a new one for the cluster.
158
When the existing vSAN Witness Host comes back online, metadata changes
are resynchronized between the 2 Node vSAN Cluster sites and the vSAN
Witness Host.
159
The amount of data that needs to be transmitted depends on a few items such
as the number of objects and the number of changes that occurred while the
vSAN Witness Host was offline. However, this amount of data is relatively small
considering it is metadata, not large objects such as virtual disks.
If there is a failure in the cluster, i.e. one of the hosts is down; new virtual
160
machines can still be provisioned. The provisioning wizard will however warn
the administrator that the virtual machine does not match its policy as follows:
In this case, when one node is offline and there is a need to provision virtual
machines, the ForceProvision capability is used to provision the VM. This
means that the virtual machine is provisioned with a
NumberOfFailuresToTolerate = 0, meaning that there is no redundancy.
Administrators will need to rectify the issues on the failing site and bring it
back online. When this is done, vSAN will automatically update the virtual
machine configuration to NumberOfFailuresToTolerate = 1, creating a second
copy of the data and any required witness components.
The vSAN Design and Sizing Guide goes into further detail about how
component availability determines access to objects. In short, each component
has a vote, and a quorum of votes must be present for an object to be
accessible. Each site will have an equal number of votes and there will be an
even distribution of votes within a site. If the total number of votes is an even
number, a random vote will be added.
In the illustration below, a 2 Node vSAN Cluster (4+4+1) has an object mirrored
across hosts.
161
162
If the Secondary Node, or the link between the Nodes, were to also fail, the
vSAN Objects would not be accessible
163
The proper mechanism to recover would be to bring the Secondary Node back
online and ensure it can communicate with the Preferred Node. With the
Preferred and Secondary Nodes online, vSAN Objects will be accessible, but
will not have policy compliance.
164
The vSAN Witness Host may be brought back online, or a new vSAN Witness
Host may be deployed.
When either the vSAN Witness Host or a replacement vSAN Witness Host is
brought online, objects may be repaired through the vSAN Health Check UI to
achieve policy compliance.
165
Should a vSAN Witness Host fail in the vSAN 2 Node Cluster, a new vSAN
Witness Host can easily be introduced to the configuration.
Replacing the vSAN Witness Host in vSAN 6.7 or higher using the vSphere Client
Navigate to Cluster > Configure > vSAN > Fault Domains & Stretched Cluster
166
Use the Change Witness Host Wizard in the same fashion as adding an initial
vSAN Witness Host.
167
Select the 10GB disk for the cache device and the 15GB for capacity
168
Replacing the vSAN Witness Host in vSAN 6.6 or higher using the vSphere Web Client
Navigate to Cluster > Configure > vSAN > Fault Domains & Stretched Cluster
169
Use the Change Witness Host Wizard in the same fashion as adding an initial
vSAN Witness Host.
Select the 10GB disk for the cache device and the 15GB for capacity
170
171
The failing witness host can be removed from the vSAN Stretched Cluster via
the UI (red X in fault domains view).
The next step is to rebuild the vSAN stretched and selecting the new witness
host. In the same view, click on the “configure stretched cluster” icon. Align
hosts to the preferred and secondary sites as before. This is quite simple to do
since the hosts are still in the original fault domain, so simply select the
secondary fault domain and move all the hosts over in a single click:
172
Create the disk group and complete the vSAN Stretched Cluster creation.
On completion, verify that the health check failures have resolved. Note that
the vSAN Object health test will continue to fail as the witness component of
VM still remains “Absent”. When CLOMD (Cluster Level Object Manager
Daemon) timer expires after a default of 60 minutes, witness components will
173
be rebuilt on new witness host. Rerun the health check tests and they should
all pass at this point, and all witness components should show as active.
Impact/Observed
VMware HA
Scenario vSAN Behaviour
Behaviour
174
Impact/Observed
VMware HA
Scenario vSAN Behaviour
Behaviour
If the VM was
running on the same
host as the failure a
HA restart of the VM
will take place if HA
is enabled.
vSAN Witness Host loss / vSAN Witness Host loss VM will continue
failed / isolated from one counts as a site (PFTT) running.
or both nodes failure, as such the
cluster will be placed in a
degraded state until the
witness comes back
online or is redeployed.
175
Impact/Observed
VMware HA
Scenario vSAN Behaviour
Behaviour
13. Appendix A
176
The vSAN Witness Appliance OVA is located on the Drivers & Tools tab of each
vSAN edition's download page.
There you will find a section called VMware vSAN Tools, Plug-ins, and
Appliances.
177
14. Appendix B
New ESXCLI commands for vSAN Stretched Cluster.
178
Available Commands:
get Get the preferred fault domain for a stretched cluster
.
set Set the preferred fault domain for a stretched cluster
.
Display, add, modify, remove vSAN VMkernel interfaces for use by vSAN
Available Namespaces:
ip Commands for configuring IP network fo
r vSAN.
ipv4 Compatibility alias for "ip"
Available Commands:
clear Clear the vSAN network configuration.
list List the network configuration current
ly in use by vSAN.
remove Remove an interface from the vSAN netw
179
ork configuration.
restore Restore the persisted vSAN network con
figuration.
Interface
VmkNic Name: vmk1
IP Protocol: IP
Interface UUID: 1f825f5b-5018-f75c-a99d-001b2193c268
Agent Group Multicast Address: 224.2.3.4
Agent Group IPv6 Multicast Address: ff19::2:3:4
Agent Group Multicast Port: 23451
Master Group Multicast Address: 224.1.2.3
180
[root@host1:~]
esxcfg-advcfg
[root@host2:~] esxcfg-advcfg
This usage
Usage: esxcfg-advcfg <options> [<adv cfg Path>]
-g|--get Get the value of the VMkernel advan
ced
configuration option
-s|--set <value> Set the value of the VMkernel advan
ced
configuration option
Get the Sparse Swap setting (Introduced in vSphere 6.0 Update 2, enabled by
default in vSphere 6.5 Update 2 or higher)
181
ed
Value of DOMOwnerForceWarmCache is 0
Set the Sparse Swap setting (Introduced in vSphere 6.0 Update 2, enabled by
default in vSphere 6.5 Update 2 or higher)
vsan.stretchedcluster.remove_witness
Remove a witness host. The name of the cluster must be provided as an
argument to the command.
vsan.stretchedcluster.witness_info
Display information about a witness host. Takes a cluster as an argument.
182
183