NetApp - MetroCluster (Design and Implement)
NetApp - MetroCluster (Design and Implement)
NetApp - MetroCluster (Design and Implement)
ABSTRACT
This document is intended to serve as a guide for architecting and deploying MetroCluster in a customer environment. This design and implementation guide describes the basic MetroCluster architecture, considerations for deployment, and best practices. As always, please refer to the latest technical publications on the NOW (NetApp on the Web) site for specific updates on processes, Data ONTAP command syntax, and the latest requirements, issues, and limitations. This document is intended for field personnel who require assistance in architecting and deploying a MetroCluster solution.
Table of Contents
1 2 3 4 5 6 7
INTRODUCTION ......................................................................................................................... 3 OVERVIEW ................................................................................................................................. 3 DEPLOYMENT PLANNING ..................................................................................................... 12 INSTALLATION AND CONFIGURATION................................................................................ 15 SITE FAILOVER AND RECOVERY ......................................................................................... 26 RELEVANT DOCUMENTATION .............................................................................................. 29 APPENDICES ........................................................................................................................... 30
1
1.1
INTRODUCTION
Intended Audience
The information in this document is intended for field personnel and administrators who are responsible for architecting and deploying successful MetroCluster high-availability and disaster-recovery configurations. A brief overview of MetroCluster is presented in order to establish baseline knowledge before discussing implementation planning, installation, configuration, and operation.
1.2
Scope
Stretched or non-switched MetroCluster (NetApp storage) Fabric or switched MetroCluster (NetApp storage)
Topics that apply only to stretch MetroCluster refer to stretch MetroCluster. Topics that are specific to NetApp storage-based fabric MetroCluster refer to Fabric MetroCluster. Topics that apply to all configurations refer simply to MetroCluster. Other than a short description, V-Series MetroCluster is not covered in this document. This document refers mostly to Fabric MetroCluster, because Fabric MetroCluster configurations introduce an increased level of complexity in design and implementation. To many, Stretch MetroCluster is simply an active-active configuration with longer cables and mirroring. The introduction of Fibre Channel switches and longer distances requires much more consideration and discussion. The operational sections (creating mirrors, forced takeovers, etc.) apply to both.
1.3
So that the methods and procedures described in this document are useful to the reader, we made the following assumptions: The reader has at least basic NetApp administration skills and has administrative access to the storage system via the command-line interface. The reader has a full understanding of clustering as it applies to the NetApp storage controller environment. The reader has at least a basic understanding of Fibre Channel switch technology and operation, along with access to the switches via command line. In the examples in this report, all administrative commands are performed at the storage system or Fibre Channel switch command line.
2
2.1
OVERVIEW
Features
MetroCluster configurations consist of a pair of active-active storage controllers configured with mirrored aggregates and extended distance capabilities to create a high-availability disaster recovery solution. The primary benefits include: Greater high availability and geographic disaster recovery protection Minimal risk of lost data, easier disaster recovery, and reduced system downtime Quicker recovery when disaster occurs Minimal disruption to users and client applications
2.2
Operation
A MetroCluster (either Stretch or Fabric) behaves in most ways just like an active-active configuration. All of the protection provided by core NetApp technology (RAID-DP, Snapshot copies, automatic controller
failover) also exists in a MetroCluster (Figure 1). However, MetroCluster adds complete synchronous mirroring along with the ability to perform a complete site failover from a storage perspective with a single command.
2.3
MetroCluster Types
Stretch MetroCluster (sometimes referred to as non-switched) is simply an active-active configuration that can extend up to 500m depending on speed and cable type. It also includes synchronous mirroring (SyncMirror) and the ability to perform a site failover with a single command. (See Figure 2.) Additional resiliency can be provided through the use of multipathing.
Stretch MetroCluster
Total Distance 500m @2Gbps (270m @4Gbps)
Primary
NetApp
FAS 3020
Secondary
NetApp
CI
FAS 3020
activity
status
power
activity
status
power
DS14
MK2
Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance
DS14
MK2
Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance
AT
P ower
AT
P ower
Fault Fault
Module A
Module B
Module A
S ys tem
Module B
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
13
12
11
10
09
08
07
06
05
04
03
02
01
00
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
13
12
11
10
09
08
07
06
05
04
03
02
01
00
Primary FC Storage
DS14
MK2
Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance
Secondary FC Storage
DS14
MK2
Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance
AT
P ower
AT
P ower
Shelf ID Shelf ID
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
13
12
11
10
09
08
07
06
05
04
03
02
01
00
13
12
11
10
09
08
07
06
05
04
03
02
01
00
Secondary FC Mirror
Figure 2) Stretch MetroCluster.
Primary FC Mirror
Fabric MetroCluster (also referred to as switched) uses four Fibre Channel switches in a dual fabric configuration and a separate cluster interconnect card to achieve an even greater distance (up to 100km depending on speed and cable type) between primary and secondary locations. (See Figure 3.)
MetroCluster
Total Distance 100Km
Primary Node
FAS6030
Secondary Node
FAS6030
NetApp
NetApp
CI
DS14
MK2
Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance
AT
P ower
Network Appliance
Network Appliance
Network Appliance
Network Appliance
Network Appliance
Network Appliance
Brocade 200E
Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance
DS14
MK2
Network Appliance
AT
P ower
Shelf ID Shelf ID
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
13
12
11
10
09
08
07
06
05
04
03
02
01
00
13
12
11
10
09
08
07
06
05
04
03
02
01
00
Primary FC Storage
DS14
MK2
Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance
Secondary FC Storage
DS14
MK2
Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance Network Appliance
AT
P ower
AT
P ower
Shelf ID Shelf ID
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
212A
13
12
11
10
09
08
07
06
05
04
03
02
01
00
13
12
11
10
09
08
07
06
05
04
03
02
01
00
Secondary FC Mirror
Primary FC Mirror
V-Series MetroCluster is simply either of the above configurations with a NetApp V-Series system. Because of the architectural differences between V-Series and a standard active-active configuration, VSeries MetroCluster has additional flexibility when it comes to the maximum number of disk spindles and the Fibre Channel switches. (See the V-Series Compatibility Guide on the NOW site for additional information.) The implementation of V-Series MetroCluster is outside the scope of this document. (See the V-Series documentation on the NOW site.) For more information about MetroCluster, see the Data ONTAP Cluster Installation and Administration Guide (for Data ONTAP 7.1) or the Data ONTAP Active-Active Configuration Guide (for Data ONTAP 7.2 or later). (For Data ONTAP 7.0 and earlier, see the Data Protection Online Backup and Recovery Guide and the NetApp Hardware Cluster Guide.) Note: When referring to the documentation just listed, use the wiring diagrams in this technical report. The diagrams in this document facilitate easier installation and expansion of MetroCluster.
2.4
Mirroring
NetApp SyncMirror, an integral part of MetroCluster, combines the disk-mirroring protection of RAID 1 with NetApp industry-leading RAID 4 and RAID-DP technology. In the event of an outagewhether it's due to a disk problem, cable break, or host bus adapter (HBA) failureSyncMirror can instantly access the mirrored data without any operator intervention or disruption to client applications. SyncMirror maintains a strict physical separation between the two copies of your mirrored data. Each of these copies is referred to as a plex. Each controllers data has its mirror at the other site.
Aggregate
Located at Partner Site
Mirror
Plex0
Plex1
Pool0
Pool1
When SyncMirror is licensed and hardware ownership is used (discussed in "Disk Ownership"), spare disks are split into two poolsPool0 and Pool1. Each plex of a mirror uses disks from these separate pools. When software ownership is used, disks are explicitly assigned to pools by the administrator. To maximize availability, Pool0 and Pool1 disks need to be on separate loops and use separate HBAs, cables, and shelves. Before enabling the SyncMirror license, ensure that disks for each pool are located on the appropriate loops that are fault isolated from each other. Best Practice: Make sure all storage is mirrored with SyncMirror. While nonmirrored storage is technically permissible, NetApp does not recommend it because the data on that storage will not be available after a site failover.
2.5
Disk Ownership
In a MetroCluster in which disk shelves on each side are mirrored to the other side and thus accessible by either controller, disk ownership comes into play. There are two methods of establishing disk ownership: hardware and software. Hardware-based ownership is the default for the 900 series and the FAS3020/3050. All other platforms (V-Series, FAS6000 series, FAS3040/3070) use software disk ownership. This capability became available in Data ONTAP 6.3.1. Please remember that while the FAS3020/3050 may be technically able to support software disk ownership, it cant be supported in a Fabric MetroCluster. A brief description of each method follows. For more details, see "Installation and Configuration" later in this document or the Data ONTAP documentation. Hardware disk ownership establishes which controller owns which disks by how the shelves are connected. For more information, see the System Configuration Guide:
HTTP://NOW.NETAPP.COM/NOW/KNOWLEDGE/DOCS/HARDWARE/NETAPP/SYSCFG/
FAS3020/3050
Note: Physical connectivity is extremely important to ensure adequate distribution of disks between the two pools. Software disk ownership allows greater flexibility in cabling by allowing disk ownership to be assigned through explicit commands. Rather than determine which controller owns each disk by hardware connection, the ownership information is written on each disk. Because disk ownership can be implemented by a system administrator, it is important to ensure that the configuration maximizes availability. For more information, see "Installation and Configuration" later in this document.
2.6
For those who are experienced with Fibre Channel technology, and storage area networks (SANs) in particular, it is worth discussing some differences and restrictions relative to how Fabric MetroCluster utilizes this technology. Fabric MetroCluster configurations use Fibre Channel switches as the means to separate the controllers by greater distances. The switches are connected between the controller heads and the disk shelves, and to each other. Each disk spindle or LUN individually logs into a Fibre Channel fabric. Except for the V-Series, the nature of this architecture requires, for performance reasons, that the two fabrics be completely dedicated to Fabric MetroCluster. Extensive testing was performed to ensure adequate performance with switches included in a Fabric MetroCluster configuration. For this reason, Fabric MetroCluster requirements prohibit the use of any other model of or any other vendors Fibre Channel switch instead of the Brocade 200E or 5000 included with the Fabric MetroCluster. Also for performance reasons the current disk spindle limit is equal to the lesser of either the model limit or 504. Higher spindle count solutions will be available in the future. Keep in mind that this is for Fabric MetroCluster only. The maximum number of disks in a Stretch MetroCluster depends solely on the NetApp model. In a traditional SAN, there is great flexibility in connecting devices to ports as long as the ports are configured correctly and any zoning requirements are met. In a MetroCluster, when using hardware ownership, Data ONTAP expects certain devices to be connected to specific ports or ranges of ports. It is therefore critical that cabling be exactly as described in the installation procedures. Also, no switch-specific functions such as trunking or zoning are currently used in a hardware ownershipbased NetApp Fabric MetroCluster (non-V-Series).
2.7
Restrictions
For further restrictions in using a Fabric MetroCluster configuration, see the Data ONTAP Cluster Installation and Administration Guide (for Data ONTAP version 7.1) or the Data ONTAP Active-Active Configuration Guide (for Data ONTAP version 7.2 or later). (For Data ONTAP 7.0 and earlier, see the Data Protection Online Backup and Recovery Guide and the NetApp Hardware Cluster Guide.)
2.8
Components
Standard A clustered pair of FAS900, 3000, or 6000 series controllers running Data ONTAP 6.4.1 or later (See the MetroCluster Compatibility Matrix on the NOW site for supported models and Data ONTAP releases.) A syncmirror_local license A cluster_remote license A cluster license Copper/Fibre converters for cluster interconnect (9xx, 30xx, 6xxx only) Associated cabling
A NetApp Fabric MetroCluster includes the following components: A clustered pair of FAS900, 3000, or 6000 series controllers running Data ONTAP 6.4.1 or later (See the MetroCluster Compatibility Matrix on the NOW site for supported models.)
Four Brocade Fibre Channel switches with supported firmware (See the MetroCluster Compatibility Matrix on the NOW site for supported models.) There is a pair at each location. Supported models may differ between locations but must be the same at each given location. Brocade Extended Distance license (if over 10km) Brocade Full-Fabric license Brocade Ports-on-Demand (POD) licenses for additional ports A VI-MC cluster adapter A syncmirror_local license A cluster_remote license A cluster license Associated cabling
2.9
As a disaster recovery solution, MetroCluster is often compared with other synchronous replication products. Even though it includes SyncMirror, MetroCluster can be differentiated by the following features: Low-level RAID mirroring (less performance impact) Site failover with a single command No extensive scripting required to make data available after failover
2.10
OPERATION OVERVIEW
The NetApp Fabric (switched) MetroCluster configuration uses four Brocade Fibre Channel switches in a dual fabric configuration to connect two active-active controllers. These switches cannot be combined with any other switch model and must be supplied by NetApp.
Fabric 1
SW1 SW3
Fabric 2
SW2 SW4
Figure 6 shows a Fabric MetroCluster configuration in which the first fabric begins at switch SW1 on the primary side and is completed by connecting the ISL cable to the switch, SW3, on the remote side. The second fabric is created by using switch SW2 on the primary side, connected through a second ISL cable
to the second switch, SW4, on the remote side. The reason for two fabrics is redundancy. The loss of a switch in a fabric or the loss of a fabric will not affect the availability of the Fabric MetroCluster. Because of the nature of the MetroCluster architecture, Fibre Channel traffic on the switch fabric includes both disk I/O traffic between the controllers and disk shelves and cluster interconnect traffic. As you can probably envision, disk problems can generate excessive traffic and result in a bottleneck for the cluster interconnect. To alleviate this potential condition, several techniques are used.
8-port Switch VC # 2 3 4 5 0, 4 1, 5 2, 6 3, 7
16-port Switch
32-port Switch
Usage
0, 4, 8, 12 1, 5, 9, 13 2, 6, 10, 14 3, 7, 11, 15
0, 4, 8, 12, 16, 20, 24, 28 1, 5, 9, 13, 17, 21, 24, 29 2, 6, 10, 14, 18, 22, 25, 30 3, 7, 11, 15, 19, 23, 26, 31
Consequently, when using software ownership it is important to isolate cluster interconnect (FC/VI) and ISL traffic from storage traffic. When the distance is greater than 10km, all virtual channels are combined into one for buffer credit purposes.
LICENSES
A number of licenses are required for the Fibre Channel switches. When ordered with a Fabric MetroCluster configuration, the switches should include all necessary licenses. For reference, they are: Extended Distance license (if over 10km) Brocade Ports-on-Demand (POD) licenses for additional ports
10
8-PORT SWITCHES As shown in Figure 8, ports 01 and 45 belong to Pool0, while ports 23 and 67 belong to Pool1. The Brocade 200E 16-port switch has ports 07 licensed and operational, as shown in Figure 8. 16-PORT SWITCHES As shown in Figure 8, ports 03 and 811 belong to Pool0, while ports 47 and 1215 belong to Pool1. The Brocade 200E 16-port switch has ports 015 licensed and operational, as shown in Figure 8. 32-PORT SWITCHES (DATA ONTAP 7.2.5 OR LATER) The Brocade 5000 is not supported using the hardware disk ownership model.
11
3
3.1
DEPLOYMENT PLANNING
Data Gathering
CUSTOMER REQUIREMENTS
To facilitate a successful MetroCluster installation, gather the following information early on.
ENVIRONMENTAL
Distance between the primary and remote sites. This information is necessary to determine which type of MetroCluster is appropriate, or even whether either version is appropriate. In calculating the effective distance, consider factors such as cable type, speed, and number of patch panels. (See Section 3.2, "Distance Considerations.") Although NetApp recommends that dedicated dark fiber be used for a MetroCluster, WDM devices are supported. Refer to the Brocade Compatibility Guide at www.brocade.com for supported devices.
SETUP
Gather the following items before you begin any MetroCluster installation: NetApp licenses Hostnames and IP addresses for each of the nodes and the Fibre Channel switches Brocade switch licenses; gathering them is extremely important since there can be a delay in obtaining them Appropriate number and type of cables with appropriate connectors
3.2
Distance Considerations
OVERVIEW
Stretch MetroCluster can support a maximum of 500 meters between nodes at a speed of 2Gbps. Fabric MetroCluster, through the use of Fibre Channel switches, extends this distance to 100km at the same speed. At 4Gbps speeds, these distances are roughly cut in half unless you use the Brocade 5000, which leaves the maximum distance at 100km. This extended distance capability gives customers greater flexibility in the physical location of the active-active controllers while maintaining the high-availability benefits of clustered failover. This section describes a number of factors that affect the overall effective distance permissible between the MetroCluster nodes. Physical distance Number of connections Desired speed Cable type
PHYSICAL DISTANCE
As stated earlier, the Stretch MetroCluster configuration can extend to a maximum of 500m (2Gbps). This distance is reduced by speed, cable type, and number of connections. A Fabric MetroCluster can extend to 100km. This distance is affected by the same factors. At a distance of 100km, latency would be around 1ms. Greater distances would obviously result in greater latencies (500km = 5ms), which may be unacceptable for an application.
12
CABLE TYPE
As shown in Table 3, the cable type affects both distance and speed. Single-mode cable is supported only for the interswitch links. Example 1: A customer has 250 meters between sites and wants to run at 4Gbps. The OM-3 cable type is required. Example 2: A customer currently has a MetroCluster configuration running at 2Gbps with a distance of 300 meters over OM2 cabling and wants to upgrade to 4Gbps speeds. Upgrading the cabling will not help, because OM-3 has a maximum of 270 meters. In this case the choices would be: Remain at 2Gbps speeds. Customers with the new ESH4 disk shelves could still use them at this distance, as long as the shelf speed is set to 2Gbps. Test current optical network infrastructure to make sure that attenuation and latency are acceptable.
Table 3) Cable types and distances.
Fiber Type
Data Rate
1Gb/s
OM-2 (50/125UM)
OM-3 (50/125UM)
2Gb/s 4Gb/s
2Gb/s 4Gb/s
*The maximum distance shown here is typically due to the standard 1310nm SFPs. Use of high-power SFPs can extend this dark fiber up to 30km. Using 1550nm high-power SFPs, you can achieve a distance of 70 100km. This topic is discussed in much greater technical detail in the following technical reports: MetroCluster Upgrade Planning Guide (TR-3517) Optical Network Installation Guide (TR-3552)
3.3
Physical Layout
DISK LOOPS
In order to optimize performance, we recommend that instead of installing full loops of disks (two shelves maximum) that you install disk shelves one per loop until you utilize all disk loop ports. You can then add add-on shelves to each shelf. (See Figure 9.)
13
CABINETS
Although it is not mandatory that you install the equipment as shown in Figure 10, this is a physically desirable configuration that facilitates ease of cabling and scalability for expansion. For ease of management and troubleshooting, it is helpful to set shelf IDs based on location. For example, all disk shelves on the primary controller are set to odd numbers (1, 3, 5) while shelves on the remote controller are set to even numbers. This makes it much easier to identify and locate shelves when using Data ONTAP utilities.
NODE A
SW1 SW2
42 42 41 41
NODE B
SW3 SW4
42 42 41 41
40
40
40
40
39
39
39
39
38
38
38
38
37
37
37
37
36
36
36
36
35
ID 5
35
35
35
34
34
34
34
33
33
33
33
32
32
32
32
31
31
31
31
30
30
30
30
29
ID 3
29
29
29
28
28
28
28
Node A Shelves
27
27
26
26
25
25
Node B Shelves
27
27
26
26
25
25
24
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
DS14
FC
Power
24
24
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
MK2
DS14
FC
Power
24
MK2
23
ID 1
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
23
23
Loop A
Loop B
System
ID 1
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
23
Loop A
Loop B
System
Shelf ID
Shelf ID
22
22
22
22
21
21
21
21
20
FAS3050
20
20
FAS3050
20
19
activity
status
power
19
19
activity
status
power
19
18
18
18
18
Node B Mirror
17
17
16
16
15
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
DS14
FC
Power
15
Node A Mirror
17
17
16
16
15
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
MK2
DS14
FC
Power
15
MK2
14
ID 2
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
14
14
Loop A
Loop B
System
ID 2
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
14
Loop A
Loop B
System
Shelf ID
Shelf ID
13
13
13
13
12
12
12
12
11
11
11
11
10
10
10
10
ID 9
ID 11
14
NODE A
SW1 SW2
42 42 41 41
NODE B
SW3 SW4
42 42 41 41
40
40
40
40
39
39
39
39
38
38
38
38
37
37
37
37
36
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
DS14
36
36
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
MK2
DS14
36
MK2
35
ID 5
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
FC
Power
Fault
35
35
Loop A
Loop B
ID 5
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
FC
Power
Fault
35
Loop A
Loop B
System
System
Shelf ID
Shelf ID
34
34
34
34
33
33
33
33
32
32
32
32
31
31
31
31
30
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
DS14
FC
Power
30
30
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
MK2
DS14
FC
Power
30
MK2
29
ID 3
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
29
29
Loop A
Loop B
System
Shelf ID
ID 3
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
29
Loop A
Loop B
System
Shelf ID
28
28
28
28
Node A Shelves
27
27
26
26
25
25
Node B Shelves
27
27
26
26
25
25
24
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
DS14
FC
Power
24
24
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
MK2
DS14
FC
Power
24
MK2
23
ID 1
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
23
23
Loop A
Loop B
System
ID 1
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
23
Loop A
Loop B
System
Shelf ID
Shelf ID
22
22
22
22
21
21
21
21
20
FAS3050
20
20
FAS3050
20
19
activity
status
power
19
19
activity
status
power
19
18
18
18
18
Node B Mirror
17
17
16
16
15
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
DS14
FC
Power
15
Node A Mirror
17
17
16
16
15
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
MK2
DS14
FC
Power
15
MK2
14
ID 2
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
14
14
Loop A
Loop B
System
ID 2
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
14
Loop A
Loop B
System
Shelf ID
Shelf ID
13
13
13
13
12
12
12
12
11
11
11
11
10
10
10
10
9
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
DS14
9
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
MK2
DS14
MK2
ID 4
72F 72F
FC
Power
Fault
Loop A
Loop B
ID 4
72F 72F 72F
FC
Power
Fault
Loop A
Loop B
System
System
Shelf ID
Shelf ID
7
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
7
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
3
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
DS14
FC
Power
3
NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance NetworkAppliance
MK2
DS14
FC
Power
MK2
ID 6
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
Loop A
Loop B
ID 6
72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F 72F
Fault
Loop A
Loop B
System
System
Shelf ID
Shelf ID
Figure 11) Typical Fabric MetroCluster rack installation (shown half populated).
3.4
The Cluster Configuration Checker is a Perl script that detects errors in the configuration of a pair of activeactive NetApp controllers. It can be run as a command from a UNIX shell or Windows prompt, but also doubles as a CGI script that can be executed by a UNIX Web server. The script uses rsh or ssh to communicate with the controllers you're checking, so you must have the appropriate permissions for rsh to run on both controllers in the cluster pair. This script detects and reports the following: Services licensed not identical on partner (some services may be unavailable on takeover) Option settings not identical on partner (some options may be changed on takeover) Network interfaces configured incorrectly (clients will disconnect during takeover) FCP cfmode settings not identical on controllers with FCP licensed /etc/rc checked on each controller to see that all interfaces have a failover set This script is available on the NOW site for download. NetApp recommends that it be run as part of the implementation process. If the controllers being implemented were part of an active-active configuration, then the configurations are probably compatible. It never hurts to run this utility anyway, just to be certain. Best Practice: NetApp strongly recommends that you complete the installation planning workbook prior to beginning the installation. A little time spent up front will expedite the installation process.
As in most successful installations, the key is planning and preparation. The collected information outlined in section 3 is essential in completing the installation. The steps of a MetroCluster installation can be divided according to primary or remote site. This chapter outlines these steps. Refer to the Active-Active Configuration Guide on the NOW site for supplementary information.
4.1
Primary Site
15
RESULT
Login prompt Prompt for password Command line prompt Look for the following to be licensed: Cluster
License
cluster_remote syncmirror_local If any of the above are not licensed, perform the following step. When adding licenses remember to add them in the following order:
COMMAND
telnet switch IPAddress Login admin password
RESULT
Login prompt Prompt for password Command line prompt Look for the following to be licensed: brcd200e_whql01:admin> licenseshow Web license Zoning license Fabric Watch license Fabric license Following two licenses only if 16-port switch: Ports-on-Demand licenseadditional 4-port upgrade Ports-on-Demand licenseadditional 4-port upgrade Following two licenses only if 32-port switch: Ports-on-Demand licenseadditional 8-port upgrade Ports-on-Demand licenseadditional 8-port upgrade Extended Fabric license (only if distance between nodes > 10km) If any of the above are not licensed, perform the following step.
licenseshow
16
Result
Login prompt Prompt for password Command line prompt The currently installed switch firmware is displayed. Check the MetroCluster Fabric Switch download page on http://now.netapp.com for the currently supported version of the switch firmware. Note: To access this page, go to the Fibre Channel switch link on the Download Software page and select Brocade from the platform list. If your switch firmware is not the supported version, complete the steps outlined in Section 7.2, Appendix B: Switch Software Upgrade Procedure. Download the switch firmware from http://now.netapp.com and install it, as described in the appropriate Brocade switch hardware reference manual.
version
Reboot the switch. Disable the switch. Clear any preexisting configuration. Configure the switch with default settings. Set the switch parameters. Set only the following parameters.
Fabric parameters = y
domain_id = XXX
As a best practice, set the domain ID according to the switch number in the configuration. For example, at the primary site, switches 1 and 2 would have domain IDs SW1 and SW2 and switches 3 and 4 at the remote would be SW3 and SW4, respectively.
17
Exit the configuration utility. Reboot the switch so that the default settings take effect. Log in to the switch and disable it.
This step is for Model 3000 switches (3200, 3800, 3250, 3850)
portCfgLPort <port#>,0,0,1
Set all ports attached to disk loops to half duplex. Note: Perform this command only for the ports to which disks are attached. Refer to the switch-to-disk tables in Cable Primary Node for the port numbers. ON is displayed in the Locked Loop HD fields only.
portCfgLPort
Set all ports attached to disk loops to half duplex. Note: Perform this command only for the ports to which disks are attached. Refer to the switch-to-disk tables in Cable Primary Node for the port numbers. Set all ports attached to disk loops to Locked L-Port. Note: Perform this command only for the ports to which disks are attached. Refer to the switch-to-disk tables in Cable Primary Node for the port numbers.
portCfgLPort <port#>,1,0,1
portCfgLPort
Configure the Inter-Switch Link (ISL) port. (0 disables trunking.) Trunking must be disabled prior to configuring the long-distance ISL port.
18
For information about configuring an ISL port for an ISL length greater than 10km, refer to Appendix C of this document. For additional compatibility information, see the Brocade Fabric Aware Compatibility Matrix: http://www.brocade.com/products/interop_and_compatibility.jsp Enable or disable ISL R_RDY mode on a port. Toggling this mode may sometimes be necessary (high error rate on port) when connecting to WDM devices. Enable the switch. Set the switch name. Loc_SW1 is the name you gave to the current switch. Use a name that is meaningful, such as location (Loc for Local, Rem for Remote) and the switch number. Best practice: To assist in documenting the installation, names can be assigned to each port. Examples:
Portname 12, ISL
switchname Loc-SW1
Portname <number>,name
Using the portshow command, this identifies port 12 as being an ISL port.
Portname 0, PRI-0a
19
VERIFY CONNECTIONS
Confirm that the disks are visible and have dual paths by entering the following command on the console: storage show disk p The output shows the disks connected to the switches, the port to which they are connected, and the disk and module to which they belong, as shown in the following example: ha16*> storage show disk -p PRIMARY PORT SECONDARY PORT --------- ---switch4:0.40 B switch3:0.41 A switch4:12.52 A SHELF ----5 5 6 BAY -0 1 4
If redundant paths are not shown for each disk, recheck the cabling. Note: When recabling or performing expansions on a MetroCluster using hardware disk ownership, it is important to triple-check that the A loop and B loop of a shelf run to the very same port number on the two local switches. This is especially important when, for example, a 16-port 3850 switch gets replaced by a 200E. Older versions of Data ONTAP immediately panic and halt when the links to a shelf are not connected to the same port number. This affects both nodes of a Fabric MetroCluster simultaneously, leading to a 100% outage for both nodes.
20
SET UP MIRRORS
To ensure highly available access to data, all data on each node must be mirrored to the other node using SyncMirror. Although it is possible to have data on one node that is not mirrored on the other, NetApp does not recommend it. Keep in mind that mirrors can exist only between like drive types (FC to FC or ATA to ATA in the case of Stretch MetroCluster). With the SyncMirror license installed, disks are divided into pools. When a mirror is created, Data ONTAP pulls disks from Pool0 for the primary data and from Pool1 for the mirror. The selection of which disks to use for the mirror can be left up to Data ONTAP or chosen specifically. For detailed information, see the Data ONTAP Online Backup and Recovery Guide. It is important to verify the correct number of disks in each pool before creating the mirrored aggregate or traditional volume. Any of these commands can be used to verify the number of drives in each pool: Sysconfig r gives the broadest information Aggr status -r Vol status -r Once the pools are verified, you can create mirrors with one of the following: { aggr | vol } create { aggrname | volname} m ndisks[@disk-size] For example, the command aggr create aggrA -m 6 creates a mirrored aggregate called aggrA with 6 drives (3 for plex 0, 3 for plex 1). Or an existing aggregate can be mirrored using the following: { aggr | vol } mirror { aggrname | volname } In both of the above cases, Data ONTAP is allowed to choose the specific drives to be used. Optionally, the user can choose the drives. Further information can be found in Data ONTAP Online Backup and Recovery.
Note: Unassigned disks are associated with a pool. To assign them to a different pool, use the -f option. However, moving individual disks between pools could result in the loss of redundancy and cause disk auto-
21
assignment to be disabled for that loop. For this reason, you should move all disks on that loop to the other pool if possible. -o -s -c ownername specifies the system that the disks are assigned to. sysid specifies the system that the disks are assigned to. specifies the checksum type (either block or zoned) for a LUN in V-Series systems.
-f must be specified if a system already owns the disk or if you want to assign a disk to a different pool. Enter man disk or refer to the Data ONTAP installation and administration guide for details.
4.2
Remote Site
VERIFY LICENSES
Follow the same procedures as performed at the primary site.
VERIFY CONNECTIONS
Follow the same procedures as performed at the primary site.
22
SET UP MIRRORS
Follow the same procedures as performed at the primary site.
4.3
Testing
VERIFY THAT EACH SITE CAN SEE THE PROPER DISKS TEST FAILOVER
Primary to remote Remote to primary
TEST GIVEBACK
Primary to remote Remote to primary
4.4
Additional disk shelves can be added to a MetroCluster configuration according to the procedures defined in the Data ONTAP Active-Active Configuration Guide. The following guidelines apply: ATA shelves are not supported on a Fabric MetroCluster. Make sure that additional shelves are added at the partner for mirroring. When adding shelves to new ports, the very same port on the other switch must be used. Otherwise, both nodes may not boot (with invalid disk configuration), or they may panic and halt at once, if this is performed online. Only like shelves can be mirrored; for example, FC to FC or ATA to ATA (Stretch MetroCluster only). There is a limit of two shelves per switch port on a Fabric MetroCluster. The total limit for a Fabric MetroCluster is 504 disks. For a Stretch MetroCluster, the limit is that of the specific platform. When adding shelves to a Fabric MetroCluster, be sure to configure the associated switch ports to which the new shelves are connected. Keep in mind that each loop must have shelves of the same speed (2- or 4Gbps) and have the same type of controller (all LRC or all ESH/ESH2/ESH4). ESH4 disk shelves have a shorter maximum distance at 4Gbps. As shown in Figures 14 and 15, you can add additional shelves with a maximum of two shelves per loop (Fabric MetroCluster).
23
Secondary
Pool 1 0 Pool 1 0
Switches 3,4
For Mirror
Primary
Switches 1,2
Pool 1 Pool 0 1 0
For Mirror
2nd Loop Mirror 3rd Loop Mirror
NOTES 1. Maximum number of disks equals lesser of platform maximum or 336. 2. At each location, each shelf must be connected to the same port on each of the two switches. 3. See Figure 3-1 for physical shelf placement
4.5
24
They should all show up as initiator. For any that do not, perform the following: Ensure that the Fibre Channel ports are offline: fcadmin config -d <adapter> Note: Adapter refers to the port ID (for example, fcadmin config -d 0c took port 0c offline). Also, more than one port can be specified. Set the onboard port to target or initiator mode: fcadmin config -t initiator <adapter> Reboot the controller and check again: fcadmin config This verifies that the onboard Fibre Channel ports are online.
CONFIGURATION OF SWITCHES
If the Brocade 200E switch is being used in the 8-port configuration, make sure that no SFPs are plugged into ports 815. Even though those ports may not be licensed or enabled, if Data ONTAP sees SFPs in those ports it treats them as 16-port switches and expects to see devices connected accordingly.
25
5
5.1
Upon determining that one of the sites has failed, the administrator must execute a specific command on the surviving node to initiate a site takeover. The command is: cf forcetakeover d The takeover is not automatic because there may be cases in an active-active configuration in which the network between sites is down and each site is still fully functional. In such cases a forced takeover might not be desirable. It is important to remember that this is the case only when a complete site is lost. In the case of a failed controller at one of the sites, a normal cluster failover occurs. Due to the operation of SyncMirror, there is also added protection from multidrive or complete shelf failures. When a storage controller fails in an active-active configuration, the partner detects the failure and automatically (if enabled) performs a takeover of the data serving responsibilities from the failed controller. Part of this process relies on the surviving controller being able to read information from the disks on the failed controller. If this quorum of disks is not available, then automatic takeover wont be performed. In a MetroCluster configuration, manually executing a single command will allow a takeover to occur in spite of the lack of a quorum of disks. This forced takeover process breaks the mirrored relationships in order to bring the failed controllers volumes online. This results in the following: Volumes have a new file system ID (FSID) in order to avoid conflict with the original volumes. LUNs (iSCSI or FCP) have a new serial number (in part derived from the FSID). Previous NFS mounts are stale and will need to be remounted. LUNs are offline in order to make sure that only the desired LUNs are brought online after the site failure. Effective in Data ONTAP 7.2.4, there is an option to preserve the original FSID, which allows LUNs to retain their original serial number and the NFS mounts to be brought online automatically. This option is called cf.takeover.change_fsid. If set to off (0) the original FSID will be preserved. This will be covered further in the implementation section of this document.
26
b. Use any application-specified method that either prevents the application from restarting at the disaster site or prevents the application clients from accessing the application servers at the disaster site. Methods can include turning off the application server, removing an application server from the network, or any other method that prevents the application server from running applications .
RECOVERY PROCESS
Although you can perform a complete site failover with a single command, there are cases in which another step or two may be necessary before data is accessible at the surviving site. If using either a release of Data ONTAP earlier than 7.2.4 or if using 7.2.4 or later and the option setting is cf.takeover.change_fsid =1: NFS volumes must be remounted. For more information about mounting volumes, see the File Access and Protocols Management Guide. iSCSI and Fibre Channel LUNS may need to be rescanned by the application if the application (i.e., VMware) relies on the LUN serial number. When a new FSID is assigned the LUN serial number changes. For iSCSI and Fibre Channel LUNS after a forced takeover, all LUNs that were being served from the failed site will be served by the surviving site. However, each of these LUNs must be brought online. For example: lun online /vol/vol1/lun0 /vol/vol1/lun1 The reason they are offline is to avoid any LUN ID conflict. For example, suppose that two LUNs with an ID of 0 mapped to the same igroup, but one of these LUNs was offline before the disaster. If the LUN that was previously offline came online first, the second LUN would not be accessible because two LUNs with the same ID mapped to the same host cannot be brought online. Both of the steps described above can be automated with some clever scripting. The main challenge for the script is to understand the state of the LUNs before the site failure so that the script brings online only LUNs that were online prior to the site failure.
GIVEBACK PROCESS
After you resolve all the problems that caused the site failure and ensure that the controller at the failed site is offline, it is time to prepare for the giveback so the sites can return to their normal operation. You can use the following procedure to resynchronize the mirrors and perform the giveback.
COMMAND
Turn on power to the disk shelves and FC switches at the disaster site. aggr status -r partner
RESULT
Validate that you can access the remote storage. If remote shelves dont show up, check connectivity. Go into partner mode on the surviving node. Determine which aggregates are at the surviving site and which aggregates are at the disaster site by entering this command. Aggregates at the disaster site show plexes that are in a failed state with an out-of-date status. Aggregates at the surviving site show plexes as online.
aggr status r
If aggregates at the disaster site are online, take them offline by entering the following command for each online aggregate:
28
6
6.1
RELEVANT DOCUMENTATION
MetroCluster Documentation
TR-3517: MetroCluster Upgrade Planning Guide
The following document can be found in the technical library: The following documents can be found on the NOW site: Active-Active Configuration Guide Data Protection Online Backup and Recovery Guide (Chapter 8, SyncMirror) MetroCluster Compatibility Matrix
29
7
7.1
APPENDICES
Appendix A: Switch Port Assignments (Hardware Ownership)
SWITCH 1 (local top) Port 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Type FC Port or HBA FC Port or HBA 9xx 5a 8a 3020/ 3050 0a 0c Pool 0 0 0 0 1 1 1 1 0 0 0 0 1 1 1 SWITCH 3 Port 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Type FC Port or HBA FC Port or HBA 9xx 5a 8a 3020/ 3050 0a 0c Pool 0 0 0 0 1 1 1 1 0 0 0 0 1 1 1 Port 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Port 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
SWITCH 2 (local bottom) Type FC Port or HBA FC Port or HBA 9xx 5b 8b 3020/ 3050 0b 0d Pool 0 0 0 0 1 1 1 1 0 0 0 0 1 1 1 SWITCH 4 Type FC Port or HBA FC Port or HBA 9xx 5b 8b 3020/ 3050 0b 0d Pool 0 0 0 0 1 1 1 1 0 0 0 0 1
Remote shelf B loop 1 Remote shelf B loop 1 Remote shelf B loop 1 Remote shelf B loop 1 Local shelf B loop 1 Local shelf B loop 2 Local shelf B loop 3 Local shelf B loop 4 Cluster Interconnect ISL
10a
1a To Switch 3
Remote shelf A loop 1 Remote shelf A loop 1 Remote shelf A loop 1 Remote shelf A loop 1 Local shelf A loop 1 Local shelf A loop 2 Local shelf A loop 3 Local shelf A loop 4 Cluster Interconnect ISL
To Switch 4
Remote shelf B loop 1 Remote shelf B loop 1 Remote shelf B loop 1 Remote shelf B loop 1 Local shelf B loop 1 Local shelf B loop 2 Local shelf B loop 3 Local shelf B loop 4 Cluster Interconnect ISL
10a
1a From Switch 1
Remote shelf A loop 1 Remote shelf A loop 1 Remote shelf A loop 1 Remote shelf A loop 1 Local shelf A loop 1 Local shelf A loop 2 Local shelf A loop 3 Local shelf A loop 4 Cluster Interconnect ISL
From Switch 2
30
7.2
In order to ensure an optimum MetroCluster environment, you may occasionally need to upgrade the firmware on the Fibre Channel switches. To determine the proper revision level, consult the MetroCluster Compatibility Matrix on the NOW site. The instructions in this section can also be found at
http://now.netapp.com/NOW/download/software/sanswitch/fcp/Brocade/mc_ontap641_fabric_200_dow nload.shtml.
To download the Brocade 5.1.0 firmware to a workstation and then install it on a Brocade 200E switch, complete the following steps: 1. Ensure that the workstation recognizes the switch on the network (using the ping utility, for example).
2. From the browser on the workstation, right-click this installation package (240.5MB) link.
3. Specify the directory to which you want to download the firmware installation package. 4. After downloading is complete, access the switch from the workstation, using a telnet session. 5. Install the firmware on the switch by issuing the firmware download command. For the syntax and a description of the firmware download command, see the Brocade Fabric OS Reference Guide. 6. Reboot the switch to activate the new firmware.
31
7.3
Distance @ 1Gbps 10km 10km 25km 50km 100km Auto to 500km 500km
Distance @ 2Gbps 5km 10km 25km 50km 100km Auto to 250km 250km
Distance @ 4Gbps 2km 10km 25km 50km 100km Auto to 100km 100km
Earliest FOS Release All 3.x, 4.x 3.1.0, 4.1.0, 4.x, 5.x All All 3.1.0, 4.1.0, 4.4.0, 5.x 5.1.0
32
7.4
The Fabric MetroCluster worksheet is a planning aid on which you can record specific cabling information about your Fabric MetroCluster. You can use this information during configuration procedures.
CUSTOMER INFORMATION
Customer Name Address City Main Phone Primary Contact Primary Phone Number Requested Install Date
Site A Installation Location (if different from above)
State/Country
State/Country
Address City Main Phone Primary Contact Primary Phone Number State/Country
33
GENERAL
As with most projects the keys to a successful deployment of a NetApp MetroCluster are planning, preparation, and documentation. In order to facilitate these activities, gather the following data early in the process.
PHYSICAL
1. 2. What is the total distance between the two locations? Please describe the fiber infrastructure between the locations. a. Fiber type b. c. 3. 4. 5. 6. 7. 8. Number of unused pairs Number of patch panels
Is this infrastructure already in place? If not, when will it be complete? Is this an upgrade to a MetroCluster? Please outline any data to be migrated to/from either of the nodes. Is there or will there be a mix of 2Gbps and 4Gbps disk shelves? If 4Gbps operation is desired, have the appropriate HBAs been ordered? (Most onboard ports are only capable of 2Gbps operation.) Which type of disks (FC, ATA, etc.) currently exist (if upgrading) or will exist? Note: ATA is not supported on a Fabric MetroCluster.
9.
34
INSTALL CHECKLIST
DESCRIPTION Appropriate power is installed/ready for Site A Appropriate power is installed/ready for Site B Distance between FAS at Site A and Switch 1please list measured distance **(this distance will determine minimum cable lengths needed) Distance between Shelves Site A and Switch 1 please list measured distance **(this distance will determine minimum cable lengths needed) Distance between FAS at Site A and Switch 2please list measured distance **(this distance will determine minimum cable lengths needed) Distance between Shelves at Site A and Switch 2please list measured distance **(this distance will determine minimum cable lengths needed) Distance between FAS at Site B and Switch 3please list measured distance **(this distance will determine minimum cable lengths needed) Distance between Shelves at Site B and Switch 3please list measured distance **(this distance will determine minimum cable lengths needed) Distance between FAS at Site B and Switch 4please list measured distance **(this distance will determine minimum cable lengths needed below) Distance between Shelves at Site B and Switch 4please list measured distance **(this distance will determine minimum cable lengths needed) Distance between Switch 1 and Switch 3please list measured distance Distance between Switch 2 and Switch 4please list measured distance Determine connection type at your demarc *(cables used between switches and demarc are the responsibility of the customer) Serial console (laptops) should be used to connect to the switches and appliances for installation, configuration, and testing. Laptops should have HyperTerm or TeraTerm applications installed. A minimum of 2 serial consoles are needed, 1 per site for the duration of the project. Request 2 per site if possible. Please advise the number of serial consoles available. Complete the NetApp System Configuration Worksheet prior to project start date Appropriate number of Ethernet network connections are available/ready Both systems have access to ALL the same networks/subnets/SAN for failover capability Complete Data Layout Design to include naming conventions for aggregates, volumes, and qtrees UNIX system administrator(s) with domain privileges dedicated for this project UNIX system administrator(s) with domain privileges dedicated for this project for any work beyond the normal 8- hour work day Windows system administrator(s) with domain privileges dedicated for this project Windows system administrator(s) with domain privileges dedicated for this project for any work beyond the normal 8-hour work day Network administrator(s) with domain privileges dedicated for this project Network administrator(s) with domain privileges dedicated for this project for any work beyond the normal 8hour work day Total number of cables: **NOTE: Any cable distances that are greater than the NetApp supplied cable lengths are the responsibility of the customer. Yes or No
35
36
Switch 1
Switch 2
Switch 3
Switch 4
37
Remote Site
Shelf ID 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 Owner System Pool Firmware Version Physical Location
38
7.5
39
7.6
www.netapp.com
2008 NetApp. All rights reserved. Specifications are subject to change without notice. NetApp, the NetApp logo, Data ONTAP, NOW, RAID-DP, Snapshot, and SyncMirror are trademarks or registered trademarks of NetApp, Inc. in the United States and/or other countries. Windows is a registered trademark of Microsoft Corporation. UNIX is a registered trademark of The Open Group. VMware is a registered trademark of VMware, Inc. All other brands or products are trademarks or registered trademarks of their respective holders and should be treated as such.
40