Cluster
cluster ha modify -configured true
HA cluster ha show
cluster quorum-service options show
cluster ring show
Cluster ring / cluster show -epsilon *
epsilon cluster modify -node < > -epsilon <false / true>
cluster statistics show
cluster ping-cluster -node node1
cluster join -clusteripaddr 192.0.1.1
cluster setup
cluster setup
/ node
cluster show
unjoin cluster unjoin -node node4 {-force}
debug vreport show
event log show -messagename scsiblade.*
cluster kernel-service show -list
debug smdb table bcomd_info show
Failover `-auto-giveback enabled
storage failover show
Node1> storage failover takeover -bynode node0 -option immediate
storage failover show-takeover
storage failover giveback -fromnode node1
storage failover giveback -ofnode < > -override-vetoes true
storage failover modify -node node0 -enabled true
storage failover show-giveback
storage failover hwassist show
storage failover hwassist stats clear -node cluster1-02
storage failover mailbox-disk show
storage failover progress-table show
Manage high-availability configuration
Show high-availability configuration status for the cluster
Display the settings for cluster quorum-service
Display cluster node member’s replication rings
Displays epsilon node
Display cluster-wide statistics -- CPU, NFS, CIFS, FCP, Cluster Interconnect Traffic
Ping remote cluster interfaces and perform RPC server check
Join an existing cluster using the specified member’s IP address or by cluster name
Setup wizard
Display cluster node members
Unjoin or remove a node from the cluster
(must be run in priv -set diag, shows WAFL and VLDB consistency)
(show that cluster is in quorum)
(in diag mode, displays in quorum information)
(displays database master / secondary for bcomd)
Display storage failover status
Take over the storage of a node’s partner
Display takeover status
Return failed-over storage to its home node
Modify storage failover attributes
Display giveback status
Display hwassist status
Clear the hwassist statistics
Display information about storage failover mailbox disks
Display status information about storage failover operations
RDB
MGWD
VIFMGR
VLDB
BCOMD
CRS
Cluster Ring
Quorum
Epsilon
Replicated DB
Management Gateway - provides access to CLI, enables cluster mgmt from nodes
Virtual Interface Mgr - stores & monitor LIF config, LIF failover policies
Volume Location DB - Index of which aggr owns which vol, which host owns which aggr
Block config & Ops mgmt - LUN map definitions, igroups config
Config replication Service - Metrocluster config
describes a group of identical processes (unit names in RDB) running on all nodes in the cluster
is a idle condition for a functionality of cluster
when a cluster is in quorum, it means that majority of nodes are healthy and can communicate with each other
is assigned to a single node in the cluster, it gives an extra fractional voting weight
which is a deciding factor when the connectivity fails
If different VLAN - firewall ports need to be opened
Same VLAN - no firewall ports needs to be opened
Mumbai
Production
Different VLAN
1
2
3
4
5
6
7
8
CFO - Controller Failover (Admin root volume)
SFO - Storage Failover (non-root data volumes)
n the cluster
ommunicate with each other
Chennai
DR Site
Different VLAN
DR1
DR2
DR3
DR4
DR5
DR6
cluster peer create -peer-addrs cluster2-d2,10.10.1.1 -username admin
cluster peer delete -cluster cluster1
cluster peer modify
cluster peer show -instance
cluster peer modify-local-name -name cluster1 -new-name cluster1A
cluster peer ping
cluster peer show {-instance}
cluster peer connection show
cluster peer health show
cluster peer offer cancel -cluster cluster2
cluster peer offer show
volume create -vserver <vserver_name> -volume vol10_mirror -aggregate <destination_aggr_name>
snapmirror create -vserver <vserver_name> -source-path sysadmincluster://vserver1/vol10 -destinati
snapmirror initialize -destination-path sysadmincluster://vserver1/vol10_mirror -type DP -foregroun
snapmirror update -source-path vserver1:vol10 -destination-path vserver2:vol10_mirror -throttle 100
snapmirror modify -source-path vserver1:vol10 -destination-path vserver2:vol10_mirror -throttle 20
snapmirror restore -source-path vserver1:vol10 -destination-path vserver2:vol10_mirror (Restore a s
snapmirror show (show snapmirror relationships and status)
snapmirror quiesce -destination-path dst_vserver:dst_c_vol
snapmirror break -destination-path src_vserver:src_c_vol
snapmirror delete -destination-path src_vserver:src_c_vol
snapmirror release dataVol20 vs1.example.com:dst_vol
SNAPVAULT
snapmirror create -source-path vserver1:vol5 -destination-path vserver2:vol5_archive -type XDP -sc
NOTE: Type DP (asynchronous), LS (load-sharing mirror), XDP (backup vault, snapvault), TDP (tr
job schedule cron create -name 15_minute_sched -minute 15
snapmirror modify -destination-path src_vserver:src_c_vol -schedule 15_minute_sched
snapmirror list-destinations -instance
Create a new cluster peer relationship
Delete a cluster peer relationship
Modify cluster peer relationships
Show cluster peer relationships
Modify the local name for a cluster peer
Initiate intercluster connectivity test
Display peer cluster information
Show current peering connections for a cluster
Check peer cluster health
Cancel the outstanding offer to authenticate with a peer cluster
Display outstanding offers to authenticate with a peer cluster
ggregate <destination_aggr_name> -type DP -size 100g (Create a destinaion Snapmirror Volume)
ncluster://vserver1/vol10 -destination -path sysadmincluster://vserver1/vol10_mirror -type DP (Create a snapm
1/vol10_mirror -type DP -foreground true (Initialize the snapmirror example)
vserver2:vol10_mirror -throttle 1000 (Snapmirror update and throttle to 1000KB/sec)
vserver2:vol10_mirror -throttle 2000 (Change the snapmirror throttle to 2000)
vserver2:vol10_mirror (Restore a snapmirror from destination to source)
erver2:vol5_archive -type XDP -schedule 5min -policy backup-vspolicy (Create snapvault relationship with 5
(backup vault, snapvault), TDP (transition), RST (transient restore)
dule 15_minute_sched
vserver peer accept -vserver < > -peer-vserver < > -peer-cluster < > -application snapmirror
vserver peer create
vserver peer delete
vserver peer modify
vserver peer modify-local-name
vserver peer reject
vserver peer repair-peer-name -vserver vs1.thesanguy.com
vserver peer resume
vserver peer show
vserver peer suspend -vserver lvs1.thesanguy.com -peer-vserver pvs1.thesanguy.com
vserver peer transition create
vserver peer transition delete
vserver peer transition modify
vserver peer transition show
destinaion Snapmirror Volume)
ver1/vol10_mirror -type DP (Create a snapmirror relationship for sysadmincluster)
example)
le to 1000KB/sec)
tle to 2000)
olicy (Create snapvault relationship with 5 min schedule using backup-vspolicy)
Accept a pending Vserver peer relationship
Create a new Vserver peer relationship
Delete a Vserver peer relationship
Modify a Vserver peer relationship
Modify the local name for a peer Vserver
Reject a Vserver peer relationship
Repair the peer vserver name that was not updated during the last rename operation
Resume a Vserver peer relationship
Display Vserver peer relationships
Suspend a Vserver peer relationship
Create a new transition peer relationship between a 7-Mode system and a Vserver.
Delete a transition peer relationship.
Modify a transition peer relationship.
Display transition peer relationships
Ports
11104 Snapmirror port1
11105 Snapmirror port2
53 DNS
443 https
139 netbios
900
7810
7824 Cluster RPC
Types of Replication
DP snapmirror
XDP snapvault
TDP Transition 7-mide to cmode
LS Load Sharing only SVM root vol
milan
operation
network ipspace create -name ipspace1
network ipspace delete -ipspace ipspace1
IP Space
network ipspace rename -ipspace ipsA -new-name ipsB
network ipspace show
network port ifgrp add-port -node <node_name> -ifgrp <vif_name> -port <port_name>
network port ifgrp create -node <node_name> -ifgrp <vif_name> -distr-func ip -mode m
Ifgrp network port ifgrp delete
network port ifgrp remove-port
network port ifgrp show
network port vlan create -node node1 -vlan-name f1d-90
VLAN network port vlan delete -node node1 -vlan-name f1d-90
network port vlan show
network port broadcast-domain add-ports
network port broadcast-domain create
network port broadcast-domain delete
network port broadcast-domain merge
Broadcast
domain network port broadcast-domain modify
network port broadcast-domain remove-ports
network port broadcast-domain rename
network port broadcast-domain show
network port broadcast-domain split
network interface failover-groups add-targets
network interface failover-groups create -failover-group < > -node < > -port <vif_name>
network interface failover-groups delete
Failover -
Groups network interface failover-groups modify
network interface failover-groups remove-targets
network interface failover-groups rename
network interface failover-groups show
Create a new IPspace
Delete an IPspace
Rename an IPspace
Display IPspace information
grp <vif_name> -port <port_name>
p <vif_name> -distr-func ip -mode multimode
Destroy a port interface group net int
Remove a port from an interface group
Display port interface groups
Create a virtual LAN
Delete a virtual LAN
Display virtual LANs
Add ports to a layer 2 broadcast domain
Create a new layer 2 broadcast domain
Delete a layer 2 broadcast domain
Merges two layer 2 broadcast domains
net port
Modify a layer 2 broadcast domain
Remove ports from a layer 2 broadcast domain
Rename a layer 2 broadcast domain
Display layer 2 broadcast domain information
Splits a layer 2 broadcast domain into two Node only
Add failover targets to a failover group Node only
roup < > -node < > -port <vif_name> Node only
Delete a failover group Cluster wide
Modify a failover group Cluster wide
Remove failover targets from a failover group
Rename a logical interface failover Group
Display logical interface failover groups
network interface create
network interface delete
network interface migrate
network interface modify -vserver < > -lif < > - address < > -netmask < >
network interface rename -vserver vs0 -lif clusterlif0 -newname clusterlif1
network interface revert -vserver * -lif *
network interface show
network interface start-cluster-check
network interface capacity show
network interface capacity details show
network interface check failover show
network interface dns-lb-stats show
network interface lif-weights show
network port delete
network port modify - node < > -port < > -mtu < > -flowcontrol-admin < >
network port show
network port show-address-filter-info
Cluster LIF
Intercluster
Node mgmt
clust management
data LIF
Create a logical interface
Delete a logical interface
Migrate a logical interface to a different port
er < > -lif < > - address < > -netmask < >
er vs0 -lif clusterlif0 -newname clusterlif1
Revert a logical interface to its home port
Display logical interfaces
Start the cluster check function
Display the number of IP data LIFs capable of being configured on the cluster.
Display details about the IP data LIFs capable of being configured on each node.
Discover if any LIFs might become inaccessible during a node outage, due to over-provisioning
Show the DNS load-balancer stats for this node
Show the load-balancer LIF weights
Delete a network port
-port < > -mtu < > -flowcontrol-admin < >
Display network port attributes
Print the port’s address filter information
carry traffic btw nodes in a cluster
cross cluster communication, backup, replication
manage a particular node in a cluster
single management interface for the entire cluster
associate with SVM used for communicating with clients
system license show -package nfs
vserver nfs create -vserver vserver_name -v3 {enabled|disabled} -v4.0 {enabled|disabled} -v4-id-domain nfsv4_id_d
vserver nfs status -vserver vserver_name
vserver nfs show -vserver vserver_name
vserver export-policy check-access Given a Volume And/or a Qtree, Check to
vserver export-policy copy Copy an export policy
vserver export-policy create -policy-name < > Create a rule set
export-policy rule create -policyname < > -clientmatch <ip/subnet/hostname> -rorule <any /none/never/krb5/krb5
vol mount -volume test1 -junction-path /test1 -active true -policy-override false
vol show -volume test1 -field junction-path
vol modify -volume test1 -policy test
vserver export-policy check-access -vserver vs1 -client-ip 1.2.3.4 -volume flex_vol -authentication-method
vserver export-policy delete Delete a rule set
vserver export-policy rename Rename an export policy
vserver export-policy show Display a list of rule sets
vserver cifs create -vserver < > -cifs-server < > -domain < > -ou < >
vol show -volume test1 -field junction-path
vserver export-policy check-access -vserver vs1 -client-ip 1.2.3.4 -volume flex_vol -authentication-method
net int create -vserver svm1-cluster1 -lif fc_test -role data -data-protocol fcp -home-node cluster1-01 -home-port 0
portset create -portset ps1 -protocol mixed -vserver svm1-cluster1 -port-name fc_test
igroup create -igroup test -protocol fcp -ostype windows -initiator - -vserver svm1-cluster1 -portset ps1
qtree create -volume test1 -qtree q1 -security-style ntfs -oplock-mode enable -vserver svm1-cluster1
lun create -path /vol/test1/q1/test -size 1g -ostype windows -space-reserve enabled -space-allocation disabled -clas
lun map -path /vol/test1/q1/test -igroup test -vserver svm1-cluster1 -lun-id 1
lun show
lun show -m
lun show -v
sabled} -v4-id-domain nfsv4_id_domain -v4-numeric-ids {enabled|disabled} -v4.1 {enabled|disabled} -v4.1-pnfs {enabled|disab
me And/or a Qtree, Check to See If the Client Is Allowed Access
ule <any /none/never/krb5/krb5i/krb5p/ntlm/sys> -rwrule < > -allow-suid <true> -allow-dev <true> -protocol <any/nfs/nfs3/nf
x_vol -authentication-method sys -protocol nfs3 -access-type read-write -qtree qt1
port policy
of rule sets
x_vol -authentication-method sys -protocol nfs3 -access-type read-write -qtree qt1
e-node cluster1-01 -home-port 0a -status-admin up
luster1 -portset ps1
ver svm1-cluster1
d -space-allocation disabled -class regular -vserver svm1-cluster1
led} -v4.1-pnfs {enabled|disabled}
ue> -protocol <any/nfs/nfs3/nfs4/cifs> -superuser any -ruleindex 1
Cluster SVM
Node SVM
data SVM
ALUA is an industry standard protocol for identifying optimized paths between a storage system and a ho
... It is beneficial because multipathing software can be developed to support any storage array.
MPIO provides more than one physical path between the controller and the initiator. ALUA determines the
AUTOSUPPORT
system node autosupport budget show -node local (In diag mode – displays current time an
system node autosupport budget modify -node local -subsystem wafl -size-limit 0 -time-limit 10m (In diag mode – m
system node autosupport show -node local -fields max-http-size,max-smtp-size (Displays m
system node autosupport modify -node local -max-http-size 0 -max-smtp-size 8MB (modification as per Netapp KB1
AGGREGATES
aggr create -aggregate <aggregate_name> -diskcount <the number of disks you wish to add> -raid
aggr offline | online (Make the aggr offline or online)
aggr rename -aggregate <aggr_name> -newname <new_aggr_name) (Change the name of an ex
aggr relocation start -node node01 -destination node02 -aggregate-list aggr1 (Relocate aggr1 from
aggr relocation show (Shows the status of an aggregate relocation job)
aggr show -space (Show used and used% for volume foot prints and aggregate metadata)
aggregate show (show all aggregates size, used% and state)
aggregate add-disks -aggregate <aggregate_name> -diskcount <number_of_disks> (Adds a numb
reallocate measure -vserver vmware -path /vol/datastore1 -once true (Test to see if the volume da
reallocate start -vserver vmware -path /vol/datastore1 -force true -once true (Run reallocate on the
deleting a stale aggregate, this can be from a foriegn shelf (Only do this is you are absolute
::> system node run -node NODE1 sysconfig -r (aggregate will be labelled as foreign)
::> set adv
::*> storage aggregate remove-stale-record -nodename NODE1 -aggregate <foreign_aggregaste_nam
a storage system and a host. ALUA enables the initiator to query the target about path attributes, such as primary p
any storage array.
iator. ALUA determines the optimized and non-optimized paths between the initiator and the LUN.
displays current time and size budgets)
me-limit 10m (In diag mode – modification as per Netapp KB1014211)
ax-smtp-size (Displays max http and smtp sizes)
odification as per Netapp KB1014211)
sks you wish to add> -raidtype raid_dp -maxraidsize 18 (Create an AGGR with X amount of disks, raid_dp a
hange the name of an existing aggr)
ggr1 (Relocate aggr1 from node01 to node02)
gregate metadata)
r_of_disks> (Adds a number of disks to the aggregate)
st to see if the volume datastore1 needs to be reallocated or not)
ue (Run reallocate on the volume datastore1 within the vmware vserver)
this is you are absolutely sure that the aggregate is foreign, best to engage NetApp support
d as foreign)
e <foreign_aggregaste_name>
attributes, such as primary path and secondary path.
amount of disks, raid_dp and raidgroup size 18)
NetApp support
BROADCAST DOMAINS
broadcast-domain create -broadcast-domain NFS -mtu 9000 -ports NODE01:e0d-5 (Create
broadcast-domain add-ports -ipspace Default -broadcast-domain NFS -ports NODE02:e0d-
broadcast-domain remove-ports -ipspace Default -broadcast-domain NFS -ports NODE02:e
broadcast-domain delete -broadcast-domain NFS -ipspace Default (Deletes a broadcast do
network port broadcast-domain merge -ipspace Default -broadcast-domain MANAGEMENT
NODE01:e0d-5 (Creates a broadcast-domain and adds the node port)
FS -ports NODE02:e0d-5 (Adds a port to a broadcast domain)
n NFS -ports NODE02:e0d-5 (Removes a port from a broadcast domain)
(Deletes a broadcast domain)
domain MANAGEMENT1 -into-broadcast-domain MANAGEMENT2 (moves all ports from MANAGE
all ports from MANAGEMENT1 broadcast domain into MANAGEMENT2 broadcast domain)
oadcast domain)
CIFS
vserver cifs modify -vserver <vserver_name> -default-site AD-DC-Site (Ontap 9.4 – Specify
vserver cifs options modify -vserver <vserver_name> -is-large-mtu-enabled false (Ontap 9.x set to false due to bug I
cifs domain discovered-servers discovery-mode modify -vserver <vserver name> -mode sit
vserver cifs create -vserver <vserver_name> -cifs-server <node_name> -domain <domain_
vserver cifs share create -share-name root -path / (Create a CIFS share called root)
vserver cifs share show
vserver cifs show
SMB
vserver cifs options modify -vserver <vserver_name>-smb2-enabled true (Enable SMB2.0 a
Site (Ontap 9.4 – Specify a Active Directory Site)
ap 9.x set to false due to bug ID: Netapp Bug ID 1139257 )
server name> -mode site (Ontap 9.3 – Set Domain Controller discover to single site)
me> -domain <domain_name> (Enable Cifs)
hare called root)
d true (Enable SMB2.0 and 2.1)
DISK SHELVES
storage shelf show (an 8.3+ command that displays the loops and shelf information)
DISKS
storage disk assign -disk 0a.00.1 -owner <node_name> (Assign a specific disk to a node) O
storage disk assign -count <number_of_disks> -owner <node_name> (Assign unallocated d
storage disk show -ownership (Show disk ownership to nodes)
storage disk show -state broken | copy | maintenance | partner | percent | reconstructing | re
storage disk modify -disk NODE1:4c.10.0 -owner NODE1 -force-owner true (Force the chan
storage disk removeowner -disk NODE1:4c.10.0 -force true (Remove ownership of a drive)
storage disk set-led -disk Node1:4c.10.0 -action blink -time 5 (Blink the led of disk 4c.10.0 f
disk show -partition-ownership 0.00.10 (Show disk partition information on disk shelf 00 dis
You can then see the partition layout by typing ::> node run NODE1 vol status -r (this will display th
storage failover mailbox-disk show (show which disks are mailbox disks. Needs to be perfo
DEDUPE AND COMPRESSION
volume efficiency on -vserver SVM1 -volume volume1 (Turns Dedupe on for this volume)
volume efficiency start -vserver SVM1 -volume volume1 -dedupe true -scan-old-data true (S
volume efficiency start -vserver SVM1 -volume volume1 -dedupe true (Starts a volume effic
volume efficiency show -op-status !idle (This will display the running volume efficiency tasks
DNS
services dns show (show DNS)
vserver services name-service dns create -vserver vs1.example.com -domains example.c
DATE AND TIME
timezone -timezone Australia/Sydney (Sets the timezone for Sydney. Type ? after -timezon
date 201307090830 (Sets date for yyyymmddhhmm)
date -node <node_name> (Displays the date and time for the node)
shelf information)
specific disk to a node) OR
e> (Assign unallocated disks to a node)
cent | reconstructing | removed | spare | unfail |zeroing (Show the state of a disk)
ner true (Force the change of ownership of a disk)
ve ownership of a drive)
the led of disk 4c.10.0 for 5 minutes. Use the blinkoff action to turn it off)
tion on disk shelf 00 disk 10)
tatus -r (this will display the aggregate along with the disk partitions)
disks. Needs to be performed in advance mode ::> set adv)
pe on for this volume)
ue -scan-old-data true (Starts a volume efficiency dedupe job on volume1, scanning old data)
ue (Starts a volume efficiency dedupe job on volume1, not scanning old data)
g volume efficiency tasks)
om -domains example.com -name-servers 192.0.2.201,192.0.2.202 -state enabled
y. Type ? after -timezone for a list)
canning old data)
FLASH CACHE
system node run -node * options flexscale.enable on (Enabling Flash Cache on each node)
system node run -node * options flexscale.lopri_blocks on (Enabling Flash Cache on each
system node run -node * options flexscale.normal_data_blocks on (Enabling Flash Cache o
node run NODENAME stats show -p flexscale (fashcache configuration)
node run NODENAME stats show -p flexscale-access (display flash cache statistics)
FLASH POOL
storage aggregate modify -hybrid-enabled true (Change the AGGR to hybrid)
storage aggregate add-disks -disktype SSD (Add SSD disks to AGGR to begin creating a fl
priority hybrid-cache set volume1 read-cache=none write-cache=none (Within node shell an
FCP
storage show adapter (Show Physical FCP adapters)
fcp adapter modify -node NODENAME -adapter 0e -state down (Take port 0e offline)
node run <nodename>fcpadmin config (Shows the config of the adapters – Initiator or Targ
node run <nodename> -t target 0a (Changes port 0a from initiator or target – You must reb
vserver fcp ping-initiator (Ping check between initiator and lifs)
vserver fcp ping-igroup (Ping check between igroup and lifs)
sh Cache on each node)
g Flash Cache on each node)
Enabling Flash Cache on each node)
h cache statistics)
to hybrid)
GR to begin creating a flash pool)
one (Within node shell and diag mode disable read and write cache on volume1)
ke port 0e offline)
apters – Initiator or Target)
or target – You must reboot the node)
JOBS
job schedule cron create -name 15_minute_sched -minute 15
Job show (Displays list of jobs)
job expunge -id xxx (Deletes a stuck job, replace xxx with the job id)
LUNS
lun show -vserver <vserver_name> (Shows all luns belonging to this specific vserver)
lun modify -vserver <vserver_name> -space-allocation enabled -path <lun_path> (Turns on
lun geometry -vserver <vserver_name> path /vol/vol1/lun1 (Displays the lun geometry)
lun mapping add-reporting-nodes -vserver <vserver_name> -volume <vol name> -lun <lun
lun mapping show -vserver <vserver name> -volume <volume name> -fields reporting-node
Before performing a vol move containing a lun with reporting-nodes:
::> lun mapping show -fields reporting-nodes (if no reporting nodes are present you can go ahead an
::> lun mapping add-reporting-nodes -vserver VSERVER1 -volume VOLUME1 -lun LUN1 -igroup
Once the volume has moved you can then remove the existing reporting-nodes (remember reporting
::> lun mapping remove-reporting-nodes -vserver VSERVER1 -volume VOLUME1 -lun LUN1 -igr
License
system license show
s specific vserver)
th <lun_path> (Turns on space allocation so you can run lun reclaims via VAAI)
s the lun geometry)
e <vol name> -lun <lun path> -igroup <igroup name> -destination-aggregate <aggregate name> (Ad
e> -fields reporting-nodes (Show reporting nodes for a specific volume)
esent you can go ahead and move the vol. If there are reporting nodes present you must add the reporting node
UME1 -lun LUN1 -igroup IGROUP1 -destination-aggregate AGGR2
odes (remember reporting nodes with Selective Lun Mapping (SLM) should be just the HA pair owning the vo
OLUME1 -lun LUN1 -igroup IGROUP1 -remote-nodes true
e <aggregate name> (Adds the igroup as reporting nodes for the lun)
must add the reporting nodes of the destination aggregate)
the HA pair owning the volume)
NTP
system services ntp server create -node <node_name> -server <ntp_server> (Adds an NTP
system services ntp config modify -enabled true (Enable ntp)
system node date modify -timezone <Area/Location Timezone> (Sets timezone for Area/Lo
node date show (Show date on all nodes)
cluster time-service ntp server create (Configure an NTP server or multiple NTP servers)
cluster time-service ntp server show (Show the current NTP servers)
cluster time-service ntp server modify (Modify the NTP server list)
cluster time-service ntp server delete (Deletes an NTP server)
cluster time-service ntp server reset (Resets configuration, removes all existing NTP server
cluster date show (Displays the cluster date)
cluster date modify (Modify the cluster date)
NODES
system node rename -node <current_node_name> -newname <new_node_name>
system node reboot -node NODENAME -reason ENTER REASON (Reboot node with a giv
NFS
vserver <vserver_name> modify -4.1 -pnfs enabled (Enable pNFS. NOTE: Cannot coexist w
NIS
vserver services nis-domain create -vserver vserver1 -domain vmlab.local -active true -serv
vserver modify -vserver vserver1 -ns-switch nis-file (Name Service Switch referencing a file
vserver services nis-domain show
p_server> (Adds an NTP server to node_name)
ets timezone for Area/Location Timezone. i.e. Australia/Sydney)
multiple NTP servers)
s all existing NTP servers)
w_node_name>
(Reboot node with a given reason. NOTE: check ha policy)
NOTE: Cannot coexist with NFSv4)
b.local -active true -servers 10.10.10.1 (Create nis-domain called vmlab.local pointing to 10.10.10.1
Switch referencing a file)
al pointing to 10.10.10.1)
PERFORMANCE
show-periodic -object volume -instance volumename -node node1 -vserver vserver1 -count
statistics show-periodic 0object nfsv3 -instance vserver1 -counter nfsv3_ops|nfsv3_read_op
sysstat -x 1 (Shows counters for CPU, NFS, CIFS, FCP, WAFL)
statistics start -object top_client -sample-id sample1 -sort-key (Need to be in advanced mod
statistics show -object top_client -sample_id sample1 -tab -counter read_ops|write_ops (Ne
vserver vserver1 -counter total_ops|avg_latency|read_ops|read_latency (Show the specific counters
fsv3_ops|nfsv3_read_ops|nfsv3_write_ops|read_avg_latency|write_avg_latency (Shows the specific
d to be in advanced mode. Options are other_ops, read_data, read_ops, total_data, total_ops, write_
read_ops|write_ops (Need to be in advanced mode. This command is used in conjunction with the a
ow the specific counters for a volume)
ency (Shows the specific nfsv3 counters for a vserver)
al_data, total_ops, write_data, write_ops) (This command starts a collection)
in conjunction with the above command, you can replace the -counter options with any options from
ns with any options from above as well)
There are no entries matching your query.
quota policy rule create -vserver < > -policy-name < > -volume < > -type <tree
ROUTING GROUPS
network interface show-routing-group (show routing groups for all vservers)
network routing-groups show -vserver vserver1 (show routing groups for vserver1)
network routing-groups route create -vserver vserver1 -routing-group 10.1.1.0/24 -destinati
ping -lif-owner vserver1 -lif data1 -destination www.google.com (ping www.google.com via
ps for vserver1)
up 10.1.1.0/24 -destination 0.0.0.0/0 -gateway 10.1.1.1 (Creates a default route on vserver1)
g www.google.com via vserver1 using the data1 port)
ute on vserver1)
SERVICE PROCESSOR
system node image get -package http://1.1.1.1:8080/SP_FW.zip -replace-package true -no
system node image get -package http://webserver/306-02765_A0_SP_3.0.1P1_SP_FW.zip
system node service-processor image update -node node1 -package 306-02765_A0_SP_3
system node service-processor show (Show the service processor firmware levels of each
system node service-processor image update-progress show (Shows the progress of a firm
service-processor reboot-sp -node NODE1 (reboot the sp of node1)
service-processor reboot-sp -node NODE1
service-processor reboot-sp -node NODE1 -image primary|backup (reboot the SP into the p
SNAPSHOTS
volume snapshot create -vserver vserver1 -volume vol1 -snapshot snapshot1 (Create a sna
volume snapshot restore -vserver vserver1 -volume vol1 -snapshot snapshot1 (Restore a s
volume snapshot show -vserver vserver1 -volume vol1 (Show snapshots on vserver1 vol1)
snap autodelete show -vserver SVM1 -enabled true (Shows which volumes have autodelet
cluster1::*> statistics
aggregate cache catalog disk lif
lun namespace node port preset
qtree samples settings show show-periodic
start stop system top volume
vserver workload
::statistics*> vserver show -interval 5 -iterations 1 -max 25 -vserver svm1-cluster1
cluster1 : 6/28/2020 05:05:44
*Total Read Write Other Read Write Latency
Vserver Ops Ops Ops Ops (Bps) (Bps) (us)
------------- ------ ---- ----- ----- ----- ----- -------
svm1-cluster1 0 0 0 0 0 0 0
security login create -user-or-group-name test -application ssh -authentication-method
eplace-package true -node NODE1 (After downloading the SP firmware for your node model, rename
SP_3.0.1P1_SP_FW.zip -replace-package true (Copies the firmware file from the webserver into the
ge 306-02765_A0_SP_3.0.1P1_SP_FW.zip -update-type differential (Installs the firmware package t
firmware levels of each node in the cluster)
ws the progress of a firmware update on the Service Processor)
(reboot the SP into the primary or backup image)
snapshot1 (Create a snapshot on vserver1, vol1 called snapshot1)
snapshot1 (Restore a snapshot on vserver1, vol1 called snapshot1)
shots on vserver1 vol1)
volumes have autodelete enabled)
authentication-method password -role admin -is-ns-switch-group no -second-authentication-m
our node model, rename the package to SP_FW.zip and place it on a HTTP server such as HFS)
m the webserver into the mroot directory on the node. ONLY USE THIS UNDER DIRECTION FROM
the firmware package to node1. ONLY USE THIS UNDER DIRECTION FROM NETAPP SUPPOR
cond-authentication-method none -vserver cluster1
server such as HFS)
DER DIRECTION FROM NETAPP SUPPORT)
ROM NETAPP SUPPORT)
UNIX
vserver services unix-user show
vserver services unix-user create -vserver vserver1 -user root -id 0 -primary-gid 0 (Create a
vserver name-mapping create -vserver vserver1 -direction win-unix -position 1 -pattern (.+)
vserver name-mapping create -vserver vserver1 -direction unix-win -position 1 -pattern (.+)
vserver name-mapping show (Show name-mappings)
UCADMIN (CAN - Converged network adapter)
ucadmin show -node NODENAME (Show CNA ports on specific node)
ucadmin -node NODENAME -adapter 0e -mode cna (Change adapter 0e from FC to CNA.
-primary-gid 0 (Create a unix user called root)
-position 1 -pattern (.+) -replacement root (Create a name mapping from windows to unix)
-position 1 -pattern (.+) -replacement sysadmin011 (Create a name mapping from unix to windows)
ter 0e from FC to CNA. NOTE: A reboot of the node is required)
ndows to unix)
g from unix to windows)
VSERVER
vserver setup (Runs the clustered ontap vserver setup wizard)
vserver create -vserver <vserver_name> -rootvolume <volume_name> (Creates a new vse
vserver show (Shows all vservers in the system)
vserver show -vserver <vserver_name> (Show information on a specific vserver)
VOLUMES
volume create -vserver <vserver_name> -volume <volume_name> -aggregate <aggr_name
volume move -vserver <vserver_name> -volume <volume_name> -destination-aggregate <
volume move -vserver <vserver_name> -volume <volume_name> -destination-aggregate <
volume move trigger-cutover -vserver <vserver_name> -volume <volume_name> (Trigger a
volume move show (shows all volume moves currently active or waiting. NOTE: You can o
system node run – node <node_name> vol size <volume_name> 400g (resize volume_nam
volume size -volume <volume_name> -new-size 400g (resize volume_name to 400GB)
volume modify -vserver <vserver_name> -filesys-size-fixed false -volume <volume_name>
volume recovery-queue purge-all (An 8.3 command that purges the volume undelete cache
volume show -vserver SVM1 -volume * -autosize true (Shows which volumes have autosize
volume show -vserver SVM1 -volume * -atime-update true (Shows which volumes have upd
volume modify -vserver SVM1 -volume volume1 -atime-update false (Turns update access
volume move abort -vserver vs0 -volume vol1 Stop a running volume move operation
volume move modify Modify parameters for a running volume
volume move show -vserver vs0 -volume vol2 Show status of a volume moving from on
volume move start Start moving a volume from one aggrega
volume move trigger-cutover Trigger cutover of a move job
volume move recommend show Display Move Recommendations
volume move target-aggr show List target aggregates compatible for vol
vserver create -vserver vserver_name -rootvolume root_volume_name -aggregate aggre
vserver show -vserver vserver_name
me> (Creates a new vserver)
ecific vserver)
-aggregate <aggr_name> -size 100GB -junction-path /eng/p7/source (Creates a Volume within a vs
destination-aggregate <aggr_name> -foreground true (Moves a Volume to a different aggregate wit
destination-aggregate <aggr_name> -cutover-action wait (Moves a Volume to a different aggregate
olume_name> (Trigger a cutover of a volume move in waiting state)
iting. NOTE: You can only do 8 volume moves at one time, more than 8 and they get queued)
00g (resize volume_name to 400GB) OR
me_name to 400GB)
olume <volume_name> (Turn off fixed file sizing on volumes)
volume undelete cache)
h volumes have autosize enabled)
which volumes have update access time enabled)
e (Turns update access time off on the volume)
volume move operation
eters for a running volume move operation
a volume moving from one aggregate to another aggregate
volume from one aggregate to another aggregate
er of a move job
Recommendations
regates compatible for volume move
olume_name -aggregate aggregate_name -rootvolume-security-style unix -language C.UTF-8 -ipspace ipspace_name
es a Volume within a vserver)
a different aggregate with high priority)
to a different aggregate with low priority but does not cutover)
d they get queued)
pace ipspace_name