NetApp

From roonics
Revision as of 09:48, 17 October 2022 by Jlambert (talk | contribs) (→‎Shutdown)
Jump to navigation Jump to search

CLI

Aggregate's

Show aggregate's

storage aggregate show

Example output:

Aggregate     Size Available Used% State   #Vols  Nodes            RAID Status
--------- -------- --------- ----- ------- ------ ---------------- ------------
netapp01_01_0_root
            2.91TB    1.52TB   48% online       1 netapp01-01    raid_dp,
                                                                   normal
netapp01_01_1_esxi
           37.83TB    6.38TB   83% online      13 netapp01-02    raid_dp,
                                                                   normal

Snap mirror

Show snapmirror progress

snapm show -fields total-progress

Example output:

source-path                            destination-path                              total-progress
-------------------------------------- --------------------------------------------- --------------
filersrv001_esxh:vol43_cluster003    filersrv002_esxh:vol43_cluster003_mirror    7.44GB
filersrv001_esxh:vol46_cluster001    filersrv002_esxh:vol46_cluster001_mirror    131.3GB
filersrv001_esxh:vol47_cluster001    filersrv003_exch:vol47_cluster001_mirror    -
filersrv001_esxh:vol49_cluster001    filersrv002_esxh:vol49_cluster001_mirror    -
filersrv001_esxh:vol51_cluster003    filersrv002_esxh:vol51_cluster003_mirror    104.1GB

Show snapmirror for volumes in transferring state

snapmirror show -status transferring

Example output:

Source            Destination Mirror  Relationship   Total             Last
Path        Type  Path        State   Status         Progress  Healthy Updated
----------- ---- ------------ ------- -------------- --------- ------- --------
filersrv001_esxh:vol43_cluster001
            DP   filersrv002_esxh:vol43_cluster001_mirror
                              Snapmirrored
                                      Transferring   8.19GB    false   06/19 15:14:34

Misc

Show number of inodes

vvolume show -vserver testsvm001 -volume testvolume01 -fields files

Example output:

vserver           volume        files
----------------- ------------- --------
testsvm001        testvolume01  84999981

Increase number of inodes

volume modify -vserver testsvm001 -volume testvolume01 <number>

Cluster shutdown

Shutdown

1. SSH To the cluster address

2. Set a maintenance window:

system node autosupport invoke -node * -type all -message "MAINT=8h Power Maintenance"

3. Get a list of all the SP's and their IP's:

system service-processor show -node * -fields address

4. Exit the cluster SSH session

5. Open up SSH sessions to all nodes in the cluster using the admin credentials

6. Enter to system console mode:

system console

7. In one of the SSH sessions halt all the nodes in the cluster:

system node halt -node * -skip-lif-migration-before-shutdown true -ignore-quorum-warnings true -inhibit-takeover true

For clusters with SnapMirror Synchronous operating in StrictSync mode:

system node halt -node * -skip-lif-migration-before-shutdown true -ignore-quorum-warnings true -inhibit-takeover true -ignore-strict-sync-warnings true

8. Respond to all the questions about confirming if you want to stop node XXXXX with y

9. Wait until all SSH sessions are at the LOADER-A> prompt ‎<comments />