Correct misaligned spare partitions
When you add partitioned disks to a local tier (aggregate), you must leave a disk with both the root and data partition available as a spare for every node. If you do not and your node experiences a disruption, ONTAP cannot dump the core to the spare data partition.
You must have both a spare data partition and a spare root partition on the same type of disk owned by the same node.
-
Using the CLI, display the spare partitions for the node:
storage aggregate show-spare-disks -original-owner node_name
Note which disk has a spare data partition (spare_data) and which disk has a spare root partition (spare_root). The spare partition will show a non-zero value under the
Local Data Usable
orLocal Root Usable
column. -
Replace the disk with a spare data partition with the disk with the spare root partition:
storage disk replace -disk spare_data -replacement spare_root -action start
You can copy the data in either direction; however, copying the root partition takes less time to complete.
-
Monitor the progress of the disk replacement:
storage aggregate show-status -aggregate aggr_name
-
After the replacement operation is complete, display the spares again to confirm that you have a full spare disk:
storage aggregate show-spare-disks -original-owner node_name
You should see a spare disk with usable space under both “Local Data Usable” and
Local Root Usable
.
You display your spare partitions for node c1-01 and see that your spare partitions are not aligned:
c1::> storage aggregate show-spare-disks -original-owner c1-01 Original Owner: c1-01 Pool0 Shared HDD Spares Local Local Data Root Physical Disk Type RPM Checksum Usable Usable Size ------- ----- ---- -------- ------- ------- -------- 1.0.1 BSAS 7200 block 753.8GB 0B 828.0GB 1.0.10 BSAS 7200 block 0B 73.89GB 828.0GB
You start the disk replacement job:
c1::> storage disk replace -disk 1.0.1 -replacement 1.0.10 -action start
While you are waiting for the replacement operation to finish, you display the progress of the operation:
c1::> storage aggregate show-status -aggregate aggr0_1 Owner Node: c1-01 Aggregate: aggr0_1 (online, raid_dp) (block checksums) Plex: /aggr0_1/plex0 (online, normal, active, pool0) RAID Group /aggr0_1/plex0/rg0 (normal, block checksums) Usable Physical Position Disk Pool Type RPM Size Size Status -------- ------- ---- ---- ----- -------- -------- ---------- shared 1.0.1 0 BSAS 7200 73.89GB 828.0GB (replacing,copy in progress) shared 1.0.10 0 BSAS 7200 73.89GB 828.0GB (copy 63% completed) shared 1.0.0 0 BSAS 7200 73.89GB 828.0GB (normal) shared 1.0.11 0 BSAS 7200 73.89GB 828.0GB (normal) shared 1.0.6 0 BSAS 7200 73.89GB 828.0GB (normal) shared 1.0.5 0 BSAS 7200 73.89GB 828.0GB (normal)
After the replacement operation is complete, confirm that you have a full spare disk:
ie2220::> storage aggregate show-spare-disks -original-owner c1-01 Original Owner: c1-01 Pool0 Shared HDD Spares Local Local Data Root Physical Disk Type RPM Checksum Usable Usable Size ------ ----- ---- -------- -------- ------- -------- 1.0.1 BSAS 7200 block 753.8GB 73.89GB 828.0GB