Skip to main content
ONTAP MetroCluster

Performing aggregate healing and restoring mirrors (MetroCluster FC configurations)

Contributors netapp-thomi netapp-pcarriga netapp-aoife netapp-folivia

After replacing hardware and assigning disks, you can perform the MetroCluster healing operations. You must then confirm that aggregates are mirrored and, if necessary, restart mirroring.

  1. Perform the two phases of healing (aggregate healing and root healing) on the disaster site:

    cluster_B::> metrocluster heal -phase aggregates
    cluster_B::> metrocluster heal -phase root-aggregates
  2. Monitor the healing and verify that the aggregates are in either the resyncing or mirrored state:

    storage aggregate show -node local

    If the aggregate shows this state…​



    No action is required. Let the aggregate complete resyncing.

    mirror degraded

    mirrored, normal

    No action is required.

    unknown, offline

    The root aggregate shows this state if all the disks on the disaster sites were replaced.

    cluster_B::> storage aggregate show -node local
     Aggregate     Size Available Used% State   #Vols  Nodes      RAID Status
     --------- -------- --------- ----- ------- ------ ---------- ------------
                227.1GB   11.00GB   95% online       1 node_B_1   raid_dp,
                430.3GB   28.02GB   93% online       2 node_B_1   raid_dp,
                812.8GB   85.37GB   89% online       5 node_B_1   raid_dp,
     3 entries were displayed.

    In the following examples, the three aggregates are each in a different state:






    mirror degraded


    mirrored, normal

  3. If one or more plexes remain offline, additional steps are required to rebuild the mirror.

    In the preceding table, the mirror for node_B_1_aggr2 must be rebuilt.

    1. View details of the aggregate to identify any failed plexes:

      storage aggregate show -r -aggregate node_B_1_aggr2

      In the following example, plex /node_B_1_aggr2/plex0 is in a failed state:

      cluster_B::> storage aggregate show -r -aggregate node_B_1_aggr2
       Owner Node: node_B_1
        Aggregate: node_B_1_aggr2 (online, raid_dp, mirror degraded) (block checksums)
         Plex: /node_B_1_aggr2/plex0 (offline, failed, inactive, pool0)
          RAID Group /node_B_1_aggr2/plex0/rg0 (partial)
                                                                     Usable Physical
            Position Disk                     Pool Type     RPM     Size     Size Status
            -------- ------------------------ ---- ----- ------ -------- -------- ----------
         Plex: /node_B_1_aggr2/plex1 (online, normal, active, pool1)
          RAID Group /node_B_1_aggr2/plex1/rg0 (normal, block checksums)
                                                                     Usable Physical
            Position Disk                     Pool Type     RPM     Size     Size Status
            -------- ------------------------ ---- ----- ------ -------- -------- ----------
            dparity  1.44.8                    1   SAS    15000  265.6GB  273.5GB (normal)
            parity   1.41.11                   1   SAS    15000  265.6GB  273.5GB (normal)
            data     1.42.8                    1   SAS    15000  265.6GB  273.5GB (normal)
            data     1.43.11                   1   SAS    15000  265.6GB  273.5GB (normal)
            data     1.44.9                    1   SAS    15000  265.6GB  273.5GB (normal)
            data     1.43.18                   1   SAS    15000  265.6GB  273.5GB (normal)
       6 entries were displayed.
    2. Delete the failed plex:

      storage aggregate plex delete -aggregate aggregate-name -plex plex

    3. Reestablish the mirror:

      storage aggregate mirror -aggregate aggregate-name

    4. Monitor the resynchronization and mirroring status of the plex until all mirrors are reestablished and all aggregates show mirrored, normal status:

      storage aggregate show