Complete the restoration and replacement process - AFF A700 and FAS9000

Contributors netapp-martyh thrisun dougthomp

You must verify the HA state of the chassis, run diagnostics, and return the failed part to NetApp, as described in the RMA instructions shipped with the kit.

Step 1: Verify and set the HA state of the chassis

You must verify the HA state of the chassis, and, if necessary, update the state to match your system configuration.

  1. In Maintenance mode, from either controller module, display the HA state of the local controller module and chassis: ha-config show

    The HA state should be the same for all components.

  2. If the displayed system state for the chassis does not match your system configuration:

    1. Set the HA state for the chassis: ha-config modify chassis HA-state

      The value for HA-state can be one of the following:

      • ha

      • mcc

      • mcc-2n

      • mccip

      • non-ha

    2. Confirm that the setting has changed: ha-config show

  3. If you have not already done so, recable the rest of your system.

  4. Exit Maintenance mode: halt

    The LOADER prompt appears.

Step 2: Running system-level diagnostics

After installing a new chassis, you should run interconnect diagnostics.

Your system must be at the LOADER prompt to start System Level Diagnostics.

All commands in the diagnostic procedures are issued from the node where the component is being replaced.

  1. If the node to be serviced is not at the LOADER prompt, perform the following steps:

    1. Select the Maintenance mode option from the displayed menu.

    2. After the node boots to Maintenance mode, halt the node: halt

      After you issue the command, you should wait until the system stops at the LOADER prompt.

      Note During the boot process, you can safely respond y to prompts:
  2. Repeat the previous step on the second node if you are in an HA configuration.

    Note Both controllers must be in Maintenance mode to run the interconnect test.
  3. At the LOADER prompt, access the special drivers specifically designed for system-level diagnostics to function properly: boot_diags

    During the boot process, you can safely respond y to the prompts until the Maintenance mode prompt (*>) appears.

  4. Enable the interconnect diagnostics tests from the Maintenance mode prompt: sldiag device modify -dev interconnect -sel enable

    The interconnect tests are disabled by default and must be enabled to run separately.

  5. Run the interconnect diagnostics test from the Maintenance mode prompt: sldiag device run -dev interconnect

    You only need to run the interconnect test from one controller.

  6. Verify that no hardware problems resulted from the replacement of the chassis: sldiag device status -dev interconnect -long -state failed

    System-level diagnostics returns you to the prompt if there are no test failures, or lists the full status of failures resulting from testing the component.

  7. Proceed based on the result of the preceding step.

    If the system-level diagnostics tests…​ Then…​

    Were completed without any failures

    1. Clear the status logs: sldiag device clearstatus

    2. Verify that the log was cleared: sldiag device status

      The following default response is displayed:

      SLDIAG: No log messages are present.
    3. Exit Maintenance mode on both controllers: halt

      The system displays the LOADER prompt.

      Note You must exit Maintenance mode on both controllers before proceeding any further.
    4. Enter the following command on both controllers at the LOADER prompt: bye

    5. Return the node to normal operation:

    With two nodes in the cluster

    Issue these commands: node::> cluster ha modify -configured true

    node::> storage failover modify -node node0 -enabled true

    With more than two nodes in the cluster

    Issue this command:node::> storage failover modify -node node0 -enabled true

    In a two-node MetroCluster configuration

    Proceed to the next step.

    The MetroCluster switchback procedure is done in the next task in the replacement process.

    In a stand-alone configuration

    You have no further steps in this particular task.

    You have completed system-level diagnostics.

    Resulted in some test failures

    Determine the cause of the problem.

    1. Exit Maintenance mode: halt

    2. Perform a clean shutdown, and then disconnect the power supplies.

    3. Verify that you have observed all of the considerations identified for running system-level diagnostics, that cables are securely connected, and that hardware components are properly installed in the storage system.

    4. Reconnect the power supplies, and then power on the storage system.

    5. Rerun the system-level diagnostics test.

Step 3: Switch back aggregates in a two-node MetroCluster configuration

After you have completed the FRU replacement in a two-node MetroCluster configuration, you can perform the MetroCluster switchback operation. This returns the configuration to its normal operating state, with the sync-source storage virtual machines (SVMs) on the formerly impaired site now active and serving data from the local disk pools.

This task only applies to two-node MetroCluster configurations.

  1. Verify that all nodes are in the enabled state: metrocluster node show

    cluster_B::>  metrocluster node show
    DR                           Configuration  DR
    Group Cluster Node           State          Mirroring Mode
    ----- ------- -------------- -------------- --------- --------------------
    1     cluster_A
                  controller_A_1 configured     enabled   heal roots completed
                  controller_B_1 configured     enabled   waiting for switchback recovery
    2 entries were displayed.
  2. Verify that resynchronization is complete on all SVMs: metrocluster vserver show

  3. Verify that any automatic LIF migrations being performed by the healing operations were completed successfully: metrocluster check lif show

  4. Perform the switchback by using the metrocluster switchback command from any node in the surviving cluster.

  5. Verify that the switchback operation has completed: metrocluster show

    The switchback operation is still running when a cluster is in the waiting-for-switchback state:

    cluster_B::> metrocluster show
    Cluster              Configuration State    Mode
    --------------------	------------------- 	---------
     Local: cluster_B configured       	switchover
    Remote: cluster_A configured       	waiting-for-switchback

    The switchback operation is complete when the clusters are in the normal state.:

    cluster_B::> metrocluster show
    Cluster              Configuration State    Mode
    --------------------	------------------- 	---------
     Local: cluster_B configured      		normal
    Remote: cluster_A configured      		normal

    If a switchback is taking a long time to finish, you can check on the status of in-progress baselines by using the metrocluster config-replication resync-status show command.

  6. Reestablish any SnapMirror or SnapVault configurations.

Step 4: Return the failed part to NetApp

Return the failed part to NetApp, as described in the RMA instructions shipped with the kit. See the Part Return & Replacements page for further information.