Replace a DIMM - AFF A700
You must replace a DIMM in the controller when your storage system encounters errors such as, excessive CECC (Correctable Error Correction Codes) errors that are based on Health Monitor alerts or uncorrectable ECC errors, typically caused by a single DIMM failure preventing the storage system from booting ONTAP.
All other components in the system must be functioning properly; if not, you must contact technical support.
You must replace the failed component with a replacement FRU component you received from your provider.
Step 1: Shut down the impaired controller
You can shut down or take over the impaired controller using different procedures, depending on the storage system hardware configuration.
To shut down the impaired controller, you must determine the status of the controller and, if necessary, take over the controller so that the healthy controller continues to serve data from the impaired controller storage.
-
If you have a SAN system, you must have checked event messages (
cluster kernel-service show
) for the impaired controller SCSI blade. Thecluster kernel-service show
command (from priv advanced mode) displays the node name, quorum status of that node, availability status of that node, and operational status of that node.Each SCSI-blade process should be in quorum with the other nodes in the cluster. Any issues must be resolved before you proceed with the replacement.
-
If you have a cluster with more than two nodes, it must be in quorum. If the cluster is not in quorum or a healthy controller shows false for eligibility and health, you must correct the issue before shutting down the impaired controller; see Synchronize a node with the cluster.
-
If AutoSupport is enabled, suppress automatic case creation by invoking an AutoSupport message:
system node autosupport invoke -node * -type all -message MAINT=<# of hours>h
The following AutoSupport message suppresses automatic case creation for two hours:
cluster1:> system node autosupport invoke -node * -type all -message MAINT=2h
-
Disable automatic giveback from the console of the healthy controller:
storage failover modify -node local -auto-giveback false
When you see Do you want to disable auto-giveback?, enter y
. -
Take the impaired controller to the LOADER prompt:
If the impaired controller is displaying… Then… The LOADER prompt
Go to the next step.
Waiting for giveback…
Press Ctrl-C, and then respond
y
when prompted.System prompt or password prompt
Take over or halt the impaired controller from the healthy controller:
storage failover takeover -ofnode impaired_node_name
When the impaired controller shows Waiting for giveback…, press Ctrl-C, and then respond
y
.
To shut down the impaired controller, you must determine the status of the controller and, if necessary, switch over the controller so that the healthy controller continues to serve data from the impaired controller storage.
-
You must leave the power supplies turned on at the end of this procedure to provide power to the healthy controller.
-
Check the MetroCluster status to determine whether the impaired controller has automatically switched over to the healthy controller:
metrocluster show
-
Depending on whether an automatic switchover has occurred, proceed according to the following table:
If the impaired controller… Then… Has automatically switched over
Proceed to the next step.
Has not automatically switched over
Perform a planned switchover operation from the healthy controller:
metrocluster switchover
Has not automatically switched over, you attempted switchover with the
metrocluster switchover
command, and the switchover was vetoedReview the veto messages and, if possible, resolve the issue and try again. If you are unable to resolve the issue, contact technical support.
-
Resynchronize the data aggregates by running the
metrocluster heal -phase aggregates
command from the surviving cluster.controller_A_1::> metrocluster heal -phase aggregates [Job 130] Job succeeded: Heal Aggregates is successful.
If the healing is vetoed, you have the option of reissuing the
metrocluster heal
command with the-override-vetoes
parameter. If you use this optional parameter, the system overrides any soft vetoes that prevent the healing operation. -
Verify that the operation has been completed by using the metrocluster operation show command.
controller_A_1::> metrocluster operation show Operation: heal-aggregates State: successful Start Time: 7/25/2016 18:45:55 End Time: 7/25/2016 18:45:56 Errors: -
-
Check the state of the aggregates by using the
storage aggregate show
command.controller_A_1::> storage aggregate show Aggregate Size Available Used% State #Vols Nodes RAID Status --------- -------- --------- ----- ------- ------ ---------------- ------------ ... aggr_b2 227.1GB 227.1GB 0% online 0 mcc1-a2 raid_dp, mirrored, normal...
-
Heal the root aggregates by using the
metrocluster heal -phase root-aggregates
command.mcc1A::> metrocluster heal -phase root-aggregates [Job 137] Job succeeded: Heal Root Aggregates is successful
If the healing is vetoed, you have the option of reissuing the
metrocluster heal
command with the -override-vetoes parameter. If you use this optional parameter, the system overrides any soft vetoes that prevent the healing operation. -
Verify that the heal operation is complete by using the
metrocluster operation show
command on the destination cluster:mcc1A::> metrocluster operation show Operation: heal-root-aggregates State: successful Start Time: 7/29/2016 20:54:41 End Time: 7/29/2016 20:54:42 Errors: -
-
On the impaired controller module, disconnect the power supplies.
Step 2: Remove the controller module
To access components inside the controller, you must first remove the controller module from the system and then remove the cover on the controller module.
-
If you are not already grounded, properly ground yourself.
-
Unplug the cables from the impaired controller module, and keep track of where the cables were connected.
-
Slide the orange button on the cam handle downward until it unlocks.
Cam handle release button
Cam handle
-
Rotate the cam handle so that it completely disengages the controller module from the chassis, and then slide the controller module out of the chassis.
Make sure that you support the bottom of the controller module as you slide it out of the chassis.
-
Place the controller module lid-side up on a stable, flat surface, press the blue button on the cover, slide the cover to the back of the controller module, and then swing the cover up and lift it off of the controller module.
Controller module cover locking button
Step 3: Replace the DIMMs
To replace the DIMMs, locate them inside the controller and follow the specific sequence of steps.
-
If you are not already grounded, properly ground yourself.
-
Locate the DIMMs on your controller module.
-
Eject the DIMM from its slot by slowly pushing apart the two DIMM ejector tabs on either side of the DIMM, and then slide the DIMM out of the slot.
Carefully hold the DIMM by the edges to avoid pressure on the components on the DIMM circuit board. DIMM ejector tabs
DIMM
-
Remove the replacement DIMM from the antistatic shipping bag, hold the DIMM by the corners, and align it to the slot.
The notch among the pins on the DIMM should line up with the tab in the socket.
-
Make sure that the DIMM ejector tabs on the connector are in the open position, and then insert the DIMM squarely into the slot.
The DIMM fits tightly in the slot, but should go in easily. If not, realign the DIMM with the slot and reinsert it.
Visually inspect the DIMM to verify that it is evenly aligned and fully inserted into the slot. -
Push carefully, but firmly, on the top edge of the DIMM until the ejector tabs snap into place over the notches at the ends of the DIMM.
-
Close the controller module cover.
Step 4: Install the controller
After you install the components into the controller module, you must install the controller module back into the system chassis and boot the operating system.
For HA pairs with two controller modules in the same chassis, the sequence in which you install the controller module is especially important because it attempts to reboot as soon as you completely seat it in the chassis.
-
If you are not already grounded, properly ground yourself.
-
If you have not already done so, replace the cover on the controller module.
-
Align the end of the controller module with the opening in the chassis, and then gently push the controller module halfway into the system.
Do not completely insert the controller module in the chassis until instructed to do so. -
Cable the management and console ports only, so that you can access the system to perform the tasks in the following sections.
You will connect the rest of the cables to the controller module later in this procedure. -
Complete the reinstallation of the controller module:
-
If you have not already done so, reinstall the cable management device.
-
Firmly push the controller module into the chassis until it meets the midplane and is fully seated.
The locking latches rise when the controller module is fully seated.
Do not use excessive force when sliding the controller module into the chassis to avoid damaging the connectors. The controller module begins to boot as soon as it is fully seated in the chassis.
-
Rotate the locking latches upward, tilting them so that they clear the locking pins, and then lower them into the locked position.
-
Step 5: Switch back aggregates in a two-node MetroCluster configuration
After you have completed the FRU replacement in a two-node MetroCluster configuration, you can perform the MetroCluster switchback operation. This returns the configuration to its normal operating state, with the sync-source storage virtual machines (SVMs) on the formerly impaired site now active and serving data from the local disk pools.
This task only applies to two-node MetroCluster configurations.
-
Verify that all nodes are in the
enabled
state:metrocluster node show
cluster_B::> metrocluster node show DR Configuration DR Group Cluster Node State Mirroring Mode ----- ------- -------------- -------------- --------- -------------------- 1 cluster_A controller_A_1 configured enabled heal roots completed cluster_B controller_B_1 configured enabled waiting for switchback recovery 2 entries were displayed.
-
Verify that resynchronization is complete on all SVMs:
metrocluster vserver show
-
Verify that any automatic LIF migrations being performed by the healing operations were completed successfully:
metrocluster check lif show
-
Perform the switchback by using the
metrocluster switchback
command from any node in the surviving cluster. -
Verify that the switchback operation has completed:
metrocluster show
The switchback operation is still running when a cluster is in the
waiting-for-switchback
state:cluster_B::> metrocluster show Cluster Configuration State Mode -------------------- ------------------- --------- Local: cluster_B configured switchover Remote: cluster_A configured waiting-for-switchback
The switchback operation is complete when the clusters are in the
normal
state.:cluster_B::> metrocluster show Cluster Configuration State Mode -------------------- ------------------- --------- Local: cluster_B configured normal Remote: cluster_A configured normal
If a switchback is taking a long time to finish, you can check on the status of in-progress baselines by using the
metrocluster config-replication resync-status show
command. -
Reestablish any SnapMirror or SnapVault configurations.
Step 6: Return the failed part to NetApp
Return the failed part to NetApp, as described in the RMA instructions shipped with the kit. See the Part Return and Replacements page for further information.