Skip to main content

ARL upgrade workflow

Contributors netapp-pcarriga netapp-aoife netapp-aherbin

Before you upgrade the nodes using ARL, you should understand how the procedure works. In this document, the procedure is broken down into several stages.

Upgrade the node pair

To upgrade the node pair, you must prepare the original nodes and then perform a series of steps on both the original and new nodes. You can then decommission the original nodes.

ARL upgrade sequence overview

During the procedure, you upgrade the original controller hardware with the replacement controller hardware, one controller at a time, taking advantage of the HA pair configuration to relocate the ownership of non-root aggregates. All non-root aggregates must undergo two relocations to reach their final destination, which is the correct upgraded node.

Each aggregate has a home owner and current owner. The home owner is the actual owner of the aggregate, and the current owner is the temporary owner.

The following illustration shows the stages of the procedure. The thick, light gray arrows represent the relocation of aggregates and the movement of LIFs, and the thinner black arrows represent the removal of the original nodes. The smaller controller images represent the original nodes, and the larger controller images represent the new nodes.

Illustration showing the stages of the ARL procedure

The following table describes the high-level tasks you perform during each stage and the state of aggregate ownership at the end of the stage. Detailed steps are provided later in the procedure:

Stage Description

Stage 1: Prepare for upgrade

During Stage 1, if required, you confirm that internal disk drives do not contain root aggregates or data aggregates, prepare the nodes for the upgrade, and run a series of prechecks. If required, you rekey disks for Storage Encryption and prepare to netboot the new controllers.

Aggregate ownership at the end of Stage 1:

  • Node1 is the home owner and current owner of the node1 aggregates.

  • Node2 is the home owner and current owner of the node2 aggregates.

Stage 2: Retire node1

During Stage 2, you relocate non-root aggregates from node1 to node2 and move non-SAN data LIFs owned by node1 to node2, including failed or vetoed aggregates. You record the necessary node1 information for use later in the procedure and then retire node1.

Aggregate ownership at the end of Stage 2:

  • Node1 is the home owner of node1 aggregates.

  • Node2 is the current owner of node1 aggregates.

  • Node2 is the home owner and current owner of node2 aggregates.

Stage 3: Install and boot node3

During Stage 3, you install and boot node3, map the cluster and node-management ports from node1 to node3, verify the node3 installation, and move data LIFs and SAN LIFs belonging to node1 from node2 to node3. You also relocate all aggregates from node2 to node3 and move the data LIFs and SAN LIFs owned by node2 to node3.

Aggregate ownership at the end of Stage 3:

  • Node2 is the home owner of node2 aggregates but not the current owner.

  • Node3 is the home owner and current owner of aggregates originally belonging to node1.

  • Node2 is the home owner and current owner of aggregates belonging to node2 but not the home owner.

Stage 4: Retire node2

During Stage 4, you record the necessary node2 information for use later in the procedure and then retire node2.

No changes occur in aggregate ownership.

Stage 5: Install and boot node4

During Stage 5, you install and boot node4, map the cluster and node-management ports from node2 to node4, verify the node4 installation, and move data LIFs and SAN LIFs belonging to node2 from node3 to node4. You also relocate node2 aggregates from node3 to node4 and move the data node2 NAS LIFs from node3 to node4.

Aggregate ownership at the end of Stage 5:

  • Node3 is the home owner and current owner of the aggregates that originally belonged to node1.

  • Node4 is the home owner and current owner of aggregates that originally belonged to node2.

Stage 6: Complete the upgrade

During Stage 6, you confirm that the new nodes are set up correctly and set up Storage Encryption or NetApp Volume Encryption if the new nodes are encryption-enabled. You should also decommission the old nodes and resume SnapMirror operations.

Note The storage virtual machine (SVM) disaster recovery updates will not be interrupted as per the schedules assigned.

No changes occur in aggregate ownership.