Install and boot node4

Contributors netapp-pcarriga Download PDF of this page

You must install node4 in the rack, transfer node2’s connections to node4, boot node4, and install ONTAP. You must then reassign any of node2’s spare disks, any disks belonging to the root volume, and any non-root aggregates that were not relocated to node3 earlier in the process, as outlined in this section.

About this task

The relocation operation is paused at the beginning of this stage. This process is mostly automated; the operation pauses to allow you to check its status. You must manually resume the operation. In addition, you must verify the NAS data LIFs have successfully moved to node4.

You need to netboot node4 if it does not have the same version of ONTAP 9 that is installed on node2. After you install node4, boot it from the ONTAP 9 image stored on the web server. You can then download the correct files to the boot media device for subsequent system boots, by following the instructions in Prepare for netboot.

Important:
Steps
  1. Make sure that node4 has sufficient rack space.

    If node4 is in a separate chassis from node2, you can put node4 in the same location as node3. If node2 and node4 are in the same chassis, then node4 is already in its appropriate rack location.

  2. Install node4 in the rack, following the instructions in the Installation and Setup Instructions for the node model.

  3. Cable node4, moving the connections from node2 to node4.

    Cable the following connections, using the instructions in the Installation and Setup Instructions or the FlexArray Virtualization Installation Requirements and Reference for the node4 platform, the appropriate disk shelf guide, and the ONTAP 9 High-Availability Configuration Guide.

    Refer to References to link to the FlexArray Virtualization Installation Requirements and Reference and the ONTAP 9 High-Availability Configuration Guide.

    • Console (remote management port)

    • Cluster ports

    • Data ports

    • Cluster and node management ports

    • Storage

    • SAN configurations: iSCSI Ethernet and FC switch ports

      You might not need to move the interconnect card/FC-VI card or interconnect/FC-VI cable connection from node2 to node4 because most platform models have unique interconnect card models.
      For the MetroCluster configuration, you must move the FC-VI cable connections from node2 to node4. If the new host does not have an FC-VI card, you might need to move the FC-VI card.
  4. Turn on the power to node4, and then interrupt the boot process by pressing Ctrl-C at the console terminal to access the boot environment prompt.

    When you boot node4, you might see the following warning message:
    WARNING: The battery is unfit to retain data during a power outage. This is likely
             because the battery is discharged but could be due to other temporary
             conditions.
             When the battery is ready, the boot process will complete
             and services will be engaged. To override this delay, press 'c' followed
             by 'Enter'
  5. If you see the warning message in Step 4, take the following actions:

    1. Check for any console messages that might indicate a problem other than a low NVRAM battery, and, if necessary, take any required corrective action.

    2. Allow the battery to charge and the boot process to complete.

      Attention: Do not override the delay; failure to allow the battery to charge could result in a loss of data.
  1. Configure the netboot connection by choosing one of the following actions.

    You should use the management port and IP as the netboot connection. Do not use a data LIF IP or a data outage might occur while the upgrade is being performed.
    If Dynamic Host Configuration Protocol (DHCP) is…​ Then…​

    Running

    Configure the connection automatically by entering the following command at the boot environment prompt:
    ifconfig e0M -auto

    Not running

    Manually configure the connection by entering the following command at the boot environment prompt:
    ifconfig e0M -addr=<filer_addr> -mask=<netmask> -gw=<gateway> dns=<dns_addr> domain=<dns_domain>
    <filer_addr> is the IP address of the storage system.
    <netmask> is the network mask of the storage system.
    <gateway> is the gateway for the storage system.
    <dns_addr> is the IP address of a name server on your network.
    <dns_domain> is the Domain Name Service (DNS) domain name. If you use this optional parameter, you do not need a fully qualified domain name in the netboot server URL; you need only the server’s host name.

    Note: Other parameters might be necessary for your interface. Enter help ifconfig at the firmware prompt for details.

  2. Perform netboot on node4:

    For…​ Then…​

    FAS/AFF8000 series systems

    netboot http://<web_server_ip/path_to_web-accessible_directory>/netboot/kernel

    All other systems

    netboot http://<web_server_ip/path_to_web-accessible_directory>/<ontap_version>_image.tgz

    The <path_to_the_web-accessible_directory> should lead to where you downloaded the <ontap_version>_image.tgz in Step 1 in the section Prepare for netboot.

    Do not interrupt the boot.
  3. From the boot menu, select option (7) Install new software first.

    This menu option downloads and installs the new ONTAP image to the boot device.

    Disregard the following message: This procedure is not supported for Non-Disruptive Upgrade on an HA pair
    The note applies to nondisruptive upgrades of ONTAP, and not upgrades of controllers.
    Always use netboot to update the new node to the desired image. If you use another method to install the image on the new controller, the wrong image might install. This issue applies to all ONTAP releases.
  4. If you are prompted to continue the procedure, enter y, and when prompted for the package, enter the URL:

    http://<web_server_ip/path_to_web-accessible_directory>/<ontap_version>_image.tgz

  5. Complete the following substeps to reboot the controller module:

    1. Enter n to skip the backup recovery when you see the following prompt:

      Do you want to restore the backup configuration now? {y|n}
    2. Reboot by entering y when you see the following prompt:

      The node must be rebooted to start using the newly installed software. Do you want to reboot now? {y|n}

      The controller module reboots but stops at the boot menu because the boot device was reformatted, and the configuration data needs to be restored.

  6. Select maintenance mode 5 from the boot menu and enter y when you are prompted to continue with the boot.

  7. Verify that the controller and chassis are configured as HA:

    ha-config show

    The following example shows the output of the ha-config show command:

    Chassis HA configuration: ha
    Controller HA configuration: ha
    System records in a PROM whether they are in an HA pair or stand-alone configuration. The state must be the same on all components within the stand-alone system or HA pair.
  8. If the controller and chassis are not configured as HA, use the following commands to correct the configuration:

    ha-config modify controller ha

    ha-config modify chassis ha

    If you have a MetroCluster configuration, use the following commands to modify the controller and chassis:

    ha-config modify controller mcc

    ha-config modify chassis mcc

  9. Exit maintenance mode:

    halt

    Interrupt the autoboot by pressing Ctrl-C at the boot environment prompt.

  10. On node3, check the system date, time, and time zone:

    date

  11. On node4, check the date by using the following command at the boot environment prompt:

    show date

  12. If necessary, set the date on node4:

    set date <mm/dd/yyyy>

  13. On node4, check the time by using the following command at the boot environment prompt:

    show time

  14. If necessary, set the time on node4:

    set time <hh:mm:ss>

  15. If necessary, set the partner system ID on node4:

    setenv partner- sysid <node2_sysid>

    1. Save the settings:

      saveenv

  16. On the new node, in boot loader, the partner- sysid parameter must be set. For node4, partner- sysid must be that of node3. Verify the partner- sysid for node3:

    printenv partner- sysid

  17. Take one of the following actions:

    If your system…​ Then…​

    Has disks and no back-end storage

    Go to Step 23.

    Is a V-Series system or a system with FlexArray Virtualization software connected to storage arrays

    1. Go to section Set the FC or UTA/UTA2 configuration on node4 and complete the subsections in this section.

    2. Return to this section and complete the remaining steps, beginning with Step 23.

    Important: You must reconfigure FC onboard ports, CNA onboard ports, and CNA cards before you boot ONTAP on the V-Series or system with FlexArray Virtualization software.

  1. Add the FC initiator ports of the new node to the switch zones.

    If required, modify the onboard ports to initiator by referring to the Configure FC ports on node4. See your storage array and zoning documentation for further instructions on zoning.

  2. Add the FC initiator ports to the storage array as new hosts, mapping the array LUNs to the new hosts.

    See your storage array and zoning documentation for instructions.

  3. Modify the worldwide port name (WWPN) values in the host or volume groups associated with array LUNs on the storage array.

    Installing a new controller module changes the WWPN values associated with each onboard FC port.

  4. If your configuration uses switch-based zoning, adjust the zoning to reflect the new WWPN values.

  5. If NetApp Storage Encryption (NSE) is in use on this configuration, the setenv bootarg.storageencryption.support command must be set to true, and the kmip.init.maxwait variable needs to be set to off to avoid a boot loop after the node2 configuration is loaded:

    setenv bootarg.storageencryption.support true

    setenv kmip.init.maxwait off

  6. Boot node into boot menu:

    boot_ontap menu

    If you do not have an FC or UTA/UTA2 configuration, execute Check and configure UTA/UTA2 ports on node4, Step 15 so that node4 can recognize node2’s disks.

  1. For MetroCluster configuration, V-Series systems and systems with FlexArray Virtualization software connected to storage arrays, go to Check and configure UTA/UTA2 ports on node4, Step 15.