Install and boot node3
You must install node3 in the rack, transfer node1's connections to node3, boot node3, and install ONTAP. You must also reassign any of node1's spare disks, any disks belonging to the root volume, and any non-root aggregates not relocated to node2 earlier.
You must netboot node3 if it does not have the same version of ONTAP 9 that is installed on node1. After you install node3, boot it from the ONTAP 9 image stored on the web server. You can then download the correct files to the boot media device for subsequent system boots. See Prepare for netboot.
However, you do not need to netboot node3 if it has the same or a later version of ONTAP 9 that is installed on node1.
If you are upgrading a V-Series system connected to storage arrays or a system with FlexArray Virtualization software that is connected to storage arrays, you must complete Step 1 through Step 5, leave this section at Step 6 and follow instructions in Configure FC ports on node3 and Check and configure UTA/UTA2 ports on node3 as needed, entering commands in maintenance mode. You must then return to this section and resume with Step 7. However, if you are upgrading a system with storage disks, you must complete this entire section and then go to Configure FC ports on node3 and Check and configure UTA/UTA2 ports on node3, entering commands at the cluster prompt. |
-
Make sure that you have rack space for node3.
If node1 and node2 were in separate chassis, you can put node3 in the same rack location as node1. However, if node1 was in the same chassis with node2, then you must put node3 into its own rack space, preferably close to the location of node1.
-
Install node3 in the rack, following the Installation and Setup Instructions for your node model.
If you are upgrading to a system with both nodes in the same chassis, install node4 in the chassis as well as node3. If you do not, when you boot node3, the node will behave as if it were in a dual-chassis configuration, and when you boot node4, the interconnect between the nodes will not come up. -
Cable node3, moving the connections from node1 to node3.
The following references help you make proper cable connections. Go to References to link to them.
-
Installation and Setup Instructions or FlexArray Virtualization Installation Requirements and Reference for the node3 platform
-
The appropriate disk shelf procedure
-
The High Availability management documentation
Cable the following connections:
-
Console (remote management port)
-
Cluster ports
-
Data ports
-
Cluster and node management ports
-
Storage
-
SAN configurations: iSCSI Ethernet and FC switch ports
You might not need to move the interconnect card or the cluster interconnect cable connection from node1 to node3 because most platform models have a unique interconnect card model. For the MetroCluster configuration, you must move the FC-VI cable connections from node1 to node3. If the new host does not have an FC-VI card, you might need to move the FC-VI card.
-
-
Turn on the power to node3, and then interrupt the boot process by pressing Ctrl-C at the console terminal to access the boot environment prompt.
If you are upgrading to a system with both nodes in the same chassis, node4 also reboots. However, you can disregard the node4 boot until later.
When you boot node3, you might see the following warning message: WARNING: The battery is unfit to retain data during a power outage. This is likely because the battery is discharged but could be due to other temporary conditions. When the battery is ready, the boot process will complete and services will be engaged. To override this delay, press 'c' followed by 'Enter'
-
If you see the warning message in Step 4, take the following actions:
-
Check for any console messages that might indicate a problem other than a low NVRAM battery, and, if necessary, take any required corrective action.
-
Allow the battery to charge and the boot process to complete.
Attention: Do not override the delay; failure to allow the battery to charge could result in a loss of data.
-
-
Take one of the following actions:
If your system… Then… Has disks and no back-end storage
Skip Step 7 through Step 12 and go to Step 13.
Is a V-Series system or a system with FlexArray Virtualization software connected to storage arrays
-
Go to Set the FC or UTA/UTA2 configuration on node3 and complete the subsections Configure FC ports on node3 and Check and configure UTA/UTA2 ports on node3, as appropriate to your system.
-
Return to this section and complete the remaining steps, beginning with Step 7.
You must reconfigure FC onboard ports, CNA onboard ports, and CNA cards before you boot ONTAP on the V-Series or system with FlexArray Virtualization software. -
-
Add the FC initiator ports of the new node to the switch zones.
If your system has a tape SAN, then you need zoning for the initiators. See your storage array and zoning documentation for instructions.
-
Add the FC initiator ports to the storage array as new hosts, mapping the array LUNs to the new hosts.
See your storage array and zoning documentation for instructions.
-
Modify the World Wide Port Name (WWPN) values in the host or volume groups associated with array LUNs on the storage array.
Installing a new controller module changes the WWPN values associated with each onboard FC port.
-
If your configuration uses switch-based zoning, adjust the zoning to reflect the new WWPN values.
-
Verify that the array LUNs are now visible to node3:
sysconfig -v
The system displays all the array LUNs visible to each of the FC initiator ports. If the array LUNs are not visible, you will not be able to reassign disks from node1 to node3 later in this section.
-
Press Ctrl-C to display the boot menu and select maintenance mode.
-
At the Maintenance mode prompt, enter the following command:
halt
The system stops at the boot environment prompt.
-
Take one of the following actions:
If the system you are upgrading to is in a… Then… Dual-chassis configuration (with controllers in different chassis)
Go to Step 15.
Single-chassis configuration (with controllers in the same chassis)
-
Switch the console cable from node3 to node4.
-
Turn on the power to node4, and then interrupt the boot process by pressing Ctrl-C at the console terminal to access the boot environment prompt.
The power should already be on if both controllers are in the same chassis.
Leave node4 at the boot environment prompt; you will return to node4 in Install and boot node4. -
If you see the warning message displayed in Step 4, follow the instructions in Step 5
-
Switch the console cable back from node4 to node3.
-
Go to Step 15.
-
-
set-defaults
-
If you have NetApp Storage Encryption (NSE) drives installed, perform the following steps.
If you have not already done so earlier in the procedure, see the Knowledge Base article How to tell if a drive is FIPS certified to determine the type of self-encrypting drives that are in use. -
Set
bootarg.storageencryption.support
totrue
orfalse
:If the following drives are in use… Then… NSE drives that conform to FIPS 140-2 Level 2 self-encryption requirements
setenv bootarg.storageencryption.support true
NetApp non-FIPS SEDs
setenv bootarg.storageencryption.support false
You cannot mix FIPS drives with other types of drives on the same node or HA pair.
You can mix SEDs with non-encrypting drives on the same node or HA pair.
-
Go to the special boot menu and select option
(10) Set Onboard Key Manager recovery secrets
.Enter the passphrase and the backup information that you recorded earlier procedure. See Manage authentication keys using the Onboard Key Manager.
-
-
If the version of ONTAP installed on node3 is the same or later than the version of ONTAP 9 installed on node1, list and reassign disks to the new node3:
boot_ontap
If this new node has ever been used in any other cluster or HA pair, you must run wipeconfig
before proceeding. Failure to do so might result in service outages or data loss. Contact technical support if the replacement controller was previously used, especially if the controllers were running ONTAP running in 7-Mode. -
Take one of the following actions:
If the system you are upgrading… Then… Does not have the correct or current ONTAP version on node3
Go to Step 20.
Has the correct or current version of ONTAP on node3
Go to Step 25.
-
Configure the netboot connection by choosing one of the following actions.
You must use the management port and IP as the netboot connection. Do not use a data LIF IP or else a data outage might occur while the upgrade is being performed. If Dynamic Host Configuration Protocol (DHCP) is… Then… Running
Configure the connection automatically by entering the following command at the boot environment prompt:
ifconfig e0M -auto
Not running
Manually configure the connection by entering the following command at the boot environment prompt:
ifconfig e0M -addr=filer_addr -mask=netmask -gw=gateway -dns=dns_addr -domain=dns_domain
filer_addr
is the IP address of the storage system (mandatory).
netmask
is the network mask of the storage system (mandatory).
gateway
is the gateway for the storage system (mandatory).
dns_addr
is the IP address of a name server on your network (optional).
dns_domain
is the Domain Name Service (DNS) domain name. If you use this optional parameter, you do not need a fully qualified domain name in the netboot server URL; you need only the server's host name.Other parameters might be necessary for your interface. Enter help ifconfig
at the firmware prompt for details. -
For… Then… FAS/AFF8000 series systems
netboot http://<web_server_ip>/<path_to_webaccessible_directory>/netboot/kernel
All other systems
netboot http://<web_server_ip>/<path_to_webaccessible_directory>/<ontap_version>_image.tgz
The
<path_to_the_web-accessible_directory>
leads to where you downloaded the<ontap_version>_image.tgz
in Step 1 in the section Prepare for netboot.Do not interrupt the boot. -
From the boot menu, select option (7) Install new software first.
This menu option downloads and installs the new ONTAP image to the boot device.
Disregard the following message:
This procedure is not supported for Non-Disruptive Upgrade on an HA pair
The note applies to nondisruptive upgrades of ONTAP, and not upgrades of controllers.
Always use netboot to update the new node to the desired image. If you use another method to install the image on the new controller, the wrong image might install. This issue applies to all releases of ONTAP. The netboot procedure combined with option (7) Install new software
wipes the boot media and places the same ONTAP version ONTAP on both image partitions. -
If you are prompted to continue the procedure, enter
y
, and when prompted for the package, enter the following URL:http://<web_server_ip>/<path_to_web-accessible_directory>/<ontap_version_image>.tgz
-
Complete the following substeps:
-
Enter
n
to skip the backup recovery when you see the following prompt:Do you want to restore the backup configuration now? {y|n}
-
Reboot by entering
y
when you see the following prompt:The node must be rebooted to start using the newly installed software. Do you want to reboot now? {y|n}
The controller module reboots but stops at the boot menu because the boot device was reformatted and the configuration data needs to be restored.
-
-
Select (5) Maintenance mode boot by entering
5
, and then entery
when prompted to continue with the boot. -
Before continuing, go to Set the FC or UTA/UTA2 configuration on node3 to make any necessary changes to the FC or UTA/UTA2 ports on the node.
Make the changes recommended in those sections, reboot the node, and go into maintenance mode.
-
disk show -a
The system displays the system ID of the node and information about its disks, as shown in the following example:
*> disk show -a Local System ID: 536881109 DISK OWNER POOL SERIAL HOME DR HOME NUMBER -------- ------------- ----- -------- ------------- ------------- 0b.02.23 nst-fas2520-2(536880939) Pool0 KPG2RK6F nst-fas2520-2(536880939) 0b.02.13 nst-fas2520-2(536880939) Pool0 KPG3DE4F nst-fas2520-2(536880939) 0b.01.13 nst-fas2520-2(536880939) Pool0 PPG4KLAA nst-fas2520-2(536880939) ...... 0a.00.0 (536881109) Pool0 YFKSX6JG (536881109) ......
You might see the message disk show: No disks match option -a.
after entering the command. This is not an error message so you can continue with the procedure. -
Reassign node1's spares, any disks belonging to the root, and any non-root aggregates that were not relocated to node2 earlier in Relocate non-root aggregates from node1 to node2.
Enter the appropriate form of the
disk reassign
command based on whether your system has shared disks:If you have shared disks, hybrid aggregates, or both on your system, you must use the correct disk reassign
command from the following table.If disk type is… Then run the command… With shared disks
disk reassign -s node1_sysid -d node3_sysid -p node2_sysid
Without shared disks
disk reassign -s node1_sysid -d node3_sysid
For the
node1_sysid
value, use the information captured in Record node1 information. To obtain the value fornode3_sysid
, use thesysconfig
command.The -p
option is only required in maintenance mode when shared disks are present.The
disk reassign
command reassigns only those disks for whichnode1_sysid
is the current owner.The system displays the following message:
Partner node must not be in Takeover mode during disk reassignment from maintenance mode. Serious problems could result!! Do not proceed with reassignment if the partner is in takeover mode. Abort reassignment (y/n)?
-
The system displays the following message:
After the node becomes operational, you must perform a takeover and giveback of the HA partner node to ensure disk reassignment is successful. Do you want to continue (y/n)?
-
The system displays the following message:
Disk ownership will be updated on all disks previously belonging to Filer with sysid <sysid>. Do you want to continue (y/n)?
-
If you are upgrading from a system with external disks to a system that supports internal and external disks (AFF A800 systems, for example), set the node1 aggregate as root to confirm that node3 boots from the root aggregate of node1.
Warning: You must perform the following substeps in the exact order shown; failure to do so might cause an outage or even data loss. The following procedure sets node3 to boot from the root aggregate of node1:
-
Check the RAID, plex, and checksum information for the node1 aggregate:
aggr status -r
-
Check the status of the node1 aggregate:
aggr status
-
Bring the node1 aggregate online, if necessary:
aggr_online root_aggr_from_node1
-
Prevent the node3 from booting from its original root aggregate:
aggr offline root_aggr_on_node3
-
Set the node1 root aggregate as the new root aggregate for node3:
aggr options aggr_from_node1 root
-
Verify that the root aggregate of node3 is offline and the root aggregate for the disks brought over from node1 is online and set to root:
aggr status
Failing to perform the previous substep might cause node3 to boot from the internal root aggregate, or it might cause the system to assume a new cluster configuration exists or prompt you to identify one. The following shows an example of the command output:
--------------------------------------------------------------- Aggr State Status Options aggr0_nst_fas8080_15 online raid_dp, aggr root, nosnap=on fast zeroed 64-bit aggr0 offline raid_dp, aggr diskroot fast zeroed 64-bit ----------------------------------------------------------------------
-
-
Verify that the controller and chassis are configured as
ha
:ha-config show
The following example shows the output of the ha-config show command:
*> ha-config show Chassis HA configuration: ha Controller HA configuration: ha
Systems record in a programmable ROM (PROM) whether they are in an HA pair or stand-alone configuration. The state must be the same on all components within the stand-alone system or HA pair.
If the controller and chassis are not configured as "ha", use the following commands to correct the configuration:
ha-config modify controller ha
ha-config modify chassis ha
If you have a MetroCluster configuration, use the following commands to modify the controller and chassis:
ha-config modify controller mcc
ha-config modify chassis mcc
-
Destroy the mailboxes on node3:
mailbox destroy local
The console displays the following message:
Destroying mailboxes forces a node to create new empty mailboxes, which clears any takeover state, removes all knowledge of out-of-date plexes of mirrored volumes, and will prevent management services from going online in 2-node cluster HA configurations. Are you sure you want to destroy the local mailboxes?
-
Enter
y
at the prompt to confirm that you want to destroy the local mailboxes. -
halt
The system stops at the boot environment prompt.
-
On node2, check the system date, time, and time zone:
date
-
On node3, check the date at the boot environment prompt:
show date
-
If necessary, set the date on node3:
set date mm/dd/yyyy
-
On node3, check the time at the boot environment prompt:
show time
-
If necessary, set the time on node3:
set time hh:mm:ss
-
Verify the partner system ID is set correctly as noted in Step 28 under -p switch:
printenv partner-sysid
-
If necessary, set the partner system ID on node3:
setenv partner-sysid node2_sysid
Save the settings:
saveenv
-
Access the boot menu at the boot environment prompt:
boot_ontap menu
-
At the boot menu, select option (6) Update flash from backup config by entering
6
at the prompt.The system displays the following message:
This will replace all flash-based configuration with the last backup to disks. Are you sure you want to continue?:
-
The boot proceeds normally, and the system then asks you to confirm the system ID mismatch.
The system might reboot twice before displaying the mismatch warning. -
Confirm the mismatch as shown in the following example:
WARNING: System id mismatch. This usually occurs when replacing CF or NVRAM cards! Override system id (y|n) ? [n] y
The node might go through one round of reboot before booting normally.