Migrate to a two-node switched cluster in FAS22xx systems with a single cluster-network connection
If you have FAS22xx systems in an existing two-node switchless cluster in which each controller module has a single, back-to-back 10 GbE connection for cluster connectivity, you can use the switchless cluster networking option and replace the direct back-to-back connectivity with switch connections.
Review requirements
Make sure you have the following:
-
Two cluster connections for migrating from a switchless configuration to a switched configuration.
-
The cluster is healthy and consists of two nodes connected with back-to-back connectivity.
-
The nodes are running ONTAP 8.2 or later.
-
The switchless cluster feature cannot be used with more than two nodes.
-
All cluster ports are in the
upstate.
Migrate the switches
This procedure is a nondisruptive procedure that removes the direct cluster connectivity in a switchless environment and replaces each connection to the switch with a connection to the partner node.
Step 1: Prepare for migration
-
Change the privilege level to advanced, entering
ywhen prompted to continue:set -privilege advancedThe advanced prompt (
*>) appears. -
Check the cluster status of the nodes at the system console of either node:
cluster showShow example
The following example displays information about the health and eligibility of the nodes in the cluster:
cluster::*> cluster show Node Health Eligibility Epsilon -------------------- ------- ------------ ------------ node1 true true false node2 true true false 2 entries were displayed.
-
Check the status of the HA pair at the system console of either node:
storage failover showShow example
The following example shows the status of node1 and node2:
Node Partner Possible State Description -------------- -------------- -------- ------------------------------------- node1 node2 true Connected to node2 node2 node1 true Connected to node1 2 entries were displayed.
-
If AutoSupport is enabled on this cluster, suppress automatic case creation by invoking an AutoSupport message:
system node autosupport invoke -node * -type all -message MAINT=xhxis the duration of the maintenance window in hours.The message notifies technical support of this maintenance task so that automatic case creation is suppressed during the maintenance window. Show example
The following command suppresses automatic case creation for two hours:
cluster::*> system node autosupport invoke -node * -type all -message MAINT=2h
-
Verify that the current state of the switchless cluster is
true, and then disable the switchless cluster mode:network options switchless-cluster modify -enabled false -
Take over the target node:
storage failover takeover -ofnode target_node_nameIt does not matter which node is the target node. When it is taken over, the target node automatically reboots and displays the
Waiting for giveback…message.The active node is now serving data for the partner (target) node that was taken over.
-
Wait for two minutes after takeover of the impaired node to confirm that the takeover was completed successfully.
-
With the target node showing the
Waiting for giveback…message, shut it down.The method you use to shut down the node depends on whether you use remote management through the node Service Processor (SP).
If SP Then… Is configured
Log in to the impaired node SP, and then power off the system:
system power offIs not configured
At the impaired node prompt, press
Ctrl-C, and then respondyto halt the node.
Step 2: Configure cables and ports
-
On each controller module, disconnect the cable that connects the 10 GbE cluster port to the switchless cluster.
-
Connect the 10 GbE cluster port to the switch on both controller modules.
-
Verify that the 10 GbE cluster ports connected on the switch are configured to be part of the same VLAN.
If you plan to connect the cluster ports on each controller module to different switches, then you must verify that the ports on which the cluster ports are connected on each switch are configured for the same VLAN and that trunking is properly configured on both switches.
-
Give back storage to the target node:
storage failover giveback -ofnode node2 -
Monitor the progress of the giveback operation:
storage failover show-giveback -
After the giveback operation is complete, confirm that the HA pair is healthy and takeover is possible:
storage failover showShow example
The output should be similar to the following:
Node Partner Possible State Description -------------- -------------- -------- ------------------------------------- node1 node2 true Connected to node2 node2 node1 true Connected to node1 2 entries were displayed.
-
Verify that the cluster port LIFs are operating correctly:
network interface show -role clusterShow example
The following example shows that the LIFs are
upon node1 and node2 and that the "Is Home" column results aretrue:cluster::*> network interface show -role cluster Logical Status Network Current Current Is Vserver Interface Admin/Oper Address/Mask Node Port Home ----------- ---------- ---------- ------------------ ------------- ------- ---- node1 clus1 up/up 192.168.177.121/24 node1 e1a true node2 clus1 up/up 192.168.177.123/24 node2 e1a true 2 entries were displayed. -
Check the cluster status of the nodes at the system console of either node:
cluster showShow example
The following example displays information about the health and eligibility of the nodes in the cluster:
cluster::*> cluster show Node Health Eligibility Epsilon -------------------- ------- ------------ ------------ node1 true true false node2 true true false 2 entries were displayed.
-
Verify the connectivity of the remote cluster interfaces:
You can use the network interface check cluster-connectivity command to start an accessibility check for cluster connectivity and then display the details:
network interface check cluster-connectivity start and network interface check cluster-connectivity show
cluster1::*> network interface check cluster-connectivity start
NOTE: Wait for a number of seconds before running the show command to display the details.
cluster1::*> network interface check cluster-connectivity show
Source Destination Packet
Node Date LIF LIF Loss
------ -------------------------- ---------------- ---------------- -----------
node1
3/5/2022 19:21:18 -06:00 node1_clus2 node2-clus1 none
3/5/2022 19:21:20 -06:00 node1_clus2 node2_clus2 none
node2
3/5/2022 19:21:18 -06:00 node2_clus2 node1_clus1 none
3/5/2022 19:21:20 -06:00 node2_clus2 node1_clus2 none
For all ONTAP releases, you can also use the cluster ping-cluster -node <name> command to check the connectivity:
cluster ping-cluster -node <name>
cluster1::*> cluster ping-cluster -node local Host is node2 Getting addresses from network interface table... Cluster node1_clus1 169.254.209.69 node1 e0a Cluster node1_clus2 169.254.49.125 node1 e0b Cluster node2_clus1 169.254.47.194 node2 e0a Cluster node2_clus2 169.254.19.183 node2 e0b Local = 169.254.47.194 169.254.19.183 Remote = 169.254.209.69 169.254.49.125 Cluster Vserver Id = 4294967293 Ping status:.... Basic connectivity succeeds on 4 path(s) Basic connectivity fails on 0 path(s) ................ Detected 9000 byte MTU on 4 path(s): Local 169.254.47.194 to Remote 169.254.209.69 Local 169.254.47.194 to Remote 169.254.49.125 Local 169.254.19.183 to Remote 169.254.209.69 Local 169.254.19.183 to Remote 169.254.49.125 Larger than PMTU communication succeeds on 4 path(s) RPC status: 2 paths up, 0 paths down (tcp check) 2 paths up, 0 paths down (udp check)
Step 3: Complete the procedure
-
If you suppressed automatic case creation, reenable it by invoking an AutoSupport message:
system node autosupport invoke -node * -type all -message MAINT=ENDShow example
cluster::*> system node autosupport invoke -node * -type all -message MAINT=END
-
Change the privilege level back to admin:
set -privilege admin