Prepare for migration from CN1610 switches to 3132Q-V switches
Follow these steps to prepare your CN1610 switches for migration to Cisco Nexus 3132Q-V switches.
-
If AutoSupport is enabled on this cluster, suppress automatic case creation by invoking an AutoSupport message:
system node autosupport invoke -node * -type all - message MAINT=xh
x is the duration of the maintenance window in hours.
The AutoSupport message notifies technical support of this maintenance task so that automatic case creation is suppressed during the maintenance window. -
Display information about the devices in your configuration:
network device-discovery show
Show example
The following example displays how many cluster interconnect interfaces have been configured in each node for each cluster interconnect switch:
cluster::> network device-discovery show Local Discovered Node Port Device Interface Platform ------ ------ ------------ ----------- ---------- n1 /cdp e0a CL1 0/1 CN1610 e0b CL2 0/1 CN1610 e0c CL2 0/2 CN1610 e0d CL1 0/2 CN1610 n2 /cdp e0a CL1 0/3 CN1610 e0b CL2 0/3 CN1610 e0c CL2 0/4 CN1610 e0d CL1 0/4 CN1610 8 entries were displayed.
-
Determine the administrative or operational status for each cluster interface.
-
Display the cluster network port attributes:
network port show
Show example
The following example displays the network port attributes on a system:
cluster::*> network port show -role Cluster (network port show) Node: n1 Broadcast Speed (Mbps) Health Ignore Port IPspace Domain Link MTU Admin/Open Status Health Status ----- --------- ---------- ----- ----- ------------ ------ ------------- e0a cluster cluster up 9000 auto/10000 - - e0b cluster cluster up 9000 auto/10000 - - e0c cluster cluster up 9000 auto/10000 - - e0d cluster cluster up 9000 auto/10000 - - Node: n2 Broadcast Speed (Mbps) Health Ignore Port IPspace Domain Link MTU Admin/Open Status Health Status ----- --------- ---------- ----- ----- ------------ ------ ------------- e0a cluster cluster up 9000 auto/10000 - - e0b cluster cluster up 9000 auto/10000 - - e0c cluster cluster up 9000 auto/10000 - - e0d cluster cluster up 9000 auto/10000 - - 8 entries were displayed.
-
Display information about the logical interfaces: +
network interface show
Show example
The following example displays the general information about all of the LIFs on your system:
cluster::*> network interface show -role Cluster (network interface show) Logical Status Network Current Current Is Vserver Interface Admin/Oper Address/Mask Node Port Home -------- ---------- ----------- -------------- -------- -------- ----- Cluster n1_clus1 up/up 10.10.0.1/24 n1 e0a true n1_clus2 up/up 10.10.0.2/24 n1 e0b true n1_clus3 up/up 10.10.0.3/24 n1 e0c true n1_clus4 up/up 10.10.0.4/24 n1 e0d true n2_clus1 up/up 10.10.0.5/24 n2 e0a true n2_clus2 up/up 10.10.0.6/24 n2 e0b true n2_clus3 up/up 10.10.0.7/24 n2 e0c true n2_clus4 up/up 10.10.0.8/24 n2 e0d true 8 entries were displayed.
-
Display information about the discovered cluster switches:
system cluster-switch show
Show example
The following example displays the cluster switches that are known to the cluster, along with their management IP addresses:
cluster::> system cluster-switch show Switch Type Address Model ----------------------------- ---------------- ------------- -------- CL1 cluster-network 10.10.1.101 CN1610 Serial Number: 01234567 Is Monitored: true Reason: Software Version: 1.2.0.7 Version Source: ISDP CL2 cluster-network 10.10.1.102 CN1610 Serial Number: 01234568 Is Monitored: true Reason: Software Version: 1.2.0.7 Version Source: ISDP 2 entries were displayed.
-
-
Set the
-auto-revert
parameter to false on cluster LIFs clus1 and clus4 on both nodes:network interface modify
Show example
cluster::*> network interface modify -vserver node1 -lif clus1 -auto-revert false cluster::*> network interface modify -vserver node1 -lif clus4 -auto-revert false cluster::*> network interface modify -vserver node2 -lif clus1 -auto-revert false cluster::*> network interface modify -vserver node2 -lif clus4 -auto-revert false
-
Verify that the appropriate RCF and image are installed on the new 3132Q-V switches as necessary for your requirements, and make any essential site customizations, such as users and passwords, network addresses, and so on.
You must prepare both switches at this time. If you need to upgrade the RCF and image, follow these steps:
-
See the Cisco Ethernet Switches page on NetApp Support Site.
-
Note your switch and the required software versions in the table on that page.
-
Download the appropriate version of the RCF.
-
Select CONTINUE on the Description page, accept the license agreement, and then follow the instructions on the Download page to download the RCF.
-
Download the appropriate version of the image software.
-
-
Migrate the LIFs associated with the second CN1610 switch to be replaced:
network interface migrate
You must migrate the cluster LIFs from a connection to the node, either through the service processor or node management interface, which owns the cluster LIF being migrated.
Show example
The following example shows n1 and n2, but LIF migration must be done on all the nodes:
cluster::*> network interface migrate -vserver Cluster -lif n1_clus2 -destination-node n1 -destination-port e0a cluster::*> network interface migrate -vserver Cluster -lif n1_clus3 -destination-node n1 -destination-port e0d cluster::*> network interface migrate -vserver Cluster -lif n2_clus2 -destination-node n2 -destination-port e0a cluster::*> network interface migrate -vserver Cluster -lif n2_clus3 -destination-node n2 -destination-port e0d
-
Verify the cluster's health:
network interface show
Show example
The following example shows the result of the previous
network interface migrate
command:cluster::*> network interface show -role Cluster (network interface show) Logical Status Network Current Current Is Vserver Interface Admin/Oper Address/Mask Node Port Home -------- ---------- ----------- --------------- -------- -------- ----- Cluster n1_clus1 up/up 10.10.0.1/24 n1 e0a true n1_clus2 up/up 10.10.0.2/24 n1 e0a false n1_clus3 up/up 10.10.0.3/24 n1 e0d false n1_clus4 up/up 10.10.0.4/24 n1 e0d true n2_clus1 up/up 10.10.0.5/24 n2 e0a true n2_clus2 up/up 10.10.0.6/24 n2 e0a false n2_clus3 up/up 10.10.0.7/24 n2 e0d false n2_clus4 up/up 10.10.0.8/24 n2 e0d true 8 entries were displayed.
-
Shut down the cluster interconnect ports that are physically connected to switch CL2:
network port modify
Show example
The following commands shut down the specified ports on n1 and n2, but the ports must be shut down on all nodes:
cluster::*> network port modify -node n1 -port e0b -up-admin false cluster::*> network port modify -node n1 -port e0c -up-admin false cluster::*> network port modify -node n2 -port e0b -up-admin false cluster::*> network port modify -node n2 -port e0c -up-admin false
-
Verify the connectivity of the remote cluster interfaces:
You can use the network interface check cluster-connectivity
command to start an accessibility check for cluster connectivity and then display the details:
network interface check cluster-connectivity start
and network interface check cluster-connectivity show
cluster1::*> network interface check cluster-connectivity start
NOTE: Wait for a number of seconds before running the show command to display the details.
cluster1::*> network interface check cluster-connectivity show Source Destination Packet Node Date LIF LIF Loss ------ -------------------------- ---------- -------------- ----------- n1 3/5/2022 19:21:18 -06:00 n1_clus2 n1_clus1 none 3/5/2022 19:21:20 -06:00 n1_clus2 n2_clus2 none n2 3/5/2022 19:21:18 -06:00 n2_clus2 n1_clus1 none 3/5/2022 19:21:20 -06:00 n2_clus2 n1_clus2 none
For all ONTAP releases, you can also use the cluster ping-cluster -node <name>
command to check the connectivity:
cluster ping-cluster -node <name>
cluster::*> cluster ping-cluster -node n1 Host is n1 Getting addresses from network interface table... Cluster n1_clus1 n1 e0a 10.10.0.1 Cluster n1_clus2 n1 e0b 10.10.0.2 Cluster n1_clus3 n1 e0c 10.10.0.3 Cluster n1_clus4 n1 e0d 10.10.0.4 Cluster n2_clus1 n2 e0a 10.10.0.5 Cluster n2_clus2 n2 e0b 10.10.0.6 Cluster n2_clus3 n2 e0c 10.10.0.7 Cluster n2_clus4 n2 e0d 10.10.0.8 Local = 10.10.0.1 10.10.0.2 10.10.0.3 10.10.0.4 Remote = 10.10.0.5 10.10.0.6 10.10.0.7 10.10.0.8 Cluster Vserver Id = 4294967293 Ping status: .... Basic connectivity succeeds on 16 path(s) Basic connectivity fails on 0 path(s) ................ Detected 1500 byte MTU on 16 path(s): Local 10.10.0.1 to Remote 10.10.0.5 Local 10.10.0.1 to Remote 10.10.0.6 Local 10.10.0.1 to Remote 10.10.0.7 Local 10.10.0.1 to Remote 10.10.0.8 Local 10.10.0.2 to Remote 10.10.0.5 Local 10.10.0.2 to Remote 10.10.0.6 Local 10.10.0.2 to Remote 10.10.0.7 Local 10.10.0.2 to Remote 10.10.0.8 Local 10.10.0.3 to Remote 10.10.0.5 Local 10.10.0.3 to Remote 10.10.0.6 Local 10.10.0.3 to Remote 10.10.0.7 Local 10.10.0.3 to Remote 10.10.0.8 Local 10.10.0.4 to Remote 10.10.0.5 Local 10.10.0.4 to Remote 10.10.0.6 Local 10.10.0.4 to Remote 10.10.0.7 Local 10.10.0.4 to Remote 10.10.0.8 Larger than PMTU communication succeeds on 16 path(s) RPC status: 4 paths up, 0 paths down (tcp check) 4 paths up, 0 paths down (udp check)
-
Shut down the ISL ports 13 through 16 on the active CN1610 switch CL1:
shutdown
Show example
The following example shows how to shut down ISL ports 13 through 16 on the CN1610 switch CL1:
(CL1)# configure (CL1)(Config)# interface 0/13-0/16 (CL1)(Interface 0/13-0/16)# shutdown (CL1)(Interface 0/13-0/16)# exit (CL1)(Config)# exit (CL1)#
-
Build a temporary ISL between CL1 and C2:
Show example
The following example builds a temporary ISL between CL1 (ports 13-16) and C2 (ports e1/24/1-4):
C2# configure C2(config)# interface port-channel 2 C2(config-if)# switchport mode trunk C2(config-if)# spanning-tree port type network C2(config-if)# mtu 9216 C2(config-if)# interface breakout module 1 port 24 map 10g-4x C2(config)# interface e1/24/1-4 C2(config-if-range)# switchport mode trunk C2(config-if-range)# mtu 9216 C2(config-if-range)# channel-group 2 mode active C2(config-if-range)# exit C2(config-if)# exit