Deployment Procedures

Contributors netapp-dorianh Download PDF of this page

This document provides details for configuring a fully redundant, highly available FlexPod Express system. To reflect this redundancy, the components being configured in each step are referred to as either component A or component B. For example, controller A and controller B identify the two NetApp storage controllers that are provisioned in this document. Switch A and switch B identify a pair of Cisco Nexus switches. Fabric Interconnect A and Fabric Interconnect B are the two Integrated Nexus Fabric Interconnects.

In addition, this document describes steps for provisioning multiple Cisco UCS hosts, which are identified sequentially as server A, server B, and so on.

To indicate that you should include information pertinent to your environment in a step, <<text>> appears as part of the command structure. See the following example for the vlan create command:

Controller01>vlan create vif0 <<mgmt_vlan_id>>

This document enables you to fully configure the FlexPod Express environment. In this process, various steps require you to insert customer-specific naming conventions, IP addresses, and virtual local area network (VLAN) schemes. The table below describes the VLANs required for deployment, as outlined in this guide. This table can be completed based on the specific site variables and used to implement the document configuration steps.

If you use separate in-band and out-of-band management VLANs, you must create a layer 3 route between them. For this validation, a common management VLAN was used.
VLAN name VLAN purpose ID used in validating this document

Management VLAN

VLAN for management interfaces

18

Native VLAN

VLAN to which untagged frames are assigned

2

NFS VLAN

VLAN for NFS traffic

104

VMware vMotion VLAN

VLAN designated for the movement of virtual machines (VMs) from one physical host to another

103

VM traffic VLAN

VLAN for VM application traffic

102

iSCSI-A-VLAN

VLAN for iSCSI traffic on fabric A

124

iSCSI-B-VLAN

VLAN for iSCSI traffic on fabric B

125

The VLAN numbers are needed throughout the configuration of FlexPod Express. The VLANs are referred to as <<var_xxxx_vlan>>, where xxxx is the purpose of the VLAN (such as iSCSI-A).

The following table lists the VMware VMs created.

VM Description Host Name

VMware vCenter Server

Seahawks-vcsa.cie.netapp.com

Cisco Nexus 31108PCV deployment procedure

This section details the Cisco Nexus 31308PCV switch configuration used in a FlexPod Express environment.

Initial setup of Cisco Nexus 31108PCV switch

This procedures describes how to configure the Cisco Nexus switches for use in a base FlexPod Express environment.

This procedure assumes that you are using a Cisco Nexus 31108PCV running NX-OS software release 7.0(3)I6(1).
  1. Upon initial boot and connection to the console port of the switch, the Cisco NX-OS setup automatically starts. This initial configuration addresses basic settings, such as the switch name, the mgmt0 interface configuration, and Secure Shell (SSH) setup.

  2. The FlexPod Express management network can be configured in multiple ways. The mgmt0 interfaces on the 31108PCV switches can be connected to an existing management network, or the mgmt0 interfaces of the 31108PCV switches can be connected in a back-to-back configuration. However, this link cannot be used for external management access such as SSH traffic.

    In this deployment guide, the FlexPod Express Cisco Nexus 31108PCV switches are connected to an existing management network.

  3. To configure the Cisco Nexus 31108PCV switches, power on the switch and follow the on-screen prompts, as illustrated here for the initial setup of both the switches, substituting the appropriate values for the switch-specific information.

    This setup utility will guide you through the basic configuration of the system. Setup configures only enough connectivity for management of the system.
    *Note: setup is mainly used for configuring the system initially, when no configuration is present. So setup always assumes system defaults and not the current system configuration values.
    Press Enter at anytime to skip a dialog. Use ctrl-c at anytime to skip the remaining dialogs.
    Would you like to enter the basic configuration dialog (yes/no): y
    Do you want to enforce secure password standard (yes/no) [y]: y
    Create another login account (yes/no) [n]: n
    Configure read-only SNMP community string (yes/no) [n]: n
    Configure read-write SNMP community string (yes/no) [n]: n
    Enter the switch name : 31108PCV-A
    Continue with Out-of-band (mgmt0) management configuration? (yes/no) [y]: y
    Mgmt0 IPv4 address : <<var_switch_mgmt_ip>>
    Mgmt0 IPv4 netmask : <<var_switch_mgmt_netmask>>
    Configure the default gateway? (yes/no) [y]: y
    IPv4 address of the default gateway : <<var_switch_mgmt_gateway>>
    Configure advanced IP options? (yes/no) [n]: n
    Enable the telnet service? (yes/no) [n]: n
    Enable the ssh service? (yes/no) [y]: y
    Type of ssh key you would like to generate (dsa/rsa) [rsa]: rsa
    Number of rsa key bits <1024-2048> [1024]: <enter>
    Configure the ntp server? (yes/no) [n]: y
    NTP server IPv4 address : <<var_ntp_ip>>
    Configure default interface layer (L3/L2) [L2]: <enter>
    Configure default switchport interface state (shut/noshut) [noshut]: <enter>
    Configure CoPP system profile (strict/moderate/lenient/dense) [strict]: <enter>
  4. A summary of your configuration is displayed and you are asked if you would like to edit the configuration. If your configuration is correct, enter n.

    Would you like to edit the configuration? (yes/no) [n]: no
  5. You are then asked if you would like to use this configuration and save it. If so, enter y.

    Use this configuration and save it? (yes/no) [y]: Enter
  6. Repeat steps 1 through 5 for Cisco Nexus switch B.

Enable advanced features

Certain advanced features must be enabled in Cisco NX-OS to provide additional configuration options.

  1. To enable the appropriate features on Cisco Nexus switch A and switch B, enter configuration mode by using the command (config t) and run the following commands:

    feature interface-vlan
    feature lacp
    feature vpc
    The default port channel load-balancing hash uses the source and destination IP addresses to determine the load-balancing algorithm across the interfaces in the port channel. You can achieve better distribution across the members of the port channel by providing more inputs to the hash algorithm beyond the source and destination IP addresses. For the same reason, NetApp highly recommends adding the source and destination TCP ports to the hash algorithm.
  2. From configuration mode (config t), run the following commands to set the global port channel load-balancing configuration on Cisco Nexus switch A and switch B:

    port-channel load-balance src-dst ip-l4port

Perform global spanning-tree configuration

The Cisco Nexus platform uses a new protection feature called bridge assurance. Bridge assurance helps protect against a unidirectional link or other software failure with a device that continues to forward data traffic when it is no longer running the spanning-tree algorithm. Ports can be placed in one of several states, including network or edge, depending on the platform.

NetApp recommends setting bridge assurance so that all ports are considered to be network ports by default. This setting forces the network administrator to review the configuration of each port. It also reveals the most common configuration errors, such as unidentified edge ports or a neighbor that does not have the bridge assurance feature enabled. In addition, it is safer to have the spanning tree block many ports rather than too few, which allows the default port state to enhance the overall stability of the network.

Pay close attention to the spanning-tree state when adding servers, storage, and uplink switches, especially if they do not support bridge assurance. In such cases, you might need to change the port type to make the ports active.

The Bridge Protocol Data Unit (BPDU) guard is enabled on edge ports by default as another layer of protection. To prevent loops in the network, this feature shuts down the port if BPDUs from another switch are seen on this interface.

From configuration mode (config t), run the following commands to configure the default spanning-tree options, including the default port type and BPDU guard, on Cisco Nexus switch A and switch B:

spanning-tree port type network default
spanning-tree port type edge bpduguard default

Define VLANs

Before individual ports with different VLANs are configured, the layer-2 VLANs must be defined on the switch. It is also a good practice to name the VLANs for easy troubleshooting in the future.

From configuration mode (config t), run the following commands to define and describe the layer 2 VLANs on Cisco Nexus switch A and switch B:

vlan <<nfs_vlan_id>>
  name NFS-VLAN
vlan <<iSCSI_A_vlan_id>>
  name iSCSI-A-VLAN
vlan <<iSCSI_B_vlan_id>>
  name iSCSI-B-VLAN
vlan <<vmotion_vlan_id>>
  name vMotion-VLAN
vlan <<vmtraffic_vlan_id>>
  name VM-Traffic-VLAN
vlan <<mgmt_vlan_id>>
  name MGMT-VLAN
vlan <<native_vlan_id>>
  name NATIVE-VLAN
exit

Configure access and management port descriptions

As is the case with assigning names to the layer-2 VLANs, setting descriptions for all the interfaces can help with both provisioning and troubleshooting.

From configuration mode (config t) in each of the switches, enter the following port descriptions for the FlexPod Express large configuration:

Cisco Nexus switch A

int eth1/1
  description AFF A220-A e0M
int eth1/2
  description Cisco UCS FI-A mgmt0
int eth1/3
  description Cisco UCS FI-A eth1/1
int eth1/4
  description Cisco UCS FI-B eth1/1
int eth1/13
  description vPC peer-link 31108PVC-B 1/13
int eth1/14
  description vPC peer-link 31108PVC-B 1/14

Cisco Nexus switch B

int eth1/1
  description AFF A220-B e0M
int eth1/2
  description Cisco UCS FI-B mgmt0
int eth1/3
  description Cisco UCS FI-A eth1/2
int eth1/4
  description Cisco UCS FI-B eth1/2
int eth1/13
  description vPC peer-link 31108PVC-B 1/13
int eth1/14
  description vPC peer-link 31108PVC-B 1/14

Configure server and storage management interfaces

The management interfaces for both the server and the storage typically use only a single VLAN. Therefore, configure the management interface ports as access ports. Define the management VLAN for each switch and change the spanning-tree port type to edge.

From configuration mode (config t), run the following commands to configure the port settings for the management interfaces of both the servers and the storage:

Cisco Nexus switch A

int eth1/1-2
  switchport mode access
  switchport access vlan <<mgmt_vlan>>
  spanning-tree port type edge
  speed 1000
exit

Cisco Nexus switch B

int eth1/1-2
  switchport mode access
  switchport access vlan <<mgmt_vlan>>
  spanning-tree port type edge
  speed 1000
exit

Add NTP distribution interface

Cisco Nexus switch A

From the global configuration mode, execute the following commands.

interface Vlan<ib-mgmt-vlan-id>
ip address <switch-a-ntp-ip>/<ib-mgmt-vlan-netmask-length>
no shutdown
exitntp peer <switch-b-ntp-ip> use-vrf default

Cisco Nexus switch B

From the global configuration mode, execute the following commands.

interface Vlan<ib-mgmt-vlan-id>
ip address <switch- b-ntp-ip>/<ib-mgmt-vlan-netmask-length>
no shutdown
exitntp peer <switch-a-ntp-ip> use-vrf default

Perform virtual port channel global configuration

A virtual port channel (vPC) enables links that are physically connected to two different Cisco Nexus switches to appear as a single port channel to a third device. The third device can be a switch, server, or any other networking device. A vPC can provide layer-2 multipathing, which allows you to create redundancy by increasing bandwidth, enabling multiple parallel paths between nodes, and load-balancing traffic where alternative paths exist.

A vPC provides the following benefits:

  • Enabling a single device to use a port channel across two upstream devices

  • Eliminating spanning-tree protocol blocked ports

  • Providing a loop-free topology

  • Using all available uplink bandwidth

  • Providing fast convergence if either the link or a device fails

  • Providing link-level resiliency

  • Helping provide high availability

The vPC feature requires some initial setup between the two Cisco Nexus switches to function properly. If you use the back-to-back mgmt0 configuration, use the addresses defined on the interfaces and verify that they can communicate by using the ping <<switch_A/B_mgmt0_ip_addr>>vrf management command.

From configuration mode (config t), run the following commands to configure the vPC global configuration for both switches:

Cisco Nexus switch A

vpc domain 1
 role priority 10
peer-keepalive destination <<switch_B_mgmt0_ip_addr>> source <<switch_A_mgmt0_ip_addr>> vrf management
  peer-gateway
  auto-recovery
  ip arp synchronize
  int eth1/13-14
  channel-group 10 mode active
int Po10description vPC peer-link
switchport
switchport mode trunkswitchport trunk native vlan <<native_vlan_id>>
switchport trunk allowed vlan <<nfs_vlan_id>>,<<vmotion_vlan_id>>, <<vmtraffic_vlan_id>>, <<mgmt_vlan>, <<iSCSI_A_vlan_id>>, <<iSCSI_B_vlan_id>> spanning-tree port type network
vpc peer-link
no shut
exit
int Po13
description vPC ucs-FI-A
switchport mode trunk
switchport trunk native vlan <<native_vlan_id>>
switchport trunk allowed vlan <<vmotion_vlan_id>>, <<vmtraffic_vlan_id>>, <<mgmt_vlan>> spanning-tree port type network
mtu 9216
vpc 13
no shut
exit
int eth1/3
  channel-group 13 mode active
int Po14
description vPC ucs-FI-B
switchport mode trunk
switchport trunk native vlan <<native_vlan_id>>
switchport trunk allowed vlan <<vmotion_vlan_id>>, <<vmtraffic_vlan_id>>, <<mgmt_vlan>> spanning-tree port type network
mtu 9216
vpc 14
no shut
exit
int eth1/4
  channel-group 14 mode active
copy run start

Cisco Nexus switch B

vpc domain 1
peer-switch
role priority 20
peer-keepalive destination <<switch_A_mgmt0_ip_addr>> source <<switch_B_mgmt0_ip_addr>> vrf management
  peer-gateway
  auto-recovery
  ip arp synchronize
  int eth1/13-14
  channel-group 10 mode active
int Po10
description vPC peer-link
switchport
switchport mode trunk
switchport trunk native vlan <<native_vlan_id>>
switchport trunk allowed vlan <<nfs_vlan_id>>,<<vmotion_vlan_id>>, <<vmtraffic_vlan_id>>, <<mgmt_vlan>>, <<iSCSI_A_vlan_id>>, <<iSCSI_B_vlan_id>> spanning-tree port type network
vpc peer-link
no shut
exit
int Po13
description vPC ucs-FI-A
switchport mode trunk
switchport trunk native vlan <<native_vlan_id>>
switchport trunk allowed vlan <<vmotion_vlan_id>>, <<vmtraffic_vlan_id>>, <<mgmt_vlan>> spanning-tree port type network
mtu 9216
vpc 13
no shut
exit
int eth1/3
  channel-group 13 mode active
int Po14
description vPC ucs-FI-B
switchport mode trunk
switchport trunk native vlan <<native_vlan_id>>
switchport trunk allowed vlan <<vmotion_vlan_id>>, <<vmtraffic_vlan_id>>, <<mgmt_vlan>> spanning-tree port type network
mtu 9216
vpc 14
no shut
exit
int eth1/4
  channel-group 14 mode active
copy run start
In this solution validation, a maximum transmission unit (MTU) of 9000 was used. However, based on application requirements, you can configure an appropriate value of MTU. It is important to set the same MTU value across the FlexPod solution. Incorrect MTU configurations between components result in packets being dropped.

Depending on the available network infrastructure, several methods and features can be used to uplink the FlexPod environment. If an existing Cisco Nexus environment is present, NetApp recommends using vPCs to uplink the Cisco Nexus 31108PVC switches included in the FlexPod environment into the infrastructure. The uplinks can be 10GbE uplinks for a 10GbE infrastructure solution or 1GbE for a 1GbE infrastructure solution if required. The previously described procedures can be used to create an uplink vPC to the existing environment. Make sure to run copy run start to save the configuration on each switch after the configuration is completed.

NetApp storage deployment procedure (part 1)

This section describes the NetApp AFF storage deployment procedure.

NetApp Storage Controller AFF2xx Series Installation

NetApp Hardware Universe

The NetApp Hardware Universe (HWU) application provides supported hardware and software components for any specific ONTAP version. It provides configuration information for all the NetApp storage appliances currently supported by ONTAP software. It also provides a table of component compatibilities.

Confirm that the hardware and software components that you would like to use are supported with the version of ONTAP that you plan to install:

  1. Access the HWU application to view the system configuration guides. Select the Compare Storage Systems tab to view the compatibility between different version of the ONTAP software and the NetApp storage appliances with your desired specifications.

  2. Alternatively, to compare components by storage appliance, click Compare Storage Systems.

Controller AFF2XX Series prerequisites

To plan the physical location of the storage systems, see the the following sections:
Electrical requirements
Supported power cords
Onboard ports and cables

Storage controllers

Follow the physical installation procedures for the controllers in the AFF A220 Documentation.

NetApp ONTAP 9.5

Configuration worksheet

Before running the setup script, complete the configuration worksheet from the product manual. The configuration worksheet is available in the ONTAP 9.5 Software Setup Guide (available in the ONTAP 9 Documentation Center). The table below illustrates ONTAP 9.5 installation and configuration information.

This system is set up in a two-node switchless cluster configuration.
Cluster Detail Cluster Detail Value

Cluster node A IP address

<<var_nodeA_mgmt_ip>>

Cluster node A netmask

<<var_nodeA_mgmt_mask>>

Cluster node A gateway

<<var_nodeA_mgmt_gateway>>

Cluster node A name

<<var_nodeA>>

Cluster node B IP address

<<var_nodeB_mgmt_ip>>

Cluster node B netmask

<<var_nodeB_mgmt_mask>>

Cluster node B gateway

<<var_nodeB_mgmt_gateway>>

Cluster node B name

<<var_nodeB>>

ONTAP 9.5 URL

<<var_url_boot_software>>

Name for cluster

<<var_clustername>>

Cluster management IP address

<<var_clustermgmt_ip>>

Cluster B gateway

<<var_clustermgmt_gateway>>

Cluster B netmask

<<var_clustermgmt_mask>>

Domain name

<<var_domain_name>>

DNS server IP (you can enter more than one)

<<var_dns_server_ip>>

NTP server A IP

<< switch-a-ntp-ip >>

NTP server B IP

<< switch-b-ntp-ip >>

Configure node A

To configure node A, complete the following steps:

  1. Connect to the storage system console port. You should see a Loader-A prompt. However, if the storage system is in a reboot loop, press Ctrl- C to exit the autoboot loop when you see this message:

    Starting AUTOBOOT press Ctrl-C to abort...
  2. Allow the system to boot.

    autoboot
  3. Press Ctrl- C to enter the Boot menu.

    If ONTAP 9. 5 is not the version of software being booted, continue with the following steps to install new software. If ONTAP 9. 5 is the version being booted, select option 8 and y to reboot the node. Then, continue with step 14.

  4. To install new software, select option 7.

  5. Enter y to perform an upgrade.

  6. Select e0M for the network port you want to use for the download.

  7. Enter y to reboot now.

  8. Enter the IP address, netmask, and default gateway for e0M in their respective places.

    <<var_nodeA_mgmt_ip>> <<var_nodeA_mgmt_mask>> <<var_nodeA_mgmt_gateway>>
  9. Enter the URL where the software can be found.

    This web server must be pingable.
  10. Press Enter for the user name, indicating no user name.

  11. Enter y to set the newly installed software as the default to be used for subsequent reboots.

  12. Enter y to reboot the node.

    When installing new software, the system might perform firmware upgrades to the BIOS and adapter cards, causing reboots and possible stops at the Loader-A prompt. If these actions occur, the system might deviate from this procedure.

  13. Press Ctrl- C to enter the Boot menu.

  14. Select option 4 for Clean Configuration and Initialize All Disks.

  15. Enter y to zero disks, reset config, and install a new file system.

  16. Enter y to erase all the data on the disks.

    The initialization and creation of the root aggregate can take 90 minutes or more to complete, depending on the number and type of disks attached. When initialization is complete, the storage system reboots. Note that SSDs take considerably less time to initialize. You can continue with the node B configuration while the disks for node A are zeroing.

  17. While node A is initializing, begin configuring node B.

Configure node B

To configure node B, complete the following steps:

  1. Connect to the storage system console port. You should see a Loader-A prompt. However, if the storage system is in a reboot loop, press Ctrl-C to exit the autoboot loop when you see this message:

    Starting AUTOBOOT press Ctrl-C to abort...
  2. Press Ctrl-C to enter the Boot menu.

    autoboot
  3. Press Ctrl-C when prompted.

    If ONTAP 9. 5 is not the version of software being booted, continue with the following steps to install new software. If ONTAP 9.4 is the version being booted, select option 8 and y to reboot the node. Then, continue with step 14.

  4. To install new software, select option 7.

  5. Enter y to perform an upgrade.

  6. Select e0M for the network port you want to use for the download.

  7. Enter y to reboot now.

  8. Enter the IP address, netmask, and default gateway for e0M in their respective places.

    <<var_nodeB_mgmt_ip>> <<var_nodeB_mgmt_ip>><<var_nodeB_mgmt_gateway>>
  9. Enter the URL where the software can be found.

    This web server must be pingable.
    <<var_url_boot_software>>
  10. Press Enter for the user name, indicating no user name

  11. Enter y to set the newly installed software as the default to be used for subsequent reboots.

  12. Enter y to reboot the node.

    When installing new software, the system might perform firmware upgrades to the BIOS and adapter cards, causing reboots and possible stops at the Loader-A prompt. If these actions occur, the system might deviate from this procedure.

  13. Press Ctrl-C to enter the Boot menu.

  14. Select option 4 for Clean Configuration and Initialize All Disks.

  15. Enter y to zero disks, reset config, and install a new file system.

  16. Enter y to erase all the data on the disks.

    The initialization and creation of the root aggregate can take 90 minutes or more to complete, depending on the number and type of disks attached. When initialization is complete, the storage system reboots. Note that SSDs take considerably less time to initialize.

Continuation node A configuration and cluster configuration

From a console port program attached to the storage controller A (node A) console port, run the node setup script. This script appears when ONTAP 9.5 boots on the node for the first time.

The node and cluster setup procedure has changed slightly in ONTAP 9.5. The cluster setup wizard is now used to configure the first node in a cluster, and System Manager is used to configure the cluster.

  1. Follow the prompts to set up node A.

    Welcome to the cluster setup wizard.
    You can enter the following commands at any time:
      "help" or "?" - if you want to have a question clarified,
      "back" - if you want to change previously answered questions, and
      "exit" or "quit" - if you want to quit the cluster setup wizard.
         Any changes you made before quitting will be saved.
    You can return to cluster setup at any time by typing "cluster setup".
    To accept a default or omit a question, do not enter a value.
    This system will send event messages and periodic reports to NetApp Technical Support. To disable this feature, enter
    autosupport modify -support disable
    within 24 hours.
    Enabling AutoSupport can significantly speed problem determination and resolution should a problem occur on your system.
    For further information on AutoSupport, see: http://support.netapp.com/autosupport/
    Type yes to confirm and continue {yes}: yes
    Enter the node management interface port [e0M]:
    Enter the node management interface IP address: <<var_nodeA_mgmt_ip>>
    Enter the node management interface netmask: <<var_nodeA_mgmt_mask>>
    Enter the node management interface default gateway: <<var_nodeA_mgmt_gateway>>
    A node management interface on port e0M with IP address <<var_nodeA_mgmt_ip>> has been created.
    Use your web browser to complete cluster setup by accessing
    https://<<var_nodeA_mgmt_ip>>
    Otherwise, press Enter to complete cluster setup using the command line interface:
  2. Navigate to the IP address of the node’s management interface.

    Cluster setup can also be performed by using the CLI. This document describes cluster setup using NetApp System Manager guided setup.
  3. Click Guided Setup to configure the cluster.

  4. Enter <<var_clustername>> for the cluster name and <<var_nodeA>> and <<var_nodeB>> for each of the nodes that you are configuring. Enter the password that you would like to use for the storage system. Select Switchless Cluster for the cluster type. Enter the cluster base license.

  5. You can also enter feature licenses for Cluster, NFS, and iSCSI.

  6. You see a status message stating the cluster is being created. This status message cycles through several statuses. This process takes several minutes.

  7. Configure the network.

    1. Deselect the IP Address Range option.

    2. Enter <<var_clustermgmt_ip>> in the Cluster Management IP Address field, <<var_clustermgmt_mask>> in the Netmask field, and <<var_clustermgmt_gateway>> in the Gateway field. Use the …​ selector in the Port field to select e0M of node A.

    3. The node management IP for node A is already populated. Enter <<var_nodeA_mgmt_ip>> for node B.

    4. Enter <<var_domain_name>> in the DNS Domain Name field. Enter <<var_dns_server_ip>> in the DNS Server IP Address field.

      You can enter multiple DNS server IP addresses.

    5. Enter <<switch-a-ntp-ip>> in the Primary NTP Server field.

      You can also enter an alternate NTP server as <<switch- b-ntp-ip>>.

  8. Configure the support information.

    1. If your environment requires a proxy to access AutoSupport, enter the URL in Proxy URL.

    2. Enter the SMTP mail host and email address for event notifications.

      You must, at a minimum, set up the event notification method before you can proceed. You can select any of the methods.

  9. When indicated that the cluster configuration has completed, click Manage Your Cluster to configure the storage.

Continuation of storage cluster configuration

After the configuration of the storage nodes and base cluster, you can continue with the configuration of the storage cluster.

Zero all spare disks

To zero all spare disks in the cluster, run the following command:

disk zerospares

Set on-board UTA2 ports personality

  1. Verify the current mode and the current type of the ports by running the ucadmin show command.

    AFFA220-Clus::> ucadmin show
                           Current  Current    Pending  Pending    Admin
    Node          Adapter  Mode     Type       Mode     Type       Status
    ------------  -------  -------  ---------  -------  ---------  -----------
    AFFA220-Clus-01
                  0c       cna      target     -        -          offline
    AFFA220-Clus-01
                  0d       cna      target     -        -          offline
    AFFA220-Clus-01
                  0e       cna      target     -        -          offline
    AFFA220-Clus-01
                  0f       cna      target     -        -          offline
    AFFA220-Clus-02
                  0c       cna      target     -        -          offline
    AFFA220-Clus-02
                  0d       cna      target     -        -          offline
    AFFA220-Clus-02
                  0e       cna      target     -        -          offline
    AFFA220-Clus-02
                  0f       cna      target     -        -          offline
    8 entries were displayed.
  2. Verify that the current mode of the ports that are in use is cna and that the current type is set to target. If not, change the port personality by running the following command:

    ucadmin modify -node <home node of the port> -adapter <port name> -mode cna -type target

    The ports must be offline to run the previous command. To take a port offline, run the following command:

    network fcp adapter modify -node <home node of the port> -adapter <port name> -state down
    If you changed the port personality, you must reboot each node for the change to take effect.

Enable Cisco Discovery Protocol

To enable the Cisco Discovery Protocol (CDP) on the NetApp storage controllers, run the following command:

node run -node * options cdpd.enable on

Enable the exchange of Link-layer Discovery Protocol (LLDP) neighbor information between the storage and network switches by running the following command. This command enables LLDP on all ports of all nodes in the cluster.

node run * options lldp.enable on

Rename management logical interfaces

To rename the management logical interfaces (LIFs), complete the following steps:

  1. Show the current management LIF names.

    network interface show –vserver <<clustername>>
  2. Rename the cluster management LIF.

    network interface rename –vserver <<clustername>> –lif cluster_setup_cluster_mgmt_lif_1 –newname cluster_mgmt
  3. Rename the node B management LIF.

    network interface rename -vserver <<clustername>> -lif cluster_setup_node_mgmt_lif_AFF A220_A_1 - newname AFF A220-01_mgmt1

Set auto-revert on cluster management

Set the auto-revert parameter on the cluster management interface.

network interface modify –vserver <<clustername>> -lif cluster_mgmt –auto-revert true

Set up service processor network interface

To assign a static IPv4 address to the service processor on each node, run the following commands:

system service-processor network modify –node <<var_nodeA>> -address-family IPv4 –enable true – dhcp none –ip-address <<var_nodeA_sp_ip>> -netmask <<var_nodeA_sp_mask>> -gateway <<var_nodeA_sp_gateway>>
system service-processor network modify –node <<var_nodeB>> -address-family IPv4 –enable true – dhcp none –ip-address <<var_nodeB_sp_ip>> -netmask <<var_nodeB_sp_mask>> -gateway <<var_nodeB_sp_gateway>>
The service processor IP addresses should be in the same subnet as the node management IP addresses.

Enable storage failover in ONTAP

To confirm that storage failover is enabled, run the following commands in a failover pair:

  1. Verify the status of storage failover.

    storage failover show

    Both <<var_nodeA>> and <<var_nodeB>> must be able to perform a takeover. Go to step 3 if the nodes can perform a takeover.

  2. Enable failover on one of the two nodes.

    storage failover modify -node <<var_nodeA>> -enabled true
  3. Verify the HA status of the two-node cluster.

    This step is not applicable for clusters with more than two nodes.
    cluster ha show
  4. Go to step 6 if high availability is configured. If high availability is configured, you see the following message upon issuing the command:

    High Availability Configured: true
  5. Enable HA mode only for the two-node cluster.

    Do not run this command for clusters with more than two nodes because it causes problems with failover.

    cluster ha modify -configured true
    Do you want to continue? {y|n}: y
  6. Verify that hardware assist is correctly configured and, if needed, modify the partner IP address.

    storage failover hwassist show

    The message Keep Alive Status : Error: did not receive hwassist keep alive alerts from partner indicates that hardware assist is not configured. Run the following commands to configure hardware assist.

    storage failover modify –hwassist-partner-ip <<var_nodeB_mgmt_ip>> -node <<var_nodeA>>
    storage failover modify –hwassist-partner-ip <<var_nodeA_mgmt_ip>> -node <<var_nodeB>>

Create jumbo frame MTU broadcast domain in ONTAP

To create a data broadcast domain with an MTU of 9000, run the following commands:

broadcast-domain create -broadcast-domain Infra_NFS -mtu 9000
broadcast-domain create -broadcast-domain Infra_iSCSI-A -mtu 9000
broadcast-domain create -broadcast-domain Infra_iSCSI-B -mtu 9000

Remove data ports from default broadcast domain

The 10GbE data ports are used for iSCSI/NFS traffic, and these ports should be removed from the default domain. Ports e0e and e0f are not used and should also be removed from the default domain.

To remove the ports from the broadcast domain, run the following command:

broadcast-domain remove-ports -broadcast-domain Default -ports <<var_nodeA>>:e0c, <<var_nodeA>>:e0d, <<var_nodeA>>:e0e, <<var_nodeA>>:e0f, <<var_nodeB>>:e0c, <<var_nodeB>>:e0d, <<var_nodeA>>:e0e, <<var_nodeA>>:e0f

Disable flow control on UTA2 ports

It is a NetApp best practice to disable flow control on all UTA2 ports that are connected to external devices. To disable flow control, run the following commands:

net port modify -node <<var_nodeA>> -port e0c -flowcontrol-admin none
Warning: Changing the network port settings will cause a several second interruption in carrier. Do you want to continue? {y|n}: y
net port modify -node <<var_nodeA>> -port e0d -flowcontrol-admin none
Warning: Changing the network port settings will cause a several second interruption in carrier. Do you want to continue? {y|n}: y
net port modify -node <<var_nodeA>> -port e0e -flowcontrol-admin none
Warning: Changing the network port settings will cause a several second interruption in carrier. Do you want to continue? {y|n}: y
net port modify -node <<var_nodeA>> -port e0f -flowcontrol-admin none
Warning: Changing the network port settings will cause a several second interruption in carrier. Do you want to continue? {y|n}: y
net port modify -node <<var_nodeB>> -port e0c -flowcontrol-admin none
Warning: Changing the network port settings will cause a several second interruption in carrier. Do you want to continue? {y|n}: y
net port modify -node <<var_nodeB>> -port e0d -flowcontrol-admin none
Warning: Changing the network port settings will cause a several second interruption in carrier. Do you want to continue? {y|n}: y
net port modify -node <<var_nodeB>> -port e0e -flowcontrol-admin none
Warning: Changing the network port settings will cause a several second interruption in carrier. Do you want to continue? {y|n}: y
net port modify -node <<var_nodeB>> -port e0f -flowcontrol-admin none
Warning: Changing the network port settings will cause a several second interruption in carrier. Do you want to continue? {y|n}: y
The Cisco UCS Mini direct connection to ONTAP does not support LACP.

Configure jumbo frames in NetApp ONTAP

To configure an ONTAP network port to use jumbo frames (that usually have an MTU of 9,000 bytes), run the following commands from the cluster shell:

AFF A220::> network port modify -node node_A -port e0e -mtu 9000
Warning: This command will cause a several second interruption of service on this network port.
Do you want to continue? {y|n}: y
AFF A220::> network port modify -node node_B -port e0e -mtu 9000
Warning: This command will cause a several second interruption of service on this network port.
Do you want to continue? {y|n}: y
AFF A220::> network port modify -node node_A -port e0f -mtu 9000
Warning: This command will cause a several second interruption of service on this network port.
Do you want to continue? {y|n}: y
AFF A220::> network port modify -node node_B -port e0f -mtu 9000
Warning: This command will cause a several second interruption of service on this network port.
Do you want to continue? {y|n}: y

Create VLANs in ONTAP

To create VLANs in ONTAP, complete the following steps:

  1. Create NFS VLAN ports and add them to the data broadcast domain.

    network port vlan create –node <<var_nodeA>> -vlan-name e0e-<<var_nfs_vlan_id>>
    network port vlan create –node <<var_nodeA>> -vlan-name e0f-<<var_nfs_vlan_id>>
    network port vlan create –node <<var_nodeB>> -vlan-name e0e-<<var_nfs_vlan_id>>
    network port vlan create –node <<var_nodeB>> -vlan-name e0f-<<var_nfs_vlan_id>>
    broadcast-domain add-ports -broadcast-domain Infra_NFS -ports <<var_nodeA>>: e0e- <<var_nfs_vlan_id>>, <<var_nodeB>>: e0e-<<var_nfs_vlan_id>> , <<var_nodeA>>:e0f- <<var_nfs_vlan_id>>, <<var_nodeB>>:e0f-<<var_nfs_vlan_id>>
  2. Create iSCSI VLAN ports and add them to the data broadcast domain.

    network port vlan create –node <<var_nodeA>> -vlan-name e0e-<<var_iscsi_vlan_A_id>>
    network port vlan create –node <<var_nodeA>> -vlan-name e0f-<<var_iscsi_vlan_B_id>>
    network port vlan create –node <<var_nodeB>> -vlan-name e0e-<<var_iscsi_vlan_A_id>>
    network port vlan create –node <<var_nodeB>> -vlan-name e0f-<<var_iscsi_vlan_B_id>>
    broadcast-domain add-ports -broadcast-domain Infra_iSCSI-A -ports <<var_nodeA>>: e0e- <<var_iscsi_vlan_A_id>>,<<var_nodeB>>: e0e-<<var_iscsi_vlan_A_id>>
    broadcast-domain add-ports -broadcast-domain Infra_iSCSI-B -ports <<var_nodeA>>: e0f- <<var_iscsi_vlan_B_id>>,<<var_nodeB>>: e0f-<<var_iscsi_vlan_B_id>>
  3. Create MGMT-VLAN ports.

    network port vlan create –node <<var_nodeA>> -vlan-name e0m-<<mgmt_vlan_id>>
    network port vlan create –node <<var_nodeB>> -vlan-name e0m-<<mgmt_vlan_id>>

Create aggregates in ONTAP

An aggregate containing the root volume is created during the ONTAP setup process. To create additional aggregates, determine the aggregate name, the node on which to create it, and the number of disks it contains.

To create aggregates, run the following commands:

aggr create -aggregate aggr1_nodeA -node <<var_nodeA>> -diskcount <<var_num_disks>>
aggr create -aggregate aggr1_nodeB -node <<var_nodeB>> -diskcount <<var_num_disks>>

Retain at least one disk (select the largest disk) in the configuration as a spare. A best practice is to have at least one spare for each disk type and size.

Start with five disks; you can add disks to an aggregate when additional storage is required.

The aggregate cannot be created until disk zeroing completes. Run the aggr show command to display the aggregate creation status. Do not proceed until aggr1_nodeA is online.

Configure time zone in ONTAP

To configure time synchronization and to set the time zone on the cluster, run the following command:

timezone <<var_timezone>>
For example, in the eastern United States, the time zone is America/New_York. After you begin typing the time zone name, press the Tab key to see available options.

Configure SNMP in ONTAP

To configure the SNMP, complete the following steps:

  1. Configure SNMP basic information, such as the location and contact. When polled, this information is visible as the sysLocation and sysContact variables in SNMP.

    snmp contact <<var_snmp_contact>>
    snmp location “<<var_snmp_location>>”
    snmp init 1
    options snmp.enable on
  2. Configure SNMP traps to send to remote hosts.

    snmp traphost add <<var_snmp_server_fqdn>>

Configure SNMPv1 in ONTAP

To configure SNMPv1, set the shared secret plain-text password called a community.

snmp community add ro <<var_snmp_community>>
Use the snmp community delete all command with caution. If community strings are used for other monitoring products, this command removes them.

Configure SNMPv3 in ONTAP

SNMPv3 requires that you define and configure a user for authentication. To configure SNMPv3, complete the following steps:

  1. Run the security snmpusers command to view the engine ID.

  2. Create a user called snmpv3user.

    security login create -username snmpv3user -authmethod usm -application snmp
  3. Enter the authoritative entity’s engine ID and select md5 as the authentication protocol.

  4. Enter an eight-character minimum-length password for the authentication protocol when prompted.

  5. Select des as the privacy protocol.

  6. Enter an eight-character minimum-length password for the privacy protocol when prompted.

Configure AutoSupport HTTPS in ONTAP

The NetApp AutoSupport tool sends support summary information to NetApp through HTTPS. To configure AutoSupport, run the following command:

system node autosupport modify -node * -state enable –mail-hosts <<var_mailhost>> -transport https -support enable -noteto <<var_storage_admin_email>>

Create a storage virtual machine

To create an infrastructure storage virtual machine (SVM), complete the following steps:

  1. Run the vserver create command.

    vserver create –vserver Infra-SVM –rootvolume rootvol –aggregate aggr1_nodeA –rootvolume- security-style unix
  2. Add the data aggregate to the infra-SVM aggregate list for the NetApp VSC.

    vserver modify -vserver Infra-SVM -aggr-list aggr1_nodeA,aggr1_nodeB
  3. Remove the unused storage protocols from the SVM, leaving NFS and iSCSI.

    vserver remove-protocols –vserver Infra-SVM -protocols cifs,ndmp,fcp
  4. Enable and run the NFS protocol in the infra-SVM SVM.

    nfs create -vserver Infra-SVM -udp disabled
  5. Turn on the SVM vstorage parameter for the NetApp NFS VAAI plug-in. Then, verify that NFS has been configured.

    vserver nfs modify –vserver Infra-SVM –vstorage enabled
    vserver nfs show
    Commands are prefaced by vserver in the command line because SVMs were previously called servers

Configure NFSv3 in ONTAP

The table below lists the information needed to complete this configuration.

Detail Detail Value

ESXi host A NFS IP address

<<var_esxi_hostA_nfs_ip>>

ESXi host B NFS IP address

<<var_esxi_hostB_nfs_ip>>

To configure NFS on the SVM, run the following commands:

  1. Create a rule for each ESXi host in the default export policy.

  2. For each ESXi host being created, assign a rule. Each host has its own rule index. Your first ESXi host has rule index 1, your second ESXi host has rule index 2, and so on.

    vserver export-policy rule create –vserver Infra-SVM -policyname default –ruleindex 1 –protocol nfs -clientmatch <<var_esxi_hostA_nfs_ip>> -rorule sys –rwrule sys -superuser sys –allow-suid falsevserver export-policy rule create –vserver Infra-SVM -policyname default –ruleindex 2 –protocol nfs -clientmatch <<var_esxi_hostB_nfs_ip>> -rorule sys –rwrule sys -superuser sys –allow-suid false
    vserver export-policy rule show
  3. Assign the export policy to the infrastructure SVM root volume.

    volume modify –vserver Infra-SVM –volume rootvol –policy default
    The NetApp VSC automatically handles export policies if you choose to install it after vSphere has been set up. If you do not install it, you must create export policy rules when additional Cisco UCS B-Series servers are added.

Create iSCSI service in ONTAP

To create the iSCSI service, complete the following step:

  1. Create the iSCSI service on the SVM. This command also starts the iSCSI service and sets the iSCSI Qualified Name (IQN) for the SVM. Verify that iSCSI has been configured.

    iscsi create -vserver Infra-SVM
    iscsi show

Create load-sharing mirror of SVM root volume in ONTAP

To create a load-sharing mirror of the SVM root volume in ONTAP, complete the following steps:

  1. Create a volume to be the load-sharing mirror of the infrastructure SVM root volume on each node.

    volume create –vserver Infra_Vserver –volume rootvol_m01 –aggregate aggr1_nodeA –size 1GB –type DPvolume create –vserver Infra_Vserver –volume rootvol_m02 –aggregate aggr1_nodeB –size 1GB –type DP
  2. Create a job schedule to update the root volume mirror relationships every 15 minutes.

    job schedule interval create -name 15min -minutes 15
  3. Create the mirroring relationships.

    snapmirror create -source-path Infra-SVM:rootvol -destination-path Infra-SVM:rootvol_m01 -type LS -schedule 15min
    snapmirror create -source-path Infra-SVM:rootvol -destination-path Infra-SVM:rootvol_m02 -type LS -schedule 15min
  4. Initialize the mirroring relationship and verify that it has been created.

    snapmirror initialize-ls-set -source-path Infra-SVM:rootvol snapmirror show

Configure HTTPS access in ONTAP

To configure secure access to the storage controller, complete the following steps:

  1. Increase the privilege level to access the certificate commands.

    set -privilege diag
    Do you want to continue? {y|n}: y
  2. Generally, a self-signed certificate is already in place. Verify the certificate by running the following command:

    security certificate show
  3. For each SVM shown, the certificate common name should match the DNS fully qualified domain name (FQDN) of the SVM. The four default certificates should be deleted and replaced by either self-signed certificates or certificates from a certificate authority.

    Deleting expired certificates before creating certificates is a best practice. Run the security certificate delete command to delete expired certificates. In the following command, use TAB completion to select and delete each default certificate.

    security certificate delete [TAB] ...
    Example: security certificate delete -vserver Infra-SVM -common-name Infra-SVM -ca Infra-SVM - type server -serial 552429A6
  4. To generate and install self-signed certificates, run the following commands as one-time commands. Generate a server certificate for the infra-SVM and the cluster SVM. Again, use TAB completion to aid in completing these commands.

    security certificate create [TAB] ...
    Example: security certificate create -common-name infra-svm.netapp.com -type server -size 2048 - country US -state "North Carolina" -locality "RTP" -organization "NetApp" -unit "FlexPod" -email- addr "abc@netapp.com" -expire-days 365 -protocol SSL -hash-function SHA256 -vserver Infra-SVM
  5. To obtain the values for the parameters required in the following step, run the security certificate show command.

  6. Enable each certificate that was just created using the –server-enabled true and –client- enabled false parameters. Again, use TAB completion.

    security ssl modify [TAB] ...
    Example: security ssl modify -vserver Infra-SVM -server-enabled true -client-enabled false -ca infra-svm.netapp.com -serial 55243646 -common-name infra-svm.netapp.com
  7. Configure and enable SSL and HTTPS access and disable HTTP access.

    system services web modify -external true -sslv3-enabled true
    Warning: Modifying the cluster configuration will cause pending web service requests to be interrupted as the web servers are restarted.
    Do you want to continue {y|n}: y
    System services firewall policy delete -policy mgmt -service http -vserver <<var_clustername>>
    It is normal for some of these commands to return an error message stating that the entry does not exist.
  8. Revert to the admin privilege level and create the setup to allow SVM to be available by the web.

    set –privilege admin
    vserver services web modify –name spi|ontapi|compat –vserver * -enabled true

Create a NetApp FlexVol volume in ONTAP

To create a NetApp FlexVol® volume, enter the volume name, size, and the aggregate on which it exists. Create two VMware datastore volumes and a server boot volume.

volume create -vserver Infra-SVM -volume infra_datastore_1 -aggregate aggr1_nodeA -size 500GB - state online -policy default -junction-path /infra_datastore_1 -space-guarantee none -percent- snapshot-space 0
volume create -vserver Infra-SVM -volume infra_datastore_2 -aggregate aggr1_nodeB -size 500GB - state online -policy default -junction-path /infra_datastore_2 -space-guarantee none -percent- snapshot-space 0
volume create -vserver Infra-SVM -volume infra_swap -aggregate aggr1_nodeA -size 100GB -state online -policy default -juntion-path /infra_swap -space-guarantee none -percent-snapshot-space 0 -snapshot-policy none
volume create -vserver Infra-SVM -volume esxi_boot -aggregate aggr1_nodeA -size 100GB -state online -policy default -space-guarantee none -percent-snapshot-space 0

Enable deduplication in ONTAP

To enable deduplication on appropriate volumes once a day, run the following commands:

volume efficiency modify –vserver Infra-SVM –volume esxi_boot –schedule sun-sat@0
volume efficiency modify –vserver Infra-SVM –volume infra_datastore_1 –schedule sun-sat@0
volume efficiency modify –vserver Infra-SVM –volume infra_datastore_2 –schedule sun-sat@0

Create LUNs in ONTAP

To create two boot logical unit numbers (LUNs), run the following commands:

lun create -vserver Infra-SVM -volume esxi_boot -lun VM-Host-Infra-A -size 15GB -ostype vmware - space-reserve disabled
lun create -vserver Infra-SVM -volume esxi_boot -lun VM-Host-Infra-B -size 15GB -ostype vmware - space-reserve disabled
When adding an extra Cisco UCS C-Series server, an extra boot LUN must be created.

Create iSCSI LIFs in ONTAP

The table below lists the information needed to complete this configuration.

Detail Detail Value

Storage node A iSCSI LIF01A

<<var_nodeA_iscsi_lif01a_ip>>

Storage node A iSCSI LIF01A network mask

<<var_nodeA_iscsi_lif01a_mask>>

Storage node A iSCSI LIF01B

<<var_nodeA_iscsi_lif01b_ip>>

Storage node A iSCSI LIF01B network mask

<<var_nodeA_iscsi_lif01b_mask>>

Storage node B iSCSI LIF01A

<<var_nodeB_iscsi_lif01a_ip>>

Storage node B iSCSI LIF01A network mask

<<var_nodeB_iscsi_lif01a_mask>>

Storage node B iSCSI LIF01B

<<var_nodeB_iscsi_lif01b_ip>>

Storage node B iSCSI LIF01B network mask

<<var_nodeB_iscsi_lif01b_mask>>

  1. Create four iSCSI LIFs, two on each node.

    network interface create -vserver Infra-SVM -lif iscsi_lif01a -role data -data-protocol iscsi - home-node <<var_nodeA>> -home-port e0e-<<var_iscsi_vlan_A_id>> -address <<var_nodeA_iscsi_lif01a_ip>> -netmask <<var_nodeA_iscsi_lif01a_mask>> –status-admin up – failover-policy disabled –firewall-policy data –auto-revert false
    network interface create -vserver Infra-SVM -lif iscsi_lif01b -role data -data-protocol iscsi - home-node <<var_nodeA>> -home-port e0f-<<var_iscsi_vlan_B_id>> -address <<var_nodeA_iscsi_lif01b_ip>> -netmask <<var_nodeA_iscsi_lif01b_mask>> –status-admin up – failover-policy disabled –firewall-policy data –auto-revert false
    network interface create -vserver Infra-SVM -lif iscsi_lif02a -role data -data-protocol iscsi - home-node <<var_nodeB>> -home-port e0e-<<var_iscsi_vlan_A_id>> -address <<var_nodeB_iscsi_lif01a_ip>> -netmask <<var_nodeB_iscsi_lif01a_mask>> –status-admin up – failover-policy disabled –firewall-policy data –auto-revert false
    network interface create -vserver Infra-SVM -lif iscsi_lif02b -role data -data-protocol iscsi - home-node <<var_nodeB>> -home-port e0f-<<var_iscsi_vlan_B_id>> -address <<var_nodeB_iscsi_lif01b_ip>> -netmask <<var_nodeB_iscsi_lif01b_mask>> –status-admin up – failover-policy disabled –firewall-policy data –auto-revert false
    network interface show

Create NFS LIFs in ONTAP

The following table lists the information needed to complete this configuration.

Detail Detail value

Storage node A NFS LIF 01 a IP

<<var_nodeA_nfs_lif_01_a_ip>>

Storage node A NFS LIF 01 a network mask

<<var_nodeA_nfs_lif_01_a_mask>>

Storage node A NFS LIF 01 b IP

<<var_nodeA_nfs_lif_01_b_ip>>

Storage node A NFS LIF 01 b network mask

<<var_nodeA_nfs_lif_01_b_mask>>

Storage node B NFS LIF 02 a IP

<<var_nodeB_nfs_lif_02_a_ip>>

Storage node B NFS LIF 02 a network mask

<<var_nodeB_nfs_lif_02_a_mask>>

Storage node B NFS LIF 02 b IP

<<var_nodeB_nfs_lif_02_b_ip>>

Storage node B NFS LIF 02 b network mask

<<var_nodeB_nfs_lif_02_b_mask>>

  1. Create an NFS LIF.

    network interface create -vserver Infra-SVM -lif nfs_lif01_a -role data -data-protocol nfs -home- node <<var_nodeA>> -home-port e0e-<<var_nfs_vlan_id>> –address <<var_nodeA_nfs_lif_01_a_ip>> - netmask << var_nodeA_nfs_lif_01_a_mask>> -status-admin up –failover-policy broadcast-domain-wide – firewall-policy data –auto-revert true
    network interface create -vserver Infra-SVM -lif nfs_lif01_b -role data -data-protocol nfs -home- node <<var_nodeA>> -home-port e0f-<<var_nfs_vlan_id>> –address <<var_nodeA_nfs_lif_01_b_ip>> - netmask << var_nodeA_nfs_lif_01_b_mask>> -status-admin up –failover-policy broadcast-domain-wide – firewall-policy data –auto-revert true
    network interface create -vserver Infra-SVM -lif nfs_lif02_a -role data -data-protocol nfs -home- node <<var_nodeB>> -home-port e0e-<<var_nfs_vlan_id>> –address <<var_nodeB_nfs_lif_02_a_ip>> - netmask << var_nodeB_nfs_lif_02_a_mask>> -status-admin up –failover-policy broadcast-domain-wide – firewall-policy data –auto-revert true
    network interface create -vserver Infra-SVM -lif nfs_lif02_b -role data -data-protocol nfs -home- node <<var_nodeB>> -home-port e0f-<<var_nfs_vlan_id>> –address <<var_nodeB_nfs_lif_02_b_ip>> - netmask << var_nodeB_nfs_lif_02_b_mask>> -status-admin up –failover-policy broadcast-domain-wide – firewall-policy data –auto-revert true
    network interface show

Add infrastructure SVM administrator

The following table lists the information needed to complete this configuration.

Detail Detail value

Vsmgmt IP

<<var_svm_mgmt_ip>>

Vsmgmt network mask

<<var_svm_mgmt_mask>>

Vsmgmt default gateway

<<var_svm_mgmt_gateway>>

To add the infrastructure SVM administrator and SVM administration LIF to the management network, complete the following steps:

  1. Run the following command:

    network interface create –vserver Infra-SVM –lif vsmgmt –role data –data-protocol none –home-node <<var_nodeB>> -home-port e0M –address <<var_svm_mgmt_ip>> -netmask <<var_svm_mgmt_mask>> - status-admin up –failover-policy broadcast-domain-wide –firewall-policy mgmt –auto-revert true
    The SVM management IP here should be in the same subnet as the storage cluster management IP.
  2. Create a default route to allow the SVM management interface to reach the outside world.

    network route create –vserver Infra-SVM -destination 0.0.0.0/0 –gateway <<var_svm_mgmt_gateway>> network route show
  3. Set a password for the SVM vsadmin user and unlock the user.

    security login password –username vsadmin –vserver Infra-SVM
    Enter a new password: <<var_password>>
    Enter it again: <<var_password>>
    security login unlock –username vsadmin –vserver

Cisco UCS server configuration

FlexPod Cisco UCS base

Perform Initial Setup of Cisco UCS 6324 Fabric Interconnect for FlexPod Environments.

This section provides detailed procedures to configure Cisco UCS for use in a FlexPod ROBO environment by using Cisco UCS Manger.

Cisco UCS fabric interconnect 6324 A

Cisco UCS uses access layer networking and servers. This high-performance, next-generation server system provides a data center with a high degree of workload agility and scalability.

Cisco UCS Manager 4.0(1b) supports the 6324 Fabric Interconnect that integrates the Fabric Interconnect into the Cisco UCS Chassis and provides an integrated solution for a smaller deployment environment. Cisco UCS Mini simplifies the system management and saves cost for the low scale deployments.

The hardware and software components support Cisco’s unified fabric, which runs multiple types of data center traffic over a single converged network adapter.

Initial system setup

The first time when you access a fabric interconnect in a Cisco UCS domain, a setup wizard prompts you for the following information required to configure the system:

  • Installation method (GUI or CLI)

  • Setup mode (restore from full system backup or initial setup)

  • System configuration type (standalone or cluster configuration)

  • System name

  • Admin password

  • Management port IPv4 address and subnet mask, or IPv6 address and prefix

  • Default gateway IPv4 or IPv6 address

  • DNS Server IPv4 or IPv6 address

  • Default domain name

The following table lists the information needed to complete the Cisco UCS initial configuration on Fabric Interconnect A

Detail Detail/value

System Name 

<<var_ucs_clustername>>

Admin Password

<<var_password>>

Management IP Address: Fabric Interconnect A

<<var_ucsa_mgmt_ip>>

Management netmask: Fabric Interconnect A

<<var_ucsa_mgmt_mask>>

Default gateway: Fabric Interconnect A

<<var_ucsa_mgmt_gateway>>

Cluster IP address

<<var_ucs_cluster_ip>>

DNS server IP address

<<var_nameserver_ip>>

Domain name

<<var_domain_name>>

To configure the Cisco UCS for use in a FlexPod environment, complete the following steps:

  1. Connect to the console port on the first Cisco UCS 6324 Fabric Interconnect A.

    Enter the configuration method. (console/gui) ? console
    
      Enter the setup mode; setup newly or restore from backup. (setup/restore) ? setup
    
      You have chosen to setup a new Fabric interconnect. Continue? (y/n): y
    
      Enforce strong password? (y/n) [y]: Enter
    
      Enter the password for "admin":<<var_password>>
      Confirm the password for "admin":<<var_password>>
    
      Is this Fabric interconnect part of a cluster(select 'no' for standalone)? (yes/no) [n]: yes
    
      Enter the switch fabric (A/B) []: A
    
      Enter the system name: <<var_ucs_clustername>>
    
      Physical Switch Mgmt0 IP address : <<var_ucsa_mgmt_ip>>
    
      Physical Switch Mgmt0 IPv4 netmask : <<var_ucsa_mgmt_mask>>
    
      IPv4 address of the default gateway : <<var_ucsa_mgmt_gateway>>
    
      Cluster IPv4 address : <<var_ucs_cluster_ip>>
    
      Configure the DNS Server IP address? (yes/no) [n]: y
    
           DNS IP address : <<var_nameserver_ip>>
    
      Configure the default domain name? (yes/no) [n]: y
    Default domain name: <<var_domain_name>>
    
      Join centralized management environment (UCS Central)? (yes/no) [n]: no
    
     NOTE: Cluster IP will be configured only after both Fabric Interconnects are initialized. UCSM will be functional only after peer FI is configured in clustering mode.
    
      Apply and save the configuration (select 'no' if you want to re-enter)? (yes/no): yes
      Applying configuration. Please wait.
    
      Configuration file - Ok
  2. Review the settings displayed on the console. If they are correct, answer yes to apply and save the configuration.

  3. Wait for the login prompt to verify that the configuration has been saved.

The following table lists the information needed to complete the Cisco UCS initial configuration on Fabric Interconnect B.

Detail Detail/value

System Name 

<<var_ucs_clustername>>

Admin Password

<<var_password>>

Management IP Address-FI B

<<var_ucsb_mgmt_ip>>

Management Netmask-FI B

<<var_ucsb_mgmt_mask>>

Default Gateway-FI B

<<var_ucsb_mgmt_gateway>>

Cluster IP Address

<<var_ucs_cluster_ip>>

DNS Server IP address

<<var_nameserver_ip>>

Domain Name

<<var_domain_name>>

  1. Connect to the console port on the second Cisco UCS 6324 Fabric Interconnect B.

     Enter the configuration method. (console/gui) ? console
    
      Installer has detected the presence of a peer Fabric interconnect. This Fabric interconnect will be added to the cluster. Continue (y/n) ? y
    
      Enter the admin password of the peer Fabric interconnect:<<var_password>>
        Connecting to peer Fabric interconnect... done
        Retrieving config from peer Fabric interconnect... done
        Peer Fabric interconnect Mgmt0 IPv4 Address: <<var_ucsb_mgmt_ip>>
        Peer Fabric interconnect Mgmt0 IPv4 Netmask: <<var_ucsb_mgmt_mask>>
        Cluster IPv4 address: <<var_ucs_cluster_address>>
    
        Peer FI is IPv4 Cluster enabled. Please Provide Local Fabric Interconnect Mgmt0 IPv4 Address
    
      Physical Switch Mgmt0 IP address : <<var_ucsb_mgmt_ip>>
    
    
      Apply and save the configuration (select 'no' if you want to re-enter)? (yes/no): yes
      Applying configuration. Please wait.
    
      Configuration file - Ok
  2. Wait for the login prompt to confirm that the configuration has been saved.

Log into Cisco UCS Manager

To log into the Cisco Unified Computing System (UCS) environment, complete the following steps:

  1. Open a web browser and navigate to the Cisco UCS Fabric Interconnect cluster address.

    You may need to wait at least 5 minutes after configuring the second fabric interconnect for Cisco UCS Manager to come up.

  2. Click the Launch UCS Manager link to launch Cisco UCS Manager.

  3. Accept the necessary security certificates.

  4. When prompted, enter admin as the user name and enter the administrator password.

  5. Click Login to log in to Cisco UCS Manager.

Cisco UCS Manager software version 4.0(1b)

This document assumes the use of Cisco UCS Manager Software version 4.0(1b). To upgrade the Cisco UCS Manager software and the Cisco UCS 6324 Fabric Interconnect software refer to  Cisco UCS Manager Install and Upgrade Guides.

Configure Cisco UCS Call Home

Cisco highly recommends that you configure Call Home in Cisco UCS Manager. Configuring Call Home accelerates the resolution of support cases. To configure Call Home, complete the following steps:

  1. In Cisco UCS Manager, click Admin on the left.

  2. Select All > Communication Management > Call Home.

  3. Change the State to On.

  4. Fill in all the fields according to your Management preferences and click Save Changes and OK to complete configuring Call Home.

Add block of IP addresses for keyboard, video, mouse access

To create a block of IP addresses for in band server keyboard, video, mouse (KVM) access in the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, click LAN on the left.

  2. Expand Pools > root > IP Pools.

  3. Right-click IP Pool ext-mgmt and select Create Block of IPv4 Addresses.

  4. Enter the starting IP address of the block, number of IP addresses required, and the subnet mask and gateway information.

    Error: Missing Graphic Image

  5. Click OK to create the block.

  6. Click OK in the confirmation message.

Synchronize Cisco UCS to NTP

To synchronize the Cisco UCS environment to the NTP servers in the Nexus switches, complete the following steps:

  1. In Cisco UCS Manager, click Admin on the left.

  2. Expand All > Time Zone Management.

  3. Select Time Zone.

  4. In the Properties pane, select the appropriate time zone in the Time Zone menu.

  5. Click Save Changes and click OK.

  6. Click Add NTP Server.

  7. Enter <switch-a-ntp-ip> or <Nexus-A-mgmt-IP> and click OK. Click OK.

    Error: Missing Graphic Image

  8. Click Add NTP Server.

  9. Enter <switch-b-ntp-ip> or <Nexus-B-mgmt-IP> and click OK. Click OK on the confirmation.

    Error: Missing Graphic Image

Edit chassis discovery policy

Setting the discovery policy simplifies the addition of Cisco UCS B-Series chassis and of additional fabric extenders for further Cisco UCS C-Series connectivity. To modify the chassis discovery policy, complete the following steps:

  1. In Cisco UCS Manager, click Equipment on the left and select Equipment in the second list.

  2. In the right pane, select the Policies tab.

  3. Under Global Policies, set the Chassis/FEX Discovery Policy to match the minimum number of uplink ports that are cabled between the chassis or fabric extenders (FEXes) and the fabric interconnects.

  4. Set the Link Grouping Preference to Port Channel. If the environment being setup contains a large amount of multicast traffic, set the Multicast Hardware Hash setting to Enabled.

  5. Click Save Changes.

  6. Click OK.

To enable server and uplink ports, complete the following steps:

  1. In Cisco UCS Manager, in the navigation pane, select the Equipment tab.

  2. Expand Equipment > Fabric Interconnects > Fabric Interconnect A > Fixed Module.

  3. Expand Ethernet Ports.

  4. Select ports 1 and 2 that are connected to the Cisco Nexus 31108 switches, right-click, and select Configure as Uplink Port.

  5. Click Yes to confirm the uplink ports and click OK.

  6. Select ports 3 and 4 that are connected to the NetApp Storage Controllers, right-click, and select Configure as Appliance Port.

  7. Click Yes to confirm the appliance ports.

  8. On the Configure as Appliance Port window, click OK. 

  9. Click OK to confirm.

  10. In the left pane, select Fixed Module under Fabric Interconnect A. 

  11. From the Ethernet Ports tab, confirm that ports have been configured correctly in the If Role column. If any port C-Series servers were configured on the Scalability port, click on it to verify port connectivity there.

    Error: Missing Graphic Image

  12. Expand Equipment > Fabric Interconnects > Fabric Interconnect B > Fixed Module.

  13. Expand Ethernet Ports.

  14. Select Ethernet ports 1 and 2 that are connected to the Cisco Nexus 31108 switches, right-click, and select Configure as Uplink Port.

  15. Click Yes to confirm the uplink ports and click OK.

  16. Select ports 3 and 4 that are connected to the NetApp Storage Controllers, right-click, and select Configure as Appliance Port.

  17. Click Yes to confirm the appliance ports.

  18. On the Configure as Appliance Port window, click OK.

  19. Click OK to confirm.

  20. In the left pane, select Fixed Module under Fabric Interconnect B. 

  21. From the Ethernet Ports tab, confirm that ports have been configured correctly in the If Role column. If any port C-Series servers were configured on the Scalability port, click it to verify port connectivity there.

    Error: Missing Graphic Image

To configure the necessary port channels in the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, select the LAN tab in the navigation pane.

    In this procedure, two port channels are created: one from Fabric A to both Cisco Nexus 31108 switches and one from Fabric B to both Cisco Nexus 31108 switches. If you are using standard switches, modify this procedure accordingly. If you are using 1 Gigabit Ethernet (1GbE) switches and GLC-T SFPs on the Fabric Interconnects, the interface speeds of Ethernet ports 1/1 and 1/2 in the Fabric Interconnects must be set to 1Gbps.
  2. Under LAN > LAN Cloud, expand the Fabric A tree.

  3. Right-click Port Channels.

  4. Select Create Port Channel.

  5. Enter 13 as the unique ID of the port channel.

  6. Enter vPC-13-Nexus as the name of the port channel.

  7. Click Next.

    Error: Missing Graphic Image

  8. Select the following ports to be added to the port channel:

    1. Slot ID 1 and port 1

    2. Slot ID 1 and port 2

  9. Click >> to add the ports to the port channel.

  10. Click Finish to create the port channel. Click OK.

  11. Under Port Channels, select the newly created port channel.

    The port channel should have an Overall Status of Up.

  12. In the navigation pane, under LAN > LAN Cloud, expand the Fabric B tree.

  13. Right-click Port Channels.

  14. Select Create Port Channel.

  15. Enter 14 as the unique ID of the port channel.

  16. Enter vPC-14-Nexus as the name of the port channel. Click Next.

  17. Select the following ports to be added to the port channel:

    1. Slot ID 1 and port 1

    2. Slot ID 1 and port 2

  18. Click >> to add the ports to the port channel.

  19. Click Finish to create the port channel. Click OK.

  20. Under Port Channels, select the newly created port-channel.

  21. The port channel should have an Overall Status of Up.

Create an organization (optional)

Organizations are used to organizing resources and restricting access to various groups within the IT organization, thereby enabling multitenancy of the compute resources.

Although this document does not assume the use of organizations, this procedure provides instructions for creating one.

To configure an organization in the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, from the New menu in the toolbar at the top of the window, select Create Organization.

  2. Enter a name for the organization.

  3. Optional: Enter a description for the organization. Click OK.

  4. Click OK in the confirmation message.

Configure storage appliance ports and storage VLANs

To configure the storage appliance ports and storage VLANs, complete the following steps:

  1. In the Cisco UCS Manager, select the LAN tab.

  2. Expand the Appliances cloud.

  3. Right-click VLANs under Appliances Cloud.

  4. Select Create VLANs.

  5. Enter NFS-VLAN as the name for the Infrastructure NFS VLAN.

  6. Leave Common/Global selected.

  7. Enter <<var_nfs_vlan_id>> for the VLAN ID.

  8. Leave Sharing Type set to None.

    Error: Missing Graphic Image

  9. Click OK, and then click OK again to create the VLAN.

  10. Right-click VLANs under Appliances Cloud.

  11. Select Create VLANs.

  12. Enter iSCSI-A-VLAN as the name for the Infrastructure iSCSI Fabric A VLAN.

  13. Leave Common/Global selected.

  14. Enter <<var_iscsi-a_vlan_id>> for the VLAN ID.

  15. Click OK, and then click OK again to create the VLAN.

  16. Right-click VLANs under Appliances Cloud.

  17. Select Create VLANs.

  18. Enter iSCSI-B-VLAN as the name for the Infrastructure iSCSI Fabric B VLAN.

  19. Leave Common/Global selected.

  20. Enter <<var_iscsi-b_vlan_id>> for the VLAN ID.

  21. Click OK, and then click OK again to create the VLAN.

  22. Right-click VLANs under Appliances Cloud.

  23. Select Create VLANs.

  24. Enter Native-VLAN as the name for the Native VLAN.

  25. Leave Common/Global selected.

  26. Enter <<var_native_vlan_id>> for the VLAN ID.

  27. Click OK, and then click OK again to create the VLAN.

    Error: Missing Graphic Image

  28. In the navigation pane, under LAN > Policies, expand Appliances and right-click Network Control Policies.

  29. Select Create Network Control Policy.

  30. Name the policy Enable_CDP_LLPD and select Enabled next to CDP.

  31. Enable the Transmit and Receive features for LLDP.

    Error: Missing Graphic Image

  32. Click OK and then click OK again to create the policy.

  33. In the navigation pane, under LAN > Appliances Cloud, expand the Fabric A tree.

  34. Expand Interfaces.

  35. Select Appliance Interface 1/3.

  36. In the User Label field, put in information indicating the storage controller port, such as <storage_controller_01_name>:e0e. Click Save Changes and OK.

  37. Select the Enable_CDP Network Control Policy and select Save Changes and OK.

  38. Under VLANs, select the iSCSI-A-VLAN, NFS VLAN, and Native VLAN. Set the Native-VLAN as the Native VLAN. Clear the default VLAN selection.

  39. Click Save Changes and OK.

    Error: Missing Graphic Image

  40. Select Appliance Interface 1/4 under Fabric A.

  41. In the User Label field, put in information indicating the storage controller port, such as <storage_controller_02_name>:e0e. Click Save Changes and OK.

  42. Select the Enable_CDP Network Control Policy and select Save Changes and OK.

  43. Under VLANs, select the iSCSI-A-VLAN, NFS VLAN, and Native VLAN.

  44. Set the Native-VLAN as the Native VLAN. 

  45. Clear the default VLAN selection.

  46. Click Save Changes and OK.

  47. In the navigation pane, under LAN > Appliances Cloud, expand the Fabric B tree.

  48. Expand Interfaces.

  49. Select Appliance Interface 1/3.

  50. In the User Label field, put in information indicating the storage controller port, such as <storage_controller_01_name>:e0f. Click Save Changes and OK.

  51. Select the Enable_CDP Network Control Policy and select Save Changes and OK.

  52. Under VLANs, select the iSCSI-B-VLAN, NFS VLAN, and Native VLAN. Set the Native-VLAN as the Native VLAN. Unselect the default VLAN.

    Error: Missing Graphic Image

  53. Click Save Changes and OK.

  54. Select Appliance Interface 1/4 under Fabric B.

  55. In the User Label field, put in information indicating the storage controller port, such as <storage_controller_02_name>:e0f. Click Save Changes and OK.

  56. Select the Enable_CDP Network Control Policy and select Save Changes and OK.

  57. Under VLANs, select the iSCSI-B-VLAN, NFS VLAN, and Native VLAN. Set the Native-VLAN as the Native VLAN. Unselect the default VLAN.

  58. Click Save Changes and OK.

Set jumbo frames in Cisco UCS fabric

To configure jumbo frames and enable quality of service in the Cisco UCS fabric, complete the following steps:

  1. In Cisco UCS Manager, in the navigation pane, click the LAN tab.

  2. Select LAN > LAN Cloud > QoS System Class.

  3. In the right pane, click the General tab.

  4. On the Best Effort row, enter 9216 in the box under the MTU column.

    Error: Missing Graphic Image

  5. Click Save Changes.

  6. Click OK.

Acknowledge Cisco UCS chassis

To acknowledge all Cisco UCS chassis, complete the following steps:

  1. In Cisco UCS Manager, select the Equipment tab, then Expand the Equipment tab on the right.

  2. Expand Equipment > Chassis.

  3. In the Actions for Chassis 1, select Acknowledge Chassis.

  4. Click OK and then click OK to complete acknowledging the chassis.

  5. Click Close to close the Properties window.

Load Cisco UCS 4.0(1b) firmware images

To upgrade the Cisco UCS Manager software and the Cisco UCS Fabric Interconnect software to version 4.0(1b) refer to Cisco UCS Manager Install and Upgrade Guides.

Create host firmware package

Firmware management policies allow the administrator to select the corresponding packages for a given server configuration. These policies often include packages for adapter, BIOS, board controller, FC adapters, host bus adapter (HBA) option ROM, and storage controller properties.

To create a firmware management policy for a given server configuration in the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, click Servers on the left.

  2. Select Policies > root.

  3. Expand Host Firmware Packages.

  4. Select default.

  5. In the Actions pane, select Modify Package Versions.

  6. Select the version 4.0(1b) for both the Blade Packages.

    Error: Missing Graphic Image

  7. Click OK then OK again to modify the host firmware package.

Create MAC address pools

To configure the necessary MAC address pools for the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, click LAN on the left.

  2. Select Pools > root.

    In this procedure, two MAC address pools are created, one for each switching fabric.

  3. Right-click MAC Pools under the root organization.

  4. Select Create MAC Pool to create the MAC address pool.

  5. Enter MAC-Pool-A as the name of the MAC pool.

  6. Optional: Enter a description for the MAC pool.

  7. Select Sequential as the option for Assignment Order. Click Next.

  8. Click Add.

  9. Specify a starting MAC address.

    For the FlexPod solution, the recommendation is to place 0A in the next-to-last octet of the starting MAC address to identify all of the MAC addresses as fabric A addresses. In our example, we have carried forward the example of also embedding the Cisco UCS domain number information giving us 00:25:B5:32:0A:00 as our first MAC address.
  10. Specify a size for the MAC address pool that is sufficient to support the available blade or server resources. Click OK.

    Error: Missing Graphic Image

  11. Click Finish.

  12. In the confirmation message, click OK.

  13. Right-click MAC Pools under the root organization.

  14. Select Create MAC Pool to create the MAC address pool.

  15. Enter MAC-Pool-B as the name of the MAC pool.

  16. Optional: Enter a description for the MAC pool.

  17. Select Sequential as the option for Assignment Order. Click Next.

  18. Click Add.

  19. Specify a starting MAC address.

    For the FlexPod solution, it is recommended to place 0B in the next to last octet of the starting MAC address to identify all the MAC addresses in this pool as fabric B addresses. Once again, we have carried forward in our example of also embedding the Cisco UCS domain number information giving us 00:25:B5:32:0B:00 as our first MAC address.
  20. Specify a size for the MAC address pool that is sufficient to support the available blade or server resources. Click OK.

  21. Click Finish.

  22. In the confirmation message, click OK.

Create iSCSI IQN pool

To configure the necessary IQN pools for the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, click SAN on the left.

  2. Select Pools > root.

  3. Right- click IQN Pools.

  4. Select Create IQN Suffix Pool to create the IQN pool.

  5. Enter IQN-Pool for the name of the IQN pool.

  6. Optional: Enter a description for the IQN pool.

  7. Enter iqn.1992-08.com.cisco as the prefix.

  8. Select Sequential for Assignment Order. Click Next.

  9. Click Add.

  10. Enter ucs-host as the suffix.

    If multiple Cisco UCS domains are being used, a more specific IQN suffix might need to be used.
  11. Enter 1 in the From field.

  12. Specify the size of the IQN block sufficient to support the available server resources. Click OK.

    Error: Missing Graphic Image

  13. Click Finish.

Create iSCSI initiator IP address pools

To configure the necessary IP pools iSCSI boot for the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, click LAN on the left.

  2. Select Pools > root.

  3. Right-click IP Pools.

  4. Select Create IP Pool.

  5. Enter iSCSI-IP-Pool-A as the name of IP pool.

  6. Optional: Enter a description for the IP pool.

  7. Select Sequential for the assignment order. Click Next.

  8. Click Add to add a block of IP address.

  9. In the From field, enter the beginning of the range to assign as iSCSI IP addresses.

  10. Set the size to enough addresses to accommodate the servers. Click OK.

  11. Click Next.

  12. Click Finish.

  13. Right-click IP Pools.

  14. Select Create IP Pool.

  15. Enter iSCSI-IP-Pool-B as the name of IP pool.

  16. Optional: Enter a description for the IP pool.

  17. Select Sequential for the assignment order. Click Next.

  18. Click Add to add a block of IP address.

  19. In the From field, enter the beginning of the range to assign as iSCSI IP addresses.

  20. Set the size to enough addresses to accommodate the servers. Click OK.

  21. Click Next.

  22. Click Finish.

Create UUID suffix pool

To configure the necessary universally unique identifier (UUID) suffix pool for the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, click Servers on the left.

  2. Select Pools > root.

  3. Right-click UUID Suffix Pools.

  4. Select Create UUID Suffix Pool.

  5. Enter UUID-Pool as the name of the UUID suffix pool.

  6. Optional: Enter a description for the UUID suffix pool.

  7. Keep the prefix at the derived option.

  8. Select Sequential for the Assignment Order.

  9. Click Next.

  10. Click Add to add a block of UUIDs.

  11. Keep the From field at the default setting.

  12. Specify a size for the UUID block that is sufficient to support the available blade or server resources. Click OK.

  13. Click Finish.

  14. Click OK.

Create server pool

To configure the necessary server pool for the Cisco UCS environment, complete the following steps:

Consider creating unique server pools to achieve the granularity that is required in your environment.
  1. In Cisco UCS Manager, click Servers on the left.

  2. Select Pools > root.

  3. Right-click Server Pools.

  4. Select Create Server Pool.

  5. Enter `Infra-Pool `as the name of the server pool.

  6. Optional: Enter a description for the server pool. Click Next.

  7. Select two (or more) servers to be used for the VMware management cluster and click >> to add them to the `Infra-Pool `server pool.

  8. Click Finish.

  9. Click OK.

To create a Network Control Policy for Cisco Discovery Protocol (CDP) and Link Layer Discovery Protocol (LLDP), complete the following steps:

  1. In Cisco UCS Manager, click LAN on the left.

  2. Select Policies > root.

  3. Right-click Network Control Policies.

  4. Select Create Network Control Policy.

  5. Enter Enable-CDP-LLDP policy name.

  6. For CDP, select the Enabled option.

  7. For LLDP, scroll down and select Enabled for both Transmit and Receive.

  8. Click OK to create the network control policy. Click OK.

    Error: Missing Graphic Image

Create power control policy

To create a power control policy for the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, click Servers tab on the left.

  2. Select Policies > root.

  3. Right-click Power Control Policies.

  4. Select Create Power Control Policy.

  5. Enter No-Power-Cap as the power control policy name.

  6. Change the power capping setting to No Cap.

  7. Click OK to create the power control policy. Click OK.

    Error: Missing Graphic Image

Create server pool qualification policy (Optional)

To create an optional server pool qualification policy for the Cisco UCS environment, complete the following steps:

This example creates a policy for Cisco UCS B-Series servers with the Intel E2660 v4 Xeon Broadwell processors.
  1. In Cisco UCS Manager, click Servers on the left.

  2. Select Policies > root.

  3. Select Server Pool Policy Qualifications.

  4. Select Create Server Pool Policy Qualification or Add.

  5. Name the policy Intel.

  6. Select Create CPU/Cores Qualifications.

  7. Select Xeon for the Processor/Architecture.

  8. Enter <UCS-CPU- PID> as the process ID (PID).

  9. Click OK to create the CPU/Core qualification.

  10. Click OK to create the policy, and then click OK for the confirmation.

    Error: Missing Graphic Image

Create server BIOS policy

To create a server BIOS policy for the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, click Servers on the left.

  2. Select Policies > root.

  3. Right-click BIOS Policies.

  4. Select Create BIOS Policy.

  5. Enter VM-Host as the BIOS policy name.

  6. Change the Quiet Boot setting to disabled.

  7. Change Consistent Device Naming to enabled.

    Error: Missing Graphic Image

  8. Select the Processor tab and set the following parameters:

    • Processor C State: disabled

    • Processor C1E: disabled

    • Processor C3 Report: disabled

    • Processor C7 Report: disabled

      Error: Missing Graphic Image

  9. Scroll down to the remaining Processor options and set the following parameters:

    • Energy Performance: performance

    • Frequency Floor Override: enabled

    • DRAM Clock Throttling: performance

      Error: Missing Graphic Image

  10. Click RAS Memory and set the following parameters:

    • LV DDR Mode: performance mode

      Error: Missing Graphic Image

  11. Click Finish to create the BIOS policy.

  12. Click OK.

Update the default maintenance policy

To update the default Maintenance Policy, complete the following steps:

  1. In Cisco UCS Manager, click Servers on the left.

  2. Select Policies > root.

  3. Select Maintenance Policies > default.

  4. Change the Reboot Policy to User Ack.

  5. Select On Next Boot to delegate maintenance windows to server administrators.

    Error: Missing Graphic Image

  6. Click Save Changes.

  7. Click OK to accept the change.

Create vNIC templates

To create multiple virtual network interface card (vNIC) templates for the Cisco UCS environment, complete the procedures described in this section.

A total of four vNIC templates are created.  

Create infrastructure vNICs

To create an infrastructure vNIC, complete the following steps:

  1. In Cisco UCS Manager, click LAN on the left.

  2. Select Policies > root.

  3. Right-click vNIC Templates.

  4. Select Create vNIC Template.

  5. Enter Site-XX-vNIC_A as the vNIC template name.

  6. Select updating-template as the Template Type.

  7. For Fabric ID, select Fabric A.

  8. Ensure that the Enable Failover option is not selected.

  9. Select Primary Template for Redundancy Type.

  10. Leave the Peer Redundancy Template set to <not set>.

  11. Under Target, make sure that only the Adapter option is selected.

  12. Set Native-VLAN as the native VLAN.

  13. Select vNIC Name for the CDN Source.

  14. For MTU, enter 9000.

  15. Under Permitted VLANs, select Native-VLAN, Site-XX-IB-MGMT, Site-XX-NFS, Site-XX-VM-Traffic, and Site-XX-vMotion. Use the Ctrl key to make this multiple selection.

  16. Click Select. These VLANs should now appear under Selected VLANs.

  17. In the MAC Pool list, select MAC_Pool_A.

  18. In the Network Control Policy list, select Pool-A.

  19. In the Network Control Policy list, select Enable-CDP-LLDP.

  20. Click OK to create the vNIC template.

  21. Click OK.

    Error: Missing Graphic Image

To create the secondary redundancy template Infra-B, complete the following steps:

  1. In Cisco UCS Manager, click LAN on the left.

  2. Select Policies > root.

  3. Right-click vNIC Templates.

  4. Select Create vNIC Template.

  5. Enter `Site-XX-vNIC_B `as the vNIC template name.

  6. Select updating-template as the Template Type.

  7. For Fabric ID, select Fabric B.

  8. Select the Enable Failover option.

    Selecting Failover is a critical step to improve link failover time by handling it at the hardware level, and to guard against any potential for NIC failure not being detected by the virtual switch.
  9. Select Primary Template for Redundancy Type.

  10. Leave the Peer Redundancy Template set to vNIC_Template_A.

  11. Under Target, make sure that only the Adapter option is selected.

  12. Set Native-VLAN as the native VLAN.

  13. Select vNIC Name for the CDN Source.

  14. For MTU, enter 9000.

  15. Under Permitted VLANs, select Native-VLAN, Site-XX-IB-MGMT, Site-XX-NFS, Site-XX-VM-Traffic, and Site-XX-vMotion. Use the Ctrl key to make this multiple selection.

  16. Click Select. These VLANs should now appear under Selected VLANs.

  17. In the MAC Pool list, select MAC_Pool_B.

  18. In the Network Control Policy list, select Pool-B.

  19. In the Network Control Policy list, select Enable-CDP-LLDP. 

  20. Click OK to create the vNIC template.

  21. Click OK.

    Error: Missing Graphic Image

Create iSCSI vNICs

To create iSCSI vNICs, complete the following steps:

  1. Select LAN on the left.

  2. Select Policies > root.

  3. Right-click vNIC Templates.

  4. Select Create vNIC Template. 

  5. Enter Site- 01-iSCSI_A as the vNIC template name.

  6. Select Fabric A. Do not select the Enable Failover option. 

  7. Leave Redundancy Type set at No Redundancy.

  8. Under Target, make sure that only the Adapter option is selected.

  9. Select Updating Template for Template Type.

  10. Under VLANs, select only Site- 01-iSCSI_A_VLAN.

  11. Select Site- 01-iSCSI_A_VLAN as the native VLAN.

  12. Leave vNIC Name set for the CDN Source. 

  13. Under MTU, enter 9000. 

  14. From the MAC Pool list, select MAC-Pool-A.

  15. From the Network Control Policy list, select Enable-CDP-LLDP.

  16. Click OK to complete creating the vNIC template.

  17. Click OK.

    Error: Missing Graphic Image

  18. Select LAN on the left.

  19. Select Policies > root.

  20. Right-click vNIC Templates.

  21. Select Create vNIC Template.

  22. Enter Site- 01-iSCSI_B as the vNIC template name.

  23. Select Fabric B. Do not select the Enable Failover option.

  24. Leave Redundancy Type set at No Redundancy.

  25. Under Target, make sure that only the Adapter option is selected.

  26. Select Updating Template for Template Type.

  27. Under VLANs, select only Site- 01-iSCSI_B_VLAN.

  28. Select Site- 01-iSCSI_B_VLAN as the native VLAN.

  29. Leave vNIC Name set for the CDN Source.

  30. Under MTU, enter 9000.

  31. From the MAC Pool list, select MAC-Pool-B

  32. From the Network Control Policy list, select Enable-CDP-LLDP.

  33. Click OK to complete creating the vNIC template.

  34. Click OK.

    Error: Missing Graphic Image

Create LAN connectivity policy for iSCSI boot

This procedure applies to a Cisco UCS environment in which two iSCSI LIFs are on cluster node 1 (iscsi_lif01a and iscsi_lif01b) and two iSCSI LIFs are on cluster node 2 (iscsi_lif02a and iscsi_lif02b). Also, it is assumed that the A LIFs are connected to Fabric A (Cisco UCS 6324 A) and the B LIFs are connected to Fabric B (Cisco UCS 6324 B).

To configure the necessary Infrastructure LAN Connectivity Policy, complete the following steps:

  1. In Cisco UCS Manager, click LAN on the left.

  2. Select LAN > Policies > root.

  3. Right-click LAN Connectivity Policies.

  4. Select Create LAN Connectivity Policy.

  5. Enter Site-XX-Fabric-A as the name of the policy.

  6. Click the upper Add option to add a vNIC.

  7. In the Create vNIC dialog box, enter Site-01-vNIC-A as the name of the vNIC.

  8. Select the Use vNIC Template option.

  9. In the vNIC Template list, select vNIC_Template_A.

  10. From the Adapter Policy drop-down list, select VMWare.

  11. Click OK to add this vNIC to the policy.

    Error: Missing Graphic Image

  12. Click the upper Add option to add a vNIC.

  13. In the Create vNIC dialog box, enter Site-01-vNIC-B as the name of the vNIC.

  14. Select the Use vNIC Template option.

  15. In the vNIC Template list, select vNIC_Template_B.

  16. From the Adapter Policy drop-down list, select VMWare.

  17. Click OK to add this vNIC to the policy.

  18. Click the upper Add option to add a vNIC.

  19. In the Create vNIC dialog box, enter Site-01- iSCSI-A as the name of the vNIC.

  20. Select the Use vNIC Template option.

  21. In the vNIC Template list, select Site-01-iSCSI-A.

  22. From the Adapter Policy drop-down list, select VMWare.

  23. Click OK to add this vNIC to the policy.

  24. Click the upper Add option to add a vNIC.

  25. In the Create vNIC dialog box, enter Site-01-iSCSI-B as the name of the vNIC.

  26. Select the Use vNIC Template option.

  27. In the vNIC Template list, select Site-01-iSCSI-B.

  28. From the Adapter Policy drop-down list, select VMWare.

  29. Click OK to add this vNIC to the policy.

  30. Expand the Add iSCSI vNICs option.

  31. Click the Lower Add option in the Add iSCSI vNICs space to add the iSCSI vNIC.

  32. In the Create iSCSI vNIC dialog box, enter Site-01-iSCSI-A as the name of the vNIC.

  33. Select the Overlay vNIC as Site-01-iSCSI-A.

  34. Leave the iSCSI Adapter Policy option to Not Set.

  35. Select the VLAN as Site-01-iSCSI-Site-A (native).

  36. Select None (used by default) as the MAC address assignment.

  37. Click OK to add the iSCSI vNIC to the policy.

    Error: Missing Graphic Image

  38. Click the Lower Add option in the Add iSCSI vNICs space to add the iSCSI vNIC.

  39. In the Create iSCSI vNIC dialog box, enter Site-01-iSCSI-B as the name of the vNIC.

  40. Select the Overlay vNIC as Site-01-iSCSI-B.

  41. Leave the iSCSI Adapter Policy option to Not Set.

  42. Select the VLAN as Site-01-iSCSI-Site-B (native).

  43. Select None(used by default) as the MAC Address Assignment.

  44. Click OK to add the iSCSI vNIC to the policy.

  45. Click Save Changes.

    Error: Missing Graphic Image

Create vMedia policy for VMware ESXi 6.7U1 install boot

In the NetApp Data ONTAP setup steps an HTTP web server is required, which is used for hosting NetApp Data ONTAP as well as VMware software. The vMedia Policy created here maps the VMware ESXi 6. 7U1 ISO to the Cisco UCS server in order to boot the ESXi installation. To create this policy, complete the following steps:

  1. In Cisco UCS Manager, select Servers on the left.

  2. Select Policies > root.

  3. Select vMedia Policies.

  4. Click Add to create new vMedia Policy.

  5. Name the policy ESXi-6.7U1-HTTP.

  6. Enter Mounts ISO for ESXi 6.7U1 in the Description field.

  7. Select Yes for Retry on Mount failure.

  8. Click Add.

  9. Name the mount ESXi-6.7U1-HTTP.

  10. Select the CDD Device Type.

  11. Select the HTTP Protocol.

  12. Enter the IP Address of the web server.

    The DNS server IPs were not entered into the KVM IP earlier, therefore, it is necessary to enter the IP of the web server instead of the hostname.
  13. Enter VMware-VMvisor-Installer-6.7.0.update01-10302608.x86_64.iso as the Remote File name.

    This VMware ESXi 6.7U1 ISO can be downloaded from VMware Downloads.

  14. Enter the web server path to the ISO file in the Remote Path field.

  15. Click OK to create the vMedia Mount.

  16. Click OK then OK again to complete creating the vMedia Policy.

    For any new servers added to the Cisco UCS environment the vMedia service profile template can be used to install the ESXi host. On first boot, the host boots into the ESXi installer since the SAN mounted disk is empty. After ESXi is installed, the vMedia is not referenced as long as the boot disk is accessible.

    Error: Missing Graphic Image

Create iSCSI boot policy

The procedure in this section applies to a Cisco UCS environment in which two iSCSI logical interfaces (LIFs) are on cluster node 1 (iscsi_lif01a and iscsi_lif01b) and two iSCSI LIFs are on cluster node 2 (iscsi_lif02a and iscsi_lif02b). Also, it is assumed that the A LIFs are connected to Fabric A (Cisco UCS Fabric Interconnect A) and the B LIFs are connected to Fabric B (Cisco UCS Fabric Interconnect B).

One boot policy is configured in this procedure. The policy configures the primary target to be iscsi_lif01a.

To create a boot policy for the Cisco UCS environment, complete the following steps:

  1. In Cisco UCS Manager, click Servers on the left.

  2. Select Policies > root.

  3. Right-click Boot Policies.

  4. Select Create Boot Policy.

  5. Enter Site-01-Fabric-A as the name of the boot policy.

  6. Optional: Enter a description for the boot policy.

  7. Keep the Reboot on Boot Order Change option cleared.

  8. Boot Mode is Legacy.

  9. Expand the Local Devices drop-down menu and select Add Remote CD/DVD.

  10. Expand the iSCSI vNICs drop-down menu and select Add iSCSI Boot.

  11. In the Add iSCSI Boot dialog box, enter Site-01-iSCSI-A. Click OK.

  12. Select Add iSCSI Boot.

  13. In the Add iSCSI Boot dialog box, enter Site-01-iSCSI-B. Click OK.

  14. Click OK to create the policy.

    Error: Missing Graphic Image

Create service profile template

In this procedure, one service profile template for Infrastructure ESXi hosts is created for Fabric A boot.

To create the service profile template, complete the following steps:

  1. In Cisco UCS Manager, click Servers on the left.

  2. Select Service Profile Templates > root.

  3. Right-click root.

  4. Select Create Service Profile Template to open the Create Service Profile Template wizard.

  5. Enter VM-Host-Infra-iSCSI-A as the name of the service profile template. This service profile template is configured to boot from storage node 1 on fabric A.

  6. Select the Updating Template option.

  7. Under UUID, select UUID_Pool as the UUID pool. Click Next.

    Error: Missing Graphic Image

Configure storage provisioning

To configure storage provisioning, complete the following steps:

  1. If you have servers with no physical disks, click Local Disk Configuration Policy and select the SAN Boot Local Storage Policy. Otherwise, select the default Local Storage Policy.

  2. Click Next.

Configure networking options

To configure the networking options, complete the following steps:

  1. Keep the default setting for Dynamic vNIC Connection Policy.

  2. Select the Use Connectivity Policy option to configure the LAN connectivity.

  3. Select iSCSI-Boot from the LAN Connectivity Policy drop-down menu.

  4. Select IQN_Pool in Initiator Name Assignment. Click Next.

    Error: Missing Graphic Image

Configure SAN connectivity

To configure SAN connectivity, complete the following steps:

  1. For the vHBAs, select No for the How Would you Like to Configure SAN Connectivity? option.

  2. Click Next.

Configure zoning

To configure zoning, simply click Next.

Configure vNIC/HBA placement

To configure vNIC/HBA placement, complete the following steps:

  1. From the Select Placement drop-down list, leave the placement policy as Let System Perform Placement.

  2. Click Next.

Configure vMedia policy

To configure the vMedia policy, complete the following steps:

  1. Do not select a vMedia Policy.

  2. Click Next.

Configure server boot order

To configure the server boot order, complete the following steps:

  1. Select Boot-Fabric-A for Boot Policy.

    Error: Missing Graphic Image

  2. In the Boor order, select Site-01- iSCSI-A.

  3. Click Set iSCSI Boot Parameters.

  4. In the Set iSCSI Boot Parameters dialog box, leave the Authentication Profile option to Not Set unless you have independently created one appropriate for your environment.

  5. Leave the Initiator Name Assignment dialog box Not Set to use the single Service Profile Initiator Name defined in the previous steps.

  6. Set iSCSI_IP_Pool_A as the Initiator IP address Policy.

  7. Select iSCSI Static Target Interface option.

  8. Click Add.

  9. Enter the iSCSI target name. To get the iSCSI target name of Infra-SVM, log in into storage cluster management interface and run the iscsi show command.

    Error: Missing Graphic Image

  10. Enter the IP address of iscsi_lif_02a for the IPv4 Address field.

    Error: Missing Graphic Image

  11. Click OK to add the iSCSI static target.

  12. Click Add.

  13. Enter the iSCSI target name.

  14. Enter the IP address of iscsi_lif_01a for the IPv4 Address field.

    Error: Missing Graphic Image

  15. Click OK to add the iSCSI static target.

    Error: Missing Graphic Image

    The target IPs were put in with the storage node 02 IP first and the storage node 01 IP second. This is assuming the boot LUN is on node 01. The host boots by using the path to node 01 if the order in this procedure is used.
  16. In the Boot order, select iSCSI-B-vNIC.

  17. Click Set iSCSI Boot Parameters.

  18. In the Set iSCSI Boot Parameters dialog box, leave the Authentication Profile option as Not Set unless you have independently created one appropriate to your environment.

  19. Leave the Initiator Name Assignment dialog box Not Set to use the single Service Profile Initiator Name defined in the previous steps.

  20. Set iSCSI_IP_Pool_B as the initiator IP address policy.

  21. Select the iSCSI Static Target Interface option.

  22. Click Add.

  23. Enter the iSCSI target name. To get the iSCSI target name of Infra-SVM, log in into storage cluster management interface and run the iscsi show command.

    Error: Missing Graphic Image

  24. Enter the IP address of iscsi_lif_02b for the IPv4 Address field.

    Error: Missing Graphic Image

  25. Click OK to add the iSCSI static target.

  26. Click Add.

  27. Enter the iSCSI target name.

  28. Enter the IP address of iscsi_lif_01b for the IPv4 Address field.

    Error: Missing Graphic Image

  29. Click OK to add the iSCSI static target.

    Error: Missing Graphic Image

  30. Click Next.

Configure maintenance policy

To configure the maintenance policy, complete the following steps:

  1. Change the maintenance policy to default.

    Error: Missing Graphic Image

  2. Click Next.

Configure server assignment

To configure the server assignment, complete the following steps:

  1. In the Pool Assignment list, select Infra-Pool.

  2. Select Down as the power state to be applied when the profile is associated with the server.

  3. Expand Firmware Management at the bottom of the page and select the default policy.

    Error: Missing Graphic Image

  4. Click Next.

Configure operational policies

To configure the operational policies, complete the following steps:

  1. From the BIOS Policy drop-down list, select VM-Host.

  2. Expand Power Control Policy Configuration and select No-Power-Cap from the Power Control Policy drop-down list.

    Error: Missing Graphic Image

  3. Click Finish to create the service profile template.

  4. Click OK in the confirmation message.

Create vMedia-enabled service profile template

To create a service profile template with vMedia enabled, complete the following steps:

  1. Connect to UCS Manager and click Servers on the left.

  2. Select Service Profile Templates > root > Service Template VM-Host-Infra-iSCSI-A.

  3. Right-click VM-Host-Infra-iSCSI-A and select Create a Clone.

  4. Name the clone VM-Host-Infra-iSCSI-A-vM.

  5. Select the newly created VM-Host-Infra-iSCSI-A-vM and select the vMedia Policy tab on the right.

  6. Click Modify vMedia Policy.

  7. Select the ESXi-6. 7U1-HTTP vMedia Policy and click OK.

  8. Click OK to confirm.

Create service profiles

To create service profiles from the service profile template, complete the following steps:

  1. Connect to Cisco UCS Manager and click Servers on the left.

  2. Expand Servers > Service Profile Templates > root > Service Template <name>.

  3. In Actions, click Create Service Profile from Template and compete the following steps:

    1. Enter Site- 01-Infra-0 as the naming prefix.

    2. Enter 2 as the number of instances to create.

    3. Select root as the org.

    4. Click OK to create the service profiles.

      Error: Missing Graphic Image

  4. Click OK in the confirmation message.

  5. Verify that the service profiles Site-01-Infra-01 and Site-01-Infra-02 have been created.

    The service profiles are automatically associated with the servers in their assigned server pools.

Storage configuration part 2: boot LUNs and initiator groups

ONTAP boot storage setup

Create initiator groups

To create initiator groups (igroups), complete the following steps:

  1. Run the following commands from the cluster management node SSH connection:

    igroup create –vserver Infra-SVM –igroup VM-Host-Infra-01 –protocol iscsi –ostype vmware –initiator <vm-host-infra-01-iqn>
    igroup create –vserver Infra-SVM –igroup VM-Host-Infra-02 –protocol iscsi –ostype vmware –initiator <vm-host-infra-02-iqn>
    igroup create –vserver Infra-SVM –igroup MGMT-Hosts –protocol iscsi –ostype vmware –initiator <vm-host-infra-01-iqn>, <vm-host-infra-02-iqn>
    Use the values listed in Table 1 and Table 2 for the IQN information.
  2. To view the three igroups just created, run the igroup show command.

Map boot LUNs to igroups

To map boot LUNs to igroups, complete the following step:

  1. From the storage cluster management SSH connection, run the following commands: 

    lun map –vserver Infra-SVM –volume esxi_boot –lun VM-Host-Infra- A –igroup VM-Host-Infra-01 –lun-id 0lun map –vserver Infra-SVM –volume esxi_boot –lun VM-Host-Infra- B –igroup VM-Host-Infra-02 –lun-id 0

VMware vSphere 6.7U1 deployment procedure

This section provides detailed procedures for installing VMware ESXi 6.7U1 in a FlexPod Express configuration. After the procedures are completed, two booted ESXi hosts are provisioned.

Several methods exist for installing ESXi in a VMware environment. These procedures focus on how to use the built-in KVM console and virtual media features in Cisco UCS Manager to map remote installation media to individual servers and connect to their boot LUNs.

Download Cisco custom image for ESXi 6.7U1

If the VMware ESXi custom image has not been downloaded, complete the following steps to complete the download:

  1. Click the following link: VMware vSphere Hypervisor (ESXi) 6.7U1.

  2. You need a user ID and password on vmware.com to download this software.

  3. Download the .iso file.

Cisco UCS Manager

The Cisco UCS IP KVM enables the administrator to begin the installation of the OS through remote media. It is necessary to log in to the Cisco UCS environment to run the IP KVM.

To log in to the Cisco UCS environment, complete the following steps:

  1. Open a web browser and enter the IP address for the Cisco UCS cluster address. This step launches the Cisco UCS Manager application.

  2. Click the Launch UCS Manager link under HTML to launch the HTML 5 UCS Manager GUI.

  3. If prompted to accept security certificates, accept as necessary.

  4. When prompted, enter admin as the user name and enter the administrative password.

  5. To log in to Cisco UCS Manager, click Login.

  6. From the main menu, click Servers on the left.

  7. Select Servers > Service Profiles > root > VM-Host-Infra-01.

  8. Right-click VM-Host-Infra-01 and select KVM Console.

  9. Follow the prompts to launch the Java-based KVM console.

  10. Select Servers > Service Profiles > root > VM-Host-Infra-02.

  11. Right-click VM-Host-Infra-02. and select KVM Console.

  12. Follow the prompts to launch the Java-based KVM console.

Set up VMware ESXi installation

ESXi Hosts VM-Host-Infra-01 and VM-Host- Infra-02

To prepare the server for the OS installation, complete the following steps on each ESXi host:

  1. In the KVM window, click Virtual Media.

  2. Click Activate Virtual Devices.

  3. If prompted to accept an Unencrypted KVM session, accept as necessary.

  4. Click Virtual Media and select Map CD/DVD.

  5. Browse to the ESXi installer ISO image file and click Open.

  6. Click Map Device. 

  7. Click the KVM tab to monitor the server boot.

Install ESXi

ESXi Hosts VM-Host-Infra-01 and VM-Host-Infra-02

To install VMware ESXi to the iSCSI-bootable LUN of the hosts, complete the following steps on each host:

  1. Boot the server by selecting Boot Server and clicking OK. Then click OK again.

  2. On reboot, the machine detects the presence of the ESXi installation media. Select the ESXi installer from the boot menu that is displayed.

  3. After the installer is finished loading, press Enter to continue with the installation.

  4. Read and accept the end-user license agreement (EULA). Press F11 to accept and continue.

  5. Select the LUN that was previously set up as the installation disk for ESXi and press Enter to continue with the installation.

  6. Select the appropriate keyboard layout and press Enter.

  7. Enter and confirm the root password and press Enter.

  8. The installer issues a warning that the selected disk will be repartitioned. Press F11 to continue with the installation.

  9. After the installation is complete, select the Virtual Media tab and clear the P mark next to the ESXi installation media. Click Yes.

    The ESXi installation image must be unmapped to make sure that the server reboots into ESXi and not into the installer.
  10. After the installation is complete, press Enter to reboot the server.

  11. In Cisco UCS Manager, bind the current service profile to the non-vMedia service profile template to prevent mounting the ESXi installation iso over HTTP.

Set up management networking for ESXi hosts

Adding a management network for each VMware host is necessary for managing the host. To add a management network for the VMware hosts, complete the following steps on each ESXi host:

ESXi Host VM-Host-Infra-01 and VM-Host-Infra-02

To configure each ESXi host with access to the management network, complete the following steps:

  1. After the server has finished rebooting, press F2 to customize the system.

  2. Log in as root, enter the corresponding password, and press Enter to log in.

  3. Select Troubleshooting Options and press Enter.

  4. Select Enable ESXi Shell and press Enter.

  5. Select Enable SSH and press Enter.

  6. Press Esc to exit the Troubleshooting Options menu.

  7. Select the Configure Management Network option and press Enter.

  8. Select Network Adapters and press Enter.

  9. Verify that the numbers in the Hardware Label field match the numbers in the Device Name field.

  10. Press Enter.

    Error: Missing Graphic Image

  11. Select the VLAN (Optional) option and press Enter.

  12. Enter the <ib-mgmt-vlan-id> and press Enter.

  13. Select IPv4 Configuration and press Enter.

  14. Select the Set Static IPv4 Address and Network Configuration option by using the space bar.

  15. Enter the IP address for managing the first ESXi host.

  16. Enter the subnet mask for the first ESXi host.

  17. Enter the default gateway for the first ESXi host.

  18. Press Enter to accept the changes to the IP configuration.

  19. Select the DNS Configuration option and press Enter.

    Because the IP address is assigned manually, the DNS information must also be entered manually.
  20. Enter the IP address of the primary DNS server.

  21. Optional: Enter the IP address of the secondary DNS server.

  22. Enter the FQDN for the first ESXi host.

  23. Press Enter to accept the changes to the DNS configuration.

  24. Press Esc to exit the Configure Management Network menu.

  25. Select Test Management Network to verify that the management network is set up correctly and press Enter.

  26. Press Enter to run the test, press Enter again once the test has completed, review environment if there is a failure.

  27. Select the Configure Management Network again and press Enter.

  28. Select the IPv6 Configuration option and press Enter.

  29. Using the spacebar, select Disable IPv6 (restart required) and press Enter.

  30. Press Esc to exit the Configure Management Network submenu.

  31. Press Y to confirm the changes and reboot the ESXi host.

Reset VMware ESXi host VMkernel port vmk0 MAC address (optional)

ESXi Host VM-Host-Infra-01 and VM-Host-Infra-02

By default, the MAC address of the management VMkernel port vmk0 is the same as the MAC address of the Ethernet port on which it is placed. If the ESXi host’s boot LUN is remapped to a different server with different MAC addresses, a MAC address conflict will occur because vmk0 retains the assigned MAC address unless the ESXi system configuration is reset. To reset the MAC address of vmk0 to a random VMware-assigned MAC address, complete the following steps:

  1. From the ESXi console menu main screen, press Ctrl-Alt-F1 to access the VMware console command line interface. In the UCSM KVM, Ctrl-Alt-F1 appears in the list of static macros.

  2. Log in as root.

  3. Type esxcfg-vmknic –l to get a detailed listing of interface vmk0. vmk0 should be a part of the Management Network port group. Note the IP address and netmask of vmk0.

  4. To remove vmk0, enter the following command:

    esxcfg-vmknic –d “Management Network”
  5. To add vmk0 again with a random MAC address, enter the following command:

    esxcfg-vmknic –a –i <vmk0-ip> -n <vmk0-netmask> “Management Network””.
  6. Verify that vmk0 has been added again with a random MAC address

    esxcfg-vmknic –l
  7. Type exit to log out of the command line interface.

  8. Press Ctrl-Alt-F2 to return to the ESXi console menu interface.

Log into VMware ESXi hosts with VMware host client

ESXi Host VM-Host-Infra-01

To log in to the VM-Host-Infra-01 ESXi host by using the VMware Host Client, complete the following steps:

  1. Open a web browser on the management workstation and navigate to the VM-Host-Infra-01 management IP address.

  2. Click Open the VMware Host Client.

  3. Enter root for the user name.

  4. Enter the root password.

  5. Click Login to connect.

  6. Repeat this process to log in to VM-Host-Infra-02 in a separate browser tab or window.

Install VMware drivers for the Cisco Virtual Interface Card (VIC)

Download and extract the offline bundle for the following VMware VIC driver to the Management workstation:

  • nenic Driver version 1.0.25.0

ESXi hosts VM-Host-Infra-01 and VM-Host-Infra-02

To install VMware VIC Drivers on the ESXi host VM-Host-Infra-01 and VM-Host-Infra-02, complete the following steps:

  1. From each Host Client, select Storage.

  2. Right-click datastore1 and select Browse.

  3. In the Datastore browser, click Upload.

  4. Navigate to the saved location for the downloaded VIC drivers and select VMW-ESX-6.7.0-nenic-1.0.25.0-offline_bundle-11271332.zip.

  5. In the Datastore browser, click Upload.

  6. Click Open to upload the file to datastore1.

  7. Make sure the file has been uploaded to both ESXi hosts.

  8. Place each host into Maintenance mode if it isn’t already.

  9. Connect to each ESXi host through ssh from a shell connection or putty terminal.

  10. Log in as root with the root password.

  11. Run the following commands on each host:

    esxcli software vib update -d /vmfs/volumes/datastore1/VMW-ESX-6.7.0-nenic-1.0.25.0-offline_bundle-11271332.zip
    reboot
  12. Log into the Host Client on each host once reboot is complete and exit Maintenance Mode.

Set up VMkernel ports and virtual switch

ESXi Host VM-Host-Infra-01 and VM-Host-Infra-02

To set up the VMkernel ports and the virtual switches on the ESXi hosts, complete the following steps:

  1. From the Host Client, select Networking on the left.

  2. In the center pane, select the Virtual switches tab.

  3. Select vSwitch0.

  4. Select Edit settings.

  5. Change the MTU to 9000.

  6. Expand NIC teaming.

  7. In the Failover order section, select vmnic1 and click Mark active.

  8. Verify that vmnic1 now has a status of Active.

  9. Click Save.

  10. Select Networking on the left.

  11. In the center pane, select the Virtual switches tab.

  12. Select iScsiBootvSwitch.

  13. Select Edit settings.

  14. Change the MTU to 9000

  15. Click Save.

  16. Select the VMkernel NICs tab.

  17. Select vmk1 iScsiBootPG.

  18. Select Edit settings.

  19. Change the MTU to 9000.

  20. Expand IPv4 settings and change the IP address to an address outside of the UCS iSCSI-IP-Pool-A.

    To avoid IP address conflicts if the Cisco UCS iSCSI IP Pool addresses should get reassigned, it is recommended to use different IP addresses in the same subnet for the iSCSI VMkernel ports.
  21. Click Save.

  22. Select the Virtual switches tab.

  23. Select the Add standard virtual switch.

  24. Provide a name of iScsciBootvSwitch-B for the vSwitch Name.

  25. Set the MTU to 9000.

  26. Select vmnic3 from the Uplink 1 drop-down menu.

  27. Click Add.

  28. In the center pane, select the VMkernel NICs tab.

  29. Select Add VMkernel NIC

  30. Specify a New port group name of iScsiBootPG-B.

  31. Select iScsciBootvSwitch-B for Virtual switch.

  32. Set the MTU to 9000. Do not enter a VLAN ID.

  33. Select Static for the IPv4 settings and expand the option to provide the Address and Subnet Mask within the Configuration.

    To avoid IP address conflicts, if the Cisco UCS iSCSI IP Pool addresses should get reassigned, it is recommended to use different IP addresses in the same subnet for the iSCSI VMkernel ports.
  34. Click Create.

  35. On the left, select Networking, then select the Port groups tab.

  36. In the center pane, right-click VM Network and select Remove.

  37. Click Remove to complete removing the port group.

  38. In the center pane, select Add port group.

  39. Name the port group Management Network and enter <ib-mgmt-vlan-id> in the VLAN ID field, and make sure Virtual switch vSwitch0 is selected.

  40. Click Add to finalize the edits for the IB-MGMT Network.

  41. At the top, select the VMkernel NICs tab.

  42. Click Add VMkernel NIC.

  43. For New port group, enter VMotion.

  44. For Virtual switch, select vSwitch0 selected.

  45. Enter <vmotion-vlan-id> for the VLAN ID.

  46. Change the MTU to 9000.

  47. Select Static IPv4 settings and expand IPv4 settings.

  48. Enter the ESXi host vMotion IP address and netmask.

  49. Select the vMotion stack TCP/IP stack.

  50. Select vMotion under Services.

  51. Click Create.

  52. Click Add VMkernel NIC.

  53. For New port group, enter NFS_Share.

  54. For Virtual switch, select vSwitch0 selected.

  55. Enter <infra-nfs-vlan-id> for the VLAN ID

  56. Change the MTU to 9000.

  57. Select Static IPv4 settings and expand IPv4 settings.

  58. Enter the ESXi host Infrastructure NFS IP address and netmask.

  59. Do not select any of the Services.

  60. Click Create.

  61. Select the Virtual Switches tab, then select vSwitch0. The properties for vSwitch0 VMkernel NICs should be similar to the following example:

    Error: Missing Graphic Image

  62. Select the VMkernel NICs tab to confirm the configured virtual adapters. The adapters listed should be similar to the following example:

    Error: Missing Graphic Image

Setup iSCSI multipathing

ESXi Hosts VM-Host-Infra-01 and VM-Host-Infra-02

To set up the iSCSI multipathing on the ESXi host VM-Host-Infra-01 and VM-Host-Infra-02, complete the following steps:

  1. From each Host Client, select Storage on the left.

  2. In the center pane, click Adapters.

  3. Select the iSCSI software adapter and click Configure iSCSI.

    Error: Missing Graphic Image

  4. Under Dynamic targets, click Add dynamic target.

  5. Enter the IP Address of iSCSI_lif01a.

  6. Repeat entering these IP addresses: iscsi_lif01b, iscsi_lif02a, and iscsi_lif02b.

  7. Click Save Configuration.

    Error: Missing Graphic Image

    To obtain all of the iscsi_lif IP addresses, log in to NetApp storage cluster management interface and run the network interface show command.

    The host automatically rescans the storage adapter and the targets are added to static targets.

Mount required datastores

ESXi Hosts VM-Host-Infra-01 and VM-Host-Infra-02

To mount the required datastores, complete the following steps on each ESXi host:

  1. From the Host Client, select Storage on the left.

  2. In the center pane, select Datastores.

  3. In the center pane, select New Datastore to add a new datastore.

  4. In the New datastore dialog box, select Mount NFS datastore and click Next.

    Error: Missing Graphic Image

  5. On the provide NFS Mount Details page, complete these steps:

    1. Enter infra_datastore_1 for the datastore name.

    2. Enter the IP address for the nfs_lif01_a LIF for the NFS server.

    3. Enter /infra_datastore_1 for the NFS share.

    4. Leave the NFS version set at NFS 3.

    5. Click Next.

      Error: Missing Graphic Image

  6. Click Finish. The datastore should now appear in the datastore list.

  7. In the center pane, select New Datastore to add a new datastore.

  8. In the New Datastore dialog box, select Mount NFS Datastore and click Next.

  9. On the provide NFS Mount Details page, complete these steps:

    1. Enter infra_datastore_2 for the datastore name.

    2. Enter the IP address for the nfs_lif02_a LIF for the NFS server.

    3. Enter /infra_datastore_2 for the NFS share.

    4. Leave the NFS version set at NFS 3.

    5. Click Next.

  10. Click Finish. The datastore should now appear in the datastore list.

    Error: Missing Graphic Image

  11. Mount both datastores on both ESXi hosts.

Configure NTP on ESXi hosts

ESXi Hosts VM-Host-Infra-01 and VM-Host-Infra-02

To configure NTP on the ESXi hosts, complete the following steps on each host:

  1. From the Host Client, select Manage on the left.

  2. In the center pane, select the Time & Date tab.

  3. Click Edit Settings.

  4. Make sure Use Network Time Protocol (enable NTP client) is selected.

  5. Use the drop-down menu to select Start and Stop with Host.

  6. Enter the two Nexus switch NTP addresses in the NTP servers box separated by a comma.

    Error: Missing Graphic Image

  7. Click Save to save the configuration changes.

  8. Select Actions > NTP service > Start.

  9. Verify that NTP service is now running and the clock is now set to approximately the correct time

    The NTP server time might vary slightly from the host time.

Configure ESXi host swap

ESXi Hosts VM-Host-Infra-01 and VM-Host-Infra-02

To configure host swap on the ESXi hosts, follow these steps on each host:

  1. Click Manage in the left navigation pane. Select System in the right pane and click Swap.

    Error: Missing Graphic Image

  2. Click Edit Settings. Select infra_swap from the Datastore options.

    Error: Missing Graphic Image

  3. Click Save.

Install the NetApp NFS Plug-in 1.1.2 for VMware VAAI

To install the NetApp NFS Plug-in 1. 1.2 for VMware VAAI, complete the following steps.

  1. Download the NetApp NFS Plug-in for VMware VAAI:

    1. Go to the NetApp software download page.

    2. Scroll down and click NetApp NFS Plug-in for VMware VAAI.

    3. Select the ESXi platform.

    4. Download either the offline bundle (.zip) or online bundle (.vib) of the most recent plug-in.

  2. The NetApp NFS plug-in for VMware VAAI is pending IMT qualification with ONTAP 9.5 and interoperability details will be posted to the NetApp IMT soon.

  3. Install the plug-in on the ESXi host by using the ESX CLI.

  4. Reboot the ESXI host.

Install VMware vCenter Server 6.7

This section provides detailed procedures for installing VMware vCenter Server 6.7 in a FlexPod Express configuration.

FlexPod Express uses the VMware vCenter Server Appliance (VCSA).

Install VMware vCenter server appliance

To install VCSA, complete the following steps:

  1. Download the VCSA. Access the download link by clicking the Get vCenter Server icon when managing the ESXi host.

    Error: Missing Graphic Image

  2. Download the VCSA from the VMware site.

    Although the Microsoft Windows vCenter Server installable is supported, VMware recommends the VCSA for new deployments.
  3. Mount the ISO image.

  4. Navigate to the vcsa-ui-installer > win32 directory. Double-click installer.exe.

  5. Click Install.

  6. Click Next on the Introduction page.

  7. Accept the EULA.

  8. Select Embedded Platform Services Controller as the deployment type.

    Error: Missing Graphic Image

    If required, the External Platform Services Controller deployment is also supported as part of the FlexPod Express solution.

  9. On the Appliance Deployment Target page, enter the IP address of an ESXi host you have deployed, the root user name, and the root password. Click Next.

    Error: Missing Graphic Image

  10. Set the appliance VM by entering VCSA as the VM name and the root password you would like to use for the VCSA. Click Next.

    Error: Missing Graphic Image

  11. Select the deployment size that best fits your environment. Click Next.

    Error: Missing Graphic Image

  12. Select the infra_datastore_1 datastore. Click Next.

    Error: Missing Graphic Image

  13. Enter the following information on the Configure Network Settings page and click Next.

    1. Select MGMT-Network as your network.

    2. Enter the FQDN or IP to be used for the VCSA.

    3. Enter the IP address to be used.

    4. Enter the subnet mask to be used.

    5. Enter the default gateway.

    6. Enter the DNS server.

      Error: Missing Graphic Image

  14. On the Ready to Complete Stage 1 page, verify that the settings you have entered are correct. Click Finish.

    The VCSA installs now. This process takes several minutes.

  15. After stage 1 completes, a message appears stating that it has completed. Click Continue to begin stage 2 configuration.

    Error: Missing Graphic Image

  16. On the Stage 2 Introduction page, click Next.

  17. Enter <<var_ntp_id>> for the NTP server address. You can enter multiple NTP IP addresses.

    If you plan to use vCenter Server high availability, make sure that SSH access is enabled.

  18. Configure the SSO domain name, password, and site name. Click Next.

    Record these values for your reference, especially if you deviate from the vsphere.local domain name.

  19. Join the VMware Customer Experience Program if desired. Click Next.

  20. View the summary of your settings. Click Finish or use the back button to edit settings.

  21. A message appears stating that you are not able to pause or stop the installation from completing after it has started. Click OK to continue.

    The appliance setup continues. This takes several minutes.

    A message appears indicating that the setup was successful.

    The links that the installer provides to access vCenter Server are clickable.

Configure VMware vCenter Server 6.7 and vSphere clustering

To configure VMware vCenter Server 6.7 and vSphere clustering, complete the following steps:

  1. Navigate to https://<<FQDN or IP of vCenter>>/vsphere-client/.

  2. Click Launch vSphere Client.

  3. Log in with the user name administrator@vsphere.local and the SSO password you entered during the VCSA setup process.

  4. Right-click the vCenter name and select New Datacenter.

  5. Enter a name for the data center and click OK.

Create vSphere Cluster.

To create a vSphere cluster, complete the following steps:

  1. Right-click the newly created data center and select New Cluster.

  2. Enter a name for the cluster.

  3. Select and enable DRS and vSphere HA options.

  4. Click OK.

    Error: Missing Graphic Image

Add ESXi Hosts to Cluster

To add ESXi hosts to the cluster, complete the following steps:

  1. Select Add Host in the Actions menu of the cluster.

    Error: Missing Graphic Image

  2. To add an ESXi host to the cluster, complete the following steps:

    1. Enter the IP or FQDN of the host. Click Next.

    2. Enter the root user name and password. Click Next.

    3. Click Yes to replace the host’s certificate with a certificate signed by the VMware certificate server.

    4. Click Next on the Host Summary page.

    5. Click the green + icon to add a license to the vSphere host.

      This step can be completed later if desired.
    6. Click Next to leave lockdown mode disabled.

    7. Click Next at the VM location page.

    8. Review the Ready to Complete page. Use the back button to make any changes or select Finish.

  3. Repeat steps 1 and 2 for Cisco UCS host B.

    This process must be completed for any additional hosts added to the FlexPod Express configuration.

Configure coredump on ESXi hosts

ESXi Dump Collector Setup for iSCSI-Booted Hosts

ESXi hosts booted with iSCSI using the VMware iSCSI software initiator need to be configured to do core dumps to the ESXi Dump Collector that is part of vCenter. The Dump Collector is not enabled by default on the vCenter Appliance. This procedure should be run at the end of the vCenter deployment section. To setup the ESXi Dump Collector, follow these steps:

  1. Log in to the vSphere Web Client as administrator@vsphere.local and select Home.

  2. In the center pane, click System Configuration.

  3. In the left pane, select Services.

  4. Under Services, click VMware vSphere ESXi Dump Collector.

  5. In the center pane, click the green start icon to start the service.

  6. In the Actions menu, click Edit Startup Type.

  7. Select Automatic.

  8. Click OK.

  9. Connect to each ESXi host using ssh as root.

  10. Run the following commands:

    esxcli system coredump network set –v vmk0 –j <vcenter-ip>
    esxcli system coredump network set –e true
    esxcli system coredump network check

    The message Verified the configured netdump server is running appears after you run the final command.

    This process must be completed for any additional hosts added to FlexPod Express.