本繁體中文版使用機器翻譯,譯文僅供參考,若與英文版本牴觸,應以英文版本為準。

將四節點MetroCluster 的靜態IP組態擴充至八節點組態

貢獻者

從更新版的支援版9.9.1開始ONTAP 、您可以將四個新節點新增到MetroCluster 作為第二個DR群組的支援版IP組態中。這會建立八節點MetroCluster 的不全功能組態。

開始之前
  • 舊節點和新節點必須執行相同版本ONTAP 的更新。

  • 您必須確保舊有和新的平台機型支援平台混合。

  • 您必須確保IP交換器同時支援新舊的平台機型。

  • 新節點必須有足夠的儲存空間來容納舊節點的資料、以及足夠的磁碟來容納根集合體和備用磁碟。

  • FAS2750或AFF FASA220系統不支援此程序。

此程序中的命名範例

此程序使用範例名稱來識別整個過程中涉及的DR群組、節點和交換器。

DR群組

叢集_A位於站台_A

站台B的叢集B

Dr_group_1-old

  • node_a_1-old

  • node_a_2-old

  • node_B_1-old

  • node_B_2-old

Dr_group_2-new

  • node_a_3-new

  • node_a_4-new

  • node_B_3-new

  • node_B_4-new

在維護前傳送自訂AutoSupport 的支援訊息

在執行維護之前、您應發出AutoSupport 一份不知所知的訊息、通知NetApp技術支援部門正在進行維護。告知技術支援部門正在進行維護、可防止他們假設發生中斷、而開啟案例。

此工作必須在每MetroCluster 個站台上執行。

步驟
  1. 為了避免自動產生支援案例、請傳送AutoSupport 一個不全的訊息、指出正在進行升級。

    1. 發出下列命令:

      「系統節點AutoSupport 不完整地叫用節點*-type all -most "MAn=10h正在將_Oold-modem_升級為_new模型」_」

      此範例指定10小時的維護時間。您可能需要額外的時間、視您的計畫而定。

      如果維護作業在時間結束之前完成、您可以叫用AutoSupport 指示維護期間結束的資訊消息:

    「系統節點AutoSupport 不完整地叫用節點*-type all -most MAn=end」

    1. 在合作夥伴叢集上重複執行命令。

驗MetroCluster 證資訊功能組態的健全狀況

在執行轉換之前、您必須先驗證MetroCluster 整個流程的健全狀況和連線能力

步驟
  1. 驗證MetroCluster 下列項目中的功能:ONTAP

    1. 檢查系統是否具有多路徑:

      「節點執行節點_norme-name_ sysconfig -A」

    2. 檢查兩個叢集上的任何健全狀況警示:

      「系統健全狀況警示顯示」

    3. 確認MetroCluster 執行功能組態、並確認操作模式正常:

      《不看》MetroCluster

    4. 執行功能檢查:MetroCluster

      《不一樣的跑程》MetroCluster

    5. 顯示MetroCluster 檢查結果:

      《不一樣的表演》MetroCluster

    6. 執行Config Advisor

    7. 執行Config Advisor 完功能後、請檢閱工具的輸出結果、並依照輸出中的建議來解決發現的任何問題。

  2. 驗證叢集是否正常:

    「叢集show -vserver叢集」

    cluster_A::> cluster show -vserver Cluster
    Node           Health  Eligibility   Epsilon
    -------------- ------  -----------   -------
    node_A_1    true    true          false
    node_A_2    true    true          false
    
    cluster_A::>
  3. 驗證所有叢集連接埠是否正常運作:

    「網路連接埠show -IPSpace叢集」

    cluster_A::> network port show -ipspace cluster
    
    Node: node_A_1-old
    
                                                      Speed(Mbps) Health
    Port      IPspace      Broadcast Domain Link MTU  Admin/Oper  Status
    --------- ------------ ---------------- ---- ---- ----------- --------
    e0a       Cluster      Cluster          up   9000  auto/10000 healthy
    e0b       Cluster      Cluster          up   9000  auto/10000 healthy
    
    Node: node_A_2-old
    
                                                      Speed(Mbps) Health
    Port      IPspace      Broadcast Domain Link MTU  Admin/Oper  Status
    --------- ------------ ---------------- ---- ---- ----------- --------
    e0a       Cluster      Cluster          up   9000  auto/10000 healthy
    e0b       Cluster      Cluster          up   9000  auto/10000 healthy
    
    4 entries were displayed.
    
    cluster_A::>
  4. 驗證所有叢集生命體是否正常運作:

    「網路介面show -vserver叢集」

    每個叢集LIF都應顯示為「真實」、並具有「狀態管理/作業者」的「正常」狀態

    cluster_A::> network interface show -vserver cluster
    
                Logical      Status     Network          Current       Current Is
    Vserver     Interface  Admin/Oper Address/Mask       Node          Port    Home
    ----------- ---------- ---------- ------------------ ------------- ------- -----
    Cluster
                node_A_1-old_clus1
                           up/up      169.254.209.69/16  node_A_1   e0a     true
                node_A_1-old_clus2
                           up/up      169.254.49.125/16  node_A_1   e0b     true
                node_A_2-old_clus1
                           up/up      169.254.47.194/16  node_A_2   e0a     true
                node_A_2-old_clus2
                           up/up      169.254.19.183/16  node_A_2   e0b     true
    
    4 entries were displayed.
    
    cluster_A::>
  5. 驗證是否已在所有叢集生命體上啟用自動還原:

    「網路介面show -vserver叢集-功能 變數自動回復」

    cluster_A::> network interface show -vserver Cluster -fields auto-revert
    
              Logical
    Vserver   Interface     Auto-revert
    --------- ------------- ------------
    Cluster
               node_A_1-old_clus1
                            true
               node_A_1-old_clus2
                            true
               node_A_2-old_clus1
                            true
               node_A_2-old_clus2
                            true
    
        4 entries were displayed.
    
    cluster_A::>

從監控應用程式移除組態

如果現有的組態是使用MetroCluster 可ONTAP 啟動切換的ESITIOR軟體、或者其他協力廠商應用程式(例如ClusterLion)來監控、則MetroCluster 在升級之前、您必須先從監控軟體中移除該組態。

步驟
  1. 移除MetroCluster 可啟動切換功能的斷路器、調解器或其他軟體中現有的物件組態。

    如果您使用…​

    使用此程序…​

    Tiebreaker

    中介者

    從資訊提示字元發出下列命令ONTAP :

    「取消組態設定中介程式」MetroCluster

    第三方應用程式

    請參閱產品文件。

  2. 從MetroCluster 任何可啟動切換的第三方應用程式移除現有的功能。

    請參閱應用程式的文件。

準備新的控制器模組

您必須準備好四MetroCluster 個新的資訊節點、並安裝正確ONTAP 的版本資訊。

此工作必須在每個新節點上執行:

  • node_a_3-new

  • node_a_4-new

  • node_B_3-new

  • node_B_4-new

在這些步驟中、您可以清除節點上的組態、並清除新磁碟機上的信箱區域。

步驟
  1. 安裝新的控制器。

  2. 將新MetroCluster 的靜態IP節點連接至IP交換器、如_Ef2 MetroCluster 安裝與組態所示

  3. 請MetroCluster 使用下列_Estrustri MetroCluster 安裝與組態區段來設定SURFIP節點

  4. 在維護模式下、發出halt命令以結束維護模式、然後發出boot_ONTAP命令以啟動系統並進入叢集設定。

    此時請勿完成叢集精靈或節點精靈。

將新節點加入叢集

您必須將四MetroCluster 個全新的靜態IP節點新增至現有MetroCluster 的靜態組態。

您必須在兩個叢集上執行此工作。

步驟
  1. 將新MetroCluster 的靜態IP節點新增至現有MetroCluster 的靜態組態。

    1. 將第一個全新MetroCluster 的「支援IP」節點(node_a_1-new)加入現有MetroCluster 的「支援IP」組態。

      Welcome to the cluster setup wizard.
      
      You can enter the following commands at any time:
        "help" or "?" - if you want to have a question clarified,
        "back" - if you want to change previously answered questions, and
        "exit" or "quit" - if you want to quit the cluster setup wizard.
           Any changes you made before quitting will be saved.
      
      You can return to cluster setup at any time by typing "cluster setup".
      To accept a default or omit a question, do not enter a value.
      
      This system will send event messages and periodic reports to NetApp Technical
      Support. To disable this feature, enter
      autosupport modify -support disable
      within 24 hours.
      
      Enabling AutoSupport can significantly speed problem determination and
      resolution, should a problem occur on your system.
      For further information on AutoSupport, see:
      http://support.netapp.com/autosupport/
      
      Type yes to confirm and continue {yes}: yes
      
      Enter the node management interface port [e0M]: 172.17.8.93
      
      172.17.8.93 is not a valid port.
      
      The physical port that is connected to the node management network. Examples of
      node management ports are "e4a" or "e0M".
      
      You can type "back", "exit", or "help" at any question.
      
      
      Enter the node management interface port [e0M]:
      Enter the node management interface IP address: 172.17.8.93
      Enter the node management interface netmask: 255.255.254.0
      Enter the node management interface default gateway: 172.17.8.1
      A node management interface on port e0M with IP address 172.17.8.93 has been created.
      
      Use your web browser to complete cluster setup by accessing https://172.17.8.93
      
      Otherwise, press Enter to complete cluster setup using the command line
      interface:
      
      
      Do you want to create a new cluster or join an existing cluster? {create, join}:
      join
      
      
      Existing cluster interface configuration found:
      
      Port    MTU     IP              Netmask
      e0c     9000    169.254.148.217 255.255.0.0
      e0d     9000    169.254.144.238 255.255.0.0
      
      Do you want to use this configuration? {yes, no} [yes]: yes
      .
      .
      .
    2. 將第二MetroCluster 個全新的靜態IP節點(node_a_2-new)加入現有MetroCluster 的靜態IP組態。

  2. 重複這些步驟、將node_B_1-new和node_B_2-new加入叢集B

設定叢集間的生命體、建立MetroCluster 支援資訊介面、以及鏡射根集合體

您必須建立叢集對等的生命、在MetroCluster 新MetroCluster 的物件節點上建立一個物件介面。

範例中使用的主連接埠是平台專屬的連接埠。您應該使用MetroCluster 特定於靜態IP節點平台的適當主連接埠。

步驟
  1. 在新MetroCluster 的SURE IP節點上、使用下列程序來設定叢集間LIF:

  2. 在每個站台上、確認已設定叢集對等:

    「叢集同儕秀」

    下列範例顯示叢集A上的叢集對等配置:

    cluster_A:> cluster peer show
    Peer Cluster Name         Cluster Serial Number Availability   Authentication
    ------------------------- --------------------- -------------- --------------
    cluster_B                 1-80-000011           Available      ok

    下列範例顯示叢集B上的叢集對等配置:

    cluster_B:> cluster peer show
    Peer Cluster Name         Cluster Serial Number Availability   Authentication
    ------------------------- --------------------- -------------- --------------
    cluster_A                 1-80-000011           Available      ok
    cluster_B::>
  3. 建立MetroCluster 適用於整個知識產權節點的DR群組:

    「組態設定DR-group create -合作 夥伴叢集」MetroCluster

    如需MetroCluster 有關「還原組態」設定和連線的詳細資訊、請參閱下列內容:

    cluster_A::> metrocluster configuration-settings dr-group create -partner-cluster
    cluster_B -local-node node_A_1-new -remote-node node_B_1-new
    [Job 259] Job succeeded: DR Group Create is successful.
    cluster_A::>
  4. 確認已建立DR群組。

    「組態設定DR群組顯示」MetroCluster

    cluster_A::> metrocluster configuration-settings dr-group show
    
    DR Group ID Cluster                    Node               DR Partner Node
    ----------- -------------------------- ------------------ ------------------
    1           cluster_A
                                           node_A_1-old        node_B_1-old
                                           node_A_2-old        node_B_2-old
                cluster_B
                                           node_B_1-old        node_A_1-old
                                           node_B_2-old        node_A_2-old
    2           cluster_A
                                           node_A_1-new        node_B_1-new
                                           node_A_2-new        node_B_2-new
                cluster_B
                                           node_B_1-new        node_A_1-new
                                           node_B_2-new        node_A_2-new
    8 entries were displayed.
    
    cluster_A::>
  5. 為MetroCluster 新加入MetroCluster 的「靜態IP」節點設定「靜態IP」介面:

    「功能組態設定介面create -cluster名稱」MetroCluster

    附註
    • 某些平台會使用VLAN作為MetroCluster 支援靜態IP介面。根據預設、這兩個連接埠各自使用不同的VLAN:10和20。您也可以使用「MetroCluster 還原組態設定介面create」命令中的「-vlan-id參數」、指定高於100(介於101和4095之間)的不同(非預設)VLAN。

    • 從ONTAP 《支援範圍》9.9.1開始、如果您使用的是第3層組態、則在建立MetroCluster 支援範圍介面時、也必須指定「閘道」參數。請參閱 "第3層廣域網路的考量"

    如果使用的VLAN是10/20或大於100、則可將下列平台機型新增至現有MetroCluster 的支援功能組態。如果使用任何其他VLAN、則MetroCluster 無法將這些平台新增至現有組態、因為無法設定此介面。如果您使用任何其他平台、則VLAN組態並不相關、ONTAP 因為在介紹中並不需要。

    支援各種平台AFF

    支援各種平台FAS

    • VA220 AFF

    • VA250 AFF

    • 解答400 AFF

    • FAS2750

    • FAS500f

    • FAS8300

    • FAS8700

    附註 您可以從MetroCluster 任一叢集設定靜態IP介面。此外、從ONTAP 版本號為32的9.1.1開始、如果您使用的是第3層組態、您也必須指定「-gateway’參數來建立MetroCluster 一個靜態IP介面。請參閱 "第3層廣域網路的考量"
    cluster_A::> metrocluster configuration-settings interface create -cluster-name cluster_A -home-node node_A_1-new -home-port e1a -address 172.17.26.10 -netmask 255.255.255.0
    [Job 260] Job succeeded: Interface Create is successful.
    
    cluster_A::> metrocluster configuration-settings interface create -cluster-name cluster_A -home-node node_A_1-new -home-port e1b -address 172.17.27.10 -netmask 255.255.255.0
    [Job 261] Job succeeded: Interface Create is successful.
    
    cluster_A::> metrocluster configuration-settings interface create -cluster-name cluster_A -home-node node_A_2-new -home-port e1a -address 172.17.26.11 -netmask 255.255.255.0
    [Job 262] Job succeeded: Interface Create is successful.
    
    cluster_A::> :metrocluster configuration-settings interface create -cluster-name cluster_A -home-node node_A_2-new -home-port e1b -address 172.17.27.11 -netmask 255.255.255.0
    [Job 263] Job succeeded: Interface Create is successful.
    
    cluster_A::> metrocluster configuration-settings interface create -cluster-name cluster_B -home-node node_B_1-new -home-port e1a -address 172.17.26.12 -netmask 255.255.255.0
    [Job 264] Job succeeded: Interface Create is successful.
    
    cluster_A::> metrocluster configuration-settings interface create -cluster-name cluster_B -home-node node_B_1-new -home-port e1b -address 172.17.27.12 -netmask 255.255.255.0
    [Job 265] Job succeeded: Interface Create is successful.
    
    cluster_A::> metrocluster configuration-settings interface create -cluster-name cluster_B -home-node node_B_2-new -home-port e1a -address 172.17.26.13 -netmask 255.255.255.0
    [Job 266] Job succeeded: Interface Create is successful.
    
    cluster_A::> metrocluster configuration-settings interface create -cluster-name cluster_B -home-node node_B_2-new -home-port e1b -address 172.17.27.13 -netmask 255.255.255.0
    [Job 267] Job succeeded: Interface Create is successful.
  6. 驗證MetroCluster 是否已建立下列的靜態IP介面:

    「顯示組態設定介面」MetroCluster

    cluster_A::>metrocluster configuration-settings interface show
    
    DR                                                                    Config
    Group Cluster Node    Network Address Netmask         Gateway         State
    ----- ------- ------- --------------- --------------- --------------- ---------
    1     cluster_A
                 node_A_1-old
                     Home Port: e1a
                          172.17.26.10    255.255.255.0   -               completed
                     Home Port: e1b
                          172.17.27.10    255.255.255.0   -               completed
                  node_A_2-old
                     Home Port: e1a
                          172.17.26.11    255.255.255.0   -               completed
                     Home Port: e1b
                          172.17.27.11    255.255.255.0   -               completed
          cluster_B
                 node_B_1-old
                     Home Port: e1a
                          172.17.26.13    255.255.255.0   -               completed
                     Home Port: e1b
                          172.17.27.13    255.255.255.0   -               completed
                  node_B_1-old
                     Home Port: e1a
                          172.17.26.12    255.255.255.0   -               completed
                     Home Port: e1b
                          172.17.27.12    255.255.255.0   -               completed
    2     cluster_A
                 node_A_3-new
                     Home Port: e1a
                          172.17.28.10    255.255.255.0   -               completed
                     Home Port: e1b
                          172.17.29.10    255.255.255.0   -               completed
                  node_A_3-new
                     Home Port: e1a
                          172.17.28.11    255.255.255.0   -               completed
                     Home Port: e1b
                          172.17.29.11    255.255.255.0   -               completed
          cluster_B
                 node_B_3-new
                     Home Port: e1a
                          172.17.28.13    255.255.255.0   -               completed
                     Home Port: e1b
                          172.17.29.13    255.255.255.0   -               completed
                  node_B_3-new
                     Home Port: e1a
                          172.17.28.12    255.255.255.0   -               completed
                     Home Port: e1b
                          172.17.29.12    255.255.255.0   -               completed
    8 entries were displayed.
    
    cluster_A>
  7. 連接MetroCluster 下列的靜態IP介面:

    「組態設定連線」MetroCluster

    附註 此命令可能需要數分鐘才能完成。
    cluster_A::> metrocluster configuration-settings connection connect
    
    cluster_A::>
  8. 驗證連接是否正確建立:MetroCluster 「不實組態設定連線顯示」

    cluster_A::> metrocluster configuration-settings connection show
    
    DR                    Source          Destination
    Group Cluster Node    Network Address Network Address Partner Type Config State
    ----- ------- ------- --------------- --------------- ------------ ------------
    1     cluster_A
                  node_A_1-old
                     Home Port: e1a
                          172.17.28.10    172.17.28.11    HA Partner   completed
                     Home Port: e1a
                          172.17.28.10    172.17.28.12    DR Partner   completed
                     Home Port: e1a
                          172.17.28.10    172.17.28.13    DR Auxiliary completed
                     Home Port: e1b
                          172.17.29.10    172.17.29.11    HA Partner   completed
                     Home Port: e1b
                          172.17.29.10    172.17.29.12    DR Partner   completed
                     Home Port: e1b
                          172.17.29.10    172.17.29.13    DR Auxiliary completed
                  node_A_2-old
                     Home Port: e1a
                          172.17.28.11    172.17.28.10    HA Partner   completed
                     Home Port: e1a
                          172.17.28.11    172.17.28.13    DR Partner   completed
                     Home Port: e1a
                          172.17.28.11    172.17.28.12    DR Auxiliary completed
                     Home Port: e1b
                          172.17.29.11    172.17.29.10    HA Partner   completed
                     Home Port: e1b
                          172.17.29.11    172.17.29.13    DR Partner   completed
                     Home Port: e1b
                          172.17.29.11    172.17.29.12    DR Auxiliary completed
    
    DR                    Source          Destination
    Group Cluster Node    Network Address Network Address Partner Type Config State
    ----- ------- ------- --------------- --------------- ------------ ------------
    1     cluster_B
                  node_B_2-old
                     Home Port: e1a
                          172.17.28.13    172.17.28.12    HA Partner   completed
                     Home Port: e1a
                          172.17.28.13    172.17.28.11    DR Partner   completed
                     Home Port: e1a
                          172.17.28.13    172.17.28.10    DR Auxiliary completed
                     Home Port: e1b
                          172.17.29.13    172.17.29.12    HA Partner   completed
                     Home Port: e1b
                          172.17.29.13    172.17.29.11    DR Partner   completed
                     Home Port: e1b
                          172.17.29.13    172.17.29.10    DR Auxiliary completed
                  node_B_1-old
                     Home Port: e1a
                          172.17.28.12    172.17.28.13    HA Partner   completed
                     Home Port: e1a
                          172.17.28.12    172.17.28.10    DR Partner   completed
                     Home Port: e1a
                          172.17.28.12    172.17.28.11    DR Auxiliary completed
                     Home Port: e1b
                          172.17.29.12    172.17.29.13    HA Partner   completed
                     Home Port: e1b
                          172.17.29.12    172.17.29.10    DR Partner   completed
                     Home Port: e1b
                          172.17.29.12    172.17.29.11    DR Auxiliary completed
    
    DR                    Source          Destination
    Group Cluster Node    Network Address Network Address Partner Type Config State
    ----- ------- ------- --------------- --------------- ------------ ------------
    2     cluster_A
                  node_A_1-new**
                     Home Port: e1a
                          172.17.26.10    172.17.26.11    HA Partner   completed
                     Home Port: e1a
                          172.17.26.10    172.17.26.12    DR Partner   completed
                     Home Port: e1a
                          172.17.26.10    172.17.26.13    DR Auxiliary completed
                     Home Port: e1b
                          172.17.27.10    172.17.27.11    HA Partner   completed
                     Home Port: e1b
                          172.17.27.10    172.17.27.12    DR Partner   completed
                     Home Port: e1b
                          172.17.27.10    172.17.27.13    DR Auxiliary completed
                  node_A_2-new
                     Home Port: e1a
                          172.17.26.11    172.17.26.10    HA Partner   completed
                     Home Port: e1a
                          172.17.26.11    172.17.26.13    DR Partner   completed
                     Home Port: e1a
                          172.17.26.11    172.17.26.12    DR Auxiliary completed
                     Home Port: e1b
                          172.17.27.11    172.17.27.10    HA Partner   completed
                     Home Port: e1b
                          172.17.27.11    172.17.27.13    DR Partner   completed
                     Home Port: e1b
                          172.17.27.11    172.17.27.12    DR Auxiliary completed
    
    DR                    Source          Destination
    Group Cluster Node    Network Address Network Address Partner Type Config State
    ----- ------- ------- --------------- --------------- ------------ ------------
    2     cluster_B
                  node_B_2-new
                     Home Port: e1a
                          172.17.26.13    172.17.26.12    HA Partner   completed
                     Home Port: e1a
                          172.17.26.13    172.17.26.11    DR Partner   completed
                     Home Port: e1a
                          172.17.26.13    172.17.26.10    DR Auxiliary completed
                     Home Port: e1b
                          172.17.27.13    172.17.27.12    HA Partner   completed
                     Home Port: e1b
                          172.17.27.13    172.17.27.11    DR Partner   completed
                     Home Port: e1b
                          172.17.27.13    172.17.27.10    DR Auxiliary completed
                  node_B_1-new
                     Home Port: e1a
                          172.17.26.12    172.17.26.13    HA Partner   completed
                     Home Port: e1a
                          172.17.26.12    172.17.26.10    DR Partner   completed
                     Home Port: e1a
                          172.17.26.12    172.17.26.11    DR Auxiliary completed
                     Home Port: e1b
                          172.17.27.12    172.17.27.13    HA Partner   completed
                     Home Port: e1b
                          172.17.27.12    172.17.27.10    DR Partner   completed
                     Home Port: e1b
                          172.17.27.12    172.17.27.11    DR Auxiliary completed
    48 entries were displayed.
    
    cluster_A::>
  9. 驗證磁碟自動指派與分割:

    「展示池Pool1」

    cluster_A::> disk show -pool Pool1
                         Usable           Disk    Container   Container
    Disk                   Size Shelf Bay Type    Type        Name      Owner
    ---------------- ---------- ----- --- ------- ----------- --------- --------
    1.10.4                    -    10   4 SAS     remote      -         node_B_2
    1.10.13                   -    10  13 SAS     remote      -         node_B_2
    1.10.14                   -    10  14 SAS     remote      -         node_B_1
    1.10.15                   -    10  15 SAS     remote      -         node_B_1
    1.10.16                   -    10  16 SAS     remote      -         node_B_1
    1.10.18                   -    10  18 SAS     remote      -         node_B_2
    ...
    2.20.0              546.9GB    20   0 SAS     aggregate   aggr0_rha1_a1 node_a_1
    2.20.3              546.9GB    20   3 SAS     aggregate   aggr0_rha1_a2 node_a_2
    2.20.5              546.9GB    20   5 SAS     aggregate   rha1_a1_aggr1 node_a_1
    2.20.6              546.9GB    20   6 SAS     aggregate   rha1_a1_aggr1 node_a_1
    2.20.7              546.9GB    20   7 SAS     aggregate   rha1_a2_aggr1 node_a_2
    2.20.10             546.9GB    20  10 SAS     aggregate   rha1_a1_aggr1 node_a_1
    ...
    43 entries were displayed.
    
    cluster_A::>
  10. 鏡射根Aggregate:

    「torage Aggregate mirror -Aggregate aggr0_node_a_1-new'」

    附註 您必須在每MetroCluster 個環節上完成此步驟。
    cluster_A::> aggr mirror -aggregate aggr0_node_A_1-new
    
    Info: Disks would be added to aggregate "aggr0_node_A_1-new"on node "node_A_1-new"
          in the following manner:
    
          Second Plex
    
            RAID Group rg0, 3 disks (block checksum, raid_dp)
                                                                Usable Physical
              Position   Disk                      Type           Size     Size
              ---------- ------------------------- ---------- -------- --------
              dparity    4.20.0                    SAS               -        -
              parity     4.20.3                    SAS               -        -
              data       4.20.1                    SAS         546.9GB  558.9GB
    
          Aggregate capacity available forvolume use would be 467.6GB.
    
    Do you want to continue? {y|n}: y
    
    cluster_A::>
  11. 驗證根集合體是否為鏡射:

    《集合體展》

    cluster_A::> aggr show
    
    Aggregate     Size Available Used% State   #Vols  Nodes            RAID Status
    --------- -------- --------- ----- ------- ------ ---------------- ------------
    aggr0_node_A_1-old
               349.0GB   16.84GB   95% online       1 node_A_1-old      raid_dp,
                                                                       mirrored,
                                                                       normal
    aggr0_node_A_2-old
               349.0GB   16.84GB   95% online       1 node_A_2-old      raid_dp,
                                                                       mirrored,
                                                                       normal
    aggr0_node_A_1-new
               467.6GB   22.63GB   95% online       1 node_A_1-new      raid_dp,
                                                                       mirrored,
                                                                       normal
    aggr0_node_A_2-new
               467.6GB   22.62GB   95% online       1 node_A_2-new      raid_dp,
                                                                       mirrored,
                                                                       normal
    aggr_data_a1
                1.02TB    1.01TB    1% online       1 node_A_1-old      raid_dp,
                                                                       mirrored,
                                                                       normal
    aggr_data_a2
                1.02TB    1.01TB    1% online       1 node_A_2-old      raid_dp,
                                                                       mirrored,

完成新節點的新增作業

您必須將新的DR群組整合到MetroCluster 「更新」組態中、並在新節點上建立鏡射的資料集合體。

步驟
  1. 在每MetroCluster 個新的EFlash節點上建立鏡射的資料集合體:

    「torage Aggregate create -Aggregate gregate名稱-node-name_-diskcount_no of磁碟_-mirror true」

    附註 每個站台至少必須建立一個鏡射資料Aggregate。建議MetroCluster 每個站台在支援每個站台的兩個鏡射資料集合體、以裝載MDV磁碟區、不過每個站台只支援一個集合體(但不建議)。支援MetroCluster 的是、其中一個站台具有單一鏡射資料集合體、另一個站台則有多個鏡射資料集合體。

    下列範例顯示在node_a_1-new上建立Aggregate。

    cluster_A::> storage aggregate create -aggregate data_a3 -node node_A_1-new -diskcount 10 -mirror t
    
    Info: The layout for aggregate "data_a3" on node "node_A_1-new" would be:
    
          First Plex
    
            RAID Group rg0, 5 disks (block checksum, raid_dp)
                                                                Usable Physical
              Position   Disk                      Type           Size     Size
              ---------- ------------------------- ---------- -------- --------
              dparity    5.10.15                   SAS               -        -
              parity     5.10.16                   SAS               -        -
              data       5.10.17                   SAS         546.9GB  547.1GB
              data       5.10.18                   SAS         546.9GB  558.9GB
              data       5.10.19                   SAS         546.9GB  558.9GB
    
          Second Plex
    
            RAID Group rg0, 5 disks (block checksum, raid_dp)
                                                                Usable Physical
              Position   Disk                      Type           Size     Size
              ---------- ------------------------- ---------- -------- --------
              dparity    4.20.17                   SAS               -        -
              parity     4.20.14                   SAS               -        -
              data       4.20.18                   SAS         546.9GB  547.1GB
              data       4.20.19                   SAS         546.9GB  547.1GB
              data       4.20.16                   SAS         546.9GB  547.1GB
    
          Aggregate capacity available for volume use would be 1.37TB.
    
    Do you want to continue? {y|n}: y
    [Job 440] Job succeeded: DONE
    
    cluster_A::>
  2. 重新整理MetroCluster 此功能的組態:

    1. 進入進階權限模式:

      "進階權限"

    2. 在MetroCluster 其中一個新節點上重新整理此功能:

      《靜態組態》MetroCluster

      以下範例顯示MetroCluster 兩個DR群組上重新整理的支援功能組態:

    cluster_A::*> metrocluster configure -refresh true
    
    [Job 726] Job succeeded: Configure is successful.
    1. 返回管理權限模式:

      「et -priv. admin」

  3. 確認節點已新增至其DR群組。

    cluster_A::*> metrocluster node show
    
    DR                               Configuration  DR
    Group Cluster Node               State          Mirroring Mode
    ----- ------- ------------------ -------------- --------- --------------------
    1     cluster_A
                  node_A_1-old        configured     enabled   normal
                  node_A_2-old        configured     enabled   normal
          cluster_B
                  node_B_1-old        configured     enabled   normal
                  node_B_2-old        configured     enabled   normal
    2     cluster_A
                  node_A_3-new        configured     enabled   normal
                  node_A_4-new        configured     enabled   normal
          cluster_B
                  node_B_3-new        configured     enabled   normal
                  node_B_4-new        configured     enabled   normal
    8 entries were displayed.
    
    cluster_A::*>
  4. 以進階權限將MDV_CRS磁碟區從舊節點移至新節點。

    1. 顯示用於識別MDV磁碟區的磁碟區:

      附註 如果每個站台有單一鏡射資料Aggregate、請將兩個MDV磁碟區移到此單一Aggregate。如果您有兩個以上的鏡射資料集合體、請將每個MDV磁碟區移至不同的集合體。

      以下範例顯示「Volume show」輸出中的MDV磁碟區:

      cluster_A::> volume show
      Vserver   Volume       Aggregate    State      Type       Size  Available Used%
      --------- ------------ ------------ ---------- ---- ---------- ---------- -----
      ...
      
      cluster_A   MDV_CRS_2c78e009ff5611e9b0f300a0985ef8c4_A
                             aggr_b1      -          RW            -          -     -
      cluster_A   MDV_CRS_2c78e009ff5611e9b0f300a0985ef8c4_B
                             aggr_b2      -          RW            -          -     -
      cluster_A   MDV_CRS_d6b0b313ff5611e9837100a098544e51_A
                             aggr_a1      online     RW         10GB     9.50GB    0%
      cluster_A   MDV_CRS_d6b0b313ff5611e9837100a098544e51_B
                             aggr_a2      online     RW         10GB     9.50GB    0%
      ...
      11 entries were displayed.mple
    2. 設定進階權限層級:

      "進階權限"

    3. 一次移動一個MDV磁碟區:

      「Volume move start -volume MDV-volume _-destination-aggregate _agger-on-new節點-vserver vserver-name

      下列範例顯示將「MDV_CRS_d6b0b313ff5611e9837100a098544e51_A」移至「node_a_3」上的Aggregate「data_A3」的命令和輸出。

      cluster_A::> vol move start -volume MDV_CRS_d6b0b313ff5611e9837100a098544e51_A -destination-aggregate data_a3 -vserver cluster_A
      
      Warning: You are about to modify the system volume
               "MDV_CRS_d6b0b313ff5611e9837100a098544e51_A". This might cause severe
               performance or stability problems. Do not proceed unless directed to
               do so by support. Do you want to proceed? {y|n}: y
      [Job 494] Job is queued: Move "MDV_CRS_d6b0b313ff5611e9837100a098544e51_A" in Vserver "cluster_A" to aggregate "data_a3". Use the "volume move show -vserver cluster_A -volume MDV_CRS_d6b0b313ff5611e9837100a098544e51_A" command to view the status of this operation.
    4. 使用volume show命令檢查是否已成功移動MDV磁碟區:

      「Volume show MDV-name

      下列輸出顯示已成功移動MDV Volume。

    cluster_A::> vol show MDV_CRS_d6b0b313ff5611e9837100a098544e51_B
    Vserver     Volume       Aggregate    State      Type       Size  Available Used%
    ---------   ------------ ------------ ---------- ---- ---------- ---------- -----
    cluster_A   MDV_CRS_d6b0b313ff5611e9837100a098544e51_B
                           aggr_a2      online     RW         10GB     9.50GB    0%
    1. 返回管理模式:

      「et -priv. admin」

  5. 將epsilon從舊節點移至新節點:

    1. 識別目前有epsilon的節點:

      「叢集展示-欄位epsilon」

      cluster_B::> cluster show -fields epsilon
      node             epsilon
      ---------------- -------
      node_A_1-old      true
      node_A_2-old      false
      node_A_3-new      false
      node_A_4-new      false
      4 entries were displayed.
    2. 將舊節點(node_a_1-old)上的epsilon設為假:

      「叢集修改-node-old-nod_-epsilon假*」

    3. 在新節點(node_a_3-new)上將epsilon設為true:

      「cluster modify -node-new節點_-epsilon true」

    4. 確認epsilon已移至正確的節點:

      「叢集展示-欄位epsilon」

    cluster_A::> cluster show -fields epsilon
    node             epsilon
    ---------------- -------
    node_A_1-old      false
    node_A_2-old      false
    node_A_3-new      true
    node_A_4-new      false
    4 entries were displayed.