跳转到主内容

共享存储池不健康

Views:
56
Visibility:
Public
Votes:
0
Category:
ontap-9
Specialty:
core
Last Updated:

可不使用  

适用于

ONTAP

解答

当存储池运行不正常时会显示以下消息。

Dec 14 04:06:21 [cluster01-01:raid.sp.unhealthy:notice]: Storage pool sp1 is unhealthy. Reason: One of the aggregates belonging to the storage pool is not in normal state.

发生此错误时,请检查存储池中是否存在 SSD 磁盘故障:重建完成后

cluster01::> storage pool show -storage-pool sp1 -instance

                        Storage Pool Name: sp1
                     UUID of Storage Pool: 84afe3e1-a215-11e5-ac48-00a09854bc10
           Nodes Sharing the Storage Pool: cluster01-01, cluster01-02
          Number of Disks in Storage Pool: 22
                     Allocation Unit Size: 1023GB
      Allocation Unit Data Size for RAID4: 976.6GB
    Allocation Unit Data Size for RAID-DP: 930.1GB
   Allocation Unit Data Size for RAID-TEC: 883.6GB
                             Storage Type: SSD
                 Storage Pool Usable Size: 2.00TB
                  Storage Pool Total Size: 4.00TB
                         Is Pool Healthy?: false
                State of the Storage Pool: degraded
 Reason for Storage Pool Being Unhealthy: One of the aggregates belonging to the storage pool is not in normal state.
Job ID of the Currently Running Operation: -
               Is Allocation Unit Broken?: false

cluster01::>


cluster01::storage pool*> run local aggr status -r aggr1
Aggregate aggr1 (online, raid_dp, reconstruct, hybrid) (block checksums)
  Plex /aggr1/plex0 (online, normal, active, pool0)
    RAID group /aggr1/plex0/rg0 (normal, block checksums)

      RAID Disk Device          HA  SHELF BAY CHAN Pool Type  RPM  Used (MB/blks)    Phys (MB/blks)
      --------- ------          ------------- ---- ---- ---- ----- --------------    --------------
      dparity   2b.64           2b    4   0   FC:B   0  FCAL 15000 272000/557056000  280104/573653840
      parity    2a.50           2a    3   2   FC:A   0  FCAL 15000 272000/557056000  280104/573653840
      data      2a.34           2a    2   2   FC:A   0  FCAL 15000 272000/557056000  274845/562884296
      data      2a.18           2a    1   2   FC:A   0  FCAL 15000 272000/557056000  274845/562884296
      data      2a.65           2a    4   1   FC:A   0  FCAL 15000 272000/557056000  280104/573653840

    RAID group /aggr1/plex0/rg1 (reconstruction 74% completed, block checksums)

      RAID Disk Device          HA  SHELF BAY CHAN Pool Type  RPM  Used (MB/blks)    Phys (MB/blks)
      --------- ------          ------------- ---- ---- ---- ----- --------------    --------------
      dparity   0a.30.4P1       0a    30  4   SA:A   0   SSD   N/A 47619/97525248    47627/97541632
      parity    0a.30.11P1      0a    30  11  SA:A   0   SSD   N/A 47619/97525248    47627/97541632
      data      0b.10.22P1      0b    10  22  SA:B   0   SSD   N/A 47619/97525248    47627/97541632 (reconstruction 74% completed)
      data      0a.30.5P1       0a    30  5   SA:A   0   SSD   N/A 47619/97525248    47627/97541632



,不健康状态应更改为正常。


但是,如果将损坏的磁盘作为分区 SSD 放回系统中、 然后,存储池在以下状态下仍将保持不健康状态:

cluster01::storage pool*> show -storage-pool sp1 -instance

                        Storage Pool Name: sp1
                     UUID of Storage Pool: 84afe3e1-a215-11e5-ac48-00a09854bc10
           Nodes Sharing the Storage Pool: cluster01-01, cluster01-02
          Number of Disks in Storage Pool: 22
                     Allocation Unit Size: 1023GB
      Allocation Unit Data Size for RAID4: 976.6GB
    Allocation Unit Data Size for RAID-DP: 930.1GB
   Allocation Unit Data Size for RAID-TEC: 883.6GB
                             Storage Type: SSD
                 Storage Pool Usable Size: 2.00TB
                  Storage Pool Total Size: 4.00TB
                         Is Pool Healthy?: false
                State of the Storage Pool: degraded
 Reason for Storage Pool Being Unhealthy: Storage pool has more number of disks than expected.
Job ID of the Currently Running Operation: -
               Is Allocation Unit Broken?: false

cluster01::storage pool*>


将显示以下消息:

cluster01::storage pool*> Dec 14 04:39:44 [cluster01-01:raid.sp.unhealthy:notice]: Storage pool sp1 is unhealthy. Reason: Storage pool has more number of disks than expected.

Storage Pool will will will 会保留在此状态中 状态,直到先前发生故障并更换的 SSD 从系统中物理移除或未分区。

其他信息

执行以下步骤以手动取消从存储池中更换的 SSD 驱动器的分区。
注意:只有当某个磁盘未被任何聚合使用时,才能将其从存储池中删除。


在执行以下步骤之前,请格外小心或联系 NetApp 技术支持。

  1. 运行storage pool show-disks -storage-pool <sp name>命令以查找存储池中的现有驱动器。
  2. 选择已被重构替换、需要从存储池中删除的驱动器。
  3. 运行storage disk show -disk <disk name>" -fields diskpathnames, owner该命令以确定共享驱动器的所有者节点以及属主节点上驱动器的本地名称。diskpathnames该字段将以hostname:localname格式指定名称。
  4. 运行storage disk partition show -container-disk <disk name> -fields owner-node-name该命令以了解此共享驱动器中的所有分区名称及其所有者节点。
  5. 对于与磁盘具有不同所有者的分区、通过运行storage disk partition removeowner -partition <partition name>storage disk partition assign -partition <partition name> -owner <disk owner>命令更改所有权。
    执行此步骤后、磁盘及其所有分区的所有者应相同。
  6. 删除到所有者节点的节点 shell 。
  7. 运行disk unpartition <disk name>该命令
    ,一旦磁盘未分区、它将作为备用磁盘返回。
    磁盘变为备用磁盘后、存储池应报告状态正常。

 

NetApp provides no representations or warranties regarding the accuracy or reliability or serviceability of any information or recommendations provided in this publication or with respect to any results that may be obtained by the use of the information or observance of any recommendations provided herein. The information in this document is distributed AS IS and the use of this information or the implementation of any recommendations or techniques herein is a customer's responsibility and depends on the customer's ability to evaluate and integrate them into the customer's operational environment. This document and the information contained herein may be used solely in connection with the NetApp products discussed in this document.