我最近买了一台 QNap NAS,决定试用几个 NAS 操作系统,其中包括 TrueNAS。在 TrueNAS 上创建 ZFS 池并传输一些文件后,我对 TrueNAS 感到不满意,并决定试用 QNap 的操作系统 (QuTS)。这两个操作系统都安装在 NAS 上。我进行了 QuTS 初始化,发现 QuTS 不会自动导入 ZFS 池,并且对 QuTS 的界面感到不满意。之后我决定重新安装 TrueNAS 和 QuTS(QuTS 已擦除安装 TrueNAS 的磁盘,该磁盘与 ZFS 池中的磁盘不同)。这就是目前的情况。

有问题的 ZFS 池包含 3 个磁盘,每个磁盘容量为 5To。TrueNAS 在其界面中显示该池,但无法挂载它。该池名为main-pool。以下是我尝试过的不同命令及其结果:

# zpool import -a
no pools available to import

# zpool status -v
  pool: boot-pool
 state: ONLINE
  scan: scrub repaired 0B in 00:00:08 with 0 errors on Sun Sep 22 03:45:10 2024
config:

        NAME         STATE     READ WRITE CKSUM
        boot-pool    ONLINE       0     0     0
          nvme0n1p3  ONLINE       0     0     0

errors: No known data errors

# zpool status -v main-pool
cannot open 'main-pool': no such pool

# zdb -l /dev/sd{b,c,e}    
failed to unpack label 0
failed to unpack label 1
------------------------------------
LABEL 2 (Bad label cksum)
------------------------------------
    version: 5000
    name: 'main-pool'
    state: 0
    txg: 4
    pool_guid: 8298165464761202083
    errata: 0
    hostid: 1555077055
    hostname: 'nas'
    top_guid: 4887568585043273647
    guid: 12714426885291094564
    vdev_children: 1
    vdev_tree:
        type: 'raidz'
        id: 0
        guid: 4887568585043273647
        nparity: 1
        metaslab_array: 128
        metaslab_shift: 34
        ashift: 12
        asize: 15002922123264
        is_log: 0
        create_txg: 4
        children[0]:
            type: 'disk'
            id: 0
            guid: 12714426885291094564
            path: '/dev/disk/by-partuuid/8b808c7a-cacd-46d1-b400-9f8c71d51b30'
            whole_disk: 0
            create_txg: 4
        children[1]:
            type: 'disk'
            id: 1
            guid: 822100609061977128
            path: '/dev/disk/by-partuuid/8965dbb6-9484-4da5-ba1b-d33303b19ae5'
            whole_disk: 0
            create_txg: 4
        children[2]:
            type: 'disk'
            id: 2
            guid: 1731788408784132969
            path: '/dev/disk/by-partuuid/fae9eb3e-e782-4115-bbc9-cde4dc72c408'
            whole_disk: 0
            create_txg: 4
    features_for_read:
        com.delphix:hole_birth
        com.delphix:embedded_data
        com.klarasystems:vdev_zaps_v2
    labels = 2 3 

奇怪的是,似乎我可以通过开发路径分别导入每个磁盘:

# zpool import -d /dev/sdb                               
   pool: main-pool
     id: 8298165464761202083
  state: UNAVAIL
status: The pool was last accessed by another system.
 action: The pool cannot be imported due to damaged devices or data.
   see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-EY
 config:

        main-pool                                 UNAVAIL  insufficient replicas
          raidz1-0                                UNAVAIL  insufficient replicas
            sdb                                   ONLINE
            8965dbb6-9484-4da5-ba1b-d33303b19ae5  UNAVAIL
            fae9eb3e-e782-4115-bbc9-cde4dc72c408  UNAVAIL

但并非全部都在一起:

# zpool import -d /dev/sdb -d /dev/sdc -d /dev/sde                   
   pool: main-pool
     id: 8298165464761202083
  state: UNAVAIL
status: The pool was last accessed by another system.
 action: The pool cannot be imported due to damaged devices or data.
   see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-EY
 config:

        main-pool   UNAVAIL  insufficient replicas
          raidz1-0  UNAVAIL  insufficient replicas
            sdb     UNAVAIL  invalid label
            sdc     UNAVAIL  invalid label
            sde     UNAVAIL  invalid label

那么…我的 ZFS 池丢失了吗?

如果是这样,那么当系统在 ZFS 池正确导出之前崩溃时,人们如何恢复 ZFS 池?例如,如果包含系统的磁盘发生故障?

编辑:下面还有更多信息。

以下是输出blkid

# blkid
/dev/mapper/nvme0n1p4: UUID="ce0e9fc8-60b8-4c80-9a3b-d06e45d06ba3" TYPE="swap"
/dev/nvme0n1p1: PARTUUID="4633d564-6e46-4db6-9579-b01358dd0264"
/dev/nvme0n1p2: LABEL_FATBOOT="EFI" LABEL="EFI" UUID="6D4A-8FC7" BLOCK_SIZE="512" TYPE="vfat" PARTUUID="1be2d0e3-7bd4-4066-863f-4db859cefd35"
/dev/nvme0n1p3: LABEL="boot-pool" UUID="17872462125275668883" UUID_SUB="9643558866770636482" BLOCK_SIZE="4096" TYPE="zfs_member" PARTUUID="efb94924-9001-4825-86d3-1285ee7a0701"
/dev/nvme0n1p4: PARTUUID="c8537bb8-6d57-4246-9e13-667680e23bf6"
/dev/sda1: UUID="bc485df0-39b6-10b6-7c50-b6f458a74999" UUID_SUB="e9118230-67c8-ed89-4a74-ad184d8cdea6" LABEL="9" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="73fbca5c-3870-4ed3-a7c6-dd22e58a193d"
/dev/sda2: PARTLABEL="primary" PARTUUID="31dd8d57-ac07-4128-ae67-19bda9b7be17"
/dev/sda3: PARTLABEL="primary" PARTUUID="4a1ebaf5-ef92-40ff-872b-b4281bf6f88f"
/dev/sda4: UUID="8a5a3234-6059-95c0-06aa-b64147247264" UUID_SUB="3402db43-204c-2354-97ae-9a17d51a4ce1" LABEL="13" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="41405202-70c9-44dc-a2a1-de773b808279"
/dev/sda5: UUID="32e6353a-e9ac-afda-3e0f-36e8b3d075ca" UUID_SUB="bf2c22d4-581c-5c16-461f-bcf048e22062" LABEL="322" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="b68d9269-1024-468d-b021-821e7e68150a"
/dev/sdb1: UUID="bc485df0-39b6-10b6-7c50-b6f458a74999" UUID_SUB="15681faf-51c3-dfce-ad94-cc40fd5925c5" LABEL="9" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="4ddc315a-dd88-477e-bd2f-a145286d8596"
/dev/sdb2: PARTLABEL="primary" PARTUUID="e9222737-18b8-4a3c-84cf-fefd9a5a6acf"
/dev/sdb3: PARTLABEL="primary" PARTUUID="320b0e51-d289-44e0-b7d9-c7a7d8448e53"
/dev/sdb4: UUID="8a5a3234-6059-95c0-06aa-b64147247264" UUID_SUB="1375ce0f-1448-cbc0-36d3-bdef9382d4f7" LABEL="13" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="3426a0c2-9c13-4451-a109-106d6e1aa913"
/dev/sdb5: UUID="32e6353a-e9ac-afda-3e0f-36e8b3d075ca" UUID_SUB="d4fe128b-cd0a-ef96-df43-32562bd22186" LABEL="322" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="48cfc3db-425d-4ecc-9d6f-060c78d16f82"
/dev/sdc1: UUID="bc485df0-39b6-10b6-7c50-b6f458a74999" UUID_SUB="5aa0e2f6-d6f3-f902-dbcf-fa23cf312d43" LABEL="9" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="3eba061c-a2d0-4059-ac47-b1ec9f79eadc"
/dev/sdc2: PARTLABEL="primary" PARTUUID="b508be5e-5bc3-4278-99ae-d73d6f2ad28c"
/dev/sdc3: PARTLABEL="primary" PARTUUID="47b34f9c-1dff-4c04-a5b5-4403727ec19b"
/dev/sdc4: UUID="8a5a3234-6059-95c0-06aa-b64147247264" UUID_SUB="bd86e25c-5a7b-af5a-d904-4bd2de8e64d7" LABEL="13" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="73a12459-f90d-4929-b905-22774e366a6c"
/dev/sdc5: UUID="32e6353a-e9ac-afda-3e0f-36e8b3d075ca" UUID_SUB="94f8107f-8d17-6357-d788-3b242fb13c92" LABEL="322" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="50508432-d0c8-41de-b656-da97b6eb67e3"
/dev/sdd1: UUID="bc485df0-39b6-10b6-7c50-b6f458a74999" UUID_SUB="fdb00b87-7805-ab91-c815-d18c2270b6f3" LABEL="9" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="0356a6fb-5aee-4a30-a9c3-d248d38dd33c"
/dev/sdd2: PARTLABEL="primary" PARTUUID="a6596854-3bcf-4ac6-96d1-9558369bab1c"
/dev/sdd3: PARTLABEL="primary" PARTUUID="94cf24c3-5de8-4106-a71d-b2bbe16c9344"
/dev/sdd4: UUID="8a5a3234-6059-95c0-06aa-b64147247264" UUID_SUB="5834801a-a5cf-03eb-6e9e-24123cab09f4" LABEL="13" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="5135b9b2-ce67-442b-b7eb-711a9825b084"
/dev/sdd5: UUID="32e6353a-e9ac-afda-3e0f-36e8b3d075ca" UUID_SUB="f60ec16c-97c7-f84e-e02c-dbba79d0a943" LABEL="322" TYPE="linux_raid_member" PARTLABEL="primary" PARTUUID="4b76e000-b45d-48ad-aaf7-934363b13aa5"
/dev/sde1: UUID="9333eb40-8071-460b-972f-a3192d483667" BLOCK_SIZE="1024" TYPE="ext2" PARTUUID="b66e61e9-01"
/dev/sde2: LABEL="QTS_BOOT_PART2" UUID="9cb067c5-0ede-4482-950d-cdf662995f85" BLOCK_SIZE="1024" TYPE="ext2" PARTUUID="b66e61e9-02"
/dev/sde3: LABEL="QTS_BOOT_PART3" UUID="a2e549dc-4a48-4ab4-bed8-9fcd14845f13" BLOCK_SIZE="1024" TYPE="ext2" PARTUUID="b66e61e9-03"
/dev/sde5: UUID="9015507c-d233-41e0-8480-0fe81ff64be2" BLOCK_SIZE="1024" TYPE="ext2" PARTUUID="b66e61e9-05"
/dev/sde6: UUID="bbeb321e-525b-4ffb-96a3-028432d3dea8" BLOCK_SIZE="1024" TYPE="ext2" PARTUUID="b66e61e9-06"
/dev/sde7: UUID="1e9f878e-af23-48b9-96e4-1fabe75c0c81" BLOCK_SIZE="1024" TYPE="ext2" PARTUUID="b66e61e9-07"

2

  • 您的 SCSI 设备名称是否发生了变化? 的输出是什么blkid


    – 


  • 不幸的是,是的,他们改变了。例如/dev/disk/by-partuuid/8b808c7a-cacd-46d1-b400-9f8c71d51b30不再存在。我用结果更新了我的问题blkid


    – 

0