I had proxmox set up with an SSD as the boot drive and later added two HDDs as additional storage as a mirrored zfs pool. After an electrical storm and a hard reboot, one of the zfs drives has started clicking badly and on booting the machine gets stuck very early in the process at
I booted the server from a usb key with a live recovery environment and an fsck shows pve-root to be clean, but the zfs pool is borked
I guess my question is, why is a non-boot drive pool halting machine startup, and if I replace the failed drive using the recovery live usb environment should that get booting up again?
Code:
/dev/mapper/pve-root: clean, 116132/6291456 files, 2831205/25165824 blocks
I booted the server from a usb key with a live recovery environment and an fsck shows pve-root to be clean, but the zfs pool is borked
Code:
zpool status -x
pool: zfs_mirror_0
state: DEGRADED
status: One or more devices could not be used because the label is missing or
invalid. Sufficient replicas exist for the pool to continue
functioning in a degraded state.
action: Replace the device using 'zpool replace'.
see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-4J
scan: scrub repaired 0B in 03:04:09 with 0 errors on Sun May 14 01:28:10 2023
config:
NAME STATE READ WRITE CKSUM
zfs_mirror_0 DEGRADED 0 0 0
mirror-0 DEGRADED 0 0 0
ata-HUH721010ALE601_7JGH6YYC ONLINE 0 0 0
17036352222114924081 UNAVAIL 0 0 0 was /dev/disk/by-id/ata-HUH721010ALE601_7JH2D71K-part1
errors: No known data errors
I guess my question is, why is a non-boot drive pool halting machine startup, and if I replace the failed drive using the recovery live usb environment should that get booting up again?