Hi,
I've recently encountered a failing drive in a ZFS pool. The pool is striped, with no redundancy. Currently, the only affected files appear to be ones that can be easily be replaced, and I'm hoping to keep it that way to avoid having to take the whole server offline to restore from backups. I tried to remove the failing drive from the pool using
I didn't cancel it myself, but I have noticed the number of read errors has increased since starting the removal. Is it the read errors that likely caused the removal to be cancelled?
How should I proceed to remove this drive with the least possible disruption? I should mention that I don't plan to replace it any time soon, as there's still plenty of space on the remaining drives for the forseeable future.
Thanks in advance!
Alex
I've recently encountered a failing drive in a ZFS pool. The pool is striped, with no redundancy. Currently, the only affected files appear to be ones that can be easily be replaced, and I'm hoping to keep it that way to avoid having to take the whole server offline to restore from backups. I tried to remove the failing drive from the pool using
zpool remove
(there's plenty of space on other drives for the contents), but after a few hours, I noticed in zpool status
:
Code:
remove: Removal of /dev/disk/by-partlabel/AAAE-03 canceled on Mon Dec 11 20:05:50 2023
364K memory used for removed device mappings
...
NAME STATE READ WRITE CKSUM
AAAE-03 DEGRADED 724 3 641 too many errors
I didn't cancel it myself, but I have noticed the number of read errors has increased since starting the removal. Is it the read errors that likely caused the removal to be cancelled?
How should I proceed to remove this drive with the least possible disruption? I should mention that I don't plan to replace it any time soon, as there's still plenty of space on the remaining drives for the forseeable future.
Thanks in advance!
Alex