i have a Proxmox node running on version 7.1-7 with 4 SSD and 2 HDD storage configured with ZFS. Everything was working perfectly. Until yesterday, i rebooted node (whis was not my first reboot after setup), and now one of the SSD storages named 'zfs_ssd_1' is not activated. The GUI shows that it is enabled but not activated. When I run
I have checked similar forum posts (https://forum.proxmox.com/threads/zfs-i-o-error.33670/) and ran a
Are there any other solutions or methods I can try to resolve this issue?
zpool status
blkid
lsblk
after i ran systemctl test
zpool status
, it does not show 'zfs_ssd_1'. However, when I runblkid
, it shows 'zfs_ssd_1' on '/dev/sdb1'. When I run lsblk
, it confirms that '/dev/sdb1' is mounted.I have checked similar forum posts (https://forum.proxmox.com/threads/zfs-i-o-error.33670/) and ran a
smartctl
test, which completed with no errors. However, I was unable to clear errors with zpool clear zfs_ssd_1
, which gave the error 'cannot open 'zfs_ssd_1': no such pool'. I believe the problem is with ZFS since other system commands can detect the storage with no problem.Are there any other solutions or methods I can try to resolve this issue?
zpool status
Code:
zpool status
pool: zfs_hdd_1
state: ONLINE
scan: scrub repaired 0B in 14:02:51 with 0 errors on Sun Apr 9 14:26:52 2023
config:
NAME STATE READ WRITE CKSUM
zfs_hdd_1 ONLINE 0 0 0
ata-ST16000NE000-2WX103_ZR50J8HZ ONLINE 0 0 0
errors: No known data errors
pool: zfs_hdd_2
state: ONLINE
scan: scrub repaired 0B in 17:14:09 with 0 errors on Sun Apr 9 17:38:12 2023
config:
NAME STATE READ WRITE CKSUM
zfs_hdd_2 ONLINE 0 0 0
ata-ST16000NM001G-2KK103_ZL2GJ7GS ONLINE 0 0 0
errors: No known data errors
pool: zfs_nvme_1
state: ONLINE
scan: scrub repaired 0B in 00:39:49 with 0 errors on Sun Apr 9 01:03:54 2023
config:
NAME STATE READ WRITE CKSUM
zfs_nvme_1 ONLINE 0 0 0
nvme-eui.6479a73f202002a7-part1 ONLINE 0 0 0
errors: No known data errors
pool: zfs_ssd_2
state: ONLINE
scan: scrub repaired 0B in 02:32:11 with 0 errors on Sun Apr 9 02:56:18 2023
config:
NAME STATE READ WRITE CKSUM
zfs_ssd_2 ONLINE 0 0 0
ata-Samsung_SSD_870_QVO_4TB_S5STNG0NC01902Z ONLINE 0 0 0
errors: No known data errors
pool: zfs_ssd_3
state: ONLINE
scan: scrub repaired 0B in 01:32:30 with 0 errors on Sun Apr 9 01:56:38 2023
config:
NAME STATE READ WRITE CKSUM
zfs_ssd_3 ONLINE 0 0 0
ata-Samsung_SSD_870_QVO_4TB_S5STNG0NC01843T ONLINE 0 0 0
blkid
Code:
blkid
/dev/nvme1n1p1: LABEL="zfs_nvme_1" UUID="11974131303405226487" UUID_SUB="3233527513562740301" BLOCK_SIZE="4096" TYPE="zfs_member" PARTLABEL="zfs-a3fa31dd8248b3be" PARTUUID="c8f3520b-492e-ae42-9297-ca4a43033d3c"
/dev/nvme0n1p2: UUID="3CEC-091F" BLOCK_SIZE="512" TYPE="vfat" PARTUUID="529e330a-9598-4e94-98f0-29610d14c440"
/dev/nvme0n1p3: UUID="1g3mOx-PfHQ-psa2-GfP9-ydUg-2i1R-YFQJIr" TYPE="LVM2_member" PARTUUID="d1354f69-a08c-4bd8-a7df-41195ce19c3c"
/dev/mapper/pve-swap: UUID="3458894b-613d-4ff6-9d50-fb2494342ba0" TYPE="swap"
/dev/mapper/pve-root: UUID="2d740743-96f8-4058-8d05-1f681e942e6f" BLOCK_SIZE="4096" TYPE="ext4"
/dev/sda1: UUID="d701877a-a835-40c0-8564-1449a579d609" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="af544811-8def-4582-be32-35ddb8df4abe"
/dev/sdb1: LABEL="zfs_ssd_1" UUID="13532520614518614830" UUID_SUB="15735742699506282199" BLOCK_SIZE="4096" TYPE="zfs_member" PARTLABEL="zfs-f14f52327cf1b1e2" PARTUUID="7f9c12b3-d668-3446-abef-bb31896e125e"
/dev/sdf1: LABEL="zfs_ssd_3" UUID="4438819123320962145" UUID_SUB="17266877199441579046" BLOCK_SIZE="4096" TYPE="zfs_member" PARTLABEL="zfs-a8b753fc7c283da9" PARTUUID="a4beaccf-2e15-f04f-b353-697aae991b89"
/dev/sdc1: LABEL="zfs_hdd_1" UUID="14137249408840784235" UUID_SUB="16108780857322988310" BLOCK_SIZE="4096" TYPE="zfs_member" PARTLABEL="zfs-f602d675bc0d0c59" PARTUUID="dc8db521-f2f0-1948-aa1b-fade319c6dc5"
/dev/sdd1: LABEL="zfs_hdd_2" UUID="3064075094854740590" UUID_SUB="8331361168905152812" BLOCK_SIZE="4096" TYPE="zfs_member" PARTLABEL="zfs-06a7e839d48182ca" PARTUUID="e2a6369b-f039-7e49-84c1-d557292034d2"
/dev/sde1: LABEL="zfs_ssd_2" UUID="7460628251768455693" UUID_SUB="5418112381992914531" BLOCK_SIZE="4096" TYPE="zfs_member" PARTLABEL="zfs-54ed532b4af9a5d4" PARTUUID="6535f9df-a239-0841-a9be-d26dc2443608"
/dev/zd48p2: UUID="d7149358-5b59-480b-ac69-3302698d3c15" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="90d459d2-94eb-412a-a46c-aaa708e21b17"
/dev/zd0: LABEL="storage" UUID="9921686d-58ac-40c6-9d0c-b9fdcc859253" BLOCK_SIZE="4096" TYPE="ext4"
/dev/zd16p1: UUID="87baa5b5-7ee5-4947-83a6-28e029af9847" BLOCK_SIZE="4096" TYPE="ext4" PARTLABEL="storage" PARTUUID="982e4c47-d1af-422e-8897-9b901e414196"
/dev/zd80p2: UUID="f16de96f-610b-4abc-97d0-8e745521a37b" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="0e57feb7-8f41-43e8-a343-c10ee0b82497"
/dev/zd112: UUID="2566b6bf-a4c9-4435-a98f-c7a5b387a42d" BLOCK_SIZE="4096" TYPE="ext4"
/dev/zd128p2: UUID="ff9fd2fc-5e17-49bb-8acb-f1dab0f25fef" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="9c8f318a-ed6b-48b9-ba54-e91099d50841"
/dev/zd96p2: UUID="ababb4e9-46f0-4084-af4f-f703da8f6764" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="c79d76bc-b8b7-493e-a022-f3cdb12aaff9"
/dev/zd144p2: UUID="2a6f8edf-c9ab-4f46-a736-697c241341a2" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="44895e25-7535-4c11-9ce4-93dc9215d5d8"
/dev/zd144p3: UUID="lB0Qr6-hBvK-CNFT-8lG7-xEKw-mcZ0-LVxOIS" TYPE="LVM2_member" PARTUUID="649b890d-65b7-43e2-9c23-508ff783f088"
/dev/zd160p2: UUID="d5a3ed9c-a3fe-437d-b1b3-3c7b1f77945f" BLOCK_SIZE="4096" TYPE="ext4" PARTUUID="32aa0079-3748-4a0b-9513-6e2f845c8023"
/dev/zd32: LABEL="storage" UUID="27ba33f9-9c65-4342-9aa2-877766d1dbb2" BLOCK_SIZE="4096" TYPE="ext4"
/dev/zd64: LABEL="cold_storage" UUID="7b2da8cf-9024-403a-93f1-1d28026f567c" BLOCK_SIZE="4096" TYPE="ext4"
/dev/zd176: UUID="2c3c1929-f9d1-41d3-af4a-82b4bf6bd3e5" BLOCK_SIZE="4096" TYPE="ext4"
/dev/zd192: UUID="046a408f-2919-4a09-a28c-19abe52e17b0" BLOCK_SIZE="4096" TYPE="ext4"
/dev/zd224: UUID="666d2953-1acd-497e-a462-5d05aa91d121" BLOCK_SIZE="4096" TYPE="ext4"
/dev/zd208: LABEL="storage" UUID="9921686d-58ac-40c6-9d0c-b9fdcc859253" BLOCK_SIZE="4096" TYPE="ext4"
/dev/zd240: LABEL="storage" UUID="9921686d-58ac-40c6-9d0c-b9fdcc859253" BLOCK_SIZE="4096" TYPE="ext4"
/dev/nvme1n1p9: PARTUUID="c9710ed6-cc8b-3e4e-b39c-e6c49338b2fc"
/dev/nvme0n1p1: PARTUUID="d0edf39e-b685-434f-b671-d40319f36ec4"
/dev/sdb9: PARTUUID="51d17a08-4106-4f40-977e-95ce92023f2d"
/dev/sdf9: PARTUUID="7a85b0e9-52af-e741-bdf5-58455575deee"
/dev/sdc9: PARTUUID="5b3e3d8d-8478-4344-a5c6-7cc913f1f641"
/dev/sdd9: PARTUUID="5532a9a5-6a82-ae48-91bf-a07b23ea18e9"
/dev/sde9: PARTUUID="05ebb001-1a4d-cd44-a321-2e21083ebefb"
/dev/zd48p1: PARTUUID="51d05779-7b3b-44d8-a1ee-4e7af3c02fef"
/dev/zd80p1: PARTUUID="6263d773-fa78-450e-bb09-6334961e287f"
/dev/zd128p1: PARTUUID="c4a0a16f-eab5-4562-859f-d40ea49d8ef0"
/dev/zd96p1: PARTUUID="b3918558-cde6-42e3-9e11-037927955d64"
/dev/zd144p1: PARTUUID="babfee28-4550-431a-beaf-f1187992b4b4"
/dev/zd160p1: PARTUUID="28fc7126-c7f5-4fa1-bfc9-e58336da6643"
lsblk
Code:
lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sda 8:0 0 3.6T 0 disk
└─sda1 8:1 0 3.6T 0 part /mnt/pve/tempssd
sdb 8:16 0 3.6T 0 disk
├─sdb1 8:17 0 3.6T 0 part
└─sdb9 8:25 0 8M 0 part
sdc 8:32 0 14.6T 0 disk
├─sdc1 8:33 0 14.6T 0 part
└─sdc9 8:41 0 8M 0 part
sdd 8:48 0 14.6T 0 disk
├─sdd1 8:49 0 14.6T 0 part
└─sdd9 8:57 0 8M 0 part
sde 8:64 0 3.6T 0 disk
├─sde1 8:65 0 3.6T 0 part
└─sde9 8:73 0 8M 0 part
sdf 8:80 0 3.6T 0 disk
├─sdf1 8:81 0 3.6T 0 part
└─sdf9 8:89 0 8M 0 part
zd0 230:0 0 3.4T 0 disk
zd16 230:16 0 2T 0 disk
└─zd16p1 230:17 0 2T 0 part
zd32 230:32 0 7.8T 0 disk
zd48 230:48 0 300G 0 disk
├─zd48p1 230:49 0 1M 0 part
└─zd48p2 230:50 0 300G 0 part
zd64 230:64 0 5.9T 0 disk
zd80 230:80 0 1.2T 0 disk
├─zd80p1 230:81 0 1M 0 part
└─zd80p2 230:82 0 1.2T 0 part
zd96 230:96 0 300G 0 disk
├─zd96p1 230:97 0 1M 0 part
└─zd96p2 230:98 0 300G 0 part
zd112 230:112 0 1000G 0 disk
zd128 230:128 0 250G 0 disk
├─zd128p1 230:129 0 1M 0 part
└─zd128p2 230:130 0 250G 0 part
zd144 230:144 0 250G 0 disk
├─zd144p1 230:145 0 1M 0 part
├─zd144p2 230:146 0 1G 0 part
└─zd144p3 230:147 0 249G 0 part
zd160 230:160 0 300G 0 disk
├─zd160p1 230:161 0 1M 0 part
└─zd160p2 230:162 0 300G 0 part
zd176 230:176 0 1.5T 0 disk
zd192 230:192 0 1000G 0 disk
zd208 230:208 0 3.4T 0 disk
zd224 230:224 0 1000G 0 disk
zd240 230:240 0 5.4T 0 disk
nvme1n1 259:0 0 3.6T 0 disk
├─nvme1n1p1 259:1 0 3.6T 0 part
└─nvme1n1p9 259:2 0 8M 0 part
nvme0n1 259:3 0 476.9G 0 disk
├─nvme0n1p1 259:4 0 1007K 0 part
├─nvme0n1p2 259:5 0 512M 0 part /boot/efi
└─nvme0n1p3 259:6 0 476.4G 0 part
├─pve-swap 253:0 0 8G 0 lvm [SWAP]
├─pve-root 253:1 0 96G 0 lvm /
├─pve-data_tmeta 253:2 0 3.6G 0 lvm
│ └─pve-data-tpool 253:4 0 349.3G 0 lvm
│ └─pve-data 253:5 0 349.3G 1 lvm
└─pve-data_tdata 253:3 0 349.3G 0 lvm
└─pve-data-tpool 253:4 0 349.3G 0 lvm
└─pve-data 253:5 0 349.3G 1 lvm
after i ran systemctl test
Code:
root@deepmia-cluster2:~# smartctl --all /dev/sdb1
smartctl 7.2 2020-12-30 r5155 [x86_64-linux-5.13.19-2-pve] (local build)
Copyright (C) 2002-20, Bruce Allen, Christian Franke, www.smartmontools.org
=== START OF INFORMATION SECTION ===
Model Family: Samsung based SSDs
Device Model: Samsung SSD 870 QVO 4TB
Serial Number: S5STNG0NC01892H
LU WWN Device Id: 5 002538 f70c096d6
Firmware Version: SVQ01B6Q
User Capacity: 4,000,787,030,016 bytes [4.00 TB]
Sector Size: 512 bytes logical/physical
Rotation Rate: Solid State Device
Form Factor: 2.5 inches
TRIM Command: Available, deterministic, zeroed
Device is: In smartctl database [for details use: -P show]
ATA Version is: ACS-4 T13/BSR INCITS 529 revision 5
SATA Version is: SATA 3.3, 6.0 Gb/s (current: 6.0 Gb/s)
Local Time is: Sat Apr 29 12:52:37 2023 +03
SMART support is: Available - device has SMART capability.
SMART support is: Enabled
=== START OF READ SMART DATA SECTION ===
SMART overall-health self-assessment test result: PASSED
General SMART Values:
Offline data collection status: (0x00) Offline data collection activity
was never started.
Auto Offline Data Collection: Disabled.
Self-test execution status: ( 25) The self-test routine was aborted by
the host.
Total time to complete Offline
data collection: ( 0) seconds.
Offline data collection
capabilities: (0x53) SMART execute Offline immediate.
Auto Offline data collection on/off support.
Suspend Offline collection upon new
command.
No Offline surface scan supported.
Self-test supported.
No Conveyance Self-test supported.
Selective Self-test supported.
SMART capabilities: (0x0003) Saves SMART data before entering
power-saving mode.
Supports SMART auto save timer.
Error logging capability: (0x01) Error logging supported.
General Purpose Logging supported.
Short self-test routine
recommended polling time: ( 2) minutes.
Extended self-test routine
recommended polling time: ( 320) minutes.
SCT capabilities: (0x003d) SCT Status supported.
SCT Error Recovery Control supported.
SCT Feature Control supported.
SCT Data Table supported.
SMART Attributes Data Structure revision number: 1
Vendor Specific SMART Attributes with Thresholds:
ID# ATTRIBUTE_NAME FLAG VALUE WORST THRESH TYPE UPDATED WHEN_FAILED RAW_VALUE
5 Reallocated_Sector_Ct 0x0033 100 100 010 Pre-fail Always - 0
9 Power_On_Hours 0x0032 096 096 000 Old_age Always - 19232
12 Power_Cycle_Count 0x0032 099 099 000 Old_age Always - 138
177 Wear_Leveling_Count 0x0013 097 097 000 Pre-fail Always - 28
179 Used_Rsvd_Blk_Cnt_Tot 0x0013 100 100 010 Pre-fail Always - 0
181 Program_Fail_Cnt_Total 0x0032 100 100 010 Old_age Always - 0
182 Erase_Fail_Count_Total 0x0032 100 100 010 Old_age Always - 0
183 Runtime_Bad_Block 0x0013 100 100 010 Pre-fail Always - 0
187 Uncorrectable_Error_Cnt 0x0032 100 100 000 Old_age Always - 0
190 Airflow_Temperature_Cel 0x0032 069 047 000 Old_age Always - 31
195 ECC_Error_Rate 0x001a 200 200 000 Old_age Always - 0
199 CRC_Error_Count 0x003e 100 100 000 Old_age Always - 0
235 POR_Recovery_Count 0x0012 099 099 000 Old_age Always - 118
241 Total_LBAs_Written 0x0032 099 099 000 Old_age Always - 127290026072
SMART Error Log Version: 1
No Errors Logged
SMART Self-test log structure revision number 1
Num Test_Description Status Remaining LifeTime(hours) LBA_of_first_error
# 1 Extended offline Aborted by host 90% 19170 -
# 2 Extended offline Completed without error 00% 19167 -
# 3 Extended offline Aborted by host 90% 19162 -
SMART Selective self-test log data structure revision number 1
SPAN MIN_LBA MAX_LBA CURRENT_TEST_STATUS
1 0 0 Not_testing
2 0 0 Not_testing
3 0 0 Not_testing
4 0 0 Not_testing
5 0 0 Not_testing
256 0 65535 Read_scanning was never started
Selective self-test flags (0x0):
After scanning selected spans, do NOT read-scan remainder of disk.
If Selective self-test is pending on power-up, resume after 0 minute delay.
Last edited: