I'm running Proxmox on a DeskMini H470 with the following disk layout:
The other day, Proxmox was completely hung and I was unable to SSH into it or get it to respond. All of my VMs were also unresponsive, all of their services were offline, and I was unable to ping any of them. So, I powered off my Proxmox box manually. Now, I'm noticing my zpool is degraded.
What worries me is that my missing drive isn't being seen by
I'm also not seeing
- 1x Samsung 970 PRO 512GB (for Proxmox installation)
- 2x Intel D3-S4510 960GB (for ZFS mirror - these are "enterprise" drives with less than 8000 power on hours)
The other day, Proxmox was completely hung and I was unable to SSH into it or get it to respond. All of my VMs were also unresponsive, all of their services were offline, and I was unable to ping any of them. So, I powered off my Proxmox box manually. Now, I'm noticing my zpool is degraded.
Code:
root@proxmox02:~# zpool status -P
pool: intel_mirror
state: DEGRADED
status: One or more devices could not be used because the label is missing or
invalid. Sufficient replicas exist for the pool to continue
functioning in a degraded state.
action: Replace the device using 'zpool replace'.
see: https://openzfs.github.io/openzfs-docs/msg/ZFS-8000-4J
scan: scrub repaired 0B in 00:36:53 with 0 errors on Sun Jul 9 01:00:54 2023
config:
NAME STATE READ WRITE CKSUM
intel_mirror DEGRADED 0 0 0
mirror-0 DEGRADED 0 0 0
3372824492195189717 UNAVAIL 0 0 0 was /dev/disk/by-id/ata-INTEL_SSDSC2KB960G8_BTYF2060050M960CGN-part1
/dev/disk/by-id/ata-INTEL_SSDSC2KB960G8_BTYF206005VJ960CGN-part1 ONLINE 0 0 0
What worries me is that my missing drive isn't being seen by
lsblk
.
Code:
root@proxmox02:~# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS
sda 8:0 0 894.3G 0 disk
├─sda1 8:1 0 894.2G 0 part
└─sda9 8:9 0 8M 0 part
zd0 230:0 0 8G 0 disk
├─zd0p1 230:1 0 7G 0 part
├─zd0p2 230:2 0 1K 0 part
└─zd0p5 230:5 0 975M 0 part
zd16 230:16 0 1M 0 disk
zd32 230:32 0 120G 0 disk
└─zd32p1 230:33 0 120G 0 part
zd48 230:48 0 75G 0 disk
├─zd48p1 230:49 0 74G 0 part
├─zd48p2 230:50 0 1K 0 part
└─zd48p5 230:53 0 975M 0 part
zd64 230:64 0 16G 0 disk
├─zd64p1 230:65 0 32M 0 part
├─zd64p2 230:66 0 24M 0 part
├─zd64p3 230:67 0 256M 0 part
├─zd64p4 230:68 0 24M 0 part
├─zd64p5 230:69 0 256M 0 part
├─zd64p6 230:70 0 8M 0 part
├─zd64p7 230:71 0 96M 0 part
└─zd64p8 230:72 0 15.3G 0 part
zd80 230:80 0 8G 0 disk
├─zd80p1 230:81 0 7G 0 part
├─zd80p2 230:82 0 1K 0 part
└─zd80p5 230:85 0 975M 0 part
zd96 230:96 0 55G 0 disk
├─zd96p1 230:97 0 549M 0 part
└─zd96p2 230:98 0 54.5G 0 part
nvme0n1 259:0 0 476.9G 0 disk
├─nvme0n1p1 259:1 0 1007K 0 part
├─nvme0n1p2 259:2 0 512M 0 part /boot/efi
└─nvme0n1p3 259:3 0 476.4G 0 part
├─pve-swap 253:0 0 8G 0 lvm [SWAP]
├─pve-root 253:1 0 96G 0 lvm /
├─pve-data_tmeta 253:2 0 3.6G 0 lvm
│ └─pve-data-tpool 253:4 0 349.3G 0 lvm
│ └─pve-data 253:5 0 349.3G 1 lvm
└─pve-data_tdata 253:3 0 349.3G 0 lvm
└─pve-data-tpool 253:4 0 349.3G 0 lvm
└─pve-data 253:5 0 349.3G 1 lvm
I'm also not seeing
sdb
in dmesg
. Am I in trouble? I'm guessing this drive is dead?
Code:
root@proxmox02:~# dmesg | grep -E "sda|sdb"
[ 1.941969] sd 3:0:0:0: [sda] 1875385008 512-byte logical blocks: (960 GB/894 GiB)
[ 1.941972] sd 3:0:0:0: [sda] 4096-byte physical blocks
[ 1.941980] sd 3:0:0:0: [sda] Write Protect is off
[ 1.941981] sd 3:0:0:0: [sda] Mode Sense: 00 3a 00 00
[ 1.942041] sd 3:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA
[ 1.942083] sd 3:0:0:0: [sda] Preferred minimum I/O size 4096 bytes
[ 1.958554] sda: sda1 sda9
[ 1.958641] sd 3:0:0:0: [sda] Attached SCSI disk
Last edited: