[SOLVED] Mirrored boot pool degraded due to PVE not recognizing one of the drives though lsblk recognizes both the drives.

Alongside4550

New Member
Jan 9, 2026
3
1
3
I have my PVE boot drive setup as mirrored zfs pool. My boot pool (rpool) is showing as degraded and the "Disks" panel is showing only one NVMe drive while LSBLK is showing both the drives (see output below) . Both these are new drives and same brand/model. Recent SMART report didn't show any issues. I am new to Proxmox so can someone help me troubleshoot this. I have only two M.2 slots in my motherboard and both are populated with these two drives.

ZFS Status.png

1767975355642.png

Output of LSBLK. Note the line nvme1n1 259:4 0 0B 0 disk

Code:
nvme0n1     259:0    0 953.9G  0 disk
├─nvme0n1p1 259:1    0  1007K  0 part
├─nvme0n1p2 259:2    0     1G  0 part
└─nvme0n1p3 259:3    0   952G  0 part
nvme1n1     259:4    0     0B  0 disk
├─nvme1n1p1 259:5    0  1007K  0 part
├─nvme1n1p2 259:6    0     1G  0 part
└─nvme1n1p3 259:7    0   952G  0 part

Output of proxmox-boot-tool status

Code:
root@pve:~# proxmox-boot-tool status
Re-executing '/usr/sbin/proxmox-boot-tool' in new private mount namespace..
System currently booted with uefi
2E41-A85D is configured with: uefi (versions: 6.14.11-5-pve, 6.17.2-2-pve, 6.17.4-2-pve)
mount: /var/tmp/espmounts/2E41-D98C: can't read superblock on /dev/nvme1n1p2.
       dmesg(1) may have more information after failed mount system call.
mount of /dev/disk/by-uuid/2E41-D98C failed - skipping

Output of efibootmgr -v

Code:
root@pve:~# efibootmgr -v
BootCurrent: 0003
Timeout: 1 seconds
BootOrder: 0003,0002,0004,0005
Boot0002* Linux Boot Manager    HD(2,GPT,251009e7-49fb-4bc4-9c1f-b2add098686c,0x800,0x200000)/File(\EFI\systemd\systemd-bootx64.efi)
      dp: 04 01 2a 00 02 00 00 00 00 08 00 00 00 00 00 00 00 00 20 00 00 00 00 00 e7 09 10 25 fb 49 c4 4b 9c 1f b2 ad d0 98 68 6c 02 02 / 04 04 46 00 5c 00 45 00 46 00 49 00 5c 00 73 00 79 00 73 00 74 00 65 00 6d 00 64 00 5c 00 73 00 79 00 73 00 74 00 65 00 6d 00 64 00 2d 00 62 00 6f 00 6f 00 74 00 78 00 36 00 34 00 2e 00 65 00 66 00 69 00 00 00 / 7f ff 04 00
Boot0003* Linux Boot Manager    HD(2,GPT,be3306b2-82b0-4845-9c7b-97eef0a2289e,0x800,0x200000)/File(\EFI\systemd\systemd-bootx64.efi)
      dp: 04 01 2a 00 02 00 00 00 00 08 00 00 00 00 00 00 00 00 20 00 00 00 00 00 b2 06 33 be b0 82 45 48 9c 7b 97 ee f0 a2 28 9e 02 02 / 04 04 46 00 5c 00 45 00 46 00 49 00 5c 00 73 00 79 00 73 00 74 00 65 00 6d 00 64 00 5c 00 73 00 79 00 73 00 74 00 65 00 6d 00 64 00 2d 00 62 00 6f 00 6f 00 74 00 78 00 36 00 34 00 2e 00 65 00 66 00 69 00 00 00 / 7f ff 04 00
Boot0004* UEFI OS       HD(2,GPT,be3306b2-82b0-4845-9c7b-97eef0a2289e,0x800,0x200000)/File(\EFI\BOOT\BOOTX64.EFI)0000424f
      dp: 04 01 2a 00 02 00 00 00 00 08 00 00 00 00 00 00 00 00 20 00 00 00 00 00 b2 06 33 be b0 82 45 48 9c 7b 97 ee f0 a2 28 9e 02 02 / 04 04 30 00 5c 00 45 00 46 00 49 00 5c 00 42 00 4f 00 4f 00 54 00 5c 00 42 00 4f 00 4f 00 54 00 58 00 36 00 34 00 2e 00 45 00 46 00 49 00 00 00 / 7f ff 04 00
    data: 00 00 42 4f
Boot0005* UEFI OS       HD(2,GPT,251009e7-49fb-4bc4-9c1f-b2add098686c,0x800,0x200000)/File(\EFI\BOOT\BOOTX64.EFI)0000424f
      dp: 04 01 2a 00 02 00 00 00 00 08 00 00 00 00 00 00 00 00 20 00 00 00 00 00 e7 09 10 25 fb 49 c4 4b 9c 1f b2 ad d0 98 68 6c 02 02 / 04 04 30 00 5c 00 45 00 46 00 49 00 5c 00 42 00 4f 00 4f 00 54 00 5c 00 42 00 4f 00 4f 00 54 00 58 00 36 00 34 00 2e 00 45 00 46 00 49 00 00 00 / 7f ff 04 00
    data: 00 00 42 4f
 
Good day Alongside4550,

I am fairly new here as well, but wondered if it could be as simple as trying to follow the suggested command the Status Rpool is suggesting of "Online the device using zpool online"

I apologize if I have misread your post, but have you tried using that command?
Perhaps something like this?
sudo zpool online <pool_name> <device_path>
 
I did try zpool online command but it didn't help. Planning to just reformat that drive and do zpool replace if there are no other better ways.
 
I noticed that the disk shows as 0B. What does nvme list and nvme smart-log /dev/nvme1n1 and journalctl -b -k say?
 
if the disk shows a 0 byte, it has a very high chance of beeing dead.

you can install the nvme-cli package and run nvme list to give output like this:

Code:
root@pve2:~# nvme list
Node                  Generic               SN                   Model                                    Namespace  Usage                      Format           FW Rev
--------------------- --------------------- -------------------- ---------------------------------------- ---------- -------------------------- ---------------- --------
/dev/nvme0n1          /dev/ng0n1            AA000000000000000686 XrayDisk 1TB SSD                         0x1          1.00  TB /   1.00  TB    512   B +  0 B   VC2S038E
/dev/nvme1n1          /dev/ng1n1            FXS500PRO230207362   Fanxiang S500PRO 1TB                     0x1          1.02  TB /   1.02  TB    512   B +  0 B   SN09273

would be interesting to see yours. i fear that the drive will show up as 0 bytes from here as well.

another good command is nvme smart-log /dev/nvme1

it will show output like this:

Code:
root@pve2:~# nvme smart-log /dev/nvme0
Smart Log for NVME device:nvme0 namespace-id:ffffffff
critical_warning                        : 0
temperature                             : 129 °F (327 K)
available_spare                         : 100%
available_spare_threshold               : 32%
percentage_used                         : 0%
endurance group critical warning summary: 0
Data Units Read                         : 25645508 (13.13 TB)
Data Units Written                      : 13425922 (6.87 TB)
host_read_commands                      : 212581494
host_write_commands                     : 300289109
controller_busy_time                    : 0
power_cycles                            : 117
power_on_hours                          : 5948
unsafe_shutdowns                        : 63
media_errors                            : 0
num_err_log_entries                     : 0
Warning Temperature Time                : 0
Critical Composite Temperature Time     : 0
Thermal Management T1 Trans Count       : 0
Thermal Management T2 Trans Count       : 0
Thermal Management T1 Total Time        : 0
Thermal Management T2 Total Time        : 0

might give us more of an idea if the drive is actually dead.

edit: completely overlooked that @Impact had basically written the same thing already :-/
 
Last edited: