I do have a ZFS pool (R1_1.6TB_SSD_EVO860) where I store my vm os disk images.
Recently, after I had to reboot the host, some vm were not starting (while others from the same zfs pool did) and I got the following log events in syslog (see below).
"timeout: no zvol device link for 'vm-1002-disk-0' found after 300 sec found."
The thing is I can see the disk images in the gui and some vms from the same zpool are booting without any issue.
I tried rebooting again and some times one of the vm will find it's way mount the disks some other times it wont.
The only way I could "workaround" this problem is by detaching the problematic disks from the vm and re-attach them. After that the vm will boot.
But that workaround does not seem to be persistent after a reboot.
The only difference the problematic vm had over the time was that a few weeks ago, I had to migrate temporarly the disk images to another zfs pool before migrating them back. At the time, everything went smoothly.
Anybody had the same problems ? ideas ?
Physical disks seems to be sane and I didn't see any zfs error
In the logs below, the problematic disk image is "R1_1.6TB_SSD_EVO860/vm-1002-disk-0"
Recently, after I had to reboot the host, some vm were not starting (while others from the same zfs pool did) and I got the following log events in syslog (see below).
"timeout: no zvol device link for 'vm-1002-disk-0' found after 300 sec found."
The thing is I can see the disk images in the gui and some vms from the same zpool are booting without any issue.
I tried rebooting again and some times one of the vm will find it's way mount the disks some other times it wont.
The only way I could "workaround" this problem is by detaching the problematic disks from the vm and re-attach them. After that the vm will boot.
But that workaround does not seem to be persistent after a reboot.
The only difference the problematic vm had over the time was that a few weeks ago, I had to migrate temporarly the disk images to another zfs pool before migrating them back. At the time, everything went smoothly.
Anybody had the same problems ? ideas ?
Physical disks seems to be sane and I didn't see any zfs error
In the logs below, the problematic disk image is "R1_1.6TB_SSD_EVO860/vm-1002-disk-0"
tail -f /var/log/syslog
Mar 11 06:25:40 pve zvol_wait[4215]: R1_1.6TB_SSD_EVO860/vm-1002-disk-0
Mar 11 06:25:40 pve zvol_wait[4215]: R1_1.6TB_SSD_EVO860/vm-1004-disk-0
Mar 11 06:25:40 pve zvol_wait[4215]: RZ2-2_5-8_2TB/vm-1002-disk-0
Mar 11 06:26:10 pve zvol_wait[4215]: Still waiting on 3 zvol links ...
Mar 11 06:26:10 pve zvol_wait[4215]: No progress since last loop.
Mar 11 06:26:10 pve zvol_wait[4215]: Checking if any zvols were deleted.
Mar 11 06:26:10 pve zvol_wait[4215]: Remaining zvols:
Mar 11 06:26:10 pve zvol_wait[4215]: R1_1.6TB_SSD_EVO860/vm-1002-disk-0
Mar 11 06:26:10 pve zvol_wait[4215]: R1_1.6TB_SSD_EVO860/vm-1004-disk-0
Mar 11 06:26:10 pve zvol_wait[4215]: RZ2-2_5-8_2TB/vm-1002-disk-0
Mar 11 06:27:19 pve pve-guests[16140]: timeout: no zvol device link for 'vm-1002-disk-0' found after 300 sec found.
zfs get mounted,mountpoint,canmount
NAME PROPERTY VALUE SOURCE
R1_1.6TB_SSD_EVO860 mounted yes -
R1_1.6TB_SSD_EVO860 mountpoint /R1_1.6TB_SSD_EVO860 default
R1_1.6TB_SSD_EVO860 canmount on default
R1_1.6TB_SSD_EVO860/base-2000-disk-0 mounted - -
R1_1.6TB_SSD_EVO860/base-2000-disk-0 mountpoint - -
R1_1.6TB_SSD_EVO860/base-2000-disk-0 canmount - -
R1_1.6TB_SSD_EVO860/base-2000-disk-0@__base__ mounted - -
R1_1.6TB_SSD_EVO860/base-2000-disk-0@__base__ mountpoint - -
R1_1.6TB_SSD_EVO860/base-2000-disk-0@__base__ canmount - -
R1_1.6TB_SSD_EVO860/subvol-100-disk-0 mounted yes -
R1_1.6TB_SSD_EVO860/subvol-100-disk-0 mountpoint /R1_1.6TB_SSD_EVO860/subvol-100-disk-0 default
R1_1.6TB_SSD_EVO860/subvol-100-disk-0 canmount on default
R1_1.6TB_SSD_EVO860/subvol-101-disk-0 mounted yes -
R1_1.6TB_SSD_EVO860/subvol-101-disk-0 mountpoint /R1_1.6TB_SSD_EVO860/subvol-101-disk-0 default
R1_1.6TB_SSD_EVO860/subvol-101-disk-0 canmount on default
R1_1.6TB_SSD_EVO860/vm-1001-disk-0 mounted - -
R1_1.6TB_SSD_EVO860/vm-1001-disk-0 mountpoint - -
R1_1.6TB_SSD_EVO860/vm-1001-disk-0 canmount - -
R1_1.6TB_SSD_EVO860/vm-1002-disk-0 mounted - -
R1_1.6TB_SSD_EVO860/vm-1002-disk-0 mountpoint - -
R1_1.6TB_SSD_EVO860/vm-1002-disk-0 canmount - -
R1_1.6TB_SSD_EVO860/vm-1003-disk-0 mounted - -
R1_1.6TB_SSD_EVO860/vm-1003-disk-0 mountpoint - -
R1_1.6TB_SSD_EVO860/vm-1003-disk-0 canmount - -
R1_1.6TB_SSD_EVO860/vm-1004-disk-0 mounted - -
R1_1.6TB_SSD_EVO860/vm-1004-disk-0 mountpoint - -
R1_1.6TB_SSD_EVO860/vm-1004-disk-0 canmount - -
R1_1.6TB_SSD_EVO860/vm-1004-disk-1 mounted - -
R1_1.6TB_SSD_EVO860/vm-1004-disk-1 mountpoint - -
R1_1.6TB_SSD_EVO860/vm-1004-disk-1 canmount - -
R1_1.6TB_SSD_EVO860/vm-1005-disk-0 mounted - -
R1_1.6TB_SSD_EVO860/vm-1005-disk-0 mountpoint - -
R1_1.6TB_SSD_EVO860/vm-1005-disk-0 canmount - -
R1_1.6TB_SSD_EVO860/vm-102-disk-0 mounted - -
R1_1.6TB_SSD_EVO860/vm-102-disk-0 mountpoint - -
R1_1.6TB_SSD_EVO860/vm-102-disk-0 canmount - -
zfs list
NAME USED AVAIL REFER MOUNTPOINT
R1_1.6TB_SSD_EVO860 273G 1.31T 96K /R1_1.6TB_SSD_EVO860
R1_1.6TB_SSD_EVO860/base-2000-disk-0 57.9G 1.35T 16.6G -
R1_1.6TB_SSD_EVO860/subvol-100-disk-0 1.19G 830M 1.19G /R1_1.6TB_SSD_EVO860/subvol-100-disk-0
R1_1.6TB_SSD_EVO860/subvol-101-disk-0 1.22G 6.78G 1.22G /R1_1.6TB_SSD_EVO860/subvol-101-disk-0
R1_1.6TB_SSD_EVO860/vm-1001-disk-0 33.0G 1.31T 30.6G -
R1_1.6TB_SSD_EVO860/vm-1002-disk-0 61.9G 1.34T 33.8G -
R1_1.6TB_SSD_EVO860/vm-1003-disk-0 33.0G 1.34T 4.59G -
R1_1.6TB_SSD_EVO860/vm-1004-disk-0 3M 1.31T 192K -
R1_1.6TB_SSD_EVO860/vm-1004-disk-1 33.0G 1.33T 8.42G -
R1_1.6TB_SSD_EVO860/vm-1005-disk-0 10.3G 1.31T 4.72G -
R1_1.6TB_SSD_EVO860/vm-102-disk-0 41.3G 1.33T 16.4G -
RZ2-1_1-4_4TB 419G 6.51T 140K /RZ2-1_1-4_4TB
RZ2-1_1-4_4TB/vm-1003-disk-0 55.3G 6.53T 36.7G -
RZ2-1_1-4_4TB/vm-1003-disk-1 363G 6.51T 363G -
RZ2-2_5-8_2TB 2.16T 1.25T 140K /RZ2-2_5-8_2TB
RZ2-2_5-8_2TB/vm-1002-disk-0 2.16T 1.94T 1.47T -
find /dev | grep 1002
/dev/RZ2-2_5-8_2TB/vm-1002-disk-0-part2
/dev/RZ2-2_5-8_2TB/vm-1002-disk-0-part1
/dev/RZ2-2_5-8_2TB/vm-1002-disk-0
/dev/R1_1.6TB_SSD_EVO860/vm-1002-disk-0-part1
/dev/R1_1.6TB_SSD_EVO860/vm-1002-disk-0
/dev/R1_1.6TB_SSD_EVO860/vm-1002-disk-0-part3
/dev/R1_1.6TB_SSD_EVO860/vm-1002-disk-0-part2
/dev/zvol/RZ2-2_5-8_2TB/vm-1002-disk-0-part2
/dev/zvol/RZ2-2_5-8_2TB/vm-1002-disk-0-part1
/dev/zvol/RZ2-2_5-8_2TB/vm-1002-disk-0
/dev/zvol/R1_1.6TB_SSD_EVO860/vm-1002-disk-0-part1
/dev/zvol/R1_1.6TB_SSD_EVO860/vm-1002-disk-0
/dev/zvol/R1_1.6TB_SSD_EVO860/vm-1002-disk-0-part3
/dev/zvol/R1_1.6TB_SSD_EVO860/vm-1002-disk-0-part2
zpool list
NAME SIZE ALLOC FREE CKPOINT EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT
R1_1.6TB_SSD_EVO860 1.62T 118G 1.51T - - 16% 7% 1.00x ONLINE -
RZ2-1_1-4_4TB 14.5T 826G 13.7T - - 0% 5% 1.00x ONLINE -
RZ2-2_5-8_2TB 7.27T 3.04T 4.23T - - 2% 41% 1.00x ONLINE -
zpool status -v R1_1.6TB_SSD_EVO860
pool: R1_1.6TB_SSD_EVO860
state: ONLINE
scan: scrub repaired 0B in 00:02:40 with 0 errors on Sun Feb 13 00:26:41 2022
config:
NAME STATE READ WRITE CKSUM
R1_1.6TB_SSD_EVO860 ONLINE 0 0 0
mirror-0 ONLINE 0 0 0
ata-Samsung_SSD_860_EVO_2TB_S597NJ0NB19827A ONLINE 0 0 0
ata-Samsung_SSD_860_EVO_2TB_S597NJ0NB19834W ONLINE 0 0 0
errors: No known data errors
nano /etc/pve/qemu-server/1002.conf
agent: 1
bios: ovmf
boot: order=scsi0
cores: 16
hotplug: disk,network,usb,memory,cpu
machine: pc-q35-6.0
memory: 16384
name: BlueIris
net0: virtio=HIDDEN,bridge=vmbr0
numa: 1
onboot: 1
ostype: win10
scsi0: R1_1.6TB_SSD_EVO860:vm-1002-disk-0,size=60G
scsi1: RZ2-2_5-8_2TB:vm-1002-disk-0,backup=0,size=2000G
scsihw: virtio-scsi-pci
smbios1: uuid=HIDDEN
sockets: 1
startup: order=3,up=0
vga: memory=8
vmgenid: HIDDEN