Every time i try to start this VM after rebooting the PVE host i get
Everything is updated.
I have another win10 VM and that one starts without any issue. (Disk is on
I suspect it have something to do with
Each time I have to restore from backup and then it will start. (Restore on the same storage pool)
I searched everywhere but I couldn't fine any answer. I will try to attach as much info as possible here:
TASK ERROR: timeout: no zvol device link for 'vm-205-disk-0' found after 300 sec found.
Everything is updated.
I have another win10 VM and that one starts without any issue. (Disk is on
rpool
)I suspect it have something to do with
nvme-pool
zpool put I scrubbed it and I found nothing wrong. Each time I have to restore from backup and then it will start. (Restore on the same storage pool)
zvol_wait
Loops saying it's waiting for a zfs link forever.I searched everywhere but I couldn't fine any answer. I will try to attach as much info as possible here:
Code:
root@proxmox:~# cat /etc/pve/qemu-server/205.conf
agent: 1,fstrim_cloned_disks=1
balloon: 1024
boot: order=scsi0;net0
cores: 4
cpu: host
machine: q35
memory: 4096
name: Ubuntu
net0: virtio=<REMOVED>,bridge=vmbr0,tag=30
numa: 0
ostype: l26
scsi0: nvme-pool:vm-205-disk-0,size=32G
scsihw: virtio-scsi-pci
smbios1: uuid=<REMOVED>
sockets: 1
vga: qxl
vmgenid: <REMOVED>
Code:
root@proxmox:~# zfs get all nvme-pool
NAME PROPERTY VALUE SOURCE
nvme-pool type filesystem -
nvme-pool creation Thu Mar 3 0:10 2022 -
nvme-pool used 320G -
nvme-pool available 130G -
nvme-pool referenced 287G -
nvme-pool compressratio 1.07x -
nvme-pool mounted yes -
nvme-pool quota none default
nvme-pool reservation none default
nvme-pool recordsize 128K default
nvme-pool mountpoint /nvme-pool default
nvme-pool sharenfs off default
nvme-pool checksum on default
nvme-pool compression zstd local
nvme-pool atime off local
nvme-pool devices on default
nvme-pool exec on default
nvme-pool setuid on default
nvme-pool readonly off default
nvme-pool zoned off default
nvme-pool snapdir hidden default
nvme-pool aclmode discard default
nvme-pool aclinherit restricted default
nvme-pool createtxg 1 -
nvme-pool canmount on default
nvme-pool xattr on default
nvme-pool copies 1 default
nvme-pool version 5 -
nvme-pool utf8only off -
nvme-pool normalization none -
nvme-pool casesensitivity sensitive -
nvme-pool vscan off default
nvme-pool nbmand off default
nvme-pool sharesmb off default
nvme-pool refquota none default
nvme-pool refreservation none default
nvme-pool guid <REMOVED> -
nvme-pool primarycache all default
nvme-pool secondarycache all default
nvme-pool usedbysnapshots 0B -
nvme-pool usedbydataset 287G -
nvme-pool usedbychildren 33.1G -
nvme-pool usedbyrefreservation 0B -
nvme-pool logbias latency default
nvme-pool objsetid 54 -
nvme-pool dedup off default
nvme-pool mlslabel none default
nvme-pool sync standard default
nvme-pool dnodesize legacy default
nvme-pool refcompressratio 1.05x -
nvme-pool written 287G -
nvme-pool logicalused 333G -
nvme-pool logicalreferenced 303G -
nvme-pool volmode default default
nvme-pool filesystem_limit none default
nvme-pool snapshot_limit none default
nvme-pool filesystem_count none default
nvme-pool snapshot_count none default
nvme-pool snapdev hidden default
nvme-pool acltype off default
nvme-pool context none default
nvme-pool fscontext none default
nvme-pool defcontext none default
nvme-pool rootcontext none default
nvme-pool relatime off default
nvme-pool redundant_metadata all default
nvme-pool overlay on default
nvme-pool encryption off default
nvme-pool keylocation none default
nvme-pool keyformat none default
nvme-pool pbkdf2iters 0 default
nvme-pool special_small_blocks 0 default
nvme-pool com.sun:auto-snapshot false local
Code:
root@proxmox:~# journalctl -b
-- Journal begins at Wed 2022-03-02 23:34:40 CET, ends at Mon 2022-08-08 00:59:40 CEST. --
Aug 08 00:32:20 proxmox kernel: Linux version 5.15.39-3-pve (build@proxmox) (gcc (Debian 10.2.1-6) 10.2.1 20210110, GNU ld (GNU Binutils for >
Aug 08 00:32:20 proxmox kernel: Command line: initrd=\EFI\proxmox\5.15.39-3-pve\initrd.img-5.15.39-3-pve root=ZFS=rpool/ROOT/pve-1 boot=zfs
Aug 08 00:32:20 proxmox kernel: KERNEL supported cpus:
Aug 08 00:32:20 proxmox kernel: Intel GenuineIntel
Aug 08 00:32:20 proxmox kernel: AMD AuthenticAMD
Aug 08 00:32:20 proxmox kernel: Hygon HygonGenuine
Aug 08 00:32:20 proxmox kernel: Centaur CentaurHauls
Aug 08 00:32:20 proxmox kernel: zhaoxin Shanghai
Aug 08 00:32:20 proxmox kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers'
Aug 08 00:32:20 proxmox kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers'
Aug 08 00:32:20 proxmox kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers'
Aug 08 00:32:20 proxmox kernel: x86/fpu: Supporting XSAVE feature 0x008: 'MPX bounds registers'
Aug 08 00:32:20 proxmox kernel: x86/fpu: Supporting XSAVE feature 0x010: 'MPX CSR'
Aug 08 00:32:20 proxmox kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256
Aug 08 00:32:20 proxmox kernel: x86/fpu: xstate_offset[3]: 832, xstate_sizes[3]: 64
Aug 08 00:32:20 proxmox kernel: x86/fpu: xstate_offset[4]: 896, xstate_sizes[4]: 64
Aug 08 00:32:20 proxmox kernel: x86/fpu: Enabled xstate features 0x1f, context size is 960 bytes, using 'compacted' format.
Aug 08 00:32:20 proxmox kernel: signal: max sigframe size: 2032
Aug 08 00:32:20 proxmox kernel: BIOS-provided physical RAM map:
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009dfff] usable
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x000000000009e000-0x000000000009efff] reserved
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x000000000009f000-0x000000000009ffff] usable
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x00000000000a0000-0x00000000000fffff] reserved
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000885fefff] usable
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x00000000885ff000-0x0000000088bfefff] reserved
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x0000000088bff000-0x0000000088cfefff] ACPI data
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x0000000088cff000-0x0000000088efefff] ACPI NVS
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x0000000088eff000-0x0000000089bfefff] reserved
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x0000000089bff000-0x0000000089bfffff] usable
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x0000000089c00000-0x000000008f7fffff] reserved
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x00000000e0000000-0x00000000efffffff] reserved
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x00000000fe000000-0x00000000fe010fff] reserved
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x00000000fec00000-0x00000000fec00fff] reserved
Aug 08 00:32:20 proxmox kernel: BIOS-e820: [mem 0x00000000fed00000-0x00000000fed00fff] reserved
Code:
root@proxmox:~# zpool status
pool: BACKUP
state: ONLINE
scan: scrub repaired 0B in 01:20:22 with 0 errors on Sun Jul 10 01:44:25 2022
config:
NAME STATE READ WRITE CKSUM
BACKUP ONLINE 0 0 0
ata-TOSHIBA_MQ<REMOVED>_<REMOVED> ONLINE 0 0 0
errors: No known data errors
pool: RAIDZ12TB
state: ONLINE
scan: scrub repaired 0B in 04:01:52 with 0 errors on Sun Jul 10 04:25:59 2022
config:
NAME STATE READ WRITE CKSUM
RAIDZ12TB ONLINE 0 0 0
raidz1-0 ONLINE 0 0 0
ata-ST4000VN008-<REMOVED> ONLINE 0 0 0
ata-ST4000VN008-<REMOVED> ONLINE 0 0 0
ata-ST4000VN008-<REMOVED> ONLINE 0 0 0
ata-ST4000VN008-<REMOVED> ONLINE 0 0 0
errors: No known data errors
pool: SINGLE1TB
state: ONLINE
scan: scrub repaired 0B in 02:51:01 with 0 errors on Sun Jul 10 03:15:14 2022
config:
NAME STATE READ WRITE CKSUM
SINGLE1TB ONLINE 0 0 0
ata-HGST_HTS54<REMOVED>_JA<REMOVED> ONLINE 0 0 0
errors: No known data errors
pool: SINGLE4TB
state: ONLINE
scan: scrub repaired 0B in 06:27:10 with 0 errors on Sun Jul 10 06:51:26 2022
config:
NAME STATE READ WRITE CKSUM
SINGLE4TB ONLINE 0 0 0
ata-WDC_WD40EZAZ-<REMOVED>_WD-WX32<REMOVED> ONLINE 0 0 0
errors: No known data errors
pool: nvme-pool
state: ONLINE
scan: scrub repaired 0B in 00:02:59 with 0 errors on Tue Jul 26 20:48:31 2022
config:
NAME STATE READ WRITE CKSUM
nvme-pool ONLINE 0 0 0
nvme-WDC_WDS500G2B0C-00<REMOVED> ONLINE 0 0 0
errors: No known data errors
pool: rpool
state: ONLINE
scan: scrub repaired 0B in 00:04:24 with 0 errors on Tue Jul 26 20:49:45 2022
config:
NAME STATE READ WRITE CKSUM
rpool ONLINE 0 0 0
nvme-eui.1911<REMOVED>3f00-part3 ONLINE 0 0 0
errors: No known data errors
Code:
root@proxmox:~# pveversion -v
proxmox-ve: 7.2-1 (running kernel: 5.15.39-3-pve)
pve-manager: 7.2-7 (running version: 7.2-7/d0dd0e85)
pve-kernel-5.15: 7.2-8
pve-kernel-helper: 7.2-8
pve-kernel-5.13: 7.1-9
pve-kernel-5.15.39-3-pve: 5.15.39-3
pve-kernel-5.15.39-2-pve: 5.15.39-2
pve-kernel-5.15.39-1-pve: 5.15.39-1
pve-kernel-5.13.19-6-pve: 5.13.19-15
pve-kernel-5.13.19-2-pve: 5.13.19-4
ceph-fuse: 15.2.15-pve1
corosync: 3.1.5-pve2
criu: 3.15-1+pve-1
glusterfs-client: 9.2-1
ifupdown2: 3.1.0-1+pmx3
ksm-control-daemon: 1.4-1
libjs-extjs: 7.0.0-1
libknet1: 1.24-pve1
libproxmox-acme-perl: 1.4.2
libproxmox-backup-qemu0: 1.3.1-1
libpve-access-control: 7.2-4
libpve-apiclient-perl: 3.2-1
libpve-common-perl: 7.2-2
libpve-guest-common-perl: 4.1-2
libpve-http-server-perl: 4.1-3
libpve-storage-perl: 7.2-7
libspice-server1: 0.14.3-2.1
lvm2: 2.03.11-2.1
lxc-pve: 5.0.0-3
lxcfs: 4.0.12-pve1
novnc-pve: 1.3.0-3
proxmox-backup-client: 2.2.5-1
proxmox-backup-file-restore: 2.2.5-1
proxmox-mini-journalreader: 1.3-1
proxmox-widget-toolkit: 3.5.1
pve-cluster: 7.2-2
pve-container: 4.2-2
pve-docs: 7.2-2
pve-edk2-firmware: 3.20210831-2
pve-firewall: 4.2-5
pve-firmware: 3.5-1
pve-ha-manager: 3.4.0
pve-i18n: 2.7-2
pve-qemu-kvm: 6.2.0-11
pve-xtermjs: 4.16.0-1
qemu-server: 7.2-3
smartmontools: 7.2-pve3
spiceterm: 3.2-2
swtpm: 0.7.1~bpo11+1
vncterm: 1.7-1
zfsutils-linux: 2.1.5-pve1