Snapshots error - qmp command 'blockdev-snapshot-internal-sync' failed - Cannot find device='drive-tpmstate0'

matto

New Member
Dec 19, 2024
10
0
1
Hi All,

Hope you are having a great day!

I have today come across an interesting error with snapshots on our proxmox cluster. There were updates on the cluster overnight and i fear this has caused this issue.

Currently whenever we attempt to snapshot a VM whilst it is running, the following error happens after the "CLEANUP" line in the GUI: qmp command 'blockdev-snapshot-internal-sync' failed - Cannot find device='drive-tpmstate0' nor node-name='drive-tpmstate0'

This error is replicable on all of our VM's (approx 70ish) no matter their age.

Interestingly enough if the VM is not running (Shutdown or Stopped) then the snapshot succeeds with no issues.

Has anyone ever come across this issue before and know what it might be?

Many thanks for any and all help!

I will put the pve versions / vm configs etc down below:

PVEVERSION -V

Bash:
root@S-1121:~# pveversion -v
proxmox-ve: 9.1.0 (running kernel: 6.17.9-1-pve)
pve-manager: 9.1.9 (running version: 9.1.9/ee7bad0a3d1546c9)
proxmox-kernel-helper: 9.0.4
proxmox-kernel-7.0: 7.0.0-3
proxmox-kernel-7.0.0-3-pve-signed: 7.0.0-3
proxmox-kernel-6.17: 6.17.13-6
proxmox-kernel-6.17.13-6-pve-signed: 6.17.13-6
proxmox-kernel-6.17.9-1-pve-signed: 6.17.9-1
proxmox-kernel-6.8: 6.8.12-18
proxmox-kernel-6.8.12-18-pve-signed: 6.8.12-18
proxmox-kernel-6.8.4-2-pve-signed: 6.8.4-2
ceph: 20.2.1-pve1
ceph-fuse: 20.2.1-pve1
corosync: 3.1.10-pve2
criu: 4.1.1-1
dnsmasq: 2.91-1
frr-pythontools: 10.4.1-1+pve1
ifupdown2: 3.3.0-1+pmx12
ksm-control-daemon: 1.5-1
libjs-extjs: 7.0.0-5
libproxmox-acme-perl: 1.7.1
libproxmox-backup-qemu0: 2.0.2
libproxmox-rs-perl: 0.4.1
libpve-access-control: 9.0.7
libpve-apiclient-perl: 3.4.2
libpve-cluster-api-perl: 9.1.2
libpve-cluster-perl: 9.1.2
libpve-common-perl: 9.1.11
libpve-guest-common-perl: 6.0.2
libpve-http-server-perl: 6.0.5
libpve-network-perl: 1.3.0
libpve-notify-perl: 9.1.2
libpve-rs-perl: 0.13.0
libpve-storage-perl: 9.1.2
libspice-server1: 0.15.2-1+b1
lvm2: 2.03.31-2+pmx1
lxc-pve: 6.0.5-4
lxcfs: 6.0.4-pve1
novnc-pve: 1.6.0-4
openvswitch-switch: 3.5.0-1+b1
proxmox-backup-client: 4.2.0-1
proxmox-backup-file-restore: 4.2.0-1
proxmox-backup-restore-image: 1.0.0
proxmox-firewall: 1.2.2
proxmox-kernel-helper: 9.0.4
proxmox-mail-forward: 1.0.3
proxmox-mini-journalreader: 1.6
proxmox-offline-mirror-helper: 0.7.3
proxmox-widget-toolkit: 5.1.9
pve-cluster: 9.1.2
pve-container: 6.1.5
pve-docs: 9.1.2
pve-edk2-firmware: 4.2025.05-2
pve-esxi-import-tools: 1.0.1
pve-firewall: 6.0.4
pve-firmware: 3.18-3
pve-ha-manager: 5.2.0
pve-i18n: 3.7.1
pve-qemu-kvm: 10.1.2-7
pve-xtermjs: 5.5.0-3
qemu-server: 9.1.9
smartmontools: 7.4-pve1
spiceterm: 3.4.2
swtpm: 0.8.0+pve3
vncterm: 1.9.2
zfsutils-linux: 2.4.1-pve1

PVE STORAGE CONFIGURATION

Bash:
root@S-1121:~# cat /etc/pve/storage.cfg
dir: local
        disable
        path /var/lib/vz
        content vztmpl,iso,backup
        shared 0

zfspool: local-zfs
        disable
        pool rpool/data
        content images,rootdir
        sparse 0

rbd: ceph-nvme
        content rootdir,images
        krbd 0
        pool ceph-nvme

cephfs: ceph-isostore
        path /mnt/pve/ceph-isostore
        content vztmpl,backup,iso
        fs-name ceph-isostore

pbs: PBS
        datastore Backups
        server [REDACTED]
        content backup
        fingerprint [REDACTED]
        prune-backups keep-all=1
        username [REDACTED]

VM CONFIGURATION

Bash:
agent: 1
bios: ovmf
boot: order=sata0;ide0
cores: 12
cpu: host
efidisk0: ceph-nvme:vm-517-disk-0,efitype=4m,pre-enrolled-keys=1,size=1M
ide0: none,media=cdrom
machine: pc-q35-9.2+pve1
memory: 49152
meta: creation-qemu=9.2.0,ctime=1750265172
name: DMZ-4-UAT-AWS
net0: virtio=BC:24:11:9D:8E:E4,bridge=vmbr2,firewall=1,tag=3569
numa: 0
ostype: win11
parent: Burgas_HF7_260427
sata0: ceph-nvme:vm-517-disk-1,discard=on,size=80G,ssd=1
scsihw: virtio-scsi-single
smbios1: uuid=0356df6a-d2d1-40f2-9466-fded870edb7f
sockets: 1
tpmstate0: ceph-nvme:vm-517-disk-4,size=4M,version=v2.0
unused0: ceph-nvme:vm-517-disk-2
vga: virtio,memory=16
virtio1: ceph-nvme:vm-517-disk-3,iothread=1,size=80G
vmgenid: 23526ef7-3bf4-496b-9892-1ea6ba9bbeb8

PS OF ONE OF THE EFFECTED MACHINES

Bash:
root@S-1121:~# ps -ax | grep -- 'id 517'
 807333 ?        Sl    32:23 /usr/bin/kvm -id 517 -name DMZ-4-UAT-AWS,debug-threads=on -no-shutdown -chardev socket,id=qmp,path=/var/run/qemu-server/517.qmp,server=on,wait=off -mon chardev=qmp,mode=control -chardev socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect-ms=5000 -mon chardev=qmp-event,mode=control -pidfile /var/run/qemu-server/517.pid -daemonize -smbios type=1,uuid=0356df6a-d2d1-40f2-9466-fded870edb7f -drive if=pflash,unit=0,format=raw,readonly=on,file=/usr/share/pve-edk2-firmware//OVMF_CODE_4M.secboot.fd -drive if=pflash,unit=1,id=drive-efidisk0,cache=writeback,format=raw,file=rbd:ceph-nvme/vm-517-disk-0:conf=/etc/pve/ceph.conf:id=admin:keyring=/etc/pve/priv/ceph/ceph-nvme.keyring:rbd_cache_policy=writeback,size=540672 -smp 12,sockets=1,cores=12,maxcpus=12 -nodefaults -boot menu=on,strict=on,reboot-timeout=1000,splash=/usr/share/qemu-server/bootsplash.jpg -vnc unix:/var/run/qemu-server/517.vnc,password=on -global kvm-pit.lost_tick_policy=discard -cpu host,hv_ipi,hv_relaxed,hv_reset,hv_runtime,hv_spinlocks=0x1fff,hv_stimer,hv_synic,hv_time,hv_vapic,hv_vpindex,+kvm_pv_eoi,+kvm_pv_unhalt -m 49152 -object iothread,id=iothread-virtio1 -global ICH9-LPC.disable_s3=1 -global ICH9-LPC.disable_s4=1 -readconfig /usr/share/qemu-server/pve-q35-4.0.cfg -device vmgenid,guid=23526ef7-3bf4-496b-9892-1ea6ba9bbeb8 -device usb-tablet,id=tablet,bus=ehci.0,port=1 -chardev socket,id=tpmchar,path=/var/run/qemu-server/517.swtpm -tpmdev emulator,id=tpmdev,chardev=tpmchar -device tpm-tis,tpmdev=tpmdev -device virtio-vga,id=vga,max_hostmem=16777216,bus=pcie.0,addr=0x1 -chardev socket,path=/var/run/qemu-server/517.qga,server=on,wait=off,id=qga0 -device virtio-serial,id=qga0,bus=pci.0,addr=0x8 -device virtserialport,chardev=qga0,name=org.qemu.guest_agent.0 -device virtio-serial,id=spice,bus=pci.0,addr=0x9 -chardev spicevmc,id=vdagent,name=vdagent -device virtserialport,chardev=vdagent,name=com.redhat.spice.0 -spice tls-port=61002,addr=127.0.0.1,tls-ciphers=HIGH,seamless-migration=on -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x3,free-page-reporting=on -iscsi initiator-name=iqn.1993-08.org.debian:01:ee6eef56d51 -drive if=none,id=drive-ide0,media=cdrom,aio=io_uring -device ide-cd,bus=ide.0,unit=0,drive=drive-ide0,id=ide0,bootindex=101 -drive file=rbd:ceph-nvme/vm-517-disk-3:conf=/etc/pve/ceph.conf:id=admin:keyring=/etc/pve/priv/ceph/ceph-nvme.keyring,if=none,id=drive-virtio1,format=raw,cache=none,aio=io_uring,detect-zeroes=on -device virtio-blk-pci,drive=drive-virtio1,id=virtio1,bus=pci.0,addr=0xb,iothread=iothread-virtio1 -device ahci,id=ahci0,multifunction=on,bus=pci.0,addr=0x7 -drive file=rbd:ceph-nvme/vm-517-disk-1:conf=/etc/pve/ceph.conf:id=admin:keyring=/etc/pve/priv/ceph/ceph-nvme.keyring,if=none,id=drive-sata0,discard=on,format=raw,cache=none,aio=io_uring,detect-zeroes=unmap -device ide-hd,bus=ahci0.0,drive=drive-sata0,id=sata0,rotation_rate=1,bootindex=100 -netdev type=tap,id=net0,ifname=tap517i0,script=/usr/libexec/qemu-server/pve-bridge,downscript=/usr/libexec/qemu-server/pve-bridgedown,vhost=on -device virtio-net-pci,mac=BC:24:11:9D:8E:E4,netdev=net0,bus=pci.0,addr=0x12,id=net0,rx_queue_size=1024,tx_queue_size=256 -rtc driftfix=slew,base=localtime -machine hpet=off,type=pc-q35-9.2+pve1

SCREENSHOT OF GUI WHEN DOING A SNAP WITH THE VM POWERED ON (AND RAM UNITCKED)
Snap when powered on.png



SCREENSHOT OF GUI WHEN DOING A SNAP WITH THE VM POWERED OFF

1777985735396.png
 

Attachments

  • Snap when powered on.png
    Snap when powered on.png
    46.5 KB · Views: 0
Hi,
this is a regression and is fixed in qemu-server >= 9.1.10, which currently is going through internal testing and should reach the pve-test and no-subscription repositories in the following days (the enterprise repository has qemu-server=9.1.8 and is not affected by the regression).