error while deleting snapshot

raj

Renowned Member
Sep 17, 2011
222
4
83
www.businessparksolutions.com
Hi Team,

I am getting th efollowing error while trying to delete snapshots from a kvm server. The kvm is cent os 7 and is switch off.
Errors :

qemu-img: Could not delete snapshot 'snapshot6': Failed to free the cluster and L1 table: Invalid argument
TASK ERROR: command '/usr/bin/qemu-img snapshot -d snapshot6 /mnt/pve/Datastore/images/101/vm-101-disk-1.qcow2' failed: exit code 1

Storage is from NFS share.

Proxmox is up to date.

proxmox-ve: 4.4-84 (running kernel: 4.4.44-1-pve)
pve-manager: 4.4-12 (running version: 4.4-12/e71b7a74)
pve-kernel-4.4.44-1-pve: 4.4.44-84
pve-kernel-4.4.19-1-pve: 4.4.19-66
lvm2: 2.02.116-pve3
corosync-pve: 2.4.2-2~pve4+1
libqb0: 1.0-1
pve-cluster: 4.0-48
qemu-server: 4.0-109
pve-firmware: 1.1-10
libpve-common-perl: 4.0-92
libpve-access-control: 4.0-23
libpve-storage-perl: 4.0-76
pve-libspice-server1: 0.12.8-2
vncterm: 1.3-1
pve-docs: 4.4-3
pve-qemu-kvm: 2.7.1-4
pve-container: 1.0-94
pve-firewall: 2.0-33
pve-ha-manager: 1.0-40
ksm-control-daemon: 1.2-1
glusterfs-client: 3.5.2-2+deb8u3
lxc-pve: 2.0.7-3
lxcfs: 2.0.6-pve1
criu: 1.6.0-1
novnc-pve: 0.5-8
smartmontools: 6.5+svn4324-1~pve80
zfsutils: 0.6.5.9-pve15~bpo80

Could anyone please advise

Rajbps
 
Hi,

please update your system and try again.
 
i have the same problem with a centos7 VM on pve ( pve-manager/7.3-3/c3928077 (running kernel: 5.15.74-1-pve) , running in production

did the update help ?
 
on second try , after powering on the VM again i'm now getting
TASK ERROR: VM 159 qmp command 'blockdev-snapshot-delete-internal-sync' failed - Snapshot with id 'null' and name 'vor_update_9_9_lts' does not exist on device 'drive-scsi0'

powered down again and in vm configuration, the snapshot section contains "snapstate: delete" and quemu-img snapshot does not list any snapshot

root@pve5:/ssdpool/qcow2/images/159# qm listsnapshot 159
`-> vor_update_9_9_lts 2023-09-20 19:03:16 no-description
`-> current You are here!

root@pve5:/ssdpool/qcow2/images/159# qemu-img snapshot -l vm-159-disk-0.qcow2
#

that VM is totally low IO

i have found these older ones, but i cannot remember that the VM had been suspended in any way, especially not since the snapshot. the snapshot is without memory.

https://bugs.launchpad.net/qemu/+bug/1846427
https://bugzilla.redhat.com/show_bug.cgi?id=1772321

any clue?

there is not much information for this on the net and snapshot corruption issue on a pve host with 393 days of uptime and no other problems so far is making me nervous a little bit....


addon:
i restored that VM from zfs snapshots from 2 weeks ago on a PVE pve-manager/7.4-17/513c62be (running kernel: 5.15.126-1-pve) system and it's also corrupted there
 
Last edited:
i got another VM with this error

qemu-img: Could not delete snapshot 'vor_os_update': Failed to free the cluster and L1 table: Invalid argument
TASK ERROR: command '/usr/bin/qemu-img snapshot -d vor_os_update /rpool/vms-files-zstd/images/106/vm-106-disk-0.qcow2' failed: exit code 1
 
qemu-img: Could not delete snapshot 'vor_os_update': Failed to free the cluster and L1 table: Invalid argument
TASK ERROR: command '/usr/bin/qemu-img snapshot -d vor_os_update /rpool/vms-files-zstd/images/106/vm-106-disk-0.qcow2' failed: exit code 1
The error indicates that QCOW file structure is corrupted/damaged.
There is a similarity to this bug report, however, the message is just a generic symptom and could be caused by various types of corruption.
It's possible that the damage was somehow induced by PVE processes. However, it could be completely storage-driven as well, or a Kernel bug.
For such issues, reproduction is the best way to get to the bottom of it, as the likely cause of the current state could be long gone.


Blockbridge : Ultra low latency all-NVME shared storage for Proxmox - https://www.blockbridge.com/proxmox