Hi,
We recently had an issue with our proxmox cluster after upgrading to the newest version that was most likely related to the kernel bug issue mentioned in github.com/lxc/lxc/issues/2141 which eventually forced us to reboot the updated hypervisor after it went unresponsive when we had to restart some containers.
pveversion
pve-manager/5.1-46/ae8241d4 (running kernel: 4.13.13-6-pve)
Now there's one container which is unable to do backups, (some names from the output have been changed due to security reasons)
INFO: starting new backup job: vzdump 1246 --node hypervisor1 --remove 0 --mode snapshot --storage local-backup2 --compress lzo
INFO: Starting Backup of VM 1246 (lxc)
INFO: status = running
INFO: CT Name: container1
INFO: backup mode: snapshot
INFO: ionice priority: 7
INFO: create storage snapshot 'vzdump'
Using default stripesize 64.00 KiB.
snapshot create failed: starting cleanup
no lock found trying to remove 'backup' lock
ERROR: Backup of VM 1246 failed - lvcreate snapshot 'pve/snap_vm-1246-disk-1_vzdump' error: Logical Volume "snap_vm-1246-disk-1_vzdump" already exists in volume group "pve"
INFO: Backup job finished with errors
TASK ERROR: job errors
There is an LV with that name in the system, created when the system went down due to the earlier mentioned problem;
lvdisplay /dev/pve/snap_vm-1246-disk-1_vzdump
--- Logical volume ---
LV Path /dev/pve/snap_vm-1246-disk-1_vzdump
LV Name snap_vm-1246-disk-1_vzdump
VG Name pve
LV UUID 4wgqeR-OpvP-xVo9-1JkD-yazS-4Ugb-lKZ2an
LV Write Access read only
LV Creation host, time hypervisor1, 2018-03-19 22:58:26 +0200
LV Pool name data
LV Thin origin name vm-1246-disk-1
LV Status NOT available
LV Size 250.00 GiB
Current LE 64000
Segments 1
Allocation inherit
Read ahead sectors auto
Should it be safe to just lvremove that volume or is there an another way to deal with such issue?
We recently had an issue with our proxmox cluster after upgrading to the newest version that was most likely related to the kernel bug issue mentioned in github.com/lxc/lxc/issues/2141 which eventually forced us to reboot the updated hypervisor after it went unresponsive when we had to restart some containers.
pveversion
pve-manager/5.1-46/ae8241d4 (running kernel: 4.13.13-6-pve)
Now there's one container which is unable to do backups, (some names from the output have been changed due to security reasons)
INFO: starting new backup job: vzdump 1246 --node hypervisor1 --remove 0 --mode snapshot --storage local-backup2 --compress lzo
INFO: Starting Backup of VM 1246 (lxc)
INFO: status = running
INFO: CT Name: container1
INFO: backup mode: snapshot
INFO: ionice priority: 7
INFO: create storage snapshot 'vzdump'
Using default stripesize 64.00 KiB.
snapshot create failed: starting cleanup
no lock found trying to remove 'backup' lock
ERROR: Backup of VM 1246 failed - lvcreate snapshot 'pve/snap_vm-1246-disk-1_vzdump' error: Logical Volume "snap_vm-1246-disk-1_vzdump" already exists in volume group "pve"
INFO: Backup job finished with errors
TASK ERROR: job errors
There is an LV with that name in the system, created when the system went down due to the earlier mentioned problem;
lvdisplay /dev/pve/snap_vm-1246-disk-1_vzdump
--- Logical volume ---
LV Path /dev/pve/snap_vm-1246-disk-1_vzdump
LV Name snap_vm-1246-disk-1_vzdump
VG Name pve
LV UUID 4wgqeR-OpvP-xVo9-1JkD-yazS-4Ugb-lKZ2an
LV Write Access read only
LV Creation host, time hypervisor1, 2018-03-19 22:58:26 +0200
LV Pool name data
LV Thin origin name vm-1246-disk-1
LV Status NOT available
LV Size 250.00 GiB
Current LE 64000
Segments 1
Allocation inherit
Read ahead sectors auto
Should it be safe to just lvremove that volume or is there an another way to deal with such issue?