LXC backup stuck on "starting final sync"

Por12

Member
Mar 6, 2023
59
3
13
Hi,

I'm having issues with a new LXC I have created. I am trying to back-it up to a HDD connected to the same machine. The backup process will start fine and finish the file sync, but will get stuck on this part: INFO: starting final sync /proc/31416/root/ to /var/tmp/vzdumptmp1232345_304.

Here's an example. In this particular case I cancelled the backup but yesterday it was stuck in the step for more than 4h (until I woke up and cancelled it).

Any ideas on what could be happening?

Thanks

Code:
INFO: starting new backup job: vzdump 304 --compress zstd --mode snapshot --remove 0 --notes-template '{{guestname}}' --node zeus --storage local-backups-ferraz
INFO: filesystem type on dumpdir is 'zfs' -using /var/tmp/vzdumptmp1232345_304 for temporary files
INFO: Starting Backup of VM 304 (lxc)
INFO: Backup started at 2023-09-14 13:07:27
INFO: status = running
INFO: CT Name: frigate
INFO: including mount point rootfs ('/') in backup
INFO: mode failure - some volumes do not support snapshots
INFO: trying 'suspend' mode instead
INFO: backup mode: suspend
INFO: ionice priority: 7
INFO: CT Name: frigate
INFO: including mount point rootfs ('/') in backup
INFO: starting first sync /proc/31416/root/ to /var/tmp/vzdumptmp1232345_304
INFO: first sync finished - transferred 3.34G bytes in 29s
INFO: suspending guest
INFO: starting final sync /proc/31416/root/ to /var/tmp/vzdumptmp1232345_304
INFO: resume vm
INFO: guest is online again after 14 seconds
ERROR: Backup of VM 304 failed - command 'rsync --stats -h -X -A --numeric-ids -aH --delete --no-whole-file --sparse --one-file-system --relative '--exclude=/tmp/?*' '--exclude=/var/tmp/?*' '--exclude=/var/run/?*.pid' /proc/31416/root//./ /var/tmp/vzdumptmp1232345_304' failed: interrupted by signal
INFO: Failed at 2023-09-14 13:08:12
INFO: Backup job finished with errors
TASK ERROR: job errors
 
I am experiencing the same on a (just now created) new LXC container.

Running latest PVE8:
Code:
proxmox-ve: 8.0.2 (running kernel: 6.2.16-15-pve)
pve-manager: 8.0.4 (running version: 8.0.4/d258a813cfa6b390)
pve-kernel-6.2: 8.0.5
proxmox-kernel-helper: 8.0.3
pve-kernel-5.15: 7.4-4
pve-kernel-5.13: 7.1-9
proxmox-kernel-6.2.16-15-pve: 6.2.16-15
proxmox-kernel-6.2: 6.2.16-15
proxmox-kernel-6.2.16-14-pve: 6.2.16-14
pve-kernel-6.2.16-5-pve: 6.2.16-6
pve-kernel-5.15.108-1-pve: 5.15.108-2
pve-kernel-5.15.74-1-pve: 5.15.74-1
pve-kernel-5.15.64-1-pve: 5.15.64-1
pve-kernel-5.15.60-2-pve: 5.15.60-2
pve-kernel-5.15.60-1-pve: 5.15.60-1
pve-kernel-5.15.53-1-pve: 5.15.53-1
pve-kernel-5.15.39-4-pve: 5.15.39-4
pve-kernel-5.15.39-3-pve: 5.15.39-3
pve-kernel-5.15.39-2-pve: 5.15.39-2
pve-kernel-5.15.39-1-pve: 5.15.39-1
pve-kernel-5.15.35-3-pve: 5.15.35-6
pve-kernel-5.15.35-2-pve: 5.15.35-5
pve-kernel-5.15.35-1-pve: 5.15.35-3
pve-kernel-5.15.30-2-pve: 5.15.30-3
pve-kernel-5.13.19-6-pve: 5.13.19-15
pve-kernel-5.13.19-5-pve: 5.13.19-13
pve-kernel-5.13.19-4-pve: 5.13.19-9
ceph-fuse: 16.2.11+ds-2
corosync: 3.1.7-pve3
criu: 3.17.1-2
gfs2-utils: 3.5.0-2
glusterfs-client: 10.3-5
ifupdown: not correctly installed
ifupdown2: 3.2.0-1+pmx5
libjs-extjs: 7.0.0-4
libknet1: 1.26-pve1
libproxmox-acme-perl: 1.4.6
libproxmox-backup-qemu0: 1.4.0
libproxmox-rs-perl: 0.3.1
libpve-access-control: 8.0.5
libpve-apiclient-perl: 3.3.1
libpve-common-perl: 8.0.9
libpve-guest-common-perl: 5.0.5
libpve-http-server-perl: 5.0.4
libpve-rs-perl: 0.8.5
libpve-storage-perl: 8.0.2
libspice-server1: 0.15.1-1
lvm2: 2.03.16-2
lxc-pve: 5.0.2-4
lxcfs: 5.0.3-pve3
novnc-pve: 1.4.0-2
proxmox-backup-client: 3.0.3-1
proxmox-backup-file-restore: 3.0.3-1
proxmox-kernel-helper: 8.0.3
proxmox-mail-forward: 0.2.0
proxmox-mini-journalreader: 1.4.0
proxmox-offline-mirror-helper: 0.6.2
proxmox-widget-toolkit: 4.0.9
pve-cluster: 8.0.4
pve-container: 5.0.4
pve-docs: 8.0.5
pve-edk2-firmware: 3.20230228-4
pve-firewall: 5.0.3
pve-firmware: 3.8-2
pve-ha-manager: 4.0.2
pve-i18n: 3.0.7
pve-qemu-kvm: 8.0.2-6
pve-xtermjs: 4.16.0-3
qemu-server: 8.0.7
smartmontools: 7.3-pve1
spiceterm: 3.3.0
swtpm: 0.8.0+pve1
vncterm: 1.8.0
zfsutils-linux: 2.1.13-pve1

I aborted the backup on the new LXC, but the same behaviour was seen:
Code:
INFO: status = running
INFO: CT Name: docker.mydomain.tld
INFO: including mount point rootfs ('/') in backup
INFO: mode failure - some volumes do not support snapshots
INFO: trying 'suspend' mode instead
INFO: backup mode: suspend
INFO: ionice priority: 7
INFO: CT Name: docker.mydomain.tld
INFO: including mount point rootfs ('/') in backup
INFO: starting first sync /proc/1492511/root/ to /var/tmp/vzdumptmp1633458_1006
INFO: first sync finished - transferred 1.45G bytes in 18s
INFO: suspending guest
INFO: starting final sync /proc/1492511/root/ to /var/tmp/vzdumptmp1633458_1006
INFO: resume vm
INFO: guest is online again after 1925 seconds
ERROR: Backup of VM 1006 failed - command 'rsync --stats -h -X -A --numeric-ids -aH --delete --no-whole-file --sparse --one-file-system --relative '--exclude=/tmp/?*' '--exclude=/var/tmp/?*' '--exclude=/var/run/?*.pid' /proc/1492511/root//./ /var/tmp/vzdumptmp1633458_1006' failed: interrupted by signal
INFO: Failed at 2023-10-14 01:06:34
INFO: filesystem type on dumpdir is 'gfs2' -using /var/tmp/vzdumptmp1633458_1008 for temporary files
 
Turns out it was the FUSE feature being in use on the LXC.
As soon as i took it off - or downed the affected host - after reading Forum posts and docs and the issues with it - backups went fine.

For reference as to where i got my answers : This forum Post

In my case we are talking about a docker LXC based on a RockyLinux 9.2 template.
 
Last edited: