Backup Errors

frantek

Renowned Member
May 30, 2009
176
7
83
Hi,

on one of my systems I see the following errors when backups fail:

Code:
[I]atus[/I]


INFO:  starting new backup job: vzdump --quiet 1 --mailto noc@mhcsoftware.de  --mode snapshot --compress 1 --maxfiles 1 --storage backup --all 1
INFO: Starting Backup of VM 100 (qemu)
INFO: status = running
INFO: backup mode: snapshot
INFO: ionice priority: 7
INFO: trying to remove stale snapshot '/dev/pve/vzsnap-bs-0'
INFO: umount: /mnt/vzsnap0: not mounted
ERROR: command ''umount /mnt/vzsnap0'' failed: exit code 1
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064523776: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064581120: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 0: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 4096: Input/output error
INFO:   Logical volume "vzsnap-bs-0" already exists in volume group "pve"
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
ERROR:  Backup of VM 100 failed - command ''lvcreate --size 1024M --snapshot  --name '\''vzsnap-bs-0'\'' '\''/dev/pve/data'\'' failed: exit code 5
INFO: Starting Backup of VM 101 (qemu)
INFO: status = running
INFO: backup mode: snapshot
INFO: ionice priority: 7
INFO: trying to remove stale snapshot '/dev/pve/vzsnap-bs-0'
INFO: umount: /mnt/vzsnap0: not mounted
ERROR: command ''umount /mnt/vzsnap0'' failed: exit code 1
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064523776: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064581120: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 0: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 4096: Input/output error
INFO:   Logical volume "vzsnap-bs-0" already exists in volume group "pve"
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
ERROR:  Backup of VM 101 failed - command ''lvcreate --size 1024M --snapshot  --name '\''vzsnap-bs-0'\'' '\''/dev/pve/data'\'' failed: exit code 5
INFO: Starting Backup of VM 102 (qemu)
INFO: status = running
INFO: backup mode: snapshot
INFO: ionice priority: 7
INFO: trying to remove stale snapshot '/dev/pve/vzsnap-bs-0'
INFO: umount: /mnt/vzsnap0: not mounted
ERROR: command ''umount /mnt/vzsnap0'' failed: exit code 1
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064523776: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064581120: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 0: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 4096: Input/output error
INFO:   Logical volume "vzsnap-bs-0" already exists in volume group "pve"
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
ERROR:  Backup of VM 102 failed - command ''lvcreate --size 1024M --snapshot  --name '\''vzsnap-bs-0'\'' '\''/dev/pve/data'\'' failed: exit code 5
INFO: Starting Backup of VM 103 (qemu)
INFO: status = running
INFO: backup mode: snapshot
INFO: ionice priority: 7
INFO: trying to remove stale snapshot '/dev/pve/vzsnap-bs-0'
INFO: umount: /mnt/vzsnap0: not mounted
ERROR: command ''umount /mnt/vzsnap0'' failed: exit code 1
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064523776: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064581120: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 0: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 4096: Input/output error
INFO:   Logical volume "vzsnap-bs-0" already exists in volume group "pve"
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
ERROR:  Backup of VM 103 failed - command ''lvcreate --size 1024M --snapshot  --name '\''vzsnap-bs-0'\'' '\''/dev/pve/data'\'' failed: exit code 5
INFO: Starting Backup of VM 104 (qemu)
INFO: status = running
INFO: backup mode: snapshot
INFO: ionice priority: 7
INFO: trying to remove stale snapshot '/dev/pve/vzsnap-bs-0'
INFO: umount: /mnt/vzsnap0: not mounted
ERROR: command ''umount /mnt/vzsnap0'' failed: exit code 1
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064523776: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 2885064581120: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 0: Input/output error
INFO:   /dev/pve/vzsnap-bs-0: read failed after 0 of 4096 at 4096: Input/output error
INFO:   Logical volume "vzsnap-bs-0" already exists in volume group "pve"
INFO: lvremove failed - trying again in 8 seconds
INFO: lvremove failed - trying again in 16 seconds
INFO: lvremove failed - trying again in 32 seconds
ERROR: command 'lvremove -f /dev/pve/vzsnap-bs-0' failed: Not a CODE reference at /usr/share/perl5/PVE/Tools.pm line 352.
ERROR:  Backup of VM 104 failed - command ''lvcreate --size 1024M --snapshot  --name '\''vzsnap-bs-0'\'' '\''/dev/pve/data'\'' failed: exit code 5
INFO: Backup job finished with errors
TASK ERROR: job errors

At the moment I'm NOT looking for a solution but I want to know what causes these I/O errors. After all backups failed I can remove the snapshot manually. When I do this I see the same I/O errors.

TIA
Matthias
 
also post the output of 'pveversion -v'
 
I have the same problem with the backup.

Code:
/dev/raven1-1/vzsnap-hawk2-0: read failed after 0 of 4096 at 161061208064: Input/output error
  /dev/raven1-1/vzsnap-hawk2-0: read failed after 0 of 4096 at 161061265408: Input/output error
  /dev/raven1-1/vzsnap-hawk2-0: read failed after 0 of 4096 at 0: Input/output error
  /dev/raven1-1/vzsnap-hawk2-0: read failed after 0 of 4096 at 4096: Input/output error

screen lvs
Code:
LV             VG       Attr     LSize   Pool Origin        Data%  Move Log Copy%  Convert
 vm-105-disk-1  raven1-1 owi-a-s- 150.00g                                                  
vzsnap-hawk2-0 raven1-1 Swi-I-s-   1.00g      vm-105-disk-1 100.00

screen lvscan

Code:
/dev/raven1-1/vzsnap-hawk2-0: read failed after 0 of 4096 at 161061208064: Input/output error
  /dev/raven1-1/vzsnap-hawk2-0: read failed after 0 of 4096 at 161061265408: Input/output error
  /dev/raven1-1/vzsnap-hawk2-0: read failed after 0 of 4096 at 0: Input/output error
  /dev/raven1-1/vzsnap-hawk2-0: read failed after 0 of 4096 at 4096: Input/output error
 inactive Original '/dev/raven1-1/vm-105-disk-1' [150.00 GiB] inherit
inactive Snapshot '/dev/raven1-1/vzsnap-hawk2-0' [1.00 GiB] inherit

Code:
pveversion pve-manager/2.1/f9b0f63a
root@hawk2:~# pveversion -v
pve-manager: 2.1-1 (pve-manager/2.1/f9b0f63a)
running kernel: 2.6.32-12-pve
proxmox-ve-2.6.32: 2.1-68
pve-kernel-2.6.32-11-pve: 2.6.32-66
pve-kernel-2.6.32-12-pve: 2.6.32-68
lvm2: 2.02.95-1pve2
clvm: 2.02.95-1pve2
corosync-pve: 1.4.3-1
openais-pve: 1.1.4-2
libqb: 0.10.1-2
redhat-cluster-pve: 3.1.8-3
resource-agents-pve: 3.9.2-3
fence-agents-pve: 3.1.7-2
pve-cluster: 1.0-26
qemu-server: 2.0-39
pve-firmware: 1.0-16
libpve-common-perl: 1.0-27
libpve-access-control: 1.0-21
libpve-storage-perl: 2.0-18
vncterm: 1.0-2
vzctl: 3.0.30-2pve5
vzprocps: 2.0.11-2
vzquota: 3.0.12-3
pve-qemu-kvm: 1.0-9
ksm-control-daemon: 1.1-1

lock on the backup - removed
how to fix the error?
 
My very first unupdated V2 setup ... I've no physical access and no backups so I will not do updates which require a reboot :-)

Code:
pve-manager: 2.0-18 (pve-manager/2.0/16283a5a)
running kernel: 2.6.32-6-pve
proxmox-ve-2.6.32: 2.0-55
pve-kernel-2.6.32-6-pve: 2.6.32-55
lvm2: 2.02.88-2pve1
clvm: 2.02.88-2pve1
corosync-pve: 1.4.1-1
openais-pve: 1.1.4-1
libqb: 0.6.0-1
redhat-cluster-pve: 3.1.8-3
pve-cluster: 1.0-17
qemu-server: 2.0-13
pve-firmware: 1.0-14
libpve-common-perl: 1.0-11
libpve-access-control: 1.0-5
libpve-storage-perl: 2.0-9
vncterm: 1.0-2
vzctl: 3.0.29-3pve8
vzprocps: 2.0.11-2
vzquota: 3.0.12-3
pve-qemu-kvm: 1.0-1
ksm-control-daemon: 1.1-1
 
I will as soon as I'm able to schedule downtime for off line backups, but I also want to know the reason for the errors in the current setup as it used to work.