several backup problems

udi

Active Member
Apr 1, 2011
76
0
26
hi,
i had these problems with backup:

1. when i add a job it's not listed in gui, unless i select include all vms.
however these jobs are trying to start so tey're there somewhere. where?

2. the jobs result to errors like this:

vzdump 102 --quiet 1 --mailto www@xxxx.eu --mode snapshot --compress 1 --maxfiles 2 --storage backup 102: Nov 29 09:30:05 INFO: Starting Backup of VM 102 (qemu) 102: Nov 29 09:30:05 INFO: status = running 102: Nov 29 09:30:07 INFO: backup mode: snapshot 102: Nov 29 09:30:07 INFO: ionice priority: 7 102: Nov 29 09:30:07 INFO: suspend vm to make snapshot 102: Nov 29 09:30:07 INFO: Logical volume "vzsnap-genya-0" created 102: Nov 29 09:30:07 INFO: trying to remove stale snapshot '/dev/vgvirt/vzsnap-genya-1' 102: Nov 29 09:30:08 INFO: Unable to deactivate open vgvirt-vm--102--disk--1-real (253:5) 102: Nov 29 09:30:08 INFO: Failed to resume vm-102-disk-1. 102: Nov 29 09:30:08 ERROR: command 'lvremove -f '/dev/vgvirt/vzsnap-genya-1'' failed: exit code 5 102: Nov 29 09:30:08 INFO: Logical volume "vzsnap-genya-1" already exists in volume group "vgvirt" 102: Nov 29 09:30:08 INFO: Unable to deactivate open vgvirt-vm--102--disk--2-real (253:9) 102: Nov 29 09:30:08 INFO: Failed to resume vm-102-disk-2. 102: Nov 29 09:30:08 ERROR: command 'lvremove -f '/dev/vgvirt/vzsnap-genya-0'' failed: exit code 5 102: Nov 29 09:30:08 INFO: Logical volume "vzsnap-genya-1" successfully removed 102: Nov 29 09:30:08 INFO: resume vm 102: Nov 29 09:30:09 INFO: vm is online again after 2 seconds 102: Nov 29 09:30:09 ERROR: Backup of VM 102 failed - command 'lvcreate --size 1024M --snapshot --name 'vzsnap-genya-1' '/dev/vgvirt/vm-102-disk-1'' failed: exit code 5


3. yesterday i added a job to backup all vms at night but this morning i found my proxmox system unresponsive, i had to reset the machine.
i found that it finished with the first vm but hunged at second, left it locked.

thank you
u.
 
we will upload new package with a lot of fixes regarding backup soon (tomorrow).
 
thank you Tom,

please help me how to clean all backup jobs, i don't see them in the gui
 
after this mornings upgrade, I can now see the backup vzdump jobs, but can not edit or delete.

on pve web page: "unable to open file '/etc/pve/vzdump.cron.tmp.4323' - Permission denied (500)"

and cli listing or /etc/pve shows perms :
Code:
-r--r-----  1 root www-data 1048 Nov 26 07:40 vzdump.cron

version info:
Code:
pveversion -v
pve-manager: 2.0-12 (pve-manager/2.0/784729f4)
running kernel: 2.6.32-6-pve
proxmox-ve-2.6.32: 2.0-53
pve-kernel-2.6.32-6-pve: 2.6.32-53
lvm2: 2.02.86-1pve2
clvm: 2.02.86-1pve2
corosync-pve: 1.4.1-1
openais-pve: 1.1.4-1
libqb: 0.6.0-1
redhat-cluster-pve: 3.1.7-1
pve-cluster: 1.0-12
qemu-server: 2.0-9
pve-firmware: 1.0-13
libpve-common-perl: 1.0-8
libpve-access-control: 1.0-2
libpve-storage-perl: 2.0-8
vncterm: 1.0-2
vzctl: 3.0.29-3pve3
vzprocps: 2.0.11-2
vzquota: 3.0.12-3
pve-qemu-kvm: 0.15.0-1
ksm-control-daemon: 1.1-1
 
what is the output of

# pvecm status

(maybe you do not have quorum?)


yes there was this at fbc158 console at boot:
Code:
Starting cluster:
   Checking if cluster has been disabled at boot... [  OK  ]
   Checking Network Manager... [  OK  ]
   Global setup... [  OK  ]
   Loading kernel modules... [  OK  ]
   Mounting configfs... [  OK  ]
   Starting cman... [  OK  ]
   Waiting for quorum... Timed-out waiting for cluster
[FAILED]
Starting OpenVZ: ..done

then the data you asked for:
Code:
root@homenet-fbc158 /etc/lvm # pvecm status
Version: 6.2.0
Config Version: 4
Cluster Name: fbcandover
Cluster Id: 37452
Cluster Member: Yes
Cluster Generation: 6656
Membership state: Cluster-Member
Nodes: 1
Expected votes: 2
Total votes: 1
Node votes: 1
Quorum: 2 Activity blocked
Active subsystems: 1
Flags: 
Ports Bound: 0  
Node name: fbc158
Node ID: 1
Multicast addresses: 239.192.146.222 
Node addresses: 10.100.100.158

and on the other node fbc10:
Code:
root@homenet-fbc10 /etc/pve # pvecm status
Version: 6.2.0
Config Version: 4
Cluster Name: fbcandover
Cluster Id: 37452
Cluster Member: Yes
Cluster Generation: 4
Membership state: Cluster-Member
Nodes: 1
Expected votes: 2
Total votes: 1
Node votes: 1
Quorum: 2 Activity blocked
Active subsystems: 1
Flags: 
Ports Bound: 0  
Node name: fbc10
Node ID: 2
Multicast addresses: 239.192.146.222 
Node addresses: 10.100.100.6

I see 10.100.100.6 , which is not the main address.. I added that to interfaces :
Code:
auto vmbr0
iface vmbr0 inet static
        address  10.100.100.10
        netmask  255.255.255.0
        gateway  10.100.100.2
        bridge_ports eth0
        bridge_stp off
        bridge_fd 0



# 2011-11-26  dhcp needs
auto eth0:1
iface eth0:1  inet static
    address 10.100.100.6
    netmask 255.255.0.0

so I'll remove the 10.100.100.6 address for now. I added it to use the same dns server address on 2 different setups [ in order to try to use the same dhcpd.conf , but I'll configure dhcp differently...].