several backup problems

udi

Active Member
Apr 1, 2011
76
0
26
hi,
i had these problems with backup:

1. when i add a job it's not listed in gui, unless i select include all vms.
however these jobs are trying to start so tey're there somewhere. where?

2. the jobs result to errors like this:

vzdump 102 --quiet 1 --mailto www@xxxx.eu --mode snapshot --compress 1 --maxfiles 2 --storage backup 102: Nov 29 09:30:05 INFO: Starting Backup of VM 102 (qemu) 102: Nov 29 09:30:05 INFO: status = running 102: Nov 29 09:30:07 INFO: backup mode: snapshot 102: Nov 29 09:30:07 INFO: ionice priority: 7 102: Nov 29 09:30:07 INFO: suspend vm to make snapshot 102: Nov 29 09:30:07 INFO: Logical volume "vzsnap-genya-0" created 102: Nov 29 09:30:07 INFO: trying to remove stale snapshot '/dev/vgvirt/vzsnap-genya-1' 102: Nov 29 09:30:08 INFO: Unable to deactivate open vgvirt-vm--102--disk--1-real (253:5) 102: Nov 29 09:30:08 INFO: Failed to resume vm-102-disk-1. 102: Nov 29 09:30:08 ERROR: command 'lvremove -f '/dev/vgvirt/vzsnap-genya-1'' failed: exit code 5 102: Nov 29 09:30:08 INFO: Logical volume "vzsnap-genya-1" already exists in volume group "vgvirt" 102: Nov 29 09:30:08 INFO: Unable to deactivate open vgvirt-vm--102--disk--2-real (253:9) 102: Nov 29 09:30:08 INFO: Failed to resume vm-102-disk-2. 102: Nov 29 09:30:08 ERROR: command 'lvremove -f '/dev/vgvirt/vzsnap-genya-0'' failed: exit code 5 102: Nov 29 09:30:08 INFO: Logical volume "vzsnap-genya-1" successfully removed 102: Nov 29 09:30:08 INFO: resume vm 102: Nov 29 09:30:09 INFO: vm is online again after 2 seconds 102: Nov 29 09:30:09 ERROR: Backup of VM 102 failed - command 'lvcreate --size 1024M --snapshot --name 'vzsnap-genya-1' '/dev/vgvirt/vm-102-disk-1'' failed: exit code 5


3. yesterday i added a job to backup all vms at night but this morning i found my proxmox system unresponsive, i had to reset the machine.
i found that it finished with the first vm but hunged at second, left it locked.

thank you
u.
 
we will upload new package with a lot of fixes regarding backup soon (tomorrow).
 
thank you Tom,

please help me how to clean all backup jobs, i don't see them in the gui
 
after this mornings upgrade, I can now see the backup vzdump jobs, but can not edit or delete.

on pve web page: "unable to open file '/etc/pve/vzdump.cron.tmp.4323' - Permission denied (500)"

and cli listing or /etc/pve shows perms :
Code:
-r--r-----  1 root www-data 1048 Nov 26 07:40 vzdump.cron

version info:
Code:
pveversion -v
pve-manager: 2.0-12 (pve-manager/2.0/784729f4)
running kernel: 2.6.32-6-pve
proxmox-ve-2.6.32: 2.0-53
pve-kernel-2.6.32-6-pve: 2.6.32-53
lvm2: 2.02.86-1pve2
clvm: 2.02.86-1pve2
corosync-pve: 1.4.1-1
openais-pve: 1.1.4-1
libqb: 0.6.0-1
redhat-cluster-pve: 3.1.7-1
pve-cluster: 1.0-12
qemu-server: 2.0-9
pve-firmware: 1.0-13
libpve-common-perl: 1.0-8
libpve-access-control: 1.0-2
libpve-storage-perl: 2.0-8
vncterm: 1.0-2
vzctl: 3.0.29-3pve3
vzprocps: 2.0.11-2
vzquota: 3.0.12-3
pve-qemu-kvm: 0.15.0-1
ksm-control-daemon: 1.1-1
 
what is the output of

# pvecm status

(maybe you do not have quorum?)


yes there was this at fbc158 console at boot:
Code:
Starting cluster:
   Checking if cluster has been disabled at boot... [  OK  ]
   Checking Network Manager... [  OK  ]
   Global setup... [  OK  ]
   Loading kernel modules... [  OK  ]
   Mounting configfs... [  OK  ]
   Starting cman... [  OK  ]
   Waiting for quorum... Timed-out waiting for cluster
[FAILED]
Starting OpenVZ: ..done

then the data you asked for:
Code:
root@homenet-fbc158 /etc/lvm # pvecm status
Version: 6.2.0
Config Version: 4
Cluster Name: fbcandover
Cluster Id: 37452
Cluster Member: Yes
Cluster Generation: 6656
Membership state: Cluster-Member
Nodes: 1
Expected votes: 2
Total votes: 1
Node votes: 1
Quorum: 2 Activity blocked
Active subsystems: 1
Flags: 
Ports Bound: 0  
Node name: fbc158
Node ID: 1
Multicast addresses: 239.192.146.222 
Node addresses: 10.100.100.158

and on the other node fbc10:
Code:
root@homenet-fbc10 /etc/pve # pvecm status
Version: 6.2.0
Config Version: 4
Cluster Name: fbcandover
Cluster Id: 37452
Cluster Member: Yes
Cluster Generation: 4
Membership state: Cluster-Member
Nodes: 1
Expected votes: 2
Total votes: 1
Node votes: 1
Quorum: 2 Activity blocked
Active subsystems: 1
Flags: 
Ports Bound: 0  
Node name: fbc10
Node ID: 2
Multicast addresses: 239.192.146.222 
Node addresses: 10.100.100.6

I see 10.100.100.6 , which is not the main address.. I added that to interfaces :
Code:
auto vmbr0
iface vmbr0 inet static
        address  10.100.100.10
        netmask  255.255.255.0
        gateway  10.100.100.2
        bridge_ports eth0
        bridge_stp off
        bridge_fd 0



# 2011-11-26  dhcp needs
auto eth0:1
iface eth0:1  inet static
    address 10.100.100.6
    netmask 255.255.0.0

so I'll remove the 10.100.100.6 address for now. I added it to use the same dns server address on 2 different setups [ in order to try to use the same dhcpd.conf , but I'll configure dhcp differently...].
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!