live migrations failes

felipe

Well-Known Member
Oct 28, 2013
222
6
58
hi,

i tried online migration but:

Aug 22 15:43:18 starting migration of VM 208 to node 'ceph2' (192.168.11.32)
Aug 22 15:43:18 copying disk images
Aug 22 15:43:18 starting VM 208 on remote node 'ceph2'
Aug 22 15:43:20 starting ssh migration tunnel
Aug 22 15:43:20 starting online/live migration on localhost:60000
Aug 22 15:43:20 migrate_set_speed: 8589934592
Aug 22 15:43:20 migrate_set_downtime: 0.1
Aug 22 15:43:22 ERROR: online migrate failure - aborting
Aug 22 15:43:22 aborting phase 2 - cleanup resources
Aug 22 15:43:22 migrate_cancel
Aug 22 15:43:23 ERROR: migration finished with problems (duration 00:00:05)
TASK ERROR: migration problems

some weaks ago it worked.

pveversion --verbose
proxmox-ve-2.6.32: 3.2-129 (running kernel: 3.10.0-3-pve)
pve-manager: 3.2-4 (running version: 3.2-4/e24a91c1)
pve-kernel-2.6.32-30-pve: 2.6.32-130
pve-kernel-2.6.32-29-pve: 2.6.32-126
pve-kernel-3.10.0-3-pve: 3.10.0-11
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.5-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.5-1
pve-cluster: 3.0-12
qemu-server: 3.1-16
pve-firmware: 1.1-3
libpve-common-perl: 3.0-18
libpve-access-control: 3.0-11
libpve-storage-perl: 3.0-19
pve-libspice-server1: 0.12.4-3
vncterm: 1.1-6
vzctl: 4.0-1pve5
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 1.7-8
ksm-control-daemon: 1.1-1
glusterfs-client: 3.4.2-1

---- identical on all 3 hosts. shared storage ceph
 
i dont know what exactly you mean with join.
i can see all ceph images on the target node. i can create a new vm on ceph on the target node. and i can also clone a vm from another node to this node. i can also migrate (NOT LIVE!)
only live migration fails. but the strange thing is that the same vms i allready live migrated some weeks before. and now i get this error....
 
todo i can confirm that live migration suddently work!
i did not change anything? just pressig the same button live migration.
will test it again the next days.... but what coult break it temoporary?
 
[QUOTE ]but what coult break it temoporary? [/QUOTE]

I think this can happen if your target node can't join ceph (temporary ?).
Because livemigration start a kvm process on target node, and if it can't join ceph, it'll die.

Double check your network to see if you don't have problem.

 
i will wait for the next time this occurs and check network... at the moment it is working again...