Hello Everyone!
My PVE cluster has been running great for many months but i just got around to updating the all the pve nodes to the latest version of PVE 5.4-3 now.
I started live migrating all my VM's to another node and that worked great.
I then set node-out on the the first node and i live migrated all the VM's from it and then i set each OSD to out on that node and waited for backfill.
Once backfill was complete i updated the all the nodes and rebooted the node that was empty.
The node came back online fine and quorate is perfect and the cluster is healthy.
BUT now when i try to live migrate the VM's back the the updated node i receive a error and live migration fails.
Here is the error
I have tried live migration on other VM's on the other nodes and everyone fails with the same error.
Any ideas on how to correct this?
I am running the latest PVE version on ALL nodes. 2 of the 3 have NOT been rebooted since update.
My PVE cluster has been running great for many months but i just got around to updating the all the pve nodes to the latest version of PVE 5.4-3 now.
I started live migrating all my VM's to another node and that worked great.
I then set node-out on the the first node and i live migrated all the VM's from it and then i set each OSD to out on that node and waited for backfill.
Once backfill was complete i updated the all the nodes and rebooted the node that was empty.
The node came back online fine and quorate is perfect and the cluster is healthy.
BUT now when i try to live migrate the VM's back the the updated node i receive a error and live migration fails.
Here is the error
Code:
2019-04-12 16:21:20 starting migration of VM 104 to node 'he-s08-r01-pve02' (23.136.0-hidden)
2019-04-12 16:21:21 copying disk images
2019-04-12 16:21:21 starting VM 104 on remote node 'he-s08-r01-pve02'
2019-04-12 16:21:22 error with cfs lock 'storage-VM-STOR2-PVE02': rbd create vm-104-cloudinit' error: rbd: create error: (17) File exists
2019-04-12 16:21:22 ERROR: online migrate failure - command '/usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=he-s08-r01-pve02' root@23.136.0.11 qm start 104 --skiplock --migratedfrom he-s07-r01-pve02 --migration_type secure --stateuri unix --machine pc-i440fx-2.12' failed: exit code 255
2019-04-12 16:21:22 aborting phase 2 - cleanup resources
2019-04-12 16:21:22 migrate_cancel
2019-04-12 16:21:23 ERROR: migration finished with problems (duration 00:00:03)
TASK ERROR: migration problems
I have tried live migration on other VM's on the other nodes and everyone fails with the same error.
Any ideas on how to correct this?
I am running the latest PVE version on ALL nodes. 2 of the 3 have NOT been rebooted since update.
Code:
# pveversion --verbose proxmox-ve: 5.4-1 (running kernel: 4.15.18-12-pve) pve-manager: 5.4-3 (running version: 5.4-3/0a6eaa62) pve-kernel-4.15: 5.3-3 pve-kernel-4.15.18-12-pve: 4.15.18-35 pve-kernel-4.15.18-10-pve: 4.15.18-32 pve-kernel-4.15.18-9-pve: 4.15.18-30 pve-kernel-4.15.17-1-pve: 4.15.17-9 ceph: 12.2.11-pve1 corosync: 2.4.4-pve1 criu: 2.11.1-1~bpo90 glusterfs-client: 3.8.8-1 ksm-control-daemon: 1.2-2 libjs-extjs: 6.0.1-2
Last edited: