Migration Problems Proxmox

rcmp16

New Member
Feb 7, 2020
5
0
1
29
Hi, I'm having a problem with 2 of 4 vms during migration process. I dont know what to do, I tried via CLI and the issue still happenning.

Bash:
root@hipervisor2:~# qm migrate 100 hipervisor1  --online --with-local-disks
2020-02-07 10:41:14 starting migration of VM 100 to node 'hipervisor1' (192.168.10.240)
2020-02-07 10:41:15 found local disk 'MV:vm-100-disk-0' (in current VM config)
2020-02-07 10:41:15 copying disk images
2020-02-07 10:41:15 starting VM 100 on remote node 'hipervisor1'
2020-02-07 10:41:16 start failed: command '/usr/bin/kvm -id 100 -name ServerAppTSSCH -chardev 'socket,id=qmp,path=/var/run/qemu-server/100.qmp,server,nowait' -mon 'chardev=qmp,mode=control' -chardev 'socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5' -mon 'chardev=qmp-event,mode=control' -pidfile /var/run/qemu-server/100.pid -daemonize -smbios 'type=1,serial=2M293103Q8,product=ProLiantDL380Gen10,family=ProLiant,uuid=1a0fc209-0f42-492d-a62c-6c227e0c44c4,sku=P06421-B21,manufacturer=HPE,version=U30' -smp '4,sockets=2,cores=2,maxcpus=4' -nodefaults -boot 'menu=on,strict=on,reboot-timeout=1000,splash=/usr/share/qemu-server/bootsplash.jpg' -vnc unix:/var/run/qemu-server/100.vnc,password -no-hpet -cpu 'kvm64,+lahf_lm,+sep,+kvm_pv_unhalt,+kvm_pv_eoi,hv_spinlocks=0x1fff,hv_vapic,hv_time,hv_reset,hv_vpindex,hv_runtime,hv_relaxed,hv_synic,hv_stimer,hv_tlbflush,hv_ipi,enforce' -m 16384 -device 'pci-bridge,id=pci.1,chassis_nr=1,bus=pci.0,addr=0x1e' -device 'pci-bridge,id=pci.2,chassis_nr=2,bus=pci.0,addr=0x1f' -device 'vmgenid,guid=4c23e552-06b2-4293-b2ce-4eca60bf02ee' -device 'piix3-usb-uhci,id=uhci,bus=pci.0,addr=0x1.0x2' -device 'usb-tablet,id=tablet,bus=uhci.0,port=1' -device 'VGA,id=vga,bus=pci.0,addr=0x2' -device 'virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x3' -iscsi 'initiator-name=iqn.1993-08.org.debian:01:75d4b580e567' -drive 'file=/dev/zvol/rpool/vm-100-disk-1,if=none,id=drive-ide0,format=raw,cache=none,aio=native,detect-zeroes=on' -device 'ide-hd,bus=ide.0,unit=0,drive=drive-ide0,id=ide0,bootindex=100' -drive 'if=none,id=drive-ide2,media=cdrom,aio=threads' -device 'ide-cd,bus=ide.1,unit=0,drive=drive-ide2,id=ide2,bootindex=200' -netdev 'type=tap,id=net0,ifname=tap100i0,script=/var/lib/qemu-server/pve-bridge,downscript=/var/lib/qemu-server/pve-bridgedown' -device 'e1000,mac=36:FA:5F:1D:FB:85,netdev=net0,bus=pci.0,addr=0x12,id=net0,bootindex=300' -rtc 'driftfix=slew,base=localtime' -machine 'type=pc-i440fx-4.0' -global 'kvm-pit.lost_tick_policy=discard' -incoming unix:/run/qemu-server/100.migrate -S' failed: exit code 1
2020-02-07 10:41:16 ERROR: online migrate failure - command '/usr/bin/ssh -e none -o 'BatchMode=yes' -o 'HostKeyAlias=hipervisor1' root@192.168.10.240 qm start 100 --skiplock --migratedfrom hipervisor2 --migration_type secure --stateuri unix --machine pc-i440fx-4.0 --targetstorage 1' failed: exit code 255
2020-02-07 10:41:16 aborting phase 2 - cleanup resources
2020-02-07 10:41:16 migrate_cancel
2020-02-07 10:41:17 ERROR: migration finished with problems (duration 00:00:03)
migration problems


Hipervisor2 version
Bash:
root@hipervisor2:~# pveversion -t
Unknown option: t
USAGE: pveversion [--verbose]
root@hipervisor2:~# pveversion -v
proxmox-ve: 6.0-2 (running kernel: 5.0.15-1-pve)
pve-manager: 6.0-4 (running version: 6.0-4/2a719255)
pve-kernel-5.0: 6.0-5
pve-kernel-helper: 6.0-5
pve-kernel-5.0.15-1-pve: 5.0.15-1
ceph-fuse: 12.2.11+dfsg1-2.1
corosync: 3.0.2-pve2
criu: 3.11-3
glusterfs-client: 5.5-3
ksm-control-daemon: 1.3-1
libjs-extjs: 6.0.1-10
libknet1: 1.10-pve1
libpve-access-control: 6.0-2
libpve-apiclient-perl: 3.0-2
libpve-common-perl: 6.0-2
libpve-guest-common-perl: 3.0-1
libpve-http-server-perl: 3.0-2
libpve-storage-perl: 6.0-5
libqb0: 1.0.5-1
lvm2: 2.03.02-pve3
lxc-pve: 3.1.0-61
lxcfs: 3.0.3-pve60
novnc-pve: 1.0.0-60
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.0-5
pve-cluster: 6.0-4
pve-container: 3.0-3
pve-docs: 6.0-4
pve-edk2-firmware: 2.20190614-1
pve-firewall: 4.0-5
pve-firmware: 3.0-2
pve-ha-manager: 3.0-2
pve-i18n: 2.0-2
pve-qemu-kvm: 4.0.0-3
pve-xtermjs: 3.13.2-1
qemu-server: 6.0-5
smartmontools: 7.0-pve2
spiceterm: 3.1-1
vncterm: 1.6-1
zfsutils-linux: 0.8.1-pve1

Hipervisor1 (The one that i want migrate the VM)
Bash:
root@hipervisor1:~# pveversion -v
proxmox-ve: 6.0-2 (running kernel: 5.0.15-1-pve)
pve-manager: 6.0-4 (running version: 6.0-4/2a719255)
pve-kernel-5.0: 6.0-5
pve-kernel-helper: 6.0-5
pve-kernel-5.0.15-1-pve: 5.0.15-1
ceph: 14.2.4-pve1
ceph-fuse: 14.2.4-pve1
corosync: 3.0.2-pve2
criu: 3.11-3
glusterfs-client: 5.5-3
ksm-control-daemon: 1.3-1
libjs-extjs: 6.0.1-10
libknet1: 1.10-pve1
libpve-access-control: 6.0-2
libpve-apiclient-perl: 3.0-2
libpve-common-perl: 6.0-2
libpve-guest-common-perl: 3.0-1
libpve-http-server-perl: 3.0-2
libpve-storage-perl: 6.0-5
libqb0: 1.0.5-1
lvm2: 2.03.02-pve3
lxc-pve: 3.1.0-61
lxcfs: 3.0.3-pve60
novnc-pve: 1.0.0-60
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.0-5
pve-cluster: 6.0-4
pve-container: 3.0-3
pve-docs: 6.0-4
pve-edk2-firmware: 2.20190614-1
pve-firewall: 4.0-5
pve-firmware: 3.0-2
pve-ha-manager: 3.0-2
pve-i18n: 2.0-2
pve-qemu-kvm: 4.0.0-3
pve-xtermjs: 3.13.2-1
qemu-server: 6.0-5
smartmontools: 7.0-pve2
spiceterm: 3.1-1
vncterm: 1.6-1
zfsutils-linux: 0.8.1-pve1

Thanks in advance!
 
What's your storage config? (cat /etc/pve/storage.cfg) Is the storage available on both nodes?

I'd recommend upgrading to the latest version. There have been lots of improvements and bug fixes.
 
What's your storage config? (cat /etc/pve/storage.cfg) Is the storage available on both nodes?

I'd recommend upgrading to the latest version. There have been lots of improvements and bug fixes.

Yes, they are available. I tried with two other vms that i have and they migrate without any problems but with the productions vms happens that I wrote above.

Hipervisor2 cat /etc/pve/storage.cfg
Code:
 cat /etc/pve/storage.cfg
dir: local
        path /var/lib/vz
        content backup,vztmpl,iso

zfspool: local-zfs
        pool rpool/data
        content images,rootdir
        sparse 1

zfspool: MV
        pool rpool
        content rootdir,images
        nodes hipervisor2,hipervisor3,hipervisor1
        sparse 0

dir: ISO
        path /ISO
        content iso
        nodes hipervisor2,hipervisor3,hipervisor1
        shared 1

dir: BACKUP
        path /BACKUP
        content images,backup,rootdir
        maxfiles 4
        nodes hipervisor1,hipervisor2,hipervisor3
        shared 0

dir: PLANTILLAS
        path /PLANTILLAS
        content vztmpl
        nodes hipervisor1,hipervisor2,hipervisor3
        shared 1

nfs: nfs
        export /mnt/Vol1
        path /mnt/pve/nfs
        server 192.168.10.48
        content backup
        maxfiles 100

dir: DISCO_EXTERNO
        path /DISCO_EXTERNO
        content images
        shared 0

Hipervisor1 cat /etc/pve/storage.cfg
Code:
root@hipervisor1:~# cat /etc/pve/storage.cfg
dir: local
        path /var/lib/vz
        content backup,vztmpl,iso

zfspool: local-zfs
        pool rpool/data
        content images,rootdir
        sparse 1

zfspool: MV
        pool rpool
        content rootdir,images
        nodes hipervisor2,hipervisor3,hipervisor1
        sparse 0

dir: ISO
        path /ISO
        content iso
        nodes hipervisor2,hipervisor3,hipervisor1
        shared 1

dir: BACKUP
        path /BACKUP
        content images,backup,rootdir
        maxfiles 4
        nodes hipervisor1,hipervisor2,hipervisor3
        shared 0

dir: PLANTILLAS
        path /PLANTILLAS
        content vztmpl
        nodes hipervisor1,hipervisor2,hipervisor3
        shared 1

nfs: nfs
        export /mnt/Vol1
        path /mnt/pve/nfs
        server 192.168.10.48
        content backup
        maxfiles 100

dir: DISCO_EXTERNO
        path /DISCO_EXTERNO
        content images
        shared 0
 
Please post the VM config (qm config <vmid>) and the config of another VM that works.
 
Please post the VM config (qm config <vmid>) and the config of another VM that works.

VM config that migration works:
Code:
root@hipervisor1:~# qm config 301
bootdisk: ide0
cores: 2
ide0: MV:vm-301-disk-0,format=raw,size=50G
ide2: none,media=cdrom
memory: 5120
name: Linux
net0: e1000=36:3D:9B:C0:EF:D7,bridge=vmbr3,firewall=1
numa: 0
ostype: other
scsihw: lsi
smbios1: uuid=f153cf32-e053-404c-b7e9-7e6e5c097c9e
sockets: 3
vga: std,memory=16
vmgenid: 8f9b85c8-37c8-4a11-9168-cc708ecbc388

VMs configs that migration is not working:


Code:
root@hipervisor2:~# qm config 100
bootdisk: ide0
cores: 2
description: smbios1%3A uuid=13019b47-8f90-4fe1-a1ee-a3d991f7be62
ide0: MV:vm-100-disk-0,size=500G
ide2: none,media=cdrom
memory: 16384
name: ServerAppTSSCH
net0: e1000=36:FA:5F:1D:FB:85,bridge=vmbr4,firewall=1
numa: 0
onboot: 1
ostype: win10
smbios1: base64=1,family=UHJvTGlhbnQ=,manufacturer=SFBF,product=UHJvTGlhbnRETDM4MEdlbjEw,serial=Mk0yOTMxMDNROA==,sku=UDA2NDIxLUIyMQ==,uuid=1a0fc209-0f42-492d-a62c-6c227e0c44c4,version=VTMw
sockets: 2
vmgenid: 4c23e552-06b2-4293-b2ce-4eca60bf02ee


Code:
root@hipervisor2:~# qm config 200
bootdisk: ide0
cores: 5
description: smbios1%3A uuid=13019b47-8f90-4fe1-a1ee-a3d991f7be62
ide0: MV:vm-200-disk-1,size=1000G
ide2: none,media=cdrom
memory: 16384
name: ServerSQLTSSCH
net0: e1000=F6:04:9D:2F:B9:90,bridge=vmbr4,firewall=1
numa: 0
onboot: 1
ostype: win10
parent: Cunrrently
smbios1: base64=1,family=UHJvTGlhbnQ=,manufacturer=SFBF,product=UHJvTGlhbnRETDM4MEdlbjEw,serial=Mk0yOTMxMDNROA==,sku=UDA2NDIxLUIyMQ==,uuid=f61ef628-89c2-4bd2-925b-064af467ade9,version=VTMw
sockets: 2
usb0: spice
vmgenid: 95b2efa8-8b77-4ee4-a137-a51867fe3257
 
Is there enough free RAM on the node you want to migrate to? Does it work to the 3rd node in the cluster?
 
Is there enough free RAM on the node you want to migrate to? Does it work to the 3rd node in the cluster?


Tried and throw the same migration error.

Yes there is enough free RAM I think.

Proxmox Metrics of the node that I want to migrate:
RAM usage
55.70% (34.85 GiB of 62.56 GiB)
 
Does offline migration work when the VM is then started on the other node?
 
It looks like the error happens when the VM is started on the other node. So to test that it is possible to start the VM, the offline migration followed by a VM can help narrow it down. Did you also start the VM on the other node after the offline migration?
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!