Hi
My setup includes:
3 PVE hyper-visors (pve1, pve2 and pve3) installed as nested VM in Hyper-V on Windows 10. All join to cluster.
2 of them have CEPH OSD 60G each forming 120G pool and they are forming ha group (pve1 and pve2).
The CEPH Nautilus is installed and set on all the 3 members.
Ceph includes 3 monitors and 3 managers (1 on each cluster member respectively).
3 MSD as well in same manner.
I have 1 VM running on the cluster, member of HA group and can run on pve1 or pve2 HA group members.
The PVE3 installed only to form a 3 peers quotum.
I can successfully migrate machines between PVE1 and PVE2, takes seconds with ~1 ping packet loss.
Also tried disk move - works as well.
When I am trying to disconnect pve server that the VM currently running on it, quorum votes to move the VM to another pve ha cluster member.
It starts the process but after a while getting "timeout".
What am I doing wrong ?
The error I am getting is :
--------------------------------------------------
task started by HA resource agent
TASK ERROR: start failed: command '/usr/bin/kvm -id 100 -name Centos-ceph -chardev 'socket,id=qmp,path=/var/run/qemu-server/100.qmp,server,nowait' -mon 'chardev=qmp,mode=control' -chardev 'socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5' -mon 'chardev=qmp-event,mode=control' -pidfile /var/run/qemu-server/100.pid -daemonize -smbios 'type=1,uuid=7504c566-e10f-4cdf-bf50-1217d99fbd2a' -smp '1,sockets=1,cores=1,maxcpus=1' -nodefaults -boot 'menu=on,strict=on,reboot-timeout=1000,splash=/usr/share/qemu-server/bootsplash.jpg' -vnc unix:/var/run/qemu-server/100.vnc,password -cpu qemu64 -m 1024 -device 'pci-bridge,id=pci.1,chassis_nr=1,bus=pci.0,addr=0x1e' -device 'pci-bridge,id=pci.2,chassis_nr=2,bus=pci.0,addr=0x1f' -device 'vmgenid,guid=6f56f77e-235b-4bbf-a43f-3d28ec1595b7' -device 'piix3-usb-uhci,id=uhci,bus=pci.0,addr=0x1.0x2' -device 'usb-tablet,id=tablet,bus=uhci.0,port=1' -device 'VGA,id=vga,bus=pci.0,addr=0x2' -chardev 'socket,path=/var/run/qemu-server/100.qga,server,nowait,id=qga0' -device 'virtio-serial,id=qga0,bus=pci.0,addr=0x8' -device 'virtserialport,chardev=qga0,name=org.qemu.guest_agent.0' -device 'virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x3' -iscsi 'initiator-name=iqn.1993-08.org.debian:01:241650b23c9' -drive 'if=none,id=drive-ide2,media=cdrom,aio=threads' -device 'ide-cd,bus=ide.1,unit=0,drive=drive-ide2,id=ide2,bootindex=200' -device 'virtio-scsi-pci,id=scsihw0,bus=pci.0,addr=0x5' -drive 'file=rbd:ceph_pool_1/vm-100-disk-0:conf=/etc/pve/ceph.conf:id=admin:keyring=/etc/pve/priv/ceph/ceph_pool_1.keyring,if=none,id=drive-scsi0,format=raw,cache=none,aio=native,detect-zeroes=on' -device 'scsi-hd,bus=scsihw0.0,channel=0,scsi-id=0,lun=0,drive=drive-scsi0,id=scsi0,bootindex=100' -netdev 'type=tap,id=net0,ifname=tap100i0,script=/var/lib/qemu-server/pve-bridge,downscript=/var/lib/qemu-server/pve-bridgedown' -device 'e1000,mac=C6:43:59:82:CA:65,netdev=net0,bus=pci.0,addr=0x12,id=net0,bootindex=300' -machine 'accel=tcg,type=pc+pve1'' failed: got timeout
------------------------------------------------------------------------
My setup includes:
3 PVE hyper-visors (pve1, pve2 and pve3) installed as nested VM in Hyper-V on Windows 10. All join to cluster.
2 of them have CEPH OSD 60G each forming 120G pool and they are forming ha group (pve1 and pve2).
The CEPH Nautilus is installed and set on all the 3 members.
Ceph includes 3 monitors and 3 managers (1 on each cluster member respectively).
3 MSD as well in same manner.
I have 1 VM running on the cluster, member of HA group and can run on pve1 or pve2 HA group members.
The PVE3 installed only to form a 3 peers quotum.
I can successfully migrate machines between PVE1 and PVE2, takes seconds with ~1 ping packet loss.
Also tried disk move - works as well.
When I am trying to disconnect pve server that the VM currently running on it, quorum votes to move the VM to another pve ha cluster member.
It starts the process but after a while getting "timeout".
What am I doing wrong ?
The error I am getting is :
--------------------------------------------------
task started by HA resource agent
TASK ERROR: start failed: command '/usr/bin/kvm -id 100 -name Centos-ceph -chardev 'socket,id=qmp,path=/var/run/qemu-server/100.qmp,server,nowait' -mon 'chardev=qmp,mode=control' -chardev 'socket,id=qmp-event,path=/var/run/qmeventd.sock,reconnect=5' -mon 'chardev=qmp-event,mode=control' -pidfile /var/run/qemu-server/100.pid -daemonize -smbios 'type=1,uuid=7504c566-e10f-4cdf-bf50-1217d99fbd2a' -smp '1,sockets=1,cores=1,maxcpus=1' -nodefaults -boot 'menu=on,strict=on,reboot-timeout=1000,splash=/usr/share/qemu-server/bootsplash.jpg' -vnc unix:/var/run/qemu-server/100.vnc,password -cpu qemu64 -m 1024 -device 'pci-bridge,id=pci.1,chassis_nr=1,bus=pci.0,addr=0x1e' -device 'pci-bridge,id=pci.2,chassis_nr=2,bus=pci.0,addr=0x1f' -device 'vmgenid,guid=6f56f77e-235b-4bbf-a43f-3d28ec1595b7' -device 'piix3-usb-uhci,id=uhci,bus=pci.0,addr=0x1.0x2' -device 'usb-tablet,id=tablet,bus=uhci.0,port=1' -device 'VGA,id=vga,bus=pci.0,addr=0x2' -chardev 'socket,path=/var/run/qemu-server/100.qga,server,nowait,id=qga0' -device 'virtio-serial,id=qga0,bus=pci.0,addr=0x8' -device 'virtserialport,chardev=qga0,name=org.qemu.guest_agent.0' -device 'virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x3' -iscsi 'initiator-name=iqn.1993-08.org.debian:01:241650b23c9' -drive 'if=none,id=drive-ide2,media=cdrom,aio=threads' -device 'ide-cd,bus=ide.1,unit=0,drive=drive-ide2,id=ide2,bootindex=200' -device 'virtio-scsi-pci,id=scsihw0,bus=pci.0,addr=0x5' -drive 'file=rbd:ceph_pool_1/vm-100-disk-0:conf=/etc/pve/ceph.conf:id=admin:keyring=/etc/pve/priv/ceph/ceph_pool_1.keyring,if=none,id=drive-scsi0,format=raw,cache=none,aio=native,detect-zeroes=on' -device 'scsi-hd,bus=scsihw0.0,channel=0,scsi-id=0,lun=0,drive=drive-scsi0,id=scsi0,bootindex=100' -netdev 'type=tap,id=net0,ifname=tap100i0,script=/var/lib/qemu-server/pve-bridge,downscript=/var/lib/qemu-server/pve-bridgedown' -device 'e1000,mac=C6:43:59:82:CA:65,netdev=net0,bus=pci.0,addr=0x12,id=net0,bootindex=300' -machine 'accel=tcg,type=pc+pve1'' failed: got timeout
------------------------------------------------------------------------