Hello,
The flapping interface issue have been resolved. It was a misconfiguration of the interfaces as well as a problem of ring redundancy on corosync that is now resolved.
The HA problem persists thought and the syslog of the node that tries to start the HA VM shows possible problems on the...
Here is the full syslog between the moment when the initial host is stopped and when it is back on:
Mar 27 11:10:32 srv-pve3 corosync[2197]: error [TOTEM ] Marking ringid 1 interface 192.168.1.103 FAULTY
Mar 27 11:10:32 srv-pve3 corosync[2197]: [TOTEM ] Marking ringid 1 interface...
Hello,
You might have found the culprit.
On all nodes, one of the interfaces flaps:
Mar 27 10:46:16 srv-pve2 corosync[2333]: error [TOTEM ] Marking ringid 1 interface 192.168.1.102 FAULTY
Mar 27 10:46:16 srv-pve2 corosync[2333]: [TOTEM ] Marking ringid 1 interface 192.168.1.102 FAULTY
Mar...
For more information, the cluster fails sometime to start a VM, when migrating it between 2 nodes:
()
task started by HA resource agent
2019-03-26 10:45:25 use dedicated network address for sending migration traffic (10.0.0.102)
2019-03-26 10:45:26 starting migration of VM 106 to node 'srv-pve2'...
Thanks for you reply :)
The qm monitor is not successful:
root@srv-pve3:~# qm monitor 106
Entering Qemu Monitor for VM 106 - type 'help' for help
qm> stop
ERROR: VM 106 qmp command 'human-monitor-command' failed - unable to connect to VM 106 qmp socket - timeout after 31 retries
Trying to...
I've got the exact same issue for a different VM on a different HA Group for a different node:
()
task started by HA resource agent
TASK ERROR: start failed: command '/usr/bin/kvm -id 106 -name SRV-APT-REPO -chardev 'socket,id=qmp,path=/var/run/qemu-server/106.qmp,server,nowait' -mon...
Hello all,
I'm experimenting with HA clustering and I ran into a possible bug.
VM is configured with HA group to preferably run on the node 3:
root@srv-pve1:~# cat /etc/pve/ha/groups.cfg
group: PVE3_First
comment PVE3 is prefered for this group
nodes srv-pve3
nofailback 0...
Hi all,
VM migration fails at an inconstant rate between nodes:
2019-03-19 13:59:14 use dedicated network address for sending migration traffic (10.0.0.102)
2019-03-19 13:59:14 starting migration of VM 108 to node 'srv-pve2' (10.0.0.102)
2019-03-19 13:59:14 copying disk images
2019-03-19...
Okay, great that makes sense!
I've got another question, related to the former subject.
How should I configure Proxmox if I want to make a VM available at all time, even if the underlying node gets interrupted?
The replication on the host doesn't seem to work with Ceph storage (no replicable...
Hello all,
I'm wondering why the VM Migration between nodes on a 3 node cluster in a Ceph environment is taking time at all.
If I understand the underlying technology correctly, Ceph replicates the data on all osd and nodes to be available at all time on all nodes.
Nevertheless, when I...
Hello,
after careful analysis, it seems impossible to expose the Ceph cluster to Windows in the current version of Proxmox as it requires to run a Luminous on top of a Linux kernel v4.17 which is not the production version of Proxmox at the time being.
Said differently, would it be secure to add the Ceph repositories to a vanilla installation of Proxmox in order to access the binaries required (ceph-iscsi-cli) to expose the cluster to Windows clients?
Hi folks,
I might need some storage space on a Windows server and my Ceph cluster currently has 110Tb available.
I'd like to expose this free space to the Windows server using the ceph-iscsi-cli as described in the official Ceph documentation.
Since it is based on a vanilla install of Ceph...
Hello,
I've got a cluster that is running smoothly but, in some cases, mass migration is not working properly.
The VM, once migrated on the destination node won't boot up with the following error:
It happened on 4 out of 8 VM that I've migrated with no explanation.
The migrated VM can be...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.