Would the hardware watchdog work in conjunction with shared storage fencing, e.g how Pacemaker uses sbd?All new Mainboards have hardware watchdogs so there is normally no need for extra device. If not softdog is a part of the kernel and works.
In any of these cases, how do the other nodes know for certain that the rogue node has been killed?
We use a distributed locking mechanism, combined with the watchdog feature.
wolfgang said:It doesn't matter if the the node is not reachable from the cluster or off.
consider if the rest cluster has quorum then the cluster know everything is ok and the one node what is missing is not ok.
[/COLOR]The concern I have is if the rogue node were to "recover" from the failed state and then start writing bad data to the shared storage (e.g DRBD, ceph, NFS, etc) at the same time as the new node that took over its VMs. If this were to occur, wouldn't it result in two nodes writing data to the same VM image files at once and thus corrupting them?
The mechanism that it uses to solve that problem is generic in that it can work with any shared storage backend (DRBD, NFS, Ceph, Gluster, iSCSI, etc)? Does it intercept I/O requests to the shared storage or how does it work?Our software solves exactly that problem. Or what do you think the software is for?
Does it intercept I/O requests to the shared storage or how does it work?
It is possible to place LXC container on Ceph RBD storage ?
I see in the docs that the minimum number of nodes for an ha confiuration is 3. Is a 2 node cluster no longer supported?
I see in the docs that the minimum number of nodes for an ha confiuration is 3. Is a 2 node cluster no longer supported?