Proxmox Staff Member
- Oct 1, 2014
All new Mainboards have hardware watchdogs so there is normally no need for extra device. If not softdog is a part of the kernel and works.
Would the hardware watchdog work in conjunction with shared storage fencing, e.g how Pacemaker uses sbd?All new Mainboards have hardware watchdogs so there is normally no need for extra device. If not softdog is a part of the kernel and works.
We use a distributed locking mechanism, combined with the watchdog feature.
wolfgang said:It doesn't matter if the the node is not reachable from the cluster or off.
consider if the rest cluster has quorum then the cluster know everything is ok and the one node what is missing is not ok.
Our software solves exactly that problem. Or what do you think the software is for?[/COLOR]The concern I have is if the rogue node were to "recover" from the failed state and then start writing bad data to the shared storage (e.g DRBD, ceph, NFS, etc) at the same time as the new node that took over its VMs. If this were to occur, wouldn't it result in two nodes writing data to the same VM image files at once and thus corrupting them?
The mechanism that it uses to solve that problem is generic in that it can work with any shared storage backend (DRBD, NFS, Ceph, Gluster, iSCSI, etc)? Does it intercept I/O requests to the shared storage or how does it work?Our software solves exactly that problem. Or what do you think the software is for?
Just to note, a two-node cluster was never a recommended or fully supported configuration.I see in the docs that the minimum number of nodes for an ha confiuration is 3. Is a 2 node cluster no longer supported?