hi community, i have currently setup a cluster with 2 nodes. today i ran into a issue because i had to reboot one of the nodes for "maintenance". After making changes and booting the macine back up. i ran into no quorum
now i DO understand that
temporary fixes the no quorum issue but here is MY problem
the node that i was doing maintenance on has a pcie network card passthourgh into a pfsense vm. wich ofc results in no quorum after boot since it needs quorum to start the vm's. but because the vm cant start there is no network availability. so my dumb idear was to just remove the cluster wich ended up screwing my time
i ended up running the following script that i found on https://gist.github.com/ianchen06/73acc392c72d6680099b7efac1351f56
(^dont do this^)
without checking what was in the script i soon realised i made a verry horrible mistake. not having backups of configs i started coppying raw images and finding configs that were not effected. now 8 hours later i kinda got things working with some data loss. now you might think i dont want to make a cluster. but i dont give up so here is my question.
how can i make it that the node starts the vm's even tho there is no quorum and i am not typing in pvecm expected 1 through ssh/etc
thx for your time and as feature request i would like to have a unbind/remove all nodes from cluster/ thing in the web ui
now i DO understand that
Code:
pvecm expected 1
the node that i was doing maintenance on has a pcie network card passthourgh into a pfsense vm. wich ofc results in no quorum after boot since it needs quorum to start the vm's. but because the vm cant start there is no network availability. so my dumb idear was to just remove the cluster wich ended up screwing my time
i ended up running the following script that i found on https://gist.github.com/ianchen06/73acc392c72d6680099b7efac1351f56
Code:
echo -e "stopping Services:\npvestatd"
systemctl stop pvestatd.service
echo "pvedaemon"
systemctl stop pvedaemon.service
echo "pve-cluster.service"
systemctl stop pve-cluster.service
echo "pve-corosync"
systemctl stop corosync
echo "pve-cluster"
systemctl stop pve-cluster
echo "deleting data from db..."
#echo "select * from tree where name = 'corosync.conf';"| sqlite3 /var/lib/pve-cluster/config.db
echo "delete from tree where name = 'corosync.conf';"| sqlite3 /var/lib/pve-cluster/config.db
echo "select * from tree where name = 'corosync.conf';"| sqlite3 /var/lib/pve-cluster/config.db
#Remove directories
pmxcfs -l
rm -f /var/lib/pve-cluster/.pmxcfs.lockfile
rm -f /etc/pve/corosync.conf
rm -f /etc/corosync/*
rm -f /var/lib/corosync/*
rm -rf /etc/pve/nodes/$1/*
echo "Staring services ..."
echo "pvestatd"
systemctl start pvestatd.service
echo "pvedaemon"
systemctl start pvedaemon.service
echo "pve-cluster.service"
systemctl start pve-cluster.service
echo "pve-corosync"
systemctl start corosync
echo "pve-cluster"
systemctl restart pve-cluster
else
echo "Deleting node name missing or not existing. EXIT!"
exit
fi
without checking what was in the script i soon realised i made a verry horrible mistake. not having backups of configs i started coppying raw images and finding configs that were not effected. now 8 hours later i kinda got things working with some data loss. now you might think i dont want to make a cluster. but i dont give up so here is my question.
how can i make it that the node starts the vm's even tho there is no quorum and i am not typing in pvecm expected 1 through ssh/etc
thx for your time and as feature request i would like to have a unbind/remove all nodes from cluster/ thing in the web ui