Hello,
Our problems started this week after we tried to upgrade a node in a 4-node cluster from 4.4. to 5.0
The upgrade itself went fine. After first reboot the whole cluster went offline due to fencing.
This repeated every single time we tried to bring back this node online again.
So, I...
Hello,
We have a failed Proxmox 4.4->5.0 upgraded node that makes all cluster nodes (remaining 4) to fence themselves whenever it comes online. Seems I will have to reinstall the node for whatever reason.
I've just noticed something strange in corosync.conf while backing up the configuration...
That's what I reckoned as well, but why warnings of apocalypse from the documentation?
So, if I want to reinstall or add a new hardware node with previously "blacklisted" hostname, all I have to do is remove keys from /root/.ssh?
It would be good to have some info on impact of using the same hostname in the same subnet for something else (we, for instance, recycle the names and name hosts sequentially). Also, what if I want to reinstall Proxmox host for whatever reason and rejoin under same name?
All my NIC's are in use and I was thinking of a more temporary solution till we replace the switch. And I still need HA. This issue happens twice per month.
Hello,
We are currently having issues with a switch stack connecting our cluster nodes. The switch restarts intermittently and cluster nodes fence themselves and restart. This loss of communications lasts between 60-120 seconds.
Until we replace the switch, is there any way to temporarily...
It is strange to be sure.
journalctl:
Jun 08 09:46:50 03 kernel: EXT4-fs (dm-7): mounted filesystem with ordered data mode. Opts: (null)
Jun 08 09:46:50 03 kernel: IPv6: ADDRCONF(NETDEV_UP): veth114i0: link is not ready
Jun 08 09:46:51 03 kernel: device veth114i0 entered promiscuous mode...
Ok, so maybe a bit of background. We have a 4 node cluster. This is happening on one of them since a recent reboot of the host. Before that reboot there were no issues with containers.
These containers init properly when migrated to other hosts. Whenever I migrate a CT over to the host in...
Hello,
As of couple of days ago, lxc containers won't run their designated runlevel (centos 6.x container):
USER PID %CPU %MEM VSZ RSS TTY STAT START TIME COMMAND
root 1 0.0 0.0 19292 2324 ? Ss 10:45 0:00 /sbin/init
root 98 0.0 0.0 11500...
Hello,
Working daily with LXC in Proxmox cluster, there's really one thing I would like to see implemented, and that's to be able to use commands like pct list/enter/whatever on any lxc container regardless on which host you are currently logged on.
For instance, a global pct list (showing all...
Ok, thanks.
Any way to assure that no quorum shenanigans occur, since I cannot ensure no race condition happens? Something like pvecm expect 1 before shut? Does that make sense?
Hi,
We have a 4-node cluster running LXC over shared storage (LVM based). All VM's are HA-enabled
So, if I do a simple "init 0" on all hosts, all VM's will shut down without crazy stuff (VM restarts, migration hangs) going on because of quorum loss and shutdown race conditions etc?
Hello,
As stated in title - what's the best way to shutdown a Proxmox cluster gracefully either by running a script on each node or preferably from a central location?
I have to implement and test an emergency shutdown scenario in case of severe power outage. The script should be triggered by...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.