I am relatively new to Proxmox and have a cluster running with 3 nodes, everything is currently working fine, cluster is up HA is running fine. The issue I face currently is that in case the cluster link goes down for let's say more than 10s, watchdog kicks in and reboots the server, this causes...
Hello together,
I'd like to build a PVE HA Cluster out of 2 PVE Nodes and 1 QDev to get quorum.
In order to get a nice and stable Corosync Link, I've a dedicated 1G NIC via Crossover LAN between the 2 PVE nodes.
The QDev VM is a external hosted system and can't be connected via LAN.
The Plan...
Hi everyone,
I'm currently facing a problem when adding a QDevice to a 2-node cluster.
Current cluster: 2x Proxmox VE 8.2.2
QDevice: 1x Proxmox Backup-Server 3.2-2
- I can remote from both nodes into PBS as root
- corosync-qdevice is installed on both nodes
- corosync-qnetd and...
Good morning,
since setting up a cluster with 3 nodes (all running PVE 8.2.2), we've been experiencing an issue where they no longer communicate with each other after a maximum of 24 hours. The hosts are directly connected to each other via a switch. There is no load on the ports in the switch...
I especially want to know what protection mechanism the PVE cluster has to allow the host to automatically restart.
Environment: There are 13 hosts in the cluster: node1-13
Version: pve-manager/6.4-4/337d6701 (running kernel: 5.4.106-1-pve)
Web environment:
There are two switches A and B...
Hi, i am using Proxmox with ZFS replication and HA, i have set the /etc/pve/corsync.conf:
quorum_votes: 1
two_nodes: yes
but everytime the pve1 crashes the pve2 stays on "waiting for quorum", when pve1 is online again the option is back to default, how to solve this?
Hallo zusammen,
ich habe ein Cluster konfiguriert, wo nicht alle Nodes die gleiche Anzahl an Votes haben soll.
Grund dafür ist deren Standort. Soweit so gut.
Allerdings werden nach aktueller Konfiguration 17 Votes benötigt um Quorum zu erreichen.
Aufgrund der Konstellation reichen allerdings...
Hi all,
I would like to cluster some VMs within my Proxmox Homer Lab, using Pacemaker and Corosync. i world then like to create a couple of Shared Storage Devices between the nodes, all sharing the same respective "physical" storage devices underneath. The idea is to create two, one being a...
So, as the title says, I am deploying all new Proxmox servers to replace our aging fleet of 2U Dells. Currently, I have a 10G trunk for all of my normal VLANs and a separate 10G connection specific to only Corosync VLAN traffic. My new servers have 4 x 10G NICs and 2 x 100G NICs each.
I was...
I'm having issues getting corosync to start up, causing my node to be unable to connect to the other.
Diagnostics so far
I've tested the basics like pinging one node from the other and it works fine.\
Results from journalctl -xeu pve-cluster.service
Jan 15 18:15:49 pve847 pmxcfs[122430]...
There's a great deal of misleading piece of argument to be found in currently official PVE docs on QDevices [1] and then some conscious effort to take it even further.
Under "Supported setups" [1] the following is advised (emphasis mine):
It continues to provide an absurd piece of reasoning...
I have a small cluster of just 2 devices, I will be adding a 3rd in the not too distant future.
One of the nodes has 10gb uplink to external, the other only has 1gb. This means that synching between the two is limited to 1GB of course. I've added a 10gb network card to the node with 1GB...
It's nowhere in the PVE official docs, but corosync does support last_man_standing and when used with HA it is suggested to also set wait_for_all. I found some previous threads, but not in relation to HA.
Now I understand the official PVE endorsed way would be to just use a qdevice, but this...
Hi folks!
I have a small homelab and I had 4 PVE nodes: pve0, pve1, pve2 and pve3 in 1 datacenter/cluster. I was looking to turn pve3 into PBS, so I simply disconnected the hardware and installed PBS there.
When I login to GUI on pve0/pve1/pve2 I still can see pve3 and I'd like to get rid of...
Hello,
I faced with a problem, where I had to rejoin same node to the cluster, but the following errors appeared:
Please enter superuser (root) password for 'IP ADDR': **************
detected the following error(s):
* authentication key '/etc/corosync/authkey' already exists
* cluster config...
What is the best practice for the corosync network? I can't imagine it having much bandwidth requirements, and I'd hate to use a seperate nic just for this. seems to just need low latency.
Since I have ceph public and cluster networks separated on separate 10G Nics, Can I put corosync on the...
Hello Proxmox Community,
I hope this post finds you well. I'm currently facing an issue with my Proxmox setup, and I'm seeking advice on how to gracefully remove a node from a Proxmox cluster.
Issue Summary: I accidentally created a Proxmox cluster with only one node, and now I'd like to...
Hi guys,
In the company where I'm working, we have a Cluster composed of 8 nodes that are connected through a 1Gb ethernet network directly to a Switch.
On the nodes we mostly run Kubernetes, DNS Servers and some other services. The problem is that on each node, the process corosync is using...
I tried to rename two of my nodes in my lab, and as I expected bricked the cluster configuration.
Then I removed the cluster configs and created a new one, unfortunately I can't join to it..... what could be the reason for that?
root@sofx1010pve3302.home.lan:~# pvecm add 192.168.30.7 -use_ssh...
Hello,
we have a 3 node PVE cluster (7.4-16), with separate cluster interfaces. The cluster interfaces are in a bond (LACP). The cluster interfaces are connected to two Cisco Nexus switches.
Over the weekend our complete HA cluster failed. According to the logs of the other servers, the...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.