Hello all, I'm hoping you maybe able to help me as I seem a bit in over my head this time.
I have three physical servers that I am working to build into an HA cluster. Two of those servers will be the processing power (hereafter referred to as P1 and P2) of the cluster and one is the storage...
Hi,
We're testing proxmox 5.2 running the latest enterprise version. We have a few LXC containers running on the hosts and managed through HA. When we reboot a host, all of the containers are started, but their network connections do not work. The only way to re-establish network connection to...
Hi everybody,
I'm new to this forum and this is my first post.
I am stuck in a High Availability cluster setup. I have the following configuration:
- two servers
- two storage nodes (qnap nas) with iSCSI support.
I want to create an ha-setup where:
- if one of server fails the vm still runs...
Hi,
I've a cluster of 9 nodes in 3 datacenters that looks like this:
DC 3 DC 1
+--------+ LINK 1->3 +---------------+
| P-ARB |--------------------| P1 P2 P3 P10 |
+--------+ +--------+------+
|...
I have 5-node proxmox in HA cluster. I wanna know how do the proxmox calculate the resource of Node?
Example when my 1st node down, the VM move to the highest-priority node. So when do the VM move to the next-priority node ? ( 80% RAM or 80% CPU of highest-priority node are used, etc....)
Thanks...
set CT to "ignore" in the HA web GUI
delete CT
At that point you have an entry in ha config without a node, which cannot be removed, and keeps logging errors.
Is there a way to restrict the HA relocations to target only a set of nodes?
Let's say: the VM1 can only run on SERVER1, SERVER4 and SERVER6, but not on SERVER2, SERVER3 or SERVER5.
Hello experts,
I am having this problem not 4th time in same month, my master node in 4 node cluster goes down and even though I have HA setup and nodes migrate but vms are still not reachable. Also is there any way I can also find reason why this is happening. I had to manually hard power...
Hi,
I am currently setting up a proxmox (v. 5.1-41) cluster. Since the default behaviour of nodes which can not reach other members is rebooting themselves, it would be interesting to know where I can configure this. Especially setting a limit of how many reboots a server is allowed to make...
Good morning everyone,
i just configured my back interfaces (10GbE) of my three new servers.
To provide enought redundancy and availability i connected my three servers the following way:
Hardware Config:
- Every Node has 2 10GbE SFP+ Ports that should be used only for Cluster communication...
Hello all,
I am new to proxmox community. I have 2 proxmox clusters with 2 nodes each setup with ceph storage in same local network with Public IPs.
Recently one node on one of the cluster died but the secondary node on that cluster did not bring up the VMs.
I have done all the research and I...
Hi every one
Is it possible to configure HA cluster with three nodes (2 identical physical machines, one VM) and use Ceph only in the two physical nodes
So Proxmox 5 will be installed in the three nodes and Ceph only in two nodes
I appreciate any help, thank you :)
we have a 3 nodes cluster which are correctly working with high availability,
Last night, one of our nodes had an outage, so the HA-Manager migrated all VMs to another node as expected.
But I noticed that all of our VMs get rebooted after migration, this caused some issue with VMs, since some of...
Hello,
I'm looking Solution PVE storage types.
Setup Proxmox Cluster(HA) with SAN Storage Fibre Channel.
I Setup LVM Storage types missing problem : Doesn't over-commit in the LVM space VM.
I Setup LVM-THin Storage types missing problem : Doesn't Migration VM
What should i use Storage...
Back when we were running PVE 4.3, I had received some advice not to run HA when running a hyper converged CEPH cluster because of the effects of Fencing and HA on a CEPH cluster. Now that we are running 5.1 and CEPH is now a first class citizen, I wanted to know if this is still a...
Hello,
i'm testing HA on a 4 node cluster Proxmox 5 last ver. ZFS local storage + replication.
When one node go down the vps is migrated correctly on the second node but when the failed node come back online i have this error on the failback:
task started by HA resource agent
2017-11-14...
Hi all.
Some backround on the environment first:
Four single CPU equipped DELL PowerEdge R620's.
Currently running two single ESXi nodes each with local storage and no HA or replication, one physical file server / Domain Controller and a physical backup server with tape drive.
Almost pure...
Hi,
4 HA Vms were not able to reboot properly this morning at 4:30 pm (reboot via cron).
These vms are running Centos 7 64 bits with qemu-guest-agent (hosted by different nodes).
Seems that my 3 nodes cluster is up to date :
proxmox-ve: 5.1-25 (running kernel: 4.13.4-1-pve)
pve-manager...
Hi
I know that's not officially supported, but I have no budget to spend on additional third server to provide HA
I want to prevent split-brain related problems with using bonded cross-connected (no switch between servers) network interfaces. I think that's well enough to prevent split-brain...
It would be neat to have a HA function on the gui to actually clear the error status of the given CT/VM.
HA have the bad tendency to put stuff in error when it had timeout, or generally holding some grudge against me. It'd do no harm: if it's still bad, it'll go into error again anyway...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.