[High availability] VM not moved

M

Mirage

Guest
Hi,

I set up a cluster of 2 PVE 2.0 servers.

* I have an iSCSI shared storage between the 2 nodes.
* Fencing is configured and working.
* I can use fence_node on one node to reboot the other node.
* ACPI is disabled at kernel boot (acpi=off)
* There is no option to disable ACPI in bios, but nmi button is enabled (servers are Dell PE R710)


The problem occurs with running VMs not being moved to the other node when the host is fenced :
1. VM running on node 1
2. node 2 issue 'fence_node node1'
3. node 1 reset (looks like power outlet are removed, no services are properly stopped), VM is keept on node1
4. ASAP node 1 is back alive, VM is rebooted on node 2.

cluser.conf :
Code:
<?xml version="1.0"?>
<cluster config_version="16" name="PROXMOX">
  <cman keyfile="/var/lib/pve-cluster/corosync.authkey"/>
  <fencedevices>
    <fencedevice agent="fence_drac5" cmd_prompt="admin1->" ipaddr="172.20.70.123" login="fence" name="PROX01" passwd="password" secure="1"/>
    <fencedevice agent="fence_drac5" cmd_prompt="admin1->" ipaddr="172.20.70.124" login="fence" name="PROX02" passwd="password" secure="1"/>
  </fencedevices>
  <clusternodes>
    <clusternode name="PROX01" nodeid="1" votes="1">
      <fence>
        <method name="1">
          <device name="PROX01"/>
        </method>
      </fence>
    </clusternode>
    <clusternode name="PROX02" nodeid="2" votes="1">
      <fence>
        <method name="1">
          <device name="PROX02"/>
        </method>
      </fence>
    </clusternode>
  </clusternodes>
  <rm>
    <pvevm autostart="1" vmid="101"/>
  </rm>
</cluster>


What makes my VM not moving the the node still alive ?
 
Thank you Dietmar,

Can you tell me what is the mecanism that choose on which node a VM is moved when its hosting node is failing ?
Is the new hosting node choose regarding its load? its number of active VMs? Can it be user-defined ?
 
Last edited by a moderator:
Can you tell me what is the mecanism that choose on which node a VM is moved when its hosting node is failing ?

You have only two node, so there is no choice. (I guess I do not really understand your question)
 
Last edited:
I'm sorry I was not clear enough.

As I can't have HA the way I want with my 2-nodes setup, I've planned to add a 3rd node soon. That way all my VMs will be moved to another node if their hosting node is failing (Am I correct on this point ?)

My question was : In a more-than-2-nodes setup, when one is failing, how does PVE choose on which node a VM shoul go ?
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!