ha-manager migrated too much load to same node

tri-dp

Member
Nov 11, 2022
17
0
6
Hi,

Today we tried rebooting our nodes one per one. We used the ha-manager crm-command node-maintenance enable like we did several times before. However, this time it seemed to migrate too many vms to the same node which caused this node to crash due to it's memory usage. I just read some things about the CRS settings, but according to the documentation this should not have happened as it already should check the node resources while migrating vms.

We have the following related settings:
- Cluster Resource Scheduliging: ha-rebalance-on-start=1, ha=static
- PVE version: 8.3.2
- HA status:
1736335733332.png

Did anyone perhaps have had the same issue before? Do we need to tweak or recheck any settings?
Please let me know if more information is needed.

Thanks in advance!

Regards,
Demi
 
We have ha groups defined as prefered first, prefered second, sometimes prefered third and prefered last (=prio definitions ti hosts run on)
and so never had unexpected behavior when using maintenance mode for pve updates. But that's not a must it's a can be ... :)

Edit: We do so because of different (smaller/older and bigger/newer) pve nodes.
 
Last edited: