I have a cluster with 6 nodes members. each nodes has 2 bonding interfaces which look like this:
Bond0
vmbr0v450 << vlan 450 for cluster and management
IPs : 10.100.0.0/24
vmbr0v451 <<vlan 451 for ceph storage
IPs: 10.1.1.0/24
Bond1
vmbr1 << for Guest VM public interface
Latest accident was when i want to move bond0 link on 10.100.0.102 to another switch. before doing that, i move all VMs on that node to another nodes then plug off the cable. Not long after. suddenly all nodes are rebooted. I dont know what causing nodes goes to reboot since only 1 node is on problem.
Bond0
vmbr0v450 << vlan 450 for cluster and management
IPs : 10.100.0.0/24
vmbr0v451 <<vlan 451 for ceph storage
IPs: 10.1.1.0/24
Bond1
vmbr1 << for Guest VM public interface
Code:
root@pve05:~# pvecm status
Cluster information
-------------------
Name: rajamitra
Config Version: 8
Transport: knet
Secure auth: on
Quorum information
------------------
Date: Mon Feb 19 22:14:56 2024
Quorum provider: corosync_votequorum
Nodes: 6
Node ID: 0x00000005
Ring ID: 1.216a
Quorate: Yes
Votequorum information
----------------------
Expected votes: 6
Highest expected: 6
Total votes: 6
Quorum: 4
Flags: Quorate
Membership information
----------------------
Nodeid Votes Name
0x00000001 1 10.100.0.100
0x00000002 1 10.100.0.101
0x00000003 1 10.100.0.102
0x00000004 1 10.100.0.103
0x00000005 1 10.100.0.104 (local)
0x00000006 1 10.100.0.125
Latest accident was when i want to move bond0 link on 10.100.0.102 to another switch. before doing that, i move all VMs on that node to another nodes then plug off the cable. Not long after. suddenly all nodes are rebooted. I dont know what causing nodes goes to reboot since only 1 node is on problem.