[SOLVED] Can a netwok loop reboot all cluster nodes?

mailinglists

Renowned Member
Mar 14, 2012
641
68
93
Hi,

a coworker made a network loop for 60 seconds and it turned out that some PM cluster which uses WAN side for clustering rebooted itself - all nodes.

It made me wonder, can network loop cause PM to reboot itself?
Will it reboot, even if no HA resources are defined on cluster?
Under what circumstances would PM reboot itself?
Surely loss of quorum should not trigger reboot, when there are no HA resources defined, right?

Here are some relevant logs I got from the admins of this cluster.
Code:
Time42 serverXYZ kernel: [89005319.786954] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005319.837135] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005319.887327] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005319.937518] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005319.987698] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005320.037872] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005320.088052] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time43 serverXYZ kernel: [89005320.138132] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time43 serverXYZ kernel: [89005320.188331] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time43 serverXYZ kernel: [89005320.238454] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time43 serverXYZ corosync[32844]: warning [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time43 serverXYZ corosync[32844]: [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time44 serverXYZ corosync[32844]: warning [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time44 serverXYZ corosync[32844]: [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time46 serverXYZ corosync[32844]: warning [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time46 serverXYZ corosync[32844]: [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time47 serverXYZ kernel: [89005324.838662] net_ratelimit: 91 callbacks suppressed
Time47 serverXYZ kernel: [89005324.838699] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ kernel: [89005324.888853] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ kernel: [89005324.939026] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ kernel: [89005324.981935] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ kernel: [89005325.032129] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ corosync[32844]: warning [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time47 serverXYZ corosync[32844]: [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time47 serverXYZ kernel: [89005325.082298] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time48 serverXYZ kernel: [89005325.132463] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time48 serverXYZ kernel: [89005325.182608] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time48 serverXYZ kernel: [89005325.232796] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time48 serverXYZ kernel: [89005325.282962] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time32 serverXYZ systemd-modules-load[634]: Inserted module 'iscsi_tcp'
Time32 serverXYZ kernel: [ 0.000000] Linux version 4.15.18-30-pve (root@nora) (gcc version 6.3.0 20170516 (Debian 6.3.0-18+deb9u1)) #1 SMP PVE 4.15.18-58 (Fri, 12 Jun 2020 13:53:01 +0200) ()
Time32 serverXYZ systemd-modules-load[634]: Inserted module 'ib_iser'
Time32 serverXYZ kernel: [ 0.000000] Command line: BOOT_IMAGE=/boot/vmlinuz-4.15.18-30-pve root=/dev/mapper/pve-root ro quiet
 
Hi,

a coworker made a network loop for 60 seconds and it turned out that some PM cluster which uses WAN side for clustering rebooted itself - all nodes.

It made me wonder, can network loop cause PM to reboot itself?
on its own it shouldn't. but of course it CAN cause a loss of quorum, which CAN trigger a reboot
Will it reboot, even if no HA resources are defined on cluster?
if HA was never active, a node should not get fenced even when quorum is lost.
Under what circumstances would PM reboot itself?
Surely loss of quorum should not trigger reboot, when there are no HA resources defined, right?

if HA was ever active on a node, a watchdog will be running. if that watchdog expires, the node will fence itself (and reboot, if it still manages to do that). the usual reason for the watchdog expiring is that pmxcfs is not writable, usually because of loss of quorum.

Here are some relevant logs I got from the admins of this cluster.
Code:
Time42 serverXYZ kernel: [89005319.786954] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005319.837135] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005319.887327] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005319.937518] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005319.987698] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005320.037872] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time42 serverXYZ kernel: [89005320.088052] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time43 serverXYZ kernel: [89005320.138132] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time43 serverXYZ kernel: [89005320.188331] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time43 serverXYZ kernel: [89005320.238454] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time43 serverXYZ corosync[32844]: warning [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time43 serverXYZ corosync[32844]: [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time44 serverXYZ corosync[32844]: warning [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time44 serverXYZ corosync[32844]: [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time46 serverXYZ corosync[32844]: warning [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time46 serverXYZ corosync[32844]: [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time47 serverXYZ kernel: [89005324.838662] net_ratelimit: 91 callbacks suppressed
Time47 serverXYZ kernel: [89005324.838699] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ kernel: [89005324.888853] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ kernel: [89005324.939026] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ kernel: [89005324.981935] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ kernel: [89005325.032129] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time47 serverXYZ corosync[32844]: warning [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time47 serverXYZ corosync[32844]: [MAIN ] Totem is unable to form a cluster because of an operating system or network fault (reason: totem is continuously in gather state). The most common cause of this message is that the local firewall is configured improperly.
Time47 serverXYZ kernel: [89005325.082298] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time48 serverXYZ kernel: [89005325.132463] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time48 serverXYZ kernel: [89005325.182608] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time48 serverXYZ kernel: [89005325.232796] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time48 serverXYZ kernel: [89005325.282962] vmbr0: received packet on eno1 with own address as source address (addr:MA:C:AD:DD:ES, vlan:0)
Time32 serverXYZ systemd-modules-load[634]: Inserted module 'iscsi_tcp'
Time32 serverXYZ kernel: [ 0.000000] Linux version 4.15.18-30-pve (root@nora) (gcc version 6.3.0 20170516 (Debian 6.3.0-18+deb9u1)) #1 SMP PVE 4.15.18-58 (Fri, 12 Jun 2020 13:53:01 +0200) ()
Time32 serverXYZ systemd-modules-load[634]: Inserted module 'ib_iser'
Time32 serverXYZ kernel: [ 0.000000] Command line: BOOT_IMAGE=/boot/vmlinuz-4.15.18-30-pve root=/dev/mapper/pve-root ro quiet
that just documents that quorum was lost followed by a new start - we'd need more information to verify whether it was HA that triggered a fence or some other error condition that caused a reboot/crash.
 
Thank you for all your answers.

So if HA was ever defined, even if the HA resources have since been deleted, there is a watchdog running which will reboot node with loss of quorum. Do I understand correct?

If that is so, how can I check if such watch dog is running and how can I disable it (after deleting all HA resources)?
 
Thank you for all your answers.

So if HA was ever defined, even if the HA resources have since been deleted, there is a watchdog running which will reboot node with loss of quorum. Do I understand correct?
yes
If that is so, how can I check if such watch dog is running and how can I disable it (after deleting all HA resources)?
ha-manager status on each node. IFF no HA resources are configured, restarting the pve-ha-lrm service should disable the watchdog.
 
  • Like
Reactions: mailinglists
Thank you for your answer.

Will restarting pve-ha-lrm result in restarting of any virtual instances or host itself?

So, after restart it should look like this:
Code:
root@p35:~# ha-manager status
quorum OK
root@p35:~# ha-manager config
root@p35:~#

And not have any lrm or any other type resources listed, correct?
 
Thank you for your answer.

Will restarting pve-ha-lrm result in restarting of any virtual instances or host itself?
no (it happens on every upgrade of the HA packages!)
So, after restart it should look like this:
Code:
root@p35:~# ha-manager status
quorum OK
root@p35:~# ha-manager config
root@p35:~#

And not have any lrm or any other type resources listed, correct?

HA active with an active resource:
Code:
$ ha-manager status
quorum OK
master nora (active, Tue Mar 23 09:02:40 2021)
lrm nora (active, Tue Mar 23 09:02:40 2021)
service ct:100023 (nora, starting)

HA (and watchdog!) active but all resources removed:
Code:
$ ha-manager status
quorum OK
master nora (active, Tue Mar 23 09:02:50 2021)
lrm nora (active, Tue Mar 23 09:02:50 2021)

HA idle after LRM restart:
Code:
$ ha-manager status
quorum OK
master nora (active, Tue Mar 23 09:05:00 2021)
lrm nora (idle, Tue Mar 23 09:05:02 2021)
 
  • Like
Reactions: mailinglists
there is no issue to be fixed. if you enable HA on a node (by having a HA-enabled resource active there) it will be "armed" until the node is rebooted or the HA services restarted.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!