Strange Ceph behavior

Gudkoff

Member
Sep 25, 2021
7
0
6
36
Russia
Hi everyone!

I'm having this strange issue with Ceph on my Proxmox cluster.
1632580466760.png
The *ceph* nodes have OSD, mon, mds and mgr on each of them.
The *pve* nodes have only OSD on them.

Recently every Saturday exactly at 06:00 (±couple minutes) ceph goes to Warning state.
Logs show that one of the tree monitors goes down and up (not exactly flapping) and other mons constantly calling for election.
And there are slow ops. VMs starts to freeze and lag.

Here is the link to Google Drive with zipped logs from last boot 10-Sep-2021 (covers three Saturdays).

After 6-7 hours cluster fixes itself and everything working as intended. VMs unfreezes.
During this issue ceph commands does not work (timeout) or take VERY long time to execute.
Monitor logs flooded with 2021-09-11 06:03:22.162 7f8450475700 1 mon.hds01-pipcephn1@0(electing) e11 handle_auth_request failed to assign global_id messages.
Googling it does not help me.
Maybe I misconfigured something, maybe some hidden background job messes things up. I already don't know where to look.
Please help!

proxmox-ve: 6.4-1 (running kernel: 5.4.128-1-pve)
pve-manager: 6.4-13 (running version: 6.4-13/9f411e79)
pve-kernel-5.4: 6.4-5
pve-kernel-helper: 6.4-5
pve-kernel-5.3: 6.1-6
pve-kernel-5.4.128-1-pve: 5.4.128-2
pve-kernel-5.4.114-1-pve: 5.4.114-1
pve-kernel-5.4.78-2-pve: 5.4.78-2
pve-kernel-5.4.78-1-pve: 5.4.78-1
pve-kernel-5.3.18-3-pve: 5.3.18-3
ceph: 14.2.22-pve1
ceph-fuse: 14.2.22-pve1
corosync: 3.1.2-pve1
criu: 3.11-3
glusterfs-client: 5.5-3
ifupdown: 0.8.35+pve1
libjs-extjs: 6.0.1-10
libknet1: 1.20-pve1
libproxmox-acme-perl: 1.1.0
libproxmox-backup-qemu0: 1.1.0-1
libpve-access-control: 6.4-3
libpve-apiclient-perl: 3.1-3
libpve-common-perl: 6.4-3
libpve-guest-common-perl: 3.1-5
libpve-http-server-perl: 3.2-3
libpve-storage-perl: 6.4-1
libqb0: 1.0.5-1
libspice-server1: 0.14.2-4~pve6+1
lvm2: 2.03.02-pve4
lxc-pve: 4.0.6-2
lxcfs: 4.0.6-pve1
novnc-pve: 1.1.0-1
openvswitch-switch: 2.12.3-1
proxmox-backup-client: 1.1.13-2
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.6-1
pve-cluster: 6.4-1
pve-container: 3.3-6
pve-docs: 6.4-2
pve-edk2-firmware: 2.20200531-1
pve-firewall: 4.1-4
pve-firmware: 3.2-4
pve-ha-manager: 3.1-1
pve-i18n: 2.3-1
pve-qemu-kvm: 5.2.0-6
pve-xtermjs: 4.7.0-3
qemu-server: 6.4-2
smartmontools: 7.2-pve2
spiceterm: 3.1-1
vncterm: 1.6-2
zfsutils-linux: 2.0.5-pve1~bpo10+1

JSON:
{
    "election_epoch": 300546,
    "quorum": [
        0,
        1,
        2
    ],
    "quorum_names": [
        "hds01-pipcephn1",
        "hds01-pipcephn3",
        "hds01-pipcephn2"
    ],
    "quorum_leader_name": "hds01-pipcephn1",
    "quorum_age": 16182,
    "monmap": {
        "epoch": 11,
        "fsid": "ebf3a0fb-7ac7-4ee4-b8b0-c9426f497ee0",
        "modified": "2021-02-20 16:25:40.801051",
        "created": "2020-03-18 00:33:16.936610",
        "min_mon_release": 14,
        "min_mon_release_name": "nautilus",
        "features": {
            "persistent": [
                "kraken",
                "luminous",
                "mimic",
                "osdmap-prune",
                "nautilus"
            ],
            "optional": []
        },
        "mons": [
            {
                "rank": 0,
                "name": "hds01-pipcephn1",
                "public_addrs": {
                    "addrvec": [
                        {
                            "type": "v2",
                            "addr": "192.168.63.11:3300",
                            "nonce": 0
                        },
                        {
                            "type": "v1",
                            "addr": "192.168.63.11:6789",
                            "nonce": 0
                        }
                    ]
                },
                "addr": "192.168.63.11:6789/0",
                "public_addr": "192.168.63.11:6789/0"
            },
            {
                "rank": 1,
                "name": "hds01-pipcephn3",
                "public_addrs": {
                    "addrvec": [
                        {
                            "type": "v2",
                            "addr": "192.168.63.13:3300",
                            "nonce": 0
                        },
                        {
                            "type": "v1",
                            "addr": "192.168.63.13:6789",
                            "nonce": 0
                        }
                    ]
                },
                "addr": "192.168.63.13:6789/0",
                "public_addr": "192.168.63.13:6789/0"
            },
            {
                "rank": 2,
                "name": "hds01-pipcephn2",
                "public_addrs": {
                    "addrvec": [
                        {
                            "type": "v2",
                            "addr": "192.168.63.12:3300",
                            "nonce": 0
                        },
                        {
                            "type": "v1",
                            "addr": "192.168.63.12:6789",
                            "nonce": 0
                        }
                    ]
                },
                "addr": "192.168.63.12:6789/0",
                "public_addr": "192.168.63.12:6789/0"
            }
        ]
    }
}

P.S. Excuse my poor English.
 
Last edited:
Just answering based off of weird stuff I've noticed myself. Is there any chance you're running backup tasks or whatnot at that time and how does your network topology look like?
Hi!
No, there are no backups configured. Almost all VMs are terminal farm nodes for work from home thingy.
Network is build on LACP bond and Open vSwitch.
*ceph* nodes have 1 bond - for Ceph.
*pve* nodes have 2 bonds - one for Ceph and one for VMs.
vlans are isolated, only cluster nodes in them.
Code:
# LACP bond on LOM 10 Gbit/s interfaces
allow-vmbr0 bond0
iface bond0 inet manual
        ovs_bonds eno1 eno2
        ovs_type OVSBond
        ovs_bridge vmbr0
        ovs_options other_config:lacp-time=fast lacp=active bond_mode=balance-tcp
        pre-up ( ifconfig eno1 mtu 9000; ifconfig eno2 mtu 9000  )
        ovs_mtu 9000


#LACP bond on PCI-e 10 Gbit/s interfaces
allow-vmbr1 bond1
iface bond1 inet manual
        ovs_bonds ens1f0 ens1f1
        ovs_type OVSBond
        ovs_bridge vmbr1
        ovs_options other_config:lacp-time=fast lacp=active bond_mode=balance-tcp
        pre-up ( ifconfig ens1f0 mtu 9000; ifconfig ens1f1 mtu 9000  )
        ovs_mtu 9000


auto lo
iface lo inet loopback

iface eno1 inet manual

iface eno2 inet manual

iface eno3 inet manual

iface eno4 inet manual

iface ens1f0 inet manual

iface ens1f1 inet manual


# Public interface untagget vlan 3062
allow-vmbr0 vlan3062
iface vlan3062 inet static
        address 192.168.62.1
        netmask 24
        # Gateway is a VM [10.209.6.32] as a proxy to vlan 4 with mascarade rules
        gateway 192.168.62.254
        ovs_type OVSIntPort
        ovs_bridge vmbr0
        # 'access' is for untagget
        ovs_options vlan_mode=access
        ovs_extra set interface ${IFACE} external-ids:iface-id=$(hostname -s)-${IFACE}-vif
        ovs_mtu 1500
        dns-nameservers 10.209.8.1 10.209.8.4
        dns-search xxxxxxx


# Private Ceth interface on tagget vlan 3063
allow-vmbr1 vlan3063
iface vlan3063 inet static
        address 192.168.63.1
        netmask 24
        ovs_type OVSIntPort
        ovs_bridge vmbr1
        ovs_extra set interface ${IFACE} external-ids:iface-id=$(hostname -s)-${IFACE}-vif
        ovs_mtu 9000
        ovs_options tag=3063


# Interface for cluster communication on tagget vlan 3064
allow-vmbr0 vlan3064
iface vlan3064 inet static
        address 192.168.64.1
        netmask 24
        ovs_type OVSIntPort
        ovs_bridge vmbr0
        ovs_extra set interface ${IFACE} external-ids:iface-id=$(hostname -s)-${IFACE}-vif
        ovs_mtu 9000
        ovs_options tag=3064


# Interface for VM migration on tagget vlan 3065
allow-vmbr0 vlan3065
iface vlan3065 inet static
        address 192.168.65.1
        netmask 24
        ovs_type OVSIntPort
        ovs_bridge vmbr0
        ovs_extra set interface ${IFACE} external-ids:iface-id=$(hostname -s)-${IFACE}-vif
        ovs_mtu 9000
        ovs_options tag=3065


# Bridge(vSwitch) for public, VMs, migration, ets.
allow-ovs vmbr0
iface vmbr0 inet manual
        ovs_type OVSBridge
        ovs_ports bond0 vlan3062 vlan3064 vlan3065
        ovs_mtu 9000


# Bridge(vSwitch) for Ceph storage
allow-ovs vmbr1
iface vmbr1 inet manual
        ovs_type OVSBridge
        ovs_ports bond1 vlan3063
        ovs_mtu 9000

I think I stumble upon this thread on my troubleshooting. Time is sync with chrony to domain controller and there are plenty of space for logs.
 
Last edited:
It is Saturday again...
And Ceph has become unresponsive.
And monitor logs flooded with this:
Code:
2021-10-02 11:57:50.439 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:50.463 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.043 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.099 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.175 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.263 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.299 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.375 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.539 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.539 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.699 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.739 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.739 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.771 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.775 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:51.971 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.139 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.139 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.375 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.499 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.575 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.727 7f4692ac3700 -1 mon.hds01-pipcephn2@2(electing) e11 get_health_metrics reporting 27384 slow ops, oldest is mgrbeacon mgr.hds01-pipcephn3(ebf3a0fb-7ac7-4ee4-b8b0-c9426f497ee0,330974131, , 0)
2021-10-02 11:57:52.739 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.863 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.939 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.939 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:52.939 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:53.063 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:53.175 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:53.339 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:53.463 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:54.139 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id
2021-10-02 11:57:54.175 7f468f2bc700  1 mon.hds01-pipcephn2@2(electing) e11 handle_auth_request failed to assign global_id


I tried to restart monitor services on each node with systemctl restart ceph-mon@<name>.service.
Restarted monitor was bumped from election, which is legit.
mon.hds01-pipcephn1@0(electing).elector(300705) init, last seen epoch 300705, mid-election, bumping

So no luck...
 
Than all of a sudden at 12:45 everything become good again.


Strange thing I notice in mon logs - mon1 forms quorum with mon2 and mon3.
And mon3 forms quorum with mon2. Mon2 never become leader.

2021-10-02 06:09:55.996 7f08706bd700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 06:10:08.448 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:11:08.073 7f08706bd700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 06:11:21.157 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:12:29.461 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:13:02.970 7f08706bd700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 06:13:23.854 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:14:01.002 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:14:34.691 7f08706bd700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 06:15:15.655 7f08706bd700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 06:15:22.455 7f08706bd700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 06:15:37.439 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:16:19.700 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:16:35.616 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:17:23.733 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:19:22.846 7f08706bd700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 06:19:40.626 7f08706bd700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 06:20:06.547 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:20:38.407 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:21:10.659 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:21:23.340 7f086deb8700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 06:22:09.136 7f08706bd700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
<SHORTENED>
2021-10-02 12:29:11.117 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:29:49.197 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:30:00.781 7ff3262d2700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 12:31:06.642 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 12:32:26.567 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 12:33:16.328 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:33:41.652 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 12:34:13.817 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:36:07.371 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:38:06.389 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 12:38:49.886 7ff3262d2700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 12:39:54.871 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:40:20.415 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:40:44.619 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 12:41:07.776 7ff3262d2700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 12:42:04.292 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 12:42:23.284 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:42:54.313 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:43:46.845 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:44:30.290 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 12:44:56.198 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:45:10.946 7ff3262d2700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
2021-10-02 12:45:43.618 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn2 in quorum (ranks 0,2)
2021-10-02 12:45:57.295 7ff328ad7700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3 in quorum (ranks 0,1)
2021-10-02 12:46:07.223 7ff3262d2700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn1 is new leader, mons hds01-pipcephn1,hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 0,1,2)
<empty>
2021-10-02 06:10:52.035 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:12:09.684 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:13:42.800 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:13:57.684 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:17:17.154 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:25:51.243 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:26:38.491 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:28:11.472 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:30:04.717 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:33:50.271 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 06:35:40.472 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
<SHORTENED>
2021-10-02 11:37:05.915 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 11:43:28.042 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 11:45:24.801 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 11:46:16.961 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 11:47:17.106 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 11:52:06.401 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 11:52:32.753 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:01:20.350 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:04:03.433 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:04:29.077 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:08:10.675 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:08:55.483 7f69fb801700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:23:53.317 7f3d25969700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:24:47.722 7f3d25969700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:27:06.679 7f3d25969700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:27:42.639 7f3d25969700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:31:39.805 7f3d25969700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)
2021-10-02 12:44:06.533 7f3d25969700 0 log_channel(cluster) log [INF] : mon.hds01-pipcephn3 is new leader, mons hds01-pipcephn3,hds01-pipcephn2 in quorum (ranks 1,2)


Is something terribly wrong in my cluster? I tried to leave most things by default.

I've checked our Zabbix, it seems this behavior was since the beginning...
1633176765823.png

At 06:00 every Saturday...
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!