CEPH monitor running on node but showing ? in GUI

unsichtbarre

New Member
Oct 1, 2024
17
3
3
Howdy, NOOB again. With a little help I got POC cluster going with CEPH. Unfortunately, I had to re-install one of the nodes. I removed from cluster (pvecm delnode pve102) and wiped all disks before re-installation.

After re-installation and re-creation of OSDs on all disks, the CEPH cluster returned to semi-health. I created a manager and a monitor on the re-installed node without error.

Unfortunately the monitor is showing stopped (in the GUI) and clicking start doesn't help. I did some googleing and checked the manager status at CLI and it shows running.

Here is some info, what do I do to get GUI to show complete health (or to get complete health?
Code:
root@pve102:~# systemctl status ceph-mon@pve102.service
● ceph-mon@pve102.service - Ceph cluster monitor daemon
     Loaded: loaded (/lib/systemd/system/ceph-mon@.service; enabled; preset: enabled)
    Drop-In: /usr/lib/systemd/system/ceph-mon@.service.d
             └─ceph-after-pve-cluster.conf
     Active: active (running) since Wed 2024-10-02 16:57:40 PDT; 4min 12s ago
   Main PID: 12526 (ceph-mon)
      Tasks: 25
     Memory: 136.5M
        CPU: 1.260s
     CGroup: /system.slice/system-ceph\x2dmon.slice/ceph-mon@pve102.service
             └─12526 /usr/bin/ceph-mon -f --cluster ceph --id pve102 --setuser ceph --setgroup ceph

Oct 02 16:57:40 pve102 systemd[1]: Started ceph-mon@pve102.service - Ceph cluster monitor daemon.
root@pve102:~# pvecm status
Cluster information
-------------------
Name:             sv4-pve-c1
Config Version:   7
Transport:        knet
Secure auth:      on

Quorum information
------------------
Date:             Wed Oct  2 17:11:27 2024
Quorum provider:  corosync_votequorum
Nodes:            3
Node ID:          0x00000002
Ring ID:          1.6c
Quorate:          Yes

Votequorum information
----------------------
Expected votes:   3
Highest expected: 3
Total votes:      3
Quorum:           2
Flags:            Quorate

Membership information
----------------------
    Nodeid      Votes Name
0x00000001          1 172.31.30.101
0x00000002          1 172.31.30.102 (local)
0x00000003          1 172.31.30.103
root@pve102:~# ceph -s
  cluster:
    id:     1101c540-2741-48d9-b64d-189700d0b84f
    health: HEALTH_WARN
            3 daemons have recently crashed

  services:
    mon: 2 daemons, quorum pve101,pve103 (age 4h)
    mgr: pve101(active, since 4h), standbys: pve103, pve102
    osd: 34 osds: 34 up (since 17m), 34 in (since 99m)

  data:
    pools:   2 pools, 33 pgs
    objects: 4.78k objects, 19 GiB
    usage:   61 GiB used, 113 TiB / 114 TiB avail
    pgs:     33 active+clean

  io:
    client:   3.3 KiB/s wr, 0 op/s rd, 0 op/s wr

root@pve103:/etc/pve# cat /etc/pve/ceph.conf
[global]
        auth_client_required = cephx
        auth_cluster_required = cephx
        auth_service_required = cephx
        cluster_network = 10.0.201.1/16
        fsid = 1101c540-2741-48d9-b64d-189700d0b84f
        mon_allow_pool_delete = true
        mon_host = 10.0.201.1 10.0.203.1 10.0.202.1
        ms_bind_ipv4 = true
        ms_bind_ipv6 = false
        osd_pool_default_min_size = 2
        osd_pool_default_size = 3
        public_network = 10.0.201.1/16

[client]
        keyring = /etc/pve/priv/$cluster.$name.keyring

[client.crash]
        keyring = /etc/pve/ceph/$cluster.$name.keyring

[mon.pve101]
        public_addr = 10.0.201.1

[mon.pve102]
        public_addr = 10.0.202.1

[mon.pve103]
        public_addr = 10.0.203.1

root@pve103:/etc/pve#
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!