Help moving ceph network

chupacabra71

New Member
Jan 18, 2024
5
0
1
I tried to move my ceph network to another subnet and now all osd's are not picking up the new network and staying down. I fear i may have hosed ceph, but as a learning experience, would like to see if i can recover them. this is what they look like in the cluster:

1740325207283.png

this is the contents of the new ceph.conf
Code:
[global]
        auth_client_required = cephx
        auth_cluster_required = cephx
        auth_service_required = cephx
        fsid = b6005ee3-c72c-4fad-8c48-3b4be1a3cb07
        mon_host = 10.10.99.6 10.10.99.7 10.10.99.8
        mon_allow_pool_delete = true
        ms_bind_ipv4 = true
        ms_bind_ipv6 = false
        osd_pool_default_min_size = 2
        osd_pool_default_size = 3
        public_network = 10.10.99.0/24
[client]
        keyring = /etc/pve/priv/$cluster.$name.keyring

[client.crash]
        keyring = /etc/pve/ceph/$cluster.$name.keyring
[mon.pve01]
    public_addr = 10.10.99.6

[mon.pve02]
    public_addr = 10.10.99.7

[mon.pve03]
    public_addr = 10.10.99.8

If i do a `ceph mon dump` it looks like the monitors are healthy and picked up the new addresses:
Code:
epoch 2
fsid b6005ee3-c72c-4fad-8c48-3b4be1a3cb07
last_changed 2025-02-22T11:16:32.323477-0500
created 2025-02-22T11:00:06.278940-0500
min_mon_release 17 (quincy)
election_strategy: 1
0: [v2:10.10.99.6:3300/0,v1:10.10.99.6:6789/0] mon.pve01
1: [v2:10.10.99.7:3300/0,v1:10.10.99.7:6789/0] mon.pve02
2: [v2:10.10.99.8:3300/0,v1:10.10.99.8:6789/0] mon.pve03
dumped monmap epoch 2

it looks like the ceph quorum (`ceph quorum_status`) is ok:
Code:
{
    "election_epoch": 50,
    "quorum": [
        0,
        1,
        2
    ],
    "quorum_names": [
        "pve01",
        "pve02",
        "pve03"
    ],
    "quorum_leader_name": "pve01",
    "quorum_age": 536,
    "features": {
        "quorum_con": "4540138320759226367",
        "quorum_mon": [
            "kraken",
            "luminous",
            "mimic",
            "osdmap-prune",
            "nautilus",
            "octopus",
            "pacific",
            "elector-pinging",
            "quincy"
        ]
    },
    "monmap": {
        "epoch": 2,
        "fsid": "b6005ee3-c72c-4fad-8c48-3b4be1a3cb07",
        "modified": "2025-02-22T16:16:32.323477Z",
        "created": "2025-02-22T16:00:06.278940Z",
        "min_mon_release": 17,
        "min_mon_release_name": "quincy",
        "election_strategy": 1,
        "disallowed_leaders: ": "",
        "stretch_mode": false,
        "tiebreaker_mon": "",
        "removed_ranks: ": "",
        "features": {
            "persistent": [
                "kraken",
                "luminous",
                "mimic",
                "osdmap-prune",
                "nautilus",
                "octopus",
                "pacific",
                "elector-pinging",
                "quincy"
            ],
            "optional": []
        },
        "mons": [
            {
                "rank": 0,
                "name": "pve01",
                "public_addrs": {
                    "addrvec": [
                        {
                            "type": "v2",
                            "addr": "10.10.99.6:3300",
                            "nonce": 0
                        },
                        {
                            "type": "v1",
                            "addr": "10.10.99.6:6789",
                            "nonce": 0
                        }
                    ]
                },
                "addr": "10.10.99.6:6789/0",
                "public_addr": "10.10.99.6:6789/0",
                "priority": 0,
                "weight": 0,
                "crush_location": "{}"
            },
            {
                "rank": 1,
                "name": "pve02",
                "public_addrs": {
                    "addrvec": [
                        {
                            "type": "v2",
                            "addr": "10.10.99.7:3300",
                            "nonce": 0
                        },
                        {
                            "type": "v1",
                            "addr": "10.10.99.7:6789",
                            "nonce": 0
                        }
                    ]
                },
                "addr": "10.10.99.7:6789/0",
                "public_addr": "10.10.99.7:6789/0",
                "priority": 0,
                "weight": 0,
                "crush_location": "{}"
            },
            {
                "rank": 2,
                "name": "pve03",
                "public_addrs": {
                    "addrvec": [
                        {
                            "type": "v2",
                            "addr": "10.10.99.8:3300",
                            "nonce": 0
                        },
                        {
                            "type": "v1",
                            "addr": "10.10.99.8:6789",
                            "nonce": 0
                        }
                    ]
                },
                "addr": "10.10.99.8:6789/0",
                "public_addr": "10.10.99.8:6789/0",
                "priority": 0,
                "weight": 0,
                "crush_location": "{}"
            }
        ]
    }
}

but when i look at the OSDs wiht `ceph osd tree`, they are all down:
Code:
ID  CLASS  WEIGHT    TYPE NAME       STATUS  REWEIGHT  PRI-AFF
-1         10.47949  root default                            
-4          3.49316      host pve01                          
 8    ssd   0.87329          osd.8     down   1.00000  1.00000
 9    ssd   0.87329          osd.9     down   1.00000  1.00000
10    ssd   0.87329          osd.10    down   1.00000  1.00000
11    ssd   0.87329          osd.11    down   1.00000  1.00000
-3          3.49316      host pve02                          
 0    ssd   0.87329          osd.0     down   1.00000  1.00000
 4    ssd   0.87329          osd.4     down   1.00000  1.00000
12    ssd   0.87329          osd.12    down   1.00000  1.00000
13    ssd   0.87329          osd.13    down   1.00000  1.00000
-2          3.49316      host pve03                          
 2    ssd   0.87329          osd.2     down   1.00000  1.00000
 3    ssd   0.87329          osd.3     down   1.00000  1.00000
 6    ssd   0.87329          osd.6     down   1.00000  1.00000
 7    ssd   0.87329          osd.7     down   1.00000  1.00000

if i try to look into one of the OSDs, it looks like they are not picking up the network:
Code:
{
    "osd": 8,
    "addrs": {
        "addrvec": []
    },
    "osd_fsid": "57895705-d901-400b-a681-61077859a5ed",
    "crush_location": {
        "host": "pve01",
        "root": "default"
    }
}

in the journal i don't see much:
Code:
Feb 23 10:13:18 pve01 systemd[1]: ceph-osd@8.service: Deactivated successfully.
Feb 23 10:13:18 pve01 systemd[1]: Stopped ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 23 10:13:18 pve01 systemd[1]: ceph-osd@8.service: Consumed 2.353s CPU time.
Feb 23 10:13:18 pve01 systemd[1]: Starting ceph-osd@8.service - Ceph object storage daemon osd.8...
Feb 23 10:13:18 pve01 systemd[1]: Started ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 23 10:13:18 pve01 ceph-osd[2487834]: 2025-02-23T10:13:18.757-0500 744e2f7b13c0 -1 Falling back to public interface
Feb 23 10:13:22 pve01 ceph-osd[2487834]: 2025-02-23T10:13:22.265-0500 744e2f7b13c0 -1 osd.8 9061 log_to_monitors true
Feb 23 10:13:23 pve01 ceph-osd[2487834]: 2025-02-23T10:13:23.256-0500 744e210006c0 -1 osd.8 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory


i've tried setting the via `ceph config set osd.8 public_addr 10.10.99.6:0`
and this is what the config (ceph config dump) looks like now:

Code:
WHO     MASK  LEVEL     OPTION                                 VALUE              RO
mon           advanced  auth_allow_insecure_global_id_reclaim  false              
osd           advanced  public_network                         10.10.99.0/24      *
osd           advanced  public_network_interface               vlan99             *
osd.0         basic     osd_mclock_max_capacity_iops_ssd       55377.221418        
osd.0         advanced  public_bind_addr                       v2:10.10.99.7:0/0  *
osd.10        basic     osd_mclock_max_capacity_iops_ssd       42485.869165        
osd.10        basic     public_addr                            v2:10.10.99.6:0/0  *
osd.11        basic     osd_mclock_max_capacity_iops_ssd       43151.652267        
osd.11        basic     public_addr                            v2:10.10.99.6:0/0  *
osd.12        basic     osd_mclock_max_capacity_iops_ssd       62003.970817        
osd.12        advanced  public_bind_addr                       v2:10.10.99.7:0/0  *
osd.13        basic     osd_mclock_max_capacity_iops_ssd       58325.014080        
osd.13        advanced  public_bind_addr                       v2:10.10.99.7:0/0  *
osd.2         basic     osd_mclock_max_capacity_iops_ssd       67670.304502        
osd.2         advanced  public_bind_addr                       v2:10.10.99.8:0/0  *
osd.3         basic     osd_mclock_max_capacity_iops_ssd       53038.204249        
osd.3         advanced  public_bind_addr                       v2:10.10.99.8:0/0  *
osd.4         basic     osd_mclock_max_capacity_iops_ssd       63817.463674        
osd.4         advanced  public_bind_addr                       v2:10.10.99.7:0/0  *
osd.6         basic     osd_mclock_max_capacity_iops_ssd       60684.431791        
osd.6         advanced  public_bind_addr                       v2:10.10.99.8:0/0  *
osd.7         basic     osd_mclock_max_capacity_iops_ssd       52488.209705        
osd.7         advanced  public_bind_addr                       v2:10.10.99.8:0/0  *
osd.8         basic     osd_mclock_max_capacity_iops_ssd       44762.641171        
osd.8         basic     public_addr                            v2:10.10.99.6:0/0  *
osd.8         advanced  public_bind_addr                       v2:10.10.99.6:0/0  *
osd.9         basic     osd_mclock_max_capacity_iops_ssd       42679.547164        
osd.9         basic     public_addr                            v2:10.10.99.6:0/0  *

here is the dump of the osds:
Code:
osd.0 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 7e86ab26-fabf-44cd-a88c-8112c91b12f8
osd.2 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new dc3cd8d5-d094-429c-86da-080aa14d3091
osd.3 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new a4e08ff6-9a45-4e58-9435-bec4acd78b7a
osd.4 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new b9c80f90-f4e8-4d4b-8beb-c1fe95292786
osd.6 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new bb269c34-5b25-453f-b7a1-98b271d1e961
osd.7 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new dea54476-b602-4d1e-bafb-f6d13a35a69d
osd.8 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 57895705-d901-400b-a681-61077859a5ed
osd.9 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new fdb7aece-4bfb-4803-926a-0a53cbe8fc14
osd.10 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 9629d126-23c4-4b1a-9451-8d6853851e0f
osd.11 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new cc6fc173-0e28-4c00-9540-0a43e39ccf65
osd.12 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 278152ba-c85c-418c-b935-b50cd5c200d0
osd.13 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 2b6245b0-657d-46d5-b327-f53b0a86f073

not sure what else to try... I am certain it is something relatively simple that i am missing, but to be honest i've been lookint at this for several hours and it seems that i am going around in circles. i do have backups if i need to reset the whole cluster, but I think it is recoverable and would be a great learning experience for me.
 
Last edited:
Hmm, usually we also have the cluster_network line in the global section. It can be the same IP subnet as the public.
Make sure that the network is configured onthe hosts and that they can reach each other via this network.


The simplest way to change this without any downtime would be to:
  1. make sure new network works (ping, also check MTU if large MTU is used: ping -M do -s 8972 {target host} (MTU - 28 bytes overhead for ICMP + IPv4)
  2. Change cluster_network and public_network settings
  3. Restart OSDs on one host at a time: systemctl restart ceph-osd.target Wait for Ceph to be healthy again and repeat on the next host.
  4. For the MONs, destroy one MON and recreate it. This will automatically remove the old and add the new (in the new IP subnet) to the monmap and the mon_host line in the ceph.conf file.
  5. Repeat for each MON, one MONs at a time and wait for the Cluster to be healthy.
  6. Similar for the MGR and MDS.
You can use ss -tulpn | grep ceph to check on which IPs the services are listening on.
 
Hmm, usually we also have the cluster_network line in the global section. It can be the same IP subnet as the public.
Make sure that the network is configured onthe hosts and that they can reach each other via this network.


The simplest way to change this without any downtime would be to:
  1. make sure new network works (ping, also check MTU if large MTU is used: ping -M do -s 8972 {target host} (MTU - 28 bytes overhead for ICMP + IPv4)
  2. Change cluster_network and public_network settings
  3. Restart OSDs on one host at a time: systemctl restart ceph-osd.target Wait for Ceph to be healthy again and repeat on the next host.
  4. For the MONs, destroy one MON and recreate it. This will automatically remove the old and add the new (in the new IP subnet) to the monmap and the mon_host line in the ceph.conf file.
  5. Repeat for each MON, one MONs at a time and wait for the Cluster to be healthy.
  6. Similar for the MGR and MDS.
You can use ss -tulpn | grep ceph to check on which IPs the services are listening on.
thanks for the tips. it is really strange. no matter what i do, i can't get the OSD to pick up any IP.

Code:
root@pve01:~# 
root@pve01:~# 
root@pve01:~# 
root@pve01:~# systemctl reset-failed 
root@pve01:~# ceph osd find 8
{
    "osd": 8,
    "addrs": {
        "addrvec": []
    },
    "osd_fsid": "57895705-d901-400b-a681-61077859a5ed",
    "crush_location": {
        "host": "pve01",
        "root": "default"
    }
}
root@pve01:~# ceph config set osd.8 public_addr 10.10.99.6:0
root@pve01:~# systemctl restart ceph-osd@8.service 
root@pve01:~# systemctl status ceph-osd@* 
● ceph-osd@11.service - Ceph object storage daemon osd.11
     Loaded: loaded (/lib/systemd/system/ceph-osd@.service; enabled-runtime; preset: enabled)
    Drop-In: /usr/lib/systemd/system/ceph-osd@.service.d
             └─ceph-after-pve-cluster.conf
     Active: active (running) since Mon 2025-02-24 06:41:13 EST; 43min ago
    Process: 819905 ExecStartPre=/usr/libexec/ceph/ceph-osd-prestart.sh --cluster ${CLUSTER} --id 11 (code=exited, status=0/SUCCESS)
   Main PID: 819918 (ceph-osd)
      Tasks: 76
     Memory: 306.1M
        CPU: 8.530s
     CGroup: /system.slice/system-ceph\x2dosd.slice/ceph-osd@11.service
             └─819918 /usr/bin/ceph-osd -f --cluster ceph --id 11 --setuser ceph --setgroup ceph

Feb 24 06:41:13 pve01 systemd[1]: Starting ceph-osd@11.service - Ceph object storage daemon osd.11...
Feb 24 06:41:13 pve01 systemd[1]: Started ceph-osd@11.service - Ceph object storage daemon osd.11.
Feb 24 06:41:17 pve01 ceph-osd[819918]: 2025-02-24T06:41:17.267-0500 7ee52e2a93c0 -1 osd.11 9061 log_to_monitors true
Feb 24 06:41:17 pve01 ceph-osd[819918]: 2025-02-24T06:41:17.714-0500 7ee51fc006c0 -1 osd.11 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory

● ceph-osd@10.service - Ceph object storage daemon osd.10
     Loaded: loaded (/lib/systemd/system/ceph-osd@.service; enabled-runtime; preset: enabled)
    Drop-In: /usr/lib/systemd/system/ceph-osd@.service.d
             └─ceph-after-pve-cluster.conf
     Active: active (running) since Mon 2025-02-24 06:41:13 EST; 43min ago
    Process: 819947 ExecStartPre=/usr/libexec/ceph/ceph-osd-prestart.sh --cluster ${CLUSTER} --id 10 (code=exited, status=0/SUCCESS)
   Main PID: 819972 (ceph-osd)
      Tasks: 76
     Memory: 285.4M
        CPU: 8.432s
     CGroup: /system.slice/system-ceph\x2dosd.slice/ceph-osd@10.service
             └─819972 /usr/bin/ceph-osd -f --cluster ceph --id 10 --setuser ceph --setgroup ceph

Feb 24 06:41:13 pve01 systemd[1]: Starting ceph-osd@10.service - Ceph object storage daemon osd.10...
Feb 24 06:41:13 pve01 systemd[1]: Started ceph-osd@10.service - Ceph object storage daemon osd.10.
Feb 24 06:41:17 pve01 ceph-osd[819972]: 2025-02-24T06:41:17.381-0500 7a54010df3c0 -1 osd.10 9061 log_to_monitors true
Feb 24 06:41:17 pve01 ceph-osd[819972]: 2025-02-24T06:41:17.714-0500 7a53f2a006c0 -1 osd.10 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory

● ceph-osd@9.service - Ceph object storage daemon osd.9
     Loaded: loaded (/lib/systemd/system/ceph-osd@.service; enabled-runtime; preset: enabled)
    Drop-In: /usr/lib/systemd/system/ceph-osd@.service.d
             └─ceph-after-pve-cluster.conf
     Active: active (running) since Mon 2025-02-24 06:41:13 EST; 43min ago
    Process: 819928 ExecStartPre=/usr/libexec/ceph/ceph-osd-prestart.sh --cluster ${CLUSTER} --id 9 (code=exited, status=0/SUCCESS)
   Main PID: 819971 (ceph-osd)
      Tasks: 76
     Memory: 325.4M
        CPU: 8.589s
     CGroup: /system.slice/system-ceph\x2dosd.slice/ceph-osd@9.service
             └─819971 /usr/bin/ceph-osd -f --cluster ceph --id 9 --setuser ceph --setgroup ceph

Feb 24 06:41:13 pve01 systemd[1]: Starting ceph-osd@9.service - Ceph object storage daemon osd.9...
Feb 24 06:41:13 pve01 systemd[1]: Started ceph-osd@9.service - Ceph object storage daemon osd.9.
Feb 24 06:41:17 pve01 ceph-osd[819971]: 2025-02-24T06:41:17.461-0500 74bd0c71a3c0 -1 osd.9 9061 log_to_monitors true
Feb 24 06:41:17 pve01 ceph-osd[819971]: 2025-02-24T06:41:17.714-0500 74bcfe0006c0 -1 osd.9 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory

● ceph-osd@8.service - Ceph object storage daemon osd.8
     Loaded: loaded (/lib/systemd/system/ceph-osd@.service; enabled; preset: enabled)
    Drop-In: /usr/lib/systemd/system/ceph-osd@.service.d
             └─ceph-after-pve-cluster.conf
     Active: active (running) since Mon 2025-02-24 07:24:12 EST; 3s ago
    Process: 873262 ExecStartPre=/usr/libexec/ceph/ceph-osd-prestart.sh --cluster ${CLUSTER} --id 8 (code=exited, status=0/SUCCESS)
   Main PID: 873266 (ceph-osd)
      Tasks: 32
root@pve01:~# 
root@pve01:~# ceph osd find 8
{
    "osd": 8,
    "addrs": {
        "addrvec": []
    },
    "osd_fsid": "57895705-d901-400b-a681-61077859a5ed",
    "crush_location": {
        "host": "pve01",
        "root": "default"
    }
}
root@pve01:~# ceph osd dump
epoch 135
fsid b6005ee3-c72c-4fad-8c48-3b4be1a3cb07
created 2025-02-22T11:11:46.615916-0500
modified 2025-02-24T07:24:17.403323-0500
flags sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit
crush_version 108
full_ratio 0.95
backfillfull_ratio 0.9
nearfull_ratio 0.85
require_min_compat_client luminous
min_compat_client jewel
require_osd_release quincy
stretch_mode_enabled false
max_osd 14
osd.0 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 7e86ab26-fabf-44cd-a88c-8112c91b12f8
osd.2 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new dc3cd8d5-d094-429c-86da-080aa14d3091
osd.3 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new a4e08ff6-9a45-4e58-9435-bec4acd78b7a
osd.4 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new b9c80f90-f4e8-4d4b-8beb-c1fe95292786
osd.6 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new bb269c34-5b25-453f-b7a1-98b271d1e961
osd.7 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new dea54476-b602-4d1e-bafb-f6d13a35a69d
osd.8 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 57895705-d901-400b-a681-61077859a5ed
osd.9 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new fdb7aece-4bfb-4803-926a-0a53cbe8fc14
osd.10 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 9629d126-23c4-4b1a-9451-8d6853851e0f
osd.11 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new cc6fc173-0e28-4c00-9540-0a43e39ccf65
osd.12 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 278152ba-c85c-418c-b935-b50cd5c200d0
osd.13 down in  weight 1 up_from 0 up_thru 0 down_at 0 last_clean_interval [0,0)   exists,new 2b6245b0-657d-46d5-b327-f53b0a86f073
root@pve01:~# ceph config dump
WHO     MASK  LEVEL     OPTION                                 VALUE              RO
mon           advanced  auth_allow_insecure_global_id_reclaim  false                
osd           advanced  public_network                         10.10.99.0/24      * 
osd           advanced  public_network_interface               vlan99             * 
osd.0         basic     osd_mclock_max_capacity_iops_ssd       55377.221418         
osd.0         advanced  public_bind_addr                       v2:10.10.99.7:0/0  * 
osd.10        basic     osd_mclock_max_capacity_iops_ssd       42485.869165         
osd.10        basic     public_addr                            v2:10.10.99.6:0/0  * 
osd.11        basic     osd_mclock_max_capacity_iops_ssd       43151.652267         
osd.11        basic     public_addr                            v2:10.10.99.6:0/0  * 
osd.12        basic     osd_mclock_max_capacity_iops_ssd       62003.970817         
osd.12        advanced  public_bind_addr                       v2:10.10.99.7:0/0  * 
osd.13        basic     osd_mclock_max_capacity_iops_ssd       58325.014080         
osd.13        advanced  public_bind_addr                       v2:10.10.99.7:0/0  * 
osd.2         basic     osd_mclock_max_capacity_iops_ssd       67670.304502         
osd.2         advanced  public_bind_addr                       v2:10.10.99.8:0/0  * 
osd.3         basic     osd_mclock_max_capacity_iops_ssd       53038.204249         
osd.3         advanced  public_bind_addr                       v2:10.10.99.8:0/0  * 
osd.4         basic     osd_mclock_max_capacity_iops_ssd       63817.463674         
osd.4         advanced  public_bind_addr                       v2:10.10.99.7:0/0  * 
osd.6         basic     osd_mclock_max_capacity_iops_ssd       60684.431791         
osd.6         advanced  public_bind_addr                       v2:10.10.99.8:0/0  * 
osd.7         basic     osd_mclock_max_capacity_iops_ssd       52488.209705         
osd.7         advanced  public_bind_addr                       v2:10.10.99.8:0/0  * 
osd.8         basic     osd_mclock_max_capacity_iops_ssd       44762.641171         
osd.8         basic     public_addr                            v2:10.10.99.6:0/0  * 
osd.8         advanced  public_bind_addr                       v2:10.10.99.6:0/0  * 
osd.9         basic     osd_mclock_max_capacity_iops_ssd       42679.547164         
osd.9         basic     public_addr                            v2:10.10.99.6:0/0  * 
root@pve01:~#
 
forgot to add that ping test also show up fine:


Code:
root@pve01:~# ping -M do -s 8972 10.10.99.7 -c 4
PING 10.10.99.7 (10.10.99.7) 8972(9000) bytes of data.
8980 bytes from 10.10.99.7: icmp_seq=1 ttl=64 time=0.194 ms
8980 bytes from 10.10.99.7: icmp_seq=2 ttl=64 time=0.314 ms
8980 bytes from 10.10.99.7: icmp_seq=3 ttl=64 time=0.201 ms
8980 bytes from 10.10.99.7: icmp_seq=4 ttl=64 time=0.205 ms

--- 10.10.99.7 ping statistics ---
4 packets transmitted, 4 received, 0% packet loss, time 3071ms
rtt min/avg/max/mdev = 0.194/0.228/0.314/0.049 ms
root@pve01:~# ping -M do -s 8972 10.10.99.7 -I vlan99 -c 4
PING 10.10.99.7 (10.10.99.7) from 10.10.99.6 vlan99: 8972(9000) bytes of data.
8980 bytes from 10.10.99.7: icmp_seq=1 ttl=64 time=0.213 ms
8980 bytes from 10.10.99.7: icmp_seq=2 ttl=64 time=0.223 ms
8980 bytes from 10.10.99.7: icmp_seq=3 ttl=64 time=0.298 ms
8980 bytes from 10.10.99.7: icmp_seq=4 ttl=64 time=0.282 ms

--- 10.10.99.7 ping statistics ---
4 packets transmitted, 4 received, 0% packet loss, time 3053ms
rtt min/avg/max/mdev = 0.213/0.254/0.298/0.036 ms
root@pve01:~#

and i added the cluster_network setting as well:

Code:
root@pve01:~# cat /etc/ceph/ceph.conf
[global]
        auth_client_required = cephx
        auth_cluster_required = cephx
        auth_service_required = cephx
        fsid = b6005ee3-c72c-4fad-8c48-3b4be1a3cb07
        mon_host = 10.10.99.6 10.10.99.7 10.10.99.8
        mon_allow_pool_delete = true
        ms_bind_ipv4 = true
        ms_bind_ipv6 = false
        osd_pool_default_min_size = 2
        osd_pool_default_size = 3
        public_network = 10.10.99.0/24
        public_network_interface = vlan99
        cluster_network = 10.10.99.0/24
        cluster_network_interface = vlan99
[client]
        keyring = /etc/pve/priv/$cluster.$name.keyring

[client.crash]
        keyring = /etc/pve/ceph/$cluster.$name.keyring
[mon.pve01]
    public_addr = 10.10.99.6

[mon.pve02]
    public_addr = 10.10.99.7

[mon.pve03]
    public_addr = 10.10.99.8

root@pve01:~#
 
Last edited:
Do you see anything else in the journal of the OSD? The status output only has the last few lines.

How are the permissions of the /etc/pve/ceph.conf file?
Should be:
Code:
-rw-r----- 1 root www-data 849 Nov 12 12:41 /etc/pve/ceph.conf
With owner=root and group=www-data
 
oh, forgot, the OSDs are listening on the correct interface and i can telnet to the ports
Do you see anything else in the journal of the OSD? The status output only has the last few lines.

How are the permissions of the /etc/pve/ceph.conf file?
Should be:
Code:
-rw-r----- 1 root www-data 849 Nov 12 12:41 /etc/pve/ceph.conf
With owner=root and group=www-data

that seems correct:
Code:
root@pve01:~# ls -la /etc/pve/ceph.conf
-rw-r----- 1 root www-data 705 Feb 24 06:39 /etc/pve/ceph.conf

nothing much in the journal that i can discern. there are some failures in there, but that's from me trying different things:
Code:
root@pve01:~# journalctl -u ceph-osd@8 --no-pager --lines=50
Feb 24 06:41:17 pve01 ceph-osd[819924]: 2025-02-24T06:41:17.435-0500 7ee448b303c0 -1 osd.8 9061 log_to_monitors true
Feb 24 06:41:17 pve01 ceph-osd[819924]: 2025-02-24T06:41:17.714-0500 7ee43a4006c0 -1 osd.8 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
Feb 24 06:47:16 pve01 systemd[1]: Stopping ceph-osd@8.service - Ceph object storage daemon osd.8...
Feb 24 06:47:16 pve01 ceph-osd[819924]: 2025-02-24T06:47:16.638-0500 7ee4444006c0 -1 received  signal: Terminated from /sbin/init  (PID: 1) UID: 0
Feb 24 06:47:16 pve01 ceph-osd[819924]: 2025-02-24T06:47:16.638-0500 7ee4444006c0 -1 osd.8 9061 *** Got signal Terminated ***
Feb 24 06:47:16 pve01 ceph-osd[819924]: 2025-02-24T06:47:16.638-0500 7ee4444006c0 -1 osd.8 9061 *** Immediate shutdown (osd_fast_shutdown=true) ***
Feb 24 06:47:22 pve01 systemd[1]: ceph-osd@8.service: Deactivated successfully.
Feb 24 06:47:22 pve01 systemd[1]: Stopped ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 24 06:47:22 pve01 systemd[1]: ceph-osd@8.service: Consumed 2.097s CPU time.
Feb 24 06:47:22 pve01 systemd[1]: Starting ceph-osd@8.service - Ceph object storage daemon osd.8...
Feb 24 06:47:22 pve01 systemd[1]: Started ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 24 06:47:26 pve01 ceph-osd[829942]: 2025-02-24T06:47:26.395-0500 729b4652f3c0 -1 osd.8 9061 log_to_monitors true
Feb 24 06:47:26 pve01 ceph-osd[829942]: 2025-02-24T06:47:26.915-0500 729b37e006c0 -1 osd.8 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
Feb 24 06:51:59 pve01 systemd[1]: Stopping ceph-osd@8.service - Ceph object storage daemon osd.8...
Feb 24 06:51:59 pve01 ceph-osd[829942]: 2025-02-24T06:51:59.465-0500 729b41e006c0 -1 received  signal: Terminated from /sbin/init  (PID: 1) UID: 0
Feb 24 06:51:59 pve01 ceph-osd[829942]: 2025-02-24T06:51:59.465-0500 729b41e006c0 -1 osd.8 9061 *** Got signal Terminated ***
Feb 24 06:51:59 pve01 ceph-osd[829942]: 2025-02-24T06:51:59.465-0500 729b41e006c0 -1 osd.8 9061 *** Immediate shutdown (osd_fast_shutdown=true) ***
Feb 24 06:52:05 pve01 systemd[1]: ceph-osd@8.service: Deactivated successfully.
Feb 24 06:52:05 pve01 systemd[1]: Stopped ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 24 06:52:05 pve01 systemd[1]: ceph-osd@8.service: Consumed 2.002s CPU time.
Feb 24 06:52:05 pve01 systemd[1]: ceph-osd@8.service: Start request repeated too quickly.
Feb 24 06:52:05 pve01 systemd[1]: ceph-osd@8.service: Failed with result 'start-limit-hit'.
Feb 24 06:52:05 pve01 systemd[1]: Failed to start ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 24 07:24:12 pve01 systemd[1]: Starting ceph-osd@8.service - Ceph object storage daemon osd.8...
Feb 24 07:24:12 pve01 systemd[1]: Started ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 24 07:24:16 pve01 ceph-osd[873266]: 2025-02-24T07:24:16.433-0500 7ee3d19033c0 -1 osd.8 9061 log_to_monitors true
Feb 24 07:24:17 pve01 ceph-osd[873266]: 2025-02-24T07:24:17.413-0500 7ee3c32006c0 -1 osd.8 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
Feb 24 07:28:00 pve01 ceph-osd[873266]: 2025-02-24T07:28:00.577-0500 7ee3cd2006c0 -1 received  signal: Terminated from /sbin/init  (PID: 1) UID: 0
Feb 24 07:28:00 pve01 ceph-osd[873266]: 2025-02-24T07:28:00.577-0500 7ee3cd2006c0 -1 osd.8 9061 *** Got signal Terminated ***
Feb 24 07:28:00 pve01 ceph-osd[873266]: 2025-02-24T07:28:00.577-0500 7ee3cd2006c0 -1 osd.8 9061 *** Immediate shutdown (osd_fast_shutdown=true) ***
Feb 24 07:28:00 pve01 systemd[1]: Stopping ceph-osd@8.service - Ceph object storage daemon osd.8...
Feb 24 07:28:00 pve01 ceph-osd[873266]: 2025-02-24T07:28:00.585-0500 7ee3bd8006c0 -1 osd.8 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
Feb 24 07:28:06 pve01 systemd[1]: ceph-osd@8.service: Deactivated successfully.
Feb 24 07:28:06 pve01 systemd[1]: Stopped ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 24 07:28:06 pve01 systemd[1]: ceph-osd@8.service: Consumed 1.696s CPU time.
Feb 24 07:28:06 pve01 systemd[1]: Starting ceph-osd@8.service - Ceph object storage daemon osd.8...
Feb 24 07:28:06 pve01 systemd[1]: Started ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 24 07:28:10 pve01 ceph-osd[879498]: 2025-02-24T07:28:10.160-0500 72ae797943c0 -1 osd.8 9061 log_to_monitors true
Feb 24 07:28:10 pve01 ceph-osd[879498]: 2025-02-24T07:28:10.375-0500 72ae6b0006c0 -1 osd.8 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
Feb 24 07:31:42 pve01 ceph-osd[879498]: 2025-02-24T07:31:42.182-0500 72ae750006c0 -1 received  signal: Terminated from /sbin/init  (PID: 1) UID: 0
Feb 24 07:31:42 pve01 ceph-osd[879498]: 2025-02-24T07:31:42.182-0500 72ae750006c0 -1 osd.8 9061 *** Got signal Terminated ***
Feb 24 07:31:42 pve01 ceph-osd[879498]: 2025-02-24T07:31:42.182-0500 72ae750006c0 -1 osd.8 9061 *** Immediate shutdown (osd_fast_shutdown=true) ***
Feb 24 07:31:42 pve01 systemd[1]: Stopping ceph-osd@8.service - Ceph object storage daemon osd.8...
Feb 24 07:31:47 pve01 systemd[1]: ceph-osd@8.service: Deactivated successfully.
Feb 24 07:31:47 pve01 systemd[1]: Stopped ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 24 07:31:47 pve01 systemd[1]: ceph-osd@8.service: Consumed 1.627s CPU time.
Feb 24 07:32:31 pve01 systemd[1]: Starting ceph-osd@8.service - Ceph object storage daemon osd.8...
Feb 24 07:32:31 pve01 systemd[1]: Started ceph-osd@8.service - Ceph object storage daemon osd.8.
Feb 24 07:32:36 pve01 ceph-osd[888175]: 2025-02-24T07:32:36.216-0500 7b48d414b3c0 -1 osd.8 9061 log_to_monitors true
Feb 24 07:32:36 pve01 ceph-osd[888175]: 2025-02-24T07:32:36.967-0500 7b48c5a006c0 -1 osd.8 9061 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory
root@pve01:~#

i doubleckecked the network and it seems fine. the OSDs seem to be listening on the correct addresses,
Code:
root@pve01:~# netstat -tulnp | grep ceph
tcp        0      0 10.10.99.6:6832         0.0.0.0:*               LISTEN      879193/ceph-mgr     
tcp        0      0 10.10.99.6:6833         0.0.0.0:*               LISTEN      879193/ceph-mgr     
tcp        0      0 10.10.99.6:6816         0.0.0.0:*               LISTEN      888176/ceph-osd     
tcp        0      0 10.10.99.6:6817         0.0.0.0:*               LISTEN      888176/ceph-osd     
tcp        0      0 10.10.99.6:6818         0.0.0.0:*               LISTEN      888176/ceph-osd     
tcp        0      0 10.10.99.6:6819         0.0.0.0:*               LISTEN      888175/ceph-osd     
tcp        0      0 10.10.99.6:6820         0.0.0.0:*               LISTEN      888176/ceph-osd     
tcp        0      0 10.10.99.6:6821         0.0.0.0:*               LISTEN      888175/ceph-osd     
tcp        0      0 10.10.99.6:6822         0.0.0.0:*               LISTEN      888176/ceph-osd     
tcp        0      0 10.10.99.6:6823         0.0.0.0:*               LISTEN      888176/ceph-osd     
tcp        0      0 10.10.99.6:6824         0.0.0.0:*               LISTEN      888175/ceph-osd     
tcp        0      0 10.10.99.6:6825         0.0.0.0:*               LISTEN      888176/ceph-osd     
tcp        0      0 10.10.99.6:6826         0.0.0.0:*               LISTEN      888175/ceph-osd     
tcp        0      0 10.10.99.6:6827         0.0.0.0:*               LISTEN      888176/ceph-osd     
tcp        0      0 10.10.99.6:6828         0.0.0.0:*               LISTEN      888175/ceph-osd     
tcp        0      0 10.10.99.6:6829         0.0.0.0:*               LISTEN      888175/ceph-osd     
tcp        0      0 10.10.99.6:6830         0.0.0.0:*               LISTEN      888175/ceph-osd     
tcp        0      0 10.10.99.6:6831         0.0.0.0:*               LISTEN      888175/ceph-osd     
tcp        0      0 10.10.99.6:6800         0.0.0.0:*               LISTEN      888174/ceph-osd     
tcp        0      0 10.10.99.6:6801         0.0.0.0:*               LISTEN      888174/ceph-osd     
tcp        0      0 10.10.99.6:6802         0.0.0.0:*               LISTEN      888174/ceph-osd     
tcp        0      0 10.10.99.6:6803         0.0.0.0:*               LISTEN      888174/ceph-osd     
tcp        0      0 10.10.99.6:6804         0.0.0.0:*               LISTEN      888174/ceph-osd     
tcp        0      0 10.10.99.6:6805         0.0.0.0:*               LISTEN      888174/ceph-osd     
tcp        0      0 10.10.99.6:6806         0.0.0.0:*               LISTEN      888174/ceph-osd     
tcp        0      0 10.10.99.6:6807         0.0.0.0:*               LISTEN      888174/ceph-osd     
tcp        0      0 10.10.99.6:6808         0.0.0.0:*               LISTEN      888167/ceph-osd     
tcp        0      0 10.10.99.6:6809         0.0.0.0:*               LISTEN      888167/ceph-osd     
tcp        0      0 10.10.99.6:6810         0.0.0.0:*               LISTEN      888167/ceph-osd     
tcp        0      0 10.10.99.6:6811         0.0.0.0:*               LISTEN      888167/ceph-osd     
tcp        0      0 10.10.99.6:6812         0.0.0.0:*               LISTEN      888167/ceph-osd     
tcp        0      0 10.10.99.6:6813         0.0.0.0:*               LISTEN      888167/ceph-osd     
tcp        0      0 10.10.99.6:6814         0.0.0.0:*               LISTEN      888167/ceph-osd     
tcp        0      0 10.10.99.6:6815         0.0.0.0:*               LISTEN      888167/ceph-osd     
tcp        0      0 10.10.99.6:6789         0.0.0.0:*               LISTEN      879220/ceph-mon     
tcp        0      0 10.10.99.6:3300         0.0.0.0:*               LISTEN      879220/ceph-mon

I can telnet to both the monitor ports:
Code:
root@pve01:~#  telnet 10.10.99.7 3300
Trying 10.10.99.7...
Connected to 10.10.99.7.
Escape character is '^]'.
ceph v2
^]
telnet> quit
Connection closed.
root@pve01:~#  telnet 10.10.99.7 6789
Trying 10.10.99.7...
Connected to 10.10.99.7.
Escape character is '^]'.
ceph v027�

c�:

c^]
telnet> quit
Connection closed.

And from another node i can telnet into one of the OSD ports on pve01:
Code:
root@pve03:~# telnet 10.10.99.6 6823
Trying 10.10.99.6...
Connected to 10.10.99.6.
Escape character is '^]'.
�eph v027p�

c�(

^[^]
Just really strange...