Hello,
I'm trying to import two OSDs from a brand new node.
I can see the OSDs in the interface, but when I start the OSD, I get an error:
systemctl status ceph-osd@2.service shows me this:
journalctl -xeu ceph-osd@2.service also has no useful information.
This is my ceph.conf
ceph-volume inventory /dev/sdb
Where can I find any log that tells me why the OSDs are not starting? What can I possibly be doing wrong? Thanks!
I'm trying to import two OSDs from a brand new node.
I can see the OSDs in the interface, but when I start the OSD, I get an error:
Code:
Job for ceph-osd@2.service failed because the control process exited with error code.
See "systemctl status ceph-osd@2.service" and "journalctl -xeu ceph-osd@2.service" for details.
TASK ERROR: command '/bin/systemctl start ceph-osd@2' failed: exit code 1
systemctl status ceph-osd@2.service shows me this:
Code:
ceph-osd@2.service - Ceph object storage daemon osd.2
Loaded: loaded (/lib/systemd/system/ceph-osd@.service; disabled; preset: enabled)
Drop-In: /usr/lib/systemd/system/ceph-osd@.service.d
└─ceph-after-pve-cluster.conf
Active: failed (Result: exit-code) since Wed 2024-03-13 21:17:05 PDT; 3min 24s ago
Process: 2878 ExecStartPre=/usr/libexec/ceph/ceph-osd-prestart.sh --cluster ${CLUST>
CPU: 13ms
Mar 13 21:17:04 pve systemd[1]: Failed to start ceph-osd@2.service - Ceph object storag>
Mar 13 21:17:05 pve systemd[1]: ceph-osd@2.service: Start request repeated too quickly.
Mar 13 21:17:05 pve systemd[1]: ceph-osd@2.service: Failed with result 'exit-code'.
Mar 13 21:17:05 pve systemd[1]: Failed to start ceph-osd@2.service - Ceph object storag>
Mar 13 21:18:25 pve systemd[1]: ceph-osd@2.service: Start request repeated too quickly.
Mar 13 21:18:25 pve systemd[1]: ceph-osd@2.service: Failed with result 'exit-code'.
Mar 13 21:18:25 pve systemd[1]: Failed to start ceph-osd@2.service - Ceph object stora
journalctl -xeu ceph-osd@2.service also has no useful information.
This is my ceph.conf
Code:
[global]
auth_client_required = cephx
auth_cluster_required = cephx
auth_service_required = cephx
cluster_network = 10.0.20.4/24
err_to_syslog = true
fsid = 5ce42d57-4371-475a-94fb-eac8acefe72e
mon_allow_pool_delete = true
mon_allow_pool_size_one = false
# mon_cluster_log_file_level = info
# mon_cluster_log_to_file = false
mon_host = 10.0.20.3 10.0.20.4 10.0.20.1
ms_bind_ipv4 = true
ms_bind_ipv6 = false
osd_deep_scrub_interval = 1209600
osd_pool_default_min_size = 2
osd_pool_default_size = 3
osd_scrub_begin_hour = 23
osd_scrub_end_hour = 7
osd_scrub_sleep = 0.1
public_network = 10.0.20.4/24
[client]
keyring = /etc/pve/priv/$cluster.$name.keyring
[mds]
keyring = /var/lib/ceph/mds/ceph-$id/keyring
[mds.pve-ceph1-1]
host = pve-ceph1
mds_standby_for_name = pve
[mds.pve-ceph1-2]
host = pve-ceph1
mds_standby_for_name = pve
[mds.pve-ceph1-3]
host = pve-ceph1
mds_standby_for_name = pve
[mds.pve2-1]
host = pve2
mds_standby_for_name = pve
[mds.pve3-1]
host = pve3
mds_standby_for_name = pve
[mon.pve]
public_addr = 10.0.20.1
[mon.pve-ceph1]
# debug_mon = 0/5
public_addr = 10.0.20.4
[mon.pve3]
# debug_mon = 0/5
public_addr = 10.0.20.3
ceph-volume inventory /dev/sdb
Code:
====== Device report /dev/sdb ======
path /dev/sdb
ceph device True
lsm data {}
available False
rejected reasons LVM detected, Has a FileSystem, Insufficient space (<10 extents) on vgs
device id CT4000MX500SSD1_2339E879E476
removable 0
ro 0
vendor ATA
model CT4000MX500SSD1
sas address
rotational 0
actuators None
scheduler mode mq-deadline
human readable size 3.64 TB
--- Logical Volume ---
name osd-block-f1e4ea87-d73e-47ca-8648-c6373110f6ea
osd id 2
cluster name ceph
type block
osd fsid f1e4ea87-d73e-47ca-8648-c6373110f6ea
cluster fsid 5ce42d57-4371-475a-94fb-eac8acefe72e
osdspec affinity
block uuid Ub2tvB-R9Hu-bQHW-92Aa-Hcyq-QP7P-dwZzi0
Where can I find any log that tells me why the OSDs are not starting? What can I possibly be doing wrong? Thanks!
Last edited: