Hello everyone.
I have a 3 node cluster that has been running well for a couple of years. Recently one of the OSDs has stopped working, and I'm unable to start it again.
The OSD doesn't start when I press the 'Start' button, and behaves differently when the OSD in 'In' or 'Out' of the cluster.
The ceph health status is below:
The last few lines of the ceph-osd.0.log is below, but being inexperienced with ceph the only thing that stands out is the last line "ERROR: osd init failed: (22) Invalid argument". However, I'm unsure what argument it's referring to, and I haven't made any changes to the ceph configuration since it was built (although I did upgrade from Proxmox 6.x to 8.2.2 a few weeks ago).
Any help would be much appreciated.
I have a 3 node cluster that has been running well for a couple of years. Recently one of the OSDs has stopped working, and I'm unable to start it again.
The OSD doesn't start when I press the 'Start' button, and behaves differently when the OSD in 'In' or 'Out' of the cluster.
- Out: The Proxmox log states: SRV osd.0 - Start but nothing seems to happen
- In: the following error appears
The ceph health status is below:
The last few lines of the ceph-osd.0.log is below, but being inexperienced with ceph the only thing that stands out is the last line "ERROR: osd init failed: (22) Invalid argument". However, I'm unsure what argument it's referring to, and I haven't made any changes to the ceph configuration since it was built (although I did upgrade from Proxmox 6.x to 8.2.2 a few weeks ago).
Code:
2024-06-27T20:56:30.102+1000 721e93c006c0 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1719485790103248, "job": 5, "event": "compaction_started", "compaction_reason": "LevelL0FilesNum", "files_L0": [18806, 18794, 18779, 18761], "files_L1": [18742], "score": 1, "input_data_size": 44726858}
2024-06-27T20:56:30.107+1000 721ea5e8f3c0 -1 bluestore(/var/lib/ceph/osd/ceph-0) _verify_csum bad crc32c/0x1000 checksum at blob offset 0x0, got 0xa2190997, expected 0x19221daa, device location [0x2000~1000], logical extent 0x0~1000, object #-1:7b3f43c4:::osd_superblock:0#
2024-06-27T20:56:30.107+1000 721ea5e8f3c0 -1 bluestore(/var/lib/ceph/osd/ceph-0) _verify_csum bad crc32c/0x1000 checksum at blob offset 0x0, got 0xa2190997, expected 0x19221daa, device location [0x2000~1000], logical extent 0x0~1000, object #-1:7b3f43c4:::osd_superblock:0#
2024-06-27T20:56:30.108+1000 721ea5e8f3c0 -1 bluestore(/var/lib/ceph/osd/ceph-0) _verify_csum bad crc32c/0x1000 checksum at blob offset 0x0, got 0xa2190997, expected 0x19221daa, device location [0x2000~1000], logical extent 0x0~1000, object #-1:7b3f43c4:::osd_superblock:0#
2024-06-27T20:56:30.108+1000 721ea5e8f3c0 -1 bluestore(/var/lib/ceph/osd/ceph-0) _verify_csum bad crc32c/0x1000 checksum at blob offset 0x0, got 0xa2190997, expected 0x19221daa, device location [0x2000~1000], logical extent 0x0~1000, object #-1:7b3f43c4:::osd_superblock:0#
2024-06-27T20:56:30.108+1000 721ea5e8f3c0 -1 osd.0 0 OSD::init() : unable to read osd superblock
2024-06-27T20:56:30.108+1000 721e91e006c0 0 bluestore(/var/lib/ceph/osd/ceph-0) allocation stats probe 0: cnt: 0 frags: 0 size: 0
2024-06-27T20:56:30.108+1000 721e91e006c0 0 bluestore(/var/lib/ceph/osd/ceph-0) probe -1: 0, 0, 0
2024-06-27T20:56:30.108+1000 721e91e006c0 0 bluestore(/var/lib/ceph/osd/ceph-0) probe -2: 0, 0, 0
2024-06-27T20:56:30.108+1000 721e91e006c0 0 bluestore(/var/lib/ceph/osd/ceph-0) probe -4: 0, 0, 0
2024-06-27T20:56:30.108+1000 721e91e006c0 0 bluestore(/var/lib/ceph/osd/ceph-0) probe -8: 0, 0, 0
2024-06-27T20:56:30.108+1000 721e91e006c0 0 bluestore(/var/lib/ceph/osd/ceph-0) probe -16: 0, 0, 02024-06-27T20:56:30.108+1000 721e91e006c0 0 bluestore(/var/lib/ceph/osd/ceph-0) ------------
2024-06-27T20:56:30.112+1000 721ea5e8f3c0 4 rocksdb: [db/db_impl/db_impl.cc:446] Shutdown: canceling all background work2024-06-27T20:56:30.119+1000 721e93c006c0 4 rocksdb: (Original Log Time 2024/06/27-20:56:30.120256) [db/compaction/compaction_job.cc:812] [p-1] compacted to: files[4 1 0 0 0 0 0] max score 0.00, MB/sec: 2632.7 rd, 0.0 wr, level 1, files in(4, 1) out(0) MB in(7.6, 35.1) out(0.0), read-write-amplify(5.6) write-amplify(0.0) Shutdown in progress: Database shutdown, records in: 243557, records dropped: 243557 output_compression: NoCompression
2024-06-27T20:56:30.119+1000 721e93c006c0 4 rocksdb: (Original Log Time 2024/06/27-20:56:30.120274) EVENT_LOG_v1 {"time_micros": 1719485790120268, "job": 5, "event": "compaction_finished", "compaction_time_micros": 16989, "compaction_time_cpu_micros": 3893, "output_level": 1, "num_output_files": 0, "total_output_size": 0, "num_input_records": 243557, "num_output_records": 0, "num_subcompactions": 1, "output_compression": "NoCompression", "num_single_delete_mismatches": 0, "num_single_delete_fallthrough": 0, "lsm_state": [4, 1, 0, 0, 0, 0, 0]}
2024-06-27T20:56:30.120+1000 721ea5e8f3c0 4 rocksdb: [db/db_impl/db_impl.cc:625] Shutdown complete
2024-06-27T20:56:30.203+1000 721ea5e8f3c0 1 bluefs umount
2024-06-27T20:56:30.203+1000 721ea5e8f3c0 1 bdev(0x59373c281000 /var/lib/ceph/osd/ceph-0/block) close2024-06-27T20:56:30.468+1000 721ea5e8f3c0 1 freelist shutdown
2024-06-27T20:56:30.484+1000 721ea5e8f3c0 1 bdev(0x59373c159800 /var/lib/ceph/osd/ceph-0/block) close
2024-06-27T20:56:30.623+1000 721ea5e8f3c0 -1 ^[[0;31m ** ERROR: osd init failed: (22) Invalid argument^[[0m
~
"ceph-osd.0.log" 48184 lines, 4386240 bytes
Any help would be much appreciated.
Last edited: