I have a three-node up-to-date (no subscription) Proxmox cluster home lab with AMD 8-core CPUs, RAID card X8 sata in JBOD and 1 SSD for the system. Network is 1Gbps on unmanaged switches. Each node is also used as a 4-OSD Ceph node with an isolated 2 x 1Gbps NICs bonded Round-Robin redundant link using separate switches with STP active. Bandwidth is 1.7 Gbps on the Bond and connectivity is fine to all three Ceph nodes. Each OSD is a WD Red 1TB 2.5" SATA drive but my first node also has 3 other drives used for testing BTRFS (separate of Ceph) and 2 unused connected drives.
I have 3 Ceph pools configured: data (2/2) is 25% full, pve (3/2) is 10% full and metadata (3/2) is minimal. I use the data pool for CephFS testing and the pve pool for rbd. PGs and PGPs is configured equally at 512 for all pools. When all drives are Up/In and the Ceph shows as Healthy, access is acceptable. I can easily use data on CephFS to serve mp4 or mkv movies remotely in HD quality (LAN) and the KVM VM functionality is fine.
I had to replace one OSD on node 1 with another unused drive connected to the same controller in order to get the Ceph system stable. I had to try two drives before I could get Ceph stable though.
This is what happened:
This is what I tried:
I can not understand why these two drives do not work as OSDs. The behavior is repeatable and re-balance usually starts between 20 and 30%. Ceph logs show slow progress with many pages unhealthy.
Can anyone give me a hint on where to look?
Serge
I have 3 Ceph pools configured: data (2/2) is 25% full, pve (3/2) is 10% full and metadata (3/2) is minimal. I use the data pool for CephFS testing and the pve pool for rbd. PGs and PGPs is configured equally at 512 for all pools. When all drives are Up/In and the Ceph shows as Healthy, access is acceptable. I can easily use data on CephFS to serve mp4 or mkv movies remotely in HD quality (LAN) and the KVM VM functionality is fine.
I had to replace one OSD on node 1 with another unused drive connected to the same controller in order to get the Ceph system stable. I had to try two drives before I could get Ceph stable though.
This is what happened:
- One of the original Ceph OSD failed after a few hours of having a Ceph Healthy system. The first spare drive (Fujitsu 1 TB 3.5") I tried also fails. The second unused drive (also a Fujitsu 1 TB 3.5") is working.
This is what I tried:
- The two failing OSD drives pass all error checks (SMART, gparted, gdisk) and nothing shows up in any logs.
- Every time I attempt to add them as OSD (through the Proxmox web interface), creation is successful. They are Up/In and the entire Ceph system attempts to re-balance.
- After a few hours, the re-balance is almost completed but stops at less than 1% remaining and the OSD goes Down/Out. Any attempt to start it fails with an error 1 dialog.
- I have attempted to re-configure these two drives as OSD at least 3 times with zapping the drives in-between attempts with no change.
I can not understand why these two drives do not work as OSDs. The behavior is repeatable and re-balance usually starts between 20 and 30%. Ceph logs show slow progress with many pages unhealthy.
Can anyone give me a hint on where to look?
Serge