Ceph OSD disk replacement

Paspao

Active Member
Aug 1, 2017
69
2
28
56
Hello,

I am looking for updated documentation on correct procedure for Ceph OSD disk replacement.

Current Proxmox docs only cover OSD creation but is lacking of management procedures using pveceph commands.

I found this Ceph document but some commands are not giving the same output (ie:
service ceph status ) and I am not sure if it is 100% Proxmox version compatible.

In general can we mix procedures using traditional ceph commands with pveceph ?

Thanks to everyone who can spread some light.

P.
 
In general can we mix procedures using traditional ceph commands with pveceph ?
yes. pveceph isnt an actual command binary, its a wrapper for ceph commands.

replacing a disk is super simple, and can be performed even from the gui:
upload_2019-5-23_11-31-42.png
1. down (stop) and out an osd (will probably already be in this state for a failed drive)
2. remove it from the tree and crush map ("destroy" in the gui)
3. replace disk
4. create new osd
5. profit :)
 
I found this Ceph document but some commands are not giving the same output (ie:
service ceph status ) and I am not sure if it is 100% Proxmox version compatible.
We use upstream packages (+ some extra patches) and just provide our management stack around it. In general the ceph tools should work as expected.
 
So glad I found this post. I was working on all the command lines for this. This was so much easier.

A note from my experience. The disk was failed but the mount point still existed on my server. Running 'destroy' from gui did not unmount the drive. After swapping the drive it was going to set the new mount point as sdm. I did a umount sdf, popped the new drive out and back in. Then adding the disk in GUI used the sdf mount point.
 
Last edited:
  • Like
Reactions: tuathan and dmulk
So glad I found this post. I was working on all the command lines for this. This was so much easier.

A note from my experience. The disk was failed but the mount point still existed on my server. Swapping the drive it was going to set the new mount point as sdm. I did a umount sdf, popped the new drive out and back in. Then adding the disk in GUI used the sdf mount point.

Seems like this is a bug with Proxmox's GUI wrapper...? Seems like "Destroy" should also unmount the OSD's mount point, no?
 
Hi David!

Just noticed nobody answered your question.
Yes you can use this to swop to larger drives.
And there is a trick, remapping takes quite a long time, and you can greatly speedup the process by :

$ceph tell 'osd.*' injectargs '--osd-max-backfills 16'

All the best

Ras
 
  • Like
Reactions: Adam Smith
The command is only persistent if you write it into the config. Otherwise you have to set the value again after restarting the specific service.
 
  • Like
Reactions: Adam Smith
The command is only persistent if you write it into the config. Otherwise you have to set the value again after restarting the specific service.
Perfect. Thank you.

Also, I did a little lookup. I had more luck with
Code:
# ceph tell 'osd.*' injectargs --osd-max-backfills 3 --osd-recovery-max 9

as described here.
 
  • Like
Reactions: takeokun