Hello,
I want to enlarge a Ceph cluster which uses filestore backend (want to keep that, it is running fine for like 5 years). We have 4 ceph nodes, each with 8 OSD á 3TB. The cluster is sometimes filled up to 85% (WARNING) and I have to manually intervene and free some storage space.
Each OSD in a node uses a 20GB journal partition on a 200GB Intel Optane DC P4801X NVMe. Got here now 24 x 6TB HDD WD Red which will replace the 3TB HDDs. I just want to make sure that I am doing all the steps required for the replacement correctly.
Edit:
Steps to replace OSDs one by one, node by node:
1. Down OSD [N] via GUI
2. Out OSD [N] via GUI
3. Destroy OSD [N] via GUI
4. Replace HDD in Server and check for correct detection
5. Now on console, create the new OSD with the former journal partition (check correct journal partition, blkid, cat /var/lib/ceph/osd/ceph-[N]/journal_uuid)
-> ceph-volume lvm create --filestore --data /dev/sd[X] --journal /dev/nvme0n1p[Y]
6. In OSD [N] via GUI
7. Start OSD [N] via GUI
8. Wait till cluster is healthy
9. Repeat from step 1. with next OSD
Am I doing it right? Do I miss something? I would appreciate it if someone can confirm my approach, please?
Thank you and have a nice week everyone !
Kind Regards
ekin06
I want to enlarge a Ceph cluster which uses filestore backend (want to keep that, it is running fine for like 5 years). We have 4 ceph nodes, each with 8 OSD á 3TB. The cluster is sometimes filled up to 85% (WARNING) and I have to manually intervene and free some storage space.
Each OSD in a node uses a 20GB journal partition on a 200GB Intel Optane DC P4801X NVMe. Got here now 24 x 6TB HDD WD Red which will replace the 3TB HDDs. I just want to make sure that I am doing all the steps required for the replacement correctly.
Edit:
Code:
# ceph -v
ceph version 14.2.22 (877fa256043e4743620f4677e72dee5e738d1226) nautilus (stable)
# pveversion -v
proxmox-ve: 6.4-1 (running kernel: 5.4.157-1-pve)
pve-manager: 6.4-13 (running version: 6.4-13/9f411e79)
pve-kernel-5.4: 6.4-11
pve-kernel-helper: 6.4-11
pve-kernel-5.4.157-1-pve: 5.4.157-1
pve-kernel-5.4.114-1-pve: 5.4.114-1
pve-kernel-5.4.65-1-pve: 5.4.65-1
pve-kernel-4.15: 5.4-19
pve-kernel-4.15.18-30-pve: 4.15.18-58
pve-kernel-4.15.18-12-pve: 4.15.18-36
pve-kernel-4.15.18-4-pve: 4.15.18-23
pve-kernel-4.15.17-3-pve: 4.15.17-14
pve-kernel-4.15.17-1-pve: 4.15.17-9
pve-kernel-4.13.13-6-pve: 4.13.13-42
pve-kernel-4.13.13-5-pve: 4.13.13-38
pve-kernel-4.13.13-2-pve: 4.13.13-33
ceph: 14.2.22-pve1
ceph-fuse: 14.2.22-pve1
corosync: 3.1.5-pve2~bpo10+1
criu: 3.11-3
glusterfs-client: 5.5-3
ifupdown: 0.8.35+pve1
ksm-control-daemon: 1.3-1
libjs-extjs: 6.0.1-10
libknet1: 1.22-pve2~bpo10+1
libproxmox-acme-perl: 1.1.0
libproxmox-backup-qemu0: 1.1.0-1
libpve-access-control: 6.4-3
libpve-apiclient-perl: 3.1-3
libpve-common-perl: 6.4-4
libpve-guest-common-perl: 3.1-5
libpve-http-server-perl: 3.2-3
libpve-storage-perl: 6.4-1
libqb0: 1.0.5-1
libspice-server1: 0.14.2-4~pve6+1
lvm2: 2.03.02-pve4
lxc-pve: 4.0.6-2
lxcfs: 4.0.6-pve1
novnc-pve: 1.1.0-1
proxmox-backup-client: 1.1.13-2
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.6-1
pve-cluster: 6.4-1
pve-container: 3.3-6
pve-docs: 6.4-2
pve-edk2-firmware: 2.20200531-1
pve-firewall: 4.1-4
pve-firmware: 3.3-2
pve-ha-manager: 3.1-1
pve-i18n: 2.3-1
pve-qemu-kvm: 5.2.0-6
pve-xtermjs: 4.7.0-3
qemu-server: 6.4-2
smartmontools: 7.2-pve2
spiceterm: 3.1-1
vncterm: 1.6-2
zfsutils-linux: 2.0.6-pve1~bpo10+1
Steps to replace OSDs one by one, node by node:
1. Down OSD [N] via GUI
2. Out OSD [N] via GUI
3. Destroy OSD [N] via GUI
4. Replace HDD in Server and check for correct detection
5. Now on console, create the new OSD with the former journal partition (check correct journal partition, blkid, cat /var/lib/ceph/osd/ceph-[N]/journal_uuid)
-> ceph-volume lvm create --filestore --data /dev/sd[X] --journal /dev/nvme0n1p[Y]
6. In OSD [N] via GUI
7. Start OSD [N] via GUI
8. Wait till cluster is healthy
9. Repeat from step 1. with next OSD
Am I doing it right? Do I miss something? I would appreciate it if someone can confirm my approach, please?
Thank you and have a nice week everyone !
Kind Regards
ekin06
Last edited: