Greetings!
We have a 3 node Proxmox cluster with 7 disk each.
All disk are Intel SSD DC S4510 that are 1.92TB capacity. All running the same firmware.
Each node is connected via a 25G MLAG pair via a Mellanox ConnectX-4.
Recently ceph has triggered the following alert:
For whatever reason it doesn't appear the OSD's are super balanced? Looks like we have some room to spare in the cluster. I did not setup the cluster orginally so I'm not sure if everything is set to defaults or some settings have been adjusted.
Here is some of the helpful output:
There's quite a few knobs to turn but it isn't apparent what is best (if that exists). Any suggestions?
Thanks!
We have a 3 node Proxmox cluster with 7 disk each.
All disk are Intel SSD DC S4510 that are 1.92TB capacity. All running the same firmware.
Each node is connected via a 25G MLAG pair via a Mellanox ConnectX-4.
Recently ceph has triggered the following alert:
Code:
# ceph health
HEALTH_WARN 1 backfillfull osd(s); 4 pool(s) backfillfull
For whatever reason it doesn't appear the OSD's are super balanced? Looks like we have some room to spare in the cluster. I did not setup the cluster orginally so I'm not sure if everything is set to defaults or some settings have been adjusted.
Here is some of the helpful output:
Code:
# pveversion -v
proxmox-ve: 7.4-1 (running kernel: 5.15.126-1-pve)
pve-manager: 7.4-17 (running version: 7.4-17/513c62be)
pve-kernel-5.15: 7.4-7
pve-kernel-5.15.126-1-pve: 5.15.126-1
pve-kernel-5.15.108-1-pve: 5.15.108-2
pve-kernel-5.15.83-1-pve: 5.15.83-1
pve-kernel-5.15.74-1-pve: 5.15.74-1
ceph: 17.2.5-pve1
ceph-fuse: 17.2.5-pve1
corosync: 3.1.7-pve1
criu: 3.15-1+pve-1
glusterfs-client: 10.2-ubuntu1~focal1
ifupdown2: 3.1.0-1+pmx4
ksm-control-daemon: 1.4-1
libjs-extjs: 7.0.0-1
libknet1: 1.24-pve2
libproxmox-acme-perl: 1.4.4
libproxmox-backup-qemu0: 1.3.1-1
libproxmox-rs-perl: 0.2.1
libpve-access-control: 7.4.1
libpve-apiclient-perl: 3.2-1
libpve-common-perl: 7.4-2
libpve-guest-common-perl: 4.2-4
libpve-http-server-perl: 4.2-3
libpve-rs-perl: 0.7.7
libpve-storage-perl: 7.4-3
libspice-server1: 0.14.3-2.1
lvm2: 2.03.11-2.1
lxc-pve: 5.0.2-2
lxcfs: 5.0.3-pve1
novnc-pve: 1.4.0-1
proxmox-backup-client: 2.4.3-1
proxmox-backup-file-restore: 2.4.3-1
proxmox-kernel-helper: 7.4-1
proxmox-mail-forward: 0.1.1-1
proxmox-mini-journalreader: 1.3-1
proxmox-widget-toolkit: 3.7.3
pve-cluster: 7.3-3
pve-container: 4.4-6
pve-docs: 7.4-2
pve-edk2-firmware: 3.20230228-4~bpo11+1
pve-firewall: 4.3-5
pve-firmware: 3.6-5
pve-ha-manager: 3.6.1
pve-i18n: 2.12-1
pve-qemu-kvm: 7.2.0-8
pve-xtermjs: 4.16.0-2
qemu-server: 7.4-4
smartmontools: 7.2-pve3
spiceterm: 3.2-2
swtpm: 0.8.0~bpo11+3
vncterm: 1.7-1
zfsutils-linux: 2.1.11-pve1
Code:
# ceph version
ceph version 17.2.5 (e04241aa9b639588fa6c864845287d2824cb6b55) quincy (stable)
Code:
# ceph status
cluster:
id: b7565e52-6907-49f9-85b9-526c3ce94676
health: HEALTH_WARN
1 backfillfull osd(s)
4 pool(s) backfillfull
services:
mon: 3 daemons, quorum hv01,hv02,hv03 (age 3M)
mgr: hv01(active, since 3M), standbys: hv02, hv03
mds: 1/1 daemons up, 2 standby
osd: 20 osds: 20 up (since 3M), 20 in (since 8M)
data:
volumes: 1/1 healthy
pools: 4 pools, 577 pgs
objects: 2.16M objects, 8.2 TiB
usage: 24 TiB used, 11 TiB / 35 TiB avail
pgs: 577 active+clean
io:
client: 450 MiB/s rd, 12 MiB/s wr, 2.02k op/s rd, 321 op/s wr
Code:
# ceph health detail
HEALTH_WARN 1 backfillfull osd(s); 4 pool(s) backfillfull
[WRN] OSD_BACKFILLFULL: 1 backfillfull osd(s)
osd.5 is backfill full
[WRN] POOL_BACKFILLFULL: 4 pool(s) backfillfull
pool '.mgr' is backfillfull
pool 'cephfs_data' is backfillfull
pool 'cephfs_metadata' is backfillfull
pool 'ceph-vm' is backfillfull
Code:
# ceph df detail
--- RAW STORAGE ---
CLASS SIZE AVAIL USED RAW USED %RAW USED
ssd 35 TiB 11 TiB 24 TiB 24 TiB 69.81
TOTAL 35 TiB 11 TiB 24 TiB 24 TiB 69.81
--- POOLS ---
POOL ID PGS STORED (DATA) (OMAP) OBJECTS USED (DATA) (OMAP) %USED MAX AVAIL QUOTA OBJECTS QUOTA BYTES DIRTY USED COMPR UNDER COMPR
.mgr 1 1 264 MiB 264 MiB 0 B 67 792 MiB 792 MiB 0 B 0.05 518 GiB N/A N/A N/A 0 B 0 B
cephfs_data 3 32 75 GiB 75 GiB 0 B 19.33k 226 GiB 226 GiB 0 B 12.73 518 GiB N/A N/A N/A 0 B 0 B
cephfs_metadata 4 32 36 MiB 36 MiB 5.2 KiB 31 108 MiB 108 MiB 15 KiB 0 518 GiB N/A N/A N/A 0 B 0 B
ceph-vm 6 512 8.0 TiB 8.0 TiB 109 KiB 2.14M 24 TiB 24 TiB 326 KiB 94.08 518 GiB N/A N/A N/A 0 B 0 B
Code:
# ceph osd df tree
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META AVAIL %USE VAR PGS STATUS TYPE NAME
-1 34.93195 - 35 TiB 24 TiB 24 TiB 4.3 MiB 70 GiB 11 TiB 69.81 1.00 - root default
-3 12.22618 - 12 TiB 8.1 TiB 8.1 TiB 1.5 MiB 24 GiB 4.1 TiB 66.49 0.95 - host hv01
0 ssd 1.74660 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 214 KiB 3.2 GiB 662 GiB 62.98 0.90 77 up osd.0
3 ssd 1.74660 1.00000 1.7 TiB 1.2 TiB 1.2 TiB 213 KiB 3.3 GiB 570 GiB 68.11 0.98 85 up osd.3
6 ssd 1.74660 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 201 KiB 3.5 GiB 615 GiB 65.59 0.94 81 up osd.6
9 ssd 1.74660 1.00000 1.7 TiB 1.3 TiB 1.3 TiB 230 KiB 3.6 GiB 474 GiB 73.49 1.05 94 up osd.9
15 ssd 1.74660 1.00000 1.7 TiB 1.2 TiB 1.2 TiB 232 KiB 3.6 GiB 551 GiB 69.19 0.99 89 up osd.15
16 ssd 1.74660 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 204 KiB 3.5 GiB 633 GiB 64.59 0.93 76 up osd.16
17 ssd 1.74660 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 201 KiB 3.3 GiB 688 GiB 61.51 0.88 75 up osd.17
-5 12.22618 - 12 TiB 8.1 TiB 8.1 TiB 1.4 MiB 24 GiB 4.1 TiB 66.49 0.95 - host hv02
1 ssd 1.74660 1.00000 1.7 TiB 1.2 TiB 1.2 TiB 221 KiB 3.5 GiB 557 GiB 68.84 0.99 84 up osd.1
4 ssd 1.74660 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 204 KiB 3.4 GiB 642 GiB 64.09 0.92 83 up osd.4
7 ssd 1.74660 1.00000 1.7 TiB 1.2 TiB 1.2 TiB 223 KiB 3.4 GiB 569 GiB 68.17 0.98 86 up osd.7
10 ssd 1.74660 1.00000 1.7 TiB 1.3 TiB 1.3 TiB 224 KiB 3.7 GiB 490 GiB 72.60 1.04 86 up osd.10
14 ssd 1.74660 1.00000 1.7 TiB 1.2 TiB 1.2 TiB 211 KiB 3.1 GiB 592 GiB 66.91 0.96 79 up osd.14
18 ssd 1.74660 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 201 KiB 3.3 GiB 647 GiB 63.83 0.91 81 up osd.18
19 ssd 1.74660 1.00000 1.7 TiB 1.1 TiB 1.1 TiB 199 KiB 3.3 GiB 697 GiB 61.00 0.87 78 up osd.19
-7 10.47958 - 10 TiB 8.1 TiB 8.1 TiB 1.4 MiB 22 GiB 2.4 TiB 77.56 1.11 - host hv03
2 ssd 1.74660 1.00000 1.7 TiB 1.3 TiB 1.3 TiB 243 KiB 3.5 GiB 433 GiB 75.82 1.09 93 up osd.2
5 ssd 1.74660 1.00000 1.7 TiB 1.6 TiB 1.6 TiB 285 KiB 4.1 GiB 167 GiB 90.66 1.30 109 up osd.5
8 ssd 1.74660 1.00000 1.7 TiB 1.4 TiB 1.4 TiB 231 KiB 3.7 GiB 393 GiB 78.04 1.12 95 up osd.8
11 ssd 1.74660 1.00000 1.7 TiB 1.3 TiB 1.3 TiB 245 KiB 3.7 GiB 413 GiB 76.91 1.10 96 up osd.11
12 ssd 1.74660 1.00000 1.7 TiB 1.3 TiB 1.3 TiB 234 KiB 3.5 GiB 484 GiB 72.92 1.04 93 up osd.12
13 ssd 1.74660 1.00000 1.7 TiB 1.2 TiB 1.2 TiB 238 KiB 3.5 GiB 519 GiB 71.01 1.02 91 up osd.13
TOTAL 35 TiB 24 TiB 24 TiB 4.4 MiB 70 GiB 11 TiB 69.81
Code:
# ceph osd dump | grep ratio
full_ratio 0.95
backfillfull_ratio 0.9
nearfull_ratio 0.85
There's quite a few knobs to turn but it isn't apparent what is best (if that exists). Any suggestions?
Thanks!