I'm a bit confused about the autoscaler and PGs.
This cluster has Ceph 19.2.1, 18 OSDs, default 3/2 replicas and default target 100 PGs per OSD. BULK is false. Capacity is just under 18000G. A while back we set a target size of 1500G and we've been gradually approaching that, currently around 1300G SIZE (4.3 TB total Usage in GUI for all pools).
The autoscaler still has the pool set to 128 PG. Multiple calculators show a recommendation of 512 PG for the pool, using 100 PG/OSD, or 1024 with 200/OSD. Ceph says the default target is 100 PG per OSD but also immediately follows that sentence with, "For all but the very smallest deployments a value of 200 is recommended." What is a "very small deployment"?
The number of PGs per OSD varies with drive size (weight) of course but most drives are about 1 TB and so contain 30-35 PG, divided basically equally across nodes and factoring in disk size.
By default the autoscaler only triggers if there is a 3x difference from its recommendation. Is that 384 or 512 in this case? Setting
as mentioned in the docs doesn't immediately change the recommendation. Does this take a while to recalculate? Column NEW PG_NUM remains blank.
On one hand, I understand that "auto" means one might just walk away
, but I am confused why the calculators give a much higher number. Is it just that the pool is only ~25% full and it hasn't gotten around to creating more PGs yet?
Thanks in advance.
This cluster has Ceph 19.2.1, 18 OSDs, default 3/2 replicas and default target 100 PGs per OSD. BULK is false. Capacity is just under 18000G. A while back we set a target size of 1500G and we've been gradually approaching that, currently around 1300G SIZE (4.3 TB total Usage in GUI for all pools).
The autoscaler still has the pool set to 128 PG. Multiple calculators show a recommendation of 512 PG for the pool, using 100 PG/OSD, or 1024 with 200/OSD. Ceph says the default target is 100 PG per OSD but also immediately follows that sentence with, "For all but the very smallest deployments a value of 200 is recommended." What is a "very small deployment"?
The number of PGs per OSD varies with drive size (weight) of course but most drives are about 1 TB and so contain 30-35 PG, divided basically equally across nodes and factoring in disk size.
By default the autoscaler only triggers if there is a 3x difference from its recommendation. Is that 384 or 512 in this case? Setting
Code:
ceph osd pool set threshold 2.0
On one hand, I understand that "auto" means one might just walk away

Thanks in advance.
Last edited: