Alert messages in CEPH

The message will still be there, as it still has a 'target ratio'. Run ceph osd pool set rbd target_size_ratio 0 to remove the ratio, then the message should disappear. The cluster may rebalance.
 
Hi Alwin,

both messages still appear after executing: ceph osd pool set ceph-vm target size ratio 0

Do i have to do something else?
 

Attachments

  • Captura5.PNG
    Captura5.PNG
    3.9 KB · Views: 7
  • Captura6.PNG
    Captura6.PNG
    26.7 KB · Views: 7
How does the status look now, ceph osd pool autoscale-status?
 
POOL SIZE TARGET SIZE RATE RAW CAPACITY RATIO TARGET RATIO BIAS PG_NUM NEW PG_NUM AUTOSCALE
ceph-vm 5173G 3.0 8379G 1.8523 1.0 256
 

Attachments

  • Captura7.PNG
    Captura7.PNG
    19.2 KB · Views: 17
For me the message went away after I removed the ratios. But maybe you need to disable the autoscaler, to make it work.
 
Not sure what to tell you now. Maybe restarting the MGR(s) helps. But besides that a post on the ceph mailing list might bring new ideas.
 
For those who may come after me... I had the same problem and was able to fix it as follows. First run the command:

ceph osd pool autoscale-status

It gave:
POOL|SIZE|TARGET SIZE|RATE|RAW CAPACITY|RATIO|TARGET RATIO|EFFECTIVE RATIO|BIAS|PG_NUM|NEW PG_NUM|AUTOSCALE|PROFILE
device_health_metrics|36087k|blank |3.0|59616G|0.0000|blank |blank |1.0|1|blank|on|scale-up
CephPool|713.5G|20000G|3.0|59616G|1.0064|blank|blank|1.0| 32|blank|on|scale-up

I noticed that if you take 59616, the raw capacity, and divide it by three (that's the number of replicas I have and the default value if I remember right), that's how much space you really have avaible. Divide that into the target size of 20,000G and you get the ratio of 1.0064. Hence the target size needs to be reduced. How much? 20000G/1.0064 should do the trick. I calculated mine based on what I know the raw disk space really is and came up with 19868G for target size. Changing that in the WebGUI->node->Ceph->pool, select pool name and edit, set target size to lower value calculated for your case and save. Be patient, it takes a few minutes for the system to catch up and remove the warning. For me it took about 5 minutes.

A second look at the status produced:

POOL|SIZE|TARGET SIZE|RATE|RAW CAPACITY|RATIO|TARGET RATIO|EFFECTIVE RATIO|BIAS|PG_NUM|NEW PG_NUM|AUTOSCALE|PROFILE
...
CephPool|713.5G|19868G|3.0||59616G|0.9998|1.0|32|256|on|scale-up

and the warning is gone!
 
  • Like
Reactions: Neobin

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!