I don't know how to fix this. I'm just starting out with ceph. It just keeps on showing active+clean+remapped. It doesn't fix it over time. How do I fix this? I just use the default replication rule for my pools.
I am running a cluster with 6 nodes on VE 5.4-3. Suddenly ceph through a health warning and now whole cluster is unusable.
I am not quite sure why and haven't found similar info on site. Can anyone help please?
Hi,
Recent updates have made ceph started to act very weird:
we keep loosing one OSD with following from syslog:
2020-10-17 04:28:21.922478 mon.n02-sxb-pve01 (mon.0) 912 : cluster [INF] osd.6 [v2:172.17.1.2:6814/308596,v1:172.17.1.2:6817/308596] boot
2020-10-17 04:28:23.919914...
Hi all,
I have a problem with my ceph installation, where my crush table changes after a reboot.
I have configured my ceph crush like this (i am giving the relevant config):
host px1 {
item osd.1 weight 1.637
item osd.2 weight 1.637
item osd.3 weight 1.637
item osd.4 weight...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.