Hi,
We increased our PG count from 256 to 512 over the weekend to accommodate our growing CEPH cluster. The cluster has been in a healthy state and everything appears to be ok, except we have noticed 2 PG have been deep scrubbing for over 24 hours now.
My questions:
1). Could this be Ceph...
We are currently running a 8 node cluster - 5 nodes are for computing, 3 nodes are for ceph. It's all in HA on a 10GB network - backups to NFS. Our data transfer speeds are around 40MB/s which is unacceptable for our needs. Curious if migration our backup solution to PBS improves this...
We're currently running a 8 node cluster - 3 nodes are dedicated to CEPH. All our QCOW2 files reside on the CEPH cluster and we do a vma.zst backup to an NFS share. Our network is all 10GBIT and we're seeing only 40MB/S backups which is becoming a big issue as our client base grows on this...
Figured it out.
I need to review my switch config. Server is setup on a bonded interface running LACP. When I shutdown the port on one of two switches it resolves the OSD flapping and Ceph errors.
Hi All,
I've done my best to read documentation and research the forums / web. I keep having these health errors pop up. I've tried a ton of different PG configurations including 128, 1024, 2000, etc. I can't seem to nail this setup - I've tried using calculators as well.
I have 3 nodes, 8...
One disk on one of our servers is labeled as a 'partition' (not sure why - clean install and all other nodes don't have the same issue) - as a result not able to add the disk as an OSD to our Ceph cluster. Anyone have this issue and know how to fix?
TIA
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.