Can someone please suggest me what is the optimal configuration for the network, having 2 NICs at 10G and 1G for each node?
The network is not your problem. I doubt you are saturating your 10g network with the current setup.
Few things can be contributing to the issue and need answers,
- How many PGs?
- How many Pools?
- How many replica?
- Customized CrushMAP?
- How is the current health of ceph? (ceph -s)
- Scrub running continuously?
- Performed any tweaks on Ceph after deployment?
Ceph is not designed with small cluster in mind. The larger a Ceph cluster becomes, the faster it gets. There is no other storage with this character. But, Ceph does work in small environment such as 3 nodes, but there are things you must keep in mind so the expectation is within the boundary. To give an example, simply going from 4 nodes to 5, it adds roughly 20%-30% performance.
Replica is very very important when a small Ceph in question. You must not use 3 replica. That will kill performance more than anything. Replica 2 is the way to go. You may be using default 3 replica when created the pool. If you want to use small Ceph, the tradeoffs must be accepted. The replica count is one of them. PG count also affects performance quite a bit. Too low or too high, neither is good.
The initial cluster probably was fast before you started loading all of your data. As more and more data gets stored on Ceph, more replicas get created and the need of PG distribution increases. Do not forget, with replica 3, each incoming data gets written 3 times.
With larger spinning drives, performance does degrade a bit. As others have suggested using SSD as journal drive you can mitigate that issue easily. Yes you can enterprise grade SSD. But, there is nothing wrong with using good consumer grade SSDs to add performance. Specially if you add 2 SSDs in mirror to hold the DB/WAL, you can add performance without breaking bank. Lexar NS100 512GB or T-force Vulcan Z 1TB both are cheap viable option. The reason I mention these ones, I have used them in production Ceph after extensive testing.