Ceph performance

lifeboy

Renowned Member
I need some input on tuning performance on a new cluster I have setup

The new cluster has 2 pool (one for HDDs and one for SSDs). For now it's only three nodes.
I have separate networks: 1 x 1Gb/s NIC for corosync, 2 x bonded 1GB/s NICs for ceph and 1 x 1GB/s NIC for the Proxmox bridged VM's and LXC's.

The SDD pool has the following performance:

Code:
# rados bench -p fast 60 rand -t 1
Total time run:       60.067481
Total reads made:     2506
Read size:            4194304
Object size:          4194304
Bandwidth (MB/sec):   166.879
Average IOPS:         41
Stddev IOPS:          4
Max IOPS:             53
Min IOPS:             33
Average Latency(s):   0.0235738
Max latency(s):       0.0469368
Min latency(s):       0.0011793

Now, I also have an old Proxmox 4 jewel-based ceph cluster with old SAS HDD's that gives me this performance:
Code:
# rados bench -p IMB-test 60 rand -t 1
Total time run:       60.179736
Total reads made:     788
Read size:            4194304
Object size:          4194304
Bandwidth (MB/sec):   52.3764
Average IOPS:         13
Stddev IOPS:          2
Max IOPS:             19
Min IOPS:             7
Average Latency(s):   0.0756719
Max latency(s):       0.341299
Min latency(s):       0.00517604

What is happening, or am I hitting the 1Gb/s network bottleneck here?
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!