Hi Proxmox Community,
We are currently running Proxmox 8.3.1 with multiple kubernetes clusters on top, accessing ceph using the ceph-csi plugin. The VMs running on Proxmox have access to the ceph public network. In one of our environments, we have 3 Proxmox nodes, each with 8x800GB SSDs in Dell R6615s, presented through Perc H755s in non-raid mode. The server profile is set to performance.
While this is working quite well for us, we have a huge amount of replication redundancy in our database volumes (and I believe it is causing issues with IOPS and throughput):
Given the above, what are the community's thoughts on this? We are also keen to look into anything else we can tune to improve our I/O.
Thanks
Chris
We are currently running Proxmox 8.3.1 with multiple kubernetes clusters on top, accessing ceph using the ceph-csi plugin. The VMs running on Proxmox have access to the ceph public network. In one of our environments, we have 3 Proxmox nodes, each with 8x800GB SSDs in Dell R6615s, presented through Perc H755s in non-raid mode. The server profile is set to performance.
While this is working quite well for us, we have a huge amount of replication redundancy in our database volumes (and I believe it is causing issues with IOPS and throughput):
- Elasticsearch
- We have all of our indices configured with a replication of 1, so each write into Elasticsearch translates into 6 writes (two Elastic nodes write to disk, 3 nodes)
- MySQL
- We have galera clusters, all in master mode, so each write translates into 9 writes (three nodes write to disk, 3 nodes)
- Redis
- This is less important, but we also have Redis doing the same thing - the Redis cluster is 3 masters and 3 slaves, so in theory every background write task causes 6 writes to the underlying disks.
Given the above, what are the community's thoughts on this? We are also keen to look into anything else we can tune to improve our I/O.
Thanks
Chris