Hi guys,
Need some bright minds in the Proxmox community
I setup a storage cluster using Ceph (as the title suggests). It's a fairly large cluster consisting of over 200 osds. When I bench the cluster using rados bench, I get exactly the kind of performance I was expecting to get... doing dd to a single drive on the node I get on average 205MB/s .. writing to the cluster via rados I get 400-500MB/s writes need to be at least in the area of that in my VMs.
I installed Ubuntu 16.04x64 in KVM with a 4TB disk partition with ext4. Same dd test there I was able to get 25MB/s write on a 1G chunk and 15-20MB (varying) on a 2G chunk. I am more interested in the sustained write as that's what my usage is going to be. Experimented with different cache settings on the disk and with everything that wasn't No Cache the write dropped.
A few details about the setup:
Cluster Network: 10.20.20.0/24 -> 10Gb
Public Network: 10.1.101.0/24 -> 1Gb
3 Monitor/VM hosts:
vmh1
vmh2
vmh3
8 Storage Nodes:
stor1
stor2
stor3
stor4
stor5
stor6
stor7
stor8
I installed the Proxmox VE on all the nodes, and ceph on all the nodes for easier management. My VMs (which I won't have that many) will all run on the vmh nodes. I noticed last night that my /etc/hosts file had the IPs of the public network for vmh1 vmh2 vmh3 and the Cluster network for the storage nodes. Not sure how proxmox compiles the hosts file, maybe someone can give me an insight on that. My thought is that maybe the Ubuntu VM was writing to the cluster through the 1Gb network instead of the 10Gb... but even then 25MB/s is only 200Mbit so any suggestions in some steps to isolate where the bottleneck is would be greatly apprciated.
I just downloaded an Ubuntu 14.04 LXC and will deploy that to do the same test there just to see if there is any difference.
Thanks in advance!
Pasha
Need some bright minds in the Proxmox community
I setup a storage cluster using Ceph (as the title suggests). It's a fairly large cluster consisting of over 200 osds. When I bench the cluster using rados bench, I get exactly the kind of performance I was expecting to get... doing dd to a single drive on the node I get on average 205MB/s .. writing to the cluster via rados I get 400-500MB/s writes need to be at least in the area of that in my VMs.
I installed Ubuntu 16.04x64 in KVM with a 4TB disk partition with ext4. Same dd test there I was able to get 25MB/s write on a 1G chunk and 15-20MB (varying) on a 2G chunk. I am more interested in the sustained write as that's what my usage is going to be. Experimented with different cache settings on the disk and with everything that wasn't No Cache the write dropped.
A few details about the setup:
Cluster Network: 10.20.20.0/24 -> 10Gb
Public Network: 10.1.101.0/24 -> 1Gb
3 Monitor/VM hosts:
vmh1
vmh2
vmh3
8 Storage Nodes:
stor1
stor2
stor3
stor4
stor5
stor6
stor7
stor8
I installed the Proxmox VE on all the nodes, and ceph on all the nodes for easier management. My VMs (which I won't have that many) will all run on the vmh nodes. I noticed last night that my /etc/hosts file had the IPs of the public network for vmh1 vmh2 vmh3 and the Cluster network for the storage nodes. Not sure how proxmox compiles the hosts file, maybe someone can give me an insight on that. My thought is that maybe the Ubuntu VM was writing to the cluster through the 1Gb network instead of the 10Gb... but even then 25MB/s is only 200Mbit so any suggestions in some steps to isolate where the bottleneck is would be greatly apprciated.
I just downloaded an Ubuntu 14.04 LXC and will deploy that to do the same test there just to see if there is any difference.
Thanks in advance!
Pasha