Hello,
We have three dedicated OVHCloud SCALE-3 servers. The first in the Roubaix (France) data center (RBX8), the other two in the Strasbourg (France also) data center (SBG3).
We have taken the private 12 Gbps in / out bandwidth option on these three SCALE-3 dedicated servers.
Reference: https://www.ovhcloud.com/en/bare-metal/scale/scale-3/
The GNU / Linux distribution Proxmox 7.0 is installed on all servers (but same problem with Proxmox 6).
All three servers are in the same vRack (reference: https://us.ovhcloud.com/products/networking/vrack-private-network).
The problem is: we are unable to achieve effective 12Gbps private bandwidth between our various servers.
After many discussions with OVH support and numerous interventions by their teams on site, it turns out that the problem is not hardware and the on-site network configuration is correct.
The problem would therefore be software side, more precisely a network configuration fault in our Proxmox system.
OVH support managed to achieve the right speed (12Gbps incoming / outgoing) via their rescue mode (which is a Debian 8 base…) using their parameters (which were not communicated to us in detail…). The test they used to confim it is a simple:
iperf3 -c 192.168.1.10 -i 10 -t 100 -P 16 | grep "SUM"
Here is the solution offered by OVH support:
“Set the MTU to 9000 and optimize the TCP Windows Size according to latency and bandwidth.
You need to know the latency between your servers on average.
Here is an example calculation for a bandwidth of 10 Gbps and a latency of 80.1 ms:
- Binding bandwidth in bits * latency = window size / 8 = perfect size in bytes.
- 10737418240 bits * 0.0801 = 860067201/8 = 107508400 bytes. (102MBytes)
Under Linux, you will need to modify the "TCP buffer" settings via "sysctl tcp_rmem, tcp_wmem and tcp_mem".
/! \ I invite you to make a backup of your data before any manipulation for security reasons.
After modifying the MTU as well as the TCP buffer, your bandwidth should be much greater. "
We have done several tests by changing these values as indicated, indeed it modulates the bandwidth between the servers well, but still cannot stably reach the bandwidth of 12Gbps guaranteed by our option on the SCALE-3.
Here are the details of our tests (as a reminder the 3 SCALE-3 servers are in the same vRack we have subscribed to the 12Gbps option on the three):
Between the SCALE-3 RBX8 and the SCALE-3 SBG3 we have a very poor throughput of 2Gbps incoming / outgoing. All our tests to modulate the bandwidth have failed (it remains below 12Gbps and is unstable).
Between the two SCALE-3 SBG3s we have a standard 6Gbps speed (so still not the 12Gbps “guaranteed” by our option) BUT with the modifications made by OVH a theoretical bandwidth of 12Gbps could be reached (via iperf3 test).
I say theoretical because in practice the transfers of files and VMs within our Proxmox remain slow (do not use 12Gbps).
For the SCALE-3 RBX8, the only time it reaches 12Gbps is on the OVH support side, in rescue mode, with their parameter and their iperf3 theoretical test. In practice on our Proxmox, despite the various tests, the bandwidth remains catastrophic (even the standard 6Gbps is not reached, we remain at mediocre 2.2Gbps).
We are starting to lose our minds a bit, so if someone has faced this scenario before (not specificaly on OVHCloud dedicated servers):
Would it be possible to know the exact network configuration to apply to reach the private 12Gbps bandwidth between several dedicated servers using Proxmox (so based on Debian Linux) in the same vRack?
Thank you.
We have three dedicated OVHCloud SCALE-3 servers. The first in the Roubaix (France) data center (RBX8), the other two in the Strasbourg (France also) data center (SBG3).
We have taken the private 12 Gbps in / out bandwidth option on these three SCALE-3 dedicated servers.
Reference: https://www.ovhcloud.com/en/bare-metal/scale/scale-3/
The GNU / Linux distribution Proxmox 7.0 is installed on all servers (but same problem with Proxmox 6).
All three servers are in the same vRack (reference: https://us.ovhcloud.com/products/networking/vrack-private-network).
The problem is: we are unable to achieve effective 12Gbps private bandwidth between our various servers.
After many discussions with OVH support and numerous interventions by their teams on site, it turns out that the problem is not hardware and the on-site network configuration is correct.
The problem would therefore be software side, more precisely a network configuration fault in our Proxmox system.
OVH support managed to achieve the right speed (12Gbps incoming / outgoing) via their rescue mode (which is a Debian 8 base…) using their parameters (which were not communicated to us in detail…). The test they used to confim it is a simple:
iperf3 -c 192.168.1.10 -i 10 -t 100 -P 16 | grep "SUM"
Here is the solution offered by OVH support:
“Set the MTU to 9000 and optimize the TCP Windows Size according to latency and bandwidth.
You need to know the latency between your servers on average.
Here is an example calculation for a bandwidth of 10 Gbps and a latency of 80.1 ms:
- Binding bandwidth in bits * latency = window size / 8 = perfect size in bytes.
- 10737418240 bits * 0.0801 = 860067201/8 = 107508400 bytes. (102MBytes)
Under Linux, you will need to modify the "TCP buffer" settings via "sysctl tcp_rmem, tcp_wmem and tcp_mem".
/! \ I invite you to make a backup of your data before any manipulation for security reasons.
After modifying the MTU as well as the TCP buffer, your bandwidth should be much greater. "
We have done several tests by changing these values as indicated, indeed it modulates the bandwidth between the servers well, but still cannot stably reach the bandwidth of 12Gbps guaranteed by our option on the SCALE-3.
Here are the details of our tests (as a reminder the 3 SCALE-3 servers are in the same vRack we have subscribed to the 12Gbps option on the three):
Between the SCALE-3 RBX8 and the SCALE-3 SBG3 we have a very poor throughput of 2Gbps incoming / outgoing. All our tests to modulate the bandwidth have failed (it remains below 12Gbps and is unstable).
Between the two SCALE-3 SBG3s we have a standard 6Gbps speed (so still not the 12Gbps “guaranteed” by our option) BUT with the modifications made by OVH a theoretical bandwidth of 12Gbps could be reached (via iperf3 test).
I say theoretical because in practice the transfers of files and VMs within our Proxmox remain slow (do not use 12Gbps).
For the SCALE-3 RBX8, the only time it reaches 12Gbps is on the OVH support side, in rescue mode, with their parameter and their iperf3 theoretical test. In practice on our Proxmox, despite the various tests, the bandwidth remains catastrophic (even the standard 6Gbps is not reached, we remain at mediocre 2.2Gbps).
We are starting to lose our minds a bit, so if someone has faced this scenario before (not specificaly on OVHCloud dedicated servers):
Would it be possible to know the exact network configuration to apply to reach the private 12Gbps bandwidth between several dedicated servers using Proxmox (so based on Debian Linux) in the same vRack?
Thank you.
Last edited: