Hi everybody,
I am planning to implement a home-lab environment that requires fault tolerance to avoid downtime. I don't have a large budget compared to the cost of a "professional" installation for high availability.
I was able to get 3 second-hand servers SuperMicro SuperServer 5019-SL :
For the network I plan to use Mellanox infiniband equipment, which seems to have the best bandwidth/cost ratio, with this in mind I thought of two solutions :
Option 1 : Full mesh network with my 3 nodes :
Option 2 : With interconnection switch :
My first problem is that the motherboard has the following pci lines (and CPU is 16 pcie 3.0 lanes limited) :
According to my research and especially this post, the pcie 3.0 8x port will not be sufficient to guarantee the maximum bandwidth of the dual QSFP+ mellanox cards.
Q1 : Given this limitation, is one of the two envisaged architectures (mesh or switch) preferable, or does it not matter ?
Q2 : If we assume that I manage to obtain a bandwidth of 40Gbs between my 3 nodes, will the configuration envisaged allow me to have correct performance in your opinion or should I abandon my plans for HA ?
Thank you for your advice.
I am planning to implement a home-lab environment that requires fault tolerance to avoid downtime. I don't have a large budget compared to the cost of a "professional" installation for high availability.
I was able to get 3 second-hand servers SuperMicro SuperServer 5019-SL :
- PSU : 200W 80+ Gold
- MB : SuperMicro X11SSLF
- Dual I210-AT Gigabit Ethernet LAN ports
- CPU : E3-1220 v5 @ 3.00GHz
- RAM : 64 GB DDR4 @ 2133Mhz
- 3x Samsung 980 pro nvme (nvme- 1To)
- 3x Crucial CT1000P1SSD8 (nvme - 1To)
- 3x Samsung 870 EVO (sata - 500Go)
For the network I plan to use Mellanox infiniband equipment, which seems to have the best bandwidth/cost ratio, with this in mind I thought of two solutions :
Option 1 : Full mesh network with my 3 nodes :
- 3x Mellanox MCX354A-FCBT pcie card
- 6x MC2207130-001cables
Pros :
+ more cost-effective
Cons :
- not scalable architecture
Option 2 : With interconnection switch :
- 2x Mellanox SX6036
Pros :
+ the architecture is scalable and I could use dedicated storage nodes in the future and keep the current servers as computing node.
Cons :
- I have to deal with the licence to set the switch to ethernet mod
- increases the total cost
My first problem is that the motherboard has the following pci lines (and CPU is 16 pcie 3.0 lanes limited) :
- 1 PCIe 3.0 x8 (in x16 slot) => I plan to install one mellanox MCX354A-FCBT card
- 1 PCIe 3.0 x8 => I plan to install one Crucial CT1000P1SSD8 with a pcie to m.2 adapter
- 1 PCIe 3.0 x4 (in x8 slot) => I plan to install one Samsung 980 pro with a pcie to m.2 adapter
According to my research and especially this post, the pcie 3.0 8x port will not be sufficient to guarantee the maximum bandwidth of the dual QSFP+ mellanox cards.
Q1 : Given this limitation, is one of the two envisaged architectures (mesh or switch) preferable, or does it not matter ?
Q2 : If we assume that I manage to obtain a bandwidth of 40Gbs between my 3 nodes, will the configuration envisaged allow me to have correct performance in your opinion or should I abandon my plans for HA ?
Thank you for your advice.