its an innocent enough question- but there are a lot of gotchas you need to consider.
Clusters are made up of 3 elements- compute, storage, and networking. lets touch on each.
COMPUTE:
- Dell R630 is a 10 year old platform. as such, it offers pretty poor performance/watt. Do you already know where you are deploying this solution? how much power and cooling are provided?
- Ignore the VM count for a moment. How much cpu load will the TOTAL cluster workload be in terms of core-ghz? you need to account for typical and max, and excess capacity for failover.
- Add a core and 4G of ram per every OSD since it appears you intend to use this as an HCI.
- once you add it all up, you'll have an idea of how many servers you will be deploying.
STORAGE:
- What is the minimum required usable capacity? you should be prepared for 4x RAW. a smallish number of high capacity OSDs could work, but you gain better performance with higher OSD count.
- Dell R630s support up to 4 NVMEs but only on the 10 drive models. Since you cant get these new, be aware that most 10 drive models you will find in the wild dont actually have NVME support and you will have to buy and install it separately.
NETWORKING:
- ideally you need separate interfaces for ceph, ceph private, cluster, and service networks. BMC too. Be aware that R630 is a PCIe gen3 platform which means your maximum practical link speed is 100gbe, and some of your pci lanes will be consumed by your nvmes (16 lanes total) so 4x25g is a good practical configuration for this generation of hardware. you'll want to get your port count based on your node count, and then provision 2 switches that can accommodate half each.
There is a lot more to consider, but this should give you a starting point.
Thank you for the advise.
COMPUTE:
- Dell R630 is a 10 year old platform. as such, it offers pretty poor performance/watt. Do you already know where you are deploying this solution? how much power and cooling are provided?
Answer : All the nodes will be put in 52u Racks in the tier-3 datacenter with a 5 - 10kw power.
- Ignore the VM count for a moment. How much cpu load will the TOTAL cluster workload be in terms of core-ghz? you need to account for typical and max, and excess capacity for failover.
Answer: I planned to use 2 x Intel Xeon E5 v4 2699
- Add a core and 4G of ram per every OSD since it appears you intend to use this as an HCI.
ANswer: This will be 384GB to 512GB RAM on each node
- once you add it all up, you'll have an idea of how many servers you will be deploying.
STORAGE:
- What is the minimum required usable capacity? you should be prepared for 4x RAW. a smallish number of high capacity OSDs could work, but you gain better performance with higher OSD count.
Answer: Yes, this will be u.2 Nvme but the storage will be 1.92Tb to 15.85TB NVMe u.2. But I dont have any idea on how many units required
- Dell R630s support up to 4 NVMEs but only on the 10 drive models. Since you cant get these new, be aware that most 10 drive models you will find in the wild dont actually have NVME support and you will have to buy and install it separately.
Answer : Yes, I aware of this. for the 10 bays it will be use for up 4 x NVMe U.2 with the card riser and extender.
NETWORKING:
- ideally you need separate interfaces for ceph, ceph private, cluster, and service networks. BMC too. Be aware that R630 is a PCIe gen3 platform which means your maximum practical link speed is 100gbe, and some of your pci lanes will be consumed by your nvmes (16 lanes total) so 4x25g is a good practical configuration for this generation of hardware. you'll want to get your port count based on your node count, and then provision 2 switches that can accommodate half each.
Answer : Yes, 10Gg, 25Gb or 40Gb is the plans. Cluster will be 1GB. any related to ceph will be use 10,25 or 40GB.
The thing is, I am not sure either to use a Pure HCI meanse compute and ceph will be in a same server or separate the Nodes and Ceph( means 7 Compute Nodes + 3 Ceph Nodes.
What is your opinion?