Correct. You do not need to route any traffic between 2 switches. VMs traffic will be restricted to one switch only.
This will really come down to the amount of VMs you have and number of people your cluster serves. While a 10gb network will certainly increase Ceph cluster bandwidth, if your VM cluster is small you may be ok to just use single network for both Proxmox and Ceph traffic. Or use 10gb network for Ceph and use existing 1gb network for Proxmox VMs. I prefer to separate networks. As mir suggested, you can also go with network bonding way.
Each port does have its own dedicated path. What you have think is how much traffic each port going to handle. If you are putting Proxmox and Ceph both on lets say 1gb port, then it is going to get consumed by both traffic on first come first basis. Meaning if Ceph using up 700mbps that will leave with 300mbps for Proxmox traffic. If the Proxmox traffic demands 500mbps, there are none left. Same goes for 10gb traffic. During Ceph self-healing it is possible that almost all 10gb bandwidth may be consumed.
This is from my personal experience. So dont take it as etched on stone. Ceph do not need beefed up node. I run my ceph cluster with dual core i3 and 16gb ram. Even on a bad day when cluster rebalances, it performs just fine. I could possibly get away with 8gb RAM, but sometimes i move VMs to Ceph nodes temporarily so additional 8GB helps.
I have the following specs per Ceph nodes:
Motherboard: Intel Server SBL1200BTLR
CPU: i3-3220
RAM: 16GB
NIC: Intel Pro 1gb
RAID: Intel RS2WC040
Expander Card: Intel 24port Expander
OSD: 10 SATA Segate Baracuda 2TB
Total Ceph nodes : 3
Total OSDs : 30
Replica : 3
Total PG : 2432[/QUOTE]
Thank you!!! You and Mir are totally awesome. Man, I was in the dark about this whole ordeal but begin to see the bright lights and am going to config all afternoon long.
So regardless, in my situation with 5 nodes it's better to run two separate 24 port switches rather than just one 48 ports switch. From the way you described, Even if I separate the ceph nodes and created bonding, it is better to be on a separate switch.
The VMs itself doesn't need bonding as 1GB connection is adequate for medium usages (4-6VMs medium to low traffic), correct? I have some heavy windows users. Should I bond the nics on those VM nodes?
I notice your Total PG is 2432. But according to the calculation, it supposed to be 1024 (30 x 100 / 3 = 1000 -> 1024). Even next power up to 2 would be 2048. But you set it at 2432? Is there a reason for that?
I always thought more PGs is better as you plan for growth. But in my previous post, it was suggested by another person in the forum to increase only as needed.
What chasis do you use that can house a 10 drives? Is it a 4U?
Your 2 cents are always utmost appreciated. Thank you.