[SOLVED] NVME/100GB Ceph network config advice needed

Jan 28, 2021
9
0
1
41
Hello,
I was looking into Proxmox set up with Ceph on my full NVME servers. At first, I was looking into 40GB but that wasn't enough with SSD.
Used following documents as guide line but wanted to get some feedback on my setup/settings (not implemented yet)

https://forum.proxmox.com/threads/proxmox-ve-ceph-benchmark-2020-09-hyper-converged-with-nvme.76516/
https://pve.proxmox.com/wiki/Full_Mesh_Network_for_Ceph_Server

Here is the config

OSD node(1-5)
ens3f0 & eno3f1 Bond0 - 192.168.1.101 to 105 (Mesh - 2 x 100GB) - Connected to other nodes directly
eno1 - 10.123.123.101 to 105 (proxmox mgmt - 1GB) - Connected to 1GB switch
ens5f0np0 & ens5f0np0 Bond1 - 10.123.122.101 to 105 (proxmox cluster - 2 x 10GB) - Connected to 10GB switch
ens1f0np0 & ens1f1np1 Bond2 - 10.123.0.0/24 (VM network - 2 x 10GB) - Connected to 10GB switch

MON node(1-3)
eno3 - 10.123.123.201 to203 (proxmox mgmt - 1GB) - Connected to 1GB switch
eno1 & eno2 - Bond0 - 10.123.122.201 to 203 (proxmox cluster - 2 x 10GB) - Connected to 10GB switch
 
Hello,
I was looking into Proxmox set up with Ceph on my full NVME servers. At first, I was looking into 40GB but that wasn't enough with SSD.
Used following documents as guide line but wanted to get some feedback on my setup/settings (not implemented yet)

https://forum.proxmox.com/threads/proxmox-ve-ceph-benchmark-2020-09-hyper-converged-with-nvme.76516/
https://pve.proxmox.com/wiki/Full_Mesh_Network_for_Ceph_Server

Here is the config

OSD node(1-5)
ens3f0 & eno3f1 Bond0 - 192.168.1.101 to 105 (Mesh - 2 x 100GB) - Connected to other nodes directly
eno1 - 10.123.123.101 to 105 (proxmox mgmt - 1GB) - Connected to 1GB switch
ens5f0np0 & ens5f0np0 Bond1 - 10.123.122.101 to 105 (proxmox cluster - 2 x 10GB) - Connected to 10GB switch
ens1f0np0 & ens1f1np1 Bond2 - 10.123.0.0/24 (VM network - 2 x 10GB) - Connected to 10GB switch

MON node(1-3)
eno3 - 10.123.123.201 to203 (proxmox mgmt - 1GB) - Connected to 1GB switch
eno1 & eno2 - Bond0 - 10.123.122.201 to 203 (proxmox cluster - 2 x 10GB) - Connected to 10GB switch
For a Mesh you need 4 Ports per node (not bonded!) for 5 nodes (with two ports bonded you need 8 ports per node)
it is not necessary to run MON's on separate nodes, load is minimal, I usually run them on every node
 
You will add more latency by daisy chaining, than by a switch

-> Switch: two hops, but modern switches can forward already when they have read the header
so in most cases latency of one hop

-> daisy chain: openvswitch with RSTP will find shortest path, in most cases 2 hops
-> as it is no dedicated switching hardware packets will probably be received completly before forwarding
-> added load to the server, as the server has to do the packet forwarding instead of a dedicated switching engine
so latency can vary widely

-> full mesh: always one hop
 
  • Like
Reactions: Sake3665

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!