Are there any Hints, Tips, Nice to know's, or Gotcha's one should know before planning, sizing and purchasing the cluster hardware ?
Things that i am curious about (Proxmox 6.2):
Background:
I am creating 2 production clusters at two different locations (3+ Nodes per cluster), seperate networks (openvswitch) for clients (4x 1G or 2x10G), corosync (2x1G) and 2x 10G or 1x 40G for storage/backup traffic. I might eventually use the Proxmox Backup Server for remote syncs (if it turns out to be working stable enough for my needs (bandwith [up] is an issue at both locations))
History:
- I have used Proxmox (clusters) extensively until 5.x with Ceph (standalone) as a storage backend . Currently unsure wether to use ZFS or Ceph as a storage backend (my experience is with 40+ HDD's backed by SSDs for journals per node, not small amount of SSD based OSDs)
- Only have used proxmox 6 and 6.2 on standalone servers (hetzner/OVH) since then for non-production projects.
- i just did a deepdive through the wiki as a refresher.
Things that i am curious about (Proxmox 6.2):
- Can you mix and match CPU-vendor and Models inside the same cluster? (e.g. 2x Intel Cascade Lake-SP + 2x Amd Epic 7000 + 1x Threadripper 3000) If i remember correctly, last time i sized one (4+ years ago) this was not recommended.
- Are there any issues with specific 1G and 10G Nic-models ? I used to have a couple Hetzner Servers that had massive issues with a specific Intel network card model. Is there a list of known to (not) be working models ?
- Can proxmox utilize NVME SSD's (m.2/pcie) with ZFS or Ceph properly ? (last time i tried to install proxmox on ZFS raid1 it wasn't supported), considering the following option(s):
- Sata 6 GB/s (or M.2 via 2.5" Sata Adapter) on controller in HBA mode
- SAS 12 GB/s based SSDs using a SAS-Raid controller in HBA mode.
- M.2 NVME via Add-in Card (e.g. Asus Hyper)
- 2.5" + NVME backplane combo from a OEM.
- Is there any noticeable speed difference by using Ext4 Raid1 vs ZFS Raid1 for the Host-OS, or does this not matter anymore ?
- Any specific Server vendors/models (dell, hp, fujitzu, terra, ...) that are known to be "iffy" when using proxmox on them (driver-support) ?
- is there a list of mainboards that work well (or is that not an issue anymore ?) I found this list (not updated since 2016?) https://pve.proxmox.com/wiki/Mainboards. Not sure if doing a self-build or going with a vendor like e.g. dell.
- How many CPU-Cycles do you typically need for CEPH when running CEPH and proxmox 6.2 on the same node in order not to bottleneck it ? (a formula would also suffice (x Cpu cycles per OSD))
- Not sure yet if i want to use Ceph or ZFS for performance (12 TB total Storage capacity requiered - would be at least 3x12 OSDs on Ceph). HA it not really requiered, but would be nice to have.
Background:
I am creating 2 production clusters at two different locations (3+ Nodes per cluster), seperate networks (openvswitch) for clients (4x 1G or 2x10G), corosync (2x1G) and 2x 10G or 1x 40G for storage/backup traffic. I might eventually use the Proxmox Backup Server for remote syncs (if it turns out to be working stable enough for my needs (bandwith [up] is an issue at both locations))
History:
- I have used Proxmox (clusters) extensively until 5.x with Ceph (standalone) as a storage backend . Currently unsure wether to use ZFS or Ceph as a storage backend (my experience is with 40+ HDD's backed by SSDs for journals per node, not small amount of SSD based OSDs)
- Only have used proxmox 6 and 6.2 on standalone servers (hetzner/OVH) since then for non-production projects.
- i just did a deepdive through the wiki as a refresher.