We're looking to migrate away from a large OnApp installation, and Proxmox is looking to be our solution. We have quite a large budget to get this done properly, so we were hoping if someone would be able to give us some best practices.
Our biggest concerns have been around Ceph within PVE. We're going to need about 60TB of usable capacity (after redundancy) within the next year so hyper-converged is not going to be likely although we may start with it on day one. Our current distributed storage platform stores 3 copies of all data. One on SSD for performance, and two for redundancy on HDDs. It has worked well and storage performance isn't really a reason we need to move away from OnApp... but we'd like to ensure we're not taking a step backwards.
Does anyone have any experience with running a large PVE cluster with an external Ceph or PVE-Ceph installation, and do you have any recommendations for it?
All our nodes will have 2x10GbE LACP for Ceph, 2x10GbE LACP for VM networking, and 2x1GbE LACP for sync and management. We typically use Juniper QFX5100s for 10GbE and EX4300s for 1GbE.
I had a crazy idea of running the PVE internal Ceph instance with all SSDs in the hypervisor (6 OSDs per hypervisor) and a few additional "storage bricks" that had only spinning disk at the bottom of the rack. We would then be able to serve the first copy of data from SSDs on the hypervisors, and then copies #2 and #3 would be on the slow storage servers at the bottom. How bad of an idea is this?
I do like the PVECeph GUI and management - is there any way to have storage only nodes within a PVECeph cluster? I'd like to have host systems that won't have a VMs assigned to it via the cluster.
If there is a better way to get answers on this via enterprise support via Proxmox, we'd be willing to go that route.
Our biggest concerns have been around Ceph within PVE. We're going to need about 60TB of usable capacity (after redundancy) within the next year so hyper-converged is not going to be likely although we may start with it on day one. Our current distributed storage platform stores 3 copies of all data. One on SSD for performance, and two for redundancy on HDDs. It has worked well and storage performance isn't really a reason we need to move away from OnApp... but we'd like to ensure we're not taking a step backwards.
Does anyone have any experience with running a large PVE cluster with an external Ceph or PVE-Ceph installation, and do you have any recommendations for it?
All our nodes will have 2x10GbE LACP for Ceph, 2x10GbE LACP for VM networking, and 2x1GbE LACP for sync and management. We typically use Juniper QFX5100s for 10GbE and EX4300s for 1GbE.
I had a crazy idea of running the PVE internal Ceph instance with all SSDs in the hypervisor (6 OSDs per hypervisor) and a few additional "storage bricks" that had only spinning disk at the bottom of the rack. We would then be able to serve the first copy of data from SSDs on the hypervisors, and then copies #2 and #3 would be on the slow storage servers at the bottom. How bad of an idea is this?
I do like the PVECeph GUI and management - is there any way to have storage only nodes within a PVECeph cluster? I'd like to have host systems that won't have a VMs assigned to it via the cluster.
If there is a better way to get answers on this via enterprise support via Proxmox, we'd be willing to go that route.
Last edited: