Hi, I am hoping to just make a sanity check for guidance to see if I'm overlooking something. Feedback is greatly appreciated!
I'm hoping to setup a modest size cluster for a client, project is ramping up with 'fairly modest' requirements and hopes the business will grow and it needs to scale
Short term they don't want to invest in a HA multi-storage device pool (ie, 2+ bricks with dual controller iscsi san "equallogic" style gear). The storage needs are small enough that a ceph cluster does not make sense. (ie, they hope for 2 proxmox physical nodes plus 1 PBS server physical node who also acts as the cluster quorum 3rd member for 'cluster up status sanity check vote'). The actul resources needed are approx 60Gb ram for VM Guests/ 1Tb for disk storage for VMs / 24 cores of CPU (approx).
I have the impression right now my options are
(a) use ZFS replication between proxmox1<>2 nodes. This will be only async, as there appears to be no realtime CoW/Sync replication support? ie, we can only schedule (replicate data every 5? or 1? minutes, for example). (?) If there is a way to do 'realtime' sync replication with ZFS data replication between nodes and I'm just missing it somehow? please do kick me along in right direction?
(b) use ceph with a highly constrained node setup. ie, I have to push client for 3x physical proxmox nodes with sufficient disk to make it functional. Dedicated 10gig for ceph storage interconnect. Maybe we have a modest size HW Raid1 Mirror Disk for the base proxmox volume on each node, then approx 4 x SSD drives present in each proxmox node / for example 4 x 1Tb SSD. Then with ceph we might get a total of 3-4? Tb usable space (2N+1 copies of data spread over 3 nodes) for the ceph HA storage pool. (I generally have the feeling that Ceph wants to run - on 'bigger more serious' deployments, ie, 6-10+ nodes, where you want multi-multi TBS of storage, have lots and lots of VMs needed, etc, ie, not such a great fit for "yeah, we have maybe 8-ish VMs and ~=<1Tb of storage footprint needed maybe kinda". But maybe I'm not being proper-fair in my assessment?).
(c) use DRBD is not really option I want to explore; I tested this a version or two back on proxmox and general feeling I had was that - it is kind of painful to get it working reliably and is a bit more fragile than I like.
(D) just forget about the true HA storage, go with a 'robust modest decently fault tolerant' QNAP NFS Filer as shared storage target (redundant power, disks, but non-redundant controller basically) - so we have single point of failure here on the qnap, which is not great, but generally these are solid and run well, so outages will be few. Regular backups on PBS cluster mean worst-case we can do a bare-VM-restore from PBS into a local storage in case we have a fail on the backing QNAP storage device and we want to get production VMs back online 'quickly, with human intervention, but clearly not an HA-style auto-recovery' sort of thing.
(e) use an external ZFS HA storage solution of some kind, let it deal with the HA storage and let proxmox just use that as the 'trusted storage tank'. I recently was given a heads up about this ZFS stack: https://haossoft.com/requirements/ - appears to be pure ZFS standard-based under the hood / with the team on this project putting together the 'glue' to 'make it work'. But they clearly say, "hey, this is not production ready, we are testing / asking for feedback'. So I am not positive I should 'test, deploy' in production.
Anyhoo. Figured I would put this out there in case anyone wanted to give me a kick in the right direction , or remind me of obvious things I am overlooking.
Many thanks!
Tim
I'm hoping to setup a modest size cluster for a client, project is ramping up with 'fairly modest' requirements and hopes the business will grow and it needs to scale
Short term they don't want to invest in a HA multi-storage device pool (ie, 2+ bricks with dual controller iscsi san "equallogic" style gear). The storage needs are small enough that a ceph cluster does not make sense. (ie, they hope for 2 proxmox physical nodes plus 1 PBS server physical node who also acts as the cluster quorum 3rd member for 'cluster up status sanity check vote'). The actul resources needed are approx 60Gb ram for VM Guests/ 1Tb for disk storage for VMs / 24 cores of CPU (approx).
I have the impression right now my options are
(a) use ZFS replication between proxmox1<>2 nodes. This will be only async, as there appears to be no realtime CoW/Sync replication support? ie, we can only schedule (replicate data every 5? or 1? minutes, for example). (?) If there is a way to do 'realtime' sync replication with ZFS data replication between nodes and I'm just missing it somehow? please do kick me along in right direction?
(b) use ceph with a highly constrained node setup. ie, I have to push client for 3x physical proxmox nodes with sufficient disk to make it functional. Dedicated 10gig for ceph storage interconnect. Maybe we have a modest size HW Raid1 Mirror Disk for the base proxmox volume on each node, then approx 4 x SSD drives present in each proxmox node / for example 4 x 1Tb SSD. Then with ceph we might get a total of 3-4? Tb usable space (2N+1 copies of data spread over 3 nodes) for the ceph HA storage pool. (I generally have the feeling that Ceph wants to run - on 'bigger more serious' deployments, ie, 6-10+ nodes, where you want multi-multi TBS of storage, have lots and lots of VMs needed, etc, ie, not such a great fit for "yeah, we have maybe 8-ish VMs and ~=<1Tb of storage footprint needed maybe kinda". But maybe I'm not being proper-fair in my assessment?).
(c) use DRBD is not really option I want to explore; I tested this a version or two back on proxmox and general feeling I had was that - it is kind of painful to get it working reliably and is a bit more fragile than I like.
(D) just forget about the true HA storage, go with a 'robust modest decently fault tolerant' QNAP NFS Filer as shared storage target (redundant power, disks, but non-redundant controller basically) - so we have single point of failure here on the qnap, which is not great, but generally these are solid and run well, so outages will be few. Regular backups on PBS cluster mean worst-case we can do a bare-VM-restore from PBS into a local storage in case we have a fail on the backing QNAP storage device and we want to get production VMs back online 'quickly, with human intervention, but clearly not an HA-style auto-recovery' sort of thing.
(e) use an external ZFS HA storage solution of some kind, let it deal with the HA storage and let proxmox just use that as the 'trusted storage tank'. I recently was given a heads up about this ZFS stack: https://haossoft.com/requirements/ - appears to be pure ZFS standard-based under the hood / with the team on this project putting together the 'glue' to 'make it work'. But they clearly say, "hey, this is not production ready, we are testing / asking for feedback'. So I am not positive I should 'test, deploy' in production.
Anyhoo. Figured I would put this out there in case anyone wanted to give me a kick in the right direction , or remind me of obvious things I am overlooking.
Many thanks!
Tim