I have decades of experience with Hpe Proliant and Dell PowerEdge and each of their hardware RAID controllers on the server, but very little knowledge with Ceph or ZFS etc.
With VMware's recent BS I am looking to switch all clients I have VMware systems installed at to Proxmox. I've created a little 3 note test system and followed some online videos etc to create the Ceph/OSD disks and monitors etc and it all went ok. Was able to create Vm's enabled for HA and did some simulations to live move running Vm's and it did that without issue or the vm needing rebooting. I failed the PVE node and also the OSD disk and the VM did come back up after a few minutes on another PVE node. I'm curious how to handle a single hard disk failure without the whole node having to move the VM to another PVE node. When creating the OSD you had to pick a single 'disk' and you get the warning they don't support hardware RAID disks. So how can we create redundant storage on each PVE node for Vm's to handle simple disk in the pool failure without having to fail it to another PVE node / cause a reboot? I created a ZFS drive using 2 drives as a simple mirror and created Vm's on that and simulated a failed drive and it seem to keep running... But couldn't find a way to create the "HA" aspect for the VM that it could be booted on another PVE node if the whole node failed.
Most clients are single VMware hosts with direct attached storage on hardware RAID controller but a few have SAN's shared storage with multiple hosts and the Vm's are setup in HA so that if a host goes down they do keep running on the other host without even needing reboot. (Hpe MSA2000 array with 12Gbps redundant SAS connections to each Proliant G10 host) Can Proxmox leverage shared SAN storage for HA?
Did a bunch of searching but not really finding what I am looking for. What is the best practice for multi node setup with local attached storage that can be redundant to survive disk failures without having to move everything to another node?
With VMware's recent BS I am looking to switch all clients I have VMware systems installed at to Proxmox. I've created a little 3 note test system and followed some online videos etc to create the Ceph/OSD disks and monitors etc and it all went ok. Was able to create Vm's enabled for HA and did some simulations to live move running Vm's and it did that without issue or the vm needing rebooting. I failed the PVE node and also the OSD disk and the VM did come back up after a few minutes on another PVE node. I'm curious how to handle a single hard disk failure without the whole node having to move the VM to another PVE node. When creating the OSD you had to pick a single 'disk' and you get the warning they don't support hardware RAID disks. So how can we create redundant storage on each PVE node for Vm's to handle simple disk in the pool failure without having to fail it to another PVE node / cause a reboot? I created a ZFS drive using 2 drives as a simple mirror and created Vm's on that and simulated a failed drive and it seem to keep running... But couldn't find a way to create the "HA" aspect for the VM that it could be booted on another PVE node if the whole node failed.
Most clients are single VMware hosts with direct attached storage on hardware RAID controller but a few have SAN's shared storage with multiple hosts and the Vm's are setup in HA so that if a host goes down they do keep running on the other host without even needing reboot. (Hpe MSA2000 array with 12Gbps redundant SAS connections to each Proliant G10 host) Can Proxmox leverage shared SAN storage for HA?
Did a bunch of searching but not really finding what I am looking for. What is the best practice for multi node setup with local attached storage that can be redundant to survive disk failures without having to move everything to another node?