After a year of using freeNAS with LVM over iSCSI with 30 nodes and almost 100 VMs, that solution finally revealed its shortcomings. As a stopgap to keep the VMs running, I moved the VM images back onto each node's local hard drives. I'm now looking to setup a three node ceph cluster, running on the three most powerful nodes I have. I am using PVE5 with the luminous test repository.
I have created two of the nodes, each with four 2-TB hard drives. I created four montors and created the 8 osds. For the OSDs I put the journals on partitions in a separate SSD. I then created a pool with a size/min of 2/2 with 512 pgs. The cluster shows as "HEALTH_OK".
I need to start using this storage so I can move the VM images from the third ceph node so I can bring it down to configure the hardware and add it to the cluster. Even after reading the documentation, I'm not sure of the following things:
I have created two of the nodes, each with four 2-TB hard drives. I created four montors and created the 8 osds. For the OSDs I put the journals on partitions in a separate SSD. I then created a pool with a size/min of 2/2 with 512 pgs. The cluster shows as "HEALTH_OK".
I need to start using this storage so I can move the VM images from the third ceph node so I can bring it down to configure the hardware and add it to the cluster. Even after reading the documentation, I'm not sure of the following things:
- Should I have created the pool with szie/min of 3/2, looking forward to the full number of OSDs (the PGCalc says the it would still be 512pgs)?
- What is the danger of data unavailability or loss of the two-node cluster if one of those nodes becomes inaccessible for any reason (network, reboot, software failure, etc) while I'm getting the third node online?
- When the third node is ready, how do I add the OSDs to the existing pool? I'm guessing that would involve bumping the size from 2 to 3.
- Once the third node is online and the cluster is healthy, I will need to take down the original two nodes one at a time to reconfigure the SSD drive where the journals are kept.
- I haven't found any good advice as to what size the journal partitions should be. It looks like Proxmox is creating those partitions. Any thoughts on moving the OS from the current separate HD to the same SSD used for the journals (in order to free up another disk slot in the chassis)?