hmm , you could pack drdb underneath for replication, but it has its own pitfalls.
can you spread out the disks equally to the nodes? Then CEPH would be an option.
What is the target for backup?`I had some troubles with NFS as target, the NFS Client in the Linux OS tends to eat up resources.
It looks like CIFS works more reliable in this case.
yes, use the other 1 GBit for corosync and/or management, don't use it for migration traffic (corosync likes low latency)
ahh, one more point: Does your RAID Controller support real HBA Mode? I hope its not one of the MegaRaid things, which force you to use a Raid0 for a single disk.
HBA mode...
> All the VMs will be public facing we dont need local networking inside the VM'S. So i dont see the need for a 10GBit for public wan traffic since for > this cluster i am only giving a wan 100mbit uplink that will be a 1gbit when i fill the 100mb up.
ok, so 1 GBit for outside net is fine.
>...
Sorry, hit the post button before editing was complete:
-> in the minimum configuration use 2 VLAN's for CEPH ( public + private), so you can segregate the traffic physically later easily with out hard reconfiguration
the part with the minimum config is doubled, as I wanted to move it up...
Sounds good, but the following recommendations:
In a ideal world you would:
-> Segregate CEPH Traffic physically
-> Segregate corosync Traffic physically (1 GBit is ok)
-> If possible segregate also Traffic for other storage protocols (NFS/CIFS) from other traffic physically
-> If possible...
Hmm, I didn't tried it this way until now, I always had all the VLAN's available over the locations (with very low latencies), so it was like a reboot (just had to set noout for CEPH OSD's to avoid the CEPH rebalancing).
But as long as the Cluster IP's do not change it sounds that your proposal...
Did you have the same IP's on the new location and are the two locations connected with not too much latency?
Then its easy, just move node for node without changing anything.
If IP's will change, and/or latency is too large for proper cluster operation, then its more complicated, as corosync...
you have 1 OSD per Node, probably just HDD? And 3 nodes, so just 3 OSD's?
Which network speed?
CEPH needs both low latency, and at best many OSD's. The slow filesystem would be no surprise.
Replace the HDD's with Enterprise SSD's and use more OSD's per Node, and at least 10 GBit in the backend...
Hi,
we see very often the problem, that pvestatd erroneously says the a CIFS Storage (we use it for Backup and ISO) is offline:
root@gar-ha-cfw01a:/usr/share/perl5/PVE/Storage# systemctl status pvestatd
● pvestatd.service - PVE Status Daemon
Loaded: loaded...
LACP works well, of course the switches have to be configured appropriately.
We run two clusters with LACP without any quirk.
(what is not working: any bonding types other than 1 or 4 -> depending on switches you can run into any network hell you can think of)
Some more points:
At least if...
With CEPH objects will be stored 3 times (Erasure Coding is currently not supported with proxmox as I remember)
Also you should not fill up pools with more than 60 percent, so you will have for images:
(12 TByte / 3 ) * 0.6 = 2.4 Tbyte SSD Storage
(48 TByte / 3 ) * 0.6 = 9.6 Tbyte SATA Storage...
Ceph is very latency dependent. I would avoid a S-ATA only Pool, it is badly slow. Plan at least SSD journals for the S-ATA OSD's
It is possible to create separate pools, but you have to edit the Crushmap per Hand.
Both solutions are ok, but installing proxmox on the storage nodes would give you a nice simple management and installation for CEPH as well as a common GUI. If you put the storage nodes also in the cluster you earn:
-> common GUI
-> possibility to use the storage nodes for some VM's with...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.