Of cause it is easier concerning scalability to use switches in the corosync link. (In all links)
But in the end with a mesh network there will be always a scalability problem. If you want to add nodes you will also need a storrage connection..
It is just a question of what you have planned. If...
Where is the recomandation to seperate ceph cluster and ceph public network? Proxmox says, "as an optional step you can go even further and separate the OSD replication & heartbeat traffic as well. This will relieve the public network and could lead to significant performance improvements...
This is, what the Wiki says (https://pve.proxmox.com/wiki/Deploy_Hyper-Converged_Ceph_Cluster):
Public Network: You should setup a dedicated network for Ceph, this setting is required. Separating your Ceph traffic is highly recommended, because it could lead to troubles with other latency...
This is the structure proxmox recomands:
The most essential part in a Proxmox HA Cluster is the Corosync link. Corosync should have minimum 1 (better 2) seperate physical link (not shared with any other traffic) because of latency. Corosync links do not need a high bandwidth connection, but a...
Time syncronization works with ntp
https://pve.proxmox.com/wiki/Time_Synchronization
Maybe this is helpful,otherwise you should search in the forum concerning time and ntp.
Could it be, that the Port 123 is blocked via a firewall?
Can you please check what your configuration concerning osd_pool_default_min_size and osd_pool_default_size is?
You can find this under Node > Ceph > Configuration > and there under [global]
I do not really understand why there is a degreded pg, when there is just one osd missing. Normally it...
There should be one. It is the one which is down and out.
Look at the OSD tabs on the nodes. There should be one with the status down and out. You should try to bring it up and in again.
Clock skew, as written comes from an problem with the time syncronization. Mayby looking in the forum after...
I do not know, if this is in any relation. Does this happens long time after your problems started, while ceph was recovering?
I had written a comment in another thread some days ago, where I found somthing concerning the nf_conntrack:
Check nf_conntrack: This connection tracking and limiting...
You you could try to to restart the proxmox cluster service with
systemctl restart corosync
node by node starting with that one wich is now seperated.
Have you checked that Date and time is synchronized?
Can you post your network tab of one node?
What is in your /etc/pve/corosync.conf?
Or posting the output of:
cat /etc/pve/corosync.conf
and
cat /etc/network/interfaces
Does ceph report a health warning?
Are this really seperate network cards (6), or cards with more ports? E.g. 3 Network Cards with 2 Ports for the 6 connections?
Here you should also mix the connections on different cards. E.g. If you have one network card with 2 Ports linked to 2 switches with lacp (e.g. for storage traffic) you...
Yes, storage frontend and backend from ceph is in this version on one link. Seperation of this two makes sense, but is not as essential as seperating corosync. If ceph starts to rebalance the pg's it might have an impact on performance on the frontend (to the vm's).
For this case you can set...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.