Which log would tell you that a monitor is blocking the communication from one client node to the ceph cluster? What's the best way to see whether the ceph monitors are blocking these two nodes that can't communicate? Again, there are no firewalls between nodes unless Proxmox inserted...
netcat is timing out trying to connect to the monitor node but .. pinging to that same address works fine ... there are no firewalls in between any nodes. The Ceph network has it's own switch and all nodes talk over that switch for Ceph communications
As a side note, we switched all Ceph networking over from 10Gb copper to 25Gb fiber .. this was done on July 5th, 2021 .. it switched over smoothly ..
Has anyone had issues using fiber? Has anyone had issues with Mellanox fiber NICs and a Dell 48port 25Gb switch?
So, the other nodes do just fine .. again, they mention the ceph.conf parsing error but then list the virtual disks available on that ceph pool
For your example of connecting using netcat, is there something special to do with IPv6 addresses? This ceph cluster is IPv6 only.
This is the next error came up
2021-07-22 09:24:10.478 7f346cfc00c0 0 monclient(hunting): authenticate timed out after 300
This makes no sense to me though .. I provided the same IPs as the other nodes that ARE connecting use and I also provided the same keyring the other clients use
version is the same
ceph version 14.2.22 (877fa256043e4743620f4677e72dee5e738d1226) nautilus (stable)
Trying the rbd -p command now ... so far it's just hanging ... only errors so far are about parsing config file which is the same on all client nodes because there is no Ceph config file on...
Please bear in mind what I mentioned in my first post. All machines were running good. All machines (12) have been updated with the latest Proxmox software. Only two have been rebooted and now the two that have been rebooted no longer can talk to Ceph. The only thing that has changed is the...
We are having more or less this same problem. Just updated all nodes to latest Ceph 14.2.22 with the latest kernel. Already restarted services on Ceph nodes (5 nodes are Ceph, 7 nodes are only for running VMs with Ceph client) After restarting 2 of the 7 nodes that only run the latest Ceph...
That's kind of you to be helpful and offer another solution, thank you...the problem is that backups are an integral part of Proxmox since its inception and frankly the problems with the speed need to be addressed within Proxmox itself instead of throwing the problems off to a third party script...
I can say for us it's identical: As long as we've been using Proxmox with Ceph the backup speeds are between 30 - 50MB per second ... no change
We keep up with all updates
I am subscribed to the pve-devel list and haven't seen any work being done on this so it makes sense that the backup...
Since Ceph Luminous was integrated with Proxmox 5.0 can we assume that the jump from Proxmox 4.4 to Proxmox 5.1 will be as safe and simple as from 4.4 to 5.0? We have a 10 node cluster with 5 nodes being Ceph Jewel and were simply waiting for Luminous 12.2 which the Ceph Team labels as the...
No, that's definitely not the answer that we're looking for. If you read the rest of the thread you'll see that was already mentioned and the response to it. Proxmox already does back ups, it's a matter of fixing the issues with backing up Ceph images within Proxmox so that it's done at full speed.
@fips
What kind of NAS are you using? I only get about a third of what you are getting to your NFS (you showed 128 MB/s)
All those throughputs have to be on 10Gb ethernet but is the NAS a generic piece of hardware like a Dell with standards Linux on it or is it a Proprietary NAS solution...
Fabian, thank you. I apologize for the "poking" as you put it. This is the first answer I've seen that nicely describes the difficulty you guys are having with getting this issue resolved. If my words here seemed excessive or lacking in tact at times, again I apologize. I am a back office admin...
I want to make something clear here ... before we were users of Proxmox, we were users of Citrix XenServer. Why did we get off of XenServer? A couple of reasons ... XenServer had no built-in way to create VM backups and XenServer performed very poorly on disk IO. As you can see, our number 1...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.