You can always increase the numbers but not decrease so only way to fix this is create a new storage pool and migrate the VMs and containers to it. Then destroy the old pool to reclaim the disk space. As always make a backup of your VMs and containers on a separate disk first before doing...
I just noticed in the ceph.conf that it was missing pve7 server to be listed as one of the 4 monitors even though I do see it as being active in the WebGUI. Don't know if adding this would have made a difference.
[mon.pve7]
host = pve7
mon_addr = 10.50.10.242:6789,10.40.10.242:6789
I had a scare with my cluster when I was doing firmware updates on one of my Dell PowerEdge R7415 servers which took 30 mins to complete. I had expected the Ceph cluster to recover on it's own but it caused some VMs to stall and two nodes became inaccessible. Once the downed node was back up...
I am having the same issue. I have 4 nodes running PVE 6.0-7 and once in awhile the server would freeze and then reboot on it's own. Looking through the log I see the same errors as you posted above. It's a Dell PowerEdge R7415 with non-raid Dell HBA330 Mini (Embedded) Firmware 16.17.00.03...
Another thing you may want to is remove the machine-id in /etc/machine-id so it'll be unique in each clone generation.
cat /etc/machine-id
rm /etc/machine-id
touch /etc/machine-id
I don't know if cloud-init does that automatically but least doing this wouldn't hurt before converting the image...
Whenever I upgrade the PVE nodes I manually move the live VMs onto another node and then upgrade the empty node. I know some just upgrade and reboot to let HA handle the migrations but I only let HA handle it if the node actually failed unexpectedly. This way I make sure the migrations are...
I can tell you both PVE 5.4 and 6.0-7 are running great with CPU(s) 64 x AMD EPYC 7551P 32-Core Processor (1 Socket). Couple of my servers have 512GB of ram and smaller RAM in other servers with total 4 in a cluster.
Figured out the problems and mostly my fault for not paying close attention to certain issues after the Proxmox upgrade from version 5.4 to 6. Even though my test environment upgraded without issues but it didn't have the 10 gig network cards which is what being used for Ceph-Cluster. I...
I've upgraded my test environment which worked without issues. Then upgraded the production environment and getting this error
auth: unable to find a keyring on /etc/pve/priv/ceph.mon.pve4.keyring: (13) Permission denied and ceph daemon isn't starting. It's happening on all of my 3 nodes...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.