Exactly the same problem here on an Intel NUC 12. I have tested extensively with BIOS settings, but that makes no difference. It is also not a Proxmox only problem, because a fresh installation of Debian 12.2 produces the same problem and even trixie/sid has the same problem. I don't use HW...
/dev/sdb2 was just an example, in your case it’s /dev/cciss/c0d1p1.
You mean when you stop and remove the OSD using GUI it time out? The OSD is mounted (and visible to the server), so I think no hardware issue, probably just a faulty filesystem. When you have not much knowhow about this kind of...
Are you sure the disk is still available in the server? Check with “lsblk”. If you see the disk with “lsblk” (in this example as /dev/sdb) and the disk is not mounted (not showing up with “df -h”), try to manually mount the disk:
# mount /dev/sdb2 /var/lib/ceph/osd/ceph-0
# start ceph-osd id=0...
Yes, seems to be a bad disk to me when this problems occurs that fast. I only work with WD RAID Edition and WD Gold for HDDs and in the few times I had issues like this (with new drives or drives in warranty), it was never a problem to replace the drive (when using CC even with advance...
For SSDs some bad/dead sectors, especially after some time of using, are “normal”. The pending sectors will be (after some time) reallocated on the disk and then disappear (you can find them under reallocated sectors after they are reallocated). From my experience: for normal HDDs (like your...
Probably a network issue. Do you use bonding? And if so, what mode? Maybe after the updelay another uplink goes active, can’t talk via multicast to other servers on other switch (caused by wrong config on switches) and gives issues. Try to omping the other nodes just after reboot (when it’s...
Good job guys! I did the upgrade on a 3 nodes cluster from Ceph Jewel (10.2.10) to Luminous (12.2.1) and PVE 4.4 to 5.1 and also changed from Ceph Filestore to Bluestore. No serious problems found so far.
Today I installed some minor updates on a small PVE 4.4 cluster, since the updates I noticed the Software Watchdog Initial Countdown has changed from 120 seconds to 600 seconds. Isn't this value a bit high? Is it possible for us to change it back to 120 seconds?
Timer Use...
If I want to use software watchdog, is it a good idea to disable NMI Watchdog? Like you should disable NMI Watchdog when using a hardware watchdog? Why I'm asking:
I use the software watchdog (everything default) and sometimes when I do a reboot of the node, it comes back online with the...
We had 5 exactly the same physical nodes for our PVE 4.4 cluster. Now we needed to expand our capacity, so we bought 2 more servers, but this new servers aren't the same as the first servers. All 7 servers have dual Intel Xeon CPU's, but different models. Clockspeed of the first servers is...
Okay, well, it works for me. I have tested with 91 GB max. and boots without any problem now.
When I move the VM to a host without this change applied, it will not boot. So seems to be the solution. :)
Situation:
Physical host has 2 CPU-sockets with each 6-core's and HT-enabled. So, 2 x 6 x 2 = 24 vCPU's
Each CPU have 48 GB of memory installed in dual-channel configuration:
[ 16 GB ] [ 8 GB ]
[ 16 GB ] [ 8 GB ]
So, total system memory is 96 GB. Host is running PVE 4.3-9.
When I have a VM...
No, this isn't possible. Like Udo wrote: a node can only be member of one cluster. But offcourse you can make bigger clusters than only 3 nodes (this is only a minimal requirement). Always keep an odd number of nodes in a cluster (3, 5, 7, 9, 11 etc).
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.