For stretch PVE + Ceph clusters we recommend a full PVE install for the tie-breaker node. See the newly published guide: https://pve.proxmox.com/wiki/Stretch_Cluster
its an innocent enough question- but there are a lot of gotchas you need to consider.
Clusters are made up of 3 elements- compute, storage, and networking. lets touch on each.
COMPUTE:
- Dell R630 is a 10 year old platform. as such, it offers...
Where did you get that idea?! ceph doesnt care about ANY of those things. Just be aware that the overall performance of the cluster will be as slow as the slowest monitor.
ceph doesnt need or perform any live migration. As for memory, nodes...
Hi riahc3
There is no offical PVE HCL but PVE uses a modified Ubuntu Kernel you can check the Ubuntu HCL. PVE9 currently uses Kernel version 6.14 (Ubuntu HWE) and you can opt-in for a 6.17 Kernel.
Just be aware that the Ubuntu HCL, as you...
I think you were correct. It hasn't crashed since i swapped out the entire system about 50 hours ago, i was getting daily crashes of ceph and hard lockups on the system...
the ceph documentation has a dedicated page for pg troubleshooting: https://docs.ceph.com/en/latest/rados/troubleshooting/troubleshooting-pg/
this is also a good resource: https://ceph.io/geen-categorie/ceph-manually-repair-object/
generally...
Jein, die Standardsachen deckt die GUI ab, aber den kompletten Umfang was mit dem Linux-Unterbau möglich ( inklusive potentiell gefährlicher Aktionen) ist, kann und wird die GUI nie ganz abdecken.
Eigentlich nicht...
Tell Ceph to benchmark those drives again on OSD start and restart the service when appropriate:
ceph config set osd osd_mclock_force_run_benchmark_on_init true
There's also another ceph tell like command to run a benchmark right now, but I...
Because there's no need for such a plan when you can simply download the software for free and start using it. Would you really prefer having to go through a registration process, handing over your email address and personal data, only to get a...
Sorry, but you seem to be deliberately mixing things up, probably just because you’re frustrated that no one was able to help you in the other thread.
With the Community Subscription, you’re paying for access to the Enterprise Repositories. The...
@Ruprecht, I have been operating proxmox for 10 years without a subscription. Still using it today and I'm gonna buy the small business subscription for one of our work servers running proxmox.
No, you don't have to buy it! You can stay on the...
Aber du kannst eine vm auch erstmal im .vmdk format auf einem pve benutzen (initial +last-sync) und später on-the-fly im pve zu raw/qcow2 umziehen.
Bei einer Optimierung eines Vollumstiegs sollte man sich immer erstmal bewußt machen, wo der...
Hello Johannes, gurubert,
thank you for your insights.
I do have a third gigabit link reserved for corosync only, I probably should have mentioned that, but left it out since I only was interested in the ceph behaviour. Looks like I will run...
That's just not how it works.
"Failure domain" is host. You could have shutdown one host with all 4 OSDs going down. That would have worked fine.
All data is distributed between those three nodes. You do not know onto which OSD. (Well, you...
Yes, most probably you have lost data. As you are having 3 copies of every PG, it can happen that PG was located on those three disks, so when you have removed them at the same time it became unavailable. As you have wiped OSDs, I dont think...