Search results

  1. P

    24 Scrub Errors, 3pgs inconsistent

    Sorry, I had missed that. I edited my post above accordingly.
  2. P

    24 Scrub Errors, 3pgs inconsistent

    Yes, they were on another node that I have since removed but missed the window to remove the OSDs first... I haven't got around to delete them from <whererever>. They don't play a role here. ceph health detail HEALTH_ERR 26 scrub errors; Possible data damage: 3 pgs inconsistent [ERR]...
  3. P

    24 Scrub Errors, 3pgs inconsistent

    So, scrubbing has continued and apparently, it identified two more errors: ceph -s cluster: id: xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx health: HEALTH_ERR 26 scrub errors Possible data damage: 3 pgs inconsistent services: mon: 3 daemons, quorum...
  4. P

    24 Scrub Errors, 3pgs inconsistent

    Hi, On my 3 node home lab cluster, Ceph tells me that it has discovered 24 scrub errors and that 3 pgs are inconsistent. That does not sound overly promising. More importantly, however, I have no idea what to do with this information... I take it that the nature of this error prevents Ceph...
  5. P

    Help configuring vGPU?

    Hmm. So disabling the card's displayport seems to have worked (the benefit of this is that I got my KVM video feed back because the card does not override the onboard KVM anymore) - yay! I then removed the blacklist entry from GRUB and nvidia was loaded again. But lspci -d 10de: still does...
  6. P

    Help configuring vGPU?

    Ah, got it disabled by adding module_blacklist=nvidia to the GRUB_CMDLINE_LINUX_DEFAULT line in /etc/default/grub cross my fingers that disabling the displayport is going to do the trick...
  7. P

    Help configuring vGPU?

    Hi, I am trying to get vGPU to work by following this guide: https://pve.proxmox.com/wiki/NVIDIA_vGPU_on_Proxmox_VE_7.x#cite_ref-3 My card is an RTX A5000, the same as used in the guide, so should work. I went through with the setup but it doesn't work. lspci -d 10de: results in a very short...
  8. P

    Trying to understand nVidia vGPU

    Hi, so far, I have only used PCIe passthrough to assin a single graphics card to a single VM - that works well as long as the CPU supports IOMMU, which isn't the case with my budget Xeon E3-12xx servers. But I understand there is now another way: vGPU where I can assign (parts of) the same GPU...
  9. P

    Mailcow + PMG make sense?

    If I am not mistaken, then the temporary down time of your mail server should not result in missing (or even the loss of) emails. The sending mail server should keep trying for a while (up to two days or so) before giving up.
  10. P

    Mailcow + PMG make sense?

    And where do you see the benefits of combining the two (they seem to be duplicating some functionality between them)?
  11. P

    rbd error: rbd: listing images failed: (2) No such file or directory (500)

    Hi, I have a three node PVE cluster with Ceph with (currently) three pools. This is about one of them. When I click in the GUI on one of the pools and want to view the VM Disks or CT Volumes, I just get the error message "rbd error: rbd: listing images failed: (2) No such file or directory...
  12. P

    Ceph: Balancing disk space unequally!?!?!?!

    One more question for my understanding please: So when I have two disks on a node and one drive goes down, Ceph will try to push its content to the other drive (and if that one isn't large enough, I have a problem). If I only have one drive on a node that doesn't happen. Understood. But what...
  13. P

    [SOLVED] PBS Metrics?

    Bonus question: Are there any compatible grafana dashboard? I can only find two in the grafana dashboard library. One requires Prometheus and the other a custom script on the host. Anything that works with the PBS metrics server (out of the box)?
  14. P

    High iodelay on one of three identical nodes

    The PVE GUI Yeah, that is something that would be useful. Cool. Thanks. Will try that. But in the case at hand, my issue does not seem to be coming from a VM, because there are no VMs on the node and the VMs running on the other node do not affect the iodelay of the other nodes (so badly)...
  15. P

    Ceph: Balancing disk space unequally!?!?!?!

    Yes, I had one disk (per type) per node for the longest time. Huh, and there was I thinking that adding an HDD per node to the HDD pool would actually improve operational safety... At the moment, I have approx. 14TB worth of data across the two HDDs per node. What you are telling me, if I...
  16. P

    Ceph: Balancing disk space unequally!?!?!?!

    Yes. Maybe there was a misunderstanding. I am not wondering why the 14TB and the 4TB HDDs get a different number of PGs. That is expected, as you explain. My issue is that on Host 2 and Host 3 the 14TB HDD gets filled only to 76% whereas the 4TB HDD gets filled to 92% capacity while on Host 1...
  17. P

    Ceph: Balancing disk space unequally!?!?!?!

    -1 79.43169 - 65 TiB 44 TiB 44 TiB 446 KiB 79 GiB 21 TiB 67.88 1.00 - root default -3 21.61177 - 22 TiB 15 TiB 15 TiB 153 KiB 26 GiB 6.9 TiB 67.88 1.00 - host 1 0 hdd 12.73340 1.00000 13 TiB 10...
  18. P

    Cluster getting really ssssllllloooooowwwwww :-(((((((((((((((((((((((((

    Agreed. Originally I thought it was the same topic, so I posted the questions here. But then I realized that it is a different issue, so I started a new thread.
  19. P

    High iodelay on one of three identical nodes

    I'd say it's definitely no local storage (because there are no VMs running on the node anymore). It could be the HDDs for sure. But shouldn't that affect the other nodes as well? (They are all practically identical). I will take a snapshot later and post the technical details.