Search results

  1. P

    Sync log only partially shown in GUI

    There is a hard stop. It does not seem to be loading/hanging. The log of the current sync job spans 2 1/2 hours and the last entry is about 2 days old.
  2. P

    Sync log only partially shown in GUI

    So I have a PBS running onsite and another one remotely that is syncing the local PBS's data for additional safety. In the remote PBS' GUI I can see the running sync job's log but while the job has been running for 1 day and three hours, the log shown only spans less than one hour. Is there a...
  3. P

    NIC keeps changing interfaces on reboots

    I am having a similar issue: On my Lenovo P700 with two built in NICs (eth1 and eno1) and under a new installation of PVE8, the two built in NICs keep swapping their names at every boot. I tried this (naming the NICs eth0 and eth1) but it does not work consistently for me: Yesterday, for...
  4. P

    setting up NVidia vGPU is driving me mad

    Yeah, this thing remains a mystery too me. I have wasted more hours on this than on any other thing in my home lab ever. NVidia support is going to be my next stop. But I thought I'd try here first, given that the Proxmox guys have experimented with vGPU (albeit not necessarily Cuda) with the...
  5. P

    vGPU driver can't find the kernel

    Sorry, my bad. The error I was getting was the same as in the OP. I misread the comment about not having to specify any options as not having to install the headers. After installing the headers, the error message went away.
  6. P

    setting up NVidia vGPU is driving me mad

    I am trying to set up an Nvidia vGPU for AI workloads and I am following this guide: https://pve.proxmox.com/wiki/NVIDIA_vGPU_on_Proxmox_VE_7.x# My card is an RTX A5000. I obtained the NVidia grid host driver (535.129) and installed it on the PVE host. I activated SR-IOV and now have several...
  7. P

    vGPU driver can't find the kernel

    Hmm. I have a normal pve kernel and I am getting exactly this error message trying to install the vGPU driver. pveversion -v: proxmox-ve: 8.1.0 (running kernel: 6.5.11-4-pve) pve-manager: 8.1.3 (running version: 8.1.3/b46aac3b42da5d15) proxmox-kernel-helper: 8.0.9...
  8. P

    24 Scrub Errors, 3pgs inconsistent

    Response: Health ok Response: a super long list of all my PGs. I did find one that is - apparently - being deep scrubbed. It says the scrubbing has been going on for 143 seconds so far. I will check back a couple of times to see whether anything changes. But there are also a few that have been...
  9. P

    24 Scrub Errors, 3pgs inconsistent

    So I have been waiting for the last 1 pg to come out of deep scrubbing before I replace the next disk(s). But it doesn't. Since I wrote this above, I have constantly had 1 pg being deep scrubbed (without knowing whether this is even true and, if it is, whether this is the same 1 pg all the time...
  10. P

    24 Scrub Errors, 3pgs inconsistent

    3.13 is now gone as well. (1 pg is still being deep scrubbed) So (other than the 1 pg being deep scrubbed) all is green again - woohoo! And now I will shake things up a bit again by replacing the two smaller HDDs in Node2 with the one new 22TB HDD (and hope this goes smoothly). Thanks for...
  11. P

    24 Scrub Errors, 3pgs inconsistent

    3.36 is now gone as well. 3.13 is active, clean and inconsistent. (But there is one pg still being deep scrubbed - so maybe that is 3.13 and it will be gone soon, too?) I'll report back. BTW: Today, the second 22TB HDD for this pool arrived and I can replace the OSDs in one more node. But I...
  12. P

    nVIDIA vGPU mdev setup not working (as per wiki)

    So I just tried to pass one of the vGPU devices to a VM but I get this error message: TASK ERROR: Cannot bind 0000:02:00.0 to vfio Interestingly, after this error, mdevctl types returns nothing. Before the error, it returns a list of mdevs. So something is off. But what?
  13. P

    Ceph ghost OSDs in configuration database?

    Hi, I have a three node home lab PVE + Ceph cluster. Recently, I replaced one node with another server. While I removed the old node from the PVE cluster config, I missed removing the OSDs from Ceph first. So for a while they were shown (toghether with the old node ("bucket")) as down...
  14. P

    24 Scrub Errors, 3pgs inconsistent

    I just went through upgrading my cluster to PVE version 8. After the upgrade, OSD.3 did not come back online again (it kept being shown as "in" but "down", irrespective of how often I started it or the entire node). I then checked the SMART values for the drive again and this time I did find...
  15. P

    nVIDIA vGPU mdev setup not working (as per wiki)

    Hi, I am trying to set up a NVIDIA A5000 as vGPU as per this wiki article: https://pve.proxmox.com/wiki/NVIDIA_vGPU_on_Proxmox_VE_7.x Got to the point where I enabled SRIOV and can list the virtual functions via lspci. But then it says under guest configuration that I should pick an mdev in...
  16. P

    24 Scrub Errors, 3pgs inconsistent

    No scrub information available for pg 3.36 error 2: (2) No such file or directory No scrub information available for pg 3.48 error 2: (2) No such file or directory Correct. Thank you for the recommendation. I am aware of this problem and am in the process of rectifying. But this being a...
  17. P

    24 Scrub Errors, 3pgs inconsistent

    "data_digest": "0xec9c58e1", "omap_digest": "0xffffffff", "expected_object_size": 4194304, "expected_write_size": 4194304, "alloc_hint_flags": 0, "manifest": { "type": 0 }...
  18. P

    24 Scrub Errors, 3pgs inconsistent

    { "object": { "name": "rbd_data.453d0387ce7766.00000000001028e4", "nspace": "", "locator": "", "snap": "head", "version": 371918 }, "errors": [], "union_shard_errors"...
  19. P

    24 Scrub Errors, 3pgs inconsistent

    Done. Sorry, I don't understand. Where/when wasn't I paying attention. And what do you mean that "'inactive' obiviously doesn't fit? Sorry, my Ceph knowledge is at "noob" level... PG_STAT STATE UP UP_PRIMARY ACTING ACTING_PRIMARY 3.76...