Unfortunately it seems like it is a hard requirement that your GPU supports SR-IOV in order for this new method of handling vGPUs on kernel 6.8 to work. The Tesla P4, which I am using, does not support SR-IOV. According to the documentation though, the latest versions of the v16 drivers should...
Yes the 535.154.02 (16.3) version of the host drivers did not work with the 6.8 kernel. I believe it was either 535.161.05 (16.5) or 535.183.04 (16.6) version of the host drivers that started installing and working successfully on the 6.8 kernel. I am currently pinning the 6.5 kernel on my vGPU...
As of v16.6 (or possibly v16.5), the kernel module for the R535 drivers does build correctly for the 6.8 kernels. I was asking specifically about this new way of handling vGPUs. As I recently found, booting into the 6.8 kernel with the v16.6 kernel module seems to work correctly except that no...
What about if you have hardware that is not supported by the R550 drivers? Are users of Pascal generation hardware permanently stuck on the 6.5 kernel?
I have two proxmox nodes that have Nvidia Tesla P4s in them. I have been stuck having to pin kernel version 6.5 on these nodes for awhile now due to the vGPU kernel modules not building successfully for newer kernels.
Today I noticed when updating the official vGPU host drivers to version 16.6...
I am having the exact same issue here. Every time I shutdown a VM with a vGPU attached to it, the vGPU process is still running according to nvidia-smi and I see the same errors in dmesg as was posted in the OP. Running a currently fully up-to-date proxmox 7.3-6 with nvidia vgpu driver version...
I recently rebuilt my proxmox cluster. I have 2 proxmox nodes, and this time I am using a raspberry pi as the qnetd server. I followed the guide.
For the purposes of this post, I am going to use these names:
1st proxmox node: NODE1
2nd proxmox node: NODE2
cluster name: CLUSTER
As soon as I...
Hello all,
Last night when my automated backup jobs ran, the job for backing up one of my LXC containers failed. Earlier that day I had updated all my VMs, containers, and Proxmox hosts and rebooted my hosts (which also rebooted my VMs and containers). I did not make any changes to this...
I figured it out! Replying to myself in case this is useful to anyone.
While reading the Ubuntu manpage for corosync-qdevice, I noticed a section that says this:
That section didn't exactly explain my issue but I wasn't aware that there is an NSS database for both the qnetd server as well as...
Hello all,
A few weeks ago I upgraded my proxmox homelab setup to include a cluster. I have a primary node that stays online 24/7, I have a secondary node that is turned on only when reboots are needed or in case of issues with the primary node, and I have a docker container deployed on my NAS...
Hi all,
I have been troubleshooting issues with kernel panics on Proxmox 7.0-13 for the past week but I was unsure of the cause due to lots of configuring and setting up prior to noticing the issues. I decided to reinstall Proxmox and start from scratch today and was able to prove that...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.