Unfortunately these types of bugs have been what's preventing me from paying for support as I need stable GVT support for Coffee Lake. It really sucks since I'm a fan of Proxmox's Ceph hyper-converged concept.
I am currently experiencing random kernel panics on Intel GVT vGPU VMs at random times. I've noticed that even though I get these kernel panics, the non-vGPU VMs that are running on the same host seem to be unaffected. I've set the following settings in my /etc/sysctl.conf
kernel.panic = 120...
I'm running into the same issues as the poster here https://github.com/intel/gvt-linux/issues/153
Running newest version of Proxmox with all latest non-subscription updates with Windows VMs with Intel GVT-g active. VMs seem to crash at random times with crash frequency increasing on heavy GPU...
Those errors are normal. There is a kernel flag to not print those but I can't remember the option name right now.
What modules are you loading? It seems like you're missing some.
Please output the contents of the files in /etc/modules.d and /etc/modules
You need to have these modules loaded...
It's i915.enable_gvt=1 not i915_enable.gvt=1
Also you want kvm.ignore_msrs=1 - otherwise a whole plethora of apps in your VM could crash all of the GPU-enabled VMs on your host and cause a kernel panic.
What modules and grub cmdline settings are you using? I have gvt-g working, but I get random gvt page fault errors under load with Windows VMs that take down the entire host and all the gvt VMs running on it. Is it that much more stable with Linux VMs?
Are you sure that's your problem? Install the intel-gpu-tools package on your host and run "intel_gpu_frequency" while your CPU is under load to check what frequency your Intel GPU is running at. You can also use "intel_gpu_top" to see how much load you're placing on the Intel GPU.
I'm running...
Attempting to get CARP and even DHCP working between two OPNsense firewalls. It seems to work reliability when both are on the same node or if I"m using normal Linux bridges. However, when the two instances are on different nodes and OVS is being used, communication does not work properly. I've...
I've already tried this but it seems something may be lingering in the proxmox configuration database. Either that or there is some other place that's keeping a monmap from a previous Ceph installation that's being shared in the cluster fs, as something is populating the UI with monitors I've...
I'm hoping this gets rid of problems I've been having with the most recent 5.3 kernel and random page faults arising from the use of Intel GVT-g on a Coffee Lake ER Xeon processor.
Just an update - tried completely removing the ceph packages, including an 'apt purge' to remove any lingering configs. Also removed all of the keyring and ceph configuration files from /etc/ceph and /etc/pve/ceph.conf and /etc/pve/priv/ceph*.keyring.
Then I also removed all of the ceph...
What is the best way to clean up a bad ceph config and start from scratch without rebuilding the Proxmox server (as everything else works fine)? (This is Proxmox VE 6.1)
Configured a ceph cluster that was working, although monitors for some reason were showing up twice in the Proxmox GUI, one...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.