[SOLVED] Problems after upgrade to 7.2: Guest Agent not running / No network when PCI passthrough is enabled

MisterDeeds

Active Member
Nov 11, 2021
143
33
33
35
Hello together

Since the upgrade to version 7.2 I have the following problems:

- The Guest Agent is running on the VM, but the web gui shows it as "Guest Agent not running". For example, the shutdown function via web gui does not work.
Guest Agent.png

- As soon as I add a PCI device, the network connection no longer works. As soon as the PCI device is removed again, it works without any problems. However, this is only after the driver has been installed. As long as the PCI device has no driver, the VM starts, shows the device correctly in the device manager simply as unknown.
passthrough.png

Does anyone have a similar problem?

root@PVE01:~# pveversion -v
proxmox-ve: 7.2-1 (running kernel: 5.15.30-2-pve)
pve-manager: 7.2-3 (running version: 7.2-3/c743d6c1)
pve-kernel-helper: 7.2-2
pve-kernel-5.15: 7.2-1
pve-kernel-5.13: 7.1-9
pve-kernel-5.15.30-2-pve: 5.15.30-3
pve-kernel-5.13.19-6-pve: 5.13.19-15
pve-kernel-5.13.19-2-pve: 5.13.19-4
ceph-fuse: 15.2.15-pve1
corosync: 3.1.5-pve2
criu: 3.15-1+pve-1
glusterfs-client: 9.2-1
ifupdown2: 3.1.0-1+pmx3
ksm-control-daemon: 1.4-1
libjs-extjs: 7.0.0-1
libknet1: 1.22-pve2
libproxmox-acme-perl: 1.4.2
libproxmox-backup-qemu0: 1.2.0-1
libpve-access-control: 7.1-8
libpve-apiclient-perl: 3.2-1
libpve-common-perl: 7.1-6
libpve-guest-common-perl: 4.1-2
libpve-http-server-perl: 4.1-1
libpve-storage-perl: 7.2-2
libspice-server1: 0.14.3-2.1
lvm2: 2.03.11-2.1
lxc-pve: 4.0.12-1
lxcfs: 4.0.12-pve1
novnc-pve: 1.3.0-3
proxmox-backup-client: 2.1.8-1
proxmox-backup-file-restore: 2.1.8-1
proxmox-mini-journalreader: 1.3-1
proxmox-widget-toolkit: 3.4-10
pve-cluster: 7.2-1
pve-container: 4.2-1
pve-docs: 7.2-2
pve-edk2-firmware: 3.20210831-2
pve-firewall: 4.2-5
pve-firmware: 3.4-1
pve-ha-manager: 3.3-4
pve-i18n: 2.7-1
pve-qemu-kvm: 6.2.0-5
pve-xtermjs: 4.16.0-1
qemu-server: 7.2-2
smartmontools: 7.2-pve3
spiceterm: 3.2-2
swtpm: 0.7.1~bpo11+1
vncterm: 1.7-1
zfsutils-linux: 2.1.4-pve1
root@PVE01:~#

Thanks and greetings
 
Hello,

Please provide us with the VM config, qm config 100, and I would try to change the Machine type to 6.2 by going to VM ⇾ Hardware ⇾ Machine click edit and select the latest one.
 
Hi Moayad

Thank you for the answer. Unfortunately it has no effect when the machine type is set to 6.2. The behavior is identical. Here still the config:

root@PVE01:~# qm config 100
agent: 1,fstrim_cloned_disks=1
balloon: 0
bios: ovmf
boot: order=scsi0;ide2;net0
cores: 8
cpu: host
efidisk0: NAS01-Vm:100/vm-100-disk-1.qcow2,efitype=4m,pre-enrolled-keys=1,size=528K
hostpci0: 0000:08:00,pcie=1,x-vga=1
ide2: none,media=cdrom
machine: pc-q35-6.2
memory: 16384
meta: creation-qemu=6.1.0,ctime=1641474635
name: vPC01
net0: virtio=36:35:6D:87:D0:77,bridge=vmbr0,firewall=1
numa: 1
onboot: 1
ostype: win10
scsi0: NAS01-Vm:100/vm-100-disk-0.qcow2,cache=writeback,discard=on,size=50G,ssd=1
scsihw: virtio-scsi-pci
smbios1: uuid=13eb2130-a64f-4034-aee9-2af719a27454
sockets: 1
vmgenid: 96f0455f-e1db-4de8-8c2e-ed29bdb98731
root@PVE01:~#
 
  • Like
Reactions: bsinha
Yes exactly, this is the device. This is an Nvidia graphics card

IOMMU is also correctly activated.

Code:
root@PVE01:~# dmesg | grep -e DMAR -e IOMMU
[    0.847165] pci 0000:00:00.2: AMD-Vi: IOMMU performance counters supported
[    0.847815] pci 0000:00:00.2: AMD-Vi: Found IOMMU cap 0x40
[    0.858850] perf/amd_iommu: Detected AMD IOMMU #0 (2 banks, 4 counters/bank).
root@PVE01:~#
 
Last edited:
Thank you! I would check for the output of lspci maybe you're missing the decimal portion from the PCI device?
 
Thanks for the answer. That actually looks good too:

Code:
root@PVE01:~# lspci -n -s 08:00
08:00.0 0300: 10de:1fb0 (rev a1)
08:00.1 0403: 10de:10fa (rev a1)
root@PVE01:~# cat /etc/modprobe.d/vfio.conf
options vfio-pci ids=10de:1fb0,10de:10fa disable_vga=1
root@PVE01:~#

I have an identical host, same installation same graphics card. There I have created a new VM from scratch. However, the phenomenon is identical. As soon as the driver is installed, the VM loses the network connection.
 
I found out. It seems to be a problem with the latest Linux kernel (5.15.30-2-pve)

Code:
proxmox-boot-tool kernel pin 5.13.19-6-pve

reboot

solved it for me. Unfortunately, the NVIDIA drivers are a bit tricky. The same is for NVIDIA's VGPU driver. The driver 510.47.03 only works under the kernel 5.13.19-6-pve
 
I found out. It seems to be a problem with the latest Linux kernel (5.15.30-2-pve)

Code:
proxmox-boot-tool kernel pin 5.13.19-6-pve

reboot

solved it for me. Unfortunately, the NVIDIA drivers are a bit tricky. The same is for NVIDIA's VGPU driver. The driver 510.47.03 only works under the kernel 5.13.19-6-pve
Thank you very much brother for your help i have same problem after update your post help me thanks again
 
  • Like
Reactions: MisterDeeds
The latest NVIDIA vGPU driver (14.1) 510.73.06 now also works with the latest kernel version 5.15.39-1-pve

Kernel.PNG

Best regards
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!