I'm still getting the same "Detected Hardware Unit Hang" errors sporadically when using PVE kernel 5.4.
Mar 19 20:11:15 pve-host1.local kernel: [30377.339967] e1000e 0000:00:1f.6 eno1: Detected Hardware Unit Hang:
I recall there was previously some advice around setting: ethtool -K <ADAPTER>...
Great Thanks! I'll take a look at testing the PVE 5.4 kernel! Will report back but it would be good to know here if others have tried the PVE 5.4 kernel and what the results were for this issue...
Thanks!
Has anyone tested this yet and any results they can share?
@spirit - Could you provide details of the patch along with how we would installed the patched kernel to test?
What’s your configuration? Are you passing through the entire host GPU to a single VM or have you tried the gvt-g (mediated devices) method (if indeed this is available for Gemini Lake - It works on my Coffee Lake headless VMs)
Hi There,
I'm running latest pve and using ZFS replication between local-zfs on 2 nodes. Every time the replication runs it seems to output the following type of messages in syslog without any clear meaning what they are referring to or whether they are just unnecessary noise?
Feb 28 07:29:01...
I'm getting this same "eno1: Detected Hardware Unit Hang" in the syslog regularly. I haven't yet experimented with disabling the features of the NIC but I did notice that my NIC (on a fairly new Intel NUC8I5BEH) is listed as follows:
# lspci -v | grep Ethernet
00:1f.6 Ethernet controller: Intel...
Yes. I have one NUC using a single NVMe and one NUC using a single SSD. Both are configured to use the ZFS file system and are clustered together (with a qdevice as a 3rd node for quorum)...
I replicate the local ZFS based VMs between the nodes but also back them up daily to shared NFS storage...
I have Docker running inside a VM on the PVE cluster. Inside Docker runs the TIG (Telegraf, InfluxDB and Grafana) containers. I then installed SNMP server on all PVE physical hosts and use Telegraf (Docker VM) to poll the PVE hosts for generic CPU/Mem/Disk/Network SMNP metrics, write the...
Hi There,
I currently have a 2-node Proxmox 6.1 cluster (3rd node is qdevice) where both nodes are the same hardware spec and support Intel GVT-g (mediated device) Passthrough.
I know live migration between nodes is not possible with standard GPU device Passthrough but, if GVT-g enables the...
I'm seeing the same issue in testing with passing through the IGP using Intel GVT-d ... It looks like this issue was also previously seen when passing through the full IGP (not as mediated device) and was fixed for that scenario but looks like the mediated device passthrough still exhibits this...
What happens if you remove the args: -device vfio-pci,host=00:02.0,addr=0x18,x-igd-opregion=on" from the VM conf and simply pass the GPU through using the webgui (which results in the line hostpci0: 00:02.0 being added to the config)?
This is the way i do it for a VM with IGP passthrough but I...
So does this mean we don’t have to enable the test repository any more to install the 5.3 kernel and it can be installed using “apt install” if the no-subscription repository is enabled? I don't see the 5.3 Kernel listed as part of the "Available Updates" panel in the Web GUI?
Regarding the...
The following is the output of pve-efiboot-tool kernel list
root@pve6-test1:~# pve-efiboot-tool kernel list
Manually selected kernels:
None.
Automatically selected kernels:
5.0.15-1-pve
5.0.21-3-pve
5.3.7-1-pve
So, I then ran the following commands as you stated above:
root@pve6-test1:~#...
Hi @fabian
I tried the work around above and the following is the output. On rebooting the 5.3 Kernel is still shown in the EFI Boot Menu (as default) and is loaded despite the output from the commands above?
Thanks!
root@pve6-test1:~# uname -a
Linux pve6-test1 5.3.7-1-pve #1 SMP PVE 5.3.7-1...
Hi @fabian
I see you’ve logged a bug report for this so I’ll also monitor the updates there.
https://bugzilla.proxmox.com/show_bug.cgi?id=2448
Is there anyway currently to work around this issue prior to it occurring (ie before removing the currently booted kernel) or to fix the module...
lsmod nls_iso88591-1 command just errors with "Usage: lsmod" (I think there is a mistake in your commands above?)
lsmod (run jus as "lsmod") output as follows:
uname -a output as follows:
Thanks!
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.