Rasdaemon Errors when using GPU Passthrough

tom_atx

New Member
Oct 29, 2023
2
0
1
Hi All,

I'm completely new to Proxmox. I built a budget homelab with scavenged parts which include the following components:

256 GB Registered Micron Memory (32x8) @2133MT/s (purchased used)
2x 18core Xeon e5-2696 v3 Processors (purchased used)
Machinist D8 Max X99 2011-3 motherboard (new off Ebay)
2x 2TB NVMe
1x 500GB NVMe (using a PCIe adapter)
1x 8TB Seagate NAS drive
3x 12TB Seagate NAS drives
NVidia GT730 low profile GPU
NVidia GTX 1660 Ti

Using the latest version of Proxmox at the time of writing: 8.0.4

All memory passed after a single pass using the memtest tool included in the Proxmox install ISO.

PXL_20231028_153247688.jpg

Since the motherboard doesn't have onboard video, I reserved the low profile GT730 for the PVE host, so I didn't blacklist 'nvidia' in /etc/modprobe.d/blacklist.conf. Perhaps there's a better way to isolate the GPUs that I'm unaware of.

Successfully installed Win Server 2022, turned Ballooning off as this is the VM I'm using to passthrough. The VM seems stable - everything works fine, running for two days without issue and verified the GTX 1660 Ti is transcoding videos just fine; however, I'm getting a sh*t-ton of rasdaemon errors. This occurs at least 35 times every 2-4 minutes based on what I'm seeing in the syslog. Note: This only occurs when the VM I'm using for passthrough is turned-on.

From PVE -> syslog
Code:
Nov 03 14:53:02 pve rasdaemon[1518]:            <...>-278270 [000]     0.015316: mce_record:           2023-11-03 12:11:03 -0500 bank=5, status= c80000c000310e0f, Rx detected CRC error - successful LLR wihout Phy re-init, mci=Error_overflow Corrected_error, mca=BUS Level-3 Generic Generic Other-transaction Request-did-not-timeout Error, cpu_type= Intel Xeon v3 (Haswell) EP/EX, cpu= 0, socketid= 0, misc= 1dd87b000d9eff, mcgstatus=0, mcgcap= 7000c16, apicid= 0
Nov 03 14:53:02 pve rasdaemon[1518]: rasdaemon: mce_record store: 0x55788c8c0868
Nov 03 14:53:02 pve rasdaemon[1518]: rasdaemon: register inserted at db
Nov 03 14:53:02 pve rasdaemon[1518]:            <...>-278270 [000]     0.015316: mce_record:           2023-11-03 12:11:04 -0500 bank=5, status= 8800004000310e0f, Rx detected CRC error - successful LLR wihout Phy re-init, mci=Corrected_error, mca=BUS Level-3 Generic Generic Other-transaction Request-did-not-timeout Error, cpu_type= Intel Xeon v3 (Haswell) EP/EX, cpu= 0, socketid= 0, misc= 1df87b000d9eff, mcgstatus=0, mcgcap= 7000c16, apicid= 0
Nov 03 14:53:02 pve rasdaemon[1518]: rasdaemon: mce_record store: 0x55788c8c0868
Nov 03 14:53:02 pve rasdaemon[1518]: rasdaemon: register inserted at db
Nov 03 14:53:48 pve kernel: mce: [Hardware Error]: Machine check events logged
Nov 03 14:53:50 pve kernel: mce: [Hardware Error]: Machine check events logged
Nov 03 14:53:54 pve rasdaemon[1518]: rasdaemon: mce_record store: 0x55788c8c0868
Nov 03 15:42:08 pve kernel: mce_notify_irq: 29 callbacks suppressed

Here's a snapshot of the syslog in the web gui to give you some perspective.
2023-11-04 00_11_06-pve - Proxmox Virtual Environment.png

I have passthrough configured as follows:

/etc/default/grub
Code:
# If you change this file, run 'update-grub' afterwards to update
# /boot/grub/grub.cfg.
# For full documentation of the options in this file, see:
#   info -f grub -n 'Simple configuration'

GRUB_DEFAULT=0
GRUB_TIMEOUT=5
GRUB_DISTRIBUTOR=`lsb_release -i -s 2> /dev/null || echo Debian`
GRUB_CMDLINE_LINUX_DEFAULT="quiet intel_iommu=on video=efifb:off"
GRUB_CMDLINE_LINUX=""

# Everything else is commented out...

/etc/modules
Code:
# /etc/modules: kernel modules to load at boot time.
#
# This file contains the names of kernel modules that should be loaded
# at boot time, one per line. Lines beginning with "#" are ignored.
# Parameters can be specified after the module name.

vfio
vfio_iommu_type1
vfio_pci
vfio_virqfd

Below are results of lspci for the GTX 1660 Ti I'm using for passthrough:
Code:
03:00.0 VGA compatible controller: NVIDIA Corporation TU116 [GeForce GTX 1660 Ti] (rev a1) (prog-if 00 [VGA controller])
03:00.1 Audio device: NVIDIA Corporation TU116 High Definition Audio Controller (rev a1)
03:00.2 USB controller: NVIDIA Corporation TU116 USB 3.1 Host Controller (rev a1) (prog-if 30 [XHCI])
03:00.3 Serial bus controller: NVIDIA Corporation TU116 USB Type-C UCSI Controller (rev a1)

root@pve:~# lspci -n -s 03:00 -v
03:00.0 0300: 10de:2182 (rev a1) (prog-if 00 [VGA controller])
        Subsystem: 1458:3fc3
        Physical Slot: 6
        Flags: bus master, fast devsel, latency 0, IRQ 24, NUMA node 0, IOMMU group 56
        Memory at c6000000 (32-bit, non-prefetchable) [size=16M]
        Memory at b0000000 (64-bit, prefetchable) [size=256M]
        Memory at c0000000 (64-bit, prefetchable) [size=32M]
        I/O ports at 6000 [size=128]
        Expansion ROM at c7000000 [disabled] [size=512K]
        Capabilities: [60] Power Management version 3
        Capabilities: [68] MSI: Enable- Count=1/1 Maskable- 64bit+
        Capabilities: [78] Express Legacy Endpoint, MSI 00
        Capabilities: [100] Virtual Channel
        Capabilities: [258] L1 PM Substates
        Capabilities: [128] Power Budgeting <?>
        Capabilities: [420] Advanced Error Reporting
        Capabilities: [600] Vendor Specific Information: ID=0001 Rev=1 Len=024 <?>
        Capabilities: [900] Secondary PCI Express
        Capabilities: [bb0] Physical Resizable BAR
        Kernel driver in use: vfio-pci
        Kernel modules: nvidiafb, nouveau

03:00.1 0403: 10de:1aeb (rev a1)
        Subsystem: 1458:3fc3
        Physical Slot: 6
        Flags: bus master, fast devsel, latency 0, IRQ 38, NUMA node 0, IOMMU group 56
        Memory at c7080000 (32-bit, non-prefetchable) [size=16K]
        Capabilities: [60] Power Management version 3
        Capabilities: [68] MSI: Enable- Count=1/1 Maskable- 64bit+
        Capabilities: [78] Express Endpoint, MSI 00
        Capabilities: [100] Advanced Error Reporting
        Kernel driver in use: vfio-pci
        Kernel modules: snd_hda_intel

03:00.2 0c03: 10de:1aec (rev a1) (prog-if 30 [XHCI])
        Subsystem: 1458:3fc3
        Physical Slot: 6
        Flags: bus master, fast devsel, latency 0, IRQ 213, NUMA node 0, IOMMU group 56
        Memory at c2000000 (64-bit, prefetchable) [size=256K]
        Memory at c2040000 (64-bit, prefetchable) [size=64K]
        Capabilities: [68] MSI: Enable+ Count=1/1 Maskable- 64bit+
        Capabilities: [78] Express Endpoint, MSI 00
        Capabilities: [b4] Power Management version 3
        Capabilities: [100] Advanced Error Reporting
        Kernel driver in use: vfio-pci
        Kernel modules: xhci_pci

03:00.3 0c80: 10de:1aed (rev a1)
        Subsystem: 1458:3fc3
        Physical Slot: 6
        Flags: bus master, fast devsel, latency 0, IRQ 222, NUMA node 0, IOMMU group 56
        Memory at c7084000 (32-bit, non-prefetchable) [size=4K]
        Capabilities: [68] MSI: Enable+ Count=1/1 Maskable- 64bit+
        Capabilities: [78] Express Endpoint, MSI 00
        Capabilities: [b4] Power Management version 3
        Capabilities: [100] Advanced Error Reporting
        Kernel driver in use: vfio-pci
        Kernel modules: i2c_nvidia_gpu

/etc/modprobe.d/vfio.conf
Code:
options vfio-pci ids=10de:2182,10de:1aeb,10de:1aec,10de:1aed disable_vga=1

2023-11-03 16_02_22-pve - Proxmox Virtual Environment.png

2023-11-03 16_03_18-pve - Proxmox Virtual Environment.png

I hope I provided enough information. I mean, like I said, the VM itself seems to be running just fine, but I'm worried about all of the errors exponentially filling the journal.

Code:
root@pve:~# journalctl --disk-usage
Archived and active journals take up 194.3M in the file system.
root@pve:~# journalctl --verify
PASS: /var/log/journal/8cef86116857454faa3b558da6c69a46/system@00060911828b5d38-6b7ff2941c94eb44.journal~                                  
PASS: /var/log/journal/8cef86116857454faa3b558da6c69a46/system.journal                                                                    
PASS: /var/log/journal/8cef86116857454faa3b558da6c69a46/user-1000.journal                                                                  
PASS: /var/log/journal/8cef86116857454faa3b558da6c69a46/user-1000@f4dee43b694e47a4a728257e008e6d09-0000000000001100-000608703aab58d6.journal
PASS: /var/log/journal/8cef86116857454faa3b558da6c69a46/system@00060911201b87f1-441d44f8604fb010.journal~                                  
PASS: /var/log/journal/8cef86116857454faa3b558da6c69a46/system@3735a9fc4310409eb2bd8de63e9215cf-0000000000011b1a-0006091182892e9d.journal  
PASS: /var/log/journal/8cef86116857454faa3b558da6c69a46/system@000608bb81e81d81-7754a38525570a35.journal~

I'm hoping someone out there can identify what I'm doing wrong. Thanks!!

(Edit: Added snapshot of syslog from web gui)
 
Last edited:
Still looking for answers. I have 32x8 ECC DDR4 memory. Are these memory errors? While Googling, found someone who was trying to sort a similar issue but no response. How can I further troubleshoot?

dmesg:
Code:
root@pve:~ dmesg
...
[269476.377525] mce_notify_irq: 36 callbacks suppressed
[269476.377529] mce: [Hardware Error]: Machine check events logged
[269479.353579] mce: [Hardware Error]: Machine check events logged
[269540.377967] mce_notify_irq: 35 callbacks suppressed
[269540.377971] mce: [Hardware Error]: Machine check events logged
[269542.357974] mce: [Hardware Error]: Machine check events logged
[269601.366370] mce_notify_irq: 29 callbacks suppressed
[269601.366373] mce: [Hardware Error]: Machine check events logged
[269603.350390] mce: [Hardware Error]: Machine check events logged
[269662.362780] mce_notify_irq: 36 callbacks suppressed
[269662.362784] mce: [Hardware Error]: Machine check events logged
[269663.354789] mce: [Hardware Error]: Machine check events logged
[269723.351187] mce_notify_irq: 33 callbacks suppressed
[269723.351193] mce: [Hardware Error]: Machine check events logged
[269725.367193] mce: [Hardware Error]: Machine check events logged
[269783.355632] mce_notify_irq: 37 callbacks suppressed
[269783.355638] mce: [Hardware Error]: Machine check events logged
[269784.379592] mce: [Hardware Error]: Machine check events logged
[269844.376014] mce_notify_irq: 36 callbacks suppressed
[269844.376021] mce: [Hardware Error]: Machine check events logged
[269845.372010] mce: [Hardware Error]: Machine check events logged

rasdaemon:
Code:
root@pve:/# ras-mc-ctl --errors | tail
158857 2023-11-04 18:03:33 -0500 error: Rx detected CRC error - successful LLR wihout Phy re-init, mcg mcgstatus=0, mci Corrected_error, mcgcap=0x07000c16, status=0x8800004000310e0f, misc=0x1df87b000d9eff, walltime=0x6546cdc5, cpuid=0x000306f2, bank=0x00000005
158858 2023-11-04 18:03:35 -0500 error: Rx detected CRC error - successful LLR wihout Phy re-init, mcg mcgstatus=0, mci Error_overflow Corrected_error, mcgcap=0x07000c16, status=0xc800008000310e0f, misc=0x1df87b000d9eff, walltime=0x6546cdc7, cpuid=0x000306f2, bank=0x00000005
158859 2023-11-04 18:03:36 -0500 error: Rx detected CRC error - successful LLR wihout Phy re-init, mcg mcgstatus=0, mci Corrected_error, mcgcap=0x07000c16, status=0x8800004000310e0f, misc=0x1df87b000d9eff, walltime=0x6546cdc8, cpuid=0x000306f2, bank=0x00000005
158860 2023-11-04 18:03:37 -0500 error: Rx detected CRC error - successful LLR wihout Phy re-init, mcg mcgstatus=0, mci Corrected_error, mcgcap=0x07000c16, status=0x8800004000310e0f, misc=0x1cf87b000d9eff, walltime=0x6546cdc9, cpuid=0x000306f2, bank=0x00000005
158861 2023-11-04 18:03:39 -0500 error: Rx detected CRC error - successful LLR wihout Phy re-init, mcg mcgstatus=0, mci Corrected_error, mcgcap=0x07000c16, status=0x8800004000310e0f, misc=0x1de87b000d9eff, walltime=0x6546cdcb, cpuid=0x000306f2, bank=0x00000005
158862 2023-11-04 18:03:42 -0500 error: Rx detected CRC error - successful LLR wihout Phy re-init, mcg mcgstatus=0, mci Error_overflow Corrected_error, mcgcap=0x07000c16, status=0xc80000c000310e0f, misc=0x1ff87b000d9eff, walltime=0x6546cdce, cpuid=0x000306f2, bank=0x00000005
158863 2023-11-04 18:03:43 -0500 error: Rx detected CRC error - successful LLR wihout Phy re-init, mcg mcgstatus=0, mci Corrected_error, mcgcap=0x07000c16, status=0x8800004000310e0f, misc=0x1cf87b000d9eff, walltime=0x6546cdcf, cpuid=0x000306f2, bank=0x00000005
158864 2023-11-04 18:03:46 -0500 error: Rx detected CRC error - successful LLR wihout Phy re-init, mcg mcgstatus=0, mci Corrected_error, mcgcap=0x07000c16, status=0x8800004000310e0f, misc=0x1df87b000c9eff, walltime=0x6546cdd2, cpuid=0x000306f2, bank=0x00000005
158865 2023-11-04 18:03:48 -0500 error: Rx detected CRC error - successful LLR wihout Phy re-init, mcg mcgstatus=0, mci Error_overflow Corrected_error, mcgcap=0x07000c16, status=0xc800010000310e0f, misc=0x1df87b000d9eff, walltime=0x6546cdd4, cpuid=0x000306f2, bank=0x00000005

I didn't label the DIMMS, which is why it shows "no DIMM info", but to my understanding the below shows that there are no memory errors:
Code:
root@pve:/# edac-util -v
mc0: 0 Uncorrected Errors with no DIMM info
mc0: 0 Corrected Errors with no DIMM info
mc1: 0 Uncorrected Errors with no DIMM info
mc1: 0 Corrected Errors with no DIMM info
mc2: 0 Uncorrected Errors with no DIMM info
mc2: 0 Corrected Errors with no DIMM info
mc3: 0 Uncorrected Errors with no DIMM info
mc3: 0 Corrected Errors with no DIMM info
edac-util: No errors to report.

Code:
root@pve:/sys/devices/system/edac/mc# tail -n +1 mc*/ce_* mc*/dimm*/dimm_ce_count
==> mc0/ce_count <==
0

==> mc0/ce_noinfo_count <==
0

==> mc1/ce_count <==
0

==> mc1/ce_noinfo_count <==
0

==> mc2/ce_count <==
0

==> mc2/ce_noinfo_count <==
0

==> mc3/ce_count <==
0

==> mc3/ce_noinfo_count <==
0

==> mc0/dimm0/dimm_ce_count <==
0

==> mc0/dimm3/dimm_ce_count <==
0

==> mc1/dimm0/dimm_ce_count <==
0

==> mc1/dimm3/dimm_ce_count <==
0

==> mc2/dimm0/dimm_ce_count <==
0

==> mc2/dimm3/dimm_ce_count <==
0

==> mc3/dimm0/dimm_ce_count <==
0

==> mc3/dimm3/dimm_ce_count <==
0

I have two processors. To isolate whether the issue is with the memory controller on one of the processors or if it's an actual memory issue, is there a way to tell Proxmox to assign a VM to use a particular CPU?

numactl --hardware
Code:
root@pve:/sys/devices/system/edac/mc# numactl --hardware
available: 2 nodes (0-1)
node 0 cpus: 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53
node 0 size: 128852 MB
node 0 free: 22207 MB
node 1 cpus: 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71
node 1 size: 128957 MB
node 1 free: 15553 MB
node distances:
node   0   1
  0:  10  21
  1:  21  10
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!