kernel: Detected Hardware Unit Hang

saeft_3004

Active Member
Mar 4, 2019
50
1
28
36
Hallo,

ich habe das Problem das meine iobroker VM gelegentlich die Verbindung verliert. Immer wenn das passiert sind sehr viele kernel Einträge im syslog von proxmox.

Leider verstehe ich diese nicht wirklich, ich vermute ein Treiber Problem. Ist das bekannt, oder gibt es sogar ein workaround?

Code:
Dec  3 16:31:45 pve systemd[1]: user@0.service: Succeeded.
Dec  3 16:31:45 pve systemd[1]: Stopped User Manager for UID 0.
Dec  3 16:31:45 pve systemd[1]: Stopping User Runtime Directory /run/user/0...
Dec  3 16:31:45 pve systemd[1]: run-user-0.mount: Succeeded.
Dec  3 16:31:45 pve systemd[1]: user-runtime-dir@0.service: Succeeded.
Dec  3 16:31:45 pve systemd[1]: Stopped User Runtime Directory /run/user/0.
Dec  3 16:31:45 pve systemd[1]: Removed slice User Slice of UID 0.
Dec  3 16:32:00 pve systemd[1]: Starting Proxmox VE replication runner...
Dec  3 16:32:01 pve systemd[1]: pvesr.service: Succeeded.
Dec  3 16:32:01 pve systemd[1]: Started Proxmox VE replication runner.
Dec  3 16:32:04 pve kernel: [67860.869955] e1000e 0000:00:1f.6 eno1: Detected Hardware Unit Hang:
Dec  3 16:32:04 pve kernel: [67860.869955]   TDH                  <3>
Dec  3 16:32:04 pve kernel: [67860.869955]   TDT                  <16>
Dec  3 16:32:04 pve kernel: [67860.869955]   next_to_use          <16>
Dec  3 16:32:04 pve kernel: [67860.869955]   next_to_clean        <2>
Dec  3 16:32:04 pve kernel: [67860.869955] buffer_info[next_to_clean]:
Dec  3 16:32:04 pve kernel: [67860.869955]   time_stamp           <10101b77d>
Dec  3 16:32:04 pve kernel: [67860.869955]   next_to_watch        <3>
Dec  3 16:32:04 pve kernel: [67860.869955]   jiffies              <10101ba58>
Dec  3 16:32:04 pve kernel: [67860.869955]   next_to_watch.status <0>
Dec  3 16:32:04 pve kernel: [67860.869955] MAC Status             <40080083>
Dec  3 16:32:04 pve kernel: [67860.869955] PHY Status             <796d>
Dec  3 16:32:04 pve kernel: [67860.869955] PHY 1000BASE-T Status  <3800>
Dec  3 16:32:04 pve kernel: [67860.869955] PHY Extended Status    <3000>
Dec  3 16:32:04 pve kernel: [67860.869955] PCI Status             <10>
Dec  3 16:32:06 pve kernel: [67862.885867] e1000e 0000:00:1f.6 eno1: Detected Hardware Unit Hang:
Dec  3 16:32:06 pve kernel: [67862.885867]   TDH                  <3>
Dec  3 16:32:06 pve kernel: [67862.885867]   TDT                  <16>
Dec  3 16:32:06 pve kernel: [67862.885867]   next_to_use          <16>
Dec  3 16:32:06 pve kernel: [67862.885867]   next_to_clean        <2>
Dec  3 16:32:06 pve kernel: [67862.885867] buffer_info[next_to_clean]:
Dec  3 16:32:06 pve kernel: [67862.885867]   time_stamp           <10101b77d>
Dec  3 16:32:06 pve kernel: [67862.885867]   next_to_watch        <3>
Dec  3 16:32:06 pve kernel: [67862.885867]   jiffies              <10101bc50>
Dec  3 16:32:06 pve kernel: [67862.885867]   next_to_watch.status <0>
Dec  3 16:32:06 pve kernel: [67862.885867] MAC Status             <40080083>
Dec  3 16:32:06 pve kernel: [67862.885867] PHY Status             <796d>
Dec  3 16:32:06 pve kernel: [67862.885867] PHY 1000BASE-T Status  <3800>
Dec  3 16:32:06 pve kernel: [67862.885867] PHY Extended Status    <3000>
Dec  3 16:32:06 pve kernel: [67862.885867] PCI Status             <10>
Dec  3 16:32:08 pve kernel: [67864.873662] ------------[ cut here ]------------
Dec  3 16:32:08 pve kernel: [67864.873665] NETDEV WATCHDOG: eno1 (e1000e): transmit queue 0 timed out
Dec  3 16:32:08 pve kernel: [67864.873679] WARNING: CPU: 0 PID: 0 at net/sched/sch_generic.c:473 dev_watchdog+0x264/0x270
Dec  3 16:32:08 pve kernel: [67864.873680] Modules linked in: dm_snapshot veth rpcsec_gss_krb5 auth_rpcgss nfsv4 nfs lockd grace fscache ebtable_filter ebtables ip_set ip6table_raw iptable_raw ip6table_filter ip6_tables iptable_filter bpfilter softdog nfnetlink_log nfnetlink snd_hda_codec_hdmi snd_hda_codec_realtek snd_hda_codec_generic snd_sof_pci snd_sof_intel_hda_common snd_soc_hdac_hda snd_sof_intel_hda snd_sof_intel_byt snd_sof_intel_ipc intel_rapl_msr snd_sof snd_sof_xtensa_dsp intel_rapl_common x86_pkg_temp_thermal intel_powerclamp snd_hda_ext_core snd_soc_acpi_intel_match snd_soc_acpi iwlmvm mei_hdcp ledtrig_audio coretemp btusb kvm_intel kvm irqbypass i915 mac80211 snd_soc_core btrtl crct10dif_pclmul libarc4 snd_compress btbcm btintel drm_kms_helper crc32_pclmul ac97_bus drm bluetooth i2c_algo_bit iwlwifi cp210x snd_pcm_dmaengine fb_sys_fops syscopyarea ghash_clmulni_intel ecdh_generic aesni_intel snd_hda_intel crypto_simd snd_intel_dspcfg snd_hda_codec snd_hda_core snd_hwdep cryptd snd_pcm glue_helper
Dec  3 16:32:08 pve kernel: [67864.873714]  snd_timer rapl snd mei_me intel_cstate pcspkr wmi_bmof intel_wmi_thunderbolt soundcore mei cdc_acm usbserial ecc cfg80211 sysfillrect sysimgblt mac_hid acpi_tad acpi_pad zfs(PO) zunicode(PO) zzstd(O) zlua(O) zavl(PO) icp(PO) zcommon(PO) znvpair(PO) spl(O) vhost_net vhost tap ib_iser rdma_cm iw_cm ib_cm ib_core iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi sunrpc ip_tables x_tables autofs4 btrfs xor zstd_compress raid6_pq dm_thin_pool dm_persistent_data dm_bio_prison dm_bufio libcrc32c hid_generic usbhid hid sdhci_pci cqhci sdhci e1000e i2c_i801 xhci_pci ahci thunderbolt intel_lpss_pci libahci intel_lpss idma64 xhci_hcd virt_dma wmi video pinctrl_cannonlake pinctrl_intel
Dec  3 16:32:08 pve kernel: [67864.873733] CPU: 0 PID: 0 Comm: swapper/0 Tainted: P           O      5.4.151-1-pve #1
Dec  3 16:32:08 pve kernel: [67864.873734] Hardware name: Intel(R) Client Systems NUC10i3FNH/NUC10i3FNB, BIOS FNCML357.0047.2020.1118.1629 11/18/2020
Dec  3 16:32:08 pve kernel: [67864.873735] RIP: 0010:dev_watchdog+0x264/0x270
Dec  3 16:32:08 pve kernel: [67864.873737] Code: 48 85 c0 75 e6 eb a0 4c 89 ef c6 05 22 86 ef 00 01 e8 60 b7 fa ff 89 d9 4c 89 ee 48 c7 c7 40 72 23 98 48 89 c2 e8 70 65 15 00 <0f> 0b eb 82 0f 1f 84 00 00 00 00 00 0f 1f 44 00 00 55 48 89 e5 41
Dec  3 16:32:08 pve kernel: [67864.873737] RSP: 0018:ffffa13580003e58 EFLAGS: 00010282
Dec  3 16:32:08 pve kernel: [67864.873738] RAX: 0000000000000000 RBX: 0000000000000000 RCX: 0000000000000006
Dec  3 16:32:08 pve kernel: [67864.873738] RDX: 0000000000000007 RSI: 0000000000000082 RDI: ffff91abf0c178c0
Dec  3 16:32:08 pve kernel: [67864.873739] RBP: ffffa13580003e88 R08: 000000000000047b R09: 0000000000000004
Dec  3 16:32:08 pve kernel: [67864.873739] R10: 0000000000000000 R11: 0000000000000001 R12: 0000000000000001
Dec  3 16:32:08 pve kernel: [67864.873739] R13: ffff91abe17d4000 R14: ffff91abe17d4480 R15: ffff91abe1787c80
Dec  3 16:32:08 pve kernel: [67864.873740] FS:  0000000000000000(0000) GS:ffff91abf0c00000(0000) knlGS:0000000000000000
Dec  3 16:32:08 pve kernel: [67864.873740] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
Dec  3 16:32:08 pve kernel: [67864.873741] CR2: 00007ffccc3fdb88 CR3: 000000002160a006 CR4: 00000000003626f0
Dec  3 16:32:08 pve kernel: [67864.873741] Call Trace:
Dec  3 16:32:08 pve kernel: [67864.873742]  <IRQ>
Dec  3 16:32:08 pve kernel: [67864.873745]  ? pfifo_fast_enqueue+0x160/0x160
Dec  3 16:32:08 pve kernel: [67864.873747]  call_timer_fn+0x32/0x130
Dec  3 16:32:08 pve kernel: [67864.873748]  run_timer_softirq+0x1a5/0x430
Dec  3 16:32:08 pve kernel: [67864.873749]  ? enqueue_hrtimer+0x3c/0x90
Dec  3 16:32:08 pve kernel: [67864.873750]  ? ktime_get+0x3c/0xa0
Dec  3 16:32:08 pve kernel: [67864.873751]  ? lapic_next_deadline+0x2c/0x40
Dec  3 16:32:08 pve kernel: [67864.873753]  ? clockevents_program_event+0x93/0xf0
Dec  3 16:32:08 pve kernel: [67864.873755]  __do_softirq+0xdc/0x2d4
Dec  3 16:32:08 pve kernel: [67864.873756]  irq_exit+0xa9/0xb0
Dec  3 16:32:08 pve kernel: [67864.873757]  smp_apic_timer_interrupt+0x79/0x130
Dec  3 16:32:08 pve kernel: [67864.873758]  apic_timer_interrupt+0xf/0x20
Dec  3 16:32:08 pve kernel: [67864.873759]  </IRQ>
Dec  3 16:32:08 pve kernel: [67864.873761] RIP: 0010:cpuidle_enter_state+0xbd/0x450
Dec  3 16:32:08 pve kernel: [67864.873761] Code: ff e8 f7 41 88 ff 80 7d c7 00 74 17 9c 58 0f 1f 44 00 00 f6 c4 02 0f 85 63 03 00 00 31 ff e8 da 50 8e ff fb 66 0f 1f 44 00 00 <45> 85 ed 0f 88 8d 02 00 00 49 63 cd 48 8b 75 d0 48 2b 75 c8 48 8d
Dec  3 16:32:08 pve kernel: [67864.873762] RSP: 0018:ffffffff98403de8 EFLAGS: 00000246 ORIG_RAX: ffffffffffffff13
Dec  3 16:32:08 pve kernel: [67864.873762] RAX: ffff91abf0c2ae40 RBX: ffffffff98557da0 RCX: 000000000000001f
Dec  3 16:32:08 pve kernel: [67864.873763] RDX: 00003db905afd05c RSI: 000000003161faed RDI: 0000000000000000
Dec  3 16:32:08 pve kernel: [67864.873763] RBP: ffffffff98403e28 R08: 0000000000000002 R09: 000000000002a6c0
Dec  3 16:32:08 pve kernel: [67864.873764] R10: 0000a004ca71c794 R11: ffff91abf0c29ae0 R12: ffff91abf0c36c00
Dec  3 16:32:08 pve kernel: [67864.873764] R13: 0000000000000004 R14: ffffffff98557f38 R15: ffffffff98557f20
Dec  3 16:32:08 pve kernel: [67864.873766]  ? cpuidle_enter_state+0x99/0x450
Dec  3 16:32:08 pve kernel: [67864.873767]  cpuidle_enter+0x2e/0x40
Dec  3 16:32:08 pve kernel: [67864.873768]  call_cpuidle+0x23/0x40
Dec  3 16:32:08 pve kernel: [67864.873769]  do_idle+0x22c/0x270
Dec  3 16:32:08 pve kernel: [67864.873770]  cpu_startup_entry+0x1d/0x20
Dec  3 16:32:08 pve kernel: [67864.873772]  rest_init+0xae/0xb0
Dec  3 16:32:08 pve kernel: [67864.873773]  arch_call_rest_init+0xe/0x1b
Dec  3 16:32:08 pve kernel: [67864.873774]  start_kernel+0x54c/0x56e
Dec  3 16:32:08 pve kernel: [67864.873775]  x86_64_start_reservations+0x24/0x26
Dec  3 16:32:08 pve kernel: [67864.873776]  x86_64_start_kernel+0x74/0x77
Dec  3 16:32:08 pve kernel: [67864.873777]  secondary_startup_64+0xa4/0xb0
Dec  3 16:32:08 pve kernel: [67864.873778] ---[ end trace 32f771dd5fd2b1f8 ]---
Dec  3 16:32:08 pve kernel: [67864.873793] e1000e 0000:00:1f.6 eno1: Reset adapter unexpectedly
Dec  3 16:32:08 pve kernel: [67864.963212] vmbr0: port 1(eno1) entered disabled state
Dec  3 16:32:12 pve kernel: [67868.620238] e1000e: eno1 NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None
Dec  3 16:32:12 pve kernel: [67868.620335] vmbr0: port 1(eno1) entered blocking state
Dec  3 16:32:12 pve kernel: [67868.620338] vmbr0: port 1(eno1) entered forwarding state
Dec  3 16:32:19 pve pvestatd[1010]: status update time (7.459 seconds)
Dec  3 16:32:35 pve systemd[1]: Created slice User Slice of UID 0.
Dec  3 16:32:35 pve systemd[1]: Starting User Runtime Directory /run/user/0...
Dec  3 16:32:35 pve systemd[1]: Started User Runtime Directory /run/user/0.
Dec  3 16:32:35 pve systemd[1]: Starting User Manager for UID 0...
Dec  3 16:32:35 pve systemd[25320]: Listening on GnuPG cryptographic agent and passphrase cache.
Dec  3 16:32:35 pve systemd[25320]: Listening on GnuPG cryptographic agent and passphrase cache (access for web browsers).
Dec  3 16:32:35 pve systemd[25320]: Listening on GnuPG network certificate management daemon.
Dec  3 16:32:35 pve systemd[25320]: Listening on GnuPG cryptographic agent (ssh-agent emulation).
Dec  3 16:32:35 pve systemd[25320]: Listening on GnuPG cryptographic agent and passphrase cache (restricted).
Dec  3 16:32:35 pve systemd[25320]: Reached target Sockets.
Dec  3 16:32:35 pve systemd[25320]: Reached target Paths.
Dec  3 16:32:35 pve systemd[25320]: Reached target Timers.
Dec  3 16:32:35 pve systemd[25320]: Reached target Basic System.
Dec  3 16:32:35 pve systemd[25320]: Reached target Default.
Dec  3 16:32:35 pve systemd[25320]: Startup finished in 31ms.
Dec  3 16:32:35 pve systemd[1]: Started User Manager for UID 0.
 
Hi,

schau mal hier: https://forum.proxmox.com/threads/detected-hardware-unit-hang.100534/

Das sieht nach dem gleichen Fehler aus.

Die Einträge kannst du in die /etc/network/interfaces mit aufnehmen und damit sind die nach jedem Reboot auch wieder vorhanden.
Muss das unter vmbr0 oder unter eno1

auto lo
iface lo inet loopback

iface eno1 inet manual
offload-gso off
offload-gro off
offload-tso off
offload-rx off
offload-tx off
offload-rxvlan off
offload-txvlan off
offload-sg off
offload-ufo off
offload-lro off

auto vmbr0
iface vmbr0 inet static
offload-gso off
offload-gro off
offload-tso off
offload-rx off
offload-tx off
offload-rxvlan off
offload-txvlan off
offload-sg off
offload-ufo off
offload-lro off
address 192.168.178.101/24
gateway 192.168.178.1
bridge_ports eno1
bridge_stp off
bridge_fd 0
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!