NIC drops and comes back multiple times per day

deanfourie

Member
Jan 28, 2023
58
1
13
My NIC keeps dropping, then coming back online. This can happen up to 5 times a day and all VMs as a result loose connectivity briefly. Is there a way I can resolve this or even begin to troubleshoot this?

Not 100% sure where to start with drilling down into this one.

Thank you.

Code:
lspci
0:1f.6 Ethernet controller: Intel Corporation Ethernet Connection (7) I219-LM (rev 10)

Code:
Jan  3 14:12:41 vm pveproxy[1155]: worker 804284 started
Jan  3 14:16:04 vm kernel: [174327.159215] e1000e 0000:00:1f.6 eno1: Detected Hardware Unit Hang:
Jan  3 14:16:04 vm kernel: [174327.159215]   TDH                  <a9>
Jan  3 14:16:04 vm kernel: [174327.159215]   TDT                  <e1>
Jan  3 14:16:04 vm kernel: [174327.159215]   next_to_use          <e1>
Jan  3 14:16:04 vm kernel: [174327.159215]   next_to_clean        <a8>
Jan  3 14:16:04 vm kernel: [174327.159215] buffer_info[next_to_clean]:
Jan  3 14:16:04 vm kernel: [174327.159215]   time_stamp           <10297d9e9>
Jan  3 14:16:04 vm kernel: [174327.159215]   next_to_watch        <a9>
Jan  3 14:16:04 vm kernel: [174327.159215]   jiffies              <10297dc00>
Jan  3 14:16:04 vm kernel: [174327.159215]   next_to_watch.status <0>
Jan  3 14:16:04 vm kernel: [174327.159215] MAC Status             <80083>
Jan  3 14:16:04 vm kernel: [174327.159215] PHY Status             <796d>
Jan  3 14:16:04 vm kernel: [174327.159215] PHY 1000BASE-T Status  <3800>
Jan  3 14:16:04 vm kernel: [174327.159215] PHY Extended Status    <3000>
Jan  3 14:16:04 vm kernel: [174327.159215] PCI Status             <10>
Jan  3 14:16:06 vm kernel: [174329.143039] e1000e 0000:00:1f.6 eno1: Detected Hardware Unit Hang:
Jan  3 14:16:06 vm kernel: [174329.143039]   TDH                  <a9>
Jan  3 14:16:06 vm kernel: [174329.143039]   TDT                  <e1>
Jan  3 14:16:06 vm kernel: [174329.143039]   next_to_use          <e1>
Jan  3 14:16:06 vm kernel: [174329.143039]   next_to_clean        <a8>
Jan  3 14:16:06 vm kernel: [174329.143039] buffer_info[next_to_clean]:
Jan  3 14:16:06 vm kernel: [174329.143039]   time_stamp           <10297d9e9>
Jan  3 14:16:06 vm kernel: [174329.143039]   next_to_watch        <a9>
Jan  3 14:16:06 vm kernel: [174329.143039]   jiffies              <10297ddf0>
Jan  3 14:16:06 vm kernel: [174329.143039]   next_to_watch.status <0>
Jan  3 14:16:06 vm kernel: [174329.143039] MAC Status             <80083>
Jan  3 14:16:06 vm kernel: [174329.143039] PHY Status             <796d>
Jan  3 14:16:06 vm kernel: [174329.143039] PHY 1000BASE-T Status  <3800>
Jan  3 14:16:06 vm kernel: [174329.143039] PHY Extended Status    <3000>
Jan  3 14:16:06 vm kernel: [174329.143039] PCI Status             <10>
Jan  3 14:16:08 vm kernel: [174331.159329] e1000e 0000:00:1f.6 eno1: Detected Hardware Unit Hang:
Jan  3 14:16:08 vm kernel: [174331.159329]   TDH                  <a9>
Jan  3 14:16:08 vm kernel: [174331.159329]   TDT                  <e1>
Jan  3 14:16:08 vm kernel: [174331.159329]   next_to_use          <e1>
Jan  3 14:16:08 vm kernel: [174331.159329]   next_to_clean        <a8>
Jan  3 14:16:08 vm kernel: [174331.159329] buffer_info[next_to_clean]:
Jan  3 14:16:08 vm kernel: [174331.159329]   time_stamp           <10297d9e9>
Jan  3 14:16:08 vm kernel: [174331.159329]   next_to_watch        <a9>
Jan  3 14:16:08 vm kernel: [174331.159329]   jiffies              <10297dfe8>
Jan  3 14:16:08 vm kernel: [174331.159329]   next_to_watch.status <0>
Jan  3 14:16:08 vm kernel: [174331.159329] MAC Status             <80083>
Jan  3 14:16:08 vm kernel: [174331.159329] PHY Status             <796d>
Jan  3 14:16:08 vm kernel: [174331.159329] PHY 1000BASE-T Status  <3800>
Jan  3 14:16:08 vm kernel: [174331.159329] PHY Extended Status    <3000>
Jan  3 14:16:08 vm kernel: [174331.159329] PCI Status             <10>
Jan  3 14:16:10 vm kernel: [174333.143097] e1000e 0000:00:1f.6 eno1: Detected Hardware Unit Hang:
Jan  3 14:16:10 vm kernel: [174333.143097]   TDH                  <a9>
Jan  3 14:16:10 vm kernel: [174333.143097]   TDT                  <e1>
Jan  3 14:16:10 vm kernel: [174333.143097]   next_to_use          <e1>
Jan  3 14:16:10 vm kernel: [174333.143097]   next_to_clean        <a8>
Jan  3 14:16:10 vm kernel: [174333.143097] buffer_info[next_to_clean]:
Jan  3 14:16:10 vm kernel: [174333.143097]   time_stamp           <10297d9e9>
Jan  3 14:16:10 vm kernel: [174333.143097]   next_to_watch        <a9>
Jan  3 14:16:10 vm kernel: [174333.143097]   jiffies              <10297e1d8>
Jan  3 14:16:10 vm kernel: [174333.143097]   next_to_watch.status <0>
Jan  3 14:16:10 vm kernel: [174333.143097] MAC Status             <80083>
Jan  3 14:16:10 vm kernel: [174333.143097] PHY Status             <796d>
Jan  3 14:16:10 vm kernel: [174333.143097] PHY 1000BASE-T Status  <3800>
Jan  3 14:16:10 vm kernel: [174333.143097] PHY Extended Status    <3000>
Jan  3 14:16:10 vm kernel: [174333.143097] PCI Status             <10>
Jan  3 14:16:11 vm kernel: [174333.302961] e1000e 0000:00:1f.6 eno1: Reset adapter unexpectedly
Jan  3 14:16:11 vm kernel: [174333.392649] vmbr0: port 1(eno1) entered disabled state
Jan  3 14:16:12 vm kernel: [174334.422921] vmbr3: port 1(vmbr0.101) entered disabled state
Jan  3 14:16:12 vm kernel: [174334.422967] vmbr1: port 1(vmbr0.102) entered disabled state
Jan  3 14:16:12 vm kernel: [174334.422997] vmbr2: port 1(vmbr0.103) entered disabled state
Jan  3 14:16:14 vm kernel: [174337.125734] e1000e 0000:00:1f.6 eno1: NIC Link is Up 1000 Mbps Full Duplex, Flow Control: None
Jan  3 14:16:14 vm kernel: [174337.125789] vmbr0: port 1(eno1) entered blocking state
Jan  3 14:16:14 vm kernel: [174337.125790] vmbr0: port 1(eno1) entered forwarding state
Jan  3 14:16:14 vm kernel: [174337.125914] vmbr3: port 1(vmbr0.101) entered blocking state
Jan  3 14:16:14 vm kernel: [174337.125917] vmbr3: port 1(vmbr0.101) entered forwarding state
Jan  3 14:16:14 vm kernel: [174337.125951] vmbr1: port 1(vmbr0.102) entered blocking state
Jan  3 14:16:14 vm kernel: [174337.125952] vmbr1: port 1(vmbr0.102) entered forwarding state
Jan  3 14:16:14 vm kernel: [174337.125969] vmbr2: port 1(vmbr0.103) entered blocking state
Jan  3 14:16:14 vm kernel: [174337.125970] vmbr2: port 1(vmbr0.103) entered forwarding state
 
I had the same problem with an intelNuc-8.
Inserting this (post-up) line solved the problem.

nano /etc/network/interfaces

Code:
iface eno1 inet manual
        post-up /sbin/ethtool -K eno1 tso off gso off

Name of the network card (eno1) is just an example from my configuration.
 
I had the same problem with an intelNuc-8.
Inserting this (post-up) line solved the problem.

nano /etc/network/interfaces

Code:
iface eno1 inet manual
        post-up /sbin/ethtool -K eno1 tso off gso off

Name of the network card (eno1) is just an example from my configuration.
Great thanks! I did read the post not long after posting.

Can you please explain what the post-up part does. All I did was run the command


Code:
ethtool -K eno1 tso off gso off

Thanks
 
The command must be entered after every restart.

The entry in /etc/network/interfaces, executes the command automatically after every network restart (including reboots).