The Intel X710 LAN connection goes down and up all the time

Jun 10, 2024
2
1
1
I have a server with 2 10G-baset Intel X710 LAN ports and is unable to maintain active connection with the switch. Link go up and down continuously.
With the concomitant transition to the enterprise repository and upgrade to version 8.2 the problem has worsened making the server unusable.

I upgraded the system and network drivers, restarted servers and switches, replaced the network cables with 1 meter cables cat.7, without success.
After a complete shutdown of the server, followed by a reboot, if there are not too many virtual machines active, the network connection remains active for a few minutes without falling.
Could it be a driver? or firmware problem?

My PVE version is pve-manager/8.2.2/9355359cd7afbae4 (running kernel: 6.8.4-3-pve).

Network hardware:
Code:
  *-network:0
       description: Ethernet interface
       product: Ethernet Controller X710 for 10GBASE-T
       vendor: Intel Corporation
       physical id: 0
       bus info: pci@0000:04:00.0
       logical name: eth1
       version: 02
       serial: a0:36:bc:c7:b8:71
       capacity: 10Gbit/s
       width: 64 bits
       clock: 33MHz
       capabilities: pm msi msix pciexpress vpd bus_master cap_list rom ethernet physical tp 100bt-fd 1000bt-fd 10000bt-fd autonegotiation
       configuration: autonegotiation=on broadcast=yes driver=i40e driverversion=6.8.4-3-pve firmware=8.15 0x8000a3e0 1.2829.0 latency=0 link=no multicast=yes port=twisted pair slave=yes
       resources: irq:16 memory:a6000000-a6ffffff memory:a7108000-a710ffff memory:a7080000-a70fffff memory:4000100000-40004fffff memory:4000900000-40009fffff
  *-network:1
       description: Ethernet interface
       product: Ethernet Controller X710 for 10GBASE-T
       vendor: Intel Corporation
       physical id: 0.1
       bus info: pci@0000:04:00.1
       logical name: eth2
       version: 02
       serial: a0:36:bc:c7:b8:71
       capacity: 10Gbit/s
       width: 64 bits
       clock: 33MHz
       capabilities: pm msi msix pciexpress vpd bus_master cap_list rom ethernet physical tp 100bt-fd 1000bt-fd 10000bt-fd autonegotiation
       configuration: autonegotiation=on broadcast=yes driver=i40e driverversion=6.8.4-3-pve firmware=8.15 0x8000a3e0 1.2829.0 latency=0 link=no multicast=yes port=twisted pair slave=yes
       resources: irq:16 memory:a5000000-a5ffffff memory:a7100000-a7107fff memory:a7000000-a707ffff memory:4000500000-40008fffff memory:4000a00000-4000afffff
  *-network
       description: Ethernet interface
       physical id: 6
       bus info: usb@1:8.3
       logical name: eth0
       serial: ba:97:c1:b4:08:06
       capabilities: ethernet physical
       configuration: autonegotiation=off broadcast=yes driver=cdc_ether driverversion=6.8.4-3-pve duplex=half firmware=CDC Ethernet Device ip=10.1.0.21 link=yes multicast=yes port=twisted pair

Network addressing links:
Code:
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group default qlen 1000
    link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
    inet 127.0.0.1/8 scope host lo
       valid_lft forever preferred_lft forever
    inet6 ::1/128 scope host noprefixroute 
       valid_lft forever preferred_lft forever
2: eth1: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq master bond0 state UP group default qlen 1000
    link/ether a0:36:bc:c7:b8:71 brd ff:ff:ff:ff:ff:ff
3: eth2: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc mq master bond0 state UP group default qlen 1000
    link/ether a0:36:bc:c7:b8:71 brd ff:ff:ff:ff:ff:ff permaddr a0:36:bc:c7:b8:72
4: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UNKNOWN group default qlen 1000
    link/ether ba:97:c1:b4:08:06 brd ff:ff:ff:ff:ff:ff
    inet 10.1.0.21/24 scope global eth0
       valid_lft forever preferred_lft forever
    inet6 fe80::b897:c1ff:feb4:806/64 scope link 
       valid_lft forever preferred_lft forever
5: bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue master vmbr0 state UP group default qlen 1000
    link/ether a0:36:bc:c7:b8:71 brd ff:ff:ff:ff:ff:ff
6: vmbr0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default qlen 1000
    link/ether a0:36:bc:c7:b8:71 brd ff:ff:ff:ff:ff:ff
    inet 10.1.0.11/24 scope global vmbr0
       valid_lft forever preferred_lft forever
    inet6 fe80::a236:bcff:fec7:b871/64 scope link 
       valid_lft forever preferred_lft forever
7: tap102i0: <BROADCAST,MULTICAST,PROMISC,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master fwbr102i0 state UNKNOWN group default qlen 1000
    link/ether 4a:05:e8:f5:c4:ea brd ff:ff:ff:ff:ff:ff
8: fwbr102i0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP group default qlen 1000
    link/ether ce:81:2e:e7:e2:75
Network config interfaces:
Code:
# network interface settings; autogenerated
# Please do NOT modify this file directly, unless you know what
# you're doing.
#
# If you want to manage parts of the network configuration manually,
# please utilize the 'source' or 'source-directory' directives to do
# so.
# PVE will preserve these directives, but will NOT read its network
# configuration from sourced files, so do not attempt to move any of
# the PVE managed interfaces into external files!

auto lo
iface lo inet loopback

auto eth0
iface eth0 inet manual
    address 10.1.0.21/24
#base port 1G

iface eth1 inet manual
#port 10G-base-t

iface eth2 inet manual
#port 10G-base-t

auto bond0
iface bond0 inet manual
    bond-slaves eth1 eth2
    bond-miimon 100
    bond-mode balance-rr
#aggregate

auto vmbr0
iface vmbr0 inet static
    address 10.1.0.11/24
    gateway 10.1.0.1
    bridge-ports bond0
    bridge-stp off
    bridge-fd 0
#Main bridge 

source /etc/network/interfaces.d/*

Driver Info lan i40e:
Code:
filename:       /lib/modules/6.8.4-3-pve/kernel/drivers/net/ethernet/intel/i40e/i40e.ko
license:        GPL v2
description:    Intel(R) Ethernet Connection XL710 Network Driver
author:         Intel Corporation, <e1000-devel@lists.sourceforge.net>
srcversion:     576504FF0D0B285FE36E3EB
alias:          pci:v00008086d0000158Bsv*sd*bc*sc*i*
alias:          pci:v00008086d0000158Asv*sd*bc*sc*i*
alias:          pci:v00008086d00000D58sv*sd*bc*sc*i*
alias:          pci:v00008086d00000CF8sv*sd*bc*sc*i*
alias:          pci:v00008086d00001588sv*sd*bc*sc*i*
alias:          pci:v00008086d00001587sv*sd*bc*sc*i*
alias:          pci:v00008086d00000DDAsv*sd*bc*sc*i*
alias:          pci:v00008086d000037D3sv*sd*bc*sc*i*
alias:          pci:v00008086d000037D2sv*sd*bc*sc*i*
alias:          pci:v00008086d000037D1sv*sd*bc*sc*i*
alias:          pci:v00008086d000037D0sv*sd*bc*sc*i*
alias:          pci:v00008086d000037CFsv*sd*bc*sc*i*
alias:          pci:v00008086d000037CEsv*sd*bc*sc*i*
alias:          pci:v00008086d0000104Fsv*sd*bc*sc*i*
alias:          pci:v00008086d0000104Esv*sd*bc*sc*i*
alias:          pci:v00008086d000015FFsv*sd*bc*sc*i*
alias:          pci:v00008086d00001589sv*sd*bc*sc*i*
alias:          pci:v00008086d00001586sv*sd*bc*sc*i*
alias:          pci:v00008086d00000DD2sv*sd*bc*sc*i*
alias:          pci:v00008086d00001585sv*sd*bc*sc*i*
alias:          pci:v00008086d00001584sv*sd*bc*sc*i*
alias:          pci:v00008086d00001583sv*sd*bc*sc*i*
alias:          pci:v00008086d00001581sv*sd*bc*sc*i*
alias:          pci:v00008086d00001580sv*sd*bc*sc*i*
alias:          pci:v00008086d00001574sv*sd*bc*sc*i*
alias:          pci:v00008086d00001572sv*sd*bc*sc*i*
depends:        
retpoline:      Y
intree:         Y
name:           i40e
vermagic:       6.8.4-3-pve SMP preempt mod_unload modversions 
sig_id:         PKCS#7
signer:         Build time autogenerated kernel key
sig_key:        71:06:6D:D6:05:97:A1:46:92:06:37:D1:16:E9:D4:8C:C3:3C:D3:93
sig_hashalgo:   sha512
signature:      81:48:51:E7:52:51:81:6D:69:C2:81:28:5B:B5:B1:59:38:D0:1B:0F:
        72:BC:F6:0C:68:07:DB:F1:A4:C7:37:5F:C7:AE:83:9D:48:C2:E4:F9:
        EE:C0:39:42:7A:D0:08:50:F7:CA:C8:61:09:F0:2F:FD:98:25:CA:F3:
        5A:FA:21:79:74:58:48:00:AA:EE:61:AF:13:73:22:A0:03:EE:A5:74:
        52:A7:E8:54:36:0C:C0:43:CA:20:AB:B9:84:A3:17:F9:3E:A9:A3:6B:
        8C:43:74:40:2B:85:3C:FE:21:D2:46:69:4F:EC:FF:DD:08:12:B4:E5:
        96:20:C9:7F:16:16:18:25:D6:2D:02:D7:34:47:18:05:8A:D2:92:DA:
        7F:3F:FB:F3:4F:67:59:6D:31:01:6A:1F:62:03:80:18:1B:81:5A:83:
        10:0F:CA:A9:8A:91:A3:CC:DF:0D:AF:BF:EC:BB:01:1E:85:9E:F0:55:
        27:D5:68:2C:AF:AD:E8:34:97:83:65:A9:74:FE:51:A3:F6:81:40:8E:
        74:DF:84:9D:3E:BD:F1:34:52:85:20:DC:E6:E4:D4:65:42:48:C7:6B:
        A1:7E:A6:4E:F0:32:6E:BA:95:35:CF:26:FA:84:BC:9E:0A:8A:74:8B:
        9C:14:C0:8D:39:5F:06:D5:0A:F1:2C:04:EA:E6:D1:60:D4:F8:39:13:
        8B:54:E2:E5:6D:D8:65:C3:FD:72:C7:42:ED:3C:8E:2F:A9:C7:D4:DB:
        41:CB:FB:DD:3D:42:62:EA:75:EA:73:47:2E:70:A4:1B:A3:9E:C0:BC:
        37:23:29:71:97:41:2A:11:B7:CE:7A:3F:0D:6B:61:5E:C1:39:A0:D8:
        06:78:3A:9E:D4:C3:64:9A:55:E2:7A:70:A2:04:02:A1:A7:28:0B:B9:
        14:DF:56:4D:F7:F7:69:A0:76:32:CC:24:FF:48:C6:EE:D1:33:92:BC:
        AA:2B:54:94:0B:29:73:A7:EC:00:3A:F8:6F:82:FD:29:E8:20:DD:41:
        DB:11:EF:4A:AF:65:71:1E:17:B9:4C:19:F7:20:67:7F:25:B6:AA:8C:
        C1:01:DE:3B:81:FE:CF:0D:20:48:52:00:F5:6D:79:8F:24:26:7D:51:
        11:67:15:8B:EC:76:77:76:55:7A:9A:01:ED:F3:DA:ED:8A:8E:1C:A2:
        1D:6C:67:6B:6E:51:8B:05:D8:13:8A:05:B5:D2:13:B4:B7:4B:14:22:
        6D:E1:5E:13:7D:07:C9:2A:A4:B5:52:39:A3:08:4B:AB:95:8B:49:6B:
        8E:21:F5:F3:DD:73:4C:68:DD:3E:AD:DC:6E:56:6A:DB:3F:97:73:CB:
        EA:54:E6:C7:DD:7C:17:47:70:38:B6:6F
parm:           debug:Debug level (0=none,...,16=all), Debug mask (0x8XXXXXXX) (uint)

Output of dmesg command is attched.

Tank you in advance.
 

Attachments

  • dmesg-output.txt
    88.6 KB · Views: 3
  • Like
Reactions: mikeone
Possible tests:
- direct connect without switch + iperf test
- use 802.3ad mode and test vs switch or vs direct connected hosts
 
Possible tests:
- direct connect without switch + iperf test
- use 802.3ad mode and test vs switch or vs direct connected hosts
I have already tried using 802.3ad bond mode and also tried a different brand switch with the same results.
I haven’t tried direct connection without concentrators yet.
 
I have the exact same issue, every 1s the link goes down/up. Same NICs, i40 module. Using DAC cables.
 

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!