pve-kernel-4.15.18-1-pve + IXGBE driver

Discussion in 'Proxmox VE: Installation and configuration' started by udo, Jul 16, 2018.

  1. udo

    udo Well-Known Member
    Proxmox Subscriber

    Joined:
    Apr 22, 2009
    Messages:
    5,835
    Likes Received:
    159
    Hi,
    I have installed last week two new nodes on an cluster with the kernel 4.15.17-3-pve.

    The ixgbe driver is in use.
    This evening I saw the update to 4.15.18-1 with the note "drop out-of-tree IXGBE driver".

    Must I worry about the nodes, which running the 4.15.17-3 kernel? The hosts are in production and not easy rebootable.

    Udo
     
  2. Symbol

    Symbol Member
    Proxmox Subscriber

    Joined:
    Mar 1, 2017
    Messages:
    42
    Likes Received:
    4
    Updated yesterday (while upgrading from 4x1G to 2x10G LACP LAG) with 4.15.18-1. Running so far ; you can have a look at the logs below...

    Code:
    [    2.022749] ixgbe: Intel(R) 10 Gigabit PCI Express Network Driver - version 5.1.0-k
    [    2.022750] ixgbe: Copyright (c) 1999-2016 Intel Corporation.
    [    2.188628] ixgbe 0000:04:00.0: Multiqueue Enabled: Rx Queue count = 12, Tx Queue count = 12 XDP Queue count = 0
    [    2.188757] ixgbe 0000:04:00.0: PCI Express bandwidth of 32GT/s available
    [    2.188759] ixgbe 0000:04:00.0: (Speed:5.0GT/s, Width: x8, Encoding Loss:20%)
    [    2.189084] ixgbe 0000:04:00.0: MAC: 2, PHY: 20, SFP+: 6, PBA No: G42955-016
    [    2.189088] ixgbe 0000:04:00.0: 48:df:37:xx:xx:xx
    [    2.327448] ixgbe 0000:04:00.0: Intel(R) 10 Gigabit Network Connection
    [    2.484734] ixgbe 0000:04:00.1: Multiqueue Enabled: Rx Queue count = 12, Tx Queue count = 12 XDP Queue count = 0
    [    2.484861] ixgbe 0000:04:00.1: PCI Express bandwidth of 32GT/s available
    [    2.484863] ixgbe 0000:04:00.1: (Speed:5.0GT/s, Width: x8, Encoding Loss:20%)
    [    2.485188] ixgbe 0000:04:00.1: MAC: 2, PHY: 20, SFP+: 5, PBA No: G42955-016
    [    2.485191] ixgbe 0000:04:00.1: 48:df:37:xx:xx:xx
    [    2.488698] ixgbe 0000:04:00.1: Intel(R) 10 Gigabit Network Connection
    [    2.492251] ixgbe 0000:04:00.1 eno50: renamed from eth1
    [    2.512244] ixgbe 0000:04:00.0 eno49: renamed from eth0
    [  174.798152] ixgbe 0000:04:00.0: registered PHC device on eno49
    [  174.976071] ixgbe 0000:04:00.0 eno49: detected SFP+: 6
    [  175.060409] ixgbe 0000:04:00.1: registered PHC device on eno50
    [  175.244087] ixgbe 0000:04:00.0 eno49: NIC Link is Up 10 Gbps, Flow Control: RX/TX
    [  175.244974] Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)
    [  175.264976] ixgbe 0000:04:00.0: removed PHC on eno49
    [  175.312061] ixgbe 0000:04:00.1 eno50: detected SFP+: 5
    [  175.350471] 8021q: adding VLAN 0 to HW filter on device bond0
    [  175.402692] ixgbe 0000:04:00.0: registered PHC device on eno49
    [  175.509753] bond0: Enslaving eno49 as a backup interface with a down link
    [  175.512948] ixgbe 0000:04:00.1: removed PHC on eno50
    [  175.643109] vmbr0: port 1(bond0.1071) entered blocking state
    [  175.643112] vmbr0: port 1(bond0.1071) entered disabled state
    [  175.643211] device bond0.1071 entered promiscuous mode
    [  175.644035] ixgbe 0000:04:00.0 eno49: detected SFP+: 6
    [  175.696715] ixgbe 0000:04:00.1: registered PHC device on eno50
    [  175.805626] bond0: Enslaving eno50 as a backup interface with a down link
    [  175.807512] device bond0 entered promiscuous mode
    [  175.908081] ixgbe 0000:04:00.0 eno49: NIC Link is Up 10 Gbps, Flow Control: RX/TX
    [  175.976059] ixgbe 0000:04:00.1 eno50: detected SFP+: 5
    [  176.000076] bond0: Warning: No 802.3ad response from the link partner for any adapters in the bond
    [  176.000126] bond0: link status definitely up for interface eno49, 10000 Mbps full duplex
    [  176.000129] bond0: first active interface up!
    [  176.000200] vmbr0: port 1(bond0.1071) entered blocking state
    [  176.000203] vmbr0: port 1(bond0.1071) entered forwarding state
    [  176.026909] vmbr100: port 1(bond0.100) entered blocking state
    [  176.026912] vmbr100: port 1(bond0.100) entered disabled state
    [  176.026988] device bond0.100 entered promiscuous mode
    [  176.036694] vmbr100: port 1(bond0.100) entered blocking state
    [  176.036697] vmbr100: port 1(bond0.100) entered forwarding state
    [  176.240169] ixgbe 0000:04:00.1 eno50: NIC Link is Up 10 Gbps, Flow Control: RX/TX
    [  176.320108] bond0: link status definitely up for interface eno50, 10000 Mbps full duplex
     
  3. udo

    udo Well-Known Member
    Proxmox Subscriber

    Joined:
    Apr 22, 2009
    Messages:
    5,835
    Likes Received:
    159
    Hi,
    thanks for the input, but I worry more about the 4.15.17-3 Kernel.

    Such a short lifetime in the enterprise repo looks not as best sign...

    Udo
     
  4. Talion

    Talion New Member

    Joined:
    Jun 19, 2018
    Messages:
    15
    Likes Received:
    0
    Indeed, you may worry. I updated and had to boot via 4.15.17-3-pve again.

    I notice this but still ixgbe not working.

    root@pmxn3:~# update-initramfs -u -v | grep ixgbe
    Adding module /lib/modules/4.15.18-1-pve/kernel/drivers/net/ethernet/intel/ixgbevf/ixgbevf.ko
    Adding module /lib/modules/4.15.18-1-pve/kernel/drivers/net/ethernet/intel/ixgbe/ixgbe.ko
    Adding config /etc/modprobe.d/ixgbe.conf

    Best Regards,

    Talion

     
  5. juniper

    juniper Member

    Joined:
    Oct 21, 2013
    Messages:
    50
    Likes Received:
    0
    with 4.15.18-1 doesn't works ixgbe? It's a big big problem...
     
  6. udo

    udo Well-Known Member
    Proxmox Subscriber

    Joined:
    Apr 22, 2009
    Messages:
    5,835
    Likes Received:
    159
    Hi,
    ixgbe work for me on 4.15.18-1-pve and 4.15.17-3-pve without issues til now.

    I use "Intel Corporation Ethernet 10G 2P X520 Adapter" with actual firmware (from dell in this case).

    Udo
     
  1. This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
    By continuing to use this site, you are consenting to our use of cookies.
    Dismiss Notice