LACP 802.3ad / Over than 1 server

timonych

Well-Known Member
Aug 8, 2017
63
20
48
34
Hello,

I have bought Netis ST3310GF and have successfully configured LACP on Mikrotik hEx and 1 Proxmox Server. It works perfectly. After I have bought another NIC with 2 ports and I have found that LACP not working for 2 Server.

1 Server NIC:
Intel Corporation 82575EB Gigabit Network Connection

2 Server NIC
Intel Corporation 82576 Gigabit Network Connection


After some time I have turned LACP on 1 Server. LACP starts work on 2 Server normally. I have found that there is Link Aggregation Group in Partner's Info has similliar value - 9. If i will turn two server simultaneously - one of them will have status of Aggragation - False.
2019-08-20 11_17_10-152. win8pers.png

I have tryed to find solution to force change this Group in Proxmox but no success.

Is it posssible?

P.S.
2 Server mode now - balance-xor that's why there is no info in Partner's view.
 
I do not understand your question?
From quick googling the Netis ST3310GF seems to be a small managed switch (no explicit mentioning of LACP, but Link Aggregation)?
The microtik is a 5-port router
* Where did you configure LACP ? (between router and switch, between switch and server)?
* The screenshot indicates that Aggregation works?!

* balance-xor is not LACP - but a independent bonding mode - LACP is 802.3ad - see https://www.kernel.org/doc/Documentation/networking/bonding.txt

please post some logs of the PVE-node and explain what issue you have and what you want to achieve

Thanks!
 
I do not understand your question?
From quick googling the Netis ST3310GF seems to be a small managed switch (no explicit mentioning of LACP, but Link Aggregation)?
In Netis ST3310GF LACP was added in last Firmware.

Снимок.PNG

I ahve configure LACP between RouterBOARD 750G r3 (MikroTik) and Netis ST3310GF - Have added 7 and 8 port in Trunk. (Trunk Group 3)
Also have configured LACP between Netis ST3310GF and my Server - Have added 3 and 4 port in Trunk. (Trunk Group 2)

I have tryed to configure LACP on 2 Server, but no success because group ID is similliar to 1 Server.

Снимок1.PNG

I need to configure it in Proxmox node, but didn't find where I could do it.

So my 2 Server works only in Balance XOR with Link Aggregation type. (Trunk Group 1)

Снимок2.PNG
 
because group ID is similliar to 1 Server.
no experience with NetIs switches - but it's new to me that you'd need to set an explicit group id on the server side for LACP...

please post the errors of the PVE-node you're seeing and your network configuration (/etc/network/interfaces)
You can see various bonding parameters in /sys/class/net/bondX (where bondX is the name of the bond-interface)


I would also ask NetIs support for the issue...

I hope this helps!
 
Bash:
auto lo
iface lo inet loopback

allow-hotplug enp31s0

auto enp32s0
iface enp32s0 inet manual

auto enp30s0f0
iface enp30s0f0 inet manual

auto enp30s0f1
iface enp30s0f1 inet manual

auto bond0
iface bond0 inet manual
        bond-slaves enp30s0f0 enp30s0f1
        bond-miimon 100
        bond-mode 802.3ad
        bond-xmit-hash-policy layer2+3

auto bond1
iface bond1 inet manual
        bond-slaves enp32s0
        bond-miimon 100
        bond-mode active-backup

auto vmbr0
iface vmbr0 inet static
        address  192.168.1.12
        netmask  24
        gateway  192.168.1.1
        bridge-ports bond0 bond1
        bridge-stp off
        bridge-fd 0

Bash:
Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)

Bonding Mode: IEEE 802.3ad Dynamic link aggregation
Transmit Hash Policy: layer2+3 (2)
MII Status: up
MII Polling Interval (ms): 100
Up Delay (ms): 0
Down Delay (ms): 0

802.3ad info
LACP rate: slow
Min links: 0
Aggregator selection policy (ad_select): stable
System priority: 65535
System MAC address: 00:1b:21:31:0c:ea
Active Aggregator Info:
        Aggregator ID: 1
        Number of ports: 2
        Actor Key: 9
        Partner Key: 1
        Partner Mac Address: 08:10:79:b3:ad:34

Slave Interface: enp30s0f0
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 0
Permanent HW addr: 00:1b:21:31:0c:ea
Slave queue ID: 0
Aggregator ID: 1
Actor Churn State: none
Partner Churn State: churned
Actor Churned Count: 0
Partner Churned Count: 1
details actor lacp pdu:
    system priority: 65535
    system mac address: 00:1b:21:31:0c:ea
    port key: 9
    port priority: 255
    port number: 1
    port state: 13
details partner lacp pdu:
    system priority: 1
    system mac address: 08:10:79:b3:ad:34
    oper key: 1
    port priority: 1
    port number: 1
    port state: 5

Slave Interface: enp30s0f1
MII Status: up
Speed: 1000 Mbps
Duplex: full
Link Failure Count: 0
Permanent HW addr: 00:1b:21:31:0c:eb
Slave queue ID: 0
Aggregator ID: 1
Actor Churn State: none
Partner Churn State: churned
Actor Churned Count: 0
Partner Churned Count: 1
details actor lacp pdu:
    system priority: 65535
    system mac address: 00:1b:21:31:0c:ea
    port key: 9
    port priority: 255
    port number: 2
    port state: 13
details partner lacp pdu:
    system priority: 1
    system mac address: 08:10:79:b3:ad:34
    oper key: 1
    port priority: 1
    port number: 2
    port state: 5

Bash:
[    7.561146] bonding: bond0 is being created...
[    7.934251] bond0: Enslaving enp30s0f0 as a backup interface with a down link
[    8.302225] bond0: Enslaving enp30s0f1 as a backup interface with a down link
[    8.523969] vmbr0: port 1(bond0) entered blocking state
[    8.523971] vmbr0: port 1(bond0) entered disabled state
[    8.524129] device bond0 entered promiscuous mode
[   10.288855] bond0: link status definitely up for interface enp30s0f0, 1000 Mbps full duplex
[   10.288859] bond0: Warning: No 802.3ad response from the link partner for any adapters in the bond
[   10.288865] bond0: first active interface up!
[   10.289162] vmbr0: port 1(bond0) entered blocking state
[   10.289164] vmbr0: port 1(bond0) entered forwarding state
[   10.704876] bond0: link status definitely up for interface enp30s0f1, 1000 Mbps full duplex
 

Attachments

  • 2019-08-29 12_24_03-ST3310GF.png
    2019-08-29 12_24_03-ST3310GF.png
    52.4 KB · Views: 26
  • 2019-08-29 12_25_32-ST3310GF.png
    2019-08-29 12_25_32-ST3310GF.png
    17.9 KB · Views: 23