SDN VLAN Zone with VNet Error

Mar 27, 2023
2
1
3
I am having trouble configuring a VLAN VNet for my virtual machines using Proxmox SDN. After adding a zone and vnet I get the following error on some hosts: "error vmbr0"

For example, the VMNet has the error on this host:
1742329749693.png

But it succeeds on this host:
1742330826546.png

All my hosts have the referenced vmbr0 linux bridge configured. For example, the host with the error has this configuration:
1742330285391.png

The configuration is not any different from a host where the vnet is successfully configured:
1742330947718.png

I have created a zone called HomeNet that uses the vmbr0 bridge:
1742329866169.png

Within that zone I created a vnet with a vlan tag:
1742330756682.png


I don't know what's causing the error and I can't find any error logs that would point me in the right direction. Any help would be appreciated.
 

Attachments

  • 1742329995135.png
    1742329995135.png
    42.1 KB · Views: 3
  • Like
Reactions: akiuni
Hello ArmorerJP

I have exactly the same problem here (with PVE 8.3.5)
The Apply succeed on all nodes excepted one with same error (error vmbr0). configurations are the same.

Unfortunately, I don't have the solution but I have a strong clue.

All my nodes have keepalived installed to share a virtual ip address with VRRP. This virtual ip is switching to other nodes when the SDN apply occures because it restarts the networking layer and the "faulty" node (the one with the zone in error) is always the one which got the virtual ip address.

To confirm that, I've tried to stop keepalived on all nodes excepted one and this latter always becomes the faulty node, whatever the node I choose.

So I think that this looks like a conflict with another netwoking tool (keepalived in my case).

Could you confirm you have a similar situation ?
 
  • Like
Reactions: ArmoredJP
Thank you akiuni, that is indeed the case! I do have keepalived running on all the hosts as they share a couple of ip addresses for ceph ingress. Unfortunately I can't remove the virtual ip addresses since they are in use by ceph :(.

Seems like a bug so maybe proxmox can fix it?
 
Hello,

Fyi I've bypassed this error with the following patch on all my proxmox nodes.
This is not really a good solution because it may have border effectes but it's working.

Code:
--- /usr/share/ifupdown2/lib/nlcache.py-dist     2025-04-01 09:34:30.411085540 +0200
+++ /usr/share/ifupdown2/lib/nlcache.py  2025-04-01 09:38:41.727038475 +0200
@@ -1949,8 +1949,8 @@
             if addr in config_addrs:
                 managed_addresses.append(addr)
 
-            elif not addr.scope & Route.RT_SCOPE_LINK:
-                managed_addresses.append(addr)
+            # elif not addr.scope & Route.RT_SCOPE_LINK:
+            #     managed_addresses.append(addr)
 
         return managed_addresses