Hi, I known how to fix it, I'll send a patch soon.
I have 2 possiblity:
- Allow vlan inside the vxlan tunnel. (I need to patch ifupdown2, It's an arbitrary limitation)
- Map vlan to vxlan with new kernel feature. (vlanid 1000 become vxlanid 1000 when going to vxlan interface, and rettaged...
Personnaly, I always upgrade my nodes each month to last version.
(you can upgrade live your nodes without any impact on the running vms/ct. Reboot is only needed if you need to upgrade kernel for security)
You can keep them mixed, but if a vm has been started on a new qemu version, you'll...
I don't think you can change the corosync.conf with no quorum.
at least /etc/pve/corosync.conf , don't touch it.
when you edit /etc/pve/corosync.conf, and change the config_version, proxmox will generate on each node, a new /etc/corosync/corosync.conf + restart of corosync.
so , maybe you...
maybe the flood come from pmxcfs retries, but I think this is because of corosync bug.
could be great to have corosync logs of each nodes (really all the nodes to try to compare and find if it's coming from a specific node) , and if you can enable debug in /etc/pve/corosync.conf it could be...
you'll not be able to boot on virtio if you don't have the driver installed.
this is the same if you use "scsi0 + scsihw: virtio-scsi-pci", (this is virtio-scsi).
Best method, is to boot with ide0 disk, then install virtio drivers.
Thanks for the report!
I'll look inside ifupdown2, I known that they limited vlan inside vxlan, because last kernel version have vlan-to-vxlan mapping.
But technically, it's possible to do vlans inside vxlan tunnel.
I'll keep you in touch today.
Do you have tried to...
you need some spanning tree here to avoid loop.
BTW,I'm currently working on a new sdn feature for proxmox, where you can use vxlan tunnel for this
(and you can add any number of hosts)
It's currenly beta (and it's using...
maybe this could help you :
I'm currently working on a new sdn feature where you can use vxlan for tunneling
(It's currently beta, but I think it could be ready soon, and shouldn't change too much)
Also, if you really able to reproduce soon, you can enable corosync debug log
and increase config_version to applied it
That's a lof of log, so be carefull of disk space. But this could give really some great...
yes, it was the same for me
proxmox6 use corosync3 with libknet protocol, so it's totally different now
I remember to have done some test the traffic shapping (tc ...), it was working fine, but I don't find the sample
I don't think it's a proxmox bug here, but more a corosync bug. (it's more...
What I remember when I had this problem, is that I had some cluster split (you can see it in different host with corosync logs),
then it was like traffic was still send everywhere, and node refuse the packet with "non member" (because it didn't see the other sending memberr in his own cluster)...
mmm, I think I have this once, with 1 corosync process flooded (corosync process was 100% cpu), and I had a lot in rx.
it was 6 months ago with older corosync version, and I never have been able to reproduce it.
what is your corosync && libknet packages versions ? can you send your...