Search results

  1. PVE 5.4-11 + Corosync 3.x: major issues

    seem related to a corosync crash do you have a generated /var/lib/corosync/fdata* on this node ? if yes, could you send it to bugzilla : https://bugzilla.proxmox.com/show_bug.cgi?id=2326
  2. PVE 5.4-11 + Corosync 3.x: major issues

    Hi, I'm able to reproduce the link down and some bad other behaviour, when the rx (ingress traffic) is saturated. (3 nodes cluster, launch an iperf from node2 to node3, node3 is link down from node1 and I had once a corosync segfault). This is with a gigabit link, when I'm reaching around...
  3. PVE 5.4-11 + Corosync 3.x: major issues

    if someone want to test, I have build libknet with 2 upstream patchs: https://github.com/kronosnet/kronosnet/commit/f45e4c67902b95bcd212275f5f6081fa31311793.patch https://github.com/kronosnet/kronosnet/commit/0f67ee86745d52d68f376c92e96e1dd6661e9f5d.patch the deb is here (need to install it on...
  4. PVE 5.4-11 + Corosync 3.x: major issues

    looking at knet code, I'm seem the link down message only at 2 places (/libknet/threads_heartbeat.c) static void _handle_check_each { ... if (dst_link->transport_connected == 0) { _link_down(knet_h, dst_host, dst_link); return; } (I think this one is when the node is really...
  5. PVE 5.4-11 + Corosync 3.x: major issues

    maybe this libknet patch could help for a node blocking other node? https://github.com/kronosnet/kronosnet/commit/0f67ee86745d52d68f376c92e96e1dd6661e9f5d
  6. PVE 5.4-11 + Corosync 3.x: major issues

    you do have something in /var/log/kern.log or #dmesg related to nic ? nic model / driver ? do you use ovs or linux bridge ? do you host locally or in a public provider (ovh, heizner,....) (maybe a public googlesheet to centralize all differents user config could help to compare setup ?)
  7. [SOLVED] So is OpenVSwitch bonding just broken on PVE 6? What's going on?

    BTW, about mtu, last openvswitch have a new option "ovs_mtu ..." , I think it's already implemented in debian buster. they are example here: https://metadata.ftp-master.debian.org/changelogs/main/o/openvswitch/testing_openvswitch-switch.README.Debian
  8. proxmox 5.4 - LAN - SPICE

    They are some scripts posted in the forum, do remember where exactly, calling the proxmox api to get the ticket, then auto open virt-viewer.
  9. ZFS send without snapshot?

    no, you need snasphot for zfs send/receive.
  10. [SOLVED] So is OpenVSwitch bonding just broken on PVE 6? What's going on?

    can you look inside /etc/network/if-pre-up.d/openvswitch and /etc/network/if-post-down.d/openvswitch and check if you have ifconfig reference inside them. normally, you should not. (from a clean install of openvswitch package 2.10.0+2018.08.28+git.8ca7c82b7d+ds1-12 on proxmox6) (if you have...
  11. [SOLVED] So is OpenVSwitch bonding just broken on PVE 6? What's going on?

    hi, you should remove "auto ..." for openvswitch interface. (remove "auto vmbr2", and keep "allow-ovs vmbr2")
  12. 3-node Proxmox/Ceph cluster - how to automatically distribute VMs among nodes?

    @tokala Thanks for your project. I'm currently looking to implement this is proxmox (but with perl, as proxmox is perl code). I have fast looked your code, do you manage nodes with différents kind of cpu (frequencies, or number or cores ?). I wonder how to manage that ? (maybe use bogomips are...
  13. [SOLVED] systemctl status networking.service down / network works

    what do you want to achieve with your 4 network interfaces ? aggregate them to have more bandwith ? if yes, you need to do a bond. about vlan, do your hardware switch force a specific vlan for all servers ? (on cisco, this is "access vlan"). in this case, you don't need to setup any vlan on...
  14. PVE 5.4-11 + Corosync 3.x: major issues

    have you already done last libknet update (on both proxmox5 (with corosync3 repo) && proxmox6 nodes ?) you need to edit /etc/pve/corosync.conf, and when you save the file, each node will copy the file locally in /etc/corosync.conf, and then node auto reload. but if you have your cluster...
  15. Cluster daily broken

    no reboot, but restart corosync/pve-cluster on each server. (no impact)
  16. [SOLVED] systemctl status networking.service down / network works

    you can't have multiple default gw in /etc/network/interfaces. it'll try to add for each interface "ip route add default via xxxx", and you can only have 1 route like this in kernel. so it'll fail on other interfaces. it's possible to have multiple ecmp gateway, with manually add "post-up ip...
  17. PVE 5.4-11 + Corosync 3.x: major issues

    @bofh looking at the logs, I'm seeing "Process pause detected for xxx ms, flushing membership messages." in corosync code, this code from https://github.com/corosync/corosync/blob/master/exec/totemsrp.c static int pause_flush (struct totemsrp_instance *instance) { ... if ((now_msec -...
  18. PVE 5.4-11 + Corosync 3.x: major issues

    do you have corosync log ? # pveversion -v ?
  19. Cluster daily broken

    can you try to do last updates (mainly for libknet1_1.11-pve1)
  20. PVE 5.4-11 + Corosync 3.x: major issues

    yes, indeed, running since 6months with corosync3 beta (on proxmox 5 with kernel 4.15). no problem until now. (I'm using mellanox connect-x4 card, and 2x10gb lacp bonding). Cluster have 16nodes, without any special tuning of corosync configuration.

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE and Proxmox Mail Gateway. We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get your own in 60 seconds.

Buy now!