Search results

  1. A

    [SOLVED] Any more NUC users with unusable watchdog here?

    Yes, but I have a HA cluser ... so they should simply move then and the machine should reboot ... :-) So the question is how fencing would work when only softdog is used? If fencing and moving vms/cts works basically also in this case that it "hangs" then i could work around Yes and no ... I...
  2. A

    [SOLVED] Any more NUC users with unusable watchdog here?

    ;-)) exactly this is the risk :-) it is really bad from Intel that they do not manage to fix the watchdog in the newer devices :-( It's a shame
  3. A

    [SOLVED] Any more NUC users with unusable watchdog here?

    iTCO ... only available ... i know and I also use the softdog at the moment ... but no idea how reliable it is :-)
  4. A

    ext4_multi_mount_protect Delaying Container Migration

    I have the same effect with containers on glusterfs (directory storage on mounted glusterfs mountpoint). I hae this also when starting containers. After some time it fails and gets restarted and so it works in the end, but delayed. And yes with proxmox 6
  5. A

    e1000 driver hang

    Linux is not working that way ;) So yes new Kernel is the way to go there
  6. A

    e1000 driver hang

    basically the host was running without reboot since I entered the command ... How and where I would need to add it? Could you please tell me that and I will try it.
  7. A

    [SOLVED] Any more NUC users with unusable watchdog here?

    hey and sorry for the "a bit off topic" post here. I use Proxmox successfully and were on NUC5PPYH so far and happy. My HA setup is working great. But NUC5PPYH is limited to 8GB RAM and in fact EOL so I decided to start upgrading and so I ended up on NUC8i5BEH2 ... but here the watchdog is not...
  8. A

    e1000 driver hang

    Sure. 00:1f.6 Ethernet controller [0200]: Intel Corporation Ethernet Connection (6) I219-V [8086:15be] (rev 30)
  9. A

    e1000 driver hang

    I had such a "hang" again today ... so also "ethtool -K eno1 tso off gso off " did not helped with me :-( kernel 5.3.10-1-pve
  10. A

    e1000 driver hang

    I'm now on 5.3.10 kernel too with pve 6.1 ... before that I had again such a case with not only the messages but also a "ethernet restart" ... lets see if it is different now
  11. A

    e1000 driver hang

    The bugs seems fixed in 5.2.2 ... but pve6 is on 5.0.x ... maybe the Proxmox guys could patch it themself in their kernel version? Maybe open an issue in their Bugtracker?
  12. A

    e1000 driver hang

    puuhhh ... but better then a crash I think
  13. A

    e1000 driver hang

    I also run into this once with my NUC8i5BEH :-( How much the speed degraded?
  14. A

    PVE with glusterfs as shored storage and "Filesystem gets read only"/On Boot

    I've already upgraded 3 of my 7 nodes to PVE6 ... rest to come the next days ... then I could check it again maybe ...
  15. A

    PVE with glusterfs as shored storage and "Filesystem gets read only"/On Boot

    PS: My problem was that after getting the i/o errors a stop of the KVM did not succeeded
  16. A

    PVE with glusterfs as shored storage and "Filesystem gets read only"/On Boot

    Could it be that the updated Qemu, kvm and such in PVE 6 reacts better to such cases as the ones included in PVE 5? root@pm1:~# gluster volume info Volume Name: gv0 Type: Distributed-Replicate Volume ID: 64651501-6df2-4106-b330-fdb3e1fbcdf4 Status: Started Snapshot Count: 0 Number of Bricks...
  17. A

    PVE with glusterfs as shored storage and "Filesystem gets read only"/On Boot

    Sure: Make sure the glusterfs looses client-quorum :-) So when you have 3 node cluster (setup as replica 2 or such) and then turn off two machines ... then it looses client quorum and everything should get blocked. Then (in my cases because it were just watchdog reboots but bad timing) the...
  18. A

    PVE with glusterfs as shored storage and "Filesystem gets read only"/On Boot

    Yes correct. I tried "Stop" ... also "Reset" after some time ... but the process was not ended till I manually killed it (or 20-30 mins was not enough time) And BTW: I use HA feature, so HA manager is involved too on stopping the vm
  19. A

    [SOLVED] PVE 5.4-11 + Corosync 3.x: major issues

    So I have 24h of stability so far ... and only 3 cases of "Token Retransmit list" cases the wole day with the new version and settings (was much more with old config and 2.x)
  20. A

    [SOLVED] PVE 5.4-11 + Corosync 3.x: major issues

    So, I also upgraded again to corosync3 on my pve5 system ... if it stays stable till sunday I will upgrade the first host to pve6 :-)