That was it - the pve-cluster had crashed on those two servers! I can't believe I didn't check that!
But, I'm still curious if you might have any thoughts on what might be causing the pvedaemon (or in this case the pve-cluster) to crash on a somewhat regular basis (maybe once a week or so?)
I...
This started appearing before the switch change.
All the nodes show the exact same results for 'pvecm status' (with the exception of which is the local node of course).
I've been having an issue with my cluster for a little while now where a random node would drop out of the GUI, and then would either come back, or if I restarted the pvesr and pvedaemon, it would come back. However, now I have two nodes that don't want to come back, and it's a bit of a...
Unfortunately, I never did find a solution. In good news however, it has appeared to have stopped for me. I did update & reboot all the physical nodes, and it has not come back since then. Are you still experiencing it on a current version, or do you have an older version of PVE that is...
@czechsys, that is interesting, and does look like my problem that has arisen after the upgrade.
@Symbol, that does seem to have fixed it. Thank you for that! ^_^
This is what lsmod says:
Module Size Used by
bonding 147456 0
When it boots, it does have this (regardless of whether it boots successfully or not):
[ 7.325589] Ethernet Channel Bonding Driver: v3.7.1 (April 27, 2011)
And these are all the files under...
It was a dist-upgraded according to the how-to on the wiki, and I can't think of anything that would cause it, I keep rc.local default. I added your line just after the bond0 declaration, and unfortunately there is no difference :-(
I checked dmesg for bond0 messages, and saw this when it does...
Because I didn't have anything else to try, I went ahead & removed all the bond0.x configs, but still get the same problem.
Any other thoughts anyone? :-D
Hi!
I have a small cluster of Proxmox machines, and I am in the process of upgrading them to 5.0 from 4.4. The two that I have converted have this problem of every few reboots, the network simply doesn't work. I can log in via the console and run /etc/init.d/networking restart and that makes...
That's an interesting idea, and I think you might be onto something!
I checked grub on the identical machine with the working grub, and everything is identical (including the incorrect sizes). I ran grub-install /dev/sda, update-grub2, and update-initramfs -u on the working machine several...
Sure!
grub> insmod zfs
grub> ls (hd0)
Device hd0: No known filesystem detected - Sector size 512B - Total size 1362276352KiB
grub> ls (hd0,gpt1)
Partition hd0,gpt1: No known filesystem detected - Partition start at 17KiB - Total size 1007KiB
grub> ls (hd0,gpt2)
Partition...
It means that whatever drive you're looking at is under a RAID controller - if the user selects ZFS on that drive, at the very least you can put up a warning that says "You might not want to install ZFS on a HW raid controller, proceed at your own risk".
Interesting update here: I had left the machine booted up from the live cd for a few days so I could copy all the data off it, and the next time I tried to boot off the drives after running the same commands, grub actually loaded (after sitting for ~60 seconds), but then it fails with:
Loading...
As I mentioned, I was able to boot with the v3.4 and import and scrub the zpool without any errors, but after updating and installing grub, I still get dumped back to the same "error: unknown filesystem" on reboot.
If it is such a liability, why doesn't the Proxmox installer mention anything about that then?
My problem is still with Grub and ZFS, not data loss or performance problems - for all intents and purposes this is no different from trying to boot off a single drive.
Hardware raid controller - an LSI SAS controller, with 8 SAS disks attached in a RAID-6
I had read that ZFS can have issues with trying to boot off a RAID device with more than a few disks, so it seemed prudent to stick with a single virtual disk.
Sorry for the delay - was dealing with some other issues. I just started a new thread for it: https://forum.proxmox.com/threads/crashes-with-zfs-root-and-stuck-on-grub-rescue-prompt.34172/
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.