Hi!
Can you share the specific VM config with the expected and actual behavior?
If there is no given boot config property, then the VM will fallback to the legacy boot order default, which is CD-ROMs, bootable hard disks (i.e. set with the...
Hi!
How did you try to gather the logs? Are there no logs at the end for e.g. journalctl -b <boot-nr>? If this is block-related in any way it might happen, because the kernel panics and cannot write to any storage media anymore. In these cases...
Hi!
Proxmox VE abstracts away paths with storages, so that users can store VM images, backups, etc. on various storage types (e.g. ext4, zfs, LVM, RBD, iSCSI, etc.). There's a help command to show you the correct command line usage of the...
Hi!
Thanks for the input!
Unfortunately, this fails on any device, which does not have the capabilities. We usually only put values in the time series, which are collectable by any system independent from the hardware configuration (e.g...
Hey!
I believe I set it up correctly at least I hope I did. I ran modprobe netconsole netconsole=@192.168.50.70/vmbr0,@192.168.50.133/ on the host and nc -l -u -p 6666 on my main machine, I'll let you know if anything comes up when it freezes next.
Hm, unfortunately it doesn't seem like there's anything valuable in those logs either. I guess that 4ba997d4d49c4298937fd85fb41c866a was the boot which ended in the kernel stack trace above (at least from the timestamps and the abrupt end of...
Hi!
Seems like there was something in the way of writing it to disk, you can try to gather the log on a separate console, e.g. netconsole [0], so that the information persists if there block io is unavailable.
[0]...
It would be helpful for the kernel stacktrace to be a bit longer, e.g. upload a whole boot and crash log (journalctl can also export a log from a whole boot). I can only see that ZFS seems involved, maybe some ZFS I/O worker panicked? Which CPU...
Either way, I'd first check if your hardware is alright, e.g. all cables and the RAM modules are connected correctly and aren't loose, there BIOS configuration is reset (no overlocking, etc.), and so forth. If it doesn't fix the problem, then a...
Hi!
journalctl with PAGER set will usually open in that pager (typically it's less) and therefore you need to page down for journalctl --since="1 hour ago". For a better view of the log from the end of the last boot, you can run journalctl -b -1...
From the provided log in log-pve540-1.zip it also seems like there's quite a few connection issues, e.g. to the InfluxDB right at the start. I assume that proxmoxW540 is one of the nodes that has the problems starting up? Is the network stable...
Are these all nodes in the cluster? What does pvecm status output? Keep in mind that it is not recommended to have a even amount of nodes in a cluster, see e.g. [0]. This seems like a clustering issue first of all.
Are the HA resources with node...
Hi!
x86-64-v4 is not a CPU model configuration that QEMU directly provides, so it only works by specifying it for the qemu-server's cpu config, but not in args.
What do you mean by "some flags aren't found"? It could be that these features...
Hi!
In general, the guest's operating system will usually pin each queue to a vCPU core. Therefore, there's not really a performance benefit to use more queues in total than vCPUs as then a vCPU must handle multiple queues then.
Multiqueues are...
Hi!
It's best to show your interest in a specific Bugzilla entry so that we can track the interest in certain features. There's for example [0] for the general interest of adding any SMBIOS type information and/or [1] for automatically...
Hi!
With GET /cluster/ha/rules and GET /cluster/ha/rules/{ruleid} you receive a digest value (of the whole config). If you provide this digest as an argument to PUT /cluster/ha/rules/{ruleid}, then the API handler will refuse to make changes to...
This fix is already done upstream [0] and for our patched ifupdown2 packages as well before Proxmox VE 9 was released, so this can only happen if the ifupdown2 package is somehow held back (e.g. using apt upgrade instead of apt dist-upgrade...