When the first crash occoured I had this Problem:
[ 5880.143371] igb 0000:05:00.0 nic1: NETDEV WATCHDOG: CPU: 6: transmit queue 1 timed out 5021 ms
I switched to proxmox two weeks ago.
Before proxmox i had OMV (Debian) with 6.1.0 on the system...
Greentigs,
I have put a Proxmox training manual together for my team (and others) and I would like to offer it here for review/use by the community. This is not intended to be a technical deep-dive, rather a practical point-to-point to help...
Laut Aussage eines Proxmox Entwicklers hat man darüber schon nachgedacht und ist zum Schluss gekommen, dass das konterproduktiv ist:
- Es gab schon früher mal eine Möglichkeit zu spenden, die dabei entstehenden Einnahmen (bzw. deren Mangel)...
@RodolfoRibeiro If you want more direct assistance, post the content of (from both hosts)
lsblk
multipath -ll -v2
if you have system logs available from the point in time when your vm became corrupted, would be good to look at what happened.
Hi @RodolfoRibeiro , thank you for clarifying. This matches my initial understanding of your situation. SAS and iSCSI are transfer and connectivity protocols. While the article I suggested is using iSCSI as example, once you are beyond basic...
For the generations of hardware where iscsi and SAS were offered as available SKUs there was no meaningful performance difference- 16G FC simply had more headroom to fill cache. When 25GB iscsi product started shipping, THOSE were faster (even...
Although they were about replicated pools (so no ec) following reads might serve as a hint why (outside of experiments/lab setups) it's not a good idea to go against the recommendations...
K=6,M=2 results in 6 data strips per 8 total. 6/8=0.75
in replication you have 1 data strips per 3 total. 1/3=0.33
its not exactly the "same" availability because survivability in a replication group is much higher; you need one living osd per...
Hi, @sidoni
I believe that the first warning can be explained in
https://pve.proxmox.com/pve-docs/chapter-qm.html#qm_system_settings
- especially in the sections "Machine Version" up to "Update to a Newer Machine Version".
As for the second...
K=6,M=2 results in 6 data strips per 8 total. 6/8=0.75
in replication you have 1 data strips per 3 total. 1/3=0.33
its not exactly the "same" availability because survivability in a replication group is much higher; you need one living osd per...
Hello wondimu,
May we know more about the Disks? Or Perhaps get a PVE report? It is possible that only some of the Disk are stored on media that allows replication.
Report can be generated by going to DataCenter > Node > Subscription > Click...
Although they were about replicated pools (so no ec) following reads might serve as a hint why (outside of experiments/lab setups) it's not a good idea to go against the recommendations...
@x509
corosync is on 1G "private" link
no redundancy for 25G - single node failure is accepted as datacenter in which those servers are has 24h service and spare parts - faulty node will be up in matter of minutes / hours
there is total of 70 VMs...
Take this with a huge grain of salt. I don't know you or your customers :)
IMHO you probably don't need HA. Redudant PSU and local storage is more than enough. And your next part is a good explanation why.
That is not automatically real HA...
I see that we have replication enabled between the two nodes, but one of the VMs is only replicating half of its disks. Is there a solution to fix this?”
Hello everyone,
I’m currently working on an automation project to provision virtual machines in Proxmox 9 using Terraform (BPG provider) together with cloud-init. The virtual machines are based on Ubuntu 24.04 cloud images.
The provisioning...
Just upgraded myself. Went just fine no issues. 3 OSDs. I got some interesting data after upgrading:
Ceph Squid → Tentacle Upgrade Benchmark Summary
Cluster: 3-node Proxmox (Intel NUC14, NVMe)
Pool: ceph-vms (replicated, size 2 / min_size 2)...