This is my original EL9 templates hardware layout
After cloning (full) via the following command, the layout changes
qm clone $VM_TEMPLATE $VM_ID --name $VM_HOSTNAME --storage ${VM_STORAGE:-data} -full
disk-0 and disk-1 switched places. I assume this is due to the way the cloning is...
We could not solve this issue reliably and ended up moving away from ZFS for primary VM storage. At this point we have switched to Ceph with an all-flash configuration. The investment in terms of EUR was quite substantial but the results are pretty good - i dare say "worth it". After roughly one...
In the GUI Datacenter -> Summary -> Nodes it shows percentage based graphs. Is there any official way to configure the color/percentage values?
Example: I would want it to start turning red when memory reaches 95% instead of 90% which seems to be the case currently.
@JamesT We didn't solve it.
For a while we used a bandwidth limit of 125 MiB/s. That made the issue almost unnoticeable at the cost of longer migration durations.
However, since then we moved away from ZFS and started using ceph flash only.
It's worth mentioning that the "more/remove" dialog in the PRX GUI has a checkbox that is misleading when using PBS.
I understand why its there and what it does for legacy dumps- but that is not the case when using PBS. It is not removing anything. I propose to make this button forget all...
Once the above patch was applied migrations caused no further problems. A quick test of backups also caused no problems.
However - we currently do not have the same test setup (with 40 big VMs) that we had before. We tested with ~5 VMs and saw no issues.
Can we expect this patch to make its way...
I have to correct this: While the patch fixes the backups - it kills migrations. The patched PVE node can no longer recieve online migrations.
Log output:
Migrations are successfull again if we revert the patch and restart pve* services.
Hi Wolfgang,
i suppose the last patch fixed the issue. We have done 8 full backup runs, ~14h/30vm's each - no error. All good.
Can you confirm the patch will make it into upstream release at some point?
That exact line gives syntax errors. I will be testing with one less bracket
my $bus = Net::DBus->system("private", 1, "nomainloop", 1 );
Will report back :)
The term backup chain does not apply to the way in which PBS works. There simply are no chains. There is chunks and indexes.
The basic assumption of "What if my data become corrupt" however is still a good thought. PBS counters this by using ZFS to make sure that no silent corruption of data can...
I edited this file (on PVE nodes) as per request, rebooted PVE nodes to be safe and re-started the Backup-Job.
After 2 successfull VM's the 3rd and all following VMs recieved the following error:
INFO: Starting Backup of VM 2026 (qemu)
INFO: Backup started at 2020-09-10 14:44:01
INFO: status =...
I understand. But: This is a testing scenario and the 40 VMs are offline. There is almost nothing else consuming any resources - and there are plenty of resources.
Having said that: Here is the atop files and some context:
prx003 is the pve node currently home to ~40 offline VMs.
pbx-backup is...
I find it unlikely that CPU is experiencing issues but i can certainly make sure. Will create atop file for both the pve node as well as the pbs node and report back here.
We moved all 40 VMs to a new/vanilla pve host. The first couple backup runs where completing without errors. All VMs where correctly processed and it took roughly 15 hours to backup all the VMs.
However, recently the same error ERROR: Backup of VM 2056 failed - start failed...
This site uses cookies to help personalise content, tailor your experience and to keep you logged in if you register.
By continuing to use this site, you are consenting to our use of cookies.