Yes, a backup job will run in parallel on each node (one backup per node at a time).How to do that? (reduce number of concurrent backups) ? In my cluster there is only 1 backup job and if I'm not mistaken PVE sterilize backups from all nodes within cluster)
What error message do you get in Linux?P.S. right, Vm hangs after backup (when another backup in progress). Win and Linux vm affected
Not in my case. 40Gbe link. There is powerful CPU and HW RAID with SSD disks on PBS server
CEPH datastore performance is a key limited factor (of backup speed)
And one more thing:
VM "hangs" (with different event to windows syslog + ID 129 as well) when another VM is backing up (that VM is located on another node)
Yes, exactly that setting. That should limit the background copying of the image during backup, so Ceph might not get overwhelmed and guest writes might have a better chance to get through.Is there any way to limit read/write bandwidth of PBS client (something like bwlimit in vdump.conf) ?
That's unfortunately not possible without using theHow should I change my Vm config to incorporate this tune?
P.S. there is another thread: https://github.com/virtio-win/kvm-guest-drivers-windows/issues/756
args
override to specify arbitrary CLI arguments.What error message do you get in Linux?
Yes, exactly that setting. That should limit the background copying of the image during backup, so Ceph might not get overwhelmed and guest writes might have a better chance to get through.
But what about the problems I have described? The error only occurs a few minutes after the backup. And not just on one system.
> Yes, a backup job will run in parallel on each node (one backup per node at a time).
and that may be just too much in larger clusters
https://bugzilla.proxmox.com/show_bug.cgi?id=3086
Have you tried tuning virtio-scsi parameters as advised by virtio devs?
- HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Services\vioscsi\Parameters\Device
- IoTimeoutValue = 0x5a (90)
- HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Services\vioscsi\Parameters\Device
- PhysicalBreaks = 0x3f (63)
Yes.2) Let me make it clear. PBS respects bwlimit parameter in vdump.conf? In other words: if I set bwlimit in vdump.conf PBS client will limit backup speed? Am I correct?
Yes, that is the setting you should use.3) And what kind of limit will stratify the most:
- a) bwlimit in vdump.conf
If you limit on the PBS-side, then guest writes during backup will also suffer from that. It's better to try and limit the reading side, i.e. Proxmox VE.
- b) traffic control in PBS
Those are various cluster-wide defaults for certain operations, but there is none for backups.
- c) Datacenter - Options - Bandwidth Limits
The ionice setting doesn't usually take any effect, in particular with VM backups to PBS it doesn't: https://pve.proxmox.com/pve-docs/chapter-vzdump.html#vzdump_configurationI will try the following:
- tune virtio-scsi driver settings via windows registry
- set ionice=8 in vzdump.conf on all nodes in cluster
I think that it can't be that I have to set timeouts to such a high value so that it doesn't lead to an error. Or that I have to limit bandwidths just so that something runs without errors.
I have never had to change such values in environments with Hyper-V and Veeam. That's just unfamiliar for me at first. Of course I will test this and will be happy to provide feedback. With such findings, perhaps the whole thing can be taken further in the right direction.What do mean by "such high values"?
Defaults are: 2Mb with 60s timeout, tuned: 256Kb with 90s timeout
You are free to test just one of them anyway
I have never had to change such values in environments with Hyper-V and Veeam. That's just unfamiliar for me at first. Of course I will test this and will be happy to provide feedback. With such findings, perhaps the whole thing can be taken further in the right direction.
The ionice setting doesn't usually take any effect, in particular with VM backups to PBS it doesn't: https://pve.proxmox.com/pve-docs/chapter-vzdump.html#vzdump_configuration
bwlimit: 150000
ionice: 8
Yes, it will be set for the vzdump process and child processes. But it only takes and effect when you use the BFQ scheduler. And for VMs, the backup is done via the QEMU process to which the ionice setting is not applied (that would also affect guest writes). And when using PBS, vzdump doesn't spawn a compressor child process. So in the case VM+PBS or if not using BFQ scheduler, the setting is essentially without effect.However with ionice=8 set in vzdump.conf I can see it in backup log
no, that is not true. If your PBS connection is too slow or gets lost, the issue described in this thread is still present.This should be fixed with pve-qemu-kvm package updated to 8.1.2-5.
So backup fleecing doesn't help...? Or only in some scenarios?Hi,This should be fixed with pve-qemu-kvm package updated to 8.1.2-5.
no, that is not true. If your PBS connection is too slow or gets lost, the issue described in this thread is still present.
chkdsk
, resulting in more IO, which exacerbates the problem. This was often reported during initial patching following Windows setup. These were ultimately unresolved because users were leaving the workaround in place (using IDE or SATA instead - I believe the default LSI works too), e.g. as mentioned here. Updating just one VM could bring the whole server down even if I shut down most of my other VMs on the affected host. In my case, I think several issues were at play which I have attempted to provide some details of in the thread I started here.ok, switching to threads does not help either. and again the errors only occur a few minutes after the backup has been completed.
Have you tried tuning virtio-scsi parameters as advised by virtio devs?
- HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Services\vioscsi\Parameters\Device
- IoTimeoutValue = 0x5a (90)
- HKEY_LOCAL_MACHINE\SYSTEM\CurrentControlSet\Services\vioscsi\Parameters\Device
- PhysicalBreaks = 0x3f (63)
grep "" /sys/block/*/queue/max_segments
.We use essential cookies to make this site work, and optional cookies to enhance your experience.