...
FSYNCS/SECOND: 21.73
...
[/CODE]
looks quite bad, this is the reason. should be between 1000 and 5000. you have no raid controller with raid cache enabled (with BBU).
more details about your storage config?
...
FSYNCS/SECOND: 21.73
...
[/CODE]
looks quite bad, this is the reason. should be between 1000 and 5000. you have no raid controller with raid cache enabled (with BBU).
more details about your storage config?
virtio0: local:126/vm-126-disk-3.raw,cache=unsafe,backup=no
VERY interesting result from last night: On a hunch, rather than moving bwlimit down to 10000, I changed our size parameter (the maximum snapshot diff size) from 32768 to 8192. We had increased the size parameter a while ago because we were hitting the max of default on some VMs that used to be quite i/o intensive during the night.
Interestingly, this reduced load by 2.5x (resulting in acceptable performance) throughout the entire backup period.
I believe we've found our culprit. I will re-increase bwlimit tonight (to 50000) and see what happens to load. I don't know much about LVM internals, but apparently snapshot size has a dramatic effect of performance.
Or, perhaps, do any of the Proxmox folks have an idea of what may have changed between these versions to cause load issues?
Or is it perhaps a kernel changed that caused this?
Unlike the other user (atran), I don't have enough data / test machines to perfectly correlate to a specific version. All of our machines showing major load during backup are running newest pvestest kernel 2.6.32 built Nov 24.