Super slow, timeout, and VM stuck while backing up, after updated to PVE 9.1.1 and PBS 4.0.20

... Does everyone who has the problem have LACP aggregation?
Me too. All my PVE nodes run with a bond lacp 803.2ad layer 3+4 with MTU 1500, on two 10g cards.
The PBS is a VM that runs on one of the nodes, even though the virtual disks are qcow2 files on a NAS via NFS.
 
you could try to downgrade kernel also on PVE host and report it, maybe help, but its not related to specific manufacter driver.

Actually Im running Intel Corporation Ethernet Controller X710 for 10GbE SFP+ on 8.4.14 hosts
and
BCM5719 on 9.1.1 test host

but no issues on both scenario restoring VMs.

On the other hand Im on Intel Corporation 82599ES 10-Gigabit SFI/SFP+ Network Connection (rev 01) on PBS 4.1 , and every backup was a nightmare before I reverted to good old 6.14
So, with a 9.1.1 node and 6.14.11-4-pve kernel, I can restore without any problem.
PS: i had not problems with mass live migration, and I use always the same bond of 4x10Gbit with vlans and mtu 9000
 
I've seen several mention the Intel 82599ES NIC which is also what I'm running in my dev PBS Server (and PVE Hosts). 20GB LAGG's as well.

Given the drop in Disk I/O it is certainly pointing towards a networking issue on the surface.
 
So, with a 9.1.1 node and 6.14.11-4-pve kernel, I can restore without any problem.
PS: i had not problems with mass live migration, and I use always the same bond of 4x10Gbit with vlans and mtu 9000

I have downgraded kernel on two clusters and PBS, so they are to the last software available (no subscription) but kernel 6.14.11-4-pve.
One backup was normal, another one is still running, very slow; linux vm with guest agent:
INFO: 32% (19.2 GiB of 60.0 GiB) in 23m 32s, read: 16.0 MiB/s, write: 15.8 MiB/s
Incremental, new dirty-map, Really too slow....
so downgrading kernel doesn't resolve the issue for me,.