I have PBS setup as a VM with the specs below
Guest:
CPU - 2 cores
RAM - 6GB
Bios - OVMF
Machine - q35
Controller - VirtIO SCSI Single
OS Drive - 32GB QCOW SSD Discard on
Backup Drive - WD 4TB RE 7200RPM HDD, passthroughed to VM using SCSI and iothread = 1
PBS version - 2.3-3 Linux 5.15.85-1-pve #1 SMP PVE 5.15.85-1 (2023-02-01T00:00Z
The host is
CPU - 8 Core E3-1275 V2
RAM - 32GB
Proxmox has its own dedicated SSD while the VMs use another SSD.
I have the PBS storage mounted on all my hosts.
I've done a iperf3 test from the PBS host to every host and VM, speeds are saturating the 1Gbs link between all of them as expected.
I've done a naive drive test on the backup drive on PBS, the script wrote a thousand or so 1MB files to the drive without issues.
This setup was running perfectly fine for about a year, but now every backup freezes up at some point and the PBS guest shows 100% cpu usage and 100% I/O wait.
I've seen threads saying a single HDD isn't fast enough for PBS, but I've always had a single HDD like this, at one point it was even a USB external drive. Did something change recently that PBS needs faster storage? I don't mind swapping over to a SSD, but that's going to suck space wise.
All the guest VMs virutal drives are of the qcow2 type.
The host with PBS is also running BlueIris with its own three HDDs passedthrough and that VM is working fine.
I did notice the host was low on RAM, using up 30 out of 32GB, so I turned off some of the VMs and that seemed to solve the issue temporarily. Maybe with small guests it can backup, but soon after the problems showed up again.
Unfortunately I can't find anything else that would help me pin down the issue. During these issues the VM console shows messages like
INFO: task jbd2/sdb1-8:483 blocked for more than 120 seconds.
sdb1 is the backup drive partition.
output of
Guest:
CPU - 2 cores
RAM - 6GB
Bios - OVMF
Machine - q35
Controller - VirtIO SCSI Single
OS Drive - 32GB QCOW SSD Discard on
Backup Drive - WD 4TB RE 7200RPM HDD, passthroughed to VM using SCSI and iothread = 1
PBS version - 2.3-3 Linux 5.15.85-1-pve #1 SMP PVE 5.15.85-1 (2023-02-01T00:00Z
The host is
CPU - 8 Core E3-1275 V2
RAM - 32GB
Proxmox has its own dedicated SSD while the VMs use another SSD.
I have the PBS storage mounted on all my hosts.
I've done a iperf3 test from the PBS host to every host and VM, speeds are saturating the 1Gbs link between all of them as expected.
I've done a naive drive test on the backup drive on PBS, the script wrote a thousand or so 1MB files to the drive without issues.
This setup was running perfectly fine for about a year, but now every backup freezes up at some point and the PBS guest shows 100% cpu usage and 100% I/O wait.
I've seen threads saying a single HDD isn't fast enough for PBS, but I've always had a single HDD like this, at one point it was even a USB external drive. Did something change recently that PBS needs faster storage? I don't mind swapping over to a SSD, but that's going to suck space wise.
All the guest VMs virutal drives are of the qcow2 type.
The host with PBS is also running BlueIris with its own three HDDs passedthrough and that VM is working fine.
I did notice the host was low on RAM, using up 30 out of 32GB, so I turned off some of the VMs and that seemed to solve the issue temporarily. Maybe with small guests it can backup, but soon after the problems showed up again.
Unfortunately I can't find anything else that would help me pin down the issue. During these issues the VM console shows messages like
INFO: task jbd2/sdb1-8:483 blocked for more than 120 seconds.
sdb1 is the backup drive partition.
output of
Bash:
df -h
Bash:
Filesystem Size Used Avail Use% Mounted on
udev 2.9G 0 2.9G 0% /dev
tmpfs 590M 856K 590M 1% /run
/dev/mapper/pbs-root 29G 3.3G 25G 12% /
tmpfs 2.9G 164K 2.9G 1% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
/dev/sda2 511M 336K 511M 1% /boot/efi
/dev/sdb1 3.6T 338G 3.1T 10% /mnt/datastore/backup
tmpfs 590M 0 590M 0% /run/user/0