Bug? Extreme io load on disks from unknown source killing runnings VMs

VolkerK

New Member
Dec 11, 2016
14
0
1
47
Hi all,

I'm experiencing a very strange problem and could really need some substantial help since I'm running out of ideas where to look for a solution to this problem:

  • I'm running a newly installed PVE, version 4.4-12 with newest upgrade installed and brand new hardware
  • I'm running the system on 2 x 3GB SATA disks and a NVMe SSD 512GB
  • I'm running ZFS with mirror on SATA and 256GB cache and 256GB log on the SSD

    Here's Info on my ZFS:

    zpool list:
    NAME SIZE ALLOC FREE EXPANDSZ FRAG CAP DEDUP HEALTH ALTROOT
    rpool 1.81T 225G 1.59T - 10% 12% 1.00x ONLINE -


    zpool status:
    pool: rpool
    state: ONLINE
    scan: none requested
    config:

    NAME STATE READ WRITE CKSUM
    rpool ONLINE 0 0 0
    mirror-0 ONLINE 0 0 0
    sda2 ONLINE 0 0 0
    sdb2 ONLINE 0 0 0
    logs
    nvme0n1p1 ONLINE 0 0 0
    cache
    nvme0n1p2 ONLINE 0 0 0

    errors: No known data errors


    zfs list:
    NAME USED AVAIL REFER MOUNTPOINT
    rpool 233G 1.53T 96K /rpool
    rpool/ROOT 3.15G 1.53T 96K /rpool/ROOT
    rpool/ROOT/pve-1 3.15G 1.53T 3.15G /
    rpool/data 221G 1.53T 96K /rpool/data
    rpool/data/base-101-disk-1 2.49G 1.53T 2.49G -
    rpool/data/subvol-100-disk-1 1.12G 10.9G 1.12G /rpool/data/subvol-100-disk-1
    rpool/data/subvol-108-disk-1 2.01G 5.99G 2.01G /rpool/data/subvol-108-disk-1
    rpool/data/subvol-109-disk-1 934M 19.1G 934M /rpool/data/subvol-109-disk-1
    rpool/data/vm-102-disk-1 3.45G 1.53T 3.45G -
    rpool/data/vm-102-disk-2 76.0G 1.53T 76.0G -
    rpool/data/vm-103-disk-1 27.1G 1.53T 27.1G -
    rpool/data/vm-104-disk-1 18.3G 1.53T 18.3G -
    rpool/data/vm-105-disk-1 2.38G 1.53T 2.38G -
    rpool/data/vm-106-disk-1 83.9G 1.53T 83.9G -
    rpool/data/vm-107-disk-1 3.56G 1.53T 3.56G -
    rpool/swap 8.50G 1.54T 568M -

  • I'm running vzdump nightly on all VMs (LXC and Qemu) in "snapshot" mode, qemu-agent is installed
  • The dump is written to a samba share, that the host automounts via autofs over a 1GBit network
  • Usually it takes about 1,5 hours to backup ~120GB
    This is how a "normal" report looks like:

    upload_2017-2-9_18-40-1.png
  • During a normal backup, the IO load is low and the system is fully responsive

  • Unfortunately several times a week, vzdump suddenly causes an extreme io load on the disks sda and sdb
  • The load saturates both disks and atop shows "busy 94-99%"
    This is a snapshot of atop:

    upload_2017-2-9_18-39-12.png
  • The io-wait increases to 15% (unusual high) and the disk io shown is > 1 GBytes/s
    See these diagrams taken during the backup with the crazy load:

    upload_2017-2-9_18-42-10.png
    upload_2017-2-9_18-42-19.png
  • The reason for this is definitly not backup data written, since the network speed is not high. It rather drops to very low values
    upload_2017-2-9_18-50-42.png
  • The network is not the reason. It's not saturated during these times

  • The result is: completely unresponsive systems, VMs "hanging", backup of subsequent VMs very slow (only some kB/s instead of dozens of MB/s)
    Examples of some problems resulting from the crazy io situation:

    upload_2017-2-9_18-45-31.png

    upload_2017-2-9_18-45-49.png

    upload_2017-2-9_18-46-48.png

  • The most problematic fact is: after the end of vzdump (which in this case takes more than 12 hours instead of 1,5) the disks still go crazy and show 99% of load
    Here's what the IO delay looks like when the backup finished: you see it drops from about 15% to about 8% - but the two disks still go like crazy and this system is still not usable.

    upload_2017-2-9_18-53-11.png
  • The problem can only be solved by rebooting the whole machine!!
  • Right now, the backup is not usable and I disabled it (not good).
  • I can't see any processes in top, atop etc. that cause this. The io load "comes out of the universe".


    Let me resume: I'm not talking about bad performance because of vzdump (writing on slow NFS etc.). The vzdump triggers something causing sda and sdb go crazy, saturating the SATA-bus without even transmitting data. The effect keeps going on even after vzdump is finished and can only be stopped by rebooting.


    I'm running out of ideas where too search for the problem. Maybe this is a bug.
    Could it be a problem with ZFS and the cache/log SSD?
 

Attachments

  • upload_2017-2-9_18-43-23.png
    upload_2017-2-9_18-43-23.png
    47.6 KB · Views: 3
Update:
About an hour ago, the situation reoccured. The reason was not vzdump. It was not running.
The server just hat normal "low) io load.

upload_2017-2-10_11-40-24.png

Again, atop showed very high io load on sda and sdb:

upload_2017-2-10_11-42-9.png

I could'nt find the reason. No VM was causing this load. Noting in atop or iotop or iostat.
Only txg_sync show up about every 5secs with 99% io load.

I removed cache and log from the zpool. Removing log took about 30 secs.
I didn't help immediately.
But after 10 minutes or so the behavior went away.

I re-added cache and log and the problem didn't not occur.
 

Attachments

  • upload_2017-2-10_11-41-53.png
    upload_2017-2-10_11-41-53.png
    7.1 KB · Views: 5
  • upload_2017-2-10_11-42-3.png
    upload_2017-2-10_11-42-3.png
    7.1 KB · Views: 5

About

The Proxmox community has been around for many years and offers help and support for Proxmox VE, Proxmox Backup Server, and Proxmox Mail Gateway.
We think our community is one of the best thanks to people like you!

Get your subscription!

The Proxmox team works very hard to make sure you are running the best software and getting stable updates and security enhancements, as well as quick enterprise support. Tens of thousands of happy customers have a Proxmox subscription. Get yours easily in our online shop.

Buy now!