Slow "MoveDisk" between two SSDs

naimo

Member
Dec 1, 2022
7
0
6
Bavaria
github.com
Hey Guys,

Sorry for my Englisch...

I have a strange behaviour here... I reused a previous ZFS Raid SSD, formated it, "wiped" it via the PVE GUI and wanted it to use as a LVM-Thin.

"Create ThinPool" via GUI doesn't work for me, it seems to do nothing and after a few minutes, all VMs are greyed out. Only logging via SHH is possible anymore. And Only a reboot seems to work, but the pool is not initialized correctly. So wipe again and create it via Console:

pvcreate /dev/sda
vgcreate vgrp1 /dev/sda
lvcreate -l 100%FREE --thinpool thpl1 vgrp1 -Zn

and adding a Storage via GUI works.

BUT: If I want to move a disk of a VM (online) from "data2" to "vgrp1-thpl1", it starts with 500MB/s for a few seconds, and falls down to 1-2 MB/s.

If tested it with serveral VM Disks, all the same...

Bildschirmfoto 2024-02-13 um 14.13.02.png

Bildschirmfoto 2024-02-13 um 22.08.40.png


iotop during coping disk of VM id 101, but it's the same for id 201 from above:
Bildschirmfoto 2024-02-13 um 14.15.05.png

qm config 201

Code:
agent: 1
bios: seabios
boot: c
cores: 2
cpu: kvm64
description: Ubuntu Server jammy Image
hotplug: network,disk,usb
ide0: local-lvm:vm-201-cloudinit,media=cdrom,size=4M
ide2: none,media=cdrom
ipconfig0: ip=10.0.0.14/24,gw=10.0.0.254
kvm: 1
memory: 1024
meta: creation-qemu=7.2.0,ctime=1687976277
name: adguard2
net0: virtio=12:99:C9:BF:C9:1B,bridge=vmbr0
numa: 0
onboot: 1
ostype: other
scsihw: virtio-scsi-pci
smbios1: uuid=41610610-5260-4311-88f3-fba1fe8fffa9
sockets: 1
tablet: 1
unused0: data2:vm-201-disk-0
virtio0: vmimages:vm-201-disk-0,iothread=0,size=32G
vmgenid: de76adea-b72d-449a-a4fe-abefeebf91b4

lsblk

Code:
NAME                           MAJ:MIN RM   SIZE RO TYPE MOUNTPOINTS
sda                              8:0    0 931.5G  0 disk
|-vgrp1-thpl1_tmeta            252:6    0   120M  0 lvm
| `-vgrp1-thpl1-tpool          252:22   0 931.3G  0 lvm
|   |-vgrp1-thpl1              252:23   0 931.3G  1 lvm
|   |-vgrp1-vm--115--disk--0   252:24   0    32G  0 lvm
|   `-vgrp1-vm--201--disk--0   252:25   0    32G  0 lvm
`-vgrp1-thpl1_tdata            252:21   0 931.3G  0 lvm
  `-vgrp1-thpl1-tpool          252:22   0 931.3G  0 lvm
    |-vgrp1-thpl1              252:23   0 931.3G  1 lvm
    |-vgrp1-vm--115--disk--0   252:24   0    32G  0 lvm
    `-vgrp1-vm--201--disk--0   252:25   0    32G  0 lvm
sdb                              8:16   0 238.5G  0 disk
|-data2-data2_tmeta            252:4    0   2.4G  0 lvm
| `-data2-data2-tpool          252:12   0 233.6G  0 lvm
|   |-data2-data2              252:13   0 233.6G  1 lvm
|   |-data2-vm--113--disk--0   252:14   0    32G  0 lvm
|   |-data2-vm--114--disk--0   252:15   0    20G  0 lvm
|   |-data2-vm--201--disk--0   252:16   0    32G  0 lvm
|   |-data2-vm--300--cloudinit 252:17   0     4M  0 lvm
|   |-data2-vm--300--disk--0   252:18   0    32G  0 lvm
|   |-data2-vm--101--disk--0   252:19   0    50G  0 lvm
|   `-data2-vm--110--disk--0   252:20   0    50G  0 lvm
`-data2-data2_tdata            252:5    0 233.6G  0 lvm
  `-data2-data2-tpool          252:12   0 233.6G  0 lvm
    |-data2-data2              252:13   0 233.6G  1 lvm
    |-data2-vm--113--disk--0   252:14   0    32G  0 lvm
    |-data2-vm--114--disk--0   252:15   0    20G  0 lvm
    |-data2-vm--201--disk--0   252:16   0    32G  0 lvm
    |-data2-vm--300--cloudinit 252:17   0     4M  0 lvm
    |-data2-vm--300--disk--0   252:18   0    32G  0 lvm
    |-data2-vm--101--disk--0   252:19   0    50G  0 lvm
    `-data2-vm--110--disk--0   252:20   0    50G  0 lvm
sdc                              8:32   0 931.5G  0 disk
nvme0n1                        259:0    0 238.5G  0 disk
|-nvme0n1p1                    259:1    0  1007K  0 part
|-nvme0n1p2                    259:2    0   512M  0 part
`-nvme0n1p3                    259:3    0   238G  0 part
  |-pve-swap                   252:0    0     8G  0 lvm  [SWAP]
  |-pve-root                   252:1    0  69.5G  0 lvm  /
  |-pve-data_tmeta             252:2    0   1.4G  0 lvm
  | `-pve-data-tpool           252:7    0 141.6G  0 lvm
  |   |-pve-data               252:8    0 141.6G  1 lvm
  |   |-pve-vm--102--disk--0   252:9    0    32G  0 lvm
  |   |-pve-vm--110--disk--0   252:10   0    50G  0 lvm
  |   `-pve-vm--201--cloudinit 252:11   0     4M  0 lvm
  `-pve-data_tdata             252:3    0 141.6G  0 lvm
    `-pve-data-tpool           252:7    0 141.6G  0 lvm
      |-pve-data               252:8    0 141.6G  1 lvm
      |-pve-vm--102--disk--0   252:9    0    32G  0 lvm
      |-pve-vm--110--disk--0   252:10   0    50G  0 lvm
      `-pve-vm--201--cloudinit 252:11   0     4M  0 lvm

SMART Tests are PASSED

After the Disk is moved, the VM is also incredible slow.

If I create a new VM on the Same storage, its normally fast.


I don't have any idea, why the move it soooo slow? Do you?

I'll appreciate any help or hint, you can give me. If I can provide more informations, please let me know.

Thanks a lot in advance.

Greets,
Benjamin
 

Attachments

  • Bildschirmfoto 2024-02-13 um 14.13.15.png
    Bildschirmfoto 2024-02-13 um 14.13.15.png
    105.7 KB · Views: 3
  • Bildschirmfoto 2024-02-13 um 16.29.07.png
    Bildschirmfoto 2024-02-13 um 16.29.07.png
    44 KB · Views: 3
Last edited:
Good Morning in the morning,

today the second, newly created, VM is also normally fast... Is it the IO Delay after moving?

A second slow behaviour: also migrating from another node to this specific SSD is more than slow. If the target storage is one of the other SSDs its transfering with roundabout 1Gbit.

qm status --verbose

Code:
balloon: 4294967296
ballooninfo:
    actual: 4294967296
    free_mem: 621457408
    last_update: 1707898719
    major_page_faults: 48599
    max_mem: 4294967296
    mem_swapped_in: 630206464
    mem_swapped_out: 713998336
    minor_page_faults: 156480940
    total_mem: 4101513216
blockstat:
    ide0:
        account_failed: 1
        account_invalid: 1
        failed_flush_operations: 0
        failed_rd_operations: 0
        failed_unmap_operations: 0
        failed_wr_operations: 0
        flush_operations: 0
        flush_total_time_ns: 0
        idle_time_ns: 228697240840957
        invalid_flush_operations: 0
        invalid_rd_operations: 0
        invalid_unmap_operations: 0
        invalid_wr_operations: 0
        rd_bytes: 418022
        rd_merged: 0
        rd_operations: 116
        rd_total_time_ns: 53126220
        timed_stats:
        unmap_bytes: 0
        unmap_merged: 0
        unmap_operations: 0
        unmap_total_time_ns: 0
        wr_bytes: 0
        wr_highest_offset: 0
        wr_merged: 0
        wr_operations: 0
        wr_total_time_ns: 0
    virtio0:
        account_failed: 1
        account_invalid: 1
        failed_flush_operations: 0
        failed_rd_operations: 0
        failed_unmap_operations: 0
        failed_wr_operations: 0
        flush_operations: 109616
        flush_total_time_ns: 162009094410
        idle_time_ns: 2080298191
        invalid_flush_operations: 0
        invalid_rd_operations: 0
        invalid_unmap_operations: 0
        invalid_wr_operations: 0
        rd_bytes: 4736989184
        rd_merged: 1189
        rd_operations: 109039
        rd_total_time_ns: 65570784594
        timed_stats:
        unmap_bytes: 0
        unmap_merged: 0
        unmap_operations: 0
        unmap_total_time_ns: 0
        wr_bytes: 10357277696
        wr_highest_offset: 33982115840
        wr_merged: 2120
        wr_operations: 1019114
        wr_total_time_ns: 1428435469280
cpus: 4
disk: 0
diskread: 4737407206
diskwrite: 10357277696
freemem: 621457408
lock: migrate
maxdisk: 0
maxmem: 4294967296
mem: 3480055808
name: k8snode01
netin: 1922044973
netout: 495358485
nics:
    tap301i0:
        netin: 1922044973
        netout: 495358485
pid: 2058
proxmox-support:
    backup-max-workers: 1
    pbs-dirty-bitmap: 1
    pbs-dirty-bitmap-migration: 1
    pbs-dirty-bitmap-savevm: 1
    pbs-library-version: 1.3.1 (4d450bb294cac5316d2f23bf087c4b02c0543d79)
    pbs-masterkey: 1
    query-bitmap-info: 1
qmpstatus: running
running-machine: pc-i440fx-7.2+pve0
running-qemu: 7.2.0
status: running
uptime: 228716
vmid: 301

Bildschirmfoto 2024-02-14 um 09.27.17.png
 
Last edited:
Hello, what kind of SSD disks are you using? It is known that consumer grade SSDs can become slower than a HDD when they fill their cache, which can happen rather quickly when moving VM disks.
 
Hello Maximiliano,

indeed I'm using for my "homelab" consumer grade SSDs.

The "fast" one is a Crucial MX100 256GB.
The "slow" one is a SanDisk SSD Plus 1TB.

Thanks a lot for the fast reponse.

Best regards,
Benjamin