Lately we have seen that disk move from ceph to zfs is very slow.
I just tested trying to move a 12GB kvm disk.
the kvm was stopped prior to move.
move started at 20:14 . I am posting progress as it occurs.
20:24 13%
20:38 14%
20:43 14%
it is going very slow.
iotop does not most of the time show move activity.
log:
the zpool is a mirror of two good data center model intel ssd's:
ceph network 10G .
the pve host has 2CPU's , 40 cores , 200GB+ memory.
ceph -s :
i'll try letting the move continue and post the result.
the move finished at 23:57 . 3+ hours to move 12GB - Something is wrong with our configuration or there is a bug.
Does anyone have a suggestion to get disk move back to normal speed?
I just tested trying to move a 12GB kvm disk.
the kvm was stopped prior to move.
move started at 20:14 . I am posting progress as it occurs.
20:24 13%
20:38 14%
20:43 14%
it is going very slow.
iotop does not most of the time show move activity.
log:
Code:
less UPID:pve3:002F43FB:019B5E32:5BF9F759:qmmove:2165:root*
create full clone of drive scsi0 (ceph_vm:vm-2165-disk-1)
transferred: 0 bytes remaining: 12884901888 bytes total: 12884901888 bytes progression: 0.00 %
transferred: 130137509 bytes remaining: 12754764379 bytes total: 12884901888 bytes progression: 1.01 %
transferred: 260275018 bytes remaining: 12624626870 bytes total: 12884901888 bytes progression: 2.02 %
transferred: 390412527 bytes remaining: 12494489361 bytes total: 12884901888 bytes progression: 3.03 %
transferred: 520550036 bytes remaining: 12364351852 bytes total: 12884901888 bytes progression: 4.04 %
transferred: 650687545 bytes remaining: 12234214343 bytes total: 12884901888 bytes progression: 5.05 %
transferred: 779536564 bytes remaining: 12105365324 bytes total: 12884901888 bytes progression: 6.05 %
transferred: 909674073 bytes remaining: 11975227815 bytes total: 12884901888 bytes progression: 7.06 %
transferred: 1039811582 bytes remaining: 11845090306 bytes total: 12884901888 bytes progression: 8.07 %
transferred: 1169949091 bytes remaining: 11714952797 bytes total: 12884901888 bytes progression: 9.08 %
transferred: 1300086600 bytes remaining: 11584815288 bytes total: 12884901888 bytes progression: 10.09 %
transferred: 1430224109 bytes remaining: 11454677779 bytes total: 12884901888 bytes progression: 11.10 %
transferred: 1560361618 bytes remaining: 11324540270 bytes total: 12884901888 bytes progression: 12.11 %
transferred: 1690499127 bytes remaining: 11194402761 bytes total: 12884901888 bytes progression: 13.12 %
transferred: 1820636636 bytes remaining: 11064265252 bytes total: 12884901888 bytes progression: 14.13 %
the zpool is a mirror of two good data center model intel ssd's:
Code:
Model Family: Intel 730 and DC S35x0/3610/3700 Series SSDs
Device Model: INTEL SSDSC2BP480G4
ceph network 10G .
the pve host has 2CPU's , 40 cores , 200GB+ memory.
ceph -s :
Code:
cluster:
id: 220b9a53-4556-48e3-a73c-28deff665e45
health: HEALTH_WARN
noout flag(s) set
services:
mon: 3 daemons, quorum pve3,sys8,pve10
mgr: pve3(active), standbys: sys8, pve10
osd: 65 osds: 65 up, 65 in
flags noout
data:
pools: 2 pools, 1088 pgs
objects: 105.93k objects, 373GiB
usage: 1.10TiB used, 24.4TiB / 25.5TiB avail
pgs: 1088 active+clean
io:
client: 194KiB/s wr, 0op/s rd, 35op/s wr
i'll try letting the move continue and post the result.
the move finished at 23:57 . 3+ hours to move 12GB - Something is wrong with our configuration or there is a bug.
Does anyone have a suggestion to get disk move back to normal speed?