Hi all,
I'm following this forum from a couple of years up to now, getting all the best knowledge from the community and solving all the small problems that I've encountered in Proxmox VE.
Now it's my turn to submit a new issue on our new installation over an HP DL380 gen 8 server with this specification:
The idea was to use the 120gb SSD as system disk and the array of 4x SSD in raid 10 configuration for the VM datastore.
Since we have the P822 raid card the easiest approach will be to build the RAID array in the controller and use it directly from PVE, but we would like to give a try with ZFS since from many topics in the net can give us many advantages in terms of reliability and performance.
So we started to create single disk volumes on the P822 card to let proxmox see individually each single drive (No SMART info unfortunately), we created the ZFS pool named VM-ZFS and we started our test with one Ubuntu VM and another Windows 10 VM with all disk using write back cache.
On Ubuntu VM we started to monitor the performance using fio command tool:
And here is were we first found issue on our configuration, we didn't expect crazy numbers but neither a result slower than single disk performance!
We performed the same random read test on the PVE host with much different results
Second message with detail due to limit on 15k char per message
I'm following this forum from a couple of years up to now, getting all the best knowledge from the community and solving all the small problems that I've encountered in Proxmox VE.
Now it's my turn to submit a new issue on our new installation over an HP DL380 gen 8 server with this specification:
Chassis
0 x HP|DL380p G8|2U|16*SFF-HS|SAS|HSP
Processor(s)
2 x Intel Xeon E5-2670 V2 2.50Ghz Ten (10) Core CPU
Heatsink(s)
2 x HP ProLiant DL380p Gen8 High Power 130W+ Heatsink
Fan(s)
2 x HP (667855-B21) ProLiant DL380p, DL380e Gen8, DL385p Gen8 Fan (662520-001)
Memory (RAM)
8 x 16GB - DDR3 1866MHz (PC3-14900R, 2Rx4)
RAID
1 x HP|P822 FH 2GB FBWC|FBOM
Hard Drive Caddy(s) & Blank(s)
8 x HP ProLiant Gen8, Gen9 SFF Hot-Swap Caddy
8 x HP ProLiant Gen8, Gen9, Gen10 SFF Hot-Swap Blank
Hard Drives
1x SSD on P420i KINGSTON A400 120Gb
4x SSD on P822 KINGSTON SEDC450 480Gb
Network Connectivity (FLOM)
1 x HP 331FLR Quad Port - 1GbE RJ45 FLR Ethernet
Full-Height Expansion Card(s)
1 x Qlogic QLE3044 Quad Port - 1GbE RJ45 Full Height PCIe-x4 Ethernet
Power Supply(s)
2 x HP Common Slot HS PSU 1200W High Eff.
0 x HP|DL380p G8|2U|16*SFF-HS|SAS|HSP
Processor(s)
2 x Intel Xeon E5-2670 V2 2.50Ghz Ten (10) Core CPU
Heatsink(s)
2 x HP ProLiant DL380p Gen8 High Power 130W+ Heatsink
Fan(s)
2 x HP (667855-B21) ProLiant DL380p, DL380e Gen8, DL385p Gen8 Fan (662520-001)
Memory (RAM)
8 x 16GB - DDR3 1866MHz (PC3-14900R, 2Rx4)
RAID
1 x HP|P822 FH 2GB FBWC|FBOM
Hard Drive Caddy(s) & Blank(s)
8 x HP ProLiant Gen8, Gen9 SFF Hot-Swap Caddy
8 x HP ProLiant Gen8, Gen9, Gen10 SFF Hot-Swap Blank
Hard Drives
1x SSD on P420i KINGSTON A400 120Gb
4x SSD on P822 KINGSTON SEDC450 480Gb
Network Connectivity (FLOM)
1 x HP 331FLR Quad Port - 1GbE RJ45 FLR Ethernet
Full-Height Expansion Card(s)
1 x Qlogic QLE3044 Quad Port - 1GbE RJ45 Full Height PCIe-x4 Ethernet
Power Supply(s)
2 x HP Common Slot HS PSU 1200W High Eff.
The idea was to use the 120gb SSD as system disk and the array of 4x SSD in raid 10 configuration for the VM datastore.
Since we have the P822 raid card the easiest approach will be to build the RAID array in the controller and use it directly from PVE, but we would like to give a try with ZFS since from many topics in the net can give us many advantages in terms of reliability and performance.
So we started to create single disk volumes on the P822 card to let proxmox see individually each single drive (No SMART info unfortunately), we created the ZFS pool named VM-ZFS and we started our test with one Ubuntu VM and another Windows 10 VM with all disk using write back cache.
oot@pve2:/# ssacli ctrl slot=2 show config
Smart Array P822 in Slot 2 (sn: PDVTF0CRHAX01H)
Internal Drive Cage at Port 5I, Box 2, OK
Internal Drive Cage at Port 6I, Box 0, OK
Port Name: 1E
Port Name: 2E
Port Name: 3E
Port Name: 4E
Port Name: 5I
Port Name: 6I
Array A (Solid State SATA, Unused Space: 0 MB)
logicaldrive 1 (447.10 GB, RAID 0, OK)
physicaldrive 5I:2:1 (port 5I:box 2:bay 1, SATA SSD, 480 GB, OK)
Array B (Solid State SATA, Unused Space: 0 MB)
logicaldrive 2 (447.10 GB, RAID 0, OK)
physicaldrive 5I:2:2 (port 5I:box 2:bay 2, SATA SSD, 480 GB, OK)
Array C (Solid State SATA, Unused Space: 0 MB)
logicaldrive 3 (447.10 GB, RAID 0, OK)
physicaldrive 5I:2:3 (port 5I:box 2:bay 3, SATA SSD, 480 GB, OK)
Array D (Solid State SATA, Unused Space: 0 MB)
logicaldrive 4 (447.10 GB, RAID 0, OK)
physicaldrive 5I:2:4 (port 5I:box 2:bay 4, SATA SSD, 480 GB, OK)
SEP (Vendor ID PMCSIERA, Model SRCv24x6G) 380 (WWID: 50014380292FD51F)
Smart Array P822 in Slot 2 (sn: PDVTF0CRHAX01H)
Internal Drive Cage at Port 5I, Box 2, OK
Internal Drive Cage at Port 6I, Box 0, OK
Port Name: 1E
Port Name: 2E
Port Name: 3E
Port Name: 4E
Port Name: 5I
Port Name: 6I
Array A (Solid State SATA, Unused Space: 0 MB)
logicaldrive 1 (447.10 GB, RAID 0, OK)
physicaldrive 5I:2:1 (port 5I:box 2:bay 1, SATA SSD, 480 GB, OK)
Array B (Solid State SATA, Unused Space: 0 MB)
logicaldrive 2 (447.10 GB, RAID 0, OK)
physicaldrive 5I:2:2 (port 5I:box 2:bay 2, SATA SSD, 480 GB, OK)
Array C (Solid State SATA, Unused Space: 0 MB)
logicaldrive 3 (447.10 GB, RAID 0, OK)
physicaldrive 5I:2:3 (port 5I:box 2:bay 3, SATA SSD, 480 GB, OK)
Array D (Solid State SATA, Unused Space: 0 MB)
logicaldrive 4 (447.10 GB, RAID 0, OK)
physicaldrive 5I:2:4 (port 5I:box 2:bay 4, SATA SSD, 480 GB, OK)
SEP (Vendor ID PMCSIERA, Model SRCv24x6G) 380 (WWID: 50014380292FD51F)
On Ubuntu VM we started to monitor the performance using fio command tool:
oot@ubuntu:~# fio --name TEST --eta-newline=5s --filename=fio-tempfile.dat --rw=read --size=500m --io_size=5g --blocksize=1024k --ioengine=libaio --fsync=10000 --iodepth=32 --direct=1 --numjobs=1 --runtime=60 --group_reporting
TEST: (g=0): rw=read, bs=(R) 1024KiB-1024KiB, (W) 1024KiB-1024KiB, (T) 1024KiB-1024KiB, ioengine=libaio, iodepth=32
fio-3.1
Starting 1 process
TEST: Laying out IO file (1 file / 500MiB)
TEST: (groupid=0, jobs=1): err= 0: pid=1311: Sat Nov 21 12:48:51 2020
read: IOPS=6066, BW=6066MiB/s (6361MB/s)(5120MiB/844msec)
slat (usec): min=36, max=4307, avg=133.65, stdev=82.54
clat (usec): min=545, max=20800, avg=4914.40, stdev=1221.92
lat (usec): min=735, max=21788, avg=5052.24, stdev=1250.03
clat percentiles (usec):
| 1.00th=[ 1598], 5.00th=[ 3490], 10.00th=[ 4178], 20.00th=[ 4490],
| 30.00th=[ 4621], 40.00th=[ 4686], 50.00th=[ 4686], 60.00th=[ 4817],
| 70.00th=[ 5211], 80.00th=[ 5538], 90.00th=[ 5735], 95.00th=[ 5932],
| 99.00th=[ 9372], 99.50th=[10814], 99.90th=[18482], 99.95th=[19792],
| 99.99th=[20841]
bw ( MiB/s): min= 5798, max= 5798, per=95.58%, avg=5798.00, stdev= 0.00, samples=1
iops : min= 5798, max= 5798, avg=5798.00, stdev= 0.00, samples=1
lat (usec) : 750=0.12%, 1000=0.25%
lat (msec) : 2=1.07%, 4=6.41%, 10=91.50%, 20=0.61%, 50=0.04%
cpu : usr=11.74%, sys=85.05%, ctx=75, majf=0, minf=8202
IO depths : 1=0.2%, 2=0.4%, 4=0.9%, 8=1.7%, 16=3.4%, 32=93.3%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.2%, 64=0.0%, >=64=0.0%
issued rwt: total=5120,0,0, short=0,0,0, dropped=0,0,0
latency : target=0, window=0, percentile=100.00%, depth=32
Run status group 0 (all jobs):
READ: bw=6066MiB/s (6361MB/s), 6066MiB/s-6066MiB/s (6361MB/s-6361MB/s), io=5120MiB (5369MB), run=844-844msec
Disk stats (read/write):
sda: ios=4718/0, merge=0/0, ticks=4260/0, in_queue=3652, util=85.39%
TEST: (g=0): rw=read, bs=(R) 1024KiB-1024KiB, (W) 1024KiB-1024KiB, (T) 1024KiB-1024KiB, ioengine=libaio, iodepth=32
fio-3.1
Starting 1 process
TEST: Laying out IO file (1 file / 500MiB)
TEST: (groupid=0, jobs=1): err= 0: pid=1311: Sat Nov 21 12:48:51 2020
read: IOPS=6066, BW=6066MiB/s (6361MB/s)(5120MiB/844msec)
slat (usec): min=36, max=4307, avg=133.65, stdev=82.54
clat (usec): min=545, max=20800, avg=4914.40, stdev=1221.92
lat (usec): min=735, max=21788, avg=5052.24, stdev=1250.03
clat percentiles (usec):
| 1.00th=[ 1598], 5.00th=[ 3490], 10.00th=[ 4178], 20.00th=[ 4490],
| 30.00th=[ 4621], 40.00th=[ 4686], 50.00th=[ 4686], 60.00th=[ 4817],
| 70.00th=[ 5211], 80.00th=[ 5538], 90.00th=[ 5735], 95.00th=[ 5932],
| 99.00th=[ 9372], 99.50th=[10814], 99.90th=[18482], 99.95th=[19792],
| 99.99th=[20841]
bw ( MiB/s): min= 5798, max= 5798, per=95.58%, avg=5798.00, stdev= 0.00, samples=1
iops : min= 5798, max= 5798, avg=5798.00, stdev= 0.00, samples=1
lat (usec) : 750=0.12%, 1000=0.25%
lat (msec) : 2=1.07%, 4=6.41%, 10=91.50%, 20=0.61%, 50=0.04%
cpu : usr=11.74%, sys=85.05%, ctx=75, majf=0, minf=8202
IO depths : 1=0.2%, 2=0.4%, 4=0.9%, 8=1.7%, 16=3.4%, 32=93.3%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.2%, 64=0.0%, >=64=0.0%
issued rwt: total=5120,0,0, short=0,0,0, dropped=0,0,0
latency : target=0, window=0, percentile=100.00%, depth=32
Run status group 0 (all jobs):
READ: bw=6066MiB/s (6361MB/s), 6066MiB/s-6066MiB/s (6361MB/s-6361MB/s), io=5120MiB (5369MB), run=844-844msec
Disk stats (read/write):
sda: ios=4718/0, merge=0/0, ticks=4260/0, in_queue=3652, util=85.39%
root@ubuntu:~# fio --name TEST --eta-newline=5s --filename=fio-tempfile.dat --rw=write --size=500m --io_size=5g --blocksize=1024k --ioengine=libaio --fsync=10000 --iodepth=32 --direct=1 --numjobs=1 --runtime=60 --group_reporting
TEST: (g=0): rw=write, bs=(R) 1024KiB-1024KiB, (W) 1024KiB-1024KiB, (T) 1024KiB-1024KiB, ioengine=libaio, iodepth=32
fio-3.1
Starting 1 process
Jobs: 1 (f=1)
TEST: (groupid=0, jobs=1): err= 0: pid=1314: Sat Nov 21 12:50:45 2020
write: IOPS=4604, BW=4604MiB/s (4828MB/s)(5120MiB/1112msec)
slat (usec): min=56, max=862, avg=199.01, stdev=46.97
clat (usec): min=404, max=9098, avg=6535.65, stdev=1266.94
lat (usec): min=583, max=9334, avg=6737.31, stdev=1285.84
clat percentiles (usec):
| 1.00th=[ 1418], 5.00th=[ 5211], 10.00th=[ 5932], 20.00th=[ 6063],
| 30.00th=[ 6128], 40.00th=[ 6194], 50.00th=[ 6259], 60.00th=[ 6325],
| 70.00th=[ 6521], 80.00th=[ 7963], 90.00th=[ 8356], 95.00th=[ 8586],
| 99.00th=[ 8717], 99.50th=[ 8848], 99.90th=[ 8979], 99.95th=[ 9110],
| 99.99th=[ 9110]
bw ( MiB/s): min= 4098, max= 4980, per=98.58%, avg=4539.00, stdev=623.67, samples=2
iops : min= 4098, max= 4980, avg=4539.00, stdev=623.67, samples=2
lat (usec) : 500=0.08%, 750=0.21%, 1000=0.25%
lat (msec) : 2=1.04%, 4=2.15%, 10=96.27%
cpu : usr=68.77%, sys=30.96%, ctx=44, majf=0, minf=13
IO depths : 1=0.2%, 2=0.4%, 4=0.9%, 8=1.7%, 16=3.4%, 32=93.3%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.2%, 64=0.0%, >=64=0.0%
issued rwt: total=0,5120,0, short=0,0,0, dropped=0,0,0
latency : target=0, window=0, percentile=100.00%, depth=32
Run status group 0 (all jobs):
WRITE: bw=4604MiB/s (4828MB/s), 4604MiB/s-4604MiB/s (4828MB/s-4828MB/s), io=5120MiB (5369MB), run=1112-1112msec
Disk stats (read/write):
sda: ios=0/5006, merge=0/0, ticks=0/3788, in_queue=2992, util=87.28%
TEST: (g=0): rw=write, bs=(R) 1024KiB-1024KiB, (W) 1024KiB-1024KiB, (T) 1024KiB-1024KiB, ioengine=libaio, iodepth=32
fio-3.1
Starting 1 process
Jobs: 1 (f=1)
TEST: (groupid=0, jobs=1): err= 0: pid=1314: Sat Nov 21 12:50:45 2020
write: IOPS=4604, BW=4604MiB/s (4828MB/s)(5120MiB/1112msec)
slat (usec): min=56, max=862, avg=199.01, stdev=46.97
clat (usec): min=404, max=9098, avg=6535.65, stdev=1266.94
lat (usec): min=583, max=9334, avg=6737.31, stdev=1285.84
clat percentiles (usec):
| 1.00th=[ 1418], 5.00th=[ 5211], 10.00th=[ 5932], 20.00th=[ 6063],
| 30.00th=[ 6128], 40.00th=[ 6194], 50.00th=[ 6259], 60.00th=[ 6325],
| 70.00th=[ 6521], 80.00th=[ 7963], 90.00th=[ 8356], 95.00th=[ 8586],
| 99.00th=[ 8717], 99.50th=[ 8848], 99.90th=[ 8979], 99.95th=[ 9110],
| 99.99th=[ 9110]
bw ( MiB/s): min= 4098, max= 4980, per=98.58%, avg=4539.00, stdev=623.67, samples=2
iops : min= 4098, max= 4980, avg=4539.00, stdev=623.67, samples=2
lat (usec) : 500=0.08%, 750=0.21%, 1000=0.25%
lat (msec) : 2=1.04%, 4=2.15%, 10=96.27%
cpu : usr=68.77%, sys=30.96%, ctx=44, majf=0, minf=13
IO depths : 1=0.2%, 2=0.4%, 4=0.9%, 8=1.7%, 16=3.4%, 32=93.3%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.2%, 64=0.0%, >=64=0.0%
issued rwt: total=0,5120,0, short=0,0,0, dropped=0,0,0
latency : target=0, window=0, percentile=100.00%, depth=32
Run status group 0 (all jobs):
WRITE: bw=4604MiB/s (4828MB/s), 4604MiB/s-4604MiB/s (4828MB/s-4828MB/s), io=5120MiB (5369MB), run=1112-1112msec
Disk stats (read/write):
sda: ios=0/5006, merge=0/0, ticks=0/3788, in_queue=2992, util=87.28%
root@ubuntu:~# fio --name TEST --eta-newline=5s --filename=fio-tempfile.dat --rw=randread --size=500m --io_size=5g --blocksize=4k --ioengine=libaio --fsync=1 --iodepth=1 --direct=1 --numjobs=1 --runtime=60 --group_reporting
TEST: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1
fio-3.1
Starting 1 process
Jobs: 1 (f=1): [r(1)][11.7%][r=19.0MiB/s,w=0KiB/s][r=4876,w=0 IOPS][eta 00m:53s]
Jobs: 1 (f=1): [r(1)][21.7%][r=19.7MiB/s,w=0KiB/s][r=5033,w=0 IOPS][eta 00m:47s]
Jobs: 1 (f=1): [r(1)][31.7%][r=23.2MiB/s,w=0KiB/s][r=5942,w=0 IOPS][eta 00m:41s]
Jobs: 1 (f=1): [r(1)][41.7%][r=22.4MiB/s,w=0KiB/s][r=5735,w=0 IOPS][eta 00m:35s]
Jobs: 1 (f=1): [r(1)][51.7%][r=17.5MiB/s,w=0KiB/s][r=4487,w=0 IOPS][eta 00m:29s]
Jobs: 1 (f=1): [r(1)][61.7%][r=33.1MiB/s,w=0KiB/s][r=8478,w=0 IOPS][eta 00m:23s]
Jobs: 1 (f=1): [r(1)][71.7%][r=17.5MiB/s,w=0KiB/s][r=4484,w=0 IOPS][eta 00m:17s]
Jobs: 1 (f=1): [r(1)][81.7%][r=17.3MiB/s,w=0KiB/s][r=4419,w=0 IOPS][eta 00m:11s]
Jobs: 1 (f=1): [r(1)][91.7%][r=21.4MiB/s,w=0KiB/s][r=5469,w=0 IOPS][eta 00m:05s]
Jobs: 1 (f=1): [r(1)][100.0%][r=17.4MiB/s,w=0KiB/s][r=4449,w=0 IOPS][eta 00m:00s]
TEST: (groupid=0, jobs=1): err= 0: pid=1318: Sat Nov 21 12:52:08 2020
read: IOPS=5789, BW=22.6MiB/s (23.7MB/s)(1357MiB/60001msec)
slat (usec): min=18, max=543, avg=40.05, stdev=12.66
clat (usec): min=4, max=5360, avg=117.61, stdev=50.93
lat (usec): min=59, max=5409, avg=160.72, stdev=59.17
clat percentiles (usec):
| 1.00th=[ 50], 5.00th=[ 54], 10.00th=[ 58], 20.00th=[ 63],
| 30.00th=[ 92], 40.00th=[ 103], 50.00th=[ 116], 60.00th=[ 133],
| 70.00th=[ 161], 80.00th=[ 165], 90.00th=[ 169], 95.00th=[ 176],
| 99.00th=[ 186], 99.50th=[ 194], 99.90th=[ 494], 99.95th=[ 586],
| 99.99th=[ 701]
bw ( KiB/s): min=16592, max=46336, per=100.00%, avg=23203.89, stdev=8263.27, samples=119
iops : min= 4148, max=11584, avg=5800.97, stdev=2065.82, samples=119
lat (usec) : 10=0.01%, 20=0.01%, 50=1.30%, 100=31.81%, 250=66.74%
lat (usec) : 500=0.05%, 750=0.09%, 1000=0.01%
lat (msec) : 2=0.01%, 4=0.01%, 10=0.01%
cpu : usr=15.32%, sys=32.82%, ctx=347387, majf=0, minf=11
IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
issued rwt: total=347383,0,0, short=0,0,0, dropped=0,0,0
latency : target=0, window=0, percentile=100.00%, depth=1
Run status group 0 (all jobs):
READ: bw=22.6MiB/s (23.7MB/s), 22.6MiB/s-22.6MiB/s (23.7MB/s-23.7MB/s), io=1357MiB (1423MB), run=60001-60001msec
Disk stats (read/write):
sda: ios=346906/3, merge=0/1, ticks=37988/0, in_queue=4, util=0.01%
TEST: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1
fio-3.1
Starting 1 process
Jobs: 1 (f=1): [r(1)][11.7%][r=19.0MiB/s,w=0KiB/s][r=4876,w=0 IOPS][eta 00m:53s]
Jobs: 1 (f=1): [r(1)][21.7%][r=19.7MiB/s,w=0KiB/s][r=5033,w=0 IOPS][eta 00m:47s]
Jobs: 1 (f=1): [r(1)][31.7%][r=23.2MiB/s,w=0KiB/s][r=5942,w=0 IOPS][eta 00m:41s]
Jobs: 1 (f=1): [r(1)][41.7%][r=22.4MiB/s,w=0KiB/s][r=5735,w=0 IOPS][eta 00m:35s]
Jobs: 1 (f=1): [r(1)][51.7%][r=17.5MiB/s,w=0KiB/s][r=4487,w=0 IOPS][eta 00m:29s]
Jobs: 1 (f=1): [r(1)][61.7%][r=33.1MiB/s,w=0KiB/s][r=8478,w=0 IOPS][eta 00m:23s]
Jobs: 1 (f=1): [r(1)][71.7%][r=17.5MiB/s,w=0KiB/s][r=4484,w=0 IOPS][eta 00m:17s]
Jobs: 1 (f=1): [r(1)][81.7%][r=17.3MiB/s,w=0KiB/s][r=4419,w=0 IOPS][eta 00m:11s]
Jobs: 1 (f=1): [r(1)][91.7%][r=21.4MiB/s,w=0KiB/s][r=5469,w=0 IOPS][eta 00m:05s]
Jobs: 1 (f=1): [r(1)][100.0%][r=17.4MiB/s,w=0KiB/s][r=4449,w=0 IOPS][eta 00m:00s]
TEST: (groupid=0, jobs=1): err= 0: pid=1318: Sat Nov 21 12:52:08 2020
read: IOPS=5789, BW=22.6MiB/s (23.7MB/s)(1357MiB/60001msec)
slat (usec): min=18, max=543, avg=40.05, stdev=12.66
clat (usec): min=4, max=5360, avg=117.61, stdev=50.93
lat (usec): min=59, max=5409, avg=160.72, stdev=59.17
clat percentiles (usec):
| 1.00th=[ 50], 5.00th=[ 54], 10.00th=[ 58], 20.00th=[ 63],
| 30.00th=[ 92], 40.00th=[ 103], 50.00th=[ 116], 60.00th=[ 133],
| 70.00th=[ 161], 80.00th=[ 165], 90.00th=[ 169], 95.00th=[ 176],
| 99.00th=[ 186], 99.50th=[ 194], 99.90th=[ 494], 99.95th=[ 586],
| 99.99th=[ 701]
bw ( KiB/s): min=16592, max=46336, per=100.00%, avg=23203.89, stdev=8263.27, samples=119
iops : min= 4148, max=11584, avg=5800.97, stdev=2065.82, samples=119
lat (usec) : 10=0.01%, 20=0.01%, 50=1.30%, 100=31.81%, 250=66.74%
lat (usec) : 500=0.05%, 750=0.09%, 1000=0.01%
lat (msec) : 2=0.01%, 4=0.01%, 10=0.01%
cpu : usr=15.32%, sys=32.82%, ctx=347387, majf=0, minf=11
IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
issued rwt: total=347383,0,0, short=0,0,0, dropped=0,0,0
latency : target=0, window=0, percentile=100.00%, depth=1
Run status group 0 (all jobs):
READ: bw=22.6MiB/s (23.7MB/s), 22.6MiB/s-22.6MiB/s (23.7MB/s-23.7MB/s), io=1357MiB (1423MB), run=60001-60001msec
Disk stats (read/write):
sda: ios=346906/3, merge=0/1, ticks=37988/0, in_queue=4, util=0.01%
And here is were we first found issue on our configuration, we didn't expect crazy numbers but neither a result slower than single disk performance!
We performed the same random read test on the PVE host with much different results
root@pve2:/VM-ZFS# fio --name TEST --eta-newline=5s --filename=fio-tempfile.dat --rw=randread --size=500m --io_size=10g --blocksize=4k --ioengine=libaio --fsync=1 --iodepth=1 --direct=1 --numjobs=1 --runtime=60 --group_reporting
TEST: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1
fio-3.12
Starting 1 process
Jobs: 1 (f=1): [r(1)][31.8%][r=511MiB/s][r=131k IOPS][eta 00m:15s]
Jobs: 1 (f=1): [r(1)][61.9%][r=509MiB/s][r=130k IOPS][eta 00m:08s]
Jobs: 1 (f=1): [r(1)][90.5%][r=510MiB/s][r=131k IOPS][eta 00m:02s]
Jobs: 1 (f=1): [r(1)][100.0%][r=508MiB/s][r=130k IOPS][eta 00m:00s]
TEST: (groupid=0, jobs=1): err= 0: pid=16467: Sat Nov 21 13:06:31 2020
read: IOPS=125k, BW=489MiB/s (513MB/s)(10.0GiB/20930msec)
slat (usec): min=4, max=719, avg= 6.09, stdev= 7.91
clat (nsec): min=850, max=401984, avg=949.30, stdev=864.11
lat (usec): min=5, max=727, avg= 7.19, stdev= 8.07
clat percentiles (nsec):
| 1.00th=[ 892], 5.00th=[ 900], 10.00th=[ 908], 20.00th=[ 908],
| 30.00th=[ 916], 40.00th=[ 916], 50.00th=[ 924], 60.00th=[ 924],
| 70.00th=[ 932], 80.00th=[ 940], 90.00th=[ 948], 95.00th=[ 964],
| 99.00th=[ 1912], 99.50th=[ 1944], 99.90th=[ 2672], 99.95th=[ 2864],
| 99.99th=[ 6368]
bw ( KiB/s): min=72104, max=524296, per=99.92%, avg=500596.68, stdev=79430.22, samples=41
iops : min=18026, max=131074, avg=125149.17, stdev=19857.55, samples=41
lat (nsec) : 1000=97.19%
lat (usec) : 2=2.58%, 4=0.20%, 10=0.03%, 20=0.01%, 50=0.01%
lat (usec) : 250=0.01%, 500=0.01%
cpu : usr=26.71%, sys=72.48%, ctx=389, majf=0, minf=17
IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
issued rwts: total=2621440,0,0,0 short=0,0,0,0 dropped=0,0,0,0
latency : target=0, window=0, percentile=100.00%, depth=1
Run status group 0 (all jobs):
READ: bw=489MiB/s (513MB/s), 489MiB/s-489MiB/s (513MB/s-513MB/s), io=10.0GiB (10.7GB), run=20930-20930msec
TEST: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1
fio-3.12
Starting 1 process
Jobs: 1 (f=1): [r(1)][31.8%][r=511MiB/s][r=131k IOPS][eta 00m:15s]
Jobs: 1 (f=1): [r(1)][61.9%][r=509MiB/s][r=130k IOPS][eta 00m:08s]
Jobs: 1 (f=1): [r(1)][90.5%][r=510MiB/s][r=131k IOPS][eta 00m:02s]
Jobs: 1 (f=1): [r(1)][100.0%][r=508MiB/s][r=130k IOPS][eta 00m:00s]
TEST: (groupid=0, jobs=1): err= 0: pid=16467: Sat Nov 21 13:06:31 2020
read: IOPS=125k, BW=489MiB/s (513MB/s)(10.0GiB/20930msec)
slat (usec): min=4, max=719, avg= 6.09, stdev= 7.91
clat (nsec): min=850, max=401984, avg=949.30, stdev=864.11
lat (usec): min=5, max=727, avg= 7.19, stdev= 8.07
clat percentiles (nsec):
| 1.00th=[ 892], 5.00th=[ 900], 10.00th=[ 908], 20.00th=[ 908],
| 30.00th=[ 916], 40.00th=[ 916], 50.00th=[ 924], 60.00th=[ 924],
| 70.00th=[ 932], 80.00th=[ 940], 90.00th=[ 948], 95.00th=[ 964],
| 99.00th=[ 1912], 99.50th=[ 1944], 99.90th=[ 2672], 99.95th=[ 2864],
| 99.99th=[ 6368]
bw ( KiB/s): min=72104, max=524296, per=99.92%, avg=500596.68, stdev=79430.22, samples=41
iops : min=18026, max=131074, avg=125149.17, stdev=19857.55, samples=41
lat (nsec) : 1000=97.19%
lat (usec) : 2=2.58%, 4=0.20%, 10=0.03%, 20=0.01%, 50=0.01%
lat (usec) : 250=0.01%, 500=0.01%
cpu : usr=26.71%, sys=72.48%, ctx=389, majf=0, minf=17
IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
issued rwts: total=2621440,0,0,0 short=0,0,0,0 dropped=0,0,0,0
latency : target=0, window=0, percentile=100.00%, depth=1
Run status group 0 (all jobs):
READ: bw=489MiB/s (513MB/s), 489MiB/s-489MiB/s (513MB/s-513MB/s), io=10.0GiB (10.7GB), run=20930-20930msec
Second message with detail due to limit on 15k char per message