Hi, we did some PVE Ceph performance testing, here is the result:
- random 4K write on PVE host OS: IOSP = 121K, BW = 472MiB/s, Storage: 100GB block device on Ceph
- random 4K write inside PVE VM: IOPS = 23K, BW = 90.3 MiB/s, Storage: 500GB virtual disk on the same Ceph, virtio-scsi/no cache/Kernel Native AIO/CFQ
- random 4K write inside ESXi VM: IOPS = 66.4K, BW = 259 MiB/s : 100GB virtual disk on local storage (single Micro 5100Max 1.92TB SSD, RAID pass-through mode)
My questions are:
1. Is my testing result normal?
2. IOPS is lower inside VM than on Hypervisor, is there anything to optimize it?
3. IOPS is lower inside PVE VM than inside ESXi VM, is it normal?
PVE Ceph Environment: 6 servers, each is Dell R730xd, 2 x 2696v3 CPU, 384GB DDR4 2133MHz Memory, 1*10Gb Ethernet inter-connect, 6 * Micro 5100Max 1.92TB SSD.
ESXi Environment: 1 server, same configuration as PVE Ceph server.
Test 1: on hypervisor :
fio -direct=1 -iodepth=128 -rw=randwrite -ioengine=libaio -bs=4k size=50G -numjobs=48 -runtime=120 -group_reporting -filename=/mnt/rbd/iotest name=Rand_Write_IOPS_Test
Test 2: in PVE VM:
fio -bs=4k -ioengine=libaio -iodepth=32 -numjobs=16 -direct=1 -rw=randwrite -thread -time_based -runtime=60 -refill_buffers -norandommap -randrepeat=0 -group_reporting -name=fio-randread-lat -size=500G -filename=/dev/sdb
Test 3: in ESXi VM:
fio -bs=4k -ioengine=libaio -iodepth=32 -numjobs=16 -direct=1 rw=randwrite -thread -time_based -runtime=120 -refill_buffers -norandommap -randrepeat=0 group_reporting -name=fio-randread-lat -size=50G -filename=/dev/sdb
- random 4K write on PVE host OS: IOSP = 121K, BW = 472MiB/s, Storage: 100GB block device on Ceph
- random 4K write inside PVE VM: IOPS = 23K, BW = 90.3 MiB/s, Storage: 500GB virtual disk on the same Ceph, virtio-scsi/no cache/Kernel Native AIO/CFQ
- random 4K write inside ESXi VM: IOPS = 66.4K, BW = 259 MiB/s : 100GB virtual disk on local storage (single Micro 5100Max 1.92TB SSD, RAID pass-through mode)
My questions are:
1. Is my testing result normal?
2. IOPS is lower inside VM than on Hypervisor, is there anything to optimize it?
3. IOPS is lower inside PVE VM than inside ESXi VM, is it normal?
PVE Ceph Environment: 6 servers, each is Dell R730xd, 2 x 2696v3 CPU, 384GB DDR4 2133MHz Memory, 1*10Gb Ethernet inter-connect, 6 * Micro 5100Max 1.92TB SSD.
ESXi Environment: 1 server, same configuration as PVE Ceph server.
Test 1: on hypervisor :
fio -direct=1 -iodepth=128 -rw=randwrite -ioengine=libaio -bs=4k size=50G -numjobs=48 -runtime=120 -group_reporting -filename=/mnt/rbd/iotest name=Rand_Write_IOPS_Test
Test 2: in PVE VM:
fio -bs=4k -ioengine=libaio -iodepth=32 -numjobs=16 -direct=1 -rw=randwrite -thread -time_based -runtime=60 -refill_buffers -norandommap -randrepeat=0 -group_reporting -name=fio-randread-lat -size=500G -filename=/dev/sdb
Test 3: in ESXi VM:
fio -bs=4k -ioengine=libaio -iodepth=32 -numjobs=16 -direct=1 rw=randwrite -thread -time_based -runtime=120 -refill_buffers -norandommap -randrepeat=0 group_reporting -name=fio-randread-lat -size=50G -filename=/dev/sdb