I am currently working on testing for a new cluster I am building with three compute nodes and 3 Ceph nodes. This is my 1st go with the 4.x branch, but I have been working with Proxmox for some time (1.x-3.x). In compute node testing (compute nodes will consist of a X10SLL-F board with a single hard drive plugged directly in as AHCI) I found something very troubling. When installing onto a test drive (Samsung 1TB 7200) with both ext3 or 4 the FSYNC numbers are terrible. Buffered reads seem correct for that test drive @ 120 or so, but the FSYNC numbers are less than 60 (in ext3 it is even worse @ 25-ish). It all felt very wrong (and I wondered if it was hardware related), so I grabbed the 3.4 iso and installed it on the same hardware and the ext3 numbers were orders of magnitude better (details below).
I now have setup the following 3 identical nodes for testing:
tn01 - 4.1 ext4
tn02 - 4.1 ext3
tn03 - 3.4 ext3
I should mention all of my fstab mounts are using the defaults.
Has anyone seen this with 4.1 or have any idea what might be going on? Any help would be greatly appreciated.
I now have setup the following 3 identical nodes for testing:
tn01 - 4.1 ext4
Code:
root@tn01:~# pveversion -v
proxmox-ve: 4.1-26 (running kernel: 4.2.6-1-pve)
pve-manager: 4.1-1 (running version: 4.1-1/2f9650d4)
pve-kernel-4.2.6-1-pve: 4.2.6-26
lvm2: 2.02.116-pve2
corosync-pve: 2.3.5-2
libqb0: 0.17.2-1
pve-cluster: 4.0-29
qemu-server: 4.0-41
pve-firmware: 1.1-7
libpve-common-perl: 4.0-41
libpve-access-control: 4.0-10
libpve-storage-perl: 4.0-38
pve-libspice-server1: 0.12.5-2
vncterm: 1.2-1
pve-qemu-kvm: 2.4-17
pve-container: 1.0-32
pve-firewall: 2.0-14
pve-ha-manager: 1.0-14
ksm-control-daemon: 1.2-1
glusterfs-client: 3.5.2-2+deb8u1
lxc-pve: 1.1.5-5
lxcfs: 0.13-pve1
cgmanager: 0.39-pve1
criu: 1.6.0-1
zfsutils: 0.6.5-pve6~jessie
Code:
root@tn01:~# pveperf /var/lib/vz
CPU BOGOMIPS: 54399.60
REGEX/SECOND: 2891995
HD SIZE: 860.54 GB (/dev/mapper/pve-data)
BUFFERED READS: 128.17 MB/sec
AVERAGE SEEK TIME: 13.59 ms
FSYNCS/SECOND: 44.87
DNS EXT: 125.84 ms
DNS INT: 129.98 ms (dev.lmbx.net)
tn02 - 4.1 ext3
Code:
root@tn02:~# pveversion -v
proxmox-ve: 4.1-26 (running kernel: 4.2.6-1-pve)
pve-manager: 4.1-1 (running version: 4.1-1/2f9650d4)
pve-kernel-4.2.6-1-pve: 4.2.6-26
lvm2: 2.02.116-pve2
corosync-pve: 2.3.5-2
libqb0: 0.17.2-1
pve-cluster: 4.0-29
qemu-server: 4.0-41
pve-firmware: 1.1-7
libpve-common-perl: 4.0-41
libpve-access-control: 4.0-10
libpve-storage-perl: 4.0-38
pve-libspice-server1: 0.12.5-2
vncterm: 1.2-1
pve-qemu-kvm: 2.4-17
pve-container: 1.0-32
pve-firewall: 2.0-14
pve-ha-manager: 1.0-14
ksm-control-daemon: 1.2-1
glusterfs-client: 3.5.2-2+deb8u1
lxc-pve: 1.1.5-5
lxcfs: 0.13-pve1
cgmanager: 0.39-pve1
criu: 1.6.0-1
zfsutils: 0.6.5-pve6~jessie
Code:
root@tn02:~# pveperf /var/lib/vz
CPU BOGOMIPS: 54400.80
REGEX/SECOND: 2917401
HD SIZE: 860.54 GB (/dev/mapper/pve-data)
BUFFERED READS: 105.81 MB/sec
AVERAGE SEEK TIME: 15.50 ms
FSYNCS/SECOND: 23.63
DNS EXT: 161.34 ms
DNS INT: 121.51 ms (dev.lmbx.net)
tn03 - 3.4 ext3
Code:
root@tn03:~# pveversion -v
proxmox-ve-2.6.32: 3.4-156 (running kernel: 2.6.32-39-pve)
pve-manager: 3.4-6 (running version: 3.4-6/102d4547)
pve-kernel-2.6.32-39-pve: 2.6.32-156
lvm2: 2.02.98-pve4
clvm: 2.02.98-pve4
corosync-pve: 1.4.7-1
openais-pve: 1.1.4-3
libqb0: 0.11.1-2
redhat-cluster-pve: 3.2.0-2
resource-agents-pve: 3.9.2-4
fence-agents-pve: 4.0.10-2
pve-cluster: 3.0-17
qemu-server: 3.4-6
pve-firmware: 1.1-4
libpve-common-perl: 3.0-24
libpve-access-control: 3.0-16
libpve-storage-perl: 3.0-33
pve-libspice-server1: 0.12.4-3
vncterm: 1.1-8
vzctl: 4.0-1pve6
vzprocps: 2.0.11-2
vzquota: 3.1-2
pve-qemu-kvm: 2.2-10
ksm-control-daemon: 1.1-1
glusterfs-client: 3.5.2-1
Code:
root@tn03:~# pveperf /var/lib/vz
CPU BOGOMIPS: 54400.40
REGEX/SECOND: 2652530
HD SIZE: 860.67 GB (/dev/mapper/pve-data)
BUFFERED READS: 112.90 MB/sec
AVERAGE SEEK TIME: 18.49 ms
FSYNCS/SECOND: 2136.83
DNS EXT: 109.51 ms
DNS INT: 77.84 ms (dev.lmbx.net)
I should mention all of my fstab mounts are using the defaults.
Has anyone seen this with 4.1 or have any idea what might be going on? Any help would be greatly appreciated.