Hi,
Since we have installing our new Ceph Cluster, we have frequently high apply latency on OSDs (near 200 ms to 1500 ms), while commit latency is continuously at 0 ms !
In Ceph documentation, when you run the command "ceph osd perf", the fs_commit_latency is generally higher than fs_apply_latency. For us it's the opposite.
The phenomenon has increased since we changed the Ceph version (migrate Giant 0.87.1 to Hammer 0.94.1)
The consequence is that our Windows VMs are very slow.
Does anyone could tell us if our configuration is good or not, and in what direction investigate ?
Different informations in our configuration :
- Proxmox 3.4-6
- kernel : 3.10.0-10-pve
- CEPH :
- Hammer 0.94.1
- 3 hosts with 3 OSDs of 4 TB (9 OSDs) + 1 SSD of 500 GB per host for journals
- 1 host with 4 OSDs of 300 GB (4 OSDs) + 1 SSD of 500 GB for journals
- OSD Tree :
- ceph.conf
Thanks.
Best regards
Since we have installing our new Ceph Cluster, we have frequently high apply latency on OSDs (near 200 ms to 1500 ms), while commit latency is continuously at 0 ms !
In Ceph documentation, when you run the command "ceph osd perf", the fs_commit_latency is generally higher than fs_apply_latency. For us it's the opposite.
The phenomenon has increased since we changed the Ceph version (migrate Giant 0.87.1 to Hammer 0.94.1)
The consequence is that our Windows VMs are very slow.
Does anyone could tell us if our configuration is good or not, and in what direction investigate ?
Code:
# ceph osd perf
osd fs_commit_latency(ms) fs_apply_latency(ms)
0 0 62
1 0 193
2 0 88
3 0 269
4 0 1055
5 0 322
6 0 272
7 0 116
8 0 653
9 0 4
10 0 1
11 0 7
12 0 4
Different informations in our configuration :
- Proxmox 3.4-6
- kernel : 3.10.0-10-pve
- CEPH :
- Hammer 0.94.1
- 3 hosts with 3 OSDs of 4 TB (9 OSDs) + 1 SSD of 500 GB per host for journals
- 1 host with 4 OSDs of 300 GB (4 OSDs) + 1 SSD of 500 GB for journals
- OSD Tree :
Code:
# ceph osd tree
ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY
-1 33.83995 root default
-6 22.91995 room salle-dr
-2 10.92000 host ceph01
0 3.64000 osd.0 up 1.00000 1.00000
2 3.64000 osd.2 up 1.00000 1.00000
1 3.64000 osd.1 up 1.00000 1.00000
-3 10.92000 host ceph02
3 3.64000 osd.3 up 1.00000 1.00000
4 3.64000 osd.4 up 1.00000 1.00000
5 3.64000 osd.5 up 1.00000 1.00000
-5 1.07996 host ceph06
9 0.26999 osd.9 up 1.00000 1.00000
10 0.26999 osd.10 up 1.00000 1.00000
11 0.26999 osd.11 up 1.00000 1.00000
12 0.26999 osd.12 up 1.00000 1.00000
-7 10.92000 room salle-log
-4 10.92000 host ceph03
6 3.64000 osd.6 up 1.00000 1.00000
7 3.64000 osd.7 up 1.00000 1.00000
8 3.64000 osd.8 up 1.00000 1.00000
- ceph.conf
Code:
[global]
auth client required = cephx
auth cluster required = cephx
auth service required = cephx
auth supported = cephx
cluster network = 10.10.1.0/24
filestore xattr use omap = true
fsid = 2dbbec32-a464-4bc5-bb2b-983695d1d0c6
keyring = /etc/pve/priv/$cluster.$name.keyring
mon osd adjust heartbeat grace = true
mon osd down out subtree limit = host
osd disk threads = 24
osd heartbeat grace = 10
osd journal size = 5120
osd max backfills = 1
osd op threads = 24
osd pool default min size = 1
osd recovery max active = 1
public network = 192.168.80.0/24
[osd]
keyring = /var/lib/ceph/osd/ceph-$id/keyring
[mon.0]
host = ceph01
mon addr = 192.168.80.41:6789
[mon.1]
host = ceph02
mon addr = 192.168.80.42:6789
[mon.2]
host = ceph03
mon addr = 192.168.80.43:6789
Thanks.
Best regards