Hello folks,
we are working some time fine with a proxmox cluster 3.4 and ceph hammer.
Cluster have four nodes (each server = dell R620 2x xeon 26202 2x12 thread, 128 GB ram, net 2x1GB + 2x10GB Intel 720, 4x2TB spin disk on hw raid5 used for os and nas/nfs, 2x2TB spin disk for OSD, 1x800GB SSD Intel for ceph cache, 1x80GB Intel for journaling).
When we want to configure a complex VLAN configuration we encountered problem on communication between different VM on different nodes (we have found a thread on VLAN problem with debian kernel 2.6).
No deep troubleshooting activity has resolved the problem until we have upgraded proxmox to 4.0 (therefore upgraded the kernel): different VM on dofferent VLAN talk correctly to only the expected target.
Ok, then we worked to optimze the performance with Ceph with the help of the SSD (because the 80 GB SSD has been added after upgrading proxmox/ceph)
Before upgrade we have obtained good performance, for example near 800 MB/s read and near 150 MB/s write from within VM win2k8r2 using CrystalBench and similar benchmarking sw.
After upgrade we have much lower performance, for example 100-120 MB/s read and write.
Well, while troubleshooting this performance degradation we decide to update ceph to latest release to hopefully solve the problem arised.
Finally, we faced a new problem on OSD creation on SSD.
Follow ssh session to explain what happen:
----------------------
root-chiopve1:~# pveceph createosd /dev/sdc
create OSD on /dev/sdc (xfs)
Caution: invalid backup GPT header, but valid main header; regenerating
backup header from main header.
****************************************************************************
Caution: Found protective or hybrid MBR and corrupt GPT. Using GPT, but disk
verification and recovery are STRONGLY recommended.
****************************************************************************
GPT data structures destroyed! You may now partition the disk using fdisk or
other utilities.
Creating new GPT entries.
The operation has completed successfully.
Setting name!
partNum is 1
REALLY setting name!
The operation has completed successfully.
Setting name!
partNum is 0
REALLY setting name!
The operation has completed successfully.
existing superblock read failed: Input/output error
mkfs.xfs: pwrite64 failed: Input/output error
meta-data=/dev/sdc1 isize=2048 agcount=4, agsize=48510517 blks
= sectsz=4096 attr=2, projid32bit=1
= crc=0 finobt=0
data = bsize=4096 blocks=194042065, imaxpct=25
= sunit=0 swidth=0 blks
naming =version 2 bsize=4096 ascii-ci=0 ftype=0
log =internal log bsize=4096 blocks=94747, version=2
= sectsz=4096 sunit=1 blks, lazy-count=1
realtime =none extsz=4096 blocks=0, rtextents=0
ceph-disk: Error: Command '['/sbin/mkfs', '-t', 'xfs', '-f', '-i', 'size=2048', '--', '/dev/sdc1']' returned non-zero exit status 1
command 'ceph-disk prepare --zap-disk --fs-type xfs --cluster ceph --cluster-uuid 307c09a3-1643-4422-b483-d0205d36d90d /dev/sdc' failed: exit code 1
root-chiopve1:~# parted
GNU Parted 3.2
Using /dev/sda
Welcome to GNU Parted! Type 'help' to view a list of commands.
(parted) select /dev/sdc
Warning: Error fsyncing/closing /dev/sdc1: Input/output error
Retry/Ignore?
Warning: Error fsyncing/closing /dev/sdc2: Input/output error
Retry/Ignore?
----------------------
We think there isn't any hw problem with SSD drive because the same problem happen on every nodes.
Anyone has an idea to suggest to troubleshoot the problem ?
Anyone is using latest proxmox with latest ceph (0.94-7) and all work fine ?
Because web interface don't work well on Ceph configuration we suspect an incompatibility problem between proxmox and ceph.
Thanks for help !
Paul
we are working some time fine with a proxmox cluster 3.4 and ceph hammer.
Cluster have four nodes (each server = dell R620 2x xeon 26202 2x12 thread, 128 GB ram, net 2x1GB + 2x10GB Intel 720, 4x2TB spin disk on hw raid5 used for os and nas/nfs, 2x2TB spin disk for OSD, 1x800GB SSD Intel for ceph cache, 1x80GB Intel for journaling).
When we want to configure a complex VLAN configuration we encountered problem on communication between different VM on different nodes (we have found a thread on VLAN problem with debian kernel 2.6).
No deep troubleshooting activity has resolved the problem until we have upgraded proxmox to 4.0 (therefore upgraded the kernel): different VM on dofferent VLAN talk correctly to only the expected target.
Ok, then we worked to optimze the performance with Ceph with the help of the SSD (because the 80 GB SSD has been added after upgrading proxmox/ceph)
Before upgrade we have obtained good performance, for example near 800 MB/s read and near 150 MB/s write from within VM win2k8r2 using CrystalBench and similar benchmarking sw.
After upgrade we have much lower performance, for example 100-120 MB/s read and write.
Well, while troubleshooting this performance degradation we decide to update ceph to latest release to hopefully solve the problem arised.
Finally, we faced a new problem on OSD creation on SSD.
Follow ssh session to explain what happen:
----------------------
root-chiopve1:~# pveceph createosd /dev/sdc
create OSD on /dev/sdc (xfs)
Caution: invalid backup GPT header, but valid main header; regenerating
backup header from main header.
****************************************************************************
Caution: Found protective or hybrid MBR and corrupt GPT. Using GPT, but disk
verification and recovery are STRONGLY recommended.
****************************************************************************
GPT data structures destroyed! You may now partition the disk using fdisk or
other utilities.
Creating new GPT entries.
The operation has completed successfully.
Setting name!
partNum is 1
REALLY setting name!
The operation has completed successfully.
Setting name!
partNum is 0
REALLY setting name!
The operation has completed successfully.
existing superblock read failed: Input/output error
mkfs.xfs: pwrite64 failed: Input/output error
meta-data=/dev/sdc1 isize=2048 agcount=4, agsize=48510517 blks
= sectsz=4096 attr=2, projid32bit=1
= crc=0 finobt=0
data = bsize=4096 blocks=194042065, imaxpct=25
= sunit=0 swidth=0 blks
naming =version 2 bsize=4096 ascii-ci=0 ftype=0
log =internal log bsize=4096 blocks=94747, version=2
= sectsz=4096 sunit=1 blks, lazy-count=1
realtime =none extsz=4096 blocks=0, rtextents=0
ceph-disk: Error: Command '['/sbin/mkfs', '-t', 'xfs', '-f', '-i', 'size=2048', '--', '/dev/sdc1']' returned non-zero exit status 1
command 'ceph-disk prepare --zap-disk --fs-type xfs --cluster ceph --cluster-uuid 307c09a3-1643-4422-b483-d0205d36d90d /dev/sdc' failed: exit code 1
root-chiopve1:~# parted
GNU Parted 3.2
Using /dev/sda
Welcome to GNU Parted! Type 'help' to view a list of commands.
(parted) select /dev/sdc
Warning: Error fsyncing/closing /dev/sdc1: Input/output error
Retry/Ignore?
Warning: Error fsyncing/closing /dev/sdc2: Input/output error
Retry/Ignore?
----------------------
We think there isn't any hw problem with SSD drive because the same problem happen on every nodes.
Anyone has an idea to suggest to troubleshoot the problem ?
Anyone is using latest proxmox with latest ceph (0.94-7) and all work fine ?
Because web interface don't work well on Ceph configuration we suspect an incompatibility problem between proxmox and ceph.
Thanks for help !
Paul