Hello,
I have a small test-setup consisting of three physical nodes, which I have equipped with four SSDs and 10GB Ethernet each. I would like to test the viability of Ceph as a storage backend for our HA-Clusters, which we are running for several customers. Currently we are using a SAN backend with a shared LVM exposed over iSCSI. I want to test two scenarios: 1) One Node goes down 2) a single SSD goes down. 2) does not seem to be a problem. Ceph recognizes the loss and starts rebalancing. I had to remove the OSD from the CRUSH map and add the replacement SSD as a new storage device, but in general everything worked as expected.
Scenario 1) is a bit more difficult. I have powered down the server instantly - no regular shutdown procedure - and after a few minutes the HA VMs on the server start on the other nodes in the clusters. Now Ceph is all red in the dashboard, but it seems I can write to it - I tested writing to the block devices with dd on the VMs and I downloaded a new iso into CephFS. Still, there is no rebalancing in process, presumably because it can't produce three replicas on different cluster nodes as initially configured. If this assumption is correct I do not understand why Ceph is still writeable - shouldn't it go into read-only mode in this condition?
Output of ceph status, anonymized:
ceph status
cluster:
id: 7f3e6609-43f4-4e6d-a6aa-e4badcbccb9a
health: HEALTH_WARN
1/3 mons down, quorum freya,odin
Degraded data redundancy: 45475/136425 objects degraded (33.333%), 90 pgs degraded, 97 pgs undersized
services:
mon: 3 daemons, quorum freya,odin(age 23m), out of quorum: thor
mgr: freya(active, since 23m), standbys: odin
mds: 1/1 daemons up, 1 standby
osd: 12 osds: 8 up (since 23m), 8 in (since 13m)
data:
volumes: 1/1 healthy
pools: 4 pools, 97 pgs
objects: 45.48k objects, 172 GiB
usage: 329 GiB used, 6.7 TiB / 7.0 TiB avail
pgs: 45475/136425 objects degraded (33.333%)
90 active+undersized+degraded
7 active+undersized
io:
client: 19 KiB/s wr, 0 op/s rd, 0 op/s wr
Yours sincerely
Stefan Malte Schumacher
I have a small test-setup consisting of three physical nodes, which I have equipped with four SSDs and 10GB Ethernet each. I would like to test the viability of Ceph as a storage backend for our HA-Clusters, which we are running for several customers. Currently we are using a SAN backend with a shared LVM exposed over iSCSI. I want to test two scenarios: 1) One Node goes down 2) a single SSD goes down. 2) does not seem to be a problem. Ceph recognizes the loss and starts rebalancing. I had to remove the OSD from the CRUSH map and add the replacement SSD as a new storage device, but in general everything worked as expected.
Scenario 1) is a bit more difficult. I have powered down the server instantly - no regular shutdown procedure - and after a few minutes the HA VMs on the server start on the other nodes in the clusters. Now Ceph is all red in the dashboard, but it seems I can write to it - I tested writing to the block devices with dd on the VMs and I downloaded a new iso into CephFS. Still, there is no rebalancing in process, presumably because it can't produce three replicas on different cluster nodes as initially configured. If this assumption is correct I do not understand why Ceph is still writeable - shouldn't it go into read-only mode in this condition?
Output of ceph status, anonymized:
ceph status
cluster:
id: 7f3e6609-43f4-4e6d-a6aa-e4badcbccb9a
health: HEALTH_WARN
1/3 mons down, quorum freya,odin
Degraded data redundancy: 45475/136425 objects degraded (33.333%), 90 pgs degraded, 97 pgs undersized
services:
mon: 3 daemons, quorum freya,odin(age 23m), out of quorum: thor
mgr: freya(active, since 23m), standbys: odin
mds: 1/1 daemons up, 1 standby
osd: 12 osds: 8 up (since 23m), 8 in (since 13m)
data:
volumes: 1/1 healthy
pools: 4 pools, 97 pgs
objects: 45.48k objects, 172 GiB
usage: 329 GiB used, 6.7 TiB / 7.0 TiB avail
pgs: 45475/136425 objects degraded (33.333%)
90 active+undersized+degraded
7 active+undersized
io:
client: 19 KiB/s wr, 0 op/s rd, 0 op/s wr
Yours sincerely
Stefan Malte Schumacher