I set up a test environment and started over a few times.
What's strange is each time I restart the Ceph network, even after writing 0's to all the osd's to make sure things were cleared out - I end up with:
HEALTH_WARN 1 pgs degraded; 1 pigs stuck degraded; 64 pgs stuck unclean; 1pgs stuck undersized; 1 pgs undersized;
My Ceph Status shows:
cluster 53d58faa-b12b-4fdb-a131-21370562f573
health HEALTH_WARN
1 pgs degraded
1 pgs stuck degraded
64 pgs stuck unclean
1 pgs stuck undersized
1 pgs undersized
monmap e5: 5 mons at {0=172.16.0.60:6789/0,1=172.16.0.61:6789/0,2=172.16.0.48:6789/0,3=172.16.0.47:6789/0,4=172.16.0.46:6789/0}
election epoch 34, quorum 0,1,2,3,4 4,3,2,0,1
osdmap e65: 8 osds: 8 up, 8 in; 63 remapped pgs
pgmap v3879: 576 pgs, 2 pools, 0 bytes data, 1 objects
307 MB used, 22304 GB / 22305 GB avail
512 active+clean
41 active+remapped
22 active
1 active+undersized+degraded
What I don't understand is that I haven't populated the CephFS with any data yet - but still have a dirty FS.
For the test environment I have 3 compute nodes, 2 ceph storage nodes and 4 osd's in each storage node. All 5 nodes are active as monitors.
For my Pool I've got:
Size 2/2 | pg_num 512
All OSD's are up, All Monitors are up. Crush Map couldn't be any simpler.
Where is the 1 object coming from? Can it be removed? What can I do to get back to a Healthy Status?
What's strange is each time I restart the Ceph network, even after writing 0's to all the osd's to make sure things were cleared out - I end up with:
HEALTH_WARN 1 pgs degraded; 1 pigs stuck degraded; 64 pgs stuck unclean; 1pgs stuck undersized; 1 pgs undersized;
My Ceph Status shows:
cluster 53d58faa-b12b-4fdb-a131-21370562f573
health HEALTH_WARN
1 pgs degraded
1 pgs stuck degraded
64 pgs stuck unclean
1 pgs stuck undersized
1 pgs undersized
monmap e5: 5 mons at {0=172.16.0.60:6789/0,1=172.16.0.61:6789/0,2=172.16.0.48:6789/0,3=172.16.0.47:6789/0,4=172.16.0.46:6789/0}
election epoch 34, quorum 0,1,2,3,4 4,3,2,0,1
osdmap e65: 8 osds: 8 up, 8 in; 63 remapped pgs
pgmap v3879: 576 pgs, 2 pools, 0 bytes data, 1 objects
307 MB used, 22304 GB / 22305 GB avail
512 active+clean
41 active+remapped
22 active
1 active+undersized+degraded
What I don't understand is that I haven't populated the CephFS with any data yet - but still have a dirty FS.
For the test environment I have 3 compute nodes, 2 ceph storage nodes and 4 osd's in each storage node. All 5 nodes are active as monitors.
For my Pool I've got:
Size 2/2 | pg_num 512
All OSD's are up, All Monitors are up. Crush Map couldn't be any simpler.
Where is the 1 object coming from? Can it be removed? What can I do to get back to a Healthy Status?