Big fail on ceph

Orionis

Active Member
Sep 2, 2021
34
0
26
41
hello all,

first i have a cluster of to node ,
my ceph work with 6 OSD see img
Capture d’écran 2022-05-17 225030.png


today one of my osd is down and i can restart it

in fact there 3 month ago, i have replace osd 5 by a bigger ( before 1TO, now 3To); but i see my pool no upgrade, the second imgCapture d’écran 2022-05-17 225310.png

you see 5.25 to but with new 3To i will have 7 approximately. but no...
now i have exceed 5.25 i suppose this is why my osd down .

with the command ceph -w

i have this

Code:
 cluster:
    id:     112e2652-f242-4b6d-857f-ccee86a01041
    health: HEALTH_WARN
            nodown flag(s) set
            2 backfillfull osd(s)
            Low space hindering backfill (add storage if this doesn't resolve itself): 95 pgs backfill_toofull
            Degraded data redundancy: 861538/2249117 objects degraded (38.306%), 99 pgs degraded, 99 pgs undersized
            99 pgs not deep-scrubbed in time
            99 pgs not scrubbed in time
            1 pool(s) do not have an application enabled
            2 pool(s) backfillfull

  services:
    mon: 2 daemons, quorum bsg-galactica,bsg-pegasus (age 50m)
    mgr: bsg-galactica(active, since 3w)
    osd: 6 osds: 5 up (since 48m), 5 in (since 18m); 95 remapped pgs
         flags nodown

  data:
    pools:   2 pools, 129 pgs
    objects: 1.12M objects, 4.3 TiB
    usage:   5.3 TiB used, 4.0 TiB / 9.2 TiB avail
    pgs:     861538/2249117 objects degraded (38.306%)
             95 active+undersized+degraded+remapped+backfill_toofull
             30 active+clean
             4  active+undersized+degraded

my question, is i can remove the osd down, destroy and recreate it , without losing data,

or how i can restart the osd for backup and restart a 0

Thanks