Hello!
I'm testing Ceph setup on single node, 12 HDDs are connected via SAS9211-4i contoller:
The problem is that even one PG in default pool .mgr is in inactive state because OSD reports slow ops:
I tried to move out such OSDs from bucket, and it seems that all 12 OSDs have slow ops. They definitely cannot be all faulty.
Any help is kindly appreciated! Thanks.
I'm testing Ceph setup on single node, 12 HDDs are connected via SAS9211-4i contoller:
Code:
root@pve-1:~# ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 32.74786 root default
-3 32.74786 host pve-1
0 hdd 2.72899 osd.0 up 0 1.00000
1 hdd 2.72899 osd.1 up 1.00000 1.00000
2 hdd 2.72899 osd.2 up 1.00000 1.00000
3 hdd 2.72899 osd.3 up 1.00000 1.00000
4 hdd 2.72899 osd.4 up 1.00000 1.00000
5 hdd 2.72899 osd.5 up 1.00000 1.00000
6 hdd 2.72899 osd.6 up 1.00000 1.00000
7 hdd 2.72899 osd.7 up 1.00000 1.00000
8 hdd 2.72899 osd.8 up 0 1.00000
9 hdd 2.72899 osd.9 up 0 1.00000
10 hdd 2.72899 osd.10 up 0 1.00000
11 hdd 2.72899 osd.11 up 1.00000 1.00000
The problem is that even one PG in default pool .mgr is in inactive state because OSD reports slow ops:
Code:
root@pve-1:~# ceph -s
cluster:
id: 02e2a493-9957-45f8-a634-f2be25df74fc
health: HEALTH_WARN
Reduced data availability: 1 pg inactive
1 slow ops, oldest one blocked for 385 sec, osd.3 has slow ops
services:
mon: 1 daemons, quorum pve-1 (age 9m)
mgr: pve-1(active, since 9m)
osd: 12 osds: 12 up (since 8m), 8 in (since 6m)
data:
pools: 1 pools, 1 pgs
objects: 0 objects, 0 B
usage: 248 MiB used, 22 TiB / 22 TiB avail
pgs: 100.000% pgs unknown
1 unknown
I tried to move out such OSDs from bucket, and it seems that all 12 OSDs have slow ops. They definitely cannot be all faulty.
Any help is kindly appreciated! Thanks.