Cepth HEALTH_WARN - mon.pve241 has 21% avail

fpausp

Renowned Member
Aug 31, 2010
676
50
93
Austria near Vienna
So wie es aussieht geht mir der Platz aus...

upload_2018-11-21_13-25-32.png

Code:
2018-11-21 07:00:00.000166 mon.pve241 mon.0 10.10.10.241:6789/0 94986 : cluster [WRN] overall HEALTH_WARN mon pve241 is low on available space
2018-11-21 07:45:55.803736 mon.pve241 mon.0 10.10.10.241:6789/0 95535 : cluster [WRN] reached concerning levels of available space on local monitor storage (27% free)
2018-11-21 07:46:55.804154 mon.pve241 mon.0 10.10.10.241:6789/0 95546 : cluster [WRN] reached concerning levels of available space on local monitor storage (26% free)
2018-11-21 08:00:00.000155 mon.pve241 mon.0 10.10.10.241:6789/0 95697 : cluster [WRN] overall HEALTH_WARN mon pve241 is low on available space
2018-11-21 08:03:55.809793 mon.pve241 mon.0 10.10.10.241:6789/0 95749 : cluster [WRN] reached concerning levels of available space on local monitor storage (27% free)
2018-11-21 08:04:55.810185 mon.pve241 mon.0 10.10.10.241:6789/0 95762 : cluster [WRN] reached concerning levels of available space on local monitor storage (26% free)
2018-11-21 09:00:00.000231 mon.pve241 mon.0 10.10.10.241:6789/0 96415 : cluster [WRN] overall HEALTH_WARN mon pve241 is low on available space
2018-11-21 09:13:55.834528 mon.pve241 mon.0 10.10.10.241:6789/0 96566 : cluster [WRN] reached concerning levels of available space on local monitor storage (27% free)
2018-11-21 09:14:55.834901 mon.pve241 mon.0 10.10.10.241:6789/0 96577 : cluster [WRN] reached concerning levels of available space on local monitor storage (26% free)
2018-11-21 09:42:55.844388 mon.pve241 mon.0 10.10.10.241:6789/0 96914 : cluster [WRN] reached concerning levels of available space on local monitor storage (27% free)
2018-11-21 09:43:55.844759 mon.pve241 mon.0 10.10.10.241:6789/0 96927 : cluster [WRN] reached concerning levels of available space on local monitor storage (26% free)
2018-11-21 09:57:55.849326 mon.pve241 mon.0 10.10.10.241:6789/0 97079 : cluster [WRN] reached concerning levels of available space on local monitor storage (27% free)
2018-11-21 09:58:55.849718 mon.pve241 mon.0 10.10.10.241:6789/0 97091 : cluster [WRN] reached concerning levels of available space on local monitor storage (26% free)
2018-11-21 10:00:00.000136 mon.pve241 mon.0 10.10.10.241:6789/0 97102 : cluster [WRN] overall HEALTH_WARN mon pve241 is low on available space
2018-11-21 10:28:31.506744 mon.pve241 mon.0 10.10.10.241:6789/0 97397 : cluster [INF] osd.4 marked itself down
2018-11-21 10:28:31.506829 mon.pve241 mon.0 10.10.10.241:6789/0 97398 : cluster [INF] osd.5 marked itself down
2018-11-21 10:28:31.557928 mon.pve241 mon.0 10.10.10.241:6789/0 97399 : cluster [WRN] Health check failed: 2 osds down (OSD_DOWN)
2018-11-21 10:28:31.557974 mon.pve241 mon.0 10.10.10.241:6789/0 97400 : cluster [WRN] Health check failed: 1 host (2 osds) down (OSD_HOST_DOWN)
2018-11-21 10:28:31.557928 mon.pve241 mon.0 10.10.10.241:6789/0 97399 : cluster [WRN] Health check failed: 2 osds down (OSD_DOWN)
2018-11-21 10:28:31.557974 mon.pve241 mon.0 10.10.10.241:6789/0 97400 : cluster [WRN] Health check failed: 1 host (2 osds) down (OSD_HOST_DOWN)
2018-11-21 10:40:29.084942 mon.pve241 mon.0 10.10.10.241:6789/0 22 : cluster [INF] osd.0 10.10.10.241:6800/2028 boot
2018-11-21 10:40:30.780553 mon.pve242 mon.1 10.10.10.242:6789/0 12 : cluster [INF] mon.pve242 calling monitor election
2018-11-21 10:40:30.953322 mon.pve243 mon.2 10.10.10.243:6789/0 1 : cluster [INF] mon.pve243 calling monitor election
2018-11-21 10:40:31.056160 mon.pve241 mon.0 10.10.10.241:6789/0 34 : cluster [INF] mon.pve241 calling monitor election
2018-11-21 10:40:34.934117 mon.pve242 mon.1 10.10.10.242:6789/0 13 : cluster [WRN] message from mon.0 was stamped 0.269715s in the future, clocks not synchronized
2018-11-21 10:40:35.109841 mon.pve243 mon.2 10.10.10.243:6789/0 2 : cluster [WRN] message from mon.0 was stamped 0.094069s in the future, clocks not synchronized
2018-11-21 10:40:35.186459 mon.pve241 mon.0 10.10.10.241:6789/0 35 : cluster [INF] mon.pve241 is new leader, mons pve241,pve242,pve243 in quorum (ranks 0,1,2)
2018-11-21 10:40:35.194794 mon.pve241 mon.0 10.10.10.241:6789/0 36 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.2685s > max 0.05s
2018-11-21 10:40:35.206384 mon.pve241 mon.0 10.10.10.241:6789/0 41 : cluster [INF] Health check cleared: MON_DOWN (was: 1/3 mons down, quorum pve241,pve242)
2018-11-21 10:40:35.211665 mon.pve241 mon.0 10.10.10.241:6789/0 42 : cluster [WRN] mon.2 10.10.10.243:6789/0 clock skew 0.0830282s > max 0.05s
2018-11-21 10:40:35.252098 mon.pve241 mon.0 10.10.10.241:6789/0 45 : cluster [WRN] overall HEALTH_WARN 3 osds down; 1 host (2 osds) down; clock skew detected on mon.pve242; mon pve241 is low on available space
2018-11-21 10:40:35.275035 mon.pve241 mon.0 10.10.10.241:6789/0 46 : cluster [WRN] Health check update: 2 osds down (OSD_DOWN)
2018-11-21 10:40:35.289401 mon.pve241 mon.0 10.10.10.241:6789/0 48 : cluster [INF] osd.3 10.10.10.242:6800/1930 boot
2018-11-21 10:40:36.241472 mon.pve241 mon.0 10.10.10.241:6789/0 51 : cluster [WRN] Health check update: clock skew detected on mon.pve242, mon.pve243 (MON_CLOCK_SKEW)
2018-11-21 10:40:36.322278 mon.pve241 mon.0 10.10.10.241:6789/0 52 : cluster [WRN] Health check failed: Reduced data availability: 95 pgs inactive, 188 pgs peering (PG_AVAILABILITY)
2018-11-21 10:40:37.338576 mon.pve241 mon.0 10.10.10.241:6789/0 54 : cluster [INF] osd.2 10.10.10.242:6804/2096 boot
2018-11-21 10:40:40.935485 mon.pve243 mon.2 10.10.10.243:6789/0 7 : cluster [WRN] message from mon.0 was stamped 0.190638s in the future, clocks not synchronized
2018-11-21 10:40:41.111081 mon.pve241 mon.0 10.10.10.241:6789/0 58 : cluster [WRN] Health check failed: Degraded data redundancy: 1100/7329 objects degraded (15.009%), 56 pgs degraded (PG_DEGRADED)
2018-11-21 10:40:43.179529 mon.pve241 mon.0 10.10.10.241:6789/0 63 : cluster [WRN] Health check update: 1 osds down (OSD_DOWN)
2018-11-21 10:40:43.179569 mon.pve241 mon.0 10.10.10.241:6789/0 64 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (2 osds) down)
2018-11-21 10:40:43.194391 mon.pve241 mon.0 10.10.10.241:6789/0 65 : cluster [INF] osd.5 10.10.10.243:6800/1913 boot
2018-11-21 10:40:45.206530 mon.pve241 mon.0 10.10.10.241:6789/0 69 : cluster [WRN] message from mon.1 was stamped 0.093323s in the future, clocks not synchronized
2018-11-21 10:40:46.209890 mon.pve241 mon.0 10.10.10.241:6789/0 72 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down)
2018-11-21 10:40:46.231773 mon.pve241 mon.0 10.10.10.241:6789/0 73 : cluster [INF] osd.4 10.10.10.243:6804/2066 boot
2018-11-21 10:40:46.244388 mon.pve241 mon.0 10.10.10.241:6789/0 76 : cluster [WRN] Health check update: Reduced data availability: 128 pgs peering (PG_AVAILABILITY)
2018-11-21 10:40:46.244465 mon.pve241 mon.0 10.10.10.241:6789/0 77 : cluster [WRN] Health check update: Degraded data redundancy: 796/7329 objects degraded (10.861%), 41 pgs degraded (PG_DEGRADED)
2018-11-21 10:40:51.244973 mon.pve241 mon.0 10.10.10.241:6789/0 79 : cluster [WRN] Health check update: Degraded data redundancy: 246/7329 objects degraded (3.357%), 25 pgs degraded (PG_DEGRADED)
2018-11-21 10:40:56.245465 mon.pve241 mon.0 10.10.10.241:6789/0 83 : cluster [WRN] Health check update: Degraded data redundancy: 10/7329 objects degraded (0.136%), 9 pgs degraded (PG_DEGRADED)
2018-11-21 10:40:59.122613 mon.pve241 mon.0 10.10.10.241:6789/0 84 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 4/7329 objects degraded (0.055%), 4 pgs degraded)
2018-11-21 10:41:01.124500 mon.pve241 mon.0 10.10.10.241:6789/0 85 : cluster [WRN] Health check failed: 1 slow requests are blocked > 32 sec. Implicated osds 1 (REQUEST_SLOW)
2018-11-21 10:41:05.216166 mon.pve241 mon.0 10.10.10.241:6789/0 89 : cluster [WRN] mon.2 10.10.10.243:6789/0 clock skew 0.0754391s > max 0.05s
2018-11-21 10:41:06.073402 mon.pve243 mon.2 10.10.10.243:6789/0 17 : cluster [WRN] message from mon.0 was stamped 0.065261s in the future, clocks not synchronized
2018-11-21 10:41:06.246205 mon.pve241 mon.0 10.10.10.241:6789/0 90 : cluster [WRN] Health check update: 11 slow requests are blocked > 32 sec. Implicated osds 1 (REQUEST_SLOW)
2018-11-21 10:41:06.246498 mon.pve241 mon.0 10.10.10.241:6789/0 91 : cluster [WRN] Health check update: clock skew detected on mon.pve243 (MON_CLOCK_SKEW)
2018-11-21 10:41:11.218016 mon.pve241 mon.0 10.10.10.241:6789/0 92 : cluster [WRN] reached concerning levels of available space on local monitor storage (26% free)
2018-11-21 10:41:29.142943 mon.pve241 mon.0 10.10.10.241:6789/0 100 : cluster [WRN] Health check update: Reduced data availability: 40 pgs inactive, 128 pgs peering (PG_AVAILABILITY)
2018-11-21 10:41:37.158541 mon.pve241 mon.0 10.10.10.241:6789/0 103 : cluster [WRN] Health check update: Reduced data availability: 63 pgs inactive, 128 pgs peering (PG_AVAILABILITY)
2018-11-21 10:41:43.164426 mon.pve241 mon.0 10.10.10.241:6789/0 105 : cluster [WRN] Health check update: Reduced data availability: 96 pgs inactive, 128 pgs peering (PG_AVAILABILITY)
2018-11-21 10:41:51.251223 mon.pve241 mon.0 10.10.10.241:6789/0 106 : cluster [WRN] Health check update: Reduced data availability: 128 pgs inactive, 128 pgs peering (PG_AVAILABILITY)
2018-11-21 10:42:06.252540 mon.pve241 mon.0 10.10.10.241:6789/0 110 : cluster [INF] Health check cleared: MON_CLOCK_SKEW (was: clock skew detected on mon.pve243)
2018-11-21 10:42:35.189323 mon.pve241 mon.0 10.10.10.241:6789/0 120 : cluster [WRN] Health check update: 12 slow requests are blocked > 32 sec. Implicated osds 1 (REQUEST_SLOW)
2018-11-21 11:00:00.000218 mon.pve241 mon.0 10.10.10.241:6789/0 366 : cluster [WRN] overall HEALTH_WARN Reduced data availability: 128 pgs inactive, 128 pgs peering; 12 slow requests are blocked > 32 sec. Implicated osds 1; mon pve241 is low on available space
2018-11-21 11:10:36.411496 mon.pve241 mon.0 10.10.10.241:6789/0 538 : cluster [INF] osd.1 marked down after no beacon for 900.076937 seconds
2018-11-21 11:10:36.412382 mon.pve241 mon.0 10.10.10.241:6789/0 539 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN)
2018-11-21 11:10:38.425771 mon.pve241 mon.0 10.10.10.241:6789/0 542 : cluster [WRN] Health check update: Reduced data availability: 109 pgs inactive, 96 pgs peering (PG_AVAILABILITY)
2018-11-21 11:10:38.425807 mon.pve241 mon.0 10.10.10.241:6789/0 543 : cluster [WRN] Health check failed: Degraded data redundancy: 349/7329 objects degraded (4.762%), 10 pgs degraded (PG_DEGRADED)
2018-11-21 11:10:38.425832 mon.pve241 mon.0 10.10.10.241:6789/0 544 : cluster [INF] Health check cleared: REQUEST_SLOW (was: 12 slow requests are blocked > 32 sec. Implicated osds 1)
2018-11-21 11:10:40.459929 mon.pve241 mon.0 10.10.10.241:6789/0 545 : cluster [WRN] Health check failed: 12 slow requests are blocked > 32 sec. Implicated osds 1 (REQUEST_SLOW)
2018-11-21 11:10:44.613457 mon.pve241 mon.0 10.10.10.241:6789/0 548 : cluster [WRN] Health check update: Degraded data redundancy: 1359/7329 objects degraded (18.543%), 72 pgs degraded (PG_DEGRADED)
2018-11-21 11:10:44.613512 mon.pve241 mon.0 10.10.10.241:6789/0 549 : cluster [INF] Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 13 pgs inactive)
2018-11-21 11:10:51.415162 mon.pve241 mon.0 10.10.10.241:6789/0 550 : cluster [WRN] Health check update: Degraded data redundancy: 1354/7329 objects degraded (18.475%), 72 pgs degraded (PG_DEGRADED)
2018-11-21 11:11:08.593148 mon.pve241 mon.0 10.10.10.241:6789/0 560 : cluster [INF] Health check cleared: REQUEST_SLOW (was: 12 slow requests are blocked > 32 sec. Implicated osds 1)
2018-11-21 11:11:10.603147 mon.pve241 mon.0 10.10.10.241:6789/0 561 : cluster [WRN] Health check failed: 12 slow requests are blocked > 32 sec. Implicated osds 1 (REQUEST_SLOW)
2018-11-21 11:11:38.494856 mon.pve241 mon.0 10.10.10.241:6789/0 568 : cluster [WRN] Health check update: Degraded data redundancy: 1354/7329 objects degraded (18.475%), 72 pgs degraded, 128 pgs undersized (PG_DEGRADED)
2018-11-21 11:13:11.228719 mon.pve241 mon.0 10.10.10.241:6789/0 588 : cluster [WRN] reached concerning levels of available space on local monitor storage (25% free)
2018-11-21 11:14:11.229110 mon.pve241 mon.0 10.10.10.241:6789/0 598 : cluster [WRN] reached concerning levels of available space on local monitor storage (21% free)
2018-11-21 11:15:11.827837 mon.pve241 mon.0 10.10.10.241:6789/0 613 : cluster [INF] osd.0 marked itself down
2018-11-21 11:15:12.667166 mon.pve241 mon.0 10.10.10.241:6789/0 614 : cluster [WRN] Health check update: 2 osds down (OSD_DOWN)
2018-11-21 11:15:12.667205 mon.pve241 mon.0 10.10.10.241:6789/0 615 : cluster [WRN] Health check failed: 1 host (2 osds) down (OSD_HOST_DOWN)
2018-11-21 11:15:24.664517 mon.pve242 mon.1 10.10.10.242:6789/0 427 : cluster [INF] mon.pve242 calling monitor election
2018-11-21 11:15:24.706179 mon.pve243 mon.2 10.10.10.243:6789/0 385 : cluster [INF] mon.pve243 calling monitor election
2018-11-21 11:15:29.668765 mon.pve242 mon.1 10.10.10.242:6789/0 428 : cluster [INF] mon.pve242 is new leader, mons pve242,pve243 in quorum (ranks 1,2)
2018-11-21 11:15:29.676424 mon.pve242 mon.1 10.10.10.242:6789/0 433 : cluster [WRN] Health check failed: 1/3 mons down, quorum pve242,pve243 (MON_DOWN)
2018-11-21 11:15:29.676470 mon.pve242 mon.1 10.10.10.242:6789/0 434 : cluster [INF] Health check cleared: MON_DISK_LOW (was: mon pve241 is low on available space)
2018-11-21 11:15:29.684371 mon.pve242 mon.1 10.10.10.242:6789/0 436 : cluster [WRN] overall HEALTH_WARN 2 osds down; 1 host (2 osds) down; Degraded data redundancy: 1354/7329 objects degraded (18.475%), 72 pgs degraded, 128 pgs undersized; 12 slow requests are blocked > 32 sec. Implicated osds 1; 1/3 mons down, quorum pve242,pve243
2018-11-21 11:15:52.279739 mon.pve242 mon.1 10.10.10.242:6789/0 447 : cluster [INF] mon.pve242 calling monitor election
2018-11-21 11:15:52.321092 mon.pve243 mon.2 10.10.10.243:6789/0 391 : cluster [INF] mon.pve243 calling monitor election
2018-11-21 11:15:52.322070 mon.pve242 mon.1 10.10.10.242:6789/0 448 : cluster [WRN] message from mon.0 was stamped 0.680893s in the future, clocks not synchronized
2018-11-21 11:15:52.362662 mon.pve243 mon.2 10.10.10.243:6789/0 392 : cluster [WRN] message from mon.0 was stamped 0.640419s in the future, clocks not synchronized
2018-11-21 11:15:52.925478 mon.pve241 mon.0 10.10.10.241:6789/0 1 : cluster [INF] mon.pve241 calling monitor election
2018-11-21 11:15:52.966399 mon.pve241 mon.0 10.10.10.241:6789/0 2 : cluster [INF] mon.pve241 calling monitor election
2018-11-21 11:15:52.975798 mon.pve241 mon.0 10.10.10.241:6789/0 3 : cluster [INF] mon.pve241 is new leader, mons pve241,pve242,pve243 in quorum (ranks 0,1,2)
2018-11-21 11:15:52.992420 mon.pve241 mon.0 10.10.10.241:6789/0 4 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.681076s > max 0.05s
2018-11-21 11:15:52.992657 mon.pve241 mon.0 10.10.10.241:6789/0 5 : cluster [WRN] mon.2 10.10.10.243:6789/0 clock skew 0.639708s > max 0.05s
2018-11-21 11:15:53.010933 mon.pve241 mon.0 10.10.10.241:6789/0 10 : cluster [WRN] Health check failed: clock skew detected on mon.pve242, mon.pve243 (MON_CLOCK_SKEW)
2018-11-21 11:15:53.011039 mon.pve241 mon.0 10.10.10.241:6789/0 11 : cluster [WRN] Health check failed: mon pve241 is low on available space (MON_DISK_LOW)
2018-11-21 11:15:53.011283 mon.pve241 mon.0 10.10.10.241:6789/0 12 : cluster [INF] Health check cleared: MON_DOWN (was: 1/3 mons down, quorum pve242,pve243)
2018-11-21 11:15:53.015085 mon.pve241 mon.0 10.10.10.241:6789/0 13 : cluster [INF] Active manager daemon pve241 restarted
2018-11-21 11:15:53.015212 mon.pve241 mon.0 10.10.10.241:6789/0 14 : cluster [INF] Activating manager daemon pve241
2018-11-21 11:15:53.027367 mon.pve241 mon.0 10.10.10.241:6789/0 15 : cluster [WRN] overall HEALTH_WARN 2 osds down; 1 host (2 osds) down; Degraded data redundancy: 1354/7329 objects degraded (18.475%), 72 pgs degraded, 128 pgs undersized; 12 slow requests are blocked > 32 sec. Implicated osds 1; clock skew detected on mon.pve242, mon.pve243; mon pve241 is low on available space
2018-11-21 11:15:54.677312 mon.pve241 mon.0 10.10.10.241:6789/0 18 : cluster [INF] Manager daemon pve241 is now available
2018-11-21 11:15:56.123928 mon.pve241 mon.0 10.10.10.241:6789/0 20 : cluster [WRN] Health check update: Degraded data redundancy: 2443/7329 objects degraded (33.333%), 128 pgs degraded, 128 pgs undersized (PG_DEGRADED)
2018-11-21 11:15:56.123967 mon.pve241 mon.0 10.10.10.241:6789/0 21 : cluster [INF] Health check cleared: REQUEST_SLOW (was: 12 slow requests are blocked > 32 sec. Implicated osds 1)
2018-11-21 11:15:57.921107 mon.pve241 mon.0 10.10.10.241:6789/0 22 : cluster [INF] Manager daemon pve241 is unresponsive.  No standby daemons available.
2018-11-21 11:15:57.921198 mon.pve241 mon.0 10.10.10.241:6789/0 23 : cluster [WRN] Health check failed: no active mgr (MGR_DOWN)
2018-11-21 11:15:59.434449 mon.pve241 mon.0 10.10.10.241:6789/0 25 : cluster [INF] Activating manager daemon pve243
2018-11-21 11:15:59.486501 mon.pve241 mon.0 10.10.10.241:6789/0 26 : cluster [INF] Health check cleared: MGR_DOWN (was: no active mgr)
2018-11-21 11:15:59.647215 mon.pve241 mon.0 10.10.10.241:6789/0 28 : cluster [INF] Manager daemon pve243 is now available
2018-11-21 11:16:01.726555 mon.pve241 mon.0 10.10.10.241:6789/0 33 : cluster [WRN] Health check update: 1 osds down (OSD_DOWN)
2018-11-21 11:16:01.726604 mon.pve241 mon.0 10.10.10.241:6789/0 34 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (2 osds) down)
2018-11-21 11:16:01.814382 mon.pve241 mon.0 10.10.10.241:6789/0 35 : cluster [INF] osd.0 10.10.10.241:6801/1984 boot
2018-11-21 11:16:02.814073 mon.pve241 mon.0 10.10.10.241:6789/0 41 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down)
2018-11-21 11:16:02.853923 mon.pve241 mon.0 10.10.10.241:6789/0 42 : cluster [INF] osd.1 10.10.10.241:6805/2143 boot
2018-11-21 11:16:07.484744 mon.pve241 mon.0 10.10.10.241:6789/0 47 : cluster [WRN] Health check update: Degraded data redundancy: 495/7329 objects degraded (6.754%), 63 pgs degraded, 40 pgs undersized (PG_DEGRADED)
2018-11-21 11:16:12.922177 mon.pve241 mon.0 10.10.10.241:6789/0 50 : cluster [WRN] Health check update: Degraded data redundancy: 113/7329 objects degraded (1.542%), 49 pgs degraded (PG_DEGRADED)
2018-11-21 11:16:17.922606 mon.pve241 mon.0 10.10.10.241:6789/0 52 : cluster [WRN] Health check update: Degraded data redundancy: 87/7329 objects degraded (1.187%), 37 pgs degraded (PG_DEGRADED)
2018-11-21 11:16:22.923194 mon.pve241 mon.0 10.10.10.241:6789/0 55 : cluster [WRN] Health check update: Degraded data redundancy: 54/7329 objects degraded (0.737%), 23 pgs degraded (PG_DEGRADED)
2018-11-21 11:16:27.923609 mon.pve241 mon.0 10.10.10.241:6789/0 58 : cluster [WRN] Health check update: Degraded data redundancy: 34/7329 objects degraded (0.464%), 14 pgs degraded (PG_DEGRADED)
2018-11-21 11:16:27.923927 mon.pve241 mon.0 10.10.10.241:6789/0 59 : cluster [INF] Health check cleared: MON_CLOCK_SKEW (was: clock skew detected on mon.pve242, mon.pve243)
2018-11-21 11:16:32.925684 mon.pve241 mon.0 10.10.10.241:6789/0 61 : cluster [WRN] Health check update: Degraded data redundancy: 11/7329 objects degraded (0.150%), 4 pgs degraded (PG_DEGRADED)
2018-11-21 11:16:37.093427 mon.pve241 mon.0 10.10.10.241:6789/0 63 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 11/7329 objects degraded (0.150%), 4 pgs degraded)
2018-11-21 11:16:52.899859 mon.pve241 mon.0 10.10.10.241:6789/0 67 : cluster [WRN] reached concerning levels of available space on local monitor storage (21% free)
2018-11-21 11:19:58.949369 mon.pve241 mon.0 10.10.10.241:6789/0 103 : cluster [WRN] message from mon.2 was stamped 0.050202s in the future, clocks not synchronized
2018-11-21 11:20:04.686406 mon.pve241 mon.0 10.10.10.241:6789/0 105 : cluster [WRN] message from mon.2 was stamped 0.050403s in the future, clocks not synchronized
2018-11-21 11:25:10.430307 mon.pve241 mon.0 10.10.10.241:6789/0 162 : cluster [INF] osd.3 marked itself down
2018-11-21 11:25:10.430546 mon.pve241 mon.0 10.10.10.241:6789/0 163 : cluster [INF] osd.2 marked itself down
2018-11-21 11:25:10.482542 mon.pve241 mon.0 10.10.10.241:6789/0 164 : cluster [WRN] Health check failed: 2 osds down (OSD_DOWN)
2018-11-21 11:25:10.482580 mon.pve241 mon.0 10.10.10.241:6789/0 165 : cluster [WRN] Health check failed: 1 host (2 osds) down (OSD_HOST_DOWN)
2018-11-21 11:25:11.648484 mon.pve241 mon.0 10.10.10.241:6789/0 169 : cluster [WRN] Health check failed: Reduced data availability: 12 pgs inactive, 59 pgs peering (PG_AVAILABILITY)
2018-11-21 11:25:22.703740 mon.pve241 mon.0 10.10.10.241:6789/0 172 : cluster [INF] mon.pve241 calling monitor election
2018-11-21 11:25:22.744516 mon.pve243 mon.2 10.10.10.243:6789/0 509 : cluster [INF] mon.pve243 calling monitor election
2018-11-21 11:25:27.717729 mon.pve241 mon.0 10.10.10.241:6789/0 173 : cluster [INF] mon.pve241 is new leader, mons pve241,pve243 in quorum (ranks 0,2)
2018-11-21 11:25:27.732035 mon.pve241 mon.0 10.10.10.241:6789/0 178 : cluster [WRN] Health check failed: 1/3 mons down, quorum pve241,pve243 (MON_DOWN)
2018-11-21 11:25:27.742095 mon.pve241 mon.0 10.10.10.241:6789/0 179 : cluster [WRN] overall HEALTH_WARN 2 osds down; 1 host (2 osds) down; Reduced data availability: 12 pgs inactive, 59 pgs peering; Degraded data redundancy: 671/7329 objects degraded (9.155%), 35 pgs degraded; mon pve241 is low on available space; 1/3 mons down, quorum pve241,pve243
2018-11-21 11:25:28.733047 mon.pve241 mon.0 10.10.10.241:6789/0 181 : cluster [WRN] Health check update: Degraded data redundancy: 2443/7329 objects degraded (33.333%), 128 pgs degraded (PG_DEGRADED)
2018-11-21 11:25:28.733097 mon.pve241 mon.0 10.10.10.241:6789/0 182 : cluster [INF] Health check cleared: PG_AVAILABILITY (was: Reduced data availability: 12 pgs inactive, 59 pgs peering)
2018-11-21 11:25:47.276909 mon.pve241 mon.0 10.10.10.241:6789/0 190 : cluster [INF] mon.pve241 calling monitor election
2018-11-21 11:25:47.304152 mon.pve241 mon.0 10.10.10.241:6789/0 191 : cluster [INF] mon.pve241 is new leader, mons pve241,pve242,pve243 in quorum (ranks 0,1,2)
2018-11-21 11:25:47.320215 mon.pve241 mon.0 10.10.10.241:6789/0 196 : cluster [INF] Health check cleared: MON_DOWN (was: 1/3 mons down, quorum pve241,pve243)
2018-11-21 11:25:47.322227 mon.pve241 mon.0 10.10.10.241:6789/0 197 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 1.12593s > max 0.05s
2018-11-21 11:25:47.323349 mon.pve241 mon.0 10.10.10.241:6789/0 198 : cluster [WRN] message from mon.1 was stamped 1.146361s in the future, clocks not synchronized
2018-11-21 11:25:47.323405 mon.pve243 mon.2 10.10.10.243:6789/0 516 : cluster [INF] mon.pve243 calling monitor election
2018-11-21 11:25:47.333755 mon.pve241 mon.0 10.10.10.241:6789/0 199 : cluster [WRN] overall HEALTH_WARN 2 osds down; 1 host (2 osds) down; Degraded data redundancy: 2443/7329 objects degraded (33.333%), 128 pgs degraded; mon pve241 is low on available space
2018-11-21 11:25:47.973007 mon.pve241 mon.0 10.10.10.241:6789/0 200 : cluster [WRN] Health check failed: clock skew detected on mon.pve242 (MON_CLOCK_SKEW)
2018-11-21 11:25:48.418753 mon.pve242 mon.1 10.10.10.242:6789/0 1 : cluster [INF] mon.pve242 calling monitor election
2018-11-21 11:25:48.780695 mon.pve241 mon.0 10.10.10.241:6789/0 201 : cluster [INF] osd.5 marked itself down
2018-11-21 11:25:48.784734 mon.pve241 mon.0 10.10.10.241:6789/0 202 : cluster [INF] osd.4 marked itself down
2018-11-21 11:25:49.056407 mon.pve241 mon.0 10.10.10.241:6789/0 203 : cluster [WRN] Health check update: 4 osds down (OSD_DOWN)
2018-11-21 11:25:49.056446 mon.pve241 mon.0 10.10.10.241:6789/0 204 : cluster [WRN] Health check update: 2 hosts (4 osds) down (OSD_HOST_DOWN)
2018-11-21 11:26:29.671249 mon.pve241 mon.0 10.10.10.241:6789/0 233 : cluster [INF] Active manager daemon pve243 restarted
2018-11-21 11:26:29.671334 mon.pve241 mon.0 10.10.10.241:6789/0 234 : cluster [INF] Activating manager daemon pve243
2018-11-21 11:26:30.132758 mon.pve241 mon.0 10.10.10.241:6789/0 236 : cluster [INF] mon.pve241 calling monitor election
2018-11-21 11:26:30.140227 mon.pve241 mon.0 10.10.10.241:6789/0 237 : cluster [INF] mon.pve241 is new leader, mons pve241,pve242,pve243 in quorum (ranks 0,1,2)
2018-11-21 11:26:30.148586 mon.pve242 mon.1 10.10.10.242:6789/0 11 : cluster [INF] mon.pve242 calling monitor election
2018-11-21 11:26:30.153863 mon.pve241 mon.0 10.10.10.241:6789/0 242 : cluster [INF] Health check cleared: MON_CLOCK_SKEW (was: clock skew detected on mon.pve242)
2018-11-21 11:26:30.153921 mon.pve241 mon.0 10.10.10.241:6789/0 243 : cluster [INF] Health check cleared: MON_DOWN (was: 1/3 mons down, quorum pve241,pve242)
2018-11-21 11:26:30.156311 mon.pve241 mon.0 10.10.10.241:6789/0 244 : cluster [WRN] mon.2 10.10.10.243:6789/0 clock skew 0.837373s > max 0.05s
2018-11-21 11:26:30.163692 mon.pve241 mon.0 10.10.10.241:6789/0 245 : cluster [WRN] overall HEALTH_WARN 2 osds down; 1 host (2 osds) down; Degraded data redundancy: 2443/7329 objects degraded (33.333%), 128 pgs degraded; mon pve241 is low on available space
2018-11-21 11:26:30.984903 mon.pve243 mon.2 10.10.10.243:6789/0 1 : cluster [INF] mon.pve243 calling monitor election
2018-11-21 11:26:32.053158 mon.pve241 mon.0 10.10.10.241:6789/0 249 : cluster [INF] Manager daemon pve243 is now available
2018-11-21 11:26:32.986266 mon.pve241 mon.0 10.10.10.241:6789/0 250 : cluster [WRN] Health check failed: clock skew detected on mon.pve243 (MON_CLOCK_SKEW)
2018-11-21 11:26:39.958960 mon.pve241 mon.0 10.10.10.241:6789/0 255 : cluster [WRN] Health check update: 1 osds down (OSD_DOWN)
2018-11-21 11:26:39.959001 mon.pve241 mon.0 10.10.10.241:6789/0 256 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (2 osds) down)
2018-11-21 11:26:39.966404 mon.pve241 mon.0 10.10.10.241:6789/0 257 : cluster [INF] osd.5 10.10.10.243:6801/1925 boot
2018-11-21 11:26:42.042172 mon.pve241 mon.0 10.10.10.241:6789/0 263 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down)
2018-11-21 11:26:42.053604 mon.pve241 mon.0 10.10.10.241:6789/0 264 : cluster [INF] osd.4 10.10.10.243:6805/2084 boot
2018-11-21 11:26:44.059419 mon.pve241 mon.0 10.10.10.241:6789/0 267 : cluster [WRN] Health check update: Degraded data redundancy: 912/7329 objects degraded (12.444%), 48 pgs degraded (PG_DEGRADED)
2018-11-21 11:26:47.945184 mon.pve241 mon.0 10.10.10.241:6789/0 269 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 507/7329 objects degraded (6.918%), 26 pgs degraded)
2018-11-21 11:26:52.907256 mon.pve241 mon.0 10.10.10.241:6789/0 271 : cluster [WRN] reached concerning levels of available space on local monitor storage (21% free)
2018-11-21 11:27:02.990380 mon.pve241 mon.0 10.10.10.241:6789/0 274 : cluster [INF] Health check cleared: MON_CLOCK_SKEW (was: clock skew detected on mon.pve243)
2018-11-21 11:34:22.826795 mon.pve241 mon.0 10.10.10.241:6789/0 380 : cluster [INF] osd.4 marked itself down
2018-11-21 11:34:22.830751 mon.pve241 mon.0 10.10.10.241:6789/0 381 : cluster [INF] osd.5 marked itself down
2018-11-21 11:34:22.878192 mon.pve241 mon.0 10.10.10.241:6789/0 382 : cluster [WRN] Health check failed: 2 osds down (OSD_DOWN)
2018-11-21 11:34:22.878241 mon.pve241 mon.0 10.10.10.241:6789/0 383 : cluster [WRN] Health check failed: 1 host (2 osds) down (OSD_HOST_DOWN)
2018-11-21 11:39:32.720127 mon.pve242 mon.1 10.10.10.242:6789/0 18 : cluster [INF] mon.pve242 calling monitor election
2018-11-21 11:39:32.733422 mon.pve241 mon.0 10.10.10.241:6789/0 58 : cluster [INF] mon.pve241 calling monitor election
2018-11-21 11:39:32.750802 mon.pve243 mon.2 10.10.10.243:6789/0 1 : cluster [INF] mon.pve243 calling monitor election
2018-11-21 11:39:35.995324 mon.pve241 mon.0 10.10.10.241:6789/0 59 : cluster [INF] mon.pve241 is new leader, mons pve241,pve242,pve243 in quorum (ranks 0,1,2)
2018-11-21 11:39:36.003491 mon.pve241 mon.0 10.10.10.241:6789/0 64 : cluster [INF] Health check cleared: MON_DOWN (was: 1/3 mons down, quorum pve241,pve242)
2018-11-21 11:39:36.037494 mon.pve241 mon.0 10.10.10.241:6789/0 67 : cluster [WRN] overall HEALTH_WARN 2 osds down; 1 host (2 osds) down; Degraded data redundancy: 2443/7329 objects degraded (33.333%), 128 pgs degraded; mon pve241 is low on available space
2018-11-21 11:39:37.114141 mon.pve241 mon.0 10.10.10.241:6789/0 70 : cluster [WRN] Health check update: 1 osds down (OSD_DOWN)
2018-11-21 11:39:37.114203 mon.pve241 mon.0 10.10.10.241:6789/0 71 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (2 osds) down)
2018-11-21 11:39:37.161817 mon.pve241 mon.0 10.10.10.241:6789/0 72 : cluster [INF] osd.4 10.10.10.243:6804/2068 boot
2018-11-21 11:39:38.140934 mon.pve241 mon.0 10.10.10.241:6789/0 76 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down)
2018-11-21 11:39:38.142279 mon.pve241 mon.0 10.10.10.241:6789/0 77 : cluster [WRN] Health check update: Degraded data redundancy: 2443/7329 objects degraded (33.333%), 128 pgs degraded, 256 pgs undersized (PG_DEGRADED)
2018-11-21 11:39:38.147556 mon.pve241 mon.0 10.10.10.241:6789/0 78 : cluster [INF] osd.5 10.10.10.243:6800/1907 boot
2018-11-21 11:39:40.075837 mon.pve241 mon.0 10.10.10.241:6789/0 81 : cluster [WRN] reached concerning levels of available space on local monitor storage (21% free)
2018-11-21 11:39:44.256138 mon.pve241 mon.0 10.10.10.241:6789/0 88 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 912/7329 objects degraded (12.444%), 48 pgs degraded, 91 pgs undersized)
2018-11-21 12:00:00.000176 mon.pve241 mon.0 10.10.10.241:6789/0 505 : cluster [WRN] overall HEALTH_WARN mon pve241 is low on available space
2018-11-21 12:58:34.931859 mon.pve242 mon.1 10.10.10.242:6789/0 1030 : cluster [WRN] message from mon.0 was stamped 0.050089s in the future, clocks not synchronized
2018-11-21 12:59:36.003885 mon.pve241 mon.0 10.10.10.241:6789/0 1361 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.0520737s > max 0.05s
2018-11-21 12:59:40.777796 mon.pve241 mon.0 10.10.10.241:6789/0 1362 : cluster [WRN] Health check failed: clock skew detected on mon.pve242 (MON_CLOCK_SKEW)
2018-11-21 13:00:00.000191 mon.pve241 mon.0 10.10.10.241:6789/0 1371 : cluster [WRN] overall HEALTH_WARN clock skew detected on mon.pve242; mon pve241 is low on available space
2018-11-21 13:00:06.005106 mon.pve241 mon.0 10.10.10.241:6789/0 1374 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.052572s > max 0.05s
2018-11-21 13:00:40.203750 mon.pve242 mon.1 10.10.10.242:6789/0 1053 : cluster [WRN] message from mon.0 was stamped 0.051901s in the future, clocks not synchronized
2018-11-21 13:01:06.006274 mon.pve241 mon.0 10.10.10.241:6789/0 1385 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.0535985s > max 0.05s
2018-11-21 13:02:36.007438 mon.pve241 mon.0 10.10.10.241:6789/0 1411 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.0552215s > max 0.05s
2018-11-21 13:04:36.008662 mon.pve241 mon.0 10.10.10.241:6789/0 1440 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.057396s > max 0.05s
2018-11-21 13:07:06.009907 mon.pve241 mon.0 10.10.10.241:6789/0 1479 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.0602925s > max 0.05s
2018-11-21 13:10:06.010965 mon.pve241 mon.0 10.10.10.241:6789/0 1521 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.0639492s > max 0.05s
2018-11-21 13:11:05.508035 mon.pve242 mon.1 10.10.10.242:6789/0 1166 : cluster [WRN] message from mon.0 was stamped 0.063312s in the future, clocks not synchronized
2018-11-21 13:13:36.012478 mon.pve241 mon.0 10.10.10.241:6789/0 1576 : cluster [WRN] mon.1 10.10.10.242:6789/0 clock skew 0.0678975s > max 0.05s
Logs

Wie sollte man das angehen ?
 
Hi,

ich hatte das mal auf einem Testcluster.
Da waren aber nicht die OSDs voll sondern eine Partition.

Was sagt denn: ceph osd df?
Oder df -lh?

Hab dann alte Kernel weggeschmissen und es lief wieder. Schau erstmal.
Gruß Thomas
 
ceph osd df
Code:
root@pve241:~# ceph osd df
ID CLASS WEIGHT  REWEIGHT SIZE    USE     AVAIL   %USE VAR  PGS
 0   hdd 0.97609  1.00000 1000GiB 4.50GiB  995GiB 0.45 0.91 128
 1   hdd 0.97609  1.00000 1000GiB 5.40GiB  994GiB 0.54 1.09 128
 2   hdd 0.97609  1.00000 1000GiB 5.03GiB  994GiB 0.50 1.02 134
 3   hdd 0.97609  1.00000 1000GiB 4.87GiB  995GiB 0.49 0.98 122
 4   hdd 0.97609  1.00000 1000GiB 5.44GiB  994GiB 0.54 1.10 133
 5   hdd 0.97609  1.00000 1000GiB 4.46GiB  995GiB 0.45 0.90 123
                    TOTAL 5.86TiB 29.7GiB 5.83TiB 0.50
MIN/MAX VAR: 0.90/1.10  STDDEV: 0.04

Code:
root@pve241:~# df -lh
Filesystem            Size  Used Avail Use% Mounted on
udev                  5.8G     0  5.8G   0% /dev
tmpfs                 1.2G  8.8M  1.2G   1% /run
/dev/mapper/pve-root  7.6G  5.6G  1.7G  77% /
tmpfs                 5.8G   63M  5.7G   2% /dev/shm
tmpfs                 5.0M     0  5.0M   0% /run/lock
tmpfs                 5.8G     0  5.8G   0% /sys/fs/cgroup
/dev/fuse              30M   28K   30M   1% /etc/pve
/dev/sdc1            1000G  4.5G  996G   1% /var/lib/ceph/osd/ceph-0
/dev/sdd1            1000G  5.5G  995G   1% /var/lib/ceph/osd/ceph-1
tmpfs                 1.2G     0  1.2G   0% /run/user/0
 
Das sieht mir eher nach alten rumliegenden Kerneln aus.
Lass Dir mal anzeigen, welche noch installiert sind.
dpkg -l | grep linux-image

lass Dir den aktuellen anzeigen. Der muss erhalten bleiben!
uname -r

Und Du kannst wahrscheinlich alte deinstallieren über:
apt remove <kernel>

Das schafft Platz.
 
Habs jetzt mal durch das löschen von ISO-Files am pve241 Node vorerst gelöst... Wie man sieht ist das rootfs nicht sehr groß, vermutlich wird das Problem wieder auftauchen. Wie pflegt man denn eigentlich den Ceph ?

upload_2018-11-21_15-21-37.png
 
dpkg -l | grep linux-image
Code:
root@pve241:/# dpkg -l | grep pve-kernel-
ii  pve-kernel-4.15                      5.2-12                         all          Latest Proxmox VE Kernel Image
ii  pve-kernel-4.15.17-1-pve             4.15.17-9                      amd64        The Proxmox PVE Kernel Image
ii  pve-kernel-4.15.18-2-pve             4.15.18-21                     amd64        The Proxmox PVE Kernel Image
ii  pve-kernel-4.15.18-4-pve             4.15.18-23                     amd64        The Proxmox PVE Kernel Image
ii  pve-kernel-4.15.18-5-pve             4.15.18-24                     amd64        The Proxmox PVE Kernel Image
ii  pve-kernel-4.15.18-7-pve             4.15.18-27                     amd64        The Proxmox PVE Kernel Image
ii  pve-kernel-4.15.18-8-pve             4.15.18-28                     amd64        The Proxmox PVE Kernel Image
ii  pve-kernel-4.15.18-9-pve             4.15.18-30                     amd64        The Proxmox PVE Kernel Image

apt remove <kernel>
Wieviele kernel sollte man drauf lassen ?
 
Hab die letzten drei Kernel drauf gelassen...

Das schaut ja gleich viel besser aus:
Code:
root@pve241:~# df -h
Filesystem            Size  Used Avail Use% Mounted on
udev                  5.8G     0  5.8G   0% /dev
tmpfs                 1.2G  8.8M  1.2G   1% /run
/dev/mapper/pve-root  7.6G  2.6G  4.6G  37% /
tmpfs                 5.8G   36M  5.7G   1% /dev/shm
tmpfs                 5.0M     0  5.0M   0% /run/lock
tmpfs                 5.8G     0  5.8G   0% /sys/fs/cgroup
/dev/fuse              30M   28K   30M   1% /etc/pve
/dev/sdd1            1000G  5.4G  995G   1% /var/lib/ceph/osd/ceph-1
tmpfs                 1.2G     0  1.2G   0% /run/user/0
/dev/sdc1            1000G  4.5G  996G   1% /var/lib/ceph/osd/ceph-0

Danke für die Infos !