Trying to gauge how much wiggle room I have on this cluster before new HW is required. We are already getting pretty close I think. OSD's are all 2TB.
	
	
	
		
	
	
	
		
This is running 7.4 and Ceph 17.x. I need to leave myself enough space to patch and eventually uplift to 8.x soon.. Can I safely run this up to 80% without causing issues during a patching/reboot, etc?
What useable with Ceph is something Im still learning (without putting myself into a corner).
Thanks!
				
			
		Code:
	
	# pveceph pool ls
┌─────────────────┬──────┬──────────┬────────┬─────────────┬────────────────┬───────────────────┬──────────────────────────┬───────────────────────────┬─────────────────┬──────────────────────┬────────────────┐
│ Name            │ Size │ Min Size │ PG Num │ min. PG Num │ Optimal PG Num │ PG Autoscale Mode │ PG Autoscale Target Size │ PG Autoscale Target Ratio │ Crush Rule Name │               %-Used │           Used │
╞═════════════════╪══════╪══════════╪════════╪═════════════╪════════════════╪═══════════════════╪══════════════════════════╪═══════════════════════════╪═════════════════╪══════════════════════╪════════════════╡
│ .mgr            │    3 │        2 │      1 │           1 │              1 │ on                │                          │                           │ replicated_rule │  0.00012499802687671 │     1113931776 │
├─────────────────┼──────┼──────────┼────────┼─────────────┼────────────────┼───────────────────┼──────────────────────────┼───────────────────────────┼─────────────────┼──────────────────────┼────────────────┤
│ ceph-vm         │    3 │        2 │    512 │             │            512 │ on                │                          │                           │ replicated_rule │    0.753570139408112 │ 27247801425767 │
├─────────────────┼──────┼──────────┼────────┼─────────────┼────────────────┼───────────────────┼──────────────────────────┼───────────────────────────┼─────────────────┼──────────────────────┼────────────────┤
│ cephfs_data     │    3 │        2 │     32 │             │             32 │ on                │                          │                           │ replicated_rule │   0.0265619959682226 │   243138416640 │
├─────────────────┼──────┼──────────┼────────┼─────────────┼────────────────┼───────────────────┼──────────────────────────┼───────────────────────────┼─────────────────┼──────────────────────┼────────────────┤
│ cephfs_metadata │    3 │        2 │     32 │          16 │             16 │ on                │                          │                           │ replicated_rule │ 1.26695904327789e-05 │      112893583 │
└─────────────────┴──────┴──────────┴────────┴─────────────┴────────────────┴───────────────────┴──────────────────────────┴───────────────────────────┴─────────────────┴──────────────────────┴────────────────┘
		Code:
	
	# ceph osd df tree
ID  CLASS  WEIGHT    REWEIGHT  SIZE     RAW USE  DATA     OMAP     META     AVAIL    %USE   VAR   PGS  STATUS  TYPE NAME
-1         36.67854         -   37 TiB   25 TiB   25 TiB  4.3 MiB   72 GiB   12 TiB  68.36  1.00    -          root default
-3         12.22618         -   12 TiB  8.4 TiB  8.3 TiB  1.4 MiB   24 GiB  3.9 TiB  68.36  1.00    -              host hv01
 0    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  215 KiB  3.4 GiB  595 GiB  66.72  0.98   80      up          osd.0
 3    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  199 KiB  3.5 GiB  568 GiB  68.25  1.00   83      up          osd.3
 6    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  208 KiB  3.5 GiB  551 GiB  69.18  1.01   82      up          osd.6
 9    ssd   1.74660   1.00000  1.7 TiB  1.3 TiB  1.3 TiB  209 KiB  3.9 GiB  487 GiB  72.78  1.06   91      up          osd.9
15    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  214 KiB  3.0 GiB  600 GiB  66.45  0.97   84      up          osd.15
16    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  203 KiB  3.6 GiB  567 GiB  68.29  1.00   78      up          osd.16
17    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  210 KiB  3.2 GiB  592 GiB  66.87  0.98   79      up          osd.17
-5         12.22618         -   12 TiB  8.4 TiB  8.3 TiB  1.4 MiB   25 GiB  3.9 TiB  68.37  1.00    -              host hv02
 1    ssd   1.74660   1.00000  1.7 TiB  1.3 TiB  1.3 TiB  213 KiB  3.7 GiB  488 GiB  72.69  1.06   87      up          osd.1
 4    ssd   1.74660   1.00000  1.7 TiB  1.1 TiB  1.1 TiB  209 KiB  3.5 GiB  627 GiB  64.94  0.95   81      up          osd.4
 7    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  226 KiB  3.6 GiB  554 GiB  69.00  1.01   84      up          osd.7
10    ssd   1.74660   1.00000  1.7 TiB  1.3 TiB  1.3 TiB  204 KiB  3.7 GiB  502 GiB  71.92  1.05   83      up          osd.10
14    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  226 KiB  3.1 GiB  541 GiB  69.73  1.02   80      up          osd.14
18    ssd   1.74660   1.00000  1.7 TiB  1.1 TiB  1.1 TiB  194 KiB  3.5 GiB  615 GiB  65.60  0.96   81      up          osd.18
19    ssd   1.74660   1.00000  1.7 TiB  1.1 TiB  1.1 TiB  195 KiB  3.5 GiB  631 GiB  64.70  0.95   81      up          osd.19
-7         12.22618         -   12 TiB  8.4 TiB  8.3 TiB  1.4 MiB   23 GiB  3.9 TiB  68.36  1.00    -              host hv03
 2    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  217 KiB  3.4 GiB  558 GiB  68.82  1.01   81      up          osd.2
 5    ssd   1.74660   1.00000  1.7 TiB  1.3 TiB  1.3 TiB  207 KiB  3.6 GiB  485 GiB  72.91  1.07   88      up          osd.5
 8    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  218 KiB  3.8 GiB  536 GiB  70.05  1.02   84      up          osd.8
11    ssd   1.74660   1.00000  1.7 TiB  1.1 TiB  1.1 TiB  198 KiB  3.2 GiB  613 GiB  65.73  0.96   78      up          osd.11
12    ssd   1.74660   1.00000  1.7 TiB  1.1 TiB  1.1 TiB  195 KiB  3.0 GiB  621 GiB  65.26  0.95   80      up          osd.12
13    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  202 KiB  3.1 GiB  601 GiB  66.42  0.97   82      up          osd.13
20    ssd   1.74660   1.00000  1.7 TiB  1.2 TiB  1.2 TiB  223 KiB  2.9 GiB  549 GiB  69.30  1.01   84      up          osd.20
                        TOTAL   37 TiB   25 TiB   25 TiB  4.3 MiB   72 GiB   12 TiB  68.36
MIN/MAX VAR: 0.95/1.07  STDDEV: 2.58This is running 7.4 and Ceph 17.x. I need to leave myself enough space to patch and eventually uplift to 8.x soon.. Can I safely run this up to 80% without causing issues during a patching/reboot, etc?
What useable with Ceph is something Im still learning (without putting myself into a corner).
Thanks!
 
	 
	 
 
		