I have just finished migrating my VMs on the cluster from a hdd pool to an ssd pool, but now that there are no vm disks or other proxmox related items left on the pool, it still is using ~7.3TiB on what i assume is orphaned data? This cluster is currently running PVE 8 with Ceph 18, but has been upgraded from PVE 5 with Ceph 12 over the years. Would it be safe to remove the data and pool entirely?
Bash:
# pvesm list ceph-hdd
Volid Format Type Size VMID
Bash:
# ceph df
--- RAW STORAGE ---
CLASS SIZE AVAIL USED RAW USED %RAW USED
hdd 49 TiB 42 TiB 7.3 TiB 7.3 TiB 14.75
hdd7200 176 TiB 113 TiB 64 TiB 64 TiB 36.05
ssd 196 TiB 120 TiB 76 TiB 76 TiB 38.90
TOTAL 422 TiB 275 TiB 147 TiB 147 TiB 34.89
--- POOLS ---
POOL ID PGS STORED OBJECTS USED %USED MAX AVAIL
ceph-ssd 3 2048 26 TiB 6.82M 76 TiB 45.37 31 TiB
ceph-hdd 5 128 1.8 TiB 914.93k 5.3 TiB 12.25 13 TiB
ceph-backup 8 256 21 TiB 5.47M 62 TiB 45.35 25 TiB
.mgr 9 1 1.0 GiB 253 3.1 GiB 0 31 TiB
Bash:
# rados -p ceph-hdd ls | cut -d. -f2 | sort | uniq -c
1095 0756139d8b5a1d
12 163391af52760b
17 196cac6b8b4567
1 2508fde9eb01b4
404 261da9a6d5e4ac
406792 314c206b8b4567
112 324fc695e2826e
857 392810f65ebff8
4378 3c280d6b8b4567
3 47a5d9d650eb2a
11 4b6cf3d813b381
8 4ce0f3c9987cd2
38 64ccad53c1c3fa
26 6580a1f3b9dfc7
15 7358c6940cc263
269 79455b8d0b1ae4
1106 7acf583455b9d9
682 815d8d6b8b4567
2399 8cb29960f33726
1065 8d3ebeb5c0beb6
401 c92b578bd6ead0
6848 cb7e866dfe4ee7
1823 d00922b6797e2d
10962 df1fffd4c1a80b
813 e0cd326b8b4567
2613 e618d9c943b4a2
4416 eaacd16b8b4567
1 rbd_children
1 rbd_directory
1 rbd_info
1 rbd_trash