After an upgrade to PVE 6 and CEPH to 14.2.4 I enabled pool mirroring to independent node (following PVE wiki)
From that time my pool usage is growing up constantly even-though no VM disk changes are made
Could anybody help to sort out where my space is flowing out?
Pool usage size is going to became critical very shortly(((
The total size of all disks of all VMs ~8Tb
However it uses almost triple size of total VM disks (with replica = 2)
Pool is defied as following:
4 nodes: 3x6Tb disks each (class hdd) + cache tier 3x480Gb SSD in writeback mode.
Ceph cluster is healthy:
on remote node mirroring is working:
How ever on cluser nodes with the same command I get:
An can see lots of journal_data on main rbd pool
Any ideas? Would very appreciate any assistance!
From that time my pool usage is growing up constantly even-though no VM disk changes are made
Could anybody help to sort out where my space is flowing out?
Pool usage size is going to became critical very shortly(((
The total size of all disks of all VMs ~8Tb
However it uses almost triple size of total VM disks (with replica = 2)
Pool is defied as following:
4 nodes: 3x6Tb disks each (class hdd) + cache tier 3x480Gb SSD in writeback mode.
Ceph cluster is healthy:
Code:
root@pve-node1:~# ceph -s
cluster:
id: c2d639ef-c720-4c85-ac77-2763ecaa0a5e
health: HEALTH_OK
services:
mon: 4 daemons, quorum pve-node3,pve-node4,pve-node2,pve-node1 (age 8d)
mgr: pve-node2(active, since 8d), standbys: pve-node3, pve-node4, pve-node1
mds: cephfs:1 {0=pve-node3=up:active}
osd: 33 osds: 33 up (since 4w), 33 in (since 4w)
data:
pools: 4 pools, 1344 pgs
objects: 5.43M objects, 28 TiB
usage: 57 TiB used, 31 TiB / 88 TiB avail
pgs: 1343 active+clean
1 active+clean+scrubbing+deep
io:
client: 3.0 MiB/s rd, 123 MiB/s wr, 253 op/s rd, 338 op/s wr
cache: 156 MiB/s flush, 9.4 MiB/s evict, 2 op/s promote
on remote node mirroring is working:
Code:
root@pve-backup:~# rbd mirror pool status rbd
health: OK
images: 18 total
18 replaying
How ever on cluser nodes with the same command I get:
Code:
root@pve-node1:~# rbd mirror pool status rbd --verbose
health: WARNING
images: 18 total
18 unknown
vm-100-disk-0:
global_id: 1beee4c3-331e-48bc-8926-fc21ff4cf00f
state: down+unknown
description: status not found
last_update:
vm-101-disk-0:
global_id: 3ec05a6e-70c5-48dc-bd6a-df3d6d3a4dc9
state: down+unknown
description: status not found
last_update:
vm-101-disk-1:
global_id: 1c32375c-28e0-4d81-aced-d10d58934ae7
state: down+unknown
description: status not found
last_update:
vm-102-disk-0:
global_id: efbd6c50-b27b-490e-95cf-10229f29a3ff
state: down+unknown
description: status not found
last_update:
vm-103-disk-0:
global_id: b62600d6-d8d0-4896-94cd-c74cc5dd4e66
state: down+unknown
description: status not found
last_update:
vm-104-disk-0:
global_id: 76adbfe9-9ca1-46cf-b40b-c75999204a41
state: down+unknown
description: status not found
last_update:
vm-104-disk-1:
global_id: 4de05037-c917-4ed1-98f5-b3d775481938
state: down+unknown
description: status not found
last_update:
vm-104-disk-2:
global_id: 38cf89e3-0c2f-4f08-ab1a-89fb44c5acc4
state: down+unknown
description: status not found
last_update:
vm-104-disk-3:
global_id: 9a5b345d-2450-4f64-9dd8-3306632a5ef8
state: down+unknown
description: status not found
last_update:
vm-105-disk-0:
global_id: 1850b05c-b54e-4218-b055-f74d9e1dfac4
state: down+unknown
description: status not found
last_update:
vm-105-disk-1:
global_id: 61ae2168-a4f3-48a8-8be7-194614e998fc
state: down+unknown
description: status not found
last_update:
vm-105-disk-2:
global_id: 3cf1311b-fa74-498a-9a8b-c60236ad1b0e
state: down+unknown
description: status not found
last_update:
vm-105-disk-3:
global_id: fb8d0ad2-962d-43f6-af81-0d1a6abeb9f6
state: down+unknown
description: status not found
last_update:
vm-106-disk-0:
global_id: 7ce1b1c3-59e4-4f1e-a934-12ace60a570c
state: down+unknown
description: status not found
last_update:
vm-107-disk-0:
global_id: c0aa5873-9e65-4b89-a53f-49a3fc351716
state: down+unknown
description: status not found
last_update:
vm-108-disk-0:
global_id: e1857775-cb9d-4bfa-b45c-5ed4ad49694e
state: down+unknown
description: status not found
last_update:
vm-108-disk-1:
global_id: fadc0e51-3f0c-4470-a3c2-6073079a1f91
state: down+unknown
description: status not found
last_update:
vm-111-disk-1:
global_id: e5a3aa06-db0e-498d-afbe-602ed0a28b53
state: down+unknown
description: status not found
last_update:
root@pve-node1:~# ceph osd pool ls detail
pool 13 'rbd' replicated size 2 min_size 1 crush_rule 1 object_hash rjenkins pg_num 512 pgp_num 512 autoscale_mode warn last_change 14784 lfor 13285/13285/13643 flags hashpspool,selfmanaged_snaps tiers 15 read_tier 15 write_tier 15 stripe_width 0 application rbd
removed_snaps [1~2d]
pool 15 'cache' replicated size 2 min_size 1 crush_rule 2 object_hash rjenkins pg_num 512 pgp_num 512 autoscale_mode warn last_change 14784 lfor 13285/13285/13285 flags hashpspool,incomplete_clones,selfmanaged_snaps tier_of 13 cache_mode writeback target_bytes 1869169767219 hit_set bloom{false_positive_probability: 0.05, target_size: 0, seed: 0} 0s x0 decay_rate 0 search_last_n 0 stripe_width 0 application rbd
removed_snaps [1~2d]
pool 16 'cephfs_data' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 256 pgp_num 256 autoscale_mode warn last_change 13340 flags hashpspool stripe_width 0 application cephfs
pool 17 'cephfs_metadata' replicated size 2 min_size 1 crush_rule 0 object_hash rjenkins pg_num 64 pgp_num 64 autoscale_mode warn last_change 13344 flags hashpspool stripe_width 0 pg_autoscale_bias 4 pg_num_min 16 recovery_priority 5 application cephfs
An can see lots of journal_data on main rbd pool
Code:
root@pve-node1:~# rados -p rbd ls
...
rbd_data.12760829ba8dbe.0000000000021599
journal_data.13.2449a28e62bde7.1684855
journal_data.13.2449a28e62bde7.1068770
rbd_data.241461564eae34.0000000000015997
journal_data.13.2449a28e62bde7.946935
rbd_data.2449a28e62bde7.00000000000047fa
journal_data.13.2449a28e62bde7.848572
journal_data.13.2449a28e62bde7.1621718
journal_data.13.2449a28e62bde7.1116442
rbd_data.12760829ba8dbe.00000000000115dc
rbd_data.12760829ba8dbe.000000000000b103
journal_data.13.2449a28e62bde7.563886
rbd_data.2449a28e62bde7.000000000004c201
rbd_data.2449a28e62bde7.000000000002c7d4
journal_data.13.2449a28e62bde7.1639724
rbd_data.23bd5e38499a26.0000000000006fb8
journal_data.13.2449a28e62bde7.1481473
journal_data.13.2449a28e62bde7.1355715
rbd_data.23bd5e38499a26.0000000000032e72
...
Any ideas? Would very appreciate any assistance!