VM gets trimmed once every day. All filesystems are trimmed.
Your wish may be fullfilled:
[root@ ~]# zpool status -v
pool: rpool
state: ONLINE
scan: scrub repaired 0B in 00:00:29 with 0 errors on Sun Dec 14 00:24:30 2025
config...
I was thinking the exact same and i'm still not certain if i figured out the correct reason just yet.
Used to be 545G throughout the whole time. Then i filled the partition up and it just kept growing nonstop.
raidz needs to store parity information somewhere (in blocks of size at least the pool's ashift, which usually nowadays is 12 so 4k). zvols by default use a very small blocksize (8k). if you write a single 8k block to the zvol, and use raidz2...
Looks totally fine to me. You're using roughly 60GB without L2ARC, rest is L2ARC usage (~ 170GB) + some remaining free GB (~ 20). I see absolutely nothing wrong here.
The graphs have different color. If you hover over these graphs it will tell...
Heya!
I've just had one of my VMs stalled because of a Proxmox "io-error". Well, seems the pool usage is 100%.
[root@~]# zfs get -p volsize,used,logicalused,compressratio vms/vm-11000-disk-0
NAME PROPERTY VALUE SOURCE...
If you have a pull request for Proxmox please be so kind to link it here so i can review/improve it before there is a chance Proxmox will merge it.
I'd add the option into the CEPH pool configuration UI because its linked on a per-pool-basis and...
rbd config pool set POOLNAME rbd_read_from_replica_policy localize
Regarding the post mortem: I had to delay my work on that because i have to deal with lots of other stuff with higher inhouse priority at the moment. Hopefully i'll be able to...