in zulip, you enter "sender:me" into the search box for displaying your own posts - and you're done
in zulip, you enter "sender:me" into the search box for displaying your own posts - and you're done
no, backups 2-5 are also all the chunks needed to recreate the data at that point time. there are three optimizations/points of deduplication in place:Hi y'all,
Like several people above, I'm also trying to get used to the chunk & 'every backup is a full backup' concepts.
Let's say I have a VM in my PVE of which I have taken five backups, snapshot mode, to a PBS.
The first one is what's comparable to a full backup in the days of yore: the entire disk of the VM has been turned into chunks, which have been deduplicated where possible.
Backups 2-5 are only the chunks which contain parts of the VM disk that changed since the previous backup, again deduplicated.
if you purge the snapshots 2, 3 and 4, then you only remove the metadata of those snapshots (including the indices that tell PBS which chunks are making up the backed-up data). if a chunk is then no longer referenced by *any* snapshot, the next Garbage Collection run will remove the chunk.What I'm wondering about is: What happens if I were to delete backups 2,3 & 4? And does it make any difference whether I delete them from the client (PVE) or from PBS?
for VMs (and block devices, and other "image" type backups) the chunk size is fixed (4M of input data). for containers/directory backups, the chunk size is determined by a sliding window algorithm (that tries to create re-usable boundaries, so that the next run will find similar chunks if the input stream hasn't changed too much).(side-question: Not all chunks are the same size, right? How is chunk size determined?)
And all chunks are compressed via zstd. So the actual resulting size on the storage might be way smaller, depending on how well that data is compressible. Not uncommon that those 4MB of data of the chunk actually only needs something like 2MB on the datastore.for VMs (and block devices, and other "image" type backups) the chunk size is fixed (4M of input data). for containers/directory backups, the chunk size is determined by a sliding window algorithm (that tries to create re-usable boundaries, so that the next run will find similar chunks if the input stream hasn't changed too much).