i guess we need an XFS megathread. I hope it turns into a long love letter instead of a horror story
Official documentation is thin right now, mostly just a block here:
https://helpcenter.veeam.com/docs/backu ... =100#linux
Maybe we can get official statements or community-based feedback this way!
So i'll start it off with some questions regarding deduplication:
Deduplication
Is XFS + deduplication supported - especially in combination with FastClone?
There are no limitations mentioned @ https://helpcenter.veeam.com/docs/backu ... =100#linux
But that doc was for ReFS and just got updated with the XFS block.
I've just opened Case #04034676 to ask this question, will update this post when i hear from them.
Why am I interested in deduplication?
15+ month retention for ~40TB. Running Incremental forever doesn't feel safe with the VMware CBT bugs and everything else that could cause corruption.
I'd love to run active fulls monthly and dedup afterwards.
How to dedup?
This was my starting point: https://strugglers.net/~andy/blog/2017/ ... ation-demo
The tool I used is this - are there other tools or better ways?
https://github.com/markfasheh/duperemove
It's in the official repos (tested on Ubuntu 18.04.4 and 19.10). Here's how I did some quick testing (deduplication worked GREAT, but (time constrained) i had no chance for running incrementals or restores with those deduplicated backups afterwards ).When given the -d option, duperemove will submit those extents for deduplication using the Linux kernel extent-same ioctl.
Code: Select all
~# apt install duperemove
~# duperemove -hdr /veeam-test-repo/Testbackup
Dedup-Performance?
It ran at a speed of >2GB/s (even touching 3GB/s) on a Cisco S3260, 42x 10TB HDD @ Raid60, Ubuntu 19.10
So yeah, should be easy to let it run once a month after the active fulls.
And because you can run duperemove with which files or folders to act on, it could be run specifically for jobs (folders) after you ran an active full.