I have directories with old incremental backups and they are full of redundant copies of various files. My plan was to use ZFS that handles checksums of files and prevent redundancy.
So a model situation:
cd /poolname/zalohy
zfs list -p poolname
NAME USED AVAIL REFER MOUNTPOINT
poolname 995328 374734901248 98304 /poolname
for i in {0..10}; do echo {1..99999} >file$i.txt; done # I create eleven identical files of the size 588888 bytes.
zfs list -p poolname
NAME USED AVAIL REFER MOUNTPOINT
poolname 5677056 374730219520 98304 /poolname
374734901248 - 374730219520 = 4 681 728, i.e. cca 5MB.
I expected that 11 identical files (with the same checksum) would take a bit more than 588888 bytes, so ten times less.
Where is the problem. How to handle this redundancy? Is there a better file system than ZFS for this purpose?
Thanks a lot for the help.