From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-io0-f175.google.com ([209.85.223.175]:39149 "EHLO mail-io0-f175.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752092AbeEUPi3 (ORCPT ); Mon, 21 May 2018 11:38:29 -0400 Received: by mail-io0-f175.google.com with SMTP id r9-v6so14804376iod.6 for ; Mon, 21 May 2018 08:38:29 -0700 (PDT) Subject: Re: Any chance to get snapshot-aware defragmentation? To: Timofey Titovets Cc: darkbasic@linuxsystems.it, David Sterba , linux-btrfs References: <4428b2eb-796a-4c1b-8527-a05532436da4@linuxsystems.it> <20180518162051.GS6649@twin.jikos.cz> <99d57070-a1df-45ef-8f7e-df832bd7ad92@linuxsystems.it> <161bea23-f1ea-4f01-b3ea-2c5e706102a7@linuxsystems.it> <07cf6f9b-7d67-93b9-2a6f-6d031ccf5468@gmail.com> From: "Austin S. Hemmelgarn" Message-ID: <7ba873b7-48f3-ba2f-3e92-3a472e2d59f5@gmail.com> Date: Mon, 21 May 2018 11:38:28 -0400 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Sender: linux-btrfs-owner@vger.kernel.org List-ID: On 2018-05-21 09:42, Timofey Titovets wrote: > пн, 21 мая 2018 г. в 16:16, Austin S. Hemmelgarn : > >> On 2018-05-19 04:54, Niccolò Belli wrote: >>> On venerdì 18 maggio 2018 20:33:53 CEST, Austin S. Hemmelgarn wrote: >>>> With a bit of work, it's possible to handle things sanely. You can >>>> deduplicate data from snapshots, even if they are read-only (you need >>>> to pass the `-A` option to duperemove and run it as root), so it's >>>> perfectly reasonable to only defrag the main subvolume, and then >>>> deduplicate the snapshots against that (so that they end up all being >>>> reflinks to the main subvolume). Of course, this won't work if you're >>>> short on space, but if you're dealing with snapshots, you should have >>>> enough space that this will work (because even without defrag, it's >>>> fully possible for something to cause the snapshots to suddenly take >>>> up a lot more space). >>> >>> Been there, tried that. Unfortunately even if I skip the defreg a simple >>> >>> duperemove -drhA --dedupe-options=noblock --hashfile=rootfs.hash rootfs >>> >>> is going to eat more space than it was previously available (probably >>> due to autodefrag?). >> It's not autodefrag (that doesn't trigger on use of the EXTENT_SAME >> ioctl). There's two things involved here: > >> * BTRFS has somewhat odd and inefficient handling of partial extents. >> When part of an extent becomes unused (because of a CLONE ioctl, or an >> EXTENT_SAME ioctl, or something similar), that part stays allocated >> until the whole extent would be unused. >> * You're using the default deduplication block size (128k), which is >> larger than your filesystem block size (which is at most 64k, most >> likely 16k, but might be 4k if it's an old filesystem), so deduplicating >> can split extents. > > That's a metadata node leaf != fs block size. > btrfs fs block size == machine page size currently. You're right, I keep forgetting about that (probably because BTRFS is pretty much the only modern filesystem that doesn't let you change the block size). > >> Because of this, if a duplicate region happens to overlap the front of >> an already shared extent, and the end of said shared extent isn't >> aligned with the deduplication block size, the EXTENT_SAME call will >> deduplicate the first part, creating a new shared extent, but not the >> tail end of the existing shared region, and all of that original shared >> region will stick around, taking up extra space that it wasn't before. > >> Additionally, if only part of an extent is duplicated, then that area of >> the extent will stay allocated, because the rest of the extent is still >> referenced (so you won't necessarily see any actual space savings). > >> You can mitigate this by telling duperemove to use the same block size >> as your filesystem using the `-b` option. Note that using a smaller >> block size will also slow down the deduplication process and greatly >> increase the size of the hash file. > > duperemove -b control "how hash data", not more or less and only support > 4KiB..1MiB And you can only deduplicate the data at the granularity you hashed it at. In particular: * The total size of a region being deduplicated has to be an exact multiple of the hash block size (what you pass to `-b`). So for the default 128k size, you can only deduplicate regions that are multiples of 128k long (128k, 256k, 384k, 512k, etc). This is a simple limit derived from how blocks are matched for deduplication. * Because duperemove uses fixed hash blocks (as opposed to using a rolling hash window like many file synchronization tools do), the regions being deduplicated also have to be exactly aligned to the hash block size. So, with the default 128k size, you can only deduplicate regions starting at 0k, 128k, 256k, 384k, 512k, etc, but not ones starting at, for example, 64k into the file. > > And size of block for dedup will change efficiency of deduplication, > when count of hash-block pairs, will change hash file size and time > complexity. > > Let's assume that: 'A' - 1KiB of data 'AAAA' - 4KiB with repeated pattern. > > So, example, you have 2 of 2x4KiB blocks: > 1: 'AAAABBBB' > 2: 'BBBBAAAA' > > With -b 8KiB hash of first block not same as second. > But with -b 4KiB duperemove will see both 'AAAA' and 'BBBB' > And then that blocks will be deduped. This supports what I'm saying though. Your deduplication granularity is bounded by your hash granularity. If in addition to the above you have a file that looks like: AABBBAA It would not get deduplicated against the first two at either `-b 4k` or `-b 8k` despite the middle 4k of the file being an exact duplicate of the final 4k of the first file and first 4k of the second one. If instead you have: AABBBBBB And the final 6k is a single on-disk extent, that extent will get split when you go to deduplicate against the first two files with a 4k block size because only the final 4k can be deduplicated, and the entire 6k original extent will stay completely allocated. > > Even, duperemove have 2 modes of deduping: > 1. By extents > 2. By blocks Yes, you can force it to not collapse runs of duplicate blocks into single extents, but that doesn't matter for this at all, you are still limited by your hash granularity.