From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-io0-f171.google.com ([209.85.223.171]:45118 "EHLO mail-io0-f171.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1754768AbeAIMXL (ORCPT ); Tue, 9 Jan 2018 07:23:11 -0500 Received: by mail-io0-f171.google.com with SMTP id e20so18327342iof.12 for ; Tue, 09 Jan 2018 04:23:10 -0800 (PST) Subject: Re: Recommendations for balancing as part of regular maintenance? To: Tom Worster , linux-btrfs@vger.kernel.org References: From: "Austin S. Hemmelgarn" Message-ID: <33ee8802-df5a-b153-6e24-2b38be597846@gmail.com> Date: Tue, 9 Jan 2018 07:23:06 -0500 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Sender: linux-btrfs-owner@vger.kernel.org List-ID: On 2018-01-08 16:43, Tom Worster wrote: > On 01/08/2018 04:55 PM, Austin S. Hemmelgarn wrote: > >> On 2018-01-08 11:20, ein wrote: >> >> > On 01/08/2018 04:55 PM, Austin S. Hemmelgarn wrote: >> > >> > > [...] >> > > >> > > And here's the FAQ entry: >> > > >> > > Q: Do I need to run a balance regularly? >> > > >> > > A: In general usage, no. A full unfiltered balance typically takes a >> > > long time, and will rewrite huge amounts of data unnecessarily. >> You may >> > > wish to run a balance on metadata only (see Balance_Filters) if >> you find >> > > you have very large amounts of metadata space allocated but >> unused, but >> > > this should be a last resort. >> > >> > IHMO three more sentencens and the answer would be more useful: >> > 1. BTRFS balance command example with note check the man first. >> > 2. What use case may cause 'large amounts of metadata space allocated >> > but unused'. >> >> That's kind of what I was thinking as well, but I'm hesitant to get >> too heavily into stuff along the lines of 'for use case X, do 1, for >> use case Y, do 2, etc', as that tends to result in pigeonholing >> (people just go with what sounds closest to their use case instead of >> trying to figure out what actually is best for their use case). >> >> Ideally, I think it should be as generic as reasonably possible, >> possibly something along the lines of: >> >> A: While not strictly necessary, running regular filtered balances >> (for example `btrfs balance start -dusage=50 -dlimit=2 -musage=50 >> -mlimit=4`, see `man btrfs-balance` for more info on what the options >> mean) can help keep a volume healthy by mitigating the things that >> typically cause ENOSPC errors. Full balances by contrast are long and >> expensive operations, and should be done only as a last resort. > > As the BTRFS noob who started the conversation on netdata's Github > issues, I'd like to describe my experience. > > I got an alert that unallocated space on a BTRFS filesystem on one host > was low. A netdata caption suggested btrfs-balance and directed me to > its man page. But I found it hard to understand since I don't know how > BTRFS works or its particular terminology. The FAQ was easier to > understand but didn't help me find a solution to my problem. > > It's a 420GiB NVMe with single data and metadata. It has a MariaDB > datadir with an OLTP workload and a small GlusterFS brick for > replicating filesystem with little activity. I recall that unallocated > space was under 2G, metadata allocation was low, a few G and about 1/3 > used. Data allocation was very large, almost everything else, with ~25% > used. > > Given the documentation and the usage stats, I did not know what options > to use with balance. I spent some time reading and researching and > trying to understand the filters and how they should relate to my > situation. Eventually I abandoned that effort and ran balance without > options. Hopefully the explanation I gave on the filters in the Github issue helped some. In this case though, it sounds like running a filtered balance probably wouldn't have saved you much over a full one. > > While general recommendations about running balance would be welcome, > what I needed was a dummy's guide to what the output of btrfs usage > _means_ and how to use balance to tackle problems with it. This really is a great point. Our documentation does a decent job as a reference for people who already have some idea what they're doing, but it really is worthless for people who have no prior experience. > > The other mystery is how the data allocation became so large. The most common case is that you had a lot of data on the device, and then deleted most of it. Unless a chunk becomes completely empty (either because the data that was in it becomes completely unused, or because a balance moved all the data), it won't be automatically deleted by the kernel, so it's not unusual for filesystems that have been very active (especially if they have the 'ssd' mount option set, which happens automatically on most SSD's and a lot of other things the kernel marks as not being rotational media) to have a reasonably large amount of empty space scattered around the data chunks.