linux-btrfs.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Duncan <1i5t5.duncan@cox.net>
To: linux-btrfs@vger.kernel.org
Subject: Re: fstrim silently does nothing on dev add/dev rem'd filesystem
Date: Mon, 28 Sep 2015 07:52:39 +0000 (UTC)	[thread overview]
Message-ID: <pan$aca60$e905a348$c86187f4$a2bdc65@cox.net> (raw)
In-Reply-To: CAGfcS_n7czPgSomGo8wFFn8Zxoi4QWA0L+NtK+LreA0j3aewnQ@mail.gmail.com

Rich Freeman posted on Sun, 27 Sep 2015 23:08:49 -0400 as excerpted:

> On Sun, Sep 27, 2015 at 10:45 PM, Duncan <1i5t5.duncan@cox.net> wrote:
>> But I think part of reasoning behind the relatively low priority this
>> issue has received is that it's a low visibility issue not really
>> affecting most people running btrfs, either because they're not running
>> on ssd or because they simply don't have a particularly high awareness
>> of what trim does and thus about how it's failing to work here and what
>> that means to them.  If we get a rash of people posting on-list that
>> it's affecting them, that relative priority is likely to go up, and
>> with it the patch testing and integration schedule for the affected
>> patches.
> 
> I've never actually seen fstrim do anything on btrfs (0 bytes trimmed). 
> I stopped using it a few months ago when the news came out about all the
> issues with its implementation, and I believe my drive is still
> blacklisted anyway.
> 
> It really should be fixed, but right now that goes all around - if btrfs
> fixed it tomorrow I'd still be stuck until somebody figures out how to
> reliably do it on a Samsung 850.

FWIW, just checked here, and fstrim -v says it trims rather more than I 
expected, actually -- tho I should mention that I *am* on 4.3-rc2+, so I 
should be running those patches mentioned as going into 4.3 that at least 
partially address the problem.

On my home filesystem (actually /h here):

Btrfs raid1 on two 20-gig partitions:

# btrfs fi sh /h
Label: 'hm0238gcnx+35l0'  uuid: c9f93efc-5ae1-4d89-b501-2d7944856e98
        Total devices 2 FS bytes used 13.24GiB
        devid    1 size 20.00GiB used 14.78GiB path /dev/sda6
        devid    2 size 20.00GiB used 14.78GiB path /dev/sdb6

# btrfs fi usage /h

Overall:
    Device size:                  40.00GiB
    Device allocated:             29.56GiB
    Device unallocated:           10.44GiB
    Device missing:                  0.00B
    Used:                         26.48GiB
    Free (estimated):              6.44GiB      (min: 6.44GiB)
    Data ratio:                       2.00
    Metadata ratio:                   2.00
    Global reserve:              160.00MiB      (used: 0.00B)

Data,RAID1: Size:14.00GiB, Used:12.78GiB
   /dev/sda6      14.00GiB
   /dev/sdb6      14.00GiB

Metadata,RAID1: Size:768.00MiB, Used:476.02MiB
   /dev/sda6     768.00MiB
   /dev/sdb6     768.00MiB

System,RAID1: Size:32.00MiB, Used:16.00KiB
   /dev/sda6      32.00MiB
   /dev/sdb6      32.00MiB

Unallocated:
   /dev/sda6       5.22GiB
   /dev/sdb6       5.22GiB

# fstrim -v /h
/h: 13.3 GiB (14255448064 bytes) trimmed

OK, so I'm used to thinking in per-device sizes, IOW ratio two, here.  
And I guess that 13.3 GiB trimmed as reported by fstrim isn't divided by 
two, which is why it looks so unexpectedly large to me on a filesystem of 
20 GiB capacity, 2*20 GiB partitions in raid1.  That'd be 6.5+ GiB per 
partition trimmed, on a 20 GiB partition that's 5.22 GiB unallocated.  
That's not to bad, actually. =:^)

At first I was thinking wait a minute, 20 GiB per partition, 12.78 GiB of 
actual data on each, and 13.3 GiB trimmed.  WTF?  That doesn't add up!  
Of course not, because all those figures but trimmed are cut in half, per 
partition numbers, while trimmed is total for the filesystem.  So either 
trimmed needs cut in half too, or the other numbers need doubled to 
reflect filesystem totals.  /Then/ it makes sense! =:^)

FWIW, the SSDs are both Corsair Neutrons, 256 GB (238.5 GiB) capacity, 
partitioned up with the largest partitions 24 GiB each, but only just 
under 51% (121 GiB of the 238.5) actually partitioned, 117.5 GiB of 
unpartitioned free space.  (Upthread I said about 55% partitioned, seems 
I was a bit conservative.)

As I said upthread, that leaves the FTL plenty of room to do wear-
leveling, even without active fstrim, so working fstrim is simply "nice 
to have" for me, not anything like necessary.

So obviously my Corsair Neutrons aren't blacklisted. =:^)  I've actually 
been rather pleasantly surprised indeed with 'em, as I didn't know half 
what I do now about SSDs when I purchased them, but it seems I made a 
pretty good choice.  The controllers don't do the invisible dedup/
compression that sandforce controllers do so I can rely on what I write, 
including dup-mode copies, actually being written; they're aimed at the 
server market where consistent speed is more important than burst speed 
and the behavior demonstrates exactly that, etc.

The one bad point, and it's probably luck of the draw, is that of the 
three identical units I purchased together, from Fry's Electronics, one 
was apparently returned by someone else and I too really should have 
returned it, as it gave me problems almost from the beginning, but I 
didn't.

OTOH, I ran for several months with the bad one and one of the good ones 
paired in the raid1, using btrfs' data integrity features and scrub to 
repair the data as sectors went bad and the scrub forced the FTL to mark 
a bad sector and replace it with one of the spares, and thus have far 
more practical experience now with btrfs raid1 in failing-data 
conditions, btrfs scrub restoring from the good copy, and the behavior of 
ssds as more and more sectors go bad, along with how smartctl -A reports 
the damage, than I ever would have gotten had I simply returned the 
defective device.

In fact, I was able to watch as over time, smartctl -A dropped that stat 
line cooked value from 255 "clean", to 100 immediately as the first spare 
sector was swapped in, down thru several hundred sectors per percentage 
point, until I finally decided I had enough, at 85 percent (with a 
threshold of 36% for that value, so I was still nowhere near threshold 
fail for that value, tho the more immediate per-session read-retry-rate 
would go thru the roof and drop that cooked value to 1, with a threshold 
of 6 so definitely failing right then, any time I did a big update or 
balance and then a scrub).

I just replaced it a few days ago, less than a week, I think, with the 
still unused third one of the three, originally bought for a netbook I 
never installed it in and eventually lost, with the third one so far 
seemed every bit as healthy as the good one of the pair, no bad blocks 
reported by scrub or spare sectors reported swapped in by smartctl -A, on 
either one, so far.  Which of course was why I was able to let the bad 
one go so long, since the good one of the pair was absolutely reliable. 
=:^)

-- 
Duncan - List replies preferred.   No HTML msgs.
"Every nonfree program has a lord, a master --
and if you use the program, he is your master."  Richard Stallman


      reply	other threads:[~2015-09-28  7:52 UTC|newest]

Thread overview: 5+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2015-09-27 17:52 fstrim silently does nothing on dev add/dev rem'd filesystem Chris West
2015-09-27 18:46 ` Lutz Euler
2015-09-28  2:45 ` Duncan
2015-09-28  3:08   ` Rich Freeman
2015-09-28  7:52     ` Duncan [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='pan$aca60$e905a348$c86187f4$a2bdc65@cox.net' \
    --to=1i5t5.duncan@cox.net \
    --cc=linux-btrfs@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).