public inbox for linux-ext4@vger.kernel.org
 help / color / mirror / Atom feed
From: Hisashi Hifumi <hifumi.hisashi@oss.ntt.co.jp>
To: Nick Piggin <npiggin@suse.de>,
	Christoph Hellwig <hch@infradead.org>, Jan Kara <jack@ucw.cz>,
	linux-ext4@vger.kernel.org,
	Andrew Morton <akpm@linux-foundation.org>,
	Linus Torvalds <t
Subject: Re: [patch] fs: revert 8ab22b9a
Date: Wed, 10 Sep 2008 17:47:00 +0900	[thread overview]
Message-ID: <6.0.0.20.2.20080910170208.05de1730@172.19.0.2> (raw)
In-Reply-To: <20080910045209.GA27092@wotan.suse.de>


At 13:52 08/09/10, Nick Piggin wrote:
>
>Patch 8ab22b9a, "vfs: pagecache usage optimization for pagesize!=blocksize",
>introduces a data race that might cause uninitialized data to be exposed to
>userland. The race is conceptually the same as the one fixed for page
>uptodateness, fixed by 0ed361de.
>
>The problem is that a buffer_head flags will be set uptodate after the
>stores to bring its pagecache data uptodate[*]. This patch introduces a
>possibility to read that pagecache data if the buffer_head flag has been
>found uptodate. The problem is there are no barriers or locks ordering
>the store/store vs the load/load.
>
>To illustrate:
> CPU0: write(2) (1024 bytes)           CPU1: read(2) (1024 bytes)
> 1. allocate new pagecache page        A. locate page, not fully uptodate
> 2. copy_from_user to part of page     B. partially uptodate? load bh flags
> 3. mark that buffer uptodate          C. if yes, then copy_to_user
>
>So if the store 3 is allowed to execute before the store 2, and/or the
>load in C is allowed to execute before the load in B, then we can wind
>up loading !uptodate data.
>

>
>One way to solve this is to add barriers to the buffer head operations
>similarly to the fix for the page issue. The problem is that, unlike the
>page race, we don't actually *need* to do that if we decide not to support
>this functionality. The barriers are quite heavyweight on some
>architectures, and we haven't seen really compelling numbers in favour of
>this patch yet (a best-case microbenchmark showed some improvement of
>course, but with memory barriers we could also produce a worst-case bench
>that shows some slowdown on many architectures).

I think that adding wmb/rmb to all buffer_uptodate/set_buffer_uptodate is heavy
on some architectures using BUFFER_FNS macros, but it can be possible
to mitigate performance slowdown by minimizing memory barrier utilization.
The patch "vfs: pagecache usage optimization for pagesize!=blocksize" is now 
just for ext2/3/4, so is it not sufficient to solve the above uninitialized data
exposure problem that adding one rmb to block_is_partially_uptodate() 
and wmb to __block_commit_write() ?


  reply	other threads:[~2008-09-10  8:49 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2008-09-10  4:52 [patch] fs: revert 8ab22b9a Nick Piggin
2008-09-10  8:47 ` Hisashi Hifumi [this message]
2008-09-10 10:19   ` Nick Piggin
2008-09-11  5:28     ` Hisashi Hifumi

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=6.0.0.20.2.20080910170208.05de1730@172.19.0.2 \
    --to=hifumi.hisashi@oss.ntt.co.jp \
    --cc=akpm@linux-foundation.org \
    --cc=hch@infradead.org \
    --cc=jack@ucw.cz \
    --cc=linux-ext4@vger.kernel.org \
    --cc=npiggin@suse.de \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox