From: David Hildenbrand <david@redhat.com>
To: Joanne Koong <joannelkoong@gmail.com>, miklos@szeredi.hu
Cc: linux-fsdevel@vger.kernel.org, jlayton@kernel.org,
jefflexu@linux.alibaba.com, josef@toxicpanda.com,
bernd.schubert@fastmail.fm, willy@infradead.org,
kernel-team@meta.com
Subject: Re: [PATCH v5 10/11] fuse: optimize direct io large folios processing
Date: Fri, 4 Jul 2025 12:24:01 +0200 [thread overview]
Message-ID: <6d9c08dd-c1d0-48bd-aacb-b4300f87d525@redhat.com> (raw)
In-Reply-To: <20250426000828.3216220-11-joannelkoong@gmail.com>
On 26.04.25 02:08, Joanne Koong wrote:
> Optimize processing folios larger than one page size for the direct io
> case. If contiguous pages are part of the same folio, collate the
> processing instead of processing each page in the folio separately.
>
> Signed-off-by: Joanne Koong <joannelkoong@gmail.com>
> Reviewed-by: Jeff Layton <jlayton@kernel.org>
> ---
> fs/fuse/file.c | 55 +++++++++++++++++++++++++++++++++++++-------------
> 1 file changed, 41 insertions(+), 14 deletions(-)
>
> diff --git a/fs/fuse/file.c b/fs/fuse/file.c
> index 9a31f2a516b9..61eaec1c993b 100644
> --- a/fs/fuse/file.c
> +++ b/fs/fuse/file.c
> @@ -1490,7 +1490,8 @@ static int fuse_get_user_pages(struct fuse_args_pages *ap, struct iov_iter *ii,
> }
>
> while (nbytes < *nbytesp && nr_pages < max_pages) {
> - unsigned nfolios, i;
> + struct folio *prev_folio = NULL;
> + unsigned npages, i;
> size_t start;
>
> ret = iov_iter_extract_pages(ii, &pages,
> @@ -1502,23 +1503,49 @@ static int fuse_get_user_pages(struct fuse_args_pages *ap, struct iov_iter *ii,
>
> nbytes += ret;
>
> - nfolios = DIV_ROUND_UP(ret + start, PAGE_SIZE);
> + npages = DIV_ROUND_UP(ret + start, PAGE_SIZE);
>
> - for (i = 0; i < nfolios; i++) {
> - struct folio *folio = page_folio(pages[i]);
> - unsigned int offset = start +
> - (folio_page_idx(folio, pages[i]) << PAGE_SHIFT);
> - unsigned int len = min_t(unsigned int, ret, PAGE_SIZE - start);
> + /*
> + * We must check each extracted page. We can't assume every page
> + * in a large folio is used. For example, userspace may mmap() a
> + * file PROT_WRITE, MAP_PRIVATE, and then store to the middle of
> + * a large folio, in which case the extracted pages could be
> + *
> + * folio A page 0
> + * folio A page 1
> + * folio B page 0
> + * folio A page 3
> + *
> + * where folio A belongs to the file and folio B is an anonymous
> + * COW page.
> + */
> + for (i = 0; i < npages && ret; i++) {
> + struct folio *folio;
> + unsigned int offset;
> + unsigned int len;
> +
> + WARN_ON(!pages[i]);
> + folio = page_folio(pages[i]);
> +
> + len = min_t(unsigned int, ret, PAGE_SIZE - start);
> +
> + if (folio == prev_folio && pages[i] != pages[i - 1]) {
I don't really understand the "pages[i] != pages[i - 1]" part.
Why would you have to equal page pointers in there?
Something that might be simpler to understand and implement would be using
num_pages_contiguous()
from
https://lore.kernel.org/all/20250704062602.33500-2-lizhe.67@bytedance.com/T/#u
and then just making sure that we don't exceed the current folio, if we
ever get contiguous pages that cross a folio.
--
Cheers,
David / dhildenb
next prev parent reply other threads:[~2025-07-04 10:24 UTC|newest]
Thread overview: 31+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-04-26 0:08 [PATCH v5 00/11] fuse: support large folios Joanne Koong
2025-04-26 0:08 ` [PATCH v5 01/11] fuse: support copying " Joanne Koong
2025-05-04 18:05 ` Bernd Schubert
2025-04-26 0:08 ` [PATCH v5 02/11] fuse: support large folios for retrieves Joanne Koong
2025-05-04 18:07 ` Bernd Schubert
2025-04-26 0:08 ` [PATCH v5 03/11] fuse: refactor fuse_fill_write_pages() Joanne Koong
2025-04-28 5:32 ` Dan Carpenter
2025-04-28 22:10 ` Joanne Koong
2025-05-04 18:08 ` Bernd Schubert
2025-04-26 0:08 ` [PATCH v5 04/11] fuse: support large folios for writethrough writes Joanne Koong
2025-05-04 18:40 ` Bernd Schubert
2025-05-05 21:36 ` Joanne Koong
2025-04-26 0:08 ` [PATCH v5 05/11] fuse: support large folios for folio reads Joanne Koong
2025-05-04 18:58 ` Bernd Schubert
2025-04-26 0:08 ` [PATCH v5 06/11] fuse: support large folios for symlinks Joanne Koong
2025-05-04 19:04 ` Bernd Schubert
2025-04-26 0:08 ` [PATCH v5 07/11] fuse: support large folios for stores Joanne Koong
2025-04-26 0:08 ` [PATCH v5 08/11] fuse: support large folios for queued writes Joanne Koong
2025-05-04 19:08 ` Bernd Schubert
2025-04-26 0:08 ` [PATCH v5 09/11] fuse: support large folios for readahead Joanne Koong
2025-05-04 19:13 ` Bernd Schubert
2025-05-05 14:40 ` Darrick J. Wong
2025-05-05 15:23 ` Bernd Schubert
2025-05-05 22:05 ` Joanne Koong
2025-04-26 0:08 ` [PATCH v5 10/11] fuse: optimize direct io large folios processing Joanne Koong
2025-05-04 19:15 ` Bernd Schubert
2025-07-04 10:24 ` David Hildenbrand [this message]
2025-07-07 23:27 ` Joanne Koong
2025-07-08 16:05 ` David Hildenbrand
2025-07-08 23:14 ` Joanne Koong
2025-04-26 0:08 ` [PATCH v5 11/11] fuse: support large folios for writeback Joanne Koong
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=6d9c08dd-c1d0-48bd-aacb-b4300f87d525@redhat.com \
--to=david@redhat.com \
--cc=bernd.schubert@fastmail.fm \
--cc=jefflexu@linux.alibaba.com \
--cc=jlayton@kernel.org \
--cc=joannelkoong@gmail.com \
--cc=josef@toxicpanda.com \
--cc=kernel-team@meta.com \
--cc=linux-fsdevel@vger.kernel.org \
--cc=miklos@szeredi.hu \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).