From: Pankaj Raghav <p.raghav@samsung.com>
To: Hannes Reinecke <hare@suse.de>
Cc: Matthew Wilcox <willy@infradead.org>,
<linux-fsdevel@vger.kernel.org>, <linux-block@vger.kernel.org>,
Andrew Morton <akpm@linux-foundation.org>,
Christoph Hellwig <hch@lst.de>,
Luis Chamberlain <mcgrof@kernel.org>, <p.raghav@samsung.com>,
<gost.dev@samsung.com>
Subject: Re: [PATCH 6/7] mm/filemap: allocate folios with mapping blocksize
Date: Mon, 19 Jun 2023 10:08:57 +0200 [thread overview]
Message-ID: <20230619080857.qxx5c7uaz6pm4h3m@localhost> (raw)
In-Reply-To: <20230614114637.89759-7-hare@suse.de>
[-- Attachment #1: Type: text/plain, Size: 2490 bytes --]
Hi Hannes,
On Wed, Jun 14, 2023 at 01:46:36PM +0200, Hannes Reinecke wrote:
> The mapping has an underlying blocksize (by virtue of
> mapping->host->i_blkbits), so if the mapping blocksize
> is larger than the pagesize we should allocate folios
> in the correct order.
>
Network filesystems such as 9pfs set the blkbits to be maximum data it
wants to transfer leading to unnecessary memory pressure as we will try
to allocate higher order folios(Order 5 in my setup). Isn't it better
for each filesystem to request the minimum folio order it needs for its
page cache early on? Block devices can do the same for its block cache.
I have prototype along those lines and I will it soon. This is also
something willy indicated before in a mailing list conversation.
> Signed-off-by: Hannes Reinecke <hare@suse.de>
> ---
> diff --git a/mm/readahead.c b/mm/readahead.c
> index 47afbca1d122..031935b78af7 100644
> --- a/mm/readahead.c
> +++ b/mm/readahead.c
> @@ -245,7 +245,7 @@ void page_cache_ra_unbounded(struct readahead_control *ractl,
> continue;
> }
>
> - folio = filemap_alloc_folio(gfp_mask, 0);
> + folio = filemap_alloc_folio(gfp_mask, mapping_get_order(mapping));
> if (!folio)
> break;
> if (filemap_add_folio(mapping, folio, index + i,
Did you turn on CONFIG_DEBUG_VM while testing? I don't think we are
incrementing the counter in this function correctly as this function
assumes order 0. We might need something like this:
- ractl->_nr_pages++;
+ ractl->_nr_pages += folio_nr_pages(folio);
+ i += folio_nr_pages(folio) - 1;
> @@ -806,7 +806,7 @@ void readahead_expand(struct readahead_control *ractl,
> if (folio && !xa_is_value(folio))
> return; /* Folio apparently present */
>
> - folio = filemap_alloc_folio(gfp_mask, 0);
> + folio = filemap_alloc_folio(gfp_mask, mapping_get_order(mapping));
> if (!folio)
> return;
> if (filemap_add_folio(mapping, folio, index, gfp_mask) < 0) {
> @@ -833,7 +833,7 @@ void readahead_expand(struct readahead_control *ractl,
> if (folio && !xa_is_value(folio))
> return; /* Folio apparently present */
Same here:
- ractl->_nr_pages++;
+ ractl->_nr_pages += folio_nr_pages(folio);
>
> - folio = filemap_alloc_folio(gfp_mask, 0);
> + folio = filemap_alloc_folio(gfp_mask, mapping_get_order(mapping));
> if (!folio)
> return;
> if (filemap_add_folio(mapping, folio, index, gfp_mask) < 0) {
> --
> 2.35.3
>
[-- Attachment #2: Type: text/plain, Size: 0 bytes --]
next prev parent reply other threads:[~2023-06-19 8:10 UTC|newest]
Thread overview: 40+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-06-14 11:46 [PATCH 0/7] RFC: high-order folio support for I/O Hannes Reinecke
2023-06-14 11:46 ` [PATCH 1/7] brd: use XArray instead of radix-tree to index backing pages Hannes Reinecke
2023-06-14 12:45 ` Matthew Wilcox
2023-06-14 12:50 ` Pankaj Raghav
2023-06-14 13:03 ` Hannes Reinecke
2023-06-14 11:46 ` [PATCH 2/7] brd: convert to folios Hannes Reinecke
2023-06-14 13:45 ` Matthew Wilcox
2023-06-14 13:50 ` Hannes Reinecke
2023-06-14 11:46 ` [PATCH 3/7] brd: abstract page_size conventions Hannes Reinecke
2023-06-14 11:46 ` [PATCH 4/7] brd: make sector size configurable Hannes Reinecke
2023-06-14 12:55 ` Matthew Wilcox
2023-06-14 13:02 ` Hannes Reinecke
2023-06-15 2:17 ` Dave Chinner
2023-06-15 5:55 ` Christoph Hellwig
2023-06-15 6:33 ` Hannes Reinecke
2023-06-15 6:23 ` Hannes Reinecke
2023-06-14 11:46 ` [PATCH 5/7] brd: make logical " Hannes Reinecke
2023-06-14 11:46 ` [PATCH 6/7] mm/filemap: allocate folios with mapping blocksize Hannes Reinecke
2023-06-19 8:08 ` Pankaj Raghav [this message]
2023-06-19 8:42 ` Hannes Reinecke
2023-06-19 22:57 ` Dave Chinner
2023-06-20 0:00 ` Matthew Wilcox
2023-06-20 5:57 ` Hannes Reinecke
2023-06-14 11:46 ` [PATCH 7/7] mm/readahead: align readahead down to " Hannes Reinecke
2023-06-14 13:17 ` [PATCH 0/7] RFC: high-order folio support for I/O Hannes Reinecke
2023-06-14 13:53 ` Matthew Wilcox
2023-06-14 15:06 ` Hannes Reinecke
2023-06-14 15:35 ` Hannes Reinecke
2023-06-14 17:46 ` Matthew Wilcox
2023-06-14 23:53 ` Dave Chinner
2023-06-15 6:21 ` Hannes Reinecke
2023-06-15 8:51 ` Dave Chinner
2023-06-16 16:06 ` Kent Overstreet
2023-06-15 3:44 ` Dave Chinner
2023-06-14 13:48 ` [PATCH 1/2] highmem: Add memcpy_to_folio() Matthew Wilcox (Oracle)
2023-06-14 18:38 ` kernel test robot
2023-06-14 19:30 ` kernel test robot
2023-06-15 5:58 ` Christoph Hellwig
2023-06-15 12:16 ` Matthew Wilcox
2023-06-14 13:48 ` [PATCH 2/2] highmem: Add memcpy_from_folio() Matthew Wilcox (Oracle)
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20230619080857.qxx5c7uaz6pm4h3m@localhost \
--to=p.raghav@samsung.com \
--cc=akpm@linux-foundation.org \
--cc=gost.dev@samsung.com \
--cc=hare@suse.de \
--cc=hch@lst.de \
--cc=linux-block@vger.kernel.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=mcgrof@kernel.org \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox