From: "Darrick J. Wong" <djwong@kernel.org>
To: Christoph Hellwig <hch@lst.de>
Cc: Christian Brauner <brauner@kernel.org>,
Joanne Koong <joannelkoong@gmail.com>,
linux-xfs@vger.kernel.org, linux-fsdevel@vger.kernel.org,
linux-doc@vger.kernel.org, linux-block@vger.kernel.org,
gfs2@lists.linux.dev
Subject: Re: [PATCH 12/12] iomap: build the writeback code without CONFIG_BLOCK
Date: Wed, 2 Jul 2025 11:20:10 -0700 [thread overview]
Message-ID: <20250702182010.GM10009@frogsfrogsfrogs> (raw)
In-Reply-To: <20250627070328.975394-13-hch@lst.de>
On Fri, Jun 27, 2025 at 09:02:45AM +0200, Christoph Hellwig wrote:
> Allow fuse to use the iomap writeback code even when CONFIG_BLOCK is
> not enabled. Do this with an ifdef instead of a separate file to keep
> the iomap_folio_state local to buffered-io.c.
>
> Signed-off-by: Christoph Hellwig <hch@lst.de>
Seems reasonable to me...
Reviewed-by: "Darrick J. Wong" <djwong@kernel.org>
--D
> ---
> fs/iomap/Makefile | 6 +--
> fs/iomap/buffered-io.c | 113 ++++++++++++++++++++++-------------------
> 2 files changed, 64 insertions(+), 55 deletions(-)
>
> diff --git a/fs/iomap/Makefile b/fs/iomap/Makefile
> index 69e8ebb41302..f7e1c8534c46 100644
> --- a/fs/iomap/Makefile
> +++ b/fs/iomap/Makefile
> @@ -9,9 +9,9 @@ ccflags-y += -I $(src) # needed for trace events
> obj-$(CONFIG_FS_IOMAP) += iomap.o
>
> iomap-y += trace.o \
> - iter.o
> -iomap-$(CONFIG_BLOCK) += buffered-io.o \
> - direct-io.o \
> + iter.o \
> + buffered-io.o
> +iomap-$(CONFIG_BLOCK) += direct-io.o \
> ioend.o \
> fiemap.o \
> seek.o
> diff --git a/fs/iomap/buffered-io.c b/fs/iomap/buffered-io.c
> index 1a9ade77aeeb..6ceeb0e2fc13 100644
> --- a/fs/iomap/buffered-io.c
> +++ b/fs/iomap/buffered-io.c
> @@ -284,6 +284,46 @@ static void iomap_adjust_read_range(struct inode *inode, struct folio *folio,
> *lenp = plen;
> }
>
> +static inline bool iomap_block_needs_zeroing(const struct iomap_iter *iter,
> + loff_t pos)
> +{
> + const struct iomap *srcmap = iomap_iter_srcmap(iter);
> +
> + return srcmap->type != IOMAP_MAPPED ||
> + (srcmap->flags & IOMAP_F_NEW) ||
> + pos >= i_size_read(iter->inode);
> +}
> +
> +/**
> + * iomap_read_inline_data - copy inline data into the page cache
> + * @iter: iteration structure
> + * @folio: folio to copy to
> + *
> + * Copy the inline data in @iter into @folio and zero out the rest of the folio.
> + * Only a single IOMAP_INLINE extent is allowed at the end of each file.
> + * Returns zero for success to complete the read, or the usual negative errno.
> + */
> +static int iomap_read_inline_data(const struct iomap_iter *iter,
> + struct folio *folio)
> +{
> + const struct iomap *iomap = iomap_iter_srcmap(iter);
> + size_t size = i_size_read(iter->inode) - iomap->offset;
> + size_t offset = offset_in_folio(folio, iomap->offset);
> +
> + if (folio_test_uptodate(folio))
> + return 0;
> +
> + if (WARN_ON_ONCE(size > iomap->length))
> + return -EIO;
> + if (offset > 0)
> + ifs_alloc(iter->inode, folio, iter->flags);
> +
> + folio_fill_tail(folio, offset, iomap->inline_data, size);
> + iomap_set_range_uptodate(folio, offset, folio_size(folio) - offset);
> + return 0;
> +}
> +
> +#ifdef CONFIG_BLOCK
> static void iomap_finish_folio_read(struct folio *folio, size_t off,
> size_t len, int error)
> {
> @@ -323,45 +363,6 @@ struct iomap_readpage_ctx {
> struct readahead_control *rac;
> };
>
> -/**
> - * iomap_read_inline_data - copy inline data into the page cache
> - * @iter: iteration structure
> - * @folio: folio to copy to
> - *
> - * Copy the inline data in @iter into @folio and zero out the rest of the folio.
> - * Only a single IOMAP_INLINE extent is allowed at the end of each file.
> - * Returns zero for success to complete the read, or the usual negative errno.
> - */
> -static int iomap_read_inline_data(const struct iomap_iter *iter,
> - struct folio *folio)
> -{
> - const struct iomap *iomap = iomap_iter_srcmap(iter);
> - size_t size = i_size_read(iter->inode) - iomap->offset;
> - size_t offset = offset_in_folio(folio, iomap->offset);
> -
> - if (folio_test_uptodate(folio))
> - return 0;
> -
> - if (WARN_ON_ONCE(size > iomap->length))
> - return -EIO;
> - if (offset > 0)
> - ifs_alloc(iter->inode, folio, iter->flags);
> -
> - folio_fill_tail(folio, offset, iomap->inline_data, size);
> - iomap_set_range_uptodate(folio, offset, folio_size(folio) - offset);
> - return 0;
> -}
> -
> -static inline bool iomap_block_needs_zeroing(const struct iomap_iter *iter,
> - loff_t pos)
> -{
> - const struct iomap *srcmap = iomap_iter_srcmap(iter);
> -
> - return srcmap->type != IOMAP_MAPPED ||
> - (srcmap->flags & IOMAP_F_NEW) ||
> - pos >= i_size_read(iter->inode);
> -}
> -
> static int iomap_readpage_iter(struct iomap_iter *iter,
> struct iomap_readpage_ctx *ctx)
> {
> @@ -554,6 +555,27 @@ void iomap_readahead(struct readahead_control *rac, const struct iomap_ops *ops)
> }
> EXPORT_SYMBOL_GPL(iomap_readahead);
>
> +static int iomap_read_folio_range(const struct iomap_iter *iter,
> + struct folio *folio, loff_t pos, size_t len)
> +{
> + const struct iomap *srcmap = iomap_iter_srcmap(iter);
> + struct bio_vec bvec;
> + struct bio bio;
> +
> + bio_init(&bio, srcmap->bdev, &bvec, 1, REQ_OP_READ);
> + bio.bi_iter.bi_sector = iomap_sector(srcmap, pos);
> + bio_add_folio_nofail(&bio, folio, len, offset_in_folio(folio, pos));
> + return submit_bio_wait(&bio);
> +}
> +#else
> +static int iomap_read_folio_range(const struct iomap_iter *iter,
> + struct folio *folio, loff_t pos, size_t len)
> +{
> + WARN_ON_ONCE(1);
> + return -EIO;
> +}
> +#endif /* CONFIG_BLOCK */
> +
> /*
> * iomap_is_partially_uptodate checks whether blocks within a folio are
> * uptodate or not.
> @@ -667,19 +689,6 @@ iomap_write_failed(struct inode *inode, loff_t pos, unsigned len)
> pos + len - 1);
> }
>
> -static int iomap_read_folio_range(const struct iomap_iter *iter,
> - struct folio *folio, loff_t pos, size_t len)
> -{
> - const struct iomap *srcmap = iomap_iter_srcmap(iter);
> - struct bio_vec bvec;
> - struct bio bio;
> -
> - bio_init(&bio, srcmap->bdev, &bvec, 1, REQ_OP_READ);
> - bio.bi_iter.bi_sector = iomap_sector(srcmap, pos);
> - bio_add_folio_nofail(&bio, folio, len, offset_in_folio(folio, pos));
> - return submit_bio_wait(&bio);
> -}
> -
> static int __iomap_write_begin(const struct iomap_iter *iter,
> const struct iomap_write_ops *write_ops, size_t len,
> struct folio *folio)
> --
> 2.47.2
>
>
prev parent reply other threads:[~2025-07-02 18:20 UTC|newest]
Thread overview: 49+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-06-27 7:02 refactor the iomap writeback code v3 Christoph Hellwig
2025-06-27 7:02 ` [PATCH 01/12] iomap: pass more arguments using the iomap writeback context Christoph Hellwig
2025-06-27 15:12 ` Brian Foster
2025-06-30 5:44 ` Christoph Hellwig
2025-06-30 12:41 ` Brian Foster
2025-07-02 18:18 ` Darrick J. Wong
2025-07-02 22:00 ` Joanne Koong
2025-07-02 22:23 ` Darrick J. Wong
2025-07-02 18:22 ` Darrick J. Wong
2025-06-27 7:02 ` [PATCH 02/12] iomap: cleanup the pending writeback tracking in iomap_writepage_map_blocks Christoph Hellwig
2025-06-27 15:12 ` Brian Foster
2025-07-02 18:23 ` Darrick J. Wong
2025-06-27 7:02 ` [PATCH 03/12] iomap: refactor the writeback interface Christoph Hellwig
2025-06-27 8:23 ` Damien Le Moal
2025-06-27 15:14 ` Brian Foster
2025-06-30 5:42 ` Christoph Hellwig
2025-06-30 12:39 ` Brian Foster
2025-07-02 18:24 ` Darrick J. Wong
2025-06-27 7:02 ` [PATCH 04/12] iomap: hide ioends from the generic writeback code Christoph Hellwig
2025-06-27 8:26 ` Damien Le Moal
2025-06-27 15:14 ` Brian Foster
2025-06-28 3:09 ` Randy Dunlap
2025-07-02 18:25 ` Darrick J. Wong
2025-06-27 7:02 ` [PATCH 05/12] iomap: add public helpers for uptodate state manipulation Christoph Hellwig
2025-06-27 15:14 ` Brian Foster
2025-07-02 18:25 ` Darrick J. Wong
2025-06-27 7:02 ` [PATCH 06/12] iomap: move all ioend handling to ioend.c Christoph Hellwig
2025-06-27 15:15 ` Brian Foster
2025-06-30 5:44 ` Christoph Hellwig
2025-07-02 18:26 ` Darrick J. Wong
2025-06-27 7:02 ` [PATCH 07/12] iomap: rename iomap_writepage_map to iomap_writeback_folio Christoph Hellwig
2025-06-27 16:38 ` Brian Foster
2025-07-02 18:26 ` Darrick J. Wong
2025-06-27 7:02 ` [PATCH 08/12] iomap: move folio_unlock out of iomap_writeback_folio Christoph Hellwig
2025-06-27 16:38 ` Brian Foster
2025-06-30 5:45 ` Christoph Hellwig
2025-06-30 12:39 ` Brian Foster
2025-06-27 7:02 ` [PATCH 09/12] iomap: export iomap_writeback_folio Christoph Hellwig
2025-07-02 18:27 ` Darrick J. Wong
2025-06-27 7:02 ` [PATCH 10/12] iomap: replace iomap_folio_ops with iomap_write_ops Christoph Hellwig
2025-06-27 8:29 ` Damien Le Moal
2025-06-27 19:18 ` Brian Foster
2025-06-30 5:43 ` Christoph Hellwig
2025-07-02 18:28 ` Darrick J. Wong
2025-06-27 7:02 ` [PATCH 11/12] iomap: add read_folio_range() handler for buffered writes Christoph Hellwig
2025-06-27 19:18 ` Brian Foster
2025-06-30 5:47 ` Christoph Hellwig
2025-06-27 7:02 ` [PATCH 12/12] iomap: build the writeback code without CONFIG_BLOCK Christoph Hellwig
2025-07-02 18:20 ` Darrick J. Wong [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250702182010.GM10009@frogsfrogsfrogs \
--to=djwong@kernel.org \
--cc=brauner@kernel.org \
--cc=gfs2@lists.linux.dev \
--cc=hch@lst.de \
--cc=joannelkoong@gmail.com \
--cc=linux-block@vger.kernel.org \
--cc=linux-doc@vger.kernel.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-xfs@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).