From: "Darrick J. Wong" <djwong@kernel.org>
To: miklos@szeredi.hu
Cc: joannelkoong@gmail.com, neal@gompa.dev,
linux-fsdevel@vger.kernel.org, bernd@bsbernd.com,
fuse-devel@lists.linux.dev
Subject: Re: [PATCH 16/33] fuse: implement buffered IO with iomap
Date: Wed, 13 May 2026 12:18:21 -0700 [thread overview]
Message-ID: <20260513191821.GN9544@frogsfrogsfrogs> (raw)
In-Reply-To: <177747205494.4101881.4417232720664103981.stgit@frogsfrogsfrogs>
On Wed, Apr 29, 2026 at 07:27:51AM -0700, Darrick J. Wong wrote:
> From: Darrick J. Wong <djwong@kernel.org>
>
> Implement pagecache IO with iomap, complete with hooks into truncate and
> fallocate so that the fuse server needn't implement disk block zeroing
> of post-EOF and unaligned punch/zero regions.
>
> Signed-off-by: "Darrick J. Wong" <djwong@kernel.org>
> ---
> fs/fuse/fuse_i.h | 7
> fs/fuse/fuse_iomap.h | 11 +
> include/uapi/linux/fuse.h | 5
> fs/fuse/dir.c | 23 +
> fs/fuse/file.c | 61 +++-
> fs/fuse/fuse_iomap.c | 699 ++++++++++++++++++++++++++++++++++++++++++++-
> 6 files changed, 775 insertions(+), 31 deletions(-)
>
>
> diff --git a/fs/fuse/fuse_i.h b/fs/fuse/fuse_i.h
> index ea95615fbff598..23212ca1b6871e 100644
> --- a/fs/fuse/fuse_i.h
> +++ b/fs/fuse/fuse_i.h
> @@ -192,6 +192,11 @@ struct fuse_inode {
> #ifdef CONFIG_FUSE_IOMAP
> /* file size as reported by fuse server */
> loff_t i_disk_size;
> +
> + /* pending io completions */
> + spinlock_t ioend_lock;
> + struct work_struct ioend_work;
> + struct list_head ioend_list;
> #endif
> };
>
> @@ -1738,4 +1743,6 @@ extern void fuse_sysctl_unregister(void);
> #define fuse_sysctl_unregister() do { } while (0)
> #endif /* CONFIG_SYSCTL */
>
> +sector_t fuse_bmap(struct address_space *mapping, sector_t block);
> +
> #endif /* _FS_FUSE_I_H */
> diff --git a/fs/fuse/fuse_iomap.h b/fs/fuse/fuse_iomap.h
> index 0076e1cb12e8b0..be37ddac2f1e25 100644
> --- a/fs/fuse/fuse_iomap.h
> +++ b/fs/fuse/fuse_iomap.h
> @@ -54,6 +54,13 @@ int fuse_iomap_setsize_finish(struct inode *inode, loff_t newsize);
>
> ssize_t fuse_iomap_read_iter(struct kiocb *iocb, struct iov_iter *to);
> ssize_t fuse_iomap_write_iter(struct kiocb *iocb, struct iov_iter *from);
> +
> +int fuse_iomap_mmap(struct file *file, struct vm_area_struct *vma);
> +int fuse_iomap_setsize_start(struct inode *inode, loff_t newsize);
> +int fuse_iomap_fallocate(struct file *file, int mode, loff_t offset,
> + loff_t length, loff_t new_size);
> +int fuse_iomap_flush_unmap_range(struct inode *inode, loff_t pos,
> + loff_t endpos);
> #else
> # define fuse_iomap_enabled(...) (false)
> # define fuse_has_iomap(...) (false)
> @@ -73,6 +80,10 @@ ssize_t fuse_iomap_write_iter(struct kiocb *iocb, struct iov_iter *from);
> # define fuse_iomap_setsize_finish(...) (-ENOSYS)
> # define fuse_iomap_read_iter(...) (-ENOSYS)
> # define fuse_iomap_write_iter(...) (-ENOSYS)
> +# define fuse_iomap_mmap(...) (-ENOSYS)
> +# define fuse_iomap_setsize_start(...) (-ENOSYS)
> +# define fuse_iomap_fallocate(...) (-ENOSYS)
> +# define fuse_iomap_flush_unmap_range(...) (-ENOSYS)
> #endif /* CONFIG_FUSE_IOMAP */
>
> #endif /* _FS_FUSE_IOMAP_H */
> diff --git a/include/uapi/linux/fuse.h b/include/uapi/linux/fuse.h
> index 543965b2f8fb37..71b216262c84cb 100644
> --- a/include/uapi/linux/fuse.h
> +++ b/include/uapi/linux/fuse.h
> @@ -1373,6 +1373,9 @@ struct fuse_uring_cmd_req {
> #define FUSE_IOMAP_OP_ATOMIC (1U << 9)
> #define FUSE_IOMAP_OP_DONTCACHE (1U << 10)
>
> +/* pagecache writeback operation */
> +#define FUSE_IOMAP_OP_WRITEBACK (1U << 31)
> +
> #define FUSE_IOMAP_NULL_ADDR (-1ULL) /* addr is not valid */
>
> struct fuse_iomap_io {
> @@ -1422,6 +1425,8 @@ struct fuse_iomap_end_in {
> #define FUSE_IOMAP_IOEND_DIRECT (1U << 3)
> /* is append ioend */
> #define FUSE_IOMAP_IOEND_APPEND (1U << 4)
> +/* is pagecache writeback */
> +#define FUSE_IOMAP_IOEND_WRITEBACK (1U << 5)
>
> struct fuse_iomap_ioend_in {
> uint32_t flags; /* FUSE_IOMAP_IOEND_* */
> diff --git a/fs/fuse/dir.c b/fs/fuse/dir.c
> index 944ed53852444c..5b10ddf9b8077a 100644
> --- a/fs/fuse/dir.c
> +++ b/fs/fuse/dir.c
> @@ -2229,7 +2229,10 @@ int fuse_do_setattr(struct mnt_idmap *idmap, struct dentry *dentry,
> is_truncate = true;
> }
>
> - if (FUSE_IS_DAX(inode) && is_truncate) {
> + if (is_iomap && is_truncate) {
> + filemap_invalidate_lock(mapping);
> + fault_blocked = true;
> + } else if (FUSE_IS_DAX(inode) && is_truncate) {
> filemap_invalidate_lock(mapping);
> fault_blocked = true;
> err = fuse_dax_break_layouts(inode, 0, -1);
> @@ -2244,6 +2247,18 @@ int fuse_do_setattr(struct mnt_idmap *idmap, struct dentry *dentry,
> WARN_ON(!(attr->ia_valid & ATTR_SIZE));
> WARN_ON(attr->ia_size != 0);
> if (fc->atomic_o_trunc) {
> + if (is_iomap) {
> + /*
> + * fuse_open already set the size to zero and
> + * truncated the pagecache, and we've since
> + * cycled the inode locks. Another thread
> + * could have performed an appending write, so
> + * we don't want to touch the file further.
> + */
> + filemap_invalidate_unlock(mapping);
> + return 0;
> + }
> +
> /*
> * No need to send request to userspace, since actual
> * truncation has already been done by OPEN. But still
> @@ -2277,6 +2292,12 @@ int fuse_do_setattr(struct mnt_idmap *idmap, struct dentry *dentry,
> set_bit(FUSE_I_SIZE_UNSTABLE, &fi->state);
> if (trust_local_cmtime && attr->ia_size != inode->i_size)
> attr->ia_valid |= ATTR_MTIME | ATTR_CTIME;
> +
> + if (is_iomap) {
> + err = fuse_iomap_setsize_start(inode, attr->ia_size);
> + if (err)
> + goto error;
> + }
> }
>
> memset(&inarg, 0, sizeof(inarg));
> diff --git a/fs/fuse/file.c b/fs/fuse/file.c
> index ad28bcd36d15e6..6471952489aa2e 100644
> --- a/fs/fuse/file.c
> +++ b/fs/fuse/file.c
> @@ -408,7 +408,7 @@ static int fuse_release(struct inode *inode, struct file *file)
> * Dirty pages might remain despite write_inode_now() call from
> * fuse_flush() due to writes racing with the close.
> */
> - if (fc->writeback_cache)
> + if (fc->writeback_cache || fuse_inode_has_iomap(inode))
> write_inode_now(inode, 1);
>
> fuse_release_common(file, false);
> @@ -2400,6 +2400,9 @@ static int fuse_file_mmap(struct file *file, struct vm_area_struct *vma)
> struct inode *inode = file_inode(file);
> int rc;
>
> + if (fuse_inode_has_iomap(inode))
> + return fuse_iomap_mmap(file, vma);
> +
> /* DAX mmap is superior to direct_io mmap */
> if (FUSE_IS_DAX(inode))
> return fuse_dax_mmap(file, vma);
> @@ -2598,7 +2601,7 @@ static int fuse_file_flock(struct file *file, int cmd, struct file_lock *fl)
> return err;
> }
>
> -static sector_t fuse_bmap(struct address_space *mapping, sector_t block)
> +sector_t fuse_bmap(struct address_space *mapping, sector_t block)
> {
> struct inode *inode = mapping->host;
> struct fuse_mount *fm = get_fuse_mount(inode);
> @@ -2952,8 +2955,12 @@ fuse_direct_IO(struct kiocb *iocb, struct iov_iter *iter)
>
> static int fuse_writeback_range(struct inode *inode, loff_t start, loff_t end)
> {
> - int err = filemap_write_and_wait_range(inode->i_mapping, start, LLONG_MAX);
> + int err;
>
> + if (fuse_inode_has_iomap(inode))
> + return fuse_iomap_flush_unmap_range(inode, start, end);
> +
> + err = filemap_write_and_wait_range(inode->i_mapping, start, LLONG_MAX);
> if (!err)
> fuse_sync_writes(inode);
>
> @@ -2989,7 +2996,10 @@ static long fuse_file_fallocate(struct file *file, int mode, loff_t offset,
> return -EOPNOTSUPP;
>
> inode_lock(inode);
> - if (block_faults) {
> + if (is_iomap) {
> + filemap_invalidate_lock(inode->i_mapping);
> + block_faults = true;
> + } else if (block_faults) {
> filemap_invalidate_lock(inode->i_mapping);
> err = fuse_dax_break_layouts(inode, 0, -1);
> if (err)
> @@ -3004,6 +3014,17 @@ static long fuse_file_fallocate(struct file *file, int mode, loff_t offset,
> goto out;
> }
>
> + /*
> + * If we are using iomap for file IO, fallocate must wait for all AIO
> + * to complete before we continue as AIO can change the file size on
> + * completion without holding any locks we currently hold. We must do
> + * this first because AIO can update the in-memory inode size, and the
> + * operations that follow require the in-memory size to be fully
> + * up-to-date.
> + */
> + if (is_iomap)
> + inode_dio_wait(inode);
> +
> if (!(mode & FALLOC_FL_KEEP_SIZE) &&
> offset + length > i_size_read(inode)) {
> err = inode_newsize_ok(inode, offset + length);
> @@ -3032,21 +3053,23 @@ static long fuse_file_fallocate(struct file *file, int mode, loff_t offset,
> if (err)
> goto out;
>
> - /* we could have extended the file */
> - if (!(mode & FALLOC_FL_KEEP_SIZE)) {
> - if (is_iomap && newsize > 0) {
> - err = fuse_iomap_setsize_finish(inode, newsize);
> - if (err)
> - goto out;
> + if (is_iomap) {
> + err = fuse_iomap_fallocate(file, mode, offset, length,
> + newsize);
> + if (err)
> + goto out;
> + } else {
> + /* we could have extended the file */
> + if (!(mode & FALLOC_FL_KEEP_SIZE)) {
> + if (fuse_write_update_attr(inode, newsize, length))
> + file_update_time(file);
> }
>
> - if (fuse_write_update_attr(inode, offset + length, length))
> - file_update_time(file);
> + if (mode & (FALLOC_FL_PUNCH_HOLE | FALLOC_FL_ZERO_RANGE))
> + truncate_pagecache_range(inode, offset,
> + offset + length - 1);
> }
>
> - if (mode & (FALLOC_FL_PUNCH_HOLE | FALLOC_FL_ZERO_RANGE))
> - truncate_pagecache_range(inode, offset, offset + length - 1);
> -
> fuse_invalidate_attr_mask(inode, FUSE_STATX_MODSIZE);
>
> out:
> @@ -3090,6 +3113,7 @@ static ssize_t __fuse_copy_file_range(struct file *file_in, loff_t pos_in,
> ssize_t err;
> /* mark unstable when write-back is not used, and file_out gets
> * extended */
> + const bool is_iomap = fuse_inode_has_iomap(inode_out);
> bool is_unstable = (!fc->writeback_cache) &&
> ((pos_out + len) > inode_out->i_size);
>
> @@ -3133,6 +3157,10 @@ static ssize_t __fuse_copy_file_range(struct file *file_in, loff_t pos_in,
> if (err)
> goto out;
>
> + /* See inode_dio_wait comment in fuse_file_fallocate */
> + if (is_iomap)
> + inode_dio_wait(inode_out);
> +
> if (is_unstable)
> set_bit(FUSE_I_SIZE_UNSTABLE, &fi_out->state);
>
> @@ -3173,7 +3201,8 @@ static ssize_t __fuse_copy_file_range(struct file *file_in, loff_t pos_in,
> goto out;
> }
>
> - truncate_inode_pages_range(inode_out->i_mapping,
> + if (!is_iomap)
> + truncate_inode_pages_range(inode_out->i_mapping,
> ALIGN_DOWN(pos_out, PAGE_SIZE),
> ALIGN(pos_out + bytes_copied, PAGE_SIZE) - 1);
>
> diff --git a/fs/fuse/fuse_iomap.c b/fs/fuse/fuse_iomap.c
> index acc7ae0f56a2bf..7a7dfc4f665d8e 100644
> --- a/fs/fuse/fuse_iomap.c
> +++ b/fs/fuse/fuse_iomap.c
> @@ -5,6 +5,8 @@
> */
> #include <linux/iomap.h>
> #include <linux/fiemap.h>
> +#include <linux/pagemap.h>
> +#include <linux/falloc.h>
> #include "fuse_i.h"
> #include "fuse_trace.h"
> #include "fuse_iomap.h"
> @@ -204,7 +206,7 @@ static inline uint16_t fuse_iomap_flags_from_server(uint16_t fuse_f_flags)
> ret |= FUSE_IOMAP_OP_##word
> static inline uint32_t fuse_iomap_op_to_server(unsigned iomap_op_flags)
> {
> - uint32_t ret = 0;
> + uint32_t ret = iomap_op_flags & FUSE_IOMAP_OP_WRITEBACK;
>
> XMAP(WRITE);
> XMAP(ZERO);
> @@ -370,7 +372,8 @@ fuse_iomap_begin_validate(const struct inode *inode,
>
> static inline bool fuse_is_iomap_file_write(unsigned int opflags)
> {
> - return opflags & (IOMAP_WRITE | IOMAP_ZERO | IOMAP_UNSHARE);
> + return opflags & (IOMAP_WRITE | IOMAP_ZERO | IOMAP_UNSHARE |
> + FUSE_IOMAP_OP_WRITEBACK);
> }
>
> static inline struct fuse_backing *
> @@ -744,12 +747,7 @@ void fuse_iomap_unmount(struct fuse_mount *fm)
> fuse_send_destroy(fm);
> }
>
> -static inline void fuse_inode_set_iomap(struct inode *inode)
> -{
> - struct fuse_inode *fi = get_fuse_inode(inode);
> -
> - set_bit(FUSE_I_IOMAP, &fi->state);
> -}
> +static inline void fuse_inode_set_iomap(struct inode *inode);
>
> static inline void fuse_inode_clear_iomap(struct inode *inode)
> {
> @@ -976,17 +974,107 @@ static const struct iomap_dio_ops fuse_iomap_dio_write_ops = {
> .end_io = fuse_iomap_dio_write_end_io,
> };
>
> +static const struct iomap_write_ops fuse_iomap_write_ops = {
> +};
> +
> +static int
> +fuse_iomap_zero_range(
> + struct inode *inode,
> + loff_t pos,
> + loff_t len,
> + bool *did_zero)
> +{
> + return iomap_zero_range(inode, pos, len, did_zero, &fuse_iomap_ops,
> + &fuse_iomap_write_ops, NULL);
> +}
> +
> +/* Take care of zeroing post-EOF blocks when they might exist. */
> +static ssize_t
> +fuse_iomap_write_zero_eof(
> + struct kiocb *iocb,
> + struct iov_iter *from,
> + bool *drained_dio)
> +{
> + struct inode *inode = file_inode(iocb->ki_filp);
> + struct fuse_inode *fi = get_fuse_inode(inode);
> + struct address_space *mapping = iocb->ki_filp->f_mapping;
> + loff_t isize;
> + int error;
> +
> + /*
> + * We need to serialise against EOF updates that occur in IO
> + * completions here. We want to make sure that nobody is changing the
> + * size while we do this check until we have placed an IO barrier (i.e.
> + * hold i_rwsem exclusively) that prevents new IO from being
> + * dispatched. The spinlock effectively forms a memory barrier once we
> + * have i_rwsem exclusively so we are guaranteed to see the latest EOF
> + * value and hence be able to correctly determine if we need to run
> + * zeroing.
> + */
> + spin_lock(&fi->lock);
> + isize = i_size_read(inode);
> + if (iocb->ki_pos <= isize) {
> + spin_unlock(&fi->lock);
> + return 0;
> + }
> + spin_unlock(&fi->lock);
> +
> + if (iocb->ki_flags & IOCB_NOWAIT)
> + return -EAGAIN;
> +
> + if (!(*drained_dio)) {
> + /*
> + * We now have an IO submission barrier in place, but AIO can
> + * do EOF updates during IO completion and hence we now need to
> + * wait for all of them to drain. Non-AIO DIO will have
> + * drained before we are given the exclusive i_rwsem, and so
> + * for most cases this wait is a no-op.
> + */
> + inode_dio_wait(inode);
> + *drained_dio = true;
> + return 1;
> + }
> +
> + filemap_invalidate_lock(mapping);
> + error = fuse_iomap_zero_range(inode, isize, iocb->ki_pos - isize, NULL);
> + filemap_invalidate_unlock(mapping);
> +
> + return error;
> +}
> +
> static ssize_t
> fuse_iomap_write_checks(
> struct kiocb *iocb,
> struct iov_iter *from)
> {
> + struct inode *inode = iocb->ki_filp->f_mapping->host;
> ssize_t error;
> + bool drained_dio = false;
>
> +restart:
> error = generic_write_checks(iocb, from);
> if (error <= 0)
> return error;
>
> + /*
> + * If the offset is beyond the size of the file, we need to zero all
> + * blocks that fall between the existing EOF and the start of this
> + * write.
> + *
> + * We can do an unlocked check for i_size here safely as I/O completion
> + * can only extend EOF. Truncate is locked out at this point, so the
> + * EOF cannot move backwards, only forwards. Hence we only need to take
> + * the slow path when we are at or beyond the current EOF.
> + */
> + if (fuse_inode_has_iomap(inode) &&
> + iocb->ki_pos > i_size_read(inode)) {
> + error = fuse_iomap_write_zero_eof(iocb, from, &drained_dio);
> + if (error == 1)
> + goto restart;
> + if (error)
> + return error;
> + }
> +
> return kiocb_modified(iocb);
> }
>
> @@ -1060,6 +1148,366 @@ void fuse_iomap_open_truncate(struct inode *inode)
> spin_unlock(&fi->lock);
> }
>
> +struct fuse_writepage_ctx {
> + struct iomap_writepage_ctx ctx;
> +};
> +
> +static void fuse_iomap_end_ioend(struct iomap_ioend *ioend)
> +{
> + struct inode *inode = ioend->io_inode;
> + unsigned int ioendflags = FUSE_IOMAP_IOEND_WRITEBACK;
> + unsigned int nofs_flag;
> + int error = blk_status_to_errno(ioend->io_bio.bi_status);
> +
> + ASSERT(fuse_inode_has_iomap(inode));
> +
> + /* We still have to clean up the ioend even if the inode is dead */
> + if (!error && fuse_is_bad(inode))
> + error = -EIO;
> +
> + if (ioend->io_flags & IOMAP_IOEND_SHARED)
> + ioendflags |= FUSE_IOMAP_IOEND_SHARED;
> + if (ioend->io_flags & IOMAP_IOEND_UNWRITTEN)
> + ioendflags |= FUSE_IOMAP_IOEND_UNWRITTEN;
> +
> + /*
> + * We can allocate memory here while doing writeback on behalf of
> + * memory reclaim. To avoid memory allocation deadlocks set the
> + * task-wide nofs context for the following operations.
> + */
> + nofs_flag = memalloc_nofs_save();
> + fuse_iomap_ioend(inode, ioend->io_offset, ioend->io_size, error,
> + ioendflags, ioend->io_bio.bi_bdev, ioend->io_sector);
Codex noticed that if the bio succeeds but the ioend mapping update
fails, we drop the new error code and complete the rest of the ioends as
if nothing went wrong. Obviously we don't want to drop writeback
errors, so we need to do the usual
err2 = fuse_iomap_ioend(...);
if (err2 && !error)
error = err2;
trick.
> + iomap_finish_ioends(ioend, error);
> + memalloc_nofs_restore(nofs_flag);
> +}
> +
> +/*
> + * Finish all pending IO completions that require transactional modifications.
> + *
> + * We try to merge physical and logically contiguous ioends before completion to
> + * minimise the number of transactions we need to perform during IO completion.
> + * Both unwritten extent conversion and COW remapping need to iterate and modify
> + * one physical extent at a time, so we gain nothing by merging physically
> + * discontiguous extents here.
> + *
> + * The ioend chain length that we can be processing here is largely unbound in
> + * length and we may have to perform significant amounts of work on each ioend
> + * to complete it. Hence we have to be careful about holding the CPU for too
> + * long in this loop.
> + */
> +static void fuse_iomap_end_io(struct work_struct *work)
> +{
> + struct fuse_inode *fi =
> + container_of(work, struct fuse_inode, ioend_work);
> + struct iomap_ioend *ioend;
> + struct list_head tmp;
> + unsigned long flags;
> +
> + spin_lock_irqsave(&fi->ioend_lock, flags);
> + list_replace_init(&fi->ioend_list, &tmp);
> + spin_unlock_irqrestore(&fi->ioend_lock, flags);
Minor optimization: if other writeback ioends come in while the ioend
worker is running, we could jump back to the top here to process the
new ioends instead of waiting for fuse_iomap_end_io to get called again.
--D
> + iomap_sort_ioends(&tmp);
> + while ((ioend = list_first_entry_or_null(&tmp, struct iomap_ioend,
> + io_list))) {
> + list_del_init(&ioend->io_list);
> + iomap_ioend_try_merge(ioend, &tmp);
> + fuse_iomap_end_ioend(ioend);
> + cond_resched();
> + }
> +}
> +
> +static void fuse_iomap_end_bio(struct bio *bio)
> +{
> + struct iomap_ioend *ioend = iomap_ioend_from_bio(bio);
> + struct inode *inode = ioend->io_inode;
> + struct fuse_inode *fi = get_fuse_inode(inode);
> + unsigned long flags;
> +
> + ASSERT(fuse_inode_has_iomap(inode));
> +
> + spin_lock_irqsave(&fi->ioend_lock, flags);
> + if (list_empty(&fi->ioend_list))
> + WARN_ON_ONCE(!queue_work(system_unbound_wq, &fi->ioend_work));
> + list_add_tail(&ioend->io_list, &fi->ioend_list);
> + spin_unlock_irqrestore(&fi->ioend_lock, flags);
> +}
> +
> +/*
> + * Fast revalidation of the cached writeback mapping. Return true if the current
> + * mapping is valid, false otherwise.
> + */
> +static bool fuse_iomap_revalidate_writeback(struct iomap_writepage_ctx *wpc,
> + loff_t offset)
> +{
> + if (offset < wpc->iomap.offset ||
> + offset >= wpc->iomap.offset + wpc->iomap.length)
> + return false;
> +
> + /* XXX actually use revalidation cookie */
> + return true;
> +}
> +
> +/*
> + * If the folio has delalloc blocks on it, the caller is asking us to punch them
> + * out. If we don't, we can leave a stale delalloc mapping covered by a clean
> + * page that needs to be dirtied again before the delalloc mapping can be
> + * converted. This stale delalloc mapping can trip up a later direct I/O read
> + * operation on the same region.
> + *
> + * We prevent this by truncating away the delalloc regions on the folio. Because
> + * they are delalloc, we can do this without needing a transaction. Indeed - if
> + * we get ENOSPC errors, we have to be able to do this truncation without a
> + * transaction as there is no space left for block reservation (typically why
> + * we see a ENOSPC in writeback).
> + */
> +static void fuse_iomap_discard_folio(struct folio *folio, loff_t pos, int error)
> +{
> + struct inode *inode = folio->mapping->host;
> + struct fuse_inode *fi = get_fuse_inode(inode);
> + loff_t end = folio_pos(folio) + folio_size(folio);
> +
> + if (fuse_is_bad(inode))
> + return;
> +
> + ASSERT(fuse_inode_has_iomap(inode));
> +
> + printk_ratelimited(KERN_ERR
> + "page discard on page %px, inode 0x%llx, pos %llu.",
> + folio, fi->orig_ino, pos);
> +
> + /* Userspace may need to remove delayed allocations */
> + fuse_iomap_ioend(inode, pos, end - pos, error, 0, NULL,
> + FUSE_IOMAP_NULL_ADDR);
> +}
> +
> +static ssize_t fuse_iomap_writeback_range(struct iomap_writepage_ctx *wpc,
> + struct folio *folio, u64 offset,
> + unsigned int len, u64 end_pos)
> +{
> + struct inode *inode = wpc->inode;
> + struct iomap write_iomap, dontcare;
> + ssize_t ret;
> +
> + if (fuse_is_bad(inode)) {
> + ret = -EIO;
> + goto discard_folio;
> + }
> +
> + ASSERT(fuse_inode_has_iomap(inode));
> +
> + if (!fuse_iomap_revalidate_writeback(wpc, offset)) {
> + ret = fuse_iomap_begin(inode, offset, len,
> + FUSE_IOMAP_OP_WRITEBACK,
> + &write_iomap, &dontcare);
> + if (ret)
> + goto discard_folio;
> +
> + /*
> + * Landed in a hole or beyond EOF? Send that to iomap, it'll
> + * skip writing back the file range.
> + */
> + if (write_iomap.offset > offset) {
> + write_iomap.length = write_iomap.offset - offset;
> + write_iomap.offset = offset;
> + write_iomap.type = IOMAP_HOLE;
> + }
> +
> + memcpy(&wpc->iomap, &write_iomap, sizeof(struct iomap));
> + }
> +
> + ret = iomap_add_to_ioend(wpc, folio, offset, end_pos, len);
> + if (ret < 0)
> + goto discard_folio;
> +
> + return ret;
> +discard_folio:
> + fuse_iomap_discard_folio(folio, offset, ret);
> + return ret;
> +}
> +
> +static int fuse_iomap_writeback_submit(struct iomap_writepage_ctx *wpc,
> + int error)
> +{
> + struct iomap_ioend *ioend = wpc->wb_ctx;
> +
> + ASSERT(fuse_inode_has_iomap(ioend->io_inode));
> +
> + /* always call our ioend function, even if we cancel the bio */
> + ioend->io_bio.bi_end_io = fuse_iomap_end_bio;
> + return iomap_ioend_writeback_submit(wpc, error);
> +}
> +
> +static const struct iomap_writeback_ops fuse_iomap_writeback_ops = {
> + .writeback_range = fuse_iomap_writeback_range,
> + .writeback_submit = fuse_iomap_writeback_submit,
> +};
> +
> +static int fuse_iomap_writepages(struct address_space *mapping,
> + struct writeback_control *wbc)
> +{
> + struct fuse_writepage_ctx wpc = {
> + .ctx = {
> + .inode = mapping->host,
> + .wbc = wbc,
> + .ops = &fuse_iomap_writeback_ops,
> + },
> + };
> +
> + ASSERT(fuse_inode_has_iomap(mapping->host));
> +
> + return iomap_writepages(&wpc.ctx);
> +}
> +
> +static int fuse_iomap_read_folio(struct file *file, struct folio *folio)
> +{
> + ASSERT(fuse_inode_has_iomap(file_inode(file)));
> +
> + iomap_bio_read_folio(folio, &fuse_iomap_ops);
> + return 0;
> +}
> +
> +static void fuse_iomap_readahead(struct readahead_control *rac)
> +{
> + ASSERT(fuse_inode_has_iomap(file_inode(rac->file)));
> +
> + iomap_bio_readahead(rac, &fuse_iomap_ops);
> +}
> +
> +static const struct address_space_operations fuse_iomap_aops = {
> + .read_folio = fuse_iomap_read_folio,
> + .readahead = fuse_iomap_readahead,
> + .writepages = fuse_iomap_writepages,
> + .dirty_folio = iomap_dirty_folio,
> + .release_folio = iomap_release_folio,
> + .invalidate_folio = iomap_invalidate_folio,
> + .migrate_folio = filemap_migrate_folio,
> + .is_partially_uptodate = iomap_is_partially_uptodate,
> + .error_remove_folio = generic_error_remove_folio,
> +
> + /* These aren't pagecache operations per se */
> + .bmap = fuse_bmap,
> +};
> +
> +static inline void fuse_inode_set_iomap(struct inode *inode)
> +{
> + struct fuse_inode *fi = get_fuse_inode(inode);
> +
> + inode->i_data.a_ops = &fuse_iomap_aops;
> +
> + INIT_WORK(&fi->ioend_work, fuse_iomap_end_io);
> + INIT_LIST_HEAD(&fi->ioend_list);
> + spin_lock_init(&fi->ioend_lock);
> + set_bit(FUSE_I_IOMAP, &fi->state);
> +}
> +
> +/*
> + * Locking for serialisation of IO during page faults. This results in a lock
> + * ordering of:
> + *
> + * mmap_lock (MM)
> + * sb_start_pagefault(vfs, freeze)
> + * invalidate_lock (vfs - truncate serialisation)
> + * page_lock (MM)
> + * i_lock (FUSE - extent map serialisation)
> + */
> +static vm_fault_t fuse_iomap_page_mkwrite(struct vm_fault *vmf)
> +{
> + struct inode *inode = file_inode(vmf->vma->vm_file);
> + struct address_space *mapping = vmf->vma->vm_file->f_mapping;
> + vm_fault_t ret;
> +
> + ASSERT(fuse_inode_has_iomap(inode));
> +
> + sb_start_pagefault(inode->i_sb);
> + file_update_time(vmf->vma->vm_file);
> +
> + filemap_invalidate_lock_shared(mapping);
> + ret = iomap_page_mkwrite(vmf, &fuse_iomap_ops, NULL);
> + filemap_invalidate_unlock_shared(mapping);
> +
> + sb_end_pagefault(inode->i_sb);
> + return ret;
> +}
> +
> +static const struct vm_operations_struct fuse_iomap_vm_ops = {
> + .fault = filemap_fault,
> + .map_pages = filemap_map_pages,
> + .page_mkwrite = fuse_iomap_page_mkwrite,
> +};
> +
> +int fuse_iomap_mmap(struct file *file, struct vm_area_struct *vma)
> +{
> + ASSERT(fuse_inode_has_iomap(file_inode(file)));
> +
> + file_accessed(file);
> + vma->vm_ops = &fuse_iomap_vm_ops;
> + return 0;
> +}
> +
> +static ssize_t fuse_iomap_buffered_read(struct kiocb *iocb, struct iov_iter *to)
> +{
> + struct inode *inode = file_inode(iocb->ki_filp);
> + ssize_t ret;
> +
> + ASSERT(fuse_inode_has_iomap(inode));
> +
> + if (!iov_iter_count(to))
> + return 0; /* skip atime */
> +
> + ret = fuse_iomap_ilock_iocb(iocb, SHARED);
> + if (ret)
> + return ret;
> + ret = generic_file_read_iter(iocb, to);
> + if (ret > 0)
> + file_accessed(iocb->ki_filp);
> + inode_unlock_shared(inode);
> +
> + return ret;
> +}
> +
> +static ssize_t fuse_iomap_buffered_write(struct kiocb *iocb,
> + struct iov_iter *from)
> +{
> + struct inode *inode = file_inode(iocb->ki_filp);
> + struct fuse_inode *fi = get_fuse_inode(inode);
> + loff_t pos = iocb->ki_pos;
> + ssize_t ret;
> +
> + ASSERT(fuse_inode_has_iomap(inode));
> +
> + if (!iov_iter_count(from))
> + return 0;
> +
> + ret = fuse_iomap_ilock_iocb(iocb, EXCL);
> + if (ret)
> + return ret;
> +
> + ret = fuse_iomap_write_checks(iocb, from);
> + if (ret)
> + goto out_unlock;
> +
> + if (inode->i_size < pos + iov_iter_count(from))
> + set_bit(FUSE_I_SIZE_UNSTABLE, &fi->state);
> +
> + ret = iomap_file_buffered_write(iocb, from, &fuse_iomap_ops,
> + &fuse_iomap_write_ops, NULL);
> +
> + if (ret > 0)
> + fuse_write_update_attr(inode, pos + ret, ret);
> + clear_bit(FUSE_I_SIZE_UNSTABLE, &fi->state);
> +
> +out_unlock:
> + inode_unlock(inode);
> +
> + if (ret > 0) {
> + /* Handle various SYNC-type writes */
> + ret = generic_write_sync(iocb, ret);
> + }
> + return ret;
> +}
> +
> static inline bool fuse_iomap_force_directio(const struct kiocb *iocb)
> {
> struct fuse_file *ff = iocb->ki_filp->private_data;
> @@ -1073,9 +1521,33 @@ ssize_t fuse_iomap_read_iter(struct kiocb *iocb, struct iov_iter *to)
>
> ASSERT(fuse_inode_has_iomap(file_inode(iocb->ki_filp)));
>
> - if ((iocb->ki_flags & IOCB_DIRECT) || force_directio)
> - return fuse_iomap_direct_read(iocb, to);
> - return -EIO;
> + if ((iocb->ki_flags & IOCB_DIRECT) || force_directio) {
> + ssize_t ret = fuse_iomap_direct_read(iocb, to);
> +
> + switch (ret) {
> + case -ENOTBLK:
> + case -ENOSYS:
> + /*
> + * We fall back to a buffered read if:
> + *
> + * - ENOTBLK means iomap told us to do it
> + * - ENOSYS means the fuse server wants it
> + *
> + * Don't fall back if we were forced to do it.
> + */
> + if (force_directio)
> + return -EIO;
> + break;
> + default:
> + /* errors, no progress, or partial progress */
> + return ret;
> + }
> +
> + /* do not let generic_file_read_iter fall into ->direct_IO */
> + iocb->ki_flags &= ~IOCB_DIRECT;
> + }
> +
> + return fuse_iomap_buffered_read(iocb, to);
> }
>
> ssize_t fuse_iomap_write_iter(struct kiocb *iocb, struct iov_iter *from)
> @@ -1084,7 +1556,206 @@ ssize_t fuse_iomap_write_iter(struct kiocb *iocb, struct iov_iter *from)
>
> ASSERT(fuse_inode_has_iomap(file_inode(iocb->ki_filp)));
>
> - if ((iocb->ki_flags & IOCB_DIRECT) || force_directio)
> - return fuse_iomap_direct_write(iocb, from);
> - return -EIO;
> + if ((iocb->ki_flags & IOCB_DIRECT) || force_directio) {
> + ssize_t ret = fuse_iomap_direct_write(iocb, from);
> +
> + switch (ret) {
> + case -ENOTBLK:
> + case -ENOSYS:
> + /*
> + * We fall back to a buffered write if:
> + *
> + * - ENOTBLK means iomap told us to do it
> + * - ENOSYS means the fuse server wants it
> + *
> + * Either way, try the write again as a synchronous
> + * buffered write unless we were forced to do directio.
> + */
> + if (force_directio)
> + return -EIO;
> + iocb->ki_flags |= IOCB_SYNC;
> + break;
> + default:
> + /* errors, no progress, or partial progress */
> + return ret;
> + }
> + }
> +
> + return fuse_iomap_buffered_write(iocb, from);
> +}
> +
> +static int
> +fuse_iomap_truncate_page(
> + struct inode *inode,
> + loff_t pos,
> + bool *did_zero)
> +{
> + return iomap_truncate_page(inode, pos, did_zero, &fuse_iomap_ops,
> + &fuse_iomap_write_ops, NULL);
> +}
> +/*
> + * Truncate pagecache for a file before sending the truncate request to
> + * userspace. Must have write permission and not be a directory.
> + *
> + * Caution: The caller of this function is responsible for calling
> + * setattr_prepare() or otherwise verifying the change is fine.
> + */
> +int
> +fuse_iomap_setsize_start(
> + struct inode *inode,
> + loff_t newsize)
> +{
> + loff_t oldsize = i_size_read(inode);
> + int error;
> + bool did_zeroing = false;
> +
> + rwsem_assert_held_write(&inode->i_rwsem);
> + rwsem_assert_held_write(&inode->i_mapping->invalidate_lock);
> + ASSERT(S_ISREG(inode->i_mode));
> +
> + /*
> + * Wait for all direct I/O to complete.
> + */
> + inode_dio_wait(inode);
> +
> + /*
> + * File data changes must be complete and flushed to disk before we
> + * call userspace to modify the inode.
> + *
> + * Start with zeroing any data beyond EOF that we may expose on file
> + * extension, or zeroing out the rest of the block on a downward
> + * truncate.
> + */
> + if (newsize > oldsize)
> + error = fuse_iomap_zero_range(inode, oldsize, newsize - oldsize,
> + &did_zeroing);
> + else
> + error = fuse_iomap_truncate_page(inode, newsize, &did_zeroing);
> + if (error)
> + return error;
> +
> + /*
> + * We've already locked out new page faults, so now we can safely
> + * remove pages from the page cache knowing they won't get refaulted
> + * until we drop the mapping invalidation lock after the extent
> + * manipulations are complete. The truncate_setsize() call also cleans
> + * folios spanning EOF on extending truncates and hence ensures
> + * sub-page block size filesystems are correctly handled, too.
> + *
> + * And we update in-core i_size and truncate page cache beyond newsize
> + * before writing back the whole file, so we're guaranteed not to write
> + * stale data past the new EOF on truncate down.
> + */
> + truncate_setsize(inode, newsize);
> +
> + /*
> + * Flush the entire pagecache to ensure the fuse server logs the inode
> + * size change and all dirty data that might be associated with it.
> + * We don't know the ondisk inode size, so we only have this clumsy
> + * hammer.
> + */
> + return filemap_write_and_wait(inode->i_mapping);
> +}
> +
> +/*
> + * Prepare for a file data block remapping operation by flushing and unmapping
> + * all pagecache for the entire range.
> + */
> +int fuse_iomap_flush_unmap_range(struct inode *inode, loff_t pos,
> + loff_t endpos)
> +{
> + loff_t start, end;
> + unsigned int rounding;
> + int error;
> +
> + /*
> + * Make sure we extend the flush out to extent alignment boundaries so
> + * any extent range overlapping the start/end of the modification we
> + * are about to do is clean and idle.
> + */
> + rounding = max_t(unsigned int, i_blocksize(inode), PAGE_SIZE);
> + start = round_down(pos, rounding);
> + end = round_up(endpos + 1, rounding) - 1;
> +
> + error = filemap_write_and_wait_range(inode->i_mapping, start, end);
> + if (error)
> + return error;
> + truncate_pagecache_range(inode, start, end);
> + return 0;
> +}
> +
> +static int fuse_iomap_punch_range(struct inode *inode, loff_t offset,
> + loff_t length)
> +{
> + loff_t isize = i_size_read(inode);
> + int error;
> +
> + /*
> + * Now that we've unmap all full blocks we'll have to zero out any
> + * partial block at the beginning and/or end. iomap_zero_range is
> + * smart enough to skip holes and unwritten extents, including those we
> + * just created, but we must take care not to zero beyond EOF, which
> + * would enlarge i_size.
> + */
> + if (offset >= isize)
> + return 0;
> + if (offset + length > isize)
> + length = isize - offset;
> + error = fuse_iomap_zero_range(inode, offset, length, NULL);
> + if (error)
> + return error;
> +
> + /*
> + * If we zeroed right up to EOF and EOF straddles a page boundary we
> + * must make sure that the post-EOF area is also zeroed because the
> + * page could be mmap'd and iomap_zero_range doesn't do that for us.
> + * Writeback of the eof page will do this, albeit clumsily.
> + */
> + if (offset + length >= isize && offset_in_page(offset + length) > 0) {
> + error = filemap_write_and_wait_range(inode->i_mapping,
> + round_down(offset + length, PAGE_SIZE),
> + LLONG_MAX);
> + }
> +
> + return error;
> +}
> +
> +int
> +fuse_iomap_fallocate(
> + struct file *file,
> + int mode,
> + loff_t offset,
> + loff_t length,
> + loff_t new_size)
> +{
> + struct inode *inode = file_inode(file);
> + int error;
> +
> + ASSERT(fuse_inode_has_iomap(inode));
> +
> + /*
> + * If we unmapped blocks from the file range, then we zero the
> + * pagecache for those regions and push them to disk rather than make
> + * the fuse server manually zero the disk blocks.
> + */
> + if (mode & (FALLOC_FL_PUNCH_HOLE | FALLOC_FL_ZERO_RANGE)) {
> + error = fuse_iomap_punch_range(inode, offset, length);
> + if (error)
> + return error;
> + }
> +
> + /*
> + * If this is an extending write, we need to zero the bytes beyond the
> + * new EOF and bounce the new size out to userspace.
> + */
> + if (new_size) {
> + error = fuse_iomap_setsize_start(inode, new_size);
> + if (error)
> + return error;
> +
> + fuse_write_update_attr(inode, new_size, length);
> + }
> +
> + file_update_time(file);
> + return 0;
> }
>
>
next prev parent reply other threads:[~2026-05-13 19:18 UTC|newest]
Thread overview: 206+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-04-29 14:12 [PATCHBLIZZARD v8] fuse/libfuse/e2fsprogs: faster file IO for containerized ext4 servers Darrick J. Wong
2026-04-29 14:16 ` [PATCHSET v8 1/8] fuse: general bug fixes Darrick J. Wong
2026-04-29 14:21 ` [PATCH 1/4] fuse: flush pending FUSE_RELEASE requests before sending FUSE_DESTROY Darrick J. Wong
2026-04-29 14:22 ` [PATCH 2/4] fuse: implement file attributes mask for statx Darrick J. Wong
2026-04-29 14:22 ` [PATCH 3/4] fuse: update file mode when updating acls Darrick J. Wong
2026-04-30 13:48 ` Joanne Koong
2026-04-30 20:57 ` Darrick J. Wong
2026-05-01 9:53 ` Joanne Koong
2026-05-01 16:15 ` Darrick J. Wong
2026-04-29 14:22 ` [PATCH 4/4] fuse: propagate default and file acls on creation Darrick J. Wong
2026-05-01 11:11 ` Joanne Koong
2026-05-01 16:57 ` Darrick J. Wong
2026-04-29 14:16 ` [PATCHSET v8 2/8] iomap: cleanups ahead of adding fuse support Darrick J. Wong
2026-04-29 14:22 ` [PATCH 1/2] iomap: allow directio callers to supply _COMP_WORK Darrick J. Wong
2026-04-29 14:23 ` [PATCH 2/2] iomap: allow NULL swap info bdev when activating swapfile Darrick J. Wong
2026-05-08 9:06 ` Christoph Hellwig
2026-05-08 23:41 ` Darrick J. Wong
2026-04-29 14:17 ` [PATCHSET v8 3/8] fuse: cleanups ahead of adding fuse support Darrick J. Wong
2026-04-29 14:23 ` [PATCH 1/2] fuse: move the passthrough-specific code back to passthrough.c Darrick J. Wong
2026-04-29 14:23 ` [PATCH 2/2] fuse_trace: " Darrick J. Wong
2026-04-29 14:17 ` [PATCHSET v8 4/8] fuse: allow servers to use iomap for better file IO performance Darrick J. Wong
2026-04-29 14:23 ` [PATCH 01/33] fuse: implement the basic iomap mechanisms Darrick J. Wong
2026-05-13 17:38 ` Darrick J. Wong
2026-04-29 14:24 ` [PATCH 02/33] fuse_trace: " Darrick J. Wong
2026-04-29 14:24 ` [PATCH 03/33] fuse: make debugging configurable at runtime Darrick J. Wong
2026-05-13 17:41 ` Darrick J. Wong
2026-05-13 17:46 ` Bernd Schubert
2026-05-13 18:35 ` Darrick J. Wong
2026-05-13 19:44 ` Bernd Schubert
2026-05-14 14:35 ` Darrick J. Wong
2026-04-29 14:24 ` [PATCH 04/33] fuse: adapt FUSE_DEV_IOC_BACKING_{OPEN,CLOSE} to add new iomap devices Darrick J. Wong
2026-05-13 18:04 ` Darrick J. Wong
2026-04-29 14:24 ` [PATCH 05/33] fuse_trace: " Darrick J. Wong
2026-04-29 14:25 ` [PATCH 06/33] fuse: enable SYNCFS and ensure we flush everything before sending DESTROY Darrick J. Wong
2026-04-29 14:25 ` [PATCH 07/33] fuse: clean up per-file type inode initialization Darrick J. Wong
2026-04-29 14:25 ` [PATCH 08/33] fuse: create a per-inode flag for setting exclusive mode Darrick J. Wong
2026-05-13 18:09 ` Darrick J. Wong
2026-04-29 14:26 ` [PATCH 09/33] fuse: create a per-inode flag for toggling iomap Darrick J. Wong
2026-04-29 14:26 ` [PATCH 10/33] fuse_trace: " Darrick J. Wong
2026-04-29 14:26 ` [PATCH 11/33] fuse: isolate the other regular file IO paths from iomap Darrick J. Wong
2026-05-13 18:18 ` Darrick J. Wong
2026-04-29 14:26 ` [PATCH 12/33] fuse: implement basic iomap reporting such as FIEMAP and SEEK_{DATA,HOLE} Darrick J. Wong
2026-05-13 18:30 ` Darrick J. Wong
2026-04-29 14:27 ` [PATCH 13/33] fuse_trace: " Darrick J. Wong
2026-04-29 14:27 ` [PATCH 14/33] fuse: implement direct IO with iomap Darrick J. Wong
2026-05-13 18:44 ` Darrick J. Wong
2026-04-29 14:27 ` [PATCH 15/33] fuse_trace: " Darrick J. Wong
2026-04-29 14:27 ` [PATCH 16/33] fuse: implement buffered " Darrick J. Wong
2026-05-13 19:18 ` Darrick J. Wong [this message]
2026-04-29 14:28 ` [PATCH 17/33] fuse_trace: " Darrick J. Wong
2026-05-13 19:25 ` Darrick J. Wong
2026-04-29 14:28 ` [PATCH 18/33] fuse: use an unrestricted backing device with iomap pagecache io Darrick J. Wong
2026-05-13 21:24 ` Darrick J. Wong
2026-04-29 14:28 ` [PATCH 19/33] fuse: implement large folios for iomap pagecache files Darrick J. Wong
2026-04-29 14:28 ` [PATCH 20/33] fuse: advertise support for iomap Darrick J. Wong
2026-04-29 14:29 ` [PATCH 21/33] fuse: query filesystem geometry when using iomap Darrick J. Wong
2026-05-13 21:05 ` Darrick J. Wong
2026-04-29 14:29 ` [PATCH 22/33] fuse_trace: " Darrick J. Wong
2026-04-29 14:29 ` [PATCH 23/33] fuse: implement fadvise for iomap files Darrick J. Wong
2026-04-29 14:29 ` [PATCH 24/33] fuse: invalidate ranges of block devices being used for iomap Darrick J. Wong
2026-05-13 22:13 ` Darrick J. Wong
2026-04-29 14:30 ` [PATCH 25/33] fuse_trace: " Darrick J. Wong
2026-04-29 14:30 ` [PATCH 26/33] fuse: implement inline data file IO via iomap Darrick J. Wong
2026-05-13 20:03 ` Darrick J. Wong
2026-04-29 14:30 ` [PATCH 27/33] fuse_trace: " Darrick J. Wong
2026-05-13 20:09 ` Darrick J. Wong
2026-04-29 14:31 ` [PATCH 28/33] fuse: allow more statx fields Darrick J. Wong
2026-04-29 14:31 ` [PATCH 29/33] fuse: support atomic writes with iomap Darrick J. Wong
2026-04-29 14:31 ` [PATCH 30/33] fuse_trace: " Darrick J. Wong
2026-04-29 14:31 ` [PATCH 31/33] fuse: disable direct fs reclaim for any fuse server that uses iomap Darrick J. Wong
2026-05-13 21:27 ` Darrick J. Wong
2026-04-29 14:32 ` [PATCH 32/33] fuse: enable swapfile activation on iomap Darrick J. Wong
2026-05-14 14:42 ` Darrick J. Wong
2026-04-29 14:32 ` [PATCH 33/33] fuse: implement freeze and shutdowns for iomap filesystems Darrick J. Wong
2026-04-29 14:17 ` [PATCHSET v8 5/8] fuse: allow servers to specify root node id Darrick J. Wong
2026-04-29 14:32 ` [PATCH 1/3] fuse: make the root nodeid dynamic Darrick J. Wong
2026-04-29 14:32 ` [PATCH 2/3] fuse_trace: " Darrick J. Wong
2026-04-29 14:33 ` [PATCH 3/3] fuse: allow setting of root nodeid Darrick J. Wong
2026-04-29 14:17 ` [PATCHSET v8 6/8] fuse: handle timestamps and ACLs correctly when iomap is enabled Darrick J. Wong
2026-04-29 14:33 ` [PATCH 1/9] fuse: enable caching of timestamps Darrick J. Wong
2026-04-29 14:33 ` [PATCH 2/9] fuse: force a ctime update after a fileattr_set call when in iomap mode Darrick J. Wong
2026-04-29 14:33 ` [PATCH 3/9] fuse: allow local filesystems to set some VFS iflags Darrick J. Wong
2026-04-29 14:34 ` [PATCH 4/9] fuse_trace: " Darrick J. Wong
2026-04-29 14:34 ` [PATCH 5/9] fuse: cache atime when in iomap mode Darrick J. Wong
2026-04-29 14:34 ` [PATCH 6/9] fuse: let the kernel handle KILL_SUID/KILL_SGID for iomap filesystems Darrick J. Wong
2026-04-29 14:34 ` [PATCH 7/9] fuse_trace: " Darrick J. Wong
2026-04-29 14:35 ` [PATCH 8/9] fuse: update ctime when updating acls on an iomap inode Darrick J. Wong
2026-04-29 14:35 ` [PATCH 9/9] fuse: always cache ACLs when using iomap Darrick J. Wong
2026-04-29 14:18 ` [PATCHSET v8 7/8] fuse: cache iomap mappings for even better file IO performance Darrick J. Wong
2026-04-29 14:35 ` [PATCH 01/12] fuse: cache iomaps Darrick J. Wong
2026-04-29 14:35 ` [PATCH 02/12] fuse_trace: " Darrick J. Wong
2026-04-29 14:36 ` [PATCH 03/12] fuse: use the iomap cache for iomap_begin Darrick J. Wong
2026-04-29 14:36 ` [PATCH 04/12] fuse_trace: " Darrick J. Wong
2026-04-29 14:36 ` [PATCH 05/12] fuse: invalidate iomap cache after file updates Darrick J. Wong
2026-04-29 14:36 ` [PATCH 06/12] fuse_trace: " Darrick J. Wong
2026-04-29 14:37 ` [PATCH 07/12] fuse: enable iomap cache management Darrick J. Wong
2026-04-29 14:37 ` [PATCH 08/12] fuse_trace: " Darrick J. Wong
2026-04-29 14:37 ` [PATCH 09/12] fuse: overlay iomap inode info in struct fuse_inode Darrick J. Wong
2026-04-29 14:38 ` [PATCH 10/12] fuse: constrain iomap mapping cache size Darrick J. Wong
2026-04-29 14:38 ` [PATCH 11/12] fuse_trace: " Darrick J. Wong
2026-04-29 14:38 ` [PATCH 12/12] fuse: enable iomap Darrick J. Wong
2026-04-29 14:18 ` [PATCHSET v8 8/8] fuse: run fuse-iomap servers as a contained service Darrick J. Wong
2026-04-29 14:38 ` [PATCH 1/2] fuse: allow privileged mount helpers to pre-approve iomap usage Darrick J. Wong
2026-05-13 21:33 ` Darrick J. Wong
2026-04-29 14:39 ` [PATCH 2/2] fuse: set iomap backing device block size Darrick J. Wong
2026-04-29 14:18 ` [PATCHSET v8 1/6] libfuse: allow servers to use iomap for better file IO performance Darrick J. Wong
2026-04-29 14:39 ` [PATCH 01/25] libfuse: bump kernel and library ABI versions Darrick J. Wong
2026-04-29 14:39 ` [PATCH 02/25] libfuse: wait in do_destroy until all open files are closed Darrick J. Wong
2026-04-29 14:39 ` [PATCH 03/25] libfuse: add kernel gates for FUSE_IOMAP Darrick J. Wong
2026-04-29 14:40 ` [PATCH 04/25] libfuse: add fuse commands for iomap_begin and end Darrick J. Wong
2026-04-29 14:40 ` [PATCH 05/25] libfuse: add upper level iomap commands Darrick J. Wong
2026-04-29 14:40 ` [PATCH 06/25] libfuse: add a lowlevel notification to add a new device to iomap Darrick J. Wong
2026-04-29 14:40 ` [PATCH 07/25] libfuse: add upper-level iomap add device function Darrick J. Wong
2026-04-29 14:41 ` [PATCH 08/25] libfuse: add iomap ioend low level handler Darrick J. Wong
2026-04-29 14:41 ` [PATCH 09/25] libfuse: add upper level iomap ioend commands Darrick J. Wong
2026-04-29 14:41 ` [PATCH 10/25] libfuse: add a reply function to send FUSE_ATTR_* to the kernel Darrick J. Wong
2026-04-29 14:41 ` [PATCH 11/25] libfuse: connect high level fuse library to fuse_reply_attr_iflags Darrick J. Wong
2026-04-29 14:42 ` [PATCH 12/25] libfuse: support enabling exclusive mode for files Darrick J. Wong
2026-04-29 14:42 ` [PATCH 13/25] libfuse: support direct I/O through iomap Darrick J. Wong
2026-04-29 14:42 ` [PATCH 14/25] libfuse: don't allow hardlinking of iomap files in the upper level fuse library Darrick J. Wong
2026-04-29 14:42 ` [PATCH 15/25] libfuse: allow discovery of the kernel's iomap capabilities Darrick J. Wong
2026-04-29 14:43 ` [PATCH 16/25] libfuse: add lower level iomap_config implementation Darrick J. Wong
2026-04-29 14:43 ` [PATCH 17/25] libfuse: add upper " Darrick J. Wong
2026-04-29 14:43 ` [PATCH 18/25] libfuse: add low level code to invalidate iomap block device ranges Darrick J. Wong
2026-04-29 14:44 ` [PATCH 19/25] libfuse: add upper-level API to invalidate parts of an iomap block device Darrick J. Wong
2026-04-29 14:44 ` [PATCH 20/25] libfuse: add atomic write support Darrick J. Wong
2026-04-29 14:44 ` [PATCH 21/25] libfuse: allow disabling of fs memory reclaim and write throttling Darrick J. Wong
2026-04-29 14:44 ` [PATCH 22/25] libfuse: create a helper to transform an open regular file into an open loopdev Darrick J. Wong
2026-04-29 14:45 ` [PATCH 23/25] libfuse: add swapfile support for iomap files Darrick J. Wong
2026-04-29 14:45 ` [PATCH 24/25] libfuse: add lower-level filesystem freeze, thaw, and shutdown requests Darrick J. Wong
2026-04-29 14:45 ` [PATCH 25/25] libfuse: add upper-level filesystem freeze, thaw, and shutdown events Darrick J. Wong
2026-04-29 14:19 ` [PATCHSET v8 2/6] libfuse: allow servers to specify root node id Darrick J. Wong
2026-04-29 14:45 ` [PATCH 1/1] libfuse: allow root_nodeid mount option Darrick J. Wong
2026-04-29 14:19 ` [PATCHSET v8 3/6] libfuse: implement syncfs Darrick J. Wong
2026-04-29 14:46 ` [PATCH 1/2] libfuse: add strictatime/lazytime mount options Darrick J. Wong
2026-04-29 14:46 ` [PATCH 2/2] libfuse: set sync, immutable, and append when loading files Darrick J. Wong
2026-04-29 14:19 ` [PATCHSET v8 4/6] libfuse: add some service helper commands for iomap Darrick J. Wong
2026-04-29 14:46 ` [PATCH 1/3] mount_service: delegate iomap privilege from mount.service to fuse services Darrick J. Wong
2026-04-29 14:46 ` [PATCH 2/3] libfuse: enable setting iomap block device block size Darrick J. Wong
2026-04-29 14:47 ` [PATCH 3/3] mount_service: create loop devices for regular files Darrick J. Wong
2026-04-29 14:19 ` [PATCHSET v8 5/6] fuse: add sample iomap fuse servers Darrick J. Wong
2026-04-29 14:47 ` [PATCH 1/7] example/iomap_ll: create a simple iomap server Darrick J. Wong
2026-04-29 14:47 ` [PATCH 2/7] example/iomap_ll: track block state Darrick J. Wong
2026-04-29 14:47 ` [PATCH 3/7] example/iomap_ll: implement atomic writes Darrick J. Wong
2026-04-29 14:48 ` [PATCH 4/7] example/iomap_inline_ll: create a simple server to test inlinedata Darrick J. Wong
2026-04-29 14:48 ` [PATCH 5/7] example/iomap_ow_ll: create a simple iomap out of place write server Darrick J. Wong
2026-04-29 14:48 ` [PATCH 6/7] example/iomap_ow_ll: implement atomic writes Darrick J. Wong
2026-04-29 14:48 ` [PATCH 7/7] example/iomap_service_ll: create a sample systemd service fuse server Darrick J. Wong
2026-04-29 14:20 ` [PATCHSET v8 6/6] libfuse: cache iomap mappings for even better file IO performance Darrick J. Wong
2026-04-29 14:49 ` [PATCH 1/9] libfuse: enable iomap cache management for lowlevel fuse Darrick J. Wong
2026-04-29 14:49 ` [PATCH 2/9] libfuse: add upper-level iomap cache management Darrick J. Wong
2026-04-29 14:49 ` [PATCH 3/9] libfuse: allow constraining of iomap mapping cache size Darrick J. Wong
2026-04-29 14:50 ` [PATCH 4/9] libfuse: add upper-level iomap mapping cache constraint code Darrick J. Wong
2026-04-29 14:50 ` [PATCH 5/9] libfuse: enable iomap Darrick J. Wong
2026-04-29 14:50 ` [PATCH 6/9] example/iomap_ll: cache mappings for later Darrick J. Wong
2026-04-29 14:50 ` [PATCH 7/9] example/iomap_inline_ll: cache iomappings in the kernel Darrick J. Wong
2026-04-29 14:51 ` [PATCH 8/9] example/iomap_ow_ll: " Darrick J. Wong
2026-04-29 14:51 ` [PATCH 9/9] example/iomap_service_ll: " Darrick J. Wong
2026-04-29 14:20 ` [PATCHSET v8 1/6] libext2fs: refactoring for fuse2fs iomap support Darrick J. Wong
2026-04-29 14:51 ` [PATCH 1/5] libext2fs: invalidate cached blocks when freeing them Darrick J. Wong
2026-04-29 14:51 ` [PATCH 2/5] libext2fs: only flush affected blocks in unix_write_byte Darrick J. Wong
2026-04-29 14:52 ` [PATCH 3/5] libext2fs: allow unix_write_byte when the write would be aligned Darrick J. Wong
2026-04-29 14:52 ` [PATCH 4/5] libext2fs: allow clients to ask to write full superblocks Darrick J. Wong
2026-04-29 14:52 ` [PATCH 5/5] libext2fs: allow callers to disallow I/O to file data blocks Darrick J. Wong
2026-04-29 14:20 ` [PATCHSET v8 2/6] fuse2fs: use fuse iomap data paths for better file I/O performance Darrick J. Wong
2026-04-29 14:52 ` [PATCH 01/19] fuse2fs: implement bare minimum iomap for file mapping reporting Darrick J. Wong
2026-04-29 14:53 ` [PATCH 02/19] fuse2fs: add iomap= mount option Darrick J. Wong
2026-04-29 14:53 ` [PATCH 03/19] fuse2fs: implement iomap configuration Darrick J. Wong
2026-04-29 14:53 ` [PATCH 04/19] fuse2fs: register block devices for use with iomap Darrick J. Wong
2026-04-29 14:53 ` [PATCH 05/19] fuse2fs: implement directio file reads Darrick J. Wong
2026-04-29 14:54 ` [PATCH 06/19] fuse2fs: add extent dump function for debugging Darrick J. Wong
2026-04-29 14:54 ` [PATCH 07/19] fuse2fs: implement direct write support Darrick J. Wong
2026-04-29 14:54 ` [PATCH 08/19] fuse2fs: turn on iomap for pagecache IO Darrick J. Wong
2026-04-29 14:54 ` [PATCH 09/19] fuse2fs: don't zero bytes in punch hole Darrick J. Wong
2026-04-29 14:55 ` [PATCH 10/19] fuse2fs: don't do file data block IO when iomap is enabled Darrick J. Wong
2026-04-29 14:55 ` [PATCH 11/19] fuse2fs: try to create loop device when ext4 device is a regular file Darrick J. Wong
2026-04-29 14:55 ` [PATCH 12/19] fuse2fs: enable file IO to inline data files Darrick J. Wong
2026-04-29 14:56 ` [PATCH 13/19] fuse2fs: set iomap-related inode flags Darrick J. Wong
2026-04-29 14:56 ` [PATCH 14/19] fuse2fs: configure block device block size Darrick J. Wong
2026-04-29 14:56 ` [PATCH 15/19] fuse4fs: separate invalidation Darrick J. Wong
2026-04-29 14:56 ` [PATCH 16/19] fuse2fs: implement statx Darrick J. Wong
2026-04-29 14:57 ` [PATCH 17/19] fuse2fs: enable atomic writes Darrick J. Wong
2026-04-29 14:57 ` [PATCH 18/19] fuse4fs: disable fs reclaim and write throttling Darrick J. Wong
2026-04-29 14:57 ` [PATCH 19/19] fuse2fs: implement freeze and shutdown requests Darrick J. Wong
2026-04-29 14:20 ` [PATCHSET v8 3/6] fuse4fs: adapt iomap for fuse services Darrick J. Wong
2026-04-29 14:57 ` [PATCH 1/3] fuse4fs: configure iomap when running as a service Darrick J. Wong
2026-04-29 14:58 ` [PATCH 2/3] fuse4fs: set iomap backing device blocksize Darrick J. Wong
2026-04-29 14:58 ` [PATCH 3/3] fuse4fs: ask for loop devices when opening via fuservicemount Darrick J. Wong
2026-04-29 14:21 ` [PATCHSET v8 4/6] fuse4fs: specify the root node id Darrick J. Wong
2026-04-29 14:58 ` [PATCH 1/1] fuse4fs: don't use inode number translation when possible Darrick J. Wong
2026-04-29 14:21 ` [PATCHSET v8 5/6] fuse2fs: handle timestamps and ACLs correctly when iomap is enabled Darrick J. Wong
2026-04-29 14:58 ` [PATCH 01/10] fuse2fs: add strictatime/lazytime mount options Darrick J. Wong
2026-04-29 14:59 ` [PATCH 02/10] fuse2fs: skip permission checking on utimens when iomap is enabled Darrick J. Wong
2026-04-29 14:59 ` [PATCH 03/10] fuse2fs: let the kernel tell us about acl/mode updates Darrick J. Wong
2026-04-29 14:59 ` [PATCH 04/10] fuse2fs: better debugging for file mode updates Darrick J. Wong
2026-04-29 14:59 ` [PATCH 05/10] fuse2fs: debug timestamp updates Darrick J. Wong
2026-04-29 15:00 ` [PATCH 06/10] fuse2fs: use coarse timestamps for iomap mode Darrick J. Wong
2026-04-29 15:00 ` [PATCH 07/10] fuse2fs: add tracing for retrieving timestamps Darrick J. Wong
2026-04-29 15:00 ` [PATCH 08/10] fuse2fs: enable syncfs Darrick J. Wong
2026-04-29 15:00 ` [PATCH 09/10] fuse2fs: set sync, immutable, and append at file load time Darrick J. Wong
2026-04-29 15:01 ` [PATCH 10/10] fuse4fs: increase attribute timeout in iomap mode Darrick J. Wong
2026-04-29 14:21 ` [PATCHSET v8 6/6] fuse2fs: cache iomap mappings for even better file IO performance Darrick J. Wong
2026-04-29 15:01 ` [PATCH 1/4] fuse2fs: enable caching of iomaps Darrick J. Wong
2026-04-29 15:01 ` [PATCH 2/4] fuse2fs: constrain iomap mapping cache size Darrick J. Wong
2026-04-29 15:02 ` [PATCH 3/4] fuse4fs: upsert first file mapping to kernel on open Darrick J. Wong
2026-04-29 15:02 ` [PATCH 4/4] fuse2fs: enable iomap Darrick J. Wong
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20260513191821.GN9544@frogsfrogsfrogs \
--to=djwong@kernel.org \
--cc=bernd@bsbernd.com \
--cc=fuse-devel@lists.linux.dev \
--cc=joannelkoong@gmail.com \
--cc=linux-fsdevel@vger.kernel.org \
--cc=miklos@szeredi.hu \
--cc=neal@gompa.dev \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox