linux-btrfs.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH] Btrfs: unlock extent when the dio completes for reads
@ 2012-07-31 20:33 Josef Bacik
  2012-07-31 21:30 ` Zach Brown
  0 siblings, 1 reply; 3+ messages in thread
From: Josef Bacik @ 2012-07-31 20:33 UTC (permalink / raw)
  To: linux-btrfs

A deadlock in xfstests 113 was uncovered by commit

d187663ef24cd3d033f0cbf2867e70b36a3a90b8

This is because we would not return EIOCBQUEUED for short AIO reads, instead
we'd wait for the DIO to complete and then return the amount of data we
transferred, which would allow our stuff to unlock the remaning amount.  But
with this change this no longer happens, so if we have a short AIO read (for
example if we try to read past EOF), we could leave the section from EOF to
the end of where we tried to read locked.  To fix this we need to store the
end of the region we've locked in our locked state bit and then only unlock
when the dio has finished completely.  This makes sure we always unlock the
entire range we locked when reading.  Before this patch we would hang when
running 113, no this no longer happens.  Thanks to Zach Brown for the idea,

Signed-off-by: Josef Bacik <jbacik@fusionio.com>
---
 fs/btrfs/inode.c |   35 +++++++++++++++++++++++++++++------
 1 files changed, 29 insertions(+), 6 deletions(-)

diff --git a/fs/btrfs/inode.c b/fs/btrfs/inode.c
index 48bdfd2..c2f9300 100644
--- a/fs/btrfs/inode.c
+++ b/fs/btrfs/inode.c
@@ -5810,9 +5810,6 @@ static int btrfs_get_blocks_direct(struct inode *inode, sector_t iblock,
 	if (!create && (em->block_start == EXTENT_MAP_HOLE ||
 			test_bit(EXTENT_FLAG_PREALLOC, &em->flags))) {
 		free_extent_map(em);
-		/* DIO will do one hole at a time, so just unlock a sector */
-		unlock_extent(&BTRFS_I(inode)->io_tree, start,
-			      start + root->sectorsize - 1);
 		return 0;
 	}
 
@@ -5961,8 +5958,6 @@ static void btrfs_endio_direct_read(struct bio *bio, int err)
 		bvec++;
 	} while (bvec <= bvec_end);
 
-	unlock_extent(&BTRFS_I(inode)->io_tree, dip->logical_offset,
-		      dip->logical_offset + dip->bytes - 1);
 	bio->bi_private = dip->private;
 
 	kfree(dip->csums);
@@ -6340,6 +6335,26 @@ static ssize_t check_direct_IO(struct btrfs_root *root, int rw, struct kiocb *io
 out:
 	return retval;
 }
+
+static void btrfs_direct_read_iodone(struct kiocb *iocb, loff_t offset,
+				     ssize_t bytes, void *private, int ret,
+				     bool is_async)
+{
+	struct inode *inode = iocb->ki_filp->f_mapping->host;
+	u64 lockend;
+
+	if (get_state_private(&BTRFS_I(inode)->io_tree, offset, &lockend)) {
+		WARN_ON(1);
+		goto out;
+	}
+
+	unlock_extent(&BTRFS_I(inode)->io_tree, offset, lockend);
+out:
+	if (is_async)
+		aio_complete(iocb, ret, 0);
+	inode_dio_done(inode);
+}
+
 static ssize_t btrfs_direct_IO(int rw, struct kiocb *iocb,
 			const struct iovec *iov, loff_t offset,
 			unsigned long nr_segs)
@@ -6353,6 +6368,7 @@ static ssize_t btrfs_direct_IO(int rw, struct kiocb *iocb,
 	int writing = rw & WRITE;
 	int write_bits = 0;
 	size_t count = iov_length(iov, nr_segs);
+	dio_iodone_t *iodone = NULL;
 
 	if (check_direct_IO(BTRFS_I(inode)->root, rw, iocb, iov,
 			    offset, nr_segs)) {
@@ -6438,6 +6454,9 @@ static ssize_t btrfs_direct_IO(int rw, struct kiocb *iocb,
 					 1, 0, &cached_state, GFP_NOFS);
 			goto out;
 		}
+	} else {
+		set_state_private(&BTRFS_I(inode)->io_tree, lockstart, lockend);
+		iodone = btrfs_direct_read_iodone;
 	}
 
 	free_extent_state(cached_state);
@@ -6445,9 +6464,13 @@ static ssize_t btrfs_direct_IO(int rw, struct kiocb *iocb,
 
 	ret = __blockdev_direct_IO(rw, iocb, inode,
 		   BTRFS_I(inode)->root->fs_info->fs_devices->latest_bdev,
-		   iov, offset, nr_segs, btrfs_get_blocks_direct, NULL,
+		   iov, offset, nr_segs, btrfs_get_blocks_direct, iodone,
 		   btrfs_submit_direct, 0);
 
+	/* If we're reading we will be unlocked by the iodone call */
+	if (!writing)
+		goto out;
+
 	if (ret < 0 && ret != -EIOCBQUEUED) {
 		clear_extent_bit(&BTRFS_I(inode)->io_tree, offset,
 			      offset + iov_length(iov, nr_segs) - 1,
-- 
1.7.7.6


^ permalink raw reply related	[flat|nested] 3+ messages in thread

* Re: [PATCH] Btrfs: unlock extent when the dio completes for reads
  2012-07-31 20:33 [PATCH] Btrfs: unlock extent when the dio completes for reads Josef Bacik
@ 2012-07-31 21:30 ` Zach Brown
  2012-08-01 12:10   ` Josef Bacik
  0 siblings, 1 reply; 3+ messages in thread
From: Zach Brown @ 2012-07-31 21:30 UTC (permalink / raw)
  To: Josef Bacik; +Cc: linux-btrfs

> +static void btrfs_direct_read_iodone(struct kiocb *iocb, loff_t offset,
> +				     ssize_t bytes, void *private, int ret,
> +				     bool is_async)
> +{
> +	struct inode *inode = iocb->ki_filp->f_mapping->host;
> +	u64 lockend;
> +
> +	if (get_state_private(&BTRFS_I(inode)->io_tree, offset, &lockend)) {
> +		WARN_ON(1);
> +		goto out;
> +	}
> +
> +	unlock_extent(&BTRFS_I(inode)->io_tree, offset, lockend);
> +out:
> +	if (is_async)
> +		aio_complete(iocb, ret, 0);
> +	inode_dio_done(inode);
> +}

Hmm.  I'm worried that this can be called from interrupt context and the
state/extent calls there seem to use bare spin locks.  Did you run this
with lockdep?  Am I confused?

- z

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [PATCH] Btrfs: unlock extent when the dio completes for reads
  2012-07-31 21:30 ` Zach Brown
@ 2012-08-01 12:10   ` Josef Bacik
  0 siblings, 0 replies; 3+ messages in thread
From: Josef Bacik @ 2012-08-01 12:10 UTC (permalink / raw)
  To: Zach Brown; +Cc: Josef Bacik, linux-btrfs@vger.kernel.org

On Tue, Jul 31, 2012 at 03:30:49PM -0600, Zach Brown wrote:
> > +static void btrfs_direct_read_iodone(struct kiocb *iocb, loff_t offset,
> > +				     ssize_t bytes, void *private, int ret,
> > +				     bool is_async)
> > +{
> > +	struct inode *inode = iocb->ki_filp->f_mapping->host;
> > +	u64 lockend;
> > +
> > +	if (get_state_private(&BTRFS_I(inode)->io_tree, offset, &lockend)) {
> > +		WARN_ON(1);
> > +		goto out;
> > +	}
> > +
> > +	unlock_extent(&BTRFS_I(inode)->io_tree, offset, lockend);
> > +out:
> > +	if (is_async)
> > +		aio_complete(iocb, ret, 0);
> > +	inode_dio_done(inode);
> > +}
> 
> Hmm.  I'm worried that this can be called from interrupt context and the
> state/extent calls there seem to use bare spin locks.  Did you run this
> with lockdep?  Am I confused?
> 

Our read completions are run in a helper thread since we need to do things like
clear the extent state stuff and such so it's safe in this regard.  Course I hit
a lockup with xfstests 91 because there are some cases where we can return
errors without doing the dio complete (rightly so since we haven't setup the dio
yet), so now I have to figure out how to know when we've been unlocked via dio
complete or that I've gotten an error back and already unlocked it...

Josef

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2012-08-01 12:10 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2012-07-31 20:33 [PATCH] Btrfs: unlock extent when the dio completes for reads Josef Bacik
2012-07-31 21:30 ` Zach Brown
2012-08-01 12:10   ` Josef Bacik

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).