From: Matthew Wilcox <willy@infradead.org>
To: linux-kernel@vger.kernel.org
Cc: Matthew Wilcox <mawilcox@microsoft.com>,
Ross Zwisler <ross.zwisler@linux.intel.com>,
David Howells <dhowells@redhat.com>, Shaohua Li <shli@kernel.org>,
Jens Axboe <axboe@kernel.dk>, Rehas Sachdeva <aquannie@gmail.com>,
Marc Zyngier <marc.zyngier@arm.com>,
linux-mm@kvack.org, linux-fsdevel@vger.kernel.org,
linux-f2fs-devel@lists.sourceforge.net,
linux-nilfs@vger.kernel.org, linux-btrfs@vger.kernel.org,
linux-xfs@vger.kernel.org, linux-usb@vger.kernel.org,
linux-raid@vger.kernel.org
Subject: [PATCH v5 60/78] dax: More XArray conversion
Date: Fri, 15 Dec 2017 14:04:32 -0800 [thread overview]
Message-ID: <20171215220450.7899-61-willy@infradead.org> (raw)
In-Reply-To: <20171215220450.7899-1-willy@infradead.org>
From: Matthew Wilcox <mawilcox@microsoft.com>
This time, we want to convert get_unlocked_mapping_entry() to use the
XArray. That has a ripple effect, causing us to change the waitqueues
to hash on the address of the xarray rather than the address of the
mapping (functionally equivalent), and create a lot of on-the-stack
xa_state which are only used as a container for passing the xarray and
the index down to deeper function calls.
Also rename dax_wake_mapping_entry_waiter() to dax_wake_entry().
Signed-off-by: Matthew Wilcox <mawilcox@microsoft.com>
---
fs/dax.c | 72 +++++++++++++++++++++++++++++-----------------------------------
1 file changed, 33 insertions(+), 39 deletions(-)
diff --git a/fs/dax.c b/fs/dax.c
index 46a4d83b1b46..2629ffa70575 100644
--- a/fs/dax.c
+++ b/fs/dax.c
@@ -120,7 +120,7 @@ static int dax_is_empty_entry(void *entry)
* DAX radix tree locking
*/
struct exceptional_entry_key {
- struct address_space *mapping;
+ struct xarray *xa;
pgoff_t entry_start;
};
@@ -129,9 +129,10 @@ struct wait_exceptional_entry_queue {
struct exceptional_entry_key key;
};
-static wait_queue_head_t *dax_entry_waitqueue(struct address_space *mapping,
- pgoff_t index, void *entry, struct exceptional_entry_key *key)
+static wait_queue_head_t *dax_entry_waitqueue(struct xa_state *xas,
+ void *entry, struct exceptional_entry_key *key)
{
+ unsigned long index = xas->xa_index;
unsigned long hash;
/*
@@ -142,10 +143,10 @@ static wait_queue_head_t *dax_entry_waitqueue(struct address_space *mapping,
if (dax_is_pmd_entry(entry))
index &= ~PG_PMD_COLOUR;
- key->mapping = mapping;
+ key->xa = xas->xa;
key->entry_start = index;
- hash = hash_long((unsigned long)mapping ^ index, DAX_WAIT_TABLE_BITS);
+ hash = hash_long((unsigned long)xas->xa ^ index, DAX_WAIT_TABLE_BITS);
return wait_table + hash;
}
@@ -156,7 +157,7 @@ static int wake_exceptional_entry_func(wait_queue_entry_t *wait, unsigned int mo
struct wait_exceptional_entry_queue *ewait =
container_of(wait, struct wait_exceptional_entry_queue, wait);
- if (key->mapping != ewait->key.mapping ||
+ if (key->xa != ewait->key.xa ||
key->entry_start != ewait->key.entry_start)
return 0;
return autoremove_wake_function(wait, mode, sync, NULL);
@@ -167,13 +168,12 @@ static int wake_exceptional_entry_func(wait_queue_entry_t *wait, unsigned int mo
* The important information it's conveying is whether the entry at
* this index used to be a PMD entry.
*/
-static void dax_wake_mapping_entry_waiter(struct address_space *mapping,
- pgoff_t index, void *entry, bool wake_all)
+static void dax_wake_entry(struct xa_state *xas, void *entry, bool wake_all)
{
struct exceptional_entry_key key;
wait_queue_head_t *wq;
- wq = dax_entry_waitqueue(mapping, index, entry, &key);
+ wq = dax_entry_waitqueue(xas, entry, &key);
/*
* Checking for locked entry and prepare_to_wait_exclusive() happens
@@ -205,10 +205,9 @@ static inline void *lock_slot(struct xa_state *xas)
*
* Must be called with xa_lock held.
*/
-static void *get_unlocked_mapping_entry(struct address_space *mapping,
- pgoff_t index, void ***slotp)
+static void *get_unlocked_mapping_entry(struct xa_state *xas)
{
- void *entry, **slot;
+ void *entry;
struct wait_exceptional_entry_queue ewait;
wait_queue_head_t *wq;
@@ -216,22 +215,19 @@ static void *get_unlocked_mapping_entry(struct address_space *mapping,
ewait.wait.func = wake_exceptional_entry_func;
for (;;) {
- entry = __radix_tree_lookup(&mapping->pages, index, NULL,
- &slot);
- if (!entry ||
- WARN_ON_ONCE(!xa_is_value(entry)) || !dax_locked(entry)) {
- if (slotp)
- *slotp = slot;
+ entry = xas_load(xas);
+ if (!entry || WARN_ON_ONCE(!xa_is_value(entry)) ||
+ !dax_locked(entry))
return entry;
- }
- wq = dax_entry_waitqueue(mapping, index, entry, &ewait.key);
+ wq = dax_entry_waitqueue(xas, entry, &ewait.key);
prepare_to_wait_exclusive(wq, &ewait.wait,
TASK_UNINTERRUPTIBLE);
- xa_unlock_irq(&mapping->pages);
+ xas_pause(xas);
+ xas_unlock_irq(xas);
schedule();
finish_wait(wq, &ewait.wait);
- xa_lock_irq(&mapping->pages);
+ xas_lock_irq(xas);
}
}
@@ -251,7 +247,7 @@ static void dax_unlock_mapping_entry(struct address_space *mapping,
xas_store(&xas, entry);
/* Safe to not call xas_pause here -- we don't touch the array after */
xas_unlock_irq(&xas);
- dax_wake_mapping_entry_waiter(mapping, index, entry, false);
+ dax_wake_entry(&xas, entry, false);
}
static void put_locked_mapping_entry(struct address_space *mapping,
@@ -264,14 +260,13 @@ static void put_locked_mapping_entry(struct address_space *mapping,
* Called when we are done with radix tree entry we looked up via
* get_unlocked_mapping_entry() and which we didn't lock in the end.
*/
-static void put_unlocked_mapping_entry(struct address_space *mapping,
- pgoff_t index, void *entry)
+static void put_unlocked_mapping_entry(struct xa_state *xas, void *entry)
{
if (!entry)
return;
/* We have to wake up next waiter for the radix tree entry lock */
- dax_wake_mapping_entry_waiter(mapping, index, entry, false);
+ dax_wake_entry(xas, entry, false);
}
/*
@@ -308,7 +303,7 @@ static void *grab_mapping_entry(struct address_space *mapping, pgoff_t index,
restart:
xa_lock_irq(&mapping->pages);
- entry = get_unlocked_mapping_entry(mapping, index, &slot);
+ entry = get_unlocked_mapping_entry(&xas);
if (WARN_ON_ONCE(entry && !xa_is_value(entry))) {
entry = ERR_PTR(-EIO);
@@ -318,8 +313,7 @@ static void *grab_mapping_entry(struct address_space *mapping, pgoff_t index,
if (entry) {
if (size_flag & DAX_PMD) {
if (dax_is_pte_entry(entry)) {
- put_unlocked_mapping_entry(mapping, index,
- entry);
+ put_unlocked_mapping_entry(&xas, entry);
entry = ERR_PTR(-EEXIST);
goto out_unlock;
}
@@ -382,8 +376,7 @@ static void *grab_mapping_entry(struct address_space *mapping, pgoff_t index,
if (pmd_downgrade) {
radix_tree_delete(&mapping->pages, index);
mapping->nrexceptional--;
- dax_wake_mapping_entry_waiter(mapping, index, entry,
- true);
+ dax_wake_entry(&xas, entry, true);
}
entry = dax_radix_locked_entry(0, size_flag | DAX_EMPTY);
@@ -417,12 +410,13 @@ static void *grab_mapping_entry(struct address_space *mapping, pgoff_t index,
static int __dax_invalidate_mapping_entry(struct address_space *mapping,
pgoff_t index, bool trunc)
{
+ XA_STATE(xas, &mapping->pages, index);
int ret = 0;
void *entry;
struct radix_tree_root *pages = &mapping->pages;
xa_lock_irq(&mapping->pages);
- entry = get_unlocked_mapping_entry(mapping, index, NULL);
+ entry = get_unlocked_mapping_entry(&xas);
if (!entry || WARN_ON_ONCE(!xa_is_value(entry)))
goto out;
if (!trunc &&
@@ -433,7 +427,7 @@ static int __dax_invalidate_mapping_entry(struct address_space *mapping,
mapping->nrexceptional--;
ret = 1;
out:
- put_unlocked_mapping_entry(mapping, index, entry);
+ put_unlocked_mapping_entry(&xas, entry);
xa_unlock_irq(&mapping->pages);
return ret;
}
@@ -641,7 +635,7 @@ static int dax_writeback_one(struct block_device *bdev,
{
struct radix_tree_root *pages = &mapping->pages;
XA_STATE(xas, pages, index);
- void *entry2, **slot, *kaddr;
+ void *entry2, *kaddr;
long ret = 0, id;
sector_t sector;
pgoff_t pgoff;
@@ -656,7 +650,7 @@ static int dax_writeback_one(struct block_device *bdev,
return -EIO;
xa_lock_irq(&mapping->pages);
- entry2 = get_unlocked_mapping_entry(mapping, index, &slot);
+ entry2 = get_unlocked_mapping_entry(&xas);
/* Entry got punched out / reallocated? */
if (!entry2 || WARN_ON_ONCE(!xa_is_value(entry2)))
goto put_unlocked;
@@ -734,7 +728,7 @@ static int dax_writeback_one(struct block_device *bdev,
return ret;
put_unlocked:
- put_unlocked_mapping_entry(mapping, index, entry2);
+ put_unlocked_mapping_entry(&xas, entry2);
xa_unlock_irq(&mapping->pages);
return ret;
}
@@ -1504,16 +1498,16 @@ static int dax_insert_pfn_mkwrite(struct vm_fault *vmf,
struct address_space *mapping = vmf->vma->vm_file->f_mapping;
pgoff_t index = vmf->pgoff;
XA_STATE(xas, &mapping->pages, index);
- void *entry, **slot;
+ void *entry;
int vmf_ret, error;
xa_lock_irq(&mapping->pages);
- entry = get_unlocked_mapping_entry(mapping, index, &slot);
+ entry = get_unlocked_mapping_entry(&xas);
/* Did we race with someone splitting entry or so? */
if (!entry ||
(pe_size == PE_SIZE_PTE && !dax_is_pte_entry(entry)) ||
(pe_size == PE_SIZE_PMD && !dax_is_pmd_entry(entry))) {
- put_unlocked_mapping_entry(mapping, index, entry);
+ put_unlocked_mapping_entry(&xas, entry);
xa_unlock_irq(&mapping->pages);
trace_dax_insert_pfn_mkwrite_no_entry(mapping->host, vmf,
VM_FAULT_NOPAGE);
--
2.15.1
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2017-12-15 22:04 UTC|newest]
Thread overview: 95+ messages / expand[flat|nested] mbox.gz Atom feed top
2017-12-15 22:03 [PATCH v5 00/78] XArray v5 Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 01/78] xfs: Rename xa_ elements to ail_ Matthew Wilcox
2018-01-03 1:01 ` Darrick J. Wong
2017-12-15 22:03 ` [PATCH v5 02/78] fscache: Use appropriate radix tree accessors Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 03/78] xarray: Add the xa_lock to the radix_tree_root Matthew Wilcox
2017-12-26 16:54 ` Kirill A. Shutemov
2017-12-27 3:43 ` Matthew Wilcox
2017-12-27 3:58 ` Matthew Wilcox
2017-12-27 10:18 ` Kirill A. Shutemov
2018-01-02 18:01 ` Darrick J. Wong
2018-01-02 22:41 ` Matthew Wilcox
2017-12-27 10:17 ` Kirill A. Shutemov
2017-12-15 22:03 ` [PATCH v5 04/78] page cache: Use xa_lock Matthew Wilcox
2017-12-26 16:56 ` Kirill A. Shutemov
2017-12-15 22:03 ` [PATCH v5 05/78] xarray: Replace exceptional entries Matthew Wilcox
2017-12-26 17:15 ` Kirill A. Shutemov
2017-12-27 3:05 ` Matthew Wilcox
2017-12-27 10:24 ` Kirill A. Shutemov
2017-12-15 22:03 ` [PATCH v5 06/78] xarray: Change definition of sibling entries Matthew Wilcox
2017-12-26 17:21 ` Kirill A. Shutemov
2017-12-27 3:13 ` Matthew Wilcox
2017-12-27 10:26 ` Kirill A. Shutemov
2017-12-15 22:03 ` [PATCH v5 07/78] xarray: Add definition of struct xarray Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 08/78] xarray: Define struct xa_node Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 09/78] xarray: Add documentation Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 10/78] xarray: Add xa_load Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 11/78] xarray: Add xa_get_tag, xa_set_tag and xa_clear_tag Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 12/78] xarray: Add xa_store Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 13/78] xarray: Add xa_cmpxchg Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 14/78] xarray: Add xa_for_each Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 15/78] xarray: Add xas_for_each_tag Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 16/78] xarray: Add xa_get_entries, xa_get_tagged and xa_get_maybe_tag Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 17/78] xarray: Add xa_destroy Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 18/78] xarray: Add xas_next and xas_prev Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 19/78] xarray: Add xas_create_range Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 20/78] xarray: Add MAINTAINERS entry Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 21/78] xarray: Add ability to store errno values Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 22/78] idr: Convert to XArray Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 23/78] ida: " Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 24/78] page cache: Convert hole search " Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 25/78] page cache: Add page_cache_range_empty function Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 26/78] page cache: Add and replace pages using the XArray Matthew Wilcox
2017-12-15 22:03 ` [PATCH v5 27/78] page cache: Convert page deletion to XArray Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 28/78] page cache: Convert page cache lookups " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 29/78] page cache: Convert delete_batch " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 30/78] page cache: Remove stray radix comment Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 31/78] mm: Convert page-writeback to XArray Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 32/78] mm: Convert workingset " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 33/78] mm: Convert truncate " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 34/78] mm: Convert add_to_swap_cache " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 35/78] mm: Convert delete_from_swap_cache " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 36/78] mm: Convert __do_page_cache_readahead " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 37/78] mm: Convert page migration " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 38/78] mm: Convert huge_memory " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 39/78] mm: Convert collapse_shmem " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 40/78] mm: Convert khugepaged_scan_shmem " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 41/78] pagevec: Use xa_tag_t Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 42/78] shmem: Convert replace to XArray Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 43/78] shmem: Convert shmem_confirm_swap " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 44/78] shmem: Convert find_swap_entry " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 45/78] shmem: Convert shmem_tag_pins " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 46/78] shmem: Convert shmem_wait_for_pins " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 47/78] shmem: Convert shmem_add_to_page_cache " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 48/78] shmem: Convert shmem_alloc_hugepage " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 49/78] shmem: Convert shmem_free_swap " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 50/78] shmem: Convert shmem_partial_swap_usage " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 51/78] shmem: Comment fixups Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 52/78] btrfs: Convert page cache to XArray Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 53/78] fs: Convert buffer " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 54/78] fs: Convert writeback " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 55/78] nilfs2: Convert " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 56/78] f2fs: " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 57/78] lustre: " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 58/78] dax: Convert dax_unlock_mapping_entry " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 59/78] dax: Convert lock_slot " Matthew Wilcox
2017-12-15 22:04 ` Matthew Wilcox [this message]
2017-12-15 22:04 ` [PATCH v5 61/78] dax: Convert __dax_invalidate_mapping_entry " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 62/78] dax: Convert dax_writeback_one " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 63/78] dax: Convert dax_insert_pfn_mkwrite " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 64/78] dax: Convert dax_insert_mapping_entry " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 65/78] dax: Convert grab_mapping_entry " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 66/78] dax: Fix sparse warning Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 67/78] page cache: Finish XArray conversion Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 68/78] mm: Convert cgroup writeback to XArray Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 69/78] vmalloc: Convert " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 70/78] brd: " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 71/78] xfs: Convert m_perag_tree " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 72/78] xfs: Convert pag_ici_root " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 73/78] xfs: Convert xfs dquot " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 74/78] xfs: Convert mru cache " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 75/78] usb: Convert xhci-mem " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 76/78] md: Convert raid5-cache " Matthew Wilcox
2017-12-15 22:04 ` [PATCH v5 77/78] irqdomain: Convert " Matthew Wilcox
2017-12-16 10:51 ` Marc Zyngier
2017-12-15 22:04 ` [PATCH v5 78/78] fscache: " Matthew Wilcox
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20171215220450.7899-61-willy@infradead.org \
--to=willy@infradead.org \
--cc=aquannie@gmail.com \
--cc=axboe@kernel.dk \
--cc=dhowells@redhat.com \
--cc=linux-btrfs@vger.kernel.org \
--cc=linux-f2fs-devel@lists.sourceforge.net \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=linux-nilfs@vger.kernel.org \
--cc=linux-raid@vger.kernel.org \
--cc=linux-usb@vger.kernel.org \
--cc=linux-xfs@vger.kernel.org \
--cc=marc.zyngier@arm.com \
--cc=mawilcox@microsoft.com \
--cc=ross.zwisler@linux.intel.com \
--cc=shli@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).