From: Matthew Wilcox <willy@infradead.org>
To: Andrew Morton <akpm@linux-foundation.org>
Cc: Matthew Wilcox <mawilcox@microsoft.com>,
linux-kernel@vger.kernel.org, linux-mm@kvack.org,
linux-fsdevel@vger.kernel.org
Subject: [PATCH v7 22/61] xarray: Add xas_next and xas_prev
Date: Mon, 19 Feb 2018 11:45:17 -0800 [thread overview]
Message-ID: <20180219194556.6575-23-willy@infradead.org> (raw)
In-Reply-To: <20180219194556.6575-1-willy@infradead.org>
From: Matthew Wilcox <mawilcox@microsoft.com>
These two functions move the xas index by one position, and adjust the
rest of the iterator state to match it. This is more efficient than
calling xas_set() as it keeps the iterator at the leaves of the tree
instead of walking the iterator from the root each time.
Signed-off-by: Matthew Wilcox <mawilcox@microsoft.com>
---
include/linux/xarray.h | 67 +++++++++
lib/xarray.c | 74 ++++++++++
tools/testing/radix-tree/xarray-test.c | 261 ++++++++++++++++++++++++++++++++-
3 files changed, 401 insertions(+), 1 deletion(-)
diff --git a/include/linux/xarray.h b/include/linux/xarray.h
index 96773f83ae03..c8a0ddc1b3df 100644
--- a/include/linux/xarray.h
+++ b/include/linux/xarray.h
@@ -683,6 +683,12 @@ static inline bool xas_not_node(struct xa_node *node)
return ((unsigned long)node & 3) || !node;
}
+/* True if the node represents RESTART or an error */
+static inline bool xas_frozen(struct xa_node *node)
+{
+ return (unsigned long)node & 2;
+}
+
/* True if the node represents head-of-tree, RESTART or BOUNDS */
static inline bool xas_top(struct xa_node *node)
{
@@ -940,6 +946,67 @@ enum {
for (entry = xas_find_tag(xas, max, tag); entry; \
entry = xas_next_tag(xas, max, tag))
+void *__xas_next(struct xa_state *);
+void *__xas_prev(struct xa_state *);
+
+/**
+ * xas_prev() - Move iterator to previous index.
+ * @xas: XArray operation state.
+ *
+ * If the @xas was in an error state, it will remain in an error state
+ * and this function will return %NULL. If the @xas has never been walked,
+ * it will have the effect of calling xas_load(). Otherwise one will be
+ * subtracted from the index and the state will be walked to the correct
+ * location in the array for the next operation.
+ *
+ * If the iterator was referencing index 0, this function wraps
+ * around to %ULONG_MAX.
+ *
+ * Return: The entry at the new index. This may be %NULL or an internal
+ * entry, although it should never be a node entry.
+ */
+static inline void *xas_prev(struct xa_state *xas)
+{
+ struct xa_node *node = xas->xa_node;
+
+ if (unlikely(xas_not_node(node) || node->shift ||
+ xas->xa_offset == 0))
+ return __xas_prev(xas);
+
+ xas->xa_index--;
+ xas->xa_offset--;
+ return xa_entry(xas->xa, node, xas->xa_offset);
+}
+
+/**
+ * xas_next() - Move state to next index.
+ * @xas: XArray operation state.
+ *
+ * If the @xas was in an error state, it will remain in an error state
+ * and this function will return %NULL. If the @xas has never been walked,
+ * it will have the effect of calling xas_load(). Otherwise one will be
+ * added to the index and the state will be walked to the correct
+ * location in the array for the next operation.
+ *
+ * If the iterator was referencing index %ULONG_MAX, this function wraps
+ * around to 0.
+ *
+ * Return: The entry at the new index. This may be %NULL or an internal
+ * entry, although it should never be a node entry.
+ */
+static inline void *xas_next(struct xa_state *xas)
+{
+ struct xa_node *node = xas->xa_node;
+
+ if (unlikely(xas_not_node(node) || node->shift ||
+ xas->xa_offset == XA_CHUNK_MASK))
+ return __xas_next(xas);
+
+ xas->xa_index++;
+ xas->xa_offset++;
+ return xa_entry(xas->xa, node, xas->xa_offset);
+}
+
/* Internal functions, mostly shared between radix-tree.c, xarray.c and idr.c */
void xas_destroy(struct xa_state *);
diff --git a/lib/xarray.c b/lib/xarray.c
index 080ed0fc3feb..7cf195b6e740 100644
--- a/lib/xarray.c
+++ b/lib/xarray.c
@@ -838,6 +838,80 @@ void xas_pause(struct xa_state *xas)
}
EXPORT_SYMBOL_GPL(xas_pause);
+/*
+ * __xas_prev() - Find the previous entry in the XArray.
+ * @xas: XArray operation state.
+ *
+ * Helper function for xas_prev() which handles all the complex cases
+ * out of line.
+ */
+void *__xas_prev(struct xa_state *xas)
+{
+ void *entry;
+
+ if (!xas_frozen(xas->xa_node))
+ xas->xa_index--;
+ if (xas_not_node(xas->xa_node))
+ return xas_load(xas);
+
+ if (xas->xa_offset != get_offset(xas->xa_index, xas->xa_node))
+ xas->xa_offset--;
+
+ while (xas->xa_offset == 255) {
+ xas->xa_offset = xas->xa_node->offset - 1;
+ xas->xa_node = xa_parent(xas->xa, xas->xa_node);
+ if (!xas->xa_node)
+ return set_bounds(xas);
+ }
+
+ for (;;) {
+ entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset);
+ if (!xa_is_node(entry))
+ return entry;
+
+ xas->xa_node = xa_to_node(entry);
+ xas_set_offset(xas);
+ }
+}
+EXPORT_SYMBOL_GPL(__xas_prev);
+
+/*
+ * __xas_next() - Find the next entry in the XArray.
+ * @xas: XArray operation state.
+ *
+ * Helper function for xas_next() which handles all the complex cases
+ * out of line.
+ */
+void *__xas_next(struct xa_state *xas)
+{
+ void *entry;
+
+ if (!xas_frozen(xas->xa_node))
+ xas->xa_index++;
+ if (xas_not_node(xas->xa_node))
+ return xas_load(xas);
+
+ if (xas->xa_offset != get_offset(xas->xa_index, xas->xa_node))
+ xas->xa_offset++;
+
+ while (xas->xa_offset == XA_CHUNK_SIZE) {
+ xas->xa_offset = xas->xa_node->offset + 1;
+ xas->xa_node = xa_parent(xas->xa, xas->xa_node);
+ if (!xas->xa_node)
+ return set_bounds(xas);
+ }
+
+ for (;;) {
+ entry = xa_entry(xas->xa, xas->xa_node, xas->xa_offset);
+ if (!xa_is_node(entry))
+ return entry;
+
+ xas->xa_node = xa_to_node(entry);
+ xas_set_offset(xas);
+ }
+}
+EXPORT_SYMBOL_GPL(__xas_next);
+
/**
* xas_find() - Find the next present entry in the XArray.
* @xas: XArray operation state.
diff --git a/tools/testing/radix-tree/xarray-test.c b/tools/testing/radix-tree/xarray-test.c
index 26b25be81656..ce20c22ad55d 100644
--- a/tools/testing/radix-tree/xarray-test.c
+++ b/tools/testing/radix-tree/xarray-test.c
@@ -49,6 +49,147 @@ void check_xa_tag(struct xarray *xa)
assert(xa_get_tag(xa, 0, XA_TAG_0) == false);
}
+/* Check that putting the xas into an error state works correctly */
+void check_xas_error(struct xarray *xa)
+{
+ XA_STATE(xas, xa, 0);
+
+ assert(xa_store(xa, 1, xa_mk_value(1), GFP_KERNEL) == 0);
+ assert(xa_load(xa, 1) == xa_mk_value(1));
+
+ assert(xas_error(&xas) == 0);
+
+ xas_set_err(&xas, -ENOTTY);
+ assert(xas_error(&xas) == -ENOTTY);
+
+ xas_set_err(&xas, -ENOSPC);
+ assert(xas_error(&xas) == -ENOSPC);
+
+ xas_set_err(&xas, -ENOMEM);
+ assert(xas_error(&xas) == -ENOMEM);
+
+ assert(xas_load(&xas) == NULL);
+ assert(xas_store(&xas, &xas) == NULL);
+ assert(xas_load(&xas) == NULL);
+
+ assert(xas.xa_index == 0);
+ assert(xas_next(&xas) == NULL);
+ assert(xas.xa_index == 0);
+
+ assert(xas_prev(&xas) == NULL);
+ assert(xas.xa_index == 0);
+
+ xas_retry(&xas, XA_RETRY_ENTRY);
+ assert(xas_error(&xas) == 0);
+
+ assert(xas_find(&xas, ULONG_MAX) == xa_mk_value(1));
+ assert(xas.xa_index == 1);
+ assert(xas_error(&xas) == 0);
+
+ assert(xas_find(&xas, ULONG_MAX) == NULL);
+ assert(xas.xa_index > 1);
+ assert(xas_error(&xas) == 0);
+ assert(xas.xa_node == XAS_BOUNDS);
+}
+
+void check_xas_pause(struct xarray *xa)
+{
+ XA_STATE(xas, xa, 0);
+ void *entry;
+ unsigned int seen;
+
+ xa_store(xa, 0, xa_mk_value(0), GFP_KERNEL);
+ xa_set_tag(xa, 0, XA_TAG_0);
+
+ seen = 0;
+ rcu_read_lock();
+ xas_for_each_tag(&xas, entry, ULONG_MAX, XA_TAG_0) {
+ if (!seen++) {
+ xa_store(xa, 1, xa_mk_value(1), GFP_KERNEL);
+ xa_set_tag(xa, 1, XA_TAG_0);
+ }
+ }
+ rcu_read_unlock();
+ /* We don't see an entry that was added after we started */
+ assert(seen == 1);
+
+ seen = 0;
+ xas_set(&xas, 0);
+ rcu_read_lock();
+ xas_for_each_tag(&xas, entry, ULONG_MAX, XA_TAG_0) {
+ if (!seen++)
+ xa_erase(xa, 1);
+ }
+ rcu_read_unlock();
+ assert(seen == 1);
+
+ seen = 0;
+ xas_set(&xas, 0);
+ rcu_read_lock();
+ xas_for_each(&xas, entry, ULONG_MAX) {
+ if (!seen++)
+ xa_store(xa, 1, xa_mk_value(1), GFP_KERNEL);
+ }
+ rcu_read_unlock();
+ assert(seen == 1);
+
+ seen = 0;
+ xas_set(&xas, 0);
+ rcu_read_lock();
+ xas_for_each(&xas, entry, ULONG_MAX) {
+ if (!seen++)
+ xa_erase(xa, 1);
+ }
+ rcu_read_unlock();
+ assert(seen == 1);
+
+ seen = 0;
+ xas_set(&xas, 0);
+ rcu_read_lock();
+ for (entry = xas_load(&xas); entry; entry = xas_next(&xas)) {
+ if (!seen++)
+ xa_store(xa, 1, xa_mk_value(1), GFP_KERNEL);
+ }
+ rcu_read_unlock();
+ assert(seen == 2);
+
+ seen = 0;
+ xas_set(&xas, 0);
+ rcu_read_lock();
+ for (entry = xas_load(&xas); entry; entry = xas_next(&xas)) {
+ if (!seen++)
+ xa_erase(xa, 1);
+ }
+ rcu_read_unlock();
+ assert(seen == 1);
+
+ xa_store(xa, 1, xa_mk_value(1), GFP_KERNEL);
+ seen = 0;
+ xas_set(&xas, 0);
+ xas_for_each(&xas, entry, ULONG_MAX) {
+ if (!seen++)
+ xas_pause(&xas);
+ }
+ assert(seen == 2);
+
+ seen = 0;
+ xas_set(&xas, 0);
+ for (entry = xas_load(&xas); entry; entry = xas_next(&xas)) {
+ if (!seen++)
+ xas_pause(&xas);
+ }
+ assert(seen == 2);
+
+ seen = 0;
+ xas_set(&xas, 0);
+ xa_set_tag(xa, 1, XA_TAG_0);
+ xas_for_each_tag(&xas, entry, ULONG_MAX, XA_TAG_0) {
+ if (!seen++)
+ xas_pause(&xas);
+ }
+ assert(seen == 2);
+}
+
void check_xas_retry(struct xarray *xa)
{
XA_STATE(xas, xa, 0);
@@ -61,7 +202,7 @@ void check_xas_retry(struct xarray *xa)
assert(xa_is_retry(xas_reload(&xas)));
assert(!xas_retry(&xas, NULL));
assert(!xas_retry(&xas, xa_mk_value(0)));
- assert(xas_retry(&xas, XA_RETRY_ENTRY));
+ xas_reset(&xas);
assert(xas.xa_node == XAS_RESTART);
assert(xas_next_entry(&xas, ULONG_MAX) == xa_mk_value(0));
assert(xas.xa_node == NULL);
@@ -257,9 +398,108 @@ void check_xas_delete(struct xarray *xa)
}
}
+void check_move_small(struct xarray *xa, unsigned long idx)
+{
+ XA_STATE(xas, xa, 0);
+ unsigned long i;
+
+ xa_store(xa, 0, xa_mk_value(0), GFP_KERNEL);
+ xa_store(xa, idx, xa_mk_value(idx), GFP_KERNEL);
+
+ for (i = 0; i < idx * 4; i++) {
+ void *entry = xas_next(&xas);
+ if (i <= idx)
+ assert(xas.xa_node != XAS_RESTART);
+ assert(xas.xa_index == i);
+ if (i == 0 || i == idx)
+ assert(entry == xa_mk_value(i));
+ else
+ assert(entry == NULL);
+ }
+ xas_next(&xas);
+ assert(xas.xa_index == i);
+
+ do {
+ void *entry = xas_prev(&xas);
+ i--;
+ if (i <= idx)
+ assert(xas.xa_node != XAS_RESTART);
+ assert(xas.xa_index == i);
+ if (i == 0 || i == idx)
+ assert(entry == xa_mk_value(i));
+ else
+ assert(entry == NULL);
+ } while (i > 0);
+
+ xas_set(&xas, ULONG_MAX);
+ assert(xas_next(&xas) == NULL);
+ assert(xas.xa_index == ULONG_MAX);
+ assert(xas_next(&xas) == xa_mk_value(0));
+ assert(xas.xa_index == 0);
+ assert(xas_prev(&xas) == NULL);
+ assert(xas.xa_index == ULONG_MAX);
+}
+
+void check_move(struct xarray *xa)
+{
+ XA_STATE(xas, xa, (1 << 16) - 1);
+ unsigned long i;
+
+ for (i = 0; i < (1 << 16); i++) {
+ xa_store(xa, i, xa_mk_value(i), GFP_KERNEL);
+ }
+
+ do {
+ void *entry = xas_prev(&xas);
+ i--;
+ assert(entry == xa_mk_value(i));
+ assert(i == xas.xa_index);
+ } while (i != 0);
+
+ assert(xas_prev(&xas) == NULL);
+ assert(xas.xa_index == ULONG_MAX);
+
+ do {
+ void *entry = xas_next(&xas);
+ assert(entry == xa_mk_value(i));
+ assert(i == xas.xa_index);
+ i++;
+ } while (i < (1 << 16));
+
+ for (i = (1 << 8); i < (1 << 15); i++) {
+ xa_erase(xa, i);
+ }
+
+ i = xas.xa_index;
+
+ do {
+ void *entry = xas_prev(&xas);
+ i--;
+ if ((i < (1 << 8)) || (i >= (1 << 15)))
+ assert(entry == xa_mk_value(i));
+ else
+ assert(entry == NULL);
+ assert(i == xas.xa_index);
+ } while (i != 0);
+
+ assert(xas_prev(&xas) == NULL);
+ assert(xas.xa_index == ULONG_MAX);
+
+ do {
+ void *entry = xas_next(&xas);
+ if ((i < (1 << 8)) || (i >= (1 << 15)))
+ assert(entry == xa_mk_value(i));
+ else
+ assert(entry == NULL);
+ assert(i == xas.xa_index);
+ i++;
+ } while (i < (1 << 16));
+}
+
void xarray_checks(void)
{
DEFINE_XARRAY(array);
+ unsigned long i;
check_xa_err(&array);
item_kill_tree(&array);
@@ -267,9 +507,15 @@ void xarray_checks(void)
check_xa_tag(&array);
item_kill_tree(&array);
+ check_xas_error(&array);
+ item_kill_tree(&array);
+
check_xas_retry(&array);
item_kill_tree(&array);
+ check_xas_pause(&array);
+ item_kill_tree(&array);
+
check_xa_load(&array);
item_kill_tree(&array);
@@ -283,6 +529,19 @@ void xarray_checks(void)
check_find(&array);
check_xas_delete(&array);
item_kill_tree(&array);
+
+ for (i = 0; i < 16; i++) {
+ check_move_small(&array, 1UL << i);
+ item_kill_tree(&array);
+ }
+
+ for (i = 2; i < 16; i++) {
+ check_move_small(&array, (1UL << i) - 1);
+ item_kill_tree(&array);
+ }
+
+ check_move(&array);
+ item_kill_tree(&array);
}
int __weak main(void)
--
2.16.1
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2018-02-19 19:46 UTC|newest]
Thread overview: 72+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-02-19 19:44 [PATCH v7 00/61] XArray version 7 -- for merging Matthew Wilcox
2018-02-19 19:44 ` [PATCH v7 01/61] radix tree test suite: Check reclaim bit Matthew Wilcox
2018-02-19 19:44 ` [PATCH v7 02/61] radix tree: Use bottom four bits of gfp_t for flags Matthew Wilcox
2018-03-03 12:44 ` Jeff Layton
2018-03-03 17:09 ` Matthew Wilcox
2018-02-19 19:44 ` [PATCH v7 03/61] arm64: Turn flush_dcache_mmap_lock into a no-op Matthew Wilcox
2018-02-19 19:44 ` [PATCH v7 04/61] unicore32: " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 05/61] Export __set_page_dirty Matthew Wilcox
2018-03-03 13:56 ` Jeff Layton
2018-02-19 19:45 ` [PATCH v7 06/61] xfs: Rename xa_ elements to ail_ Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 07/61] fscache: Use appropriate radix tree accessors Matthew Wilcox
2018-03-03 13:57 ` Jeff Layton
2018-02-19 19:45 ` [PATCH v7 08/61] xarray: Add the xa_lock to the radix_tree_root Matthew Wilcox
2018-03-03 14:19 ` Jeff Layton
2018-03-03 14:55 ` Jeff Layton
2018-03-03 16:57 ` Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 09/61] page cache: Use xa_lock Matthew Wilcox
2018-03-03 14:27 ` Jeff Layton
2018-02-19 19:45 ` [PATCH v7 10/61] xarray: Replace exceptional entries Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 11/61] xarray: Change definition of sibling entries Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 12/61] xarray: Add definition of struct xarray Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 13/61] xarray: Define struct xa_node Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 14/61] xarray: Add documentation Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 15/61] xarray: Add xa_load Matthew Wilcox
2018-02-20 7:34 ` Philippe Ombredanne
2018-02-20 12:47 ` Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 16/61] xarray: Add xa_get_tag, xa_set_tag and xa_clear_tag Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 17/61] xarray: Add xa_store Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 18/61] xarray: Add xa_cmpxchg and xa_insert Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 19/61] xarray: Add xa_for_each Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 20/61] xarray: Add xa_extract Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 21/61] xarray: Add xa_destroy Matthew Wilcox
2018-02-19 19:45 ` Matthew Wilcox [this message]
2018-02-19 19:45 ` [PATCH v7 23/61] xarray: Add xas_create_range Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 24/61] xarray: Add MAINTAINERS entry Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 25/61] page cache: Convert hole search to XArray Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 26/61] page cache: Add page_cache_range_empty function Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 27/61] page cache: Add and replace pages using the XArray Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 28/61] page cache: Convert page deletion to XArray Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 29/61] page cache: Convert page cache lookups " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 30/61] page cache: Convert delete_batch " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 31/61] page cache: Remove stray radix comment Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 32/61] page cache: Convert filemap_range_has_page to XArray Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 33/61] mm: Convert page-writeback " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 34/61] mm: Convert workingset " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 35/61] mm: Convert truncate " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 36/61] mm: Convert add_to_swap_cache " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 37/61] mm: Convert delete_from_swap_cache " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 38/61] mm: Convert __do_page_cache_readahead " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 39/61] mm: Convert page migration " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 40/61] mm: Convert huge_memory " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 41/61] mm: Convert collapse_shmem " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 42/61] mm: Convert khugepaged_scan_shmem " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 43/61] pagevec: Use xa_tag_t Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 44/61] shmem: Convert replace to XArray Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 45/61] shmem: Convert shmem_confirm_swap " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 46/61] shmem: Convert find_swap_entry " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 47/61] shmem: Convert shmem_tag_pins " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 48/61] shmem: Convert shmem_wait_for_pins " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 49/61] shmem: Convert shmem_add_to_page_cache " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 50/61] shmem: Convert shmem_alloc_hugepage " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 51/61] shmem: Convert shmem_free_swap " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 52/61] shmem: Convert shmem_partial_swap_usage " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 53/61] shmem: Comment fixups Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 54/61] btrfs: Convert page cache to XArray Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 55/61] fs: Convert buffer " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 56/61] fs: Convert writeback " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 57/61] nilfs2: Convert " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 58/61] f2fs: " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 59/61] lustre: " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 60/61] dax: " Matthew Wilcox
2018-02-19 19:45 ` [PATCH v7 61/61] page cache: Finish XArray conversion Matthew Wilcox
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20180219194556.6575-23-willy@infradead.org \
--to=willy@infradead.org \
--cc=akpm@linux-foundation.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mawilcox@microsoft.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).