From: David Howells <dhowells@redhat.com>
To: Viacheslav Dubeyko <slava@dubeyko.com>,
Alex Markuze <amarkuze@redhat.com>
Cc: David Howells <dhowells@redhat.com>,
Ilya Dryomov <idryomov@gmail.com>,
Jeff Layton <jlayton@kernel.org>,
Dongsheng Yang <dongsheng.yang@easystack.cn>,
ceph-devel@vger.kernel.org, linux-fsdevel@vger.kernel.org,
linux-block@vger.kernel.org, linux-kernel@vger.kernel.org
Subject: [RFC PATCH 16/35] libceph: Convert req_page of ceph_osdc_call() to ceph_databuf
Date: Thu, 13 Mar 2025 23:33:08 +0000 [thread overview]
Message-ID: <20250313233341.1675324-17-dhowells@redhat.com> (raw)
In-Reply-To: <20250313233341.1675324-1-dhowells@redhat.com>
Convert the request data (req_page) of ceph_osdc_call() to ceph_databuf.
Signed-off-by: David Howells <dhowells@redhat.com>
cc: Viacheslav Dubeyko <slava@dubeyko.com>
cc: Alex Markuze <amarkuze@redhat.com>
cc: Ilya Dryomov <idryomov@gmail.com>
cc: ceph-devel@vger.kernel.org
cc: linux-fsdevel@vger.kernel.org
---
drivers/block/rbd.c | 53 +++++++++++-----------
include/linux/ceph/osd_client.h | 2 +-
net/ceph/cls_lock_client.c | 78 ++++++++++++++++++---------------
net/ceph/osd_client.c | 25 ++---------
4 files changed, 74 insertions(+), 84 deletions(-)
diff --git a/drivers/block/rbd.c b/drivers/block/rbd.c
index dd22cea7ae89..ec09d578b0b0 100644
--- a/drivers/block/rbd.c
+++ b/drivers/block/rbd.c
@@ -1789,7 +1789,7 @@ static int __rbd_object_map_load(struct rbd_device *rbd_dev)
rbd_object_map_name(rbd_dev, rbd_dev->spec->snap_id, &oid);
ret = ceph_osdc_call(osdc, &oid, &rbd_dev->header_oloc,
"rbd", "object_map_load", CEPH_OSD_FLAG_READ,
- NULL, 0, reply);
+ NULL, reply);
if (ret)
goto out;
@@ -4553,8 +4553,8 @@ static int rbd_obj_method_sync(struct rbd_device *rbd_dev,
size_t inbound_size)
{
struct ceph_osd_client *osdc = &rbd_dev->rbd_client->client->osdc;
- struct ceph_databuf *reply;
- struct page *req_page = NULL;
+ struct ceph_databuf *request = NULL, *reply;
+ void *p;
int ret;
/*
@@ -4568,32 +4568,33 @@ static int rbd_obj_method_sync(struct rbd_device *rbd_dev,
if (outbound_size > PAGE_SIZE)
return -E2BIG;
- req_page = alloc_page(GFP_KERNEL);
- if (!req_page)
+ request = ceph_databuf_req_alloc(0, outbound_size, GFP_KERNEL);
+ if (!request)
return -ENOMEM;
- memcpy(page_address(req_page), outbound, outbound_size);
+ p = kmap_ceph_databuf_page(request, 0);
+ memcpy(p, outbound, outbound_size);
+ kunmap_local(p);
+ ceph_databuf_added_data(request, outbound_size);
}
reply = ceph_databuf_reply_alloc(1, inbound_size, GFP_KERNEL);
if (!reply) {
- if (req_page)
- __free_page(req_page);
- return -ENOMEM;
+ ret = -ENOMEM;
+ goto out;
}
ret = ceph_osdc_call(osdc, oid, oloc, RBD_DRV_NAME, method_name,
- CEPH_OSD_FLAG_READ, req_page, outbound_size,
- reply);
+ CEPH_OSD_FLAG_READ, request, reply);
if (!ret) {
ret = ceph_databuf_len(reply);
if (copy_from_iter(inbound, ret, &reply->iter) != ret)
ret = -EIO;
}
- if (req_page)
- __free_page(req_page);
ceph_databuf_release(reply);
+out:
+ ceph_databuf_release(request);
return ret;
}
@@ -5513,7 +5514,7 @@ static int decode_parent_image_spec(void **p, void *end,
}
static int __get_parent_info(struct rbd_device *rbd_dev,
- struct page *req_page,
+ struct ceph_databuf *request,
struct ceph_databuf *reply,
struct parent_image_info *pii)
{
@@ -5524,7 +5525,7 @@ static int __get_parent_info(struct rbd_device *rbd_dev,
ret = ceph_osdc_call(osdc, &rbd_dev->header_oid, &rbd_dev->header_oloc,
"rbd", "parent_get", CEPH_OSD_FLAG_READ,
- req_page, sizeof(u64), reply);
+ request, reply);
if (ret)
return ret == -EOPNOTSUPP ? 1 : ret;
@@ -5539,7 +5540,7 @@ static int __get_parent_info(struct rbd_device *rbd_dev,
ret = ceph_osdc_call(osdc, &rbd_dev->header_oid, &rbd_dev->header_oloc,
"rbd", "parent_overlap_get", CEPH_OSD_FLAG_READ,
- req_page, sizeof(u64), reply);
+ request, reply);
if (ret)
return ret;
@@ -5563,7 +5564,7 @@ static int __get_parent_info(struct rbd_device *rbd_dev,
* The caller is responsible for @pii.
*/
static int __get_parent_info_legacy(struct rbd_device *rbd_dev,
- struct page *req_page,
+ struct ceph_databuf *request,
struct ceph_databuf *reply,
struct parent_image_info *pii)
{
@@ -5573,7 +5574,7 @@ static int __get_parent_info_legacy(struct rbd_device *rbd_dev,
ret = ceph_osdc_call(osdc, &rbd_dev->header_oid, &rbd_dev->header_oloc,
"rbd", "get_parent", CEPH_OSD_FLAG_READ,
- req_page, sizeof(u64), reply);
+ request, reply);
if (ret)
return ret;
@@ -5604,29 +5605,29 @@ static int __get_parent_info_legacy(struct rbd_device *rbd_dev,
static int rbd_dev_v2_parent_info(struct rbd_device *rbd_dev,
struct parent_image_info *pii)
{
- struct ceph_databuf *reply;
- struct page *req_page;
+ struct ceph_databuf *request, *reply;
void *p;
int ret = -ENOMEM;
- req_page = alloc_page(GFP_KERNEL);
- if (!req_page)
+ request = ceph_databuf_req_alloc(0, sizeof(__le64), GFP_KERNEL);
+ if (!request)
goto out;
reply = ceph_databuf_reply_alloc(1, PAGE_SIZE, GFP_KERNEL);
if (!reply)
goto out_free;
- p = kmap_local_page(req_page);
+ p = kmap_ceph_databuf_page(request, 0);
ceph_encode_64(&p, rbd_dev->spec->snap_id);
kunmap_local(p);
- ret = __get_parent_info(rbd_dev, req_page, reply, pii);
+ ceph_databuf_added_data(request, sizeof(__le64));
+ ret = __get_parent_info(rbd_dev, request, reply, pii);
if (ret > 0)
- ret = __get_parent_info_legacy(rbd_dev, req_page, reply, pii);
+ ret = __get_parent_info_legacy(rbd_dev, request, reply, pii);
ceph_databuf_release(reply);
out_free:
- __free_page(req_page);
+ ceph_databuf_release(request);
out:
return ret;
}
diff --git a/include/linux/ceph/osd_client.h b/include/linux/ceph/osd_client.h
index 9182aa5075b2..d31e59bd128c 100644
--- a/include/linux/ceph/osd_client.h
+++ b/include/linux/ceph/osd_client.h
@@ -568,7 +568,7 @@ int ceph_osdc_call(struct ceph_osd_client *osdc,
struct ceph_object_locator *oloc,
const char *class, const char *method,
unsigned int flags,
- struct page *req_page, size_t req_len,
+ struct ceph_databuf *request,
struct ceph_databuf *response);
/* watch/notify */
diff --git a/net/ceph/cls_lock_client.c b/net/ceph/cls_lock_client.c
index 37bb8708e8bb..6c8608aabe5f 100644
--- a/net/ceph/cls_lock_client.c
+++ b/net/ceph/cls_lock_client.c
@@ -34,7 +34,7 @@ int ceph_cls_lock(struct ceph_osd_client *osdc,
int tag_len = strlen(tag);
int desc_len = strlen(desc);
void *p, *end;
- struct page *lock_op_page;
+ struct ceph_databuf *lock_op_req;
struct timespec64 mtime;
int ret;
@@ -49,11 +49,11 @@ int ceph_cls_lock(struct ceph_osd_client *osdc,
if (lock_op_buf_size > PAGE_SIZE)
return -E2BIG;
- lock_op_page = alloc_page(GFP_NOIO);
- if (!lock_op_page)
+ lock_op_req = ceph_databuf_req_alloc(0, lock_op_buf_size, GFP_NOIO);
+ if (!lock_op_req)
return -ENOMEM;
- p = page_address(lock_op_page);
+ p = kmap_ceph_databuf_page(lock_op_req, 0);
end = p + lock_op_buf_size;
/* encode cls_lock_lock_op struct */
@@ -69,15 +69,16 @@ int ceph_cls_lock(struct ceph_osd_client *osdc,
ceph_encode_timespec64(p, &mtime);
p += sizeof(struct ceph_timespec);
ceph_encode_8(&p, flags);
+ kunmap_local(p);
+ ceph_databuf_added_data(lock_op_req, lock_op_buf_size);
dout("%s lock_name %s type %d cookie %s tag %s desc %s flags 0x%x\n",
__func__, lock_name, type, cookie, tag, desc, flags);
ret = ceph_osdc_call(osdc, oid, oloc, "lock", "lock",
- CEPH_OSD_FLAG_WRITE, lock_op_page,
- lock_op_buf_size, NULL);
+ CEPH_OSD_FLAG_WRITE, lock_op_req, NULL);
dout("%s: status %d\n", __func__, ret);
- __free_page(lock_op_page);
+ ceph_databuf_release(lock_op_req);
return ret;
}
EXPORT_SYMBOL(ceph_cls_lock);
@@ -99,7 +100,7 @@ int ceph_cls_unlock(struct ceph_osd_client *osdc,
int name_len = strlen(lock_name);
int cookie_len = strlen(cookie);
void *p, *end;
- struct page *unlock_op_page;
+ struct ceph_databuf *unlock_op_req;
int ret;
unlock_op_buf_size = name_len + sizeof(__le32) +
@@ -108,11 +109,11 @@ int ceph_cls_unlock(struct ceph_osd_client *osdc,
if (unlock_op_buf_size > PAGE_SIZE)
return -E2BIG;
- unlock_op_page = alloc_page(GFP_NOIO);
- if (!unlock_op_page)
+ unlock_op_req = ceph_databuf_req_alloc(0, unlock_op_buf_size, GFP_NOIO);
+ if (!unlock_op_req)
return -ENOMEM;
- p = page_address(unlock_op_page);
+ p = kmap_ceph_databuf_page(unlock_op_req, 0);
end = p + unlock_op_buf_size;
/* encode cls_lock_unlock_op struct */
@@ -120,14 +121,15 @@ int ceph_cls_unlock(struct ceph_osd_client *osdc,
unlock_op_buf_size - CEPH_ENCODING_START_BLK_LEN);
ceph_encode_string(&p, end, lock_name, name_len);
ceph_encode_string(&p, end, cookie, cookie_len);
+ kunmap_local(p);
+ ceph_databuf_added_data(unlock_op_req, unlock_op_buf_size);
dout("%s lock_name %s cookie %s\n", __func__, lock_name, cookie);
ret = ceph_osdc_call(osdc, oid, oloc, "lock", "unlock",
- CEPH_OSD_FLAG_WRITE, unlock_op_page,
- unlock_op_buf_size, NULL);
+ CEPH_OSD_FLAG_WRITE, unlock_op_req, NULL);
dout("%s: status %d\n", __func__, ret);
- __free_page(unlock_op_page);
+ ceph_databuf_release(unlock_op_req);
return ret;
}
EXPORT_SYMBOL(ceph_cls_unlock);
@@ -150,7 +152,7 @@ int ceph_cls_break_lock(struct ceph_osd_client *osdc,
int break_op_buf_size;
int name_len = strlen(lock_name);
int cookie_len = strlen(cookie);
- struct page *break_op_page;
+ struct ceph_databuf *break_op_req;
void *p, *end;
int ret;
@@ -161,11 +163,11 @@ int ceph_cls_break_lock(struct ceph_osd_client *osdc,
if (break_op_buf_size > PAGE_SIZE)
return -E2BIG;
- break_op_page = alloc_page(GFP_NOIO);
- if (!break_op_page)
+ break_op_req = ceph_databuf_req_alloc(0, break_op_buf_size, GFP_NOIO);
+ if (!break_op_req)
return -ENOMEM;
- p = page_address(break_op_page);
+ p = kmap_ceph_databuf_page(break_op_req, 0);
end = p + break_op_buf_size;
/* encode cls_lock_break_op struct */
@@ -174,15 +176,16 @@ int ceph_cls_break_lock(struct ceph_osd_client *osdc,
ceph_encode_string(&p, end, lock_name, name_len);
ceph_encode_copy(&p, locker, sizeof(*locker));
ceph_encode_string(&p, end, cookie, cookie_len);
+ kunmap_local(p);
+ ceph_databuf_added_data(break_op_req, break_op_buf_size);
dout("%s lock_name %s cookie %s locker %s%llu\n", __func__, lock_name,
cookie, ENTITY_NAME(*locker));
ret = ceph_osdc_call(osdc, oid, oloc, "lock", "break_lock",
- CEPH_OSD_FLAG_WRITE, break_op_page,
- break_op_buf_size, NULL);
+ CEPH_OSD_FLAG_WRITE, break_op_req, NULL);
dout("%s: status %d\n", __func__, ret);
- __free_page(break_op_page);
+ ceph_databuf_release(break_op_req);
return ret;
}
EXPORT_SYMBOL(ceph_cls_break_lock);
@@ -199,7 +202,7 @@ int ceph_cls_set_cookie(struct ceph_osd_client *osdc,
int tag_len = strlen(tag);
int new_cookie_len = strlen(new_cookie);
void *p, *end;
- struct page *cookie_op_page;
+ struct ceph_databuf *cookie_op_req;
int ret;
cookie_op_buf_size = name_len + sizeof(__le32) +
@@ -210,11 +213,11 @@ int ceph_cls_set_cookie(struct ceph_osd_client *osdc,
if (cookie_op_buf_size > PAGE_SIZE)
return -E2BIG;
- cookie_op_page = alloc_page(GFP_NOIO);
- if (!cookie_op_page)
+ cookie_op_req = ceph_databuf_req_alloc(0, cookie_op_buf_size, GFP_NOIO);
+ if (!cookie_op_req)
return -ENOMEM;
- p = page_address(cookie_op_page);
+ p = kmap_ceph_databuf_page(cookie_op_req, 0);
end = p + cookie_op_buf_size;
/* encode cls_lock_set_cookie_op struct */
@@ -225,15 +228,16 @@ int ceph_cls_set_cookie(struct ceph_osd_client *osdc,
ceph_encode_string(&p, end, old_cookie, old_cookie_len);
ceph_encode_string(&p, end, tag, tag_len);
ceph_encode_string(&p, end, new_cookie, new_cookie_len);
+ kunmap_local(p);
+ ceph_databuf_added_data(cookie_op_req, cookie_op_buf_size);
dout("%s lock_name %s type %d old_cookie %s tag %s new_cookie %s\n",
__func__, lock_name, type, old_cookie, tag, new_cookie);
ret = ceph_osdc_call(osdc, oid, oloc, "lock", "set_cookie",
- CEPH_OSD_FLAG_WRITE, cookie_op_page,
- cookie_op_buf_size, NULL);
+ CEPH_OSD_FLAG_WRITE, cookie_op_req, NULL);
dout("%s: status %d\n", __func__, ret);
- __free_page(cookie_op_page);
+ ceph_databuf_release(cookie_op_req);
return ret;
}
EXPORT_SYMBOL(ceph_cls_set_cookie);
@@ -340,7 +344,7 @@ int ceph_cls_lock_info(struct ceph_osd_client *osdc,
struct ceph_databuf *reply;
int get_info_op_buf_size;
int name_len = strlen(lock_name);
- struct page *get_info_op_page;
+ struct ceph_databuf *get_info_op_req;
void *p, *end;
int ret;
@@ -349,28 +353,30 @@ int ceph_cls_lock_info(struct ceph_osd_client *osdc,
if (get_info_op_buf_size > PAGE_SIZE)
return -E2BIG;
- get_info_op_page = alloc_page(GFP_NOIO);
- if (!get_info_op_page)
+ get_info_op_req = ceph_databuf_req_alloc(0, get_info_op_buf_size,
+ GFP_NOIO);
+ if (!get_info_op_req)
return -ENOMEM;
reply = ceph_databuf_reply_alloc(1, PAGE_SIZE, GFP_NOIO);
if (!reply) {
- __free_page(get_info_op_page);
+ ceph_databuf_release(get_info_op_req);
return -ENOMEM;
}
- p = page_address(get_info_op_page);
+ p = kmap_ceph_databuf_page(get_info_op_req, 0);
end = p + get_info_op_buf_size;
/* encode cls_lock_get_info_op struct */
ceph_start_encoding(&p, 1, 1,
get_info_op_buf_size - CEPH_ENCODING_START_BLK_LEN);
ceph_encode_string(&p, end, lock_name, name_len);
+ kunmap_local(p);
+ ceph_databuf_added_data(get_info_op_req, get_info_op_buf_size);
dout("%s lock_name %s\n", __func__, lock_name);
ret = ceph_osdc_call(osdc, oid, oloc, "lock", "get_info",
- CEPH_OSD_FLAG_READ, get_info_op_page,
- get_info_op_buf_size, reply);
+ CEPH_OSD_FLAG_READ, get_info_op_req, reply);
dout("%s: status %d\n", __func__, ret);
if (ret >= 0) {
@@ -381,8 +387,8 @@ int ceph_cls_lock_info(struct ceph_osd_client *osdc,
kunmap_local(p);
}
- __free_page(get_info_op_page);
ceph_databuf_release(reply);
+ ceph_databuf_release(get_info_op_req);
return ret;
}
EXPORT_SYMBOL(ceph_cls_lock_info);
diff --git a/net/ceph/osd_client.c b/net/ceph/osd_client.c
index 405ccf7e7a91..c4525feb8e26 100644
--- a/net/ceph/osd_client.c
+++ b/net/ceph/osd_client.c
@@ -264,7 +264,7 @@ void osd_req_op_cls_request_databuf(struct ceph_osd_request *osd_req,
BUG_ON(!ceph_databuf_len(dbuf));
osd_data = osd_req_op_data(osd_req, which, cls, request_data);
- ceph_osd_databuf_init(osd_data, dbuf);
+ ceph_osd_databuf_init(osd_data, ceph_databuf_get(dbuf));
osd_req->r_ops[which].cls.indata_len += ceph_databuf_len(dbuf);
osd_req->r_ops[which].indata_len += ceph_databuf_len(dbuf);
}
@@ -283,19 +283,6 @@ void osd_req_op_cls_request_data_pagelist(
}
EXPORT_SYMBOL(osd_req_op_cls_request_data_pagelist);
-static void osd_req_op_cls_request_data_pages(struct ceph_osd_request *osd_req,
- unsigned int which, struct page **pages, u64 length,
- u32 offset, bool pages_from_pool, bool own_pages)
-{
- struct ceph_osd_data *osd_data;
-
- osd_data = osd_req_op_data(osd_req, which, cls, request_data);
- ceph_osd_data_pages_init(osd_data, pages, length, offset,
- pages_from_pool, own_pages);
- osd_req->r_ops[which].cls.indata_len += length;
- osd_req->r_ops[which].indata_len += length;
-}
-
void osd_req_op_cls_response_databuf(struct ceph_osd_request *osd_req,
unsigned int which,
struct ceph_databuf *dbuf)
@@ -5089,15 +5076,12 @@ int ceph_osdc_call(struct ceph_osd_client *osdc,
struct ceph_object_locator *oloc,
const char *class, const char *method,
unsigned int flags,
- struct page *req_page, size_t req_len,
+ struct ceph_databuf *request,
struct ceph_databuf *response)
{
struct ceph_osd_request *req;
int ret;
- if (req_len > PAGE_SIZE)
- return -E2BIG;
-
req = ceph_osdc_alloc_request(osdc, NULL, 1, false, GFP_NOIO);
if (!req)
return -ENOMEM;
@@ -5110,9 +5094,8 @@ int ceph_osdc_call(struct ceph_osd_client *osdc,
if (ret)
goto out_put_req;
- if (req_page)
- osd_req_op_cls_request_data_pages(req, 0, &req_page, req_len,
- 0, false, false);
+ if (request)
+ osd_req_op_cls_request_databuf(req, 0, request);
if (response)
osd_req_op_cls_response_databuf(req, 0, response);
next prev parent reply other threads:[~2025-03-13 23:34 UTC|newest]
Thread overview: 72+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-03-13 23:32 [RFC PATCH 00/35] ceph, rbd, netfs: Make ceph fully use netfslib David Howells
2025-03-13 23:32 ` [RFC PATCH 01/35] ceph: Fix incorrect flush end position calculation David Howells
2025-03-13 23:32 ` [RFC PATCH 02/35] libceph: Rename alignment to offset David Howells
2025-03-14 19:04 ` Viacheslav Dubeyko
2025-03-14 20:01 ` David Howells
2025-03-13 23:32 ` [RFC PATCH 03/35] libceph: Add a new data container type, ceph_databuf David Howells
2025-03-14 20:06 ` Viacheslav Dubeyko
2025-03-17 11:27 ` David Howells
2025-03-13 23:32 ` [RFC PATCH 04/35] ceph: Convert ceph_mds_request::r_pagelist to a databuf David Howells
2025-03-14 22:27 ` slava
2025-03-17 11:52 ` David Howells
2025-03-20 20:34 ` Viacheslav Dubeyko
2025-03-20 22:01 ` David Howells
2025-03-13 23:32 ` [RFC PATCH 05/35] libceph: Add functions to add ceph_databufs to requests David Howells
2025-03-13 23:32 ` [RFC PATCH 06/35] rbd: Use ceph_databuf for rbd_obj_read_sync() David Howells
2025-03-17 19:08 ` Viacheslav Dubeyko
2025-04-11 13:48 ` David Howells
2025-03-13 23:32 ` [RFC PATCH 07/35] libceph: Change ceph_osdc_call()'s reply to a ceph_databuf David Howells
2025-03-17 19:41 ` Viacheslav Dubeyko
2025-03-17 22:12 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 08/35] libceph: Unexport osd_req_op_cls_request_data_pages() David Howells
2025-03-13 23:33 ` [RFC PATCH 09/35] libceph: Remove osd_req_op_cls_response_data_pages() David Howells
2025-03-13 23:33 ` [RFC PATCH 10/35] libceph: Convert notify_id_pages to a ceph_databuf David Howells
2025-03-13 23:33 ` [RFC PATCH 11/35] ceph: Use ceph_databuf in DIO David Howells
2025-03-17 20:03 ` Viacheslav Dubeyko
2025-03-17 22:26 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 12/35] libceph: Bypass the messenger-v1 Tx loop for databuf/iter data blobs David Howells
2025-03-13 23:33 ` [RFC PATCH 13/35] rbd: Switch from using bvec_iter to iov_iter David Howells
2025-03-18 19:38 ` Viacheslav Dubeyko
2025-03-18 22:13 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 14/35] libceph: Remove bvec and bio data container types David Howells
2025-03-13 23:33 ` [RFC PATCH 15/35] libceph: Make osd_req_op_cls_init() use a ceph_databuf and map it David Howells
2025-03-13 23:33 ` David Howells [this message]
2025-03-13 23:33 ` [RFC PATCH 17/35] libceph, rbd: Use ceph_databuf encoding start/stop David Howells
2025-03-18 19:59 ` Viacheslav Dubeyko
2025-03-18 22:19 ` David Howells
2025-03-20 21:45 ` Viacheslav Dubeyko
2025-03-13 23:33 ` [RFC PATCH 18/35] libceph, rbd: Convert some page arrays to ceph_databuf David Howells
2025-03-18 20:02 ` Viacheslav Dubeyko
2025-03-18 22:25 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 19/35] libceph, ceph: Convert users of ceph_pagelist " David Howells
2025-03-18 20:09 ` Viacheslav Dubeyko
2025-03-18 22:27 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 20/35] libceph: Remove ceph_pagelist David Howells
2025-03-13 23:33 ` [RFC PATCH 21/35] libceph: Make notify code use ceph_databuf_enc_start/stop David Howells
2025-03-18 20:12 ` Viacheslav Dubeyko
2025-03-18 22:36 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 22/35] libceph, rbd: Convert ceph_osdc_notify() reply to ceph_databuf David Howells
2025-03-19 0:08 ` Viacheslav Dubeyko
2025-03-20 14:44 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 23/35] rbd: Use ceph_databuf_enc_start/stop() David Howells
2025-03-19 0:32 ` Viacheslav Dubeyko
2025-03-20 14:59 ` Why use plain numbers and totals rather than predef'd constants for RPC sizes? David Howells
2025-03-20 21:48 ` Viacheslav Dubeyko
2025-03-13 23:33 ` [RFC PATCH 24/35] ceph: Make ceph_calc_file_object_mapping() return size as size_t David Howells
2025-03-13 23:33 ` [RFC PATCH 25/35] ceph: Wrap POSIX_FADV_WILLNEED to get caps David Howells
2025-03-13 23:33 ` [RFC PATCH 26/35] ceph: Kill ceph_rw_context David Howells
2025-03-13 23:33 ` [RFC PATCH 27/35] netfs: Pass extra write context to write functions David Howells
2025-03-13 23:33 ` [RFC PATCH 28/35] netfs: Adjust group handling David Howells
2025-03-19 18:57 ` Viacheslav Dubeyko
2025-03-20 15:22 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 29/35] netfs: Allow fs-private data to be handed through to request alloc David Howells
2025-03-13 23:33 ` [RFC PATCH 30/35] netfs: Make netfs_page_mkwrite() use folio_mkwrite_check_truncate() David Howells
2025-03-13 23:33 ` [RFC PATCH 31/35] netfs: Fix netfs_unbuffered_read() to return ssize_t rather than int David Howells
2025-03-13 23:33 ` [RFC PATCH 32/35] netfs: Add some more RMW support for ceph David Howells
2025-03-19 19:14 ` Viacheslav Dubeyko
2025-03-20 15:25 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 33/35] ceph: Use netfslib [INCOMPLETE] David Howells
2025-03-19 19:54 ` Viacheslav Dubeyko
2025-03-20 15:38 ` David Howells
2025-03-13 23:33 ` [RFC PATCH 34/35] ceph: Enable multipage folios for ceph files David Howells
2025-03-13 23:33 ` [RFC PATCH 35/35] ceph: Remove old I/O API bits David Howells
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250313233341.1675324-17-dhowells@redhat.com \
--to=dhowells@redhat.com \
--cc=amarkuze@redhat.com \
--cc=ceph-devel@vger.kernel.org \
--cc=dongsheng.yang@easystack.cn \
--cc=idryomov@gmail.com \
--cc=jlayton@kernel.org \
--cc=linux-block@vger.kernel.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=slava@dubeyko.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox