From: Keith Busch <kbusch@meta.com>
To: <linux-block@vger.kernel.org>, <linux-nvme@lists.infradead.org>,
<hch@lst.de>
Cc: <axboe@kernel.dk>, <leonro@nvidia.com>, Keith Busch <kbusch@kernel.org>
Subject: [PATCHv3 1/7] blk-mq: introduce blk_map_iter
Date: Tue, 29 Jul 2025 07:34:36 -0700 [thread overview]
Message-ID: <20250729143442.2586575-2-kbusch@meta.com> (raw)
In-Reply-To: <20250729143442.2586575-1-kbusch@meta.com>
From: Keith Busch <kbusch@kernel.org>
Create a type that fully captures the lower level physical address
iteration.
Signed-off-by: Keith Busch <kbusch@kernel.org>
---
block/blk-mq-dma.c | 93 +++++++++++++++++---------------------
include/linux/blk-mq-dma.h | 9 +++-
2 files changed, 50 insertions(+), 52 deletions(-)
diff --git a/block/blk-mq-dma.c b/block/blk-mq-dma.c
index ad283017caef2..61fbdb715220f 100644
--- a/block/blk-mq-dma.c
+++ b/block/blk-mq-dma.c
@@ -5,13 +5,7 @@
#include <linux/blk-mq-dma.h>
#include "blk.h"
-struct phys_vec {
- phys_addr_t paddr;
- u32 len;
-};
-
-static bool blk_map_iter_next(struct request *req, struct req_iterator *iter,
- struct phys_vec *vec)
+static bool blk_map_iter_next(struct request *req, struct blk_map_iter *iter)
{
unsigned int max_size;
struct bio_vec bv;
@@ -19,8 +13,8 @@ static bool blk_map_iter_next(struct request *req, struct req_iterator *iter,
if (req->rq_flags & RQF_SPECIAL_PAYLOAD) {
if (!iter->bio)
return false;
- vec->paddr = bvec_phys(&req->special_vec);
- vec->len = req->special_vec.bv_len;
+ iter->paddr = bvec_phys(&req->special_vec);
+ iter->len = req->special_vec.bv_len;
iter->bio = NULL;
return true;
}
@@ -29,8 +23,8 @@ static bool blk_map_iter_next(struct request *req, struct req_iterator *iter,
return false;
bv = mp_bvec_iter_bvec(iter->bio->bi_io_vec, iter->iter);
- vec->paddr = bvec_phys(&bv);
- max_size = get_max_segment_size(&req->q->limits, vec->paddr, UINT_MAX);
+ iter->paddr = bvec_phys(&bv);
+ max_size = get_max_segment_size(&req->q->limits, iter->paddr, UINT_MAX);
bv.bv_len = min(bv.bv_len, max_size);
bio_advance_iter_single(iter->bio, &iter->iter, bv.bv_len);
@@ -39,12 +33,11 @@ static bool blk_map_iter_next(struct request *req, struct req_iterator *iter,
* one could be merged into it. This typically happens when moving to
* the next bio, but some callers also don't pack bvecs tight.
*/
- while (!iter->iter.bi_size || !iter->iter.bi_bvec_done) {
+ while (!iter->iter.bi_size ||
+ (!iter->iter.bi_bvec_done && iter->bio->bi_next)) {
struct bio_vec next;
if (!iter->iter.bi_size) {
- if (!iter->bio->bi_next)
- break;
iter->bio = iter->bio->bi_next;
iter->iter = iter->bio->bi_iter;
}
@@ -58,7 +51,7 @@ static bool blk_map_iter_next(struct request *req, struct req_iterator *iter,
bio_advance_iter_single(iter->bio, &iter->iter, next.bv_len);
}
- vec->len = bv.bv_len;
+ iter->len = bv.bv_len;
return true;
}
@@ -77,29 +70,29 @@ static inline bool blk_can_dma_map_iova(struct request *req,
dma_get_merge_boundary(dma_dev));
}
-static bool blk_dma_map_bus(struct blk_dma_iter *iter, struct phys_vec *vec)
+static bool blk_dma_map_bus(struct blk_dma_iter *iter)
{
- iter->addr = pci_p2pdma_bus_addr_map(&iter->p2pdma, vec->paddr);
- iter->len = vec->len;
+ iter->addr = pci_p2pdma_bus_addr_map(&iter->p2pdma, iter->iter.paddr);
+ iter->len = iter->iter.len;
return true;
}
static bool blk_dma_map_direct(struct request *req, struct device *dma_dev,
- struct blk_dma_iter *iter, struct phys_vec *vec)
+ struct blk_dma_iter *iter)
{
- iter->addr = dma_map_page(dma_dev, phys_to_page(vec->paddr),
- offset_in_page(vec->paddr), vec->len, rq_dma_dir(req));
- if (dma_mapping_error(dma_dev, iter->addr)) {
+ iter->addr = dma_map_page(dma_dev, phys_to_page(iter->iter.paddr),
+ offset_in_page(iter->iter.paddr), iter->iter.len,
+ rq_dma_dir(req));
+ if (dma_mapping_error(dma_dev, iter->iter.paddr)) {
iter->status = BLK_STS_RESOURCE;
return false;
}
- iter->len = vec->len;
+ iter->len = iter->iter.len;
return true;
}
static bool blk_rq_dma_map_iova(struct request *req, struct device *dma_dev,
- struct dma_iova_state *state, struct blk_dma_iter *iter,
- struct phys_vec *vec)
+ struct dma_iova_state *state, struct blk_dma_iter *iter)
{
enum dma_data_direction dir = rq_dma_dir(req);
unsigned int mapped = 0;
@@ -109,12 +102,12 @@ static bool blk_rq_dma_map_iova(struct request *req, struct device *dma_dev,
iter->len = dma_iova_size(state);
do {
- error = dma_iova_link(dma_dev, state, vec->paddr, mapped,
- vec->len, dir, 0);
+ error = dma_iova_link(dma_dev, state, iter->iter.paddr, mapped,
+ iter->iter.len, dir, 0);
if (error)
break;
- mapped += vec->len;
- } while (blk_map_iter_next(req, &iter->iter, vec));
+ mapped += iter->iter.len;
+ } while (blk_map_iter_next(req, &iter->iter));
error = dma_iova_sync(dma_dev, state, 0, mapped);
if (error) {
@@ -151,10 +144,10 @@ bool blk_rq_dma_map_iter_start(struct request *req, struct device *dma_dev,
struct dma_iova_state *state, struct blk_dma_iter *iter)
{
unsigned int total_len = blk_rq_payload_bytes(req);
- struct phys_vec vec;
+ struct blk_map_iter *map_iter = &iter->iter;
- iter->iter.bio = req->bio;
- iter->iter.iter = req->bio->bi_iter;
+ map_iter->bio = req->bio;
+ map_iter->iter = req->bio->bi_iter;
memset(&iter->p2pdma, 0, sizeof(iter->p2pdma));
iter->status = BLK_STS_OK;
@@ -162,14 +155,14 @@ bool blk_rq_dma_map_iter_start(struct request *req, struct device *dma_dev,
* Grab the first segment ASAP because we'll need it to check for P2P
* transfers.
*/
- if (!blk_map_iter_next(req, &iter->iter, &vec))
+ if (!blk_map_iter_next(req, map_iter))
return false;
if (IS_ENABLED(CONFIG_PCI_P2PDMA) && (req->cmd_flags & REQ_P2PDMA)) {
switch (pci_p2pdma_state(&iter->p2pdma, dma_dev,
- phys_to_page(vec.paddr))) {
+ phys_to_page(map_iter->paddr))) {
case PCI_P2PDMA_MAP_BUS_ADDR:
- return blk_dma_map_bus(iter, &vec);
+ return blk_dma_map_bus(iter);
case PCI_P2PDMA_MAP_THRU_HOST_BRIDGE:
/*
* P2P transfers through the host bridge are treated the
@@ -184,9 +177,9 @@ bool blk_rq_dma_map_iter_start(struct request *req, struct device *dma_dev,
}
if (blk_can_dma_map_iova(req, dma_dev) &&
- dma_iova_try_alloc(dma_dev, state, vec.paddr, total_len))
- return blk_rq_dma_map_iova(req, dma_dev, state, iter, &vec);
- return blk_dma_map_direct(req, dma_dev, iter, &vec);
+ dma_iova_try_alloc(dma_dev, state, map_iter->paddr, total_len))
+ return blk_rq_dma_map_iova(req, dma_dev, state, iter);
+ return blk_dma_map_direct(req, dma_dev, iter);
}
EXPORT_SYMBOL_GPL(blk_rq_dma_map_iter_start);
@@ -211,14 +204,12 @@ EXPORT_SYMBOL_GPL(blk_rq_dma_map_iter_start);
bool blk_rq_dma_map_iter_next(struct request *req, struct device *dma_dev,
struct dma_iova_state *state, struct blk_dma_iter *iter)
{
- struct phys_vec vec;
-
- if (!blk_map_iter_next(req, &iter->iter, &vec))
+ if (!blk_map_iter_next(req, &iter->iter))
return false;
if (iter->p2pdma.map == PCI_P2PDMA_MAP_BUS_ADDR)
- return blk_dma_map_bus(iter, &vec);
- return blk_dma_map_direct(req, dma_dev, iter, &vec);
+ return blk_dma_map_bus(iter);
+ return blk_dma_map_direct(req, dma_dev, iter);
}
EXPORT_SYMBOL_GPL(blk_rq_dma_map_iter_next);
@@ -246,20 +237,20 @@ blk_next_sg(struct scatterlist **sg, struct scatterlist *sglist)
int __blk_rq_map_sg(struct request *rq, struct scatterlist *sglist,
struct scatterlist **last_sg)
{
- struct req_iterator iter = {
- .bio = rq->bio,
+ struct bio *bio = rq->bio;
+ struct blk_map_iter iter = {
+ .bio = bio,
};
- struct phys_vec vec;
int nsegs = 0;
/* the internal flush request may not have bio attached */
- if (iter.bio)
- iter.iter = iter.bio->bi_iter;
+ if (bio)
+ iter.iter = bio->bi_iter;
- while (blk_map_iter_next(rq, &iter, &vec)) {
+ while (blk_map_iter_next(rq, &iter)) {
*last_sg = blk_next_sg(last_sg, sglist);
- sg_set_page(*last_sg, phys_to_page(vec.paddr), vec.len,
- offset_in_page(vec.paddr));
+ sg_set_page(*last_sg, phys_to_page(iter.paddr), iter.len,
+ offset_in_page(iter.paddr));
nsegs++;
}
diff --git a/include/linux/blk-mq-dma.h b/include/linux/blk-mq-dma.h
index c26a01aeae006..1e5988afdb978 100644
--- a/include/linux/blk-mq-dma.h
+++ b/include/linux/blk-mq-dma.h
@@ -5,6 +5,13 @@
#include <linux/blk-mq.h>
#include <linux/pci-p2pdma.h>
+struct blk_map_iter {
+ phys_addr_t paddr;
+ u32 len;
+ struct bvec_iter iter;
+ struct bio *bio;
+};
+
struct blk_dma_iter {
/* Output address range for this iteration */
dma_addr_t addr;
@@ -14,7 +21,7 @@ struct blk_dma_iter {
blk_status_t status;
/* Internal to blk_rq_dma_map_iter_* */
- struct req_iterator iter;
+ struct blk_map_iter iter;
struct pci_p2pdma_map_state p2pdma;
};
--
2.47.3
next prev parent reply other threads:[~2025-07-29 14:35 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-07-29 14:34 [PATCHv3 0/7] blk dma iter for integrity metadata Keith Busch
2025-07-29 14:34 ` Keith Busch [this message]
2025-07-30 6:52 ` [PATCHv3 1/7] blk-mq: introduce blk_map_iter Kanchan Joshi
2025-07-30 8:18 ` Kanchan Joshi
2025-07-30 15:18 ` Keith Busch
2025-07-31 5:05 ` Kanchan Joshi
2025-07-29 14:34 ` [PATCHv3 2/7] blk-mq-dma: provide the bio_vec list being iterated Keith Busch
2025-07-29 20:55 ` Keith Busch
2025-07-30 12:15 ` Kanchan Joshi
2025-07-30 14:18 ` Keith Busch
2025-07-29 14:34 ` [PATCHv3 3/7] blk-mq-dma: require unmap caller provide p2p map type Keith Busch
2025-07-29 14:34 ` [PATCHv3 4/7] blk-mq: remove REQ_P2PDMA flag Keith Busch
2025-07-29 14:34 ` [PATCHv3 5/7] blk-mq-dma: move common dma start code to a helper Keith Busch
2025-07-29 14:34 ` [PATCHv3 6/7] blk-mq-dma: add support for mapping integrity metadata Keith Busch
2025-07-29 14:34 ` [PATCHv3 7/7] nvme: convert metadata mapping to dma iter Keith Busch
2025-07-29 18:56 ` Keith Busch
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250729143442.2586575-2-kbusch@meta.com \
--to=kbusch@meta.com \
--cc=axboe@kernel.dk \
--cc=hch@lst.de \
--cc=kbusch@kernel.org \
--cc=leonro@nvidia.com \
--cc=linux-block@vger.kernel.org \
--cc=linux-nvme@lists.infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox