From: Zhu Yanjun <zyjzyj2000@gmail.com>
To: Boshi Yu <boshiyu@alibaba-inc.com>, jgg@ziepe.ca, leon@kernel.org
Cc: linux-rdma@vger.kernel.org, chengyou@linux.alibaba.com,
KaiShen@linux.alibaba.com
Subject: Re: [PATCH for-next 1/3] RDMA/erdma: Allocate doorbell records from dma pool
Date: Thu, 14 Mar 2024 11:23:40 +0100 [thread overview]
Message-ID: <2ee0f7b0-21fc-43d2-a126-850e73ff84fd@gmail.com> (raw)
In-Reply-To: <09da3e92-3da5-405a-9f55-60533e966f09@linux.dev>
On 13.03.24 17:20, Zhu Yanjun wrote:
> On 11.03.24 12:38, Boshi Yu wrote:
>> From: Boshi Yu <boshiyu@linux.alibaba.com>
>>
>> Currently, the 8 byte doorbell record is allocated along with the queue
>> buffer, which may result in waste of dma space when the queue buffer is
>> page aligned. To address this issue, we introduce a dma pool named
>> db_pool and allocate doorbell record from it.
>>
>> Reviewed-by: Cheng Xu <chengyou@linux.alibaba.com>
>> Signed-off-by: Boshi Yu <boshiyu@linux.alibaba.com>
>> ---
>> drivers/infiniband/hw/erdma/erdma.h | 7 +-
>> drivers/infiniband/hw/erdma/erdma_cmdq.c | 102 +++++++++++++---------
>> drivers/infiniband/hw/erdma/erdma_eq.c | 55 +++++++-----
>> drivers/infiniband/hw/erdma/erdma_main.c | 15 +++-
>> drivers/infiniband/hw/erdma/erdma_verbs.c | 85 ++++++++++--------
>> drivers/infiniband/hw/erdma/erdma_verbs.h | 4 +
>> 6 files changed, 167 insertions(+), 101 deletions(-)
>>
>> diff --git a/drivers/infiniband/hw/erdma/erdma.h
>> b/drivers/infiniband/hw/erdma/erdma.h
>> index 5df401a30cb9..e116263a608f 100644
>> --- a/drivers/infiniband/hw/erdma/erdma.h
>> +++ b/drivers/infiniband/hw/erdma/erdma.h
>> @@ -34,6 +34,7 @@ struct erdma_eq {
>> void __iomem *db;
>> u64 *db_record;
>> + dma_addr_t db_record_dma_addr;
>> };
>> struct erdma_cmdq_sq {
>> @@ -49,6 +50,7 @@ struct erdma_cmdq_sq {
>> u16 wqebb_cnt;
>> u64 *db_record;
>> + dma_addr_t db_record_dma_addr;
>> };
>> struct erdma_cmdq_cq {
>> @@ -62,6 +64,7 @@ struct erdma_cmdq_cq {
>> u32 cmdsn;
>> u64 *db_record;
>> + dma_addr_t db_record_dma_addr;
>> atomic64_t armed_num;
>> };
>> @@ -177,9 +180,6 @@ enum {
>> ERDMA_RES_CNT = 2,
>> };
>> -#define ERDMA_EXTRA_BUFFER_SIZE ERDMA_DB_SIZE
>> -#define WARPPED_BUFSIZE(size) ((size) + ERDMA_EXTRA_BUFFER_SIZE)
>> -
>> struct erdma_dev {
>> struct ib_device ibdev;
>> struct net_device *netdev;
>> @@ -213,6 +213,7 @@ struct erdma_dev {
>> atomic_t num_ctx;
>> struct list_head cep_list;
>> + struct dma_pool *db_pool;
>> struct dma_pool *resp_pool;
>> };
>> diff --git a/drivers/infiniband/hw/erdma/erdma_cmdq.c
>> b/drivers/infiniband/hw/erdma/erdma_cmdq.c
>> index a151a7bdd504..c2c666040949 100644
>> --- a/drivers/infiniband/hw/erdma/erdma_cmdq.c
>> +++ b/drivers/infiniband/hw/erdma/erdma_cmdq.c
>> @@ -89,20 +89,19 @@ static int erdma_cmdq_sq_init(struct erdma_dev *dev)
>> {
>> struct erdma_cmdq *cmdq = &dev->cmdq;
>> struct erdma_cmdq_sq *sq = &cmdq->sq;
>> - u32 buf_size;
>> sq->wqebb_cnt = SQEBB_COUNT(ERDMA_CMDQ_SQE_SIZE);
>> sq->depth = cmdq->max_outstandings * sq->wqebb_cnt;
>> - buf_size = sq->depth << SQEBB_SHIFT;
>> -
>> - sq->qbuf =
>> - dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
>> - &sq->qbuf_dma_addr, GFP_KERNEL);
>> + sq->qbuf = dma_alloc_coherent(&dev->pdev->dev, sq->depth <<
>> SQEBB_SHIFT,
>> + &sq->qbuf_dma_addr, GFP_KERNEL);
>> if (!sq->qbuf)
>> return -ENOMEM;
>> - sq->db_record = (u64 *)(sq->qbuf + buf_size);
>> + sq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
>> + &sq->db_record_dma_addr);
>> + if (!sq->db_record)
>> + goto err_out;
>> spin_lock_init(&sq->lock);
>> @@ -112,29 +111,35 @@ static int erdma_cmdq_sq_init(struct erdma_dev
>> *dev)
>> lower_32_bits(sq->qbuf_dma_addr));
>> erdma_reg_write32(dev, ERDMA_REGS_CMDQ_DEPTH_REG, sq->depth);
>> erdma_reg_write64(dev, ERDMA_CMDQ_SQ_DB_HOST_ADDR_REG,
>> - sq->qbuf_dma_addr + buf_size);
>> + sq->db_record_dma_addr);
>> return 0;
>> +
>> +err_out:
>> + dma_free_coherent(&dev->pdev->dev, sq->depth << SQEBB_SHIFT,
>> + sq->qbuf, sq->qbuf_dma_addr);
>> +
>> + return -ENOMEM;
>> }
>> static int erdma_cmdq_cq_init(struct erdma_dev *dev)
>> {
>> struct erdma_cmdq *cmdq = &dev->cmdq;
>> struct erdma_cmdq_cq *cq = &cmdq->cq;
>> - u32 buf_size;
>> cq->depth = cmdq->sq.depth;
>> - buf_size = cq->depth << CQE_SHIFT;
>> -
>> - cq->qbuf =
>> - dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
>> - &cq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
>> + cq->qbuf = dma_alloc_coherent(&dev->pdev->dev, cq->depth <<
>> CQE_SHIFT,
>> + &cq->qbuf_dma_addr,
>> + GFP_KERNEL | __GFP_ZERO);
>
> <...>
>
>> if (!cq->qbuf)
>> return -ENOMEM;
>> spin_lock_init(&cq->lock);
>> - cq->db_record = (u64 *)(cq->qbuf + buf_size);
>> + cq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
>> + &cq->db_record_dma_addr);
>> + if (!cq->db_record)
>> + goto err_out;
>> atomic64_set(&cq->armed_num, 0);
>> @@ -143,23 +148,26 @@ static int erdma_cmdq_cq_init(struct erdma_dev
>> *dev)
>> erdma_reg_write32(dev, ERDMA_REGS_CMDQ_CQ_ADDR_L_REG,
>> lower_32_bits(cq->qbuf_dma_addr));
>> erdma_reg_write64(dev, ERDMA_CMDQ_CQ_DB_HOST_ADDR_REG,
>> - cq->qbuf_dma_addr + buf_size);
>> + cq->db_record_dma_addr);
>> return 0;
>> +
>> +err_out:
>> + dma_free_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT, cq->qbuf,
>> + cq->qbuf_dma_addr);
>> +
>> + return -ENOMEM;
>> }
>> static int erdma_cmdq_eq_init(struct erdma_dev *dev)
>> {
>> struct erdma_cmdq *cmdq = &dev->cmdq;
>> struct erdma_eq *eq = &cmdq->eq;
>> - u32 buf_size;
>> eq->depth = cmdq->max_outstandings;
>> - buf_size = eq->depth << EQE_SHIFT;
>> -
>> - eq->qbuf =
>> - dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
>> - &eq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
>> + eq->qbuf = dma_alloc_coherent(&dev->pdev->dev, eq->depth <<
>> EQE_SHIFT,
>> + &eq->qbuf_dma_addr,
>> + GFP_KERNEL | __GFP_ZERO);
> <...>
>> if (!eq->qbuf)
>> return -ENOMEM;
>> @@ -167,7 +175,10 @@ static int erdma_cmdq_eq_init(struct erdma_dev *dev)
>> atomic64_set(&eq->event_num, 0);
>> eq->db = dev->func_bar + ERDMA_REGS_CEQ_DB_BASE_REG;
>> - eq->db_record = (u64 *)(eq->qbuf + buf_size);
>> + eq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
>> + &eq->db_record_dma_addr);
>> + if (!eq->db_record)
>> + goto err_out;
>> erdma_reg_write32(dev, ERDMA_REGS_CMDQ_EQ_ADDR_H_REG,
>> upper_32_bits(eq->qbuf_dma_addr));
>> @@ -175,9 +186,15 @@ static int erdma_cmdq_eq_init(struct erdma_dev *dev)
>> lower_32_bits(eq->qbuf_dma_addr));
>> erdma_reg_write32(dev, ERDMA_REGS_CMDQ_EQ_DEPTH_REG, eq->depth);
>> erdma_reg_write64(dev, ERDMA_CMDQ_EQ_DB_HOST_ADDR_REG,
>> - eq->qbuf_dma_addr + buf_size);
>> + eq->db_record_dma_addr);
>> return 0;
>> +
>> +err_out:
>> + dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT, eq->qbuf,
>> + eq->qbuf_dma_addr);
>> +
>> + return -ENOMEM;
>> }
>> int erdma_cmdq_init(struct erdma_dev *dev)
>> @@ -211,17 +228,19 @@ int erdma_cmdq_init(struct erdma_dev *dev)
>> return 0;
>> err_destroy_cq:
>> - dma_free_coherent(&dev->pdev->dev,
>> - (cmdq->cq.depth << CQE_SHIFT) +
>> - ERDMA_EXTRA_BUFFER_SIZE,
>> + dma_free_coherent(&dev->pdev->dev, cmdq->cq.depth << CQE_SHIFT,
>> cmdq->cq.qbuf, cmdq->cq.qbuf_dma_addr);
>> + dma_pool_free(dev->db_pool, cmdq->cq.db_record,
>> + cmdq->cq.db_record_dma_addr);
>> +
>> err_destroy_sq:
>> - dma_free_coherent(&dev->pdev->dev,
>> - (cmdq->sq.depth << SQEBB_SHIFT) +
>> - ERDMA_EXTRA_BUFFER_SIZE,
>> + dma_free_coherent(&dev->pdev->dev, cmdq->sq.depth << SQEBB_SHIFT,
>> cmdq->sq.qbuf, cmdq->sq.qbuf_dma_addr);
>> + dma_pool_free(dev->db_pool, cmdq->sq.db_record,
>> + cmdq->sq.db_record_dma_addr);
>> +
>> return err;
>> }
>> @@ -238,18 +257,23 @@ void erdma_cmdq_destroy(struct erdma_dev *dev)
>> clear_bit(ERDMA_CMDQ_STATE_OK_BIT, &cmdq->state);
>> - dma_free_coherent(&dev->pdev->dev,
>> - (cmdq->eq.depth << EQE_SHIFT) +
>> - ERDMA_EXTRA_BUFFER_SIZE,
>> + dma_free_coherent(&dev->pdev->dev, cmdq->eq.depth << EQE_SHIFT,
>> cmdq->eq.qbuf, cmdq->eq.qbuf_dma_addr);
>> - dma_free_coherent(&dev->pdev->dev,
>> - (cmdq->sq.depth << SQEBB_SHIFT) +
>> - ERDMA_EXTRA_BUFFER_SIZE,
>> +
>> + dma_pool_free(dev->db_pool, cmdq->eq.db_record,
>> + cmdq->eq.db_record_dma_addr);
>> +
>> + dma_free_coherent(&dev->pdev->dev, cmdq->sq.depth << SQEBB_SHIFT,
>> cmdq->sq.qbuf, cmdq->sq.qbuf_dma_addr);
>> - dma_free_coherent(&dev->pdev->dev,
>> - (cmdq->cq.depth << CQE_SHIFT) +
>> - ERDMA_EXTRA_BUFFER_SIZE,
>> +
>> + dma_pool_free(dev->db_pool, cmdq->sq.db_record,
>> + cmdq->sq.db_record_dma_addr);
>> +
>> + dma_free_coherent(&dev->pdev->dev, cmdq->cq.depth << CQE_SHIFT,
>> cmdq->cq.qbuf, cmdq->cq.qbuf_dma_addr);
>> +
>> + dma_pool_free(dev->db_pool, cmdq->cq.db_record,
>> + cmdq->cq.db_record_dma_addr);
>> }
>> static void *get_next_valid_cmdq_cqe(struct erdma_cmdq *cmdq)
>> diff --git a/drivers/infiniband/hw/erdma/erdma_eq.c
>> b/drivers/infiniband/hw/erdma/erdma_eq.c
>> index ea47cb21fdb8..809c33628f38 100644
>> --- a/drivers/infiniband/hw/erdma/erdma_eq.c
>> +++ b/drivers/infiniband/hw/erdma/erdma_eq.c
>> @@ -83,14 +83,12 @@ void erdma_aeq_event_handler(struct erdma_dev *dev)
>> int erdma_aeq_init(struct erdma_dev *dev)
>> {
>> struct erdma_eq *eq = &dev->aeq;
>> - u32 buf_size;
>> eq->depth = ERDMA_DEFAULT_EQ_DEPTH;
>> - buf_size = eq->depth << EQE_SHIFT;
>> - eq->qbuf =
>> - dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
>> - &eq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
>> + eq->qbuf = dma_alloc_coherent(&dev->pdev->dev, eq->depth <<
>> EQE_SHIFT,
>> + &eq->qbuf_dma_addr,
>> + GFP_KERNEL | __GFP_ZERO);
>
> <...>
>
>> if (!eq->qbuf)
>> return -ENOMEM;
>> @@ -99,7 +97,10 @@ int erdma_aeq_init(struct erdma_dev *dev)
>> atomic64_set(&eq->notify_num, 0);
>> eq->db = dev->func_bar + ERDMA_REGS_AEQ_DB_REG;
>> - eq->db_record = (u64 *)(eq->qbuf + buf_size);
>> + eq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
>> + &eq->db_record_dma_addr);
>> + if (!eq->db_record)
>> + goto err_out;
>> erdma_reg_write32(dev, ERDMA_REGS_AEQ_ADDR_H_REG,
>> upper_32_bits(eq->qbuf_dma_addr));
>> @@ -107,18 +108,25 @@ int erdma_aeq_init(struct erdma_dev *dev)
>> lower_32_bits(eq->qbuf_dma_addr));
>> erdma_reg_write32(dev, ERDMA_REGS_AEQ_DEPTH_REG, eq->depth);
>> erdma_reg_write64(dev, ERDMA_AEQ_DB_HOST_ADDR_REG,
>> - eq->qbuf_dma_addr + buf_size);
>> + eq->db_record_dma_addr);
>> return 0;
>> +
>> +err_out:
>> + dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT, eq->qbuf,
>> + eq->qbuf_dma_addr);
>> +
>> + return -ENOMEM;
>> }
>> void erdma_aeq_destroy(struct erdma_dev *dev)
>> {
>> struct erdma_eq *eq = &dev->aeq;
>> - dma_free_coherent(&dev->pdev->dev,
>> - WARPPED_BUFSIZE(eq->depth << EQE_SHIFT), eq->qbuf,
>> + dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT, eq->qbuf,
>> eq->qbuf_dma_addr);
>> +
>> + dma_pool_free(dev->db_pool, eq->db_record, eq->db_record_dma_addr);
>> }
>> void erdma_ceq_completion_handler(struct erdma_eq_cb *ceq_cb)
>> @@ -209,7 +217,6 @@ static void erdma_free_ceq_irq(struct erdma_dev
>> *dev, u16 ceqn)
>> static int create_eq_cmd(struct erdma_dev *dev, u32 eqn, struct
>> erdma_eq *eq)
>> {
>> struct erdma_cmdq_create_eq_req req;
>> - dma_addr_t db_info_dma_addr;
>> erdma_cmdq_build_reqhdr(&req.hdr, CMDQ_SUBMOD_COMMON,
>> CMDQ_OPCODE_CREATE_EQ);
>> @@ -219,9 +226,8 @@ static int create_eq_cmd(struct erdma_dev *dev,
>> u32 eqn, struct erdma_eq *eq)
>> req.qtype = ERDMA_EQ_TYPE_CEQ;
>> /* Vector index is the same as EQN. */
>> req.vector_idx = eqn;
>> - db_info_dma_addr = eq->qbuf_dma_addr + (eq->depth << EQE_SHIFT);
>> - req.db_dma_addr_l = lower_32_bits(db_info_dma_addr);
>> - req.db_dma_addr_h = upper_32_bits(db_info_dma_addr);
>> + req.db_dma_addr_l = lower_32_bits(eq->db_record_dma_addr);
>> + req.db_dma_addr_h = upper_32_bits(eq->db_record_dma_addr);
>> return erdma_post_cmd_wait(&dev->cmdq, &req, sizeof(req), NULL,
>> NULL);
>> }
>> @@ -229,12 +235,12 @@ static int create_eq_cmd(struct erdma_dev *dev,
>> u32 eqn, struct erdma_eq *eq)
>> static int erdma_ceq_init_one(struct erdma_dev *dev, u16 ceqn)
>> {
>> struct erdma_eq *eq = &dev->ceqs[ceqn].eq;
>> - u32 buf_size = ERDMA_DEFAULT_EQ_DEPTH << EQE_SHIFT;
>> int ret;
>> - eq->qbuf =
>> - dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
>> - &eq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
>> + eq->depth = ERDMA_DEFAULT_EQ_DEPTH;
>> + eq->qbuf = dma_alloc_coherent(&dev->pdev->dev, eq->depth <<
>> EQE_SHIFT,
>> + &eq->qbuf_dma_addr,
>> + GFP_KERNEL | __GFP_ZERO);
>
> The patch in this link
> (https://lore.kernel.org/all/20181214082515.14835-1-hch@lst.de/T/#m70c723c646004445713f31b7837f7e9d910c06f5) makes sure that dma_alloc_coherent(xxx) always returns zeroed memory.
>
> So __GFP_ZERO is necessary? can we remove them?
Sorry. Please ignore this mail. In patch 3, these __GFP_ZEROs are removed.
Zhu Yanjun
>
> Zhu Yanjun
>
>> if (!eq->qbuf)
>> return -ENOMEM;
>> @@ -242,10 +248,17 @@ static int erdma_ceq_init_one(struct erdma_dev
>> *dev, u16 ceqn)
>> atomic64_set(&eq->event_num, 0);
>> atomic64_set(&eq->notify_num, 0);
>> - eq->depth = ERDMA_DEFAULT_EQ_DEPTH;
>> eq->db = dev->func_bar + ERDMA_REGS_CEQ_DB_BASE_REG +
>> (ceqn + 1) * ERDMA_DB_SIZE;
>> - eq->db_record = (u64 *)(eq->qbuf + buf_size);
>> +
>> + eq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
>> + &eq->db_record_dma_addr);
>> + if (!eq->db_record) {
>> + dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT,
>> + eq->qbuf, eq->qbuf_dma_addr);
>> + return -ENOMEM;
>> + }
>> +
>> eq->ci = 0;
>> dev->ceqs[ceqn].dev = dev;
>> @@ -259,7 +272,6 @@ static int erdma_ceq_init_one(struct erdma_dev
>> *dev, u16 ceqn)
>> static void erdma_ceq_uninit_one(struct erdma_dev *dev, u16 ceqn)
>> {
>> struct erdma_eq *eq = &dev->ceqs[ceqn].eq;
>> - u32 buf_size = ERDMA_DEFAULT_EQ_DEPTH << EQE_SHIFT;
>> struct erdma_cmdq_destroy_eq_req req;
>> int err;
>> @@ -276,8 +288,9 @@ static void erdma_ceq_uninit_one(struct erdma_dev
>> *dev, u16 ceqn)
>> if (err)
>> return;
>> - dma_free_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
>> eq->qbuf,
>> + dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT, eq->qbuf,
>> eq->qbuf_dma_addr);
>> + dma_pool_free(dev->db_pool, eq->db_record, eq->db_record_dma_addr);
>> }
>> int erdma_ceqs_init(struct erdma_dev *dev)
>> diff --git a/drivers/infiniband/hw/erdma/erdma_main.c
>> b/drivers/infiniband/hw/erdma/erdma_main.c
>> index 472939172f0c..7080f8a71ec4 100644
>> --- a/drivers/infiniband/hw/erdma/erdma_main.c
>> +++ b/drivers/infiniband/hw/erdma/erdma_main.c
>> @@ -178,16 +178,26 @@ static int erdma_device_init(struct erdma_dev
>> *dev, struct pci_dev *pdev)
>> if (!dev->resp_pool)
>> return -ENOMEM;
>> + dev->db_pool = dma_pool_create("erdma_db_pool", &pdev->dev,
>> + ERDMA_DB_SIZE, ERDMA_DB_SIZE, 0);
>> + if (!dev->db_pool) {
>> + ret = -ENOMEM;
>> + goto destroy_resp_pool;
>> + }
>> +
>> ret = dma_set_mask_and_coherent(&pdev->dev,
>> DMA_BIT_MASK(ERDMA_PCI_WIDTH));
>> if (ret)
>> - goto destroy_pool;
>> + goto destroy_db_pool;
>> dma_set_max_seg_size(&pdev->dev, UINT_MAX);
>> return 0;
>> -destroy_pool:
>> +destroy_db_pool:
>> + dma_pool_destroy(dev->db_pool);
>> +
>> +destroy_resp_pool:
>> dma_pool_destroy(dev->resp_pool);
>> return ret;
>> @@ -195,6 +205,7 @@ static int erdma_device_init(struct erdma_dev
>> *dev, struct pci_dev *pdev)
>> static void erdma_device_uninit(struct erdma_dev *dev)
>> {
>> + dma_pool_destroy(dev->db_pool);
>> dma_pool_destroy(dev->resp_pool);
>> }
>> diff --git a/drivers/infiniband/hw/erdma/erdma_verbs.c
>> b/drivers/infiniband/hw/erdma/erdma_verbs.c
>> index 23dfc01603f8..b78ddca1483e 100644
>> --- a/drivers/infiniband/hw/erdma/erdma_verbs.c
>> +++ b/drivers/infiniband/hw/erdma/erdma_verbs.c
>> @@ -76,10 +76,8 @@ static int create_qp_cmd(struct erdma_ucontext
>> *uctx, struct erdma_qp *qp)
>> req.rq_buf_addr = qp->kern_qp.rq_buf_dma_addr;
>> req.sq_buf_addr = qp->kern_qp.sq_buf_dma_addr;
>> - req.sq_db_info_dma_addr = qp->kern_qp.sq_buf_dma_addr +
>> - (qp->attrs.sq_size << SQEBB_SHIFT);
>> - req.rq_db_info_dma_addr = qp->kern_qp.rq_buf_dma_addr +
>> - (qp->attrs.rq_size << RQE_SHIFT);
>> + req.sq_db_info_dma_addr = qp->kern_qp.sq_db_info_dma_addr;
>> + req.rq_db_info_dma_addr = qp->kern_qp.rq_db_info_dma_addr;
>> } else {
>> user_qp = &qp->user_qp;
>> req.sq_cqn_mtt_cfg = FIELD_PREP(
>> @@ -209,8 +207,7 @@ static int create_cq_cmd(struct erdma_ucontext
>> *uctx, struct erdma_cq *cq)
>> ERDMA_MR_MTT_0LEVEL);
>> req.first_page_offset = 0;
>> - req.cq_db_info_addr =
>> - cq->kern_cq.qbuf_dma_addr + (cq->depth << CQE_SHIFT);
>> + req.cq_db_info_addr = cq->kern_cq.db_record_dma_addr;
>> } else {
>> mem = &cq->user_cq.qbuf_mem;
>> req.cfg0 |=
>> @@ -482,16 +479,24 @@ static void free_kernel_qp(struct erdma_qp *qp)
>> vfree(qp->kern_qp.rwr_tbl);
>> if (qp->kern_qp.sq_buf)
>> - dma_free_coherent(
>> - &dev->pdev->dev,
>> - WARPPED_BUFSIZE(qp->attrs.sq_size << SQEBB_SHIFT),
>> - qp->kern_qp.sq_buf, qp->kern_qp.sq_buf_dma_addr);
>> + dma_free_coherent(&dev->pdev->dev,
>> + qp->attrs.sq_size << SQEBB_SHIFT,
>> + qp->kern_qp.sq_buf,
>> + qp->kern_qp.sq_buf_dma_addr);
>> +
>> + if (qp->kern_qp.sq_db_info)
>> + dma_pool_free(dev->db_pool, qp->kern_qp.sq_db_info,
>> + qp->kern_qp.sq_db_info_dma_addr);
>> if (qp->kern_qp.rq_buf)
>> - dma_free_coherent(
>> - &dev->pdev->dev,
>> - WARPPED_BUFSIZE(qp->attrs.rq_size << RQE_SHIFT),
>> - qp->kern_qp.rq_buf, qp->kern_qp.rq_buf_dma_addr);
>> + dma_free_coherent(&dev->pdev->dev,
>> + qp->attrs.rq_size << RQE_SHIFT,
>> + qp->kern_qp.rq_buf,
>> + qp->kern_qp.rq_buf_dma_addr);
>> +
>> + if (qp->kern_qp.rq_db_info)
>> + dma_pool_free(dev->db_pool, qp->kern_qp.rq_db_info,
>> + qp->kern_qp.rq_db_info_dma_addr);
>> }
>> static int init_kernel_qp(struct erdma_dev *dev, struct erdma_qp *qp,
>> @@ -516,20 +521,27 @@ static int init_kernel_qp(struct erdma_dev *dev,
>> struct erdma_qp *qp,
>> if (!kqp->swr_tbl || !kqp->rwr_tbl)
>> goto err_out;
>> - size = (qp->attrs.sq_size << SQEBB_SHIFT) + ERDMA_EXTRA_BUFFER_SIZE;
>> + size = qp->attrs.sq_size << SQEBB_SHIFT;
>> kqp->sq_buf = dma_alloc_coherent(&dev->pdev->dev, size,
>> &kqp->sq_buf_dma_addr, GFP_KERNEL);
>> if (!kqp->sq_buf)
>> goto err_out;
>> - size = (qp->attrs.rq_size << RQE_SHIFT) + ERDMA_EXTRA_BUFFER_SIZE;
>> + kqp->sq_db_info = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
>> + &kqp->sq_db_info_dma_addr);
>> + if (!kqp->sq_db_info)
>> + goto err_out;
>> +
>> + size = qp->attrs.rq_size << RQE_SHIFT;
>> kqp->rq_buf = dma_alloc_coherent(&dev->pdev->dev, size,
>> &kqp->rq_buf_dma_addr, GFP_KERNEL);
>> if (!kqp->rq_buf)
>> goto err_out;
>> - kqp->sq_db_info = kqp->sq_buf + (qp->attrs.sq_size << SQEBB_SHIFT);
>> - kqp->rq_db_info = kqp->rq_buf + (qp->attrs.rq_size << RQE_SHIFT);
>> + kqp->rq_db_info = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
>> + &kqp->rq_db_info_dma_addr);
>> + if (!kqp->rq_db_info)
>> + goto err_out;
>> return 0;
>> @@ -1237,9 +1249,10 @@ int erdma_destroy_cq(struct ib_cq *ibcq, struct
>> ib_udata *udata)
>> return err;
>> if (rdma_is_kernel_res(&cq->ibcq.res)) {
>> - dma_free_coherent(&dev->pdev->dev,
>> - WARPPED_BUFSIZE(cq->depth << CQE_SHIFT),
>> + dma_free_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT,
>> cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr);
>> + dma_pool_free(dev->db_pool, cq->kern_cq.db_record,
>> + cq->kern_cq.db_record_dma_addr);
>> } else {
>> erdma_unmap_user_dbrecords(ctx, &cq->user_cq.user_dbr_page);
>> put_mtt_entries(dev, &cq->user_cq.qbuf_mem);
>> @@ -1279,16 +1292,7 @@ int erdma_destroy_qp(struct ib_qp *ibqp, struct
>> ib_udata *udata)
>> wait_for_completion(&qp->safe_free);
>> if (rdma_is_kernel_res(&qp->ibqp.res)) {
>> - vfree(qp->kern_qp.swr_tbl);
>> - vfree(qp->kern_qp.rwr_tbl);
>> - dma_free_coherent(
>> - &dev->pdev->dev,
>> - WARPPED_BUFSIZE(qp->attrs.rq_size << RQE_SHIFT),
>> - qp->kern_qp.rq_buf, qp->kern_qp.rq_buf_dma_addr);
>> - dma_free_coherent(
>> - &dev->pdev->dev,
>> - WARPPED_BUFSIZE(qp->attrs.sq_size << SQEBB_SHIFT),
>> - qp->kern_qp.sq_buf, qp->kern_qp.sq_buf_dma_addr);
>> + free_kernel_qp(qp);
>> } else {
>> put_mtt_entries(dev, &qp->user_qp.sq_mem);
>> put_mtt_entries(dev, &qp->user_qp.rq_mem);
>> @@ -1600,19 +1604,27 @@ static int erdma_init_kernel_cq(struct
>> erdma_cq *cq)
>> struct erdma_dev *dev = to_edev(cq->ibcq.device);
>> cq->kern_cq.qbuf =
>> - dma_alloc_coherent(&dev->pdev->dev,
>> - WARPPED_BUFSIZE(cq->depth << CQE_SHIFT),
>> + dma_alloc_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT,
>> &cq->kern_cq.qbuf_dma_addr, GFP_KERNEL);
>> if (!cq->kern_cq.qbuf)
>> return -ENOMEM;
>> - cq->kern_cq.db_record =
>> - (u64 *)(cq->kern_cq.qbuf + (cq->depth << CQE_SHIFT));
>> + cq->kern_cq.db_record = dma_pool_zalloc(
>> + dev->db_pool, GFP_KERNEL, &cq->kern_cq.db_record_dma_addr);
>> + if (!cq->kern_cq.db_record)
>> + goto err_out;
>> +
>> spin_lock_init(&cq->kern_cq.lock);
>> /* use default cqdb addr */
>> cq->kern_cq.db = dev->func_bar + ERDMA_BAR_CQDB_SPACE_OFFSET;
>> return 0;
>> +
>> +err_out:
>> + dma_free_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT,
>> + cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr);
>> +
>> + return -ENOMEM;
>> }
>> int erdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr
>> *attr,
>> @@ -1676,9 +1688,10 @@ int erdma_create_cq(struct ib_cq *ibcq, const
>> struct ib_cq_init_attr *attr,
>> erdma_unmap_user_dbrecords(ctx, &cq->user_cq.user_dbr_page);
>> put_mtt_entries(dev, &cq->user_cq.qbuf_mem);
>> } else {
>> - dma_free_coherent(&dev->pdev->dev,
>> - WARPPED_BUFSIZE(depth << CQE_SHIFT),
>> + dma_free_coherent(&dev->pdev->dev, depth << CQE_SHIFT,
>> cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr);
>> + dma_pool_free(dev->db_pool, cq->kern_cq.db_record,
>> + cq->kern_cq.db_record_dma_addr);
>> }
>> err_out_xa:
>> diff --git a/drivers/infiniband/hw/erdma/erdma_verbs.h
>> b/drivers/infiniband/hw/erdma/erdma_verbs.h
>> index db6018529ccc..b02ffdc8c811 100644
>> --- a/drivers/infiniband/hw/erdma/erdma_verbs.h
>> +++ b/drivers/infiniband/hw/erdma/erdma_verbs.h
>> @@ -170,6 +170,9 @@ struct erdma_kqp {
>> void *sq_db_info;
>> void *rq_db_info;
>> + dma_addr_t sq_db_info_dma_addr;
>> + dma_addr_t rq_db_info_dma_addr;
>> +
>> u8 sig_all;
>> };
>> @@ -247,6 +250,7 @@ struct erdma_kcq_info {
>> spinlock_t lock;
>> u8 __iomem *db;
>> u64 *db_record;
>> + dma_addr_t db_record_dma_addr;
>> };
>> struct erdma_ucq_info {
>
next prev parent reply other threads:[~2024-03-14 10:23 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-03-11 11:38 [PATCH for-next 0/3] RDMA/erdma: A series of fixes for the erdma driver Boshi Yu
2024-03-11 11:38 ` [PATCH for-next 1/3] RDMA/erdma: Allocate doorbell records from dma pool Boshi Yu
2024-03-13 16:20 ` Zhu Yanjun
2024-03-14 10:23 ` Zhu Yanjun [this message]
2024-03-11 11:38 ` [PATCH for-next 2/3] RDMA/erdma: Unify the names related to doorbell records Boshi Yu
2024-03-11 11:38 ` [PATCH for-next 3/3] RDMA/erdma: Remove unnecessary __GFP_ZERO flag Boshi Yu
2024-03-12 10:53 ` Leon Romanovsky
2024-03-13 2:40 ` Boshi Yu
2024-03-13 9:31 ` Leon Romanovsky
2024-03-12 10:54 ` [PATCH for-next 0/3] RDMA/erdma: A series of fixes for the erdma driver Leon Romanovsky
2024-04-01 11:46 ` Leon Romanovsky
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=2ee0f7b0-21fc-43d2-a126-850e73ff84fd@gmail.com \
--to=zyjzyj2000@gmail.com \
--cc=KaiShen@linux.alibaba.com \
--cc=boshiyu@alibaba-inc.com \
--cc=chengyou@linux.alibaba.com \
--cc=jgg@ziepe.ca \
--cc=leon@kernel.org \
--cc=linux-rdma@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox