public inbox for linux-rdma@vger.kernel.org
 help / color / mirror / Atom feed
From: Zhu Yanjun <zyjzyj2000@gmail.com>
To: Boshi Yu <boshiyu@alibaba-inc.com>, jgg@ziepe.ca, leon@kernel.org
Cc: linux-rdma@vger.kernel.org, chengyou@linux.alibaba.com,
	KaiShen@linux.alibaba.com
Subject: Re: [PATCH for-next 1/3] RDMA/erdma: Allocate doorbell records from dma pool
Date: Wed, 13 Mar 2024 17:20:03 +0100	[thread overview]
Message-ID: <09da3e92-3da5-405a-9f55-60533e966f09@linux.dev> (raw)
In-Reply-To: <20240311113821.22482-2-boshiyu@alibaba-inc.com>

On 11.03.24 12:38, Boshi Yu wrote:
> From: Boshi Yu <boshiyu@linux.alibaba.com>
> 
> Currently, the 8 byte doorbell record is allocated along with the queue
> buffer, which may result in waste of dma space when the queue buffer is
> page aligned. To address this issue, we introduce a dma pool named
> db_pool and allocate doorbell record from it.
> 
> Reviewed-by: Cheng Xu <chengyou@linux.alibaba.com>
> Signed-off-by: Boshi Yu <boshiyu@linux.alibaba.com>
> ---
>   drivers/infiniband/hw/erdma/erdma.h       |   7 +-
>   drivers/infiniband/hw/erdma/erdma_cmdq.c  | 102 +++++++++++++---------
>   drivers/infiniband/hw/erdma/erdma_eq.c    |  55 +++++++-----
>   drivers/infiniband/hw/erdma/erdma_main.c  |  15 +++-
>   drivers/infiniband/hw/erdma/erdma_verbs.c |  85 ++++++++++--------
>   drivers/infiniband/hw/erdma/erdma_verbs.h |   4 +
>   6 files changed, 167 insertions(+), 101 deletions(-)
> 
> diff --git a/drivers/infiniband/hw/erdma/erdma.h b/drivers/infiniband/hw/erdma/erdma.h
> index 5df401a30cb9..e116263a608f 100644
> --- a/drivers/infiniband/hw/erdma/erdma.h
> +++ b/drivers/infiniband/hw/erdma/erdma.h
> @@ -34,6 +34,7 @@ struct erdma_eq {
>   
>   	void __iomem *db;
>   	u64 *db_record;
> +	dma_addr_t db_record_dma_addr;
>   };
>   
>   struct erdma_cmdq_sq {
> @@ -49,6 +50,7 @@ struct erdma_cmdq_sq {
>   	u16 wqebb_cnt;
>   
>   	u64 *db_record;
> +	dma_addr_t db_record_dma_addr;
>   };
>   
>   struct erdma_cmdq_cq {
> @@ -62,6 +64,7 @@ struct erdma_cmdq_cq {
>   	u32 cmdsn;
>   
>   	u64 *db_record;
> +	dma_addr_t db_record_dma_addr;
>   
>   	atomic64_t armed_num;
>   };
> @@ -177,9 +180,6 @@ enum {
>   	ERDMA_RES_CNT = 2,
>   };
>   
> -#define ERDMA_EXTRA_BUFFER_SIZE ERDMA_DB_SIZE
> -#define WARPPED_BUFSIZE(size) ((size) + ERDMA_EXTRA_BUFFER_SIZE)
> -
>   struct erdma_dev {
>   	struct ib_device ibdev;
>   	struct net_device *netdev;
> @@ -213,6 +213,7 @@ struct erdma_dev {
>   	atomic_t num_ctx;
>   	struct list_head cep_list;
>   
> +	struct dma_pool *db_pool;
>   	struct dma_pool *resp_pool;
>   };
>   
> diff --git a/drivers/infiniband/hw/erdma/erdma_cmdq.c b/drivers/infiniband/hw/erdma/erdma_cmdq.c
> index a151a7bdd504..c2c666040949 100644
> --- a/drivers/infiniband/hw/erdma/erdma_cmdq.c
> +++ b/drivers/infiniband/hw/erdma/erdma_cmdq.c
> @@ -89,20 +89,19 @@ static int erdma_cmdq_sq_init(struct erdma_dev *dev)
>   {
>   	struct erdma_cmdq *cmdq = &dev->cmdq;
>   	struct erdma_cmdq_sq *sq = &cmdq->sq;
> -	u32 buf_size;
>   
>   	sq->wqebb_cnt = SQEBB_COUNT(ERDMA_CMDQ_SQE_SIZE);
>   	sq->depth = cmdq->max_outstandings * sq->wqebb_cnt;
>   
> -	buf_size = sq->depth << SQEBB_SHIFT;
> -
> -	sq->qbuf =
> -		dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
> -				   &sq->qbuf_dma_addr, GFP_KERNEL);
> +	sq->qbuf = dma_alloc_coherent(&dev->pdev->dev, sq->depth << SQEBB_SHIFT,
> +				      &sq->qbuf_dma_addr, GFP_KERNEL);
>   	if (!sq->qbuf)
>   		return -ENOMEM;
>   
> -	sq->db_record = (u64 *)(sq->qbuf + buf_size);
> +	sq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
> +					&sq->db_record_dma_addr);
> +	if (!sq->db_record)
> +		goto err_out;
>   
>   	spin_lock_init(&sq->lock);
>   
> @@ -112,29 +111,35 @@ static int erdma_cmdq_sq_init(struct erdma_dev *dev)
>   			  lower_32_bits(sq->qbuf_dma_addr));
>   	erdma_reg_write32(dev, ERDMA_REGS_CMDQ_DEPTH_REG, sq->depth);
>   	erdma_reg_write64(dev, ERDMA_CMDQ_SQ_DB_HOST_ADDR_REG,
> -			  sq->qbuf_dma_addr + buf_size);
> +			  sq->db_record_dma_addr);
>   
>   	return 0;
> +
> +err_out:
> +	dma_free_coherent(&dev->pdev->dev, sq->depth << SQEBB_SHIFT,
> +			  sq->qbuf, sq->qbuf_dma_addr);
> +
> +	return -ENOMEM;
>   }
>   
>   static int erdma_cmdq_cq_init(struct erdma_dev *dev)
>   {
>   	struct erdma_cmdq *cmdq = &dev->cmdq;
>   	struct erdma_cmdq_cq *cq = &cmdq->cq;
> -	u32 buf_size;
>   
>   	cq->depth = cmdq->sq.depth;
> -	buf_size = cq->depth << CQE_SHIFT;
> -
> -	cq->qbuf =
> -		dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
> -				   &cq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
> +	cq->qbuf = dma_alloc_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT,
> +				      &cq->qbuf_dma_addr,
> +				      GFP_KERNEL | __GFP_ZERO);

<...>

>   	if (!cq->qbuf)
>   		return -ENOMEM;
>   
>   	spin_lock_init(&cq->lock);
>   
> -	cq->db_record = (u64 *)(cq->qbuf + buf_size);
> +	cq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
> +					&cq->db_record_dma_addr);
> +	if (!cq->db_record)
> +		goto err_out;
>   
>   	atomic64_set(&cq->armed_num, 0);
>   
> @@ -143,23 +148,26 @@ static int erdma_cmdq_cq_init(struct erdma_dev *dev)
>   	erdma_reg_write32(dev, ERDMA_REGS_CMDQ_CQ_ADDR_L_REG,
>   			  lower_32_bits(cq->qbuf_dma_addr));
>   	erdma_reg_write64(dev, ERDMA_CMDQ_CQ_DB_HOST_ADDR_REG,
> -			  cq->qbuf_dma_addr + buf_size);
> +			  cq->db_record_dma_addr);
>   
>   	return 0;
> +
> +err_out:
> +	dma_free_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT, cq->qbuf,
> +			  cq->qbuf_dma_addr);
> +
> +	return -ENOMEM;
>   }
>   
>   static int erdma_cmdq_eq_init(struct erdma_dev *dev)
>   {
>   	struct erdma_cmdq *cmdq = &dev->cmdq;
>   	struct erdma_eq *eq = &cmdq->eq;
> -	u32 buf_size;
>   
>   	eq->depth = cmdq->max_outstandings;
> -	buf_size = eq->depth << EQE_SHIFT;
> -
> -	eq->qbuf =
> -		dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
> -				   &eq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
> +	eq->qbuf = dma_alloc_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT,
> +				      &eq->qbuf_dma_addr,
> +				      GFP_KERNEL | __GFP_ZERO);
<...>
>   	if (!eq->qbuf)
>   		return -ENOMEM;
>   
> @@ -167,7 +175,10 @@ static int erdma_cmdq_eq_init(struct erdma_dev *dev)
>   	atomic64_set(&eq->event_num, 0);
>   
>   	eq->db = dev->func_bar + ERDMA_REGS_CEQ_DB_BASE_REG;
> -	eq->db_record = (u64 *)(eq->qbuf + buf_size);
> +	eq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
> +					&eq->db_record_dma_addr);
> +	if (!eq->db_record)
> +		goto err_out;
>   
>   	erdma_reg_write32(dev, ERDMA_REGS_CMDQ_EQ_ADDR_H_REG,
>   			  upper_32_bits(eq->qbuf_dma_addr));
> @@ -175,9 +186,15 @@ static int erdma_cmdq_eq_init(struct erdma_dev *dev)
>   			  lower_32_bits(eq->qbuf_dma_addr));
>   	erdma_reg_write32(dev, ERDMA_REGS_CMDQ_EQ_DEPTH_REG, eq->depth);
>   	erdma_reg_write64(dev, ERDMA_CMDQ_EQ_DB_HOST_ADDR_REG,
> -			  eq->qbuf_dma_addr + buf_size);
> +			  eq->db_record_dma_addr);
>   
>   	return 0;
> +
> +err_out:
> +	dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT, eq->qbuf,
> +			  eq->qbuf_dma_addr);
> +
> +	return -ENOMEM;
>   }
>   
>   int erdma_cmdq_init(struct erdma_dev *dev)
> @@ -211,17 +228,19 @@ int erdma_cmdq_init(struct erdma_dev *dev)
>   	return 0;
>   
>   err_destroy_cq:
> -	dma_free_coherent(&dev->pdev->dev,
> -			  (cmdq->cq.depth << CQE_SHIFT) +
> -				  ERDMA_EXTRA_BUFFER_SIZE,
> +	dma_free_coherent(&dev->pdev->dev, cmdq->cq.depth << CQE_SHIFT,
>   			  cmdq->cq.qbuf, cmdq->cq.qbuf_dma_addr);
>   
> +	dma_pool_free(dev->db_pool, cmdq->cq.db_record,
> +		      cmdq->cq.db_record_dma_addr);
> +
>   err_destroy_sq:
> -	dma_free_coherent(&dev->pdev->dev,
> -			  (cmdq->sq.depth << SQEBB_SHIFT) +
> -				  ERDMA_EXTRA_BUFFER_SIZE,
> +	dma_free_coherent(&dev->pdev->dev, cmdq->sq.depth << SQEBB_SHIFT,
>   			  cmdq->sq.qbuf, cmdq->sq.qbuf_dma_addr);
>   
> +	dma_pool_free(dev->db_pool, cmdq->sq.db_record,
> +		      cmdq->sq.db_record_dma_addr);
> +
>   	return err;
>   }
>   
> @@ -238,18 +257,23 @@ void erdma_cmdq_destroy(struct erdma_dev *dev)
>   
>   	clear_bit(ERDMA_CMDQ_STATE_OK_BIT, &cmdq->state);
>   
> -	dma_free_coherent(&dev->pdev->dev,
> -			  (cmdq->eq.depth << EQE_SHIFT) +
> -				  ERDMA_EXTRA_BUFFER_SIZE,
> +	dma_free_coherent(&dev->pdev->dev, cmdq->eq.depth << EQE_SHIFT,
>   			  cmdq->eq.qbuf, cmdq->eq.qbuf_dma_addr);
> -	dma_free_coherent(&dev->pdev->dev,
> -			  (cmdq->sq.depth << SQEBB_SHIFT) +
> -				  ERDMA_EXTRA_BUFFER_SIZE,
> +
> +	dma_pool_free(dev->db_pool, cmdq->eq.db_record,
> +		      cmdq->eq.db_record_dma_addr);
> +
> +	dma_free_coherent(&dev->pdev->dev, cmdq->sq.depth << SQEBB_SHIFT,
>   			  cmdq->sq.qbuf, cmdq->sq.qbuf_dma_addr);
> -	dma_free_coherent(&dev->pdev->dev,
> -			  (cmdq->cq.depth << CQE_SHIFT) +
> -				  ERDMA_EXTRA_BUFFER_SIZE,
> +
> +	dma_pool_free(dev->db_pool, cmdq->sq.db_record,
> +		      cmdq->sq.db_record_dma_addr);
> +
> +	dma_free_coherent(&dev->pdev->dev, cmdq->cq.depth << CQE_SHIFT,
>   			  cmdq->cq.qbuf, cmdq->cq.qbuf_dma_addr);
> +
> +	dma_pool_free(dev->db_pool, cmdq->cq.db_record,
> +		      cmdq->cq.db_record_dma_addr);
>   }
>   
>   static void *get_next_valid_cmdq_cqe(struct erdma_cmdq *cmdq)
> diff --git a/drivers/infiniband/hw/erdma/erdma_eq.c b/drivers/infiniband/hw/erdma/erdma_eq.c
> index ea47cb21fdb8..809c33628f38 100644
> --- a/drivers/infiniband/hw/erdma/erdma_eq.c
> +++ b/drivers/infiniband/hw/erdma/erdma_eq.c
> @@ -83,14 +83,12 @@ void erdma_aeq_event_handler(struct erdma_dev *dev)
>   int erdma_aeq_init(struct erdma_dev *dev)
>   {
>   	struct erdma_eq *eq = &dev->aeq;
> -	u32 buf_size;
>   
>   	eq->depth = ERDMA_DEFAULT_EQ_DEPTH;
> -	buf_size = eq->depth << EQE_SHIFT;
>   
> -	eq->qbuf =
> -		dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
> -				   &eq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
> +	eq->qbuf = dma_alloc_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT,
> +				      &eq->qbuf_dma_addr,
> +				      GFP_KERNEL | __GFP_ZERO);

<...>

>   	if (!eq->qbuf)
>   		return -ENOMEM;
>   
> @@ -99,7 +97,10 @@ int erdma_aeq_init(struct erdma_dev *dev)
>   	atomic64_set(&eq->notify_num, 0);
>   
>   	eq->db = dev->func_bar + ERDMA_REGS_AEQ_DB_REG;
> -	eq->db_record = (u64 *)(eq->qbuf + buf_size);
> +	eq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
> +					&eq->db_record_dma_addr);
> +	if (!eq->db_record)
> +		goto err_out;
>   
>   	erdma_reg_write32(dev, ERDMA_REGS_AEQ_ADDR_H_REG,
>   			  upper_32_bits(eq->qbuf_dma_addr));
> @@ -107,18 +108,25 @@ int erdma_aeq_init(struct erdma_dev *dev)
>   			  lower_32_bits(eq->qbuf_dma_addr));
>   	erdma_reg_write32(dev, ERDMA_REGS_AEQ_DEPTH_REG, eq->depth);
>   	erdma_reg_write64(dev, ERDMA_AEQ_DB_HOST_ADDR_REG,
> -			  eq->qbuf_dma_addr + buf_size);
> +			  eq->db_record_dma_addr);
>   
>   	return 0;
> +
> +err_out:
> +	dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT, eq->qbuf,
> +			  eq->qbuf_dma_addr);
> +
> +	return -ENOMEM;
>   }
>   
>   void erdma_aeq_destroy(struct erdma_dev *dev)
>   {
>   	struct erdma_eq *eq = &dev->aeq;
>   
> -	dma_free_coherent(&dev->pdev->dev,
> -			  WARPPED_BUFSIZE(eq->depth << EQE_SHIFT), eq->qbuf,
> +	dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT, eq->qbuf,
>   			  eq->qbuf_dma_addr);
> +
> +	dma_pool_free(dev->db_pool, eq->db_record, eq->db_record_dma_addr);
>   }
>   
>   void erdma_ceq_completion_handler(struct erdma_eq_cb *ceq_cb)
> @@ -209,7 +217,6 @@ static void erdma_free_ceq_irq(struct erdma_dev *dev, u16 ceqn)
>   static int create_eq_cmd(struct erdma_dev *dev, u32 eqn, struct erdma_eq *eq)
>   {
>   	struct erdma_cmdq_create_eq_req req;
> -	dma_addr_t db_info_dma_addr;
>   
>   	erdma_cmdq_build_reqhdr(&req.hdr, CMDQ_SUBMOD_COMMON,
>   				CMDQ_OPCODE_CREATE_EQ);
> @@ -219,9 +226,8 @@ static int create_eq_cmd(struct erdma_dev *dev, u32 eqn, struct erdma_eq *eq)
>   	req.qtype = ERDMA_EQ_TYPE_CEQ;
>   	/* Vector index is the same as EQN. */
>   	req.vector_idx = eqn;
> -	db_info_dma_addr = eq->qbuf_dma_addr + (eq->depth << EQE_SHIFT);
> -	req.db_dma_addr_l = lower_32_bits(db_info_dma_addr);
> -	req.db_dma_addr_h = upper_32_bits(db_info_dma_addr);
> +	req.db_dma_addr_l = lower_32_bits(eq->db_record_dma_addr);
> +	req.db_dma_addr_h = upper_32_bits(eq->db_record_dma_addr);
>   
>   	return erdma_post_cmd_wait(&dev->cmdq, &req, sizeof(req), NULL, NULL);
>   }
> @@ -229,12 +235,12 @@ static int create_eq_cmd(struct erdma_dev *dev, u32 eqn, struct erdma_eq *eq)
>   static int erdma_ceq_init_one(struct erdma_dev *dev, u16 ceqn)
>   {
>   	struct erdma_eq *eq = &dev->ceqs[ceqn].eq;
> -	u32 buf_size = ERDMA_DEFAULT_EQ_DEPTH << EQE_SHIFT;
>   	int ret;
>   
> -	eq->qbuf =
> -		dma_alloc_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size),
> -				   &eq->qbuf_dma_addr, GFP_KERNEL | __GFP_ZERO);
> +	eq->depth = ERDMA_DEFAULT_EQ_DEPTH;
> +	eq->qbuf = dma_alloc_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT,
> +				      &eq->qbuf_dma_addr,
> +				      GFP_KERNEL | __GFP_ZERO);

The patch in this link 
(https://lore.kernel.org/all/20181214082515.14835-1-hch@lst.de/T/#m70c723c646004445713f31b7837f7e9d910c06f5) 
makes sure that dma_alloc_coherent(xxx) always returns zeroed memory.

So __GFP_ZERO is necessary? can we remove them?

Zhu Yanjun

>   	if (!eq->qbuf)
>   		return -ENOMEM;
>   
> @@ -242,10 +248,17 @@ static int erdma_ceq_init_one(struct erdma_dev *dev, u16 ceqn)
>   	atomic64_set(&eq->event_num, 0);
>   	atomic64_set(&eq->notify_num, 0);
>   
> -	eq->depth = ERDMA_DEFAULT_EQ_DEPTH;
>   	eq->db = dev->func_bar + ERDMA_REGS_CEQ_DB_BASE_REG +
>   		 (ceqn + 1) * ERDMA_DB_SIZE;
> -	eq->db_record = (u64 *)(eq->qbuf + buf_size);
> +
> +	eq->db_record = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
> +					&eq->db_record_dma_addr);
> +	if (!eq->db_record) {
> +		dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT,
> +				  eq->qbuf, eq->qbuf_dma_addr);
> +		return -ENOMEM;
> +	}
> +
>   	eq->ci = 0;
>   	dev->ceqs[ceqn].dev = dev;
>   
> @@ -259,7 +272,6 @@ static int erdma_ceq_init_one(struct erdma_dev *dev, u16 ceqn)
>   static void erdma_ceq_uninit_one(struct erdma_dev *dev, u16 ceqn)
>   {
>   	struct erdma_eq *eq = &dev->ceqs[ceqn].eq;
> -	u32 buf_size = ERDMA_DEFAULT_EQ_DEPTH << EQE_SHIFT;
>   	struct erdma_cmdq_destroy_eq_req req;
>   	int err;
>   
> @@ -276,8 +288,9 @@ static void erdma_ceq_uninit_one(struct erdma_dev *dev, u16 ceqn)
>   	if (err)
>   		return;
>   
> -	dma_free_coherent(&dev->pdev->dev, WARPPED_BUFSIZE(buf_size), eq->qbuf,
> +	dma_free_coherent(&dev->pdev->dev, eq->depth << EQE_SHIFT, eq->qbuf,
>   			  eq->qbuf_dma_addr);
> +	dma_pool_free(dev->db_pool, eq->db_record, eq->db_record_dma_addr);
>   }
>   
>   int erdma_ceqs_init(struct erdma_dev *dev)
> diff --git a/drivers/infiniband/hw/erdma/erdma_main.c b/drivers/infiniband/hw/erdma/erdma_main.c
> index 472939172f0c..7080f8a71ec4 100644
> --- a/drivers/infiniband/hw/erdma/erdma_main.c
> +++ b/drivers/infiniband/hw/erdma/erdma_main.c
> @@ -178,16 +178,26 @@ static int erdma_device_init(struct erdma_dev *dev, struct pci_dev *pdev)
>   	if (!dev->resp_pool)
>   		return -ENOMEM;
>   
> +	dev->db_pool = dma_pool_create("erdma_db_pool", &pdev->dev,
> +				       ERDMA_DB_SIZE, ERDMA_DB_SIZE, 0);
> +	if (!dev->db_pool) {
> +		ret = -ENOMEM;
> +		goto destroy_resp_pool;
> +	}
> +
>   	ret = dma_set_mask_and_coherent(&pdev->dev,
>   					DMA_BIT_MASK(ERDMA_PCI_WIDTH));
>   	if (ret)
> -		goto destroy_pool;
> +		goto destroy_db_pool;
>   
>   	dma_set_max_seg_size(&pdev->dev, UINT_MAX);
>   
>   	return 0;
>   
> -destroy_pool:
> +destroy_db_pool:
> +	dma_pool_destroy(dev->db_pool);
> +
> +destroy_resp_pool:
>   	dma_pool_destroy(dev->resp_pool);
>   
>   	return ret;
> @@ -195,6 +205,7 @@ static int erdma_device_init(struct erdma_dev *dev, struct pci_dev *pdev)
>   
>   static void erdma_device_uninit(struct erdma_dev *dev)
>   {
> +	dma_pool_destroy(dev->db_pool);
>   	dma_pool_destroy(dev->resp_pool);
>   }
>   
> diff --git a/drivers/infiniband/hw/erdma/erdma_verbs.c b/drivers/infiniband/hw/erdma/erdma_verbs.c
> index 23dfc01603f8..b78ddca1483e 100644
> --- a/drivers/infiniband/hw/erdma/erdma_verbs.c
> +++ b/drivers/infiniband/hw/erdma/erdma_verbs.c
> @@ -76,10 +76,8 @@ static int create_qp_cmd(struct erdma_ucontext *uctx, struct erdma_qp *qp)
>   
>   		req.rq_buf_addr = qp->kern_qp.rq_buf_dma_addr;
>   		req.sq_buf_addr = qp->kern_qp.sq_buf_dma_addr;
> -		req.sq_db_info_dma_addr = qp->kern_qp.sq_buf_dma_addr +
> -					  (qp->attrs.sq_size << SQEBB_SHIFT);
> -		req.rq_db_info_dma_addr = qp->kern_qp.rq_buf_dma_addr +
> -					  (qp->attrs.rq_size << RQE_SHIFT);
> +		req.sq_db_info_dma_addr = qp->kern_qp.sq_db_info_dma_addr;
> +		req.rq_db_info_dma_addr = qp->kern_qp.rq_db_info_dma_addr;
>   	} else {
>   		user_qp = &qp->user_qp;
>   		req.sq_cqn_mtt_cfg = FIELD_PREP(
> @@ -209,8 +207,7 @@ static int create_cq_cmd(struct erdma_ucontext *uctx, struct erdma_cq *cq)
>   				       ERDMA_MR_MTT_0LEVEL);
>   
>   		req.first_page_offset = 0;
> -		req.cq_db_info_addr =
> -			cq->kern_cq.qbuf_dma_addr + (cq->depth << CQE_SHIFT);
> +		req.cq_db_info_addr = cq->kern_cq.db_record_dma_addr;
>   	} else {
>   		mem = &cq->user_cq.qbuf_mem;
>   		req.cfg0 |=
> @@ -482,16 +479,24 @@ static void free_kernel_qp(struct erdma_qp *qp)
>   	vfree(qp->kern_qp.rwr_tbl);
>   
>   	if (qp->kern_qp.sq_buf)
> -		dma_free_coherent(
> -			&dev->pdev->dev,
> -			WARPPED_BUFSIZE(qp->attrs.sq_size << SQEBB_SHIFT),
> -			qp->kern_qp.sq_buf, qp->kern_qp.sq_buf_dma_addr);
> +		dma_free_coherent(&dev->pdev->dev,
> +				  qp->attrs.sq_size << SQEBB_SHIFT,
> +				  qp->kern_qp.sq_buf,
> +				  qp->kern_qp.sq_buf_dma_addr);
> +
> +	if (qp->kern_qp.sq_db_info)
> +		dma_pool_free(dev->db_pool, qp->kern_qp.sq_db_info,
> +			      qp->kern_qp.sq_db_info_dma_addr);
>   
>   	if (qp->kern_qp.rq_buf)
> -		dma_free_coherent(
> -			&dev->pdev->dev,
> -			WARPPED_BUFSIZE(qp->attrs.rq_size << RQE_SHIFT),
> -			qp->kern_qp.rq_buf, qp->kern_qp.rq_buf_dma_addr);
> +		dma_free_coherent(&dev->pdev->dev,
> +				  qp->attrs.rq_size << RQE_SHIFT,
> +				  qp->kern_qp.rq_buf,
> +				  qp->kern_qp.rq_buf_dma_addr);
> +
> +	if (qp->kern_qp.rq_db_info)
> +		dma_pool_free(dev->db_pool, qp->kern_qp.rq_db_info,
> +			      qp->kern_qp.rq_db_info_dma_addr);
>   }
>   
>   static int init_kernel_qp(struct erdma_dev *dev, struct erdma_qp *qp,
> @@ -516,20 +521,27 @@ static int init_kernel_qp(struct erdma_dev *dev, struct erdma_qp *qp,
>   	if (!kqp->swr_tbl || !kqp->rwr_tbl)
>   		goto err_out;
>   
> -	size = (qp->attrs.sq_size << SQEBB_SHIFT) + ERDMA_EXTRA_BUFFER_SIZE;
> +	size = qp->attrs.sq_size << SQEBB_SHIFT;
>   	kqp->sq_buf = dma_alloc_coherent(&dev->pdev->dev, size,
>   					 &kqp->sq_buf_dma_addr, GFP_KERNEL);
>   	if (!kqp->sq_buf)
>   		goto err_out;
>   
> -	size = (qp->attrs.rq_size << RQE_SHIFT) + ERDMA_EXTRA_BUFFER_SIZE;
> +	kqp->sq_db_info = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
> +					  &kqp->sq_db_info_dma_addr);
> +	if (!kqp->sq_db_info)
> +		goto err_out;
> +
> +	size = qp->attrs.rq_size << RQE_SHIFT;
>   	kqp->rq_buf = dma_alloc_coherent(&dev->pdev->dev, size,
>   					 &kqp->rq_buf_dma_addr, GFP_KERNEL);
>   	if (!kqp->rq_buf)
>   		goto err_out;
>   
> -	kqp->sq_db_info = kqp->sq_buf + (qp->attrs.sq_size << SQEBB_SHIFT);
> -	kqp->rq_db_info = kqp->rq_buf + (qp->attrs.rq_size << RQE_SHIFT);
> +	kqp->rq_db_info = dma_pool_zalloc(dev->db_pool, GFP_KERNEL,
> +					  &kqp->rq_db_info_dma_addr);
> +	if (!kqp->rq_db_info)
> +		goto err_out;
>   
>   	return 0;
>   
> @@ -1237,9 +1249,10 @@ int erdma_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata)
>   		return err;
>   
>   	if (rdma_is_kernel_res(&cq->ibcq.res)) {
> -		dma_free_coherent(&dev->pdev->dev,
> -				  WARPPED_BUFSIZE(cq->depth << CQE_SHIFT),
> +		dma_free_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT,
>   				  cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr);
> +		dma_pool_free(dev->db_pool, cq->kern_cq.db_record,
> +			      cq->kern_cq.db_record_dma_addr);
>   	} else {
>   		erdma_unmap_user_dbrecords(ctx, &cq->user_cq.user_dbr_page);
>   		put_mtt_entries(dev, &cq->user_cq.qbuf_mem);
> @@ -1279,16 +1292,7 @@ int erdma_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata)
>   	wait_for_completion(&qp->safe_free);
>   
>   	if (rdma_is_kernel_res(&qp->ibqp.res)) {
> -		vfree(qp->kern_qp.swr_tbl);
> -		vfree(qp->kern_qp.rwr_tbl);
> -		dma_free_coherent(
> -			&dev->pdev->dev,
> -			WARPPED_BUFSIZE(qp->attrs.rq_size << RQE_SHIFT),
> -			qp->kern_qp.rq_buf, qp->kern_qp.rq_buf_dma_addr);
> -		dma_free_coherent(
> -			&dev->pdev->dev,
> -			WARPPED_BUFSIZE(qp->attrs.sq_size << SQEBB_SHIFT),
> -			qp->kern_qp.sq_buf, qp->kern_qp.sq_buf_dma_addr);
> +		free_kernel_qp(qp);
>   	} else {
>   		put_mtt_entries(dev, &qp->user_qp.sq_mem);
>   		put_mtt_entries(dev, &qp->user_qp.rq_mem);
> @@ -1600,19 +1604,27 @@ static int erdma_init_kernel_cq(struct erdma_cq *cq)
>   	struct erdma_dev *dev = to_edev(cq->ibcq.device);
>   
>   	cq->kern_cq.qbuf =
> -		dma_alloc_coherent(&dev->pdev->dev,
> -				   WARPPED_BUFSIZE(cq->depth << CQE_SHIFT),
> +		dma_alloc_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT,
>   				   &cq->kern_cq.qbuf_dma_addr, GFP_KERNEL);
>   	if (!cq->kern_cq.qbuf)
>   		return -ENOMEM;
>   
> -	cq->kern_cq.db_record =
> -		(u64 *)(cq->kern_cq.qbuf + (cq->depth << CQE_SHIFT));
> +	cq->kern_cq.db_record = dma_pool_zalloc(
> +		dev->db_pool, GFP_KERNEL, &cq->kern_cq.db_record_dma_addr);
> +	if (!cq->kern_cq.db_record)
> +		goto err_out;
> +
>   	spin_lock_init(&cq->kern_cq.lock);
>   	/* use default cqdb addr */
>   	cq->kern_cq.db = dev->func_bar + ERDMA_BAR_CQDB_SPACE_OFFSET;
>   
>   	return 0;
> +
> +err_out:
> +	dma_free_coherent(&dev->pdev->dev, cq->depth << CQE_SHIFT,
> +			  cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr);
> +
> +	return -ENOMEM;
>   }
>   
>   int erdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
> @@ -1676,9 +1688,10 @@ int erdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
>   		erdma_unmap_user_dbrecords(ctx, &cq->user_cq.user_dbr_page);
>   		put_mtt_entries(dev, &cq->user_cq.qbuf_mem);
>   	} else {
> -		dma_free_coherent(&dev->pdev->dev,
> -				  WARPPED_BUFSIZE(depth << CQE_SHIFT),
> +		dma_free_coherent(&dev->pdev->dev, depth << CQE_SHIFT,
>   				  cq->kern_cq.qbuf, cq->kern_cq.qbuf_dma_addr);
> +		dma_pool_free(dev->db_pool, cq->kern_cq.db_record,
> +			      cq->kern_cq.db_record_dma_addr);
>   	}
>   
>   err_out_xa:
> diff --git a/drivers/infiniband/hw/erdma/erdma_verbs.h b/drivers/infiniband/hw/erdma/erdma_verbs.h
> index db6018529ccc..b02ffdc8c811 100644
> --- a/drivers/infiniband/hw/erdma/erdma_verbs.h
> +++ b/drivers/infiniband/hw/erdma/erdma_verbs.h
> @@ -170,6 +170,9 @@ struct erdma_kqp {
>   	void *sq_db_info;
>   	void *rq_db_info;
>   
> +	dma_addr_t sq_db_info_dma_addr;
> +	dma_addr_t rq_db_info_dma_addr;
> +
>   	u8 sig_all;
>   };
>   
> @@ -247,6 +250,7 @@ struct erdma_kcq_info {
>   	spinlock_t lock;
>   	u8 __iomem *db;
>   	u64 *db_record;
> +	dma_addr_t db_record_dma_addr;
>   };
>   
>   struct erdma_ucq_info {


  reply	other threads:[~2024-03-13 16:20 UTC|newest]

Thread overview: 11+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-03-11 11:38 [PATCH for-next 0/3] RDMA/erdma: A series of fixes for the erdma driver Boshi Yu
2024-03-11 11:38 ` [PATCH for-next 1/3] RDMA/erdma: Allocate doorbell records from dma pool Boshi Yu
2024-03-13 16:20   ` Zhu Yanjun [this message]
2024-03-14 10:23     ` Zhu Yanjun
2024-03-11 11:38 ` [PATCH for-next 2/3] RDMA/erdma: Unify the names related to doorbell records Boshi Yu
2024-03-11 11:38 ` [PATCH for-next 3/3] RDMA/erdma: Remove unnecessary __GFP_ZERO flag Boshi Yu
2024-03-12 10:53   ` Leon Romanovsky
2024-03-13  2:40     ` Boshi Yu
2024-03-13  9:31       ` Leon Romanovsky
2024-03-12 10:54 ` [PATCH for-next 0/3] RDMA/erdma: A series of fixes for the erdma driver Leon Romanovsky
2024-04-01 11:46 ` Leon Romanovsky

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=09da3e92-3da5-405a-9f55-60533e966f09@linux.dev \
    --to=zyjzyj2000@gmail.com \
    --cc=KaiShen@linux.alibaba.com \
    --cc=boshiyu@alibaba-inc.com \
    --cc=chengyou@linux.alibaba.com \
    --cc=jgg@ziepe.ca \
    --cc=leon@kernel.org \
    --cc=linux-rdma@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox