From: Leon Romanovsky <leon@kernel.org>
To: Yuval Basson <ybason@marvell.com>
Cc: dledford@redhat.com, jgg@ziepe.ca, linux-rdma@vger.kernel.org,
Michal Kalderon <mkalderon@marvell.com>
Subject: Re: [PATCH v2 rdma-next] RDMA/qedr: SRQ's bug fixes
Date: Wed, 8 Jul 2020 08:39:00 +0300 [thread overview]
Message-ID: <20200708053900.GQ207186@unreal> (raw)
In-Reply-To: <20200707171847.29352-1-ybason@marvell.com>
On Tue, Jul 07, 2020 at 08:18:47PM +0300, Yuval Basson wrote:
> QP's with the same SRQ, working on different CQs and running in parallel
> on different CPUs could lead to a race when maintaining the SRQ consumer
> count, and leads to FW running out of SRQs. Update the consumer atomically.
> Make sure the wqe_prod is updated after the sge_prod due to FW
> requirements.
>
> Fixes: 3491c9e799fb9 ("RDMA/qedr: Add support for kernel mode SRQ's")
> Signed-off-by: Michal Kalderon <mkalderon@marvell.com>
> Signed-off-by: Yuval Basson <ybason@marvell.com>
>
> Changes in v2:
"Changes ..." should come after "---".
Thanks
> Replace barrier() with dma_wmb()
> Remove redundant dma_wmb()
> ---
> drivers/infiniband/hw/qedr/qedr.h | 4 ++--
> drivers/infiniband/hw/qedr/verbs.c | 23 +++++++++++------------
> 2 files changed, 13 insertions(+), 14 deletions(-)
>
> diff --git a/drivers/infiniband/hw/qedr/qedr.h b/drivers/infiniband/hw/qedr/qedr.h
> index fdf90ec..aa33202 100644
> --- a/drivers/infiniband/hw/qedr/qedr.h
> +++ b/drivers/infiniband/hw/qedr/qedr.h
> @@ -344,10 +344,10 @@ struct qedr_srq_hwq_info {
> u32 wqe_prod;
> u32 sge_prod;
> u32 wr_prod_cnt;
> - u32 wr_cons_cnt;
> + atomic_t wr_cons_cnt;
> u32 num_elems;
>
> - u32 *virt_prod_pair_addr;
> + struct rdma_srq_producers *virt_prod_pair_addr;
> dma_addr_t phy_prod_pair_addr;
> };
>
> diff --git a/drivers/infiniband/hw/qedr/verbs.c b/drivers/infiniband/hw/qedr/verbs.c
> index 9b9e802..444537b 100644
> --- a/drivers/infiniband/hw/qedr/verbs.c
> +++ b/drivers/infiniband/hw/qedr/verbs.c
> @@ -1510,6 +1510,7 @@ int qedr_create_srq(struct ib_srq *ibsrq, struct ib_srq_init_attr *init_attr,
> srq->dev = dev;
> hw_srq = &srq->hw_srq;
> spin_lock_init(&srq->lock);
> + atomic_set(&hw_srq->wr_cons_cnt, 0);
>
> hw_srq->max_wr = init_attr->attr.max_wr;
> hw_srq->max_sges = init_attr->attr.max_sge;
> @@ -3686,7 +3687,7 @@ static u32 qedr_srq_elem_left(struct qedr_srq_hwq_info *hw_srq)
> * count and consumer count and subtract it from max
> * work request supported so that we get elements left.
> */
> - used = hw_srq->wr_prod_cnt - hw_srq->wr_cons_cnt;
> + used = hw_srq->wr_prod_cnt - (u32)atomic_read(&hw_srq->wr_cons_cnt);
>
> return hw_srq->max_wr - used;
> }
> @@ -3701,7 +3702,6 @@ int qedr_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
> unsigned long flags;
> int status = 0;
> u32 num_sge;
> - u32 offset;
>
> spin_lock_irqsave(&srq->lock, flags);
>
> @@ -3714,7 +3714,8 @@ int qedr_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
> if (!qedr_srq_elem_left(hw_srq) ||
> wr->num_sge > srq->hw_srq.max_sges) {
> DP_ERR(dev, "Can't post WR (%d,%d) || (%d > %d)\n",
> - hw_srq->wr_prod_cnt, hw_srq->wr_cons_cnt,
> + hw_srq->wr_prod_cnt,
> + atomic_read(&hw_srq->wr_cons_cnt),
> wr->num_sge, srq->hw_srq.max_sges);
> status = -ENOMEM;
> *bad_wr = wr;
> @@ -3748,22 +3749,20 @@ int qedr_post_srq_recv(struct ib_srq *ibsrq, const struct ib_recv_wr *wr,
> hw_srq->sge_prod++;
> }
>
> - /* Flush WQE and SGE information before
> + /* Update WQE and SGE information before
> * updating producer.
> */
> - wmb();
> + dma_wmb();
>
> /* SRQ producer is 8 bytes. Need to update SGE producer index
> * in first 4 bytes and need to update WQE producer in
> * next 4 bytes.
> */
> - *srq->hw_srq.virt_prod_pair_addr = hw_srq->sge_prod;
> - offset = offsetof(struct rdma_srq_producers, wqe_prod);
> - *((u8 *)srq->hw_srq.virt_prod_pair_addr + offset) =
> - hw_srq->wqe_prod;
> + srq->hw_srq.virt_prod_pair_addr->sge_prod = hw_srq->sge_prod;
> + /* Make sure sge producer is updated first */
> + dma_wmb();
> + srq->hw_srq.virt_prod_pair_addr->wqe_prod = hw_srq->wqe_prod;
>
> - /* Flush producer after updating it. */
> - wmb();
> wr = wr->next;
> }
>
> @@ -4182,7 +4181,7 @@ static int process_resp_one_srq(struct qedr_dev *dev, struct qedr_qp *qp,
> } else {
> __process_resp_one(dev, qp, cq, wc, resp, wr_id);
> }
> - srq->hw_srq.wr_cons_cnt++;
> + atomic_inc(&srq->hw_srq.wr_cons_cnt);
>
> return 1;
> }
> --
> 1.8.3.1
>
prev parent reply other threads:[~2020-07-08 5:39 UTC|newest]
Thread overview: 2+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-07-07 17:18 [PATCH v2 rdma-next] RDMA/qedr: SRQ's bug fixes Yuval Basson
2020-07-08 5:39 ` Leon Romanovsky [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200708053900.GQ207186@unreal \
--to=leon@kernel.org \
--cc=dledford@redhat.com \
--cc=jgg@ziepe.ca \
--cc=linux-rdma@vger.kernel.org \
--cc=mkalderon@marvell.com \
--cc=ybason@marvell.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).