From: Yuval Shaia <yuval.shaia@oracle.com>
To: Kamal Heib <kamalheib1@gmail.com>
Cc: qemu-devel@nongnu.org
Subject: Re: [Qemu-devel] [PATCH 2/4] hw/rdma: Add support for managing SRQ resource
Date: Wed, 3 Apr 2019 21:10:47 +0300 [thread overview]
Message-ID: <20190403181047.GB3446@lap1> (raw)
In-Reply-To: <20190403113343.26384-3-kamalheib1@gmail.com>
On Wed, Apr 03, 2019 at 02:33:41PM +0300, Kamal Heib wrote:
> Adding the required functions and definitions for support managing the
> shared receive queues (SRQs).
>
> Signed-off-by: Kamal Heib <kamalheib1@gmail.com>
> ---
> hw/rdma/rdma_rm.c | 93 ++++++++++++++++++++++++++++++++++++++++++
> hw/rdma/rdma_rm.h | 10 +++++
> hw/rdma/rdma_rm_defs.h | 8 ++++
> 3 files changed, 111 insertions(+)
>
> diff --git a/hw/rdma/rdma_rm.c b/hw/rdma/rdma_rm.c
> index b683506b8616..c4fb140dcd96 100644
> --- a/hw/rdma/rdma_rm.c
> +++ b/hw/rdma/rdma_rm.c
> @@ -544,6 +544,96 @@ void rdma_rm_dealloc_qp(RdmaDeviceResources *dev_res, uint32_t qp_handle)
> rdma_res_tbl_dealloc(&dev_res->qp_tbl, qp->qpn);
> }
>
> +RdmaRmSRQ *rdma_rm_get_srq(RdmaDeviceResources *dev_res, uint32_t srq_handle)
> +{
> + return rdma_res_tbl_get(&dev_res->srq_tbl, srq_handle);
> +}
> +
> +int rdma_rm_alloc_srq(RdmaDeviceResources *dev_res, uint32_t pd_handle,
> + uint32_t max_wr, uint32_t max_sge, uint32_t srq_limit,
> + uint32_t *srq_handle, void *opaque)
> +{
> + RdmaRmSRQ *srq;
> + RdmaRmPD *pd;
> + int rc;
> +
> + pd = rdma_rm_get_pd(dev_res, pd_handle);
> + if (!pd) {
> + return -EINVAL;
> + }
> +
> + srq = rdma_res_tbl_alloc(&dev_res->srq_tbl, srq_handle);
> + if (!srq) {
> + return -ENOMEM;
> + }
> +
> + rc = rdma_backend_create_srq(&srq->backend_srq, &pd->backend_pd,
> + max_wr, max_sge, srq_limit);
> + if (rc) {
> + rc = -EIO;
> + goto out_dealloc_srq;
> + }
> +
> + srq->opaque = opaque;
> +
> + return 0;
> +
> +out_dealloc_srq:
> + rdma_res_tbl_dealloc(&dev_res->srq_tbl, *srq_handle);
> +
> + return rc;
> +}
> +
> +int rdma_rm_query_srq(RdmaDeviceResources *dev_res, uint32_t srq_handle,
> + struct ibv_srq_attr *srq_attr)
> +{
> + RdmaRmSRQ *srq;
> +
> + srq = rdma_rm_get_srq(dev_res, srq_handle);
> + if (!srq) {
> + return -EINVAL;
> + }
> +
> + return rdma_backend_query_srq(&srq->backend_srq, srq_attr);
> +}
> +
> +int rdma_rm_modify_srq(RdmaDeviceResources *dev_res, uint32_t srq_handle,
> + struct ibv_srq_attr *srq_attr, int srq_attr_mask)
> +{
> + RdmaRmSRQ *srq;
> +
> + srq = rdma_rm_get_srq(dev_res, srq_handle);
> + if (!srq) {
> + return -EINVAL;
> + }
> +
> + if ((srq_attr_mask & IBV_SRQ_LIMIT) &&
> + (srq_attr->srq_limit == 0)) {
> + return -EINVAL;
> + }
> +
> + if ((srq_attr_mask & IBV_SRQ_MAX_WR) &&
> + (srq_attr->max_wr == 0)) {
> + return -EINVAL;
> + }
> +
> + return rdma_backend_modify_srq(&srq->backend_srq, srq_attr,
> + srq_attr_mask);
> +}
> +
> +void rdma_rm_dealloc_srq(RdmaDeviceResources *dev_res, uint32_t srq_handle)
> +{
> + RdmaRmSRQ *srq;
> +
> + srq = rdma_rm_get_srq(dev_res, srq_handle);
> + if (!srq) {
> + return;
> + }
> +
> + rdma_backend_destroy_srq(&srq->backend_srq, dev_res);
> + rdma_res_tbl_dealloc(&dev_res->srq_tbl, srq_handle);
> +}
> +
> void *rdma_rm_get_cqe_ctx(RdmaDeviceResources *dev_res, uint32_t cqe_ctx_id)
> {
> void **cqe_ctx;
> @@ -673,6 +763,8 @@ int rdma_rm_init(RdmaDeviceResources *dev_res, struct ibv_device_attr *dev_attr)
> res_tbl_init("CQE_CTX", &dev_res->cqe_ctx_tbl, dev_attr->max_qp *
> dev_attr->max_qp_wr, sizeof(void *));
> res_tbl_init("UC", &dev_res->uc_tbl, MAX_UCS, sizeof(RdmaRmUC));
> + res_tbl_init("SRQ", &dev_res->srq_tbl, dev_attr->max_srq,
> + sizeof(RdmaRmSRQ));
>
> init_ports(dev_res);
>
> @@ -691,6 +783,7 @@ void rdma_rm_fini(RdmaDeviceResources *dev_res, RdmaBackendDev *backend_dev,
>
> fini_ports(dev_res, backend_dev, ifname);
>
> + res_tbl_free(&dev_res->srq_tbl);
> res_tbl_free(&dev_res->uc_tbl);
> res_tbl_free(&dev_res->cqe_ctx_tbl);
> res_tbl_free(&dev_res->qp_tbl);
> diff --git a/hw/rdma/rdma_rm.h b/hw/rdma/rdma_rm.h
> index 4f03f9b8c5f1..e88ab95e264b 100644
> --- a/hw/rdma/rdma_rm.h
> +++ b/hw/rdma/rdma_rm.h
> @@ -65,6 +65,16 @@ int rdma_rm_query_qp(RdmaDeviceResources *dev_res, RdmaBackendDev *backend_dev,
> int attr_mask, struct ibv_qp_init_attr *init_attr);
> void rdma_rm_dealloc_qp(RdmaDeviceResources *dev_res, uint32_t qp_handle);
>
> +RdmaRmSRQ *rdma_rm_get_srq(RdmaDeviceResources *dev_res, uint32_t srq_handle);
> +int rdma_rm_alloc_srq(RdmaDeviceResources *dev_res, uint32_t pd_handle,
> + uint32_t max_wr, uint32_t max_sge, uint32_t srq_limit,
> + uint32_t *srq_handle, void *opaque);
> +int rdma_rm_query_srq(RdmaDeviceResources *dev_res, uint32_t srq_handle,
> + struct ibv_srq_attr *srq_attr);
> +int rdma_rm_modify_srq(RdmaDeviceResources *dev_res, uint32_t srq_handle,
> + struct ibv_srq_attr *srq_attr, int srq_attr_mask);
> +void rdma_rm_dealloc_srq(RdmaDeviceResources *dev_res, uint32_t srq_handle);
> +
> int rdma_rm_alloc_cqe_ctx(RdmaDeviceResources *dev_res, uint32_t *cqe_ctx_id,
> void *ctx);
> void *rdma_rm_get_cqe_ctx(RdmaDeviceResources *dev_res, uint32_t cqe_ctx_id);
> diff --git a/hw/rdma/rdma_rm_defs.h b/hw/rdma/rdma_rm_defs.h
> index e774af528022..7bdd9f291f9f 100644
> --- a/hw/rdma/rdma_rm_defs.h
> +++ b/hw/rdma/rdma_rm_defs.h
> @@ -33,6 +33,7 @@
> #define MAX_QP_RD_ATOM 16
> #define MAX_QP_INIT_RD_ATOM 16
> #define MAX_AH 64
> +#define MAX_SRQ 512
>
> #define MAX_RM_TBL_NAME 16
> #define MAX_CONSEQ_EMPTY_POLL_CQ 4096 /* considered as error above this */
> @@ -89,6 +90,12 @@ typedef struct RdmaRmQP {
> enum ibv_qp_state qp_state;
> } RdmaRmQP;
>
> +typedef struct RdmaRmSRQ {
> + RdmaBackendSRQ backend_srq;
> + uint32_t recv_cq_handle;
> + void *opaque;
> +} RdmaRmSRQ;
> +
> typedef struct RdmaRmGid {
> union ibv_gid gid;
> int backend_gid_index;
> @@ -129,6 +136,7 @@ struct RdmaDeviceResources {
> RdmaRmResTbl qp_tbl;
> RdmaRmResTbl cq_tbl;
> RdmaRmResTbl cqe_ctx_tbl;
> + RdmaRmResTbl srq_tbl;
> GHashTable *qp_hash; /* Keeps mapping between real and emulated */
> QemuMutex lock;
> RdmaRmStats stats;
For some reason v3 is omitted from subject, weird.
Anyway, looks like you took care for the comment raised.
With that - patch lgtm.
Reviewed-by: Yuval Shaia <yuval.shaia@oracle.com>
> --
> 2.20.1
>
>
next prev parent reply other threads:[~2019-04-03 18:11 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-04-03 11:33 [Qemu-devel] [PATCH v3 0/4] pvrdma: Add support for SRQ Kamal Heib
2019-04-03 11:33 ` [Qemu-devel] [PATCH v3 1/4] hw/rdma: Add SRQ support to backend layer Kamal Heib
2019-04-03 18:05 ` Yuval Shaia
2019-04-07 8:13 ` Kamal Heib
2019-04-07 8:13 ` Kamal Heib
2019-04-07 9:13 ` Yuval Shaia
2019-04-07 9:13 ` Yuval Shaia
2019-04-03 11:33 ` [Qemu-devel] [PATCH 2/4] hw/rdma: Add support for managing SRQ resource Kamal Heib
2019-04-03 18:10 ` Yuval Shaia [this message]
2019-04-03 11:33 ` [Qemu-devel] [PATCH v3 3/4] hw/rdma: Modify create/destroy QP to support SRQ Kamal Heib
2019-04-03 18:15 ` Yuval Shaia
2019-04-03 11:33 ` [Qemu-devel] [PATCH v3 4/4] hw/pvrdma: Add support for SRQ Kamal Heib
2019-04-03 18:16 ` Yuval Shaia
2019-04-03 18:19 ` [Qemu-devel] [PATCH v3 0/4] pvrdma: " Yuval Shaia
2019-04-07 9:15 ` Yuval Shaia
2019-04-07 9:15 ` Yuval Shaia
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20190403181047.GB3446@lap1 \
--to=yuval.shaia@oracle.com \
--cc=kamalheib1@gmail.com \
--cc=qemu-devel@nongnu.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).