From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from linux.microsoft.com (linux.microsoft.com [13.77.154.182]) by smtp.subspace.kernel.org (Postfix) with ESMTP id E10123AA504; Wed, 25 Mar 2026 10:18:43 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=13.77.154.182 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1774433926; cv=none; b=TfoEWQBWI68XnjuzZbZmdWs4HoAtUxolFKOAbj4MbHAp6hSqr2ifY4XEfOEULpR/n2uf/9BuQ1EdP3fg70QRJqXU0Qo8Pl2R54vg4Eyky9KlzfHXyYrQF2fmFDEL8GSH8d3ocwXSMY20gWoE19voEqUJcUYjDarakCKohL0s/rY= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1774433926; c=relaxed/simple; bh=y0Lr6Wd3WT+GVvxdWrAd62DB5wyNDMGayE2+HCHah28=; h=From:To:Cc:Subject:Date:Message-ID:MIME-Version; b=uYjEftvdx7ku1M2fZMnSsg0zl7cV7NDMhOW5H8LTSGl40pIHJy0XSoNBPWAy9JRfjC5lUKojIMKXUtF2GQ4yGkmvtyGzevs9QQz9EJzi901CcpXP0VnIIZpE70y/CMsbOxkwfhMMPtBxlpcdviRkulUnO81MgxCB7JEKH53BTKs= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.microsoft.com; spf=pass smtp.mailfrom=linux.microsoft.com; dkim=pass (1024-bit key) header.d=linux.microsoft.com header.i=@linux.microsoft.com header.b=RUluTI+p; arc=none smtp.client-ip=13.77.154.182 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=linux.microsoft.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=linux.microsoft.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=linux.microsoft.com header.i=@linux.microsoft.com header.b="RUluTI+p" Received: by linux.microsoft.com (Postfix, from userid 1186) id CB23520B6F08; Wed, 25 Mar 2026 03:18:43 -0700 (PDT) DKIM-Filter: OpenDKIM Filter v2.11.0 linux.microsoft.com CB23520B6F08 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.microsoft.com; s=default; t=1774433923; bh=esMll0L2cWbLndpQ1iAQHxHRnpr6qDp7oqCB0eDx1M0=; h=From:To:Cc:Subject:Date:From; b=RUluTI+pKUoK/G7bQYOzeXTBY1jIstdR9+fXZXipYTwzi1ZfzmUwIDxh2n7pFfcmD /BMy1VyC7j0iMPWaO9qQWEPYjsTz/5P4eyX/ckzAyoyBABtnCPqEsdm9y39aZqJgY4 NDra36hf6djkESz+C++/suVKdtiziqjYqYqAXwgg= From: Konstantin Taranov To: kotaranov@microsoft.com, shirazsaleem@microsoft.com, longli@microsoft.com, jgg@ziepe.ca, leon@kernel.org Cc: linux-rdma@vger.kernel.org, linux-kernel@vger.kernel.org Subject: [PATCH rdma-next v2 1/1] RDMA/mana_ib: UC QP support for UAPI Date: Wed, 25 Mar 2026 03:18:43 -0700 Message-ID: <20260325101843.1893866-1-kotaranov@linux.microsoft.com> X-Mailer: git-send-email 2.43.7 Precedence: bulk X-Mailing-List: linux-rdma@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: 8bit From: Konstantin Taranov Implement UC QP creation in the RNIC HW for user API. An UC QP is exposed as three work queues: send, receive, and memory management. The latter is used for bind and invalidate WQEs to support memory windows. Signed-off-by: Konstantin Taranov --- v2: removed udata check and removed enum from mana-abi.h drivers/infiniband/hw/mana/main.c | 41 ++++++++++++- drivers/infiniband/hw/mana/mana_ib.h | 41 ++++++++++++- drivers/infiniband/hw/mana/qp.c | 90 ++++++++++++++++++++++++++-- include/uapi/rdma/mana-abi.h | 11 ++++ 4 files changed, 174 insertions(+), 9 deletions(-) diff --git a/drivers/infiniband/hw/mana/main.c b/drivers/infiniband/hw/mana/main.c index ac5e75dd3..42567c30e 100644 --- a/drivers/infiniband/hw/mana/main.c +++ b/drivers/infiniband/hw/mana/main.c @@ -978,7 +978,46 @@ int mana_ib_gd_create_rc_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp, return 0; } -int mana_ib_gd_destroy_rc_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp) +int mana_ib_gd_create_uc_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp, + struct ib_qp_init_attr *attr, u32 doorbell, u64 flags) +{ + struct mana_ib_cq *send_cq = container_of(qp->ibqp.send_cq, struct mana_ib_cq, ibcq); + struct mana_ib_cq *recv_cq = container_of(qp->ibqp.recv_cq, struct mana_ib_cq, ibcq); + struct mana_ib_pd *pd = container_of(qp->ibqp.pd, struct mana_ib_pd, ibpd); + struct gdma_context *gc = mdev_to_gc(mdev); + struct mana_rnic_create_uc_qp_resp resp = {}; + struct mana_rnic_create_uc_qp_req req = {}; + int err, i; + + mana_gd_init_req_hdr(&req.hdr, MANA_IB_CREATE_UC_QP, sizeof(req), sizeof(resp)); + req.hdr.dev_id = mdev->gdma_dev->dev_id; + req.adapter = mdev->adapter_handle; + req.pd_handle = pd->pd_handle; + req.send_cq_handle = send_cq->cq_handle; + req.recv_cq_handle = recv_cq->cq_handle; + for (i = 0; i < MANA_UC_QUEUE_TYPE_MAX; i++) + req.dma_region[i] = qp->uc_qp.queues[i].gdma_region; + req.doorbell_page = doorbell; + req.max_send_wr = attr->cap.max_send_wr; + req.max_recv_wr = attr->cap.max_recv_wr; + req.max_send_sge = attr->cap.max_send_sge; + req.max_recv_sge = attr->cap.max_recv_sge; + req.flags = flags; + + err = mana_gd_send_request(gc, sizeof(req), &req, sizeof(resp), &resp); + if (err) + return err; + + qp->qp_handle = resp.qp_handle; + for (i = 0; i < MANA_UC_QUEUE_TYPE_MAX; i++) { + qp->uc_qp.queues[i].id = resp.queue_ids[i]; + /* The GDMA regions are now owned by the RNIC QP handle */ + qp->uc_qp.queues[i].gdma_region = GDMA_INVALID_DMA_REGION; + } + return 0; +} + +int mana_ib_gd_destroy_rnic_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp) { struct mana_rnic_destroy_rc_qp_resp resp = {0}; struct mana_rnic_destroy_rc_qp_req req = {0}; diff --git a/drivers/infiniband/hw/mana/mana_ib.h b/drivers/infiniband/hw/mana/mana_ib.h index a7c8c0fd7..37edd11c9 100644 --- a/drivers/infiniband/hw/mana/mana_ib.h +++ b/drivers/infiniband/hw/mana/mana_ib.h @@ -161,6 +161,17 @@ struct mana_ib_rc_qp { struct mana_ib_queue queues[MANA_RC_QUEUE_TYPE_MAX]; }; +enum mana_uc_queue_type { + MANA_UC_SEND_QUEUE_REQUESTER = 0, + MANA_UC_RECV_QUEUE_RESPONDER, + MANA_UC_SEND_QUEUE_MMQ, + MANA_UC_QUEUE_TYPE_MAX, +}; + +struct mana_ib_uc_qp { + struct mana_ib_queue queues[MANA_UC_QUEUE_TYPE_MAX]; +}; + enum mana_ud_queue_type { MANA_UD_SEND_QUEUE = 0, MANA_UD_RECV_QUEUE, @@ -179,6 +190,7 @@ struct mana_ib_qp { union { struct mana_ib_queue raw_sq; struct mana_ib_rc_qp rc_qp; + struct mana_ib_uc_qp uc_qp; struct mana_ib_ud_qp ud_qp; }; @@ -216,6 +228,7 @@ enum mana_ib_command_code { MANA_IB_CREATE_RC_QP = 0x3000a, MANA_IB_DESTROY_RC_QP = 0x3000b, MANA_IB_SET_QP_STATE = 0x3000d, + MANA_IB_CREATE_UC_QP = 0x30020, MANA_IB_QUERY_VF_COUNTERS = 0x30022, MANA_IB_QUERY_DEVICE_COUNTERS = 0x30023, }; @@ -375,6 +388,29 @@ struct mana_rnic_destroy_rc_qp_resp { struct gdma_resp_hdr hdr; }; /* HW Data */ +struct mana_rnic_create_uc_qp_req { + struct gdma_req_hdr hdr; + mana_handle_t adapter; + mana_handle_t pd_handle; + mana_handle_t send_cq_handle; + mana_handle_t recv_cq_handle; + u64 dma_region[MANA_UC_QUEUE_TYPE_MAX]; + u64 flags; + u32 doorbell_page; + u32 max_send_wr; + u32 max_recv_wr; + u32 max_send_sge; + u32 max_recv_sge; + u32 reserved; +}; /* HW Data */ + +struct mana_rnic_create_uc_qp_resp { + struct gdma_resp_hdr hdr; + mana_handle_t qp_handle; + u32 queue_ids[MANA_UC_QUEUE_TYPE_MAX]; + u32 reserved; +}; /* HW Data*/ + struct mana_rnic_create_udqp_req { struct gdma_req_hdr hdr; mana_handle_t adapter; @@ -717,8 +753,9 @@ int mana_ib_gd_destroy_cq(struct mana_ib_dev *mdev, struct mana_ib_cq *cq); int mana_ib_gd_create_rc_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp, struct ib_qp_init_attr *attr, u32 doorbell, u64 flags); -int mana_ib_gd_destroy_rc_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp); - +int mana_ib_gd_destroy_rnic_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp); +int mana_ib_gd_create_uc_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp, + struct ib_qp_init_attr *attr, u32 doorbell, u64 flags); int mana_ib_gd_create_ud_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp, struct ib_qp_init_attr *attr, u32 doorbell, u32 type); int mana_ib_gd_destroy_ud_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp); diff --git a/drivers/infiniband/hw/mana/qp.c b/drivers/infiniband/hw/mana/qp.c index f3bb1edc7..6c6230b59 100644 --- a/drivers/infiniband/hw/mana/qp.c +++ b/drivers/infiniband/hw/mana/qp.c @@ -427,13 +427,13 @@ static enum gdma_queue_type mana_ib_queue_type(struct ib_qp_init_attr *attr, u32 return type; } -static int mana_table_store_rc_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp) +static int mana_table_store_rnic_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp) { return xa_insert_irq(&mdev->qp_table_wq, qp->ibqp.qp_num, qp, GFP_KERNEL); } -static void mana_table_remove_rc_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp) +static void mana_table_remove_rnic_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp) { xa_erase_irq(&mdev->qp_table_wq, qp->ibqp.qp_num); } @@ -475,7 +475,8 @@ static int mana_table_store_qp(struct mana_ib_dev *mdev, struct mana_ib_qp *qp) switch (qp->ibqp.qp_type) { case IB_QPT_RC: - return mana_table_store_rc_qp(mdev, qp); + case IB_QPT_UC: + return mana_table_store_rnic_qp(mdev, qp); case IB_QPT_UD: case IB_QPT_GSI: return mana_table_store_ud_qp(mdev, qp); @@ -492,7 +493,8 @@ static void mana_table_remove_qp(struct mana_ib_dev *mdev, { switch (qp->ibqp.qp_type) { case IB_QPT_RC: - mana_table_remove_rc_qp(mdev, qp); + case IB_QPT_UC: + mana_table_remove_rnic_qp(mdev, qp); break; case IB_QPT_UD: case IB_QPT_GSI: @@ -576,13 +578,68 @@ static int mana_ib_create_rc_qp(struct ib_qp *ibqp, struct ib_pd *ibpd, return 0; destroy_qp: - mana_ib_gd_destroy_rc_qp(mdev, qp); + mana_ib_gd_destroy_rnic_qp(mdev, qp); destroy_queues: while (i-- > 0) mana_ib_destroy_queue(mdev, &qp->rc_qp.queues[i]); return err; } +static int mana_ib_create_uc_qp(struct ib_qp *ibqp, struct ib_pd *ibpd, + struct ib_qp_init_attr *attr, struct ib_udata *udata) +{ + struct mana_ib_dev *mdev = container_of(ibpd->device, struct mana_ib_dev, ib_dev); + struct mana_ib_qp *qp = container_of(ibqp, struct mana_ib_qp, ibqp); + struct mana_ib_create_uc_qp_resp resp = {}; + struct mana_ib_ucontext *mana_ucontext; + struct mana_ib_create_uc_qp ucmd = {}; + u64 flags = 0; + u32 doorbell; + int err, i; + + if (!udata || udata->inlen < sizeof(ucmd)) + return -EINVAL; + + mana_ucontext = rdma_udata_to_drv_context(udata, struct mana_ib_ucontext, ibucontext); + doorbell = mana_ucontext->doorbell; + + err = ib_copy_from_udata(&ucmd, udata, min(sizeof(ucmd), udata->inlen)); + if (err) + return err; + + for (i = 0; i < MANA_UC_QUEUE_TYPE_MAX; ++i) { + err = mana_ib_create_queue(mdev, ucmd.queue_buf[i], ucmd.queue_size[i], + &qp->uc_qp.queues[i]); + if (err) + goto destroy_queues; + } + + err = mana_ib_gd_create_uc_qp(mdev, qp, attr, doorbell, flags); + if (err) + goto destroy_queues; + + qp->ibqp.qp_num = qp->uc_qp.queues[MANA_UC_RECV_QUEUE_RESPONDER].id; + qp->port = attr->port_num; + + for (i = 0; i < MANA_UC_QUEUE_TYPE_MAX; ++i) + resp.queue_id[i] = qp->uc_qp.queues[i].id; + + err = ib_copy_to_udata(udata, &resp, min(sizeof(resp), udata->outlen)); + if (err) + goto destroy_qp; + + err = mana_table_store_qp(mdev, qp); + if (err) + goto destroy_qp; + return 0; +destroy_qp: + mana_ib_gd_destroy_rnic_qp(mdev, qp); +destroy_queues: + while (i-- > 0) + mana_ib_destroy_queue(mdev, &qp->uc_qp.queues[i]); + return err; +} + static void mana_add_qp_to_cqs(struct mana_ib_qp *qp) { struct mana_ib_cq *send_cq = container_of(qp->ibqp.send_cq, struct mana_ib_cq, ibcq); @@ -694,6 +751,8 @@ int mana_ib_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *attr, return mana_ib_create_qp_raw(ibqp, ibqp->pd, attr, udata); case IB_QPT_RC: return mana_ib_create_rc_qp(ibqp, ibqp->pd, attr, udata); + case IB_QPT_UC: + return mana_ib_create_uc_qp(ibqp, ibqp->pd, attr, udata); case IB_QPT_UD: case IB_QPT_GSI: return mana_ib_create_ud_qp(ibqp, ibqp->pd, attr, udata); @@ -775,6 +834,7 @@ int mana_ib_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr, { switch (ibqp->qp_type) { case IB_QPT_RC: + case IB_QPT_UC: case IB_QPT_UD: case IB_QPT_GSI: return mana_ib_gd_modify_qp(ibqp, attr, attr_mask, udata); @@ -843,13 +903,29 @@ static int mana_ib_destroy_rc_qp(struct mana_ib_qp *qp, struct ib_udata *udata) /* Ignore return code as there is not much we can do about it. * The error message is printed inside. */ - mana_ib_gd_destroy_rc_qp(mdev, qp); + mana_ib_gd_destroy_rnic_qp(mdev, qp); for (i = 0; i < MANA_RC_QUEUE_TYPE_MAX; ++i) mana_ib_destroy_queue(mdev, &qp->rc_qp.queues[i]); return 0; } +static int mana_ib_destroy_uc_qp(struct mana_ib_qp *qp, struct ib_udata *udata) +{ + struct mana_ib_dev *mdev = + container_of(qp->ibqp.device, struct mana_ib_dev, ib_dev); + int i; + + mana_table_remove_qp(mdev, qp); + /* Ignore return code as there is not much we can do about it. + * The error message is printed inside. + */ + mana_ib_gd_destroy_rnic_qp(mdev, qp); + for (i = 0; i < MANA_UC_QUEUE_TYPE_MAX; ++i) + mana_ib_destroy_queue(mdev, &qp->uc_qp.queues[i]); + return 0; +} + static int mana_ib_destroy_ud_qp(struct mana_ib_qp *qp, struct ib_udata *udata) { struct mana_ib_dev *mdev = @@ -885,6 +961,8 @@ int mana_ib_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata) return mana_ib_destroy_qp_raw(qp, udata); case IB_QPT_RC: return mana_ib_destroy_rc_qp(qp, udata); + case IB_QPT_UC: + return mana_ib_destroy_uc_qp(qp, udata); case IB_QPT_UD: case IB_QPT_GSI: return mana_ib_destroy_ud_qp(qp, udata); diff --git a/include/uapi/rdma/mana-abi.h b/include/uapi/rdma/mana-abi.h index a75bf32b8..f844afb6b 100644 --- a/include/uapi/rdma/mana-abi.h +++ b/include/uapi/rdma/mana-abi.h @@ -57,6 +57,17 @@ struct mana_ib_create_rc_qp_resp { __u32 queue_id[4]; }; +struct mana_ib_create_uc_qp { + __aligned_u64 queue_buf[3]; + __u32 queue_size[3]; + __u32 reserved; +}; + +struct mana_ib_create_uc_qp_resp { + __u32 queue_id[3]; + __u32 reserved; +}; + struct mana_ib_create_wq { __aligned_u64 wq_buf_addr; __u32 wq_buf_size; -- 2.43.0