From: Frank Li <Frank.li@nxp.com>
To: Nathan Lynch <nathan.lynch@amd.com>
Cc: Vinod Koul <vkoul@kernel.org>, Wei Huang <wei.huang2@amd.com>,
Mario Limonciello <mario.limonciello@amd.com>,
Bjorn Helgaas <bhelgaas@google.com>,
Jonathan Cameron <jonathan.cameron@huawei.com>,
Stephen Bates <Stephen.Bates@amd.com>,
PradeepVineshReddy.Kodamati@amd.com, John.Kariuki@amd.com,
linux-pci@vger.kernel.org, linux-kernel@vger.kernel.org,
dmaengine@vger.kernel.org
Subject: Re: [PATCH 11/23] dmaengine: sdxi: Add client context alloc and release APIs
Date: Mon, 20 Apr 2026 04:19:28 -0400 [thread overview]
Message-ID: <aeXhkIZgwGttlJB0@lizhi-Precision-Tower-5810> (raw)
In-Reply-To: <20260410-sdxi-base-v1-11-1d184cb5c60a@amd.com>
On Fri, Apr 10, 2026 at 08:07:21AM -0500, Nathan Lynch wrote:
> Expose sdxi_cxt_new() and sdxi_cxt_exit(), which are the rest of the
> driver's entry points to creating and releasing SDXI contexts.
>
> Track client contexts in a device-wide allocating xarray, mapping
> context ID to the context object. The admin context always has ID 0,
> so begin allocations at 1. Define a local class for ID allocation to
> facilitate automatic release of an ID if an error is encountered when
> "publishing" a context to the L1 table.
>
> Introduce new code to invalidate a context's entry in the L1 table on
> deallocation.
>
> Support for starting and stopping contexts will be added in changes to
> follow.
>
> The only expected user of sdxi_cxt_new() and sdxi_cxt_exit() at this
> point is the DMA engine provider code where a client context per
> channel will be created.
>
> Co-developed-by: Wei Huang <wei.huang2@amd.com>
> Signed-off-by: Wei Huang <wei.huang2@amd.com>
> Signed-off-by: Nathan Lynch <nathan.lynch@amd.com>
> ---
> drivers/dma/sdxi/context.c | 111 +++++++++++++++++++++++++++++++++++++++++++++
> drivers/dma/sdxi/context.h | 13 ++++++
> drivers/dma/sdxi/device.c | 8 ++++
> drivers/dma/sdxi/sdxi.h | 2 +
> 4 files changed, 134 insertions(+)
>
> diff --git a/drivers/dma/sdxi/context.c b/drivers/dma/sdxi/context.c
> index 934c487b4774..7cae140c0a20 100644
> --- a/drivers/dma/sdxi/context.c
> +++ b/drivers/dma/sdxi/context.c
> @@ -155,6 +155,16 @@ static int configure_cxt_ctl(struct sdxi_cxt_ctl *ctl, const struct sdxi_cxt_ctl
> return 0;
> }
>
> +static void invalidate_cxtl_ctl(struct sdxi_cxt_ctl *ctl)
> +{
> + u64 ds_ring_ptr = le64_to_cpu(ctl->ds_ring_ptr);
> +
> + FIELD_MODIFY(SDXI_CXT_CTL_VL, &ds_ring_ptr, 0);
> + WRITE_ONCE(ctl->ds_ring_ptr, cpu_to_le64(ds_ring_ptr));
> + dma_wmb();
> + *ctl = (typeof(*ctl)) { 0 };
> +}
> +
> /*
> * Logical representation of CXT_L1_ENT subfields.
> */
> @@ -209,6 +219,16 @@ static int configure_L1_entry(struct sdxi_cxt_L1_ent *ent,
> return 0;
> }
>
> +static void invalidate_L1_entry(struct sdxi_cxt_L1_ent *ent)
> +{
> + u64 cxt_ctl_ptr = le64_to_cpu(ent->cxt_ctl_ptr);
> +
> + FIELD_MODIFY(SDXI_CXT_L1_ENT_VL, &cxt_ctl_ptr, 0);
> + WRITE_ONCE(ent->cxt_ctl_ptr, cpu_to_le64(cxt_ctl_ptr));
> + dma_wmb();
> + *ent = (typeof(*ent)) { 0 };
> +}
> +
> /*
> * Make the context control structure hierarchy valid from the POV of
> * the SDXI implementation. This may eventually involve allocation of
> @@ -259,6 +279,17 @@ static int sdxi_publish_cxt(const struct sdxi_cxt *cxt)
> /* todo: need to send DSC_CXT_UPD to admin */
> }
>
> +/* Invalidate a context. */
> +static void sdxi_rescind_cxt(struct sdxi_cxt *cxt)
> +{
> + u8 l1_idx = ID_TO_L1_INDEX(cxt->id);
> + struct sdxi_cxt_L1_ent *ent = &cxt->sdxi->L1_table->entry[l1_idx];
> +
> + invalidate_L1_entry(ent);
> + invalidate_cxtl_ctl(cxt->cxt_ctl);
> + /* todo: need to send DSC_CXT_UPD to admin */
> +}
> +
> void sdxi_cxt_push_doorbell(struct sdxi_cxt *cxt, u64 index)
> {
> /* Ensure preceding write index increment is visible. */
> @@ -266,6 +297,61 @@ void sdxi_cxt_push_doorbell(struct sdxi_cxt *cxt, u64 index)
> iowrite64(index, cxt->db);
> }
>
> +/* Enable automatic cleanup of an allocated context ID */
> +struct __class_sdxi_cxt_id {
> + struct sdxi_dev *sdxi;
> + s32 id;
> +};
> +
> +#define sdxi_cxt_id_null ((struct __class_sdxi_cxt_id){ NULL, -1 })
> +#define take_sdxi_cxt_id(x) __get_and_null(x, sdxi_cxt_id_null)
> +
> +DEFINE_CLASS(sdxi_alloc_cxt_id, struct __class_sdxi_cxt_id,
> + if (_T.id >= 0)
> + xa_erase(&_T.sdxi->client_cxts, _T.id),
> + ((struct __class_sdxi_cxt_id){
> + .sdxi = sdxi,
> + .id = ({
> + struct xa_limit limit = XA_LIMIT(1, sdxi->max_cxtid);
> + u32 id;
> + int err = xa_alloc(&sdxi->client_cxts, &id, cxt,
> + limit, GFP_KERNEL);
> + err ? err : id;
> + }),
> + }),
> + struct sdxi_dev *sdxi, struct sdxi_cxt *cxt)
> +
> +/*
> + * Allocate the context ID; link the context back to the device;
> + * perform some final initialization of the context based on the ID
> + * allocated; update the context tables.
> + */
> +static int register_cxt(struct sdxi_dev *sdxi, struct sdxi_cxt *cxt)
> +{
> + int err;
> +
> + CLASS(sdxi_alloc_cxt_id, slot)(sdxi, cxt);
> + if (slot.id < 0)
> + return slot.id;
I like use cleanup to do this. define error macro, like cleanup.h
* ACQUIRE(pci_dev_try, lock)(dev);
* rc = ACQUIRE_ERR(pci_dev_try, &lock);
so hidden detail "id" in sdxi_alloc_cxt_id.
Or you can refer runtime pm method, save necceary information to "cxt"
DEFINE_GUARD_COND(pm_runtime_active, _try,
pm_runtime_get_active(_T, RPM_TRANSPARENT), _RET == 0)
> +
> + cxt->sdxi = sdxi;
> + cxt->id = slot.id;
> + cxt->db = sdxi->dbs + slot.id * sdxi->db_stride;
> +
> + err = sdxi_publish_cxt(cxt);
> + if (err)
> + return err;
> +
> + take_sdxi_cxt_id(slot);
I undestand try to keep id to avoid call xa_erase. but it hidden too much
detail.
If only one error branch, using cleanup here have not bring too much beneafit.
Idealy logic
id = __free(your_xa_erase) your_xa_alloc()
...
ctx->id = no_free_ptr(id).
Frank
> + return 0;
> +}
> +
> +static void unregister_cxt(struct sdxi_cxt *cxt)
> +{
> + sdxi_rescind_cxt(cxt);
> + xa_erase(&cxt->sdxi->client_cxts, cxt->id);
> +}
> +
> static void free_admin_cxt(void *ptr)
> {
> struct sdxi_dev *sdxi = ptr;
> @@ -296,3 +382,28 @@ int sdxi_admin_cxt_init(struct sdxi_dev *sdxi)
>
> return devm_add_action_or_reset(sdxi_to_dev(sdxi), free_admin_cxt, sdxi);
> }
> +
> +/*
> + * Allocate a context for in-kernel use. Starting the context is the
> + * caller's responsibility.
> + */
> +struct sdxi_cxt *sdxi_cxt_new(struct sdxi_dev *sdxi)
> +{
> + struct sdxi_cxt *cxt __free(sdxi_cxt) = sdxi_alloc_cxt(sdxi);
> + if (!cxt)
> + return NULL;
> +
> + if (register_cxt(sdxi, cxt))
> + return NULL;
> +
> + return_ptr(cxt);
> +}
> +
> +void sdxi_cxt_exit(struct sdxi_cxt *cxt)
> +{
> + if (WARN_ON(sdxi_cxt_is_admin(cxt)))
> + return;
> +
> + unregister_cxt(cxt);
> + sdxi_free_cxt(cxt);
> +}
> diff --git a/drivers/dma/sdxi/context.h b/drivers/dma/sdxi/context.h
> index c34acd730acb..5cd78e883c8d 100644
> --- a/drivers/dma/sdxi/context.h
> +++ b/drivers/dma/sdxi/context.h
> @@ -58,6 +58,19 @@ struct sdxi_cxt {
>
> int sdxi_admin_cxt_init(struct sdxi_dev *sdxi);
>
> +struct sdxi_cxt *sdxi_cxt_new(struct sdxi_dev *sdxi);
> +void sdxi_cxt_exit(struct sdxi_cxt *cxt);
> +
> +static inline struct sdxi_cxt *to_admin_cxt(const struct sdxi_cxt *cxt)
> +{
> + return cxt->sdxi->admin_cxt;
> +}
> +
> +static inline bool sdxi_cxt_is_admin(const struct sdxi_cxt *cxt)
> +{
> + return cxt == to_admin_cxt(cxt);
> +}
> +
> void sdxi_cxt_push_doorbell(struct sdxi_cxt *cxt, u64 index);
>
> #endif /* DMA_SDXI_CONTEXT_H */
> diff --git a/drivers/dma/sdxi/device.c b/drivers/dma/sdxi/device.c
> index 15f61d1ce490..aaff6b15325a 100644
> --- a/drivers/dma/sdxi/device.c
> +++ b/drivers/dma/sdxi/device.c
> @@ -12,6 +12,7 @@
> #include <linux/dmapool.h>
> #include <linux/log2.h>
> #include <linux/slab.h>
> +#include <linux/xarray.h>
>
> #include "context.h"
> #include "hw.h"
> @@ -302,6 +303,7 @@ int sdxi_register(struct device *dev, const struct sdxi_bus_ops *ops)
>
> sdxi->dev = dev;
> sdxi->bus_ops = ops;
> + xa_init_flags(&sdxi->client_cxts, XA_FLAGS_ALLOC1);
> dev_set_drvdata(dev, sdxi);
>
> err = sdxi->bus_ops->init(sdxi);
> @@ -314,6 +316,12 @@ int sdxi_register(struct device *dev, const struct sdxi_bus_ops *ops)
> void sdxi_unregister(struct device *dev)
> {
> struct sdxi_dev *sdxi = dev_get_drvdata(dev);
> + struct sdxi_cxt *cxt;
> + unsigned long index;
> +
> + xa_for_each(&sdxi->client_cxts, index, cxt)
> + sdxi_cxt_exit(cxt);
> + xa_destroy(&sdxi->client_cxts);
>
> sdxi_dev_stop(sdxi);
> }
> diff --git a/drivers/dma/sdxi/sdxi.h b/drivers/dma/sdxi/sdxi.h
> index 426101875334..da33719735ab 100644
> --- a/drivers/dma/sdxi/sdxi.h
> +++ b/drivers/dma/sdxi/sdxi.h
> @@ -12,6 +12,7 @@
> #include <linux/dev_printk.h>
> #include <linux/io-64-nonatomic-lo-hi.h>
> #include <linux/types.h>
> +#include <linux/xarray.h>
>
> #include "mmio.h"
>
> @@ -61,6 +62,7 @@ struct sdxi_dev {
> struct dma_pool *cst_blk_pool;
>
> struct sdxi_cxt *admin_cxt;
> + struct xarray client_cxts; /* context id -> (struct sdxi_cxt *) */
>
> const struct sdxi_bus_ops *bus_ops;
> };
>
> --
> 2.53.0
>
next prev parent reply other threads:[~2026-04-20 8:19 UTC|newest]
Thread overview: 48+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-04-10 13:07 [PATCH 00/23] dmaengine: Smart Data Accelerator Interface (SDXI) basic support Nathan Lynch via B4 Relay
2026-04-10 13:07 ` [PATCH 01/23] PCI: Add SNIA SDXI accelerator sub-class Nathan Lynch via B4 Relay
2026-04-10 13:07 ` [PATCH 02/23] MAINTAINERS: Add entry for SDXI driver Nathan Lynch via B4 Relay
2026-04-10 13:07 ` [PATCH 03/23] dmaengine: sdxi: Add PCI initialization Nathan Lynch via B4 Relay
2026-04-20 6:36 ` Frank Li
2026-04-20 15:29 ` Lynch, Nathan
2026-04-10 13:07 ` [PATCH 04/23] dmaengine: sdxi: Feature discovery and initial configuration Nathan Lynch via B4 Relay
2026-04-20 6:51 ` Frank Li
2026-04-20 19:31 ` Lynch, Nathan
2026-04-10 13:07 ` [PATCH 05/23] dmaengine: sdxi: Configure context tables Nathan Lynch via B4 Relay
2026-04-10 13:07 ` [PATCH 06/23] dmaengine: sdxi: Allocate DMA pools Nathan Lynch via B4 Relay
2026-04-20 7:03 ` Frank Li
2026-04-20 22:16 ` Lynch, Nathan
2026-04-10 13:07 ` [PATCH 07/23] dmaengine: sdxi: Allocate administrative context Nathan Lynch via B4 Relay
2026-04-20 7:11 ` Frank Li
2026-04-10 13:07 ` [PATCH 08/23] dmaengine: sdxi: Install " Nathan Lynch via B4 Relay
2026-04-20 7:16 ` Frank Li
2026-04-10 13:07 ` [PATCH 09/23] dmaengine: sdxi: Start functions on probe, stop on remove Nathan Lynch via B4 Relay
2026-04-20 7:18 ` Frank Li
2026-04-10 13:07 ` [PATCH 10/23] dmaengine: sdxi: Complete administrative context jump start Nathan Lynch via B4 Relay
2026-04-20 7:33 ` Frank Li
2026-04-10 13:07 ` [PATCH 11/23] dmaengine: sdxi: Add client context alloc and release APIs Nathan Lynch via B4 Relay
2026-04-20 8:19 ` Frank Li [this message]
2026-04-10 13:07 ` [PATCH 12/23] dmaengine: sdxi: Add descriptor ring management Nathan Lynch via B4 Relay
2026-04-20 8:32 ` Frank Li
2026-04-20 22:42 ` Lynch, Nathan
2026-04-10 13:07 ` [PATCH 13/23] dmaengine: sdxi: Add unit tests for descriptor ring reservations Nathan Lynch via B4 Relay
2026-04-10 13:07 ` [PATCH 14/23] dmaengine: sdxi: Attach descriptor ring state to contexts Nathan Lynch via B4 Relay
2026-04-20 8:34 ` Frank Li
2026-04-10 13:07 ` [PATCH 15/23] dmaengine: sdxi: Per-context access key (AKey) table entry allocator Nathan Lynch via B4 Relay
2026-04-20 8:39 ` Frank Li
2026-04-10 13:07 ` [PATCH 16/23] dmaengine: sdxi: Generic descriptor manipulation helpers Nathan Lynch via B4 Relay
2026-04-20 8:40 ` Frank Li
2026-04-10 13:07 ` [PATCH 17/23] dmaengine: sdxi: Add completion status block API Nathan Lynch via B4 Relay
2026-04-20 8:44 ` Frank Li
2026-04-10 13:07 ` [PATCH 18/23] dmaengine: sdxi: Encode context start, stop, and sync descriptors Nathan Lynch via B4 Relay
2026-04-20 8:48 ` Frank Li
2026-04-10 13:07 ` [PATCH 19/23] dmaengine: sdxi: Provide context start and stop APIs Nathan Lynch via B4 Relay
2026-04-20 8:50 ` Frank Li
2026-04-21 19:54 ` Lynch, Nathan
2026-04-10 13:07 ` [PATCH 20/23] dmaengine: sdxi: Encode nop, copy, and interrupt descriptors Nathan Lynch via B4 Relay
2026-04-20 8:52 ` Frank Li
2026-04-10 13:07 ` [PATCH 21/23] dmaengine: sdxi: Add unit tests for descriptor encoding Nathan Lynch via B4 Relay
2026-04-10 13:07 ` [PATCH 22/23] dmaengine: sdxi: MSI/MSI-X vector allocation and mapping Nathan Lynch via B4 Relay
2026-04-20 8:54 ` Frank Li
2026-04-10 13:07 ` [PATCH 23/23] dmaengine: sdxi: Add DMA engine provider Nathan Lynch via B4 Relay
2026-04-20 9:08 ` Frank Li
2026-04-21 20:04 ` Lynch, Nathan
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aeXhkIZgwGttlJB0@lizhi-Precision-Tower-5810 \
--to=frank.li@nxp.com \
--cc=John.Kariuki@amd.com \
--cc=PradeepVineshReddy.Kodamati@amd.com \
--cc=Stephen.Bates@amd.com \
--cc=bhelgaas@google.com \
--cc=dmaengine@vger.kernel.org \
--cc=jonathan.cameron@huawei.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-pci@vger.kernel.org \
--cc=mario.limonciello@amd.com \
--cc=nathan.lynch@amd.com \
--cc=vkoul@kernel.org \
--cc=wei.huang2@amd.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox