From: Matthew Brost <matthew.brost@intel.com>
To: "Christian König" <ckoenig.leichtzumerken@gmail.com>
Cc: <phasta@mailbox.org>, <alexdeucher@gmail.com>,
<simona.vetter@ffwll.ch>, <tursulin@ursulin.net>,
<dri-devel@lists.freedesktop.org>,
<amd-gfx@lists.freedesktop.org>
Subject: Re: [PATCH 05/15] dma-buf: inline spinlock for fence protection
Date: Thu, 23 Oct 2025 11:09:56 -0700 [thread overview]
Message-ID: <aPpvdINe59gYfr2v@lstrano-desk.jf.intel.com> (raw)
In-Reply-To: <20251013143502.1655-6-christian.koenig@amd.com>
On Mon, Oct 13, 2025 at 03:48:32PM +0200, Christian König wrote:
> Allow implementations to not give a spinlock to protect the fence
> internal state, instead a spinlock embedded into the fence structure
> itself is used in this case.
>
> Apart from simplifying the handling for containers and the stub fence
> this has the advantage of allowing implementations to issue fences
> without caring about theit spinlock lifetime.
>
> That in turn is necessary for independent fences who outlive the module
> who originally issued them.
>
One thing Xe really wants to do is use a shared lock for HW fences,
since our IRQ handler walks the pending fence list under the shared lock
and signals the fences. I don’t think it would be desirable to split
this into an IRQ handler list lock and individual fence locks.
It would be great if we could come up with a way to support this model,
ensure it's safe for module unload, and document it clearly.
I’ve thought of a few possible ideas:
- After a fence signals, the dma-fence core should not be allowed to touch
the external lock.
- Only export HW fences to the DRM scheduler, which guarantees that after
HW signaling, it won’t perform any dangerous operations (e.g., making a
call that takes the external lock). This really isn't generic solution
though.
- Create an embedded dma-fence spinlock structure with a refcount and a
dma-fence flag indicating that the external lock is refcounted. In
dma_fence_free, drop the lock’s refcount.
In all of the above cases, the rule is that all module fences must be
signaled before unloading.
Thoughts?
Matt
> Signed-off-by: Christian König <christian.koenig@amd.com>
> ---
> drivers/dma-buf/dma-fence.c | 54 ++++++++++++------------
> drivers/dma-buf/sw_sync.c | 14 +++---
> drivers/dma-buf/sync_debug.h | 2 +-
> drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c | 4 +-
> drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c | 4 +-
> drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h | 12 +++---
> drivers/gpu/drm/drm_crtc.c | 2 +-
> drivers/gpu/drm/drm_writeback.c | 2 +-
> drivers/gpu/drm/nouveau/nouveau_drm.c | 5 ++-
> drivers/gpu/drm/nouveau/nouveau_fence.c | 3 +-
> drivers/gpu/drm/qxl/qxl_release.c | 3 +-
> drivers/gpu/drm/vmwgfx/vmwgfx_fence.c | 3 +-
> drivers/gpu/drm/xe/xe_hw_fence.c | 3 +-
> drivers/gpu/drm/xe/xe_sched_job.c | 4 +-
> include/linux/dma-fence.h | 42 +++++++++++++++++-
> 15 files changed, 99 insertions(+), 58 deletions(-)
>
> diff --git a/drivers/dma-buf/dma-fence.c b/drivers/dma-buf/dma-fence.c
> index 39f73edf3a33..a0b328fdd90d 100644
> --- a/drivers/dma-buf/dma-fence.c
> +++ b/drivers/dma-buf/dma-fence.c
> @@ -343,7 +343,6 @@ void __dma_fence_might_wait(void)
> }
> #endif
>
> -
> /**
> * dma_fence_signal_timestamp_locked - signal completion of a fence
> * @fence: the fence to signal
> @@ -368,7 +367,7 @@ int dma_fence_signal_timestamp_locked(struct dma_fence *fence,
> struct dma_fence_cb *cur, *tmp;
> struct list_head cb_list;
>
> - lockdep_assert_held(fence->lock);
> + lockdep_assert_held(dma_fence_spinlock(fence));
>
> if (unlikely(test_and_set_bit(DMA_FENCE_FLAG_SIGNALED_BIT,
> &fence->flags)))
> @@ -421,9 +420,9 @@ int dma_fence_signal_timestamp(struct dma_fence *fence, ktime_t timestamp)
> if (WARN_ON(!fence))
> return -EINVAL;
>
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> ret = dma_fence_signal_timestamp_locked(fence, timestamp);
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
>
> return ret;
> }
> @@ -475,9 +474,9 @@ int dma_fence_signal(struct dma_fence *fence)
>
> tmp = dma_fence_begin_signalling();
>
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> ret = dma_fence_signal_timestamp_locked(fence, ktime_get());
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
>
> dma_fence_end_signalling(tmp);
>
> @@ -579,10 +578,10 @@ void dma_fence_release(struct kref *kref)
> * don't leave chains dangling. We set the error flag first
> * so that the callbacks know this signal is due to an error.
> */
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> fence->error = -EDEADLK;
> dma_fence_signal_locked(fence);
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
> }
>
> ops = rcu_dereference(fence->ops);
> @@ -612,7 +611,7 @@ static bool __dma_fence_enable_signaling(struct dma_fence *fence)
> const struct dma_fence_ops *ops;
> bool was_set;
>
> - lockdep_assert_held(fence->lock);
> + lockdep_assert_held(dma_fence_spinlock(fence));
>
> was_set = test_and_set_bit(DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT,
> &fence->flags);
> @@ -648,9 +647,9 @@ void dma_fence_enable_sw_signaling(struct dma_fence *fence)
> {
> unsigned long flags;
>
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> __dma_fence_enable_signaling(fence);
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
> }
> EXPORT_SYMBOL(dma_fence_enable_sw_signaling);
>
> @@ -690,8 +689,7 @@ int dma_fence_add_callback(struct dma_fence *fence, struct dma_fence_cb *cb,
> return -ENOENT;
> }
>
> - spin_lock_irqsave(fence->lock, flags);
> -
> + dma_fence_lock(fence, flags);
> if (__dma_fence_enable_signaling(fence)) {
> cb->func = func;
> list_add_tail(&cb->node, &fence->cb_list);
> @@ -699,8 +697,7 @@ int dma_fence_add_callback(struct dma_fence *fence, struct dma_fence_cb *cb,
> INIT_LIST_HEAD(&cb->node);
> ret = -ENOENT;
> }
> -
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
>
> return ret;
> }
> @@ -723,9 +720,9 @@ int dma_fence_get_status(struct dma_fence *fence)
> unsigned long flags;
> int status;
>
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> status = dma_fence_get_status_locked(fence);
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
>
> return status;
> }
> @@ -755,13 +752,11 @@ dma_fence_remove_callback(struct dma_fence *fence, struct dma_fence_cb *cb)
> unsigned long flags;
> bool ret;
>
> - spin_lock_irqsave(fence->lock, flags);
> -
> + dma_fence_lock(fence, flags);
> ret = !list_empty(&cb->node);
> if (ret)
> list_del_init(&cb->node);
> -
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
>
> return ret;
> }
> @@ -800,8 +795,7 @@ dma_fence_default_wait(struct dma_fence *fence, bool intr, signed long timeout)
> unsigned long flags;
> signed long ret = timeout ? timeout : 1;
>
> - spin_lock_irqsave(fence->lock, flags);
> -
> + dma_fence_lock(fence, flags);
> if (test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags))
> goto out;
>
> @@ -824,11 +818,11 @@ dma_fence_default_wait(struct dma_fence *fence, bool intr, signed long timeout)
> __set_current_state(TASK_INTERRUPTIBLE);
> else
> __set_current_state(TASK_UNINTERRUPTIBLE);
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
>
> ret = schedule_timeout(ret);
>
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> if (ret > 0 && intr && signal_pending(current))
> ret = -ERESTARTSYS;
> }
> @@ -838,7 +832,7 @@ dma_fence_default_wait(struct dma_fence *fence, bool intr, signed long timeout)
> __set_current_state(TASK_RUNNING);
>
> out:
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
> return ret;
> }
> EXPORT_SYMBOL(dma_fence_default_wait);
> @@ -1046,7 +1040,6 @@ static void
> __dma_fence_init(struct dma_fence *fence, const struct dma_fence_ops *ops,
> spinlock_t *lock, u64 context, u64 seqno, unsigned long flags)
> {
> - BUG_ON(!lock);
> BUG_ON(!ops || !ops->get_driver_name || !ops->get_timeline_name);
>
> kref_init(&fence->refcount);
> @@ -1057,10 +1050,15 @@ __dma_fence_init(struct dma_fence *fence, const struct dma_fence_ops *ops,
> */
> RCU_INIT_POINTER(fence->ops, ops);
> INIT_LIST_HEAD(&fence->cb_list);
> - fence->lock = lock;
> fence->context = context;
> fence->seqno = seqno;
> fence->flags = flags;
> + if (lock) {
> + fence->extern_lock = lock;
> + } else {
> + spin_lock_init(&fence->inline_lock);
> + fence->flags |= BIT(DMA_FENCE_FLAG_INLINE_LOCK_BIT);
> + }
> fence->error = 0;
>
> trace_dma_fence_init(fence);
> diff --git a/drivers/dma-buf/sw_sync.c b/drivers/dma-buf/sw_sync.c
> index 3c20f1d31cf5..8f48529214a4 100644
> --- a/drivers/dma-buf/sw_sync.c
> +++ b/drivers/dma-buf/sw_sync.c
> @@ -155,12 +155,12 @@ static void timeline_fence_release(struct dma_fence *fence)
> struct sync_timeline *parent = dma_fence_parent(fence);
> unsigned long flags;
>
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> if (!list_empty(&pt->link)) {
> list_del(&pt->link);
> rb_erase(&pt->node, &parent->pt_tree);
> }
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
>
> sync_timeline_put(parent);
> dma_fence_free(fence);
> @@ -178,7 +178,7 @@ static void timeline_fence_set_deadline(struct dma_fence *fence, ktime_t deadlin
> struct sync_pt *pt = dma_fence_to_sync_pt(fence);
> unsigned long flags;
>
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> if (test_bit(SW_SYNC_HAS_DEADLINE_BIT, &fence->flags)) {
> if (ktime_before(deadline, pt->deadline))
> pt->deadline = deadline;
> @@ -186,7 +186,7 @@ static void timeline_fence_set_deadline(struct dma_fence *fence, ktime_t deadlin
> pt->deadline = deadline;
> __set_bit(SW_SYNC_HAS_DEADLINE_BIT, &fence->flags);
> }
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
> }
>
> static const struct dma_fence_ops timeline_fence_ops = {
> @@ -427,13 +427,13 @@ static int sw_sync_ioctl_get_deadline(struct sync_timeline *obj, unsigned long a
> goto put_fence;
> }
>
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> if (!test_bit(SW_SYNC_HAS_DEADLINE_BIT, &fence->flags)) {
> ret = -ENOENT;
> goto unlock;
> }
> data.deadline_ns = ktime_to_ns(pt->deadline);
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
>
> dma_fence_put(fence);
>
> @@ -446,7 +446,7 @@ static int sw_sync_ioctl_get_deadline(struct sync_timeline *obj, unsigned long a
> return 0;
>
> unlock:
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
> put_fence:
> dma_fence_put(fence);
>
> diff --git a/drivers/dma-buf/sync_debug.h b/drivers/dma-buf/sync_debug.h
> index 02af347293d0..c49324505b20 100644
> --- a/drivers/dma-buf/sync_debug.h
> +++ b/drivers/dma-buf/sync_debug.h
> @@ -47,7 +47,7 @@ struct sync_timeline {
>
> static inline struct sync_timeline *dma_fence_parent(struct dma_fence *fence)
> {
> - return container_of(fence->lock, struct sync_timeline, lock);
> + return container_of(fence->extern_lock, struct sync_timeline, lock);
> }
>
> /**
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c
> index 5ec5c3ff22bb..fcc7a3fb93b3 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_ring.c
> @@ -468,10 +468,10 @@ bool amdgpu_ring_soft_recovery(struct amdgpu_ring *ring, unsigned int vmid,
> if (amdgpu_sriov_vf(ring->adev) || !ring->funcs->soft_recovery || !fence)
> return false;
>
> - spin_lock_irqsave(fence->lock, flags);
> + dma_fence_lock(fence, flags);
> if (!dma_fence_is_signaled_locked(fence))
> dma_fence_set_error(fence, -ENODATA);
> - spin_unlock_irqrestore(fence->lock, flags);
> + dma_fence_unlock(fence, flags);
>
> while (!dma_fence_is_signaled(fence) &&
> ktime_to_ns(ktime_sub(deadline, ktime_get())) > 0)
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> index db66b4232de0..db6516ce8335 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
> @@ -2774,8 +2774,8 @@ void amdgpu_vm_fini(struct amdgpu_device *adev, struct amdgpu_vm *vm)
> dma_fence_put(vm->last_unlocked);
> dma_fence_wait(vm->last_tlb_flush, false);
> /* Make sure that all fence callbacks have completed */
> - spin_lock_irqsave(vm->last_tlb_flush->lock, flags);
> - spin_unlock_irqrestore(vm->last_tlb_flush->lock, flags);
> + dma_fence_lock(vm->last_tlb_flush, flags);
> + dma_fence_unlock(vm->last_tlb_flush, flags);
> dma_fence_put(vm->last_tlb_flush);
>
> list_for_each_entry_safe(mapping, tmp, &vm->freed, list) {
> diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> index 77207f4e448e..4fc7f66b7d13 100644
> --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
> @@ -631,20 +631,20 @@ bool amdgpu_vm_is_bo_always_valid(struct amdgpu_vm *vm, struct amdgpu_bo *bo);
> */
> static inline uint64_t amdgpu_vm_tlb_seq(struct amdgpu_vm *vm)
> {
> + struct dma_fence *fence;
> unsigned long flags;
> - spinlock_t *lock;
>
> /*
> * Workaround to stop racing between the fence signaling and handling
> - * the cb. The lock is static after initially setting it up, just make
> - * sure that the dma_fence structure isn't freed up.
> + * the cb.
> */
> rcu_read_lock();
> - lock = vm->last_tlb_flush->lock;
> + fence = dma_fence_get_rcu(vm->last_tlb_flush);
> rcu_read_unlock();
>
> - spin_lock_irqsave(lock, flags);
> - spin_unlock_irqrestore(lock, flags);
> + dma_fence_lock(fence, flags);
> + dma_fence_unlock(fence, flags);
> + dma_fence_put(fence);
>
> return atomic64_read(&vm->tlb_seq);
> }
> diff --git a/drivers/gpu/drm/drm_crtc.c b/drivers/gpu/drm/drm_crtc.c
> index 46655339003d..ad47f58cd159 100644
> --- a/drivers/gpu/drm/drm_crtc.c
> +++ b/drivers/gpu/drm/drm_crtc.c
> @@ -159,7 +159,7 @@ static const struct dma_fence_ops drm_crtc_fence_ops;
> static struct drm_crtc *fence_to_crtc(struct dma_fence *fence)
> {
> BUG_ON(fence->ops != &drm_crtc_fence_ops);
> - return container_of(fence->lock, struct drm_crtc, fence_lock);
> + return container_of(fence->extern_lock, struct drm_crtc, fence_lock);
> }
>
> static const char *drm_crtc_fence_get_driver_name(struct dma_fence *fence)
> diff --git a/drivers/gpu/drm/drm_writeback.c b/drivers/gpu/drm/drm_writeback.c
> index 95b8a2e4bda6..624a4e8b6c99 100644
> --- a/drivers/gpu/drm/drm_writeback.c
> +++ b/drivers/gpu/drm/drm_writeback.c
> @@ -81,7 +81,7 @@
> * From userspace, this property will always read as zero.
> */
>
> -#define fence_to_wb_connector(x) container_of(x->lock, \
> +#define fence_to_wb_connector(x) container_of(x->extern_lock, \
> struct drm_writeback_connector, \
> fence_lock)
>
> diff --git a/drivers/gpu/drm/nouveau/nouveau_drm.c b/drivers/gpu/drm/nouveau/nouveau_drm.c
> index 1527b801f013..2956ed2ec073 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_drm.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_drm.c
> @@ -156,12 +156,13 @@ nouveau_name(struct drm_device *dev)
> static inline bool
> nouveau_cli_work_ready(struct dma_fence *fence)
> {
> + unsigned long flags;
> bool ret = true;
>
> - spin_lock_irq(fence->lock);
> + dma_fence_lock(fence, flags);
> if (!dma_fence_is_signaled_locked(fence))
> ret = false;
> - spin_unlock_irq(fence->lock);
> + dma_fence_unlock(fence, flags);
>
> if (ret == true)
> dma_fence_put(fence);
> diff --git a/drivers/gpu/drm/nouveau/nouveau_fence.c b/drivers/gpu/drm/nouveau/nouveau_fence.c
> index d5654e26d5bc..272b492c4d7c 100644
> --- a/drivers/gpu/drm/nouveau/nouveau_fence.c
> +++ b/drivers/gpu/drm/nouveau/nouveau_fence.c
> @@ -47,7 +47,8 @@ from_fence(struct dma_fence *fence)
> static inline struct nouveau_fence_chan *
> nouveau_fctx(struct nouveau_fence *fence)
> {
> - return container_of(fence->base.lock, struct nouveau_fence_chan, lock);
> + return container_of(fence->base.extern_lock, struct nouveau_fence_chan,
> + lock);
> }
>
> static bool
> diff --git a/drivers/gpu/drm/qxl/qxl_release.c b/drivers/gpu/drm/qxl/qxl_release.c
> index 05204a6a3fa8..1d346822c1f7 100644
> --- a/drivers/gpu/drm/qxl/qxl_release.c
> +++ b/drivers/gpu/drm/qxl/qxl_release.c
> @@ -60,7 +60,8 @@ static long qxl_fence_wait(struct dma_fence *fence, bool intr,
> struct qxl_device *qdev;
> unsigned long cur, end = jiffies + timeout;
>
> - qdev = container_of(fence->lock, struct qxl_device, release_lock);
> + qdev = container_of(fence->extern_lock, struct qxl_device,
> + release_lock);
>
> if (!wait_event_timeout(qdev->release_event,
> (dma_fence_is_signaled(fence) ||
> diff --git a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
> index c2294abbe753..346761172c1b 100644
> --- a/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
> +++ b/drivers/gpu/drm/vmwgfx/vmwgfx_fence.c
> @@ -47,7 +47,8 @@ struct vmw_event_fence_action {
> static struct vmw_fence_manager *
> fman_from_fence(struct vmw_fence_obj *fence)
> {
> - return container_of(fence->base.lock, struct vmw_fence_manager, lock);
> + return container_of(fence->base.extern_lock, struct vmw_fence_manager,
> + lock);
> }
>
> static void vmw_fence_obj_destroy(struct dma_fence *f)
> diff --git a/drivers/gpu/drm/xe/xe_hw_fence.c b/drivers/gpu/drm/xe/xe_hw_fence.c
> index b2a0c46dfcd4..3456bec93c70 100644
> --- a/drivers/gpu/drm/xe/xe_hw_fence.c
> +++ b/drivers/gpu/drm/xe/xe_hw_fence.c
> @@ -144,7 +144,8 @@ static struct xe_hw_fence *to_xe_hw_fence(struct dma_fence *fence);
>
> static struct xe_hw_fence_irq *xe_hw_fence_irq(struct xe_hw_fence *fence)
> {
> - return container_of(fence->dma.lock, struct xe_hw_fence_irq, lock);
> + return container_of(fence->dma.extern_lock, struct xe_hw_fence_irq,
> + lock);
> }
>
> static const char *xe_hw_fence_get_driver_name(struct dma_fence *dma_fence)
> diff --git a/drivers/gpu/drm/xe/xe_sched_job.c b/drivers/gpu/drm/xe/xe_sched_job.c
> index d21bf8f26964..ea7038475b4b 100644
> --- a/drivers/gpu/drm/xe/xe_sched_job.c
> +++ b/drivers/gpu/drm/xe/xe_sched_job.c
> @@ -187,11 +187,11 @@ static bool xe_fence_set_error(struct dma_fence *fence, int error)
> unsigned long irq_flags;
> bool signaled;
>
> - spin_lock_irqsave(fence->lock, irq_flags);
> + dma_fence_lock(fence, irq_flags);
> signaled = test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &fence->flags);
> if (!signaled)
> dma_fence_set_error(fence, error);
> - spin_unlock_irqrestore(fence->lock, irq_flags);
> + dma_fence_unlock(fence, irq_flags);
>
> return signaled;
> }
> diff --git a/include/linux/dma-fence.h b/include/linux/dma-fence.h
> index e1ba1d53de88..fb416f500664 100644
> --- a/include/linux/dma-fence.h
> +++ b/include/linux/dma-fence.h
> @@ -34,7 +34,8 @@ struct seq_file;
> * @ops: dma_fence_ops associated with this fence
> * @rcu: used for releasing fence with kfree_rcu
> * @cb_list: list of all callbacks to call
> - * @lock: spin_lock_irqsave used for locking
> + * @extern_lock: external spin_lock_irqsave used for locking
> + * @inline_lock: alternative internal spin_lock_irqsave used for locking
> * @context: execution context this fence belongs to, returned by
> * dma_fence_context_alloc()
> * @seqno: the sequence number of this fence inside the execution context,
> @@ -48,6 +49,7 @@ struct seq_file;
> * atomic ops (bit_*), so taking the spinlock will not be needed most
> * of the time.
> *
> + * DMA_FENCE_FLAG_INLINE_LOCK_BIT - use inline spinlock instead of external one
> * DMA_FENCE_FLAG_SIGNALED_BIT - fence is already signaled
> * DMA_FENCE_FLAG_TIMESTAMP_BIT - timestamp recorded for fence signaling
> * DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT - enable_signaling might have been called
> @@ -65,7 +67,10 @@ struct seq_file;
> * been completed, or never called at all.
> */
> struct dma_fence {
> - spinlock_t *lock;
> + union {
> + spinlock_t *extern_lock;
> + spinlock_t inline_lock;
> + };
> const struct dma_fence_ops __rcu *ops;
> /*
> * We clear the callback list on kref_put so that by the time we
> @@ -98,6 +103,7 @@ struct dma_fence {
> };
>
> enum dma_fence_flag_bits {
> + DMA_FENCE_FLAG_INLINE_LOCK_BIT,
> DMA_FENCE_FLAG_SEQNO64_BIT,
> DMA_FENCE_FLAG_SIGNALED_BIT,
> DMA_FENCE_FLAG_TIMESTAMP_BIT,
> @@ -351,6 +357,38 @@ dma_fence_get_rcu_safe(struct dma_fence __rcu **fencep)
> } while (1);
> }
>
> +/**
> + * dma_fence_spinlock - return pointer to the spinlock protecting the fence
> + * @fence: the fence to get the lock from
> + *
> + * Return either the pointer to the embedded or the external spin lock.
> + */
> +static inline spinlock_t *dma_fence_spinlock(struct dma_fence *fence)
> +{
> + return test_bit(DMA_FENCE_FLAG_INLINE_LOCK_BIT, &fence->flags) ?
> + &fence->inline_lock : fence->extern_lock;
> +}
> +
> +/**
> + * dma_fence_lock - irqsave lock the fence
> + * @fence: the fence to lock
> + * @flags: where to store the CPU flags.
> + *
> + * Lock the fence, preventing it from changing to the signaled state.
> + */
> +#define dma_fence_lock(fence, flags) \
> + spin_lock_irqsave(dma_fence_spinlock(fence), flags)
> +
> +/**
> + * dma_fence_unlock - unlock the fence and irqrestore
> + * @fence: the fence to unlock
> + * @flags the CPU flags to restore
> + *
> + * Unlock the fence, allowing it to change it's state to signaled again.
> + */
> +#define dma_fence_unlock(fence, flags) \
> + spin_unlock_irqrestore(dma_fence_spinlock(fence), flags)
> +
> #ifdef CONFIG_LOCKDEP
> bool dma_fence_begin_signalling(void);
> void dma_fence_end_signalling(bool cookie);
> --
> 2.43.0
>
next prev parent reply other threads:[~2025-10-23 18:10 UTC|newest]
Thread overview: 47+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-10-13 13:48 Independence for dma_fences! Christian König
2025-10-13 13:48 ` [PATCH 01/15] dma-buf: cleanup dma_fence_describe Christian König
2025-10-14 14:37 ` Tvrtko Ursulin
2025-10-23 3:45 ` Matthew Brost
2025-10-13 13:48 ` [PATCH 02/15] dma-buf: rework stub fence initialisation Christian König
2025-10-14 15:03 ` Tvrtko Ursulin
2025-10-24 7:29 ` Tvrtko Ursulin
2025-10-13 13:48 ` [PATCH 03/15] dma-buf: protected fence ops by RCU Christian König
2025-10-16 18:04 ` Tvrtko Ursulin
2025-10-31 10:35 ` Tvrtko Ursulin
2025-10-13 13:48 ` [PATCH 04/15] dma-buf: detach fence ops on signal Christian König
2025-10-16 8:56 ` Tvrtko Ursulin
2025-10-16 15:57 ` Tvrtko Ursulin
2025-10-23 4:23 ` Matthew Brost
2025-10-23 4:44 ` Matthew Brost
2025-10-30 13:52 ` Christian König
2025-10-31 10:31 ` Tvrtko Ursulin
2025-10-17 9:14 ` Philipp Stanner
2025-10-30 15:05 ` Christian König
2025-10-13 13:48 ` [PATCH 05/15] dma-buf: inline spinlock for fence protection Christian König
2025-10-16 9:26 ` Tvrtko Ursulin
2025-11-03 13:07 ` Philipp Stanner
2025-10-23 18:09 ` Matthew Brost [this message]
2025-10-30 15:14 ` Christian König
2025-10-13 13:48 ` [PATCH 06/15] dma-buf: use inline lock for the stub fence Christian König
2025-10-13 13:48 ` [PATCH 07/15] dma-buf: use inline lock for the dma-fence-array Christian König
2025-10-13 13:48 ` [PATCH 08/15] dma-buf: use inline lock for the dma-fence-chain Christian König
2025-10-13 13:48 ` [PATCH 09/15] drm/sched: use inline locks for the drm-sched-fence Christian König
2025-10-13 13:48 ` [PATCH 10/15] drm/amdgpu: fix KFD eviction fence enable_signaling path Christian König
2025-10-13 13:48 ` [PATCH 11/15] drm/amdgpu: independence for the amdgpu_fence! Christian König
2025-10-13 13:48 ` [PATCH 12/15] drm/amdgpu: independence for the amdgpu_eviction_fence! Christian König
2025-10-13 13:48 ` [PATCH 13/15] drm/amdgpu: independence for the amdgpu_vm_tlb_fence! Christian König
2025-10-13 13:48 ` [PATCH 14/15] drm/amdgpu: independence for the amdkfd_fence! Christian König
2025-10-17 22:22 ` Felix Kuehling
2025-10-30 15:07 ` Christian König
2025-10-30 20:04 ` Felix Kuehling
2025-10-13 13:48 ` [PATCH 15/15] drm/amdgpu: independence for the amdgpu_userq__fence! Christian König
2025-10-13 14:54 ` Independence for dma_fences! Philipp Stanner
2025-10-14 15:54 ` Christian König
2025-10-17 8:32 ` Philipp Stanner
2025-10-28 14:06 ` Christian König
2025-10-29 20:53 ` Matthew Brost
2025-10-30 10:59 ` Christian König
2025-10-31 17:44 ` Matthew Brost
2025-11-03 11:43 ` Christian König
2025-11-03 19:32 ` Matthew Brost
2025-10-15 0:51 ` Dave Airlie
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aPpvdINe59gYfr2v@lstrano-desk.jf.intel.com \
--to=matthew.brost@intel.com \
--cc=alexdeucher@gmail.com \
--cc=amd-gfx@lists.freedesktop.org \
--cc=ckoenig.leichtzumerken@gmail.com \
--cc=dri-devel@lists.freedesktop.org \
--cc=phasta@mailbox.org \
--cc=simona.vetter@ffwll.ch \
--cc=tursulin@ursulin.net \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox