public inbox for intel-xe@lists.freedesktop.org
 help / color / mirror / Atom feed
From: "Christian König" <christian.koenig@amd.com>
To: Boris Brezillon <boris.brezillon@collabora.com>,
	Matthew Brost <matthew.brost@intel.com>
Cc: intel-xe@lists.freedesktop.org, dri-devel@lists.freedesktop.org,
	Tvrtko Ursulin <tvrtko.ursulin@igalia.com>,
	Philipp Stanner <phasta@kernel.org>
Subject: Re: [PATCH 1/2] dma-buf: Assign separate lockdep class to chain lock
Date: Mon, 2 Mar 2026 16:42:28 +0100	[thread overview]
Message-ID: <470ae9bb-e955-4773-b5b5-cc97b5bda20a@amd.com> (raw)
In-Reply-To: <20260302162812.15a0614b@fedora>

On 3/2/26 16:28, Boris Brezillon wrote:
> On Tue, 24 Feb 2026 09:55:43 -0800
> Matthew Brost <matthew.brost@intel.com> wrote:
> 
>> dma_fence_chain_enable_signaling() runs while holding the chain
>> inline_lock and may add callbacks to underlying fences, which takes
>> their inline_lock.
>>
>> Since both locks share the same lockdep class, this valid nesting
>> triggers a recursive locking warning. Assign a distinct lockdep class
>> to the chain inline_lock so lockdep can correctly model the hierarchy.
>>
>> Fixes: a408c0ca0c41 ("dma-buf: use inline lock for the dma-fence-chain")
>> Cc: Christian König <christian.koenig@amd.com>
>> Cc: Tvrtko Ursulin <tvrtko.ursulin@igalia.com>
>> Cc: Philipp Stanner <phasta@kernel.org>
>> Cc: Boris Brezillon <boris.brezillon@collabora.com>
>> Signed-off-by: Matthew Brost <matthew.brost@intel.com>
>> ---
>>  drivers/dma-buf/dma-fence-chain.c | 17 +++++++++++++++++
>>  1 file changed, 17 insertions(+)
>>
>> diff --git a/drivers/dma-buf/dma-fence-chain.c b/drivers/dma-buf/dma-fence-chain.c
>> index a707792b6025..4c2a9f2ce126 100644
>> --- a/drivers/dma-buf/dma-fence-chain.c
>> +++ b/drivers/dma-buf/dma-fence-chain.c
>> @@ -242,6 +242,9 @@ void dma_fence_chain_init(struct dma_fence_chain *chain,
>>  			  struct dma_fence *fence,
>>  			  uint64_t seqno)
>>  {
>> +#if IS_ENABLED(CONFIG_LOCKDEP)
>> +	static struct lock_class_key dma_fence_chain_lock_key;
>> +#endif
>>  	struct dma_fence_chain *prev_chain = to_dma_fence_chain(prev);
>>  	uint64_t context;
>>  
>> @@ -263,6 +266,20 @@ void dma_fence_chain_init(struct dma_fence_chain *chain,
>>  	dma_fence_init64(&chain->base, &dma_fence_chain_ops, NULL,
>>  			 context, seqno);
>>  
>> +#if IS_ENABLED(CONFIG_LOCKDEP)
>> +	/*
>> +	 * dma_fence_chain_enable_signaling() is invoked while holding
>> +	 * chain->base.inline_lock and may call dma_fence_add_callback()
>> +	 * on the underlying fences, which takes their inline_lock.
>> +	 *
>> +	 * Since both locks share the same lockdep class, this legitimate
>> +	 * nesting confuses lockdep and triggers a recursive locking
>> +	 * warning. Assign a separate lockdep class to the chain lock
>> +	 * to model this hierarchy correctly.
>> +	 */
>> +	lockdep_set_class(&chain->base.inline_lock, &dma_fence_chain_lock_key);
>> +#endif
> 
> If we're going to recommend the use of this inline_lock for all new
> dma_fence_ops implementers, as the commit message seems to imply
> 
>> Shared spinlocks have the problem that implementations need to guarantee
>> that the lock lives at least as long all fences referencing them.
>>
>> Using a per-fence spinlock allows completely decoupling spinlock
>> producer and consumer life times, simplifying the handling in most use
>> cases.
> 
> maybe we should have the lock_class_key at the dma_buf_ops level and
> have this lockdep_set_class() automated in __dma_fence_init().

The dma_fence_chain() and dma_fence_array() containers are the only ones who are allowed to nest the lock with other dma_fences. E.g. we have WARN_ON()s in place which fire when you try to stitch together something which won't work.

So everybody else should get a lockdep warning when they try to do nasty things like this because you really can't guarantee lock order between different dma_fence implementations.

Regards,
Christian.

> 
>> +
>>  	/*
>>  	 * Chaining dma_fence_chain container together is only allowed through
>>  	 * the prev fence and not through the contained fence.
> 


  reply	other threads:[~2026-03-02 15:42 UTC|newest]

Thread overview: 8+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-02-24 17:55 [PATCH 1/2] dma-buf: Assign separate lockdep class to chain lock Matthew Brost
2026-02-24 17:55 ` [PATCH 2/2] dma-buf: Assign separate lockdep class to array lock Matthew Brost
2026-02-25  8:26 ` [PATCH 1/2] dma-buf: Assign separate lockdep class to chain lock Christian König
2026-03-02 15:28 ` Boris Brezillon
2026-03-02 15:42   ` Christian König [this message]
2026-03-02 16:39     ` Boris Brezillon
2026-03-02 20:05       ` Matthew Brost
2026-03-03  8:36         ` Boris Brezillon

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=470ae9bb-e955-4773-b5b5-cc97b5bda20a@amd.com \
    --to=christian.koenig@amd.com \
    --cc=boris.brezillon@collabora.com \
    --cc=dri-devel@lists.freedesktop.org \
    --cc=intel-xe@lists.freedesktop.org \
    --cc=matthew.brost@intel.com \
    --cc=phasta@kernel.org \
    --cc=tvrtko.ursulin@igalia.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox