From: Rob Clark <robdclark@gmail.com>
To: dri-devel@lists.freedesktop.org
Cc: freedreno@lists.freedesktop.org, linux-arm-msm@vger.kernel.org,
Connor Abbott <cwabbott0@gmail.com>,
Rob Clark <robdclark@chromium.org>,
Maarten Lankhorst <maarten.lankhorst@linux.intel.com>,
Maxime Ripard <mripard@kernel.org>,
Thomas Zimmermann <tzimmermann@suse.de>,
David Airlie <airlied@gmail.com>, Simona Vetter <simona@ffwll.ch>,
Rob Clark <robdclark@gmail.com>,
Abhinav Kumar <quic_abhinavk@quicinc.com>,
Dmitry Baryshkov <lumag@kernel.org>, Sean Paul <sean@poorly.run>,
Marijn Suijten <marijn.suijten@somainline.org>,
linux-kernel@vger.kernel.org (open list)
Subject: [PATCH v4 03/40] drm/gem: Add ww_acquire_ctx support to drm_gem_lru_scan()
Date: Wed, 14 May 2025 10:53:17 -0700 [thread overview]
Message-ID: <20250514175527.42488-4-robdclark@gmail.com> (raw)
In-Reply-To: <20250514175527.42488-1-robdclark@gmail.com>
From: Rob Clark <robdclark@chromium.org>
If the callback is going to have to attempt to grab more locks, it is
useful to have an ww_acquire_ctx to avoid locking order problems.
Why not use the drm_exec helper instead? Mainly because (a) where
ww_acquire_init() is called is awkward, and (b) we don't really
need to retry after backoff, we can just move on to the next object.
Signed-off-by: Rob Clark <robdclark@chromium.org>
---
drivers/gpu/drm/drm_gem.c | 14 +++++++++++---
drivers/gpu/drm/msm/msm_gem_shrinker.c | 24 +++++++++++++-----------
include/drm/drm_gem.h | 10 ++++++----
3 files changed, 30 insertions(+), 18 deletions(-)
diff --git a/drivers/gpu/drm/drm_gem.c b/drivers/gpu/drm/drm_gem.c
index ee811764c3df..9e3db9a864f8 100644
--- a/drivers/gpu/drm/drm_gem.c
+++ b/drivers/gpu/drm/drm_gem.c
@@ -1460,12 +1460,14 @@ EXPORT_SYMBOL(drm_gem_lru_move_tail);
* @nr_to_scan: The number of pages to try to reclaim
* @remaining: The number of pages left to reclaim, should be initialized by caller
* @shrink: Callback to try to shrink/reclaim the object.
+ * @ticket: Optional ww_acquire_ctx context to use for locking
*/
unsigned long
drm_gem_lru_scan(struct drm_gem_lru *lru,
unsigned int nr_to_scan,
unsigned long *remaining,
- bool (*shrink)(struct drm_gem_object *obj))
+ bool (*shrink)(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket),
+ struct ww_acquire_ctx *ticket)
{
struct drm_gem_lru still_in_lru;
struct drm_gem_object *obj;
@@ -1498,17 +1500,20 @@ drm_gem_lru_scan(struct drm_gem_lru *lru,
*/
mutex_unlock(lru->lock);
+ if (ticket)
+ ww_acquire_init(ticket, &reservation_ww_class);
+
/*
* Note that this still needs to be trylock, since we can
* hit shrinker in response to trying to get backing pages
* for this obj (ie. while it's lock is already held)
*/
- if (!dma_resv_trylock(obj->resv)) {
+ if (!ww_mutex_trylock(&obj->resv->lock, ticket)) {
*remaining += obj->size >> PAGE_SHIFT;
goto tail;
}
- if (shrink(obj)) {
+ if (shrink(obj, ticket)) {
freed += obj->size >> PAGE_SHIFT;
/*
@@ -1522,6 +1527,9 @@ drm_gem_lru_scan(struct drm_gem_lru *lru,
dma_resv_unlock(obj->resv);
+ if (ticket)
+ ww_acquire_fini(ticket);
+
tail:
drm_gem_object_put(obj);
mutex_lock(lru->lock);
diff --git a/drivers/gpu/drm/msm/msm_gem_shrinker.c b/drivers/gpu/drm/msm/msm_gem_shrinker.c
index 07ca4ddfe4e3..de185fc34084 100644
--- a/drivers/gpu/drm/msm/msm_gem_shrinker.c
+++ b/drivers/gpu/drm/msm/msm_gem_shrinker.c
@@ -44,7 +44,7 @@ msm_gem_shrinker_count(struct shrinker *shrinker, struct shrink_control *sc)
}
static bool
-purge(struct drm_gem_object *obj)
+purge(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket)
{
if (!is_purgeable(to_msm_bo(obj)))
return false;
@@ -58,7 +58,7 @@ purge(struct drm_gem_object *obj)
}
static bool
-evict(struct drm_gem_object *obj)
+evict(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket)
{
if (is_unevictable(to_msm_bo(obj)))
return false;
@@ -79,21 +79,21 @@ wait_for_idle(struct drm_gem_object *obj)
}
static bool
-active_purge(struct drm_gem_object *obj)
+active_purge(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket)
{
if (!wait_for_idle(obj))
return false;
- return purge(obj);
+ return purge(obj, ticket);
}
static bool
-active_evict(struct drm_gem_object *obj)
+active_evict(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket)
{
if (!wait_for_idle(obj))
return false;
- return evict(obj);
+ return evict(obj, ticket);
}
static unsigned long
@@ -102,7 +102,7 @@ msm_gem_shrinker_scan(struct shrinker *shrinker, struct shrink_control *sc)
struct msm_drm_private *priv = shrinker->private_data;
struct {
struct drm_gem_lru *lru;
- bool (*shrink)(struct drm_gem_object *obj);
+ bool (*shrink)(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket);
bool cond;
unsigned long freed;
unsigned long remaining;
@@ -122,8 +122,9 @@ msm_gem_shrinker_scan(struct shrinker *shrinker, struct shrink_control *sc)
continue;
stages[i].freed =
drm_gem_lru_scan(stages[i].lru, nr,
- &stages[i].remaining,
- stages[i].shrink);
+ &stages[i].remaining,
+ stages[i].shrink,
+ NULL);
nr -= stages[i].freed;
freed += stages[i].freed;
remaining += stages[i].remaining;
@@ -164,7 +165,7 @@ msm_gem_shrinker_shrink(struct drm_device *dev, unsigned long nr_to_scan)
static const int vmap_shrink_limit = 15;
static bool
-vmap_shrink(struct drm_gem_object *obj)
+vmap_shrink(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket)
{
if (!is_vunmapable(to_msm_bo(obj)))
return false;
@@ -192,7 +193,8 @@ msm_gem_shrinker_vmap(struct notifier_block *nb, unsigned long event, void *ptr)
unmapped += drm_gem_lru_scan(lrus[idx],
vmap_shrink_limit - unmapped,
&remaining,
- vmap_shrink);
+ vmap_shrink,
+ NULL);
}
*(unsigned long *)ptr += unmapped;
diff --git a/include/drm/drm_gem.h b/include/drm/drm_gem.h
index fdae947682cd..0e2c476df731 100644
--- a/include/drm/drm_gem.h
+++ b/include/drm/drm_gem.h
@@ -555,10 +555,12 @@ void drm_gem_lru_init(struct drm_gem_lru *lru, struct mutex *lock);
void drm_gem_lru_remove(struct drm_gem_object *obj);
void drm_gem_lru_move_tail_locked(struct drm_gem_lru *lru, struct drm_gem_object *obj);
void drm_gem_lru_move_tail(struct drm_gem_lru *lru, struct drm_gem_object *obj);
-unsigned long drm_gem_lru_scan(struct drm_gem_lru *lru,
- unsigned int nr_to_scan,
- unsigned long *remaining,
- bool (*shrink)(struct drm_gem_object *obj));
+unsigned long
+drm_gem_lru_scan(struct drm_gem_lru *lru,
+ unsigned int nr_to_scan,
+ unsigned long *remaining,
+ bool (*shrink)(struct drm_gem_object *obj, struct ww_acquire_ctx *ticket),
+ struct ww_acquire_ctx *ticket);
int drm_gem_evict(struct drm_gem_object *obj);
--
2.49.0
next prev parent reply other threads:[~2025-05-14 17:56 UTC|newest]
Thread overview: 66+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-05-14 17:53 [PATCH v4 00/40] drm/msm: sparse / "VM_BIND" support Rob Clark
2025-05-14 17:53 ` [PATCH v4 01/40] drm/gpuvm: Don't require obj lock in destructor path Rob Clark
2025-05-15 8:54 ` Danilo Krummrich
2025-05-15 9:06 ` Danilo Krummrich
2025-05-15 17:35 ` Rob Clark
2025-05-15 17:55 ` Danilo Krummrich
2025-05-15 21:57 ` Rob Clark
2025-05-16 9:01 ` Danilo Krummrich
2025-05-16 16:20 ` Rob Clark
2025-05-20 21:25 ` Dave Airlie
2025-05-20 21:52 ` Rob Clark
2025-05-20 22:31 ` Dave Airlie
2025-05-20 22:56 ` Rob Clark
2025-05-23 2:51 ` Rob Clark
2025-05-23 6:28 ` Danilo Krummrich
2025-05-14 17:53 ` [PATCH v4 02/40] drm/gpuvm: Allow VAs to hold soft reference to BOs Rob Clark
2025-05-15 9:00 ` Danilo Krummrich
2025-05-15 14:59 ` Rob Clark
2025-05-15 15:30 ` Danilo Krummrich
2025-05-15 17:34 ` Rob Clark
2025-05-15 17:51 ` Danilo Krummrich
2025-05-15 20:10 ` Rob Clark
2025-05-14 17:53 ` Rob Clark [this message]
2025-05-14 17:53 ` [PATCH v4 04/40] drm/sched: Add enqueue credit limit Rob Clark
2025-05-14 17:53 ` [PATCH v4 05/40] iommu/io-pgtable-arm: Add quirk to quiet WARN_ON() Rob Clark
2025-05-15 14:33 ` Will Deacon
2025-05-15 14:48 ` Rob Clark
2025-05-20 11:31 ` Will Deacon
2025-05-20 13:06 ` Robin Murphy
2025-05-20 14:06 ` Will Deacon
2025-05-14 17:53 ` [PATCH v4 06/40] drm/msm: Rename msm_file_private -> msm_context Rob Clark
2025-05-14 17:53 ` [PATCH v4 07/40] drm/msm: Improve msm_context comments Rob Clark
2025-05-14 17:53 ` [PATCH v4 08/40] drm/msm: Rename msm_gem_address_space -> msm_gem_vm Rob Clark
2025-05-14 17:53 ` [PATCH v4 09/40] drm/msm: Remove vram carveout support Rob Clark
2025-05-14 17:53 ` [PATCH v4 10/40] drm/msm: Collapse vma allocation and initialization Rob Clark
2025-05-14 17:53 ` [PATCH v4 11/40] drm/msm: Collapse vma close and delete Rob Clark
2025-05-14 17:53 ` [PATCH v4 12/40] drm/msm: Don't close VMAs on purge Rob Clark
2025-05-14 17:53 ` [PATCH v4 13/40] drm/msm: drm_gpuvm conversion Rob Clark
2025-05-14 17:53 ` [PATCH v4 14/40] drm/msm: Convert vm locking Rob Clark
2025-05-14 17:53 ` [PATCH v4 15/40] drm/msm: Use drm_gpuvm types more Rob Clark
2025-05-14 17:53 ` [PATCH v4 16/40] drm/msm: Split out helper to get iommu prot flags Rob Clark
2025-05-14 17:53 ` [PATCH v4 17/40] drm/msm: Add mmu support for non-zero offset Rob Clark
2025-05-14 17:53 ` [PATCH v4 18/40] drm/msm: Add PRR support Rob Clark
2025-05-14 17:53 ` [PATCH v4 19/40] drm/msm: Rename msm_gem_vma_purge() -> _unmap() Rob Clark
2025-05-14 17:53 ` [PATCH v4 20/40] drm/msm: Drop queued submits on lastclose() Rob Clark
2025-05-14 17:53 ` [PATCH v4 21/40] drm/msm: Lazily create context VM Rob Clark
2025-05-14 17:53 ` [PATCH v4 22/40] drm/msm: Add opt-in for VM_BIND Rob Clark
2025-05-14 17:53 ` [PATCH v4 23/40] drm/msm: Mark VM as unusable on GPU hangs Rob Clark
2025-05-14 17:53 ` [PATCH v4 24/40] drm/msm: Add _NO_SHARE flag Rob Clark
2025-05-14 17:53 ` [PATCH v4 25/40] drm/msm: Crashdump prep for sparse mappings Rob Clark
2025-05-14 17:53 ` [PATCH v4 26/40] drm/msm: rd dumping " Rob Clark
2025-05-14 17:53 ` [PATCH v4 27/40] drm/msm: Crashdec support for sparse Rob Clark
2025-05-14 17:53 ` [PATCH v4 28/40] drm/msm: rd dumping " Rob Clark
2025-05-14 17:53 ` [PATCH v4 29/40] drm/msm: Extract out syncobj helpers Rob Clark
2025-05-14 17:53 ` [PATCH v4 30/40] drm/msm: Use DMA_RESV_USAGE_BOOKKEEP/KERNEL Rob Clark
2025-05-14 17:53 ` [PATCH v4 31/40] drm/msm: Add VM_BIND submitqueue Rob Clark
2025-05-14 17:53 ` [PATCH v4 32/40] drm/msm: Support IO_PGTABLE_QUIRK_NO_WARN_ON Rob Clark
2025-05-14 17:53 ` [PATCH v4 33/40] drm/msm: Support pgtable preallocation Rob Clark
2025-05-14 17:53 ` [PATCH v4 34/40] drm/msm: Split out map/unmap ops Rob Clark
2025-05-14 17:53 ` [PATCH v4 35/40] drm/msm: Add VM_BIND ioctl Rob Clark
2025-05-14 17:53 ` [PATCH v4 36/40] drm/msm: Add VM logging for VM_BIND updates Rob Clark
2025-05-14 17:53 ` [PATCH v4 37/40] drm/msm: Add VMA unmap reason Rob Clark
2025-05-14 17:53 ` [PATCH v4 38/40] drm/msm: Add mmu prealloc tracepoint Rob Clark
2025-05-14 17:53 ` [PATCH v4 39/40] drm/msm: use trylock for debugfs Rob Clark
2025-05-14 17:53 ` [PATCH v4 40/40] drm/msm: Bump UAPI version Rob Clark
-- strict thread matches above, loose matches on Subject: below --
2025-05-14 16:58 [PATCH v4 00/40] drm/msm: sparse / "VM_BIND" support Rob Clark
2025-05-14 16:59 ` [PATCH v4 03/40] drm/gem: Add ww_acquire_ctx support to drm_gem_lru_scan() Rob Clark
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20250514175527.42488-4-robdclark@gmail.com \
--to=robdclark@gmail.com \
--cc=airlied@gmail.com \
--cc=cwabbott0@gmail.com \
--cc=dri-devel@lists.freedesktop.org \
--cc=freedreno@lists.freedesktop.org \
--cc=linux-arm-msm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=lumag@kernel.org \
--cc=maarten.lankhorst@linux.intel.com \
--cc=marijn.suijten@somainline.org \
--cc=mripard@kernel.org \
--cc=quic_abhinavk@quicinc.com \
--cc=robdclark@chromium.org \
--cc=sean@poorly.run \
--cc=simona@ffwll.ch \
--cc=tzimmermann@suse.de \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox