From: "Thomas Hellström" <thomas.hellstrom@linux.intel.com>
To: "Yadav, Arvind" <arvind.yadav@intel.com>,
Matthew Brost <matthew.brost@intel.com>
Cc: intel-xe@lists.freedesktop.org, himal.prasad.ghimiray@intel.com,
pallavi.mishra@intel.com
Subject: Re: [PATCH v4 5/8] drm/xe/vm: Prevent binding of purged buffer objects
Date: Fri, 23 Jan 2026 13:37:19 +0100 [thread overview]
Message-ID: <bedb36f03c7e82ffca688a1ba69929a343bc4f37.camel@linux.intel.com> (raw)
In-Reply-To: <7614efa2-b126-4fa6-8213-a8cd764e4419@intel.com>
On Fri, 2026-01-23 at 11:11 +0530, Yadav, Arvind wrote:
>
> On 20-01-2026 22:57, Matthew Brost wrote:
> > On Tue, Jan 20, 2026 at 11:38:51AM +0530, Arvind Yadav wrote:
> > > Add check_purged parameter to vma_lock_and_validate() to block
> > > new mapping operations on purged BOs while allowing cleanup
> > > operations to proceed.
> > >
> > > Purged BOs have their backing pages freed by the kernel. New
> > > mapping operations (MAP, PREFETCH, REMAP) must be rejected with
> > > -EINVAL to prevent GPU access to invalid memory. Cleanup
> > > operations (UNMAP) must be allowed so applications can release
> > > resources after detecting purge via the retained field.
> > >
> > > REMAP operations require mixed handling - reject new prev/next
> > > VMAs if the BO is purged, but allow the unmap portion to proceed
> > > for cleanup.
> > >
> > > The check_purged parameter distinguishes between these cases:
> > > true for new mappings (must reject), false for cleanup (allow).
> > >
> > > v2:
> > > - Clarify that purged BOs are permanently invalid (i915
> > > semantics)
> > > - Remove incorrect claim about madvise(WILLNEED) restoring
> > > purged BOs
> > >
> > > v3:
> > > - Move xe_bo_is_purged check under vma_lock_and_validate
> > > (Matthew Brost)
> > > - Add check_purged parameter to distinguish new mappings from
> > > cleanup
> > > - Allow UNMAP operations to prevent resource leaks
> > > - Handle REMAP operation's dual nature (cleanup + new
> > > mappings)
> > >
> > > Cc: Matthew Brost<matthew.brost@intel.com>
> > > Cc: Thomas Hellström<thomas.hellstrom@linux.intel.com>
> > > Cc: Himal Prasad Ghimiray<himal.prasad.ghimiray@intel.com>
> > > Signed-off-by: Arvind Yadav<arvind.yadav@intel.com>
> > > ---
> > > drivers/gpu/drm/xe/xe_vm.c | 20 +++++++++++++-------
> > > 1 file changed, 13 insertions(+), 7 deletions(-)
> > >
> > > diff --git a/drivers/gpu/drm/xe/xe_vm.c
> > > b/drivers/gpu/drm/xe/xe_vm.c
> > > index c3a5fe76ff96..f250daae3012 100644
> > > --- a/drivers/gpu/drm/xe/xe_vm.c
> > > +++ b/drivers/gpu/drm/xe/xe_vm.c
> > > @@ -2883,7 +2883,7 @@ static void vm_bind_ioctl_ops_unwind(struct
> > > xe_vm *vm,
> > > }
> > >
> > > static int vma_lock_and_validate(struct drm_exec *exec, struct
> > > xe_vma *vma,
> > > - bool res_evict, bool validate)
> > > + bool res_evict, bool validate,
> > > bool check_purged)
> > It probably time to add something like this to avoid transposing
> > arguments.
> >
> > struct lock_and_validate_flags {
> > bool res_evict;
> > bool validate;
> > bool check_purged;
> > };
> >
> > Logic in the patch looks correct though.
>
>
> Noted, I will add "struct xe_lock_and_validate_flags" Thanks, Arvind
Note that if you follow the pattern of struct xe_bo_shink_flags,
passing the struct as a const and using
struct lock_and_validate_flags {
u32 res_evict : 1;
u32 validate : 1;
u32 check_purged : 1;
};
This will be more or less equivalent to passing a bit-field with type-
checking.
Some reviewers frown on using "bool" in compound types although we
accept that in the xe driver.
Otherwise patch LGTM as well.
/Thomas
>
> >
> > Matt
> >
> > > {
> > > struct xe_bo *bo = xe_vma_bo(vma);
> > > struct xe_vm *vm = xe_vma_vm(vma);
> > > @@ -2892,6 +2892,11 @@ static int vma_lock_and_validate(struct
> > > drm_exec *exec, struct xe_vma *vma,
> > > if (bo) {
> > > if (!bo->vm)
> > > err = drm_exec_lock_obj(exec, &bo-
> > > >ttm.base);
> > > +
> > > + /* Reject new mappings to purged BOs; allow
> > > cleanup operations */
> > > + if (!err && check_purged && xe_bo_is_purged(bo))
> > > + err = -EINVAL;
> > > +
> > > if (!err && validate)
> > > err = xe_bo_validate(bo, vm,
> > >
> > > !xe_vm_in_preempt_fence_mode(vm) &&
> > > @@ -2990,7 +2995,8 @@ static int op_lock_and_prep(struct drm_exec
> > > *exec, struct xe_vm *vm,
> > > err = vma_lock_and_validate(exec, op-
> > > >map.vma,
> > > res_evict,
> > >
> > > !xe_vm_in_fault_mode(vm) ||
> > > - op-
> > > >map.immediate);
> > > + op-
> > > >map.immediate,
> > > + true);
> > > break;
> > > case DRM_GPUVA_OP_REMAP:
> > > err = check_ufence(gpuva_to_vma(op-
> > > >base.remap.unmap->va));
> > > @@ -2999,13 +3005,13 @@ static int op_lock_and_prep(struct
> > > drm_exec *exec, struct xe_vm *vm,
> > >
> > > err = vma_lock_and_validate(exec,
> > > gpuva_to_vma(op-
> > > >base.remap.unmap->va),
> > > - res_evict, false);
> > > + res_evict, false,
> > > false);
> > > if (!err && op->remap.prev)
> > > err = vma_lock_and_validate(exec, op-
> > > >remap.prev,
> > > - res_evict,
> > > true);
> > > + res_evict,
> > > true, true);
> > > if (!err && op->remap.next)
> > > err = vma_lock_and_validate(exec, op-
> > > >remap.next,
> > > - res_evict,
> > > true);
> > > + res_evict,
> > > true, true);
> > > break;
> > > case DRM_GPUVA_OP_UNMAP:
> > > err = check_ufence(gpuva_to_vma(op-
> > > >base.unmap.va));
> > > @@ -3014,7 +3020,7 @@ static int op_lock_and_prep(struct drm_exec
> > > *exec, struct xe_vm *vm,
> > >
> > > err = vma_lock_and_validate(exec,
> > > gpuva_to_vma(op-
> > > >base.unmap.va),
> > > - res_evict, false);
> > > + res_evict, false,
> > > false);
> > > break;
> > > case DRM_GPUVA_OP_PREFETCH:
> > > {
> > > @@ -3029,7 +3035,7 @@ static int op_lock_and_prep(struct drm_exec
> > > *exec, struct xe_vm *vm,
> > >
> > > err = vma_lock_and_validate(exec,
> > > gpuva_to_vma(op-
> > > >base.prefetch.va),
> > > - res_evict, false);
> > > + res_evict, false,
> > > true);
> > > if (!err && !xe_vma_has_no_bo(vma))
> > > err = xe_bo_migrate(xe_vma_bo(vma),
> > >
> > > region_to_mem_type[region],
> > > --
> > > 2.43.0
next prev parent reply other threads:[~2026-01-23 12:37 UTC|newest]
Thread overview: 37+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-01-20 6:08 [PATCH v4 0/8] drm/xe/madvise: Add support for purgeable buffer objects Arvind Yadav
2026-01-20 6:08 ` [PATCH v4 1/8] drm/xe/uapi: Add UAPI " Arvind Yadav
2026-01-20 17:20 ` Matthew Brost
2026-01-21 18:42 ` Vivi, Rodrigo
2026-01-20 6:08 ` [PATCH v4 2/8] drm/xe/bo: Add purgeable bo state tracking and field madv to xe_bo Arvind Yadav
2026-01-20 17:45 ` Matthew Brost
2026-01-21 5:30 ` Yadav, Arvind
2026-01-22 15:05 ` Thomas Hellström
2026-01-20 6:08 ` [PATCH v4 3/8] drm/xe/madvise: Implement purgeable buffer object support Arvind Yadav
2026-01-20 16:58 ` Matthew Brost
2026-01-20 17:15 ` Matthew Brost
2026-01-21 8:24 ` Yadav, Arvind
2026-01-22 15:30 ` Thomas Hellström
2026-01-30 8:13 ` Yadav, Arvind
2026-01-20 17:44 ` Matthew Brost
2026-01-20 6:08 ` [PATCH v4 4/8] drm/xe/bo: Handle CPU faults on purged buffer objects Arvind Yadav
2026-01-20 17:23 ` Matthew Brost
2026-01-22 15:54 ` Thomas Hellström
2026-01-20 6:08 ` [PATCH v4 5/8] drm/xe/vm: Prevent binding of " Arvind Yadav
2026-01-20 17:27 ` Matthew Brost
2026-01-23 5:41 ` Yadav, Arvind
2026-01-23 12:37 ` Thomas Hellström [this message]
2026-01-30 8:17 ` Yadav, Arvind
2026-01-20 6:08 ` [PATCH v4 6/8] drm/xe/madvise: Implement per-VMA purgeable state tracking Arvind Yadav
2026-01-20 17:41 ` Matthew Brost
2026-01-21 5:11 ` Yadav, Arvind
2026-01-23 13:07 ` Thomas Hellström
2026-01-20 6:08 ` [PATCH v4 7/8] drm/xe/madvise: Block imported and exported dma-bufs Arvind Yadav
2026-01-20 17:51 ` Matthew Brost
2026-01-23 13:31 ` Thomas Hellström
2026-01-30 8:22 ` Yadav, Arvind
2026-01-30 8:59 ` Thomas Hellström
2026-01-20 6:08 ` [PATCH v4 8/8] drm/xe/bo: Add purgeable shrinker state helpers Arvind Yadav
2026-01-20 17:58 ` Matthew Brost
2026-01-23 13:42 ` Thomas Hellström
2026-01-20 6:14 ` ✗ CI.checkpatch: warning for drm/xe/madvise: Add support for purgeable buffer objects (rev5) Patchwork
2026-01-20 6:16 ` ✗ CI.KUnit: failure " Patchwork
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=bedb36f03c7e82ffca688a1ba69929a343bc4f37.camel@linux.intel.com \
--to=thomas.hellstrom@linux.intel.com \
--cc=arvind.yadav@intel.com \
--cc=himal.prasad.ghimiray@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=matthew.brost@intel.com \
--cc=pallavi.mishra@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox