From: Matthew Brost <matthew.brost@intel.com>
To: "Summers, Stuart" <stuart.summers@intel.com>
Cc: "intel-xe@lists.freedesktop.org" <intel-xe@lists.freedesktop.org>,
"Ghimiray, Himal Prasad" <himal.prasad.ghimiray@intel.com>,
"Yadav, Arvind" <arvind.yadav@intel.com>,
"thomas.hellstrom@linux.intel.com"
<thomas.hellstrom@linux.intel.com>,
"Dugast, Francois" <francois.dugast@intel.com>
Subject: Re: [PATCH v3 09/25] drm/xe: Add struct xe_pt_job_ops
Date: Tue, 3 Mar 2026 15:28:16 -0800 [thread overview]
Message-ID: <aadukHCYIPthpwrg@lstrano-desk.jf.intel.com> (raw)
In-Reply-To: <15113d16313e71e158e39c8db7dbc83ae8bcc76c.camel@intel.com>
On Tue, Mar 03, 2026 at 04:26:01PM -0700, Summers, Stuart wrote:
> On Fri, 2026-02-27 at 17:34 -0800, Matthew Brost wrote:
> > Add struct xe_pt_job_ops, a dynamically refcounted object that
> > contains
> > the information required to issue a CPU bind via a job after the
> > initial
> > bind IOCTL returns.
> >
> > Signed-off-by: Matthew Brost <matthew.brost@intel.com>
> > ---
> > drivers/gpu/drm/xe/xe_migrate.c | 10 +--
> > drivers/gpu/drm/xe/xe_pt.c | 132 +++++++++++++++++++++++++++--
> > --
> > drivers/gpu/drm/xe/xe_pt.h | 4 +
> > drivers/gpu/drm/xe/xe_pt_types.h | 27 +++++--
> > drivers/gpu/drm/xe/xe_vm.c | 10 +--
> > 5 files changed, 149 insertions(+), 34 deletions(-)
> >
> > diff --git a/drivers/gpu/drm/xe/xe_migrate.c
> > b/drivers/gpu/drm/xe/xe_migrate.c
> > index 69e6e3135ec6..cd6802642ef3 100644
> > --- a/drivers/gpu/drm/xe/xe_migrate.c
> > +++ b/drivers/gpu/drm/xe/xe_migrate.c
> > @@ -1771,7 +1771,7 @@ xe_migrate_update_pgtables_cpu(struct
> > xe_migrate *m,
> > }
> >
> > xe_migrate_update_pgtables_cpu_execute(vm, m->tile, ops,
> > - pt_update_ops->ops,
> > + pt_update_ops-
> > >pt_job_ops->ops,
> > pt_update_ops-
> > >num_ops);
> >
> > return dma_fence_get_stub();
> > @@ -1798,7 +1798,7 @@ __xe_migrate_update_pgtables(struct xe_migrate
> > *m,
> > bool usm = is_migrate && xe->info.has_usm;
> >
> > for (i = 0; i < pt_update_ops->num_ops; ++i) {
> > - struct xe_vm_pgtable_update_op *pt_op =
> > &pt_update_ops->ops[i];
> > + struct xe_vm_pgtable_update_op *pt_op =
> > &pt_update_ops->pt_job_ops->ops[i];
> > struct xe_vm_pgtable_update *updates = pt_op-
> > >entries;
> >
> > num_updates += pt_op->num_entries;
> > @@ -1867,7 +1867,7 @@ __xe_migrate_update_pgtables(struct xe_migrate
> > *m,
> >
> > for (; i < pt_update_ops->num_ops; ++i) {
> > struct xe_vm_pgtable_update_op *pt_op
> > =
> > - &pt_update_ops->ops[i];
> > + &pt_update_ops->pt_job_ops-
> > >ops[i];
> > struct xe_vm_pgtable_update *updates
> > = pt_op->entries;
> >
> > for (; j < pt_op->num_entries; ++j,
> > ++current_update, ++idx) {
> > @@ -1904,7 +1904,7 @@ __xe_migrate_update_pgtables(struct xe_migrate
> > *m,
> > (page_ofs / sizeof(u64)) * XE_PAGE_SIZE;
> > for (i = 0; i < pt_update_ops->num_ops; ++i) {
> > struct xe_vm_pgtable_update_op *pt_op =
> > - &pt_update_ops->ops[i];
> > + &pt_update_ops->pt_job_ops->ops[i];
> > struct xe_vm_pgtable_update *updates = pt_op-
> > >entries;
> >
> > for (j = 0; j < pt_op->num_entries; ++j) {
> > @@ -1922,7 +1922,7 @@ __xe_migrate_update_pgtables(struct xe_migrate
> > *m,
> >
> > for (i = 0; i < pt_update_ops->num_ops; ++i) {
> > struct xe_vm_pgtable_update_op *pt_op =
> > - &pt_update_ops->ops[i];
> > + &pt_update_ops->pt_job_ops->ops[i];
> > struct xe_vm_pgtable_update *updates = pt_op-
> > >entries;
> >
> > for (j = 0; j < pt_op->num_entries; ++j)
> > diff --git a/drivers/gpu/drm/xe/xe_pt.c b/drivers/gpu/drm/xe/xe_pt.c
> > index 1f24eff75185..6b56e62a35c1 100644
> > --- a/drivers/gpu/drm/xe/xe_pt.c
> > +++ b/drivers/gpu/drm/xe/xe_pt.c
> > @@ -204,7 +204,9 @@ unsigned int xe_pt_shift(unsigned int level)
> > * and finally frees @pt. TODO: Can we remove the @flags argument?
> > */
> > void xe_pt_destroy(struct xe_pt *pt, u32 flags, struct llist_head
> > *deferred)
> > +
>
> Extra line..
>
+1
> > {
> > + bool added = false;
> > int i;
> >
> > if (!pt)
> > @@ -212,7 +214,20 @@ void xe_pt_destroy(struct xe_pt *pt, u32 flags,
> > struct llist_head *deferred)
> >
> > XE_WARN_ON(!list_empty(&pt->bo->ttm.base.gpuva.list));
> > xe_bo_unpin(pt->bo);
> > - xe_bo_put_deferred(pt->bo, deferred, NULL);
> > + xe_bo_put_deferred(pt->bo, deferred, &added);
> > + if (added) {
> > + xe_assert(pt->bo->vm->xe, !kref_read(&pt->bo-
> > >ttm.base.refcount));
> > +
> > + /*
> > + * We need the VM present until the BO is destroyed
> > as it shares
> > + * a dma-resv and BO destroy is async. Reinit BO
> > refcount so
> > + * xe_bo_put_async can be used when the PT job ops
> > refcount goes
> > + * to zero.
> > + */
> > + xe_vm_get(pt->bo->vm);
> > + pt->bo->flags |= XE_BO_FLAG_PUT_VM_ASYNC;
> > + kref_init(&pt->bo->ttm.base.refcount);
> > + }
> >
> > if (pt->level > 0 && pt->num_live) {
> > struct xe_pt_dir *pt_dir = as_xe_pt_dir(pt);
> > @@ -1884,13 +1899,13 @@ xe_pt_commit_prepare_unbind(struct xe_vma
> > *vma,
> > static struct xe_vm_pgtable_update_op *
> > to_pt_op(struct xe_vm_pgtable_update_ops *pt_update_ops, u32 op_idx)
> > {
> > - return &pt_update_ops->ops[op_idx];
> > + return &pt_update_ops->pt_job_ops->ops[op_idx];
> > }
> >
> > static u32
> > get_current_op(struct xe_vm_pgtable_update_ops *pt_update_ops)
> > {
> > - return pt_update_ops->current_op;
> > + return pt_update_ops->pt_job_ops->current_op;
> > }
> >
> > static struct xe_vm_pgtable_update_op *
> > @@ -1902,7 +1917,7 @@ to_current_pt_op(struct
> > xe_vm_pgtable_update_ops *pt_update_ops)
> > static void
> > incr_current_op(struct xe_vm_pgtable_update_ops *pt_update_ops)
> > {
> > - ++pt_update_ops->current_op;
> > + ++pt_update_ops->pt_job_ops->current_op;
> > }
> >
> > static void
> > @@ -2264,7 +2279,6 @@ static int op_prepare(struct xe_vm *vm,
> > static void
> > xe_pt_update_ops_init(struct xe_vm_pgtable_update_ops
> > *pt_update_ops)
> > {
> > - init_llist_head(&pt_update_ops->deferred);
> > pt_update_ops->start = ~0x0ull;
> > pt_update_ops->last = 0x0ull;
> > xe_page_reclaim_list_init(&pt_update_ops->prl);
> > @@ -2612,7 +2626,8 @@ xe_pt_update_ops_run(struct xe_tile *tile,
> > struct xe_vma_ops *vops)
> > to_pt_op(pt_update_ops, i);
> >
> > xe_pt_commit(pt_op->vma, pt_op->entries,
> > - pt_op->num_entries, &pt_update_ops-
> > >deferred);
> > + pt_op->num_entries,
> > + &pt_update_ops->pt_job_ops->deferred);
> > pt_op->vma = NULL; /* skip in
> > xe_pt_update_ops_abort */
> > }
> >
> > @@ -2700,19 +2715,8 @@ void xe_pt_update_ops_fini(struct xe_tile
> > *tile, struct xe_vma_ops *vops)
> > {
> > struct xe_vm_pgtable_update_ops *pt_update_ops =
> > &vops->pt_update_ops[tile->id];
> > - int i;
> >
> > xe_page_reclaim_entries_put(pt_update_ops->prl.entries);
> > -
> > - lockdep_assert_held(&vops->vm->lock);
> > - xe_vm_assert_held(vops->vm);
> > -
> > - for (i = 0; i < pt_update_ops->current_op; ++i) {
> > - struct xe_vm_pgtable_update_op *pt_op =
> > &pt_update_ops->ops[i];
> > -
> > - xe_pt_free_bind(pt_op->entries, pt_op->num_entries);
> > - }
> > - xe_bo_put_commit(&vops->pt_update_ops[tile->id].deferred);
> > }
> >
> > /**
> > @@ -2749,3 +2753,97 @@ void xe_pt_update_ops_abort(struct xe_tile
> > *tile, struct xe_vma_ops *vops)
> >
> > xe_pt_update_ops_fini(tile, vops);
> > }
> > +
> > +/**
> > + * xe_pt_job_ops_alloc() - Allocate PT job ops
> > + * @num_ops: Number of VM PT update ops
> > + *
> > + * Allocate PT job ops and internal array of VM PT update ops.
> > + *
> > + * Return: Pointer to PT job ops or NULL
> > + */
> > +struct xe_pt_job_ops *xe_pt_job_ops_alloc(u32 num_ops)
> > +{
> > + struct xe_pt_job_ops *pt_job_ops;
> > +
> > + pt_job_ops = kmalloc(sizeof(*pt_job_ops), GFP_KERNEL);
> > + if (!pt_job_ops)
> > + return NULL;
> > +
> > + pt_job_ops->ops = kvmalloc_array(num_ops, sizeof(*pt_job_ops-
> > >ops),
> > + GFP_KERNEL);
> > + if (!pt_job_ops->ops) {
> > + kvfree(pt_job_ops);
>
> This should be kfree right?
>
Yes, will fix.
Matt
> Thanks,
> Stuart
>
> > + return NULL;
> > + }
> > +
> > + pt_job_ops->current_op = 0;
> > + kref_init(&pt_job_ops->refcount);
> > + init_llist_head(&pt_job_ops->deferred);
> > +
> > + return pt_job_ops;
> > +}
> > +
> > +/**
> > + * xe_pt_job_ops_get() - Get PT job ops
> > + * @pt_job_ops: PT job ops to get
> > + *
> > + * Take a reference to PT job ops
> > + *
> > + * Return: Pointer to PT job ops or NULL
> > + */
> > +struct xe_pt_job_ops *xe_pt_job_ops_get(struct xe_pt_job_ops
> > *pt_job_ops)
> > +{
> > + if (pt_job_ops)
> > + kref_get(&pt_job_ops->refcount);
> > +
> > + return pt_job_ops;
> > +}
> > +
> > +static void xe_pt_update_ops_free(struct xe_vm_pgtable_update_op
> > *pt_op,
> > + u32 num_ops)
> > +{
> > + u32 i;
> > +
> > + for (i = 0; i < num_ops; ++i, ++pt_op)
> > + xe_pt_free_bind(pt_op->entries, pt_op->num_entries);
> > +}
> > +
> > +static void xe_pt_job_ops_destroy(struct kref *ref)
> > +{
> > + struct xe_pt_job_ops *pt_job_ops =
> > + container_of(ref, struct xe_pt_job_ops, refcount);
> > + struct llist_node *freed;
> > + struct xe_bo *bo, *next;
> > +
> > + xe_pt_update_ops_free(pt_job_ops->ops,
> > + pt_job_ops->current_op);
> > +
> > + freed = llist_del_all(&pt_job_ops->deferred);
> > + if (freed) {
> > + llist_for_each_entry_safe(bo, next, freed, freed)
> > + /*
> > + * If called from run_job, we are in the dma-
> > fencing
> > + * path and cannot take dma-resv locks so use
> > an async
> > + * put.
> > + */
> > + xe_bo_put_async(bo);
> > + }
> > +
> > + kvfree(pt_job_ops->ops);
> > + kfree(pt_job_ops);
> > +}
> > +
> > +/**
> > + * xe_pt_job_ops_put() - Put PT job ops
> > + * @pt_job_ops: PT job ops to put
> > + *
> > + * Drop a reference to PT job ops
> > + */
> > +void xe_pt_job_ops_put(struct xe_pt_job_ops *pt_job_ops)
> > +{
> > + if (!pt_job_ops)
> > + return;
> > +
> > + kref_put(&pt_job_ops->refcount, xe_pt_job_ops_destroy);
> > +}
> > diff --git a/drivers/gpu/drm/xe/xe_pt.h b/drivers/gpu/drm/xe/xe_pt.h
> > index 4daeebaab5a1..5faddb8e700c 100644
> > --- a/drivers/gpu/drm/xe/xe_pt.h
> > +++ b/drivers/gpu/drm/xe/xe_pt.h
> > @@ -49,4 +49,8 @@ bool xe_pt_zap_ptes(struct xe_tile *tile, struct
> > xe_vma *vma);
> > bool xe_pt_zap_ptes_range(struct xe_tile *tile, struct xe_vm *vm,
> > struct xe_svm_range *range);
> >
> > +struct xe_pt_job_ops *xe_pt_job_ops_alloc(u32 num_ops);
> > +struct xe_pt_job_ops *xe_pt_job_ops_get(struct xe_pt_job_ops
> > *pt_job_ops);
> > +void xe_pt_job_ops_put(struct xe_pt_job_ops *pt_job_ops);
> > +
> > #endif
> > diff --git a/drivers/gpu/drm/xe/xe_pt_types.h
> > b/drivers/gpu/drm/xe/xe_pt_types.h
> > index 84b51d3762a4..92d50573ed1d 100644
> > --- a/drivers/gpu/drm/xe/xe_pt_types.h
> > +++ b/drivers/gpu/drm/xe/xe_pt_types.h
> > @@ -91,12 +91,29 @@ struct xe_vm_pgtable_update_op {
> > bool rebind;
> > };
> >
> > +/**
> > + * struct xe_pt_job_ops - Page-table update operations (dynamically
> > allocated)
> > + *
> > + * This is the portion of &struct xe_vma_ops and
> > + * &struct xe_vm_pgtable_update_ops that is dynamically allocated,
> > as it
> > + * must remain valid until the associated bind job completes. A
> > reference
> > + * count controls its lifetime.
> > + */
> > +struct xe_pt_job_ops {
> > + /** @current_op: current page-table update operation */
> > + u32 current_op;
> > + /** @refcount: reference count */
> > + struct kref refcount;
> > + /** @deferred: list of deferred PT entries to destroy */
> > + struct llist_head deferred;
> > + /** @ops: page-table update operations */
> > + struct xe_vm_pgtable_update_op *ops;
> > +};
> > +
> > /** struct xe_vm_pgtable_update_ops: page table update operations */
> > struct xe_vm_pgtable_update_ops {
> > - /** @ops: operations */
> > - struct xe_vm_pgtable_update_op *ops;
> > - /** @deferred: deferred list to destroy PT entries */
> > - struct llist_head deferred;
> > + /** @pt_job_ops: PT update operations dynamic allocation*/
> > + struct xe_pt_job_ops *pt_job_ops;
> > /** @q: exec queue for PT operations */
> > struct xe_exec_queue *q;
> > /** @prl: embedded page reclaim list */
> > @@ -107,8 +124,6 @@ struct xe_vm_pgtable_update_ops {
> > u64 last;
> > /** @num_ops: number of operations */
> > u32 num_ops;
> > - /** @current_op: current operations */
> > - u32 current_op;
> > /** @needs_svm_lock: Needs SVM lock */
> > bool needs_svm_lock;
> > /** @needs_invalidation: Needs invalidation */
> > diff --git a/drivers/gpu/drm/xe/xe_vm.c b/drivers/gpu/drm/xe/xe_vm.c
> > index 548b0769b3ef..3e2d2191b78c 100644
> > --- a/drivers/gpu/drm/xe/xe_vm.c
> > +++ b/drivers/gpu/drm/xe/xe_vm.c
> > @@ -585,11 +585,9 @@ static int xe_vma_ops_alloc(struct xe_vma_ops
> > *vops, bool array_of_binds)
> > if (!vops->pt_update_ops[i].num_ops)
> > continue;
> >
> > - vops->pt_update_ops[i].ops =
> > - kmalloc_objs(*vops->pt_update_ops[i].ops,
> > - vops->pt_update_ops[i].num_ops,
> > - GFP_KERNEL | __GFP_RETRY_MAYFAIL
> > | __GFP_NOWARN);
> > - if (!vops->pt_update_ops[i].ops)
> > + vops->pt_update_ops[i].pt_job_ops =
> > + xe_pt_job_ops_alloc(vops-
> > >pt_update_ops[i].num_ops);
> > + if (!vops->pt_update_ops[i].pt_job_ops)
> > return array_of_binds ? -ENOBUFS : -ENOMEM;
> > }
> >
> > @@ -625,7 +623,7 @@ static void xe_vma_ops_fini(struct xe_vma_ops
> > *vops)
> > xe_vma_svm_prefetch_ops_fini(vops);
> >
> > for (i = 0; i < XE_MAX_TILES_PER_DEVICE; ++i)
> > - kfree(vops->pt_update_ops[i].ops);
> > + xe_pt_job_ops_put(vops->pt_update_ops[i].pt_job_ops);
> > }
> >
> > static void xe_vma_ops_incr_pt_update_ops(struct xe_vma_ops *vops,
> > u8 tile_mask, int inc_val)
>
next prev parent reply other threads:[~2026-03-03 23:28 UTC|newest]
Thread overview: 63+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-02-28 1:34 [PATCH v3 00/25] CPU binds and ULLS on migration queue Matthew Brost
2026-02-28 1:34 ` [PATCH v3 01/25] drm/xe: Drop struct xe_migrate_pt_update argument from populate/clear vfuns Matthew Brost
2026-03-05 14:17 ` Francois Dugast
2026-02-28 1:34 ` [PATCH v3 02/25] drm/xe: Add xe_migrate_update_pgtables_cpu_execute helper Matthew Brost
2026-03-05 14:39 ` Francois Dugast
2026-02-28 1:34 ` [PATCH v3 03/25] drm/xe: Decouple exec queue idle check from LRC Matthew Brost
2026-03-02 20:50 ` Summers, Stuart
2026-03-02 21:02 ` Matthew Brost
2026-03-03 21:26 ` Summers, Stuart
2026-03-03 22:42 ` Matthew Brost
2026-03-03 22:54 ` Summers, Stuart
2026-02-28 1:34 ` [PATCH v3 04/25] drm/xe: Add job count to GuC exec queue snapshot Matthew Brost
2026-03-02 20:50 ` Summers, Stuart
2026-02-28 1:34 ` [PATCH v3 05/25] drm/xe: Update xe_bo_put_deferred arguments to include writeback flag Matthew Brost
2026-04-01 12:20 ` Francois Dugast
2026-04-01 22:39 ` Matthew Brost
2026-02-28 1:34 ` [PATCH v3 06/25] drm/xe: Add XE_BO_FLAG_PUT_VM_ASYNC Matthew Brost
2026-04-01 12:22 ` Francois Dugast
2026-04-01 22:38 ` Matthew Brost
2026-02-28 1:34 ` [PATCH v3 07/25] drm/xe: Update scheduler job layer to support PT jobs Matthew Brost
2026-03-03 22:50 ` Summers, Stuart
2026-03-03 23:00 ` Matthew Brost
2026-02-28 1:34 ` [PATCH v3 08/25] drm/xe: Add helpers to access PT ops Matthew Brost
2026-04-07 15:22 ` Francois Dugast
2026-02-28 1:34 ` [PATCH v3 09/25] drm/xe: Add struct xe_pt_job_ops Matthew Brost
2026-03-03 23:26 ` Summers, Stuart
2026-03-03 23:28 ` Matthew Brost [this message]
2026-02-28 1:34 ` [PATCH v3 10/25] drm/xe: Update GuC submission backend to run PT jobs Matthew Brost
2026-03-03 23:28 ` Summers, Stuart
2026-03-04 0:26 ` Matthew Brost
2026-03-04 20:43 ` Summers, Stuart
2026-03-04 21:53 ` Matthew Brost
2026-03-05 20:24 ` Summers, Stuart
2026-02-28 1:34 ` [PATCH v3 11/25] drm/xe: Store level in struct xe_vm_pgtable_update Matthew Brost
2026-03-03 23:44 ` Summers, Stuart
2026-02-28 1:34 ` [PATCH v3 12/25] drm/xe: Don't use migrate exec queue for page fault binds Matthew Brost
2026-02-28 1:34 ` [PATCH v3 13/25] drm/xe: Enable CPU binds for jobs Matthew Brost
2026-02-28 1:34 ` [PATCH v3 14/25] drm/xe: Remove unused arguments from xe_migrate_pt_update_ops Matthew Brost
2026-02-28 1:34 ` [PATCH v3 15/25] drm/xe: Make bind queues operate cross-tile Matthew Brost
2026-02-28 1:34 ` [PATCH v3 16/25] drm/xe: Add CPU bind layer Matthew Brost
2026-02-28 1:34 ` [PATCH v3 17/25] drm/xe: Add device flag to enable PT mirroring across tiles Matthew Brost
2026-02-28 1:34 ` [PATCH v3 18/25] drm/xe: Add xe_hw_engine_write_ring_tail Matthew Brost
2026-02-28 1:34 ` [PATCH v3 19/25] drm/xe: Add ULLS support to LRC Matthew Brost
2026-03-05 20:21 ` Francois Dugast
2026-02-28 1:34 ` [PATCH v3 20/25] drm/xe: Add ULLS migration job support to migration layer Matthew Brost
2026-03-05 23:34 ` Summers, Stuart
2026-03-09 23:11 ` Matthew Brost
2026-02-28 1:34 ` [PATCH v3 21/25] drm/xe: Add MI_SEMAPHORE_WAIT instruction defs Matthew Brost
2026-02-28 1:34 ` [PATCH v3 22/25] drm/xe: Add ULLS migration job support to ring ops Matthew Brost
2026-02-28 1:34 ` [PATCH v3 23/25] drm/xe: Add ULLS migration job support to GuC submission Matthew Brost
2026-02-28 1:35 ` [PATCH v3 24/25] drm/xe: Enter ULLS for migration jobs upon page fault or SVM prefetch Matthew Brost
2026-02-28 1:35 ` [PATCH v3 25/25] drm/xe: Add modparam to enable / disable ULLS on migrate queue Matthew Brost
2026-03-05 22:59 ` Summers, Stuart
2026-04-01 22:44 ` Matthew Brost
2026-02-28 1:43 ` ✗ CI.checkpatch: warning for CPU binds and ULLS on migration queue (rev3) Patchwork
2026-02-28 1:44 ` ✓ CI.KUnit: success " Patchwork
2026-02-28 2:32 ` ✓ Xe.CI.BAT: " Patchwork
2026-02-28 13:59 ` ✗ Xe.CI.FULL: failure " Patchwork
2026-03-02 17:54 ` Summers, Stuart
2026-03-02 18:13 ` Matthew Brost
2026-03-05 22:56 ` [PATCH v3 00/25] CPU binds and ULLS on migration queue Summers, Stuart
2026-03-10 22:17 ` Matthew Brost
2026-03-20 15:31 ` Thomas Hellström
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aadukHCYIPthpwrg@lstrano-desk.jf.intel.com \
--to=matthew.brost@intel.com \
--cc=arvind.yadav@intel.com \
--cc=francois.dugast@intel.com \
--cc=himal.prasad.ghimiray@intel.com \
--cc=intel-xe@lists.freedesktop.org \
--cc=stuart.summers@intel.com \
--cc=thomas.hellstrom@linux.intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.