From: Stanislav Kinsburskii <skinsburskii@linux.microsoft.com>
To: Michael Kelley <mhklinux@outlook.com>
Cc: "kys@microsoft.com" <kys@microsoft.com>,
"haiyangz@microsoft.com" <haiyangz@microsoft.com>,
"wei.liu@kernel.org" <wei.liu@kernel.org>,
"decui@microsoft.com" <decui@microsoft.com>,
"longli@microsoft.com" <longli@microsoft.com>,
"linux-hyperv@vger.kernel.org" <linux-hyperv@vger.kernel.org>,
"linux-kernel@vger.kernel.org" <linux-kernel@vger.kernel.org>
Subject: Re: [PATCH 3/7] mshv: Support regions with different VMAs
Date: Mon, 20 Apr 2026 09:29:11 -0700 [thread overview]
Message-ID: <aeZUVyrIkYvdAy0q@skinsburskii.localdomain> (raw)
In-Reply-To: <SN6PR02MB4157043516DEB3DC5E987AA6D4242@SN6PR02MB4157.namprd02.prod.outlook.com>
On Mon, Apr 13, 2026 at 09:08:52PM +0000, Michael Kelley wrote:
> From: Stanislav Kinsburskii <skinsburskii@linux.microsoft.com> Sent: Monday, March 30, 2026 1:04 PM
> >
> > Allow HMM fault handling across memory regions that span multiple VMAs
> > with different protection flags. The previous implementation assumed a
> > single VMA per region, which would fail when guest memory crosses VMA
> > boundaries.
> >
> > Iterate through VMAs within the range and handle each separately with
> > appropriate protection flags, enabling more flexible memory region
> > configurations for partitions.
> >
> > Signed-off-by: Stanislav Kinsburskii <skinsburskii@linux.microsoft.com>
> > ---
> > drivers/hv/mshv_regions.c | 72 +++++++++++++++++++++++++++++++++------------
> > 1 file changed, 52 insertions(+), 20 deletions(-)
> >
> > diff --git a/drivers/hv/mshv_regions.c b/drivers/hv/mshv_regions.c
> > index ed9c55841140..1bb1bfe177e2 100644
> > --- a/drivers/hv/mshv_regions.c
> > +++ b/drivers/hv/mshv_regions.c
> > @@ -492,37 +492,72 @@ int mshv_region_get(struct mshv_mem_region *region)
> > }
> >
> > /**
> > - * mshv_region_hmm_fault_and_lock - Handle HMM faults and lock the memory region
> > + * mshv_region_hmm_fault_and_lock - Handle HMM faults across VMAs and lock
> > + * the memory region
> > * @region: Pointer to the memory region structure
> > - * @range: Pointer to the HMM range structure
> > + * @start : Starting virtual address of the range to fault
> > + * @end : Ending virtual address of the range to fault (exclusive)
> > + * @pfns : Output array for page frame numbers with HMM flags
> > *
> > * This function performs the following steps:
> > * 1. Reads the notifier sequence for the HMM range.
> > * 2. Acquires a read lock on the memory map.
> > - * 3. Handles HMM faults for the specified range.
> > - * 4. Releases the read lock on the memory map.
> > - * 5. If successful, locks the memory region mutex.
> > - * 6. Verifies if the notifier sequence has changed during the operation.
> > - * If it has, releases the mutex and returns -EBUSY to match with
> > - * hmm_range_fault() return code for repeating.
> > + * 3. Iterates through VMAs in the specified range, handling each
> > + * separately with appropriate protection flags (HMM_PFN_REQ_WRITE set
> > + * based on VMA flags).
> > + * 4. Handles HMM faults for each VMA segment.
> > + * 5. Releases the read lock on the memory map.
> > + * 6. If successful, locks the memory region mutex.
> > + * 7. Verifies if the notifier sequence has changed during the operation.
> > + * If it has, releases the mutex and returns -EBUSY to signal retry.
> > + *
> > + * The function expects the range [start, end] is backed by valid VMAs.
>
> Use "[start, end)" to describe the range since end is exclusive.
>
Will do
> > + * Returns -EFAULT if any address in the range is not covered by a VMA.
> > *
> > * Return: 0 on success, a negative error code otherwise.
> > */
> > static int mshv_region_hmm_fault_and_lock(struct mshv_mem_region *region,
> > - struct hmm_range *range)
> > + unsigned long start,
> > + unsigned long end,
> > + unsigned long *pfns)
> > {
> > + struct hmm_range range = {
> > + .notifier = ®ion->mreg_mni,
> > + };
> > int ret;
> >
> > - range->notifier_seq = mmu_interval_read_begin(range->notifier);
> > + range.notifier_seq = mmu_interval_read_begin(range.notifier);
> > mmap_read_lock(region->mreg_mni.mm);
> > - ret = hmm_range_fault(range);
> > + while (start < end) {
> > + struct vm_area_struct *vma;
> > +
> > + vma = vma_lookup(current->mm, start);
>
> The mmap_read_lock() was obtained on region->mreg_mni.mm, but the
> lookup is done against current->mm. Maybe these are the same, but
> it looks wrong. (Pointed out by a Co-Pilot AI review.)
>
Yes, they arethe same, but I'll update to use the same mm for clarity.
> > + if (!vma) {
> > + ret = -EFAULT;
> > + break;
> > + }
> > +
> > + range.hmm_pfns = pfns;
> > + range.start = start;
> > + range.end = min(vma->vm_end, end);
> > + range.default_flags = HMM_PFN_REQ_FAULT;
> > + if (vma->vm_flags & VM_WRITE)
> > + range.default_flags |= HMM_PFN_REQ_WRITE;
> > +
> > + ret = hmm_range_fault(&range);
> > + if (ret)
> > + break;
> > +
> > + start = range.end + 1;
>
> Since range.end is exclusive, the +1 should not be done.
>
Is it always? I'll need to check to make sure the end passed to this
function is page aligned. If it is, then I'll remove the +1.
> > + pfns += DIV_ROUND_UP(range.end - range.start, PAGE_SIZE);
>
> Just to confirm, range.end and range.start should always be page aligned,
> right? So the ROUND_UP should never kick in.
>
Same as above: if the end passed to this function is page aligned, then
I'll remove the DIV_ROUND_UP and just do a simple division.
Thanks,
Stanislav
> > + }
> > mmap_read_unlock(region->mreg_mni.mm);
> > if (ret)
> > return ret;
> >
> > mutex_lock(®ion->mreg_mutex);
> >
> > - if (mmu_interval_read_retry(range->notifier, range->notifier_seq)) {
> > + if (mmu_interval_read_retry(range.notifier, range.notifier_seq)) {
> > mutex_unlock(®ion->mreg_mutex);
> > cond_resched();
> > return -EBUSY;
> > @@ -546,10 +581,7 @@ static int mshv_region_hmm_fault_and_lock(struct mshv_mem_region *region,
> > static int mshv_region_range_fault(struct mshv_mem_region *region,
> > u64 pfn_offset, u64 pfn_count)
> > {
> > - struct hmm_range range = {
> > - .notifier = ®ion->mreg_mni,
> > - .default_flags = HMM_PFN_REQ_FAULT | HMM_PFN_REQ_WRITE,
> > - };
> > + unsigned long start, end;
> > unsigned long *pfns;
> > int ret;
> > u64 i;
> > @@ -558,12 +590,12 @@ static int mshv_region_range_fault(struct mshv_mem_region *region,
> > if (!pfns)
> > return -ENOMEM;
> >
> > - range.hmm_pfns = pfns;
> > - range.start = region->start_uaddr + pfn_offset * HV_HYP_PAGE_SIZE;
> > - range.end = range.start + pfn_count * HV_HYP_PAGE_SIZE;
> > + start = region->start_uaddr + pfn_offset * PAGE_SIZE;
> > + end = start + pfn_count * PAGE_SIZE;
> >
> > do {
> > - ret = mshv_region_hmm_fault_and_lock(region, &range);
> > + ret = mshv_region_hmm_fault_and_lock(region, start, end,
> > + pfns);
> > } while (ret == -EBUSY);
> >
> > if (ret)
> >
> >
>
next prev parent reply other threads:[~2026-04-20 16:29 UTC|newest]
Thread overview: 20+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-03-30 20:04 [PATCH 0/7] mshv: Refactor memory region management and map pages at creation Stanislav Kinsburskii
2026-03-30 20:04 ` [PATCH 1/7] mshv: Convert from page pointers to PFNs Stanislav Kinsburskii
2026-04-13 21:08 ` Michael Kelley
2026-04-20 16:21 ` Stanislav Kinsburskii
2026-04-20 17:18 ` Michael Kelley
2026-04-20 23:45 ` Stanislav Kinsburskii
2026-03-30 20:04 ` [PATCH 2/7] mshv: Add support to address range holes remapping Stanislav Kinsburskii
2026-04-13 21:08 ` Michael Kelley
2026-04-20 16:24 ` Stanislav Kinsburskii
2026-03-30 20:04 ` [PATCH 3/7] mshv: Support regions with different VMAs Stanislav Kinsburskii
2026-04-13 21:08 ` Michael Kelley
2026-04-20 16:29 ` Stanislav Kinsburskii [this message]
2026-03-30 20:04 ` [PATCH 4/7] mshv: Move pinned region setup to mshv_regions.c Stanislav Kinsburskii
2026-03-30 20:04 ` [PATCH 5/7] mshv: Map populated pages on movable region creation Stanislav Kinsburskii
2026-04-13 21:09 ` Michael Kelley
2026-04-20 16:35 ` Stanislav Kinsburskii
2026-03-30 20:04 ` [PATCH 6/7] mshv: Extract MMIO region mapping into separate function Stanislav Kinsburskii
2026-03-30 20:04 ` [PATCH 7/7] mshv: Add tracepoint for map GPA hypercall Stanislav Kinsburskii
2026-04-13 21:07 ` [PATCH 0/7] mshv: Refactor memory region management and map pages at creation Michael Kelley
2026-04-20 16:40 ` Stanislav Kinsburskii
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=aeZUVyrIkYvdAy0q@skinsburskii.localdomain \
--to=skinsburskii@linux.microsoft.com \
--cc=decui@microsoft.com \
--cc=haiyangz@microsoft.com \
--cc=kys@microsoft.com \
--cc=linux-hyperv@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=longli@microsoft.com \
--cc=mhklinux@outlook.com \
--cc=wei.liu@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox