From: Nicolin Chen <nicolinc@nvidia.com>
To: <jgg@nvidia.com>, <kevin.tian@intel.com>, <corbet@lwn.net>,
<will@kernel.org>
Cc: <bagasdotme@gmail.com>, <robin.murphy@arm.com>, <joro@8bytes.org>,
<thierry.reding@gmail.com>, <vdumpa@nvidia.com>,
<jonathanh@nvidia.com>, <shuah@kernel.org>, <jsnitsel@redhat.com>,
<nathan@kernel.org>, <peterz@infradead.org>, <yi.l.liu@intel.com>,
<mshavit@google.com>, <praan@google.com>,
<zhangzekun11@huawei.com>, <iommu@lists.linux.dev>,
<linux-doc@vger.kernel.org>, <linux-kernel@vger.kernel.org>,
<linux-arm-kernel@lists.infradead.org>,
<linux-tegra@vger.kernel.org>, <linux-kselftest@vger.kernel.org>,
<patches@lists.linux.dev>, <mochs@nvidia.com>,
<alok.a.tiwari@oracle.com>, <vasant.hegde@amd.com>
Subject: [PATCH v4 09/23] iommufd: Abstract iopt_pin_pages and iopt_unpin_pages helpers
Date: Thu, 8 May 2025 20:02:30 -0700 [thread overview]
Message-ID: <88a0b17337897de8ca170afa1e9be26ff7442437.1746757630.git.nicolinc@nvidia.com> (raw)
In-Reply-To: <cover.1746757630.git.nicolinc@nvidia.com>
The new HW QUEUE object will be added for HW to access the guest queue for
HW-accelerated virtualization feature. It needs to ensure the guest memory
pages are pinned when HW accesses them and they are contiguous in physical
address space.
This is very like the existing iommufd_access_pin_pages() that outputs the
pinned page list for the caller to test its contiguity.
Move those code from iommufd_access_pin/unpin_pages() and related function
for a pair of iopt helpers that can be shared with the HW QUEUE allocator.
Rename check_area_prot() to align with the existing iopt_area helpers, and
inline it to the header since iommufd_access_rw() still uses it.
Reviewed-by: Pranjal Shrivastava <praan@google.com>
Signed-off-by: Nicolin Chen <nicolinc@nvidia.com>
---
drivers/iommu/iommufd/io_pagetable.h | 8 ++
drivers/iommu/iommufd/iommufd_private.h | 6 ++
drivers/iommu/iommufd/device.c | 117 ++----------------------
drivers/iommu/iommufd/io_pagetable.c | 95 +++++++++++++++++++
4 files changed, 117 insertions(+), 109 deletions(-)
diff --git a/drivers/iommu/iommufd/io_pagetable.h b/drivers/iommu/iommufd/io_pagetable.h
index 10c928a9a463..4288a2b1a90f 100644
--- a/drivers/iommu/iommufd/io_pagetable.h
+++ b/drivers/iommu/iommufd/io_pagetable.h
@@ -114,6 +114,14 @@ static inline unsigned long iopt_area_iova_to_index(struct iopt_area *area,
return iopt_area_start_byte(area, iova) / PAGE_SIZE;
}
+static inline bool iopt_area_check_prot(struct iopt_area *area,
+ unsigned int flags)
+{
+ if (flags & IOMMUFD_ACCESS_RW_WRITE)
+ return area->iommu_prot & IOMMU_WRITE;
+ return area->iommu_prot & IOMMU_READ;
+}
+
#define __make_iopt_iter(name) \
static inline struct iopt_##name *iopt_##name##_iter_first( \
struct io_pagetable *iopt, unsigned long start, \
diff --git a/drivers/iommu/iommufd/iommufd_private.h b/drivers/iommu/iommufd/iommufd_private.h
index 8d96aa514033..79160b039bc7 100644
--- a/drivers/iommu/iommufd/iommufd_private.h
+++ b/drivers/iommu/iommufd/iommufd_private.h
@@ -130,6 +130,12 @@ int iopt_cut_iova(struct io_pagetable *iopt, unsigned long *iovas,
void iopt_enable_large_pages(struct io_pagetable *iopt);
int iopt_disable_large_pages(struct io_pagetable *iopt);
+int iopt_pin_pages(struct io_pagetable *iopt, unsigned long iova,
+ unsigned long length, struct page **out_pages,
+ unsigned int flags);
+void iopt_unpin_pages(struct io_pagetable *iopt, unsigned long iova,
+ unsigned long length);
+
struct iommufd_ucmd {
struct iommufd_ctx *ictx;
void __user *ubuffer;
diff --git a/drivers/iommu/iommufd/device.c b/drivers/iommu/iommufd/device.c
index 2111bad72c72..a5c6be164254 100644
--- a/drivers/iommu/iommufd/device.c
+++ b/drivers/iommu/iommufd/device.c
@@ -1240,58 +1240,17 @@ void iommufd_access_notify_unmap(struct io_pagetable *iopt, unsigned long iova,
void iommufd_access_unpin_pages(struct iommufd_access *access,
unsigned long iova, unsigned long length)
{
- struct iopt_area_contig_iter iter;
- struct io_pagetable *iopt;
- unsigned long last_iova;
- struct iopt_area *area;
-
- if (WARN_ON(!length) ||
- WARN_ON(check_add_overflow(iova, length - 1, &last_iova)))
- return;
-
- mutex_lock(&access->ioas_lock);
+ guard(mutex)(&access->ioas_lock);
/*
* The driver must be doing something wrong if it calls this before an
* iommufd_access_attach() or after an iommufd_access_detach().
*/
- if (WARN_ON(!access->ioas_unpin)) {
- mutex_unlock(&access->ioas_lock);
+ if (WARN_ON(!access->ioas_unpin))
return;
- }
- iopt = &access->ioas_unpin->iopt;
-
- down_read(&iopt->iova_rwsem);
- iopt_for_each_contig_area(&iter, area, iopt, iova, last_iova)
- iopt_area_remove_access(
- area, iopt_area_iova_to_index(area, iter.cur_iova),
- iopt_area_iova_to_index(
- area,
- min(last_iova, iopt_area_last_iova(area))));
- WARN_ON(!iopt_area_contig_done(&iter));
- up_read(&iopt->iova_rwsem);
- mutex_unlock(&access->ioas_lock);
+ iopt_unpin_pages(&access->ioas_unpin->iopt, iova, length);
}
EXPORT_SYMBOL_NS_GPL(iommufd_access_unpin_pages, "IOMMUFD");
-static bool iopt_area_contig_is_aligned(struct iopt_area_contig_iter *iter)
-{
- if (iopt_area_start_byte(iter->area, iter->cur_iova) % PAGE_SIZE)
- return false;
-
- if (!iopt_area_contig_done(iter) &&
- (iopt_area_start_byte(iter->area, iopt_area_last_iova(iter->area)) %
- PAGE_SIZE) != (PAGE_SIZE - 1))
- return false;
- return true;
-}
-
-static bool check_area_prot(struct iopt_area *area, unsigned int flags)
-{
- if (flags & IOMMUFD_ACCESS_RW_WRITE)
- return area->iommu_prot & IOMMU_WRITE;
- return area->iommu_prot & IOMMU_READ;
-}
-
/**
* iommufd_access_pin_pages() - Return a list of pages under the iova
* @access: IOAS access to act on
@@ -1315,76 +1274,16 @@ int iommufd_access_pin_pages(struct iommufd_access *access, unsigned long iova,
unsigned long length, struct page **out_pages,
unsigned int flags)
{
- struct iopt_area_contig_iter iter;
- struct io_pagetable *iopt;
- unsigned long last_iova;
- struct iopt_area *area;
- int rc;
-
/* Driver's ops don't support pin_pages */
if (IS_ENABLED(CONFIG_IOMMUFD_TEST) &&
WARN_ON(access->iova_alignment != PAGE_SIZE || !access->ops->unmap))
return -EINVAL;
- if (!length)
- return -EINVAL;
- if (check_add_overflow(iova, length - 1, &last_iova))
- return -EOVERFLOW;
-
- mutex_lock(&access->ioas_lock);
- if (!access->ioas) {
- mutex_unlock(&access->ioas_lock);
+ guard(mutex)(&access->ioas_lock);
+ if (!access->ioas)
return -ENOENT;
- }
- iopt = &access->ioas->iopt;
-
- down_read(&iopt->iova_rwsem);
- iopt_for_each_contig_area(&iter, area, iopt, iova, last_iova) {
- unsigned long last = min(last_iova, iopt_area_last_iova(area));
- unsigned long last_index = iopt_area_iova_to_index(area, last);
- unsigned long index =
- iopt_area_iova_to_index(area, iter.cur_iova);
-
- if (area->prevent_access ||
- !iopt_area_contig_is_aligned(&iter)) {
- rc = -EINVAL;
- goto err_remove;
- }
-
- if (!check_area_prot(area, flags)) {
- rc = -EPERM;
- goto err_remove;
- }
-
- rc = iopt_area_add_access(area, index, last_index, out_pages,
- flags);
- if (rc)
- goto err_remove;
- out_pages += last_index - index + 1;
- }
- if (!iopt_area_contig_done(&iter)) {
- rc = -ENOENT;
- goto err_remove;
- }
-
- up_read(&iopt->iova_rwsem);
- mutex_unlock(&access->ioas_lock);
- return 0;
-
-err_remove:
- if (iova < iter.cur_iova) {
- last_iova = iter.cur_iova - 1;
- iopt_for_each_contig_area(&iter, area, iopt, iova, last_iova)
- iopt_area_remove_access(
- area,
- iopt_area_iova_to_index(area, iter.cur_iova),
- iopt_area_iova_to_index(
- area, min(last_iova,
- iopt_area_last_iova(area))));
- }
- up_read(&iopt->iova_rwsem);
- mutex_unlock(&access->ioas_lock);
- return rc;
+ return iopt_pin_pages(&access->ioas->iopt, iova, length, out_pages,
+ flags);
}
EXPORT_SYMBOL_NS_GPL(iommufd_access_pin_pages, "IOMMUFD");
@@ -1431,7 +1330,7 @@ int iommufd_access_rw(struct iommufd_access *access, unsigned long iova,
goto err_out;
}
- if (!check_area_prot(area, flags)) {
+ if (!iopt_area_check_prot(area, flags)) {
rc = -EPERM;
goto err_out;
}
diff --git a/drivers/iommu/iommufd/io_pagetable.c b/drivers/iommu/iommufd/io_pagetable.c
index 8a790e597e12..4dfe14dcbb51 100644
--- a/drivers/iommu/iommufd/io_pagetable.c
+++ b/drivers/iommu/iommufd/io_pagetable.c
@@ -1472,3 +1472,98 @@ int iopt_table_enforce_dev_resv_regions(struct io_pagetable *iopt,
up_write(&iopt->iova_rwsem);
return rc;
}
+
+static bool iopt_area_contig_is_aligned(struct iopt_area_contig_iter *iter)
+{
+ if (iopt_area_start_byte(iter->area, iter->cur_iova) % PAGE_SIZE)
+ return false;
+
+ if (!iopt_area_contig_done(iter) &&
+ (iopt_area_start_byte(iter->area, iopt_area_last_iova(iter->area)) %
+ PAGE_SIZE) != (PAGE_SIZE - 1))
+ return false;
+ return true;
+}
+
+int iopt_pin_pages(struct io_pagetable *iopt, unsigned long iova,
+ unsigned long length, struct page **out_pages,
+ unsigned int flags)
+{
+ struct iopt_area_contig_iter iter;
+ unsigned long last_iova;
+ struct iopt_area *area;
+ int rc;
+
+ if (!length)
+ return -EINVAL;
+ if (check_add_overflow(iova, length - 1, &last_iova))
+ return -EOVERFLOW;
+
+ down_read(&iopt->iova_rwsem);
+ iopt_for_each_contig_area(&iter, area, iopt, iova, last_iova) {
+ unsigned long last = min(last_iova, iopt_area_last_iova(area));
+ unsigned long last_index = iopt_area_iova_to_index(area, last);
+ unsigned long index =
+ iopt_area_iova_to_index(area, iter.cur_iova);
+
+ if (area->prevent_access ||
+ !iopt_area_contig_is_aligned(&iter)) {
+ rc = -EINVAL;
+ goto err_remove;
+ }
+
+ if (!iopt_area_check_prot(area, flags)) {
+ rc = -EPERM;
+ goto err_remove;
+ }
+
+ rc = iopt_area_add_access(area, index, last_index, out_pages,
+ flags);
+ if (rc)
+ goto err_remove;
+ out_pages += last_index - index + 1;
+ }
+ if (!iopt_area_contig_done(&iter)) {
+ rc = -ENOENT;
+ goto err_remove;
+ }
+
+ up_read(&iopt->iova_rwsem);
+ return 0;
+
+err_remove:
+ if (iova < iter.cur_iova) {
+ last_iova = iter.cur_iova - 1;
+ iopt_for_each_contig_area(&iter, area, iopt, iova, last_iova)
+ iopt_area_remove_access(
+ area,
+ iopt_area_iova_to_index(area, iter.cur_iova),
+ iopt_area_iova_to_index(
+ area, min(last_iova,
+ iopt_area_last_iova(area))));
+ }
+ up_read(&iopt->iova_rwsem);
+ return rc;
+}
+
+void iopt_unpin_pages(struct io_pagetable *iopt, unsigned long iova,
+ unsigned long length)
+{
+ struct iopt_area_contig_iter iter;
+ unsigned long last_iova;
+ struct iopt_area *area;
+
+ if (WARN_ON(!length) ||
+ WARN_ON(check_add_overflow(iova, length - 1, &last_iova)))
+ return;
+
+ down_read(&iopt->iova_rwsem);
+ iopt_for_each_contig_area(&iter, area, iopt, iova, last_iova)
+ iopt_area_remove_access(
+ area, iopt_area_iova_to_index(area, iter.cur_iova),
+ iopt_area_iova_to_index(
+ area,
+ min(last_iova, iopt_area_last_iova(area))));
+ WARN_ON(!iopt_area_contig_done(&iter));
+ up_read(&iopt->iova_rwsem);
+}
--
2.43.0
next prev parent reply other threads:[~2025-05-09 3:03 UTC|newest]
Thread overview: 106+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-05-09 3:02 [PATCH v4 00/23] iommufd: Add vIOMMU infrastructure (Part-4 HW QUEUE) Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 01/23] iommufd/viommu: Add driver-allocated vDEVICE support Nicolin Chen
2025-05-15 5:42 ` Tian, Kevin
2025-05-15 16:55 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 02/23] iommu: Pass in a driver-level user data structure to viommu_alloc op Nicolin Chen
2025-05-15 5:44 ` Tian, Kevin
2025-05-09 3:02 ` [PATCH v4 03/23] iommufd/viommu: Allow driver-specific user data for a vIOMMU object Nicolin Chen
2025-05-15 5:45 ` Tian, Kevin
2025-05-15 16:56 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 04/23] iommu: Add iommu_copy_struct_to_user helper Nicolin Chen
2025-05-15 5:46 ` Tian, Kevin
2025-05-09 3:02 ` [PATCH v4 05/23] iommufd/driver: Let iommufd_viommu_alloc helper save ictx to viommu->ictx Nicolin Chen
2025-05-14 17:06 ` Jason Gunthorpe
2025-05-16 2:05 ` Nicolin Chen
2025-05-16 13:28 ` Jason Gunthorpe
2025-05-16 20:56 ` Nicolin Chen
2025-05-26 13:30 ` Jason Gunthorpe
2025-05-27 18:41 ` Nicolin Chen
2025-05-30 18:27 ` Jason Gunthorpe
2025-05-30 18:34 ` Nicolin Chen
2025-05-15 5:48 ` Tian, Kevin
2025-05-09 3:02 ` [PATCH v4 06/23] iommufd/driver: Add iommufd_struct_destroy to revert iommufd_viommu_alloc Nicolin Chen
2025-05-14 18:26 ` Jason Gunthorpe
2025-05-14 19:21 ` Nicolin Chen
2025-05-15 12:49 ` Jason Gunthorpe
2025-05-15 16:55 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 07/23] iommufd/selftest: Support user_data in mock_viommu_alloc Nicolin Chen
2025-05-15 5:49 ` Tian, Kevin
2025-05-09 3:02 ` [PATCH v4 08/23] iommufd/selftest: Add covearge for viommu data Nicolin Chen
2025-05-15 5:50 ` Tian, Kevin
2025-05-09 3:02 ` Nicolin Chen [this message]
2025-05-14 18:45 ` [PATCH v4 09/23] iommufd: Abstract iopt_pin_pages and iopt_unpin_pages helpers Jason Gunthorpe
2025-05-15 5:54 ` Tian, Kevin
2025-05-09 3:02 ` [PATCH v4 10/23] iommufd/viommu: Introduce IOMMUFD_OBJ_HW_QUEUE and its related struct Nicolin Chen
2025-05-15 5:58 ` Tian, Kevin
2025-05-15 17:14 ` Nicolin Chen
2025-05-16 2:30 ` Nicolin Chen
2025-05-16 2:59 ` Tian, Kevin
2025-05-19 17:05 ` Vasant Hegde
2025-05-15 15:39 ` Jason Gunthorpe
2025-05-15 17:17 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 11/23] iommufd/viommu: Add IOMMUFD_CMD_HW_QUEUE_ALLOC ioctl Nicolin Chen
2025-05-15 6:30 ` Tian, Kevin
2025-05-15 18:44 ` Nicolin Chen
2025-05-16 2:49 ` Tian, Kevin
2025-05-16 3:16 ` Nicolin Chen
2025-05-16 3:52 ` Tian, Kevin
2025-05-16 4:05 ` Nicolin Chen
2025-05-18 15:19 ` Nicolin Chen
2025-05-15 16:06 ` Jason Gunthorpe
2025-05-15 18:16 ` Nicolin Chen
2025-05-15 18:59 ` Jason Gunthorpe
2025-05-15 20:32 ` Nicolin Chen
2025-05-16 13:26 ` Jason Gunthorpe
2025-05-16 2:42 ` Tian, Kevin
2025-05-16 13:25 ` Jason Gunthorpe
2025-05-19 17:29 ` Vasant Hegde
2025-05-19 18:14 ` Nicolin Chen
2025-05-20 8:38 ` Vasant Hegde
2025-05-23 1:51 ` Tian, Kevin
2025-05-26 13:29 ` Jason Gunthorpe
2025-05-09 3:02 ` [PATCH v4 12/23] iommufd/driver: Add iommufd_hw_queue_depend/undepend() helpers Nicolin Chen
2025-05-15 16:12 ` Jason Gunthorpe
2025-05-16 4:51 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 13/23] iommufd/selftest: Add coverage for IOMMUFD_CMD_HW_QUEUE_ALLOC Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 14/23] iommufd: Add mmap interface Nicolin Chen
2025-05-09 14:13 ` kernel test robot
2025-05-09 19:30 ` Nicolin Chen
2025-05-15 6:41 ` Tian, Kevin
2025-05-15 16:47 ` Jason Gunthorpe
2025-05-16 4:08 ` Tian, Kevin
2025-05-16 13:29 ` Jason Gunthorpe
2025-05-16 17:42 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 15/23] iommufd/selftest: Add coverage for the new " Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 16/23] Documentation: userspace-api: iommufd: Update HW QUEUE Nicolin Chen
2025-05-15 6:42 ` Tian, Kevin
2025-05-15 16:58 ` Jason Gunthorpe
2025-05-09 3:02 ` [PATCH v4 17/23] iommu/arm-smmu-v3-iommufd: Add vsmmu_alloc impl op Nicolin Chen
2025-05-15 7:52 ` Tian, Kevin
2025-05-15 17:19 ` Jason Gunthorpe
2025-05-15 17:32 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 18/23] iommu/arm-smmu-v3-iommufd: Support implementation-defined hw_info Nicolin Chen
2025-05-15 7:54 ` Tian, Kevin
2025-05-15 17:17 ` Jason Gunthorpe
2025-05-15 18:52 ` Nicolin Chen
2025-05-15 18:56 ` Jason Gunthorpe
2025-05-15 19:21 ` Nicolin Chen
2025-05-15 19:23 ` Jason Gunthorpe
2025-05-15 20:17 ` Nicolin Chen
2025-05-16 13:22 ` Jason Gunthorpe
2025-05-16 17:34 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 19/23] iommu/tegra241-cmdqv: Use request_threaded_irq Nicolin Chen
2025-05-15 7:57 ` Tian, Kevin
2025-05-15 17:21 ` Jason Gunthorpe
2025-05-09 3:02 ` [PATCH v4 20/23] iommu/tegra241-cmdqv: Simplify deinit flow in tegra241_cmdqv_remove_vintf() Nicolin Chen
2025-05-15 8:00 ` Tian, Kevin
2025-05-15 17:27 ` Jason Gunthorpe
2025-05-09 3:02 ` [PATCH v4 21/23] iommu/tegra241-cmdqv: Do not statically map LVCMDQs Nicolin Chen
2025-05-15 8:20 ` Tian, Kevin
2025-05-15 17:03 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 22/23] iommu/tegra241-cmdqv: Add user-space use support Nicolin Chen
2025-05-15 8:27 ` Tian, Kevin
2025-05-15 17:13 ` Nicolin Chen
2025-05-16 4:00 ` Tian, Kevin
2025-05-16 4:10 ` Nicolin Chen
2025-05-09 3:02 ` [PATCH v4 23/23] iommu/tegra241-cmdqv: Add IOMMU_VEVENTQ_TYPE_TEGRA241_CMDQV support Nicolin Chen
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=88a0b17337897de8ca170afa1e9be26ff7442437.1746757630.git.nicolinc@nvidia.com \
--to=nicolinc@nvidia.com \
--cc=alok.a.tiwari@oracle.com \
--cc=bagasdotme@gmail.com \
--cc=corbet@lwn.net \
--cc=iommu@lists.linux.dev \
--cc=jgg@nvidia.com \
--cc=jonathanh@nvidia.com \
--cc=joro@8bytes.org \
--cc=jsnitsel@redhat.com \
--cc=kevin.tian@intel.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-doc@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-kselftest@vger.kernel.org \
--cc=linux-tegra@vger.kernel.org \
--cc=mochs@nvidia.com \
--cc=mshavit@google.com \
--cc=nathan@kernel.org \
--cc=patches@lists.linux.dev \
--cc=peterz@infradead.org \
--cc=praan@google.com \
--cc=robin.murphy@arm.com \
--cc=shuah@kernel.org \
--cc=thierry.reding@gmail.com \
--cc=vasant.hegde@amd.com \
--cc=vdumpa@nvidia.com \
--cc=will@kernel.org \
--cc=yi.l.liu@intel.com \
--cc=zhangzekun11@huawei.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).