From: Yi Liu <yi.l.liu@intel.com>
To: joro@8bytes.org, alex.williamson@redhat.com, jgg@nvidia.com,
kevin.tian@intel.com, robin.murphy@arm.com,
baolu.lu@linux.intel.com
Cc: cohuck@redhat.com, eric.auger@redhat.com, nicolinc@nvidia.com,
kvm@vger.kernel.org, mjrosato@linux.ibm.com,
chao.p.peng@linux.intel.com, yi.l.liu@intel.com,
yi.y.sun@linux.intel.com, peterx@redhat.com, jasowang@redhat.com,
shameerali.kolothum.thodi@huawei.com, lulu@redhat.com,
suravee.suthikulpanit@amd.com, iommu@lists.linux.dev,
linux-kernel@vger.kernel.org, linux-kselftest@vger.kernel.org,
zhenzhong.duan@intel.com
Subject: [PATCH v2 09/11] iommufd/selftest: Add coverage for IOMMU_HWPT_ALLOC with user data
Date: Thu, 11 May 2023 07:38:42 -0700 [thread overview]
Message-ID: <20230511143844.22693-10-yi.l.liu@intel.com> (raw)
In-Reply-To: <20230511143844.22693-1-yi.l.liu@intel.com>
From: Nicolin Chen <nicolinc@nvidia.com>
The IOMMU_HWPT_ALLOC ioctl now supports passing user_data to allocate a
customized domain. Add its coverage for both a regular domain case and
a nested domain case.
Signed-off-by: Nicolin Chen <nicolinc@nvidia.com>
Signed-off-by: Yi Liu <yi.l.liu@intel.com>
---
tools/testing/selftests/iommu/iommufd.c | 114 +++++++++++++++++-
tools/testing/selftests/iommu/iommufd_utils.h | 36 ++++++
2 files changed, 149 insertions(+), 1 deletion(-)
diff --git a/tools/testing/selftests/iommu/iommufd.c b/tools/testing/selftests/iommu/iommufd.c
index fa2324741ad2..568f80f596ad 100644
--- a/tools/testing/selftests/iommu/iommufd.c
+++ b/tools/testing/selftests/iommu/iommufd.c
@@ -122,6 +122,7 @@ TEST_F(iommufd, cmd_length)
TEST_LENGTH(iommu_option, IOMMU_OPTION);
TEST_LENGTH(iommu_vfio_ioas, IOMMU_VFIO_IOAS);
TEST_LENGTH(iommu_hw_info, IOMMU_DEVICE_GET_HW_INFO);
+ TEST_LENGTH(iommu_hwpt_alloc, IOMMU_HWPT_ALLOC);
#undef TEST_LENGTH
}
@@ -194,6 +195,7 @@ FIXTURE_VARIANT(iommufd_ioas)
{
unsigned int mock_domains;
unsigned int memory_limit;
+ bool new_hwpt;
};
FIXTURE_SETUP(iommufd_ioas)
@@ -233,6 +235,12 @@ FIXTURE_VARIANT_ADD(iommufd_ioas, mock_domain)
.mock_domains = 1,
};
+FIXTURE_VARIANT_ADD(iommufd_ioas, mock_domain_hwpt)
+{
+ .mock_domains = 1,
+ .new_hwpt = true,
+};
+
FIXTURE_VARIANT_ADD(iommufd_ioas, two_mock_domain)
{
.mock_domains = 2,
@@ -260,6 +268,93 @@ TEST_F(iommufd_ioas, ioas_destroy)
}
}
+TEST_F(iommufd_ioas, hwpt_alloc)
+{
+ uint32_t new_hwpt_id = 0;
+
+ if (self->stdev_id && self->device_id) {
+ test_cmd_hwpt_alloc(self->device_id, self->ioas_id, &new_hwpt_id);
+ test_cmd_mock_domain_replace(self->stdev_id, new_hwpt_id);
+ /* hw_pagetable cannot be freed if a device is attached to it */
+ EXPECT_ERRNO(EBUSY, _test_ioctl_destroy(self->fd, new_hwpt_id));
+
+ /* Detach from the new hw_pagetable and try again */
+ test_cmd_mock_domain_replace(self->stdev_id, self->ioas_id);
+ test_ioctl_destroy(new_hwpt_id);
+ } else {
+ test_err_cmd_hwpt_alloc(ENOENT, self->device_id,
+ self->ioas_id, &new_hwpt_id);
+ test_err_mock_domain_replace(ENOENT,
+ self->stdev_id, new_hwpt_id);
+ }
+}
+
+TEST_F(iommufd_ioas, nested_hwpt_alloc)
+{
+ uint32_t nested_hwpt_id[2] = {};
+ uint32_t parent_hwpt_id = 0;
+ uint32_t test_hwpt_id = 0;
+
+ if (self->device_id) {
+ /* Negative tests */
+ test_err_cmd_hwpt_alloc(ENOENT, self->ioas_id, self->device_id,
+ &test_hwpt_id);
+ test_err_cmd_hwpt_alloc(EINVAL, self->device_id,
+ self->device_id, &test_hwpt_id);
+
+ /* Allocate two nested hwpts sharing one common parent hwpt */
+ test_cmd_hwpt_alloc(self->device_id, self->ioas_id,
+ &parent_hwpt_id);
+
+ test_cmd_hwpt_alloc_nested(self->device_id, parent_hwpt_id,
+ &nested_hwpt_id[0]);
+ test_cmd_hwpt_alloc_nested(self->device_id, parent_hwpt_id,
+ &nested_hwpt_id[1]);
+
+ /* Negative test: a nested hwpt on top of a nested hwpt */
+ test_err_cmd_hwpt_alloc_nested(EINVAL, self->device_id,
+ nested_hwpt_id[0],
+ &test_hwpt_id);
+ /* Negative test: parent hwpt now cannot be freed */
+ EXPECT_ERRNO(EBUSY,
+ _test_ioctl_destroy(self->fd, parent_hwpt_id));
+
+ /* Attach device to nested_hwpt_id[0] that then will be busy */
+ test_cmd_mock_domain_replace(self->stdev_id,
+ nested_hwpt_id[0]);
+ EXPECT_ERRNO(EBUSY,
+ _test_ioctl_destroy(self->fd, nested_hwpt_id[0]));
+
+ /* Switch from nested_hwpt_id[0] to nested_hwpt_id[1] */
+ test_cmd_mock_domain_replace(self->stdev_id,
+ nested_hwpt_id[1]);
+ EXPECT_ERRNO(EBUSY,
+ _test_ioctl_destroy(self->fd, nested_hwpt_id[1]));
+ test_ioctl_destroy(nested_hwpt_id[0]);
+
+ /* Detach from nested_hwpt_id[1] and destroy it */
+ test_cmd_mock_domain_replace(self->stdev_id, parent_hwpt_id);
+ test_ioctl_destroy(nested_hwpt_id[1]);
+
+ /* Detach from the parent hw_pagetable and destroy it */
+ test_cmd_mock_domain_replace(self->stdev_id, self->ioas_id);
+ test_ioctl_destroy(parent_hwpt_id);
+ } else {
+ test_err_cmd_hwpt_alloc(ENOENT, self->device_id, self->ioas_id,
+ &parent_hwpt_id);
+ test_err_cmd_hwpt_alloc_nested(ENOENT, self->device_id,
+ parent_hwpt_id,
+ &nested_hwpt_id[0]);
+ test_err_cmd_hwpt_alloc_nested(ENOENT, self->device_id,
+ parent_hwpt_id,
+ &nested_hwpt_id[1]);
+ test_err_mock_domain_replace(ENOENT, self->stdev_id,
+ nested_hwpt_id[0]);
+ test_err_mock_domain_replace(ENOENT, self->stdev_id,
+ nested_hwpt_id[1]);
+ }
+}
+
TEST_F(iommufd_ioas, hwpt_attach)
{
/* Create a device attached directly to a hwpt */
@@ -663,6 +758,8 @@ TEST_F(iommufd_ioas, access_pin)
MOCK_FLAGS_ACCESS_CREATE_NEEDS_PIN_PAGES);
for (npages = 1; npages < BUFFER_SIZE / PAGE_SIZE; npages++) {
+ uint32_t new_hwpt_id = 0;
+ uint32_t mock_device_id;
uint32_t mock_stdev_id;
uint32_t mock_hwpt_id;
@@ -696,12 +793,27 @@ TEST_F(iommufd_ioas, access_pin)
_IOMMU_TEST_CMD(IOMMU_TEST_OP_ACCESS_PAGES),
&access_cmd));
test_cmd_mock_domain(self->ioas_id, &mock_stdev_id,
- &mock_hwpt_id, NULL);
+ &mock_hwpt_id, &mock_device_id);
check_map_cmd.id = mock_hwpt_id;
+ if (variant->new_hwpt) {
+ test_cmd_hwpt_alloc(mock_device_id, self->ioas_id,
+ &new_hwpt_id);
+ test_cmd_mock_domain_replace(mock_stdev_id,
+ new_hwpt_id);
+ check_map_cmd.id = new_hwpt_id;
+ } else {
+ check_map_cmd.id = mock_hwpt_id;
+ }
ASSERT_EQ(0, ioctl(self->fd,
_IOMMU_TEST_CMD(IOMMU_TEST_OP_MD_CHECK_MAP),
&check_map_cmd));
+ if (variant->new_hwpt) {
+ /* Detach from the new hwpt for its destroy() */
+ test_cmd_mock_domain_replace(mock_stdev_id,
+ mock_hwpt_id);
+ test_ioctl_destroy(new_hwpt_id);
+ }
test_ioctl_destroy(mock_stdev_id);
test_cmd_destroy_access_pages(
access_cmd.id,
diff --git a/tools/testing/selftests/iommu/iommufd_utils.h b/tools/testing/selftests/iommu/iommufd_utils.h
index 8dced7ef9118..e2c26168ec89 100644
--- a/tools/testing/selftests/iommu/iommufd_utils.h
+++ b/tools/testing/selftests/iommu/iommufd_utils.h
@@ -118,6 +118,42 @@ static int _test_cmd_hwpt_alloc(int fd, __u32 device_id, __u32 pt_id,
#define test_cmd_hwpt_alloc(device_id, pt_id, hwpt_id) \
ASSERT_EQ(0, _test_cmd_hwpt_alloc(self->fd, device_id, pt_id, hwpt_id))
+#define test_err_cmd_hwpt_alloc(_errno, device_id, pt_id, hwpt_id) \
+ EXPECT_ERRNO(_errno, _test_cmd_hwpt_alloc(self->fd, device_id, \
+ pt_id, hwpt_id))
+
+static int _test_cmd_hwpt_alloc_nested(int fd, __u32 device_id, __u32 parent_id,
+ __u32 *hwpt_id)
+{
+ struct iommu_hwpt_selftest data = {
+ .flags = IOMMU_TEST_FLAG_NESTED,
+ .test_config = IOMMU_TEST_IOTLB_DEFAULT,
+ };
+ struct iommu_hwpt_alloc cmd = {
+ .size = sizeof(cmd),
+ .dev_id = device_id,
+ .pt_id = parent_id,
+ .hwpt_type = IOMMU_HWPT_TYPE_SELFTTEST,
+ .data_len = sizeof(data),
+ .data_uptr = (uint64_t)&data,
+ };
+ int ret;
+
+ ret = ioctl(fd, IOMMU_HWPT_ALLOC, &cmd);
+ if (ret)
+ return ret;
+ if (hwpt_id)
+ *hwpt_id = cmd.out_hwpt_id;
+ return 0;
+}
+
+#define test_cmd_hwpt_alloc_nested(device_id, parent_id, hwpt_id) \
+ ASSERT_EQ(0, _test_cmd_hwpt_alloc_nested(self->fd, device_id, \
+ parent_id, hwpt_id))
+#define test_err_cmd_hwpt_alloc_nested(_errno, device_id, parent_id, hwpt_id) \
+ EXPECT_ERRNO(_errno, \
+ _test_cmd_hwpt_alloc_nested(self->fd, device_id, \
+ parent_id, hwpt_id))
static int _test_cmd_access_replace_ioas(int fd, __u32 access_id,
unsigned int ioas_id)
--
2.34.1
next prev parent reply other threads:[~2023-05-11 14:38 UTC|newest]
Thread overview: 64+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-05-11 14:38 [PATCH v2 00/11] iommufd: Add nesting infrastructure Yi Liu
2023-05-11 14:38 ` [PATCH v2 01/11] iommu: Add new iommu op to create domains owned by userspace Yi Liu
2023-05-19 8:47 ` Tian, Kevin
2023-05-19 18:45 ` Nicolin Chen
2023-05-24 5:02 ` Tian, Kevin
2023-05-24 5:23 ` Nicolin Chen
2023-05-24 7:48 ` Tian, Kevin
2023-05-25 1:41 ` Nicolin Chen
2023-06-06 14:08 ` Jason Gunthorpe
2023-06-06 19:43 ` Nicolin Chen
2023-06-07 0:14 ` Jason Gunthorpe
2023-05-11 14:38 ` [PATCH v2 02/11] iommu: Add nested domain support Yi Liu
2023-05-19 8:51 ` Tian, Kevin
2023-05-19 18:49 ` Nicolin Chen
2023-05-24 5:03 ` Tian, Kevin
2023-05-24 5:28 ` Nicolin Chen
2023-05-11 14:38 ` [PATCH v2 03/11] iommufd/hw_pagetable: Use domain_alloc_user op for domain allocation Yi Liu
2023-05-19 8:56 ` Tian, Kevin
2023-05-19 18:57 ` Nicolin Chen
2023-05-24 5:04 ` Tian, Kevin
2023-05-11 14:38 ` [PATCH v2 04/11] iommufd: Pass parent hwpt and user_data to iommufd_hw_pagetable_alloc() Yi Liu
2023-05-19 9:06 ` Tian, Kevin
2023-05-19 19:09 ` Nicolin Chen
2023-05-24 5:11 ` Tian, Kevin
2023-05-24 5:31 ` Nicolin Chen
2023-05-19 9:34 ` Tian, Kevin
2023-05-19 19:12 ` Nicolin Chen
2023-05-11 14:38 ` [PATCH v2 05/11] iommufd/hw_pagetable: Do not populate user-managed hw_pagetables Yi Liu
2023-05-11 14:38 ` [PATCH v2 06/11] iommufd: IOMMU_HWPT_ALLOC allocation with user data Yi Liu
2023-05-17 3:08 ` Liu, Jingqi
2023-05-19 19:34 ` Nicolin Chen
2023-05-19 9:41 ` Tian, Kevin
2023-05-19 19:48 ` Nicolin Chen
2023-05-24 5:16 ` Tian, Kevin
2023-05-24 5:40 ` Nicolin Chen
2023-05-24 7:55 ` Tian, Kevin
2023-05-11 14:38 ` [PATCH v2 07/11] iommufd: Add IOMMU_HWPT_INVALIDATE Yi Liu
2023-05-11 14:38 ` [PATCH v2 08/11] iommufd/selftest: Add domain_alloc_user() support in iommu mock Yi Liu
2023-05-11 14:38 ` Yi Liu [this message]
2023-05-11 14:38 ` [PATCH v2 10/11] iommufd/selftest: Add IOMMU_TEST_OP_MD_CHECK_IOTLB test op Yi Liu
2023-05-11 14:38 ` [PATCH v2 11/11] iommufd/selftest: Add coverage for IOMMU_HWPT_INVALIDATE ioctl Yi Liu
2023-05-19 9:56 ` [PATCH v2 00/11] iommufd: Add nesting infrastructure Tian, Kevin
2023-05-19 11:49 ` Jason Gunthorpe
2023-05-24 3:48 ` Tian, Kevin
2023-06-06 14:18 ` Jason Gunthorpe
2023-06-16 2:43 ` Tian, Kevin
2023-06-19 12:37 ` Jason Gunthorpe
2023-06-20 1:43 ` Tian, Kevin
2023-06-20 12:47 ` Jason Gunthorpe
2023-06-21 6:02 ` Tian, Kevin
2023-06-21 7:09 ` Liu, Yi L
2023-06-21 12:04 ` Jason Gunthorpe
2023-06-26 6:32 ` Tian, Kevin
2023-06-21 17:13 ` Nicolin Chen
2023-06-26 6:42 ` Tian, Kevin
2023-06-26 13:05 ` Jason Gunthorpe
2023-06-26 17:28 ` Nicolin Chen
2023-06-27 6:02 ` Tian, Kevin
2023-06-27 16:01 ` Jason Gunthorpe
2023-06-28 2:47 ` Tian, Kevin
2023-06-28 12:36 ` Jason Gunthorpe
2023-06-29 2:16 ` Tian, Kevin
2023-06-21 8:29 ` Duan, Zhenzhong
2023-06-21 12:07 ` Jason Gunthorpe
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20230511143844.22693-10-yi.l.liu@intel.com \
--to=yi.l.liu@intel.com \
--cc=alex.williamson@redhat.com \
--cc=baolu.lu@linux.intel.com \
--cc=chao.p.peng@linux.intel.com \
--cc=cohuck@redhat.com \
--cc=eric.auger@redhat.com \
--cc=iommu@lists.linux.dev \
--cc=jasowang@redhat.com \
--cc=jgg@nvidia.com \
--cc=joro@8bytes.org \
--cc=kevin.tian@intel.com \
--cc=kvm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-kselftest@vger.kernel.org \
--cc=lulu@redhat.com \
--cc=mjrosato@linux.ibm.com \
--cc=nicolinc@nvidia.com \
--cc=peterx@redhat.com \
--cc=robin.murphy@arm.com \
--cc=shameerali.kolothum.thodi@huawei.com \
--cc=suravee.suthikulpanit@amd.com \
--cc=yi.y.sun@linux.intel.com \
--cc=zhenzhong.duan@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox