* [PATCH v2 01/12] iommu/arm-smmu-v3: Put iopf enablement in the domain attach path
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-24 5:16 ` [PATCH v2 02/12] iommu/vt-d: Check if SVA is supported when attaching the SVA domain Lu Baolu
` (11 subsequent siblings)
12 siblings, 0 replies; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Jason Gunthorpe, Lu Baolu
From: Jason Gunthorpe <jgg@nvidia.com>
SMMUv3 co-mingles FEAT_IOPF and FEAT_SVA behaviors so that fault reporting
doesn't work unless both are enabled. This is not correct and causes
problems for iommufd which does not enable FEAT_SVA for it's fault capable
domains.
These APIs are both obsolete, update SMMUv3 to use the new method like AMD
implements.
A driver should enable iopf support when a domain with an iopf_handler is
attached, and disable iopf support when the domain is removed.
Move the fault support logic to sva domain allocation and to domain
attach, refusing to create or attach fault capable domains if the HW
doesn't support it.
Move all the logic for controlling the iopf queue under
arm_smmu_attach_prepare(). Keep track of the number of domains on the
master (over all the SSIDs) that require iopf. When the first domain
requiring iopf is attached create the iopf queue, when the last domain is
detached destroy it.
Turn FEAT_IOPF and FEAT_SVA into no ops.
Remove the sva_lock, this is all protected by the group mutex.
Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
---
.../iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c | 86 +-------------
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c | 105 +++++++++++++-----
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h | 39 ++-----
3 files changed, 91 insertions(+), 139 deletions(-)
diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
index 9ba596430e7c..605d1dd0e1cc 100644
--- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
+++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3-sva.c
@@ -13,8 +13,6 @@
#include "arm-smmu-v3.h"
#include "../../io-pgtable-arm.h"
-static DEFINE_MUTEX(sva_lock);
-
static void __maybe_unused
arm_smmu_update_s1_domain_cd_entry(struct arm_smmu_domain *smmu_domain)
{
@@ -257,84 +255,6 @@ bool arm_smmu_sva_supported(struct arm_smmu_device *smmu)
return true;
}
-bool arm_smmu_master_iopf_supported(struct arm_smmu_master *master)
-{
- /* We're not keeping track of SIDs in fault events */
- if (master->num_streams != 1)
- return false;
-
- return master->stall_enabled;
-}
-
-bool arm_smmu_master_sva_supported(struct arm_smmu_master *master)
-{
- if (!(master->smmu->features & ARM_SMMU_FEAT_SVA))
- return false;
-
- /* SSID support is mandatory for the moment */
- return master->ssid_bits;
-}
-
-bool arm_smmu_master_sva_enabled(struct arm_smmu_master *master)
-{
- bool enabled;
-
- mutex_lock(&sva_lock);
- enabled = master->sva_enabled;
- mutex_unlock(&sva_lock);
- return enabled;
-}
-
-static int arm_smmu_master_sva_enable_iopf(struct arm_smmu_master *master)
-{
- struct device *dev = master->dev;
-
- /*
- * Drivers for devices supporting PRI or stall should enable IOPF first.
- * Others have device-specific fault handlers and don't need IOPF.
- */
- if (!arm_smmu_master_iopf_supported(master))
- return 0;
-
- if (!master->iopf_enabled)
- return -EINVAL;
-
- return iopf_queue_add_device(master->smmu->evtq.iopf, dev);
-}
-
-static void arm_smmu_master_sva_disable_iopf(struct arm_smmu_master *master)
-{
- struct device *dev = master->dev;
-
- if (!master->iopf_enabled)
- return;
-
- iopf_queue_remove_device(master->smmu->evtq.iopf, dev);
-}
-
-int arm_smmu_master_enable_sva(struct arm_smmu_master *master)
-{
- int ret;
-
- mutex_lock(&sva_lock);
- ret = arm_smmu_master_sva_enable_iopf(master);
- if (!ret)
- master->sva_enabled = true;
- mutex_unlock(&sva_lock);
-
- return ret;
-}
-
-int arm_smmu_master_disable_sva(struct arm_smmu_master *master)
-{
- mutex_lock(&sva_lock);
- arm_smmu_master_sva_disable_iopf(master);
- master->sva_enabled = false;
- mutex_unlock(&sva_lock);
-
- return 0;
-}
-
void arm_smmu_sva_notifier_synchronize(void)
{
/*
@@ -353,6 +273,9 @@ static int arm_smmu_sva_set_dev_pasid(struct iommu_domain *domain,
struct arm_smmu_cd target;
int ret;
+ if (!(master->smmu->features & ARM_SMMU_FEAT_SVA))
+ return -EOPNOTSUPP;
+
/* Prevent arm_smmu_mm_release from being called while we are attaching */
if (!mmget_not_zero(domain->mm))
return -EINVAL;
@@ -406,6 +329,9 @@ struct iommu_domain *arm_smmu_sva_domain_alloc(struct device *dev,
u32 asid;
int ret;
+ if (!(master->smmu->features & ARM_SMMU_FEAT_SVA))
+ return ERR_PTR(-EOPNOTSUPP);
+
smmu_domain = arm_smmu_domain_alloc();
if (IS_ERR(smmu_domain))
return ERR_CAST(smmu_domain);
diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
index 358072b4e293..4aaf9b7bf862 100644
--- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
+++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
@@ -2712,6 +2712,7 @@ static void arm_smmu_disable_pasid(struct arm_smmu_master *master)
static struct arm_smmu_master_domain *
arm_smmu_find_master_domain(struct arm_smmu_domain *smmu_domain,
+ struct iommu_domain *domain,
struct arm_smmu_master *master,
ioasid_t ssid, bool nested_ats_flush)
{
@@ -2722,6 +2723,7 @@ arm_smmu_find_master_domain(struct arm_smmu_domain *smmu_domain,
list_for_each_entry(master_domain, &smmu_domain->devices,
devices_elm) {
if (master_domain->master == master &&
+ master_domain->domain == domain &&
master_domain->ssid == ssid &&
master_domain->nested_ats_flush == nested_ats_flush)
return master_domain;
@@ -2748,6 +2750,54 @@ to_smmu_domain_devices(struct iommu_domain *domain)
return NULL;
}
+static int arm_smmu_enable_iopf(struct arm_smmu_master *master,
+ struct arm_smmu_master_domain *master_domain)
+{
+ int ret;
+
+ iommu_group_mutex_assert(master->dev);
+
+ if (!IS_ENABLED(CONFIG_ARM_SMMU_V3_SVA))
+ return -EOPNOTSUPP;
+
+ /*
+ * Drivers for devices supporting PRI or stall require iopf others have
+ * device-specific fault handlers and don't need IOPF, so this is not a
+ * failure.
+ */
+ if (!master->stall_enabled)
+ return 0;
+
+ /* We're not keeping track of SIDs in fault events */
+ if (master->num_streams != 1)
+ return -EOPNOTSUPP;
+
+ if (master->iopf_refcount) {
+ master->iopf_refcount++;
+ master_domain->using_iopf = true;
+ return 0;
+ }
+
+ ret = iopf_queue_add_device(master->smmu->evtq.iopf, master->dev);
+ if (ret)
+ return ret;
+ master->iopf_refcount = 1;
+ master_domain->using_iopf = true;
+ return 0;
+}
+
+static void arm_smmu_disable_iopf(struct arm_smmu_master *master)
+{
+ iommu_group_mutex_assert(master->dev);
+
+ if (!IS_ENABLED(CONFIG_ARM_SMMU_V3_SVA))
+ return;
+
+ master->iopf_refcount--;
+ if (master->iopf_refcount == 0)
+ iopf_queue_remove_device(master->smmu->evtq.iopf, master->dev);
+}
+
static void arm_smmu_remove_master_domain(struct arm_smmu_master *master,
struct iommu_domain *domain,
ioasid_t ssid)
@@ -2764,15 +2814,20 @@ static void arm_smmu_remove_master_domain(struct arm_smmu_master *master,
nested_ats_flush = to_smmu_nested_domain(domain)->enable_ats;
spin_lock_irqsave(&smmu_domain->devices_lock, flags);
- master_domain = arm_smmu_find_master_domain(smmu_domain, master, ssid,
- nested_ats_flush);
+ master_domain = arm_smmu_find_master_domain(smmu_domain, domain, master,
+ ssid, nested_ats_flush);
if (master_domain) {
list_del(&master_domain->devices_elm);
- kfree(master_domain);
if (master->ats_enabled)
atomic_dec(&smmu_domain->nr_ats_masters);
}
spin_unlock_irqrestore(&smmu_domain->devices_lock, flags);
+
+ if (master_domain) {
+ if (master_domain->using_iopf)
+ arm_smmu_disable_iopf(master);
+ kfree(master_domain);
+ }
}
/*
@@ -2803,6 +2858,7 @@ int arm_smmu_attach_prepare(struct arm_smmu_attach_state *state,
struct arm_smmu_domain *smmu_domain =
to_smmu_domain_devices(new_domain);
unsigned long flags;
+ int ret;
/*
* arm_smmu_share_asid() must not see two domains pointing to the same
@@ -2835,12 +2891,19 @@ int arm_smmu_attach_prepare(struct arm_smmu_attach_state *state,
master_domain = kzalloc(sizeof(*master_domain), GFP_KERNEL);
if (!master_domain)
return -ENOMEM;
+ master_domain->domain = new_domain;
master_domain->master = master;
master_domain->ssid = state->ssid;
if (new_domain->type == IOMMU_DOMAIN_NESTED)
master_domain->nested_ats_flush =
to_smmu_nested_domain(new_domain)->enable_ats;
+ if (new_domain->iopf_handler) {
+ ret = arm_smmu_enable_iopf(master, master_domain);
+ if (ret)
+ goto err_free_master_domain;
+ }
+
/*
* During prepare we want the current smmu_domain and new
* smmu_domain to be in the devices list before we change any
@@ -2860,8 +2923,8 @@ int arm_smmu_attach_prepare(struct arm_smmu_attach_state *state,
!arm_smmu_master_canwbs(master)) {
spin_unlock_irqrestore(&smmu_domain->devices_lock,
flags);
- kfree(master_domain);
- return -EINVAL;
+ ret = -EINVAL;
+ goto err_iopf;
}
if (state->ats_enabled)
@@ -2880,6 +2943,13 @@ int arm_smmu_attach_prepare(struct arm_smmu_attach_state *state,
wmb();
}
return 0;
+
+err_iopf:
+ if (master_domain && master_domain->using_iopf)
+ arm_smmu_disable_iopf(master);
+err_free_master_domain:
+ kfree(master_domain);
+ return ret;
}
/*
@@ -3475,8 +3545,7 @@ static void arm_smmu_release_device(struct device *dev)
{
struct arm_smmu_master *master = dev_iommu_priv_get(dev);
- if (WARN_ON(arm_smmu_master_sva_enabled(master)))
- iopf_queue_remove_device(master->smmu->evtq.iopf, dev);
+ WARN_ON(master->iopf_refcount);
/* Put the STE back to what arm_smmu_init_strtab() sets */
if (dev->iommu->require_direct)
@@ -3561,18 +3630,8 @@ static int arm_smmu_dev_enable_feature(struct device *dev,
switch (feat) {
case IOMMU_DEV_FEAT_IOPF:
- if (!arm_smmu_master_iopf_supported(master))
- return -EINVAL;
- if (master->iopf_enabled)
- return -EBUSY;
- master->iopf_enabled = true;
- return 0;
case IOMMU_DEV_FEAT_SVA:
- if (!arm_smmu_master_sva_supported(master))
- return -EINVAL;
- if (arm_smmu_master_sva_enabled(master))
- return -EBUSY;
- return arm_smmu_master_enable_sva(master);
+ return 0;
default:
return -EINVAL;
}
@@ -3588,16 +3647,8 @@ static int arm_smmu_dev_disable_feature(struct device *dev,
switch (feat) {
case IOMMU_DEV_FEAT_IOPF:
- if (!master->iopf_enabled)
- return -EINVAL;
- if (master->sva_enabled)
- return -EBUSY;
- master->iopf_enabled = false;
- return 0;
case IOMMU_DEV_FEAT_SVA:
- if (!arm_smmu_master_sva_enabled(master))
- return -EINVAL;
- return arm_smmu_master_disable_sva(master);
+ return 0;
default:
return -EINVAL;
}
diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h
index bd9d7c85576a..fe6b88affa4a 100644
--- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h
+++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.h
@@ -830,9 +830,8 @@ struct arm_smmu_master {
bool ats_enabled : 1;
bool ste_ats_enabled : 1;
bool stall_enabled;
- bool sva_enabled;
- bool iopf_enabled;
unsigned int ssid_bits;
+ unsigned int iopf_refcount;
};
/* SMMU private data for an IOMMU domain */
@@ -908,8 +907,14 @@ void arm_smmu_make_sva_cd(struct arm_smmu_cd *target,
struct arm_smmu_master_domain {
struct list_head devices_elm;
struct arm_smmu_master *master;
+ /*
+ * For nested domains the master_domain is threaded onto the S2 parent,
+ * this points to the IOMMU_DOMAIN_NESTED to disambiguate the masters.
+ */
+ struct iommu_domain *domain;
ioasid_t ssid;
bool nested_ats_flush : 1;
+ bool using_iopf : 1;
};
static inline struct arm_smmu_domain *to_smmu_domain(struct iommu_domain *dom)
@@ -987,11 +992,6 @@ int arm_smmu_cmdq_issue_cmdlist(struct arm_smmu_device *smmu,
#ifdef CONFIG_ARM_SMMU_V3_SVA
bool arm_smmu_sva_supported(struct arm_smmu_device *smmu);
-bool arm_smmu_master_sva_supported(struct arm_smmu_master *master);
-bool arm_smmu_master_sva_enabled(struct arm_smmu_master *master);
-int arm_smmu_master_enable_sva(struct arm_smmu_master *master);
-int arm_smmu_master_disable_sva(struct arm_smmu_master *master);
-bool arm_smmu_master_iopf_supported(struct arm_smmu_master *master);
void arm_smmu_sva_notifier_synchronize(void);
struct iommu_domain *arm_smmu_sva_domain_alloc(struct device *dev,
struct mm_struct *mm);
@@ -1001,31 +1001,6 @@ static inline bool arm_smmu_sva_supported(struct arm_smmu_device *smmu)
return false;
}
-static inline bool arm_smmu_master_sva_supported(struct arm_smmu_master *master)
-{
- return false;
-}
-
-static inline bool arm_smmu_master_sva_enabled(struct arm_smmu_master *master)
-{
- return false;
-}
-
-static inline int arm_smmu_master_enable_sva(struct arm_smmu_master *master)
-{
- return -ENODEV;
-}
-
-static inline int arm_smmu_master_disable_sva(struct arm_smmu_master *master)
-{
- return -ENODEV;
-}
-
-static inline bool arm_smmu_master_iopf_supported(struct arm_smmu_master *master)
-{
- return false;
-}
-
static inline void arm_smmu_sva_notifier_synchronize(void) {}
#define arm_smmu_sva_domain_alloc NULL
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* [PATCH v2 02/12] iommu/vt-d: Check if SVA is supported when attaching the SVA domain
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
2025-02-24 5:16 ` [PATCH v2 01/12] iommu/arm-smmu-v3: Put iopf enablement in the domain attach path Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 7:23 ` Tian, Kevin
2025-03-04 7:44 ` Yi Liu
2025-02-24 5:16 ` [PATCH v2 03/12] iommu: Remove IOMMU_DEV_FEAT_SVA Lu Baolu
` (10 subsequent siblings)
12 siblings, 2 replies; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Jason Gunthorpe, Lu Baolu
From: Jason Gunthorpe <jgg@nvidia.com>
Attach of a SVA domain should fail if SVA is not supported, move the check
for SVA support out of IOMMU_DEV_FEAT_SVA and into attach.
Also check when allocating a SVA domain to match other drivers.
Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
---
drivers/iommu/intel/iommu.c | 37 +------------------------------
drivers/iommu/intel/svm.c | 43 +++++++++++++++++++++++++++++++++++++
2 files changed, 44 insertions(+), 36 deletions(-)
diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
index cc46098f875b..a4048de66378 100644
--- a/drivers/iommu/intel/iommu.c
+++ b/drivers/iommu/intel/iommu.c
@@ -3851,41 +3851,6 @@ static struct iommu_group *intel_iommu_device_group(struct device *dev)
return generic_device_group(dev);
}
-static int intel_iommu_enable_sva(struct device *dev)
-{
- struct device_domain_info *info = dev_iommu_priv_get(dev);
- struct intel_iommu *iommu;
-
- if (!info || dmar_disabled)
- return -EINVAL;
-
- iommu = info->iommu;
- if (!iommu)
- return -EINVAL;
-
- if (!(iommu->flags & VTD_FLAG_SVM_CAPABLE))
- return -ENODEV;
-
- if (!info->pasid_enabled || !info->ats_enabled)
- return -EINVAL;
-
- /*
- * Devices having device-specific I/O fault handling should not
- * support PCI/PRI. The IOMMU side has no means to check the
- * capability of device-specific IOPF. Therefore, IOMMU can only
- * default that if the device driver enables SVA on a non-PRI
- * device, it will handle IOPF in its own way.
- */
- if (!info->pri_supported)
- return 0;
-
- /* Devices supporting PRI should have it enabled. */
- if (!info->pri_enabled)
- return -EINVAL;
-
- return 0;
-}
-
static int context_flip_pri(struct device_domain_info *info, bool enable)
{
struct intel_iommu *iommu = info->iommu;
@@ -4006,7 +3971,7 @@ intel_iommu_dev_enable_feat(struct device *dev, enum iommu_dev_features feat)
return intel_iommu_enable_iopf(dev);
case IOMMU_DEV_FEAT_SVA:
- return intel_iommu_enable_sva(dev);
+ return 0;
default:
return -ENODEV;
diff --git a/drivers/iommu/intel/svm.c b/drivers/iommu/intel/svm.c
index f5569347591f..ba93123cb4eb 100644
--- a/drivers/iommu/intel/svm.c
+++ b/drivers/iommu/intel/svm.c
@@ -110,6 +110,41 @@ static const struct mmu_notifier_ops intel_mmuops = {
.free_notifier = intel_mm_free_notifier,
};
+static int intel_iommu_sva_supported(struct device *dev)
+{
+ struct device_domain_info *info = dev_iommu_priv_get(dev);
+ struct intel_iommu *iommu;
+
+ if (!info || dmar_disabled)
+ return -EINVAL;
+
+ iommu = info->iommu;
+ if (!iommu)
+ return -EINVAL;
+
+ if (!(iommu->flags & VTD_FLAG_SVM_CAPABLE))
+ return -ENODEV;
+
+ if (!info->pasid_enabled || !info->ats_enabled)
+ return -EINVAL;
+
+ /*
+ * Devices having device-specific I/O fault handling should not
+ * support PCI/PRI. The IOMMU side has no means to check the
+ * capability of device-specific IOPF. Therefore, IOMMU can only
+ * default that if the device driver enables SVA on a non-PRI
+ * device, it will handle IOPF in its own way.
+ */
+ if (!info->pri_supported)
+ return 0;
+
+ /* Devices supporting PRI should have it enabled. */
+ if (!info->pri_enabled)
+ return -EINVAL;
+
+ return 0;
+}
+
static int intel_svm_set_dev_pasid(struct iommu_domain *domain,
struct device *dev, ioasid_t pasid,
struct iommu_domain *old)
@@ -121,6 +156,10 @@ static int intel_svm_set_dev_pasid(struct iommu_domain *domain,
unsigned long sflags;
int ret = 0;
+ ret = intel_iommu_sva_supported(dev);
+ if (ret)
+ return ret;
+
dev_pasid = domain_add_dev_pasid(domain, dev, pasid);
if (IS_ERR(dev_pasid))
return PTR_ERR(dev_pasid);
@@ -161,6 +200,10 @@ struct iommu_domain *intel_svm_domain_alloc(struct device *dev,
struct dmar_domain *domain;
int ret;
+ ret = intel_iommu_sva_supported(dev);
+ if (ret)
+ return ERR_PTR(ret);
+
domain = kzalloc(sizeof(*domain), GFP_KERNEL);
if (!domain)
return ERR_PTR(-ENOMEM);
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 02/12] iommu/vt-d: Check if SVA is supported when attaching the SVA domain
2025-02-24 5:16 ` [PATCH v2 02/12] iommu/vt-d: Check if SVA is supported when attaching the SVA domain Lu Baolu
@ 2025-02-25 7:23 ` Tian, Kevin
2025-03-04 7:44 ` Yi Liu
1 sibling, 0 replies; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 7:23 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org,
Jason Gunthorpe
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> From: Jason Gunthorpe <jgg@nvidia.com>
>
> Attach of a SVA domain should fail if SVA is not supported, move the check
> for SVA support out of IOMMU_DEV_FEAT_SVA and into attach.
>
> Also check when allocating a SVA domain to match other drivers.
>
> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 02/12] iommu/vt-d: Check if SVA is supported when attaching the SVA domain
2025-02-24 5:16 ` [PATCH v2 02/12] iommu/vt-d: Check if SVA is supported when attaching the SVA domain Lu Baolu
2025-02-25 7:23 ` Tian, Kevin
@ 2025-03-04 7:44 ` Yi Liu
1 sibling, 0 replies; 38+ messages in thread
From: Yi Liu @ 2025-03-04 7:44 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe, Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Jason Gunthorpe
On 2025/2/24 13:16, Lu Baolu wrote:
> From: Jason Gunthorpe <jgg@nvidia.com>
>
> Attach of a SVA domain should fail if SVA is not supported, move the check
> for SVA support out of IOMMU_DEV_FEAT_SVA and into attach.
>
> Also check when allocating a SVA domain to match other drivers.
>
> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
> ---
> drivers/iommu/intel/iommu.c | 37 +------------------------------
> drivers/iommu/intel/svm.c | 43 +++++++++++++++++++++++++++++++++++++
> 2 files changed, 44 insertions(+), 36 deletions(-)
>
Reviewed-by: Yi Liu <yi.l.liu@intel.com>
> diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
> index cc46098f875b..a4048de66378 100644
> --- a/drivers/iommu/intel/iommu.c
> +++ b/drivers/iommu/intel/iommu.c
> @@ -3851,41 +3851,6 @@ static struct iommu_group *intel_iommu_device_group(struct device *dev)
> return generic_device_group(dev);
> }
>
> -static int intel_iommu_enable_sva(struct device *dev)
> -{
> - struct device_domain_info *info = dev_iommu_priv_get(dev);
> - struct intel_iommu *iommu;
> -
> - if (!info || dmar_disabled)
> - return -EINVAL;
> -
> - iommu = info->iommu;
> - if (!iommu)
> - return -EINVAL;
> -
> - if (!(iommu->flags & VTD_FLAG_SVM_CAPABLE))
> - return -ENODEV;
> -
> - if (!info->pasid_enabled || !info->ats_enabled)
> - return -EINVAL;
> -
> - /*
> - * Devices having device-specific I/O fault handling should not
> - * support PCI/PRI. The IOMMU side has no means to check the
> - * capability of device-specific IOPF. Therefore, IOMMU can only
> - * default that if the device driver enables SVA on a non-PRI
> - * device, it will handle IOPF in its own way.
> - */
> - if (!info->pri_supported)
> - return 0;
> -
> - /* Devices supporting PRI should have it enabled. */
> - if (!info->pri_enabled)
> - return -EINVAL;
> -
> - return 0;
> -}
> -
> static int context_flip_pri(struct device_domain_info *info, bool enable)
> {
> struct intel_iommu *iommu = info->iommu;
> @@ -4006,7 +3971,7 @@ intel_iommu_dev_enable_feat(struct device *dev, enum iommu_dev_features feat)
> return intel_iommu_enable_iopf(dev);
>
> case IOMMU_DEV_FEAT_SVA:
> - return intel_iommu_enable_sva(dev);
> + return 0;
>
> default:
> return -ENODEV;
> diff --git a/drivers/iommu/intel/svm.c b/drivers/iommu/intel/svm.c
> index f5569347591f..ba93123cb4eb 100644
> --- a/drivers/iommu/intel/svm.c
> +++ b/drivers/iommu/intel/svm.c
> @@ -110,6 +110,41 @@ static const struct mmu_notifier_ops intel_mmuops = {
> .free_notifier = intel_mm_free_notifier,
> };
>
> +static int intel_iommu_sva_supported(struct device *dev)
> +{
> + struct device_domain_info *info = dev_iommu_priv_get(dev);
> + struct intel_iommu *iommu;
> +
> + if (!info || dmar_disabled)
> + return -EINVAL;
> +
> + iommu = info->iommu;
> + if (!iommu)
> + return -EINVAL;
> +
> + if (!(iommu->flags & VTD_FLAG_SVM_CAPABLE))
> + return -ENODEV;
> +
> + if (!info->pasid_enabled || !info->ats_enabled)
> + return -EINVAL;
> +
> + /*
> + * Devices having device-specific I/O fault handling should not
> + * support PCI/PRI. The IOMMU side has no means to check the
> + * capability of device-specific IOPF. Therefore, IOMMU can only
> + * default that if the device driver enables SVA on a non-PRI
> + * device, it will handle IOPF in its own way.
> + */
> + if (!info->pri_supported)
> + return 0;
> +
> + /* Devices supporting PRI should have it enabled. */
> + if (!info->pri_enabled)
> + return -EINVAL;
> +
> + return 0;
> +}
> +
> static int intel_svm_set_dev_pasid(struct iommu_domain *domain,
> struct device *dev, ioasid_t pasid,
> struct iommu_domain *old)
> @@ -121,6 +156,10 @@ static int intel_svm_set_dev_pasid(struct iommu_domain *domain,
> unsigned long sflags;
> int ret = 0;
>
> + ret = intel_iommu_sva_supported(dev);
> + if (ret)
> + return ret;
> +
> dev_pasid = domain_add_dev_pasid(domain, dev, pasid);
> if (IS_ERR(dev_pasid))
> return PTR_ERR(dev_pasid);
> @@ -161,6 +200,10 @@ struct iommu_domain *intel_svm_domain_alloc(struct device *dev,
> struct dmar_domain *domain;
> int ret;
>
> + ret = intel_iommu_sva_supported(dev);
> + if (ret)
> + return ERR_PTR(ret);
> +
> domain = kzalloc(sizeof(*domain), GFP_KERNEL);
> if (!domain)
> return ERR_PTR(-ENOMEM);
--
Regards,
Yi Liu
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 03/12] iommu: Remove IOMMU_DEV_FEAT_SVA
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
2025-02-24 5:16 ` [PATCH v2 01/12] iommu/arm-smmu-v3: Put iopf enablement in the domain attach path Lu Baolu
2025-02-24 5:16 ` [PATCH v2 02/12] iommu/vt-d: Check if SVA is supported when attaching the SVA domain Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 7:24 ` Tian, Kevin
2025-03-04 7:50 ` Yi Liu
2025-02-24 5:16 ` [PATCH v2 04/12] iommu/vt-d: Move scalable mode ATS enablement to probe path Lu Baolu
` (9 subsequent siblings)
12 siblings, 2 replies; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Jason Gunthorpe, Lu Baolu
From: Jason Gunthorpe <jgg@nvidia.com>
None of the drivers implement anything here anymore, remove the dead code.
Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
---
drivers/accel/amdxdna/aie2_pci.c | 13 ++-----------
drivers/dma/idxd/init.c | 8 +-------
drivers/iommu/amd/iommu.c | 2 --
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c | 2 --
drivers/iommu/intel/iommu.c | 6 ------
drivers/iommu/iommu-sva.c | 3 ---
drivers/misc/uacce/uacce.c | 9 ---------
include/linux/iommu.h | 9 +--------
8 files changed, 4 insertions(+), 48 deletions(-)
diff --git a/drivers/accel/amdxdna/aie2_pci.c b/drivers/accel/amdxdna/aie2_pci.c
index 5a058e565b01..c6cf7068d23c 100644
--- a/drivers/accel/amdxdna/aie2_pci.c
+++ b/drivers/accel/amdxdna/aie2_pci.c
@@ -512,12 +512,6 @@ static int aie2_init(struct amdxdna_dev *xdna)
goto release_fw;
}
- ret = iommu_dev_enable_feature(&pdev->dev, IOMMU_DEV_FEAT_SVA);
- if (ret) {
- XDNA_ERR(xdna, "Enable PASID failed, ret %d", ret);
- goto free_irq;
- }
-
psp_conf.fw_size = fw->size;
psp_conf.fw_buf = fw->data;
for (i = 0; i < PSP_MAX_REGS; i++)
@@ -526,14 +520,14 @@ static int aie2_init(struct amdxdna_dev *xdna)
if (!ndev->psp_hdl) {
XDNA_ERR(xdna, "failed to create psp");
ret = -ENOMEM;
- goto disable_sva;
+ goto free_irq;
}
xdna->dev_handle = ndev;
ret = aie2_hw_start(xdna);
if (ret) {
XDNA_ERR(xdna, "start npu failed, ret %d", ret);
- goto disable_sva;
+ goto free_irq;
}
ret = aie2_mgmt_fw_query(ndev);
@@ -584,8 +578,6 @@ static int aie2_init(struct amdxdna_dev *xdna)
aie2_error_async_events_free(ndev);
stop_hw:
aie2_hw_stop(xdna);
-disable_sva:
- iommu_dev_disable_feature(&pdev->dev, IOMMU_DEV_FEAT_SVA);
free_irq:
pci_free_irq_vectors(pdev);
release_fw:
@@ -601,7 +593,6 @@ static void aie2_fini(struct amdxdna_dev *xdna)
aie2_hw_stop(xdna);
aie2_error_async_events_free(ndev);
- iommu_dev_disable_feature(&pdev->dev, IOMMU_DEV_FEAT_SVA);
pci_free_irq_vectors(pdev);
}
diff --git a/drivers/dma/idxd/init.c b/drivers/dma/idxd/init.c
index b946f78f85e1..1e5038cca22c 100644
--- a/drivers/dma/idxd/init.c
+++ b/drivers/dma/idxd/init.c
@@ -633,17 +633,11 @@ static int idxd_enable_sva(struct pci_dev *pdev)
ret = iommu_dev_enable_feature(&pdev->dev, IOMMU_DEV_FEAT_IOPF);
if (ret)
return ret;
-
- ret = iommu_dev_enable_feature(&pdev->dev, IOMMU_DEV_FEAT_SVA);
- if (ret)
- iommu_dev_disable_feature(&pdev->dev, IOMMU_DEV_FEAT_IOPF);
-
- return ret;
+ return 0;
}
static void idxd_disable_sva(struct pci_dev *pdev)
{
- iommu_dev_disable_feature(&pdev->dev, IOMMU_DEV_FEAT_SVA);
iommu_dev_disable_feature(&pdev->dev, IOMMU_DEV_FEAT_IOPF);
}
diff --git a/drivers/iommu/amd/iommu.c b/drivers/iommu/amd/iommu.c
index b48a72bd7b23..e3653bdfcd7d 100644
--- a/drivers/iommu/amd/iommu.c
+++ b/drivers/iommu/amd/iommu.c
@@ -2990,7 +2990,6 @@ static int amd_iommu_dev_enable_feature(struct device *dev,
switch (feat) {
case IOMMU_DEV_FEAT_IOPF:
- case IOMMU_DEV_FEAT_SVA:
break;
default:
ret = -EINVAL;
@@ -3006,7 +3005,6 @@ static int amd_iommu_dev_disable_feature(struct device *dev,
switch (feat) {
case IOMMU_DEV_FEAT_IOPF:
- case IOMMU_DEV_FEAT_SVA:
break;
default:
ret = -EINVAL;
diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
index 4aaf9b7bf862..3a2faf4fdd09 100644
--- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
+++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
@@ -3630,7 +3630,6 @@ static int arm_smmu_dev_enable_feature(struct device *dev,
switch (feat) {
case IOMMU_DEV_FEAT_IOPF:
- case IOMMU_DEV_FEAT_SVA:
return 0;
default:
return -EINVAL;
@@ -3647,7 +3646,6 @@ static int arm_smmu_dev_disable_feature(struct device *dev,
switch (feat) {
case IOMMU_DEV_FEAT_IOPF:
- case IOMMU_DEV_FEAT_SVA:
return 0;
default:
return -EINVAL;
diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
index a4048de66378..16dd8f0de76d 100644
--- a/drivers/iommu/intel/iommu.c
+++ b/drivers/iommu/intel/iommu.c
@@ -3970,9 +3970,6 @@ intel_iommu_dev_enable_feat(struct device *dev, enum iommu_dev_features feat)
case IOMMU_DEV_FEAT_IOPF:
return intel_iommu_enable_iopf(dev);
- case IOMMU_DEV_FEAT_SVA:
- return 0;
-
default:
return -ENODEV;
}
@@ -3985,9 +3982,6 @@ intel_iommu_dev_disable_feat(struct device *dev, enum iommu_dev_features feat)
case IOMMU_DEV_FEAT_IOPF:
return intel_iommu_disable_iopf(dev);
- case IOMMU_DEV_FEAT_SVA:
- return 0;
-
default:
return -ENODEV;
}
diff --git a/drivers/iommu/iommu-sva.c b/drivers/iommu/iommu-sva.c
index 503c5d23c1ea..331be2761a75 100644
--- a/drivers/iommu/iommu-sva.c
+++ b/drivers/iommu/iommu-sva.c
@@ -63,9 +63,6 @@ static struct iommu_mm_data *iommu_alloc_mm_data(struct mm_struct *mm, struct de
* reference is taken. Caller must call iommu_sva_unbind_device()
* to release each reference.
*
- * iommu_dev_enable_feature(dev, IOMMU_DEV_FEAT_SVA) must be called first, to
- * initialize the required SVA features.
- *
* On error, returns an ERR_PTR value.
*/
struct iommu_sva *iommu_sva_bind_device(struct device *dev, struct mm_struct *mm)
diff --git a/drivers/misc/uacce/uacce.c b/drivers/misc/uacce/uacce.c
index bdc2e6fda782..2a1db2abeeca 100644
--- a/drivers/misc/uacce/uacce.c
+++ b/drivers/misc/uacce/uacce.c
@@ -479,14 +479,6 @@ static unsigned int uacce_enable_sva(struct device *parent, unsigned int flags)
dev_err(parent, "failed to enable IOPF feature! ret = %pe\n", ERR_PTR(ret));
return flags;
}
-
- ret = iommu_dev_enable_feature(parent, IOMMU_DEV_FEAT_SVA);
- if (ret) {
- dev_err(parent, "failed to enable SVA feature! ret = %pe\n", ERR_PTR(ret));
- iommu_dev_disable_feature(parent, IOMMU_DEV_FEAT_IOPF);
- return flags;
- }
-
return flags | UACCE_DEV_SVA;
}
@@ -495,7 +487,6 @@ static void uacce_disable_sva(struct uacce_device *uacce)
if (!(uacce->flags & UACCE_DEV_SVA))
return;
- iommu_dev_disable_feature(uacce->parent, IOMMU_DEV_FEAT_SVA);
iommu_dev_disable_feature(uacce->parent, IOMMU_DEV_FEAT_IOPF);
}
diff --git a/include/linux/iommu.h b/include/linux/iommu.h
index 38c65e92ecd0..1d0dde49168d 100644
--- a/include/linux/iommu.h
+++ b/include/linux/iommu.h
@@ -303,18 +303,11 @@ struct iommu_iort_rmr_data {
/**
* enum iommu_dev_features - Per device IOMMU features
- * @IOMMU_DEV_FEAT_SVA: Shared Virtual Addresses
- * @IOMMU_DEV_FEAT_IOPF: I/O Page Faults such as PRI or Stall. Generally
- * enabling %IOMMU_DEV_FEAT_SVA requires
- * %IOMMU_DEV_FEAT_IOPF, but some devices manage I/O Page
- * Faults themselves instead of relying on the IOMMU. When
- * supported, this feature must be enabled before and
- * disabled after %IOMMU_DEV_FEAT_SVA.
+ * @IOMMU_DEV_FEAT_IOPF: I/O Page Faults such as PRI or Stall.
*
* Device drivers enable a feature using iommu_dev_enable_feature().
*/
enum iommu_dev_features {
- IOMMU_DEV_FEAT_SVA,
IOMMU_DEV_FEAT_IOPF,
};
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 03/12] iommu: Remove IOMMU_DEV_FEAT_SVA
2025-02-24 5:16 ` [PATCH v2 03/12] iommu: Remove IOMMU_DEV_FEAT_SVA Lu Baolu
@ 2025-02-25 7:24 ` Tian, Kevin
2025-03-04 7:50 ` Yi Liu
1 sibling, 0 replies; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 7:24 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org,
Jason Gunthorpe
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> From: Jason Gunthorpe <jgg@nvidia.com>
>
> None of the drivers implement anything here anymore, remove the dead
> code.
>
> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 03/12] iommu: Remove IOMMU_DEV_FEAT_SVA
2025-02-24 5:16 ` [PATCH v2 03/12] iommu: Remove IOMMU_DEV_FEAT_SVA Lu Baolu
2025-02-25 7:24 ` Tian, Kevin
@ 2025-03-04 7:50 ` Yi Liu
1 sibling, 0 replies; 38+ messages in thread
From: Yi Liu @ 2025-03-04 7:50 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe, Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Jason Gunthorpe
On 2025/2/24 13:16, Lu Baolu wrote:
> From: Jason Gunthorpe <jgg@nvidia.com>
>
> None of the drivers implement anything here anymore, remove the dead code.
>
> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
> ---
> drivers/accel/amdxdna/aie2_pci.c | 13 ++-----------
> drivers/dma/idxd/init.c | 8 +-------
> drivers/iommu/amd/iommu.c | 2 --
> drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c | 2 --
> drivers/iommu/intel/iommu.c | 6 ------
> drivers/iommu/iommu-sva.c | 3 ---
> drivers/misc/uacce/uacce.c | 9 ---------
> include/linux/iommu.h | 9 +--------
> 8 files changed, 4 insertions(+), 48 deletions(-)
Reviewed-by: Yi Liu <yi.l.liu@intel.com>
with a nit:
> diff --git a/drivers/dma/idxd/init.c b/drivers/dma/idxd/init.c
> index b946f78f85e1..1e5038cca22c 100644
> --- a/drivers/dma/idxd/init.c
> +++ b/drivers/dma/idxd/init.c
> @@ -633,17 +633,11 @@ static int idxd_enable_sva(struct pci_dev *pdev)
> ret = iommu_dev_enable_feature(&pdev->dev, IOMMU_DEV_FEAT_IOPF);
> if (ret)
delete this check, and return ret.
> return ret;
> -
> - ret = iommu_dev_enable_feature(&pdev->dev, IOMMU_DEV_FEAT_SVA);
> - if (ret)
> - iommu_dev_disable_feature(&pdev->dev, IOMMU_DEV_FEAT_IOPF);
> -
> - return ret;
> + return 0;
--
Regards,
Yi Liu
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 04/12] iommu/vt-d: Move scalable mode ATS enablement to probe path
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (2 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 03/12] iommu: Remove IOMMU_DEV_FEAT_SVA Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 7:28 ` Tian, Kevin
2025-03-04 7:55 ` Yi Liu
2025-02-24 5:16 ` [PATCH v2 05/12] iommu/vt-d: Move PRI enablement in " Lu Baolu
` (8 subsequent siblings)
12 siblings, 2 replies; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Lu Baolu
Device ATS is currently enabled when a domain is attached to the device
and disabled when the domain is detached. This creates a limitation:
when the IOMMU is operating in scalable mode and IOPF is enabled, the
device's domain cannot be changed.
Remove this limitation by moving ATS enablement to the device probe path.
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
---
drivers/iommu/intel/iommu.c | 78 ++++++++++++++++++-------------------
1 file changed, 38 insertions(+), 40 deletions(-)
diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
index 16dd8f0de76d..f52602bde742 100644
--- a/drivers/iommu/intel/iommu.c
+++ b/drivers/iommu/intel/iommu.c
@@ -1172,34 +1172,6 @@ static bool dev_needs_extra_dtlb_flush(struct pci_dev *pdev)
return true;
}
-static void iommu_enable_pci_caps(struct device_domain_info *info)
-{
- struct pci_dev *pdev;
-
- if (!dev_is_pci(info->dev))
- return;
-
- pdev = to_pci_dev(info->dev);
- if (info->ats_supported && pci_ats_page_aligned(pdev) &&
- !pci_enable_ats(pdev, VTD_PAGE_SHIFT))
- info->ats_enabled = 1;
-}
-
-static void iommu_disable_pci_caps(struct device_domain_info *info)
-{
- struct pci_dev *pdev;
-
- if (!dev_is_pci(info->dev))
- return;
-
- pdev = to_pci_dev(info->dev);
-
- if (info->ats_enabled) {
- pci_disable_ats(pdev);
- info->ats_enabled = 0;
- }
-}
-
static void intel_flush_iotlb_all(struct iommu_domain *domain)
{
cache_tag_flush_all(to_dmar_domain(domain));
@@ -1556,12 +1528,22 @@ domain_context_mapping(struct dmar_domain *domain, struct device *dev)
struct device_domain_info *info = dev_iommu_priv_get(dev);
struct intel_iommu *iommu = info->iommu;
u8 bus = info->bus, devfn = info->devfn;
+ struct pci_dev *pdev;
+ int ret;
if (!dev_is_pci(dev))
return domain_context_mapping_one(domain, iommu, bus, devfn);
- return pci_for_each_dma_alias(to_pci_dev(dev),
- domain_context_mapping_cb, domain);
+ pdev = to_pci_dev(dev);
+ ret = pci_for_each_dma_alias(pdev, domain_context_mapping_cb, domain);
+ if (ret)
+ return ret;
+
+ if (info->ats_supported && pci_ats_page_aligned(pdev) &&
+ !pci_enable_ats(pdev, VTD_PAGE_SHIFT))
+ info->ats_enabled = 1;
+
+ return 0;
}
/* Return largest possible superpage level for a given mapping */
@@ -1843,8 +1825,6 @@ static int dmar_domain_attach_device(struct dmar_domain *domain,
if (ret)
goto out_block_translation;
- iommu_enable_pci_caps(info);
-
ret = cache_tag_assign_domain(domain, dev, IOMMU_NO_PASID);
if (ret)
goto out_block_translation;
@@ -3191,13 +3171,20 @@ static int domain_context_clear_one_cb(struct pci_dev *pdev, u16 alias, void *op
*/
static void domain_context_clear(struct device_domain_info *info)
{
+ struct pci_dev *pdev;
+
if (!dev_is_pci(info->dev)) {
domain_context_clear_one(info, info->bus, info->devfn);
return;
}
- pci_for_each_dma_alias(to_pci_dev(info->dev),
- &domain_context_clear_one_cb, info);
+ pdev = to_pci_dev(info->dev);
+ pci_for_each_dma_alias(pdev, &domain_context_clear_one_cb, info);
+
+ if (info->ats_enabled) {
+ pci_disable_ats(pdev);
+ info->ats_enabled = 0;
+ }
}
/*
@@ -3214,7 +3201,6 @@ void device_block_translation(struct device *dev)
if (info->domain)
cache_tag_unassign_domain(info->domain, dev, IOMMU_NO_PASID);
- iommu_disable_pci_caps(info);
if (!dev_is_real_dma_subdevice(dev)) {
if (sm_supported(iommu))
intel_pasid_tear_down_entry(iommu, dev,
@@ -3749,6 +3735,16 @@ static struct iommu_device *intel_iommu_probe_device(struct device *dev)
!pci_enable_pasid(pdev, info->pasid_supported & ~1))
info->pasid_enabled = 1;
+ if (sm_supported(iommu)) {
+ if (info->ats_supported && pci_ats_page_aligned(pdev)) {
+ ret = pci_enable_ats(pdev, VTD_PAGE_SHIFT);
+ if (ret)
+ pci_info(pdev, "Failed to enable ATS on device\n");
+ else
+ info->ats_enabled = 1;
+ }
+ }
+
return &iommu->iommu;
free_table:
intel_pasid_free_table(dev);
@@ -3765,6 +3761,11 @@ static void intel_iommu_release_device(struct device *dev)
struct device_domain_info *info = dev_iommu_priv_get(dev);
struct intel_iommu *iommu = info->iommu;
+ if (info->ats_enabled) {
+ pci_disable_ats(to_pci_dev(dev));
+ info->ats_enabled = 0;
+ }
+
if (info->pasid_enabled) {
pci_disable_pasid(to_pci_dev(dev));
info->pasid_enabled = 0;
@@ -4365,13 +4366,10 @@ static int identity_domain_attach_dev(struct iommu_domain *domain, struct device
if (dev_is_real_dma_subdevice(dev))
return 0;
- if (sm_supported(iommu)) {
+ if (sm_supported(iommu))
ret = intel_pasid_setup_pass_through(iommu, dev, IOMMU_NO_PASID);
- if (!ret)
- iommu_enable_pci_caps(info);
- } else {
+ else
ret = device_setup_pass_through(dev);
- }
return ret;
}
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 04/12] iommu/vt-d: Move scalable mode ATS enablement to probe path
2025-02-24 5:16 ` [PATCH v2 04/12] iommu/vt-d: Move scalable mode ATS enablement to probe path Lu Baolu
@ 2025-02-25 7:28 ` Tian, Kevin
2025-02-26 2:21 ` Baolu Lu
2025-03-04 7:55 ` Yi Liu
1 sibling, 1 reply; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 7:28 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> Device ATS is currently enabled when a domain is attached to the device
> and disabled when the domain is detached. This creates a limitation:
> when the IOMMU is operating in scalable mode and IOPF is enabled, the
> device's domain cannot be changed.
I got what this patch does, but not this description. Can you elaborate
about the limitation?
> @@ -1556,12 +1528,22 @@ domain_context_mapping(struct dmar_domain
> *domain, struct device *dev)
> struct device_domain_info *info = dev_iommu_priv_get(dev);
> struct intel_iommu *iommu = info->iommu;
> u8 bus = info->bus, devfn = info->devfn;
> + struct pci_dev *pdev;
> + int ret;
>
> if (!dev_is_pci(dev))
> return domain_context_mapping_one(domain, iommu, bus,
> devfn);
>
> - return pci_for_each_dma_alias(to_pci_dev(dev),
> - domain_context_mapping_cb, domain);
> + pdev = to_pci_dev(dev);
> + ret = pci_for_each_dma_alias(pdev, domain_context_mapping_cb,
> domain);
> + if (ret)
> + return ret;
> +
> + if (info->ats_supported && pci_ats_page_aligned(pdev) &&
> + !pci_enable_ats(pdev, VTD_PAGE_SHIFT))
> + info->ats_enabled = 1;
> +
> + return 0;
> }
It'd be good to add a note here for why legacy mode still requires
dynamic toggle at attach/detach time. It's not obvious w/o knowing
the story about legacy + identity.
btw the same enabling logic occurs in multiple places. Probably
you can still make a helper for that.
Otherwise,
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 04/12] iommu/vt-d: Move scalable mode ATS enablement to probe path
2025-02-25 7:28 ` Tian, Kevin
@ 2025-02-26 2:21 ` Baolu Lu
0 siblings, 0 replies; 38+ messages in thread
From: Baolu Lu @ 2025-02-26 2:21 UTC (permalink / raw)
To: Tian, Kevin, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
On 2/25/25 15:28, Tian, Kevin wrote:
>> From: Lu Baolu <baolu.lu@linux.intel.com>
>> Sent: Monday, February 24, 2025 1:16 PM
>>
>> Device ATS is currently enabled when a domain is attached to the device
>> and disabled when the domain is detached. This creates a limitation:
>> when the IOMMU is operating in scalable mode and IOPF is enabled, the
>> device's domain cannot be changed.
>
> I got what this patch does, but not this description. Can you elaborate
> about the limitation?
Okay, sure.
The previous code enables ATS when a domain is set to a device's RID and
disables it during RID domain switch. So, if a PASID is set with a
domain requiring PRI, ATS should remain enabled until the domain is
removed. During the PASID domain's lifecycle, if the RID's domain
changes, PRI will be disrupted because it depends on ATS, which is
disabled when the blocking domain is set for the device's RID.
>
>> @@ -1556,12 +1528,22 @@ domain_context_mapping(struct dmar_domain
>> *domain, struct device *dev)
>> struct device_domain_info *info = dev_iommu_priv_get(dev);
>> struct intel_iommu *iommu = info->iommu;
>> u8 bus = info->bus, devfn = info->devfn;
>> + struct pci_dev *pdev;
>> + int ret;
>>
>> if (!dev_is_pci(dev))
>> return domain_context_mapping_one(domain, iommu, bus,
>> devfn);
>>
>> - return pci_for_each_dma_alias(to_pci_dev(dev),
>> - domain_context_mapping_cb, domain);
>> + pdev = to_pci_dev(dev);
>> + ret = pci_for_each_dma_alias(pdev, domain_context_mapping_cb,
>> domain);
>> + if (ret)
>> + return ret;
>> +
>> + if (info->ats_supported && pci_ats_page_aligned(pdev) &&
>> + !pci_enable_ats(pdev, VTD_PAGE_SHIFT))
>> + info->ats_enabled = 1;
>> +
>> + return 0;
>> }
>
> It'd be good to add a note here for why legacy mode still requires
> dynamic toggle at attach/detach time. It's not obvious w/o knowing
> the story about legacy + identity.
"legacy + identity" is part of the reason. Additionally, in legacy mode,
the ATS control bit is defined as a translation type and should be set
together with the page table pointer in the context entry. Separating
ATS enablement and the translation page table into different places
would make the code more complex and fragile.
> btw the same enabling logic occurs in multiple places. Probably
> you can still make a helper for that.
Yes, I will make a helper like this,
+static void device_enable_pci_ats(struct pci_dev *pdev)
+{
+ struct device_domain_info *info = dev_iommu_priv_get(&pdev->dev);
+
+ if (!info->ats_supported)
+ return;
+
+ if (!pci_ats_page_aligned(pdev))
+ return;
+
+ if(!pci_enable_ats(pdev, VTD_PAGE_SHIFT))
+ info->ats_enabled = 1;
+}
> Otherwise,
>
> Reviewed-by: Kevin Tian <kevin.tian@intel.com>
Thanks,
baolu
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 04/12] iommu/vt-d: Move scalable mode ATS enablement to probe path
2025-02-24 5:16 ` [PATCH v2 04/12] iommu/vt-d: Move scalable mode ATS enablement to probe path Lu Baolu
2025-02-25 7:28 ` Tian, Kevin
@ 2025-03-04 7:55 ` Yi Liu
1 sibling, 0 replies; 38+ messages in thread
From: Yi Liu @ 2025-03-04 7:55 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe, Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel
On 2025/2/24 13:16, Lu Baolu wrote:
> Device ATS is currently enabled when a domain is attached to the device
> and disabled when the domain is detached. This creates a limitation:
> when the IOMMU is operating in scalable mode and IOPF is enabled, the
> device's domain cannot be changed.
>
> Remove this limitation by moving ATS enablement to the device probe path.
>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
> ---
> drivers/iommu/intel/iommu.c | 78 ++++++++++++++++++-------------------
> 1 file changed, 38 insertions(+), 40 deletions(-)
I'm ok with this patch. Just a heads up in case of anyone that is not aware
of a discussion in another threa which intends to enable ATS in domain
attach.
[1]
https://lore.kernel.org/linux-iommu/2c9ef073-fee5-43c6-8932-a8cae677970e@intel.com/
> diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
> index 16dd8f0de76d..f52602bde742 100644
> --- a/drivers/iommu/intel/iommu.c
> +++ b/drivers/iommu/intel/iommu.c
> @@ -1172,34 +1172,6 @@ static bool dev_needs_extra_dtlb_flush(struct pci_dev *pdev)
> return true;
> }
>
> -static void iommu_enable_pci_caps(struct device_domain_info *info)
> -{
> - struct pci_dev *pdev;
> -
> - if (!dev_is_pci(info->dev))
> - return;
> -
> - pdev = to_pci_dev(info->dev);
> - if (info->ats_supported && pci_ats_page_aligned(pdev) &&
> - !pci_enable_ats(pdev, VTD_PAGE_SHIFT))
> - info->ats_enabled = 1;
> -}
> -
> -static void iommu_disable_pci_caps(struct device_domain_info *info)
> -{
> - struct pci_dev *pdev;
> -
> - if (!dev_is_pci(info->dev))
> - return;
> -
> - pdev = to_pci_dev(info->dev);
> -
> - if (info->ats_enabled) {
> - pci_disable_ats(pdev);
> - info->ats_enabled = 0;
> - }
> -}
> -
> static void intel_flush_iotlb_all(struct iommu_domain *domain)
> {
> cache_tag_flush_all(to_dmar_domain(domain));
> @@ -1556,12 +1528,22 @@ domain_context_mapping(struct dmar_domain *domain, struct device *dev)
> struct device_domain_info *info = dev_iommu_priv_get(dev);
> struct intel_iommu *iommu = info->iommu;
> u8 bus = info->bus, devfn = info->devfn;
> + struct pci_dev *pdev;
> + int ret;
>
> if (!dev_is_pci(dev))
> return domain_context_mapping_one(domain, iommu, bus, devfn);
>
> - return pci_for_each_dma_alias(to_pci_dev(dev),
> - domain_context_mapping_cb, domain);
> + pdev = to_pci_dev(dev);
> + ret = pci_for_each_dma_alias(pdev, domain_context_mapping_cb, domain);
> + if (ret)
> + return ret;
> +
> + if (info->ats_supported && pci_ats_page_aligned(pdev) &&
> + !pci_enable_ats(pdev, VTD_PAGE_SHIFT))
> + info->ats_enabled = 1;
> +
> + return 0;
> }
>
> /* Return largest possible superpage level for a given mapping */
> @@ -1843,8 +1825,6 @@ static int dmar_domain_attach_device(struct dmar_domain *domain,
> if (ret)
> goto out_block_translation;
>
> - iommu_enable_pci_caps(info);
> -
> ret = cache_tag_assign_domain(domain, dev, IOMMU_NO_PASID);
> if (ret)
> goto out_block_translation;
> @@ -3191,13 +3171,20 @@ static int domain_context_clear_one_cb(struct pci_dev *pdev, u16 alias, void *op
> */
> static void domain_context_clear(struct device_domain_info *info)
> {
> + struct pci_dev *pdev;
> +
> if (!dev_is_pci(info->dev)) {
> domain_context_clear_one(info, info->bus, info->devfn);
> return;
> }
>
> - pci_for_each_dma_alias(to_pci_dev(info->dev),
> - &domain_context_clear_one_cb, info);
> + pdev = to_pci_dev(info->dev);
> + pci_for_each_dma_alias(pdev, &domain_context_clear_one_cb, info);
> +
> + if (info->ats_enabled) {
> + pci_disable_ats(pdev);
> + info->ats_enabled = 0;
> + }
> }
>
> /*
> @@ -3214,7 +3201,6 @@ void device_block_translation(struct device *dev)
> if (info->domain)
> cache_tag_unassign_domain(info->domain, dev, IOMMU_NO_PASID);
>
> - iommu_disable_pci_caps(info);
> if (!dev_is_real_dma_subdevice(dev)) {
> if (sm_supported(iommu))
> intel_pasid_tear_down_entry(iommu, dev,
> @@ -3749,6 +3735,16 @@ static struct iommu_device *intel_iommu_probe_device(struct device *dev)
> !pci_enable_pasid(pdev, info->pasid_supported & ~1))
> info->pasid_enabled = 1;
>
> + if (sm_supported(iommu)) {
> + if (info->ats_supported && pci_ats_page_aligned(pdev)) {
> + ret = pci_enable_ats(pdev, VTD_PAGE_SHIFT);
> + if (ret)
> + pci_info(pdev, "Failed to enable ATS on device\n");
> + else
> + info->ats_enabled = 1;
> + }
> + }
> +
> return &iommu->iommu;
> free_table:
> intel_pasid_free_table(dev);
> @@ -3765,6 +3761,11 @@ static void intel_iommu_release_device(struct device *dev)
> struct device_domain_info *info = dev_iommu_priv_get(dev);
> struct intel_iommu *iommu = info->iommu;
>
> + if (info->ats_enabled) {
> + pci_disable_ats(to_pci_dev(dev));
> + info->ats_enabled = 0;
> + }
> +
> if (info->pasid_enabled) {
> pci_disable_pasid(to_pci_dev(dev));
> info->pasid_enabled = 0;
> @@ -4365,13 +4366,10 @@ static int identity_domain_attach_dev(struct iommu_domain *domain, struct device
> if (dev_is_real_dma_subdevice(dev))
> return 0;
>
> - if (sm_supported(iommu)) {
> + if (sm_supported(iommu))
> ret = intel_pasid_setup_pass_through(iommu, dev, IOMMU_NO_PASID);
> - if (!ret)
> - iommu_enable_pci_caps(info);
> - } else {
> + else
> ret = device_setup_pass_through(dev);
> - }
>
> return ret;
> }
--
Regards,
Yi Liu
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 05/12] iommu/vt-d: Move PRI enablement in probe path
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (3 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 04/12] iommu/vt-d: Move scalable mode ATS enablement to probe path Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 7:37 ` Tian, Kevin
2025-02-24 5:16 ` [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present() Lu Baolu
` (7 subsequent siblings)
12 siblings, 1 reply; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Lu Baolu
Update PRI enablement to use the new method, similar to the amd iommu
driver. Enable PRI in the device probe path and disable it when the device
is released. PRI is enabled throughout the device's iommu lifecycle. The
infrastructure for the iommu subsystem to handle iopf requests is created
during iopf enablement and released during iopf disablement. All invalid
page requests from the device are automatically handled by the iommu
subsystem if iopf is not enabled. Add iopf_refcount to track the iopf
enablement.
Convert the return type of intel_iommu_disable_iopf() to void, as there
is no way to handle a failure when disabling this feature. Make
intel_iommu_enable/disable_iopf() helpers global, as they will be used
beyond the current file in the subsequent patch.
The iopf_refcount is not protected by any lock. This is acceptable, as
there is no concurrent access to it in the current code. The following
patch will address this by moving it to the domain attach/detach paths,
which are protected by the iommu group mutex.
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
---
drivers/iommu/intel/iommu.c | 145 +++++++++++-------------------------
drivers/iommu/intel/iommu.h | 4 +
drivers/iommu/intel/pasid.c | 2 +
drivers/iommu/intel/prq.c | 2 +-
4 files changed, 51 insertions(+), 102 deletions(-)
diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
index f52602bde742..91d49e2cea34 100644
--- a/drivers/iommu/intel/iommu.c
+++ b/drivers/iommu/intel/iommu.c
@@ -3743,6 +3743,16 @@ static struct iommu_device *intel_iommu_probe_device(struct device *dev)
else
info->ats_enabled = 1;
}
+
+ if (info->ats_enabled && info->pri_supported) {
+ /* PASID is required in PRG Response Message. */
+ if (info->pasid_enabled || !pci_prg_resp_pasid_required(pdev)) {
+ if (!pci_reset_pri(pdev) && !pci_enable_pri(pdev, PRQ_DEPTH))
+ info->pri_enabled = 1;
+ else
+ pci_info(pdev, "Failed to enable PRI on device\n");
+ }
+ }
}
return &iommu->iommu;
@@ -3761,6 +3771,13 @@ static void intel_iommu_release_device(struct device *dev)
struct device_domain_info *info = dev_iommu_priv_get(dev);
struct intel_iommu *iommu = info->iommu;
+ WARN_ON(info->iopf_refcount);
+
+ if (info->pri_enabled) {
+ pci_disable_pri(to_pci_dev(dev));
+ info->pri_enabled = 0;
+ }
+
if (info->ats_enabled) {
pci_disable_ats(to_pci_dev(dev));
info->ats_enabled = 0;
@@ -3852,118 +3869,43 @@ static struct iommu_group *intel_iommu_device_group(struct device *dev)
return generic_device_group(dev);
}
-static int context_flip_pri(struct device_domain_info *info, bool enable)
+int intel_iommu_enable_iopf(struct device *dev)
{
- struct intel_iommu *iommu = info->iommu;
- u8 bus = info->bus, devfn = info->devfn;
- struct context_entry *context;
- u16 did;
-
- spin_lock(&iommu->lock);
- if (context_copied(iommu, bus, devfn)) {
- spin_unlock(&iommu->lock);
- return -EINVAL;
- }
-
- context = iommu_context_addr(iommu, bus, devfn, false);
- if (!context || !context_present(context)) {
- spin_unlock(&iommu->lock);
- return -ENODEV;
- }
- did = context_domain_id(context);
-
- if (enable)
- context_set_sm_pre(context);
- else
- context_clear_sm_pre(context);
-
- if (!ecap_coherent(iommu->ecap))
- clflush_cache_range(context, sizeof(*context));
- intel_context_flush_present(info, context, did, true);
- spin_unlock(&iommu->lock);
-
- return 0;
-}
-
-static int intel_iommu_enable_iopf(struct device *dev)
-{
- struct pci_dev *pdev = dev_is_pci(dev) ? to_pci_dev(dev) : NULL;
struct device_domain_info *info = dev_iommu_priv_get(dev);
- struct intel_iommu *iommu;
+ struct intel_iommu *iommu = info->iommu;
int ret;
- if (!pdev || !info || !info->ats_enabled || !info->pri_supported)
- return -ENODEV;
-
- if (info->pri_enabled)
- return -EBUSY;
-
- iommu = info->iommu;
- if (!iommu)
- return -EINVAL;
-
- /* PASID is required in PRG Response Message. */
- if (info->pasid_enabled && !pci_prg_resp_pasid_required(pdev))
- return -EINVAL;
-
- ret = pci_reset_pri(pdev);
- if (ret)
- return ret;
-
- ret = iopf_queue_add_device(iommu->iopf_queue, dev);
- if (ret)
- return ret;
-
- ret = context_flip_pri(info, true);
- if (ret)
- goto err_remove_device;
-
- ret = pci_enable_pri(pdev, PRQ_DEPTH);
- if (ret)
- goto err_clear_pri;
-
- info->pri_enabled = 1;
-
- return 0;
-err_clear_pri:
- context_flip_pri(info, false);
-err_remove_device:
- iopf_queue_remove_device(iommu->iopf_queue, dev);
-
- return ret;
-}
-
-static int intel_iommu_disable_iopf(struct device *dev)
-{
- struct device_domain_info *info = dev_iommu_priv_get(dev);
- struct intel_iommu *iommu = info->iommu;
-
if (!info->pri_enabled)
- return -EINVAL;
+ return -ENODEV;
- /* Disable new PRI reception: */
- context_flip_pri(info, false);
+ if (info->iopf_refcount) {
+ info->iopf_refcount++;
+ return 0;
+ }
- /*
- * Remove device from fault queue and acknowledge all outstanding
- * PRQs to the device:
- */
- iopf_queue_remove_device(iommu->iopf_queue, dev);
+ ret = iopf_queue_add_device(iommu->iopf_queue, dev);
+ if (ret)
+ return ret;
- /*
- * PCIe spec states that by clearing PRI enable bit, the Page
- * Request Interface will not issue new page requests, but has
- * outstanding page requests that have been transmitted or are
- * queued for transmission. This is supposed to be called after
- * the device driver has stopped DMA, all PASIDs have been
- * unbound and the outstanding PRQs have been drained.
- */
- pci_disable_pri(to_pci_dev(dev));
- info->pri_enabled = 0;
+ info->iopf_refcount = 1;
return 0;
}
+void intel_iommu_disable_iopf(struct device *dev)
+{
+ struct device_domain_info *info = dev_iommu_priv_get(dev);
+ struct intel_iommu *iommu = info->iommu;
+
+ if (WARN_ON(!info->pri_enabled))
+ return;
+
+ if (--info->iopf_refcount)
+ return;
+
+ iopf_queue_remove_device(iommu->iopf_queue, dev);
+}
+
static int
intel_iommu_dev_enable_feat(struct device *dev, enum iommu_dev_features feat)
{
@@ -3981,7 +3923,8 @@ intel_iommu_dev_disable_feat(struct device *dev, enum iommu_dev_features feat)
{
switch (feat) {
case IOMMU_DEV_FEAT_IOPF:
- return intel_iommu_disable_iopf(dev);
+ intel_iommu_disable_iopf(dev);
+ return 0;
default:
return -ENODEV;
diff --git a/drivers/iommu/intel/iommu.h b/drivers/iommu/intel/iommu.h
index 6ea7bbe26b19..f7d78cf0778c 100644
--- a/drivers/iommu/intel/iommu.h
+++ b/drivers/iommu/intel/iommu.h
@@ -774,6 +774,7 @@ struct device_domain_info {
u8 ats_enabled:1;
u8 dtlb_extra_inval:1; /* Quirk for devices need extra flush */
u8 ats_qdep;
+ unsigned int iopf_refcount;
struct device *dev; /* it's NULL for PCIe-to-PCI bridge */
struct intel_iommu *iommu; /* IOMMU used by this device */
struct dmar_domain *domain; /* pointer to domain */
@@ -1314,6 +1315,9 @@ void intel_iommu_page_response(struct device *dev, struct iopf_fault *evt,
struct iommu_page_response *msg);
void intel_iommu_drain_pasid_prq(struct device *dev, u32 pasid);
+int intel_iommu_enable_iopf(struct device *dev);
+void intel_iommu_disable_iopf(struct device *dev);
+
#ifdef CONFIG_INTEL_IOMMU_SVM
void intel_svm_check(struct intel_iommu *iommu);
struct iommu_domain *intel_svm_domain_alloc(struct device *dev,
diff --git a/drivers/iommu/intel/pasid.c b/drivers/iommu/intel/pasid.c
index fb59a7d35958..c2742e256552 100644
--- a/drivers/iommu/intel/pasid.c
+++ b/drivers/iommu/intel/pasid.c
@@ -992,6 +992,8 @@ static int context_entry_set_pasid_table(struct context_entry *context,
context_set_sm_dte(context);
if (info->pasid_supported)
context_set_pasid(context);
+ if (info->pri_supported)
+ context_set_sm_pre(context);
context_set_fault_enable(context);
context_set_present(context);
diff --git a/drivers/iommu/intel/prq.c b/drivers/iommu/intel/prq.c
index 064194399b38..5b6a64d96850 100644
--- a/drivers/iommu/intel/prq.c
+++ b/drivers/iommu/intel/prq.c
@@ -67,7 +67,7 @@ void intel_iommu_drain_pasid_prq(struct device *dev, u32 pasid)
u16 sid, did;
info = dev_iommu_priv_get(dev);
- if (!info->pri_enabled)
+ if (!info->iopf_refcount)
return;
iommu = info->iommu;
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 05/12] iommu/vt-d: Move PRI enablement in probe path
2025-02-24 5:16 ` [PATCH v2 05/12] iommu/vt-d: Move PRI enablement in " Lu Baolu
@ 2025-02-25 7:37 ` Tian, Kevin
2025-02-26 3:53 ` Baolu Lu
0 siblings, 1 reply; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 7:37 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> @@ -3743,6 +3743,16 @@ static struct iommu_device
> *intel_iommu_probe_device(struct device *dev)
> else
> info->ats_enabled = 1;
> }
> +
> + if (info->ats_enabled && info->pri_supported) {
> + /* PASID is required in PRG Response Message. */
> + if (info->pasid_enabled
> || !pci_prg_resp_pasid_required(pdev)) {
this should be:
if (!info->pasid_enabled || pci_prg_resp_pasid_required(pdev))
> @@ -3761,6 +3771,13 @@ static void intel_iommu_release_device(struct
> device *dev)
> struct device_domain_info *info = dev_iommu_priv_get(dev);
> struct intel_iommu *iommu = info->iommu;
>
> + WARN_ON(info->iopf_refcount);
call iopf_queue_remove_device() in this case?
> +void intel_iommu_disable_iopf(struct device *dev)
> +{
> + struct device_domain_info *info = dev_iommu_priv_get(dev);
> + struct intel_iommu *iommu = info->iommu;
> +
> + if (WARN_ON(!info->pri_enabled))
> + return;
also warn on !info->iopf_refcount
^ permalink raw reply [flat|nested] 38+ messages in thread* Re: [PATCH v2 05/12] iommu/vt-d: Move PRI enablement in probe path
2025-02-25 7:37 ` Tian, Kevin
@ 2025-02-26 3:53 ` Baolu Lu
0 siblings, 0 replies; 38+ messages in thread
From: Baolu Lu @ 2025-02-26 3:53 UTC (permalink / raw)
To: Tian, Kevin, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
On 2/25/25 15:37, Tian, Kevin wrote:
>> From: Lu Baolu <baolu.lu@linux.intel.com>
>> Sent: Monday, February 24, 2025 1:16 PM
>>
>> @@ -3743,6 +3743,16 @@ static struct iommu_device
>> *intel_iommu_probe_device(struct device *dev)
>> else
>> info->ats_enabled = 1;
>> }
>> +
>> + if (info->ats_enabled && info->pri_supported) {
>> + /* PASID is required in PRG Response Message. */
>> + if (info->pasid_enabled
>> || !pci_prg_resp_pasid_required(pdev)) {
>
> this should be:
>
> if (!info->pasid_enabled || pci_prg_resp_pasid_required(pdev))
Yes, fixed.
>
>
>> @@ -3761,6 +3771,13 @@ static void intel_iommu_release_device(struct
>> device *dev)
>> struct device_domain_info *info = dev_iommu_priv_get(dev);
>> struct intel_iommu *iommu = info->iommu;
>>
>> + WARN_ON(info->iopf_refcount);
>
> call iopf_queue_remove_device() in this case?
Yes. I will make it like this,
if (WARN_ON(info->iopf_refcount))
iopf_queue_remove_device(iommu->iopf_queue, dev);
>> +void intel_iommu_disable_iopf(struct device *dev)
>> +{
>> + struct device_domain_info *info = dev_iommu_priv_get(dev);
>> + struct intel_iommu *iommu = info->iommu;
>> +
>> + if (WARN_ON(!info->pri_enabled))
>> + return;
>
> also warn on !info->iopf_refcount
Yes, added.
Thanks,
baolu
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present()
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (4 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 05/12] iommu/vt-d: Move PRI enablement in " Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 7:43 ` Tian, Kevin
2025-03-04 8:43 ` Yi Liu
2025-02-24 5:16 ` [PATCH v2 07/12] iommu/vt-d: Put iopf enablement in domain attach path Lu Baolu
` (6 subsequent siblings)
12 siblings, 2 replies; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Lu Baolu
The intel_context_flush_present() is called in places where either the
scalable mode is disabled, or scalable mode is enabled but all PASID
entries are known to be non-present. In these cases, the flush_domains
path within intel_context_flush_present() will never execute. This dead
code is therefore removed.
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
---
drivers/iommu/intel/iommu.c | 2 +-
drivers/iommu/intel/iommu.h | 3 +--
drivers/iommu/intel/pasid.c | 39 ++++++-------------------------------
3 files changed, 8 insertions(+), 36 deletions(-)
diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
index 91d49e2cea34..1d564240c977 100644
--- a/drivers/iommu/intel/iommu.c
+++ b/drivers/iommu/intel/iommu.c
@@ -1730,7 +1730,7 @@ static void domain_context_clear_one(struct device_domain_info *info, u8 bus, u8
context_clear_entry(context);
__iommu_flush_cache(iommu, context, sizeof(*context));
spin_unlock(&iommu->lock);
- intel_context_flush_present(info, context, did, true);
+ intel_context_flush_present(info, context, did);
}
int __domain_setup_first_level(struct intel_iommu *iommu,
diff --git a/drivers/iommu/intel/iommu.h b/drivers/iommu/intel/iommu.h
index f7d78cf0778c..754f6d7ade26 100644
--- a/drivers/iommu/intel/iommu.h
+++ b/drivers/iommu/intel/iommu.h
@@ -1306,8 +1306,7 @@ void cache_tag_flush_range_np(struct dmar_domain *domain, unsigned long start,
unsigned long end);
void intel_context_flush_present(struct device_domain_info *info,
- struct context_entry *context,
- u16 did, bool affect_domains);
+ struct context_entry *context, u16 did);
int intel_iommu_enable_prq(struct intel_iommu *iommu);
int intel_iommu_finish_prq(struct intel_iommu *iommu);
diff --git a/drivers/iommu/intel/pasid.c b/drivers/iommu/intel/pasid.c
index c2742e256552..a2c6be624dbf 100644
--- a/drivers/iommu/intel/pasid.c
+++ b/drivers/iommu/intel/pasid.c
@@ -932,7 +932,7 @@ static void device_pasid_table_teardown(struct device *dev, u8 bus, u8 devfn)
context_clear_entry(context);
__iommu_flush_cache(iommu, context, sizeof(*context));
spin_unlock(&iommu->lock);
- intel_context_flush_present(info, context, did, false);
+ intel_context_flush_present(info, context, did);
}
static int pci_pasid_table_teardown(struct pci_dev *pdev, u16 alias, void *data)
@@ -1119,17 +1119,15 @@ static void __context_flush_dev_iotlb(struct device_domain_info *info)
/*
* Cache invalidations after change in a context table entry that was present
- * according to the Spec 6.5.3.3 (Guidance to Software for Invalidations). If
- * IOMMU is in scalable mode and all PASID table entries of the device were
- * non-present, set flush_domains to false. Otherwise, true.
+ * according to the Spec 6.5.3.3 (Guidance to Software for Invalidations).
+ * This helper can only be used when IOMMU is working in the legacy mode or
+ * IOMMU is in scalable mode but all PASID table entries of the device are
+ * non-present.
*/
void intel_context_flush_present(struct device_domain_info *info,
- struct context_entry *context,
- u16 did, bool flush_domains)
+ struct context_entry *context, u16 did)
{
struct intel_iommu *iommu = info->iommu;
- struct pasid_entry *pte;
- int i;
/*
* Device-selective context-cache invalidation. The Domain-ID field
@@ -1152,30 +1150,5 @@ void intel_context_flush_present(struct device_domain_info *info,
return;
}
- /*
- * For scalable mode:
- * - Domain-selective PASID-cache invalidation to affected domains
- * - Domain-selective IOTLB invalidation to affected domains
- * - Global Device-TLB invalidation to affected functions
- */
- if (flush_domains) {
- /*
- * If the IOMMU is running in scalable mode and there might
- * be potential PASID translations, the caller should hold
- * the lock to ensure that context changes and cache flushes
- * are atomic.
- */
- assert_spin_locked(&iommu->lock);
- for (i = 0; i < info->pasid_table->max_pasid; i++) {
- pte = intel_pasid_get_entry(info->dev, i);
- if (!pte || !pasid_pte_is_present(pte))
- continue;
-
- did = pasid_get_domain_id(pte);
- qi_flush_pasid_cache(iommu, did, QI_PC_ALL_PASIDS, 0);
- iommu->flush.flush_iotlb(iommu, did, 0, 0, DMA_TLB_DSI_FLUSH);
- }
- }
-
__context_flush_dev_iotlb(info);
}
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present()
2025-02-24 5:16 ` [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present() Lu Baolu
@ 2025-02-25 7:43 ` Tian, Kevin
2025-02-26 3:57 ` Baolu Lu
2025-03-04 8:43 ` Yi Liu
1 sibling, 1 reply; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 7:43 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> The intel_context_flush_present() is called in places where either the
> scalable mode is disabled, or scalable mode is enabled but all PASID
> entries are known to be non-present. In these cases, the flush_domains
> path within intel_context_flush_present() will never execute. This dead
> code is therefore removed.
>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
usually a suffix "_present()" indicates that the helper can be called
on an object which is currently in-use, which is obviously not the
case here.
To avoid confusion probably just call it intel_context_flush() or
intel_context_flush_no_user() is clearer.
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present()
2025-02-25 7:43 ` Tian, Kevin
@ 2025-02-26 3:57 ` Baolu Lu
2025-02-26 5:17 ` Tian, Kevin
0 siblings, 1 reply; 38+ messages in thread
From: Baolu Lu @ 2025-02-26 3:57 UTC (permalink / raw)
To: Tian, Kevin, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
On 2/25/25 15:43, Tian, Kevin wrote:
>> From: Lu Baolu<baolu.lu@linux.intel.com>
>> Sent: Monday, February 24, 2025 1:16 PM
>>
>> The intel_context_flush_present() is called in places where either the
>> scalable mode is disabled, or scalable mode is enabled but all PASID
>> entries are known to be non-present. In these cases, the flush_domains
>> path within intel_context_flush_present() will never execute. This dead
>> code is therefore removed.
>>
>> Signed-off-by: Lu Baolu<baolu.lu@linux.intel.com>
> usually a suffix "_present()" indicates that the helper can be called
> on an object which is currently in-use, which is obviously not the
> case here.
>
> To avoid confusion probably just call it intel_context_flush() or
> intel_context_flush_no_user() is clearer.
How about intel_context_flush_no_pasid()?
^ permalink raw reply [flat|nested] 38+ messages in thread
* RE: [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present()
2025-02-26 3:57 ` Baolu Lu
@ 2025-02-26 5:17 ` Tian, Kevin
0 siblings, 0 replies; 38+ messages in thread
From: Tian, Kevin @ 2025-02-26 5:17 UTC (permalink / raw)
To: Baolu Lu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
> From: Baolu Lu <baolu.lu@linux.intel.com>
> Sent: Wednesday, February 26, 2025 11:58 AM
>
> On 2/25/25 15:43, Tian, Kevin wrote:
> >> From: Lu Baolu<baolu.lu@linux.intel.com>
> >> Sent: Monday, February 24, 2025 1:16 PM
> >>
> >> The intel_context_flush_present() is called in places where either the
> >> scalable mode is disabled, or scalable mode is enabled but all PASID
> >> entries are known to be non-present. In these cases, the flush_domains
> >> path within intel_context_flush_present() will never execute. This dead
> >> code is therefore removed.
> >>
> >> Signed-off-by: Lu Baolu<baolu.lu@linux.intel.com>
> > usually a suffix "_present()" indicates that the helper can be called
> > on an object which is currently in-use, which is obviously not the
> > case here.
> >
> > To avoid confusion probably just call it intel_context_flush() or
> > intel_context_flush_no_user() is clearer.
>
> How about intel_context_flush_no_pasid()?
yes
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present()
2025-02-24 5:16 ` [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present() Lu Baolu
2025-02-25 7:43 ` Tian, Kevin
@ 2025-03-04 8:43 ` Yi Liu
2025-03-05 2:21 ` Baolu Lu
1 sibling, 1 reply; 38+ messages in thread
From: Yi Liu @ 2025-03-04 8:43 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe, Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel
Hi Baolu,
On 2025/2/24 13:16, Lu Baolu wrote:
> The intel_context_flush_present() is called in places where either the
> scalable mode is disabled, or scalable mode is enabled but all PASID
> entries are known to be non-present. In these cases, the flush_domains
> path within intel_context_flush_present() will never execute. This dead
> code is therefore removed.
The reason for this path is the remaining caller of
intel_context_flush_present() is only the domain_context_clear_one() which
is called in legacy mode path. Is it?
If so, it seems unnecessary to keep __context_flush_dev_iotlb(info); in the
end of the new intel_context_flush_present(). Also, since this helper is
more for legacy mode, might be good to move it out of pasid.c.:)
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
> ---
> drivers/iommu/intel/iommu.c | 2 +-
> drivers/iommu/intel/iommu.h | 3 +--
> drivers/iommu/intel/pasid.c | 39 ++++++-------------------------------
> 3 files changed, 8 insertions(+), 36 deletions(-)
>
> diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
> index 91d49e2cea34..1d564240c977 100644
> --- a/drivers/iommu/intel/iommu.c
> +++ b/drivers/iommu/intel/iommu.c
> @@ -1730,7 +1730,7 @@ static void domain_context_clear_one(struct device_domain_info *info, u8 bus, u8
> context_clear_entry(context);
> __iommu_flush_cache(iommu, context, sizeof(*context));
> spin_unlock(&iommu->lock);
> - intel_context_flush_present(info, context, did, true);
> + intel_context_flush_present(info, context, did);
> }
>
> int __domain_setup_first_level(struct intel_iommu *iommu,
> diff --git a/drivers/iommu/intel/iommu.h b/drivers/iommu/intel/iommu.h
> index f7d78cf0778c..754f6d7ade26 100644
> --- a/drivers/iommu/intel/iommu.h
> +++ b/drivers/iommu/intel/iommu.h
> @@ -1306,8 +1306,7 @@ void cache_tag_flush_range_np(struct dmar_domain *domain, unsigned long start,
> unsigned long end);
>
> void intel_context_flush_present(struct device_domain_info *info,
> - struct context_entry *context,
> - u16 did, bool affect_domains);
> + struct context_entry *context, u16 did);
>
> int intel_iommu_enable_prq(struct intel_iommu *iommu);
> int intel_iommu_finish_prq(struct intel_iommu *iommu);
> diff --git a/drivers/iommu/intel/pasid.c b/drivers/iommu/intel/pasid.c
> index c2742e256552..a2c6be624dbf 100644
> --- a/drivers/iommu/intel/pasid.c
> +++ b/drivers/iommu/intel/pasid.c
> @@ -932,7 +932,7 @@ static void device_pasid_table_teardown(struct device *dev, u8 bus, u8 devfn)
> context_clear_entry(context);
> __iommu_flush_cache(iommu, context, sizeof(*context));
> spin_unlock(&iommu->lock);
> - intel_context_flush_present(info, context, did, false);
> + intel_context_flush_present(info, context, did);
> }
>
> static int pci_pasid_table_teardown(struct pci_dev *pdev, u16 alias, void *data)
> @@ -1119,17 +1119,15 @@ static void __context_flush_dev_iotlb(struct device_domain_info *info)
>
> /*
> * Cache invalidations after change in a context table entry that was present
> - * according to the Spec 6.5.3.3 (Guidance to Software for Invalidations). If
> - * IOMMU is in scalable mode and all PASID table entries of the device were
> - * non-present, set flush_domains to false. Otherwise, true.
> + * according to the Spec 6.5.3.3 (Guidance to Software for Invalidations).
> + * This helper can only be used when IOMMU is working in the legacy mode or
> + * IOMMU is in scalable mode but all PASID table entries of the device are
> + * non-present.
> */
> void intel_context_flush_present(struct device_domain_info *info,
> - struct context_entry *context,
> - u16 did, bool flush_domains)
> + struct context_entry *context, u16 did)
> {
> struct intel_iommu *iommu = info->iommu;
> - struct pasid_entry *pte;
> - int i;
>
> /*
> * Device-selective context-cache invalidation. The Domain-ID field
> @@ -1152,30 +1150,5 @@ void intel_context_flush_present(struct device_domain_info *info,
> return;
> }
>
> - /*
> - * For scalable mode:
> - * - Domain-selective PASID-cache invalidation to affected domains
> - * - Domain-selective IOTLB invalidation to affected domains
> - * - Global Device-TLB invalidation to affected functions
> - */
> - if (flush_domains) {
> - /*
> - * If the IOMMU is running in scalable mode and there might
> - * be potential PASID translations, the caller should hold
> - * the lock to ensure that context changes and cache flushes
> - * are atomic.
> - */
> - assert_spin_locked(&iommu->lock);
> - for (i = 0; i < info->pasid_table->max_pasid; i++) {
> - pte = intel_pasid_get_entry(info->dev, i);
> - if (!pte || !pasid_pte_is_present(pte))
> - continue;
> -
> - did = pasid_get_domain_id(pte);
> - qi_flush_pasid_cache(iommu, did, QI_PC_ALL_PASIDS, 0);
> - iommu->flush.flush_iotlb(iommu, did, 0, 0, DMA_TLB_DSI_FLUSH);
> - }
> - }
> -
> __context_flush_dev_iotlb(info);
> }
--
Regards,
Yi Liu
^ permalink raw reply [flat|nested] 38+ messages in thread* Re: [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present()
2025-03-04 8:43 ` Yi Liu
@ 2025-03-05 2:21 ` Baolu Lu
2025-03-05 3:34 ` Yi Liu
0 siblings, 1 reply; 38+ messages in thread
From: Baolu Lu @ 2025-03-05 2:21 UTC (permalink / raw)
To: Yi Liu, Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel
On 3/4/25 16:43, Yi Liu wrote:
> On 2025/2/24 13:16, Lu Baolu wrote:
>> The intel_context_flush_present() is called in places where either the
>> scalable mode is disabled, or scalable mode is enabled but all PASID
>> entries are known to be non-present. In these cases, the flush_domains
>> path within intel_context_flush_present() will never execute. This dead
>> code is therefore removed.
>
> The reason for this path is the remaining caller of
> intel_context_flush_present() is only the domain_context_clear_one() which
> is called in legacy mode path. Is it?
> If so, it seems unnecessary to keep __context_flush_dev_iotlb(info); in the
> end of the new intel_context_flush_present(). Also, since this helper is
> more for legacy mode, might be good to move it out of pasid.c.:)
This helper is for invalidating various caches when a context entry is
present and certain fields are changed. It is used in both legacy and
scalable modes.
In the past, this helper would work even if some PASIDs were still in
use. After the changes introduced in this series, this PASID-in-use case
is removed. So remove the dead code.
Thanks,
baolu
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present()
2025-03-05 2:21 ` Baolu Lu
@ 2025-03-05 3:34 ` Yi Liu
2025-03-05 5:36 ` Baolu Lu
0 siblings, 1 reply; 38+ messages in thread
From: Yi Liu @ 2025-03-05 3:34 UTC (permalink / raw)
To: Baolu Lu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe, Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel
On 2025/3/5 10:21, Baolu Lu wrote:
> On 3/4/25 16:43, Yi Liu wrote:
>> On 2025/2/24 13:16, Lu Baolu wrote:
>>> The intel_context_flush_present() is called in places where either the
>>> scalable mode is disabled, or scalable mode is enabled but all PASID
>>> entries are known to be non-present. In these cases, the flush_domains
>>> path within intel_context_flush_present() will never execute. This dead
>>> code is therefore removed.
>>
>> The reason for this path is the remaining caller of
>> intel_context_flush_present() is only the domain_context_clear_one() which
>> is called in legacy mode path. Is it?
>> If so, it seems unnecessary to keep __context_flush_dev_iotlb(info); in the
>> end of the new intel_context_flush_present(). Also, since this helper is
>> more for legacy mode, might be good to move it out of pasid.c.:)
>
> This helper is for invalidating various caches when a context entry is
> present and certain fields are changed. It is used in both legacy and
> scalable modes.
hmmm. the kdoc says all the pasid entries are non-present, is it necessary
to flush dev_tlb in such scenario? I suppose no present pasid entry means
no pagetable as well.
> In the past, this helper would work even if some PASIDs were still in
> use. After the changes introduced in this series, this PASID-in-use case
> is removed. So remove the dead code.
yeah, I got this part. As I mentioned, the only caller now is the
domain_context_clear_one() which is used in the legacy path. That's why I
feel like it is now more for legacy mode path now after this series.
--
Regards,
Yi Liu
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present()
2025-03-05 3:34 ` Yi Liu
@ 2025-03-05 5:36 ` Baolu Lu
0 siblings, 0 replies; 38+ messages in thread
From: Baolu Lu @ 2025-03-05 5:36 UTC (permalink / raw)
To: Yi Liu, Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel
On 3/5/25 11:34, Yi Liu wrote:
> On 2025/3/5 10:21, Baolu Lu wrote:
>> On 3/4/25 16:43, Yi Liu wrote:
>>> On 2025/2/24 13:16, Lu Baolu wrote:
>>>> The intel_context_flush_present() is called in places where either the
>>>> scalable mode is disabled, or scalable mode is enabled but all PASID
>>>> entries are known to be non-present. In these cases, the flush_domains
>>>> path within intel_context_flush_present() will never execute. This dead
>>>> code is therefore removed.
>>>
>>> The reason for this path is the remaining caller of
>>> intel_context_flush_present() is only the domain_context_clear_one()
>>> which
>>> is called in legacy mode path. Is it?
>>> If so, it seems unnecessary to keep __context_flush_dev_iotlb(info);
>>> in the
>>> end of the new intel_context_flush_present(). Also, since this helper
>>> is more for legacy mode, might be good to move it out of pasid.c.:)
>>
>> This helper is for invalidating various caches when a context entry is
>> present and certain fields are changed. It is used in both legacy and
>> scalable modes.
>
> hmmm. the kdoc says all the pasid entries are non-present, is it necessary
> to flush dev_tlb in such scenario? I suppose no present pasid entry means
> no pagetable as well.
The spec has defined the software behavior for cache invalidation in
"Table 28: Guidance to Software for Invalidations". This helper was
written according to it.
>
>> In the past, this helper would work even if some PASIDs were still in
>> use. After the changes introduced in this series, this PASID-in-use case
>> is removed. So remove the dead code.
>
> yeah, I got this part. As I mentioned, the only caller now is the
> domain_context_clear_one() which is used in the legacy path. That's why I
> feel like it is now more for legacy mode path now after this series.
It's also called by device_pasid_table_teardown().
Thanks,
baolu
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 07/12] iommu/vt-d: Put iopf enablement in domain attach path
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (5 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 06/12] iommu/vt-d: Cleanup intel_context_flush_present() Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 8:02 ` Tian, Kevin
2025-02-24 5:16 ` [PATCH v2 08/12] iommufd/selftest: " Lu Baolu
` (5 subsequent siblings)
12 siblings, 1 reply; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Lu Baolu
Update iopf enablement in the driver to use the new method, similar to
the arm-smmu-v3 driver. Enable iopf support when any domain with an
iopf_handler is attached, and disable it when the domain is removed.
Place all the logic for controlling the PRI and iopf queue in the domain
set/remove/replace paths. Keep track of the number of domains set to the
device and PASIDs that require iopf. When the first domain requiring iopf
is attached, add the device to the iopf queue and enable PRI. When the
last domain is removed, remove it from the iopf queue and disable PRI.
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
---
drivers/iommu/intel/iommu.c | 42 ++++++++++++++++++++++++++++++-----
drivers/iommu/intel/iommu.h | 43 ++++++++++++++++++++++++++++++++++++
drivers/iommu/intel/nested.c | 16 ++++++++++++--
drivers/iommu/intel/svm.c | 9 ++++++--
4 files changed, 100 insertions(+), 10 deletions(-)
diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
index 1d564240c977..20d07f5fea85 100644
--- a/drivers/iommu/intel/iommu.c
+++ b/drivers/iommu/intel/iommu.c
@@ -3223,6 +3223,9 @@ void device_block_translation(struct device *dev)
static int blocking_domain_attach_dev(struct iommu_domain *domain,
struct device *dev)
{
+ struct device_domain_info *info = dev_iommu_priv_get(dev);
+
+ iopf_for_domain_remove(info->domain ? &info->domain->domain : NULL, dev);
device_block_translation(dev);
return 0;
}
@@ -3432,7 +3435,15 @@ static int intel_iommu_attach_device(struct iommu_domain *domain,
if (ret)
return ret;
- return dmar_domain_attach_device(to_dmar_domain(domain), dev);
+ ret = iopf_for_domain_set(domain, dev);
+ if (ret)
+ return ret;
+
+ ret = dmar_domain_attach_device(to_dmar_domain(domain), dev);
+ if (ret)
+ iopf_for_domain_remove(domain, dev);
+
+ return ret;
}
static int intel_iommu_map(struct iommu_domain *domain,
@@ -3878,6 +3889,8 @@ int intel_iommu_enable_iopf(struct device *dev)
if (!info->pri_enabled)
return -ENODEV;
+ /* pri_enabled is protected by the group mutex. */
+ iommu_group_mutex_assert(dev);
if (info->iopf_refcount) {
info->iopf_refcount++;
return 0;
@@ -3900,6 +3913,7 @@ void intel_iommu_disable_iopf(struct device *dev)
if (WARN_ON(!info->pri_enabled))
return;
+ iommu_group_mutex_assert(dev);
if (--info->iopf_refcount)
return;
@@ -3911,8 +3925,7 @@ intel_iommu_dev_enable_feat(struct device *dev, enum iommu_dev_features feat)
{
switch (feat) {
case IOMMU_DEV_FEAT_IOPF:
- return intel_iommu_enable_iopf(dev);
-
+ return 0;
default:
return -ENODEV;
}
@@ -3923,7 +3936,6 @@ intel_iommu_dev_disable_feat(struct device *dev, enum iommu_dev_features feat)
{
switch (feat) {
case IOMMU_DEV_FEAT_IOPF:
- intel_iommu_disable_iopf(dev);
return 0;
default:
@@ -4004,6 +4016,7 @@ static int blocking_domain_set_dev_pasid(struct iommu_domain *domain,
{
struct device_domain_info *info = dev_iommu_priv_get(dev);
+ iopf_for_domain_remove(old, dev);
intel_pasid_tear_down_entry(info->iommu, dev, pasid, false);
domain_remove_dev_pasid(old, dev, pasid);
@@ -4077,6 +4090,10 @@ static int intel_iommu_set_dev_pasid(struct iommu_domain *domain,
if (IS_ERR(dev_pasid))
return PTR_ERR(dev_pasid);
+ ret = iopf_for_domain_replace(domain, old, dev);
+ if (ret)
+ goto out_remove_dev_pasid;
+
if (dmar_domain->use_first_level)
ret = domain_setup_first_level(iommu, dmar_domain,
dev, pasid, old);
@@ -4084,7 +4101,7 @@ static int intel_iommu_set_dev_pasid(struct iommu_domain *domain,
ret = domain_setup_second_level(iommu, dmar_domain,
dev, pasid, old);
if (ret)
- goto out_remove_dev_pasid;
+ goto out_unwind_iopf;
domain_remove_dev_pasid(old, dev, pasid);
@@ -4092,6 +4109,8 @@ static int intel_iommu_set_dev_pasid(struct iommu_domain *domain,
return 0;
+out_unwind_iopf:
+ iopf_for_domain_replace(old, domain, dev);
out_remove_dev_pasid:
domain_remove_dev_pasid(domain, dev, pasid);
return ret;
@@ -4309,6 +4328,11 @@ static int identity_domain_attach_dev(struct iommu_domain *domain, struct device
if (dev_is_real_dma_subdevice(dev))
return 0;
+ /*
+ * No PRI support with the global identity domain. No need to enable or
+ * disable PRI in this path as the iommu has been put in the blocking
+ * state.
+ */
if (sm_supported(iommu))
ret = intel_pasid_setup_pass_through(iommu, dev, IOMMU_NO_PASID);
else
@@ -4328,9 +4352,15 @@ static int identity_domain_set_dev_pasid(struct iommu_domain *domain,
if (!pasid_supported(iommu) || dev_is_real_dma_subdevice(dev))
return -EOPNOTSUPP;
+ ret = iopf_for_domain_replace(domain, old, dev);
+ if (ret)
+ return ret;
+
ret = domain_setup_passthrough(iommu, dev, pasid, old);
- if (ret)
+ if (ret) {
+ iopf_for_domain_replace(old, domain, dev);
return ret;
+ }
domain_remove_dev_pasid(old, dev, pasid);
return 0;
diff --git a/drivers/iommu/intel/iommu.h b/drivers/iommu/intel/iommu.h
index 754f6d7ade26..dfb0628fabf8 100644
--- a/drivers/iommu/intel/iommu.h
+++ b/drivers/iommu/intel/iommu.h
@@ -1317,6 +1317,49 @@ void intel_iommu_drain_pasid_prq(struct device *dev, u32 pasid);
int intel_iommu_enable_iopf(struct device *dev);
void intel_iommu_disable_iopf(struct device *dev);
+static inline int iopf_for_domain_set(struct iommu_domain *domain,
+ struct device *dev)
+{
+ if (!domain || !domain->iopf_handler)
+ return 0;
+
+ return intel_iommu_enable_iopf(dev);
+}
+
+static inline void iopf_for_domain_remove(struct iommu_domain *domain,
+ struct device *dev)
+{
+ if (!domain || !domain->iopf_handler)
+ return;
+
+ intel_iommu_disable_iopf(dev);
+}
+
+static inline int iopf_for_domain_replace(struct iommu_domain *new,
+ struct iommu_domain *old,
+ struct device *dev)
+{
+ if (!old)
+ return iopf_for_domain_set(new, dev);
+
+ if (!new) {
+ iopf_for_domain_remove(old, dev);
+ return 0;
+ }
+
+ /*
+ * Replace a non-iopf-capable domain with an iopf-capable one requires
+ * to enable PRI on the device. On the contrary, disable it.
+ */
+ if (new->iopf_handler && !old->iopf_handler)
+ return intel_iommu_enable_iopf(dev);
+
+ if (!new->iopf_handler && old->iopf_handler)
+ intel_iommu_disable_iopf(dev);
+
+ return 0;
+}
+
#ifdef CONFIG_INTEL_IOMMU_SVM
void intel_svm_check(struct intel_iommu *iommu);
struct iommu_domain *intel_svm_domain_alloc(struct device *dev,
diff --git a/drivers/iommu/intel/nested.c b/drivers/iommu/intel/nested.c
index aba92c00b427..ad307248bcae 100644
--- a/drivers/iommu/intel/nested.c
+++ b/drivers/iommu/intel/nested.c
@@ -56,10 +56,14 @@ static int intel_nested_attach_dev(struct iommu_domain *domain,
if (ret)
goto detach_iommu;
+ ret = iopf_for_domain_set(domain, dev);
+ if (ret)
+ goto unassign_tag;
+
ret = intel_pasid_setup_nested(iommu, dev,
IOMMU_NO_PASID, dmar_domain);
if (ret)
- goto unassign_tag;
+ goto disable_iopf;
info->domain = dmar_domain;
spin_lock_irqsave(&dmar_domain->lock, flags);
@@ -67,6 +71,8 @@ static int intel_nested_attach_dev(struct iommu_domain *domain,
spin_unlock_irqrestore(&dmar_domain->lock, flags);
return 0;
+disable_iopf:
+ iopf_for_domain_remove(domain, dev);
unassign_tag:
cache_tag_unassign_domain(dmar_domain, dev, IOMMU_NO_PASID);
detach_iommu:
@@ -166,14 +172,20 @@ static int intel_nested_set_dev_pasid(struct iommu_domain *domain,
if (IS_ERR(dev_pasid))
return PTR_ERR(dev_pasid);
- ret = domain_setup_nested(iommu, dmar_domain, dev, pasid, old);
+ ret = iopf_for_domain_replace(domain, old, dev);
if (ret)
goto out_remove_dev_pasid;
+ ret = domain_setup_nested(iommu, dmar_domain, dev, pasid, old);
+ if (ret)
+ goto out_unwind_iopf;
+
domain_remove_dev_pasid(old, dev, pasid);
return 0;
+out_unwind_iopf:
+ iopf_for_domain_replace(old, domain, dev);
out_remove_dev_pasid:
domain_remove_dev_pasid(domain, dev, pasid);
return ret;
diff --git a/drivers/iommu/intel/svm.c b/drivers/iommu/intel/svm.c
index ba93123cb4eb..f3da596410b5 100644
--- a/drivers/iommu/intel/svm.c
+++ b/drivers/iommu/intel/svm.c
@@ -164,18 +164,23 @@ static int intel_svm_set_dev_pasid(struct iommu_domain *domain,
if (IS_ERR(dev_pasid))
return PTR_ERR(dev_pasid);
+ ret = iopf_for_domain_replace(domain, old, dev);
+ if (ret)
+ goto out_remove_dev_pasid;
+
/* Setup the pasid table: */
sflags = cpu_feature_enabled(X86_FEATURE_LA57) ? PASID_FLAG_FL5LP : 0;
ret = __domain_setup_first_level(iommu, dev, pasid,
FLPT_DEFAULT_DID, mm->pgd,
sflags, old);
if (ret)
- goto out_remove_dev_pasid;
+ goto out_unwind_iopf;
domain_remove_dev_pasid(old, dev, pasid);
return 0;
-
+out_unwind_iopf:
+ iopf_for_domain_replace(old, domain, dev);
out_remove_dev_pasid:
domain_remove_dev_pasid(domain, dev, pasid);
return ret;
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 07/12] iommu/vt-d: Put iopf enablement in domain attach path
2025-02-24 5:16 ` [PATCH v2 07/12] iommu/vt-d: Put iopf enablement in domain attach path Lu Baolu
@ 2025-02-25 8:02 ` Tian, Kevin
0 siblings, 0 replies; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 8:02 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> Update iopf enablement in the driver to use the new method, similar to
> the arm-smmu-v3 driver. Enable iopf support when any domain with an
> iopf_handler is attached, and disable it when the domain is removed.
>
> Place all the logic for controlling the PRI and iopf queue in the domain
> set/remove/replace paths. Keep track of the number of domains set to the
> device and PASIDs that require iopf. When the first domain requiring iopf
> is attached, add the device to the iopf queue and enable PRI. When the
> last domain is removed, remove it from the iopf queue and disable PRI.
>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 08/12] iommufd/selftest: Put iopf enablement in domain attach path
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (6 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 07/12] iommu/vt-d: Put iopf enablement in domain attach path Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 8:16 ` Tian, Kevin
2025-02-24 5:16 ` [PATCH v2 09/12] dmaengine: idxd: Remove unnecessary IOMMU_DEV_FEAT_IOPF Lu Baolu
` (4 subsequent siblings)
12 siblings, 1 reply; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Lu Baolu
Update iopf enablement in the iommufd mock device driver to use the new
method, similar to the arm-smmu-v3 driver. Enable iopf support when any
domain with an iopf_handler is attached, and disable it when the domain
is removed.
Add a refcount in the mock device state structure to keep track of the
number of domains set to the device and PASIDs that require iopf.
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
---
drivers/iommu/iommufd/selftest.c | 63 ++++++++++++++++++++++++++------
1 file changed, 52 insertions(+), 11 deletions(-)
diff --git a/drivers/iommu/iommufd/selftest.c b/drivers/iommu/iommufd/selftest.c
index d40deb0a4f06..345be71b0328 100644
--- a/drivers/iommu/iommufd/selftest.c
+++ b/drivers/iommu/iommufd/selftest.c
@@ -58,6 +58,9 @@ enum {
MOCK_PFN_HUGE_IOVA = _MOCK_PFN_START << 2,
};
+static int mock_dev_enable_iopf(struct device *dev, struct iommu_domain *domain);
+static void mock_dev_disable_iopf(struct device *dev, struct iommu_domain *domain);
+
/*
* Syzkaller has trouble randomizing the correct iova to use since it is linked
* to the map ioctl's output, and it has no ide about that. So, simplify things.
@@ -164,6 +167,8 @@ struct mock_dev {
unsigned long flags;
int id;
u32 cache[MOCK_DEV_CACHE_NUM];
+ unsigned int iopf_refcount;
+ struct iommu_domain *domain;
};
static inline struct mock_dev *to_mock_dev(struct device *dev)
@@ -193,15 +198,33 @@ static int mock_domain_nop_attach(struct iommu_domain *domain,
struct device *dev)
{
struct mock_dev *mdev = to_mock_dev(dev);
+ int ret;
if (domain->dirty_ops && (mdev->flags & MOCK_FLAGS_DEVICE_NO_DIRTY))
return -EINVAL;
+ ret = mock_dev_enable_iopf(dev, domain);
+ if (ret)
+ return ret;
+
+ mock_dev_disable_iopf(dev, mdev->domain);
+ mdev->domain = domain;
+
+ return 0;
+}
+
+static int mock_domain_blocking_attach(struct iommu_domain *domain,
+ struct device *dev)
+{
+ struct mock_dev *mdev = to_mock_dev(dev);
+
+ mock_dev_disable_iopf(dev, mdev->domain);
+
return 0;
}
static const struct iommu_domain_ops mock_blocking_ops = {
- .attach_dev = mock_domain_nop_attach,
+ .attach_dev = mock_domain_blocking_attach,
};
static struct iommu_domain mock_blocking_domain = {
@@ -549,22 +572,42 @@ static void mock_domain_page_response(struct device *dev, struct iopf_fault *evt
{
}
-static int mock_dev_enable_feat(struct device *dev, enum iommu_dev_features feat)
+static int mock_dev_enable_iopf(struct device *dev, struct iommu_domain *domain)
{
- if (feat != IOMMU_DEV_FEAT_IOPF || !mock_iommu_iopf_queue)
+ struct mock_dev *mdev = to_mock_dev(dev);
+ int ret;
+
+ if (!domain || !domain->iopf_handler)
+ return 0;
+
+ if (!mock_iommu_iopf_queue)
return -ENODEV;
- return iopf_queue_add_device(mock_iommu_iopf_queue, dev);
+ if (mdev->iopf_refcount) {
+ mdev->iopf_refcount++;
+ return 0;
+ }
+
+ ret = iopf_queue_add_device(mock_iommu_iopf_queue, dev);
+ if (ret)
+ return ret;
+
+ mdev->iopf_refcount = 1;
+
+ return 0;
}
-static int mock_dev_disable_feat(struct device *dev, enum iommu_dev_features feat)
+static void mock_dev_disable_iopf(struct device *dev, struct iommu_domain *domain)
{
- if (feat != IOMMU_DEV_FEAT_IOPF || !mock_iommu_iopf_queue)
- return -ENODEV;
+ struct mock_dev *mdev = to_mock_dev(dev);
+
+ if (!domain || !domain->iopf_handler)
+ return;
+
+ if (--mdev->iopf_refcount)
+ return;
iopf_queue_remove_device(mock_iommu_iopf_queue, dev);
-
- return 0;
}
static void mock_viommu_destroy(struct iommufd_viommu *viommu)
@@ -709,8 +752,6 @@ static const struct iommu_ops mock_ops = {
.device_group = generic_device_group,
.probe_device = mock_probe_device,
.page_response = mock_domain_page_response,
- .dev_enable_feat = mock_dev_enable_feat,
- .dev_disable_feat = mock_dev_disable_feat,
.user_pasid_table = true,
.viommu_alloc = mock_viommu_alloc,
.default_domain_ops =
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 08/12] iommufd/selftest: Put iopf enablement in domain attach path
2025-02-24 5:16 ` [PATCH v2 08/12] iommufd/selftest: " Lu Baolu
@ 2025-02-25 8:16 ` Tian, Kevin
2025-02-26 4:06 ` Baolu Lu
0 siblings, 1 reply; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 8:16 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> Update iopf enablement in the iommufd mock device driver to use the new
> method, similar to the arm-smmu-v3 driver. Enable iopf support when any
> domain with an iopf_handler is attached, and disable it when the domain
> is removed.
>
> Add a refcount in the mock device state structure to keep track of the
> number of domains set to the device and PASIDs that require iopf.
>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
with a nit:
>
> + ret = mock_dev_enable_iopf(dev, domain);
> + if (ret)
> + return ret;
> +
> + mock_dev_disable_iopf(dev, mdev->domain);
You could use the same trick to simplify iopf_for_domain_replace()
in patch7?
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 08/12] iommufd/selftest: Put iopf enablement in domain attach path
2025-02-25 8:16 ` Tian, Kevin
@ 2025-02-26 4:06 ` Baolu Lu
0 siblings, 0 replies; 38+ messages in thread
From: Baolu Lu @ 2025-02-26 4:06 UTC (permalink / raw)
To: Tian, Kevin, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org
On 2/25/25 16:16, Tian, Kevin wrote:
>> + ret = mock_dev_enable_iopf(dev, domain);
>> + if (ret)
>> + return ret;
>> +
>> + mock_dev_disable_iopf(dev, mdev->domain);
> You could use the same trick to simplify iopf_for_domain_replace()
> in patch7?
Yes, simplify it like this,
static inline int iopf_for_domain_replace(struct iommu_domain *new,
struct iommu_domain *old,
struct device *dev)
{
int ret;
ret = iopf_for_domain_set(new, dev);
if (ret)
return ret;
iopf_for_domain_remove(old, dev);
return 0;
}
Thanks,
baolu
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 09/12] dmaengine: idxd: Remove unnecessary IOMMU_DEV_FEAT_IOPF
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (7 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 08/12] iommufd/selftest: " Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 8:16 ` Tian, Kevin
2025-02-24 5:16 ` [PATCH v2 10/12] uacce: " Lu Baolu
` (3 subsequent siblings)
12 siblings, 1 reply; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Lu Baolu, Jason Gunthorpe
The IOMMU_DEV_FEAT_IOPF implementation in the iommu driver is just a no-op.
It will also be removed from the iommu driver in the subsequent patch.
Remove it to avoid dead code.
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Acked-by: Vinod Koul <vkoul@kernel.org>
Reviewed-by: Dave Jiang <dave.jiang@intel.com>
Reviewed-by: Fenghua Yu <fenghuay@nvidia.com>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
---
drivers/dma/idxd/init.c | 37 ++++++-------------------------------
1 file changed, 6 insertions(+), 31 deletions(-)
diff --git a/drivers/dma/idxd/init.c b/drivers/dma/idxd/init.c
index 1e5038cca22c..d44944195807 100644
--- a/drivers/dma/idxd/init.c
+++ b/drivers/dma/idxd/init.c
@@ -626,21 +626,6 @@ static void idxd_disable_system_pasid(struct idxd_device *idxd)
idxd->pasid = IOMMU_PASID_INVALID;
}
-static int idxd_enable_sva(struct pci_dev *pdev)
-{
- int ret;
-
- ret = iommu_dev_enable_feature(&pdev->dev, IOMMU_DEV_FEAT_IOPF);
- if (ret)
- return ret;
- return 0;
-}
-
-static void idxd_disable_sva(struct pci_dev *pdev)
-{
- iommu_dev_disable_feature(&pdev->dev, IOMMU_DEV_FEAT_IOPF);
-}
-
static int idxd_probe(struct idxd_device *idxd)
{
struct pci_dev *pdev = idxd->pdev;
@@ -655,17 +640,13 @@ static int idxd_probe(struct idxd_device *idxd)
dev_dbg(dev, "IDXD reset complete\n");
if (IS_ENABLED(CONFIG_INTEL_IDXD_SVM) && sva) {
- if (idxd_enable_sva(pdev)) {
- dev_warn(dev, "Unable to turn on user SVA feature.\n");
- } else {
- set_bit(IDXD_FLAG_USER_PASID_ENABLED, &idxd->flags);
+ set_bit(IDXD_FLAG_USER_PASID_ENABLED, &idxd->flags);
- rc = idxd_enable_system_pasid(idxd);
- if (rc)
- dev_warn(dev, "No in-kernel DMA with PASID. %d\n", rc);
- else
- set_bit(IDXD_FLAG_PASID_ENABLED, &idxd->flags);
- }
+ rc = idxd_enable_system_pasid(idxd);
+ if (rc)
+ dev_warn(dev, "No in-kernel DMA with PASID. %d\n", rc);
+ else
+ set_bit(IDXD_FLAG_PASID_ENABLED, &idxd->flags);
} else if (!sva) {
dev_warn(dev, "User forced SVA off via module param.\n");
}
@@ -703,8 +684,6 @@ static int idxd_probe(struct idxd_device *idxd)
err:
if (device_pasid_enabled(idxd))
idxd_disable_system_pasid(idxd);
- if (device_user_pasid_enabled(idxd))
- idxd_disable_sva(pdev);
return rc;
}
@@ -715,8 +694,6 @@ static void idxd_cleanup(struct idxd_device *idxd)
idxd_cleanup_internals(idxd);
if (device_pasid_enabled(idxd))
idxd_disable_system_pasid(idxd);
- if (device_user_pasid_enabled(idxd))
- idxd_disable_sva(idxd->pdev);
}
/*
@@ -1248,8 +1225,6 @@ static void idxd_remove(struct pci_dev *pdev)
free_irq(irq_entry->vector, irq_entry);
pci_free_irq_vectors(pdev);
pci_iounmap(pdev, idxd->reg_base);
- if (device_user_pasid_enabled(idxd))
- idxd_disable_sva(pdev);
pci_disable_device(pdev);
destroy_workqueue(idxd->wq);
perfmon_pmu_remove(idxd);
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 09/12] dmaengine: idxd: Remove unnecessary IOMMU_DEV_FEAT_IOPF
2025-02-24 5:16 ` [PATCH v2 09/12] dmaengine: idxd: Remove unnecessary IOMMU_DEV_FEAT_IOPF Lu Baolu
@ 2025-02-25 8:16 ` Tian, Kevin
0 siblings, 0 replies; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 8:16 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org,
Jason Gunthorpe
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> The IOMMU_DEV_FEAT_IOPF implementation in the iommu driver is just a
> no-op.
> It will also be removed from the iommu driver in the subsequent patch.
> Remove it to avoid dead code.
>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
> Acked-by: Vinod Koul <vkoul@kernel.org>
> Reviewed-by: Dave Jiang <dave.jiang@intel.com>
> Reviewed-by: Fenghua Yu <fenghuay@nvidia.com>
> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 10/12] uacce: Remove unnecessary IOMMU_DEV_FEAT_IOPF
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (8 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 09/12] dmaengine: idxd: Remove unnecessary IOMMU_DEV_FEAT_IOPF Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-24 10:00 ` Zhangfei Gao
2025-02-25 8:17 ` Tian, Kevin
2025-02-24 5:16 ` [PATCH v2 11/12] iommufd: " Lu Baolu
` (2 subsequent siblings)
12 siblings, 2 replies; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Lu Baolu, Jason Gunthorpe
None of the drivers implement anything for IOMMU_DEV_FEAT_IOPF anymore,
remove it to avoid dead code.
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
---
drivers/misc/uacce/uacce.c | 31 -------------------------------
1 file changed, 31 deletions(-)
diff --git a/drivers/misc/uacce/uacce.c b/drivers/misc/uacce/uacce.c
index 2a1db2abeeca..42e7d2a2a90c 100644
--- a/drivers/misc/uacce/uacce.c
+++ b/drivers/misc/uacce/uacce.c
@@ -465,31 +465,6 @@ static void uacce_release(struct device *dev)
kfree(uacce);
}
-static unsigned int uacce_enable_sva(struct device *parent, unsigned int flags)
-{
- int ret;
-
- if (!(flags & UACCE_DEV_SVA))
- return flags;
-
- flags &= ~UACCE_DEV_SVA;
-
- ret = iommu_dev_enable_feature(parent, IOMMU_DEV_FEAT_IOPF);
- if (ret) {
- dev_err(parent, "failed to enable IOPF feature! ret = %pe\n", ERR_PTR(ret));
- return flags;
- }
- return flags | UACCE_DEV_SVA;
-}
-
-static void uacce_disable_sva(struct uacce_device *uacce)
-{
- if (!(uacce->flags & UACCE_DEV_SVA))
- return;
-
- iommu_dev_disable_feature(uacce->parent, IOMMU_DEV_FEAT_IOPF);
-}
-
/**
* uacce_alloc() - alloc an accelerator
* @parent: pointer of uacce parent device
@@ -509,8 +484,6 @@ struct uacce_device *uacce_alloc(struct device *parent,
if (!uacce)
return ERR_PTR(-ENOMEM);
- flags = uacce_enable_sva(parent, flags);
-
uacce->parent = parent;
uacce->flags = flags;
uacce->ops = interface->ops;
@@ -533,7 +506,6 @@ struct uacce_device *uacce_alloc(struct device *parent,
return uacce;
err_with_uacce:
- uacce_disable_sva(uacce);
kfree(uacce);
return ERR_PTR(ret);
}
@@ -596,9 +568,6 @@ void uacce_remove(struct uacce_device *uacce)
unmap_mapping_range(q->mapping, 0, 0, 1);
}
- /* disable sva now since no opened queues */
- uacce_disable_sva(uacce);
-
if (uacce->cdev)
cdev_device_del(uacce->cdev, &uacce->dev);
xa_erase(&uacce_xa, uacce->dev_id);
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* Re: [PATCH v2 10/12] uacce: Remove unnecessary IOMMU_DEV_FEAT_IOPF
2025-02-24 5:16 ` [PATCH v2 10/12] uacce: " Lu Baolu
@ 2025-02-24 10:00 ` Zhangfei Gao
2025-02-25 8:17 ` Tian, Kevin
1 sibling, 0 replies; 38+ messages in thread
From: Zhangfei Gao @ 2025-02-24 10:00 UTC (permalink / raw)
To: Lu Baolu
Cc: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian, Dave Jiang, Vinod Koul, Fenghua Yu, Zhou Wang, iommu,
linux-kernel, Jason Gunthorpe
On Mon, 24 Feb 2025 at 13:16, Lu Baolu <baolu.lu@linux.intel.com> wrote:
>
> None of the drivers implement anything for IOMMU_DEV_FEAT_IOPF anymore,
> remove it to avoid dead code.
>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Acked-by: Zhangfei Gao <zhangfei.gao@linaro.org>
> ---
> drivers/misc/uacce/uacce.c | 31 -------------------------------
> 1 file changed, 31 deletions(-)
>
> diff --git a/drivers/misc/uacce/uacce.c b/drivers/misc/uacce/uacce.c
> index 2a1db2abeeca..42e7d2a2a90c 100644
> --- a/drivers/misc/uacce/uacce.c
> +++ b/drivers/misc/uacce/uacce.c
> @@ -465,31 +465,6 @@ static void uacce_release(struct device *dev)
> kfree(uacce);
> }
>
> -static unsigned int uacce_enable_sva(struct device *parent, unsigned int flags)
> -{
> - int ret;
> -
> - if (!(flags & UACCE_DEV_SVA))
> - return flags;
> -
> - flags &= ~UACCE_DEV_SVA;
> -
> - ret = iommu_dev_enable_feature(parent, IOMMU_DEV_FEAT_IOPF);
> - if (ret) {
> - dev_err(parent, "failed to enable IOPF feature! ret = %pe\n", ERR_PTR(ret));
> - return flags;
> - }
> - return flags | UACCE_DEV_SVA;
> -}
> -
> -static void uacce_disable_sva(struct uacce_device *uacce)
> -{
> - if (!(uacce->flags & UACCE_DEV_SVA))
> - return;
> -
> - iommu_dev_disable_feature(uacce->parent, IOMMU_DEV_FEAT_IOPF);
> -}
> -
> /**
> * uacce_alloc() - alloc an accelerator
> * @parent: pointer of uacce parent device
> @@ -509,8 +484,6 @@ struct uacce_device *uacce_alloc(struct device *parent,
> if (!uacce)
> return ERR_PTR(-ENOMEM);
>
> - flags = uacce_enable_sva(parent, flags);
> -
> uacce->parent = parent;
> uacce->flags = flags;
> uacce->ops = interface->ops;
> @@ -533,7 +506,6 @@ struct uacce_device *uacce_alloc(struct device *parent,
> return uacce;
>
> err_with_uacce:
> - uacce_disable_sva(uacce);
> kfree(uacce);
> return ERR_PTR(ret);
> }
> @@ -596,9 +568,6 @@ void uacce_remove(struct uacce_device *uacce)
> unmap_mapping_range(q->mapping, 0, 0, 1);
> }
>
> - /* disable sva now since no opened queues */
> - uacce_disable_sva(uacce);
> -
> if (uacce->cdev)
> cdev_device_del(uacce->cdev, &uacce->dev);
> xa_erase(&uacce_xa, uacce->dev_id);
> --
> 2.43.0
>
^ permalink raw reply [flat|nested] 38+ messages in thread* RE: [PATCH v2 10/12] uacce: Remove unnecessary IOMMU_DEV_FEAT_IOPF
2025-02-24 5:16 ` [PATCH v2 10/12] uacce: " Lu Baolu
2025-02-24 10:00 ` Zhangfei Gao
@ 2025-02-25 8:17 ` Tian, Kevin
1 sibling, 0 replies; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 8:17 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org,
Jason Gunthorpe
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> None of the drivers implement anything for IOMMU_DEV_FEAT_IOPF
> anymore,
> remove it to avoid dead code.
>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 11/12] iommufd: Remove unnecessary IOMMU_DEV_FEAT_IOPF
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (9 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 10/12] uacce: " Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 8:18 ` Tian, Kevin
2025-02-24 5:16 ` [PATCH v2 12/12] iommu: Remove iommu_dev_enable/disable_feature() Lu Baolu
2025-02-24 9:58 ` [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Zhangfei Gao
12 siblings, 1 reply; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Lu Baolu, Jason Gunthorpe
The iopf enablement has been moved to the iommu drivers. It is unnecessary
for iommufd to handle iopf enablement. Remove the iopf enablement logic to
avoid duplication.
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
---
drivers/iommu/iommufd/device.c | 1 -
drivers/iommu/iommufd/fault.c | 111 ++++++------------------
drivers/iommu/iommufd/iommufd_private.h | 3 -
3 files changed, 28 insertions(+), 87 deletions(-)
diff --git a/drivers/iommu/iommufd/device.c b/drivers/iommu/iommufd/device.c
index dfd0898fb6c1..47e36456b438 100644
--- a/drivers/iommu/iommufd/device.c
+++ b/drivers/iommu/iommufd/device.c
@@ -215,7 +215,6 @@ struct iommufd_device *iommufd_device_bind(struct iommufd_ctx *ictx,
refcount_inc(&idev->obj.users);
/* igroup refcount moves into iommufd_device */
idev->igroup = igroup;
- mutex_init(&idev->iopf_lock);
/*
* If the caller fails after this success it must call
diff --git a/drivers/iommu/iommufd/fault.c b/drivers/iommu/iommufd/fault.c
index d9a937450e55..4776c632cff2 100644
--- a/drivers/iommu/iommufd/fault.c
+++ b/drivers/iommu/iommufd/fault.c
@@ -17,49 +17,6 @@
#include "../iommu-priv.h"
#include "iommufd_private.h"
-static int iommufd_fault_iopf_enable(struct iommufd_device *idev)
-{
- struct device *dev = idev->dev;
- int ret;
-
- /*
- * Once we turn on PCI/PRI support for VF, the response failure code
- * should not be forwarded to the hardware due to PRI being a shared
- * resource between PF and VFs. There is no coordination for this
- * shared capability. This waits for a vPRI reset to recover.
- */
- if (dev_is_pci(dev)) {
- struct pci_dev *pdev = to_pci_dev(dev);
-
- if (pdev->is_virtfn && pci_pri_supported(pdev))
- return -EINVAL;
- }
-
- mutex_lock(&idev->iopf_lock);
- /* Device iopf has already been on. */
- if (++idev->iopf_enabled > 1) {
- mutex_unlock(&idev->iopf_lock);
- return 0;
- }
-
- ret = iommu_dev_enable_feature(dev, IOMMU_DEV_FEAT_IOPF);
- if (ret)
- --idev->iopf_enabled;
- mutex_unlock(&idev->iopf_lock);
-
- return ret;
-}
-
-static void iommufd_fault_iopf_disable(struct iommufd_device *idev)
-{
- mutex_lock(&idev->iopf_lock);
- if (!WARN_ON(idev->iopf_enabled == 0)) {
- if (--idev->iopf_enabled == 0)
- iommu_dev_disable_feature(idev->dev, IOMMU_DEV_FEAT_IOPF);
- }
- mutex_unlock(&idev->iopf_lock);
-}
-
static int __fault_domain_attach_dev(struct iommufd_hw_pagetable *hwpt,
struct iommufd_device *idev)
{
@@ -82,20 +39,23 @@ static int __fault_domain_attach_dev(struct iommufd_hw_pagetable *hwpt,
int iommufd_fault_domain_attach_dev(struct iommufd_hw_pagetable *hwpt,
struct iommufd_device *idev)
{
- int ret;
-
if (!hwpt->fault)
return -EINVAL;
- ret = iommufd_fault_iopf_enable(idev);
- if (ret)
- return ret;
+ /*
+ * Once we turn on PCI/PRI support for VF, the response failure code
+ * should not be forwarded to the hardware due to PRI being a shared
+ * resource between PF and VFs. There is no coordination for this
+ * shared capability. This waits for a vPRI reset to recover.
+ */
+ if (dev_is_pci(idev->dev)) {
+ struct pci_dev *pdev = to_pci_dev(idev->dev);
- ret = __fault_domain_attach_dev(hwpt, idev);
- if (ret)
- iommufd_fault_iopf_disable(idev);
+ if (pdev->is_virtfn && pci_pri_supported(pdev))
+ return -EINVAL;
+ }
- return ret;
+ return __fault_domain_attach_dev(hwpt, idev);
}
static void iommufd_auto_response_faults(struct iommufd_hw_pagetable *hwpt,
@@ -155,13 +115,12 @@ void iommufd_fault_domain_detach_dev(struct iommufd_hw_pagetable *hwpt,
handle = iommufd_device_get_attach_handle(idev);
iommu_detach_group_handle(hwpt->domain, idev->igroup->group);
iommufd_auto_response_faults(hwpt, handle);
- iommufd_fault_iopf_disable(idev);
kfree(handle);
}
-static int __fault_domain_replace_dev(struct iommufd_device *idev,
- struct iommufd_hw_pagetable *hwpt,
- struct iommufd_hw_pagetable *old)
+int iommufd_fault_domain_replace_dev(struct iommufd_device *idev,
+ struct iommufd_hw_pagetable *hwpt,
+ struct iommufd_hw_pagetable *old)
{
struct iommufd_attach_handle *handle, *curr = NULL;
int ret;
@@ -170,6 +129,19 @@ static int __fault_domain_replace_dev(struct iommufd_device *idev,
curr = iommufd_device_get_attach_handle(idev);
if (hwpt->fault) {
+ /*
+ * Once we turn on PCI/PRI support for VF, the response failure code
+ * should not be forwarded to the hardware due to PRI being a shared
+ * resource between PF and VFs. There is no coordination for this
+ * shared capability. This waits for a vPRI reset to recover.
+ */
+ if (dev_is_pci(idev->dev)) {
+ struct pci_dev *pdev = to_pci_dev(idev->dev);
+
+ if (pdev->is_virtfn && pci_pri_supported(pdev))
+ return -EINVAL;
+ }
+
handle = kzalloc(sizeof(*handle), GFP_KERNEL);
if (!handle)
return -ENOMEM;
@@ -190,33 +162,6 @@ static int __fault_domain_replace_dev(struct iommufd_device *idev,
return ret;
}
-int iommufd_fault_domain_replace_dev(struct iommufd_device *idev,
- struct iommufd_hw_pagetable *hwpt,
- struct iommufd_hw_pagetable *old)
-{
- bool iopf_off = !hwpt->fault && old->fault;
- bool iopf_on = hwpt->fault && !old->fault;
- int ret;
-
- if (iopf_on) {
- ret = iommufd_fault_iopf_enable(idev);
- if (ret)
- return ret;
- }
-
- ret = __fault_domain_replace_dev(idev, hwpt, old);
- if (ret) {
- if (iopf_on)
- iommufd_fault_iopf_disable(idev);
- return ret;
- }
-
- if (iopf_off)
- iommufd_fault_iopf_disable(idev);
-
- return 0;
-}
-
void iommufd_fault_destroy(struct iommufd_object *obj)
{
struct iommufd_fault *fault = container_of(obj, struct iommufd_fault, obj);
diff --git a/drivers/iommu/iommufd/iommufd_private.h b/drivers/iommu/iommufd/iommufd_private.h
index 0b1bafc7fd99..0eb3779db156 100644
--- a/drivers/iommu/iommufd/iommufd_private.h
+++ b/drivers/iommu/iommufd/iommufd_private.h
@@ -399,9 +399,6 @@ struct iommufd_device {
/* always the physical device */
struct device *dev;
bool enforce_cache_coherency;
- /* protect iopf_enabled counter */
- struct mutex iopf_lock;
- unsigned int iopf_enabled;
};
static inline struct iommufd_device *
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 11/12] iommufd: Remove unnecessary IOMMU_DEV_FEAT_IOPF
2025-02-24 5:16 ` [PATCH v2 11/12] iommufd: " Lu Baolu
@ 2025-02-25 8:18 ` Tian, Kevin
0 siblings, 0 replies; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 8:18 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org,
Jason Gunthorpe
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> The iopf enablement has been moved to the iommu drivers. It is unnecessary
> for iommufd to handle iopf enablement. Remove the iopf enablement logic
> to
> avoid duplication.
>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
^ permalink raw reply [flat|nested] 38+ messages in thread
* [PATCH v2 12/12] iommu: Remove iommu_dev_enable/disable_feature()
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (10 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 11/12] iommufd: " Lu Baolu
@ 2025-02-24 5:16 ` Lu Baolu
2025-02-25 8:18 ` Tian, Kevin
2025-02-24 9:58 ` [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Zhangfei Gao
12 siblings, 1 reply; 38+ messages in thread
From: Lu Baolu @ 2025-02-24 5:16 UTC (permalink / raw)
To: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian
Cc: Dave Jiang, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu, linux-kernel, Lu Baolu, Jason Gunthorpe
No external drivers use these interfaces anymore. Furthermore, no existing
iommu drivers implement anything in the callbacks. Remove them to avoid
dead code.
Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
---
drivers/iommu/amd/iommu.c | 32 -------------------
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c | 34 ---------------------
drivers/iommu/intel/iommu.c | 25 ---------------
drivers/iommu/iommu.c | 32 -------------------
include/linux/iommu.h | 28 -----------------
5 files changed, 151 deletions(-)
diff --git a/drivers/iommu/amd/iommu.c b/drivers/iommu/amd/iommu.c
index e3653bdfcd7d..8d74afa552fb 100644
--- a/drivers/iommu/amd/iommu.c
+++ b/drivers/iommu/amd/iommu.c
@@ -2983,36 +2983,6 @@ static const struct iommu_dirty_ops amd_dirty_ops = {
.read_and_clear_dirty = amd_iommu_read_and_clear_dirty,
};
-static int amd_iommu_dev_enable_feature(struct device *dev,
- enum iommu_dev_features feat)
-{
- int ret = 0;
-
- switch (feat) {
- case IOMMU_DEV_FEAT_IOPF:
- break;
- default:
- ret = -EINVAL;
- break;
- }
- return ret;
-}
-
-static int amd_iommu_dev_disable_feature(struct device *dev,
- enum iommu_dev_features feat)
-{
- int ret = 0;
-
- switch (feat) {
- case IOMMU_DEV_FEAT_IOPF:
- break;
- default:
- ret = -EINVAL;
- break;
- }
- return ret;
-}
-
const struct iommu_ops amd_iommu_ops = {
.capable = amd_iommu_capable,
.blocked_domain = &blocked_domain,
@@ -3026,8 +2996,6 @@ const struct iommu_ops amd_iommu_ops = {
.get_resv_regions = amd_iommu_get_resv_regions,
.is_attach_deferred = amd_iommu_is_attach_deferred,
.def_domain_type = amd_iommu_def_domain_type,
- .dev_enable_feat = amd_iommu_dev_enable_feature,
- .dev_disable_feat = amd_iommu_dev_disable_feature,
.page_response = amd_iommu_page_response,
.default_domain_ops = &(const struct iommu_domain_ops) {
.attach_dev = amd_iommu_attach_device,
diff --git a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
index 3a2faf4fdd09..31500e25c4a5 100644
--- a/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
+++ b/drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
@@ -3620,38 +3620,6 @@ static void arm_smmu_get_resv_regions(struct device *dev,
iommu_dma_get_resv_regions(dev, head);
}
-static int arm_smmu_dev_enable_feature(struct device *dev,
- enum iommu_dev_features feat)
-{
- struct arm_smmu_master *master = dev_iommu_priv_get(dev);
-
- if (!master)
- return -ENODEV;
-
- switch (feat) {
- case IOMMU_DEV_FEAT_IOPF:
- return 0;
- default:
- return -EINVAL;
- }
-}
-
-static int arm_smmu_dev_disable_feature(struct device *dev,
- enum iommu_dev_features feat)
-{
- struct arm_smmu_master *master = dev_iommu_priv_get(dev);
-
- if (!master)
- return -EINVAL;
-
- switch (feat) {
- case IOMMU_DEV_FEAT_IOPF:
- return 0;
- default:
- return -EINVAL;
- }
-}
-
/*
* HiSilicon PCIe tune and trace device can be used to trace TLP headers on the
* PCIe link and save the data to memory by DMA. The hardware is restricted to
@@ -3684,8 +3652,6 @@ static struct iommu_ops arm_smmu_ops = {
.device_group = arm_smmu_device_group,
.of_xlate = arm_smmu_of_xlate,
.get_resv_regions = arm_smmu_get_resv_regions,
- .dev_enable_feat = arm_smmu_dev_enable_feature,
- .dev_disable_feat = arm_smmu_dev_disable_feature,
.page_response = arm_smmu_page_response,
.def_domain_type = arm_smmu_def_domain_type,
.viommu_alloc = arm_vsmmu_alloc,
diff --git a/drivers/iommu/intel/iommu.c b/drivers/iommu/intel/iommu.c
index 20d07f5fea85..aad11e76ac40 100644
--- a/drivers/iommu/intel/iommu.c
+++ b/drivers/iommu/intel/iommu.c
@@ -3920,29 +3920,6 @@ void intel_iommu_disable_iopf(struct device *dev)
iopf_queue_remove_device(iommu->iopf_queue, dev);
}
-static int
-intel_iommu_dev_enable_feat(struct device *dev, enum iommu_dev_features feat)
-{
- switch (feat) {
- case IOMMU_DEV_FEAT_IOPF:
- return 0;
- default:
- return -ENODEV;
- }
-}
-
-static int
-intel_iommu_dev_disable_feat(struct device *dev, enum iommu_dev_features feat)
-{
- switch (feat) {
- case IOMMU_DEV_FEAT_IOPF:
- return 0;
-
- default:
- return -ENODEV;
- }
-}
-
static bool intel_iommu_is_attach_deferred(struct device *dev)
{
struct device_domain_info *info = dev_iommu_priv_get(dev);
@@ -4387,8 +4364,6 @@ const struct iommu_ops intel_iommu_ops = {
.release_device = intel_iommu_release_device,
.get_resv_regions = intel_iommu_get_resv_regions,
.device_group = intel_iommu_device_group,
- .dev_enable_feat = intel_iommu_dev_enable_feat,
- .dev_disable_feat = intel_iommu_dev_disable_feat,
.is_attach_deferred = intel_iommu_is_attach_deferred,
.def_domain_type = device_def_domain_type,
.pgsize_bitmap = SZ_4K,
diff --git a/drivers/iommu/iommu.c b/drivers/iommu/iommu.c
index 60aed01e54f2..e19c04304b03 100644
--- a/drivers/iommu/iommu.c
+++ b/drivers/iommu/iommu.c
@@ -2877,38 +2877,6 @@ int iommu_fwspec_add_ids(struct device *dev, const u32 *ids, int num_ids)
}
EXPORT_SYMBOL_GPL(iommu_fwspec_add_ids);
-/*
- * Per device IOMMU features.
- */
-int iommu_dev_enable_feature(struct device *dev, enum iommu_dev_features feat)
-{
- if (dev_has_iommu(dev)) {
- const struct iommu_ops *ops = dev_iommu_ops(dev);
-
- if (ops->dev_enable_feat)
- return ops->dev_enable_feat(dev, feat);
- }
-
- return -ENODEV;
-}
-EXPORT_SYMBOL_GPL(iommu_dev_enable_feature);
-
-/*
- * The device drivers should do the necessary cleanups before calling this.
- */
-int iommu_dev_disable_feature(struct device *dev, enum iommu_dev_features feat)
-{
- if (dev_has_iommu(dev)) {
- const struct iommu_ops *ops = dev_iommu_ops(dev);
-
- if (ops->dev_disable_feat)
- return ops->dev_disable_feat(dev, feat);
- }
-
- return -EBUSY;
-}
-EXPORT_SYMBOL_GPL(iommu_dev_disable_feature);
-
/**
* iommu_setup_default_domain - Set the default_domain for the group
* @group: Group to change
diff --git a/include/linux/iommu.h b/include/linux/iommu.h
index 1d0dde49168d..127f99acab5f 100644
--- a/include/linux/iommu.h
+++ b/include/linux/iommu.h
@@ -301,16 +301,6 @@ struct iommu_iort_rmr_data {
u32 num_sids;
};
-/**
- * enum iommu_dev_features - Per device IOMMU features
- * @IOMMU_DEV_FEAT_IOPF: I/O Page Faults such as PRI or Stall.
- *
- * Device drivers enable a feature using iommu_dev_enable_feature().
- */
-enum iommu_dev_features {
- IOMMU_DEV_FEAT_IOPF,
-};
-
#define IOMMU_NO_PASID (0U) /* Reserved for DMA w/o PASID */
#define IOMMU_FIRST_GLOBAL_PASID (1U) /*starting range for allocation */
#define IOMMU_PASID_INVALID (-1U)
@@ -630,9 +620,6 @@ struct iommu_ops {
bool (*is_attach_deferred)(struct device *dev);
/* Per device IOMMU features */
- int (*dev_enable_feat)(struct device *dev, enum iommu_dev_features f);
- int (*dev_disable_feat)(struct device *dev, enum iommu_dev_features f);
-
void (*page_response)(struct device *dev, struct iopf_fault *evt,
struct iommu_page_response *msg);
@@ -1102,9 +1089,6 @@ void dev_iommu_priv_set(struct device *dev, void *priv);
extern struct mutex iommu_probe_device_lock;
int iommu_probe_device(struct device *dev);
-int iommu_dev_enable_feature(struct device *dev, enum iommu_dev_features f);
-int iommu_dev_disable_feature(struct device *dev, enum iommu_dev_features f);
-
int iommu_device_use_default_domain(struct device *dev);
void iommu_device_unuse_default_domain(struct device *dev);
@@ -1393,18 +1377,6 @@ static inline int iommu_fwspec_add_ids(struct device *dev, u32 *ids,
return -ENODEV;
}
-static inline int
-iommu_dev_enable_feature(struct device *dev, enum iommu_dev_features feat)
-{
- return -ENODEV;
-}
-
-static inline int
-iommu_dev_disable_feature(struct device *dev, enum iommu_dev_features feat)
-{
- return -ENODEV;
-}
-
static inline struct iommu_fwspec *dev_iommu_fwspec_get(struct device *dev)
{
return NULL;
--
2.43.0
^ permalink raw reply related [flat|nested] 38+ messages in thread* RE: [PATCH v2 12/12] iommu: Remove iommu_dev_enable/disable_feature()
2025-02-24 5:16 ` [PATCH v2 12/12] iommu: Remove iommu_dev_enable/disable_feature() Lu Baolu
@ 2025-02-25 8:18 ` Tian, Kevin
0 siblings, 0 replies; 38+ messages in thread
From: Tian, Kevin @ 2025-02-25 8:18 UTC (permalink / raw)
To: Lu Baolu, Joerg Roedel, Will Deacon, Robin Murphy,
Jason Gunthorpe
Cc: Jiang, Dave, Vinod Koul, Fenghua Yu, Zhangfei Gao, Zhou Wang,
iommu@lists.linux.dev, linux-kernel@vger.kernel.org,
Jason Gunthorpe
> From: Lu Baolu <baolu.lu@linux.intel.com>
> Sent: Monday, February 24, 2025 1:16 PM
>
> No external drivers use these interfaces anymore. Furthermore, no existing
> iommu drivers implement anything in the callbacks. Remove them to avoid
> dead code.
>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Reviewed-by: Kevin Tian <kevin.tian@intel.com>
^ permalink raw reply [flat|nested] 38+ messages in thread
* Re: [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF
2025-02-24 5:16 [PATCH v2 00/12] iommu: Remove IOMMU_DEV_FEAT_SVA/_IOPF Lu Baolu
` (11 preceding siblings ...)
2025-02-24 5:16 ` [PATCH v2 12/12] iommu: Remove iommu_dev_enable/disable_feature() Lu Baolu
@ 2025-02-24 9:58 ` Zhangfei Gao
12 siblings, 0 replies; 38+ messages in thread
From: Zhangfei Gao @ 2025-02-24 9:58 UTC (permalink / raw)
To: Lu Baolu
Cc: Joerg Roedel, Will Deacon, Robin Murphy, Jason Gunthorpe,
Kevin Tian, Dave Jiang, Vinod Koul, Fenghua Yu, Zhou Wang, iommu,
linux-kernel, Jason Gunthorpe
On Mon, 24 Feb 2025 at 13:16, Lu Baolu <baolu.lu@linux.intel.com> wrote:
>
> The new method for driver fault reporting support relies on the domain
> to specify a iopf_handler. The driver should detect this and setup the
> HW when fault capable domains are attached.
>
> Move SMMUv3 to use this method and have VT-D validate support during
> attach so that all three fault capable drivers have a no-op FEAT_SVA and
> _IOPF. Then remove them.
>
> This was initiated by Jason. I'm following up to remove FEAT_IOPF and
> further clean up.
>
> The whole series is also available at github:
> https://github.com/LuBaolu/intel-iommu/commits/iommu_no_feat-v2
>
> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
> Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com>
>
Thanks Baolu
Tested-by: Zhangfei Gao <zhangfei.gao@linaro.org>
^ permalink raw reply [flat|nested] 38+ messages in thread