From: Yi Liu <yi.l.liu@intel.com>
To: Baolu Lu <baolu.lu@linux.intel.com>, <joro@8bytes.org>,
<jgg@nvidia.com>, <kevin.tian@intel.com>, <will@kernel.org>
Cc: <alex.williamson@redhat.com>, <eric.auger@redhat.com>,
<nicolinc@nvidia.com>, <kvm@vger.kernel.org>,
<chao.p.peng@linux.intel.com>, <iommu@lists.linux.dev>,
<zhenzhong.duan@intel.com>, <vasant.hegde@amd.com>
Subject: Re: [PATCH v3 3/9] iommu/vt-d: Let intel_pasid_tear_down_entry() return pasid entry
Date: Tue, 22 Oct 2024 17:38:38 +0800 [thread overview]
Message-ID: <fe88f071-0d06-4838-9ce6-a5bcccf10163@intel.com> (raw)
In-Reply-To: <9d726285-730a-400d-8d45-f494b2c62205@linux.intel.com>
On 2024/10/22 17:23, Baolu Lu wrote:
> On 2024/10/21 15:24, Yi Liu wrote:
>> On 2024/10/21 14:59, Baolu Lu wrote:
>>> On 2024/10/21 14:35, Yi Liu wrote:
>>>> On 2024/10/21 14:13, Baolu Lu wrote:
>>>>> On 2024/10/18 13:53, Yi Liu wrote:
>>>>>> intel_pasid_tear_down_entry() finds the pasid entry and tears it down.
>>>>>> There are paths that need to get the pasid entry, tear it down and
>>>>>> re-configure it. Letting intel_pasid_tear_down_entry() return the pasid
>>>>>> entry can avoid duplicate codes to get the pasid entry. No functional
>>>>>> change is intended.
>>>>>>
>>>>>> Signed-off-by: Yi Liu<yi.l.liu@intel.com>
>>>>>> ---
>>>>>> drivers/iommu/intel/pasid.c | 11 ++++++++---
>>>>>> drivers/iommu/intel/pasid.h | 5 +++--
>>>>>> 2 files changed, 11 insertions(+), 5 deletions(-)
>>>>>>
>>>>>> diff --git a/drivers/iommu/intel/pasid.c b/drivers/iommu/intel/pasid.c
>>>>>> index 2898e7af2cf4..336f9425214c 100644
>>>>>> --- a/drivers/iommu/intel/pasid.c
>>>>>> +++ b/drivers/iommu/intel/pasid.c
>>>>>> @@ -239,9 +239,12 @@ devtlb_invalidation_with_pasid(struct
>>>>>> intel_iommu *iommu,
>>>>>> /*
>>>>>> * Caller can request to drain PRQ in this helper if it hasn't done
>>>>>> so,
>>>>>> * e.g. in a path which doesn't follow remove_dev_pasid().
>>>>>> + * Return the pasid entry pointer if the entry is found or NULL if no
>>>>>> + * entry found.
>>>>>> */
>>>>>> -void intel_pasid_tear_down_entry(struct intel_iommu *iommu, struct
>>>>>> device *dev,
>>>>>> - u32 pasid, u32 flags)
>>>>>> +struct pasid_entry *
>>>>>> +intel_pasid_tear_down_entry(struct intel_iommu *iommu, struct device
>>>>>> *dev,
>>>>>> + u32 pasid, u32 flags)
>>>>>> {
>>>>>> struct pasid_entry *pte;
>>>>>> u16 did, pgtt;
>>>>>> @@ -250,7 +253,7 @@ void intel_pasid_tear_down_entry(struct
>>>>>> intel_iommu *iommu, struct device *dev,
>>>>>> pte = intel_pasid_get_entry(dev, pasid);
>>>>>> if (WARN_ON(!pte) || !pasid_pte_is_present(pte)) {
>>>>>> spin_unlock(&iommu->lock);
>>>>>> - return;
>>>>>> + goto out;
>>>>>
>>>>> The pasid table entry is protected by iommu->lock. It's not reasonable
>>>>> to return the pte pointer which is beyond the lock protected range.
>>>>
>>>> Per my understanding, the iommu->lock protects the content of the entry,
>>>> so the modifications to the entry need to hold it. While, it looks not
>>>> necessary to protect the pasid entry pointer itself. The pasid table
>>>> should
>>>> exist during device probe and release. is it?
>>>
>>> The pattern of the code that modifies a pasid table entry is,
>>>
>>> spin_lock(&iommu->lock);
>>> pte = intel_pasid_get_entry(dev, pasid);
>>> ... modify the pasid table entry ...
>>> spin_unlock(&iommu->lock);
>>>
>>> Returning the pte pointer to the caller introduces a potential race
>>> condition. If the caller subsequently modifies the pte without re-
>>> acquiring the spin lock, there's a risk of data corruption or
>>> inconsistencies.
>>
>> it appears that we are on the same page about if pte pointer needs to be
>> protected or not. And I agree the modifications to the pte should be
>> protected by iommu->lock. If so, will documenting that the caller must hold
>> iommu->lock if is tries to modify the content of pte work? Also, it might
>> be helpful to add lockdep to make sure all the modifications of pte entry
>> are under protection.
>
> People will soon forget about this lock and may modify the returned pte
> pointer without locking, introducing a race condition silently.
>
>> Or any suggestion from you given a path that needs to get pte first, check
>> if it exists and then call intel_pasid_tear_down_entry(). For example the
>> intel_pasid_setup_first_level() [1], in my series, I need to call the
>> unlock iommu->lock and call intel_pasid_tear_down_entry() and then lock
>> iommu->lock and do more modifications on the pasid entry. It would invoke
>> the intel_pasid_get_entry() twice if no change to
>> intel_pasid_tear_down_entry().
>
> There is no need to check the present of a pte entry before calling into
> intel_pasid_tear_down_entry(). The helper will return directly if the
> pte is not present:
>
> spin_lock(&iommu->lock);
> pte = intel_pasid_get_entry(dev, pasid);
> if (WARN_ON(!pte) || !pasid_pte_is_present(pte)) {
> spin_unlock(&iommu->lock);
> return;
> }
>
> Does it work for you?
This is not I'm talking about. My intention is to avoid duplicated
intel_pasid_get_entry() call when calling intel_pasid_tear_down_entry() in
intel_pasid_setup_first_level(). Both the two functions call the
intel_pasid_get_entry() to get pte pointer. So I think it might be good to
save one of them.
--
Regards,
Yi Liu
next prev parent reply other threads:[~2024-10-22 9:34 UTC|newest]
Thread overview: 29+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-10-18 5:53 [PATCH v3 0/9] Make set_dev_pasid op supporting domain replacement Yi Liu
2024-10-18 5:53 ` [PATCH v3 1/9] iommu: Pass old domain to set_dev_pasid op Yi Liu
2024-10-21 5:55 ` Baolu Lu
2024-10-22 5:12 ` Nicolin Chen
2024-10-18 5:53 ` [PATCH v3 2/9] iommu/vt-d: Move intel_drain_pasid_prq() into intel_pasid_tear_down_entry() Yi Liu
2024-10-21 5:58 ` Baolu Lu
2024-10-18 5:53 ` [PATCH v3 3/9] iommu/vt-d: Let intel_pasid_tear_down_entry() return pasid entry Yi Liu
2024-10-21 6:13 ` Baolu Lu
2024-10-21 6:35 ` Yi Liu
2024-10-21 6:59 ` Baolu Lu
2024-10-21 7:24 ` Yi Liu
2024-10-22 9:23 ` Baolu Lu
2024-10-22 9:38 ` Yi Liu [this message]
2024-10-22 11:23 ` Baolu Lu
2024-10-22 13:25 ` Yi Liu
2024-10-23 1:10 ` Baolu Lu
2024-10-18 5:53 ` [PATCH v3 4/9] iommu/vt-d: Make pasid setup helpers support modifying present " Yi Liu
2024-10-18 5:53 ` [PATCH v3 5/9] iommu/vt-d: Rename prepare_domain_attach_device() Yi Liu
2024-10-21 6:18 ` Baolu Lu
2024-10-21 6:36 ` Yi Liu
2024-10-18 5:53 ` [PATCH v3 6/9] iommu/vt-d: Make intel_iommu_set_dev_pasid() to handle domain replacement Yi Liu
2024-10-18 5:54 ` [PATCH v3 7/9] iommu/vt-d: Add set_dev_pasid callback for nested domain Yi Liu
2024-10-18 5:54 ` [PATCH v3 8/9] iommu/arm-smmu-v3: Make set_dev_pasid() op support replace Yi Liu
2024-10-22 5:25 ` Nicolin Chen
2024-10-22 6:07 ` Yi Liu
2024-10-18 5:54 ` [PATCH v3 9/9] iommu: Make set_dev_pasid op support domain replacement Yi Liu
2024-10-21 6:27 ` Baolu Lu
2024-10-21 6:40 ` Yi Liu
2024-10-21 10:50 ` Vasant Hegde
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=fe88f071-0d06-4838-9ce6-a5bcccf10163@intel.com \
--to=yi.l.liu@intel.com \
--cc=alex.williamson@redhat.com \
--cc=baolu.lu@linux.intel.com \
--cc=chao.p.peng@linux.intel.com \
--cc=eric.auger@redhat.com \
--cc=iommu@lists.linux.dev \
--cc=jgg@nvidia.com \
--cc=joro@8bytes.org \
--cc=kevin.tian@intel.com \
--cc=kvm@vger.kernel.org \
--cc=nicolinc@nvidia.com \
--cc=vasant.hegde@amd.com \
--cc=will@kernel.org \
--cc=zhenzhong.duan@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox