From: Nadav Amit <namit@vmware.com>
To: Robin Murphy <robin.murphy@arm.com>
Cc: Will Deacon <will@kernel.org>,
"linux-kernel@vger.kernel.org" <linux-kernel@vger.kernel.org>,
"iommu@lists.linux-foundation.org"
<iommu@lists.linux-foundation.org>,
Jiajun Cao <caojiajun@vmware.com>
Subject: Re: [PATCH v5 5/7] iommu/amd: Tailored gather logic for AMD
Date: Tue, 13 Jul 2021 21:52:45 +0000 [thread overview]
Message-ID: <80134D97-D8CD-40ED-88C3-4FAD620E4F12@vmware.com> (raw)
In-Reply-To: <d9818e0b-681c-29a0-d9cd-e2f4dc963eb4@arm.com>
> On Jul 13, 2021, at 11:40 AM, Robin Murphy <robin.murphy@arm.com> wrote:
>
> On 2021-07-13 10:41, Nadav Amit wrote:
>> From: Nadav Amit <namit@vmware.com>
>> AMD's IOMMU can flush efficiently (i.e., in a single flush) any range.
>> This is in contrast, for instnace, to Intel IOMMUs that have a limit on
>> the number of pages that can be flushed in a single flush. In addition,
>> AMD's IOMMU do not care about the page-size, so changes of the page size
>> do not need to trigger a TLB flush.
>> So in most cases, a TLB flush due to disjoint range is not needed for
>> AMD. Yet, vIOMMUs require the hypervisor to synchronize the virtualized
>> IOMMU's PTEs with the physical ones. This process induce overheads, so
>> it is better not to cause unnecessary flushes, i.e., flushes of PTEs
>> that were not modified.
>> Implement and use amd_iommu_iotlb_gather_add_page() and use it instead
>> of the generic iommu_iotlb_gather_add_page(). Ignore disjoint regions
>> unless "non-present cache" feature is reported by the IOMMU
>> capabilities, as this is an indication we are running on a physical
>> IOMMU. A similar indication is used by VT-d (see "caching mode"). The
>> new logic retains the same flushing behavior that we had before the
>> introduction of page-selective IOTLB flushes for AMD.
>> On virtualized environments, check if the newly flushed region and the
>> gathered one are disjoint and flush if it is.
>> Cc: Joerg Roedel <joro@8bytes.org>
>> Cc: Will Deacon <will@kernel.org>
>> Cc: Jiajun Cao <caojiajun@vmware.com>
>> Cc: Lu Baolu <baolu.lu@linux.intel.com>
>> Cc: iommu@lists.linux-foundation.org
>> Cc: linux-kernel@vger.kernel.org>
>> Cc: Robin Murphy <robin.murphy@arm.com>
>> Signed-off-by: Nadav Amit <namit@vmware.com>
>> ---
>> drivers/iommu/amd/iommu.c | 23 ++++++++++++++++++++++-
>> 1 file changed, 22 insertions(+), 1 deletion(-)
>> diff --git a/drivers/iommu/amd/iommu.c b/drivers/iommu/amd/iommu.c
>> index bfae3928b98f..cc55c4c6a355 100644
>> --- a/drivers/iommu/amd/iommu.c
>> +++ b/drivers/iommu/amd/iommu.c
>> @@ -2048,6 +2048,27 @@ static int amd_iommu_map(struct iommu_domain *dom, unsigned long iova,
>> return ret;
>> }
>> +static void amd_iommu_iotlb_gather_add_page(struct iommu_domain *domain,
>> + struct iommu_iotlb_gather *gather,
>> + unsigned long iova, size_t size)
>> +{
>> + /*
>> + * AMD's IOMMU can flush as many pages as necessary in a single flush.
>> + * Unless we run in a virtual machine, which can be inferred according
>> + * to whether "non-present cache" is on, it is probably best to prefer
>> + * (potentially) too extensive TLB flushing (i.e., more misses) over
>> + * mutliple TLB flushes (i.e., more flushes). For virtual machines the
>> + * hypervisor needs to synchronize the host IOMMU PTEs with those of
>> + * the guest, and the trade-off is different: unnecessary TLB flushes
>> + * should be avoided.
>> + */
>> + if (amd_iommu_np_cache && gather->end != 0 &&
>
> iommu_iotlb_gather_is_disjoint() is also checking "gather->end != 0", so I don't think we need both. Strictly it's only necessary here since the other call from iommu_iotlb_gather_add_page() equivalently asserts that the gather is already non-empty via its gather->pgsize check, but one could argue it either way and I don't have a hugely strong preference.
You are correct (even if the compiler would have eliminated the redundancy).
I will remove the redundant check.
>
> Otherwise, I love how neat this has all ended up, thanks for persevering!
Thank you for the thorough review!
Regards,
Nadav
_______________________________________________
iommu mailing list
iommu@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/iommu
next prev parent reply other threads:[~2021-07-13 21:52 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-07-13 9:41 [PATCH v5 0/7] iommu/amd: Enable page-selective flushes Nadav Amit
2021-07-13 9:41 ` [PATCH v5 1/7] iommu/amd: Selective flush on unmap Nadav Amit
2021-07-13 9:41 ` [PATCH v5 2/7] iommu/amd: Do not use flush-queue when NpCache is on Nadav Amit
2021-07-13 9:41 ` [PATCH v5 3/7] iommu: Improve iommu_iotlb_gather helpers Nadav Amit
2021-07-13 9:41 ` [PATCH v5 4/7] iommu: Factor iommu_iotlb_gather_is_disjoint() out Nadav Amit
2021-07-13 18:25 ` Robin Murphy
2021-07-13 9:41 ` [PATCH v5 5/7] iommu/amd: Tailored gather logic for AMD Nadav Amit
2021-07-13 18:40 ` Robin Murphy
2021-07-13 21:52 ` Nadav Amit [this message]
2021-07-13 9:41 ` [PATCH v5 6/7] iommu/amd: Sync once for scatter-gather operations Nadav Amit
2021-07-13 9:41 ` [PATCH v5 7/7] iommu/amd: Use only natural aligned flushes in a VM Nadav Amit
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=80134D97-D8CD-40ED-88C3-4FAD620E4F12@vmware.com \
--to=namit@vmware.com \
--cc=caojiajun@vmware.com \
--cc=iommu@lists.linux-foundation.org \
--cc=linux-kernel@vger.kernel.org \
--cc=robin.murphy@arm.com \
--cc=will@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox