* Re: [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap
2019-06-24 9:18 ` [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap Peter Xu
@ 2019-06-24 10:09 ` Auger Eric
2019-06-24 11:10 ` Peter Xu
2019-06-24 10:12 ` Paolo Bonzini
` (3 subsequent siblings)
4 siblings, 1 reply; 14+ messages in thread
From: Auger Eric @ 2019-06-24 10:09 UTC (permalink / raw)
To: Peter Xu, qemu-devel
Cc: Paolo Bonzini, Jason Wang, Yan Zhao, Michael S . Tsirkin
Hi Peter,
On 6/24/19 11:18 AM, Peter Xu wrote:
> This is an replacement work of Yan Zhao's patch:
>
> https://www.mail-archive.com/qemu-devel@nongnu.org/msg625340.html
>
> vtd_address_space_unmap() will do proper page mask alignment to make
> sure each IOTLB message will have correct masks for notification
> messages (2^N-1), but sometimes it can be expanded to even supercede
> the registered range. That could lead to unexpected UNMAP of already
> mapped regions in some other notifiers.
>
> Instead of doing mindless expension of the start address and address
> mask, we split the range into smaller ones and guarantee that each
> small range will have correct masks (2^N-1) and at the same time we
> should also try our best to generate as less IOTLB messages as
> possible.
>
> Reported-by: Yan Zhao <yan.y.zhao@intel.com>
> Signed-off-by: Peter Xu <peterx@redhat.com>
> ---
> hw/i386/intel_iommu.c | 67 ++++++++++++++++++++++++++-----------------
> 1 file changed, 41 insertions(+), 26 deletions(-)
>
> diff --git a/hw/i386/intel_iommu.c b/hw/i386/intel_iommu.c
> index 719ce19ab3..de86f53b4e 100644
> --- a/hw/i386/intel_iommu.c
> +++ b/hw/i386/intel_iommu.c
> @@ -3363,11 +3363,28 @@ VTDAddressSpace *vtd_find_add_as(IntelIOMMUState *s, PCIBus *bus, int devfn)
> return vtd_dev_as;
> }
>
> +static uint64_t get_naturally_aligned_size(uint64_t start,
> + uint64_t size, int gaw)
> +{
> + uint64_t max_mask = 1ULL << gaw;
> + uint64_t alignment = start ? start & -start : max_mask;
> +
> + alignment = MIN(alignment, max_mask);
> + size = MIN(size, max_mask);
this does not not prevent from invalidating beyond gaw if start != 0, right?
> +
> + if (alignment <= size) {
> + /* Increase the alignment of start */
I don't really get this comment
> + return alignment;
> + } else {
> + /* Find the largest page mask from size */
> + return 1ULL << (63 - clz64(size));
> + }> +}
> +
> /* Unmap the whole range in the notifier's scope. */
> static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> {
> - IOMMUTLBEntry entry;
> - hwaddr size;
> + hwaddr size, remain;
> hwaddr start = n->start;
> hwaddr end = n->end;
> IntelIOMMUState *s = as->iommu_state;
> @@ -3388,39 +3405,37 @@ static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> }
>
> assert(start <= end);
> - size = end - start;
> + size = remain = end - start + 1;
>
> - if (ctpop64(size) != 1) {
> - /*
> - * This size cannot format a correct mask. Let's enlarge it to
> - * suite the minimum available mask.
> - */
> - int n = 64 - clz64(size);
> - if (n > s->aw_bits) {
> - /* should not happen, but in case it happens, limit it */
> - n = s->aw_bits;
> - }
> - size = 1ULL << n;
> + while (remain >= VTD_PAGE_SIZE) {
Can't we stop as soon as entry.iova exceeds gaw as well?
> + IOMMUTLBEntry entry;
> + uint64_t mask = get_naturally_aligned_size(start, remain, s->aw_bits);
> +
> + assert(mask);
> +
> + entry.iova = start;
> + entry.addr_mask = mask - 1;
> + entry.target_as = &address_space_memory;
> + entry.perm = IOMMU_NONE;
> + /* This field is meaningless for unmap */
> + entry.translated_addr = 0;
> +
> + memory_region_notify_one(n, &entry);
> +
> + start += mask;
> + remain -= mask;
> }
>
> - entry.target_as = &address_space_memory;
> - /* Adjust iova for the size */
> - entry.iova = n->start & ~(size - 1);
> - /* This field is meaningless for unmap */
> - entry.translated_addr = 0;
> - entry.perm = IOMMU_NONE;
> - entry.addr_mask = size - 1;
> + assert(!remain);
>
> trace_vtd_as_unmap_whole(pci_bus_num(as->bus),
> VTD_PCI_SLOT(as->devfn),
> VTD_PCI_FUNC(as->devfn),
> - entry.iova, size);
> + n->start, size);
>
> - map.iova = entry.iova;
> - map.size = entry.addr_mask;
> + map.iova = n->start;
> + map.size = size;
> iova_tree_remove(as->iova_tree, &map);
> -
> - memory_region_notify_one(n, &entry);
> }
>
> static void vtd_address_space_unmap_all(IntelIOMMUState *s)
>
Thanks
Eric
^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap
2019-06-24 10:09 ` Auger Eric
@ 2019-06-24 11:10 ` Peter Xu
2019-06-24 12:48 ` Auger Eric
0 siblings, 1 reply; 14+ messages in thread
From: Peter Xu @ 2019-06-24 11:10 UTC (permalink / raw)
To: Auger Eric
Cc: Paolo Bonzini, Jason Wang, Yan Zhao, qemu-devel,
Michael S . Tsirkin
On Mon, Jun 24, 2019 at 12:09:48PM +0200, Auger Eric wrote:
> Hi Peter,
>
> On 6/24/19 11:18 AM, Peter Xu wrote:
> > This is an replacement work of Yan Zhao's patch:
> >
> > https://www.mail-archive.com/qemu-devel@nongnu.org/msg625340.html
> >
> > vtd_address_space_unmap() will do proper page mask alignment to make
> > sure each IOTLB message will have correct masks for notification
> > messages (2^N-1), but sometimes it can be expanded to even supercede
> > the registered range. That could lead to unexpected UNMAP of already
> > mapped regions in some other notifiers.
> >
> > Instead of doing mindless expension of the start address and address
> > mask, we split the range into smaller ones and guarantee that each
> > small range will have correct masks (2^N-1) and at the same time we
> > should also try our best to generate as less IOTLB messages as
> > possible.
> >
> > Reported-by: Yan Zhao <yan.y.zhao@intel.com>
> > Signed-off-by: Peter Xu <peterx@redhat.com>
> > ---
> > hw/i386/intel_iommu.c | 67 ++++++++++++++++++++++++++-----------------
> > 1 file changed, 41 insertions(+), 26 deletions(-)
> >
> > diff --git a/hw/i386/intel_iommu.c b/hw/i386/intel_iommu.c
> > index 719ce19ab3..de86f53b4e 100644
> > --- a/hw/i386/intel_iommu.c
> > +++ b/hw/i386/intel_iommu.c
> > @@ -3363,11 +3363,28 @@ VTDAddressSpace *vtd_find_add_as(IntelIOMMUState *s, PCIBus *bus, int devfn)
> > return vtd_dev_as;
> > }
> >
> > +static uint64_t get_naturally_aligned_size(uint64_t start,
> > + uint64_t size, int gaw)
> > +{
> > + uint64_t max_mask = 1ULL << gaw;
> > + uint64_t alignment = start ? start & -start : max_mask;
> > +
> > + alignment = MIN(alignment, max_mask);
> > + size = MIN(size, max_mask);
> this does not not prevent from invalidating beyond gaw if start != 0, right?
Yes. But at the start of vtd_address_space_unmap(), we have:
if (end > VTD_ADDRESS_SIZE(s->aw_bits) - 1) {
/*
* Don't need to unmap regions that is bigger than the whole
* VT-d supported address space size
*/
end = VTD_ADDRESS_SIZE(s->aw_bits) - 1;
}
So we don't need to worry about (start+size) exceeding GAW?
[1]
> > +
> > + if (alignment <= size) {
> > + /* Increase the alignment of start */
> I don't really get this comment
This comment comes from Paolo, but I'll try to explain - it tries to
mean that this "alignment" will be used as an increasement to "start"
variable, so finally variable "start" will align with larger mask
size.
Better comments welcomed... :)
> > + return alignment;
> > + } else {
> > + /* Find the largest page mask from size */
> > + return 1ULL << (63 - clz64(size));
> > + }> +}
> > +
> > /* Unmap the whole range in the notifier's scope. */
> > static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> > {
> > - IOMMUTLBEntry entry;
> > - hwaddr size;
> > + hwaddr size, remain;
> > hwaddr start = n->start;
> > hwaddr end = n->end;
> > IntelIOMMUState *s = as->iommu_state;
> > @@ -3388,39 +3405,37 @@ static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> > }
> >
> > assert(start <= end);
> > - size = end - start;
> > + size = remain = end - start + 1;
> >
> > - if (ctpop64(size) != 1) {
> > - /*
> > - * This size cannot format a correct mask. Let's enlarge it to
> > - * suite the minimum available mask.
> > - */
> > - int n = 64 - clz64(size);
> > - if (n > s->aw_bits) {
> > - /* should not happen, but in case it happens, limit it */
> > - n = s->aw_bits;
> > - }
> > - size = 1ULL << n;
> > + while (remain >= VTD_PAGE_SIZE) {
> Can't we stop as soon as entry.iova exceeds gaw as well?
As explained at [1], I think we've already checked it.
Thanks,
--
Peter Xu
^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap
2019-06-24 11:10 ` Peter Xu
@ 2019-06-24 12:48 ` Auger Eric
2019-06-24 13:21 ` Paolo Bonzini
0 siblings, 1 reply; 14+ messages in thread
From: Auger Eric @ 2019-06-24 12:48 UTC (permalink / raw)
To: Peter Xu
Cc: Paolo Bonzini, Jason Wang, Yan Zhao, qemu-devel,
Michael S . Tsirkin
On 6/24/19 1:10 PM, Peter Xu wrote:
> On Mon, Jun 24, 2019 at 12:09:48PM +0200, Auger Eric wrote:
>> Hi Peter,
>>
>> On 6/24/19 11:18 AM, Peter Xu wrote:
>>> This is an replacement work of Yan Zhao's patch:
>>>
>>> https://www.mail-archive.com/qemu-devel@nongnu.org/msg625340.html
>>>
>>> vtd_address_space_unmap() will do proper page mask alignment to make
>>> sure each IOTLB message will have correct masks for notification
>>> messages (2^N-1), but sometimes it can be expanded to even supercede
>>> the registered range. That could lead to unexpected UNMAP of already
>>> mapped regions in some other notifiers.
>>>
>>> Instead of doing mindless expension of the start address and address
>>> mask, we split the range into smaller ones and guarantee that each
>>> small range will have correct masks (2^N-1) and at the same time we
>>> should also try our best to generate as less IOTLB messages as
>>> possible.
>>>
>>> Reported-by: Yan Zhao <yan.y.zhao@intel.com>
>>> Signed-off-by: Peter Xu <peterx@redhat.com>
>>> ---
>>> hw/i386/intel_iommu.c | 67 ++++++++++++++++++++++++++-----------------
>>> 1 file changed, 41 insertions(+), 26 deletions(-)
>>>
>>> diff --git a/hw/i386/intel_iommu.c b/hw/i386/intel_iommu.c
>>> index 719ce19ab3..de86f53b4e 100644
>>> --- a/hw/i386/intel_iommu.c
>>> +++ b/hw/i386/intel_iommu.c
>>> @@ -3363,11 +3363,28 @@ VTDAddressSpace *vtd_find_add_as(IntelIOMMUState *s, PCIBus *bus, int devfn)
>>> return vtd_dev_as;
>>> }
>>>
>>> +static uint64_t get_naturally_aligned_size(uint64_t start,
>>> + uint64_t size, int gaw)
>>> +{
>>> + uint64_t max_mask = 1ULL << gaw;
>>> + uint64_t alignment = start ? start & -start : max_mask;
>>> +
>>> + alignment = MIN(alignment, max_mask);
>>> + size = MIN(size, max_mask);
>> this does not not prevent from invalidating beyond gaw if start != 0, right?
>
> Yes. But at the start of vtd_address_space_unmap(), we have:
>
> if (end > VTD_ADDRESS_SIZE(s->aw_bits) - 1) {
> /*
> * Don't need to unmap regions that is bigger than the whole
> * VT-d supported address space size
> */
> end = VTD_ADDRESS_SIZE(s->aw_bits) - 1;
> }
>
> So we don't need to worry about (start+size) exceeding GAW?
Hum yes. Reviewed the previous patch with blinkers ...
>
> [1]
>
>>> +
>>> + if (alignment <= size) {
>>> + /* Increase the alignment of start */
>> I don't really get this comment
>
> This comment comes from Paolo, but I'll try to explain - it tries to
> mean that this "alignment" will be used as an increasement to "start"
> variable, so finally variable "start" will align with larger mask
> size.
>
> Better comments welcomed... :)
smallest page mask from @start or gaw?
>
>>> + return alignment;
>>> + } else {
>>> + /* Find the largest page mask from size */
>>> + return 1ULL << (63 - clz64(size));
>>> + }> +}
>>> +
>>> /* Unmap the whole range in the notifier's scope. */
>>> static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
>>> {
>>> - IOMMUTLBEntry entry;
>>> - hwaddr size;
>>> + hwaddr size, remain;
>>> hwaddr start = n->start;
>>> hwaddr end = n->end;
>>> IntelIOMMUState *s = as->iommu_state;
>>> @@ -3388,39 +3405,37 @@ static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
>>> }
>>>
>>> assert(start <= end);
>>> - size = end - start;
>>> + size = remain = end - start + 1;
>>>
>>> - if (ctpop64(size) != 1) {
>>> - /*
>>> - * This size cannot format a correct mask. Let's enlarge it to
>>> - * suite the minimum available mask.
>>> - */
>>> - int n = 64 - clz64(size);
>>> - if (n > s->aw_bits) {
>>> - /* should not happen, but in case it happens, limit it */
>>> - n = s->aw_bits;
>>> - }
>>> - size = 1ULL << n;
>>> + while (remain >= VTD_PAGE_SIZE) {
>> Can't we stop as soon as entry.iova exceeds gaw as well?
>
> As explained at [1], I think we've already checked it.
OK
Thanks
Eric
>
> Thanks,
>
^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap
2019-06-24 12:48 ` Auger Eric
@ 2019-06-24 13:21 ` Paolo Bonzini
0 siblings, 0 replies; 14+ messages in thread
From: Paolo Bonzini @ 2019-06-24 13:21 UTC (permalink / raw)
To: Auger Eric, Peter Xu
Cc: Jason Wang, Yan Zhao, qemu-devel, Michael S . Tsirkin
On 24/06/19 14:48, Auger Eric wrote:
>>>> + if (alignment <= size) {
>>>> + /* Increase the alignment of start */
>>> I don't really get this comment
>> This comment comes from Paolo, but I'll try to explain - it tries to
>> mean that this "alignment" will be used as an increasement to "start"
>> variable, so finally variable "start" will align with larger mask
>> size.
>>
>> Better comments welcomed... :)
> smallest page mask from @start or gaw?
What it means is that there will be "more 0 bits" at the beginning of
start. Perhaps "On the next iteration start will be aligned to a bigger
power of two"? I can do this when applying.
Paolo
^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap
2019-06-24 9:18 ` [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap Peter Xu
2019-06-24 10:09 ` Auger Eric
@ 2019-06-24 10:12 ` Paolo Bonzini
2019-06-25 3:02 ` Yan Zhao
` (2 subsequent siblings)
4 siblings, 0 replies; 14+ messages in thread
From: Paolo Bonzini @ 2019-06-24 10:12 UTC (permalink / raw)
To: Peter Xu, qemu-devel
Cc: Auger Eric, Jason Wang, Yan Zhao, Michael S . Tsirkin
On 24/06/19 11:18, Peter Xu wrote:
> This is an replacement work of Yan Zhao's patch:
>
> https://www.mail-archive.com/qemu-devel@nongnu.org/msg625340.html
>
> vtd_address_space_unmap() will do proper page mask alignment to make
> sure each IOTLB message will have correct masks for notification
> messages (2^N-1), but sometimes it can be expanded to even supercede
> the registered range. That could lead to unexpected UNMAP of already
> mapped regions in some other notifiers.
>
> Instead of doing mindless expension of the start address and address
> mask, we split the range into smaller ones and guarantee that each
> small range will have correct masks (2^N-1) and at the same time we
> should also try our best to generate as less IOTLB messages as
> possible.
>
> Reported-by: Yan Zhao <yan.y.zhao@intel.com>
> Signed-off-by: Peter Xu <peterx@redhat.com>
> ---
> hw/i386/intel_iommu.c | 67 ++++++++++++++++++++++++++-----------------
> 1 file changed, 41 insertions(+), 26 deletions(-)
>
> diff --git a/hw/i386/intel_iommu.c b/hw/i386/intel_iommu.c
> index 719ce19ab3..de86f53b4e 100644
> --- a/hw/i386/intel_iommu.c
> +++ b/hw/i386/intel_iommu.c
> @@ -3363,11 +3363,28 @@ VTDAddressSpace *vtd_find_add_as(IntelIOMMUState *s, PCIBus *bus, int devfn)
> return vtd_dev_as;
> }
>
> +static uint64_t get_naturally_aligned_size(uint64_t start,
> + uint64_t size, int gaw)
> +{
> + uint64_t max_mask = 1ULL << gaw;
> + uint64_t alignment = start ? start & -start : max_mask;
> +
> + alignment = MIN(alignment, max_mask);
> + size = MIN(size, max_mask);
> +
> + if (alignment <= size) {
> + /* Increase the alignment of start */
> + return alignment;
> + } else {
> + /* Find the largest page mask from size */
> + return 1ULL << (63 - clz64(size));
> + }
> +}
> +
> /* Unmap the whole range in the notifier's scope. */
> static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> {
> - IOMMUTLBEntry entry;
> - hwaddr size;
> + hwaddr size, remain;
> hwaddr start = n->start;
> hwaddr end = n->end;
> IntelIOMMUState *s = as->iommu_state;
> @@ -3388,39 +3405,37 @@ static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> }
>
> assert(start <= end);
> - size = end - start;
> + size = remain = end - start + 1;
>
> - if (ctpop64(size) != 1) {
> - /*
> - * This size cannot format a correct mask. Let's enlarge it to
> - * suite the minimum available mask.
> - */
> - int n = 64 - clz64(size);
> - if (n > s->aw_bits) {
> - /* should not happen, but in case it happens, limit it */
> - n = s->aw_bits;
> - }
> - size = 1ULL << n;
> + while (remain >= VTD_PAGE_SIZE) {
> + IOMMUTLBEntry entry;
> + uint64_t mask = get_naturally_aligned_size(start, remain, s->aw_bits);
> +
> + assert(mask);
> +
> + entry.iova = start;
> + entry.addr_mask = mask - 1;
> + entry.target_as = &address_space_memory;
> + entry.perm = IOMMU_NONE;
> + /* This field is meaningless for unmap */
> + entry.translated_addr = 0;
> +
> + memory_region_notify_one(n, &entry);
> +
> + start += mask;
> + remain -= mask;
> }
>
> - entry.target_as = &address_space_memory;
> - /* Adjust iova for the size */
> - entry.iova = n->start & ~(size - 1);
> - /* This field is meaningless for unmap */
> - entry.translated_addr = 0;
> - entry.perm = IOMMU_NONE;
> - entry.addr_mask = size - 1;
> + assert(!remain);
>
> trace_vtd_as_unmap_whole(pci_bus_num(as->bus),
> VTD_PCI_SLOT(as->devfn),
> VTD_PCI_FUNC(as->devfn),
> - entry.iova, size);
> + n->start, size);
>
> - map.iova = entry.iova;
> - map.size = entry.addr_mask;
> + map.iova = n->start;
> + map.size = size;
> iova_tree_remove(as->iova_tree, &map);
> -
> - memory_region_notify_one(n, &entry);
> }
>
> static void vtd_address_space_unmap_all(IntelIOMMUState *s)
>
Looks good, ignore my previous message.
Paolo
^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap
2019-06-24 9:18 ` [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap Peter Xu
2019-06-24 10:09 ` Auger Eric
2019-06-24 10:12 ` Paolo Bonzini
@ 2019-06-25 3:02 ` Yan Zhao
2019-06-25 7:00 ` Auger Eric
2019-07-04 5:45 ` Jason Wang
4 siblings, 0 replies; 14+ messages in thread
From: Yan Zhao @ 2019-06-25 3:02 UTC (permalink / raw)
To: Peter Xu
Cc: Paolo Bonzini, Jason Wang, qemu-devel@nongnu.org, Auger Eric,
Michael S . Tsirkin
Tested-by: Yan Zhao <yan.y.zhao@intel.com>
On Mon, Jun 24, 2019 at 05:18:11PM +0800, Peter Xu wrote:
> This is an replacement work of Yan Zhao's patch:
>
> https://www.mail-archive.com/qemu-devel@nongnu.org/msg625340.html
>
> vtd_address_space_unmap() will do proper page mask alignment to make
> sure each IOTLB message will have correct masks for notification
> messages (2^N-1), but sometimes it can be expanded to even supercede
> the registered range. That could lead to unexpected UNMAP of already
> mapped regions in some other notifiers.
>
> Instead of doing mindless expension of the start address and address
> mask, we split the range into smaller ones and guarantee that each
> small range will have correct masks (2^N-1) and at the same time we
> should also try our best to generate as less IOTLB messages as
> possible.
>
> Reported-by: Yan Zhao <yan.y.zhao@intel.com>
> Signed-off-by: Peter Xu <peterx@redhat.com>
> ---
> hw/i386/intel_iommu.c | 67 ++++++++++++++++++++++++++-----------------
> 1 file changed, 41 insertions(+), 26 deletions(-)
>
> diff --git a/hw/i386/intel_iommu.c b/hw/i386/intel_iommu.c
> index 719ce19ab3..de86f53b4e 100644
> --- a/hw/i386/intel_iommu.c
> +++ b/hw/i386/intel_iommu.c
> @@ -3363,11 +3363,28 @@ VTDAddressSpace *vtd_find_add_as(IntelIOMMUState *s, PCIBus *bus, int devfn)
> return vtd_dev_as;
> }
>
> +static uint64_t get_naturally_aligned_size(uint64_t start,
> + uint64_t size, int gaw)
> +{
> + uint64_t max_mask = 1ULL << gaw;
> + uint64_t alignment = start ? start & -start : max_mask;
> +
> + alignment = MIN(alignment, max_mask);
> + size = MIN(size, max_mask);
> +
> + if (alignment <= size) {
> + /* Increase the alignment of start */
> + return alignment;
> + } else {
> + /* Find the largest page mask from size */
> + return 1ULL << (63 - clz64(size));
> + }
> +}
> +
> /* Unmap the whole range in the notifier's scope. */
> static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> {
> - IOMMUTLBEntry entry;
> - hwaddr size;
> + hwaddr size, remain;
> hwaddr start = n->start;
> hwaddr end = n->end;
> IntelIOMMUState *s = as->iommu_state;
> @@ -3388,39 +3405,37 @@ static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> }
>
> assert(start <= end);
> - size = end - start;
> + size = remain = end - start + 1;
>
> - if (ctpop64(size) != 1) {
> - /*
> - * This size cannot format a correct mask. Let's enlarge it to
> - * suite the minimum available mask.
> - */
> - int n = 64 - clz64(size);
> - if (n > s->aw_bits) {
> - /* should not happen, but in case it happens, limit it */
> - n = s->aw_bits;
> - }
> - size = 1ULL << n;
> + while (remain >= VTD_PAGE_SIZE) {
> + IOMMUTLBEntry entry;
> + uint64_t mask = get_naturally_aligned_size(start, remain, s->aw_bits);
> +
> + assert(mask);
> +
> + entry.iova = start;
> + entry.addr_mask = mask - 1;
> + entry.target_as = &address_space_memory;
> + entry.perm = IOMMU_NONE;
> + /* This field is meaningless for unmap */
> + entry.translated_addr = 0;
> +
> + memory_region_notify_one(n, &entry);
> +
> + start += mask;
> + remain -= mask;
> }
>
> - entry.target_as = &address_space_memory;
> - /* Adjust iova for the size */
> - entry.iova = n->start & ~(size - 1);
> - /* This field is meaningless for unmap */
> - entry.translated_addr = 0;
> - entry.perm = IOMMU_NONE;
> - entry.addr_mask = size - 1;
> + assert(!remain);
>
> trace_vtd_as_unmap_whole(pci_bus_num(as->bus),
> VTD_PCI_SLOT(as->devfn),
> VTD_PCI_FUNC(as->devfn),
> - entry.iova, size);
> + n->start, size);
>
> - map.iova = entry.iova;
> - map.size = entry.addr_mask;
> + map.iova = n->start;
> + map.size = size;
> iova_tree_remove(as->iova_tree, &map);
> -
> - memory_region_notify_one(n, &entry);
> }
>
> static void vtd_address_space_unmap_all(IntelIOMMUState *s)
> --
> 2.21.0
>
^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap
2019-06-24 9:18 ` [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap Peter Xu
` (2 preceding siblings ...)
2019-06-25 3:02 ` Yan Zhao
@ 2019-06-25 7:00 ` Auger Eric
2019-07-04 5:45 ` Jason Wang
4 siblings, 0 replies; 14+ messages in thread
From: Auger Eric @ 2019-06-25 7:00 UTC (permalink / raw)
To: Peter Xu, qemu-devel
Cc: Paolo Bonzini, Jason Wang, Yan Zhao, Michael S . Tsirkin
Hi Peter,
On 6/24/19 11:18 AM, Peter Xu wrote:
> This is an replacement work of Yan Zhao's patch:
>
> https://www.mail-archive.com/qemu-devel@nongnu.org/msg625340.html
>
> vtd_address_space_unmap() will do proper page mask alignment to make
> sure each IOTLB message will have correct masks for notification
> messages (2^N-1), but sometimes it can be expanded to even supercede
> the registered range. That could lead to unexpected UNMAP of already
> mapped regions in some other notifiers.
>
> Instead of doing mindless expension of the start address and address
> mask, we split the range into smaller ones and guarantee that each
> small range will have correct masks (2^N-1) and at the same time we
> should also try our best to generate as less IOTLB messages as
> possible.
>
> Reported-by: Yan Zhao <yan.y.zhao@intel.com>
> Signed-off-by: Peter Xu <peterx@redhat.com>
Reviewed-by: Eric Auger <eric.auger@redhat.com>
Thanks
Eric
> ---
> hw/i386/intel_iommu.c | 67 ++++++++++++++++++++++++++-----------------
> 1 file changed, 41 insertions(+), 26 deletions(-)
>
> diff --git a/hw/i386/intel_iommu.c b/hw/i386/intel_iommu.c
> index 719ce19ab3..de86f53b4e 100644
> --- a/hw/i386/intel_iommu.c
> +++ b/hw/i386/intel_iommu.c
> @@ -3363,11 +3363,28 @@ VTDAddressSpace *vtd_find_add_as(IntelIOMMUState *s, PCIBus *bus, int devfn)
> return vtd_dev_as;
> }
>
> +static uint64_t get_naturally_aligned_size(uint64_t start,
> + uint64_t size, int gaw)
> +{
> + uint64_t max_mask = 1ULL << gaw;
> + uint64_t alignment = start ? start & -start : max_mask;
> +
> + alignment = MIN(alignment, max_mask);
> + size = MIN(size, max_mask);
> +
> + if (alignment <= size) {
> + /* Increase the alignment of start */
> + return alignment;
> + } else {
> + /* Find the largest page mask from size */
> + return 1ULL << (63 - clz64(size));
> + }
> +}
> +
> /* Unmap the whole range in the notifier's scope. */
> static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> {
> - IOMMUTLBEntry entry;
> - hwaddr size;
> + hwaddr size, remain;
> hwaddr start = n->start;
> hwaddr end = n->end;
> IntelIOMMUState *s = as->iommu_state;
> @@ -3388,39 +3405,37 @@ static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> }
>
> assert(start <= end);
> - size = end - start;
> + size = remain = end - start + 1;
>
> - if (ctpop64(size) != 1) {
> - /*
> - * This size cannot format a correct mask. Let's enlarge it to
> - * suite the minimum available mask.
> - */
> - int n = 64 - clz64(size);
> - if (n > s->aw_bits) {
> - /* should not happen, but in case it happens, limit it */
> - n = s->aw_bits;
> - }
> - size = 1ULL << n;
> + while (remain >= VTD_PAGE_SIZE) {
> + IOMMUTLBEntry entry;
> + uint64_t mask = get_naturally_aligned_size(start, remain, s->aw_bits);
> +
> + assert(mask);
> +
> + entry.iova = start;
> + entry.addr_mask = mask - 1;
> + entry.target_as = &address_space_memory;
> + entry.perm = IOMMU_NONE;
> + /* This field is meaningless for unmap */
> + entry.translated_addr = 0;
> +
> + memory_region_notify_one(n, &entry);
> +
> + start += mask;
> + remain -= mask;
> }
>
> - entry.target_as = &address_space_memory;
> - /* Adjust iova for the size */
> - entry.iova = n->start & ~(size - 1);
> - /* This field is meaningless for unmap */
> - entry.translated_addr = 0;
> - entry.perm = IOMMU_NONE;
> - entry.addr_mask = size - 1;
> + assert(!remain);
>
> trace_vtd_as_unmap_whole(pci_bus_num(as->bus),
> VTD_PCI_SLOT(as->devfn),
> VTD_PCI_FUNC(as->devfn),
> - entry.iova, size);
> + n->start, size);
>
> - map.iova = entry.iova;
> - map.size = entry.addr_mask;
> + map.iova = n->start;
> + map.size = size;
> iova_tree_remove(as->iova_tree, &map);
> -
> - memory_region_notify_one(n, &entry);
> }
>
> static void vtd_address_space_unmap_all(IntelIOMMUState *s)
>
^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap
2019-06-24 9:18 ` [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap Peter Xu
` (3 preceding siblings ...)
2019-06-25 7:00 ` Auger Eric
@ 2019-07-04 5:45 ` Jason Wang
2019-07-04 8:17 ` Peter Xu
4 siblings, 1 reply; 14+ messages in thread
From: Jason Wang @ 2019-07-04 5:45 UTC (permalink / raw)
To: Peter Xu, qemu-devel
Cc: Paolo Bonzini, Yan Zhao, Auger Eric, Michael S . Tsirkin
On 2019/6/24 下午5:18, Peter Xu wrote:
> This is an replacement work of Yan Zhao's patch:
>
> https://www.mail-archive.com/qemu-devel@nongnu.org/msg625340.html
>
> vtd_address_space_unmap() will do proper page mask alignment to make
> sure each IOTLB message will have correct masks for notification
> messages (2^N-1), but sometimes it can be expanded to even supercede
> the registered range.
I wonder under what circumstance that could we meet this?
Thanks
> That could lead to unexpected UNMAP of already
> mapped regions in some other notifiers.
>
> Instead of doing mindless expension of the start address and address
> mask, we split the range into smaller ones and guarantee that each
> small range will have correct masks (2^N-1) and at the same time we
> should also try our best to generate as less IOTLB messages as
> possible.
>
> Reported-by: Yan Zhao <yan.y.zhao@intel.com>
> Signed-off-by: Peter Xu <peterx@redhat.com>
> ---
> hw/i386/intel_iommu.c | 67 ++++++++++++++++++++++++++-----------------
> 1 file changed, 41 insertions(+), 26 deletions(-)
>
> diff --git a/hw/i386/intel_iommu.c b/hw/i386/intel_iommu.c
> index 719ce19ab3..de86f53b4e 100644
> --- a/hw/i386/intel_iommu.c
> +++ b/hw/i386/intel_iommu.c
> @@ -3363,11 +3363,28 @@ VTDAddressSpace *vtd_find_add_as(IntelIOMMUState *s, PCIBus *bus, int devfn)
> return vtd_dev_as;
> }
>
> +static uint64_t get_naturally_aligned_size(uint64_t start,
> + uint64_t size, int gaw)
> +{
> + uint64_t max_mask = 1ULL << gaw;
> + uint64_t alignment = start ? start & -start : max_mask;
> +
> + alignment = MIN(alignment, max_mask);
> + size = MIN(size, max_mask);
> +
> + if (alignment <= size) {
> + /* Increase the alignment of start */
> + return alignment;
> + } else {
> + /* Find the largest page mask from size */
> + return 1ULL << (63 - clz64(size));
> + }
> +}
> +
> /* Unmap the whole range in the notifier's scope. */
> static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> {
> - IOMMUTLBEntry entry;
> - hwaddr size;
> + hwaddr size, remain;
> hwaddr start = n->start;
> hwaddr end = n->end;
> IntelIOMMUState *s = as->iommu_state;
> @@ -3388,39 +3405,37 @@ static void vtd_address_space_unmap(VTDAddressSpace *as, IOMMUNotifier *n)
> }
>
> assert(start <= end);
> - size = end - start;
> + size = remain = end - start + 1;
>
> - if (ctpop64(size) != 1) {
> - /*
> - * This size cannot format a correct mask. Let's enlarge it to
> - * suite the minimum available mask.
> - */
> - int n = 64 - clz64(size);
> - if (n > s->aw_bits) {
> - /* should not happen, but in case it happens, limit it */
> - n = s->aw_bits;
> - }
> - size = 1ULL << n;
> + while (remain >= VTD_PAGE_SIZE) {
> + IOMMUTLBEntry entry;
> + uint64_t mask = get_naturally_aligned_size(start, remain, s->aw_bits);
> +
> + assert(mask);
> +
> + entry.iova = start;
> + entry.addr_mask = mask - 1;
> + entry.target_as = &address_space_memory;
> + entry.perm = IOMMU_NONE;
> + /* This field is meaningless for unmap */
> + entry.translated_addr = 0;
> +
> + memory_region_notify_one(n, &entry);
> +
> + start += mask;
> + remain -= mask;
> }
>
> - entry.target_as = &address_space_memory;
> - /* Adjust iova for the size */
> - entry.iova = n->start & ~(size - 1);
> - /* This field is meaningless for unmap */
> - entry.translated_addr = 0;
> - entry.perm = IOMMU_NONE;
> - entry.addr_mask = size - 1;
> + assert(!remain);
>
> trace_vtd_as_unmap_whole(pci_bus_num(as->bus),
> VTD_PCI_SLOT(as->devfn),
> VTD_PCI_FUNC(as->devfn),
> - entry.iova, size);
> + n->start, size);
>
> - map.iova = entry.iova;
> - map.size = entry.addr_mask;
> + map.iova = n->start;
> + map.size = size;
> iova_tree_remove(as->iova_tree, &map);
> -
> - memory_region_notify_one(n, &entry);
> }
>
> static void vtd_address_space_unmap_all(IntelIOMMUState *s)
^ permalink raw reply [flat|nested] 14+ messages in thread
* Re: [Qemu-devel] [PATCH v2 2/2] intel_iommu: Fix unexpected unmaps during global unmap
2019-07-04 5:45 ` Jason Wang
@ 2019-07-04 8:17 ` Peter Xu
0 siblings, 0 replies; 14+ messages in thread
From: Peter Xu @ 2019-07-04 8:17 UTC (permalink / raw)
To: Jason Wang
Cc: Paolo Bonzini, Yan Zhao, qemu-devel, Auger Eric,
Michael S . Tsirkin
On Thu, Jul 04, 2019 at 01:45:41PM +0800, Jason Wang wrote:
>
> On 2019/6/24 下午5:18, Peter Xu wrote:
> > This is an replacement work of Yan Zhao's patch:
> >
> > https://www.mail-archive.com/qemu-devel@nongnu.org/msg625340.html
> >
> > vtd_address_space_unmap() will do proper page mask alignment to make
> > sure each IOTLB message will have correct masks for notification
> > messages (2^N-1), but sometimes it can be expanded to even supercede
> > the registered range.
>
>
> I wonder under what circumstance that could we meet this?
Sorry I forgot to reply-all just now...
I've asked a similar question, and Yan's answer is here:
https://www.mail-archive.com/qemu-devel@nongnu.org/msg625597.html
Regards,
--
Peter Xu
^ permalink raw reply [flat|nested] 14+ messages in thread