qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
* [PATCH v2 0/2] support unaligned access for some xHCI registers
@ 2024-02-01  8:13 Tomoyuki HIROSE
  2024-02-01  8:13 ` [PATCH v2 1/2] system/memory.c: support unaligned access Tomoyuki HIROSE
  2024-02-01  8:13 ` [PATCH v2 2/2] hw/usb/hcd-xhci.c: allow unaligned access to Capability Registers Tomoyuki HIROSE
  0 siblings, 2 replies; 9+ messages in thread
From: Tomoyuki HIROSE @ 2024-02-01  8:13 UTC (permalink / raw)
  To: qemu-devel; +Cc: Tomoyuki HIROSE

v1 -> v2:
* Improved the calculation of addresses and masks in memory.c.

According to xHCI spec rev 1.2, unaligned access to xHCI Host
Controller Capability Registers are not prohibited. But current
implementation does not support unaligned access to 'MemoryRegion'.
These patches contain 2 changes:
1. support unaligned access to 'MemoryRegion' .
2. allow unaligned access to Host Controller Capability Registers.

Tomoyuki HIROSE (2):
  system/memory.c: support unaligned access
  hw/usb/hcd-xhci.c: allow unaligned access to Capability Registers

 hw/usb/hcd-xhci.c |  4 +++-
 system/memory.c   | 38 +++++++++++++++++++++++++-------------
 2 files changed, 28 insertions(+), 14 deletions(-)

-- 
2.39.2



^ permalink raw reply	[flat|nested] 9+ messages in thread

* [PATCH v2 1/2] system/memory.c: support unaligned access
  2024-02-01  8:13 [PATCH v2 0/2] support unaligned access for some xHCI registers Tomoyuki HIROSE
@ 2024-02-01  8:13 ` Tomoyuki HIROSE
  2024-02-26  7:28   ` Tomoyuki Hirose
                     ` (2 more replies)
  2024-02-01  8:13 ` [PATCH v2 2/2] hw/usb/hcd-xhci.c: allow unaligned access to Capability Registers Tomoyuki HIROSE
  1 sibling, 3 replies; 9+ messages in thread
From: Tomoyuki HIROSE @ 2024-02-01  8:13 UTC (permalink / raw)
  To: qemu-devel
  Cc: Tomoyuki HIROSE, Paolo Bonzini, Peter Xu, David Hildenbrand,
	Philippe Mathieu-Daudé

The previous code ignored 'impl.unaligned' and handled unaligned accesses
as is. But this implementation cannot emulate specific registers of some
devices that allow unaligned access such as xHCI Host Controller Capability
Registers.
This commit checks 'impl.unaligned' and if it is false, QEMU emulates
unaligned access with multiple aligned access.

Signed-off-by: Tomoyuki HIROSE <tomoyuki.hirose@igel.co.jp>
---
 system/memory.c | 38 +++++++++++++++++++++++++-------------
 1 file changed, 25 insertions(+), 13 deletions(-)

diff --git a/system/memory.c b/system/memory.c
index a229a79988..a7ca0c9f54 100644
--- a/system/memory.c
+++ b/system/memory.c
@@ -535,10 +535,17 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
                                       MemTxAttrs attrs)
 {
     uint64_t access_mask;
+    unsigned access_mask_shift;
+    unsigned access_mask_start_offset;
+    unsigned access_mask_end_offset;
     unsigned access_size;
-    unsigned i;
     MemTxResult r = MEMTX_OK;
     bool reentrancy_guard_applied = false;
+    bool is_big_endian = memory_region_big_endian(mr);
+    signed start_diff;
+    signed current_offset;
+    signed access_shift;
+    hwaddr current_addr;
 
     if (!access_size_min) {
         access_size_min = 1;
@@ -560,19 +567,24 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
         reentrancy_guard_applied = true;
     }
 
-    /* FIXME: support unaligned access? */
     access_size = MAX(MIN(size, access_size_max), access_size_min);
-    access_mask = MAKE_64BIT_MASK(0, access_size * 8);
-    if (memory_region_big_endian(mr)) {
-        for (i = 0; i < size; i += access_size) {
-            r |= access_fn(mr, addr + i, value, access_size,
-                        (size - access_size - i) * 8, access_mask, attrs);
-        }
-    } else {
-        for (i = 0; i < size; i += access_size) {
-            r |= access_fn(mr, addr + i, value, access_size, i * 8,
-                        access_mask, attrs);
-        }
+    start_diff = mr->ops->impl.unaligned ? 0 : addr & (access_size - 1);
+    current_addr = addr - start_diff;
+    for (current_offset = -start_diff; current_offset < (signed)size;
+         current_offset += access_size, current_addr += access_size) {
+        access_shift = is_big_endian
+                          ? (signed)size - (signed)access_size - current_offset
+                          : current_offset;
+        access_mask_shift = current_offset > 0 ? 0 : -current_offset;
+        access_mask_start_offset = current_offset > 0 ? current_offset : 0;
+        access_mask_end_offset = current_offset + access_size > size
+                                     ? size
+                                     : current_offset + access_size;
+        access_mask = MAKE_64BIT_MASK(access_mask_shift * 8,
+            (access_mask_end_offset - access_mask_start_offset) * 8);
+
+        r |= access_fn(mr, current_addr, value, access_size, access_shift * 8,
+                       access_mask, attrs);
     }
     if (mr->dev && reentrancy_guard_applied) {
         mr->dev->mem_reentrancy_guard.engaged_in_io = false;
-- 
2.39.2



^ permalink raw reply related	[flat|nested] 9+ messages in thread

* [PATCH v2 2/2] hw/usb/hcd-xhci.c: allow unaligned access to Capability Registers
  2024-02-01  8:13 [PATCH v2 0/2] support unaligned access for some xHCI registers Tomoyuki HIROSE
  2024-02-01  8:13 ` [PATCH v2 1/2] system/memory.c: support unaligned access Tomoyuki HIROSE
@ 2024-02-01  8:13 ` Tomoyuki HIROSE
  1 sibling, 0 replies; 9+ messages in thread
From: Tomoyuki HIROSE @ 2024-02-01  8:13 UTC (permalink / raw)
  To: qemu-devel; +Cc: Tomoyuki HIROSE, Gerd Hoffmann

According to xHCI spec rev 1.2, unaligned access to xHCI Host
Controller Capability Registers is not prohibited. In Addition, the
limit of access size is also unspecified. Actually, some real devices
allow unaligned access and 8-byte access to these registers.
This commit makes it possible to unaligned access and 8-byte access
to Host Controller Capability Registers.

Resolves: https://gitlab.com/qemu-project/qemu/-/issues/143
Signed-off-by: Tomoyuki HIROSE <tomoyuki.hirose@igel.co.jp>
---
 hw/usb/hcd-xhci.c | 4 +++-
 1 file changed, 3 insertions(+), 1 deletion(-)

diff --git a/hw/usb/hcd-xhci.c b/hw/usb/hcd-xhci.c
index ad40232eb6..9e0b24c93e 100644
--- a/hw/usb/hcd-xhci.c
+++ b/hw/usb/hcd-xhci.c
@@ -3181,9 +3181,11 @@ static const MemoryRegionOps xhci_cap_ops = {
     .read = xhci_cap_read,
     .write = xhci_cap_write,
     .valid.min_access_size = 1,
-    .valid.max_access_size = 4,
+    .valid.max_access_size = 8,
+    .valid.unaligned = true,
     .impl.min_access_size = 4,
     .impl.max_access_size = 4,
+    .impl.unaligned = false,
     .endianness = DEVICE_LITTLE_ENDIAN,
 };
 
-- 
2.39.2



^ permalink raw reply related	[flat|nested] 9+ messages in thread

* Re: [PATCH v2 1/2] system/memory.c: support unaligned access
  2024-02-01  8:13 ` [PATCH v2 1/2] system/memory.c: support unaligned access Tomoyuki HIROSE
@ 2024-02-26  7:28   ` Tomoyuki Hirose
  2024-03-18  4:34     ` Tomoyuki Hirose
  2024-03-18 16:15   ` Peter Xu
  2024-03-19 14:08   ` Peter Maydell
  2 siblings, 1 reply; 9+ messages in thread
From: Tomoyuki Hirose @ 2024-02-26  7:28 UTC (permalink / raw)
  To: qemu-devel
  Cc: Paolo Bonzini, Peter Xu, David Hildenbrand,
	Philippe Mathieu-Daudé

Hello,
I would be happy if you could give me some comments.

ping.

On Thu, Feb 1, 2024 at 5:14 PM Tomoyuki HIROSE
<tomoyuki.hirose@igel.co.jp> wrote:
>
> The previous code ignored 'impl.unaligned' and handled unaligned accesses
> as is. But this implementation cannot emulate specific registers of some
> devices that allow unaligned access such as xHCI Host Controller Capability
> Registers.
> This commit checks 'impl.unaligned' and if it is false, QEMU emulates
> unaligned access with multiple aligned access.
>
> Signed-off-by: Tomoyuki HIROSE <tomoyuki.hirose@igel.co.jp>
> ---
>  system/memory.c | 38 +++++++++++++++++++++++++-------------
>  1 file changed, 25 insertions(+), 13 deletions(-)
>
> diff --git a/system/memory.c b/system/memory.c
> index a229a79988..a7ca0c9f54 100644
> --- a/system/memory.c
> +++ b/system/memory.c
> @@ -535,10 +535,17 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
>                                        MemTxAttrs attrs)
>  {
>      uint64_t access_mask;
> +    unsigned access_mask_shift;
> +    unsigned access_mask_start_offset;
> +    unsigned access_mask_end_offset;
>      unsigned access_size;
> -    unsigned i;
>      MemTxResult r = MEMTX_OK;
>      bool reentrancy_guard_applied = false;
> +    bool is_big_endian = memory_region_big_endian(mr);
> +    signed start_diff;
> +    signed current_offset;
> +    signed access_shift;
> +    hwaddr current_addr;
>
>      if (!access_size_min) {
>          access_size_min = 1;
> @@ -560,19 +567,24 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
>          reentrancy_guard_applied = true;
>      }
>
> -    /* FIXME: support unaligned access? */
>      access_size = MAX(MIN(size, access_size_max), access_size_min);
> -    access_mask = MAKE_64BIT_MASK(0, access_size * 8);
> -    if (memory_region_big_endian(mr)) {
> -        for (i = 0; i < size; i += access_size) {
> -            r |= access_fn(mr, addr + i, value, access_size,
> -                        (size - access_size - i) * 8, access_mask, attrs);
> -        }
> -    } else {
> -        for (i = 0; i < size; i += access_size) {
> -            r |= access_fn(mr, addr + i, value, access_size, i * 8,
> -                        access_mask, attrs);
> -        }
> +    start_diff = mr->ops->impl.unaligned ? 0 : addr & (access_size - 1);
> +    current_addr = addr - start_diff;
> +    for (current_offset = -start_diff; current_offset < (signed)size;
> +         current_offset += access_size, current_addr += access_size) {
> +        access_shift = is_big_endian
> +                          ? (signed)size - (signed)access_size - current_offset
> +                          : current_offset;
> +        access_mask_shift = current_offset > 0 ? 0 : -current_offset;
> +        access_mask_start_offset = current_offset > 0 ? current_offset : 0;
> +        access_mask_end_offset = current_offset + access_size > size
> +                                     ? size
> +                                     : current_offset + access_size;
> +        access_mask = MAKE_64BIT_MASK(access_mask_shift * 8,
> +            (access_mask_end_offset - access_mask_start_offset) * 8);
> +
> +        r |= access_fn(mr, current_addr, value, access_size, access_shift * 8,
> +                       access_mask, attrs);
>      }
>      if (mr->dev && reentrancy_guard_applied) {
>          mr->dev->mem_reentrancy_guard.engaged_in_io = false;
> --
> 2.39.2
>


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2 1/2] system/memory.c: support unaligned access
  2024-02-26  7:28   ` Tomoyuki Hirose
@ 2024-03-18  4:34     ` Tomoyuki Hirose
  0 siblings, 0 replies; 9+ messages in thread
From: Tomoyuki Hirose @ 2024-03-18  4:34 UTC (permalink / raw)
  To: qemu-devel
  Cc: Paolo Bonzini, Peter Xu, David Hildenbrand,
	Philippe Mathieu-Daudé

ping.

On Mon, Feb 26, 2024 at 4:28 PM Tomoyuki Hirose
<tomoyuki.hirose@igel.co.jp> wrote:
>
> Hello,
> I would be happy if you could give me some comments.
>
> ping.
>
> On Thu, Feb 1, 2024 at 5:14 PM Tomoyuki HIROSE
> <tomoyuki.hirose@igel.co.jp> wrote:
> >
> > The previous code ignored 'impl.unaligned' and handled unaligned accesses
> > as is. But this implementation cannot emulate specific registers of some
> > devices that allow unaligned access such as xHCI Host Controller Capability
> > Registers.
> > This commit checks 'impl.unaligned' and if it is false, QEMU emulates
> > unaligned access with multiple aligned access.
> >
> > Signed-off-by: Tomoyuki HIROSE <tomoyuki.hirose@igel.co.jp>
> > ---
> >  system/memory.c | 38 +++++++++++++++++++++++++-------------
> >  1 file changed, 25 insertions(+), 13 deletions(-)
> >
> > diff --git a/system/memory.c b/system/memory.c
> > index a229a79988..a7ca0c9f54 100644
> > --- a/system/memory.c
> > +++ b/system/memory.c
> > @@ -535,10 +535,17 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
> >                                        MemTxAttrs attrs)
> >  {
> >      uint64_t access_mask;
> > +    unsigned access_mask_shift;
> > +    unsigned access_mask_start_offset;
> > +    unsigned access_mask_end_offset;
> >      unsigned access_size;
> > -    unsigned i;
> >      MemTxResult r = MEMTX_OK;
> >      bool reentrancy_guard_applied = false;
> > +    bool is_big_endian = memory_region_big_endian(mr);
> > +    signed start_diff;
> > +    signed current_offset;
> > +    signed access_shift;
> > +    hwaddr current_addr;
> >
> >      if (!access_size_min) {
> >          access_size_min = 1;
> > @@ -560,19 +567,24 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
> >          reentrancy_guard_applied = true;
> >      }
> >
> > -    /* FIXME: support unaligned access? */
> >      access_size = MAX(MIN(size, access_size_max), access_size_min);
> > -    access_mask = MAKE_64BIT_MASK(0, access_size * 8);
> > -    if (memory_region_big_endian(mr)) {
> > -        for (i = 0; i < size; i += access_size) {
> > -            r |= access_fn(mr, addr + i, value, access_size,
> > -                        (size - access_size - i) * 8, access_mask, attrs);
> > -        }
> > -    } else {
> > -        for (i = 0; i < size; i += access_size) {
> > -            r |= access_fn(mr, addr + i, value, access_size, i * 8,
> > -                        access_mask, attrs);
> > -        }
> > +    start_diff = mr->ops->impl.unaligned ? 0 : addr & (access_size - 1);
> > +    current_addr = addr - start_diff;
> > +    for (current_offset = -start_diff; current_offset < (signed)size;
> > +         current_offset += access_size, current_addr += access_size) {
> > +        access_shift = is_big_endian
> > +                          ? (signed)size - (signed)access_size - current_offset
> > +                          : current_offset;
> > +        access_mask_shift = current_offset > 0 ? 0 : -current_offset;
> > +        access_mask_start_offset = current_offset > 0 ? current_offset : 0;
> > +        access_mask_end_offset = current_offset + access_size > size
> > +                                     ? size
> > +                                     : current_offset + access_size;
> > +        access_mask = MAKE_64BIT_MASK(access_mask_shift * 8,
> > +            (access_mask_end_offset - access_mask_start_offset) * 8);
> > +
> > +        r |= access_fn(mr, current_addr, value, access_size, access_shift * 8,
> > +                       access_mask, attrs);
> >      }
> >      if (mr->dev && reentrancy_guard_applied) {
> >          mr->dev->mem_reentrancy_guard.engaged_in_io = false;
> > --
> > 2.39.2
> >


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2 1/2] system/memory.c: support unaligned access
  2024-02-01  8:13 ` [PATCH v2 1/2] system/memory.c: support unaligned access Tomoyuki HIROSE
  2024-02-26  7:28   ` Tomoyuki Hirose
@ 2024-03-18 16:15   ` Peter Xu
  2024-03-19  6:43     ` Philippe Mathieu-Daudé
  2024-03-19 14:08   ` Peter Maydell
  2 siblings, 1 reply; 9+ messages in thread
From: Peter Xu @ 2024-03-18 16:15 UTC (permalink / raw)
  To: Tomoyuki HIROSE, Peter Maydell
  Cc: qemu-devel, Paolo Bonzini, David Hildenbrand,
	Philippe Mathieu-Daudé

Hi,

On Thu, Feb 01, 2024 at 05:13:12PM +0900, Tomoyuki HIROSE wrote:
> The previous code ignored 'impl.unaligned' and handled unaligned accesses
> as is. But this implementation cannot emulate specific registers of some
> devices that allow unaligned access such as xHCI Host Controller Capability
> Registers.
> This commit checks 'impl.unaligned' and if it is false, QEMU emulates
> unaligned access with multiple aligned access.

This patch looks mostly good to me.  Just a few trivial comments.

Firstly, can we provide the USB example here (or also the bug link) so that
we can still pick up the context of why this will start to be useful when
people read about this commit separately?

> 
> Signed-off-by: Tomoyuki HIROSE <tomoyuki.hirose@igel.co.jp>
> ---
>  system/memory.c | 38 +++++++++++++++++++++++++-------------
>  1 file changed, 25 insertions(+), 13 deletions(-)
> 
> diff --git a/system/memory.c b/system/memory.c
> index a229a79988..a7ca0c9f54 100644
> --- a/system/memory.c
> +++ b/system/memory.c
> @@ -535,10 +535,17 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
>                                        MemTxAttrs attrs)
>  {
>      uint64_t access_mask;
> +    unsigned access_mask_shift;
> +    unsigned access_mask_start_offset;
> +    unsigned access_mask_end_offset;
>      unsigned access_size;
> -    unsigned i;
>      MemTxResult r = MEMTX_OK;
>      bool reentrancy_guard_applied = false;
> +    bool is_big_endian = memory_region_big_endian(mr);
> +    signed start_diff;
> +    signed current_offset;
> +    signed access_shift;
> +    hwaddr current_addr;
>  
>      if (!access_size_min) {
>          access_size_min = 1;
> @@ -560,19 +567,24 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
>          reentrancy_guard_applied = true;
>      }
>  
> -    /* FIXME: support unaligned access? */
>      access_size = MAX(MIN(size, access_size_max), access_size_min);
> -    access_mask = MAKE_64BIT_MASK(0, access_size * 8);
> -    if (memory_region_big_endian(mr)) {
> -        for (i = 0; i < size; i += access_size) {
> -            r |= access_fn(mr, addr + i, value, access_size,
> -                        (size - access_size - i) * 8, access_mask, attrs);
> -        }
> -    } else {
> -        for (i = 0; i < size; i += access_size) {
> -            r |= access_fn(mr, addr + i, value, access_size, i * 8,
> -                        access_mask, attrs);
> -        }
> +    start_diff = mr->ops->impl.unaligned ? 0 : addr & (access_size - 1);
> +    current_addr = addr - start_diff;
> +    for (current_offset = -start_diff; current_offset < (signed)size;
> +         current_offset += access_size, current_addr += access_size) {
> +        access_shift = is_big_endian
> +                          ? (signed)size - (signed)access_size - current_offset
> +                          : current_offset;
> +        access_mask_shift = current_offset > 0 ? 0 : -current_offset;
> +        access_mask_start_offset = current_offset > 0 ? current_offset : 0;
> +        access_mask_end_offset = current_offset + access_size > size
> +                                     ? size
> +                                     : current_offset + access_size;

Maybe this looks slightly easier to read?

        if (current_offset < 0) {
            access_mask_shift = -current_offset;
            access_mask_start_offset = 0;
        } else {
            access_mask_shift = 0;
            access_mask_start_offset = current_offset;
        }
        access_mask_end_offset = MIN(current_offset + access_size, size);

But I confess this can be pretty subjective..

Since PeterM used to comment, please remember to copy PeterM too in the
future post in case this got overlooked.

Peter, do you still have any other comments or concerns?

Thanks,

> +        access_mask = MAKE_64BIT_MASK(access_mask_shift * 8,
> +            (access_mask_end_offset - access_mask_start_offset) * 8);
> +
> +        r |= access_fn(mr, current_addr, value, access_size, access_shift * 8,
> +                       access_mask, attrs);
>      }
>      if (mr->dev && reentrancy_guard_applied) {
>          mr->dev->mem_reentrancy_guard.engaged_in_io = false;
> -- 
> 2.39.2
> 

-- 
Peter Xu



^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2 1/2] system/memory.c: support unaligned access
  2024-03-18 16:15   ` Peter Xu
@ 2024-03-19  6:43     ` Philippe Mathieu-Daudé
  2024-03-19  6:50       ` Philippe Mathieu-Daudé
  0 siblings, 1 reply; 9+ messages in thread
From: Philippe Mathieu-Daudé @ 2024-03-19  6:43 UTC (permalink / raw)
  To: Peter Xu, Tomoyuki HIROSE, Peter Maydell
  Cc: qemu-devel, Paolo Bonzini, David Hildenbrand, Cameron Esfahani,
	Cédric Le Goater, Andrew Jeffery, Richard Henderson

On 18/3/24 17:15, Peter Xu wrote:
> Hi,
> 
> On Thu, Feb 01, 2024 at 05:13:12PM +0900, Tomoyuki HIROSE wrote:
>> The previous code ignored 'impl.unaligned' and handled unaligned accesses
>> as is. But this implementation cannot emulate specific registers of some
>> devices that allow unaligned access such as xHCI Host Controller Capability
>> Registers.
>> This commit checks 'impl.unaligned' and if it is false, QEMU emulates
>> unaligned access with multiple aligned access.
> 
> This patch looks mostly good to me.  Just a few trivial comments.
> 
> Firstly, can we provide the USB example here (or also the bug link) so that
> we can still pick up the context of why this will start to be useful when
> people read about this commit separately?
> 
>>
>> Signed-off-by: Tomoyuki HIROSE <tomoyuki.hirose@igel.co.jp>
>> ---
>>   system/memory.c | 38 +++++++++++++++++++++++++-------------
>>   1 file changed, 25 insertions(+), 13 deletions(-)
>>
>> diff --git a/system/memory.c b/system/memory.c
>> index a229a79988..a7ca0c9f54 100644
>> --- a/system/memory.c
>> +++ b/system/memory.c
>> @@ -535,10 +535,17 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
>>                                         MemTxAttrs attrs)
>>   {
>>       uint64_t access_mask;
>> +    unsigned access_mask_shift;
>> +    unsigned access_mask_start_offset;
>> +    unsigned access_mask_end_offset;
>>       unsigned access_size;
>> -    unsigned i;
>>       MemTxResult r = MEMTX_OK;
>>       bool reentrancy_guard_applied = false;
>> +    bool is_big_endian = memory_region_big_endian(mr);
>> +    signed start_diff;
>> +    signed current_offset;
>> +    signed access_shift;
>> +    hwaddr current_addr;
>>   
>>       if (!access_size_min) {
>>           access_size_min = 1;
>> @@ -560,19 +567,24 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
>>           reentrancy_guard_applied = true;
>>       }
>>   
>> -    /* FIXME: support unaligned access? */
>>       access_size = MAX(MIN(size, access_size_max), access_size_min);
>> -    access_mask = MAKE_64BIT_MASK(0, access_size * 8);
>> -    if (memory_region_big_endian(mr)) {
>> -        for (i = 0; i < size; i += access_size) {
>> -            r |= access_fn(mr, addr + i, value, access_size,
>> -                        (size - access_size - i) * 8, access_mask, attrs);
>> -        }
>> -    } else {
>> -        for (i = 0; i < size; i += access_size) {
>> -            r |= access_fn(mr, addr + i, value, access_size, i * 8,
>> -                        access_mask, attrs);
>> -        }
>> +    start_diff = mr->ops->impl.unaligned ? 0 : addr & (access_size - 1);
>> +    current_addr = addr - start_diff;
>> +    for (current_offset = -start_diff; current_offset < (signed)size;
>> +         current_offset += access_size, current_addr += access_size) {
>> +        access_shift = is_big_endian
>> +                          ? (signed)size - (signed)access_size - current_offset
>> +                          : current_offset;
>> +        access_mask_shift = current_offset > 0 ? 0 : -current_offset;
>> +        access_mask_start_offset = current_offset > 0 ? current_offset : 0;
>> +        access_mask_end_offset = current_offset + access_size > size
>> +                                     ? size
>> +                                     : current_offset + access_size;
> 
> Maybe this looks slightly easier to read?
> 
>          if (current_offset < 0) {
>              access_mask_shift = -current_offset;
>              access_mask_start_offset = 0;
>          } else {
>              access_mask_shift = 0;
>              access_mask_start_offset = current_offset;
>          }
>          access_mask_end_offset = MIN(current_offset + access_size, size);
> 
> But I confess this can be pretty subjective..
> 
> Since PeterM used to comment, please remember to copy PeterM too in the
> future post in case this got overlooked.
> 
> Peter, do you still have any other comments or concerns?

See also this thread:
https://lore.kernel.org/qemu-devel/20200331144225.67dadl6crwd57qvi@sirius.home.kraxel.org/
->
https://www.mail-archive.com/qemu-devel@nongnu.org/msg461247.html

Also I guess remembering Richard mentioning we should unify this
code for softmmu / physmem, but I might be wrong ...

> 
> Thanks,
> 
>> +        access_mask = MAKE_64BIT_MASK(access_mask_shift * 8,
>> +            (access_mask_end_offset - access_mask_start_offset) * 8);
>> +
>> +        r |= access_fn(mr, current_addr, value, access_size, access_shift * 8,
>> +                       access_mask, attrs);
>>       }
>>       if (mr->dev && reentrancy_guard_applied) {
>>           mr->dev->mem_reentrancy_guard.engaged_in_io = false;
>> -- 
>> 2.39.2
>>
> 



^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2 1/2] system/memory.c: support unaligned access
  2024-03-19  6:43     ` Philippe Mathieu-Daudé
@ 2024-03-19  6:50       ` Philippe Mathieu-Daudé
  0 siblings, 0 replies; 9+ messages in thread
From: Philippe Mathieu-Daudé @ 2024-03-19  6:50 UTC (permalink / raw)
  To: Peter Xu, Tomoyuki HIROSE, Peter Maydell
  Cc: qemu-devel, Paolo Bonzini, David Hildenbrand, Cameron Esfahani,
	Cédric Le Goater, Andrew Jeffery, Richard Henderson

Hi Tomoyuki,

On 19/3/24 07:43, Philippe Mathieu-Daudé wrote:
> On 18/3/24 17:15, Peter Xu wrote:
>> Hi,
>>
>> On Thu, Feb 01, 2024 at 05:13:12PM +0900, Tomoyuki HIROSE wrote:
>>> The previous code ignored 'impl.unaligned' and handled unaligned 
>>> accesses
>>> as is. But this implementation cannot emulate specific registers of some
>>> devices that allow unaligned access such as xHCI Host Controller 
>>> Capability
>>> Registers.
>>> This commit checks 'impl.unaligned' and if it is false, QEMU emulates
>>> unaligned access with multiple aligned access.
>>
>> This patch looks mostly good to me.  Just a few trivial comments.
>>
>> Firstly, can we provide the USB example here (or also the bug link) so 
>> that
>> we can still pick up the context of why this will start to be useful when
>> people read about this commit separately?
>>
>>>
>>> Signed-off-by: Tomoyuki HIROSE <tomoyuki.hirose@igel.co.jp>
>>> ---
>>>   system/memory.c | 38 +++++++++++++++++++++++++-------------
>>>   1 file changed, 25 insertions(+), 13 deletions(-)
>>>
>>> diff --git a/system/memory.c b/system/memory.c
>>> index a229a79988..a7ca0c9f54 100644
>>> --- a/system/memory.c
>>> +++ b/system/memory.c
>>> @@ -535,10 +535,17 @@ static MemTxResult 
>>> access_with_adjusted_size(hwaddr addr,
>>>                                         MemTxAttrs attrs)
>>>   {
>>>       uint64_t access_mask;
>>> +    unsigned access_mask_shift;
>>> +    unsigned access_mask_start_offset;
>>> +    unsigned access_mask_end_offset;
>>>       unsigned access_size;
>>> -    unsigned i;
>>>       MemTxResult r = MEMTX_OK;
>>>       bool reentrancy_guard_applied = false;
>>> +    bool is_big_endian = memory_region_big_endian(mr);
>>> +    signed start_diff;
>>> +    signed current_offset;
>>> +    signed access_shift;
>>> +    hwaddr current_addr;
>>>       if (!access_size_min) {
>>>           access_size_min = 1;
>>> @@ -560,19 +567,24 @@ static MemTxResult 
>>> access_with_adjusted_size(hwaddr addr,
>>>           reentrancy_guard_applied = true;
>>>       }
>>> -    /* FIXME: support unaligned access? */
>>>       access_size = MAX(MIN(size, access_size_max), access_size_min);
>>> -    access_mask = MAKE_64BIT_MASK(0, access_size * 8);
>>> -    if (memory_region_big_endian(mr)) {
>>> -        for (i = 0; i < size; i += access_size) {
>>> -            r |= access_fn(mr, addr + i, value, access_size,
>>> -                        (size - access_size - i) * 8, access_mask, 
>>> attrs);
>>> -        }
>>> -    } else {
>>> -        for (i = 0; i < size; i += access_size) {
>>> -            r |= access_fn(mr, addr + i, value, access_size, i * 8,
>>> -                        access_mask, attrs);
>>> -        }
>>> +    start_diff = mr->ops->impl.unaligned ? 0 : addr & (access_size - 
>>> 1);
>>> +    current_addr = addr - start_diff;
>>> +    for (current_offset = -start_diff; current_offset < (signed)size;
>>> +         current_offset += access_size, current_addr += access_size) {
>>> +        access_shift = is_big_endian
>>> +                          ? (signed)size - (signed)access_size - 
>>> current_offset
>>> +                          : current_offset;
>>> +        access_mask_shift = current_offset > 0 ? 0 : -current_offset;
>>> +        access_mask_start_offset = current_offset > 0 ? 
>>> current_offset : 0;
>>> +        access_mask_end_offset = current_offset + access_size > size
>>> +                                     ? size
>>> +                                     : current_offset + access_size;
>>
>> Maybe this looks slightly easier to read?
>>
>>          if (current_offset < 0) {
>>              access_mask_shift = -current_offset;
>>              access_mask_start_offset = 0;
>>          } else {
>>              access_mask_shift = 0;
>>              access_mask_start_offset = current_offset;
>>          }
>>          access_mask_end_offset = MIN(current_offset + access_size, 
>> size);
>>
>> But I confess this can be pretty subjective..
>>
>> Since PeterM used to comment, please remember to copy PeterM too in the
>> future post in case this got overlooked.
>>
>> Peter, do you still have any other comments or concerns?
> 
> See also this thread:
> https://lore.kernel.org/qemu-devel/20200331144225.67dadl6crwd57qvi@sirius.home.kraxel.org/
> ->
> https://www.mail-archive.com/qemu-devel@nongnu.org/msg461247.html

Now I noticed this thread was mentioned on v1, but not all person Cc'ed
there are on v2.

What I'd like to see to get confidence and avoid regression is some
harness qtests triggering this problem. See for example:
https://lore.kernel.org/qemu-devel/20200817161853.593247-8-f4bug@amsat.org/

> 
> Also I guess remembering Richard mentioning we should unify this
> code for softmmu / physmem, but I might be wrong ...
> 
>>
>> Thanks,
>>
>>> +        access_mask = MAKE_64BIT_MASK(access_mask_shift * 8,
>>> +            (access_mask_end_offset - access_mask_start_offset) * 8);
>>> +
>>> +        r |= access_fn(mr, current_addr, value, access_size, 
>>> access_shift * 8,
>>> +                       access_mask, attrs);
>>>       }
>>>       if (mr->dev && reentrancy_guard_applied) {
>>>           mr->dev->mem_reentrancy_guard.engaged_in_io = false;
>>> -- 
>>> 2.39.2
>>>
>>
> 



^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2 1/2] system/memory.c: support unaligned access
  2024-02-01  8:13 ` [PATCH v2 1/2] system/memory.c: support unaligned access Tomoyuki HIROSE
  2024-02-26  7:28   ` Tomoyuki Hirose
  2024-03-18 16:15   ` Peter Xu
@ 2024-03-19 14:08   ` Peter Maydell
  2 siblings, 0 replies; 9+ messages in thread
From: Peter Maydell @ 2024-03-19 14:08 UTC (permalink / raw)
  To: Tomoyuki HIROSE
  Cc: qemu-devel, Paolo Bonzini, Peter Xu, David Hildenbrand,
	Philippe Mathieu-Daudé

On Thu, 1 Feb 2024 at 08:15, Tomoyuki HIROSE <tomoyuki.hirose@igel.co.jp> wrote:
>
> The previous code ignored 'impl.unaligned' and handled unaligned accesses
> as is. But this implementation cannot emulate specific registers of some
> devices that allow unaligned access such as xHCI Host Controller Capability
> Registers.
> This commit checks 'impl.unaligned' and if it is false, QEMU emulates
> unaligned access with multiple aligned access.
>
> Signed-off-by: Tomoyuki HIROSE <tomoyuki.hirose@igel.co.jp>
> ---
>  system/memory.c | 38 +++++++++++++++++++++++++-------------
>  1 file changed, 25 insertions(+), 13 deletions(-)
>
> diff --git a/system/memory.c b/system/memory.c
> index a229a79988..a7ca0c9f54 100644
> --- a/system/memory.c
> +++ b/system/memory.c
> @@ -535,10 +535,17 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
>                                        MemTxAttrs attrs)
>  {
>      uint64_t access_mask;
> +    unsigned access_mask_shift;
> +    unsigned access_mask_start_offset;
> +    unsigned access_mask_end_offset;
>      unsigned access_size;
> -    unsigned i;
>      MemTxResult r = MEMTX_OK;
>      bool reentrancy_guard_applied = false;
> +    bool is_big_endian = memory_region_big_endian(mr);
> +    signed start_diff;
> +    signed current_offset;
> +    signed access_shift;

"signed foo" is a weird way to specify this type, which we use almost
nowhere else in the codebase -- this is equivalent to "int foo".

> +    hwaddr current_addr;
>
>      if (!access_size_min) {
>          access_size_min = 1;
> @@ -560,19 +567,24 @@ static MemTxResult access_with_adjusted_size(hwaddr addr,
>          reentrancy_guard_applied = true;
>      }
>
> -    /* FIXME: support unaligned access? */
>      access_size = MAX(MIN(size, access_size_max), access_size_min);

This still has a problem I noted for the v1 patch:
we compute the access_size without thinking about the alignment,
so for an access like:
 * addr = 2, size = 4, access_size_min = 2, access_size_max = 8
we will calculate access_size = 4 and do two 4-byte accesses
(at addresses 0 and 4) when we should do two 2-byte accesses
(at addresses 2 and 4).

> -    access_mask = MAKE_64BIT_MASK(0, access_size * 8);
> -    if (memory_region_big_endian(mr)) {
> -        for (i = 0; i < size; i += access_size) {
> -            r |= access_fn(mr, addr + i, value, access_size,
> -                        (size - access_size - i) * 8, access_mask, attrs);
> -        }
> -    } else {
> -        for (i = 0; i < size; i += access_size) {
> -            r |= access_fn(mr, addr + i, value, access_size, i * 8,
> -                        access_mask, attrs);
> -        }
> +    start_diff = mr->ops->impl.unaligned ? 0 : addr & (access_size - 1);
> +    current_addr = addr - start_diff;
> +    for (current_offset = -start_diff; current_offset < (signed)size;
> +         current_offset += access_size, current_addr += access_size) {
> +        access_shift = is_big_endian
> +                          ? (signed)size - (signed)access_size - current_offset
> +                          : current_offset;
> +        access_mask_shift = current_offset > 0 ? 0 : -current_offset;
> +        access_mask_start_offset = current_offset > 0 ? current_offset : 0;
> +        access_mask_end_offset = current_offset + access_size > size
> +                                     ? size
> +                                     : current_offset + access_size;
> +        access_mask = MAKE_64BIT_MASK(access_mask_shift * 8,
> +            (access_mask_end_offset - access_mask_start_offset) * 8);

I don't understand here why the access_mask_shift and the
access_mask_start_offset are different. Aren't we trying to create
a mask value with 1s from start through to end ?

> +
> +        r |= access_fn(mr, current_addr, value, access_size, access_shift * 8,
> +                       access_mask, attrs);
>      }
>      if (mr->dev && reentrancy_guard_applied) {
>          mr->dev->mem_reentrancy_guard.engaged_in_io = false;

I agree with Philippe that we could be a lot more confident in
this change if we had some unit tests that tested whether
various combinations of unaligned accesses turned into the
right sequence of accesses to the underlying device.

thanks
-- PMM


^ permalink raw reply	[flat|nested] 9+ messages in thread

end of thread, other threads:[~2024-03-19 14:12 UTC | newest]

Thread overview: 9+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2024-02-01  8:13 [PATCH v2 0/2] support unaligned access for some xHCI registers Tomoyuki HIROSE
2024-02-01  8:13 ` [PATCH v2 1/2] system/memory.c: support unaligned access Tomoyuki HIROSE
2024-02-26  7:28   ` Tomoyuki Hirose
2024-03-18  4:34     ` Tomoyuki Hirose
2024-03-18 16:15   ` Peter Xu
2024-03-19  6:43     ` Philippe Mathieu-Daudé
2024-03-19  6:50       ` Philippe Mathieu-Daudé
2024-03-19 14:08   ` Peter Maydell
2024-02-01  8:13 ` [PATCH v2 2/2] hw/usb/hcd-xhci.c: allow unaligned access to Capability Registers Tomoyuki HIROSE

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).