qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Zang Hongyong <zanghongyong@huawei.com>
To: Cam Macdonell <cam@cs.ualberta.ca>
Cc: wusongwei@huawei.com, kvm@vger.kernel.org, hanweidong@huawei.com,
	qemu-devel@nongnu.org, louzhengwei@huawei.com,
	xiaowei.yang@huawei.com, avi@redhat.com
Subject: Re: [Qemu-devel] [PATCH] ivshmem: fix guest unable to start with ioeventfd
Date: Fri, 25 Nov 2011 17:12:11 +0800	[thread overview]
Message-ID: <4ECF5BEB.3080403@huawei.com> (raw)
In-Reply-To: <CAKjmthLPyAdF_WqAVB=nJ=z2wQP1psVcy+Xk_wrOS3C7P4pB3Q@mail.gmail.com>

于 2011/11/25,星期五 4:29, Cam Macdonell 写道:
> On Thu, Nov 24, 2011 at 3:05 AM,<zanghongyong@huawei.com>  wrote:
>> From: Hongyong Zang<zanghongyong@huawei.com>
>>
>> When a guest boots with ioeventfd, an error (by gdb) occurs:
>>   Program received signal SIGSEGV, Segmentation fault.
>>   0x00000000006009cc in setup_ioeventfds (s=0x171dc40)
>>       at /home/louzhengwei/git_source/qemu-kvm/hw/ivshmem.c:363
>>   363             for (j = 0; j<  s->peers[i].nb_eventfds; j++) {
>> The bug is due to accessing s->peers which is NULL.
> Can you share the command-line that caused the fault?
The command-lines:
1) ivshmem_server -m 4 -p /tmp/nahanni
2) qemu-system-x86_64 -smp 1 -m 1024 -boot c -drive 
file=./rhel6_ivsh1.img,if=virtio
-chardev socket,path=/tmp/nahanni,id=lzw -device 
ivshmem,chardev=lzw,size=4m,ioeventfd=on,vectors=8 -vnc :41

>> This patch uses the memory region API to replace the old one kvm_set_ioeventfd_mmio_long().
>> And this patch makes memory_region_add_eventfd() called in ivshmem_read() when qemu receives
>> eventfd information from ivshmem_server.
> Should this patch be split into two patches, to separate the bug fix
> from the other changes related to the Memory API?  Unless I
> misunderstand how the two are necessarily related.

This bug locates in setup_ioeventfds(). The setup_ioeventfds() function 
wants to call memory_region_add_eventfd()
to configure eventfd info for every s->peer, but s->peers are 
uninitialized at this moment.

When qemu receives eventfd info from "ivshmem_server", ivshmem_read() is 
called to set eventfd info.
Function ivshmem_read() calls kvm_set_ioeventfd_mmio_long() which can be 
encapsulated in the new Memory API.

So this patch uses memory_region_add_eventfd() to replace 
kvm_set_ioeventfd_mmio_long(). In this way, each qemu
has already configured the proper eventfd info, and there's no need to 
use the function setup_ioeventfds().
Furthermore, nobody uses IVShmemState's member "pcibus_t mmio_addr". So 
we can remove the member as well.

Regards,
Hongyong Zang

>
> Cam
>
>> Signed-off-by: Hongyong Zang<zanghongyong@huawei.com>
>> ---
>>   hw/ivshmem.c |   41 ++++++++++++++---------------------------
>>   1 files changed, 14 insertions(+), 27 deletions(-)
>>
>> diff --git a/hw/ivshmem.c b/hw/ivshmem.c
>> index 242fbea..be26f03 100644
>> --- a/hw/ivshmem.c
>> +++ b/hw/ivshmem.c
>> @@ -58,7 +58,6 @@ typedef struct IVShmemState {
>>      CharDriverState *server_chr;
>>      MemoryRegion ivshmem_mmio;
>>
>> -    pcibus_t mmio_addr;
>>      /* We might need to register the BAR before we actually have the memory.
>>       * So prepare a container MemoryRegion for the BAR immediately and
>>       * add a subregion when we have the memory.
>> @@ -346,8 +345,14 @@ static void close_guest_eventfds(IVShmemState *s, int posn)
>>      guest_curr_max = s->peers[posn].nb_eventfds;
>>
>>      for (i = 0; i<  guest_curr_max; i++) {
>> -        kvm_set_ioeventfd_mmio_long(s->peers[posn].eventfds[i],
>> -                    s->mmio_addr + DOORBELL, (posn<<  16) | i, 0);
>> +        if (ivshmem_has_feature(s, IVSHMEM_IOEVENTFD)) {
>> +            memory_region_del_eventfd(&s->ivshmem_mmio,
>> +                                     DOORBELL,
>> +                                     4,
>> +                                     true,
>> +                                     (posn<<  16) | i,
>> +                                     s->peers[posn].eventfds[i]);
>> +        }
>>          close(s->peers[posn].eventfds[i]);
>>      }
>>
>> @@ -355,22 +360,6 @@ static void close_guest_eventfds(IVShmemState *s, int posn)
>>      s->peers[posn].nb_eventfds = 0;
>>   }
>>
>> -static void setup_ioeventfds(IVShmemState *s) {
>> -
>> -    int i, j;
>> -
>> -    for (i = 0; i<= s->max_peer; i++) {
>> -        for (j = 0; j<  s->peers[i].nb_eventfds; j++) {
>> -            memory_region_add_eventfd(&s->ivshmem_mmio,
>> -                                      DOORBELL,
>> -                                      4,
>> -                                      true,
>> -                                      (i<<  16) | j,
>> -                                      s->peers[i].eventfds[j]);
>> -        }
>> -    }
>> -}
>> -
>>   /* this function increase the dynamic storage need to store data about other
>>   * guests */
>>   static void increase_dynamic_storage(IVShmemState *s, int new_min_size) {
>> @@ -491,10 +480,12 @@ static void ivshmem_read(void *opaque, const uint8_t * buf, int flags)
>>      }
>>
>>      if (ivshmem_has_feature(s, IVSHMEM_IOEVENTFD)) {
>> -        if (kvm_set_ioeventfd_mmio_long(incoming_fd, s->mmio_addr + DOORBELL,
>> -                        (incoming_posn<<  16) | guest_max_eventfd, 1)<  0) {
>> -            fprintf(stderr, "ivshmem: ioeventfd not available\n");
>> -        }
>> +        memory_region_add_eventfd(&s->ivshmem_mmio,
>> +                                  DOORBELL,
>> +                                  4,
>> +                                  true,
>> +                                  (incoming_posn<<  16) | guest_max_eventfd,
>> +                                  incoming_fd);
>>      }
>>
>>      return;
>> @@ -659,10 +650,6 @@ static int pci_ivshmem_init(PCIDevice *dev)
>>      memory_region_init_io(&s->ivshmem_mmio,&ivshmem_mmio_ops, s,
>>                            "ivshmem-mmio", IVSHMEM_REG_BAR_SIZE);
>>
>> -    if (ivshmem_has_feature(s, IVSHMEM_IOEVENTFD)) {
>> -        setup_ioeventfds(s);
>> -    }
>> -
>>      /* region for registers*/
>>      pci_register_bar(&s->dev, 0, PCI_BASE_ADDRESS_SPACE_MEMORY,
>>                       &s->ivshmem_mmio);
>> --
>> 1.7.1
>>
>>
> .
>

  reply	other threads:[~2011-11-25  9:14 UTC|newest]

Thread overview: 8+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2011-11-24 10:05 [Qemu-devel] [PATCH] ivshmem: fix guest unable to start with ioeventfd zanghongyong
2011-11-24 20:29 ` Cam Macdonell
2011-11-25  9:12   ` Zang Hongyong [this message]
2011-11-30  8:29 ` Zang Hongyong
2011-11-30 22:26   ` Cam Macdonell
2011-12-02  8:00     ` Cam Macdonell
2011-12-05 19:27       ` Cam Macdonell
2012-01-13 22:47 ` Cam Macdonell

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4ECF5BEB.3080403@huawei.com \
    --to=zanghongyong@huawei.com \
    --cc=avi@redhat.com \
    --cc=cam@cs.ualberta.ca \
    --cc=hanweidong@huawei.com \
    --cc=kvm@vger.kernel.org \
    --cc=louzhengwei@huawei.com \
    --cc=qemu-devel@nongnu.org \
    --cc=wusongwei@huawei.com \
    --cc=xiaowei.yang@huawei.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).