From: Alexander Graf <agraf@suse.de>
To: kvm-ppc@vger.kernel.org
Cc: "kvm@vger.kernel.org mailing list" <kvm@vger.kernel.org>,
Gleb Natapov <gleb@redhat.com>,
Marcelo Tosatti <mtosatti@redhat.com>,
Scott Wood <scottwood@freescale.com>
Subject: [PATCH 2/6] kvm/ppc/mpic: fix mmio region lists when multiple guests used
Date: Thu, 02 May 2013 13:32:47 +0000 [thread overview]
Message-ID: <1367501571-15509-3-git-send-email-agraf@suse.de> (raw)
In-Reply-To: <1367501571-15509-1-git-send-email-agraf@suse.de>
From: Scott Wood <scottwood@freescale.com>
Keeping a linked list of statically defined objects doesn't work
very well when we have multiple guests. :-P
Switch to an array of constant objects. This fixes a hang when
multiple guests are used.
Signed-off-by: Scott Wood <scottwood@freescale.com>
[agraf: remove struct list_head from mem_reg]
Signed-off-by: Alexander Graf <agraf@suse.de>
---
arch/powerpc/kvm/mpic.c | 53 +++++++++++++++++++++++++++++--------------------
1 file changed, 32 insertions(+), 21 deletions(-)
diff --git a/arch/powerpc/kvm/mpic.c b/arch/powerpc/kvm/mpic.c
index f3148f8..886ebf6 100644
--- a/arch/powerpc/kvm/mpic.c
+++ b/arch/powerpc/kvm/mpic.c
@@ -184,11 +184,14 @@ struct irq_dest {
uint32_t outputs_active[NUM_OUTPUTS];
};
+#define MAX_MMIO_REGIONS 10
+
struct openpic {
struct kvm *kvm;
struct kvm_device *dev;
struct kvm_io_device mmio;
- struct list_head mmio_regions;
+ const struct mem_reg *mmio_regions[MAX_MMIO_REGIONS];
+ int num_mmio_regions;
atomic_t users;
gpa_t reg_base;
@@ -1238,62 +1241,71 @@ static int openpic_cpu_read(void *opaque, gpa_t addr, u32 *ptr)
}
struct mem_reg {
- struct list_head list;
int (*read)(void *opaque, gpa_t addr, u32 *ptr);
int (*write)(void *opaque, gpa_t addr, u32 val);
gpa_t start_addr;
int size;
};
-static struct mem_reg openpic_gbl_mmio = {
+static const struct mem_reg openpic_gbl_mmio = {
.write = openpic_gbl_write,
.read = openpic_gbl_read,
.start_addr = OPENPIC_GLB_REG_START,
.size = OPENPIC_GLB_REG_SIZE,
};
-static struct mem_reg openpic_tmr_mmio = {
+static const struct mem_reg openpic_tmr_mmio = {
.write = openpic_tmr_write,
.read = openpic_tmr_read,
.start_addr = OPENPIC_TMR_REG_START,
.size = OPENPIC_TMR_REG_SIZE,
};
-static struct mem_reg openpic_cpu_mmio = {
+static const struct mem_reg openpic_cpu_mmio = {
.write = openpic_cpu_write,
.read = openpic_cpu_read,
.start_addr = OPENPIC_CPU_REG_START,
.size = OPENPIC_CPU_REG_SIZE,
};
-static struct mem_reg openpic_src_mmio = {
+static const struct mem_reg openpic_src_mmio = {
.write = openpic_src_write,
.read = openpic_src_read,
.start_addr = OPENPIC_SRC_REG_START,
.size = OPENPIC_SRC_REG_SIZE,
};
-static struct mem_reg openpic_msi_mmio = {
+static const struct mem_reg openpic_msi_mmio = {
.read = openpic_msi_read,
.write = openpic_msi_write,
.start_addr = OPENPIC_MSI_REG_START,
.size = OPENPIC_MSI_REG_SIZE,
};
-static struct mem_reg openpic_summary_mmio = {
+static const struct mem_reg openpic_summary_mmio = {
.read = openpic_summary_read,
.write = openpic_summary_write,
.start_addr = OPENPIC_SUMMARY_REG_START,
.size = OPENPIC_SUMMARY_REG_SIZE,
};
+static void add_mmio_region(struct openpic *opp, const struct mem_reg *mr)
+{
+ if (opp->num_mmio_regions >= MAX_MMIO_REGIONS) {
+ WARN(1, "kvm mpic: too many mmio regions\n");
+ return;
+ }
+
+ opp->mmio_regions[opp->num_mmio_regions++] = mr;
+}
+
static void fsl_common_init(struct openpic *opp)
{
int i;
int virq = MAX_SRC;
- list_add(&openpic_msi_mmio.list, &opp->mmio_regions);
- list_add(&openpic_summary_mmio.list, &opp->mmio_regions);
+ add_mmio_region(opp, &openpic_msi_mmio);
+ add_mmio_region(opp, &openpic_summary_mmio);
opp->vid = VID_REVISION_1_2;
opp->vir = VIR_GENERIC;
@@ -1330,10 +1342,10 @@ static void fsl_common_init(struct openpic *opp)
static int kvm_mpic_read_internal(struct openpic *opp, gpa_t addr, u32 *ptr)
{
- struct list_head *node;
+ int i;
- list_for_each(node, &opp->mmio_regions) {
- struct mem_reg *mr = list_entry(node, struct mem_reg, list);
+ for (i = 0; i < opp->num_mmio_regions; i++) {
+ const struct mem_reg *mr = opp->mmio_regions[i];
if (mr->start_addr > addr || addr >= mr->start_addr + mr->size)
continue;
@@ -1346,10 +1358,10 @@ static int kvm_mpic_read_internal(struct openpic *opp, gpa_t addr, u32 *ptr)
static int kvm_mpic_write_internal(struct openpic *opp, gpa_t addr, u32 val)
{
- struct list_head *node;
+ int i;
- list_for_each(node, &opp->mmio_regions) {
- struct mem_reg *mr = list_entry(node, struct mem_reg, list);
+ for (i = 0; i < opp->num_mmio_regions; i++) {
+ const struct mem_reg *mr = opp->mmio_regions[i];
if (mr->start_addr > addr || addr >= mr->start_addr + mr->size)
continue;
@@ -1660,11 +1672,10 @@ static int mpic_create(struct kvm_device *dev, u32 type)
opp->model = type;
spin_lock_init(&opp->lock);
- INIT_LIST_HEAD(&opp->mmio_regions);
- list_add(&openpic_gbl_mmio.list, &opp->mmio_regions);
- list_add(&openpic_tmr_mmio.list, &opp->mmio_regions);
- list_add(&openpic_src_mmio.list, &opp->mmio_regions);
- list_add(&openpic_cpu_mmio.list, &opp->mmio_regions);
+ add_mmio_region(opp, &openpic_gbl_mmio);
+ add_mmio_region(opp, &openpic_tmr_mmio);
+ add_mmio_region(opp, &openpic_src_mmio);
+ add_mmio_region(opp, &openpic_cpu_mmio);
switch (opp->model) {
case KVM_DEV_TYPE_FSL_MPIC_20:
--
1.8.1.4
next prev parent reply other threads:[~2013-05-02 13:32 UTC|newest]
Thread overview: 8+ messages / expand[flat|nested] mbox.gz Atom feed top
2013-05-02 13:32 [PULL 0/6] ppc patch queue 2013-05-02 for 3.10 Alexander Graf
2013-05-02 13:32 ` [PATCH 1/6] kvm/ppc/mpic: remove default routes from documentation Alexander Graf
2013-05-02 13:32 ` Alexander Graf [this message]
2013-05-02 13:32 ` [PATCH 3/6] kvm/ppc/mpic: remove users Alexander Graf
2013-05-02 13:32 ` [PATCH 4/6] kvm/ppc: Hold srcu lock when calling kvm_io_bus_read/write Alexander Graf
2013-05-02 13:32 ` [PATCH 5/6] kvm/ppc/mpic: fix missing unlock in set_base_addr() Alexander Graf
2013-05-02 13:32 ` [PATCH 6/6] KVM: PPC: Book3S: Add API for in-kernel XICS emulation Alexander Graf
2013-05-03 0:09 ` [PULL 0/6] ppc patch queue 2013-05-02 for 3.10 Marcelo Tosatti
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1367501571-15509-3-git-send-email-agraf@suse.de \
--to=agraf@suse.de \
--cc=gleb@redhat.com \
--cc=kvm-ppc@vger.kernel.org \
--cc=kvm@vger.kernel.org \
--cc=mtosatti@redhat.com \
--cc=scottwood@freescale.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox