From: Anton Blanchard <anton@samba.org>
To: benh@kernel.crashing.org, paulus@samba.org, olof@lixom.net,
michael@ellerman.id.au
Cc: linuxppc-dev@lists.ozlabs.org
Subject: [PATCH 4/5] powerpc: iommu: Push spinlock into iommu_range_alloc and __iommu_free
Date: Mon, 4 Jun 2012 15:44:25 +1000 [thread overview]
Message-ID: <20120604154425.147c3fc5@kryten> (raw)
In-Reply-To: <20120604154213.173feecb@kryten>
In preparation for IOMMU pools, push the spinlock into
iommu_range_alloc and __iommu_free.
Signed-off-by: Anton Blanchard <anton@samba.org>
---
Index: linux-build/arch/powerpc/kernel/iommu.c
===================================================================
--- linux-build.orig/arch/powerpc/kernel/iommu.c 2012-06-04 10:38:43.813334034 +1000
+++ linux-build/arch/powerpc/kernel/iommu.c 2012-06-04 10:38:45.305365604 +1000
@@ -71,6 +71,7 @@ static unsigned long iommu_range_alloc(s
int pass = 0;
unsigned long align_mask;
unsigned long boundary_size;
+ unsigned long flags;
align_mask = 0xffffffffffffffffl >> (64 - align_order);
@@ -83,6 +84,8 @@ static unsigned long iommu_range_alloc(s
return DMA_ERROR_CODE;
}
+ spin_lock_irqsave(&(tbl->it_lock), flags);
+
if (handle && *handle)
start = *handle;
else
@@ -136,6 +139,7 @@ static unsigned long iommu_range_alloc(s
goto again;
} else {
/* Third failure, give up */
+ spin_unlock_irqrestore(&(tbl->it_lock), flags);
return DMA_ERROR_CODE;
}
}
@@ -156,6 +160,7 @@ static unsigned long iommu_range_alloc(s
if (handle)
*handle = end;
+ spin_unlock_irqrestore(&(tbl->it_lock), flags);
return n;
}
@@ -165,13 +170,11 @@ static dma_addr_t iommu_alloc(struct dev
unsigned long mask, unsigned int align_order,
struct dma_attrs *attrs)
{
- unsigned long entry, flags;
+ unsigned long entry;
dma_addr_t ret = DMA_ERROR_CODE;
int build_fail;
- spin_lock_irqsave(&(tbl->it_lock), flags);
entry = iommu_range_alloc(dev, tbl, npages, NULL, mask, align_order);
- spin_unlock_irqrestore(&(tbl->it_lock), flags);
if (unlikely(entry == DMA_ERROR_CODE))
return DMA_ERROR_CODE;
@@ -232,23 +235,6 @@ static bool iommu_free_check(struct iomm
return true;
}
-static void __iommu_free_locked(struct iommu_table *tbl, dma_addr_t dma_addr,
- unsigned int npages)
-{
- unsigned long entry, free_entry;
-
- BUG_ON(!spin_is_locked(&tbl->it_lock));
-
- entry = dma_addr >> IOMMU_PAGE_SHIFT;
- free_entry = entry - tbl->it_offset;
-
- if (!iommu_free_check(tbl, dma_addr, npages))
- return;
-
- ppc_md.tce_free(tbl, entry, npages);
- bitmap_clear(tbl->it_map, free_entry, npages);
-}
-
static void __iommu_free(struct iommu_table *tbl, dma_addr_t dma_addr,
unsigned int npages)
{
@@ -287,7 +273,6 @@ int iommu_map_sg(struct device *dev, str
struct dma_attrs *attrs)
{
dma_addr_t dma_next = 0, dma_addr;
- unsigned long flags;
struct scatterlist *s, *outs, *segstart;
int outcount, incount, i, build_fail = 0;
unsigned int align;
@@ -309,8 +294,6 @@ int iommu_map_sg(struct device *dev, str
DBG("sg mapping %d elements:\n", nelems);
- spin_lock_irqsave(&(tbl->it_lock), flags);
-
max_seg_size = dma_get_max_seg_size(dev);
for_each_sg(sglist, s, nelems, i) {
unsigned long vaddr, npages, entry, slen;
@@ -393,8 +376,6 @@ int iommu_map_sg(struct device *dev, str
if (ppc_md.tce_flush)
ppc_md.tce_flush(tbl);
- spin_unlock_irqrestore(&(tbl->it_lock), flags);
-
DBG("mapped %d elements:\n", outcount);
/* For the sake of iommu_unmap_sg, we clear out the length in the
@@ -419,14 +400,13 @@ int iommu_map_sg(struct device *dev, str
vaddr = s->dma_address & IOMMU_PAGE_MASK;
npages = iommu_num_pages(s->dma_address, s->dma_length,
IOMMU_PAGE_SIZE);
- __iommu_free_locked(tbl, vaddr, npages);
+ __iommu_free(tbl, vaddr, npages);
s->dma_address = DMA_ERROR_CODE;
s->dma_length = 0;
}
if (s == outs)
break;
}
- spin_unlock_irqrestore(&(tbl->it_lock), flags);
return 0;
}
@@ -436,15 +416,12 @@ void iommu_unmap_sg(struct iommu_table *
struct dma_attrs *attrs)
{
struct scatterlist *sg;
- unsigned long flags;
BUG_ON(direction == DMA_NONE);
if (!tbl)
return;
- spin_lock_irqsave(&(tbl->it_lock), flags);
-
sg = sglist;
while (nelems--) {
unsigned int npages;
@@ -454,7 +431,7 @@ void iommu_unmap_sg(struct iommu_table *
break;
npages = iommu_num_pages(dma_handle, sg->dma_length,
IOMMU_PAGE_SIZE);
- __iommu_free_locked(tbl, dma_handle, npages);
+ __iommu_free(tbl, dma_handle, npages);
sg = sg_next(sg);
}
@@ -464,8 +441,6 @@ void iommu_unmap_sg(struct iommu_table *
*/
if (ppc_md.tce_flush)
ppc_md.tce_flush(tbl);
-
- spin_unlock_irqrestore(&(tbl->it_lock), flags);
}
static void iommu_table_clear(struct iommu_table *tbl)
next prev parent reply other threads:[~2012-06-04 5:44 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-06-04 5:42 [PATCH 1/5] powerpc/pseries: Disable interrupts around IOMMU percpu data accesses Anton Blanchard
2012-06-04 5:43 ` [PATCH 2/5] powerpc: iommu: Reduce spinlock coverage in iommu_alloc and iommu_free Anton Blanchard
2012-06-04 5:43 ` [PATCH 3/5] powerpc: iommu: Reduce spinlock coverage in iommu_free Anton Blanchard
2012-06-04 5:44 ` Anton Blanchard [this message]
2012-06-04 5:45 ` [PATCH 5/5] powerpc: iommu: Implement IOMMU pools to improve multiqueue adapter performance Anton Blanchard
2012-06-08 2:43 ` Michael Ellerman
2012-06-08 4:02 ` Anton Blanchard
2012-06-08 4:03 ` Michael Ellerman
2012-06-08 4:14 ` Anton Blanchard
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20120604154425.147c3fc5@kryten \
--to=anton@samba.org \
--cc=benh@kernel.crashing.org \
--cc=linuxppc-dev@lists.ozlabs.org \
--cc=michael@ellerman.id.au \
--cc=olof@lixom.net \
--cc=paulus@samba.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).