From: Christoph Hellwig <hch@lst.de>
To: Robin Murphy <robin.murphy@arm.com>
Cc: Joerg Roedel <joro@8bytes.org>,
Catalin Marinas <catalin.marinas@arm.com>,
Will Deacon <will.deacon@arm.com>,
Tom Lendacky <thomas.lendacky@amd.com>,
iommu@lists.linux-foundation.org,
linux-arm-kernel@lists.infradead.org,
linux-kernel@vger.kernel.org
Subject: [PATCH 13/21] dma-iommu: factor contiguous allocations into helpers
Date: Wed, 27 Mar 2019 09:04:40 +0100 [thread overview]
Message-ID: <20190327080448.5500-14-hch@lst.de> (raw)
In-Reply-To: <20190327080448.5500-1-hch@lst.de>
This keeps the code together and will simplify using it in different
ways.
Signed-off-by: Christoph Hellwig <hch@lst.de>
---
drivers/iommu/dma-iommu.c | 110 ++++++++++++++++++++------------------
1 file changed, 59 insertions(+), 51 deletions(-)
diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c
index 8ec69176673d..da2e0f4a63b6 100644
--- a/drivers/iommu/dma-iommu.c
+++ b/drivers/iommu/dma-iommu.c
@@ -458,6 +458,48 @@ static dma_addr_t __iommu_dma_map(struct device *dev, phys_addr_t phys,
return iova + iova_off;
}
+static void iommu_dma_free_contiguous(struct device *dev, size_t size,
+ struct page *page, dma_addr_t dma_handle)
+{
+ unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
+
+ __iommu_dma_unmap(iommu_get_domain_for_dev(dev), dma_handle, size);
+ if (!dma_release_from_contiguous(dev, page, count))
+ __free_pages(page, get_order(size));
+}
+
+
+static void *iommu_dma_alloc_contiguous(struct device *dev, size_t size,
+ dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs)
+{
+ bool coherent = dev_is_dma_coherent(dev);
+ int ioprot = dma_info_to_prot(DMA_BIDIRECTIONAL, coherent, attrs);
+ unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
+ unsigned int page_order = get_order(size);
+ struct page *page = NULL;
+
+ if (gfpflags_allow_blocking(gfp))
+ page = dma_alloc_from_contiguous(dev, count, page_order,
+ gfp & __GFP_NOWARN);
+
+ if (page)
+ memset(page_address(page), 0, PAGE_ALIGN(size));
+ else
+ page = alloc_pages(gfp, page_order);
+ if (!page)
+ return NULL;
+
+ *dma_handle = __iommu_dma_map(dev, page_to_phys(page), size, ioprot,
+ iommu_get_dma_domain(dev));
+ if (*dma_handle == DMA_MAPPING_ERROR) {
+ if (!dma_release_from_contiguous(dev, page, count))
+ __free_pages(page, page_order);
+ return NULL;
+ }
+
+ return page_address(page);
+}
+
static void __iommu_dma_free_pages(struct page **pages, int count)
{
while (count--)
@@ -754,19 +796,6 @@ static void iommu_dma_sync_sg_for_device(struct device *dev,
arch_sync_dma_for_device(dev, sg_phys(sg), sg->length, dir);
}
-static dma_addr_t __iommu_dma_map_page(struct device *dev, struct page *page,
- unsigned long offset, size_t size, int prot)
-{
- return __iommu_dma_map(dev, page_to_phys(page) + offset, size, prot,
- iommu_get_dma_domain(dev));
-}
-
-static void __iommu_dma_unmap_page(struct device *dev, dma_addr_t handle,
- size_t size, enum dma_data_direction dir, unsigned long attrs)
-{
- __iommu_dma_unmap(iommu_get_dma_domain(dev), handle, size);
-}
-
static dma_addr_t iommu_dma_map_page(struct device *dev, struct page *page,
unsigned long offset, size_t size, enum dma_data_direction dir,
unsigned long attrs)
@@ -991,7 +1020,6 @@ static void *iommu_dma_alloc(struct device *dev, size_t size,
dma_addr_t *handle, gfp_t gfp, unsigned long attrs)
{
bool coherent = dev_is_dma_coherent(dev);
- int ioprot = dma_info_to_prot(DMA_BIDIRECTIONAL, coherent, attrs);
size_t iosize = size;
void *addr;
@@ -1004,7 +1032,6 @@ static void *iommu_dma_alloc(struct device *dev, size_t size,
gfp |= __GFP_ZERO;
if (!gfpflags_allow_blocking(gfp)) {
- struct page *page;
/*
* In atomic context we can't remap anything, so we'll only
* get the virtually contiguous buffer we need by way of a
@@ -1013,44 +1040,27 @@ static void *iommu_dma_alloc(struct device *dev, size_t size,
if (!coherent)
return iommu_dma_alloc_pool(dev, iosize, handle, gfp,
attrs);
-
- page = alloc_pages(gfp, get_order(size));
- if (!page)
- return NULL;
-
- addr = page_address(page);
- *handle = __iommu_dma_map_page(dev, page, 0, iosize, ioprot);
- if (*handle == DMA_MAPPING_ERROR) {
- __free_pages(page, get_order(size));
- addr = NULL;
- }
+ return iommu_dma_alloc_contiguous(dev, iosize, handle, gfp,
+ attrs);
} else if (attrs & DMA_ATTR_FORCE_CONTIGUOUS) {
pgprot_t prot = arch_dma_mmap_pgprot(dev, PAGE_KERNEL, attrs);
struct page *page;
- page = dma_alloc_from_contiguous(dev, size >> PAGE_SHIFT,
- get_order(size), gfp & __GFP_NOWARN);
- if (!page)
+ addr = iommu_dma_alloc_contiguous(dev, iosize, handle, gfp,
+ attrs);
+ if (!addr)
return NULL;
+ page = virt_to_page(addr);
- *handle = __iommu_dma_map_page(dev, page, 0, iosize, ioprot);
- if (*handle == DMA_MAPPING_ERROR) {
- dma_release_from_contiguous(dev, page,
- size >> PAGE_SHIFT);
+ addr = dma_common_contiguous_remap(page, size, VM_USERMAP, prot,
+ __builtin_return_address(0));
+ if (!addr) {
+ iommu_dma_free_contiguous(dev, iosize, page, *handle);
return NULL;
}
- addr = dma_common_contiguous_remap(page, size, VM_USERMAP,
- prot,
- __builtin_return_address(0));
- if (addr) {
- if (!coherent)
- arch_dma_prep_coherent(page, iosize);
- memset(addr, 0, size);
- } else {
- __iommu_dma_unmap_page(dev, *handle, iosize, 0, attrs);
- dma_release_from_contiguous(dev, page,
- size >> PAGE_SHIFT);
- }
+
+ if (!coherent)
+ arch_dma_prep_coherent(page, iosize);
} else {
addr = iommu_dma_alloc_remap(dev, iosize, handle, gfp, attrs);
}
@@ -1077,16 +1087,14 @@ static void iommu_dma_free(struct device *dev, size_t size, void *cpu_addr,
if (dma_in_atomic_pool(cpu_addr, size)) {
iommu_dma_free_pool(dev, size, cpu_addr, handle);
} else if (attrs & DMA_ATTR_FORCE_CONTIGUOUS) {
- struct page *page = vmalloc_to_page(cpu_addr);
-
- __iommu_dma_unmap_page(dev, handle, iosize, 0, attrs);
- dma_release_from_contiguous(dev, page, size >> PAGE_SHIFT);
+ iommu_dma_free_contiguous(dev, iosize,
+ vmalloc_to_page(cpu_addr), handle);
dma_common_free_remap(cpu_addr, size, VM_USERMAP);
} else if (is_vmalloc_addr(cpu_addr)){
iommu_dma_free_remap(dev, iosize, cpu_addr, handle);
} else {
- __iommu_dma_unmap_page(dev, handle, iosize, 0, 0);
- __free_pages(virt_to_page(cpu_addr), get_order(size));
+ iommu_dma_free_contiguous(dev, iosize, virt_to_page(cpu_addr),
+ handle);
}
}
--
2.20.1
next prev parent reply other threads:[~2019-03-27 8:04 UTC|newest]
Thread overview: 97+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-03-27 8:04 implement generic dma_map_ops for IOMMUs v2 [rebase + resend] Christoph Hellwig
2019-03-27 8:04 ` [PATCH 01/21] arm64/iommu: handle non-remapped addresses in ->mmap and ->get_sgtable Christoph Hellwig
2019-04-05 17:16 ` Robin Murphy
2019-04-05 17:16 ` Robin Murphy
2019-03-27 8:04 ` [PATCH 02/21] arm64/iommu: improve mmap bounds checking Christoph Hellwig
2019-04-05 17:30 ` Robin Murphy
2019-04-05 17:30 ` Robin Murphy
2019-04-07 6:59 ` Christoph Hellwig
2019-04-07 6:59 ` Christoph Hellwig
2019-04-09 15:12 ` Robin Murphy
2019-04-09 15:12 ` Robin Murphy
2019-04-09 17:09 ` Christoph Hellwig
2019-04-09 17:09 ` Christoph Hellwig
[not found] ` <20190327080448.5500-1-hch-jcswGhMUV9g@public.gmane.org>
2019-03-27 8:04 ` [PATCH 03/21] dma-mapping: add a Kconfig symbol to indicated arch_dma_prep_coherent presence Christoph Hellwig
2019-04-05 17:41 ` Robin Murphy
2019-04-05 17:41 ` Robin Murphy
2019-03-27 8:04 ` [PATCH 07/21] dma-iommu: move the arm64 wrappers to common code Christoph Hellwig
2019-04-09 15:07 ` Robin Murphy
2019-04-09 15:07 ` Robin Murphy
2019-04-09 17:15 ` Christoph Hellwig
2019-04-09 17:15 ` Christoph Hellwig
[not found] ` <67573dd3-72c7-692d-bc1a-7edb49ff9551-5wv7dgnIgG8@public.gmane.org>
2019-04-09 17:23 ` Christoph Hellwig
2019-04-09 17:23 ` Christoph Hellwig
2019-04-09 17:33 ` Robin Murphy
2019-04-09 17:33 ` Robin Murphy
2019-03-27 8:04 ` [PATCH 04/21] dma-iommu: cleanup dma-iommu.h Christoph Hellwig
2019-04-05 17:42 ` Robin Murphy
2019-04-05 17:42 ` Robin Murphy
2019-04-09 17:10 ` Christoph Hellwig
2019-04-09 17:10 ` Christoph Hellwig
2019-03-27 8:04 ` [PATCH 05/21] dma-iommu: remove the flush_page callback Christoph Hellwig
2019-04-05 17:46 ` Robin Murphy
2019-04-05 17:46 ` Robin Murphy
2019-03-27 8:04 ` [PATCH 06/21] dma-iommu: use for_each_sg in iommu_dma_alloc Christoph Hellwig
2019-04-05 18:08 ` Robin Murphy
2019-04-05 18:08 ` Robin Murphy
2019-03-27 8:04 ` [PATCH 08/21] dma-iommu: refactor iommu_dma_mmap Christoph Hellwig
2019-04-09 15:29 ` Robin Murphy
2019-04-09 15:29 ` Robin Murphy
[not found] ` <adf139e2-a595-fa08-e287-f358c06d7e91-5wv7dgnIgG8@public.gmane.org>
2019-04-09 17:25 ` Christoph Hellwig
2019-04-09 17:25 ` Christoph Hellwig
2019-03-27 8:04 ` [PATCH 09/21] dma-iommu: refactor iommu_dma_get_sgtable Christoph Hellwig
2019-04-09 15:49 ` Robin Murphy
2019-04-09 15:49 ` Robin Murphy
2019-04-09 17:26 ` Christoph Hellwig
2019-04-09 17:26 ` Christoph Hellwig
2019-03-27 8:04 ` [PATCH 10/21] dma-iommu: move __iommu_dma_map Christoph Hellwig
2019-04-09 15:54 ` Robin Murphy
2019-04-09 15:54 ` Robin Murphy
2019-03-27 8:04 ` [PATCH 11/21] dma-iommu: refactor page array remap helpers Christoph Hellwig
2019-04-09 16:38 ` Robin Murphy
2019-04-09 16:38 ` Robin Murphy
2019-03-27 8:04 ` [PATCH 12/21] dma-iommu: factor atomic pool allocations into helpers Christoph Hellwig
2019-04-09 17:59 ` Robin Murphy
2019-04-09 17:59 ` Robin Murphy
2019-04-10 6:11 ` Christoph Hellwig
2019-04-10 6:11 ` Christoph Hellwig
2019-04-17 6:33 ` Christoph Hellwig
2019-04-17 6:33 ` Christoph Hellwig
2019-04-17 11:54 ` Robin Murphy
2019-04-17 11:54 ` Robin Murphy
2019-04-18 15:06 ` Robin Murphy
2019-04-18 15:06 ` Robin Murphy
2019-04-18 16:35 ` Christoph Hellwig
2019-04-18 16:35 ` Christoph Hellwig
2019-04-18 16:41 ` Robin Murphy
2019-04-18 16:41 ` Robin Murphy
2019-04-19 9:07 ` Christoph Hellwig
2019-04-19 9:07 ` Christoph Hellwig
2019-04-23 9:48 ` Robin Murphy
2019-04-23 9:48 ` Robin Murphy
2019-04-18 18:15 ` Robin Murphy
2019-04-18 18:15 ` Robin Murphy
2019-04-19 8:23 ` Christoph Hellwig
2019-04-19 8:23 ` Christoph Hellwig
[not found] ` <20190419082348.GA22299-jcswGhMUV9g@public.gmane.org>
2019-04-23 10:01 ` Robin Murphy
2019-04-23 10:01 ` Robin Murphy
2019-04-23 14:52 ` Christoph Hellwig
2019-04-23 14:52 ` Christoph Hellwig
2019-04-29 11:49 ` Christoph Hellwig
2019-04-29 11:49 ` Christoph Hellwig
2019-04-29 12:02 ` Robin Murphy
2019-04-29 12:02 ` Robin Murphy
2019-04-22 18:03 ` Christoph Hellwig
2019-04-22 18:03 ` Christoph Hellwig
2019-03-27 8:04 ` Christoph Hellwig [this message]
2019-03-27 8:04 ` [PATCH 14/21] dma-iommu: refactor iommu_dma_free Christoph Hellwig
2019-03-27 8:04 ` [PATCH 15/21] dma-iommu: don't remap contiguous allocations for coherent devices Christoph Hellwig
2019-03-27 8:04 ` [PATCH 16/21] dma-iommu: factor contiguous remapped allocations into helpers Christoph Hellwig
2019-03-27 8:04 ` [PATCH 17/21] dma-iommu: refactor iommu_dma_alloc Christoph Hellwig
2019-03-27 8:04 ` [PATCH 18/21] dma-iommu: don't depend on CONFIG_DMA_DIRECT_REMAP Christoph Hellwig
2019-03-27 8:04 ` [PATCH 19/21] dma-iommu: switch copyright boilerplace to SPDX Christoph Hellwig
2019-03-27 8:04 ` [PATCH 20/21] arm64: switch copyright boilerplace to SPDX in dma-mapping.c Christoph Hellwig
2019-04-01 6:28 ` Mukesh Ojha
2019-04-01 9:39 ` Robin Murphy
2019-03-27 8:04 ` [PATCH 21/21] arm64: trim includes " Christoph Hellwig
-- strict thread matches above, loose matches on Subject: below --
2019-02-13 18:28 implement generic dma_map_ops for IOMMUs v2 Christoph Hellwig
2019-02-13 18:29 ` [PATCH 13/21] dma-iommu: factor contiguous allocations into helpers Christoph Hellwig
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20190327080448.5500-14-hch@lst.de \
--to=hch@lst.de \
--cc=catalin.marinas@arm.com \
--cc=iommu@lists.linux-foundation.org \
--cc=joro@8bytes.org \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-kernel@vger.kernel.org \
--cc=robin.murphy@arm.com \
--cc=thomas.lendacky@amd.com \
--cc=will.deacon@arm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).