From: Leon Romanovsky <leon-DgEjT+Ai2ygdnm+yROfE0A@public.gmane.org>
To: "Wei Hu (Xavier)" <xavier.huwei-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
Cc: dledford-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org,
linux-rdma-u79uwXL29TY76Z2rM5mHXA@public.gmane.org,
lijun_nudt-9Onoh4P/yGk@public.gmane.org,
oulijun-hv44wF8Li93QT0dZR+AlfA@public.gmane.org,
charles.chenxin-hv44wF8Li93QT0dZR+AlfA@public.gmane.org,
liuyixian-hv44wF8Li93QT0dZR+AlfA@public.gmane.org,
xushaobo2-hv44wF8Li93QT0dZR+AlfA@public.gmane.org,
zhangxiping3-hv44wF8Li93QT0dZR+AlfA@public.gmane.org,
xavier.huwei-WVlzvzqoTvw@public.gmane.org,
linuxarm-hv44wF8Li93QT0dZR+AlfA@public.gmane.org,
linux-kernel-u79uwXL29TY76Z2rM5mHXA@public.gmane.org,
shaobohsu-9Onoh4P/yGk@public.gmane.org,
shaoboxu-WVlzvzqoTvw@public.gmane.org
Subject: Re: [PATCH for-next 2/4] RDMA/hns: Add IOMMU enable support in hip08
Date: Sat, 30 Sep 2017 19:10:23 +0300 [thread overview]
Message-ID: <20170930161023.GI2965@mtr-leonro.local> (raw)
In-Reply-To: <1506763741-81429-3-git-send-email-xavier.huwei-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
[-- Attachment #1: Type: text/plain, Size: 6802 bytes --]
On Sat, Sep 30, 2017 at 05:28:59PM +0800, Wei Hu (Xavier) wrote:
> If the IOMMU is enabled, the length of sg obtained from
> __iommu_map_sg_attrs is not 4kB. When the IOVA is set with the sg
> dma address, the IOVA will not be page continuous. and the VA
> returned from dma_alloc_coherent is a vmalloc address. However,
> the VA obtained by the page_address is a discontinuous VA. Under
> these circumstances, the IOVA should be calculated based on the
> sg length, and record the VA returned from dma_alloc_coherent
> in the struct of hem.
>
> Signed-off-by: Wei Hu (Xavier) <xavier.huwei-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
> Signed-off-by: Shaobo Xu <xushaobo2-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
> Signed-off-by: Lijun Ou <oulijun-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
> ---
Doug,
I didn't invest time in reviewing it, but having "is_vmalloc_addr" in
driver code to deal with dma_alloc_coherent is most probably wrong.
Thanks
> drivers/infiniband/hw/hns/hns_roce_alloc.c | 5 ++++-
> drivers/infiniband/hw/hns/hns_roce_hem.c | 30 +++++++++++++++++++++++++++---
> drivers/infiniband/hw/hns/hns_roce_hem.h | 6 ++++++
> drivers/infiniband/hw/hns/hns_roce_hw_v2.c | 22 +++++++++++++++-------
> 4 files changed, 52 insertions(+), 11 deletions(-)
>
> diff --git a/drivers/infiniband/hw/hns/hns_roce_alloc.c b/drivers/infiniband/hw/hns/hns_roce_alloc.c
> index 3e4c525..a69cd4b 100644
> --- a/drivers/infiniband/hw/hns/hns_roce_alloc.c
> +++ b/drivers/infiniband/hw/hns/hns_roce_alloc.c
> @@ -243,7 +243,10 @@ int hns_roce_buf_alloc(struct hns_roce_dev *hr_dev, u32 size, u32 max_direct,
> goto err_free;
>
> for (i = 0; i < buf->nbufs; ++i)
> - pages[i] = virt_to_page(buf->page_list[i].buf);
> + pages[i] =
> + is_vmalloc_addr(buf->page_list[i].buf) ?
> + vmalloc_to_page(buf->page_list[i].buf) :
> + virt_to_page(buf->page_list[i].buf);
>
> buf->direct.buf = vmap(pages, buf->nbufs, VM_MAP,
> PAGE_KERNEL);
> diff --git a/drivers/infiniband/hw/hns/hns_roce_hem.c b/drivers/infiniband/hw/hns/hns_roce_hem.c
> index 8388ae2..4a3d1d4 100644
> --- a/drivers/infiniband/hw/hns/hns_roce_hem.c
> +++ b/drivers/infiniband/hw/hns/hns_roce_hem.c
> @@ -200,6 +200,7 @@ static struct hns_roce_hem *hns_roce_alloc_hem(struct hns_roce_dev *hr_dev,
> gfp_t gfp_mask)
> {
> struct hns_roce_hem_chunk *chunk = NULL;
> + struct hns_roce_vmalloc *vmalloc;
> struct hns_roce_hem *hem;
> struct scatterlist *mem;
> int order;
> @@ -227,6 +228,7 @@ static struct hns_roce_hem *hns_roce_alloc_hem(struct hns_roce_dev *hr_dev,
> sg_init_table(chunk->mem, HNS_ROCE_HEM_CHUNK_LEN);
> chunk->npages = 0;
> chunk->nsg = 0;
> + memset(chunk->vmalloc, 0, sizeof(chunk->vmalloc));
> list_add_tail(&chunk->list, &hem->chunk_list);
> }
>
> @@ -243,7 +245,15 @@ static struct hns_roce_hem *hns_roce_alloc_hem(struct hns_roce_dev *hr_dev,
> if (!buf)
> goto fail;
>
> - sg_set_buf(mem, buf, PAGE_SIZE << order);
> + if (is_vmalloc_addr(buf)) {
> + vmalloc = &chunk->vmalloc[chunk->npages];
> + vmalloc->is_vmalloc_addr = true;
> + vmalloc->vmalloc_addr = buf;
> + sg_set_page(mem, vmalloc_to_page(buf),
> + PAGE_SIZE << order, offset_in_page(buf));
> + } else {
> + sg_set_buf(mem, buf, PAGE_SIZE << order);
> + }
> WARN_ON(mem->offset);
> sg_dma_len(mem) = PAGE_SIZE << order;
>
> @@ -262,17 +272,25 @@ static struct hns_roce_hem *hns_roce_alloc_hem(struct hns_roce_dev *hr_dev,
> void hns_roce_free_hem(struct hns_roce_dev *hr_dev, struct hns_roce_hem *hem)
> {
> struct hns_roce_hem_chunk *chunk, *tmp;
> + void *cpu_addr;
> int i;
>
> if (!hem)
> return;
>
> list_for_each_entry_safe(chunk, tmp, &hem->chunk_list, list) {
> - for (i = 0; i < chunk->npages; ++i)
> + for (i = 0; i < chunk->npages; ++i) {
> + if (chunk->vmalloc[i].is_vmalloc_addr)
> + cpu_addr = chunk->vmalloc[i].vmalloc_addr;
> + else
> + cpu_addr =
> + lowmem_page_address(sg_page(&chunk->mem[i]));
> +
> dma_free_coherent(hr_dev->dev,
> chunk->mem[i].length,
> - lowmem_page_address(sg_page(&chunk->mem[i])),
> + cpu_addr,
> sg_dma_address(&chunk->mem[i]));
> + }
> kfree(chunk);
> }
>
> @@ -774,6 +792,12 @@ void *hns_roce_table_find(struct hns_roce_dev *hr_dev,
>
> if (chunk->mem[i].length > (u32)offset) {
> page = sg_page(&chunk->mem[i]);
> + if (chunk->vmalloc[i].is_vmalloc_addr) {
> + mutex_unlock(&table->mutex);
> + return page ?
> + chunk->vmalloc[i].vmalloc_addr
> + + offset : NULL;
> + }
> goto out;
> }
> offset -= chunk->mem[i].length;
> diff --git a/drivers/infiniband/hw/hns/hns_roce_hem.h b/drivers/infiniband/hw/hns/hns_roce_hem.h
> index af28bbf..62d712a 100644
> --- a/drivers/infiniband/hw/hns/hns_roce_hem.h
> +++ b/drivers/infiniband/hw/hns/hns_roce_hem.h
> @@ -72,11 +72,17 @@ enum {
> HNS_ROCE_HEM_PAGE_SIZE = 1 << HNS_ROCE_HEM_PAGE_SHIFT,
> };
>
> +struct hns_roce_vmalloc {
> + bool is_vmalloc_addr;
> + void *vmalloc_addr;
> +};
> +
> struct hns_roce_hem_chunk {
> struct list_head list;
> int npages;
> int nsg;
> struct scatterlist mem[HNS_ROCE_HEM_CHUNK_LEN];
> + struct hns_roce_vmalloc vmalloc[HNS_ROCE_HEM_CHUNK_LEN];
> };
>
> struct hns_roce_hem {
> diff --git a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
> index b99d70a..9e19bf1 100644
> --- a/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
> +++ b/drivers/infiniband/hw/hns/hns_roce_hw_v2.c
> @@ -1093,9 +1093,11 @@ static int hns_roce_v2_write_mtpt(void *mb_buf, struct hns_roce_mr *mr,
> {
> struct hns_roce_v2_mpt_entry *mpt_entry;
> struct scatterlist *sg;
> + u64 page_addr = 0;
> u64 *pages;
> + int i = 0, j = 0;
> + int len = 0;
> int entry;
> - int i;
>
> mpt_entry = mb_buf;
> memset(mpt_entry, 0, sizeof(*mpt_entry));
> @@ -1153,14 +1155,20 @@ static int hns_roce_v2_write_mtpt(void *mb_buf, struct hns_roce_mr *mr,
>
> i = 0;
> for_each_sg(mr->umem->sg_head.sgl, sg, mr->umem->nmap, entry) {
> - pages[i] = ((u64)sg_dma_address(sg)) >> 6;
> -
> - /* Record the first 2 entry directly to MTPT table */
> - if (i >= HNS_ROCE_V2_MAX_INNER_MTPT_NUM - 1)
> - break;
> - i++;
> + len = sg_dma_len(sg) >> PAGE_SHIFT;
> + for (j = 0; j < len; ++j) {
> + page_addr = sg_dma_address(sg) +
> + (j << mr->umem->page_shift);
> + pages[i] = page_addr >> 6;
> +
> + /* Record the first 2 entry directly to MTPT table */
> + if (i >= HNS_ROCE_V2_MAX_INNER_MTPT_NUM - 1)
> + goto found;
> + i++;
> + }
> }
>
> +found:
> mpt_entry->pa0_l = cpu_to_le32(lower_32_bits(pages[0]));
> roce_set_field(mpt_entry->byte_56_pa0_h, V2_MPT_BYTE_56_PA0_H_M,
> V2_MPT_BYTE_56_PA0_H_S,
> --
> 1.9.1
>
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 833 bytes --]
next prev parent reply other threads:[~2017-09-30 16:10 UTC|newest]
Thread overview: 25+ messages / expand[flat|nested] mbox.gz Atom feed top
2017-09-30 9:28 [PATCH for-next 0/4] Add Features & Code improvements for hip08 Wei Hu (Xavier)
2017-09-30 9:28 ` [PATCH for-next 2/4] RDMA/hns: Add IOMMU enable support in hip08 Wei Hu (Xavier)
[not found] ` <1506763741-81429-3-git-send-email-xavier.huwei-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
2017-09-30 16:10 ` Leon Romanovsky [this message]
2017-10-12 12:31 ` Wei Hu (Xavier)
[not found] ` <59DF60A3.7080803-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
2017-10-12 12:59 ` Robin Murphy
2017-11-01 7:46 ` Wei Hu (Xavier)
[not found] ` <59F97BBE.5070207-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
2017-11-01 12:26 ` Robin Murphy
[not found] ` <fc7433af-4fa7-6b78-6bec-26941a427002-5wv7dgnIgG8@public.gmane.org>
2017-11-07 2:45 ` Wei Hu (Xavier)
[not found] ` <5A011E49.6060407-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
2017-11-07 6:32 ` Leon Romanovsky
[not found] ` <20171107063209.GA18825-U/DQcQFIOTAAJjI8aNfphQ@public.gmane.org>
2017-11-09 1:17 ` Wei Hu (Xavier)
2017-11-07 15:48 ` Jason Gunthorpe
[not found] ` <20171107154838.GC21466-uk2M96/98Pc@public.gmane.org>
2017-11-07 15:58 ` Christoph Hellwig
[not found] ` <20171107155805.GA24082-wEGCiKHe2LqWVfeAwA7xHQ@public.gmane.org>
2017-11-07 16:03 ` Jason Gunthorpe
2017-11-09 1:26 ` Wei Hu (Xavier)
2017-11-09 1:30 ` Wei Hu (Xavier)
2017-11-09 1:36 ` Wei Hu (Xavier)
2017-10-12 14:54 ` Leon Romanovsky
2017-10-18 8:42 ` Wei Hu (Xavier)
[not found] ` <59E713EE.5040703-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
2017-10-18 9:12 ` Wei Hu (Xavier)
[not found] ` <59E71AE2.6080202-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
2017-10-18 14:23 ` Leon Romanovsky
[not found] ` <1506763741-81429-1-git-send-email-xavier.huwei-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
2017-09-30 9:28 ` [PATCH for-next 1/4] RDMA/hns: Support WQE/CQE/PBL page size configurable feature " Wei Hu (Xavier)
2017-09-30 9:29 ` [PATCH for-next 3/4] RDMA/hns: Update the IRRL table chunk size " Wei Hu (Xavier)
[not found] ` <1506763741-81429-4-git-send-email-xavier.huwei-hv44wF8Li93QT0dZR+AlfA@public.gmane.org>
2017-10-01 5:40 ` Leon Romanovsky
2017-10-17 11:40 ` Wei Hu (Xavier)
2017-09-30 9:29 ` [PATCH for-next 4/4] RDMA/hns: Update the PD&CQE&MTT specification " Wei Hu (Xavier)
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20170930161023.GI2965@mtr-leonro.local \
--to=leon-dgejt+ai2ygdnm+yrofe0a@public.gmane.org \
--cc=charles.chenxin-hv44wF8Li93QT0dZR+AlfA@public.gmane.org \
--cc=dledford-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org \
--cc=lijun_nudt-9Onoh4P/yGk@public.gmane.org \
--cc=linux-kernel-u79uwXL29TY76Z2rM5mHXA@public.gmane.org \
--cc=linux-rdma-u79uwXL29TY76Z2rM5mHXA@public.gmane.org \
--cc=linuxarm-hv44wF8Li93QT0dZR+AlfA@public.gmane.org \
--cc=liuyixian-hv44wF8Li93QT0dZR+AlfA@public.gmane.org \
--cc=oulijun-hv44wF8Li93QT0dZR+AlfA@public.gmane.org \
--cc=shaobohsu-9Onoh4P/yGk@public.gmane.org \
--cc=shaoboxu-WVlzvzqoTvw@public.gmane.org \
--cc=xavier.huwei-WVlzvzqoTvw@public.gmane.org \
--cc=xavier.huwei-hv44wF8Li93QT0dZR+AlfA@public.gmane.org \
--cc=xushaobo2-hv44wF8Li93QT0dZR+AlfA@public.gmane.org \
--cc=zhangxiping3-hv44wF8Li93QT0dZR+AlfA@public.gmane.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox