From: Jerome Glisse <jglisse@redhat.com>
To: Dan Williams <dan.j.williams@intel.com>
Cc: akpm@linux-foundation.org, Christoph Hellwig <hch@lst.de>,
Logan Gunthorpe <logang@deltatee.com>,
alexander.h.duyck@intel.com, linux-mm@kvack.org,
linux-kernel@vger.kernel.org
Subject: Re: [PATCH v5 6/7] mm, hmm: Replace hmm_devmem_pages_create() with devm_memremap_pages()
Date: Tue, 18 Sep 2018 16:35:37 -0400 [thread overview]
Message-ID: <20180918203537.GF14689@redhat.com> (raw)
In-Reply-To: <153680535314.453305.11205770267271657025.stgit@dwillia2-desk3.amr.corp.intel.com>
On Wed, Sep 12, 2018 at 07:22:33PM -0700, Dan Williams wrote:
> Commit e8d513483300 "memremap: change devm_memremap_pages interface to
> use struct dev_pagemap" refactored devm_memremap_pages() to allow a
> dev_pagemap instance to be supplied. Passing in a dev_pagemap interface
> simplifies the design of pgmap type drivers in that they can rely on
> container_of() to lookup any private data associated with the given
> dev_pagemap instance.
>
> In addition to the cleanups this also gives hmm users multi-order-radix
> improvements that arrived with commit ab1b597ee0e4 "mm,
> devm_memremap_pages: use multi-order radix for ZONE_DEVICE lookups"
>
> As part of the conversion to the devm_memremap_pages() method of
> handling the percpu_ref relative to when pages are put, the percpu_ref
> completion needs to move to hmm_devmem_ref_exit(). See commit
> 71389703839e ("mm, zone_device: Replace {get, put}_zone_device_page...")
> for details.
>
> Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Jerome Glisse <jglisse@redhat.com>
> Cc: Logan Gunthorpe <logang@deltatee.com>
> Signed-off-by: Dan Williams <dan.j.williams@intel.com>
> ---
> mm/hmm.c | 194 ++++++++------------------------------------------------------
> 1 file changed, 26 insertions(+), 168 deletions(-)
>
> diff --git a/mm/hmm.c b/mm/hmm.c
> index ec1d9eccf176..c6cab5205b99 100644
> --- a/mm/hmm.c
> +++ b/mm/hmm.c
> @@ -938,17 +938,16 @@ static void hmm_devmem_ref_exit(void *data)
> struct hmm_devmem *devmem;
>
> devmem = container_of(ref, struct hmm_devmem, ref);
> + wait_for_completion(&devmem->completion);
> percpu_ref_exit(ref);
> }
>
> -static void hmm_devmem_ref_kill(void *data)
> +static void hmm_devmem_ref_kill(struct percpu_ref *ref)
> {
> - struct percpu_ref *ref = data;
> struct hmm_devmem *devmem;
>
> devmem = container_of(ref, struct hmm_devmem, ref);
> percpu_ref_kill(ref);
> - wait_for_completion(&devmem->completion);
> }
>
> static int hmm_devmem_fault(struct vm_area_struct *vma,
> @@ -971,152 +970,6 @@ static void hmm_devmem_free(struct page *page, void *data)
> devmem->ops->free(devmem, page);
> }
>
> -static DEFINE_MUTEX(hmm_devmem_lock);
> -static RADIX_TREE(hmm_devmem_radix, GFP_KERNEL);
> -
> -static void hmm_devmem_radix_release(struct resource *resource)
> -{
> - resource_size_t key;
> -
> - mutex_lock(&hmm_devmem_lock);
> - for (key = resource->start;
> - key <= resource->end;
> - key += PA_SECTION_SIZE)
> - radix_tree_delete(&hmm_devmem_radix, key >> PA_SECTION_SHIFT);
> - mutex_unlock(&hmm_devmem_lock);
> -}
> -
> -static void hmm_devmem_release(void *data)
> -{
> - struct hmm_devmem *devmem = data;
> - struct resource *resource = devmem->resource;
> - unsigned long start_pfn, npages;
> - struct zone *zone;
> - struct page *page;
> -
> - /* pages are dead and unused, undo the arch mapping */
> - start_pfn = (resource->start & ~(PA_SECTION_SIZE - 1)) >> PAGE_SHIFT;
> - npages = ALIGN(resource_size(resource), PA_SECTION_SIZE) >> PAGE_SHIFT;
> -
> - page = pfn_to_page(start_pfn);
> - zone = page_zone(page);
> -
> - mem_hotplug_begin();
> - if (resource->desc == IORES_DESC_DEVICE_PRIVATE_MEMORY)
> - __remove_pages(zone, start_pfn, npages, NULL);
> - else
> - arch_remove_memory(start_pfn << PAGE_SHIFT,
> - npages << PAGE_SHIFT, NULL);
> - mem_hotplug_done();
> -
> - hmm_devmem_radix_release(resource);
> -}
> -
> -static int hmm_devmem_pages_create(struct hmm_devmem *devmem)
> -{
> - resource_size_t key, align_start, align_size, align_end;
> - struct device *device = devmem->device;
> - int ret, nid, is_ram;
> - unsigned long pfn;
> -
> - align_start = devmem->resource->start & ~(PA_SECTION_SIZE - 1);
> - align_size = ALIGN(devmem->resource->start +
> - resource_size(devmem->resource),
> - PA_SECTION_SIZE) - align_start;
> -
> - is_ram = region_intersects(align_start, align_size,
> - IORESOURCE_SYSTEM_RAM,
> - IORES_DESC_NONE);
> - if (is_ram == REGION_MIXED) {
> - WARN_ONCE(1, "%s attempted on mixed region %pr\n",
> - __func__, devmem->resource);
> - return -ENXIO;
> - }
> - if (is_ram == REGION_INTERSECTS)
> - return -ENXIO;
> -
> - if (devmem->resource->desc == IORES_DESC_DEVICE_PUBLIC_MEMORY)
> - devmem->pagemap.type = MEMORY_DEVICE_PUBLIC;
> - else
> - devmem->pagemap.type = MEMORY_DEVICE_PRIVATE;
> -
> - devmem->pagemap.res = *devmem->resource;
> - devmem->pagemap.page_fault = hmm_devmem_fault;
> - devmem->pagemap.page_free = hmm_devmem_free;
> - devmem->pagemap.dev = devmem->device;
> - devmem->pagemap.ref = &devmem->ref;
> - devmem->pagemap.data = devmem;
> -
> - mutex_lock(&hmm_devmem_lock);
> - align_end = align_start + align_size - 1;
> - for (key = align_start; key <= align_end; key += PA_SECTION_SIZE) {
> - struct hmm_devmem *dup;
> -
> - dup = radix_tree_lookup(&hmm_devmem_radix,
> - key >> PA_SECTION_SHIFT);
> - if (dup) {
> - dev_err(device, "%s: collides with mapping for %s\n",
> - __func__, dev_name(dup->device));
> - mutex_unlock(&hmm_devmem_lock);
> - ret = -EBUSY;
> - goto error;
> - }
> - ret = radix_tree_insert(&hmm_devmem_radix,
> - key >> PA_SECTION_SHIFT,
> - devmem);
> - if (ret) {
> - dev_err(device, "%s: failed: %d\n", __func__, ret);
> - mutex_unlock(&hmm_devmem_lock);
> - goto error_radix;
> - }
> - }
> - mutex_unlock(&hmm_devmem_lock);
> -
> - nid = dev_to_node(device);
> - if (nid < 0)
> - nid = numa_mem_id();
> -
> - mem_hotplug_begin();
> - /*
> - * For device private memory we call add_pages() as we only need to
> - * allocate and initialize struct page for the device memory. More-
> - * over the device memory is un-accessible thus we do not want to
> - * create a linear mapping for the memory like arch_add_memory()
> - * would do.
> - *
> - * For device public memory, which is accesible by the CPU, we do
> - * want the linear mapping and thus use arch_add_memory().
> - */
> - if (devmem->pagemap.type == MEMORY_DEVICE_PUBLIC)
> - ret = arch_add_memory(nid, align_start, align_size, NULL,
> - false);
> - else
> - ret = add_pages(nid, align_start >> PAGE_SHIFT,
> - align_size >> PAGE_SHIFT, NULL, false);
> - if (ret) {
> - mem_hotplug_done();
> - goto error_add_memory;
> - }
> - move_pfn_range_to_zone(&NODE_DATA(nid)->node_zones[ZONE_DEVICE],
> - align_start >> PAGE_SHIFT,
> - align_size >> PAGE_SHIFT, NULL);
> - mem_hotplug_done();
> -
> - for (pfn = devmem->pfn_first; pfn < devmem->pfn_last; pfn++) {
> - struct page *page = pfn_to_page(pfn);
> -
> - page->pgmap = &devmem->pagemap;
> - }
> - return 0;
> -
> -error_add_memory:
> - untrack_pfn(NULL, PHYS_PFN(align_start), align_size);
> -error_radix:
> - hmm_devmem_radix_release(devmem->resource);
> -error:
> - return ret;
> -}
> -
> /*
> * hmm_devmem_add() - hotplug ZONE_DEVICE memory for device memory
> *
> @@ -1140,6 +993,7 @@ struct hmm_devmem *hmm_devmem_add(const struct hmm_devmem_ops *ops,
> {
> struct hmm_devmem *devmem;
> resource_size_t addr;
> + void *result;
> int ret;
>
> dev_pagemap_get_ops();
> @@ -1194,14 +1048,18 @@ struct hmm_devmem *hmm_devmem_add(const struct hmm_devmem_ops *ops,
> devmem->pfn_last = devmem->pfn_first +
> (resource_size(devmem->resource) >> PAGE_SHIFT);
>
> - ret = hmm_devmem_pages_create(devmem);
> - if (ret)
> - return ERR_PTR(ret);
> -
> - ret = devm_add_action_or_reset(device, hmm_devmem_release, devmem);
> - if (ret)
> - return ERR_PTR(ret);
> + devmem->pagemap.type = MEMORY_DEVICE_PRIVATE;
> + devmem->pagemap.res = *devmem->resource;
> + devmem->pagemap.page_fault = hmm_devmem_fault;
> + devmem->pagemap.page_free = hmm_devmem_free;
> + devmem->pagemap.altmap_valid = false;
> + devmem->pagemap.ref = &devmem->ref;
> + devmem->pagemap.data = devmem;
>
> + result = devm_memremap_pages(devmem->device, &devmem->pagemap,
> + hmm_devmem_ref_kill);
> + if (IS_ERR(result))
> + return result;
> return devmem;
> }
> EXPORT_SYMBOL(hmm_devmem_add);
> @@ -1211,6 +1069,7 @@ struct hmm_devmem *hmm_devmem_add_resource(const struct hmm_devmem_ops *ops,
> struct resource *res)
> {
> struct hmm_devmem *devmem;
> + void *result;
> int ret;
>
> if (res->desc != IORES_DESC_DEVICE_PUBLIC_MEMORY)
> @@ -1243,19 +1102,18 @@ struct hmm_devmem *hmm_devmem_add_resource(const struct hmm_devmem_ops *ops,
> devmem->pfn_last = devmem->pfn_first +
> (resource_size(devmem->resource) >> PAGE_SHIFT);
>
> - ret = hmm_devmem_pages_create(devmem);
> - if (ret)
> - return ERR_PTR(ret);
> -
> - ret = devm_add_action_or_reset(device, hmm_devmem_release, devmem);
> - if (ret)
> - return ERR_PTR(ret);
> -
> - ret = devm_add_action_or_reset(device, hmm_devmem_ref_kill,
> - &devmem->ref);
> - if (ret)
> - return ERR_PTR(ret);
> + devmem->pagemap.type = MEMORY_DEVICE_PUBLIC;
> + devmem->pagemap.res = *devmem->resource;
> + devmem->pagemap.page_fault = hmm_devmem_fault;
> + devmem->pagemap.page_free = hmm_devmem_free;
> + devmem->pagemap.altmap_valid = false;
> + devmem->pagemap.ref = &devmem->ref;
> + devmem->pagemap.data = devmem;
>
> + result = devm_memremap_pages(devmem->device, &devmem->pagemap,
> + hmm_devmem_ref_kill);
> + if (IS_ERR(result))
> + return result;
> return devmem;
> }
> EXPORT_SYMBOL(hmm_devmem_add_resource);
>
next prev parent reply other threads:[~2018-09-18 20:35 UTC|newest]
Thread overview: 23+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-09-13 2:22 [PATCH v5 0/7] mm: Merge hmm into devm_memremap_pages, mark GPL-only Dan Williams
2018-09-13 2:22 ` [PATCH v5 1/7] mm, devm_memremap_pages: Mark devm_memremap_pages() EXPORT_SYMBOL_GPL Dan Williams
2018-09-13 16:25 ` Logan Gunthorpe
2018-09-13 2:22 ` [PATCH v5 2/7] mm, devm_memremap_pages: Kill mapping "System RAM" support Dan Williams
2018-09-13 16:10 ` Logan Gunthorpe
2018-09-14 13:14 ` Christoph Hellwig
2018-09-14 17:40 ` Dan Williams
2018-09-18 20:28 ` Jerome Glisse
2018-09-13 2:22 ` [PATCH v5 3/7] mm, devm_memremap_pages: Fix shutdown handling Dan Williams
2018-09-14 13:16 ` Christoph Hellwig
2018-09-14 17:25 ` Dan Williams
2018-09-18 20:28 ` Jerome Glisse
2018-09-13 2:22 ` [PATCH v5 4/7] mm, devm_memremap_pages: Add MEMORY_DEVICE_PRIVATE support Dan Williams
2018-09-14 13:18 ` Christoph Hellwig
2018-09-18 20:34 ` Jerome Glisse
2018-09-13 2:22 ` [PATCH v5 5/7] mm, hmm: Use devm semantics for hmm_devmem_{add, remove} Dan Williams
2018-09-14 13:18 ` Christoph Hellwig
2018-09-14 14:16 ` Jerome Glisse
2018-09-18 20:34 ` Jerome Glisse
2018-09-13 2:22 ` [PATCH v5 6/7] mm, hmm: Replace hmm_devmem_pages_create() with devm_memremap_pages() Dan Williams
2018-09-18 20:35 ` Jerome Glisse [this message]
2018-09-19 1:24 ` Balbir Singh
2018-09-13 2:22 ` [PATCH v5 7/7] mm, hmm: Mark hmm_devmem_{add, add_resource} EXPORT_SYMBOL_GPL Dan Williams
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20180918203537.GF14689@redhat.com \
--to=jglisse@redhat.com \
--cc=akpm@linux-foundation.org \
--cc=alexander.h.duyck@intel.com \
--cc=dan.j.williams@intel.com \
--cc=hch@lst.de \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=logang@deltatee.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).