From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id B9A30C05027 for ; Fri, 10 Feb 2023 22:03:26 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S233452AbjBJWD0 (ORCPT ); Fri, 10 Feb 2023 17:03:26 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:45272 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232505AbjBJWDZ (ORCPT ); Fri, 10 Feb 2023 17:03:25 -0500 Received: from mga04.intel.com (mga04.intel.com [192.55.52.120]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 3C0585ACF0; Fri, 10 Feb 2023 14:03:24 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1676066604; x=1707602604; h=message-id:date:mime-version:subject:to:cc:references: from:in-reply-to:content-transfer-encoding; bh=DJ4H1V+6TU5G803CMbKqoAlYJI/kGtz69jyguC5gnOM=; b=UMFahUlk5rBPjEU3isY+hb1WCiNot7v6ykx8y+SO0WY6blBY7UYGXAoU 5JtzDsqokKx4CLfSWmb1H7gGPhkZbcGNc9hbrPYC/7gXo0KCxjf5tEjGN Kf5y4fx9AS7srF2JV1ELzAPbEgRavevBWl2S4fGvHxmFSlPofj4nSVLQe 9yaS5F8h+lAxzOmWlzRS8Z8doPYtnwm9Q9J31RjX8+tTIsI+ZHWQNuhuD tpsHKM8+TswKTFyGmyilkXyZfjqzWw1FBePOYiqbEh50ippLD01jx+56K x4OJF/6PlFXyWzyG/VBcfndWPCkiTzseSewXCnfkYTgsL64w73QKBx0YY A==; X-IronPort-AV: E=McAfee;i="6500,9779,10617"; a="329168193" X-IronPort-AV: E=Sophos;i="5.97,287,1669104000"; d="scan'208";a="329168193" Received: from orsmga004.jf.intel.com ([10.7.209.38]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 10 Feb 2023 14:03:23 -0800 X-IronPort-AV: E=McAfee;i="6500,9779,10617"; a="792115392" X-IronPort-AV: E=Sophos;i="5.97,287,1669104000"; d="scan'208";a="792115392" Received: from djiang5-mobl3.amr.corp.intel.com (HELO [10.213.190.133]) ([10.213.190.133]) by orsmga004-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 10 Feb 2023 14:03:22 -0800 Message-ID: <7f9303ac-1e3c-8b13-4b85-36a49baa37f1@intel.com> Date: Fri, 10 Feb 2023 15:03:21 -0700 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:102.0) Gecko/20100101 Firefox/102.0 Thunderbird/102.6.0 Subject: Re: [PATCH v2 17/20] dax/hmem: Convey the dax range via memregion_info() Content-Language: en-US To: Dan Williams , linux-cxl@vger.kernel.org Cc: Jonathan Cameron , Fan Ni , vishal.l.verma@intel.com, dave.hansen@linux.intel.com, linux-mm@kvack.org, linux-acpi@vger.kernel.org References: <167601992097.1924368.18291887895351917895.stgit@dwillia2-xfh.jf.intel.com> <167602002217.1924368.7036275892522551624.stgit@dwillia2-xfh.jf.intel.com> From: Dave Jiang In-Reply-To: <167602002217.1924368.7036275892522551624.stgit@dwillia2-xfh.jf.intel.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-cxl@vger.kernel.org On 2/10/23 2:07 AM, Dan Williams wrote: > In preparation for hmem platform devices to be unregistered, stop using > platform_device_add_resources() to convey the address range. The > platform_device_add_resources() API causes an existing "Soft Reserved" > iomem resource to be re-parented under an inserted platform device > resource. When that platform device is deleted it removes the platform > device resource and all children. > > Instead, it is sufficient to convey just the address range and let > request_mem_region() insert resources to indicate the devices active in > the range. This allows the "Soft Reserved" resource to be re-enumerated > upon the next probe event. > > Reviewed-by: Jonathan Cameron > Tested-by: Fan Ni > Link: https://lore.kernel.org/r/167564543303.847146.11045895213318648441.stgit@dwillia2-xfh.jf.intel.com > Signed-off-by: Dan Williams Reviewed-by: Dave Jiang > --- > drivers/dax/hmem/device.c | 37 ++++++++++++++----------------------- > drivers/dax/hmem/hmem.c | 14 +++----------- > include/linux/memregion.h | 2 ++ > 3 files changed, 19 insertions(+), 34 deletions(-) > > diff --git a/drivers/dax/hmem/device.c b/drivers/dax/hmem/device.c > index 20749c7fab81..b1b339bccfe5 100644 > --- a/drivers/dax/hmem/device.c > +++ b/drivers/dax/hmem/device.c > @@ -15,15 +15,8 @@ static struct resource hmem_active = { > .flags = IORESOURCE_MEM, > }; > > -void hmem_register_device(int target_nid, struct resource *r) > +void hmem_register_device(int target_nid, struct resource *res) > { > - /* define a clean / non-busy resource for the platform device */ > - struct resource res = { > - .start = r->start, > - .end = r->end, > - .flags = IORESOURCE_MEM, > - .desc = IORES_DESC_SOFT_RESERVED, > - }; > struct platform_device *pdev; > struct memregion_info info; > int rc, id; > @@ -31,55 +24,53 @@ void hmem_register_device(int target_nid, struct resource *r) > if (nohmem) > return; > > - rc = region_intersects(res.start, resource_size(&res), IORESOURCE_MEM, > - IORES_DESC_SOFT_RESERVED); > + rc = region_intersects(res->start, resource_size(res), IORESOURCE_MEM, > + IORES_DESC_SOFT_RESERVED); > if (rc != REGION_INTERSECTS) > return; > > id = memregion_alloc(GFP_KERNEL); > if (id < 0) { > - pr_err("memregion allocation failure for %pr\n", &res); > + pr_err("memregion allocation failure for %pr\n", res); > return; > } > > pdev = platform_device_alloc("hmem", id); > if (!pdev) { > - pr_err("hmem device allocation failure for %pr\n", &res); > + pr_err("hmem device allocation failure for %pr\n", res); > goto out_pdev; > } > > - if (!__request_region(&hmem_active, res.start, resource_size(&res), > + if (!__request_region(&hmem_active, res->start, resource_size(res), > dev_name(&pdev->dev), 0)) { > - dev_dbg(&pdev->dev, "hmem range %pr already active\n", &res); > + dev_dbg(&pdev->dev, "hmem range %pr already active\n", res); > goto out_active; > } > > pdev->dev.numa_node = numa_map_to_online_node(target_nid); > info = (struct memregion_info) { > .target_node = target_nid, > + .range = { > + .start = res->start, > + .end = res->end, > + }, > }; > rc = platform_device_add_data(pdev, &info, sizeof(info)); > if (rc < 0) { > - pr_err("hmem memregion_info allocation failure for %pr\n", &res); > - goto out_resource; > - } > - > - rc = platform_device_add_resources(pdev, &res, 1); > - if (rc < 0) { > - pr_err("hmem resource allocation failure for %pr\n", &res); > + pr_err("hmem memregion_info allocation failure for %pr\n", res); > goto out_resource; > } > > rc = platform_device_add(pdev); > if (rc < 0) { > - dev_err(&pdev->dev, "device add failed for %pr\n", &res); > + dev_err(&pdev->dev, "device add failed for %pr\n", res); > goto out_resource; > } > > return; > > out_resource: > - __release_region(&hmem_active, res.start, resource_size(&res)); > + __release_region(&hmem_active, res->start, resource_size(res)); > out_active: > platform_device_put(pdev); > out_pdev: > diff --git a/drivers/dax/hmem/hmem.c b/drivers/dax/hmem/hmem.c > index c7351e0dc8ff..5025a8c9850b 100644 > --- a/drivers/dax/hmem/hmem.c > +++ b/drivers/dax/hmem/hmem.c > @@ -15,25 +15,17 @@ static int dax_hmem_probe(struct platform_device *pdev) > struct memregion_info *mri; > struct dev_dax_data data; > struct dev_dax *dev_dax; > - struct resource *res; > - struct range range; > - > - res = platform_get_resource(pdev, IORESOURCE_MEM, 0); > - if (!res) > - return -ENOMEM; > > mri = dev->platform_data; > - range.start = res->start; > - range.end = res->end; > - dax_region = alloc_dax_region(dev, pdev->id, &range, mri->target_node, > - PMD_SIZE, 0); > + dax_region = alloc_dax_region(dev, pdev->id, &mri->range, > + mri->target_node, PMD_SIZE, 0); > if (!dax_region) > return -ENOMEM; > > data = (struct dev_dax_data) { > .dax_region = dax_region, > .id = -1, > - .size = region_idle ? 0 : resource_size(res), > + .size = region_idle ? 0 : range_len(&mri->range), > }; > dev_dax = devm_create_dev_dax(&data); > if (IS_ERR(dev_dax)) > diff --git a/include/linux/memregion.h b/include/linux/memregion.h > index bf83363807ac..c01321467789 100644 > --- a/include/linux/memregion.h > +++ b/include/linux/memregion.h > @@ -3,10 +3,12 @@ > #define _MEMREGION_H_ > #include > #include > +#include > #include > > struct memregion_info { > int target_node; > + struct range range; > }; > > #ifdef CONFIG_MEMREGION >