From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 049A1C4707B for ; Wed, 10 Jan 2024 23:12:47 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 7106E6B0095; Wed, 10 Jan 2024 18:12:47 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 6C0646B0096; Wed, 10 Jan 2024 18:12:47 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 587BC6B0098; Wed, 10 Jan 2024 18:12:47 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0013.hostedemail.com [216.40.44.13]) by kanga.kvack.org (Postfix) with ESMTP id 478886B0095 for ; Wed, 10 Jan 2024 18:12:47 -0500 (EST) Received: from smtpin30.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay01.hostedemail.com (Postfix) with ESMTP id 0F0C61C14DD for ; Wed, 10 Jan 2024 23:12:47 +0000 (UTC) X-FDA: 81664953174.30.61A4407 Received: from mgamail.intel.com (mgamail.intel.com [134.134.136.65]) by imf26.hostedemail.com (Postfix) with ESMTP id 8AEB314001E for ; Wed, 10 Jan 2024 23:12:44 +0000 (UTC) Authentication-Results: imf26.hostedemail.com; dkim=pass header.d=intel.com header.s=Intel header.b=P3+luxXx; spf=none (imf26.hostedemail.com: domain of tim.c.chen@linux.intel.com has no SPF policy when checking 134.134.136.65) smtp.mailfrom=tim.c.chen@linux.intel.com; dmarc=pass (policy=none) header.from=intel.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1704928365; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=E8kpmUP+4xsob87Qk0MeyOkWoo0YnuaKF/FsRR5ibaA=; b=gi4IxxtHxByUt6fooo31FbUb7NVRaO4cAWDC4XhuMFja+AoUEvPAQPIs9BN9LZKCt4K3g+ XctWbgcFjYAtovtZ2/Jl115uXCobdlr75MMwjswt57zE6qLPFYCaot5VRNg1GeYT+NBBdA /B14wu6lzb/hcYL8BwsHcVxsLF/syVI= ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1704928365; a=rsa-sha256; cv=none; b=oQVruFshi1m6cRksQ+vVfVgQGowTB6QokFNhlJhj0LTBtxnuVWW5Dcoso3sOHF8pLFfsFY bOV3SoNKtFZ6FsOD9pD5ivs9Q7YK1nUy6afWYJQm/CaNhjlsAiFph6JBzE6K3gUKj62che 1eaU+QaanSYwnAYTRfroNmGGJU89EwU= ARC-Authentication-Results: i=1; imf26.hostedemail.com; dkim=pass header.d=intel.com header.s=Intel header.b=P3+luxXx; spf=none (imf26.hostedemail.com: domain of tim.c.chen@linux.intel.com has no SPF policy when checking 134.134.136.65) smtp.mailfrom=tim.c.chen@linux.intel.com; dmarc=pass (policy=none) header.from=intel.com DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1704928364; x=1736464364; h=message-id:subject:from:to:cc:date:in-reply-to: references:content-transfer-encoding:mime-version; bh=oZto8Uu2fZ3IZ8wqMhOSNCyt5zHeJNO3Al5r6NP7YbE=; b=P3+luxXxj2bHJFE921pzI81K8q/TW2sUh4cHu0IiiXclnZlZDeOj7Rlu 7CEPUxYjz5c5NfLBZx4ZKa4gX9YWeDzJmh+vlOHyv8nKbA52PNds0oUkk ZNSHBZqxIGBDykexgOlrxCIXdggC2nzI8Tsk7aXNW6immgbJ9PMjgyE54 lekmyNXrtEeRAk7XnO3pMjFBUsH1ptRWMvbYAwxHAY5ix8eQVa6j+6QLA 96XuY5MJYNLi0+XGf8VzLpgMI+/so5RoVSQ6Ry2w8xXDcp+/mTnjS77EC mwfZP5j8DxDwOgrfQ+y6sP35OQder/pZbAISBh5BNU/yksLXFBUkICS2N Q==; X-IronPort-AV: E=McAfee;i="6600,9927,10949"; a="402451303" X-IronPort-AV: E=Sophos;i="6.04,184,1695711600"; d="scan'208";a="402451303" Received: from orsmga004.jf.intel.com ([10.7.209.38]) by orsmga103.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 10 Jan 2024 15:12:40 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6600,9927,10949"; a="905722185" X-IronPort-AV: E=Sophos;i="6.04,184,1695711600"; d="scan'208";a="905722185" Received: from phungng1-mobl.amr.corp.intel.com (HELO [10.255.229.71]) ([10.255.229.71]) by orsmga004-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 10 Jan 2024 15:12:39 -0800 Message-ID: <9b95c0ac0e5d560f1482af3753df54211a057b9d.camel@linux.intel.com> Subject: Re: [PATCH v3 2/7] hugetlb: split hugetlb_hstate_alloc_pages From: Tim Chen To: Gang Li , David Hildenbrand , David Rientjes , Mike Kravetz , Muchun Song , Andrew Morton Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, ligang.bdlg@bytedance.com Date: Wed, 10 Jan 2024 15:12:38 -0800 In-Reply-To: <20240102131249.76622-3-gang.li@linux.dev> References: <20240102131249.76622-1-gang.li@linux.dev> <20240102131249.76622-3-gang.li@linux.dev> Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable User-Agent: Evolution 3.44.4 (3.44.4-2.fc36) MIME-Version: 1.0 X-Rspamd-Queue-Id: 8AEB314001E X-Rspam-User: X-Stat-Signature: n7uk5rimj3bi3yuyx489j1egz8unji3k X-Rspamd-Server: rspam03 X-HE-Tag: 1704928364-259928 X-HE-Meta: U2FsdGVkX188otRj4Z+8C5MUR0Ns0veaBAg3PjgdCfHCYQPwxGM4enzuCUq1AqEDqOCxquBmOuw3tAMB0vWC6zF4Y7dDjoo8Y0wnDdOfSD3uIgTt+IFvzfGPEiola+/BaWrrYJXl5FyP7GYnN6eSekla650gLP5z63VrhuePBM0M7FI+gI4H6M5U6f21Oxch4hZECYq/W5yTD6z9+UIBKRRMf72Fd7fWNX4gd/fSnYYleRdeOdZvF7VS9B5G+wnEPBH+hqa3Ju7QooJ+6Gn1tJvJIDZDpHEYuZXhM1yvdPmovQVGEFch5MFaTjuEIzSyHURwMcazoiMW5Yhj315mdV4IJ3p5hQhrx5Wj7SDAQFQAXXU4MmIiyGMpvO2XCzeHLLn0ORZF0VvZrBAPsp2+uQ/WazAR+iwaGJ2l+9GZRwPiAoKp8I9EXoou9xubKR2bbOefY4rn4Pbnnxm3q4QpVMMTkKXq+9JPkKVKeCTW3a7kRyT5bJstSeThB13h6OjCLN/TZxorwU+o8/yrIsOjN/OrPUnAoz8PsBWmxM5+tyNQjG0NX3xzV1kCswdKsjAgyHQK4YnO9fbQp5J/HpR2rCNBuih1Zzs1Qt6ShZYLXpWsTI5W3D0+K7mYGF4yPtCCJPRRZ0HQU8YR17xaymoUdpmy1x/i8aCp7h2C0OGo5xetn2QUwsRjLNzd6nz6PYRD/IaqR6S5p9vnCbXv98G2jRUv9Kh030OQuxQWo0KrEFF+xaXICH5I38tQQ0iEM3120VeIgSosR6to/3eyuBk5UKXt5ubu9xSBom7OkwlVwaYBJfvwK/9dGtPVKxK4klqeLW5XMYJ06c4PacFMA+hOm3XUv99WsEj+vPE9ivDR/LMN9dzaaCwurH0pmy8/GsTmQhTV5VYj6m/8204gdiL9cjptdQ3J3LjuXWBWnFhhcisFV3KqT83gCNLhKtgmh17zd3c3qcGA+eFtqpkTUU7 twr3GX4l TwFhCSpUkA+uGNmBbjxygqE4kEn5PP4XoF5VzYEajMH7GJsZcXDYzd2YAJ3ITYvJ0EIfw12rGKj1pZ/r/DMREhVfrVdn8RZDBcs6FqtAnT3yjvlQVt6ddC3G/HFxn0+LRGY395l1SqEbo8wSabr8K+gP0u8B7iBFXSr3JDF+TRKGdSc65fSWxI8yOK7hGU/jRUkiW X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On Tue, 2024-01-02 at 21:12 +0800, Gang Li wrote: > 1G and 2M huge pages have different allocation and initialization logic, > which leads to subtle differences in parallelization. Therefore, it is > appropriate to split hugetlb_hstate_alloc_pages into gigantic and > non-gigantic. >=20 > This patch has no functional changes. >=20 > Signed-off-by: Gang Li > --- > mm/hugetlb.c | 86 +++++++++++++++++++++++++++------------------------- > 1 file changed, 45 insertions(+), 41 deletions(-) >=20 > diff --git a/mm/hugetlb.c b/mm/hugetlb.c > index 2606135ec55e6..92448e747991d 100644 > --- a/mm/hugetlb.c > +++ b/mm/hugetlb.c > @@ -3509,6 +3509,47 @@ static void __init hugetlb_hstate_alloc_pages_repo= rt(unsigned long allocated, st > } > } > =20 > +static unsigned long __init hugetlb_hstate_alloc_pages_gigantic(struct h= state *h) > +{ > + unsigned long i; > + > + for (i =3D 0; i < h->max_huge_pages; ++i) { > + /* > + * gigantic pages not added to list as they are not > + * added to pools now. > + */ This comment unnecessary as now we don't have mix gigantic and non-gigantic= code, which uses foilio list. And folio_list is not in this routine. Can be removed. Otherwise Reviewed-by: Tim Chen > + if (!alloc_bootmem_huge_page(h, NUMA_NO_NODE)) > + break; > + cond_resched(); > + } > + > + return i; > +} > + > +static unsigned long __init hugetlb_hstate_alloc_pages_non_gigantic(stru= ct hstate *h) > +{ > + unsigned long i; > + struct folio *folio; > + LIST_HEAD(folio_list); > + nodemask_t node_alloc_noretry; > + > + /* Bit mask controlling how hard we retry per-node allocations.*/ > + nodes_clear(node_alloc_noretry); > + > + for (i =3D 0; i < h->max_huge_pages; ++i) { > + folio =3D alloc_pool_huge_folio(h, &node_states[N_MEMORY], > + &node_alloc_noretry); > + if (!folio) > + break; > + list_add(&folio->lru, &folio_list); > + cond_resched(); > + } > + > + prep_and_add_allocated_folios(h, &folio_list); > + > + return i; > +} > + > /* > * NOTE: this routine is called in different contexts for gigantic and > * non-gigantic pages. > @@ -3522,10 +3563,7 @@ static void __init hugetlb_hstate_alloc_pages_repo= rt(unsigned long allocated, st > */ > static void __init hugetlb_hstate_alloc_pages(struct hstate *h) > { > - unsigned long i; > - struct folio *folio; > - LIST_HEAD(folio_list); > - nodemask_t *node_alloc_noretry; > + unsigned long allocated; > =20 > /* skip gigantic hugepages allocation if hugetlb_cma enabled */ > if (hstate_is_gigantic(h) && hugetlb_cma_size) { > @@ -3539,46 +3577,12 @@ static void __init hugetlb_hstate_alloc_pages(str= uct hstate *h) > =20 > /* below will do all node balanced alloc */ > if (!hstate_is_gigantic(h)) { > - /* > - * Bit mask controlling how hard we retry per-node allocations. > - * Ignore errors as lower level routines can deal with > - * node_alloc_noretry =3D=3D NULL. If this kmalloc fails at boot > - * time, we are likely in bigger trouble. > - */ > - node_alloc_noretry =3D kmalloc(sizeof(*node_alloc_noretry), > - GFP_KERNEL); > + allocated =3D hugetlb_hstate_alloc_pages_non_gigantic(h); > } else { > - /* allocations done at boot time */ > - node_alloc_noretry =3D NULL; > - } > - > - /* bit mask controlling how hard we retry per-node allocations */ > - if (node_alloc_noretry) > - nodes_clear(*node_alloc_noretry); > - > - for (i =3D 0; i < h->max_huge_pages; ++i) { > - if (hstate_is_gigantic(h)) { > - /* > - * gigantic pages not added to list as they are not > - * added to pools now. > - */ > - if (!alloc_bootmem_huge_page(h, NUMA_NO_NODE)) > - break; > - } else { > - folio =3D alloc_pool_huge_folio(h, &node_states[N_MEMORY], > - node_alloc_noretry); > - if (!folio) > - break; > - list_add(&folio->lru, &folio_list); > - } > - cond_resched(); > + allocated =3D hugetlb_hstate_alloc_pages_gigantic(h); > } > =20 > - /* list will be empty if hstate_is_gigantic */ > - prep_and_add_allocated_folios(h, &folio_list); > - > - hugetlb_hstate_alloc_pages_report(i, h); > - kfree(node_alloc_noretry); > + hugetlb_hstate_alloc_pages_report(allocated, h); > } > =20 > static void __init hugetlb_init_hstates(void)