From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 7774B1CD2C; Fri, 19 Dec 2025 09:33:34 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1766136814; cv=none; b=u9x7Ngv2/FzGgOVRClDMih/z9LUZK849ISXXOJNBsCBRtAAO7QtIPrFYBaPelbJIjPXHM6oH0BeVV+/ssPh2sW9MWMHV238FDW9G8pfjm2irdc9V/PRxUgndNWdDGUGmJ8oMosTU9zGPPPeH6N2ShSmDYFLPkAy+pzL4tV19t8U= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1766136814; c=relaxed/simple; bh=8gogU4puAS1Eixl5xxRff+w5BC9Gk8ktr3c7vX414Ek=; h=Date:From:To:Cc:Subject:Message-ID:References:MIME-Version: Content-Type:Content-Disposition:In-Reply-To; b=W7CsehCAJuCSYIbyCKHbAB1RwJeFLX/ABwVgCyD1WnbEJJJCCOYQe9fn7WkKgVHGwftfp16cMVHHha9Ani3eNG6I0Ig1MkBJEXjKhk5BnbcSTknfD00bCRYu1aOdi4+Q4gbAm+Sd4tQR/BFkZMcsSHfBHcxVV+g9Z8LjC7ZuFI4= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=Zuh6QYOA; arc=none smtp.client-ip=10.30.226.201 Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="Zuh6QYOA" Received: by smtp.kernel.org (Postfix) with ESMTPSA id 862DCC116D0; Fri, 19 Dec 2025 09:33:28 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1766136814; bh=8gogU4puAS1Eixl5xxRff+w5BC9Gk8ktr3c7vX414Ek=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=Zuh6QYOAWUvaiFNc1+kuSjpiYs3pjbjJ3Csu4iZZciekNaLP+wHcr3huE7B51cq7A XCy2GoY/9siUVelznccVcqkr/TOAdqEbIOjTScA7H+NXGQieAGDI8ze/HFfBK+RP3i 55TnbAvYDcT+pVGl74NAqmFDb/zu2OWf5hM7dKB5JHfZ7ZRAotDMLnCfEV9YDnYXTt brZSd34Bi5e9O+BZm1bKE/WZEQ/W38LIcPpJiwBR+eIPQSq4g1cjXqONwl/1GhC+HP tE76QBIZCXDJiQm93GUZ4FaAWIqXdm7QfhkMauUB0ng04S1xfsNoKeZ7HGARq3jpkb F1zkulauKJObA== Date: Fri, 19 Dec 2025 11:33:25 +0200 From: Mike Rapoport To: chengkaitao Cc: davem@davemloft.net, andreas@gaisler.com, akpm@linux-foundation.org, david@kernel.org, lorenzo.stoakes@oracle.com, Liam.Howlett@oracle.com, vbabka@suse.cz, surenb@google.com, mhocko@suse.com, kevin.brodsky@arm.com, dave.hansen@linux.intel.com, ziy@nvidia.com, chengkaitao@kylinos.cn, willy@infradead.org, zhengqi.arch@bytedance.com, sparclinux@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org Subject: Re: [PATCH v3 1/3] sparc: Use vmemmap_populate_hugepages for vmemmap_populate Message-ID: References: <20251218174749.45965-1-pilgrimtao@gmail.com> <20251218174749.45965-2-pilgrimtao@gmail.com> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20251218174749.45965-2-pilgrimtao@gmail.com> On Fri, Dec 19, 2025 at 01:47:47AM +0800, chengkaitao wrote: > From: Chengkaitao > > 1. In the SPARC architecture, reimplemented vmemmap_populate using > vmemmap_populate_hugepages. > 2. Allow the SPARC arch to fallback to vmemmap_populate_basepages(), > when vmemmap_alloc_block returns NULL. > > Signed-off-by: Chengkaitao > --- > arch/sparc/mm/init_64.c | 47 ++++++++++++----------------------------- > 1 file changed, 14 insertions(+), 33 deletions(-) > > diff --git a/arch/sparc/mm/init_64.c b/arch/sparc/mm/init_64.c > index df9f7c444c39..f477ab1b4f08 100644 > --- a/arch/sparc/mm/init_64.c > +++ b/arch/sparc/mm/init_64.c > @@ -2581,8 +2581,8 @@ unsigned long _PAGE_CACHE __read_mostly; > EXPORT_SYMBOL(_PAGE_CACHE); > > #ifdef CONFIG_SPARSEMEM_VMEMMAP > -int __meminit vmemmap_populate(unsigned long vstart, unsigned long vend, > - int node, struct vmem_altmap *altmap) > +void __meminit vmemmap_set_pmd(pmd_t *pmd, void *p, int node, > + unsigned long addr, unsigned long next) > { > unsigned long pte_base; > > @@ -2595,39 +2595,20 @@ int __meminit vmemmap_populate(unsigned long vstart, unsigned long vend, > > pte_base |= _PAGE_PMD_HUGE; > > - vstart = vstart & PMD_MASK; > - vend = ALIGN(vend, PMD_SIZE); > - for (; vstart < vend; vstart += PMD_SIZE) { > - pgd_t *pgd = vmemmap_pgd_populate(vstart, node); > - unsigned long pte; > - p4d_t *p4d; > - pud_t *pud; > - pmd_t *pmd; > - > - if (!pgd) > - return -ENOMEM; > - > - p4d = vmemmap_p4d_populate(pgd, vstart, node); > - if (!p4d) > - return -ENOMEM; > - > - pud = vmemmap_pud_populate(p4d, vstart, node); > - if (!pud) > - return -ENOMEM; > - > - pmd = pmd_offset(pud, vstart); > - pte = pmd_val(*pmd); > - if (!(pte & _PAGE_VALID)) { > - void *block = vmemmap_alloc_block(PMD_SIZE, node); > - > - if (!block) > - return -ENOMEM; > + pmd_val(*pmd) = pte_base | __pa(p); > +} > > - pmd_val(*pmd) = pte_base | __pa(block); > - } > - } > +int __meminit vmemmap_check_pmd(pmd_t *pmdp, int node, > + unsigned long addr, unsigned long next) > +{ > + vmemmap_verify((pte_t *)pmdp, node, addr, next); > + return 1; This should return 1 only if the pmd actually maps a large page, see x86 and powerpc implementations for example. > +} > > - return 0; > +int __meminit vmemmap_populate(unsigned long vstart, unsigned long vend, > + int node, struct vmem_altmap *altmap) > +{ > + return vmemmap_populate_hugepages(vstart, vend, node, altmap); > } > #endif /* CONFIG_SPARSEMEM_VMEMMAP */ > > -- > 2.50.1 (Apple Git-155) > -- Sincerely yours, Mike.