From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from bombadil.infradead.org (bombadil.infradead.org [198.137.202.133]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.lore.kernel.org (Postfix) with ESMTPS id DD7CEC77B75 for ; Wed, 24 May 2023 00:18:09 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=lists.infradead.org; s=bombadil.20210309; h=Sender: Content-Transfer-Encoding:Content-Type:List-Subscribe:List-Help:List-Post: List-Archive:List-Unsubscribe:List-Id:In-Reply-To:MIME-Version:References: Message-ID:Subject:Cc:To:From:Date:Reply-To:Content-ID:Content-Description: Resent-Date:Resent-From:Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID: List-Owner; bh=i1Wp3cyMSzNW3nAYAz1lbOyHtCiaXUkaIEoYdN5Z1so=; b=Ef6hcXy73h1pJp 4LiTszjoNQo47PS5jztwt/9UmCJ8T5YmrOclXMEDmppzezkwj+fUToj02/uUo2Agzd/RIV1v9Q39A 0NI8HCvpxZl5Gh3l+jSCobdk92AZ4jvJM/NhDO7PCt8+l/Sg6WAZZp8LnFOrbSeYZyeWX4CRJutza 4GiI5PzY9LPHThsi1qKc9lOUBBsCTQ4NfcU4Dcfffxli1JGuK8tkkYSpgZQKBjKYysDQe9z/R0BKQ 83ePof9IqBfwloP8tVT3ql7W7ZhJiQq/Frqvl9a4A24ie1xwYaNv2XhwS0AJR0h74aNYbcpzK78tu I3g8JZI8gif1a41CAyKg==; Received: from localhost ([::1] helo=bombadil.infradead.org) by bombadil.infradead.org with esmtp (Exim 4.96 #2 (Red Hat Linux)) id 1q1cCW-00BtAs-2S; Wed, 24 May 2023 00:18:00 +0000 Received: from dfw.source.kernel.org ([2604:1380:4641:c500::1]) by bombadil.infradead.org with esmtps (Exim 4.96 #2 (Red Hat Linux)) id 1q1cCT-00BtAS-19 for linux-riscv@lists.infradead.org; Wed, 24 May 2023 00:17:59 +0000 Received: from smtp.kernel.org (relay.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by dfw.source.kernel.org (Postfix) with ESMTPS id AD8206328D; Wed, 24 May 2023 00:17:56 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id C6650C433EF; Wed, 24 May 2023 00:17:54 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1684887476; bh=9dxLjHPo+SZEGy7KQS3cCv9yVec/7vFxtb4zBAqLmXg=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=Wr7T+5i/6DQrZylu3WCREvJ2bcO9dFx9j+flmANgbbKwhs3FVnYFrJV967oKEfeyD YaSfFfAdfwU/d4c/BDPMmayNOvc/FfiKVKHsE1Ay4B/b98MeOAAZv3MX7LVnwS1CfX Q88UbGhhrbK1+okbRb5/KhncQQ67FMZMShmeZgmhl3BrDuN0sWCn/FeDXxn5KnII0m wR8vT6zgg42KZZhLyDFnkwNsago373inq3hwEKApu1gWsnsEZlktQlfe8E3PuJm2Bd P3O3ZJLPpzsUU7TRBYQTyH8AJ9Py1nm/Hg426LOqE7Pb5CaeQ1hcY0ORJzf8MtqdLa 1YWM74O5ER+yA== Date: Wed, 24 May 2023 08:06:44 +0800 From: Jisheng Zhang To: Paul Walmsley , Palmer Dabbelt , Albert Ou Cc: linux-riscv@lists.infradead.org, linux-kernel@vger.kernel.org, Suren Baghdasaryan Subject: Re: [PATCH] riscv: mm: try VMA lock-based page fault handling first Message-ID: References: <20230523165942.2630-1-jszhang@kernel.org> MIME-Version: 1.0 Content-Disposition: inline In-Reply-To: X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20230523_171757_500127_6C373C1E X-CRM114-Status: GOOD ( 25.29 ) X-BeenThere: linux-riscv@lists.infradead.org X-Mailman-Version: 2.1.34 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Sender: "linux-riscv" Errors-To: linux-riscv-bounces+linux-riscv=archiver.kernel.org@lists.infradead.org On Wed, May 24, 2023 at 01:11:55AM +0800, Jisheng Zhang wrote: > On Wed, May 24, 2023 at 12:59:42AM +0800, Jisheng Zhang wrote: > > Attempt VMA lock-based page fault handling first, and fall back to the > > existing mmap_lock-based handling if that fails. > > > > A simple running the ebizzy benchmark on Lichee Pi 4A shows that > > PER_VMA_LOCK can improve the ebizzy benchmark by about 32.68%. In > > theory, the more CPUs, the bigger improvement, but I don't have any > > HW platform which has more than 4 CPUs. > > > > This is the riscv variant of "x86/mm: try VMA lock-based page fault > > handling first". > > > > Signed-off-by: Jisheng Zhang > > --- > > Any performance numbers are welcome! Especially the numbers on HW > > platforms with 8 or more CPUs. > > PS: run ebizzy as below: > ./ebizzy -mTt your_nr_cpus Sorry, should be ./ebizzy -mTt 2*your_nr_cpus > > > > arch/riscv/Kconfig | 1 + > > arch/riscv/mm/fault.c | 33 +++++++++++++++++++++++++++++++++ > > 2 files changed, 34 insertions(+) > > > > diff --git a/arch/riscv/Kconfig b/arch/riscv/Kconfig > > index 62e84fee2cfd..b958f67f9a12 100644 > > --- a/arch/riscv/Kconfig > > +++ b/arch/riscv/Kconfig > > @@ -42,6 +42,7 @@ config RISCV > > select ARCH_SUPPORTS_DEBUG_PAGEALLOC if MMU > > select ARCH_SUPPORTS_HUGETLBFS if MMU > > select ARCH_SUPPORTS_PAGE_TABLE_CHECK if MMU > > + select ARCH_SUPPORTS_PER_VMA_LOCK if MMU > > select ARCH_USE_MEMTEST > > select ARCH_USE_QUEUED_RWLOCKS > > select ARCH_WANT_DEFAULT_TOPDOWN_MMAP_LAYOUT if MMU > > diff --git a/arch/riscv/mm/fault.c b/arch/riscv/mm/fault.c > > index 8685f85a7474..eccdddf26f4b 100644 > > --- a/arch/riscv/mm/fault.c > > +++ b/arch/riscv/mm/fault.c > > @@ -286,6 +286,36 @@ void handle_page_fault(struct pt_regs *regs) > > flags |= FAULT_FLAG_WRITE; > > else if (cause == EXC_INST_PAGE_FAULT) > > flags |= FAULT_FLAG_INSTRUCTION; > > +#ifdef CONFIG_PER_VMA_LOCK > > + if (!(flags & FAULT_FLAG_USER)) > > + goto lock_mmap; > > + > > + vma = lock_vma_under_rcu(mm, addr); > > + if (!vma) > > + goto lock_mmap; > > + > > + if (unlikely(access_error(cause, vma))) { > > + vma_end_read(vma); > > + goto lock_mmap; > > + } > > + > > + fault = handle_mm_fault(vma, addr, flags | FAULT_FLAG_VMA_LOCK, regs); > > + vma_end_read(vma); > > + > > + if (!(fault & VM_FAULT_RETRY)) { > > + count_vm_vma_lock_event(VMA_LOCK_SUCCESS); > > + goto done; > > + } > > + count_vm_vma_lock_event(VMA_LOCK_RETRY); > > + > > + if (fault_signal_pending(fault, regs)) { > > + if (!user_mode(regs)) > > + no_context(regs, addr); > > + return; > > + } > > +lock_mmap: > > +#endif /* CONFIG_PER_VMA_LOCK */ > > + > > retry: > > mmap_read_lock(mm); > > vma = find_vma(mm, addr); > > @@ -355,6 +385,9 @@ void handle_page_fault(struct pt_regs *regs) > > > > mmap_read_unlock(mm); > > > > +#ifdef CONFIG_PER_VMA_LOCK > > +done: > > +#endif > > if (unlikely(fault & VM_FAULT_ERROR)) { > > tsk->thread.bad_cause = cause; > > mm_fault_error(regs, addr, fault); > > -- > > 2.40.1 > > > > > > _______________________________________________ > > linux-riscv mailing list > > linux-riscv@lists.infradead.org > > http://lists.infradead.org/mailman/listinfo/linux-riscv _______________________________________________ linux-riscv mailing list linux-riscv@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-riscv