From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 579E7C001DF for ; Sat, 21 Oct 2023 03:31:14 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id CB0108D0009; Fri, 20 Oct 2023 23:31:13 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id C5FE08D0008; Fri, 20 Oct 2023 23:31:13 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id B4E8E8D0009; Fri, 20 Oct 2023 23:31:13 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0012.hostedemail.com [216.40.44.12]) by kanga.kvack.org (Postfix) with ESMTP id A58548D0008 for ; Fri, 20 Oct 2023 23:31:13 -0400 (EDT) Received: from smtpin07.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay03.hostedemail.com (Postfix) with ESMTP id 751F5A04A0 for ; Sat, 21 Oct 2023 03:31:13 +0000 (UTC) X-FDA: 81368042826.07.917B412 Received: from out30-124.freemail.mail.aliyun.com (out30-124.freemail.mail.aliyun.com [115.124.30.124]) by imf05.hostedemail.com (Postfix) with ESMTP id B9605100005 for ; Sat, 21 Oct 2023 03:31:10 +0000 (UTC) Authentication-Results: imf05.hostedemail.com; dkim=none; spf=pass (imf05.hostedemail.com: domain of baolin.wang@linux.alibaba.com designates 115.124.30.124 as permitted sender) smtp.mailfrom=baolin.wang@linux.alibaba.com; dmarc=pass (policy=none) header.from=alibaba.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1697859071; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=jrKShpqdDxPt2Y1v5BLIPzc8W70ba+P4HglwWDx9uok=; b=lSjCywDXUyBUc5QsVWKUYujJ6vwfyIEWzhlH2n+Vf8XlHH66E9b7eF73NGqldI8K7SAOfo SLGNIt+BJF+oxw089yuwF6jxv8jbIyuG0Ki52oP9xNYVxV5nEZo140wTGnDvz10O/PgxDH kNbOscRkQ996syEJfK3P46x2KsqC3Ao= ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1697859071; a=rsa-sha256; cv=none; b=TX7MC9xjW35jp47Si2TEHIx0CiZXDNbto1JQKgUqpAPOVv1a5O4fWLKuOFejwDDv+HKeoz WcFcKNfC+4HAMvYGXBkVP2PZsgI+CfDCd3XkOnwLS4YvAORC+aAzoXoL2jnRNA1qzf7Ob7 KlkUaNL7YNDyXRXu7tXUpsMp+G4JU7o= ARC-Authentication-Results: i=1; imf05.hostedemail.com; dkim=none; spf=pass (imf05.hostedemail.com: domain of baolin.wang@linux.alibaba.com designates 115.124.30.124 as permitted sender) smtp.mailfrom=baolin.wang@linux.alibaba.com; dmarc=pass (policy=none) header.from=alibaba.com X-Alimail-AntiSpam:AC=PASS;BC=-1|-1;BR=01201311R911e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=ay29a033018046050;MF=baolin.wang@linux.alibaba.com;NM=1;PH=DS;RN=9;SR=0;TI=SMTPD_---0VuYXo-8_1697859064; Received: from 192.168.43.40(mailfrom:baolin.wang@linux.alibaba.com fp:SMTPD_---0VuYXo-8_1697859064) by smtp.aliyun-inc.com; Sat, 21 Oct 2023 11:31:06 +0800 Message-ID: <5f6fc244-4e48-2118-4e8c-fcd00a0943ec@linux.alibaba.com> Date: Sat, 21 Oct 2023 11:31:20 +0800 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:102.0) Gecko/20100101 Thunderbird/102.15.1 Subject: Re: [PATCH v2] mm: migrate: record the mlocked page status to remove unnecessary lru drain To: Hugh Dickins Cc: akpm@linux-foundation.org, mgorman@techsingularity.net, vbabka@suse.cz, ying.huang@intel.com, ziy@nvidia.com, fengwei.yin@intel.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org References: <163ce2c0-9c8a-3db3-26a7-4d115fb95802@google.com> From: Baolin Wang In-Reply-To: <163ce2c0-9c8a-3db3-26a7-4d115fb95802@google.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Stat-Signature: gxe9mzs5zm13stt79sznd4qcfmw47jqr X-Rspamd-Server: rspam10 X-Rspamd-Queue-Id: B9605100005 X-Rspam-User: X-HE-Tag: 1697859070-734525 X-HE-Meta: U2FsdGVkX18Vi3lb9BfLN/4Q9ZDpVUXapgodoe/rtS8ih9KHOgTDE+IAHnlZJODF7YH8uvRIIUTDe3mFEe8C2IdNinYSCWdlrNya0VJ8/H9+zkFcv6OUgGrI8ERwrfADZC9aujTvsPM4x6nNuPhaegFxvazTO0VT4ydwklWJzjB0kTmclGoNu12ZrLVjnT+lDIQjm8ZkMnhCwqi0tHgMLrqObjxraQDkQjTm/v9yFS0VO9PGR/esLyobDi2w+HTBZ3kpyiYUip3gCWiqVLjMmdFZLxWcftE/lDsjnojkzs2gDCKbfXduETxq557VwdMOGSp6aXWRBmAnUL9soIJrRBlO+x/Voz0qDDcZ2ym+RSoDTxEsU67lzcE7Zh5oIJAwYd2kY4CGucD2H+YjdvJBZ7KSIiXgcR/L/3jH65eYgH5QELi2y+2YEMNrGgabFdLGotBTR4rDLgf5Hhs+yrs/sUv7vuN6IGg/rWRLR+xQk1e8nLV7u7XdJ5MrUAGkSrcLHlqqyzQaDHxW6KMQD4bPhM8b09oVyl63vph16seH+yZ9IPqHKvxB3v+qsDlsMBpJ8vB06Txm48VNOA/tqZ89fLYwmnmfCADSDz/N9f5FbgaoOk5oMlsmT8cBLSUeghmQmg+iBqWv0fHgJQTDh0QZuZAWhLUKaZyR8UwiEfgfyyfVuPu9KwBBEHeb6LH7h1onoGX997kP7Qsj4Yp3X0w/CvFhbj5keS6ZwyZQNMVVsP4qIaJi+Mf6SShclXEDV27qEEQOtJb1OQ1T4D0zAgxtRtoyGPWdfmGLPb6QzXyAvXpSM9g1ObeLo5OADJD9YYKpaql1Gn/7qBflUtFVDy++qjuvR74rTZl8sry/Woro30E0/vrElT3P1RabUSl6liuMJwzAw85Yw1udrwxm+/T9quEdQjQ0mmDLbOvbNyY0z4gZ5JYBOERDOmFXzqBXRGYxXGwRgF9WW/mpD8Vlfli UNdQiWd0 EZyr+pEiHT3dqmmyt6z6kWWY0cADaPd/gcoJDYPq8/vCau5nCzIxB1oDzp7sW/AIxZB1Zu9GAMqv3blrOzn8Kxx0r4z6GQEDDh8SL X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On 10/20/2023 12:48 PM, Hugh Dickins wrote: > On Fri, 20 Oct 2023, Baolin Wang wrote: > >> When doing compaction, I found the lru_add_drain() is an obvious hotspot >> when migrating pages. The distribution of this hotspot is as follows: >> - 18.75% compact_zone >> - 17.39% migrate_pages >> - 13.79% migrate_pages_batch >> - 11.66% migrate_folio_move >> - 7.02% lru_add_drain >> + 7.02% lru_add_drain_cpu >> + 3.00% move_to_new_folio >> 1.23% rmap_walk >> + 1.92% migrate_folio_unmap >> + 3.20% migrate_pages_sync >> + 0.90% isolate_migratepages >> >> The lru_add_drain() was added by commit c3096e6782b7 ("mm/migrate: >> __unmap_and_move() push good newpage to LRU") to drain the newpage to LRU >> immediately, to help to build up the correct newpage->mlock_count in >> remove_migration_ptes() for mlocked pages. However, if there are no mlocked >> pages are migrating, then we can avoid this lru drain operation, especailly >> for the heavy concurrent scenarios. >> >> So we can record the source pages' mlocked status in migrate_folio_unmap(), >> and only drain the lru list when the mlocked status is set in migrate_folio_move(). >> In addition, the page was already isolated from lru when migrating, so checking >> the mlocked status is stable by folio_test_mlocked() in migrate_folio_unmap(). >> >> After this patch, I can see the hotpot of the lru_add_drain() is gone: >> - 9.41% migrate_pages_batch >> - 6.15% migrate_folio_move >> - 3.64% move_to_new_folio >> + 1.80% migrate_folio_extra >> + 1.70% buffer_migrate_folio >> + 1.41% rmap_walk >> + 0.62% folio_add_lru >> + 3.07% migrate_folio_unmap >> >> Meanwhile, the compaction latency shows some improvements when running >> thpscale: >> base patched >> Amean fault-both-1 1131.22 ( 0.00%) 1112.55 * 1.65%* >> Amean fault-both-3 2489.75 ( 0.00%) 2324.15 * 6.65%* >> Amean fault-both-5 3257.37 ( 0.00%) 3183.18 * 2.28%* >> Amean fault-both-7 4257.99 ( 0.00%) 4079.04 * 4.20%* >> Amean fault-both-12 6614.02 ( 0.00%) 6075.60 * 8.14%* >> Amean fault-both-18 10607.78 ( 0.00%) 8978.86 * 15.36%* >> Amean fault-both-24 14911.65 ( 0.00%) 11619.55 * 22.08%* >> Amean fault-both-30 14954.67 ( 0.00%) 14925.66 * 0.19%* >> Amean fault-both-32 16654.87 ( 0.00%) 15580.31 * 6.45%* >> > > Seems a sensible change with good results (I'll conceal how little of > the stats I understand, I expect everyone else understands them: in my > naivety, I'm mainly curious why rmap_walk's 1.23% didn't get a + on it). TBH, I also don't know why the rmap_walk didn't get a + on it, let me check it again. >> Signed-off-by: Baolin Wang >> --- >> Chages from v1: >> - Use separate flags in __migrate_folio_record() to avoid to pack flags >> in each call site per Ying. >> --- >> mm/migrate.c | 47 +++++++++++++++++++++++++++++++++++------------ >> 1 file changed, 35 insertions(+), 12 deletions(-) >> >> diff --git a/mm/migrate.c b/mm/migrate.c >> index 125194f5af0f..fac96139dbba 100644 >> --- a/mm/migrate.c >> +++ b/mm/migrate.c >> @@ -1027,22 +1027,39 @@ union migration_ptr { >> struct anon_vma *anon_vma; >> struct address_space *mapping; >> }; >> + >> +enum { >> + PAGE_WAS_MAPPED = 1 << 0, >> + PAGE_WAS_MLOCKED = 1 << 1, >> +}; >> + > > I was whispering to myself "I bet someone will suggest BIT()"; > and indeed that someone has turned out to be Huang, Ying. Sure. > >> static void __migrate_folio_record(struct folio *dst, >> - unsigned long page_was_mapped, >> + unsigned int page_was_mapped, >> + unsigned int page_was_mlocked, >> struct anon_vma *anon_vma) >> { >> union migration_ptr ptr = { .anon_vma = anon_vma }; >> + unsigned long page_flags = 0; > > Huang, Ying preferred a different name, me too: old_page_state? OK, sounds better to me. > >> + >> + if (page_was_mapped) >> + page_flags |= PAGE_WAS_MAPPED; >> + if (page_was_mlocked) >> + page_flags |= PAGE_WAS_MLOCKED; > > What's annoying me about the patch is all this mix of page_was_mapped and > page_was_mlocked variables, then the old_page_state bits. Can't it be > done with PAGE_WAS_ bits in old_page_state throughout, without any > page_was_mapped and page_was_mlocked variables? Yes, good point. Let me try it. Thanks for your comments.