public inbox for linux-mm@kvack.org
 help / color / mirror / Atom feed
From: "Yin, Fengwei" <fengwei.yin@intel.com>
To: Kefeng Wang <wangkefeng.wang@huawei.com>,
	Yang Shi <shy828301@gmail.com>,
	"zhangpeng (AS)" <zhangpeng362@huawei.com>,
	Aneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: <linux-mm@kvack.org>, <linux-kernel@vger.kernel.org>,
	<akpm@linux-foundation.org>, Matthew Wilcox <willy@infradead.org>,
	<lstoakes@gmail.com>, <hughd@google.com>, <david@redhat.com>,
	<vbabka@suse.cz>, <peterz@infradead.org>, <mgorman@suse.de>,
	<mingo@redhat.com>, <riel@redhat.com>, <ying.huang@intel.com>,
	<hannes@cmpxchg.org>, Nanyong Sun <sunnanyong@huawei.com>
Subject: Re: [Question]: major faults are still triggered after mlockall when numa balancing
Date: Fri, 10 Nov 2023 11:50:32 +0800	[thread overview]
Message-ID: <98479379-0fff-409a-a60d-2233da114588@intel.com> (raw)
In-Reply-To: <56e1e123-f593-443a-be5b-754cbfb0e611@huawei.com>



On 11/10/2023 11:39 AM, Kefeng Wang wrote:
> 
> 
> On 2023/11/10 9:57, Yin, Fengwei wrote:
>>
>>
>> On 11/10/2023 6:54 AM, Yang Shi wrote:
>>> On Thu, Nov 9, 2023 at 5:48 AM zhangpeng (AS) <zhangpeng362@huawei.com> wrote:
>>>>
>>>> Hi everyone,
>>>>
>>>> There is a performance issue that has been bothering us recently.
>>>> This problem can reproduce in the latest mainline version (Linux 6.6).
>>>>
>>>> We use mlockall(MCL_CURRENT | MCL_FUTURE) in the user mode process
>>>> to avoid performance problems caused by major fault.
>>>>
>>>> There is a stage in numa fault which will set pte as 0 in do_numa_page() :
>>>> ptep_modify_prot_start() will clear the vmf->pte, until
>>>> ptep_modify_prot_commit() assign a value to the vmf->pte.
>>>>
>>>> For the data segment of the user-mode program, the global variable area
>>>> is a private mapping. After the pagecache is loaded, the private
>>>> anonymous page is generated after the COW is triggered. Mlockall can
>>>> lock COW pages (anonymous pages), but the original file pages cannot
>>>> be locked and may be reclaimed. If the global variable (private anon page)
>>>> is accessed when vmf->pte is zero which is concurrently set by numa fault,
>>>> a file page fault will be triggered.
>>>>
>>>> At this time, the original private file page may have been reclaimed.
>>>> If the page cache is not available at this time, a major fault will be
>>>> triggered and the file will be read, causing additional overhead.
>>>>
>>>> Our problem scenario is as follows:
>>>>
>>>> task 1                      task 2
>>>> ------                      ------
>>>> /* scan global variables */
>>>> do_numa_page()
>>>>     spin_lock(vmf->ptl)
>>>>     ptep_modify_prot_start()
>>>>     /* set vmf->pte as null */
>>>>                               /* Access global variables */
>>>>                               handle_pte_fault()
>>>>                                 /* no pte lock */
>>>>                                 do_pte_missing()
>>>>                                   do_fault()
>>>>                                     do_read_fault()
>>>>     ptep_modify_prot_commit()
>>>>     /* ptep update done */
>>>>     pte_unmap_unlock(vmf->pte, vmf->ptl)
>>>>                                       do_fault_around()
>>>>                                       __do_fault()
>>>>                                         filemap_fault()
>>>>                                           /* page cache is not available
>>>>                                           and a major fault is triggered */
>>>>                                           do_sync_mmap_readahead()
>>>>                                           /* page_not_uptodate and goto
>>>>                                           out_retry. */
>>>>
>>>> Is there any way to avoid such a major fault?
>>>
>>> IMHO I don't think it is a bug. The man page quoted by Willy says "All
>>> mapped pages are guaranteed to be resident in RAM when the call
>>> returns successfully", but the later COW already made the file page
>>> unmapped, right? The PTE pointed to the COW'ed anon page.
>>> Hypothetically if we kept the file page mlocked and unmapped,
>>> munlock() would have not munlocked the file page at all, it would be
>>> mlocked in memory forever.
>> But in this case, even the COW page is mlocked. There is small window
>> that PTE is set to null in do_numa_page(). data segment access (it's to
>> COW page which has nothing to do with original page cache) happens in
>> this small window will trigger filemap_fault() to fault in original
>> page cache.
>>
>> I had thought to do double check whether vmf->pte is NULL in do_read_fault().
>> But it's not reliable enough.
>>
>> Matthew's idea to use protnone to block both hardware accessing and
>> do_pte_missing() looks more promising to me.
> 
> Actual, we could revert the following patch to avoid this issue,
> but this workaroud from ppc...
> 
> commit cee216a696b2004017a5ecb583366093d90b1568
> Author: Aneesh Kumar K.V <aneesh.kumar@linux.vnet.ibm.com>
> Date:   Fri Feb 24 14:59:13 2017 -0800
> 
>     mm/autonuma: don't use set_pte_at when updating protnone ptes
> 
>     Architectures like ppc64, use privilege access bit to mark pte non
>     accessible.  This implies that kernel can do a copy_to_user to an
>     address marked for numa fault.  This also implies that there can be a
>     parallel hardware update for the pte.  set_pte_at cannot be used in such
>     scenarios.  Hence switch the pte update to use ptep_get_and_clear and
>     set_pte_at combination.
Oh. This means the protnone doesn't work for PPC.

> 
>>
>>
>> Regards
>> Yin, Fengwei
>>
>>>
>>>>
>>>> -- 
>>>> Best Regards,
>>>> Peng
>>>>


  reply	other threads:[~2023-11-10  3:50 UTC|newest]

Thread overview: 23+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-11-09 13:47 [Question]: major faults are still triggered after mlockall when numa balancing zhangpeng (AS)
2023-11-09 14:11 ` Peter Zijlstra
2023-11-09 14:29   ` Matthew Wilcox
2023-11-09 15:15     ` Yin, Fengwei
2023-11-09 17:27 ` Matthew Wilcox
2023-11-10  5:32   ` Huang, Ying
2023-11-10  9:04     ` Yin, Fengwei
2023-11-13  2:02       ` Huang, Ying
2023-11-14 11:23         ` Yin, Fengwei
2023-11-15  1:46           ` Huang, Ying
2023-11-10  9:39   ` zhangpeng (AS)
2023-11-09 22:54 ` Yang Shi
2023-11-10  1:57   ` Yin, Fengwei
2023-11-10  3:39     ` Kefeng Wang
2023-11-10  3:50       ` Yin, Fengwei [this message]
2023-11-10  4:00         ` Aneesh Kumar K V
2023-11-14  1:41     ` Yang Shi
2023-11-14 11:10       ` Yin, Fengwei
2023-11-09 23:21 ` Matthew Wilcox
2023-11-10  5:04 ` Aneesh Kumar K.V
2023-11-10  8:36   ` zhangpeng (AS)
2023-11-10  8:17 ` Aneesh Kumar K.V
2023-11-10  9:50   ` zhangpeng (AS)

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=98479379-0fff-409a-a60d-2233da114588@intel.com \
    --to=fengwei.yin@intel.com \
    --cc=akpm@linux-foundation.org \
    --cc=aneesh.kumar@linux.ibm.com \
    --cc=david@redhat.com \
    --cc=hannes@cmpxchg.org \
    --cc=hughd@google.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=lstoakes@gmail.com \
    --cc=mgorman@suse.de \
    --cc=mingo@redhat.com \
    --cc=peterz@infradead.org \
    --cc=riel@redhat.com \
    --cc=shy828301@gmail.com \
    --cc=sunnanyong@huawei.com \
    --cc=vbabka@suse.cz \
    --cc=wangkefeng.wang@huawei.com \
    --cc=willy@infradead.org \
    --cc=ying.huang@intel.com \
    --cc=zhangpeng362@huawei.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox