From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 4D8BFC43334 for ; Thu, 16 Jun 2022 07:33:53 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id BBD366B0072; Thu, 16 Jun 2022 03:33:52 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id B1E496B0073; Thu, 16 Jun 2022 03:33:52 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 997FC6B0074; Thu, 16 Jun 2022 03:33:52 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0010.hostedemail.com [216.40.44.10]) by kanga.kvack.org (Postfix) with ESMTP id 845C96B0072 for ; Thu, 16 Jun 2022 03:33:52 -0400 (EDT) Received: from smtpin04.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay06.hostedemail.com (Postfix) with ESMTP id 47E4835A0A for ; Thu, 16 Jun 2022 07:33:52 +0000 (UTC) X-FDA: 79583284704.04.E955CF9 Received: from szxga02-in.huawei.com (szxga02-in.huawei.com [45.249.212.188]) by imf26.hostedemail.com (Postfix) with ESMTP id 2174C140097 for ; Thu, 16 Jun 2022 07:33:50 +0000 (UTC) Received: from canpemm500002.china.huawei.com (unknown [172.30.72.53]) by szxga02-in.huawei.com (SkyGuard) with ESMTP id 4LNv361hD4zSh0Y; Thu, 16 Jun 2022 15:30:26 +0800 (CST) Received: from [10.174.177.76] (10.174.177.76) by canpemm500002.china.huawei.com (7.192.104.244) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2375.24; Thu, 16 Jun 2022 15:33:44 +0800 Subject: Re: [PATCH 7/7] mm/khugepaged: try to free transhuge swapcache when possible To: Zach O'Keefe CC: , , , , , , , , , , , References: <20220611084731.55155-1-linmiaohe@huawei.com> <20220611084731.55155-8-linmiaohe@huawei.com> From: Miaohe Lin Message-ID: Date: Thu, 16 Jun 2022 15:33:43 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Thunderbird/78.6.0 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset="utf-8" Content-Language: en-US Content-Transfer-Encoding: 7bit X-Originating-IP: [10.174.177.76] X-ClientProxiedBy: dggems702-chm.china.huawei.com (10.3.19.179) To canpemm500002.china.huawei.com (7.192.104.244) X-CFilter-Loop: Reflected ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1655364832; a=rsa-sha256; cv=none; b=mI1sBh+auh6UP1d29B9Ci+XWXymaBznUHkw7JDoJCCakPUNF2UBzYpfdgeFTVoQVHMUb7e kLRsm8GRekbjLzXhrl5LA1ObUiZPZM7LHoYFOCI5auBiWBLmNXiQUhq5psm53j/W7E/8qz j4KAnf7Pn1Yamh+RmMvM/oGKxRibG5U= ARC-Authentication-Results: i=1; imf26.hostedemail.com; dkim=none; dmarc=pass (policy=quarantine) header.from=huawei.com; spf=pass (imf26.hostedemail.com: domain of linmiaohe@huawei.com designates 45.249.212.188 as permitted sender) smtp.mailfrom=linmiaohe@huawei.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1655364832; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=CZKfzlRcfc9afsXeiT44ul8hZK5c7D8yO2b7RCu/DYg=; b=qgv1FoiZbEhGLA/AA5/YpCmcbwPc06YvGQBqhEgo56ylmPGrkVgPUAW2M8cF3nsy47Btbx m/CjkPqgNm/GORXABIaIatVR0lnzMUGh4E+l4d8RsoH5ejeAOmW0HRXGS4lfvpC5wW0QML IJCreuv1xi9NsenmLg0qeghzyxlobNM= X-Rspamd-Queue-Id: 2174C140097 X-Rspam-User: Authentication-Results: imf26.hostedemail.com; dkim=none; dmarc=pass (policy=quarantine) header.from=huawei.com; spf=pass (imf26.hostedemail.com: domain of linmiaohe@huawei.com designates 45.249.212.188 as permitted sender) smtp.mailfrom=linmiaohe@huawei.com X-Rspamd-Server: rspam06 X-Stat-Signature: wfbaidyra93td8q8oy1mdu3dgx1ezmmx X-HE-Tag: 1655364830-274195 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On 2022/6/16 1:13, Zach O'Keefe wrote: > On 11 Jun 16:47, Miaohe Lin wrote: >> Transhuge swapcaches won't be freed in __collapse_huge_page_copy(). >> It's because release_pte_page() is not called for these pages and >> thus free_page_and_swap_cache can't grab the page lock. These pages >> won't be freed from swap cache even if we are the only user until >> next time reclaim. It shouldn't hurt indeed, but we could try to >> free these pages to save more memory for system. >> >> Signed-off-by: Miaohe Lin >> --- >> include/linux/swap.h | 5 +++++ >> mm/khugepaged.c | 1 + >> mm/swap.h | 5 ----- >> 3 files changed, 6 insertions(+), 5 deletions(-) >> >> diff --git a/include/linux/swap.h b/include/linux/swap.h >> index 8672a7123ccd..ccb83b12b724 100644 >> --- a/include/linux/swap.h >> +++ b/include/linux/swap.h >> @@ -456,6 +456,7 @@ static inline unsigned long total_swapcache_pages(void) >> return global_node_page_state(NR_SWAPCACHE); >> } >> >> +extern void free_swap_cache(struct page *page); >> extern void free_page_and_swap_cache(struct page *); >> extern void free_pages_and_swap_cache(struct page **, int); >> /* linux/mm/swapfile.c */ >> @@ -540,6 +541,10 @@ static inline void put_swap_device(struct swap_info_struct *si) >> /* used to sanity check ptes in zap_pte_range when CONFIG_SWAP=0 */ >> #define free_swap_and_cache(e) is_pfn_swap_entry(e) >> >> +static inline void free_swap_cache(struct page *page) >> +{ >> +} >> + >> static inline int add_swap_count_continuation(swp_entry_t swp, gfp_t gfp_mask) >> { >> return 0; >> diff --git a/mm/khugepaged.c b/mm/khugepaged.c >> index ee0a719c8be9..52109ad13f78 100644 >> --- a/mm/khugepaged.c >> +++ b/mm/khugepaged.c >> @@ -756,6 +756,7 @@ static void __collapse_huge_page_copy(pte_t *pte, struct page *page, >> list_for_each_entry_safe(src_page, tmp, compound_pagelist, lru) { >> list_del(&src_page->lru); >> release_pte_page(src_page); >> + free_swap_cache(src_page); >> } >> } > > Aside: in __collapse_huge_page_isolate() (and also here) why can't we just check > PageCompound(page) && page == compound_head(page) to only act on compound pages > once? AFAIK this would alleviate this compound_pagelist business.. > > Anyways, as-is, free_page_and_swap_cache() won't be able to do > try_to_free_swap(), since it can't grab page lock, put it will call put_page(). > I think (?) the last page ref might be dropped in release_pte_page(), so should > free_swap_cache() come before it? Thanks for catching this! If page is not in swapcache, the last page ref might be dropped. So it's bad to call free_swap_cache() after it. Thanks! > >> >> diff --git a/mm/swap.h b/mm/swap.h >> index 0193797b0c92..863f6086c916 100644 >> --- a/mm/swap.h >> +++ b/mm/swap.h >> @@ -41,7 +41,6 @@ void __delete_from_swap_cache(struct page *page, >> void delete_from_swap_cache(struct page *page); >> void clear_shadow_from_swap_cache(int type, unsigned long begin, >> unsigned long end); >> -void free_swap_cache(struct page *page); >> struct page *lookup_swap_cache(swp_entry_t entry, >> struct vm_area_struct *vma, >> unsigned long addr); >> @@ -81,10 +80,6 @@ static inline struct address_space *swap_address_space(swp_entry_t entry) >> return NULL; >> } >> >> -static inline void free_swap_cache(struct page *page) >> -{ >> -} >> - >> static inline void show_swap_cache_info(void) >> { >> } >> -- >> 2.23.0 >> >> > . >