From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S932155AbbCCNgM (ORCPT ); Tue, 3 Mar 2015 08:36:12 -0500 Received: from mta-out1.inet.fi ([62.71.2.227]:41701 "EHLO kirsi1.inet.fi" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752490AbbCCNgK (ORCPT ); Tue, 3 Mar 2015 08:36:10 -0500 Date: Tue, 3 Mar 2015 15:35:24 +0200 From: "Kirill A. Shutemov" To: "Aneesh Kumar K.V" Cc: Rik van Riel , "Kirill A. Shutemov" , Andrew Morton , Andrea Arcangeli , Hugh Dickins , Dave Hansen , Mel Gorman , Vlastimil Babka , Christoph Lameter , Naoya Horiguchi , Steve Capper , Johannes Weiner , Michal Hocko , Jerome Marchand , linux-kernel@vger.kernel.org, linux-mm@kvack.org Subject: Re: [PATCHv3 03/24] mm: avoid PG_locked on tail pages Message-ID: <20150303133524.GA6111@node.dhcp.inet.fi> References: <1423757918-197669-1-git-send-email-kirill.shutemov@linux.intel.com> <1423757918-197669-4-git-send-email-kirill.shutemov@linux.intel.com> <54DD054E.7000605@redhat.com> <54DD08BC.2020008@redhat.com> <87egp69pyw.fsf@linux.vnet.ibm.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <87egp69pyw.fsf@linux.vnet.ibm.com> User-Agent: Mutt/1.5.23.1 (2014-03-12) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Mar 03, 2015 at 06:51:11PM +0530, Aneesh Kumar K.V wrote: > Rik van Riel writes: > > > -----BEGIN PGP SIGNED MESSAGE----- > > Hash: SHA1 > > > > On 02/12/2015 02:55 PM, Rik van Riel wrote: > >> On 02/12/2015 11:18 AM, Kirill A. Shutemov wrote: > > > >>> @@ -490,6 +493,7 @@ extern int > >>> wait_on_page_bit_killable_timeout(struct page *page, > >> > >>> static inline int wait_on_page_locked_killable(struct page *page) > >>> { + page = compound_head(page); if (PageLocked(page)) return > >>> wait_on_page_bit_killable(page, PG_locked); return 0; @@ -510,6 > >>> +514,7 @@ static inline void wake_up_page(struct page *page, int > >>> bit) */ static inline void wait_on_page_locked(struct page *page) > >>> { + page = compound_head(page); if (PageLocked(page)) > >>> wait_on_page_bit(page, PG_locked); } > >> > >> These are all atomic operations. > >> > >> This may be a stupid question with the answer lurking somewhere in > >> the other patches, but how do you ensure you operate on the right > >> page lock during a THP collapse or split? > > > > Kirill answered that question on IRC. > > > > The VM takes a refcount on a page before attempting to take a page > > lock, which prevents the THP code from doing anything with the > > page. In other words, while we have a refcount on the page, we > > will dereference the same page lock. > > Can we explain this more ? Don't we allow a thp split to happen even if > we have page refcount ?. The patchset changes this. Have you read the cover letter? -- Kirill A. Shutemov