From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-pg0-f69.google.com (mail-pg0-f69.google.com [74.125.83.69]) by kanga.kvack.org (Postfix) with ESMTP id C68926B026D for ; Thu, 26 Jan 2017 06:58:41 -0500 (EST) Received: by mail-pg0-f69.google.com with SMTP id 75so308735271pgf.3 for ; Thu, 26 Jan 2017 03:58:41 -0800 (PST) Received: from mga07.intel.com (mga07.intel.com. [134.134.136.100]) by mx.google.com with ESMTPS id j9si1196448pfc.290.2017.01.26.03.58.40 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Thu, 26 Jan 2017 03:58:40 -0800 (PST) From: "Kirill A. Shutemov" Subject: [PATCHv6 09/37] filemap: allocate huge page in pagecache_get_page(), if allowed Date: Thu, 26 Jan 2017 14:57:51 +0300 Message-Id: <20170126115819.58875-10-kirill.shutemov@linux.intel.com> In-Reply-To: <20170126115819.58875-1-kirill.shutemov@linux.intel.com> References: <20170126115819.58875-1-kirill.shutemov@linux.intel.com> Sender: owner-linux-mm@kvack.org List-ID: To: Theodore Ts'o , Andreas Dilger , Jan Kara , Andrew Morton Cc: Alexander Viro , Hugh Dickins , Andrea Arcangeli , Dave Hansen , Vlastimil Babka , Matthew Wilcox , Ross Zwisler , linux-ext4@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, linux-block@vger.kernel.org, "Kirill A. Shutemov" Write path allocate pages using pagecache_get_page(). We should be able to allocate huge pages there, if it's allowed. As usually, fallback to small pages, if failed. Signed-off-by: Kirill A. Shutemov --- mm/filemap.c | 17 +++++++++++++++-- 1 file changed, 15 insertions(+), 2 deletions(-) diff --git a/mm/filemap.c b/mm/filemap.c index 6cba69176ea9..4e398d5e4134 100644 --- a/mm/filemap.c +++ b/mm/filemap.c @@ -1374,13 +1374,16 @@ struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset, no_page: if (!page && (fgp_flags & FGP_CREAT)) { + pgoff_t hoffset; int err; if ((fgp_flags & FGP_WRITE) && mapping_cap_account_dirty(mapping)) gfp_mask |= __GFP_WRITE; if (fgp_flags & FGP_NOFS) gfp_mask &= ~__GFP_FS; - page = __page_cache_alloc(gfp_mask); + page = page_cache_alloc_huge(mapping, offset, gfp_mask); +no_huge: if (!page) + page = __page_cache_alloc(gfp_mask); if (!page) return NULL; @@ -1391,9 +1394,19 @@ struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset, if (fgp_flags & FGP_ACCESSED) __SetPageReferenced(page); - err = add_to_page_cache_lru(page, mapping, offset, + if (PageTransHuge(page)) + hoffset = round_down(offset, HPAGE_PMD_NR); + else + hoffset = offset; + + err = add_to_page_cache_lru(page, mapping, hoffset, gfp_mask & GFP_RECLAIM_MASK); if (unlikely(err)) { + if (PageTransHuge(page)) { + put_page(page); + page = NULL; + goto no_huge; + } put_page(page); page = NULL; if (err == -EEXIST) -- 2.11.0 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org