From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from psmtp.com (na3sys010amx177.postini.com [74.125.245.177]) by kanga.kvack.org (Postfix) with SMTP id 3000E6B0070 for ; Tue, 2 Oct 2012 11:19:02 -0400 (EDT) From: "Kirill A. Shutemov" Subject: [PATCH v3 08/10] thp: setup huge zero page on non-write page fault Date: Tue, 2 Oct 2012 18:19:30 +0300 Message-Id: <1349191172-28855-9-git-send-email-kirill.shutemov@linux.intel.com> In-Reply-To: <1349191172-28855-1-git-send-email-kirill.shutemov@linux.intel.com> References: <1349191172-28855-1-git-send-email-kirill.shutemov@linux.intel.com> Sender: owner-linux-mm@kvack.org List-ID: To: Andrew Morton , linux-mm@kvack.org Cc: Andrea Arcangeli , Andi Kleen , "H. Peter Anvin" , linux-kernel@vger.kernel.org, "Kirill A. Shutemov" , "Kirill A. Shutemov" From: "Kirill A. Shutemov" All code paths seems covered. Now we can map huge zero page on read page fault. Signed-off-by: Kirill A. Shutemov Reviewed-by: Andrea Arcangeli --- mm/huge_memory.c | 10 ++++++++++ 1 files changed, 10 insertions(+), 0 deletions(-) diff --git a/mm/huge_memory.c b/mm/huge_memory.c index 3f1c59c..a5b9282 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -751,6 +751,16 @@ int do_huge_pmd_anonymous_page(struct mm_struct *mm, struct vm_area_struct *vma, return VM_FAULT_OOM; if (unlikely(khugepaged_enter(vma))) return VM_FAULT_OOM; + if (!(flags & FAULT_FLAG_WRITE)) { + pgtable_t pgtable; + pgtable = pte_alloc_one(mm, haddr); + if (unlikely(!pgtable)) + goto out; + spin_lock(&mm->page_table_lock); + set_huge_zero_page(pgtable, mm, vma, haddr, pmd); + spin_unlock(&mm->page_table_lock); + return 0; + } page = alloc_hugepage_vma(transparent_hugepage_defrag(vma), vma, haddr, numa_node_id(), 0); if (unlikely(!page)) { -- 1.7.7.6 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org