From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from psmtp.com (na3sys010amx116.postini.com [74.125.245.116]) by kanga.kvack.org (Postfix) with SMTP id 6B9C16B0073 for ; Fri, 26 Oct 2012 08:53:47 -0400 (EDT) From: "Kirill A. Shutemov" Subject: [PATCH 2/2] numa, mm: consolidate error path in do_huge_pmd_numa_page() Date: Fri, 26 Oct 2012 15:54:35 +0300 Message-Id: <1351256077-1594-2-git-send-email-kirill.shutemov@linux.intel.com> In-Reply-To: <1351256077-1594-1-git-send-email-kirill.shutemov@linux.intel.com> References: <1351256077-1594-1-git-send-email-kirill.shutemov@linux.intel.com> Sender: owner-linux-mm@kvack.org List-ID: To: linux-mm@kvack.org Cc: Will Deacon , "Kirill A. Shutemov" , Andrew Morton , Andrea Arcangeli , Xiao Guangrong , Ingo Molnar , Peter Zijlstra , linux-kernel@vger.kernel.org From: "Kirill A. Shutemov" Let's move all error path code to the end if the function. It makes code more straight-forward. Signed-off-by: Kirill A. Shutemov --- mm/huge_memory.c | 44 ++++++++++++++++++++------------------------ 1 file changed, 20 insertions(+), 24 deletions(-) diff --git a/mm/huge_memory.c b/mm/huge_memory.c index 9bb2c23..95ec485 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -759,30 +759,14 @@ void do_huge_pmd_numa_page(struct mm_struct *mm, struct vm_area_struct *vma, goto unlock; page = pmd_page(entry); - if (page) { - VM_BUG_ON(!PageCompound(page) || !PageHead(page)); - - get_page(page); - node = mpol_misplaced(page, vma, haddr); - if (node != -1) - goto migrate; - } - -fixup: - /* change back to regular protection */ - entry = pmd_modify(entry, vma->vm_page_prot); - set_pmd_at(mm, haddr, pmd, entry); - update_mmu_cache_pmd(vma, address, entry); - -unlock: - spin_unlock(&mm->page_table_lock); - if (page) { - task_numa_fault(page_to_nid(page), HPAGE_PMD_NR); - put_page(page); - } - return; + if (!page) + goto fixup; + VM_BUG_ON(!PageCompound(page) || !PageHead(page)); -migrate: + get_page(page); + node = mpol_misplaced(page, vma, haddr); + if (node == -1) + goto fixup; spin_unlock(&mm->page_table_lock); lock_page(page); @@ -871,7 +855,19 @@ alloc_fail: page = NULL; goto unlock; } - goto fixup; +fixup: + /* change back to regular protection */ + entry = pmd_modify(entry, vma->vm_page_prot); + set_pmd_at(mm, haddr, pmd, entry); + update_mmu_cache_pmd(vma, address, entry); + +unlock: + spin_unlock(&mm->page_table_lock); + if (page) { + task_numa_fault(page_to_nid(page), HPAGE_PMD_NR); + put_page(page); + } + return; } int copy_huge_pmd(struct mm_struct *dst_mm, struct mm_struct *src_mm, -- 1.7.10.4 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org