From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from psmtp.com (na3sys010amx184.postini.com [74.125.245.184]) by kanga.kvack.org (Postfix) with SMTP id 33DBE6B00E8 for ; Fri, 16 Mar 2012 13:39:58 -0400 (EDT) Received: from /spool/local by e28smtp05.in.ibm.com with IBM ESMTP SMTP Gateway: Authorized Use Only! Violators will be prosecuted for from ; Fri, 16 Mar 2012 23:09:54 +0530 Received: from d28av03.in.ibm.com (d28av03.in.ibm.com [9.184.220.65]) by d28relay05.in.ibm.com (8.13.8/8.13.8/NCO v10.0) with ESMTP id q2GHdpZQ2191442 for ; Fri, 16 Mar 2012 23:09:51 +0530 Received: from d28av03.in.ibm.com (loopback [127.0.0.1]) by d28av03.in.ibm.com (8.14.4/8.13.1/NCO v10.0 AVout) with ESMTP id q2GN95th030882 for ; Sat, 17 Mar 2012 10:09:05 +1100 From: "Aneesh Kumar K.V" Subject: [PATCH -V4 05/10] hugetlb: add charge/uncharge calls for HugeTLB alloc/free Date: Fri, 16 Mar 2012 23:09:25 +0530 Message-Id: <1331919570-2264-6-git-send-email-aneesh.kumar@linux.vnet.ibm.com> In-Reply-To: <1331919570-2264-1-git-send-email-aneesh.kumar@linux.vnet.ibm.com> References: <1331919570-2264-1-git-send-email-aneesh.kumar@linux.vnet.ibm.com> Sender: owner-linux-mm@kvack.org List-ID: To: linux-mm@kvack.org, mgorman@suse.de, kamezawa.hiroyu@jp.fujitsu.com, dhillf@gmail.com, aarcange@redhat.com, mhocko@suse.cz, akpm@linux-foundation.org, hannes@cmpxchg.org Cc: linux-kernel@vger.kernel.org, cgroups@vger.kernel.org, "Aneesh Kumar K.V" From: "Aneesh Kumar K.V" This adds necessary charge/uncharge calls in the HugeTLB code Acked-by: Hillf Danton Signed-off-by: Aneesh Kumar K.V --- mm/hugetlb.c | 21 ++++++++++++++++++++- mm/memcontrol.c | 5 +++++ 2 files changed, 25 insertions(+), 1 deletions(-) diff --git a/mm/hugetlb.c b/mm/hugetlb.c index c672187..91361a0 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -21,6 +21,8 @@ #include #include #include +#include +#include #include #include @@ -542,6 +544,9 @@ static void free_huge_page(struct page *page) BUG_ON(page_mapcount(page)); INIT_LIST_HEAD(&page->lru); + if (mapping) + mem_cgroup_hugetlb_uncharge_page(hstate_index(h), + pages_per_huge_page(h), page); spin_lock(&hugetlb_lock); if (h->surplus_huge_pages_node[nid] && huge_page_order(h) < MAX_ORDER) { update_and_free_page(h, page); @@ -1019,12 +1024,15 @@ static void vma_commit_reservation(struct hstate *h, static struct page *alloc_huge_page(struct vm_area_struct *vma, unsigned long addr, int avoid_reserve) { + int ret, idx; struct hstate *h = hstate_vma(vma); struct page *page; + struct mem_cgroup *memcg = NULL; struct address_space *mapping = vma->vm_file->f_mapping; struct inode *inode = mapping->host; long chg; + idx = hstate_index(h); /* * Processes that did not create the mapping will have no reserves and * will not have accounted against quota. Check that the quota can be @@ -1039,6 +1047,12 @@ static struct page *alloc_huge_page(struct vm_area_struct *vma, if (hugetlb_get_quota(inode->i_mapping, chg)) return ERR_PTR(-ENOSPC); + ret = mem_cgroup_hugetlb_charge_page(idx, pages_per_huge_page(h), + &memcg); + if (ret) { + hugetlb_put_quota(inode->i_mapping, chg); + return ERR_PTR(-ENOSPC); + } spin_lock(&hugetlb_lock); page = dequeue_huge_page_vma(h, vma, addr, avoid_reserve); spin_unlock(&hugetlb_lock); @@ -1046,6 +1060,9 @@ static struct page *alloc_huge_page(struct vm_area_struct *vma, if (!page) { page = alloc_buddy_huge_page(h, NUMA_NO_NODE); if (!page) { + mem_cgroup_hugetlb_uncharge_memcg(idx, + pages_per_huge_page(h), + memcg); hugetlb_put_quota(inode->i_mapping, chg); return ERR_PTR(-ENOSPC); } @@ -1054,7 +1071,9 @@ static struct page *alloc_huge_page(struct vm_area_struct *vma, set_page_private(page, (unsigned long) mapping); vma_commit_reservation(h, vma, addr); - + /* update page cgroup details */ + mem_cgroup_hugetlb_commit_charge(idx, pages_per_huge_page(h), + memcg, page); return page; } diff --git a/mm/memcontrol.c b/mm/memcontrol.c index 4b36c5e..7a9ea94 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -2901,6 +2901,11 @@ __mem_cgroup_uncharge_common(struct page *page, enum charge_type ctype) if (PageSwapCache(page)) return NULL; + /* + * HugeTLB page uncharge happen in the HugeTLB compound page destructor + */ + if (PageHuge(page)) + return NULL; if (PageTransHuge(page)) { nr_pages <<= compound_order(page); -- 1.7.9 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: email@kvack.org