linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
To: Wu Fengguang <fengguang.wu@intel.com>
Cc: kosaki.motohiro@jp.fujitsu.com, Rik van Riel <riel@redhat.com>,
	Christoph Lameter <cl@linux-foundation.org>,
	Peter Zijlstra <peterz@infradead.org>,
	Minchan Kim <minchan.kim@gmail.com>,
	Johannes Weiner <hannes@cmpxchg.org>,
	David Howells <dhowells@redhat.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	LKML <linux-kernel@vger.kernel.org>,
	"tytso@mit.edu" <tytso@mit.edu>,
	"linux-mm@kvack.org" <linux-mm@kvack.org>,
	"elladan@eskimo.com" <elladan@eskimo.com>,
	"npiggin@suse.de" <npiggin@suse.de>,
	"Barnes, Jesse" <jesse.barnes@intel.com>
Subject: Re: [PATCH] mm: count only reclaimable lru pages v2
Date: Fri, 17 Jul 2009 08:53:46 +0900 (JST)	[thread overview]
Message-ID: <20090717085108.A8FD.A69D9226@jp.fujitsu.com> (raw)
In-Reply-To: <20090716150901.GA31204@localhost>

> ---
> mm: count only reclaimable lru pages 
> 
> global_lru_pages() / zone_lru_pages() can be used in two ways:
> - to estimate max reclaimable pages in determine_dirtyable_memory()  
> - to calculate the slab scan ratio
> 
> When swap is full or not present, the anon lru lists are not reclaimable
> and also won't be scanned. So the anon pages shall not be counted in both
> usage scenarios. Also rename to _reclaimable_pages: now they are counting
> the possibly reclaimable lru pages.
> 
> It can greatly (and correctly) increase the slab scan rate under high memory
> pressure (when most file pages have been reclaimed and swap is full/absent),
> thus reduce false OOM kills.
> 
> Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
> Reviewed-by: Rik van Riel <riel@redhat.com>
> Reviewed-by: Christoph Lameter <cl@linux-foundation.org>
> Reviewed-by: Minchan Kim <minchan.kim@gmail.com>
> Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
> Signed-off-by: Wu Fengguang <fengguang.wu@intel.com>
> ---
>  include/linux/vmstat.h |   11 +-------
>  mm/page-writeback.c    |    5 ++-
>  mm/vmscan.c            |   50 ++++++++++++++++++++++++++++++---------
>  3 files changed, 44 insertions(+), 22 deletions(-)
> 
> --- linux.orig/include/linux/vmstat.h
> +++ linux/include/linux/vmstat.h
> @@ -166,15 +166,8 @@ static inline unsigned long zone_page_st
>  	return x;
>  }
>  
> -extern unsigned long global_lru_pages(void);
> -
> -static inline unsigned long zone_lru_pages(struct zone *zone)
> -{
> -	return (zone_page_state(zone, NR_ACTIVE_ANON)
> -		+ zone_page_state(zone, NR_ACTIVE_FILE)
> -		+ zone_page_state(zone, NR_INACTIVE_ANON)
> -		+ zone_page_state(zone, NR_INACTIVE_FILE));
> -}
> +extern unsigned long global_reclaimable_pages(void);
> +extern unsigned long zone_reclaimable_pages(struct zone *zone);
>  
>  #ifdef CONFIG_NUMA
>  /*
> --- linux.orig/mm/page-writeback.c
> +++ linux/mm/page-writeback.c
> @@ -380,7 +380,8 @@ static unsigned long highmem_dirtyable_m
>  		struct zone *z =
>  			&NODE_DATA(node)->node_zones[ZONE_HIGHMEM];
>  
> -		x += zone_page_state(z, NR_FREE_PAGES) + zone_lru_pages(z);
> +		x += zone_page_state(z, NR_FREE_PAGES) +
> +		     zone_reclaimable_pages(z);
>  	}
>  	/*
>  	 * Make sure that the number of highmem pages is never larger
> @@ -404,7 +405,7 @@ unsigned long determine_dirtyable_memory
>  {
>  	unsigned long x;
>  
> -	x = global_page_state(NR_FREE_PAGES) + global_lru_pages();
> +	x = global_page_state(NR_FREE_PAGES) + global_reclaimable_pages();
>  
>  	if (!vm_highmem_is_dirtyable)
>  		x -= highmem_dirtyable_memory(x);
> --- linux.orig/mm/vmscan.c
> +++ linux/mm/vmscan.c
> @@ -1735,7 +1735,7 @@ static unsigned long do_try_to_free_page
>  			if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL))
>  				continue;
>  
> -			lru_pages += zone_lru_pages(zone);
> +			lru_pages += zone_reclaimable_pages(zone);
>  		}
>  	}
>  
> @@ -1952,7 +1952,7 @@ loop_again:
>  		for (i = 0; i <= end_zone; i++) {
>  			struct zone *zone = pgdat->node_zones + i;
>  
> -			lru_pages += zone_lru_pages(zone);
> +			lru_pages += zone_reclaimable_pages(zone);
>  		}
>  
>  		/*
> @@ -1996,7 +1996,7 @@ loop_again:
>  			if (zone_is_all_unreclaimable(zone))
>  				continue;
>  			if (nr_slab == 0 && zone->pages_scanned >=
> -						(zone_lru_pages(zone) * 6))
> +					(zone_reclaimable_pages(zone) * 6))
>  					zone_set_flag(zone,
>  						      ZONE_ALL_UNRECLAIMABLE);
>  			/*
> @@ -2163,12 +2163,39 @@ void wakeup_kswapd(struct zone *zone, in
>  	wake_up_interruptible(&pgdat->kswapd_wait);
>  }
>  
> -unsigned long global_lru_pages(void)
> +/*
> + * The reclaimable count would be mostly accurate.
> + * The less reclaimable pages may be
> + * - mlocked pages, which will be moved to unevictable list when encountered
> + * - mapped pages, which may require several travels to be reclaimed
> + * - dirty pages, which is not "instantly" reclaimable
> + */
> +unsigned long global_reclaimable_pages(void)
>  {
> -	return global_page_state(NR_ACTIVE_ANON)
> -		+ global_page_state(NR_ACTIVE_FILE)
> -		+ global_page_state(NR_INACTIVE_ANON)
> -		+ global_page_state(NR_INACTIVE_FILE);
> +	int nr;
> +
> +	nr = global_page_state(NR_ACTIVE_FILE) +
> +	     global_page_state(NR_INACTIVE_FILE);
> +
> +	if (nr_swap_pages > 0)
> +		nr += global_page_state(NR_ACTIVE_ANON) +
> +		      global_page_state(NR_INACTIVE_ANON);
> +
> +	return nr;
> +}
> +
> +unsigned long zone_reclaimable_pages(struct zone *zone)
> +{
> +	int nr;
> +
> +	nr = zone_page_state(zone, NR_ACTIVE_FILE) +
> +	     zone_page_state(zone, NR_INACTIVE_FILE);
> +
> +	if (nr_swap_pages > 0)
> +		nr += zone_page_state(zone, NR_ACTIVE_ANON) +
> +		      zone_page_state(zone, NR_INACTIVE_ANON);
> +
> +	return nr;
>  }
>  
>  #ifdef CONFIG_HIBERNATION
> @@ -2240,7 +2267,7 @@ unsigned long shrink_all_memory(unsigned
>  
>  	current->reclaim_state = &reclaim_state;
>  
> -	lru_pages = global_lru_pages();
> +	lru_pages = global_reclaimable_pages();
>  	nr_slab = global_page_state(NR_SLAB_RECLAIMABLE);
>  	/* If slab caches are huge, it's better to hit them first */
>  	while (nr_slab >= lru_pages) {
> @@ -2282,7 +2309,7 @@ unsigned long shrink_all_memory(unsigned
>  
>  			reclaim_state.reclaimed_slab = 0;
>  			shrink_slab(sc.nr_scanned, sc.gfp_mask,
> -					global_lru_pages());
> +				    global_reclaimable_pages());
>  			sc.nr_reclaimed += reclaim_state.reclaimed_slab;
>  			if (sc.nr_reclaimed >= nr_pages)
>  				goto out;
> @@ -2299,7 +2326,8 @@ unsigned long shrink_all_memory(unsigned
>  	if (!sc.nr_reclaimed) {
>  		do {
>  			reclaim_state.reclaimed_slab = 0;
> -			shrink_slab(nr_pages, sc.gfp_mask, global_lru_pages());
> +			shrink_slab(nr_pages, sc.gfp_mask,
> +				    global_reclaimable_pages());
>  			sc.nr_reclaimed += reclaim_state.reclaimed_slab;
>  		} while (sc.nr_reclaimed < nr_pages &&
>  				reclaim_state.reclaimed_slab > 0);
> 

I feel like I already reviewed this patch past days..
Anyway,

	Reviewed-by: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>


--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  parent reply	other threads:[~2009-07-16 23:53 UTC|newest]

Thread overview: 20+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2009-07-16 13:34 [PATCH] mm: count only reclaimable lru pages Wu Fengguang
2009-07-16 13:40 ` Peter Zijlstra
2009-07-16 13:49 ` Rik van Riel
2009-07-16 14:00 ` Christoph Lameter
2009-07-16 14:25   ` Wu Fengguang
2009-07-16 14:28     ` Peter Zijlstra
2009-07-16 14:39       ` Christoph Lameter
2009-07-16 14:42         ` Rik van Riel
2009-07-16 15:09           ` [PATCH] mm: count only reclaimable lru pages v2 Wu Fengguang
2009-07-16 15:59             ` Johannes Weiner
2009-07-16 16:21             ` Christoph Lameter
2009-07-16 23:53             ` KOSAKI Motohiro [this message]
2009-07-16 14:17 ` [PATCH] mm: count only reclaimable lru pages Minchan Kim
2009-07-16 16:15 ` David Howells
2009-07-16 16:24   ` Rik van Riel
2009-07-16 17:11   ` David Howells
2009-07-16 18:51     ` Li, Ming Chun
2009-07-17  4:57       ` KOSAKI Motohiro
2009-07-17  6:32         ` Li, Ming Chun
2009-07-16 16:21 ` Jesse Barnes

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20090717085108.A8FD.A69D9226@jp.fujitsu.com \
    --to=kosaki.motohiro@jp.fujitsu.com \
    --cc=akpm@linux-foundation.org \
    --cc=cl@linux-foundation.org \
    --cc=dhowells@redhat.com \
    --cc=elladan@eskimo.com \
    --cc=fengguang.wu@intel.com \
    --cc=hannes@cmpxchg.org \
    --cc=jesse.barnes@intel.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=minchan.kim@gmail.com \
    --cc=npiggin@suse.de \
    --cc=peterz@infradead.org \
    --cc=riel@redhat.com \
    --cc=tytso@mit.edu \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).