linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: Mel Gorman <mel@csn.ul.ie>
To: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Cc: Linux Memory Management List <linux-mm@kvack.org>,
	Christoph Lameter <cl@linux-foundation.org>,
	Nick Piggin <npiggin@suse.de>,
	Linux Kernel Mailing List <linux-kernel@vger.kernel.org>,
	Lin Ming <ming.m.lin@intel.com>,
	Zhang Yanmin <yanmin_zhang@linux.intel.com>,
	Peter Zijlstra <peterz@infradead.org>,
	Pekka Enberg <penberg@cs.helsinki.fi>,
	Andrew Morton <akpm@linux-foundation.org>
Subject: Re: [PATCH 09/25] Calculate the migratetype for allocation only once
Date: Tue, 21 Apr 2009 09:35:13 +0100	[thread overview]
Message-ID: <20090421083513.GC12713@csn.ul.ie> (raw)
In-Reply-To: <20090421160729.F136.A69D9226@jp.fujitsu.com>

On Tue, Apr 21, 2009 at 04:37:28PM +0900, KOSAKI Motohiro wrote:
> > GFP mask is converted into a migratetype when deciding which pagelist to
> > take a page from. However, it is happening multiple times per
> > allocation, at least once per zone traversed. Calculate it once.
> > 
> > Signed-off-by: Mel Gorman <mel@csn.ul.ie>
> > ---
> >  mm/page_alloc.c |   43 ++++++++++++++++++++++++++-----------------
> >  1 files changed, 26 insertions(+), 17 deletions(-)
> > 
> > diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> > index b27bcde..f960cf5 100644
> > --- a/mm/page_alloc.c
> > +++ b/mm/page_alloc.c
> > @@ -1065,13 +1065,13 @@ void split_page(struct page *page, unsigned int order)
> >   * or two.
> >   */
> >  static struct page *buffered_rmqueue(struct zone *preferred_zone,
> > -			struct zone *zone, int order, gfp_t gfp_flags)
> > +			struct zone *zone, int order, gfp_t gfp_flags,
> > +			int migratetype)
> >  {
> >  	unsigned long flags;
> >  	struct page *page;
> >  	int cold = !!(gfp_flags & __GFP_COLD);
> >  	int cpu;
> > -	int migratetype = allocflags_to_migratetype(gfp_flags);
> 
> hmmm....
> 
> allocflags_to_migratetype() is very cheap function and buffered_rmqueue()
> and other non-inline static function isn't guranteed inlined.
> 

A later patch makes them inlined due to the fact there is only one call
site.

> I don't think this patch improve performance on x86.
> and, I have one comment to allocflags_to_migratetype.
> 
> -------------------------------------------------------------------
> /* Convert GFP flags to their corresponding migrate type */
> static inline int allocflags_to_migratetype(gfp_t gfp_flags)
> {
>         WARN_ON((gfp_flags & GFP_MOVABLE_MASK) == GFP_MOVABLE_MASK);
> 
>         if (unlikely(page_group_by_mobility_disabled))
>                 return MIGRATE_UNMOVABLE;
> 
>         /* Group based on mobility */
>         return (((gfp_flags & __GFP_MOVABLE) != 0) << 1) |
>                 ((gfp_flags & __GFP_RECLAIMABLE) != 0);
> }
> -------------------------------------------------------------------
> 
> s/WARN_ON/VM_BUG_ON/ is better?
> 

I wanted to catch out-of-tree drivers but it's been a while so maybe VM_BUG_ON
wouldn't hurt. I can add a patch that does that a pass 2 of improving the
allocator or would you prefer to see it now?

> GFP_MOVABLE_MASK makes 3. 3 mean MIGRATE_RESERVE. it seems obviously bug.
> 

Short answer;
No, GFP flags that result in MIGRATE_RESERVE is a bug. The caller should
never want to be allocating from there.

Longer answer;
The size of the MIGRATE_RESERVE depends on the number of free pages that
must be kept in the zone. Because GFP flags never result in here, the
area is only used when the alternative is to fail the allocation and the
watermarks are still met. The intention is that high-order atomic
allocations that were short lived may be allocated from here. This was
to preserve a behaviour in the allocator before MIGRATE_RESERVE was
introduced. It makes no sense for a caller to allocate directly out of
here and in fact the fallback list for MIGRATE_RESERVE are useless


> >  
> >  again:
> >  	cpu  = get_cpu();
> > @@ -1397,7 +1397,7 @@ static void zlc_mark_zone_full(struct zonelist *zonelist, struct zoneref *z)
> >  static struct page *
> >  get_page_from_freelist(gfp_t gfp_mask, nodemask_t *nodemask, unsigned int order,
> >  		struct zonelist *zonelist, int high_zoneidx, int alloc_flags,
> > -		struct zone *preferred_zone)
> > +		struct zone *preferred_zone, int migratetype)
> >  {
> >  	struct zoneref *z;
> >  	struct page *page = NULL;
> > @@ -1449,7 +1449,8 @@ zonelist_scan:
> >  			}
> >  		}
> >  
> > -		page = buffered_rmqueue(preferred_zone, zone, order, gfp_mask);
> > +		page = buffered_rmqueue(preferred_zone, zone, order,
> > +						gfp_mask, migratetype);
> >  		if (page)
> >  			break;
> >  this_zone_full:
> > @@ -1513,7 +1514,8 @@ should_alloc_retry(gfp_t gfp_mask, unsigned int order,
> >  static inline struct page *
> >  __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order,
> >  	struct zonelist *zonelist, enum zone_type high_zoneidx,
> > -	nodemask_t *nodemask, struct zone *preferred_zone)
> > +	nodemask_t *nodemask, struct zone *preferred_zone,
> > +	int migratetype)
> >  {
> >  	struct page *page;
> >  
> > @@ -1531,7 +1533,7 @@ __alloc_pages_may_oom(gfp_t gfp_mask, unsigned int order,
> >  	page = get_page_from_freelist(gfp_mask|__GFP_HARDWALL, nodemask,
> >  		order, zonelist, high_zoneidx,
> >  		ALLOC_WMARK_HIGH|ALLOC_CPUSET,
> > -		preferred_zone);
> > +		preferred_zone, migratetype);
> >  	if (page)
> >  		goto out;
> >  
> > @@ -1552,7 +1554,7 @@ static inline struct page *
> >  __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order,
> >  	struct zonelist *zonelist, enum zone_type high_zoneidx,
> >  	nodemask_t *nodemask, int alloc_flags, struct zone *preferred_zone,
> > -	unsigned long *did_some_progress)
> > +	int migratetype, unsigned long *did_some_progress)
> >  {
> >  	struct page *page = NULL;
> >  	struct reclaim_state reclaim_state;
> > @@ -1585,7 +1587,8 @@ __alloc_pages_direct_reclaim(gfp_t gfp_mask, unsigned int order,
> >  	if (likely(*did_some_progress))
> >  		page = get_page_from_freelist(gfp_mask, nodemask, order,
> >  					zonelist, high_zoneidx,
> > -					alloc_flags, preferred_zone);
> > +					alloc_flags, preferred_zone,
> > +					migratetype);
> >  	return page;
> >  }
> >  
> > @@ -1606,14 +1609,15 @@ is_allocation_high_priority(struct task_struct *p, gfp_t gfp_mask)
> >  static inline struct page *
> >  __alloc_pages_high_priority(gfp_t gfp_mask, unsigned int order,
> >  	struct zonelist *zonelist, enum zone_type high_zoneidx,
> > -	nodemask_t *nodemask, struct zone *preferred_zone)
> > +	nodemask_t *nodemask, struct zone *preferred_zone,
> > +	int migratetype)
> >  {
> >  	struct page *page;
> >  
> >  	do {
> >  		page = get_page_from_freelist(gfp_mask, nodemask, order,
> >  			zonelist, high_zoneidx, ALLOC_NO_WATERMARKS,
> > -			preferred_zone);
> > +			preferred_zone, migratetype);
> >  
> >  		if (!page && gfp_mask & __GFP_NOFAIL)
> >  			congestion_wait(WRITE, HZ/50);
> > @@ -1636,7 +1640,8 @@ void wake_all_kswapd(unsigned int order, struct zonelist *zonelist,
> >  static inline struct page *
> >  __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
> >  	struct zonelist *zonelist, enum zone_type high_zoneidx,
> > -	nodemask_t *nodemask, struct zone *preferred_zone)
> > +	nodemask_t *nodemask, struct zone *preferred_zone,
> > +	int migratetype)
> >  {
> >  	const gfp_t wait = gfp_mask & __GFP_WAIT;
> >  	struct page *page = NULL;
> > @@ -1687,14 +1692,16 @@ restart:
> >  	 */
> >  	page = get_page_from_freelist(gfp_mask, nodemask, order, zonelist,
> >  						high_zoneidx, alloc_flags,
> > -						preferred_zone);
> > +						preferred_zone,
> > +						migratetype);
> >  	if (page)
> >  		goto got_pg;
> >  
> >  	/* Allocate without watermarks if the context allows */
> >  	if (is_allocation_high_priority(p, gfp_mask))
> >  		page = __alloc_pages_high_priority(gfp_mask, order,
> > -			zonelist, high_zoneidx, nodemask, preferred_zone);
> > +			zonelist, high_zoneidx, nodemask, preferred_zone,
> > +			migratetype);
> >  	if (page)
> >  		goto got_pg;
> >  
> > @@ -1707,7 +1714,7 @@ restart:
> >  					zonelist, high_zoneidx,
> >  					nodemask,
> >  					alloc_flags, preferred_zone,
> > -					&did_some_progress);
> > +					migratetype, &did_some_progress);
> >  	if (page)
> >  		goto got_pg;
> >  
> > @@ -1719,7 +1726,8 @@ restart:
> >  		if ((gfp_mask & __GFP_FS) && !(gfp_mask & __GFP_NORETRY)) {
> >  			page = __alloc_pages_may_oom(gfp_mask, order,
> >  					zonelist, high_zoneidx,
> > -					nodemask, preferred_zone);
> > +					nodemask, preferred_zone,
> > +					migratetype);
> >  			if (page)
> >  				goto got_pg;
> >  
> > @@ -1758,6 +1766,7 @@ __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order,
> >  	enum zone_type high_zoneidx = gfp_zone(gfp_mask);
> >  	struct zone *preferred_zone;
> >  	struct page *page;
> > +	int migratetype = allocflags_to_migratetype(gfp_mask);
> >  
> >  	lockdep_trace_alloc(gfp_mask);
> >  
> > @@ -1783,11 +1792,11 @@ __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order,
> >  	/* First allocation attempt */
> >  	page = get_page_from_freelist(gfp_mask|__GFP_HARDWALL, nodemask, order,
> >  			zonelist, high_zoneidx, ALLOC_WMARK_LOW|ALLOC_CPUSET,
> > -			preferred_zone);
> > +			preferred_zone, migratetype);
> >  	if (unlikely(!page))
> >  		page = __alloc_pages_slowpath(gfp_mask, order,
> >  				zonelist, high_zoneidx, nodemask,
> > -				preferred_zone);
> > +				preferred_zone, migratetype);
> >  
> >  	return page;
> >  }
> > -- 
> > 1.5.6.5
> > 
> 
> 
> 

-- 
Mel Gorman
Part-time Phd Student                          Linux Technology Center
University of Limerick                         IBM Dublin Software Lab

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  reply	other threads:[~2009-04-21  8:35 UTC|newest]

Thread overview: 106+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2009-04-20 22:19 [PATCH 00/25] Cleanup and optimise the page allocator V6 Mel Gorman
2009-04-20 22:19 ` [PATCH 01/25] Replace __alloc_pages_internal() with __alloc_pages_nodemask() Mel Gorman
2009-04-21  1:44   ` KOSAKI Motohiro
2009-04-21  5:55   ` Pekka Enberg
2009-04-20 22:19 ` [PATCH 02/25] Do not sanity check order in the fast path Mel Gorman
2009-04-21  1:45   ` KOSAKI Motohiro
2009-04-21  5:55   ` Pekka Enberg
2009-04-20 22:19 ` [PATCH 03/25] Do not check NUMA node ID when the caller knows the node is valid Mel Gorman
2009-04-21  2:44   ` KOSAKI Motohiro
2009-04-21  6:00   ` Pekka Enberg
2009-04-21  6:33   ` Paul Mundt
2009-04-20 22:19 ` [PATCH 04/25] Check only once if the zonelist is suitable for the allocation Mel Gorman
2009-04-21  3:03   ` KOSAKI Motohiro
2009-04-21  7:09   ` Pekka Enberg
2009-04-20 22:19 ` [PATCH 05/25] Break up the allocator entry point into fast and slow paths Mel Gorman
2009-04-21  6:35   ` KOSAKI Motohiro
2009-04-21  7:13     ` Pekka Enberg
2009-04-21  9:30       ` Mel Gorman
2009-04-21  9:29     ` Mel Gorman
2009-04-21 10:44       ` KOSAKI Motohiro
2009-04-20 22:19 ` [PATCH 06/25] Move check for disabled anti-fragmentation out of fastpath Mel Gorman
2009-04-21  6:37   ` KOSAKI Motohiro
2009-04-20 22:19 ` [PATCH 07/25] Check in advance if the zonelist needs additional filtering Mel Gorman
2009-04-21  6:52   ` KOSAKI Motohiro
2009-04-21  9:47     ` Mel Gorman
2009-04-21  7:21   ` Pekka Enberg
2009-04-21  9:49     ` Mel Gorman
2009-04-20 22:19 ` [PATCH 08/25] Calculate the preferred zone for allocation only once Mel Gorman
2009-04-21  7:03   ` KOSAKI Motohiro
2009-04-21  8:23     ` Mel Gorman
2009-04-21  7:37   ` Pekka Enberg
2009-04-21  8:27     ` Mel Gorman
2009-04-21  8:29       ` Pekka Enberg
2009-04-20 22:19 ` [PATCH 09/25] Calculate the migratetype " Mel Gorman
2009-04-21  7:37   ` KOSAKI Motohiro
2009-04-21  8:35     ` Mel Gorman [this message]
2009-04-21 10:19       ` KOSAKI Motohiro
2009-04-21 10:30         ` Mel Gorman
2009-04-20 22:19 ` [PATCH 10/25] Calculate the alloc_flags " Mel Gorman
2009-04-21  9:03   ` KOSAKI Motohiro
2009-04-21 10:05     ` Mel Gorman
2009-04-21 10:12       ` KOSAKI Motohiro
2009-04-21 10:37         ` Mel Gorman
2009-04-21 10:40           ` KOSAKI Motohiro
2009-04-20 22:19 ` [PATCH 11/25] Calculate the cold parameter " Mel Gorman
2009-04-21  7:43   ` Pekka Enberg
2009-04-21  8:41     ` Mel Gorman
2009-04-21  9:07   ` KOSAKI Motohiro
2009-04-21 10:08     ` Mel Gorman
2009-04-21 14:59     ` Christoph Lameter
2009-04-21 14:58   ` Christoph Lameter
2009-04-20 22:19 ` [PATCH 12/25] Remove a branch by assuming __GFP_HIGH == ALLOC_HIGH Mel Gorman
2009-04-21  7:46   ` Pekka Enberg
2009-04-21  8:45     ` Mel Gorman
2009-04-21 10:25       ` Pekka Enberg
2009-04-21  9:08   ` KOSAKI Motohiro
2009-04-21 10:31     ` KOSAKI Motohiro
2009-04-21 10:43       ` Mel Gorman
2009-04-20 22:19 ` [PATCH 13/25] Inline __rmqueue_smallest() Mel Gorman
2009-04-21  7:58   ` Pekka Enberg
2009-04-21  8:48     ` Mel Gorman
2009-04-21  9:52   ` KOSAKI Motohiro
2009-04-21 10:11     ` Mel Gorman
2009-04-21 10:22       ` KOSAKI Motohiro
2009-04-20 22:20 ` [PATCH 14/25] Inline buffered_rmqueue() Mel Gorman
2009-04-21  9:56   ` KOSAKI Motohiro
2009-04-20 22:20 ` [PATCH 15/25] Inline __rmqueue_fallback() Mel Gorman
2009-04-21  9:56   ` KOSAKI Motohiro
2009-04-20 22:20 ` [PATCH 16/25] Save text by reducing call sites of __rmqueue() Mel Gorman
2009-04-21 10:47   ` KOSAKI Motohiro
2009-04-20 22:20 ` [PATCH 17/25] Do not call get_pageblock_migratetype() more than necessary Mel Gorman
2009-04-21 11:03   ` KOSAKI Motohiro
2009-04-21 16:12     ` Mel Gorman
2009-04-22  2:25       ` KOSAKI Motohiro
2009-04-20 22:20 ` [PATCH 18/25] Do not disable interrupts in free_page_mlock() Mel Gorman
2009-04-21  7:55   ` Pekka Enberg
2009-04-21  8:50     ` Mel Gorman
2009-04-21 15:05       ` Christoph Lameter
2009-04-22  0:13   ` KOSAKI Motohiro
2009-04-22 14:43     ` Lee Schermerhorn
2009-04-20 22:20 ` [PATCH 19/25] Do not setup zonelist cache when there is only one node Mel Gorman
2009-04-20 22:20 ` [PATCH 20/25] Do not check for compound pages during the page allocator sanity checks Mel Gorman
2009-04-22  0:20   ` KOSAKI Motohiro
2009-04-22 10:09     ` Mel Gorman
2009-04-22 10:41       ` KOSAKI Motohiro
2009-04-20 22:20 ` [PATCH 21/25] Use allocation flags as an index to the zone watermark Mel Gorman
2009-04-22  0:26   ` KOSAKI Motohiro
2009-04-22  0:41     ` David Rientjes
2009-04-22 10:21     ` Mel Gorman
2009-04-22 10:23       ` Mel Gorman
2009-04-20 22:20 ` [PATCH 22/25] Update NR_FREE_PAGES only as necessary Mel Gorman
2009-04-22  0:35   ` KOSAKI Motohiro
2009-04-20 22:20 ` [PATCH 23/25] Get the pageblock migratetype without disabling interrupts Mel Gorman
2009-04-20 22:20 ` [PATCH 24/25] Re-sort GFP flags and fix whitespace alignment for easier reading Mel Gorman
2009-04-21  8:04   ` Pekka Enberg
2009-04-21  8:52     ` Mel Gorman
2009-04-21 15:08       ` Christoph Lameter
2009-04-21 15:24         ` Mel Gorman
2009-04-20 22:20 ` [PATCH 25/25] Use a pre-calculated value instead of num_online_nodes() in fast paths Mel Gorman
2009-04-21  8:08   ` Pekka Enberg
2009-04-21  9:01     ` Mel Gorman
2009-04-21 15:09       ` Christoph Lameter
2009-04-21  8:13 ` [PATCH 00/25] Cleanup and optimise the page allocator V6 Pekka Enberg
2009-04-22 14:13   ` Mel Gorman
  -- strict thread matches above, loose matches on Subject: below --
2009-03-20 10:02 [PATCH 00/25] Cleanup and optimise the page allocator V5 Mel Gorman
2009-03-20 10:02 ` [PATCH 09/25] Calculate the migratetype for allocation only once Mel Gorman
2009-03-20 15:08   ` Christoph Lameter

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20090421083513.GC12713@csn.ul.ie \
    --to=mel@csn.ul.ie \
    --cc=akpm@linux-foundation.org \
    --cc=cl@linux-foundation.org \
    --cc=kosaki.motohiro@jp.fujitsu.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=ming.m.lin@intel.com \
    --cc=npiggin@suse.de \
    --cc=penberg@cs.helsinki.fi \
    --cc=peterz@infradead.org \
    --cc=yanmin_zhang@linux.intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).