From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from psmtp.com (na3sys010amx205.postini.com [74.125.245.205]) by kanga.kvack.org (Postfix) with SMTP id 5E5FA6B005D for ; Mon, 4 Jun 2012 16:22:27 -0400 (EDT) Received: by qabg27 with SMTP id g27so1989695qab.14 for ; Mon, 04 Jun 2012 13:22:26 -0700 (PDT) Message-ID: <4FCD18FD.5030307@gmail.com> Date: Mon, 04 Jun 2012 16:22:21 -0400 From: KOSAKI Motohiro MIME-Version: 1.0 Subject: Re: [PATCH v9] mm: compaction: handle incorrect MIGRATE_UNMOVABLE type pageblocks References: <201206041543.56917.b.zolnierkie@samsung.com> In-Reply-To: <201206041543.56917.b.zolnierkie@samsung.com> Content-Type: text/plain; charset=ISO-8859-1; format=flowed Content-Transfer-Encoding: 7bit Sender: owner-linux-mm@kvack.org List-ID: To: Bartlomiej Zolnierkiewicz Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, Minchan Kim , Hugh Dickins , Linus Torvalds , Kyungmin Park , Marek Szyprowski , Mel Gorman , Rik van Riel , Dave Jones , Andrew Morton , Cong Wang , Markus Trippelsdorf , kosaki.motohiro@gmail.com > +/* > + * Returns true if MIGRATE_UNMOVABLE pageblock can be successfully > + * converted to MIGRATE_MOVABLE type, false otherwise. > + */ > +static bool can_rescue_unmovable_pageblock(struct page *page, bool locked) > +{ > + unsigned long pfn, start_pfn, end_pfn; > + struct page *start_page, *end_page, *cursor_page; > + > + pfn = page_to_pfn(page); > + start_pfn = pfn& ~(pageblock_nr_pages - 1); > + end_pfn = start_pfn + pageblock_nr_pages - 1; > + > + start_page = pfn_to_page(start_pfn); > + end_page = pfn_to_page(end_pfn); > + > + for (cursor_page = start_page, pfn = start_pfn; cursor_page<= end_page; > + pfn++, cursor_page++) { > + struct zone *zone = page_zone(start_page); > + unsigned long flags; > + > + if (!pfn_valid_within(pfn)) > + continue; > + > + /* Do not deal with pageblocks that overlap zones */ > + if (page_zone(cursor_page) != zone) > + return false; > + > + if (!locked) > + spin_lock_irqsave(&zone->lock, flags); > + > + if (PageBuddy(cursor_page)) { > + int order = page_order(cursor_page); > > -/* Returns true if the page is within a block suitable for migration to */ > -static bool suitable_migration_target(struct page *page) > + pfn += (1<< order) - 1; > + cursor_page += (1<< order) - 1; > + > + if (!locked) > + spin_unlock_irqrestore(&zone->lock, flags); > + continue; > + } else if (page_count(cursor_page) == 0 || > + PageLRU(cursor_page)) { > + if (!locked) > + spin_unlock_irqrestore(&zone->lock, flags); > + continue; > + } > + > + if (!locked) > + spin_unlock_irqrestore(&zone->lock, flags); > + > + return false; > + } > + > + return true; > +} Minchan, are you interest this patch? If yes, can you please rewrite it? This one are not fixed our pointed issue and can_rescue_unmovable_pageblock() still has plenty bugs. We can't ack it. -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: email@kvack.org