linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [PATCH][RESEND] mm: Avoid possible deadlock caused by too_many_isolated()
@ 2012-12-10  2:48 Fengguang Wu
  2012-12-13 18:16 ` Rik van Riel
  2012-12-19  7:40 ` Simon Jeons
  0 siblings, 2 replies; 3+ messages in thread
From: Fengguang Wu @ 2012-12-10  2:48 UTC (permalink / raw)
  To: Andrew Morton
  Cc: Linux Memory Management List, Torsten Kaiser, NeilBrown,
	Minchan Kim, KOSAKI Motohiro, Rik van Riel, Li Zefan, wuqixuan,
	zengweilin, shaoyafang

Neil find that if too_many_isolated() returns true while performing
direct reclaim we can end up waiting for other threads to complete their
direct reclaim.  If those threads are allowed to enter the FS or IO to
free memory, but this thread is not, then it is possible that those
threads will be waiting on this thread and so we get a circular
deadlock.

some task enters direct reclaim with GFP_KERNEL
  => too_many_isolated() false
    => vmscan and run into dirty pages
      => pageout()
        => take some FS lock
	  => fs/block code does GFP_NOIO allocation
	    => enter direct reclaim again
	      => too_many_isolated() true
		  => waiting for others to progress, however the other
		     tasks may be circular waiting for the FS lock..

The fix is to let !__GFP_IO and !__GFP_FS direct reclaims enjoy higher
priority than normal ones, by lowering the throttle threshold for the
latter.

Allowing ~1/8 isolated pages in normal is large enough. For example,
for a 1GB LRU list, that's ~128MB isolated pages, or 1k blocked tasks
(each isolates 32 4KB pages), or 64 blocked tasks per logical CPU
(assuming 16 logical CPUs per NUMA node). So it's not likely some CPU
goes idle waiting (when it could make progress) because of this limit:
there are much more sleeping reclaim tasks than the number of CPU, so
the task may well be blocked by some low level queue/lock anyway.

Now !GFP_IOFS reclaims won't be waiting for GFP_IOFS reclaims to
progress. They will be blocked only when there are too many concurrent
!GFP_IOFS reclaims, however that's very unlikely because the IO-less
direct reclaims is able to progress much more faster, and they won't
deadlock each other. The threshold is raised high enough for them, so
that there can be sufficient parallel progress of !GFP_IOFS reclaims.

CC: Torsten Kaiser <just.for.lkml@googlemail.com>
Tested-by: NeilBrown <neilb@suse.de>
Reviewed-by: Minchan Kim <minchan.kim@gmail.com>
Acked-by: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Acked-by: Rik van Riel <riel@redhat.com>
Signed-off-by: Wu Fengguang <fengguang.wu@intel.com>
---
 mm/vmscan.c |    7 +++++++
 1 file changed, 7 insertions(+)

--- linux-next.orig/mm/vmscan.c	2012-12-10 10:43:06.474928860 +0800
+++ linux-next/mm/vmscan.c	2012-12-10 10:43:09.022928920 +0800
@@ -1202,6 +1202,13 @@ static int too_many_isolated(struct zone
 		isolated = zone_page_state(zone, NR_ISOLATED_ANON);
 	}
 
+	/*
+	 * GFP_NOIO/GFP_NOFS callers are allowed to isolate more pages, so that
+	 * they won't get blocked by normal ones and form circular deadlock.
+	 */
+	if ((sc->gfp_mask & GFP_IOFS) == GFP_IOFS)
+		inactive >>= 3;
+
 	return isolated > inactive;
 }
 

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [PATCH][RESEND] mm: Avoid possible deadlock caused by too_many_isolated()
  2012-12-10  2:48 [PATCH][RESEND] mm: Avoid possible deadlock caused by too_many_isolated() Fengguang Wu
@ 2012-12-13 18:16 ` Rik van Riel
  2012-12-19  7:40 ` Simon Jeons
  1 sibling, 0 replies; 3+ messages in thread
From: Rik van Riel @ 2012-12-13 18:16 UTC (permalink / raw)
  To: Fengguang Wu
  Cc: Andrew Morton, Linux Memory Management List, Torsten Kaiser,
	NeilBrown, Minchan Kim, KOSAKI Motohiro, Li Zefan, wuqixuan,
	zengweilin, shaoyafang

On 12/09/2012 09:48 PM, Fengguang Wu wrote:
> Neil find that if too_many_isolated() returns true while performing
> direct reclaim we can end up waiting for other threads to complete their
> direct reclaim.  If those threads are allowed to enter the FS or IO to
> free memory, but this thread is not, then it is possible that those
> threads will be waiting on this thread and so we get a circular
> deadlock.
>
> some task enters direct reclaim with GFP_KERNEL
>    => too_many_isolated() false
>      => vmscan and run into dirty pages
>        => pageout()
>          => take some FS lock
> 	  => fs/block code does GFP_NOIO allocation
> 	    => enter direct reclaim again
> 	      => too_many_isolated() true
> 		  => waiting for others to progress, however the other
> 		     tasks may be circular waiting for the FS lock..
>
> The fix is to let !__GFP_IO and !__GFP_FS direct reclaims enjoy higher
> priority than normal ones, by lowering the throttle threshold for the
> latter.
>
> Allowing ~1/8 isolated pages in normal is large enough. For example,
> for a 1GB LRU list, that's ~128MB isolated pages, or 1k blocked tasks
> (each isolates 32 4KB pages), or 64 blocked tasks per logical CPU
> (assuming 16 logical CPUs per NUMA node). So it's not likely some CPU
> goes idle waiting (when it could make progress) because of this limit:
> there are much more sleeping reclaim tasks than the number of CPU, so
> the task may well be blocked by some low level queue/lock anyway.
>
> Now !GFP_IOFS reclaims won't be waiting for GFP_IOFS reclaims to
> progress. They will be blocked only when there are too many concurrent
> !GFP_IOFS reclaims, however that's very unlikely because the IO-less
> direct reclaims is able to progress much more faster, and they won't
> deadlock each other. The threshold is raised high enough for them, so
> that there can be sufficient parallel progress of !GFP_IOFS reclaims.
>
> CC: Torsten Kaiser <just.for.lkml@googlemail.com>
> Tested-by: NeilBrown <neilb@suse.de>
> Reviewed-by: Minchan Kim <minchan.kim@gmail.com>
> Acked-by: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
> Acked-by: Rik van Riel <riel@redhat.com>
> Signed-off-by: Wu Fengguang <fengguang.wu@intel.com>

Reviewed-by: Rik van Riel <riel@redhat.com>

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: [PATCH][RESEND] mm: Avoid possible deadlock caused by too_many_isolated()
  2012-12-10  2:48 [PATCH][RESEND] mm: Avoid possible deadlock caused by too_many_isolated() Fengguang Wu
  2012-12-13 18:16 ` Rik van Riel
@ 2012-12-19  7:40 ` Simon Jeons
  1 sibling, 0 replies; 3+ messages in thread
From: Simon Jeons @ 2012-12-19  7:40 UTC (permalink / raw)
  To: Fengguang Wu
  Cc: Andrew Morton, Linux Memory Management List, Torsten Kaiser,
	NeilBrown, Minchan Kim, KOSAKI Motohiro, Rik van Riel, Li Zefan,
	wuqixuan, zengweilin, shaoyafang

On Mon, 2012-12-10 at 10:48 +0800, Fengguang Wu wrote: 
> Neil find that if too_many_isolated() returns true while performing
> direct reclaim we can end up waiting for other threads to complete their
> direct reclaim.  If those threads are allowed to enter the FS or IO to
> free memory, but this thread is not, then it is possible that those
> threads will be waiting on this thread and so we get a circular
> deadlock.
> 
> some task enters direct reclaim with GFP_KERNEL
>   => too_many_isolated() false
>     => vmscan and run into dirty pages
>       => pageout()
>         => take some FS lock
> 	  => fs/block code does GFP_NOIO allocation

Hi Fengguang,

GFP_NOIO allocation for what?

> 	    => enter direct reclaim again
> 	      => too_many_isolated() true
> 		  => waiting for others to progress, however the other
> 		     tasks may be circular waiting for the FS lock..
> 
> The fix is to let !__GFP_IO and !__GFP_FS direct reclaims enjoy higher
> priority than normal ones, by lowering the throttle threshold for the
> latter.
> 
> Allowing ~1/8 isolated pages in normal is large enough. For example,
> for a 1GB LRU list, that's ~128MB isolated pages, or 1k blocked tasks
> (each isolates 32 4KB pages), or 64 blocked tasks per logical CPU
> (assuming 16 logical CPUs per NUMA node). So it's not likely some CPU
> goes idle waiting (when it could make progress) because of this limit:
> there are much more sleeping reclaim tasks than the number of CPU, so
> the task may well be blocked by some low level queue/lock anyway.
> 
> Now !GFP_IOFS reclaims won't be waiting for GFP_IOFS reclaims to
> progress. They will be blocked only when there are too many concurrent
> !GFP_IOFS reclaims, however that's very unlikely because the IO-less

Why you said that direct reclaim is IO-less?

> direct reclaims is able to progress much more faster, and they won't
> deadlock each other. The threshold is raised high enough for them, so
> that there can be sufficient parallel progress of !GFP_IOFS reclaims.
> 
> CC: Torsten Kaiser <just.for.lkml@googlemail.com>
> Tested-by: NeilBrown <neilb@suse.de>
> Reviewed-by: Minchan Kim <minchan.kim@gmail.com>
> Acked-by: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
> Acked-by: Rik van Riel <riel@redhat.com>
> Signed-off-by: Wu Fengguang <fengguang.wu@intel.com>
> ---
>  mm/vmscan.c |    7 +++++++
>  1 file changed, 7 insertions(+)
> 
> --- linux-next.orig/mm/vmscan.c	2012-12-10 10:43:06.474928860 +0800
> +++ linux-next/mm/vmscan.c	2012-12-10 10:43:09.022928920 +0800
> @@ -1202,6 +1202,13 @@ static int too_many_isolated(struct zone
>  		isolated = zone_page_state(zone, NR_ISOLATED_ANON);
>  	}
>  
> +	/*
> +	 * GFP_NOIO/GFP_NOFS callers are allowed to isolate more pages, so that
> +	 * they won't get blocked by normal ones and form circular deadlock.
> +	 */
> +	if ((sc->gfp_mask & GFP_IOFS) == GFP_IOFS)
> +		inactive >>= 3;
> +
>  	return isolated > inactive;
>  }
>  
> 
> --
> To unsubscribe, send a message with 'unsubscribe linux-mm' in
> the body to majordomo@kvack.org.  For more info on Linux MM,
> see: http://www.linux-mm.org/ .
> Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>



--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2012-12-19  7:41 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2012-12-10  2:48 [PATCH][RESEND] mm: Avoid possible deadlock caused by too_many_isolated() Fengguang Wu
2012-12-13 18:16 ` Rik van Riel
2012-12-19  7:40 ` Simon Jeons

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).