From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from psmtp.com (na3sys010amx177.postini.com [74.125.245.177]) by kanga.kvack.org (Postfix) with SMTP id 78B186B008C for ; Wed, 14 Nov 2012 04:19:09 -0500 (EST) Received: by mail-ea0-f169.google.com with SMTP id k11so110593eaa.14 for ; Wed, 14 Nov 2012 01:19:07 -0800 (PST) From: Ingo Molnar Subject: [PATCH 0/3, v2] mprotect() and working set sampling optimizations Date: Wed, 14 Nov 2012 10:18:48 +0100 Message-Id: <1352884731-20024-1-git-send-email-mingo@kernel.org> Sender: owner-linux-mm@kvack.org List-ID: To: linux-kernel@vger.kernel.org, linux-mm@kvack.org Cc: Paul Turner , Lee Schermerhorn , Christoph Lameter , Rik van Riel , Mel Gorman , Andrew Morton , Andrea Arcangeli , Linus Torvalds , Peter Zijlstra , Thomas Gleixner , Hugh Dickins Ok, people suggested to split out the change_protection() modification into a third patch. This series implements an mprotect() optimization that also helps improve the quality of working set scanning: - working set scanning gets faster - we can scan with a touched-page rate, instead of with a virtual-memory proportional rate (within limits). This is already part of numa/core, but wanted to send it out separately as well, to get specific feedback for the mprotect() bits. Thanks, Ingo --- Ingo Molnar (1): mm: Optimize the TLB flush of sys_mprotect() and change_protection() users Peter Zijlstra (2): mm: Count the number of pages affected in change_protection() sched, numa, mm: Count WS scanning against present PTEs, not virtual memory ranges include/linux/hugetlb.h | 8 ++++++-- include/linux/mm.h | 6 +++--- kernel/sched/fair.c | 37 +++++++++++++++++++++---------------- mm/hugetlb.c | 10 ++++++++-- mm/mprotect.c | 46 ++++++++++++++++++++++++++++++++++------------ 5 files changed, 72 insertions(+), 35 deletions(-) -- 1.7.11.7 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org