From: Xishi Qiu <qiuxishi@huawei.com>
To: Li Zefan <lizefan@huawei.com>, <robin.yb@huawei.com>,
Andrew Morton <akpm@linux-foundation.org>,
Mel Gorman <mgorman@suse.de>, <riel@redhat.com>
Cc: Xishi Qiu <qiuxishi@huawei.com>, <linux-fsdevel@vger.kernel.org>,
Linux MM <linux-mm@kvack.org>,
LKML <linux-kernel@vger.kernel.org>
Subject: [PATCH] mm/fs: don't keep pages when receiving a pending SIGKILL in __get_user_pages()
Date: Wed, 15 Jan 2014 17:31:20 +0800 [thread overview]
Message-ID: <52D65568.6080106@huawei.com> (raw)
In the process IO direction, dio_refill_pages will call get_user_pages_fast
to map the page from user space. If ret is less than 0 and IO is write, the
function will create a zero page to fill data. This may work for some file
system, but in some device operate we prefer whole write or fail, not half
data half zero, e.g. fs metadata, like inode, identy.
This happens often when kill a process which is doing direct IO. Consider
the following cases, the process A is doing IO process, may enter __get_user_pages
function, if other processes send process A SIG_KILL, A will enter the
following branches
/*
* If we have a pending SIGKILL, don't keep faulting
* pages and potentially allocating memory.
*/
if (unlikely(fatal_signal_pending(current)))
return i ? i : -ERESTARTSYS;
Return current pages. direct IO will write the pages, the subsequent pages
which can’t get will use zero page instead.
This patch will modify this judgment, if receive SIG_KILL, release pages and
return an error. Direct IO will find no blocks_available and return error
direct, rather than half IO data and half zero page.
Signed-off-by: Xishi Qiu <qiuxishi@huawei.com>
Signed-off-by: Bin Yang <robin.yb@huawei.com>
---
mm/memory.c | 10 ++++++++--
1 files changed, 8 insertions(+), 2 deletions(-)
diff --git a/mm/memory.c b/mm/memory.c
index 6768ce9..0568faa 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -1799,8 +1799,14 @@ long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
* If we have a pending SIGKILL, don't keep faulting
* pages and potentially allocating memory.
*/
- if (unlikely(fatal_signal_pending(current)))
- return i ? i : -ERESTARTSYS;
+ if (unlikely(fatal_signal_pending(current))) {
+ int j;
+ for (j = 0; j < i; j++) {
+ put_page(pages[j]);
+ pages[j] = NULL;
+ }
+ return -ERESTARTSYS;
+ }
cond_resched();
while (!(page = follow_page_mask(vma, start,
--
1.7.1
--
To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
next reply other threads:[~2014-01-15 9:32 UTC|newest]
Thread overview: 4+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-01-15 9:31 Xishi Qiu [this message]
2014-01-15 23:15 ` [PATCH] mm/fs: don't keep pages when receiving a pending SIGKILL in __get_user_pages() David Rientjes
2014-01-16 12:59 ` Xishi Qiu
2014-01-27 23:59 ` Jan Kara
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=52D65568.6080106@huawei.com \
--to=qiuxishi@huawei.com \
--cc=akpm@linux-foundation.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=lizefan@huawei.com \
--cc=mgorman@suse.de \
--cc=riel@redhat.com \
--cc=robin.yb@huawei.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).