From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-pa0-f50.google.com (mail-pa0-f50.google.com [209.85.220.50]) by kanga.kvack.org (Postfix) with ESMTP id DA22B6B00DF for ; Fri, 25 Oct 2013 11:46:47 -0400 (EDT) Received: by mail-pa0-f50.google.com with SMTP id fb1so3836154pad.23 for ; Fri, 25 Oct 2013 08:46:47 -0700 (PDT) Received: from psmtp.com ([74.125.245.152]) by mx.google.com with SMTP id yj4si5425895pac.282.2013.10.25.08.46.45 for ; Fri, 25 Oct 2013 08:46:47 -0700 (PDT) Received: from /spool/local by e28smtp08.in.ibm.com with IBM ESMTP SMTP Gateway: Authorized Use Only! Violators will be prosecuted for from ; Fri, 25 Oct 2013 21:16:40 +0530 Received: from d28relay02.in.ibm.com (d28relay02.in.ibm.com [9.184.220.59]) by d28dlp03.in.ibm.com (Postfix) with ESMTP id AC0DF125805A for ; Fri, 25 Oct 2013 21:17:12 +0530 (IST) Received: from d28av05.in.ibm.com (d28av05.in.ibm.com [9.184.220.67]) by d28relay02.in.ibm.com (8.13.8/8.13.8/NCO v10.0) with ESMTP id r9PFkS9l41091294 for ; Fri, 25 Oct 2013 21:16:29 +0530 Received: from d28av05.in.ibm.com (localhost [127.0.0.1]) by d28av05.in.ibm.com (8.14.4/8.14.4/NCO v10.0 AVout) with ESMTP id r9PFkZxW022916 for ; Fri, 25 Oct 2013 21:16:35 +0530 From: Robert Jennings Subject: [PATCH v2 1/2] vmsplice: unmap gifted pages for recipient Date: Fri, 25 Oct 2013 10:46:23 -0500 Message-Id: <1382715984-10558-2-git-send-email-rcj@linux.vnet.ibm.com> In-Reply-To: <1382715984-10558-1-git-send-email-rcj@linux.vnet.ibm.com> References: <1382715984-10558-1-git-send-email-rcj@linux.vnet.ibm.com> Sender: owner-linux-mm@kvack.org List-ID: To: linux-kernel@vger.kernel.org Cc: linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, Alexander Viro , Rik van Riel , Andrea Arcangeli , Dave Hansen , Robert Jennings , Matt Helsley , Anthony Liguori , Michael Roth , Lei Li , Leonardo Garcia , Simon Jin , Vlastimil Babka From: Robert C Jennings Introduce use of the unused SPLICE_F_MOVE flag for vmsplice to zap pages. When vmsplice is called with flags (SPLICE_F_GIFT | SPLICE_F_MOVE) the writer's gift'ed pages would be zapped. This patch supports further work to move vmsplice'd pages rather than copying them. That patch has the restriction that the page must not be mapped by the source for the move, otherwise it will fall back to copying the page. Signed-off-by: Matt Helsley Signed-off-by: Robert C Jennings --- Changes since v1: - Cleanup zap coalescing in splice_to_pipe for readability - Field added to struct partial_page in v1 was unnecessary, using private field instead. --- fs/splice.c | 38 ++++++++++++++++++++++++++++++++++++++ 1 file changed, 38 insertions(+) diff --git a/fs/splice.c b/fs/splice.c index 3b7ee65..c14be6f 100644 --- a/fs/splice.c +++ b/fs/splice.c @@ -188,12 +188,18 @@ ssize_t splice_to_pipe(struct pipe_inode_info *pipe, { unsigned int spd_pages = spd->nr_pages; int ret, do_wakeup, page_nr; + struct vm_area_struct *vma; + unsigned long user_start, user_end, addr; ret = 0; do_wakeup = 0; page_nr = 0; + vma = NULL; + user_start = user_end = 0; pipe_lock(pipe); + /* mmap_sem taken for zap_page_range with SPLICE_F_MOVE */ + down_read(¤t->mm->mmap_sem); for (;;) { if (!pipe->readers) { @@ -215,6 +221,33 @@ ssize_t splice_to_pipe(struct pipe_inode_info *pipe, if (spd->flags & SPLICE_F_GIFT) buf->flags |= PIPE_BUF_FLAG_GIFT; + /* Prepare to move page sized/aligned bufs. + * Gather pages for a single zap_page_range() + * call per VMA. + */ + if (spd->flags & (SPLICE_F_GIFT | SPLICE_F_MOVE) && + !buf->offset && + (buf->len == PAGE_SIZE)) { + addr = buf->private; + + if (vma && (addr == user_end) && + (addr + PAGE_SIZE <= vma->vm_end)) { + /* Same vma, no holes */ + user_end += PAGE_SIZE; + } else { + if (vma) + zap_page_range(vma, user_start, + (user_end - user_start), + NULL); + vma = find_vma(current->mm, addr); + if (!IS_ERR_OR_NULL(vma)) { + user_start = addr; + user_end = (addr + PAGE_SIZE); + } else + vma = NULL; + } + } + pipe->nrbufs++; page_nr++; ret += buf->len; @@ -255,6 +288,10 @@ ssize_t splice_to_pipe(struct pipe_inode_info *pipe, pipe->waiting_writers--; } + if (vma) + zap_page_range(vma, user_start, (user_end - user_start), NULL); + + up_read(¤t->mm->mmap_sem); pipe_unlock(pipe); if (do_wakeup) @@ -1475,6 +1512,7 @@ static int get_iovec_page_array(const struct iovec __user *iov, partial[buffers].offset = off; partial[buffers].len = plen; + partial[buffers].private = (unsigned long)base; off = 0; len -= plen; -- 1.8.1.2 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org