From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-12.8 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY, SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 3AE4BC4363D for ; Fri, 25 Sep 2020 14:58:12 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id CEDF320715 for ; Fri, 25 Sep 2020 14:58:11 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org CEDF320715 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=intel.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id F0A686B0088; Fri, 25 Sep 2020 10:57:28 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id E94916B0089; Fri, 25 Sep 2020 10:57:28 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id D37176B008A; Fri, 25 Sep 2020 10:57:28 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0229.hostedemail.com [216.40.44.229]) by kanga.kvack.org (Postfix) with ESMTP id AA9AB6B0088 for ; Fri, 25 Sep 2020 10:57:28 -0400 (EDT) Received: from smtpin26.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay05.hostedemail.com (Postfix) with ESMTP id 76A04181AE860 for ; Fri, 25 Sep 2020 14:57:28 +0000 (UTC) X-FDA: 77301887376.26.wind82_0a0d0ce27168 Received: from filter.hostedemail.com (10.5.16.251.rfc1918.com [10.5.16.251]) by smtpin26.hostedemail.com (Postfix) with ESMTP id 56E9F1804B654 for ; Fri, 25 Sep 2020 14:57:28 +0000 (UTC) X-HE-Tag: wind82_0a0d0ce27168 X-Filterd-Recvd-Size: 12381 Received: from mga03.intel.com (mga03.intel.com [134.134.136.65]) by imf40.hostedemail.com (Postfix) with ESMTP for ; Fri, 25 Sep 2020 14:57:27 +0000 (UTC) IronPort-SDR: sjLrdZjYKTeC4R6Gp9ZJyEkzxMX8VGT3rd8wEG2VYMMymBsxzlQ9B+JaYlt0qljfeQxcytiv01 kBULzZVWqdkg== X-IronPort-AV: E=McAfee;i="6000,8403,9755"; a="161631932" X-IronPort-AV: E=Sophos;i="5.77,302,1596524400"; d="scan'208";a="161631932" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga005.jf.intel.com ([10.7.209.41]) by orsmga103.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 25 Sep 2020 07:57:27 -0700 IronPort-SDR: ZATIUGVPkvsDQ3H+ieDr8zSeF5Uj19WioKk/jVh4/FaYS5OcSqVFTs4V5vm+NYiBAP0BP63UHs iYvSNTOT3sOw== X-IronPort-AV: E=Sophos;i="5.77,302,1596524400"; d="scan'208";a="487499218" Received: from yyu32-desk.sc.intel.com ([143.183.136.146]) by orsmga005-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 25 Sep 2020 07:57:26 -0700 From: Yu-cheng Yu To: x86@kernel.org, "H. Peter Anvin" , Thomas Gleixner , Ingo Molnar , linux-kernel@vger.kernel.org, linux-doc@vger.kernel.org, linux-mm@kvack.org, linux-arch@vger.kernel.org, linux-api@vger.kernel.org, Arnd Bergmann , Andy Lutomirski , Balbir Singh , Borislav Petkov , Cyrill Gorcunov , Dave Hansen , Eugene Syromiatnikov , Florian Weimer , "H.J. Lu" , Jann Horn , Jonathan Corbet , Kees Cook , Mike Kravetz , Nadav Amit , Oleg Nesterov , Pavel Machek , Peter Zijlstra , Randy Dunlap , "Ravi V. Shankar" , Vedvyas Shanbhogue , Dave Martin , Weijiang Yang , Pengfei Xu Cc: Yu-cheng Yu , Peter Collingbourne , Andrew Morton Subject: [PATCH v13 19/26] mm: Re-introduce do_mmap_pgoff() Date: Fri, 25 Sep 2020 07:56:42 -0700 Message-Id: <20200925145649.5438-20-yu-cheng.yu@intel.com> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20200925145649.5438-1-yu-cheng.yu@intel.com> References: <20200925145649.5438-1-yu-cheng.yu@intel.com> MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: There was no more caller passing vm_flags to do_mmap(), and vm_flags was removed from the function's input by: commit 45e55300f114 ("mm: remove unnecessary wrapper function do_mmap= _pgoff()"). There is a new user now. Shadow stack allocation passes VM_SHSTK to do_mmap(). Re-introduce the vm_flags and do_mmap_pgoff(). Signed-off-by: Yu-cheng Yu Cc: Peter Collingbourne Cc: Andrew Morton Cc: Oleg Nesterov Cc: linux-mm@kvack.org --- fs/aio.c | 6 +++--- fs/hugetlbfs/inode.c | 2 +- include/linux/fs.h | 2 +- include/linux/mm.h | 12 +++++++++++- ipc/shm.c | 2 +- mm/mmap.c | 16 ++++++++-------- mm/nommu.c | 6 +++--- mm/shmem.c | 2 +- mm/util.c | 4 ++-- 9 files changed, 31 insertions(+), 21 deletions(-) diff --git a/fs/aio.c b/fs/aio.c index d5ec30385566..22d19a4ad586 100644 --- a/fs/aio.c +++ b/fs/aio.c @@ -525,9 +525,9 @@ static int aio_setup_ring(struct kioctx *ctx, unsigne= d int nr_events) return -EINTR; } =20 - ctx->mmap_base =3D do_mmap(ctx->aio_ring_file, 0, ctx->mmap_size, - PROT_READ | PROT_WRITE, - MAP_SHARED, 0, &unused, NULL); + ctx->mmap_base =3D do_mmap_pgoff(ctx->aio_ring_file, 0, ctx->mmap_size, + PROT_READ | PROT_WRITE, + MAP_SHARED, 0, &unused, NULL); mmap_write_unlock(mm); if (IS_ERR((void *)ctx->mmap_base)) { ctx->mmap_size =3D 0; diff --git a/fs/hugetlbfs/inode.c b/fs/hugetlbfs/inode.c index b5c109703daa..f936bcf02cce 100644 --- a/fs/hugetlbfs/inode.c +++ b/fs/hugetlbfs/inode.c @@ -140,7 +140,7 @@ static int hugetlbfs_file_mmap(struct file *file, str= uct vm_area_struct *vma) * already been checked by prepare_hugepage_range. If you add * any error returns here, do so after setting VM_HUGETLB, so * is_vm_hugetlb_page tests below unmap_region go the right - * way when do_mmap unwinds (may be important on powerpc + * way when do_mmap_pgoff unwinds (may be important on powerpc * and ia64). */ vma->vm_flags |=3D VM_HUGETLB | VM_DONTEXPAND; diff --git a/include/linux/fs.h b/include/linux/fs.h index 7519ae003a08..f7df4558f72c 100644 --- a/include/linux/fs.h +++ b/include/linux/fs.h @@ -538,7 +538,7 @@ static inline int mapping_mapped(struct address_space= *mapping) =20 /* * Might pages of this file have been modified in userspace? - * Note that i_mmap_writable counts all VM_SHARED vmas: do_mmap + * Note that i_mmap_writable counts all VM_SHARED vmas: do_mmap_pgoff * marks vma as VM_SHARED if it is shared, and the file was opened for * writing i.e. vma may be mprotected writable even if now readonly. * diff --git a/include/linux/mm.h b/include/linux/mm.h index e09d13699bbe..9b6a0f22cd89 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -2560,13 +2560,23 @@ extern unsigned long mmap_region(struct file *fil= e, unsigned long addr, struct list_head *uf); extern unsigned long do_mmap(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, - unsigned long pgoff, unsigned long *populate, struct list_head *uf); + vm_flags_t vm_flags, unsigned long pgoff, unsigned long *populate, + struct list_head *uf); extern int __do_munmap(struct mm_struct *, unsigned long, size_t, struct list_head *uf, bool downgrade); extern int do_munmap(struct mm_struct *, unsigned long, size_t, struct list_head *uf); extern int do_madvise(unsigned long start, size_t len_in, int behavior); =20 +static inline unsigned long +do_mmap_pgoff(struct file *file, unsigned long addr, + unsigned long len, unsigned long prot, unsigned long flags, + unsigned long pgoff, unsigned long *populate, + struct list_head *uf) +{ + return do_mmap(file, addr, len, prot, flags, 0, pgoff, populate, uf); +} + #ifdef CONFIG_MMU extern int __mm_populate(unsigned long addr, unsigned long len, int ignore_errors); diff --git a/ipc/shm.c b/ipc/shm.c index e25c7c6106bc..3131c1de6bba 100644 --- a/ipc/shm.c +++ b/ipc/shm.c @@ -1556,7 +1556,7 @@ long do_shmat(int shmid, char __user *shmaddr, int = shmflg, goto invalid; } =20 - addr =3D do_mmap(file, addr, size, prot, flags, 0, &populate, NULL); + addr =3D do_mmap_pgoff(file, addr, size, prot, flags, 0, &populate, NUL= L); *raddr =3D addr; err =3D 0; if (IS_ERR_VALUE(addr)) diff --git a/mm/mmap.c b/mm/mmap.c index 574b3f273462..81d4a00092da 100644 --- a/mm/mmap.c +++ b/mm/mmap.c @@ -1030,7 +1030,7 @@ static inline int is_mergeable_anon_vma(struct anon= _vma *anon_vma1, * anon_vmas, nor if same anon_vma is assigned but offsets incompatible. * * We don't check here for the merged mmap wrapping around the end of pa= gecache - * indices (16TB on ia32) because do_mmap() does not permit mmap's which + * indices (16TB on ia32) because do_mmap_pgoff() does not permit mmap's= which * wrap, nor mmaps which cover the final page at index -1UL. */ static int @@ -1365,11 +1365,11 @@ static inline bool file_mmap_ok(struct file *file= , struct inode *inode, */ unsigned long do_mmap(struct file *file, unsigned long addr, unsigned long len, unsigned long prot, - unsigned long flags, unsigned long pgoff, - unsigned long *populate, struct list_head *uf) + unsigned long flags, vm_flags_t vm_flags, + unsigned long pgoff, unsigned long *populate, + struct list_head *uf) { struct mm_struct *mm =3D current->mm; - vm_flags_t vm_flags; int pkey =3D 0; =20 *populate =3D 0; @@ -1431,7 +1431,7 @@ unsigned long do_mmap(struct file *file, unsigned l= ong addr, * to. we assume access permissions have been handled by the open * of the memory object, so we don't do any here. */ - vm_flags =3D calc_vm_prot_bits(prot, pkey) | calc_vm_flag_bits(flags) | + vm_flags |=3D calc_vm_prot_bits(prot, pkey) | calc_vm_flag_bits(flags) = | mm->def_flags | VM_MAYREAD | VM_MAYWRITE | VM_MAYEXEC; =20 if (flags & MAP_LOCKED) @@ -2233,7 +2233,7 @@ get_unmapped_area(struct file *file, unsigned long = addr, unsigned long len, /* * mmap_region() will call shmem_zero_setup() to create a file, * so use shmem's get_unmapped_area in case it can be huge. - * do_mmap() will clear pgoff, so match alignment. + * do_mmap_pgoff() will clear pgoff, so match alignment. */ pgoff =3D 0; get_area =3D shmem_get_unmapped_area; @@ -3006,7 +3006,7 @@ SYSCALL_DEFINE5(remap_file_pages, unsigned long, st= art, unsigned long, size, } =20 file =3D get_file(vma->vm_file); - ret =3D do_mmap(vma->vm_file, start, size, + ret =3D do_mmap_pgoff(vma->vm_file, start, size, prot, flags, pgoff, &populate, NULL); fput(file); out: @@ -3226,7 +3226,7 @@ int insert_vm_struct(struct mm_struct *mm, struct v= m_area_struct *vma) * By setting it to reflect the virtual start address of the * vma, merges and splits can happen in a seamless way, just * using the existing file pgoff checks and manipulations. - * Similarly in do_mmap and in do_brk. + * Similarly in do_mmap_pgoff and in do_brk. */ if (vma_is_anonymous(vma)) { BUG_ON(vma->anon_vma); diff --git a/mm/nommu.c b/mm/nommu.c index 75a327149af1..71a4ea828f06 100644 --- a/mm/nommu.c +++ b/mm/nommu.c @@ -1078,6 +1078,7 @@ unsigned long do_mmap(struct file *file, unsigned long len, unsigned long prot, unsigned long flags, + vm_flags_t vm_flags, unsigned long pgoff, unsigned long *populate, struct list_head *uf) @@ -1085,7 +1086,6 @@ unsigned long do_mmap(struct file *file, struct vm_area_struct *vma; struct vm_region *region; struct rb_node *rb; - vm_flags_t vm_flags; unsigned long capabilities, result; int ret; =20 @@ -1104,7 +1104,7 @@ unsigned long do_mmap(struct file *file, =20 /* we've determined that we can make the mapping, now translate what we * now know into VMA flags */ - vm_flags =3D determine_vm_flags(file, prot, flags, capabilities); + vm_flags |=3D determine_vm_flags(file, prot, flags, capabilities); =20 /* we're going to need to record the mapping */ region =3D kmem_cache_zalloc(vm_region_jar, GFP_KERNEL); @@ -1763,7 +1763,7 @@ EXPORT_SYMBOL_GPL(access_process_vm); * * Check the shared mappings on an inode on behalf of a shrinking trunca= te to * make sure that any outstanding VMAs aren't broken and then shrink the - * vm_regions that extend beyond so that do_mmap() doesn't + * vm_regions that extend beyond so that do_mmap_pgoff() doesn't * automatically grant mappings that are too large. */ int nommu_shrink_inode_mappings(struct inode *inode, size_t size, diff --git a/mm/shmem.c b/mm/shmem.c index 8e2b35ba93ad..54464c1e7414 100644 --- a/mm/shmem.c +++ b/mm/shmem.c @@ -4248,7 +4248,7 @@ EXPORT_SYMBOL_GPL(shmem_file_setup_with_mnt); =20 /** * shmem_zero_setup - setup a shared anonymous mapping - * @vma: the vma to be mmapped is prepared by do_mmap + * @vma: the vma to be mmapped is prepared by do_mmap_pgoff */ int shmem_zero_setup(struct vm_area_struct *vma) { diff --git a/mm/util.c b/mm/util.c index 5ef378a2a038..8d6280c05238 100644 --- a/mm/util.c +++ b/mm/util.c @@ -503,8 +503,8 @@ unsigned long vm_mmap_pgoff(struct file *file, unsign= ed long addr, if (!ret) { if (mmap_write_lock_killable(mm)) return -EINTR; - ret =3D do_mmap(file, addr, len, prot, flag, pgoff, &populate, - &uf); + ret =3D do_mmap_pgoff(file, addr, len, prot, flag, pgoff, + &populate, &uf); mmap_write_unlock(mm); userfaultfd_unmap_complete(mm, &uf); if (populate) --=20 2.21.0