From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-6.9 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_PASS,URIBL_BLOCKED autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 6186EC04A6B for ; Fri, 10 May 2019 05:42:22 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id DCCAC2054F for ; Fri, 10 May 2019 05:42:21 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org DCCAC2054F Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=redhat.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 483F56B0007; Fri, 10 May 2019 01:42:21 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 432796B0008; Fri, 10 May 2019 01:42:21 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 349046B000A; Fri, 10 May 2019 01:42:21 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from mail-qt1-f199.google.com (mail-qt1-f199.google.com [209.85.160.199]) by kanga.kvack.org (Postfix) with ESMTP id 138246B0007 for ; Fri, 10 May 2019 01:42:21 -0400 (EDT) Received: by mail-qt1-f199.google.com with SMTP id i5so5126804qtd.17 for ; Thu, 09 May 2019 22:42:21 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-original-authentication-results:x-gm-message-state:date:from:to :cc:message-id:in-reply-to:references:subject:mime-version :content-transfer-encoding:thread-topic:thread-index; bh=cqMgMHc7e7408FEFx2B3a93a71h+ttTQ0QpTSRUi0Gg=; b=Vh8JLEcxEV7HioMpUKlnUDFLKXrXBcuuucJEWeArkJctlI1ym7yj7YqdqqiOns32sw kAniT8Xgvr6RvBDK8ww+2RzNpehUZQlC23aQTlal18IFIiL2QdZHxJ3K9ZElXxQvENyu zSoELxuwpB9QOliw6EZmrmM5v3Pug0sXDAH5ZLjiBVBRyverPLRAtGExmCFZQr/hlBbi wVXZtNvXFymRhfgpqfi6LpulETp9U5ED711KkdoQR/oMutu/JMFHObOJPtYd1FAabHNz 27MGGthcKGgI3fjPFDuMAOG7y+1n4iF1DEpLgQDN+8JO0A4gKe1klso67gW0Alj6aGVM ot4A== X-Original-Authentication-Results: mx.google.com; spf=pass (google.com: domain of pagupta@redhat.com designates 209.132.183.28 as permitted sender) smtp.mailfrom=pagupta@redhat.com; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com X-Gm-Message-State: APjAAAXE9iZy+7jg4ugzaGZ/LyxQ13hGtIH6pGHW12PH5DQq0uYhWC4Z a8Kq5G0U89t8JiIWyfd0DLTaT8u8u1C15umt7ot820NXos30Ol3xEA+8wAqf5bZmuOV2JNmJvL+ LX7MKolwv27seDJvkdjl2KEK97a84v9XYL+Nbn2+uM8rVim0yHhrubw2aML8io99jhA== X-Received: by 2002:a0c:b993:: with SMTP id v19mr7298036qvf.58.1557466940774; Thu, 09 May 2019 22:42:20 -0700 (PDT) X-Google-Smtp-Source: APXvYqxdJ/PjzvRz+BN5e7kaBuilcli3eMzp51UAHNsAWnBvhA/lAX99OBm/OEEe+tSbXbG3PCCL X-Received: by 2002:a0c:b993:: with SMTP id v19mr7297976qvf.58.1557466939563; Thu, 09 May 2019 22:42:19 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1557466939; cv=none; d=google.com; s=arc-20160816; b=oUjvY8m8Z0N3wf94J4T2PQrDoJdOngbLZ/nMBcsjlo0W13ECrflDs2prU/hwzPc+Ge 14WRaChLK+0ISKNONADWGparRYJvFQvG0NFZ1LapOzPVW2X2bmirS96LEaFz/9w0HPHx aoRnwHE+v4r/pqAjfFD5wBgzcIAKltfD5cudLT/NQcrJWKljmbITwXmQDLNyl6jBoIsf ojo2QaMQgKnDZRrRQ5b+g5fxiIAXqLRdQTJk10WwHPykIT7X7IDMy4I0CuzTm+lMhAox GGgxY0BNG5xSB1D0ebEsCATOqwLuAw+UsvjgkOmA0kBch1j/ZdrWcWq1qPqHYTpGNyIk jg3Q== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=thread-index:thread-topic:content-transfer-encoding:mime-version :subject:references:in-reply-to:message-id:cc:to:from:date; bh=cqMgMHc7e7408FEFx2B3a93a71h+ttTQ0QpTSRUi0Gg=; b=z1zonc1eTLqip4xu6OMNjPaamuOp4bBLBaupv47EF74DqBvDDX43YmRN3G2Csdtwo6 f9OfP5BFlcbZ26/caC+AkIkAqzliqM1UTHY6ZSHiE7VP4VjHkMg16GxbQ08IyV/mjcfT RPeus9S16QTyav/Qmzc44xrGGqXSIA4Rsm5WFLNk4eas8Gr5Z1UHQXuGc7Nn7pHkRl4y 7TceA6VPDXULace2Gs2AODAa6uoMQu/+dBCgrShWICQRAK6ssIVLTeiRdsP91QX0n5Lg RgOTiqDK9xf/nk/eEovRMLu2GZa7rdNmwpvC/043eP4Um2Bf8KTWcruO86c4snvfaTdj ZAwA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of pagupta@redhat.com designates 209.132.183.28 as permitted sender) smtp.mailfrom=pagupta@redhat.com; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: from mx1.redhat.com (mx1.redhat.com. [209.132.183.28]) by mx.google.com with ESMTPS id g10si3040505qvo.35.2019.05.09.22.42.19 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Thu, 09 May 2019 22:42:19 -0700 (PDT) Received-SPF: pass (google.com: domain of pagupta@redhat.com designates 209.132.183.28 as permitted sender) client-ip=209.132.183.28; Authentication-Results: mx.google.com; spf=pass (google.com: domain of pagupta@redhat.com designates 209.132.183.28 as permitted sender) smtp.mailfrom=pagupta@redhat.com; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: from smtp.corp.redhat.com (int-mx07.intmail.prod.int.phx2.redhat.com [10.5.11.22]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 66DA9C057F47; Fri, 10 May 2019 05:42:18 +0000 (UTC) Received: from colo-mx.corp.redhat.com (colo-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.21]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 450421001E65; Fri, 10 May 2019 05:42:18 +0000 (UTC) Received: from zmail21.collab.prod.int.phx2.redhat.com (zmail21.collab.prod.int.phx2.redhat.com [10.5.83.24]) by colo-mx.corp.redhat.com (Postfix) with ESMTP id F35A341F56; Fri, 10 May 2019 05:42:17 +0000 (UTC) Date: Fri, 10 May 2019 01:42:17 -0400 (EDT) From: Pankaj Gupta To: Dan Williams Cc: akpm@linux-foundation.org, stable@vger.kernel.org, Piotr Balcer , Yan Ma , "Aneesh Kumar K.V" , Chandan Rajendra , Jan Kara , Matthew Wilcox , Souptick Joarder , linux-mm@kvack.org, linux-kernel@vger.kernel.org, linux-fsdevel@vger.kernel.org Message-ID: <1750394500.27863545.1557466937334.JavaMail.zimbra@redhat.com> In-Reply-To: <155741946350.372037.11148198430068238140.stgit@dwillia2-desk3.amr.corp.intel.com> References: <155741946350.372037.11148198430068238140.stgit@dwillia2-desk3.amr.corp.intel.com> Subject: Re: [PATCH] mm/huge_memory: Fix vmf_insert_pfn_{pmd, pud}() crash, handle unaligned addresses MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-Originating-IP: [10.65.16.148, 10.4.195.13] Thread-Topic: mm/huge_memory: Fix vmf_insert_pfn_{pmd, pud}() crash, handle unaligned addresses Thread-Index: gtFotDOw5RjI2gXScL+S1YykooYblA== X-Scanned-By: MIMEDefang 2.84 on 10.5.11.22 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.32]); Fri, 10 May 2019 05:42:18 +0000 (UTC) X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: > > Starting with commit c6f3c5ee40c1 "mm/huge_memory.c: fix modifying of > page protection by insert_pfn_pmd()" vmf_insert_pfn_pmd() internally > calls pmdp_set_access_flags(). That helper enforces a pmd aligned > @address argument via VM_BUG_ON() assertion. > > Update the implementation to take a 'struct vm_fault' argument directly > and apply the address alignment fixup internally to fix crash signatures > like: > > kernel BUG at arch/x86/mm/pgtable.c:515! > invalid opcode: 0000 [#1] SMP NOPTI > CPU: 51 PID: 43713 Comm: java Tainted: G OE 4.19.35 #1 > [..] > RIP: 0010:pmdp_set_access_flags+0x48/0x50 > [..] > Call Trace: > vmf_insert_pfn_pmd+0x198/0x350 > dax_iomap_fault+0xe82/0x1190 > ext4_dax_huge_fault+0x103/0x1f0 > ? __switch_to_asm+0x40/0x70 > __handle_mm_fault+0x3f6/0x1370 > ? __switch_to_asm+0x34/0x70 > ? __switch_to_asm+0x40/0x70 > handle_mm_fault+0xda/0x200 > __do_page_fault+0x249/0x4f0 > do_page_fault+0x32/0x110 > ? page_fault+0x8/0x30 > page_fault+0x1e/0x30 > > Cc: vger.kernel.org> > Fixes: c6f3c5ee40c1 ("mm/huge_memory.c: fix modifying of page protection by > insert_pfn_pmd()") > Reported-by: Piotr Balcer > Tested-by: Yan Ma > Cc: Aneesh Kumar K.V > Cc: Chandan Rajendra > Cc: Jan Kara suse.cz> > Cc: Andrew Morton > Cc: Matthew Wilcox infradead.org> > Cc: Souptick Joarder > Signed-off-by: Dan Williams > --- > > drivers/dax/device.c | 6 ++---- > fs/dax.c | 6 ++---- > include/linux/huge_mm.h | 6 ++---- > mm/huge_memory.c | 16 ++++++++++------ > 4 files changed, 16 insertions(+), 18 deletions(-) > > diff --git a/drivers/dax/device.c b/drivers/dax/device.c > index e428468ab661..996d68ff992a 100644 > --- a/drivers/dax/device.c > +++ b/drivers/dax/device.c > @@ -184,8 +184,7 @@ static vm_fault_t __dev_dax_pmd_fault(struct dev_dax > *dev_dax, > > *pfn = phys_to_pfn_t(phys, dax_region->pfn_flags); > > - return vmf_insert_pfn_pmd(vmf->vma, vmf->address, vmf->pmd, *pfn, > - vmf->flags & FAULT_FLAG_WRITE); > + return vmf_insert_pfn_pmd(vmf, *pfn, vmf->flags & FAULT_FLAG_WRITE); > } > > #ifdef CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD > @@ -235,8 +234,7 @@ static vm_fault_t __dev_dax_pud_fault(struct dev_dax > *dev_dax, > > *pfn = phys_to_pfn_t(phys, dax_region->pfn_flags); > > - return vmf_insert_pfn_pud(vmf->vma, vmf->address, vmf->pud, *pfn, > - vmf->flags & FAULT_FLAG_WRITE); > + return vmf_insert_pfn_pud(vmf, *pfn, vmf->flags & FAULT_FLAG_WRITE); > } > #else > static vm_fault_t __dev_dax_pud_fault(struct dev_dax *dev_dax, > diff --git a/fs/dax.c b/fs/dax.c > index e5e54da1715f..83009875308c 100644 > --- a/fs/dax.c > +++ b/fs/dax.c > @@ -1575,8 +1575,7 @@ static vm_fault_t dax_iomap_pmd_fault(struct vm_fault > *vmf, pfn_t *pfnp, > } > > trace_dax_pmd_insert_mapping(inode, vmf, PMD_SIZE, pfn, entry); > - result = vmf_insert_pfn_pmd(vma, vmf->address, vmf->pmd, pfn, > - write); > + result = vmf_insert_pfn_pmd(vmf, pfn, write); > break; > case IOMAP_UNWRITTEN: > case IOMAP_HOLE: > @@ -1686,8 +1685,7 @@ dax_insert_pfn_mkwrite(struct vm_fault *vmf, pfn_t pfn, > unsigned int order) > ret = vmf_insert_mixed_mkwrite(vmf->vma, vmf->address, pfn); > #ifdef CONFIG_FS_DAX_PMD > else if (order == PMD_ORDER) > - ret = vmf_insert_pfn_pmd(vmf->vma, vmf->address, vmf->pmd, > - pfn, true); > + ret = vmf_insert_pfn_pmd(vmf, pfn, FAULT_FLAG_WRITE); > #endif > else > ret = VM_FAULT_FALLBACK; > diff --git a/include/linux/huge_mm.h b/include/linux/huge_mm.h > index 381e872bfde0..7cd5c150c21d 100644 > --- a/include/linux/huge_mm.h > +++ b/include/linux/huge_mm.h > @@ -47,10 +47,8 @@ extern bool move_huge_pmd(struct vm_area_struct *vma, > unsigned long old_addr, > extern int change_huge_pmd(struct vm_area_struct *vma, pmd_t *pmd, > unsigned long addr, pgprot_t newprot, > int prot_numa); > -vm_fault_t vmf_insert_pfn_pmd(struct vm_area_struct *vma, unsigned long > addr, > - pmd_t *pmd, pfn_t pfn, bool write); > -vm_fault_t vmf_insert_pfn_pud(struct vm_area_struct *vma, unsigned long > addr, > - pud_t *pud, pfn_t pfn, bool write); > +vm_fault_t vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write); > +vm_fault_t vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write); > enum transparent_hugepage_flag { > TRANSPARENT_HUGEPAGE_FLAG, > TRANSPARENT_HUGEPAGE_REQ_MADV_FLAG, > diff --git a/mm/huge_memory.c b/mm/huge_memory.c > index 165ea46bf149..4310c6e9e5a3 100644 > --- a/mm/huge_memory.c > +++ b/mm/huge_memory.c > @@ -793,11 +793,13 @@ static void insert_pfn_pmd(struct vm_area_struct *vma, > unsigned long addr, > pte_free(mm, pgtable); > } > > -vm_fault_t vmf_insert_pfn_pmd(struct vm_area_struct *vma, unsigned long > addr, > - pmd_t *pmd, pfn_t pfn, bool write) > +vm_fault_t vmf_insert_pfn_pmd(struct vm_fault *vmf, pfn_t pfn, bool write) > { > + unsigned long addr = vmf->address & PMD_MASK; > + struct vm_area_struct *vma = vmf->vma; > pgprot_t pgprot = vma->vm_page_prot; > pgtable_t pgtable = NULL; > + > /* > * If we had pmd_special, we could avoid all these restrictions, > * but we need to be consistent with PTEs and architectures that > @@ -820,7 +822,7 @@ vm_fault_t vmf_insert_pfn_pmd(struct vm_area_struct *vma, > unsigned long addr, > > track_pfn_insert(vma, &pgprot, pfn); > > - insert_pfn_pmd(vma, addr, pmd, pfn, pgprot, write, pgtable); > + insert_pfn_pmd(vma, addr, vmf->pmd, pfn, pgprot, write, pgtable); > return VM_FAULT_NOPAGE; > } > EXPORT_SYMBOL_GPL(vmf_insert_pfn_pmd); > @@ -869,10 +871,12 @@ static void insert_pfn_pud(struct vm_area_struct *vma, > unsigned long addr, > spin_unlock(ptl); > } > > -vm_fault_t vmf_insert_pfn_pud(struct vm_area_struct *vma, unsigned long > addr, > - pud_t *pud, pfn_t pfn, bool write) > +vm_fault_t vmf_insert_pfn_pud(struct vm_fault *vmf, pfn_t pfn, bool write) > { > + unsigned long addr = vmf->address & PUD_MASK; > + struct vm_area_struct *vma = vmf->vma; > pgprot_t pgprot = vma->vm_page_prot; > + > /* > * If we had pud_special, we could avoid all these restrictions, > * but we need to be consistent with PTEs and architectures that > @@ -889,7 +893,7 @@ vm_fault_t vmf_insert_pfn_pud(struct vm_area_struct *vma, > unsigned long addr, > > track_pfn_insert(vma, &pgprot, pfn); > > - insert_pfn_pud(vma, addr, pud, pfn, pgprot, write); > + insert_pfn_pud(vma, addr, vmf->pud, pfn, pgprot, write); > return VM_FAULT_NOPAGE; > } > EXPORT_SYMBOL_GPL(vmf_insert_pfn_pud); > Thanks for the patch. This patch solves the issue faced while testing virtio pmem. Tested-by: Pankaj Gupta >