From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-10.6 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id A06CCC47082 for ; Thu, 3 Jun 2021 16:02:38 +0000 (UTC) Received: from bombadil.infradead.org (bombadil.infradead.org [198.137.202.133]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 65D27613E7 for ; Thu, 3 Jun 2021 16:02:38 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 65D27613E7 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=arm.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=linux-arm-kernel-bounces+linux-arm-kernel=archiver.kernel.org@lists.infradead.org DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=lists.infradead.org; s=bombadil.20210309; h=Sender: Content-Transfer-Encoding:Content-Type:List-Subscribe:List-Help:List-Post: List-Archive:List-Unsubscribe:List-Id:In-Reply-To:MIME-Version:References: Message-ID:Subject:Cc:To:From:Date:Reply-To:Content-ID:Content-Description: Resent-Date:Resent-From:Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID: List-Owner; bh=udDJIHJNGviYyGPCQiTl5KcYYzkOspRinRdPw2uoo5g=; b=T2D6An0fJcBwHF 9CK8aKV06oGyHeFn3ORFywUU0FYQdct+34nr+wzPNOfTy7aqoLsJ7Eug/sP7pNF7GCPdNr1d5tWyO IPVESu09Wd11UhOgs7V77nI5AXKQY5Hcf5QzJHSi7u7rPY5R5ge14R0wfKI1AaCTjaiXV5ga31WbK iEozseoW5f+i+kPHNiv95V7XRs3zTMlMQZV3g+2NnTFLOjvM/arguctz3dtC1Lctv4RmjAjWq9fJ4 B3lnkS+TMKUyTywRC0pukcOLsoYuwpGme7oI0YJKO1lD0WU3t7WLfPbCwuVGWbAitU3jJk0p8SVUm aj/4a24b7Pm0bAhkCIvg==; Received: from localhost ([::1] helo=bombadil.infradead.org) by bombadil.infradead.org with esmtp (Exim 4.94.2 #2 (Red Hat Linux)) id 1lopm1-009RFq-O6; Thu, 03 Jun 2021 16:00:46 +0000 Received: from mail.kernel.org ([198.145.29.99]) by bombadil.infradead.org with esmtps (Exim 4.94.2 #2 (Red Hat Linux)) id 1loplt-009RCt-8O for linux-arm-kernel@lists.infradead.org; Thu, 03 Jun 2021 16:00:41 +0000 Received: by mail.kernel.org (Postfix) with ESMTPSA id 0F074613DC; Thu, 3 Jun 2021 16:00:33 +0000 (UTC) Date: Thu, 3 Jun 2021 17:00:31 +0100 From: Catalin Marinas To: Steven Price Cc: Marc Zyngier , Will Deacon , James Morse , Julien Thierry , Suzuki K Poulose , kvmarm@lists.cs.columbia.edu, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, Dave Martin , Mark Rutland , Thomas Gleixner , qemu-devel@nongnu.org, Juan Quintela , "Dr. David Alan Gilbert" , Richard Henderson , Peter Maydell , Haibo Xu , Andrew Jones Subject: Re: [PATCH v13 4/8] KVM: arm64: Introduce MTE VM feature Message-ID: <20210603160031.GE20338@arm.com> References: <20210524104513.13258-1-steven.price@arm.com> <20210524104513.13258-5-steven.price@arm.com> MIME-Version: 1.0 Content-Disposition: inline In-Reply-To: <20210524104513.13258-5-steven.price@arm.com> User-Agent: Mutt/1.10.1 (2018-07-13) X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20210603_090037_358610_8BEE12D2 X-CRM114-Status: GOOD ( 29.49 ) X-BeenThere: linux-arm-kernel@lists.infradead.org X-Mailman-Version: 2.1.34 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Sender: "linux-arm-kernel" Errors-To: linux-arm-kernel-bounces+linux-arm-kernel=archiver.kernel.org@lists.infradead.org On Mon, May 24, 2021 at 11:45:09AM +0100, Steven Price wrote: > diff --git a/arch/arm64/kvm/mmu.c b/arch/arm64/kvm/mmu.c > index c5d1f3c87dbd..226035cf7d6c 100644 > --- a/arch/arm64/kvm/mmu.c > +++ b/arch/arm64/kvm/mmu.c > @@ -822,6 +822,42 @@ transparent_hugepage_adjust(struct kvm_memory_slot *memslot, > return PAGE_SIZE; > } > > +static int sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn, > + unsigned long size) > +{ > + if (kvm_has_mte(kvm)) { Nitpick (less indentation): if (!kvm_has_mte(kvm)) return 0; > + /* > + * The page will be mapped in stage 2 as Normal Cacheable, so > + * the VM will be able to see the page's tags and therefore > + * they must be initialised first. If PG_mte_tagged is set, > + * tags have already been initialised. > + * pfn_to_online_page() is used to reject ZONE_DEVICE pages > + * that may not support tags. > + */ > + unsigned long i, nr_pages = size >> PAGE_SHIFT; > + struct page *page = pfn_to_online_page(pfn); > + > + if (!page) > + return -EFAULT; > + > + for (i = 0; i < nr_pages; i++, page++) { > + /* > + * There is a potential (but very unlikely) race > + * between two VMs which are sharing a physical page > + * entering this at the same time. However by splitting > + * the test/set the only risk is tags being overwritten > + * by the mte_clear_page_tags() call. > + */ And I think the real risk here is when the page is writable by at least one of the VMs sharing the page. This excludes KSM, so it only leaves the MAP_SHARED mappings. > + if (!test_bit(PG_mte_tagged, &page->flags)) { > + mte_clear_page_tags(page_address(page)); > + set_bit(PG_mte_tagged, &page->flags); > + } > + } If we want to cover this race (I'd say in a separate patch), we can call mte_sync_page_tags(page, __pte(0), false, true) directly (hopefully I got the arguments right). We can avoid the big lock in most cases if kvm_arch_prepare_memory_region() sets a VM_MTE_RESET (tag clear etc.) and __alloc_zeroed_user_highpage() clears the tags on allocation (as we do for VM_MTE but the new flag would not affect the stage 1 VMM page attributes). > + } > + > + return 0; > +} > + > static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa, > struct kvm_memory_slot *memslot, unsigned long hva, > unsigned long fault_status) > @@ -971,8 +1007,13 @@ static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa, > if (writable) > prot |= KVM_PGTABLE_PROT_W; > > - if (fault_status != FSC_PERM && !device) > + if (fault_status != FSC_PERM && !device) { > + ret = sanitise_mte_tags(kvm, pfn, vma_pagesize); > + if (ret) > + goto out_unlock; Maybe it was discussed in a previous version, why do we need this in addition to kvm_set_spte_gfn()? > + > clean_dcache_guest_page(pfn, vma_pagesize); > + } > > if (exec_fault) { > prot |= KVM_PGTABLE_PROT_X; > @@ -1168,12 +1209,17 @@ bool kvm_unmap_gfn_range(struct kvm *kvm, struct kvm_gfn_range *range) > bool kvm_set_spte_gfn(struct kvm *kvm, struct kvm_gfn_range *range) > { > kvm_pfn_t pfn = pte_pfn(range->pte); > + int ret; > > if (!kvm->arch.mmu.pgt) > return 0; > > WARN_ON(range->end - range->start != 1); > > + ret = sanitise_mte_tags(kvm, pfn, PAGE_SIZE); > + if (ret) > + return false; > + > /* > * We've moved a page around, probably through CoW, so let's treat it > * just like a translation fault and clean the cache to the PoC. Otherwise the patch looks fine. -- Catalin _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel