From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 918C0C83F1A for ; Fri, 18 Jul 2025 06:09:48 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 0A0976B00B8; Fri, 18 Jul 2025 02:09:48 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 04A4A6B00BA; Fri, 18 Jul 2025 02:09:47 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id EA2486B00BB; Fri, 18 Jul 2025 02:09:47 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0013.hostedemail.com [216.40.44.13]) by kanga.kvack.org (Postfix) with ESMTP id DD7406B00B8 for ; Fri, 18 Jul 2025 02:09:47 -0400 (EDT) Received: from smtpin18.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay05.hostedemail.com (Postfix) with ESMTP id 760335871B for ; Fri, 18 Jul 2025 06:09:47 +0000 (UTC) X-FDA: 83676359214.18.B9CA5FD Received: from mgamail.intel.com (mgamail.intel.com [198.175.65.18]) by imf08.hostedemail.com (Postfix) with ESMTP id 0233F16000A for ; Fri, 18 Jul 2025 06:09:44 +0000 (UTC) Authentication-Results: imf08.hostedemail.com; dkim=pass header.d=intel.com header.s=Intel header.b=B+NCY+Vi; dmarc=pass (policy=none) header.from=intel.com; spf=pass (imf08.hostedemail.com: domain of xiaoyao.li@intel.com designates 198.175.65.18 as permitted sender) smtp.mailfrom=xiaoyao.li@intel.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1752818985; a=rsa-sha256; cv=none; b=yWbSTN8ZY5Mw6gmH44p5jzBQdaHuBn5Ja8i5UZgP7hxdAP2RlxWgSXQXLNBC0Q8SwAtD77 QxHcWL05VAnzj0IBPgIhYjhHxPl/rsXEEY7SId6A4AHLv+3R46qH7iNHo6OHuip5DvG1v/ UWk/Jkq2k69hCKwMa1iDOGZf5LeJPwA= ARC-Authentication-Results: i=1; imf08.hostedemail.com; dkim=pass header.d=intel.com header.s=Intel header.b=B+NCY+Vi; dmarc=pass (policy=none) header.from=intel.com; spf=pass (imf08.hostedemail.com: domain of xiaoyao.li@intel.com designates 198.175.65.18 as permitted sender) smtp.mailfrom=xiaoyao.li@intel.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1752818985; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=jZbC40i5eciguCH3NuyBsJkc9vhA4MccKWb0o0boyJ0=; b=xtvMpl/ImzrO02XSrnSO6a+z7dJopUk6Dikr80e8fIJy5+po7nzuJ9Q6kAspTAGxfi9sE1 Z5PPl7snPfPk45Wva7cay0/d/9p+wkqdT+Hch1+k5QFBYxZWFvvTf1DcOwqT+sxcQ1bT6i XOQyKIsmHgWh5tPW9Re8K60FiaNwOCA= DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1752818986; x=1784354986; h=message-id:date:mime-version:subject:to:cc:references: from:in-reply-to:content-transfer-encoding; bh=Muu4to5olDa5hJVh7CdT0a9aefhZ6PwtM5Jz5Cbadhk=; b=B+NCY+ViPY3b8dDVg57o3hl/1LUotBzP7K9nJcJr96h1z+enDOB7UiK4 3seyt8JnpKAc0WR74DK4MEHXVl1DOvokjwhLbZvt/GTQ3InDVOegj4xc7 pLgEJO69l9DRbP3b89/4rEd+m0WHlKG7eNSX/D9KpkclvdE8tWfA+2NFB FVhoeTWjN6jzwqieejJ1y5K9iVVPxFU0yBOIB0TCFNmGgC1747RdcTEtz F4V0YuQJjlygxjJMikrVGmLrDarwsCdtN8OCyda4cBHg8GMs2XF1jG3IE dvwdyWBkCMSo0hruDyF/WToTevKCnNlZck+Bfz4xnEpvtKnRa8O/lWmOI g==; X-CSE-ConnectionGUID: DIRHt7zbRsWhJzKF1eoD2w== X-CSE-MsgGUID: 4bQIvwyIQ4GQSfrdIS2tEg== X-IronPort-AV: E=McAfee;i="6800,10657,11495"; a="55231397" X-IronPort-AV: E=Sophos;i="6.16,320,1744095600"; d="scan'208";a="55231397" Received: from orviesa006.jf.intel.com ([10.64.159.146]) by orvoesa110.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Jul 2025 23:09:44 -0700 X-CSE-ConnectionGUID: b3Qe9VFRT1yDBJGJPtpzNw== X-CSE-MsgGUID: xpob/a9xSH2SyoR3DNf2uA== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="6.16,320,1744095600"; d="scan'208";a="157367080" Received: from xiaoyaol-hp-g830.ccr.corp.intel.com (HELO [10.124.247.1]) ([10.124.247.1]) by orviesa006-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Jul 2025 23:09:27 -0700 Message-ID: <9f8bd84d-6386-4e3c-802b-de598b8ac0c2@intel.com> Date: Fri, 18 Jul 2025 14:09:24 +0800 MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v15 13/21] KVM: x86/mmu: Handle guest page faults for guest_memfd with shared memory To: Fuad Tabba , kvm@vger.kernel.org, linux-arm-msm@vger.kernel.org, linux-mm@kvack.org, kvmarm@lists.linux.dev Cc: pbonzini@redhat.com, chenhuacai@kernel.org, mpe@ellerman.id.au, anup@brainfault.org, paul.walmsley@sifive.com, palmer@dabbelt.com, aou@eecs.berkeley.edu, seanjc@google.com, viro@zeniv.linux.org.uk, brauner@kernel.org, willy@infradead.org, akpm@linux-foundation.org, yilun.xu@intel.com, chao.p.peng@linux.intel.com, jarkko@kernel.org, amoorthy@google.com, dmatlack@google.com, isaku.yamahata@intel.com, mic@digikod.net, vbabka@suse.cz, vannapurve@google.com, ackerleytng@google.com, mail@maciej.szmigiero.name, david@redhat.com, michael.roth@amd.com, wei.w.wang@intel.com, liam.merwick@oracle.com, isaku.yamahata@gmail.com, kirill.shutemov@linux.intel.com, suzuki.poulose@arm.com, steven.price@arm.com, quic_eberman@quicinc.com, quic_mnalajal@quicinc.com, quic_tsoni@quicinc.com, quic_svaddagi@quicinc.com, quic_cvanscha@quicinc.com, quic_pderrin@quicinc.com, quic_pheragu@quicinc.com, catalin.marinas@arm.com, james.morse@arm.com, yuzenghui@huawei.com, oliver.upton@linux.dev, maz@kernel.org, will@kernel.org, qperret@google.com, keirf@google.com, roypat@amazon.co.uk, shuah@kernel.org, hch@infradead.org, jgg@nvidia.com, rientjes@google.com, jhubbard@nvidia.com, fvdl@google.com, hughd@google.com, jthoughton@google.com, peterx@redhat.com, pankaj.gupta@amd.com, ira.weiny@intel.com References: <20250717162731.446579-1-tabba@google.com> <20250717162731.446579-14-tabba@google.com> Content-Language: en-US From: Xiaoyao Li In-Reply-To: <20250717162731.446579-14-tabba@google.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Rspamd-Server: rspam12 X-Rspamd-Queue-Id: 0233F16000A X-Stat-Signature: tzup7faoe4gstjgtjbrp68k3sf5p9rnh X-Rspam-User: X-HE-Tag: 1752818984-669228 X-HE-Meta: U2FsdGVkX1+hm5RPzE8NTJXrZLtvzWYMpM2P5DUhR7iEiBYuNhwqBT4Qj2rPi+IzTN2v7nGxikeOnuMFaz0CLvu2J2ulngmKXUdQKy8L5+LLGhU//+AJwIpF1yomsztXon7gn+wYutthsevWKkPXxnbQdNHHJsf06G+jBQ8ToZHHXyaD12WGPVLSKDjaCanYXlN30F/4NSCqp53C7v6iP8OJ06c1kosy2syKw3Bcl+c9nw7vdKWnYAKj2W6rO5Rybe0AbTfBssCoi4JVje0oi0a45M6ZEEpoWgcVgstMx5K90r4vxWK41IIvb/TEBhv7ucFrppSAogaT4/5j1zZ9GcUgN/l2uJOqPRxHjncArvtF5MMTBE5BR6YRxkqyryLa4vTQklODngIQYLZLY68ShHB3QTIta7Q1JDmWKre6ew6DOZLH9P534Uj1gmsNu0KnAITVK/R7AVEyTmE/w9y7lkxOfN34EnG5fYqkqAhrG++22QYoYh9ZyGy8hgaraC3ini+15OXkpKbJ6EQaBi/ZdxPQwnXnD+l/3faJ+jUNwzh0FtK+eNM+uf0uTZRnTFwXMm6SegMaK302lL/2xRZsxp9VAvC2wnun1e5M1QydhhDjnEv9Sy7ktzZfwjq5OrUtXzm/pynqyOAHOCdSA7Hga+p73EwD+cGEbrMl+e5FRUFopNU3scYiWynqb/hH7bx48gvZD0RVQWEnU9hIplitg8/cSKbqmGL7FPywzpIWeyfIdI9O+oHY4yTb/g23ZY2M8+t9YBoCB2zwAYA1YPWty1NVl0aviPxc5/cjctCa3d7NxYE2lmvX7U8WhGs6C5+dDpnEA8fDERLe2O9G5iyo4XuNmD2oD/yZhABlOI1kHyTDlihWTg/yPyroamzNI2efX0f2uQ8k9h+c8GjF6MmKm22LP9hz1LtWNnm5cRkTtwpJX12hyVRsaVooF5MBr/8XmMEVGuvwmVnM5ns/cdp J0Xl8vj2 6rRrXZPbAw1X5sQgA7GS1TpEKPbmgkXRNP3EaOPF4RgaXgocUfm4MeysK5S7cl/TZYOR5mfHR6tenRMV7rwYgxDYcNGSn52YzNaCquFEEVxCW9E8UZujJDmUGhCk0TTOkut55vuSSoR/jBpLk79H975C1drqHUrRhpO1dJJJtUeNHlugC63tP7YRtrLCN8EAsCpb8jkGsXnp7KUp1mNLi5LNxffcVyDEFrBGUvpWDDuskiegF61oT6wfnAYUTt+VzJIU/mzzV92BaScGkHg6+3HeQegbFRHzRI2cNbv6c0lpsUWRDGjssVl1YEE008AGDoP6Hqv/zS9h0Kh/e7WOAIkboNLMWcQJCP75t+/p20CMriwuyF3g3y/cnndOdLdAWnZBr3o/k+C33Yk0= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On 7/18/2025 12:27 AM, Fuad Tabba wrote: > From: Ackerley Tng > > Update the KVM MMU fault handler to service guest page faults > for memory slots backed by guest_memfd with mmap support. For such > slots, the MMU must always fault in pages directly from guest_memfd, > bypassing the host's userspace_addr. > > This ensures that guest_memfd-backed memory is always handled through > the guest_memfd specific faulting path, regardless of whether it's for > private or non-private (shared) use cases. > > Additionally, rename kvm_mmu_faultin_pfn_private() to > kvm_mmu_faultin_pfn_gmem(), as this function is now used to fault in > pages from guest_memfd for both private and non-private memory, > accommodating the new use cases. > > Co-developed-by: David Hildenbrand > Signed-off-by: David Hildenbrand > Signed-off-by: Ackerley Tng > Co-developed-by: Fuad Tabba > Signed-off-by: Fuad Tabba Reviewed-by: Xiaoyao Li Note to myself: After this patch, it looks possible that kvm_mmu_prepare_memory_fault_exit() in kvm_mmu_faultin_pfn_gmem() might be triggered for guest_memfd with mmap support, though I'm not sure if there is real case to trigger it. This requires some change in QEMU when it adds support for guest_memfd mmap support, since current QEMU handles KVM_EXIT_MEMORY_FAULT by always converting the memory attribute. > --- > arch/x86/kvm/mmu/mmu.c | 13 +++++++++---- > 1 file changed, 9 insertions(+), 4 deletions(-) > > diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c > index 94be15cde6da..ad5f337b496c 100644 > --- a/arch/x86/kvm/mmu/mmu.c > +++ b/arch/x86/kvm/mmu/mmu.c > @@ -4511,8 +4511,8 @@ static void kvm_mmu_finish_page_fault(struct kvm_vcpu *vcpu, > r == RET_PF_RETRY, fault->map_writable); > } > > -static int kvm_mmu_faultin_pfn_private(struct kvm_vcpu *vcpu, > - struct kvm_page_fault *fault) > +static int kvm_mmu_faultin_pfn_gmem(struct kvm_vcpu *vcpu, > + struct kvm_page_fault *fault) > { > int max_order, r; > > @@ -4536,13 +4536,18 @@ static int kvm_mmu_faultin_pfn_private(struct kvm_vcpu *vcpu, > return RET_PF_CONTINUE; > } > > +static bool fault_from_gmem(struct kvm_page_fault *fault) > +{ > + return fault->is_private || kvm_memslot_is_gmem_only(fault->slot); > +} > + > static int __kvm_mmu_faultin_pfn(struct kvm_vcpu *vcpu, > struct kvm_page_fault *fault) > { > unsigned int foll = fault->write ? FOLL_WRITE : 0; > > - if (fault->is_private) > - return kvm_mmu_faultin_pfn_private(vcpu, fault); > + if (fault_from_gmem(fault)) > + return kvm_mmu_faultin_pfn_gmem(vcpu, fault); > > foll |= FOLL_NOWAIT; > fault->pfn = __kvm_faultin_pfn(fault->slot, fault->gfn, foll,