From mboxrd@z Thu Jan 1 00:00:00 1970 From: Benjamin Herrenschmidt Subject: Re: [RFC PATCH] KVM: PPC: Book3S HV: add support for page faults in VM_IO|VM_PFNMAP vmas Date: Fri, 09 Feb 2018 09:19:13 +1100 Message-ID: <1518128353.2312.217.camel@kernel.crashing.org> References: <20180208160303.19016-1-clg@kaod.org> Mime-Version: 1.0 Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: 8bit To: =?ISO-8859-1?Q?C=E9dric?= Le Goater , kvm-ppc@vger.kernel.org, kvm@vger.kernel.org, Paolo Bonzini , Paul Mackerras Return-path: Received: from gate.crashing.org ([63.228.1.57]:46740 "EHLO gate.crashing.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751094AbeBHWTq (ORCPT ); Thu, 8 Feb 2018 17:19:46 -0500 In-Reply-To: <20180208160303.19016-1-clg@kaod.org> Sender: kvm-owner@vger.kernel.org List-ID: On Thu, 2018-02-08 at 17:03 +0100, Cédric Le Goater wrote: > > +/* > + * Stolen from virt/kvm/kvm_main.c > + */ Just export it. It's annoying that we can't just ues hva_to_pfn() ... > +static int hva_to_pfn_remapped(struct vm_area_struct *vma, > + unsigned long addr, bool write_fault, > + unsigned long *p_pfn) > +{ > + unsigned long pfn; > + int r; > + > + r = follow_pfn(vma, addr, &pfn); > + if (r) { > + /* > + * get_user_pages fails for VM_IO and VM_PFNMAP vmas and does > + * not call the fault handler, so do it here. > + */ > + bool unlocked = false; > + > + r = fixup_user_fault(current, current->mm, addr, > + (write_fault ? FAULT_FLAG_WRITE : 0), > + &unlocked); > + if (unlocked) > + return -EAGAIN; > + if (r) > + return r; > + > + r = follow_pfn(vma, addr, &pfn); > + if (r) > + return r; > + } > + > + /* > + * Get a reference here because callers of *hva_to_pfn* and > + * *gfn_to_pfn* ultimately call kvm_release_pfn_clean on the > + * returned pfn. This is only needed if the VMA has VM_MIXEDMAP > + * set, but the kvm_get_pfn/kvm_release_pfn_clean pair will > + * simply do nothing for reserved pfns. > + * > + * Whoever called remap_pfn_range is also going to call e.g. > + * unmap_mapping_range before the underlying pages are freed, > + * causing a call to our MMU notifier. > + */ > + kvm_get_pfn(pfn); > + > + *p_pfn = pfn; > + return 0; > +} > + > int kvmppc_book3s_radix_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu, > unsigned long ea, unsigned long dsisr) > { > @@ -402,8 +450,15 @@ int kvmppc_book3s_radix_page_fault(struct kvm_run *run, struct kvm_vcpu *vcpu, > vma = find_vma(current->mm, hva); > if (vma && vma->vm_start <= hva && hva < vma->vm_end && > (vma->vm_flags & VM_PFNMAP)) { > - pfn = vma->vm_pgoff + > - ((hva - vma->vm_start) >> PAGE_SHIFT); > + if (vma->vm_flags & (VM_IO | VM_PFNMAP)) { > + ret = hva_to_pfn_remapped(vma, hva, writing, > + &pfn); > + if (ret == -EAGAIN) > + return RESUME_GUEST; > + } else { > + pfn = vma->vm_pgoff + > + ((hva - vma->vm_start) >> PAGE_SHIFT); > + } I don't think the else case is useful. If fact you are checking VM_PFNMAP twice... > pgflags = pgprot_val(vma->vm_page_prot); > } > up_read(¤t->mm->mmap_sem);