From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 4A2F9C433F5 for ; Fri, 1 Apr 2022 14:21:21 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S245165AbiDAOXI (ORCPT ); Fri, 1 Apr 2022 10:23:08 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:47210 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S240438AbiDAOWw (ORCPT ); Fri, 1 Apr 2022 10:22:52 -0400 Received: from mx0a-001b2d01.pphosted.com (mx0b-001b2d01.pphosted.com [148.163.158.5]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id E27C24D258; Fri, 1 Apr 2022 07:21:02 -0700 (PDT) Received: from pps.filterd (m0098419.ppops.net [127.0.0.1]) by mx0b-001b2d01.pphosted.com (8.16.1.2/8.16.1.2) with SMTP id 231EDJKB010389; Fri, 1 Apr 2022 14:21:02 GMT DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ibm.com; h=date : from : to : cc : subject : message-id : in-reply-to : references : mime-version : content-type : content-transfer-encoding; s=pp1; bh=QVk9Cb8n2dvKyFWkRBjUNBaq68qJiGxkXzSUBAqMpmQ=; b=dMFqXGoLiAfMcC/ybNHl/MWnKeZMjz6QtJDBU5sUSGPrs0hr/SFB56M5/RLrOzfhYFKK +YiCzhAwZFazTxirMDU5q3l15NJqnQBQie/IdnP+rx3TYAmoOwM5sM3Xm17RVGKSu/TJ kj+tv9GpwljMtkOe0FNakbL9juN4MEwFPzsEh9O0a8bCH4Qc+1chMWdhZv0zSmwO33EL hf8PSDJoDCcJB6wYgUbMB6XBBoCsrMKeueSotWIvgw4SfzRyoILUo9OZChlaVD7Y4hu1 CriYwBtSCX7e81SC09Xao4Z6DefZvsBn90W4ITzRb7haJ8MnKMwKO/zg7nTUQAy4zpkF hQ== Received: from pps.reinject (localhost [127.0.0.1]) by mx0b-001b2d01.pphosted.com with ESMTP id 3f62y905eb-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Fri, 01 Apr 2022 14:21:01 +0000 Received: from m0098419.ppops.net (m0098419.ppops.net [127.0.0.1]) by pps.reinject (8.16.0.43/8.16.0.43) with SMTP id 231EL1Oh004070; Fri, 1 Apr 2022 14:21:01 GMT Received: from ppma01fra.de.ibm.com (46.49.7a9f.ip4.static.sl-reverse.com [159.122.73.70]) by mx0b-001b2d01.pphosted.com with ESMTP id 3f62y905de-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Fri, 01 Apr 2022 14:21:01 +0000 Received: from pps.filterd (ppma01fra.de.ibm.com [127.0.0.1]) by ppma01fra.de.ibm.com (8.16.1.2/8.16.1.2) with SMTP id 231EF8jx012649; Fri, 1 Apr 2022 14:20:59 GMT Received: from b06cxnps4075.portsmouth.uk.ibm.com (d06relay12.portsmouth.uk.ibm.com [9.149.109.197]) by ppma01fra.de.ibm.com with ESMTP id 3f1tf8u5wu-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Fri, 01 Apr 2022 14:20:59 +0000 Received: from d06av23.portsmouth.uk.ibm.com (d06av23.portsmouth.uk.ibm.com [9.149.105.59]) by b06cxnps4075.portsmouth.uk.ibm.com (8.14.9/8.14.9/NCO v10.0) with ESMTP id 231EKuBI45351230 (version=TLSv1/SSLv3 cipher=DHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Fri, 1 Apr 2022 14:20:56 GMT Received: from d06av23.portsmouth.uk.ibm.com (unknown [127.0.0.1]) by IMSVA (Postfix) with ESMTP id 15CE5A404D; Fri, 1 Apr 2022 14:20:56 +0000 (GMT) Received: from d06av23.portsmouth.uk.ibm.com (unknown [127.0.0.1]) by IMSVA (Postfix) with ESMTP id 88F79A4051; Fri, 1 Apr 2022 14:20:55 +0000 (GMT) Received: from p-imbrenda (unknown [9.145.3.73]) by d06av23.portsmouth.uk.ibm.com (Postfix) with ESMTP; Fri, 1 Apr 2022 14:20:55 +0000 (GMT) Date: Fri, 1 Apr 2022 16:20:53 +0200 From: Claudio Imbrenda To: Janosch Frank Cc: kvm@vger.kernel.org, borntraeger@de.ibm.com, thuth@redhat.com, pasic@linux.ibm.com, david@redhat.com, linux-s390@vger.kernel.org, linux-kernel@vger.kernel.org, scgl@linux.ibm.com, mimu@linux.ibm.com, nrb@linux.ibm.com Subject: Re: [PATCH v9 01/18] KVM: s390: pv: leak the topmost page table when destroy fails Message-ID: <20220401162053.5466832d@p-imbrenda> In-Reply-To: <6b24e1f6-22ee-c0e4-5bde-9eefdccd3619@linux.ibm.com> References: <20220330122605.247613-1-imbrenda@linux.ibm.com> <20220330122605.247613-2-imbrenda@linux.ibm.com> <6b24e1f6-22ee-c0e4-5bde-9eefdccd3619@linux.ibm.com> Organization: IBM X-Mailer: Claws Mail 3.18.0 (GTK+ 2.24.33; x86_64-redhat-linux-gnu) MIME-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit X-TM-AS-GCONF: 00 X-Proofpoint-GUID: t7SywXXLLLJ8farJ8moqqH2j8zY-ynDU X-Proofpoint-ORIG-GUID: JmsvE-BSa3Cbf4nzg0YR7REfOffvvW22 X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.205,Aquarius:18.0.850,Hydra:6.0.425,FMLib:17.11.64.514 definitions=2022-04-01_05,2022-03-31_01,2022-02-23_01 X-Proofpoint-Spam-Details: rule=outbound_notspam policy=outbound score=0 malwarescore=0 spamscore=0 impostorscore=0 mlxscore=0 adultscore=0 bulkscore=0 mlxlogscore=999 priorityscore=1501 clxscore=1015 phishscore=0 suspectscore=0 lowpriorityscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2202240000 definitions=main-2204010066 Precedence: bulk List-ID: X-Mailing-List: kvm@vger.kernel.org On Thu, 31 Mar 2022 15:13:41 +0200 Janosch Frank wrote: > On 3/30/22 14:25, Claudio Imbrenda wrote: > > Each secure guest must have a unique ASCE (address space control > > element); we must avoid that new guests use the same page for their > > ASCE, to avoid errors. > > > > Since the ASCE mostly consists of the address of the topmost page table > > (plus some flags), we must not return that memory to the pool unless > > the ASCE is no longer in use. > > > > Only a successful Destroy Secure Configuration UVC will make the ASCE > > reusable again. > > > > If the Destroy Configuration UVC fails, the ASCE cannot be reused for a > > secure guest (either for the ASCE or for other memory areas). To avoid > > a collision, it must not be used again. This is a permanent error and > > the page becomes in practice unusable, so we set it aside and leak it. > > On failure we already leak other memory that belongs to the ultravisor > > (i.e. the variable and base storage for a guest) and not leaking the > > topmost page table was an oversight. > > > > This error (and thus the leakage) should not happen unless the hardware > > is broken or KVM has some unknown serious bug. > > > > Signed-off-by: Claudio Imbrenda > > Fixes: 29b40f105ec8d55 ("KVM: s390: protvirt: Add initial vm and cpu lifecycle handling") > > --- > > arch/s390/include/asm/gmap.h | 2 + > > arch/s390/kvm/pv.c | 9 +++-- > > arch/s390/mm/gmap.c | 71 ++++++++++++++++++++++++++++++++++++ > > 3 files changed, 79 insertions(+), 3 deletions(-) > > > > diff --git a/arch/s390/include/asm/gmap.h b/arch/s390/include/asm/gmap.h > > index 40264f60b0da..746e18bf8984 100644 > > --- a/arch/s390/include/asm/gmap.h > > +++ b/arch/s390/include/asm/gmap.h > > @@ -148,4 +148,6 @@ void gmap_sync_dirty_log_pmd(struct gmap *gmap, unsigned long dirty_bitmap[4], > > unsigned long gaddr, unsigned long vmaddr); > > int gmap_mark_unmergeable(void); > > void s390_reset_acc(struct mm_struct *mm); > > +void s390_remove_old_asce(struct gmap *gmap); > > +int s390_replace_asce(struct gmap *gmap); > > #endif /* _ASM_S390_GMAP_H */ > > diff --git a/arch/s390/kvm/pv.c b/arch/s390/kvm/pv.c > > index 7f7c0d6af2ce..3c59ef763dde 100644 > > --- a/arch/s390/kvm/pv.c > > +++ b/arch/s390/kvm/pv.c > > @@ -166,10 +166,13 @@ int kvm_s390_pv_deinit_vm(struct kvm *kvm, u16 *rc, u16 *rrc) > > atomic_set(&kvm->mm->context.is_protected, 0); > > KVM_UV_EVENT(kvm, 3, "PROTVIRT DESTROY VM: rc %x rrc %x", *rc, *rrc); > > WARN_ONCE(cc, "protvirt destroy vm failed rc %x rrc %x", *rc, *rrc); > > - /* Inteded memory leak on "impossible" error */ > > - if (!cc) > > + /* Intended memory leak on "impossible" error */ > > + if (!cc) { > > kvm_s390_pv_dealloc_vm(kvm); > > - return cc ? -EIO : 0; > > + return 0; > > + } > > + s390_replace_asce(kvm->arch.gmap); > > + return -EIO; > > } > > > > int kvm_s390_pv_init_vm(struct kvm *kvm, u16 *rc, u16 *rrc) > > diff --git a/arch/s390/mm/gmap.c b/arch/s390/mm/gmap.c > > index dfee0ebb2fac..3b42bf7adb77 100644 > > --- a/arch/s390/mm/gmap.c > > +++ b/arch/s390/mm/gmap.c > > @@ -2714,3 +2714,74 @@ void s390_reset_acc(struct mm_struct *mm) > > mmput(mm); > > } > > EXPORT_SYMBOL_GPL(s390_reset_acc); > > + > > +/** > > + * s390_remove_old_asce - Remove the topmost level of page tables from the > > + * list of page tables of the gmap. > > + * @gmap the gmap whose table is to be removed > > + * > > + * This means that it will not be freed when the VM is torn down, and needs > > + * to be handled separately by the caller, unless an intentional leak is > > + * intended. Notice that this function will only remove the page from the > > + * list, the page will still be used as a top level page table (and ASCE). > > + */ > > +void s390_remove_old_asce(struct gmap *gmap) > > +{ > > + struct page *old; > > + > > + old = virt_to_page(gmap->table); > > + spin_lock(&gmap->guest_table_lock); > > + list_del(&old->lru); > > + /* > > + * in case the ASCE needs to be "removed" multiple times, for example > > + * if the VM is rebooted into secure mode several times > > + * concurrently. > > + */ > > How can that happen, what are we protecting against here? for example if replace_asce fails, and we call it again later. in that case we have removed the old asce from the list ( = it won't be freed when the VM terminates), but the ASCE is still in use and still pointed to. a subsequent call to replace_asce will follow the pointer and try to remove the same page from the list _again_. Therefore it's necessary that the page of the ASCE has valid pointers, so list_del can work (and do nothing) without dereferencing stale or invalid pointers. Maybe I should improve the comment > > > + INIT_LIST_HEAD(&old->lru); > > + spin_unlock(&gmap->guest_table_lock); > > +} > > +EXPORT_SYMBOL_GPL(s390_remove_old_asce); > > + > > +/** > > + * s390_replace_asce - Try to replace the current ASCE of a gmap with > > + * another equivalent one. > > + * @gmap the gmap > > + * > > + * If the allocation of the new top level page table fails, the ASCE is not > > + * replaced. > > + * In any case, the old ASCE is always removed from the list. Therefore the > > + * caller has to make sure to save a pointer to it beforehands, unless an > > + * intentional leak is intended. > > + */ > > +int s390_replace_asce(struct gmap *gmap) > > +{ > > + unsigned long asce; > > + struct page *page; > > + void *table; > > + > > + s390_remove_old_asce(gmap); > > + > > + page = alloc_pages(GFP_KERNEL_ACCOUNT, CRST_ALLOC_ORDER); > > + if (!page) > > + return -ENOMEM; > > + table = page_to_virt(page); > > + memcpy(table, gmap->table, 1UL << (CRST_ALLOC_ORDER + PAGE_SHIFT)); > > + > > + /* > > + * The caller has to deal with the old ASCE, but here we make sure > > + * the new one is properly added to the list of page tables, so that > > + * it will be freed when the VM is torn down. > > + */ > > + spin_lock(&gmap->guest_table_lock); > > + list_add(&page->lru, &gmap->crst_list); > > + spin_unlock(&gmap->guest_table_lock); > > + > > + /* Set new table origin while preserving existing ASCE control bits */ > > + asce = (gmap->asce & ~_ASCE_ORIGIN) | __pa(table); > > + WRITE_ONCE(gmap->asce, asce); > > + WRITE_ONCE(gmap->mm->context.gmap_asce, asce); > > + WRITE_ONCE(gmap->table, table); > > + > > + return 0; > > +} > > +EXPORT_SYMBOL_GPL(s390_replace_asce); >