Received: by 2002:a05:6902:102b:0:0:0:0 with SMTP id x11csp2435422ybt; Fri, 3 Jul 2020 09:01:21 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxVzsTQVYczciw8vdEo4LDjmM4hxG8WhiC6P+Oy1Y+EMXz7niyifg3Y85lykou8cTg/PTkR X-Received: by 2002:aa7:dd8e:: with SMTP id g14mr42735914edv.208.1593792081272; Fri, 03 Jul 2020 09:01:21 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1593792081; cv=none; d=google.com; s=arc-20160816; b=iVTl84SdAOsv3sNLFBR0nZnk427dtCNoVn5f9SRzCr4D3ueRKA4AVsTVLp5Zs52Aa4 KSFff8XZ/9slJ5nBEVG9vQjDeejnDYMsnfbWeeEUuMqnJzI8t6wxWWw0pvmx25DurPYC iO4DnLezp/UXbUjUCscAwJwzLAbTh+fcWxh4l1x98YXEf5H0qdMHf71pAdwl8oOqxVXp Fzi1fsZ+UycmOv9kKESwqpco2cfvPlaFvdNXMHd8jkl+dulYBfbZdci5YI/Tp5icZek1 ayBzv06oR7F50YdtZpb+53UvjEj62J6b6c4pqYz96y3qV+nwn6s+tnvd7dLJPCmVqFkD oNEw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=BLmDre6E+VPVtQgxhLXokrCdEgQkyqQbZ6c/Arncmjk=; b=sb41mkXWz3nejsOXFdkQvMONb+mmuy6A8xYNQbcct2ABfjOxirxdN7AR6iuVYreai+ YZBieBtHuxDZAV8CvdGwAQsxFjHTQBwllz7Pb4tk3LBr6nEPW4ca3YFrYZarxGhaedOP 12YZxc6jFowcsmhFeERUWPJEn4JZaOvE3Ib21Z/XbzOAMaKoPhkNmPCQF6udeiYaHtm/ aPGaa/ydgnMIqBZZ4A4qqWnhsllgH6TvCGCpRsNeWFW4PzDf0I+YmXKuTat2ExpoG4G+ daw6w1A0EsLCgrtx/QWgjOtL+wG25EEZPD1mewom0Hc+k/WxbvH1y5LYzrFMlA8hfodw HE+A== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=ibm.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id dk5si1841859edb.116.2020.07.03.09.00.58; Fri, 03 Jul 2020 09:01:21 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=ibm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726475AbgGCP7h (ORCPT + 99 others); Fri, 3 Jul 2020 11:59:37 -0400 Received: from mx0a-001b2d01.pphosted.com ([148.163.156.1]:40082 "EHLO mx0a-001b2d01.pphosted.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726098AbgGCP7h (ORCPT ); Fri, 3 Jul 2020 11:59:37 -0400 Received: from pps.filterd (m0187473.ppops.net [127.0.0.1]) by mx0a-001b2d01.pphosted.com (8.16.0.42/8.16.0.42) with SMTP id 063FXLhO172941; Fri, 3 Jul 2020 11:59:23 -0400 Received: from ppma03ams.nl.ibm.com (62.31.33a9.ip4.static.sl-reverse.com [169.51.49.98]) by mx0a-001b2d01.pphosted.com with ESMTP id 32121rh6e7-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Fri, 03 Jul 2020 11:59:22 -0400 Received: from pps.filterd (ppma03ams.nl.ibm.com [127.0.0.1]) by ppma03ams.nl.ibm.com (8.16.0.42/8.16.0.42) with SMTP id 063FtKYh005927; Fri, 3 Jul 2020 15:59:20 GMT Received: from b06avi18878370.portsmouth.uk.ibm.com (b06avi18878370.portsmouth.uk.ibm.com [9.149.26.194]) by ppma03ams.nl.ibm.com with ESMTP id 3217b01ufy-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Fri, 03 Jul 2020 15:59:20 +0000 Received: from d06av24.portsmouth.uk.ibm.com (mk.ibm.com [9.149.105.60]) by b06avi18878370.portsmouth.uk.ibm.com (8.14.9/8.14.9/NCO v10.0) with ESMTP id 063FxHqC66060656 (version=TLSv1/SSLv3 cipher=DHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Fri, 3 Jul 2020 15:59:17 GMT Received: from d06av24.portsmouth.uk.ibm.com (unknown [127.0.0.1]) by IMSVA (Postfix) with ESMTP id 53AD642045; Fri, 3 Jul 2020 15:59:17 +0000 (GMT) Received: from d06av24.portsmouth.uk.ibm.com (unknown [127.0.0.1]) by IMSVA (Postfix) with ESMTP id E481342047; Fri, 3 Jul 2020 15:59:16 +0000 (GMT) Received: from pomme.tlslab.ibm.com (unknown [9.145.68.59]) by d06av24.portsmouth.uk.ibm.com (Postfix) with ESMTP; Fri, 3 Jul 2020 15:59:16 +0000 (GMT) From: Laurent Dufour To: linux-kernel@vger.kernel.org, kvm-ppc@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, mpe@ellerman.id.au, paulus@samba.org Cc: bharata@linux.ibm.com, bauerman@linux.ibm.com, sukadev@linux.ibm.com, sathnaga@linux.vnet.ibm.com, Ram Pai , Paul Mackerras Subject: [PATCH 2/2] KVM: PPC: Book3S HV: rework secure mem slot dropping Date: Fri, 3 Jul 2020 17:59:14 +0200 Message-Id: <20200703155914.40262-3-ldufour@linux.ibm.com> X-Mailer: git-send-email 2.27.0 In-Reply-To: <20200703155914.40262-1-ldufour@linux.ibm.com> References: <20200703155914.40262-1-ldufour@linux.ibm.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-TM-AS-GCONF: 00 X-Proofpoint-Virus-Version: vendor=fsecure engine=2.50.10434:6.0.235,18.0.687 definitions=2020-07-03_10:2020-07-02,2020-07-03 signatures=0 X-Proofpoint-Spam-Details: rule=outbound_notspam policy=outbound score=0 priorityscore=1501 phishscore=0 impostorscore=0 cotscore=-2147483648 mlxscore=0 malwarescore=0 clxscore=1015 mlxlogscore=702 spamscore=0 lowpriorityscore=0 adultscore=0 bulkscore=0 suspectscore=2 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2004280000 definitions=main-2007030106 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org When a secure memslot is dropped, all the pages backed in the secure device (aka really backed by secure memory by the Ultravisor) should be paged out to a normal page. Previously, this was achieved by triggering the page fault mechanism which is calling kvmppc_svm_page_out() on each pages. This can't work when hot unplugging a memory slot because the memory slot is flagged as invalid and gfn_to_pfn() is then not trying to access the page, so the page fault mechanism is not triggered. Since the final goal is to make a call to kvmppc_svm_page_out() it seems simpler to directly calling it instead of triggering such a mechanism. This way kvmppc_uvmem_drop_pages() can be called even when hot unplugging a memslot. Since kvmppc_uvmem_drop_pages() is already holding kvm->arch.uvmem_lock, the call to __kvmppc_svm_page_out() is made. As __kvmppc_svm_page_out needs the vma pointer to migrate the pages, the VMA is fetched in a lazy way, to not trigger find_vma() all the time. In addition, the mmap_sem is help in read mode during that time, not in write mode since the virual memory layout is not impacted, and kvm->arch.uvmem_lock prevents concurrent operation on the secure device. Cc: Ram Pai Cc: Bharata B Rao Cc: Paul Mackerras Signed-off-by: Laurent Dufour --- arch/powerpc/kvm/book3s_hv_uvmem.c | 54 ++++++++++++++++++++---------- 1 file changed, 37 insertions(+), 17 deletions(-) diff --git a/arch/powerpc/kvm/book3s_hv_uvmem.c b/arch/powerpc/kvm/book3s_hv_uvmem.c index 852cc9ae6a0b..479ddf16d18c 100644 --- a/arch/powerpc/kvm/book3s_hv_uvmem.c +++ b/arch/powerpc/kvm/book3s_hv_uvmem.c @@ -533,35 +533,55 @@ static inline int kvmppc_svm_page_out(struct vm_area_struct *vma, * fault on them, do fault time migration to replace the device PTEs in * QEMU page table with normal PTEs from newly allocated pages. */ -void kvmppc_uvmem_drop_pages(const struct kvm_memory_slot *free, +void kvmppc_uvmem_drop_pages(const struct kvm_memory_slot *slot, struct kvm *kvm, bool skip_page_out) { int i; struct kvmppc_uvmem_page_pvt *pvt; - unsigned long pfn, uvmem_pfn; - unsigned long gfn = free->base_gfn; + struct page *uvmem_page; + struct vm_area_struct *vma = NULL; + unsigned long uvmem_pfn, gfn; + unsigned long addr, end; + + down_read(&kvm->mm->mmap_sem); + + addr = slot->userspace_addr; + end = addr + (slot->npages * PAGE_SIZE); - for (i = free->npages; i; --i, ++gfn) { - struct page *uvmem_page; + gfn = slot->base_gfn; + for (i = slot->npages; i; --i, ++gfn, addr += PAGE_SIZE) { + + /* Fetch the VMA if addr is not in the latest fetched one */ + if (!vma || (addr < vma->vm_start || addr >= vma->vm_end)) { + vma = find_vma_intersection(kvm->mm, addr, end); + if (!vma || + vma->vm_start > addr || vma->vm_end < end) { + pr_err("Can't find VMA for gfn:0x%lx\n", gfn); + break; + } + } mutex_lock(&kvm->arch.uvmem_lock); - if (!kvmppc_gfn_is_uvmem_pfn(gfn, kvm, &uvmem_pfn)) { + + if (kvmppc_gfn_is_uvmem_pfn(gfn, kvm, &uvmem_pfn)) { + uvmem_page = pfn_to_page(uvmem_pfn); + pvt = uvmem_page->zone_device_data; + pvt->skip_page_out = skip_page_out; + pvt->remove_gfn = true; + + if (__kvmppc_svm_page_out(vma, addr, addr + PAGE_SIZE, + PAGE_SHIFT, kvm, pvt->gpa)) + pr_err("Can't page out gpa:0x%lx addr:0x%lx\n", + pvt->gpa, addr); + } else { + /* Remove the shared flag if any */ kvmppc_gfn_remove(gfn, kvm); - mutex_unlock(&kvm->arch.uvmem_lock); - continue; } - uvmem_page = pfn_to_page(uvmem_pfn); - pvt = uvmem_page->zone_device_data; - pvt->skip_page_out = skip_page_out; - pvt->remove_gfn = true; mutex_unlock(&kvm->arch.uvmem_lock); - - pfn = gfn_to_pfn(kvm, gfn); - if (is_error_noslot_pfn(pfn)) - continue; - kvm_release_pfn_clean(pfn); } + + up_read(&kvm->mm->mmap_sem); } unsigned long kvmppc_h_svm_init_abort(struct kvm *kvm) -- 2.27.0