Received: by 2002:a05:6358:4e97:b0:b3:742d:4702 with SMTP id ce23csp5537667rwb; Wed, 17 Aug 2022 20:49:52 -0700 (PDT) X-Google-Smtp-Source: AA6agR5sDVGWc59rFRT7v5wbL7jeVvOLYqBdRzeMgtL3sl49mubljmjL6zT7JXsIR3hNIQIVKSae X-Received: by 2002:a05:6402:d06:b0:440:3e9d:77d with SMTP id eb6-20020a0564020d0600b004403e9d077dmr740567edb.286.1660794591887; Wed, 17 Aug 2022 20:49:51 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1660794591; cv=none; d=google.com; s=arc-20160816; b=jgiMbWI+Q0SWKzWUDdareFWVcTw6NXDvUYbtUWSIJUUhqcwr4JO9owb6H6kULo21nB 2HSWqwinIbC19mIzxjNwyhN+no3CNKrBBWQ0m+faV+Tq6jfMp3egKoZsji0bpxrxess5 ufVekTcuRYMVzMrLy1MnR8wB6U0hQ500C4A5x7ON/EhLCq019TB3Jn5OwSuUgIXJTsRy KzN9sWNfdU4G+3UW/ngW37St/yE7944Og7rf7goc84ZhaIwWOtcEMXyI5II3hPam/p1X YWXKCqdJtNDFGyqKBVa5ZJLJQjQNV0ySRt/mijvC4IYx230G75mLdSZyykX6hsrH/ZL6 hGoQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=awFGwwSRocOeifmjAPEDB1aSCbnwQlgCGRRbkk5b0mw=; b=YFa5Kwq3v4msMLuDy9Rlw8q1CWOvBvnxsWAHk1gMx8/t20iKgCEDREyWv9u0XPeCb2 U7b06g1KuPj5/yQxRrL8J5O3gfLvvyBz8WlN7GHTe9wofGya32KXXaaIDZp1nrRJsMOk HyuCopIDlaouPWyHJ/+8bWoWAK6/VhKxy7bIpcCyz78oN27GduKMGpXSsaNJoEMANn2b HG47Z27qKIAL/Um2OGT8YfKBrgp9W87vYBR0aSl3esyDC7VXSg96sH4ub1v9K82QO22B cJi2Npr3BGVLrDNacTg0adzKOdYGML457wnMAgGnWcuM15Qt9ALVu6UwK392P8+I5SMo 57tA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20210112 header.b=HuuwBZom; spf=pass (google.com: domain of linux-crypto-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-crypto-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id di6-20020a170906730600b007309e910729si279526ejc.580.2022.08.17.20.49.05; Wed, 17 Aug 2022 20:49:51 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-crypto-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20210112 header.b=HuuwBZom; spf=pass (google.com: domain of linux-crypto-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-crypto-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S242850AbiHRDsz (ORCPT + 99 others); Wed, 17 Aug 2022 23:48:55 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:40888 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S237655AbiHRDsD (ORCPT ); Wed, 17 Aug 2022 23:48:03 -0400 Received: from mail-il1-x12c.google.com (mail-il1-x12c.google.com [IPv6:2607:f8b0:4864:20::12c]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id ACDC19411D for ; Wed, 17 Aug 2022 20:47:44 -0700 (PDT) Received: by mail-il1-x12c.google.com with SMTP id k3so251589ilv.11 for ; Wed, 17 Aug 2022 20:47:44 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:from:to:cc; bh=awFGwwSRocOeifmjAPEDB1aSCbnwQlgCGRRbkk5b0mw=; b=HuuwBZomHQD9jDKeuTr2VNO9olgvvd8D0ZYsRVU+dLL2RPb2ssLOWCDp2onOIcQ1PZ kfoCj9T0R0Xfbfx7lF24moS/6w5INxaWGwptve9xDhrNP2Adv4KzaBkwltgUq1z21AWm uHVDVvnl6LcFUbGVTqMFFxCD/aL6FhbjXqH7OMB/7kfR1fPv0czzY1fd//fIMposD07f b9aTXeoPVIk/RWKntVLDYbMIHvqFhziJDS0KKv60RjWzmPM4vTWW81Fe2uEsoEo0KPwq 5EJFihNKM43HercJ872rCFaYNpPZ/exlMZgcpbYowu087PnQKgRbawlj5pxDv6RAdF1V bqaA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:x-gm-message-state:from:to:cc; bh=awFGwwSRocOeifmjAPEDB1aSCbnwQlgCGRRbkk5b0mw=; b=qTWEplDBcbR4Ovy/x5PkhtfqTcRw8Bj60UY8s/Ix00z/jr0MqB/SxvrcFhhR4jD2sS zbPMz3r342BySexA+n1/f1u6uaJqIPqxkqE3YgRmBJVGXSrM1noMvuUzFfjZHwkpWbjV B9OerbJ0v+FCVtDyc+jSCXgUz6pgnVzMGd1hxk+v6/DgZGq+Ibed1YQvd226wXdIJXx3 GsobyysjQ5PcR4G9ZcnIZ72mJchJt9ZB03xMLWEXkJ4VSYrwzHYc2HWAKptm9omPYKve sjZN/nSN7bw4joBDb6oR0sChbLi6VptBEK/l2oUykykumvy5rQFxPBeigQPWPrMgHRfK EwFA== X-Gm-Message-State: ACgBeo2VBuItIPuug23FJpHf75VsF3sF3mzeI3nuhung/s68OXFevAqH tAseMx5dK90JEJZZgOpnLW6t+KH+LA5ROXDjjvO65A== X-Received: by 2002:a92:3652:0:b0:2df:4133:787 with SMTP id d18-20020a923652000000b002df41330787mr590419ilf.39.1660794463902; Wed, 17 Aug 2022 20:47:43 -0700 (PDT) MIME-Version: 1.0 References: <34246866043db7bab34a92fe22f359667ab155a0.1655761627.git.ashish.kalra@amd.com> In-Reply-To: <34246866043db7bab34a92fe22f359667ab155a0.1655761627.git.ashish.kalra@amd.com> From: Alper Gun Date: Wed, 17 Aug 2022 20:47:33 -0700 Message-ID: Subject: Re: [PATCH Part2 v6 39/49] KVM: SVM: Introduce ops for the post gfn map and unmap To: Ashish Kalra Cc: x86@kernel.org, linux-kernel@vger.kernel.org, kvm@vger.kernel.org, linux-coco@lists.linux.dev, linux-mm@kvack.org, linux-crypto@vger.kernel.org, tglx@linutronix.de, mingo@redhat.com, jroedel@suse.de, thomas.lendacky@amd.com, hpa@zytor.com, ardb@kernel.org, pbonzini@redhat.com, seanjc@google.com, vkuznets@redhat.com, jmattson@google.com, luto@kernel.org, dave.hansen@linux.intel.com, slp@redhat.com, pgonda@google.com, peterz@infradead.org, srinivas.pandruvada@linux.intel.com, rientjes@google.com, dovmurik@linux.ibm.com, tobin@ibm.com, bp@alien8.de, michael.roth@amd.com, vbabka@suse.cz, kirill@shutemov.name, ak@linux.intel.com, tony.luck@intel.com, marcorr@google.com, sathyanarayanan.kuppuswamy@linux.intel.com, dgilbert@redhat.com, jarkko@kernel.org Content-Type: text/plain; charset="UTF-8" X-Spam-Status: No, score=-17.6 required=5.0 tests=BAYES_00,DKIMWL_WL_MED, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF, ENV_AND_HDR_SPF_MATCH,RCVD_IN_DNSWL_NONE,SPF_HELO_NONE,SPF_PASS, T_SCC_BODY_TEXT_LINE,USER_IN_DEF_DKIM_WL,USER_IN_DEF_SPF_WL autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-crypto@vger.kernel.org On Mon, Jun 20, 2022 at 4:12 PM Ashish Kalra wrote: > > From: Brijesh Singh > > When SEV-SNP is enabled in the guest VM, the guest memory pages can > either be a private or shared. A write from the hypervisor goes through > the RMP checks. If hardware sees that hypervisor is attempting to write > to a guest private page, then it triggers an RMP violation #PF. > > To avoid the RMP violation with GHCB pages, added new post_{map,unmap}_gfn > functions to verify if its safe to map GHCB pages. Uses a spinlock to > protect against the page state change for existing mapped pages. > > Need to add generic post_{map,unmap}_gfn() ops that can be used to verify > that its safe to map a given guest page in the hypervisor. > > This patch will need to be revisited later after consensus is reached on > how to manage guest private memory as probably UPM private memslots will > be able to handle this page state change more gracefully. > > Signed-off-by: Brijesh Singh > Signed-off by: Ashish Kalra > --- > arch/x86/include/asm/kvm-x86-ops.h | 1 + > arch/x86/include/asm/kvm_host.h | 3 ++ > arch/x86/kvm/svm/sev.c | 48 ++++++++++++++++++++++++++++-- > arch/x86/kvm/svm/svm.c | 3 ++ > arch/x86/kvm/svm/svm.h | 11 +++++++ > 5 files changed, 64 insertions(+), 2 deletions(-) > > diff --git a/arch/x86/include/asm/kvm-x86-ops.h b/arch/x86/include/asm/kvm-x86-ops.h > index e0068e702692..2dd2bc0cf4c3 100644 > --- a/arch/x86/include/asm/kvm-x86-ops.h > +++ b/arch/x86/include/asm/kvm-x86-ops.h > @@ -130,6 +130,7 @@ KVM_X86_OP(vcpu_deliver_sipi_vector) > KVM_X86_OP_OPTIONAL_RET0(vcpu_get_apicv_inhibit_reasons); > KVM_X86_OP(alloc_apic_backing_page) > KVM_X86_OP_OPTIONAL(rmp_page_level_adjust) > +KVM_X86_OP(update_protected_guest_state) > > #undef KVM_X86_OP > #undef KVM_X86_OP_OPTIONAL > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > index 49b217dc8d7e..8abc0e724f5c 100644 > --- a/arch/x86/include/asm/kvm_host.h > +++ b/arch/x86/include/asm/kvm_host.h > @@ -1522,7 +1522,10 @@ struct kvm_x86_ops { > unsigned long (*vcpu_get_apicv_inhibit_reasons)(struct kvm_vcpu *vcpu); > > void *(*alloc_apic_backing_page)(struct kvm_vcpu *vcpu); > + > void (*rmp_page_level_adjust)(struct kvm *kvm, kvm_pfn_t pfn, int *level); > + > + int (*update_protected_guest_state)(struct kvm_vcpu *vcpu); > }; > > struct kvm_x86_nested_ops { > diff --git a/arch/x86/kvm/svm/sev.c b/arch/x86/kvm/svm/sev.c > index cb2d1bbb862b..4ed90331bca0 100644 > --- a/arch/x86/kvm/svm/sev.c > +++ b/arch/x86/kvm/svm/sev.c > @@ -341,6 +341,7 @@ static int sev_guest_init(struct kvm *kvm, struct kvm_sev_cmd *argp) > if (ret) > goto e_free; > > + spin_lock_init(&sev->psc_lock); > ret = sev_snp_init(&argp->error); > } else { > ret = sev_platform_init(&argp->error); > @@ -2828,19 +2829,28 @@ static inline int svm_map_ghcb(struct vcpu_svm *svm, struct kvm_host_map *map) > { > struct vmcb_control_area *control = &svm->vmcb->control; > u64 gfn = gpa_to_gfn(control->ghcb_gpa); > + struct kvm_vcpu *vcpu = &svm->vcpu; > > - if (kvm_vcpu_map(&svm->vcpu, gfn, map)) { > + if (kvm_vcpu_map(vcpu, gfn, map)) { > /* Unable to map GHCB from guest */ > pr_err("error mapping GHCB GFN [%#llx] from guest\n", gfn); > return -EFAULT; > } > > + if (sev_post_map_gfn(vcpu->kvm, map->gfn, map->pfn)) { > + kvm_vcpu_unmap(vcpu, map, false); > + return -EBUSY; > + } > + > return 0; > } > > static inline void svm_unmap_ghcb(struct vcpu_svm *svm, struct kvm_host_map *map) > { > - kvm_vcpu_unmap(&svm->vcpu, map, true); > + struct kvm_vcpu *vcpu = &svm->vcpu; > + > + kvm_vcpu_unmap(vcpu, map, true); > + sev_post_unmap_gfn(vcpu->kvm, map->gfn, map->pfn); > } > > static void dump_ghcb(struct vcpu_svm *svm) > @@ -3383,6 +3393,8 @@ static int __snp_handle_page_state_change(struct kvm_vcpu *vcpu, enum psc_op op, > return PSC_UNDEF_ERR; > } > > + spin_lock(&sev->psc_lock); > + > write_lock(&kvm->mmu_lock); > > rc = kvm_mmu_get_tdp_walk(vcpu, gpa, &pfn, &npt_level); > @@ -3417,6 +3429,8 @@ static int __snp_handle_page_state_change(struct kvm_vcpu *vcpu, enum psc_op op, > > write_unlock(&kvm->mmu_lock); > > + spin_unlock(&sev->psc_lock); There is a corner case where the psc_lock is not released. If kvm_mmu_get_tdp_walk fails, the lock will be kept and will cause soft lockup. > + > if (rc) { > pr_err_ratelimited("Error op %d gpa %llx pfn %llx level %d rc %d\n", > op, gpa, pfn, level, rc); > @@ -3965,3 +3979,33 @@ void sev_rmp_page_level_adjust(struct kvm *kvm, kvm_pfn_t pfn, int *level) > /* Adjust the level to keep the NPT and RMP in sync */ > *level = min_t(size_t, *level, rmp_level); > } > + > +int sev_post_map_gfn(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn) > +{ > + struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info; > + int level; > + > + if (!sev_snp_guest(kvm)) > + return 0; > + > + spin_lock(&sev->psc_lock); > + > + /* If pfn is not added as private then fail */ > + if (snp_lookup_rmpentry(pfn, &level) == 1) { > + spin_unlock(&sev->psc_lock); > + pr_err_ratelimited("failed to map private gfn 0x%llx pfn 0x%llx\n", gfn, pfn); > + return -EBUSY; > + } > + > + return 0; > +} > + > +void sev_post_unmap_gfn(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn) > +{ > + struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info; > + > + if (!sev_snp_guest(kvm)) > + return; > + > + spin_unlock(&sev->psc_lock); > +} > diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c > index b24e0171cbf2..1c8e035ba011 100644 > --- a/arch/x86/kvm/svm/svm.c > +++ b/arch/x86/kvm/svm/svm.c > @@ -4734,7 +4734,10 @@ static struct kvm_x86_ops svm_x86_ops __initdata = { > .vcpu_get_apicv_inhibit_reasons = avic_vcpu_get_apicv_inhibit_reasons, > > .alloc_apic_backing_page = svm_alloc_apic_backing_page, > + > .rmp_page_level_adjust = sev_rmp_page_level_adjust, > + > + .update_protected_guest_state = sev_snp_update_protected_guest_state, > }; > > /* > diff --git a/arch/x86/kvm/svm/svm.h b/arch/x86/kvm/svm/svm.h > index 54ff56cb6125..3fd95193ed8d 100644 > --- a/arch/x86/kvm/svm/svm.h > +++ b/arch/x86/kvm/svm/svm.h > @@ -79,19 +79,25 @@ struct kvm_sev_info { > bool active; /* SEV enabled guest */ > bool es_active; /* SEV-ES enabled guest */ > bool snp_active; /* SEV-SNP enabled guest */ > + > unsigned int asid; /* ASID used for this guest */ > unsigned int handle; /* SEV firmware handle */ > int fd; /* SEV device fd */ > + > unsigned long pages_locked; /* Number of pages locked */ > struct list_head regions_list; /* List of registered regions */ > + > u64 ap_jump_table; /* SEV-ES AP Jump Table address */ > + > struct kvm *enc_context_owner; /* Owner of copied encryption context */ > struct list_head mirror_vms; /* List of VMs mirroring */ > struct list_head mirror_entry; /* Use as a list entry of mirrors */ > struct misc_cg *misc_cg; /* For misc cgroup accounting */ > atomic_t migration_in_progress; > + > u64 snp_init_flags; > void *snp_context; /* SNP guest context page */ > + spinlock_t psc_lock; > }; > > struct kvm_svm { > @@ -702,6 +708,11 @@ void sev_es_prepare_switch_to_guest(struct sev_es_save_area *hostsa); > void sev_es_unmap_ghcb(struct vcpu_svm *svm); > struct page *snp_safe_alloc_page(struct kvm_vcpu *vcpu); > void sev_rmp_page_level_adjust(struct kvm *kvm, kvm_pfn_t pfn, int *level); > +int sev_post_map_gfn(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn); > +void sev_post_unmap_gfn(struct kvm *kvm, gfn_t gfn, kvm_pfn_t pfn); > +void handle_rmp_page_fault(struct kvm_vcpu *vcpu, gpa_t gpa, u64 error_code); > +void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu); > +int sev_snp_update_protected_guest_state(struct kvm_vcpu *vcpu); > > /* vmenter.S */ > > -- > 2.25.1 >