Received: by 2002:a05:7412:b795:b0:e2:908c:2ebd with SMTP id iv21csp423905rdb; Thu, 2 Nov 2023 07:36:01 -0700 (PDT) X-Google-Smtp-Source: AGHT+IHqxFW1QrmUIhfcb35P0pZQaPGIPyHhlZf02DXwKxuOxSXvSnlF2SjoGqyB89QLyWOJsawo X-Received: by 2002:a05:6a21:3395:b0:17a:cfce:5a30 with SMTP id yy21-20020a056a21339500b0017acfce5a30mr18578038pzb.16.1698935760809; Thu, 02 Nov 2023 07:36:00 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1698935760; cv=none; d=google.com; s=arc-20160816; b=TzBFQlMZPgsR5fQvPwKPvajvClL33McVNCWaTw7YJZHQe65+rk5YM1NFsxEvJgfrt4 7vBkFu+RZkXnlsRBUNQivgHaGou0+tMlqk7DtrhB8qIrwc2KIPoIS5XPFcqJYYoWTqtB c9MfN3rg716VxL1X7ZyUHs+hNvhkJ6Pppx6kw6q8EiwKSCeHdOCsEXvRz02gs/dBNlHP ZCgD80XEtldJVNKiYXs0HMxrVNIQj+aVruZPGxVcGPrKXbEWgt2QbJmHfXafz3uiykjz MaXpY59BG5XEuYcALsBtuSg1MlsarhFDSS6SJbkLbC1xriG+6Sr7aWhYpZ+v96GfTz4P wGag== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:cc:to:subject :message-id:date:from:in-reply-to:references:mime-version :dkim-signature; bh=/2lJd6XsyJuryPcyY7Aaw/1637wZx+D1R9nKqgAWHhc=; fh=xK+wiX2TTq7O2X4TlNj4MXVc7h6ma8eyuy1Ur/RMLHE=; b=Qzpf8tNq1gB15zKW6c8P4srYN0gAyqsVtM/uo9f9zwr9V1UFzm9hDFSOfiHB+ftd17 FdKHKql+Qp41V8Fv9VYoQQC0ZL7OL6K2AS9SjJQCJ9A1PCsE3/zLUkSu/KQ/5j0vUDo4 Iuf41lFUOEutiaZmsL0nalDewyA0VD+jnCq9dRibW7ohyUh7zRf5U9h28N1n9dzx2Kjd +uEoqfoOC17GgYBySTXGHHvtL4imstBiEfER6PQOXrjvhUbL48zxSWmoRn2yV5ToPQzX PQnWJs//EIiU+evArC5HVe85NGzlRlJs/0YmNF6NgmD0HsUZ9Of7J9e45AvFTTFYVtch zYrQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20230601 header.b=DXf4IwJA; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.33 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from lipwig.vger.email (lipwig.vger.email. [23.128.96.33]) by mx.google.com with ESMTPS id a18-20020aa794b2000000b006c320a95e8esi1400100pfl.404.2023.11.02.07.35.57 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 02 Nov 2023 07:36:00 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.33 as permitted sender) client-ip=23.128.96.33; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20230601 header.b=DXf4IwJA; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.33 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by lipwig.vger.email (Postfix) with ESMTP id CE7F9809E203; Thu, 2 Nov 2023 07:35:38 -0700 (PDT) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.10 at lipwig.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230205AbjKBOfZ (ORCPT + 99 others); Thu, 2 Nov 2023 10:35:25 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:34600 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230186AbjKBOfW (ORCPT ); Thu, 2 Nov 2023 10:35:22 -0400 Received: from mail-qv1-xf2c.google.com (mail-qv1-xf2c.google.com [IPv6:2607:f8b0:4864:20::f2c]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 910C3133 for ; Thu, 2 Nov 2023 07:35:17 -0700 (PDT) Received: by mail-qv1-xf2c.google.com with SMTP id 6a1803df08f44-66d12b3b479so5706226d6.1 for ; Thu, 02 Nov 2023 07:35:17 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20230601; t=1698935716; x=1699540516; darn=vger.kernel.org; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:from:to:cc:subject:date :message-id:reply-to; bh=/2lJd6XsyJuryPcyY7Aaw/1637wZx+D1R9nKqgAWHhc=; b=DXf4IwJAu8bEvsnw7UWI046t/VyHgiyBZCraX3/XvzB+jbzE162rL1B+8jVID/4UQa uMy9vCpY1EBqgt3EVus260jVCYgKrDKjYEMK4zZS0HJ4URMKu3bGmeh73RdIYve+z+O0 nSImY52iUTw5RNo1B/yXW7VKej2QFOgUfJJ6+XoZzZNn8/olpxfuYDZ/cdkfMGGdkxS0 Lt0FI8eNbT8W41bI/AdzxZ7YFcVDfrcKe+oGcidJl74nLL38BITI6JEXELlZ2L3SfPi7 OzjeDh7koa76/zJramWfL5VZf2aHM0EwJpCFWlw0YVe0mDgmQhtTHEvIC8Jh9D70XHRH CEVw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1698935716; x=1699540516; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=/2lJd6XsyJuryPcyY7Aaw/1637wZx+D1R9nKqgAWHhc=; b=SpgoMavtGVcg56A0HuNdbReS0GLay05pS/ZG/UOi13ISjqxJC5XW9XyoqsIdCFBNbk ukgWkNQT1seKov42oW9n2NNO6bEQ5tPCsGkPOXk8zAcOPUkZYfMdBIuDf2wNERR4wG9n HWjvhnhj1ncHFVRKrCT3x2MD7rbdfimzkqLtPbysRJLGWDm5uH8uDKbbXQi/8/d8f5rs QpNbKNLiUvxWXwIR8ZzmphuurNigpCbSMpe7wkFQ9D/s61cB9c45EK2AgB+06vgNooDH DPYv+FY6Et7SL4J4c+uWd9zYglPGpEkrbLwEovFRioLt2PrNVe9mbslGTEEQqljkTbe4 kdpA== X-Gm-Message-State: AOJu0YymE6pEMxfHThVNqtVhekMtwiWZnYCC2S/A5mcpQSMlpVnxBldR eB/C7KplQzPqpDDszl4Ksz4CHWctyP5/WpCRWKr6vw== X-Received: by 2002:ad4:5761:0:b0:671:6003:73b1 with SMTP id r1-20020ad45761000000b00671600373b1mr17968741qvx.60.1698935716453; Thu, 02 Nov 2023 07:35:16 -0700 (PDT) MIME-Version: 1.0 References: <20231027182217.3615211-1-seanjc@google.com> <20231027182217.3615211-21-seanjc@google.com> In-Reply-To: <20231027182217.3615211-21-seanjc@google.com> From: Fuad Tabba Date: Thu, 2 Nov 2023 14:34:40 +0000 Message-ID: Subject: Re: [PATCH v13 20/35] KVM: x86/mmu: Handle page fault for private memory To: Sean Christopherson Cc: Paolo Bonzini , Marc Zyngier , Oliver Upton , Huacai Chen , Michael Ellerman , Anup Patel , Paul Walmsley , Palmer Dabbelt , Albert Ou , Alexander Viro , Christian Brauner , "Matthew Wilcox (Oracle)" , Andrew Morton , kvm@vger.kernel.org, linux-arm-kernel@lists.infradead.org, kvmarm@lists.linux.dev, linux-mips@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, kvm-riscv@lists.infradead.org, linux-riscv@lists.infradead.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org, Xiaoyao Li , Xu Yilun , Chao Peng , Jarkko Sakkinen , Anish Moorthy , David Matlack , Yu Zhang , Isaku Yamahata , =?UTF-8?B?TWlja2HDq2wgU2FsYcO8bg==?= , Vlastimil Babka , Vishal Annapurve , Ackerley Tng , Maciej Szmigiero , David Hildenbrand , Quentin Perret , Michael Roth , Wang , Liam Merwick , Isaku Yamahata , "Kirill A . Shutemov" Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable X-Spam-Status: No, score=-8.4 required=5.0 tests=DKIMWL_WL_MED,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE, USER_IN_DEF_DKIM_WL autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lipwig.vger.email Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (lipwig.vger.email [0.0.0.0]); Thu, 02 Nov 2023 07:35:38 -0700 (PDT) On Fri, Oct 27, 2023 at 7:23=E2=80=AFPM Sean Christopherson wrote: > > From: Chao Peng > > Add support for resolving page faults on guest private memory for VMs > that differentiate between "shared" and "private" memory. For such VMs, > KVM_MEM_PRIVATE memslots can include both fd-based private memory and > hva-based shared memory, and KVM needs to map in the "correct" variant, > i.e. KVM needs to map the gfn shared/private as appropriate based on the > current state of the gfn's KVM_MEMORY_ATTRIBUTE_PRIVATE flag. > > For AMD's SEV-SNP and Intel's TDX, the guest effectively gets to request > shared vs. private via a bit in the guest page tables, i.e. what the gues= t > wants may conflict with the current memory attributes. To support such > "implicit" conversion requests, exit to user with KVM_EXIT_MEMORY_FAULT > to forward the request to userspace. Add a new flag for memory faults, > KVM_MEMORY_EXIT_FLAG_PRIVATE, to communicate whether the guest wants to > map memory as shared vs. private. > > Like KVM_MEMORY_ATTRIBUTE_PRIVATE, use bit 3 for flagging private memory > so that KVM can use bits 0-2 for capturing RWX behavior if/when userspace > needs such information, e.g. a likely user of KVM_EXIT_MEMORY_FAULT is to > exit on missing mappings when handling guest page fault VM-Exits. In > that case, userspace will want to know RWX information in order to > correctly/precisely resolve the fault. > > Note, private memory *must* be backed by guest_memfd, i.e. shared mapping= s > always come from the host userspace page tables, and private mappings > always come from a guest_memfd instance. > > Co-developed-by: Yu Zhang > Signed-off-by: Yu Zhang > Signed-off-by: Chao Peng > Co-developed-by: Sean Christopherson > Signed-off-by: Sean Christopherson > --- With my limited understanding of kvm-x86 mmu code: Reviewed-by: Fuad Tabba Tested the x86 code (as part of this patch series) on qemu. The x86 fault handling code was used as a guide to how it should be done for pKVM/arm64 (with similar code added there): Tested-by: Fuad Tabba Cheers, /fuad > Documentation/virt/kvm/api.rst | 8 ++- > arch/x86/kvm/mmu/mmu.c | 101 ++++++++++++++++++++++++++++++-- > arch/x86/kvm/mmu/mmu_internal.h | 1 + > include/linux/kvm_host.h | 8 ++- > include/uapi/linux/kvm.h | 1 + > 5 files changed, 110 insertions(+), 9 deletions(-) > > diff --git a/Documentation/virt/kvm/api.rst b/Documentation/virt/kvm/api.= rst > index 7f00c310c24a..38dc1fda4f45 100644 > --- a/Documentation/virt/kvm/api.rst > +++ b/Documentation/virt/kvm/api.rst > @@ -6837,6 +6837,7 @@ spec refer, https://github.com/riscv/riscv-sbi-doc. > > /* KVM_EXIT_MEMORY_FAULT */ > struct { > + #define KVM_MEMORY_EXIT_FLAG_PRIVATE (1ULL << 3) > __u64 flags; > __u64 gpa; > __u64 size; > @@ -6845,8 +6846,11 @@ spec refer, https://github.com/riscv/riscv-sbi-doc= . > KVM_EXIT_MEMORY_FAULT indicates the vCPU has encountered a memory fault = that > could not be resolved by KVM. The 'gpa' and 'size' (in bytes) describe = the > guest physical address range [gpa, gpa + size) of the fault. The 'flags= ' field > -describes properties of the faulting access that are likely pertinent. > -Currently, no flags are defined. > +describes properties of the faulting access that are likely pertinent: > + > + - KVM_MEMORY_EXIT_FLAG_PRIVATE - When set, indicates the memory fault o= ccurred > + on a private memory access. When clear, indicates the fault occurred= on a > + shared access. > > Note! KVM_EXIT_MEMORY_FAULT is unique among all KVM exit reasons in tha= t it > accompanies a return code of '-1', not '0'! errno will always be set to= EFAULT > diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c > index 4167d557c577..c4e758f0aebb 100644 > --- a/arch/x86/kvm/mmu/mmu.c > +++ b/arch/x86/kvm/mmu/mmu.c > @@ -3147,9 +3147,9 @@ static int host_pfn_mapping_level(struct kvm *kvm, = gfn_t gfn, > return level; > } > > -int kvm_mmu_max_mapping_level(struct kvm *kvm, > - const struct kvm_memory_slot *slot, gfn_t g= fn, > - int max_level) > +static int __kvm_mmu_max_mapping_level(struct kvm *kvm, > + const struct kvm_memory_slot *slot= , > + gfn_t gfn, int max_level, bool is_= private) > { > struct kvm_lpage_info *linfo; > int host_level; > @@ -3161,6 +3161,9 @@ int kvm_mmu_max_mapping_level(struct kvm *kvm, > break; > } > > + if (is_private) > + return max_level; > + > if (max_level =3D=3D PG_LEVEL_4K) > return PG_LEVEL_4K; > > @@ -3168,6 +3171,16 @@ int kvm_mmu_max_mapping_level(struct kvm *kvm, > return min(host_level, max_level); > } > > +int kvm_mmu_max_mapping_level(struct kvm *kvm, > + const struct kvm_memory_slot *slot, gfn_t g= fn, > + int max_level) > +{ > + bool is_private =3D kvm_slot_can_be_private(slot) && > + kvm_mem_is_private(kvm, gfn); > + > + return __kvm_mmu_max_mapping_level(kvm, slot, gfn, max_level, is_= private); > +} > + > void kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu, struct kvm_page_faul= t *fault) > { > struct kvm_memory_slot *slot =3D fault->slot; > @@ -3188,8 +3201,9 @@ void kvm_mmu_hugepage_adjust(struct kvm_vcpu *vcpu,= struct kvm_page_fault *fault > * Enforce the iTLB multihit workaround after capturing the reque= sted > * level, which will be used to do precise, accurate accounting. > */ > - fault->req_level =3D kvm_mmu_max_mapping_level(vcpu->kvm, slot, > - fault->gfn, fault->m= ax_level); > + fault->req_level =3D __kvm_mmu_max_mapping_level(vcpu->kvm, slot, > + fault->gfn, fault-= >max_level, > + fault->is_private)= ; > if (fault->req_level =3D=3D PG_LEVEL_4K || fault->huge_page_disal= lowed) > return; > > @@ -4261,6 +4275,55 @@ void kvm_arch_async_page_ready(struct kvm_vcpu *vc= pu, struct kvm_async_pf *work) > kvm_mmu_do_page_fault(vcpu, work->cr2_or_gpa, 0, true, NULL); > } > > +static inline u8 kvm_max_level_for_order(int order) > +{ > + BUILD_BUG_ON(KVM_MAX_HUGEPAGE_LEVEL > PG_LEVEL_1G); > + > + KVM_MMU_WARN_ON(order !=3D KVM_HPAGE_GFN_SHIFT(PG_LEVEL_1G) && > + order !=3D KVM_HPAGE_GFN_SHIFT(PG_LEVEL_2M) && > + order !=3D KVM_HPAGE_GFN_SHIFT(PG_LEVEL_4K)); > + > + if (order >=3D KVM_HPAGE_GFN_SHIFT(PG_LEVEL_1G)) > + return PG_LEVEL_1G; > + > + if (order >=3D KVM_HPAGE_GFN_SHIFT(PG_LEVEL_2M)) > + return PG_LEVEL_2M; > + > + return PG_LEVEL_4K; > +} > + > +static void kvm_mmu_prepare_memory_fault_exit(struct kvm_vcpu *vcpu, > + struct kvm_page_fault *faul= t) > +{ > + kvm_prepare_memory_fault_exit(vcpu, fault->gfn << PAGE_SHIFT, > + PAGE_SIZE, fault->write, fault->exe= c, > + fault->is_private); > +} > + > +static int kvm_faultin_pfn_private(struct kvm_vcpu *vcpu, > + struct kvm_page_fault *fault) > +{ > + int max_order, r; > + > + if (!kvm_slot_can_be_private(fault->slot)) { > + kvm_mmu_prepare_memory_fault_exit(vcpu, fault); > + return -EFAULT; > + } > + > + r =3D kvm_gmem_get_pfn(vcpu->kvm, fault->slot, fault->gfn, &fault= ->pfn, > + &max_order); > + if (r) { > + kvm_mmu_prepare_memory_fault_exit(vcpu, fault); > + return r; > + } > + > + fault->max_level =3D min(kvm_max_level_for_order(max_order), > + fault->max_level); > + fault->map_writable =3D !(fault->slot->flags & KVM_MEM_READONLY); > + > + return RET_PF_CONTINUE; > +} > + > static int __kvm_faultin_pfn(struct kvm_vcpu *vcpu, struct kvm_page_faul= t *fault) > { > struct kvm_memory_slot *slot =3D fault->slot; > @@ -4293,6 +4356,14 @@ static int __kvm_faultin_pfn(struct kvm_vcpu *vcpu= , struct kvm_page_fault *fault > return RET_PF_EMULATE; > } > > + if (fault->is_private !=3D kvm_mem_is_private(vcpu->kvm, fault->g= fn)) { > + kvm_mmu_prepare_memory_fault_exit(vcpu, fault); > + return -EFAULT; > + } > + > + if (fault->is_private) > + return kvm_faultin_pfn_private(vcpu, fault); > + > async =3D false; > fault->pfn =3D __gfn_to_pfn_memslot(slot, fault->gfn, false, fals= e, &async, > fault->write, &fault->map_writa= ble, > @@ -7173,6 +7244,26 @@ void kvm_mmu_pre_destroy_vm(struct kvm *kvm) > } > > #ifdef CONFIG_KVM_GENERIC_MEMORY_ATTRIBUTES > +bool kvm_arch_pre_set_memory_attributes(struct kvm *kvm, > + struct kvm_gfn_range *range) > +{ > + /* > + * Zap SPTEs even if the slot can't be mapped PRIVATE. KVM x86 o= nly > + * supports KVM_MEMORY_ATTRIBUTE_PRIVATE, and so it *seems* like = KVM > + * can simply ignore such slots. But if userspace is making memo= ry > + * PRIVATE, then KVM must prevent the guest from accessing the me= mory > + * as shared. And if userspace is making memory SHARED and this = point > + * is reached, then at least one page within the range was previo= usly > + * PRIVATE, i.e. the slot's possible hugepage ranges are changing= . > + * Zapping SPTEs in this case ensures KVM will reassess whether o= r not > + * a hugepage can be used for affected ranges. > + */ > + if (WARN_ON_ONCE(!kvm_arch_has_private_mem(kvm))) > + return false; > + > + return kvm_unmap_gfn_range(kvm, range); > +} > + > static bool hugepage_test_mixed(struct kvm_memory_slot *slot, gfn_t gfn, > int level) > { > diff --git a/arch/x86/kvm/mmu/mmu_internal.h b/arch/x86/kvm/mmu/mmu_inter= nal.h > index decc1f153669..86c7cb692786 100644 > --- a/arch/x86/kvm/mmu/mmu_internal.h > +++ b/arch/x86/kvm/mmu/mmu_internal.h > @@ -201,6 +201,7 @@ struct kvm_page_fault { > > /* Derived from mmu and global state. */ > const bool is_tdp; > + const bool is_private; > const bool nx_huge_page_workaround_enabled; > > /* > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > index 7de93858054d..e3223cafd7db 100644 > --- a/include/linux/kvm_host.h > +++ b/include/linux/kvm_host.h > @@ -2358,14 +2358,18 @@ static inline void kvm_account_pgtable_pages(void= *virt, int nr) > #define KVM_DIRTY_RING_MAX_ENTRIES 65536 > > static inline void kvm_prepare_memory_fault_exit(struct kvm_vcpu *vcpu, > - gpa_t gpa, gpa_t size) > + gpa_t gpa, gpa_t size, > + bool is_write, bool is_e= xec, > + bool is_private) > { > vcpu->run->exit_reason =3D KVM_EXIT_MEMORY_FAULT; > vcpu->run->memory_fault.gpa =3D gpa; > vcpu->run->memory_fault.size =3D size; > > - /* Flags are not (yet) defined or communicated to userspace. */ > + /* RWX flags are not (yet) defined or communicated to userspace. = */ > vcpu->run->memory_fault.flags =3D 0; > + if (is_private) > + vcpu->run->memory_fault.flags |=3D KVM_MEMORY_EXIT_FLAG_P= RIVATE; > } > > #ifdef CONFIG_KVM_GENERIC_MEMORY_ATTRIBUTES > diff --git a/include/uapi/linux/kvm.h b/include/uapi/linux/kvm.h > index 33d542de0a61..29e9eb51dec9 100644 > --- a/include/uapi/linux/kvm.h > +++ b/include/uapi/linux/kvm.h > @@ -527,6 +527,7 @@ struct kvm_run { > } notify; > /* KVM_EXIT_MEMORY_FAULT */ > struct { > +#define KVM_MEMORY_EXIT_FLAG_PRIVATE (1ULL << 3) > __u64 flags; > __u64 gpa; > __u64 size; > -- > 2.42.0.820.g83a721a137-goog >