Received: by 2002:a05:7412:b795:b0:e2:908c:2ebd with SMTP id iv21csp433658rdb; Thu, 2 Nov 2023 07:53:38 -0700 (PDT) X-Google-Smtp-Source: AGHT+IG/E/Nc5afHWCzHxt2380RxRQjnJAZW1Te8BHWntzGq99hqJnr4fpFW9VoKCNTBauaGeij1 X-Received: by 2002:a05:6a00:24c8:b0:6bd:3ba8:e60e with SMTP id d8-20020a056a0024c800b006bd3ba8e60emr18194744pfv.22.1698936818530; Thu, 02 Nov 2023 07:53:38 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1698936818; cv=none; d=google.com; s=arc-20160816; b=NleW5NjNQSbQ+zzcAZGECEnl+nQl92JWVvoaDCbtetQseayKo/UkI48gvF6n7NyAZa FklEgI46NCNBpSI2IOSGAyQoqiJqsHKAX9NQeyyyNPbUearFGys8NsGGD1T0AV40bbsc mAvgTG3CjtbGfAinsYFrYVAm13bQ46hdFyWT2CXAvfMBEIX0YDHuoqj9k8QYvsLCKm5a eHLyIoK2Bfd9yaV9usG3xSkUOzAiGPlT4k+Y9mdFImlsBMqE4uW+4YUucYi98HcOzWfQ DfQx+2b6kNzGNLiFieY3eZbOCFW3v5Dy4zlMP+pMZH6jcGSWOaAWF+UQ62193GjorIfx Fq9w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:cc:to:subject :message-id:date:from:in-reply-to:references:mime-version :dkim-signature; bh=8GvzXOE6ewh/7Rc6C9DonV3a0Npx2I7bSl14yjDuvxo=; fh=xK+wiX2TTq7O2X4TlNj4MXVc7h6ma8eyuy1Ur/RMLHE=; b=mbEO/TCQhZYyM5Y5NPog3knH2e0rGCHUBUEpwLZOi1JjZiFRpWd2fMj9DLcIGlurAH w2iHO6wqQW1SwbAeav7fWIZeYLg7b7CZ/h3wUjCFfCHxkHE27DVIOjTvwX+2GucENC5B 43e+Vj74RLFJn6/oAOv74EGmwVZuhzIhrySgsnSTwHZuP4/7Orpp+oxfTpffMqAOX5RH nn8CfhFSbxXVqZ+251vsbsn4GpLVpK4Wo2CUMarV9tbuV+osTAUkiyY2tZ/bl5ymAyEF 5ihG9063vFyIU7xLFX1BHOIdx0CKNL9fUEj1M5AHJU2uHcqMnwPh6Z00cYJpkDWCTVdm 13HQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20230601 header.b=wbxKDP5C; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.35 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from groat.vger.email (groat.vger.email. [23.128.96.35]) by mx.google.com with ESMTPS id o19-20020a056a0015d300b006c34752a6e8si306747pfu.81.2023.11.02.07.53.38 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 02 Nov 2023 07:53:38 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.35 as permitted sender) client-ip=23.128.96.35; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20230601 header.b=wbxKDP5C; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.35 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by groat.vger.email (Postfix) with ESMTP id 9B4388022C54; Thu, 2 Nov 2023 07:53:32 -0700 (PDT) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.10 at groat.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232145AbjKBOxN (ORCPT + 99 others); Thu, 2 Nov 2023 10:53:13 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:34308 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229596AbjKBOxM (ORCPT ); Thu, 2 Nov 2023 10:53:12 -0400 Received: from mail-qv1-xf2f.google.com (mail-qv1-xf2f.google.com [IPv6:2607:f8b0:4864:20::f2f]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id DA0B1136 for ; Thu, 2 Nov 2023 07:53:08 -0700 (PDT) Received: by mail-qv1-xf2f.google.com with SMTP id 6a1803df08f44-67089696545so5929966d6.0 for ; Thu, 02 Nov 2023 07:53:08 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20230601; t=1698936788; x=1699541588; darn=vger.kernel.org; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:from:to:cc:subject:date :message-id:reply-to; bh=8GvzXOE6ewh/7Rc6C9DonV3a0Npx2I7bSl14yjDuvxo=; b=wbxKDP5Ci6ITuNqXJaZc/4v+hCB16rOOImLWZg8H8jfbX+IBe85Gd91QiZPJ0hnJ2J ikAVQZGqmh64TlpfaJdzfka9dRtP3Pbl4SG+FX7GgEp02ngMhfwly0nwGIWgwbek0hlO I0waYrtN/tQevOHtqJ3y2kLdgpDS/CM9OlSJx3aIz3J6ChXz9ZNSkDMIYZPV1jTjb45v d6MvH5/5K7PXnR/GndyxzbC9sK5NJkgJhnI1kmcR1YT6N6UPXjvvXuWo/aBt04PtPhOJ 1z+XKpE6KANKbN/d3FFnmRF4/KMz36gSwbZ3AzLONCCCeGqie1+SlDdpI9QKf2fCuozL KuwQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1698936788; x=1699541588; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=8GvzXOE6ewh/7Rc6C9DonV3a0Npx2I7bSl14yjDuvxo=; b=EOM105m5RUJJKcSIEqfWzmZFEMp9UlmOa8vVFuBsqiDR1GRMAkcpHjV3kfn0S1mOww Cm3JfDXTSx0rjGbVlKPbf2rysSPE7d9b/u26vFbvXHR/4yNSXJQ4iuCVENOFmcxDMPl6 75sTTwNmkm+ROGrMXcD4lpTsmQxqEuTU9u3HVTFZ10zfd6uLZkiGRIgdxvAXRaygLd+d 2KDXSUQ6Ets559i0s+TMa2I87Yo0RR1Zn4c+Vt+SXsmm1VXxfjtbdj9jpCippS2DLAam f4Zq6yEo8cxq2P3SlhZcM0KD28Sigx4405wlosjcLwK5v3L7Axp0SdFASYUJH8WQRzEa Gq5A== X-Gm-Message-State: AOJu0Yy91ob9mOxQzl1gaagNTguNLbJD0WTvC38L1S71z5Wwc5/4QnjZ nROv0+HXk7yEiSdyzU1CT4pZWQXMcVwPUIqC7elJSQ== X-Received: by 2002:a05:6214:404:b0:66d:130c:bb9d with SMTP id z4-20020a056214040400b0066d130cbb9dmr25099306qvx.13.1698936787681; Thu, 02 Nov 2023 07:53:07 -0700 (PDT) MIME-Version: 1.0 References: <20231027182217.3615211-1-seanjc@google.com> <20231027182217.3615211-23-seanjc@google.com> In-Reply-To: <20231027182217.3615211-23-seanjc@google.com> From: Fuad Tabba Date: Thu, 2 Nov 2023 14:52:31 +0000 Message-ID: Subject: Re: [PATCH v13 22/35] KVM: Allow arch code to track number of memslot address spaces per VM To: Sean Christopherson Cc: Paolo Bonzini , Marc Zyngier , Oliver Upton , Huacai Chen , Michael Ellerman , Anup Patel , Paul Walmsley , Palmer Dabbelt , Albert Ou , Alexander Viro , Christian Brauner , "Matthew Wilcox (Oracle)" , Andrew Morton , kvm@vger.kernel.org, linux-arm-kernel@lists.infradead.org, kvmarm@lists.linux.dev, linux-mips@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, kvm-riscv@lists.infradead.org, linux-riscv@lists.infradead.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org, Xiaoyao Li , Xu Yilun , Chao Peng , Jarkko Sakkinen , Anish Moorthy , David Matlack , Yu Zhang , Isaku Yamahata , =?UTF-8?B?TWlja2HDq2wgU2FsYcO8bg==?= , Vlastimil Babka , Vishal Annapurve , Ackerley Tng , Maciej Szmigiero , David Hildenbrand , Quentin Perret , Michael Roth , Wang , Liam Merwick , Isaku Yamahata , "Kirill A . Shutemov" Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable X-Spam-Status: No, score=-8.4 required=5.0 tests=DKIMWL_WL_MED,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE, USER_IN_DEF_DKIM_WL autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on groat.vger.email Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (groat.vger.email [0.0.0.0]); Thu, 02 Nov 2023 07:53:32 -0700 (PDT) On Fri, Oct 27, 2023 at 7:23=E2=80=AFPM Sean Christopherson wrote: > > Let x86 track the number of address spaces on a per-VM basis so that KVM > can disallow SMM memslots for confidential VMs. Confidentials VMs are > fundamentally incompatible with emulating SMM, which as the name suggests > requires being able to read and write guest memory and register state. > > Disallowing SMM will simplify support for guest private memory, as KVM > will not need to worry about tracking memory attributes for multiple > address spaces (SMM is the only "non-default" address space across all > architectures). > > Signed-off-by: Sean Christopherson > --- Reviewed-by: Fuad Tabba Tested-by: Fuad Tabba Cheers, /fuad > arch/powerpc/kvm/book3s_hv.c | 2 +- > arch/x86/include/asm/kvm_host.h | 8 +++++++- > arch/x86/kvm/debugfs.c | 2 +- > arch/x86/kvm/mmu/mmu.c | 6 +++--- > arch/x86/kvm/x86.c | 2 +- > include/linux/kvm_host.h | 17 +++++++++++------ > virt/kvm/dirty_ring.c | 2 +- > virt/kvm/kvm_main.c | 26 ++++++++++++++------------ > 8 files changed, 39 insertions(+), 26 deletions(-) > > diff --git a/arch/powerpc/kvm/book3s_hv.c b/arch/powerpc/kvm/book3s_hv.c > index 130bafdb1430..9b0eaa17275a 100644 > --- a/arch/powerpc/kvm/book3s_hv.c > +++ b/arch/powerpc/kvm/book3s_hv.c > @@ -6084,7 +6084,7 @@ static int kvmhv_svm_off(struct kvm *kvm) > } > > srcu_idx =3D srcu_read_lock(&kvm->srcu); > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > struct kvm_memory_slot *memslot; > struct kvm_memslots *slots =3D __kvm_memslots(kvm, i); > int bkt; > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_h= ost.h > index 6702f795c862..f9e8d5642069 100644 > --- a/arch/x86/include/asm/kvm_host.h > +++ b/arch/x86/include/asm/kvm_host.h > @@ -2124,9 +2124,15 @@ enum { > #define HF_SMM_MASK (1 << 1) > #define HF_SMM_INSIDE_NMI_MASK (1 << 2) > > -# define KVM_ADDRESS_SPACE_NUM 2 > +# define KVM_MAX_NR_ADDRESS_SPACES 2 > # define kvm_arch_vcpu_memslots_id(vcpu) ((vcpu)->arch.hflags & HF_SMM_M= ASK ? 1 : 0) > # define kvm_memslots_for_spte_role(kvm, role) __kvm_memslots(kvm, (role= ).smm) > + > +static inline int kvm_arch_nr_memslot_as_ids(struct kvm *kvm) > +{ > + return KVM_MAX_NR_ADDRESS_SPACES; > +} > + > #else > # define kvm_memslots_for_spte_role(kvm, role) __kvm_memslots(kvm, 0) > #endif > diff --git a/arch/x86/kvm/debugfs.c b/arch/x86/kvm/debugfs.c > index ee8c4c3496ed..42026b3f3ff3 100644 > --- a/arch/x86/kvm/debugfs.c > +++ b/arch/x86/kvm/debugfs.c > @@ -111,7 +111,7 @@ static int kvm_mmu_rmaps_stat_show(struct seq_file *m= , void *v) > mutex_lock(&kvm->slots_lock); > write_lock(&kvm->mmu_lock); > > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > int bkt; > > slots =3D __kvm_memslots(kvm, i); > diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c > index c4e758f0aebb..baeba8fc1c38 100644 > --- a/arch/x86/kvm/mmu/mmu.c > +++ b/arch/x86/kvm/mmu/mmu.c > @@ -3755,7 +3755,7 @@ static int mmu_first_shadow_root_alloc(struct kvm *= kvm) > kvm_page_track_write_tracking_enabled(kvm)) > goto out_success; > > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > slots =3D __kvm_memslots(kvm, i); > kvm_for_each_memslot(slot, bkt, slots) { > /* > @@ -6294,7 +6294,7 @@ static bool kvm_rmap_zap_gfn_range(struct kvm *kvm,= gfn_t gfn_start, gfn_t gfn_e > if (!kvm_memslots_have_rmaps(kvm)) > return flush; > > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > slots =3D __kvm_memslots(kvm, i); > > kvm_for_each_memslot_in_gfn_range(&iter, slots, gfn_start= , gfn_end) { > @@ -6791,7 +6791,7 @@ void kvm_mmu_invalidate_mmio_sptes(struct kvm *kvm,= u64 gen) > * modifier prior to checking for a wrap of the MMIO generation s= o > * that a wrap in any address space is detected. > */ > - gen &=3D ~((u64)KVM_ADDRESS_SPACE_NUM - 1); > + gen &=3D ~((u64)kvm_arch_nr_memslot_as_ids(kvm) - 1); > > /* > * The very rare case: if the MMIO generation number has wrapped, > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > index 824b58b44382..c4d17727b199 100644 > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -12456,7 +12456,7 @@ void __user * __x86_set_memory_region(struct kvm = *kvm, int id, gpa_t gpa, > hva =3D slot->userspace_addr; > } > > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > struct kvm_userspace_memory_region2 m; > > m.slot =3D id | (i << 16); > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > index c3cfe08b1300..687589ce9f63 100644 > --- a/include/linux/kvm_host.h > +++ b/include/linux/kvm_host.h > @@ -80,8 +80,8 @@ > /* Two fragments for cross MMIO pages. */ > #define KVM_MAX_MMIO_FRAGMENTS 2 > > -#ifndef KVM_ADDRESS_SPACE_NUM > -#define KVM_ADDRESS_SPACE_NUM 1 > +#ifndef KVM_MAX_NR_ADDRESS_SPACES > +#define KVM_MAX_NR_ADDRESS_SPACES 1 > #endif > > /* > @@ -692,7 +692,12 @@ bool kvm_arch_irqchip_in_kernel(struct kvm *kvm); > #define KVM_MEM_SLOTS_NUM SHRT_MAX > #define KVM_USER_MEM_SLOTS (KVM_MEM_SLOTS_NUM - KVM_INTERNAL_MEM_SLOTS) > > -#if KVM_ADDRESS_SPACE_NUM =3D=3D 1 > +#if KVM_MAX_NR_ADDRESS_SPACES =3D=3D 1 > +static inline int kvm_arch_nr_memslot_as_ids(struct kvm *kvm) > +{ > + return KVM_MAX_NR_ADDRESS_SPACES; > +} > + > static inline int kvm_arch_vcpu_memslots_id(struct kvm_vcpu *vcpu) > { > return 0; > @@ -747,9 +752,9 @@ struct kvm { > struct mm_struct *mm; /* userspace tied to this vm */ > unsigned long nr_memslot_pages; > /* The two memslot sets - active and inactive (per address space)= */ > - struct kvm_memslots __memslots[KVM_ADDRESS_SPACE_NUM][2]; > + struct kvm_memslots __memslots[KVM_MAX_NR_ADDRESS_SPACES][2]; > /* The current active memslot set for each address space */ > - struct kvm_memslots __rcu *memslots[KVM_ADDRESS_SPACE_NUM]; > + struct kvm_memslots __rcu *memslots[KVM_MAX_NR_ADDRESS_SPACES]; > struct xarray vcpu_array; > /* > * Protected by slots_lock, but can be read outside if an > @@ -1018,7 +1023,7 @@ void kvm_put_kvm_no_destroy(struct kvm *kvm); > > static inline struct kvm_memslots *__kvm_memslots(struct kvm *kvm, int a= s_id) > { > - as_id =3D array_index_nospec(as_id, KVM_ADDRESS_SPACE_NUM); > + as_id =3D array_index_nospec(as_id, KVM_MAX_NR_ADDRESS_SPACES); > return srcu_dereference_check(kvm->memslots[as_id], &kvm->srcu, > lockdep_is_held(&kvm->slots_lock) || > !refcount_read(&kvm->users_count)); > diff --git a/virt/kvm/dirty_ring.c b/virt/kvm/dirty_ring.c > index c1cd7dfe4a90..86d267db87bb 100644 > --- a/virt/kvm/dirty_ring.c > +++ b/virt/kvm/dirty_ring.c > @@ -58,7 +58,7 @@ static void kvm_reset_dirty_gfn(struct kvm *kvm, u32 sl= ot, u64 offset, u64 mask) > as_id =3D slot >> 16; > id =3D (u16)slot; > > - if (as_id >=3D KVM_ADDRESS_SPACE_NUM || id >=3D KVM_USER_MEM_SLOT= S) > + if (as_id >=3D kvm_arch_nr_memslot_as_ids(kvm) || id >=3D KVM_USE= R_MEM_SLOTS) > return; > > memslot =3D id_to_memslot(__kvm_memslots(kvm, as_id), id); > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > index 5d1a2f1b4e94..23633984142f 100644 > --- a/virt/kvm/kvm_main.c > +++ b/virt/kvm/kvm_main.c > @@ -615,7 +615,7 @@ static __always_inline kvm_mn_ret_t __kvm_handle_hva_= range(struct kvm *kvm, > > idx =3D srcu_read_lock(&kvm->srcu); > > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > struct interval_tree_node *node; > > slots =3D __kvm_memslots(kvm, i); > @@ -1248,7 +1248,7 @@ static struct kvm *kvm_create_vm(unsigned long type= , const char *fdname) > goto out_err_no_irq_srcu; > > refcount_set(&kvm->users_count, 1); > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > for (j =3D 0; j < 2; j++) { > slots =3D &kvm->__memslots[i][j]; > > @@ -1398,7 +1398,7 @@ static void kvm_destroy_vm(struct kvm *kvm) > #endif > kvm_arch_destroy_vm(kvm); > kvm_destroy_devices(kvm); > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > kvm_free_memslots(kvm, &kvm->__memslots[i][0]); > kvm_free_memslots(kvm, &kvm->__memslots[i][1]); > } > @@ -1681,7 +1681,7 @@ static void kvm_swap_active_memslots(struct kvm *kv= m, int as_id) > * space 0 will use generations 0, 2, 4, ... while address space = 1 will > * use generations 1, 3, 5, ... > */ > - gen +=3D KVM_ADDRESS_SPACE_NUM; > + gen +=3D kvm_arch_nr_memslot_as_ids(kvm); > > kvm_arch_memslots_updated(kvm, gen); > > @@ -2051,7 +2051,7 @@ int __kvm_set_memory_region(struct kvm *kvm, > (mem->guest_memfd_offset & (PAGE_SIZE - 1) || > mem->guest_memfd_offset + mem->memory_size < mem->guest_memf= d_offset)) > return -EINVAL; > - if (as_id >=3D KVM_ADDRESS_SPACE_NUM || id >=3D KVM_MEM_SLOTS_NUM= ) > + if (as_id >=3D kvm_arch_nr_memslot_as_ids(kvm) || id >=3D KVM_MEM= _SLOTS_NUM) > return -EINVAL; > if (mem->guest_phys_addr + mem->memory_size < mem->guest_phys_add= r) > return -EINVAL; > @@ -2187,7 +2187,7 @@ int kvm_get_dirty_log(struct kvm *kvm, struct kvm_d= irty_log *log, > > as_id =3D log->slot >> 16; > id =3D (u16)log->slot; > - if (as_id >=3D KVM_ADDRESS_SPACE_NUM || id >=3D KVM_USER_MEM_SLOT= S) > + if (as_id >=3D kvm_arch_nr_memslot_as_ids(kvm) || id >=3D KVM_USE= R_MEM_SLOTS) > return -EINVAL; > > slots =3D __kvm_memslots(kvm, as_id); > @@ -2249,7 +2249,7 @@ static int kvm_get_dirty_log_protect(struct kvm *kv= m, struct kvm_dirty_log *log) > > as_id =3D log->slot >> 16; > id =3D (u16)log->slot; > - if (as_id >=3D KVM_ADDRESS_SPACE_NUM || id >=3D KVM_USER_MEM_SLOT= S) > + if (as_id >=3D kvm_arch_nr_memslot_as_ids(kvm) || id >=3D KVM_USE= R_MEM_SLOTS) > return -EINVAL; > > slots =3D __kvm_memslots(kvm, as_id); > @@ -2361,7 +2361,7 @@ static int kvm_clear_dirty_log_protect(struct kvm *= kvm, > > as_id =3D log->slot >> 16; > id =3D (u16)log->slot; > - if (as_id >=3D KVM_ADDRESS_SPACE_NUM || id >=3D KVM_USER_MEM_SLOT= S) > + if (as_id >=3D kvm_arch_nr_memslot_as_ids(kvm) || id >=3D KVM_USE= R_MEM_SLOTS) > return -EINVAL; > > if (log->first_page & 63) > @@ -2502,7 +2502,7 @@ static __always_inline void kvm_handle_gfn_range(st= ruct kvm *kvm, > gfn_range.only_private =3D false; > gfn_range.only_shared =3D false; > > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > slots =3D __kvm_memslots(kvm, i); > > kvm_for_each_memslot_in_gfn_range(&iter, slots, range->st= art, range->end) { > @@ -4857,9 +4857,11 @@ static int kvm_vm_ioctl_check_extension_generic(st= ruct kvm *kvm, long arg) > case KVM_CAP_IRQ_ROUTING: > return KVM_MAX_IRQ_ROUTES; > #endif > -#if KVM_ADDRESS_SPACE_NUM > 1 > +#if KVM_MAX_NR_ADDRESS_SPACES > 1 > case KVM_CAP_MULTI_ADDRESS_SPACE: > - return KVM_ADDRESS_SPACE_NUM; > + if (kvm) > + return kvm_arch_nr_memslot_as_ids(kvm); > + return KVM_MAX_NR_ADDRESS_SPACES; > #endif > case KVM_CAP_NR_MEMSLOTS: > return KVM_USER_MEM_SLOTS; > @@ -4967,7 +4969,7 @@ bool kvm_are_all_memslots_empty(struct kvm *kvm) > > lockdep_assert_held(&kvm->slots_lock); > > - for (i =3D 0; i < KVM_ADDRESS_SPACE_NUM; i++) { > + for (i =3D 0; i < kvm_arch_nr_memslot_as_ids(kvm); i++) { > if (!kvm_memslots_empty(__kvm_memslots(kvm, i))) > return false; > } > -- > 2.42.0.820.g83a721a137-goog >