Received: by 2002:a05:6358:3188:b0:123:57c1:9b43 with SMTP id q8csp2107307rwd; Fri, 9 Jun 2023 06:50:45 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ5p3JDD8+0UmKugL+jmcj6bxRJszzLxcrHBJmPzu7cfWWFawNDNtT6fXLqEAsQ4/4VUZRWK X-Received: by 2002:a05:6a00:1a89:b0:647:7ee8:6251 with SMTP id e9-20020a056a001a8900b006477ee86251mr1550380pfv.21.1686318645490; Fri, 09 Jun 2023 06:50:45 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1686318645; cv=none; d=google.com; s=arc-20160816; b=rikAFFEMz7w+2lv0iGXwt3ugVFl4ReWOztTRlbI1pETyxegCpPxctS8nL0rUy7iC4t VhD67FpNcfq+cIgYZgrGMFO+PewKM5QefDBn34h04lzdqZHUfBhrJTAmrssTZ5hC+SdJ QQ69RwUnmlQ03AxS0ApxjCvQdvFh5f02Duf67DmkXD/RiY84F22gJUaJZpgRLbzlAug0 U7+gjGJK3TVR4gI+0sS2yoB+2XSmaAC2W4QONTSc9WpmrZfb7yMStSOAVQCHPydXdDtE gTzxzF9gEV0IFJox5C0EgatQgZmlLqBv/4VRAanpUNiq8Hz6e9xhZ9QcArFVkRDLDk/2 oGRA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=L6t0TvZfDLoqeIOaTnZ9oEVfgP8U8M4IOz1T+NhsED0=; b=cS4AQcwGUEPO9EfV4XvnWPjBhY34l1+l1ra2zDcxBxpdTTIJGRtNU8s4OkEHFM4//Y XnTYokkXSlhjv6PePxBjtsPl3CZthDLL3upQMBkNekBVbRPNwlGSpLSuVm63PqQlKf3Y GQ6XtHJ4dQY08V5Kha3CHLjKiqYJ+A4JxhOerKz6vCVUS7yQOU7z0Upgjh6NMkBWSxTf eWTdgpwd+j9+vQJgo27TOD3LlJavbA4f5Pvp3dy1+LuR68Y0ChJ2Vq0u5louR0H7uJPN EOMIzuZkIbZxEjyKLEFzwdTqUbzjrtGP9ResTonYOnr/NVrx9tuz9L8ksFtW2EB9J0Sc l4fA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@ventanamicro.com header.s=google header.b=UDjeaXrU; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id v5-20020a637a05000000b0050309caf60esi2656565pgc.750.2023.06.09.06.50.31; Fri, 09 Jun 2023 06:50:45 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@ventanamicro.com header.s=google header.b=UDjeaXrU; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S237997AbjFINfu (ORCPT + 99 others); Fri, 9 Jun 2023 09:35:50 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:50810 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S231134AbjFINft (ORCPT ); Fri, 9 Jun 2023 09:35:49 -0400 Received: from mail-wm1-x32b.google.com (mail-wm1-x32b.google.com [IPv6:2a00:1450:4864:20::32b]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D31462D48 for ; Fri, 9 Jun 2023 06:35:43 -0700 (PDT) Received: by mail-wm1-x32b.google.com with SMTP id 5b1f17b1804b1-3f6dfc4dffaso13666075e9.0 for ; Fri, 09 Jun 2023 06:35:43 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ventanamicro.com; s=google; t=1686317742; x=1688909742; h=in-reply-to:content-disposition:mime-version:references:message-id :subject:cc:to:from:date:from:to:cc:subject:date:message-id:reply-to; bh=L6t0TvZfDLoqeIOaTnZ9oEVfgP8U8M4IOz1T+NhsED0=; b=UDjeaXrUquBlqG/iLwmk4H6WGbSpSDsoayFDZ4iD1+lhgaPzF/60UM9fxIsG9Zckh0 BfNulAZmBlKAxgU54ih5JLw3TKzrnwRsys6BsenHiX+OIT+UDtLtROmC1h6pH6cf3jRd Wt0Bz1CQQocRc9xrBg4tEh6wQfRPc1iFV3u1YiWS90y/ys2GXHktyQS3q8UHh8sCtYjJ wTMFI8r3mkQEM7pGGnQJnf5bPODBsBUB6Vwi5WdX40OubcQa9nKsoDNzYamcIykhd5EA G1rLzFJgZt2mpKNy/cJukniuhovE38Y9WrBqInN7ZVXSFaqbuItyzuLHckRe9qiaaIEH T7FA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20221208; t=1686317742; x=1688909742; h=in-reply-to:content-disposition:mime-version:references:message-id :subject:cc:to:from:date:x-gm-message-state:from:to:cc:subject:date :message-id:reply-to; bh=L6t0TvZfDLoqeIOaTnZ9oEVfgP8U8M4IOz1T+NhsED0=; b=iYO5b1dNyy4UaMBSWguZFg1wKOPHro3bf3ikonYXd/eNZ1L8Rl0swYVIhpyznV63/5 qmS0mcJ4b0j3vRUXBEndU6gVIRLz61PFISVV8wnIbD4s4OP3V/zUq6Vq2y0720DTZQYk OjMfeSdUs0NViJ2s6HOAZtlUJI9Kid4rGlQIfiJ44JZ+49tthUZ5C/rDvr6Mrc00Hjy9 CErH2YkrbH3jzwkpD4W/LB6k0Mn8Qpk0y2AbuwQFy4SpOYKnUtk3TS/nabeqRW36c/I/ l0OKZ54/+s4yG6yHyS10dQYOR8k5p5C4Vj/gpwP5Yzs+QJINOeYNIXjCJaJ9HtNZo40N YIYA== X-Gm-Message-State: AC+VfDx7kM8/raHQanfRknwS1ZvucomN/t1IGdVwd+A8D0h+8zdzLGye aJwgF0bDUATS78gqjnNeJ2MiTg== X-Received: by 2002:a7b:cd86:0:b0:3f6:be1:b8d6 with SMTP id y6-20020a7bcd86000000b003f60be1b8d6mr963244wmj.9.1686317742176; Fri, 09 Jun 2023 06:35:42 -0700 (PDT) Received: from localhost (2001-1ae9-1c2-4c00-20f-c6b4-1e57-7965.ip6.tmcz.cz. [2001:1ae9:1c2:4c00:20f:c6b4:1e57:7965]) by smtp.gmail.com with ESMTPSA id u9-20020a05600c00c900b003f72468833esm2736963wmm.26.2023.06.09.06.35.41 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 09 Jun 2023 06:35:41 -0700 (PDT) Date: Fri, 9 Jun 2023 15:35:40 +0200 From: Andrew Jones To: Haibo Xu Cc: xiaobo55x@gmail.com, maz@kernel.org, oliver.upton@linux.dev, seanjc@google.com, Paolo Bonzini , Jonathan Corbet , Anup Patel , Atish Patra , Paul Walmsley , Palmer Dabbelt , Albert Ou , Shuah Khan , James Morse , Suzuki K Poulose , Zenghui Yu , Ben Gardon , David Matlack , Vipin Sharma , Colton Lewis , kvm@vger.kernel.org, linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org, kvm-riscv@lists.infradead.org, linux-riscv@lists.infradead.org, linux-kselftest@vger.kernel.org, linux-arm-kernel@lists.infradead.org, kvmarm@lists.linux.dev Subject: Re: [PATCH v3 10/10] KVM: riscv: selftests: Add get-reg-list test Message-ID: <20230609-fba04b424a4d46574e04e587@orel> References: <8cd4ce50f5f4a639f4508085959aae222d4d4386.1686275310.git.haibo1.xu@intel.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <8cd4ce50f5f4a639f4508085959aae222d4d4386.1686275310.git.haibo1.xu@intel.com> X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE,UPPERCASE_50_75, URIBL_BLOCKED autolearn=no autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Jun 09, 2023 at 10:12:18AM +0800, Haibo Xu wrote: > get-reg-list test is used to check for KVM registers regressions > during VM migration which happens when destination host kernel > missing registers that the source host kernel has. The blessed > list registers was created by running on v6.4-rc5. > > Signed-off-by: Haibo Xu > --- > tools/testing/selftests/kvm/Makefile | 1 + > tools/testing/selftests/kvm/get-reg-list.c | 28 + > .../selftests/kvm/include/riscv/processor.h | 3 + > .../selftests/kvm/riscv/get-reg-list.c | 611 ++++++++++++++++++ > 4 files changed, 643 insertions(+) > create mode 100644 tools/testing/selftests/kvm/riscv/get-reg-list.c > > diff --git a/tools/testing/selftests/kvm/Makefile b/tools/testing/selftests/kvm/Makefile > index d90cad19c9ee..f7bcda903dd9 100644 > --- a/tools/testing/selftests/kvm/Makefile > +++ b/tools/testing/selftests/kvm/Makefile > @@ -174,6 +174,7 @@ TEST_GEN_PROGS_s390x += kvm_binary_stats_test > > TEST_GEN_PROGS_riscv += demand_paging_test > TEST_GEN_PROGS_riscv += dirty_log_test > +TEST_GEN_PROGS_riscv += get-reg-list > TEST_GEN_PROGS_riscv += kvm_create_max_vcpus > TEST_GEN_PROGS_riscv += kvm_page_table_test > TEST_GEN_PROGS_riscv += set_memory_region_test > diff --git a/tools/testing/selftests/kvm/get-reg-list.c b/tools/testing/selftests/kvm/get-reg-list.c > index abacb95c21c6..73f40e0842b8 100644 > --- a/tools/testing/selftests/kvm/get-reg-list.c > +++ b/tools/testing/selftests/kvm/get-reg-list.c > @@ -133,6 +133,34 @@ static struct kvm_vcpu *vcpu_config_get_vcpu(struct vcpu_reg_list *c, struct kvm > return vcpu; > } > #else > +static inline bool vcpu_has_ext(struct kvm_vcpu *vcpu, int ext) > +{ > + int ret; > + unsigned long value; > + > + ret = __vcpu_get_reg(vcpu, RISCV_ISA_EXT_REG(ext), &value); > + if (ret) { > + printf("Failed to get ext %d", ext); > + return false; > + } > + > + return !!value; > +} > + > +static void finalize_vcpu(struct kvm_vcpu *vcpu, struct vcpu_reg_list *c) > +{ > + struct vcpu_reg_sublist *s; > + > + for_each_sublist(c, s) { > + if (!s->feature) > + continue; Using zero to mean "not specified" means we can't test for KVM_RISCV_ISA_EXT_A, but that's probably OK, since Linux always has 'a', so we'll never need to check for it. > + > + __TEST_REQUIRE(vcpu_has_ext(vcpu, s->feature), > + "%s: %s not available, skipping tests\n", > + config_name(c), s->name); > + } > +} > + > static struct kvm_vcpu *vcpu_config_get_vcpu(struct vcpu_reg_list *c, struct kvm_vm *vm) > { > return __vm_vcpu_add(vm, 0); > diff --git a/tools/testing/selftests/kvm/include/riscv/processor.h b/tools/testing/selftests/kvm/include/riscv/processor.h > index d00d213c3805..5b62a3d2aa9b 100644 > --- a/tools/testing/selftests/kvm/include/riscv/processor.h > +++ b/tools/testing/selftests/kvm/include/riscv/processor.h > @@ -38,6 +38,9 @@ static inline uint64_t __kvm_reg_id(uint64_t type, uint64_t idx, > KVM_REG_RISCV_TIMER_REG(name), \ > KVM_REG_SIZE_U64) > > +#define RISCV_ISA_EXT_REG(idx) __kvm_reg_id(KVM_REG_RISCV_ISA_EXT, \ > + idx, KVM_REG_SIZE_ULONG) > + > /* L3 index Bit[47:39] */ > #define PGTBL_L3_INDEX_MASK 0x0000FF8000000000ULL > #define PGTBL_L3_INDEX_SHIFT 39 > diff --git a/tools/testing/selftests/kvm/riscv/get-reg-list.c b/tools/testing/selftests/kvm/riscv/get-reg-list.c > new file mode 100644 > index 000000000000..0f371d99d471 > --- /dev/null > +++ b/tools/testing/selftests/kvm/riscv/get-reg-list.c > @@ -0,0 +1,611 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Check for KVM_GET_REG_LIST regressions. > + * > + * Copyright (C) 2020, Red Hat, Inc. I don't think we need the Red Hat copyright. This is a completely new work. > + * Copyright (c) 2023 Intel Corporation > + * > + */ > +#include > +#include "kvm_util.h" > +#include "test_util.h" > +#include "processor.h" > + > +#define REG_MASK (KVM_REG_ARCH_MASK | KVM_REG_SIZE_MASK) > + > +static const char *config_id_to_str(__u64 id) > +{ > + /* reg_off is the offset into struct kvm_riscv_config */ > + __u64 reg_off = id & ~(REG_MASK | KVM_REG_RISCV_CONFIG); > + > + switch (reg_off) { > + case KVM_REG_RISCV_CONFIG_REG(isa): > + return "KVM_REG_RISCV_CONFIG_REG(isa)"; > + case KVM_REG_RISCV_CONFIG_REG(zicbom_block_size): > + return "KVM_REG_RISCV_CONFIG_REG(zicbom_block_size)"; > + case KVM_REG_RISCV_CONFIG_REG(zicboz_block_size): > + return "KVM_REG_RISCV_CONFIG_REG(zicboz_block_size)"; > + case KVM_REG_RISCV_CONFIG_REG(mvendorid): > + return "KVM_REG_RISCV_CONFIG_REG(mvendorid)"; > + case KVM_REG_RISCV_CONFIG_REG(marchid): > + return "KVM_REG_RISCV_CONFIG_REG(marchid)"; > + case KVM_REG_RISCV_CONFIG_REG(mimpid): > + return "KVM_REG_RISCV_CONFIG_REG(mimpid)"; > + } > + > + /* > + * Config regs would grow regularly with new pseudo reg added, so > + * just show raw id to indicate a new pseudo config reg. > + */ > + return strdup_printf("KVM_REG_RISCV_CONFIG_REG(%lld) /* UNKNOWN */", reg_off); > +} > + > +static const char *core_id_to_str(const char *prefix, __u64 id) > +{ > + /* reg_off is the offset into struct kvm_riscv_core */ > + __u64 reg_off = id & ~(REG_MASK | KVM_REG_RISCV_CORE); > + > + switch (reg_off) { > + case KVM_REG_RISCV_CORE_REG(regs.pc): > + return "KVM_REG_RISCV_CORE_REG(regs.pc)"; > + case KVM_REG_RISCV_CORE_REG(regs.ra): > + return "KVM_REG_RISCV_CORE_REG(regs.ra)"; > + case KVM_REG_RISCV_CORE_REG(regs.sp): > + return "KVM_REG_RISCV_CORE_REG(regs.sp)"; > + case KVM_REG_RISCV_CORE_REG(regs.gp): > + return "KVM_REG_RISCV_CORE_REG(regs.gp)"; > + case KVM_REG_RISCV_CORE_REG(regs.tp): > + return "KVM_REG_RISCV_CORE_REG(regs.tp)"; > + case KVM_REG_RISCV_CORE_REG(regs.t0) ... KVM_REG_RISCV_CORE_REG(regs.t2): > + return strdup_printf("KVM_REG_RISCV_CORE_REG(regs.t%lld)", > + reg_off - KVM_REG_RISCV_CORE_REG(regs.t0)); > + case KVM_REG_RISCV_CORE_REG(regs.s0) ... KVM_REG_RISCV_CORE_REG(regs.s1): > + return strdup_printf("KVM_REG_RISCV_CORE_REG(regs.s%lld)", > + reg_off - KVM_REG_RISCV_CORE_REG(regs.s0)); > + case KVM_REG_RISCV_CORE_REG(regs.a0) ... KVM_REG_RISCV_CORE_REG(regs.a7): > + return strdup_printf("KVM_REG_RISCV_CORE_REG(regs.a%lld)", > + reg_off - KVM_REG_RISCV_CORE_REG(regs.a0)); > + case KVM_REG_RISCV_CORE_REG(regs.s2) ... KVM_REG_RISCV_CORE_REG(regs.s11): > + return strdup_printf("KVM_REG_RISCV_CORE_REG(regs.s%lld)", > + reg_off - KVM_REG_RISCV_CORE_REG(regs.s2) + 2); > + case KVM_REG_RISCV_CORE_REG(regs.t3) ... KVM_REG_RISCV_CORE_REG(regs.t6): > + return strdup_printf("KVM_REG_RISCV_CORE_REG(regs.t%lld)", > + reg_off - KVM_REG_RISCV_CORE_REG(regs.t3) + 3); > + case KVM_REG_RISCV_CORE_REG(mode): > + return "KVM_REG_RISCV_CORE_REG(mode)"; > + } > + > + TEST_FAIL("%s: Unknown core reg id: 0x%llx", prefix, id); > + return NULL; > +} > + > +#define RISCV_CSR_GENERAL(csr) \ > + "KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(" #csr ")" > +#define RISCV_CSR_AIA(csr) \ > + "KVM_REG_RISCV_CSR_AIA | KVM_REG_RISCV_CSR_REG(" #csr ")" > + > +static const char *general_csr_id_to_str(__u64 reg_off) > +{ > + /* reg_off is the offset into struct kvm_riscv_csr */ > + switch (reg_off) { > + case KVM_REG_RISCV_CSR_REG(sstatus): > + return RISCV_CSR_GENERAL(sstatus); > + case KVM_REG_RISCV_CSR_REG(sie): > + return RISCV_CSR_GENERAL(sie); > + case KVM_REG_RISCV_CSR_REG(stvec): > + return RISCV_CSR_GENERAL(stvec); > + case KVM_REG_RISCV_CSR_REG(sscratch): > + return RISCV_CSR_GENERAL(sscratch); > + case KVM_REG_RISCV_CSR_REG(sepc): > + return RISCV_CSR_GENERAL(sepc); > + case KVM_REG_RISCV_CSR_REG(scause): > + return RISCV_CSR_GENERAL(scause); > + case KVM_REG_RISCV_CSR_REG(stval): > + return RISCV_CSR_GENERAL(stval); > + case KVM_REG_RISCV_CSR_REG(sip): > + return RISCV_CSR_GENERAL(sip); > + case KVM_REG_RISCV_CSR_REG(satp): > + return RISCV_CSR_GENERAL(satp); > + case KVM_REG_RISCV_CSR_REG(scounteren): > + return RISCV_CSR_GENERAL(scounteren); > + } > + > + TEST_FAIL("Unknown general csr reg: 0x%llx", reg_off); > + return NULL; > +} > + > +static const char *aia_csr_id_to_str(__u64 reg_off) > +{ > + /* reg_off is the offset into struct kvm_riscv_aia_csr */ > + switch (reg_off) { > + case KVM_REG_RISCV_CSR_AIA_REG(siselect): > + return RISCV_CSR_AIA(siselect); > + case KVM_REG_RISCV_CSR_AIA_REG(iprio1): > + return RISCV_CSR_AIA(iprio1); > + case KVM_REG_RISCV_CSR_AIA_REG(iprio2): > + return RISCV_CSR_AIA(iprio2); > + case KVM_REG_RISCV_CSR_AIA_REG(sieh): > + return RISCV_CSR_AIA(sieh); > + case KVM_REG_RISCV_CSR_AIA_REG(siph): > + return RISCV_CSR_AIA(siph); > + case KVM_REG_RISCV_CSR_AIA_REG(iprio1h): > + return RISCV_CSR_AIA(iprio1h); > + case KVM_REG_RISCV_CSR_AIA_REG(iprio2h): > + return RISCV_CSR_AIA(iprio2h); > + } > + > + TEST_FAIL("Unknown aia csr reg: 0x%llx", reg_off); > + return NULL; > +} > + > +static const char *csr_id_to_str(const char *prefix, __u64 id) > +{ > + __u64 reg_off = id & ~(REG_MASK | KVM_REG_RISCV_CSR); > + __u64 reg_subtype = reg_off & KVM_REG_RISCV_SUBTYPE_MASK; > + > + reg_off &= ~KVM_REG_RISCV_SUBTYPE_MASK; > + > + switch (reg_subtype) { > + case KVM_REG_RISCV_CSR_GENERAL: > + return general_csr_id_to_str(reg_off); > + case KVM_REG_RISCV_CSR_AIA: > + return aia_csr_id_to_str(reg_off); > + } > + > + TEST_FAIL("%s: Unknown csr subtype: 0x%llx", prefix, reg_subtype); > + return NULL; > +} > + > +static const char *timer_id_to_str(const char *prefix, __u64 id) > +{ > + /* reg_off is the offset into struct kvm_riscv_timer */ > + __u64 reg_off = id & ~(REG_MASK | KVM_REG_RISCV_TIMER); > + > + switch (reg_off) { > + case KVM_REG_RISCV_TIMER_REG(frequency): > + return "KVM_REG_RISCV_TIMER_REG(frequency)"; > + case KVM_REG_RISCV_TIMER_REG(time): > + return "KVM_REG_RISCV_TIMER_REG(time)"; > + case KVM_REG_RISCV_TIMER_REG(compare): > + return "KVM_REG_RISCV_TIMER_REG(compare)"; > + case KVM_REG_RISCV_TIMER_REG(state): > + return "KVM_REG_RISCV_TIMER_REG(state)"; > + } > + > + TEST_FAIL("%s: Unknown timer reg id: 0x%llx", prefix, id); > + return NULL; > +} > + > +static const char *fp_f_id_to_str(const char *prefix, __u64 id) > +{ > + /* reg_off is the offset into struct __riscv_f_ext_state */ > + __u64 reg_off = id & ~(REG_MASK | KVM_REG_RISCV_FP_F); > + > + switch (reg_off) { > + case KVM_REG_RISCV_FP_F_REG(f[0]) ... > + KVM_REG_RISCV_FP_F_REG(f[31]): > + return strdup_printf("KVM_REG_RISCV_FP_F_REG(f[%lld])", reg_off); > + case KVM_REG_RISCV_FP_F_REG(fcsr): > + return "KVM_REG_RISCV_FP_F_REG(fcsr)"; > + } > + > + TEST_FAIL("%s: Unknown fp_f reg id: 0x%llx", prefix, id); > + return NULL; > +} > + > +static const char *fp_d_id_to_str(const char *prefix, __u64 id) > +{ > + /* reg_off is the offset into struct __riscv_d_ext_state */ > + __u64 reg_off = id & ~(REG_MASK | KVM_REG_RISCV_FP_D); > + > + switch (reg_off) { > + case KVM_REG_RISCV_FP_D_REG(f[0]) ... > + KVM_REG_RISCV_FP_D_REG(f[31]): > + return strdup_printf("KVM_REG_RISCV_FP_D_REG(f[%lld])", reg_off); > + case KVM_REG_RISCV_FP_D_REG(fcsr): > + return "KVM_REG_RISCV_FP_D_REG(fcsr)"; > + } > + > + TEST_FAIL("%s: Unknown fp_d reg id: 0x%llx", prefix, id); > + return NULL; > +} > + > +static const char *isa_ext_id_to_str(__u64 id) > +{ > + /* reg_off is the offset into unsigned long kvm_isa_ext_arr[] */ > + __u64 reg_off = id & ~(REG_MASK | KVM_REG_RISCV_ISA_EXT); > + > + static const char * const kvm_isa_ext_reg_name[] = { > + "KVM_RISCV_ISA_EXT_A", > + "KVM_RISCV_ISA_EXT_C", > + "KVM_RISCV_ISA_EXT_D", > + "KVM_RISCV_ISA_EXT_F", > + "KVM_RISCV_ISA_EXT_H", > + "KVM_RISCV_ISA_EXT_I", > + "KVM_RISCV_ISA_EXT_M", > + "KVM_RISCV_ISA_EXT_SVPBMT", > + "KVM_RISCV_ISA_EXT_SSTC", > + "KVM_RISCV_ISA_EXT_SVINVAL", > + "KVM_RISCV_ISA_EXT_ZIHINTPAUSE", > + "KVM_RISCV_ISA_EXT_ZICBOM", > + "KVM_RISCV_ISA_EXT_ZICBOZ", > + "KVM_RISCV_ISA_EXT_ZBB", > + "KVM_RISCV_ISA_EXT_SSAIA", > + }; > + > + if (reg_off >= ARRAY_SIZE(kvm_isa_ext_reg_name)) { > + /* > + * isa_ext regs would grow regularly with new isa extension added, so > + * just show "reg" to indicate a new extension. > + */ > + return strdup_printf("%lld /* UNKNOWN */", reg_off); > + } > + > + return kvm_isa_ext_reg_name[reg_off]; > +} > + > +static const char *sbi_ext_single_id_to_str(__u64 reg_off) > +{ > + /* reg_off is KVM_RISCV_SBI_EXT_ID */ > + static const char * const kvm_sbi_ext_reg_name[] = { > + "KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_V01", > + "KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_TIME", > + "KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_IPI", > + "KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_RFENCE", > + "KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_SRST", > + "KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_HSM", > + "KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_PMU", > + "KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_EXPERIMENTAL", > + "KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_VENDOR", > + }; > + > + if (reg_off >= ARRAY_SIZE(kvm_sbi_ext_reg_name)) { > + /* > + * sbi_ext regs would grow regularly with new sbi extension added, so > + * just show "reg" to indicate a new extension. > + */ > + return strdup_printf("KVM_REG_RISCV_SBI_SINGLE | %lld /* UNKNOWN */", reg_off); > + } > + > + return kvm_sbi_ext_reg_name[reg_off]; > +} > + > +static const char *sbi_ext_multi_id_to_str(__u64 reg_subtype, __u64 reg_off) > +{ > + if (reg_off > KVM_REG_RISCV_SBI_MULTI_REG_LAST) { > + /* > + * sbi_ext regs would grow regularly with new sbi extension added, so > + * just show "reg" to indicate a new extension. > + */ > + return strdup_printf("%lld /* UNKNOWN */", reg_off); > + } > + > + switch (reg_subtype) { > + case KVM_REG_RISCV_SBI_MULTI_EN: > + return strdup_printf("KVM_REG_RISCV_SBI_MULTI_EN | %lld", reg_off); > + case KVM_REG_RISCV_SBI_MULTI_DIS: > + return strdup_printf("KVM_REG_RISCV_SBI_MULTI_DIS | %lld", reg_off); > + } > + > + return NULL; > +} > + > +static const char *sbi_ext_id_to_str(const char *prefix, __u64 id) > +{ > + __u64 reg_off = id & ~(REG_MASK | KVM_REG_RISCV_SBI_EXT); > + __u64 reg_subtype = reg_off & KVM_REG_RISCV_SUBTYPE_MASK; > + > + reg_off &= ~KVM_REG_RISCV_SUBTYPE_MASK; > + > + switch (reg_subtype) { > + case KVM_REG_RISCV_SBI_SINGLE: > + return sbi_ext_single_id_to_str(reg_off); > + case KVM_REG_RISCV_SBI_MULTI_EN: > + case KVM_REG_RISCV_SBI_MULTI_DIS: > + return sbi_ext_multi_id_to_str(reg_subtype, reg_off); > + } > + > + TEST_FAIL("%s: Unknown sbi ext subtype: 0x%llx", prefix, reg_subtype); > + return NULL; > +} > + > +void print_reg(const char *prefix, __u64 id) > +{ > + const char *reg_size = NULL; > + > + TEST_ASSERT((id & KVM_REG_ARCH_MASK) == KVM_REG_RISCV, > + "%s: KVM_REG_RISCV missing in reg id: 0x%llx", prefix, id); > + > + switch (id & KVM_REG_SIZE_MASK) { > + case KVM_REG_SIZE_U32: > + reg_size = "KVM_REG_SIZE_U32"; > + break; > + case KVM_REG_SIZE_U64: > + reg_size = "KVM_REG_SIZE_U64"; > + break; > + case KVM_REG_SIZE_U128: > + reg_size = "KVM_REG_SIZE_U128"; > + break; > + default: > + TEST_FAIL("%s: Unexpected reg size: 0x%llx in reg id: 0x%llx", > + prefix, (id & KVM_REG_SIZE_MASK) >> KVM_REG_SIZE_SHIFT, id); > + } > + > + switch (id & KVM_REG_RISCV_TYPE_MASK) { > + case KVM_REG_RISCV_CONFIG: > + printf("\tKVM_REG_RISCV | %s | KVM_REG_RISCV_CONFIG | %s,\n", All the work to try and use KVM_REG_SIZE_ULONG in the right places will be lost if we print a reg list and then copy+paste it as a blessed list. On 64-bit, the only thing supported now, we'll get U64, but if we ever supported 32-bit, then we'd get U32. This is unfortunate, but there's nothing we can do about it. Either we can't have a true print+copy+paste workflow or we should assume we'll only support 64-bit and only use U64 in the blessed lists (from a copy+paste). But, we've already got ULONG in there now, so we can just leave it and burn this bridge later. > + reg_size, config_id_to_str(id)); > + break; > + case KVM_REG_RISCV_CORE: > + printf("\tKVM_REG_RISCV | %s | KVM_REG_RISCV_CORE | %s,\n", > + reg_size, core_id_to_str(prefix, id)); > + break; > + case KVM_REG_RISCV_CSR: > + printf("\tKVM_REG_RISCV | %s | KVM_REG_RISCV_CSR | %s,\n", > + reg_size, csr_id_to_str(prefix, id)); > + break; > + case KVM_REG_RISCV_TIMER: > + printf("\tKVM_REG_RISCV | %s | KVM_REG_RISCV_TIMER | %s,\n", > + reg_size, timer_id_to_str(prefix, id)); > + break; > + case KVM_REG_RISCV_FP_F: > + printf("\tKVM_REG_RISCV | %s | KVM_REG_RISCV_FP_F | %s,\n", > + reg_size, fp_f_id_to_str(prefix, id)); > + break; > + case KVM_REG_RISCV_FP_D: > + printf("\tKVM_REG_RISCV | %s | KVM_REG_RISCV_FP_D | %s,\n", > + reg_size, fp_d_id_to_str(prefix, id)); > + break; > + case KVM_REG_RISCV_ISA_EXT: > + printf("\tKVM_REG_RISCV | %s | KVM_REG_RISCV_ISA_EXT | %s,\n", > + reg_size, isa_ext_id_to_str(id)); > + break; > + case KVM_REG_RISCV_SBI_EXT: > + printf("\tKVM_REG_RISCV | %s | KVM_REG_RISCV_SBI_EXT | %s,\n", > + reg_size, sbi_ext_id_to_str(prefix, id)); > + break; > + default: > + TEST_FAIL("%s: Unexpected reg type: 0x%llx in reg id: 0x%llx", prefix, > + (id & KVM_REG_RISCV_TYPE_MASK) >> KVM_REG_RISCV_TYPE_SHIFT, id); > + } > +} > + > +/* > + * The current blessed list was primed with the output of kernel version > + * v6.4-rc5 and then later updated with new registers. > + */ > +static __u64 base_regs[] = { > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CONFIG | KVM_REG_RISCV_CONFIG_REG(isa), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CONFIG | KVM_REG_RISCV_CONFIG_REG(mvendorid), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CONFIG | KVM_REG_RISCV_CONFIG_REG(marchid), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CONFIG | KVM_REG_RISCV_CONFIG_REG(mimpid), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.pc), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.ra), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.sp), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.gp), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.tp), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.t0), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.t1), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.t2), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s0), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s1), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.a0), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.a1), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.a2), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.a3), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.a4), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.a5), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.a6), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.a7), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s2), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s3), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s4), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s5), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s6), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s7), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s8), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s9), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s10), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.s11), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.t3), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.t4), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.t5), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(regs.t6), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CORE | KVM_REG_RISCV_CORE_REG(mode), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(sstatus), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(sie), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(stvec), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(sscratch), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(sepc), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(scause), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(stval), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(sip), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(satp), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_GENERAL | KVM_REG_RISCV_CSR_REG(scounteren), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_TIMER | KVM_REG_RISCV_TIMER_REG(frequency), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_TIMER | KVM_REG_RISCV_TIMER_REG(time), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_TIMER | KVM_REG_RISCV_TIMER_REG(compare), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_TIMER | KVM_REG_RISCV_TIMER_REG(state), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_A, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_C, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_D, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_F, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_I, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_M, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_V01, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_TIME, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_IPI, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_RFENCE, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_SRST, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_HSM, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_PMU, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_EXPERIMENTAL, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_SINGLE | KVM_RISCV_SBI_EXT_VENDOR, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_MULTI_EN | 0, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_SBI_EXT | KVM_REG_RISCV_SBI_MULTI_DIS | 0, > +}; > + > +/* > + * The rejects_set list registers that should skip set test. > + * - KVM_REG_RISCV_TIMER_REG(state): set would fail if it was not initialized properly. > + * - KVM_REG_RISCV_TIMER_REG(frequency): set not supported > + * - KVM_REG_RISCV_CONFIG_REG(zicbom_block_size): set not supported > + * - KVM_REG_RISCV_CONFIG_REG(zicboz_block_size): set not supported > + * - KVM_RISCV_ISA_EXT_SVPBMT: set not supported > + * - KVM_RISCV_ISA_EXT_SVINVA: set not supported > + * - KVM_RISCV_ISA_EXT_SSAIA: set not supported > + */ > +static __u64 base_rejects_set[] = { > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CONFIG | KVM_REG_RISCV_CONFIG_REG(zicbom_block_size), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CONFIG | KVM_REG_RISCV_CONFIG_REG(zicboz_block_size), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_TIMER | KVM_REG_RISCV_TIMER_REG(frequency), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_TIMER | KVM_REG_RISCV_TIMER_REG(state), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_SVPBMT, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_SVINVAL, > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_SSAIA, These aren't all base registers. I think we should divide the reject lists up too, especially considering the idea I wrote in the last patch, which is to test setting the rejects to ensure the expected error is returned. The error may be different for a rejected set of a supported register vs. that of an unsupported register. > +}; > + > +static __u64 zicbom_regs[] = { > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CONFIG | KVM_REG_RISCV_CONFIG_REG(zicbom_block_size), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_ZICBOM, > +}; > + > +static __u64 zicboz_regs[] = { > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CONFIG | KVM_REG_RISCV_CONFIG_REG(zicboz_block_size), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_ZICBOZ, > +}; > + > +static __u64 aia_csr_regs[] = { > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_AIA | KVM_REG_RISCV_CSR_AIA_REG(siselect), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_AIA | KVM_REG_RISCV_CSR_AIA_REG(iprio1), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_AIA | KVM_REG_RISCV_CSR_AIA_REG(iprio2), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_AIA | KVM_REG_RISCV_CSR_AIA_REG(sieh), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_AIA | KVM_REG_RISCV_CSR_AIA_REG(siph), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_AIA | KVM_REG_RISCV_CSR_AIA_REG(iprio1h), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_CSR | KVM_REG_RISCV_CSR_AIA | KVM_REG_RISCV_CSR_AIA_REG(iprio2h), > + KVM_REG_RISCV | KVM_REG_SIZE_ULONG | KVM_REG_RISCV_ISA_EXT | KVM_RISCV_ISA_EXT_SSAIA, > +}; > + > +static __u64 fp_f_regs[] = { > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[0]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[1]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[2]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[3]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[4]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[5]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[6]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[7]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[8]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[9]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[10]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[11]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[12]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[13]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[14]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[15]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[16]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[17]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[18]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[19]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[20]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[21]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[22]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[23]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[24]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[25]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[26]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[27]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[28]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[29]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[30]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(f[31]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_F | KVM_REG_RISCV_FP_F_REG(fcsr), > +}; > + > +static __u64 fp_d_regs[] = { > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[0]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[1]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[2]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[3]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[4]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[5]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[6]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[7]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[8]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[9]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[10]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[11]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[12]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[13]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[14]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[15]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[16]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[17]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[18]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[19]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[20]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[21]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[22]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[23]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[24]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[25]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[26]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[27]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[28]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[29]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[30]), > + KVM_REG_RISCV | KVM_REG_SIZE_U64 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(f[31]), > + KVM_REG_RISCV | KVM_REG_SIZE_U32 | KVM_REG_RISCV_FP_D | KVM_REG_RISCV_FP_D_REG(fcsr), > +}; > + > +#define BASE_SUBLIST \ > + {"base", .regs = base_regs, .regs_n = ARRAY_SIZE(base_regs), \ > + .rejects_set = base_rejects_set, .rejects_set_n = ARRAY_SIZE(base_rejects_set),} > +#define ZICBOM_REGS_SUBLIST \ > + {"zicbom", .feature = KVM_RISCV_ISA_EXT_ZICBOM, .regs = zicbom_regs, \ > + .regs_n = ARRAY_SIZE(zicbom_regs),} > +#define ZICBOZ_REGS_SUBLIST \ > + {"zicboz", .feature = KVM_RISCV_ISA_EXT_ZICBOZ, .regs = zicboz_regs, \ > + .regs_n = ARRAY_SIZE(zicboz_regs),} > +#define AIA_REGS_SUBLIST \ > + {"aia", .feature = KVM_RISCV_ISA_EXT_SSAIA, .regs = aia_csr_regs, \ > + .regs_n = ARRAY_SIZE(aia_csr_regs),} > +#define FP_F_REGS_SUBLIST \ > + {"fp_f", .feature = KVM_RISCV_ISA_EXT_F, .regs = fp_f_regs, \ > + .regs_n = ARRAY_SIZE(fp_f_regs),} > +#define FP_D_REGS_SUBLIST \ > + {"fp_d", .feature = KVM_RISCV_ISA_EXT_D, .regs = fp_d_regs, \ > + .regs_n = ARRAY_SIZE(fp_d_regs),} > + > +static struct vcpu_reg_list zicbo_config = { > + .sublists = { > + BASE_SUBLIST, > + ZICBOM_REGS_SUBLIST, > + ZICBOZ_REGS_SUBLIST, It's possible to have zicbom without zicboz and vice-versa. Since finalize_vcpu() will skip the whole test when it detects a missing feature for a config, then we won't be able to test one without the other. It's a bit annoying, but I think we may need a separate config for each independent extension. > + {0}, > + }, > +}; > + > +static struct vcpu_reg_list aia_config = { > + .sublists = { > + BASE_SUBLIST, > + AIA_REGS_SUBLIST, > + {0}, > + }, > +}; > + > +static struct vcpu_reg_list fp_f_d_config = { > + .sublists = { > + BASE_SUBLIST, > + FP_F_REGS_SUBLIST, > + FP_D_REGS_SUBLIST, > + {0}, > + }, > +}; > + > +struct vcpu_reg_list *vcpu_configs[] = { > + &zicbo_config, > + &aia_config, > + &fp_f_d_config, > +}; > +int vcpu_configs_n = ARRAY_SIZE(vcpu_configs); > -- > 2.34.1 > I see we have a bit of a problem with the configs for riscv. Since we don't disable anything we're not testing, then for any test that is missing, for example, the f and d registers, we'll get output like "There are 66 new registers. Consider adding them to the blessed reg list with the following lines:" and then a dump of all the f and d registers. The test doesn't fail, but it's messy and confusing. Ideally we'd disable all registers of all sublists not in the config, probably by starting by disabling everything and then only reenabling the ones in the config. Anything that can't be disabled is either a KVM bug, i.e. we should be able to disable it, because we can't expect every host to have it, or it needs to be in the base register sublist (meaning every host will always have it). Thanks, drew