Received: by 2002:ab2:6203:0:b0:1f5:f2ab:c469 with SMTP id o3csp2804281lqt; Tue, 23 Apr 2024 01:59:52 -0700 (PDT) X-Forwarded-Encrypted: i=3; AJvYcCWQbU9lsNLaDDVpGfZVEJP1xklWBxvo1W7hhEY6PgC47BHlsGan8hLp0WHPkuSjQ1uOY7sp+cGY89mb6ydxBOjJ2qxzVGRWwxJ/ePy7bA== X-Google-Smtp-Source: AGHT+IEw/q0OWoUSKiO4sXiB3ldLA+TtVAMXK2NG9Yxfjj6GthkgUsvIAkiNbNtjOpxel4U1xkWo X-Received: by 2002:a05:622a:511:b0:437:80b1:a076 with SMTP id l17-20020a05622a051100b0043780b1a076mr15946328qtx.53.1713862792308; Tue, 23 Apr 2024 01:59:52 -0700 (PDT) ARC-Seal: i=2; a=rsa-sha256; t=1713862792; cv=pass; d=google.com; s=arc-20160816; b=k8G33koipp114UDIbDs6/8vkhPnk1+PMTKvoYInQTCFpMEAdARy3jYM90fmy0QrudN VvN4+2j+ri/lyit120R9xQlMD5YE+8bPw37XFhkQrlN3MW0V+kYAlqOGYQ+CawM9Nj90 zqXaXs6o/xT4oP4SJryfnJEU1uXT9rIk1z/wYOq55nGSm2kcDuBgTW+ERIRDnOsVCdql u/cywhzG5xUwtZZT/a0Ocz2HUygnZnQTU3CPxpOv+miC9FSUUFkkvanrVh5O1gDDLGvi 3SVz7ANwxtTT6/zNbae/crK5VMiqPgrAukBpOBa/VfDaM0smoamfAU5tgqtjVYRcW98g iB4w== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:in-reply-to:from:content-language :references:to:subject:cc:user-agent:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:date:message-id:dkim-signature; bh=HWtqP+sUq3dBNKARyOinvS6qvCVBJ9KSZpclLiy77Yk=; fh=x2vwvesj6wNgXhNq5X1lA1s17OVWTmIVmT/TVJ11qYY=; b=b6diLhDd5WvMabj3JYXtdPoVHdJOfMr2uGxEjxRS7FirILGsuWkHYXQAMHj8PR8BHd 9rgFQtBYLzwRz80zwDS9YKaX3hxhWfOQ79n9DBbIC5cE4CC4DpOBpJKfrS7WkfLmiKpt dCmQtEzdbTLDyNWaPuBv4JOkkFqRkAzTekFV5FYDN67y6R71hq5qPWcNjkIRh9Y9jrqf /w9taMTtHzbRLySCKHb4SdVG9erC+pmm3dM/3BFfvoECHAD9qrSKeZNrw4xZNVyasuqS JV+WKJj3y5hYEdVGjjU7TYGtndGxRZMkRuoF8OrOWnONdu/NgHl8NQwWVNLOS76Ig4/z xP0Q==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@collabora.com header.s=mail header.b=ONoo2JUH; arc=pass (i=1 spf=pass spfdomain=collabora.com dkim=pass dkdomain=collabora.com dmarc=pass fromdomain=collabora.com); spf=pass (google.com: domain of linux-kernel+bounces-154741-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-154741-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=collabora.com Return-Path: Received: from ny.mirrors.kernel.org (ny.mirrors.kernel.org. [147.75.199.223]) by mx.google.com with ESMTPS id w12-20020ac857cc000000b00439f3d33732si765134qta.450.2024.04.23.01.59.52 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 23 Apr 2024 01:59:52 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel+bounces-154741-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) client-ip=147.75.199.223; Authentication-Results: mx.google.com; dkim=pass header.i=@collabora.com header.s=mail header.b=ONoo2JUH; arc=pass (i=1 spf=pass spfdomain=collabora.com dkim=pass dkdomain=collabora.com dmarc=pass fromdomain=collabora.com); spf=pass (google.com: domain of linux-kernel+bounces-154741-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-154741-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=collabora.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ny.mirrors.kernel.org (Postfix) with ESMTPS id E1A691C21C1A for ; Tue, 23 Apr 2024 08:59:51 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id AD47656768; Tue, 23 Apr 2024 08:59:38 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=collabora.com header.i=@collabora.com header.b="ONoo2JUH" Received: from madrid.collaboradmins.com (madrid.collaboradmins.com [46.235.227.194]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id DB9ED4F8BC; Tue, 23 Apr 2024 08:59:34 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=46.235.227.194 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1713862777; cv=none; b=GOcQFLaGMxgOU+Uwsn6wHNGT+NtVnmMzyQa60FmHg8Q6NEFM/Ui9fgEKreURrvfyLuTgJU0z1TZ507Jf0IwgMe24YfZUffxFNPStfqhKR/FqgFoi9HQrx4FPqLaKHmV+EdtrOIuQECr9aE9ervqYTXYiJIYSJe5eq9aFPruHI6s= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1713862777; c=relaxed/simple; bh=ej9dserSO+yCYrpv8pLU9mpbBmJu/VvlmqNcPtql9yg=; h=Message-ID:Date:MIME-Version:Cc:Subject:To:References:From: In-Reply-To:Content-Type; b=awyXkR6tNzyf7PskU/6f16XwEJen7Dray+SdJxigdmpJ5nOPcfulYoWsS2SyALR9ze0UspKabicFlTLQsorFE/PrFApeBqykIOjQ1anvSWKiwDAhYZOOmyg2I+cv28uTGgjp0Te3izgUwRbUxm/F+C7WtVG6ng1ZLi1QTQDV0LQ= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=collabora.com; spf=pass smtp.mailfrom=collabora.com; dkim=pass (2048-bit key) header.d=collabora.com header.i=@collabora.com header.b=ONoo2JUH; arc=none smtp.client-ip=46.235.227.194 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=collabora.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=collabora.com DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=collabora.com; s=mail; t=1713862773; bh=ej9dserSO+yCYrpv8pLU9mpbBmJu/VvlmqNcPtql9yg=; h=Date:Cc:Subject:To:References:From:In-Reply-To:From; b=ONoo2JUHRZuoq84An2uEDJnAYz2kFpqMLmaV0BpesejcVtTPQKGoWxcXbDnULfQCu q+pqbLlL+NuoeBcgh/xNBZRvGgGez2/O1hwc/EI4dtWimuBQ4fC7H7PP1+WBtksuj8 nkrBUqaf26WJK5JR8zUxu8m2d5JrtYzaZK5v68b/JwScgD+j8p6ssKKwQ1now3wI1y HytQGTOYD/zwC4dGLtruAO/eqEItQC73ybRg6vRctdoJly6AgwvIx6cJlvOBEuWMLS fqgnT+6wfH9dcF82Ym7cefsvh5R4nw63ihJ6HFtU7cOvWgovZAuR1LC8ACmH2JlpFM /1ejBjRSGZhFA== Received: from [10.193.1.1] (broslavsky.collaboradmins.com [68.183.210.73]) (using TLSv1.3 with cipher TLS_AES_128_GCM_SHA256 (128/128 bits) key-exchange X25519 server-signature RSA-PSS (4096 bits) server-digest SHA256) (No client certificate requested) (Authenticated sender: usama.anjum) by madrid.collaboradmins.com (Postfix) with ESMTPSA id 9A7BE3780016; Tue, 23 Apr 2024 08:59:25 +0000 (UTC) Message-ID: Date: Tue, 23 Apr 2024 13:59:54 +0500 Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Cc: Muhammad Usama Anjum , Anup Patel , Andrew Jones , Ajay Kaher , Albert Ou , Alexandre Ghiti , samuel.holland@sifive.com, Conor Dooley , Juergen Gross , kvm-riscv@lists.infradead.org, kvm@vger.kernel.org, linux-kselftest@vger.kernel.org, linux-riscv@lists.infradead.org, Mark Rutland , Palmer Dabbelt , Paolo Bonzini , Paul Walmsley , Shuah Khan , virtualization@lists.linux.dev, Will Deacon , x86@kernel.org Subject: Re: [PATCH v8 21/24] KVM: riscv: selftests: Add SBI PMU selftest To: Atish Patra , linux-kernel@vger.kernel.org References: <20240420151741.962500-1-atishp@rivosinc.com> <20240420151741.962500-22-atishp@rivosinc.com> Content-Language: en-US From: Muhammad Usama Anjum In-Reply-To: <20240420151741.962500-22-atishp@rivosinc.com> Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 7bit On 4/20/24 8:17 PM, Atish Patra wrote: > This test implements basic sanity test and cycle/instret event > counting tests. > > Reviewed-by: Anup Patel > Reviewed-by: Andrew Jones > Signed-off-by: Atish Patra LGTM Reviewed-by: Muhammad Usama Anjum > --- > tools/testing/selftests/kvm/Makefile | 1 + > .../selftests/kvm/riscv/sbi_pmu_test.c | 369 ++++++++++++++++++ > 2 files changed, 370 insertions(+) > create mode 100644 tools/testing/selftests/kvm/riscv/sbi_pmu_test.c > > diff --git a/tools/testing/selftests/kvm/Makefile b/tools/testing/selftests/kvm/Makefile > index 741c7dc16afc..1cfcd2797ee4 100644 > --- a/tools/testing/selftests/kvm/Makefile > +++ b/tools/testing/selftests/kvm/Makefile > @@ -189,6 +189,7 @@ TEST_GEN_PROGS_s390x += rseq_test > TEST_GEN_PROGS_s390x += set_memory_region_test > TEST_GEN_PROGS_s390x += kvm_binary_stats_test > > +TEST_GEN_PROGS_riscv += riscv/sbi_pmu_test > TEST_GEN_PROGS_riscv += arch_timer > TEST_GEN_PROGS_riscv += demand_paging_test > TEST_GEN_PROGS_riscv += dirty_log_test > diff --git a/tools/testing/selftests/kvm/riscv/sbi_pmu_test.c b/tools/testing/selftests/kvm/riscv/sbi_pmu_test.c > new file mode 100644 > index 000000000000..7c81691e39c5 > --- /dev/null > +++ b/tools/testing/selftests/kvm/riscv/sbi_pmu_test.c > @@ -0,0 +1,369 @@ > +// SPDX-License-Identifier: GPL-2.0-only > +/* > + * sbi_pmu_test.c - Tests the riscv64 SBI PMU functionality. > + * > + * Copyright (c) 2024, Rivos Inc. > + */ > + > +#include > +#include > +#include > +#include > +#include > +#include "kvm_util.h" > +#include "test_util.h" > +#include "processor.h" > +#include "sbi.h" > + > +/* Maximum counters(firmware + hardware) */ > +#define RISCV_MAX_PMU_COUNTERS 64 > +union sbi_pmu_ctr_info ctrinfo_arr[RISCV_MAX_PMU_COUNTERS]; > + > +/* Cache the available counters in a bitmask */ > +static unsigned long counter_mask_available; > + > +static bool illegal_handler_invoked; > + > +unsigned long pmu_csr_read_num(int csr_num) > +{ > +#define switchcase_csr_read(__csr_num, __val) {\ > + case __csr_num: \ > + __val = csr_read(__csr_num); \ > + break; } > +#define switchcase_csr_read_2(__csr_num, __val) {\ > + switchcase_csr_read(__csr_num + 0, __val) \ > + switchcase_csr_read(__csr_num + 1, __val)} > +#define switchcase_csr_read_4(__csr_num, __val) {\ > + switchcase_csr_read_2(__csr_num + 0, __val) \ > + switchcase_csr_read_2(__csr_num + 2, __val)} > +#define switchcase_csr_read_8(__csr_num, __val) {\ > + switchcase_csr_read_4(__csr_num + 0, __val) \ > + switchcase_csr_read_4(__csr_num + 4, __val)} > +#define switchcase_csr_read_16(__csr_num, __val) {\ > + switchcase_csr_read_8(__csr_num + 0, __val) \ > + switchcase_csr_read_8(__csr_num + 8, __val)} > +#define switchcase_csr_read_32(__csr_num, __val) {\ > + switchcase_csr_read_16(__csr_num + 0, __val) \ > + switchcase_csr_read_16(__csr_num + 16, __val)} > + > + unsigned long ret = 0; > + > + switch (csr_num) { > + switchcase_csr_read_32(CSR_CYCLE, ret) > + switchcase_csr_read_32(CSR_CYCLEH, ret) > + default : > + break; > + } > + > + return ret; > +#undef switchcase_csr_read_32 > +#undef switchcase_csr_read_16 > +#undef switchcase_csr_read_8 > +#undef switchcase_csr_read_4 > +#undef switchcase_csr_read_2 > +#undef switchcase_csr_read > +} > + > +static inline void dummy_func_loop(uint64_t iter) > +{ > + int i = 0; > + > + while (i < iter) { > + asm volatile("nop"); > + i++; > + } > +} > + > +static void start_counter(unsigned long counter, unsigned long start_flags, > + unsigned long ival) > +{ > + struct sbiret ret; > + > + ret = sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_START, counter, 1, start_flags, > + ival, 0, 0); > + __GUEST_ASSERT(ret.error == 0, "Unable to start counter %ld\n", counter); > +} > + > +/* This should be invoked only for reset counter use case */ > +static void stop_reset_counter(unsigned long counter, unsigned long stop_flags) > +{ > + struct sbiret ret; > + > + ret = sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_STOP, counter, 1, > + stop_flags | SBI_PMU_STOP_FLAG_RESET, 0, 0, 0); > + __GUEST_ASSERT(ret.error == SBI_ERR_ALREADY_STOPPED, > + "Unable to stop counter %ld\n", counter); > +} > + > +static void stop_counter(unsigned long counter, unsigned long stop_flags) > +{ > + struct sbiret ret; > + > + ret = sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_STOP, counter, 1, stop_flags, > + 0, 0, 0); > + __GUEST_ASSERT(ret.error == 0, "Unable to stop counter %ld error %ld\n", > + counter, ret.error); > +} > + > +static void guest_illegal_exception_handler(struct ex_regs *regs) > +{ > + __GUEST_ASSERT(regs->cause == EXC_INST_ILLEGAL, > + "Unexpected exception handler %lx\n", regs->cause); > + > + illegal_handler_invoked = true; > + /* skip the trapping instruction */ > + regs->epc += 4; > +} > + > +static unsigned long get_counter_index(unsigned long cbase, unsigned long cmask, > + unsigned long cflags, > + unsigned long event) > +{ > + struct sbiret ret; > + > + ret = sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_CFG_MATCH, cbase, cmask, > + cflags, event, 0, 0); > + __GUEST_ASSERT(ret.error == 0, "config matching failed %ld\n", ret.error); > + GUEST_ASSERT(ret.value < RISCV_MAX_PMU_COUNTERS); > + GUEST_ASSERT(BIT(ret.value) & counter_mask_available); > + > + return ret.value; > +} > + > +static unsigned long get_num_counters(void) > +{ > + struct sbiret ret; > + > + ret = sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_NUM_COUNTERS, 0, 0, 0, 0, 0, 0); > + > + __GUEST_ASSERT(ret.error == 0, "Unable to retrieve number of counters from SBI PMU"); > + __GUEST_ASSERT(ret.value < RISCV_MAX_PMU_COUNTERS, > + "Invalid number of counters %ld\n", ret.value); > + > + return ret.value; > +} > + > +static void update_counter_info(int num_counters) > +{ > + int i = 0; > + struct sbiret ret; > + > + for (i = 0; i < num_counters; i++) { > + ret = sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_GET_INFO, i, 0, 0, 0, 0, 0); > + > + /* There can be gaps in logical counter indicies*/ > + if (ret.error) > + continue; > + GUEST_ASSERT_NE(ret.value, 0); > + > + ctrinfo_arr[i].value = ret.value; > + counter_mask_available |= BIT(i); > + } > + > + GUEST_ASSERT(counter_mask_available > 0); > +} > + > +static unsigned long read_fw_counter(int idx, union sbi_pmu_ctr_info ctrinfo) > +{ > + struct sbiret ret; > + > + ret = sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_FW_READ, idx, 0, 0, 0, 0, 0); > + GUEST_ASSERT(ret.error == 0); > + return ret.value; > +} > + > +static unsigned long read_counter(int idx, union sbi_pmu_ctr_info ctrinfo) > +{ > + unsigned long counter_val = 0; > + > + __GUEST_ASSERT(ctrinfo.type < 2, "Invalid counter type %d", ctrinfo.type); > + > + if (ctrinfo.type == SBI_PMU_CTR_TYPE_HW) > + counter_val = pmu_csr_read_num(ctrinfo.csr); > + else if (ctrinfo.type == SBI_PMU_CTR_TYPE_FW) > + counter_val = read_fw_counter(idx, ctrinfo); > + > + return counter_val; > +} > + > +static void test_pmu_event(unsigned long event) > +{ > + unsigned long counter; > + unsigned long counter_value_pre, counter_value_post; > + unsigned long counter_init_value = 100; > + > + counter = get_counter_index(0, counter_mask_available, 0, event); > + counter_value_pre = read_counter(counter, ctrinfo_arr[counter]); > + > + /* Do not set the initial value */ > + start_counter(counter, 0, 0); > + dummy_func_loop(10000); > + stop_counter(counter, 0); > + > + counter_value_post = read_counter(counter, ctrinfo_arr[counter]); > + __GUEST_ASSERT(counter_value_post > counter_value_pre, > + "Event update verification failed: post [%lx] pre [%lx]\n", > + counter_value_post, counter_value_pre); > + > + /* > + * We can't just update the counter without starting it. > + * Do start/stop twice to simulate that by first initializing to a very > + * high value and a low value after that. > + */ > + start_counter(counter, SBI_PMU_START_FLAG_SET_INIT_VALUE, ULONG_MAX/2); > + stop_counter(counter, 0); > + counter_value_pre = read_counter(counter, ctrinfo_arr[counter]); > + > + start_counter(counter, SBI_PMU_START_FLAG_SET_INIT_VALUE, counter_init_value); > + stop_counter(counter, 0); > + counter_value_post = read_counter(counter, ctrinfo_arr[counter]); > + __GUEST_ASSERT(counter_value_pre > counter_value_post, > + "Counter reinitialization verification failed : post [%lx] pre [%lx]\n", > + counter_value_post, counter_value_pre); > + > + /* Now set the initial value and compare */ > + start_counter(counter, SBI_PMU_START_FLAG_SET_INIT_VALUE, counter_init_value); > + dummy_func_loop(10000); > + stop_counter(counter, 0); > + > + counter_value_post = read_counter(counter, ctrinfo_arr[counter]); > + __GUEST_ASSERT(counter_value_post > counter_init_value, > + "Event update verification failed: post [%lx] pre [%lx]\n", > + counter_value_post, counter_init_value); > + > + stop_reset_counter(counter, 0); > +} > + > +static void test_invalid_event(void) > +{ > + struct sbiret ret; > + unsigned long event = 0x1234; /* A random event */ > + > + ret = sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_CFG_MATCH, 0, > + counter_mask_available, 0, event, 0, 0); > + GUEST_ASSERT_EQ(ret.error, SBI_ERR_NOT_SUPPORTED); > +} > + > +static void test_pmu_events(void) > +{ > + int num_counters = 0; > + > + /* Get the counter details */ > + num_counters = get_num_counters(); > + update_counter_info(num_counters); > + > + /* Sanity testing for any random invalid event */ > + test_invalid_event(); > + > + /* Only these two events are guaranteed to be present */ > + test_pmu_event(SBI_PMU_HW_CPU_CYCLES); > + test_pmu_event(SBI_PMU_HW_INSTRUCTIONS); > + > + GUEST_DONE(); > +} > + > +static void test_pmu_basic_sanity(void) > +{ > + long out_val = 0; > + bool probe; > + struct sbiret ret; > + int num_counters = 0, i; > + union sbi_pmu_ctr_info ctrinfo; > + > + probe = guest_sbi_probe_extension(SBI_EXT_PMU, &out_val); > + GUEST_ASSERT(probe && out_val == 1); > + > + num_counters = get_num_counters(); > + > + for (i = 0; i < num_counters; i++) { > + ret = sbi_ecall(SBI_EXT_PMU, SBI_EXT_PMU_COUNTER_GET_INFO, i, > + 0, 0, 0, 0, 0); > + > + /* There can be gaps in logical counter indicies*/ > + if (ret.error) > + continue; > + GUEST_ASSERT_NE(ret.value, 0); > + > + ctrinfo.value = ret.value; > + > + /** > + * Accessibility check of hardware and read capability of firmware counters. > + * The spec doesn't mandate any initial value. No need to check any value. > + */ > + if (ctrinfo.type == SBI_PMU_CTR_TYPE_HW) { > + pmu_csr_read_num(ctrinfo.csr); > + GUEST_ASSERT(illegal_handler_invoked); > + } else if (ctrinfo.type == SBI_PMU_CTR_TYPE_FW) { > + read_fw_counter(i, ctrinfo); > + } > + } > + > + GUEST_DONE(); > +} > + > +static void run_vcpu(struct kvm_vcpu *vcpu) > +{ > + struct ucall uc; > + > + vcpu_run(vcpu); > + switch (get_ucall(vcpu, &uc)) { > + case UCALL_ABORT: > + REPORT_GUEST_ASSERT(uc); > + break; > + case UCALL_DONE: > + case UCALL_SYNC: > + break; > + default: > + TEST_FAIL("Unknown ucall %lu", uc.cmd); > + break; > + } > +} > + > +void test_vm_destroy(struct kvm_vm *vm) > +{ > + memset(ctrinfo_arr, 0, sizeof(union sbi_pmu_ctr_info) * RISCV_MAX_PMU_COUNTERS); > + counter_mask_available = 0; > + kvm_vm_free(vm); > +} > + > +static void test_vm_basic_test(void *guest_code) > +{ > + struct kvm_vm *vm; > + struct kvm_vcpu *vcpu; > + > + vm = vm_create_with_one_vcpu(&vcpu, guest_code); > + __TEST_REQUIRE(__vcpu_has_sbi_ext(vcpu, KVM_RISCV_SBI_EXT_PMU), > + "SBI PMU not available, skipping test"); > + vm_init_vector_tables(vm); > + /* Illegal instruction handler is required to verify read access without configuration */ > + vm_install_exception_handler(vm, EXC_INST_ILLEGAL, guest_illegal_exception_handler); > + > + vcpu_init_vector_tables(vcpu); > + run_vcpu(vcpu); > + > + test_vm_destroy(vm); > +} > + > +static void test_vm_events_test(void *guest_code) > +{ > + struct kvm_vm *vm = NULL; > + struct kvm_vcpu *vcpu = NULL; > + > + vm = vm_create_with_one_vcpu(&vcpu, guest_code); > + __TEST_REQUIRE(__vcpu_has_sbi_ext(vcpu, KVM_RISCV_SBI_EXT_PMU), > + "SBI PMU not available, skipping test"); > + run_vcpu(vcpu); > + > + test_vm_destroy(vm); > +} > + > +int main(void) > +{ > + test_vm_basic_test(test_pmu_basic_sanity); > + pr_info("SBI PMU basic test : PASS\n"); > + > + test_vm_events_test(test_pmu_events); > + pr_info("SBI PMU event verification test : PASS\n"); > + > + return 0; > +} -- BR, Muhammad Usama Anjum