Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 27252C433F5 for ; Sat, 13 Nov 2021 01:22:59 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 0C223610FC for ; Sat, 13 Nov 2021 01:22:59 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S235638AbhKMBZs (ORCPT ); Fri, 12 Nov 2021 20:25:48 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:47656 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S235599AbhKMBZq (ORCPT ); Fri, 12 Nov 2021 20:25:46 -0500 Received: from mail-pf1-x44a.google.com (mail-pf1-x44a.google.com [IPv6:2607:f8b0:4864:20::44a]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 96C1FC0613F5 for ; Fri, 12 Nov 2021 17:22:54 -0800 (PST) Received: by mail-pf1-x44a.google.com with SMTP id 184-20020a6217c1000000b0049f9aad0040so6540166pfx.21 for ; Fri, 12 Nov 2021 17:22:54 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20210112; h=date:in-reply-to:message-id:mime-version:references:subject:from:to :cc; bh=+aTc395V6StwlPMMozSaDYyuowW/pxa49nUmlm+vlGY=; b=Ok27NYYmRgN2/z3RWIZhPphOf3rQMPDTd8TDzYrQElylGFgb2NFU78LSbFWqgnQnkh WFc+Edik8kKsOPvBoB+4sYx0QDpJ7pF1WJWYAURM1Fy58LSjQo1evNB+IMdy881jtniH K2oEhsmQ0C1Q0eSCivY8r/UuNLCK8WiYbozEemo4NZkA4VH554Klz7vNEsODDiuDZNj1 pszvd7xkdEe9UluHWyTQEYrz/BTVZHYlRmUKoO5bqzPHoKTD7mauBd5ZcgIwK8aSaMQK tlnVXYUYae9kFs/MF8NHX2gZhsmKGZt5oxVEyFaHCRmtY/4eVUnuIDFy+ntPL/Z23sUV iZnQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:date:in-reply-to:message-id:mime-version :references:subject:from:to:cc; bh=+aTc395V6StwlPMMozSaDYyuowW/pxa49nUmlm+vlGY=; b=2k+8twyNh9uDblayOuvFl1Kw8lXMuQUUyrFXCOgahe02xziyQbGz03vHv/Jcm/cbIX lGetT5OlIdwOkd9SJC+IuWs0DAkKTa0MrAhCjZFfv6ZoOUq8ijTCI6NvzdFzkB5EynlE VotcIocXbhlGwtF70SHFalZc7YArmSAS/AjDoD0buHj6VzE9tN0k3oPUoSfwO7R/GaYx nzgodoZ5qxNdu0xVRxOO1WR9wV9ATgr08Uwh6vfZ87UPrBg4GGdvrZAxGOap5DrB3I9v n5kIg16Uynd9/Y7hR8ReqRTkUPb45a+0vZoN+cPJqL8+qS0YiLgFHQyUNxuyN4HnL0Yz 9OfQ== X-Gm-Message-State: AOAM532vQA8m0gYx6vZTsBgHA05VLHTb46+OTp8oEasvCvhR3JZLtsZf PmuWPiz2Uz9MabDu6SdDZPQGILW8cCzV X-Google-Smtp-Source: ABdhPJxi6EDbVldKMojd8H/bfjMtkpYbD8II0syl77qvXLNkEbiH2L/XTiDNhpaSL6TwDf1T9xDt+y18ClLS X-Received: from rananta-virt.c.googlers.com ([fda3:e722:ac3:cc00:7f:e700:c0a8:1bcc]) (user=rananta job=sendgmr) by 2002:a17:90b:1c87:: with SMTP id oo7mr17488390pjb.159.1636766574127; Fri, 12 Nov 2021 17:22:54 -0800 (PST) Date: Sat, 13 Nov 2021 01:22:27 +0000 In-Reply-To: <20211113012234.1443009-1-rananta@google.com> Message-Id: <20211113012234.1443009-5-rananta@google.com> Mime-Version: 1.0 References: <20211113012234.1443009-1-rananta@google.com> X-Mailer: git-send-email 2.34.0.rc1.387.gb447b232ab-goog Subject: [RFC PATCH v2 04/11] KVM: arm64: Setup a framework for hypercall bitmap firmware registers From: Raghavendra Rao Ananta To: Marc Zyngier , Andrew Jones , James Morse , Alexandru Elisei , Suzuki K Poulose Cc: Paolo Bonzini , Catalin Marinas , Will Deacon , Peter Shier , Ricardo Koller , Oliver Upton , Reiji Watanabe , Jing Zhang , Raghavendra Rao Anata , linux-arm-kernel@lists.infradead.org, kvmarm@lists.cs.columbia.edu, linux-kernel@vger.kernel.org, kvm@vger.kernel.org Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org KVM regularly introduces new hypercall services to the guests without any consent from the Virtual Machine Manager (VMM). This means, the guests can observe hypercall services in and out as they migrate across various host kernel versions. This could be a major problem if the guest discovered a hypercall, started using it, and after getting migrated to an older kernel realizes that it's no longer available. Depending on how the guest handles the change, there's a potential chance that the guest would just panic. As a result, there's a need for the VMM to elect the services that it wishes the guest to discover. VMM can elect these services based on the kernels spread across its (migration) fleet. To remedy this, extend the existing firmware psuedo-registers, such as KVM_REG_ARM_PSCI_VERSION, for all the hypercall services available. These firmware registers are categorized based on the service call owners, and unlike the existing firmware psuedo-registers, they hold the features supported in the form of a bitmap. During VM (vCPU) initialization, the registers shows an upper-limit of the features supported by the corresponding registers. The VMM can simply use GET_ONE_REG to discover the features. If it's unhappy with any of the features, it can simply write-back the desired feature bitmap using SET_ONE_REG. KVM allows these modification only until a VM has started. KVM also assumes that the VMM is unaware of a register if a register remains unaccessed (read/write), and would simply clear all the bits of the registers such that the guest accidently doesn't get exposed to the features. Finally, the set of bitmaps from all the registers are the services that are exposed to the guest. In order to provide backward compatibility with already existing VMMs, a new capability, KVM_CAP_ARM_HVC_FW_REG_BMAP, is introduced. To enable the bitmap firmware registers extension, the capability must be explicitly enabled. If not, the behavior is similar to the previous setup. In this patch, the framework adds the register only for ARM's standard secure services (owner value 4). Currently, this includes support only for ARM True Random Number Generator (TRNG) service, with bit-0 of the register representing mandatory features of v1.0. Other services are momentarily added in the upcoming patches. Signed-off-by: Raghavendra Rao Ananta --- arch/arm64/include/asm/kvm_host.h | 16 +++ arch/arm64/include/uapi/asm/kvm.h | 4 + arch/arm64/kvm/arm.c | 23 +++- arch/arm64/kvm/hypercalls.c | 217 +++++++++++++++++++++++++++++- arch/arm64/kvm/trng.c | 9 +- include/kvm/arm_hypercalls.h | 7 + include/uapi/linux/kvm.h | 1 + 7 files changed, 262 insertions(+), 15 deletions(-) diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h index 02dffe50a20c..1546a2f973ef 100644 --- a/arch/arm64/include/asm/kvm_host.h +++ b/arch/arm64/include/asm/kvm_host.h @@ -102,6 +102,19 @@ struct kvm_s2_mmu { struct kvm_arch_memory_slot { }; +struct hvc_fw_reg_bmap { + bool accessed; + u64 reg_id; + u64 bmap; +}; + +struct hvc_reg_desc { + spinlock_t lock; + bool fw_reg_bmap_enabled; + + struct hvc_fw_reg_bmap hvc_std_bmap; +}; + struct kvm_arch { struct kvm_s2_mmu mmu; @@ -137,6 +150,9 @@ struct kvm_arch { /* Memory Tagging Extension enabled for the guest */ bool mte_enabled; + + /* Hypercall firmware registers' descriptor */ + struct hvc_reg_desc hvc_desc; }; struct kvm_vcpu_fault_info { diff --git a/arch/arm64/include/uapi/asm/kvm.h b/arch/arm64/include/uapi/asm/kvm.h index b3edde68bc3e..d6e099ed14ef 100644 --- a/arch/arm64/include/uapi/asm/kvm.h +++ b/arch/arm64/include/uapi/asm/kvm.h @@ -281,6 +281,10 @@ struct kvm_arm_copy_mte_tags { #define KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_NOT_REQUIRED 3 #define KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2_ENABLED (1U << 4) +#define KVM_REG_ARM_STD_BMAP KVM_REG_ARM_FW_REG(3) +#define KVM_REG_ARM_STD_BIT_TRNG_V1_0 BIT(0) +#define KVM_REG_ARM_STD_BMAP_BIT_MAX 0 /* Last valid bit */ + /* SVE registers */ #define KVM_REG_ARM64_SVE (0x15 << KVM_REG_ARM_COPROC_SHIFT) diff --git a/arch/arm64/kvm/arm.c b/arch/arm64/kvm/arm.c index 0cc148211b4e..f2099e4d1109 100644 --- a/arch/arm64/kvm/arm.c +++ b/arch/arm64/kvm/arm.c @@ -81,26 +81,32 @@ int kvm_arch_check_processor_compat(void *opaque) int kvm_vm_ioctl_enable_cap(struct kvm *kvm, struct kvm_enable_cap *cap) { - int r; + int r = 0; + struct hvc_reg_desc *hvc_desc = &kvm->arch.hvc_desc; if (cap->flags) return -EINVAL; switch (cap->cap) { case KVM_CAP_ARM_NISV_TO_USER: - r = 0; kvm->arch.return_nisv_io_abort_to_user = true; break; case KVM_CAP_ARM_MTE: mutex_lock(&kvm->lock); - if (!system_supports_mte() || kvm->created_vcpus) { + if (!system_supports_mte() || kvm->created_vcpus) r = -EINVAL; - } else { - r = 0; + else kvm->arch.mte_enabled = true; - } mutex_unlock(&kvm->lock); break; + case KVM_CAP_ARM_HVC_FW_REG_BMAP: + if (kvm_vm_has_run_once(kvm)) + return -EBUSY; + + spin_lock(&hvc_desc->lock); + hvc_desc->fw_reg_bmap_enabled = true; + spin_unlock(&hvc_desc->lock); + break; default: r = -EINVAL; break; @@ -157,6 +163,8 @@ int kvm_arch_init_vm(struct kvm *kvm, unsigned long type) set_default_spectre(kvm); + kvm_arm_init_hypercalls(kvm); + return ret; out_free_stage2_pgd: kvm_free_stage2_pgd(&kvm->arch.mmu); @@ -215,6 +223,7 @@ int kvm_vm_ioctl_check_extension(struct kvm *kvm, long ext) case KVM_CAP_SET_GUEST_DEBUG: case KVM_CAP_VCPU_ATTRIBUTES: case KVM_CAP_PTP_KVM: + case KVM_CAP_ARM_HVC_FW_REG_BMAP: r = 1; break; case KVM_CAP_SET_GUEST_DEBUG2: @@ -622,6 +631,8 @@ static int kvm_vcpu_first_run_init(struct kvm_vcpu *vcpu) if (kvm_vm_is_protected(kvm)) kvm_call_hyp_nvhe(__pkvm_vcpu_init_traps, vcpu); + kvm_arm_sanitize_fw_regs(kvm); + return ret; } diff --git a/arch/arm64/kvm/hypercalls.c b/arch/arm64/kvm/hypercalls.c index 9e136d91b470..f5df7bc61146 100644 --- a/arch/arm64/kvm/hypercalls.c +++ b/arch/arm64/kvm/hypercalls.c @@ -58,6 +58,41 @@ static void kvm_ptp_get_time(struct kvm_vcpu *vcpu, u64 *val) val[3] = lower_32_bits(cycles); } +static bool +kvm_arm_fw_reg_feat_enabled(struct hvc_fw_reg_bmap *reg_bmap, u64 feat_bit) +{ + return reg_bmap->bmap & feat_bit; +} + +bool kvm_hvc_call_supported(struct kvm_vcpu *vcpu, u32 func_id) +{ + struct hvc_reg_desc *hvc_desc = &vcpu->kvm->arch.hvc_desc; + + /* + * To ensure backward compatibility, support all the service calls, + * including new additions, if the firmware registers holding the + * feature bitmaps isn't explicitly enabled. + */ + if (!hvc_desc->fw_reg_bmap_enabled) + return true; + + switch (func_id) { + case ARM_SMCCC_TRNG_VERSION: + case ARM_SMCCC_TRNG_FEATURES: + case ARM_SMCCC_TRNG_GET_UUID: + case ARM_SMCCC_TRNG_RND32: + case ARM_SMCCC_TRNG_RND64: + return kvm_arm_fw_reg_feat_enabled(&hvc_desc->hvc_std_bmap, + KVM_REG_ARM_STD_BIT_TRNG_V1_0); + default: + /* By default, allow the services that aren't listed here */ + return true; + } + + /* We shouldn't be reaching here */ + return true; +} + int kvm_hvc_call_handler(struct kvm_vcpu *vcpu) { u32 func_id = smccc_get_function(vcpu); @@ -65,6 +100,9 @@ int kvm_hvc_call_handler(struct kvm_vcpu *vcpu) u32 feature; gpa_t gpa; + if (!kvm_hvc_call_supported(vcpu, func_id)) + goto out; + switch (func_id) { case ARM_SMCCC_VERSION_FUNC_ID: val[0] = ARM_SMCCC_VERSION_1_1; @@ -143,6 +181,7 @@ int kvm_hvc_call_handler(struct kvm_vcpu *vcpu) return kvm_psci_call(vcpu); } +out: smccc_set_retval(vcpu, val[0], val[1], val[2], val[3]); return 1; } @@ -153,17 +192,178 @@ static const u64 fw_reg_ids[] = { KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2, }; +static const u64 fw_reg_bmap_ids[] = { + KVM_REG_ARM_STD_BMAP, +}; + +static void kvm_arm_fw_reg_init_hvc(struct hvc_reg_desc *hvc_desc, + struct hvc_fw_reg_bmap *fw_reg_bmap, + u64 reg_id, u64 default_map) +{ + fw_reg_bmap->reg_id = reg_id; + fw_reg_bmap->bmap = default_map; +} + +void kvm_arm_init_hypercalls(struct kvm *kvm) +{ + struct hvc_reg_desc *hvc_desc = &kvm->arch.hvc_desc; + + spin_lock_init(&hvc_desc->lock); + + kvm_arm_fw_reg_init_hvc(hvc_desc, &hvc_desc->hvc_std_bmap, + KVM_REG_ARM_STD_BMAP, ARM_SMCCC_STD_FEATURES); +} + +static void kvm_arm_fw_reg_sanitize(struct hvc_fw_reg_bmap *fw_reg_bmap) +{ + if (!fw_reg_bmap->accessed) + fw_reg_bmap->bmap = 0; +} + +/* + * kvm_arm_sanitize_fw_regs: Sanitize the hypercall firmware registers + * + * Sanitization, in the case of hypercall firmware registers, is basically + * clearing out the feature bitmaps so that the guests are not exposed to + * the services corresponding to a particular register. The registers that + * needs sanitization is decided on two factors on the user-space part: + * 1. Enablement of KVM_CAP_ARM_HVC_FW_REG_BMAP: + * If the user-space hasn't enabled the capability, it either means + * that it's unaware of its existence, or it simply doesn't want to + * participate in the arrangement and is okay with the default settings. + * The former case is to ensure backward compatibility. + * + * 2. Has the user-space accessed (read/write) the register? : + * If yes, it means that the user-space is aware of the register's + * existence and can set the bits as it sees fit for the guest. A + * read-only access from user-space indicates that the user-space is + * happy with the default settings, and doesn't wish to change it. + * + * The logic for sanitizing a register will then be: + * --------------------------------------------------------------------------- + * | CAP enabled | Accessed reg | Clear reg | Comments | + * --------------------------------------------------------------------------- + * | N | N | N | | + * | N | Y | N | -ENOENT returned during access | + * | Y | N | Y | | + * | Y | Y | N | | + * --------------------------------------------------------------------------- + */ +void kvm_arm_sanitize_fw_regs(struct kvm *kvm) +{ + struct hvc_reg_desc *hvc_desc = &kvm->arch.hvc_desc; + + spin_lock(&hvc_desc->lock); + + if (!hvc_desc->fw_reg_bmap_enabled) + goto out; + + kvm_arm_fw_reg_sanitize(&hvc_desc->hvc_std_bmap); + +out: + spin_unlock(&hvc_desc->lock); +} + +static int kvm_arm_fw_reg_get_bmap(struct kvm *kvm, + struct hvc_fw_reg_bmap *fw_reg_bmap, u64 *val) +{ + int ret = 0; + struct hvc_reg_desc *hvc_desc = &kvm->arch.hvc_desc; + + spin_lock(&hvc_desc->lock); + + if (!hvc_desc->fw_reg_bmap_enabled) { + ret = -ENOENT; + goto out; + } + + fw_reg_bmap->accessed = true; + *val = fw_reg_bmap->bmap; +out: + spin_unlock(&hvc_desc->lock); + return ret; +} + +static int kvm_arm_fw_reg_set_bmap(struct kvm *kvm, + struct hvc_fw_reg_bmap *fw_reg_bmap, u64 val) +{ + int ret = 0; + u64 fw_reg_features; + struct hvc_reg_desc *hvc_desc = &kvm->arch.hvc_desc; + + spin_lock(&hvc_desc->lock); + + if (!hvc_desc->fw_reg_bmap_enabled) { + ret = -ENOENT; + goto out; + } + + if (fw_reg_bmap->bmap == val) + goto out; + + if (kvm_vm_has_run_once(kvm)) { + ret = -EBUSY; + goto out; + } + + switch (fw_reg_bmap->reg_id) { + case KVM_REG_ARM_STD_BMAP: + fw_reg_features = ARM_SMCCC_STD_FEATURES; + break; + default: + ret = -EINVAL; + goto out; + } + + /* Check for unsupported feature bit */ + if (val & ~fw_reg_features) { + ret = -EINVAL; + goto out; + } + + fw_reg_bmap->accessed = true; + fw_reg_bmap->bmap = val; +out: + spin_unlock(&hvc_desc->lock); + return ret; +} + int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu) { - return ARRAY_SIZE(fw_reg_ids); + struct hvc_reg_desc *hvc_desc = &vcpu->kvm->arch.hvc_desc; + int n_regs = ARRAY_SIZE(fw_reg_ids); + + spin_lock(&hvc_desc->lock); + + if (hvc_desc->fw_reg_bmap_enabled) + n_regs += ARRAY_SIZE(fw_reg_bmap_ids); + + spin_unlock(&hvc_desc->lock); + + return n_regs; } int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices) { + struct hvc_reg_desc *hvc_desc = &vcpu->kvm->arch.hvc_desc; int i; for (i = 0; i < ARRAY_SIZE(fw_reg_ids); i++) { - if (put_user(fw_reg_ids[i], uindices)) + if (put_user(fw_reg_ids[i], uindices++)) + return -EFAULT; + } + + spin_lock(&hvc_desc->lock); + + if (!hvc_desc->fw_reg_bmap_enabled) { + spin_unlock(&hvc_desc->lock); + return 0; + } + + spin_unlock(&hvc_desc->lock); + + for (i = 0; i < ARRAY_SIZE(fw_reg_bmap_ids); i++) { + if (put_user(fw_reg_bmap_ids[i], uindices++)) return -EFAULT; } @@ -213,8 +413,11 @@ static int get_kernel_wa_level(u64 regid) int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) { + struct hvc_reg_desc *hvc_desc = &vcpu->kvm->arch.hvc_desc; void __user *uaddr = (void __user *)(long)reg->addr; + struct kvm *kvm = vcpu->kvm; u64 val; + int ret; switch (reg->id) { case KVM_REG_ARM_PSCI_VERSION: @@ -223,6 +426,12 @@ int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_1: case KVM_REG_ARM_SMCCC_ARCH_WORKAROUND_2: val = get_kernel_wa_level(reg->id) & KVM_REG_FEATURE_LEVEL_MASK; + break; + case KVM_REG_ARM_STD_BMAP: + ret = kvm_arm_fw_reg_get_bmap(kvm, &hvc_desc->hvc_std_bmap, &val); + if (ret) + return ret; + break; default: return -ENOENT; @@ -236,6 +445,8 @@ int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) { + struct kvm *kvm = vcpu->kvm; + struct hvc_reg_desc *hvc_desc = &kvm->arch.hvc_desc; void __user *uaddr = (void __user *)(long)reg->addr; u64 val; int wa_level; @@ -310,6 +521,8 @@ int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg) return -EINVAL; return 0; + case KVM_REG_ARM_STD_BMAP: + return kvm_arm_fw_reg_set_bmap(kvm, &hvc_desc->hvc_std_bmap, val); default: return -ENOENT; } diff --git a/arch/arm64/kvm/trng.c b/arch/arm64/kvm/trng.c index 99bdd7103c9c..6dff765f5b9b 100644 --- a/arch/arm64/kvm/trng.c +++ b/arch/arm64/kvm/trng.c @@ -60,14 +60,9 @@ int kvm_trng_call(struct kvm_vcpu *vcpu) val = ARM_SMCCC_TRNG_VERSION_1_0; break; case ARM_SMCCC_TRNG_FEATURES: - switch (smccc_get_arg1(vcpu)) { - case ARM_SMCCC_TRNG_VERSION: - case ARM_SMCCC_TRNG_FEATURES: - case ARM_SMCCC_TRNG_GET_UUID: - case ARM_SMCCC_TRNG_RND32: - case ARM_SMCCC_TRNG_RND64: + if (kvm_hvc_call_supported(vcpu, smccc_get_arg1(vcpu))) val = TRNG_SUCCESS; - } + break; case ARM_SMCCC_TRNG_GET_UUID: smccc_set_retval(vcpu, le32_to_cpu(u[0]), le32_to_cpu(u[1]), diff --git a/include/kvm/arm_hypercalls.h b/include/kvm/arm_hypercalls.h index 5d38628a8d04..8c6300d1cbaf 100644 --- a/include/kvm/arm_hypercalls.h +++ b/include/kvm/arm_hypercalls.h @@ -6,6 +6,9 @@ #include +#define ARM_SMCCC_STD_FEATURES \ + GENMASK_ULL(KVM_REG_ARM_STD_BMAP_BIT_MAX, 0) + int kvm_hvc_call_handler(struct kvm_vcpu *vcpu); static inline u32 smccc_get_function(struct kvm_vcpu *vcpu) @@ -42,9 +45,13 @@ static inline void smccc_set_retval(struct kvm_vcpu *vcpu, struct kvm_one_reg; +void kvm_arm_init_hypercalls(struct kvm *kvm); int kvm_arm_get_fw_num_regs(struct kvm_vcpu *vcpu); int kvm_arm_copy_fw_reg_indices(struct kvm_vcpu *vcpu, u64 __user *uindices); int kvm_arm_get_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); int kvm_arm_set_fw_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); +void kvm_arm_sanitize_fw_regs(struct kvm *kvm); + +bool kvm_hvc_call_supported(struct kvm_vcpu *vcpu, u32 func_id); #endif diff --git a/include/uapi/linux/kvm.h b/include/uapi/linux/kvm.h index 78f0719cc2a3..3855b7b33bb3 100644 --- a/include/uapi/linux/kvm.h +++ b/include/uapi/linux/kvm.h @@ -1130,6 +1130,7 @@ struct kvm_ppc_resize_hpt { #define KVM_CAP_BINARY_STATS_FD 203 #define KVM_CAP_EXIT_ON_EMULATION_FAILURE 204 #define KVM_CAP_ARM_MTE 205 +#define KVM_CAP_ARM_HVC_FW_REG_BMAP 206 #ifdef KVM_CAP_IRQ_ROUTING -- 2.34.0.rc1.387.gb447b232ab-goog