Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 04AEAC636D4 for ; Wed, 15 Feb 2023 22:43:17 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229712AbjBOWnP (ORCPT ); Wed, 15 Feb 2023 17:43:15 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:44600 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229866AbjBOWnM (ORCPT ); Wed, 15 Feb 2023 17:43:12 -0500 Received: from mail-yw1-x114a.google.com (mail-yw1-x114a.google.com [IPv6:2607:f8b0:4864:20::114a]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id E3D8624134 for ; Wed, 15 Feb 2023 14:43:10 -0800 (PST) Received: by mail-yw1-x114a.google.com with SMTP id 00721157ae682-52ec7c792b1so389747b3.5 for ; Wed, 15 Feb 2023 14:43:10 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20210112; h=cc:to:from:subject:message-id:references:mime-version:in-reply-to :date:from:to:cc:subject:date:message-id:reply-to; bh=cVqwUoC4GP/fOMjS8mr++nAU4yxunp7mbNFnyq/qIn8=; b=iIQH/g0b9xgWNpEM6kDdNbbbMfOVqoYx/50hTtksPVkvnVRWPK7N0Y54z76+I3ZgJN kCEvQctXT2mnhEECw8xbCZHPuyFDILI8siKAD74DW7FkmmhXgCU9ADG3TTWw6xkYE77A goF2OUeCaEETnV9f8G/epYW/sBvbXrfkwq7b13CrKfaoWmfbcZx/rVxZ2V8JqAf/2H0Y 9fuZgtNgbEckfnHGaNJuKsaYxxjiY5ZFG+xelAc07oClxPO6YMZ3Ks7ni6geqCByrjST Av1+wyqR+tCscDA8kq7Cn8kBMjAp70iHz8UwRi/FwQmpvnko1remy0ZXg5KhpDd3Q5TM zsXQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=cc:to:from:subject:message-id:references:mime-version:in-reply-to :date:x-gm-message-state:from:to:cc:subject:date:message-id:reply-to; bh=cVqwUoC4GP/fOMjS8mr++nAU4yxunp7mbNFnyq/qIn8=; b=QNj8ap+o4VOFTWsRnHqD5tw9oEmM1JLsp68vmcErlDu114+JmGc5mbqXyeQOR33yPp SYv3BDuaIXTmwj20r3cUb64ZzQHKtm4MEU0yV1e8/qLRaFd+bJFWX1XWvsDhMFMbahIS Yf/av8Y6NS40XERVa0yI8/bJqdLB2X5p+b4SgslfaeIsY3kLQxDyv52WPZeIpCd/HE1F gful3PTZk2LN7BZJtUpx3BX6LEIGDuDoTi0YB/CPIu3Nd/eNwuH0WkHzzRaRsg/kFfO6 SAVlbk7NudK8HnLPMIiMbOhfN9FkHOM/6cMihl5JufgnqUCRuhrbl0NkHdakSYb4wvvi 2xnQ== X-Gm-Message-State: AO0yUKWr6MRp+NofKUn60Bsu99r5zRM0a8QmUn20ZEp87xoXsGidqUsw 2S9zexV5suW77WIPJTyqhpy1jkWhxfI= X-Google-Smtp-Source: AK7set83BDwMNOrgS6ktEvYsmjAck1AxlN4mXYwbgpbE0KOH/4yoO6GQe2EpHX2G0ere3dR6hGktDR/8Jpk= X-Received: from zagreus.c.googlers.com ([fda3:e722:ac3:cc00:7f:e700:c0a8:5c37]) (user=seanjc job=sendgmr) by 2002:a81:ae5f:0:b0:52f:daf9:66af with SMTP id g31-20020a81ae5f000000b0052fdaf966afmr5ywk.6.1676500989866; Wed, 15 Feb 2023 14:43:09 -0800 (PST) Date: Wed, 15 Feb 2023 14:43:08 -0800 In-Reply-To: <2b5994e2-15ba-dd57-285c-fb33827a5275@amd.com> Mime-Version: 1.0 References: <20221129193717.513824-1-mlevitsk@redhat.com> <20221129193717.513824-8-mlevitsk@redhat.com> <2b5994e2-15ba-dd57-285c-fb33827a5275@amd.com> Message-ID: Subject: Re: [PATCH v2 07/11] KVM: x86: add a delayed hardware NMI injection interface From: Sean Christopherson To: Santosh Shukla Cc: Maxim Levitsky , kvm@vger.kernel.org, Sandipan Das , Paolo Bonzini , Jim Mattson , Peter Zijlstra , Dave Hansen , Borislav Petkov , Pawan Gupta , Thomas Gleixner , Ingo Molnar , Josh Poimboeuf , Daniel Sneddon , Jiaxi Chen , Babu Moger , linux-kernel@vger.kernel.org, Jing Liu , Wyes Karny , x86@kernel.org, "H. Peter Anvin" Content-Type: text/plain; charset="us-ascii" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Feb 14, 2023, Santosh Shukla wrote: > On 2/8/2023 9:36 PM, Sean Christopherson wrote: > > On Wed, Feb 08, 2023, Santosh Shukla wrote: > >> On 2/1/2023 3:58 AM, Sean Christopherson wrote: > >>> On Tue, Nov 29, 2022, Maxim Levitsky wrote: > >>>> @@ -5191,9 +5191,12 @@ static int kvm_vcpu_ioctl_x86_set_vcpu_events(struct kvm_vcpu *vcpu, > >>>> > >>>> vcpu->arch.nmi_injected = events->nmi.injected; > >>>> if (events->flags & KVM_VCPUEVENT_VALID_NMI_PENDING) > >>>> - vcpu->arch.nmi_pending = events->nmi.pending; > >>>> + atomic_add(events->nmi.pending, &vcpu->arch.nmi_queued); > >>>> + > >>>> static_call(kvm_x86_set_nmi_mask)(vcpu, events->nmi.masked); > >>>> > >>>> + process_nmi(vcpu); > >>> > >>> Argh, having two process_nmi() calls is ugly (not blaming your code, it's KVM's > >>> ABI that's ugly). E.g. if we collapse this down, it becomes: > >>> > >>> process_nmi(vcpu); > >>> > >>> if (events->flags & KVM_VCPUEVENT_VALID_NMI_PENDING) { > >>> > >>> } > >>> static_call(kvm_x86_set_nmi_mask)(vcpu, events->nmi.masked); > >>> > >>> process_nmi(vcpu); > >>> > >>> And the second mess is that V_NMI needs to be cleared. > >>> > >> > >> Can you please elaborate on "V_NMI cleared" scenario? Are you mentioning > >> about V_NMI_MASK or svm->nmi_masked? > > > > V_NMI_MASK. KVM needs to purge any pending virtual NMIs when userspace sets > > vCPU event state and KVM_VCPUEVENT_VALID_NMI_PENDING is set. > > > > As per the APM: V_NMI_MASK is managed by the processor Heh, we're running afoul over KVM's bad terminology conflicting with the APM's terminology. By V_NMI_MASK, I meant "KVM's V_NMI_MASK", a.k.a. the flag that says whether or there's a pending NMI. However... > " > V_NMI_MASK: Indicates whether virtual NMIs are masked. The processor will set V_NMI_MASK > once it takes the virtual NMI. V_NMI_MASK is cleared when the guest successfully completes an > IRET instruction or #VMEXIT occurs while delivering the virtual NMI > " > > In my initial implementation I had changed V_NMI_MASK for the SMM scenario [1], > This is also not required as HW will save the V_NMI/V_NMI_MASK on > SMM entry and restore them on RSM. > > That said the svm_{get,set}_nmi_mask will look something like: > > diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c > index a9e9bfbffd72..08911a33cf1e 100644 > --- a/arch/x86/kvm/svm/svm.c > +++ b/arch/x86/kvm/svm/svm.c > @@ -3635,13 +3635,21 @@ static int svm_nmi_allowed(struct kvm_vcpu *vcpu, bool for_injection) > > static bool svm_get_nmi_mask(struct kvm_vcpu *vcpu) > { > - return to_svm(vcpu)->nmi_masked; > + struct vcpu_svm *svm = to_svm(vcpu); > + > + if (is_vnmi_enabled(svm)) > + return svm->vmcb->control.int_ctl & V_NMI_MASK; > + else > + return svm->nmi_masked; > } > > static void svm_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked) > { > struct vcpu_svm *svm = to_svm(vcpu); > > + if (is_vnmi_enabled(svm)) > + return; > + > if (masked) { > svm->nmi_masked = true; > svm_set_iret_intercept(svm); > > is there any inputs on above approach? What happens if software clears the "NMIs are blocked" flag? If KVM can't clear the flag, then we've got problems. E.g. if KVM emulates IRET or SMI+RSM. And I I believe there are use cases that use KVM to snapshot and reload vCPU state, e.g. record+replay?, in which case KVM_SET_VCPU_EVENTS needs to be able to adjust NMI blocking too. > On top of the above, I am including your suggested change as below... > > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > index e0855599df65..437a6cea3bc7 100644 > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -5201,9 +5201,9 @@ static int kvm_vcpu_ioctl_x86_set_vcpu_events(struct kvm_vcpu *vcpu, > > vcpu->arch.nmi_injected = events->nmi.injected; > if (events->flags & KVM_VCPUEVENT_VALID_NMI_PENDING) { > - vcpu->arch.nmi_pending = events->nmi.pending; > - if (vcpu->arch.nmi_pending) > - kvm_make_request(KVM_REQ_NMI, vcpu); > + vcpu->arch.nmi_pending = 0; > + atomic_set(&vcpu->arch.nmi_queued, events->nmi.pending); > + kvm_make_request(KVM_REQ_NMI, vcpu); > } > static_call(kvm_x86_set_nmi_mask)(vcpu, events->nmi.masked); > > does that make sense? Yep!