Received: by 2002:a05:6a10:17d3:0:0:0:0 with SMTP id hz19csp3071485pxb; Mon, 19 Apr 2021 23:09:50 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxbb8rDN4Pn9Tm7j4vz9bHT395wD39EhnbZJCh+rpJNj46TdnL4IraQYbaXT80XfSav8aYM X-Received: by 2002:a17:902:f284:b029:ec:a434:b6ff with SMTP id k4-20020a170902f284b02900eca434b6ffmr9445037plc.81.1618898990022; Mon, 19 Apr 2021 23:09:50 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1618898990; cv=none; d=google.com; s=arc-20160816; b=TzICRYLeUKqAyNMeR03R59EQvPO+uEp7LWTqxzkUgZBRpZ0j3HPFGy2cyAC1DJIA1Q h38q5cl1I1R/tGyyjs4XdfPvCBx1+sc5ynf4vikieZ7CVTndAftSv2HiazcBL70BJjym GB+aBdmuQGbyOyJS6iRYA0zPv5mA+nJLQ7UetWMJiFIOwxH5bNoYHrbDhUxd6Hwb8hNE zQ/M6MbUgoJIbf0DRnixDDW+lzmfRy8x9NhfVuDZVns+ex/N+T4D0P5fIXaluN9bgF1Z SBYHn1Ud5KcjUzlu6QGNlUFkjIie0GyUM6Xzoj/0qpEW9N420OGp9uwZm/HHoAOQOY1G 1hrQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=fjuizEiuZ1OYwihnIJO4mEbU4UOFwOFWRZt/pOleI5w=; b=G1+U4wmdDbp7HuHRBxF+qsPg2eWojXVsHdxsidgFRKUTAj52G7+s3q4pyAEy8O62FO aiq+5u7sXdfFdSByyXhCDVwUYoaw6RyhDEDd5GiklHXowUG25TPpU5bu+qf85LPMNvxg d+e7E+YLMV0U3xpITIkOkXS44lbZwu86vPh9+aq/k8JwatCw9QokSn871pOsseMDC/II 3HhzxBOMP4OP5R6nAZkMhotJiBKHIem7c0xTO1422Fa4zLp0KiLLlTTY/QmY8Ulw+/KY lSG8WdAHjAt/o6yowDriYg2KWMEzLA15RrGmvAWujl3eAH1IyKkCAW3QzVZevcIXM6QD 97uQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=T1uXkT4Z; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id p23si18282145pgm.525.2021.04.19.23.09.37; Mon, 19 Apr 2021 23:09:50 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=T1uXkT4Z; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229985AbhDTGJc (ORCPT + 99 others); Tue, 20 Apr 2021 02:09:32 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:53864 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229577AbhDTGJa (ORCPT ); Tue, 20 Apr 2021 02:09:30 -0400 Received: from mail-oi1-x22e.google.com (mail-oi1-x22e.google.com [IPv6:2607:f8b0:4864:20::22e]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id CEB58C061763; Mon, 19 Apr 2021 23:08:59 -0700 (PDT) Received: by mail-oi1-x22e.google.com with SMTP id k18so32990289oik.1; Mon, 19 Apr 2021 23:08:59 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=fjuizEiuZ1OYwihnIJO4mEbU4UOFwOFWRZt/pOleI5w=; b=T1uXkT4Zg696s/bXD4aAJb4FCzs8hj7re4tttOM/ywPRlIDUYJ8pasbAklFBCw/Q1F Mv3X7uUN2zeyWl7RRN/kGGL2DU7mWkLHSHwjXxZJ2KWQ5Cw2eBVs6cHrRdSYw+Qp1tFr 3GCsjhIV8wgO/H6M5fd91htKxl1dkB0OwCRXxLuZgEmM5+HHnUkzy5W10gQPniG/CRhc dNtdCRyI8i4dtl90Q+isgaJBy9QqkR+5PZmsQGMv0l6+NDjT4YcCgLbcU9t2mojwyYwh Xi+iLUPBdKgO9eA/sw2l66cmfx+Kw89t4lvHP+h5/kl3Cevl/uQNCehJBsRO1JxoKm5i qZwg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=fjuizEiuZ1OYwihnIJO4mEbU4UOFwOFWRZt/pOleI5w=; b=pdWdYExMj46EOa3svUf4PpmiDBRIVtGVA1EvllPVOzhxskHURfHzPTHZh3vzbzp0yR sKQzqLSk1CNUtv6tttDInf8Y3Y5xXyhh9elthJQr9TVNYrQX+MBpyXhAnK2sOHLRIZUD PD3el7tBhOh9x9ek36VVZEwiwh7TUCW+o9JHatpqgxRJO0L77ZAUwllat4TS0/t0m/dM 6xjRIlHyKN7TkC3IezLXBeJaSSbtonP3Vpg2vnVAKPSqPx+vKDobbyTe9TwC3mPkns1X ooVsHcpAVVnoX6ho1vFkW24jQTrLbBQQhGVVKL2klYrq0/mz6LIL2sScZV/4/lTD9RyJ 5jgA== X-Gm-Message-State: AOAM530Mh/mwJnFlBOFNNpLKjPyjPuaNhU2JHVx9aZEfotVL0pCD6+gz oWPh45Fx2r5tVChBB40UmaLa22HvpZT7RyfDHLA= X-Received: by 2002:aca:bb09:: with SMTP id l9mr1978200oif.33.1618898939329; Mon, 19 Apr 2021 23:08:59 -0700 (PDT) MIME-Version: 1.0 References: <1618542490-14756-1-git-send-email-wanpengli@tencent.com> <9c49c6ff-d896-e6a5-c051-b6707f6ec58a@redhat.com> In-Reply-To: From: Wanpeng Li Date: Tue, 20 Apr 2021 14:08:48 +0800 Message-ID: Subject: Re: [PATCH] KVM: Boost vCPU candidiate in user mode which is delivering interrupt To: Paolo Bonzini Cc: Sean Christopherson , LKML , kvm , Vitaly Kuznetsov , Wanpeng Li , Jim Mattson , Joerg Roedel Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, 20 Apr 2021 at 14:02, Wanpeng Li wrote: > > On Tue, 20 Apr 2021 at 00:59, Paolo Bonzini wrote: > > > > On 19/04/21 18:32, Sean Christopherson wrote: > > > If false positives are a big concern, what about adding another pass to the loop > > > and only yielding to usermode vCPUs with interrupts in the second full pass? > > > I.e. give vCPUs that are already in kernel mode priority, and only yield to > > > handle an interrupt if there are no vCPUs in kernel mode. > > > > > > kvm_arch_dy_runnable() pulls in pv_unhalted, which seems like a good thing. > > > > pv_unhalted won't help if you're waiting for a kernel spinlock though, > > would it? Doing two passes (or looking for a "best" candidate that > > prefers kernel mode vCPUs to user mode vCPUs waiting for an interrupt) > > seems like the best choice overall. > > How about something like this: Sorry, should be the codes below: diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h index 6b4dd95..9bc5f87 100644 --- a/include/linux/kvm_host.h +++ b/include/linux/kvm_host.h @@ -325,10 +325,12 @@ struct kvm_vcpu { * Cpu relax intercept or pause loop exit optimization * in_spin_loop: set when a vcpu does a pause loop exit * or cpu relax intercepted. + * pending_interrupt: set when a vcpu waiting for an interrupt * dy_eligible: indicates whether vcpu is eligible for directed yield. */ struct { bool in_spin_loop; + bool pending_interrupt; bool dy_eligible; } spin_loop; #endif @@ -1427,6 +1429,12 @@ static inline void kvm_vcpu_set_in_spin_loop(struct kvm_vcpu *vcpu, bool val) { vcpu->spin_loop.in_spin_loop = val; } + +static inline void kvm_vcpu_set_pending_interrupt(struct kvm_vcpu *vcpu, bool val) +{ + vcpu->spin_loop.pending_interrupt = val; +} + static inline void kvm_vcpu_set_dy_eligible(struct kvm_vcpu *vcpu, bool val) { vcpu->spin_loop.dy_eligible = val; @@ -1438,6 +1446,10 @@ static inline void kvm_vcpu_set_in_spin_loop(struct kvm_vcpu *vcpu, bool val) { } +static inline void kvm_vcpu_set_pending_interrupt(struct kvm_vcpu *vcpu, bool val) +{ +} + static inline void kvm_vcpu_set_dy_eligible(struct kvm_vcpu *vcpu, bool val) { } diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index 529cff1..bf6f1ec 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -410,6 +410,7 @@ static void kvm_vcpu_init(struct kvm_vcpu *vcpu, struct kvm *kvm, unsigned id) INIT_LIST_HEAD(&vcpu->blocked_vcpu_list); kvm_vcpu_set_in_spin_loop(vcpu, false); + kvm_vcpu_set_pending_interrupt(vcpu, false); kvm_vcpu_set_dy_eligible(vcpu, false); vcpu->preempted = false; vcpu->ready = false; @@ -3079,14 +3080,17 @@ EXPORT_SYMBOL_GPL(kvm_vcpu_yield_to); * Helper that checks whether a VCPU is eligible for directed yield. * Most eligible candidate to yield is decided by following heuristics: * - * (a) VCPU which has not done pl-exit or cpu relax intercepted recently - * (preempted lock holder), indicated by @in_spin_loop. - * Set at the beginning and cleared at the end of interception/PLE handler. + * (a) VCPU which has not done pl-exit or cpu relax intercepted and is not + * waiting for an interrupt recently (preempted lock holder). The former + * one is indicated by @in_spin_loop, set at the beginning and cleared at + * the end of interception/PLE handler. The later one is indicated by + * @pending_interrupt, set when interrupt is delivering and cleared at + * the end of directed yield. * - * (b) VCPU which has done pl-exit/ cpu relax intercepted but did not get - * chance last time (mostly it has become eligible now since we have probably - * yielded to lockholder in last iteration. This is done by toggling - * @dy_eligible each time a VCPU checked for eligibility.) + * (b) VCPU which has done pl-exit/ cpu relax intercepted or is waiting for + * interrupt but did not get chance last time (mostly it has become eligible + * now since we have probably yielded to lockholder in last iteration. This + * is done by toggling @dy_eligible each time a VCPU checked for eligibility.) * * Yielding to a recently pl-exited/cpu relax intercepted VCPU before yielding * to preempted lock-holder could result in wrong VCPU selection and CPU @@ -3102,10 +3106,10 @@ static bool kvm_vcpu_eligible_for_directed_yield(struct kvm_vcpu *vcpu) #ifdef CONFIG_HAVE_KVM_CPU_RELAX_INTERCEPT bool eligible; - eligible = !vcpu->spin_loop.in_spin_loop || + eligible = !(vcpu->spin_loop.in_spin_loop || vcpu->spin_loop.pending_interrupt) || vcpu->spin_loop.dy_eligible; - if (vcpu->spin_loop.in_spin_loop) + if (vcpu->spin_loop.in_spin_loop || vcpu->spin_loop.pending_interrupt) kvm_vcpu_set_dy_eligible(vcpu, !vcpu->spin_loop.dy_eligible); return eligible; @@ -3137,6 +3141,16 @@ static bool vcpu_dy_runnable(struct kvm_vcpu *vcpu) return false; } +static bool kvm_has_interrupt_delivery(struct kvm_vcpu *vcpu) +{ + if (vcpu_dy_runnable(vcpu)) { + kvm_vcpu_set_pending_interrupt(vcpu, true); + return true; + } + + return false; +} + void kvm_vcpu_on_spin(struct kvm_vcpu *me, bool yield_to_kernel_mode) { struct kvm *kvm = me->kvm; @@ -3170,6 +3184,7 @@ void kvm_vcpu_on_spin(struct kvm_vcpu *me, bool yield_to_kernel_mode) !vcpu_dy_runnable(vcpu)) continue; if (READ_ONCE(vcpu->preempted) && yield_to_kernel_mode && + !kvm_has_interrupt_delivery(vcpu) && !kvm_arch_vcpu_in_kernel(vcpu)) continue; if (!kvm_vcpu_eligible_for_directed_yield(vcpu)) @@ -3177,6 +3192,7 @@ void kvm_vcpu_on_spin(struct kvm_vcpu *me, bool yield_to_kernel_mode) yielded = kvm_vcpu_yield_to(vcpu); if (yielded > 0) { + kvm_vcpu_set_pending_interrupt(vcpu, false); kvm->last_boosted_vcpu = i; break; } else if (yielded < 0) {