2019-05-30 01:07:08

by Wanpeng Li

[permalink] [raw]
Subject: [PATCH v3 2/3] KVM: X86: Implement PV sched yield hypercall

From: Wanpeng Li <[email protected]>

The target vCPUs are in runnable state after vcpu_kick and suitable
as a yield target. This patch implements the sched yield hypercall.

17% performance increasement of ebizzy benchmark can be observed in an
over-subscribe environment. (w/ kvm-pv-tlb disabled, testing TLB flush
call-function IPI-many since call-function is not easy to be trigged
by userspace workload).

Cc: Paolo Bonzini <[email protected]>
Cc: Radim Krčmář <[email protected]>
Cc: Liran Alon <[email protected]>
Signed-off-by: Wanpeng Li <[email protected]>
---
arch/x86/kvm/x86.c | 26 ++++++++++++++++++++++++++
1 file changed, 26 insertions(+)

diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index e7e57de..8575b36 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -7172,6 +7172,28 @@ void kvm_vcpu_deactivate_apicv(struct kvm_vcpu *vcpu)
kvm_x86_ops->refresh_apicv_exec_ctrl(vcpu);
}

+static void kvm_sched_yield(struct kvm *kvm, unsigned long dest_id)
+{
+ struct kvm_vcpu *target = NULL;
+ struct kvm_apic_map *map = NULL;
+
+ rcu_read_lock();
+ map = rcu_dereference(kvm->arch.apic_map);
+
+ if (unlikely(!map) || dest_id > map->max_apic_id)
+ goto out;
+
+ if (map->phys_map[dest_id]->vcpu) {
+ target = map->phys_map[dest_id]->vcpu;
+ rcu_read_unlock();
+ kvm_vcpu_yield_to(target);
+ }
+
+out:
+ if (!target)
+ rcu_read_unlock();
+}
+
int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
{
unsigned long nr, a0, a1, a2, a3, ret;
@@ -7218,6 +7240,10 @@ int kvm_emulate_hypercall(struct kvm_vcpu *vcpu)
case KVM_HC_SEND_IPI:
ret = kvm_pv_send_ipi(vcpu->kvm, a0, a1, a2, a3, op_64_bit);
break;
+ case KVM_HC_SCHED_YIELD:
+ kvm_sched_yield(vcpu->kvm, a0);
+ ret = 0;
+ break;
default:
ret = -KVM_ENOSYS;
break;
--
2.7.4


2019-06-10 14:17:50

by Radim Krčmář

[permalink] [raw]
Subject: Re: [PATCH v3 2/3] KVM: X86: Implement PV sched yield hypercall

2019-05-30 09:05+0800, Wanpeng Li:
> From: Wanpeng Li <[email protected]>
>
> The target vCPUs are in runnable state after vcpu_kick and suitable
> as a yield target. This patch implements the sched yield hypercall.
>
> 17% performance increasement of ebizzy benchmark can be observed in an
> over-subscribe environment. (w/ kvm-pv-tlb disabled, testing TLB flush
> call-function IPI-many since call-function is not easy to be trigged
> by userspace workload).
>
> Cc: Paolo Bonzini <[email protected]>
> Cc: Radim Krčmář <[email protected]>
> Cc: Liran Alon <[email protected]>
> Signed-off-by: Wanpeng Li <[email protected]>
> ---
> diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> @@ -7172,6 +7172,28 @@ void kvm_vcpu_deactivate_apicv(struct kvm_vcpu *vcpu)
> kvm_x86_ops->refresh_apicv_exec_ctrl(vcpu);
> }
>
> +static void kvm_sched_yield(struct kvm *kvm, unsigned long dest_id)
> +{
> + struct kvm_vcpu *target = NULL;
> + struct kvm_apic_map *map = NULL;
> +
> + rcu_read_lock();
> + map = rcu_dereference(kvm->arch.apic_map);
> +
> + if (unlikely(!map) || dest_id > map->max_apic_id)
> + goto out;
> +
> + if (map->phys_map[dest_id]->vcpu) {

This should check for map->phys_map[dest_id].

> + target = map->phys_map[dest_id]->vcpu;
> + rcu_read_unlock();
> + kvm_vcpu_yield_to(target);
> + }
> +
> +out:
> + if (!target)
> + rcu_read_unlock();

Also, I find the following logic clearer

{
struct kvm_vcpu *target = NULL;
struct kvm_apic_map *map;

rcu_read_lock();
map = rcu_dereference(kvm->arch.apic_map);

if (likely(map) && dest_id <= map->max_apic_id && map->phys_map[dest_id])
target = map->phys_map[dest_id]->vcpu;

rcu_read_unlock();

if (target)
kvm_vcpu_yield_to(target);
}

thanks.

2019-06-11 08:47:26

by Wanpeng Li

[permalink] [raw]
Subject: Re: [PATCH v3 2/3] KVM: X86: Implement PV sched yield hypercall

On Mon, 10 Jun 2019 at 22:17, Radim Krčmář <[email protected]> wrote:
>
> 2019-05-30 09:05+0800, Wanpeng Li:
> > From: Wanpeng Li <[email protected]>
> >
> > The target vCPUs are in runnable state after vcpu_kick and suitable
> > as a yield target. This patch implements the sched yield hypercall.
> >
> > 17% performance increasement of ebizzy benchmark can be observed in an
> > over-subscribe environment. (w/ kvm-pv-tlb disabled, testing TLB flush
> > call-function IPI-many since call-function is not easy to be trigged
> > by userspace workload).
> >
> > Cc: Paolo Bonzini <[email protected]>
> > Cc: Radim Krčmář <[email protected]>
> > Cc: Liran Alon <[email protected]>
> > Signed-off-by: Wanpeng Li <[email protected]>
> > ---
> > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
> > @@ -7172,6 +7172,28 @@ void kvm_vcpu_deactivate_apicv(struct kvm_vcpu *vcpu)
> > kvm_x86_ops->refresh_apicv_exec_ctrl(vcpu);
> > }
> >
> > +static void kvm_sched_yield(struct kvm *kvm, unsigned long dest_id)
> > +{
> > + struct kvm_vcpu *target = NULL;
> > + struct kvm_apic_map *map = NULL;
> > +
> > + rcu_read_lock();
> > + map = rcu_dereference(kvm->arch.apic_map);
> > +
> > + if (unlikely(!map) || dest_id > map->max_apic_id)
> > + goto out;
> > +
> > + if (map->phys_map[dest_id]->vcpu) {
>
> This should check for map->phys_map[dest_id].

Yeah, make a mistake here.

>
> > + target = map->phys_map[dest_id]->vcpu;
> > + rcu_read_unlock();
> > + kvm_vcpu_yield_to(target);
> > + }
> > +
> > +out:
> > + if (!target)
> > + rcu_read_unlock();
>
> Also, I find the following logic clearer
>
> {
> struct kvm_vcpu *target = NULL;
> struct kvm_apic_map *map;
>
> rcu_read_lock();
> map = rcu_dereference(kvm->arch.apic_map);
>
> if (likely(map) && dest_id <= map->max_apic_id && map->phys_map[dest_id])
> target = map->phys_map[dest_id]->vcpu;
>
> rcu_read_unlock();
>
> if (target)
> kvm_vcpu_yield_to(target);
> }

More better, thanks.

Regards,
Wanpeng Li