2020-04-19 07:52:41

by Tianjia Zhang

[permalink] [raw]
Subject: [PATCH 1/7] KVM: s390: clean up redundant 'kvm_run' parameters

In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu'
structure. Earlier than historical reasons, many kvm-related function
parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time.
This patch does a unified cleanup of these remaining redundant parameters.

Signed-off-by: Tianjia Zhang <[email protected]>
---
arch/s390/kvm/kvm-s390.c | 127 +++++++++++++++++++++------------------
1 file changed, 67 insertions(+), 60 deletions(-)

diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
index 443af3ead739..cf420d013ba3 100644
--- a/arch/s390/kvm/kvm-s390.c
+++ b/arch/s390/kvm/kvm-s390.c
@@ -4173,24 +4173,25 @@ static int __vcpu_run(struct kvm_vcpu *vcpu)
return rc;
}

-static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
+static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
{
+ struct kvm_run *run = vcpu->run;
struct runtime_instr_cb *riccb;
struct gs_cb *gscb;

- riccb = (struct runtime_instr_cb *) &kvm_run->s.regs.riccb;
- gscb = (struct gs_cb *) &kvm_run->s.regs.gscb;
- vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
- vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
- if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
- vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
- vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
- vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
- }
- if (kvm_run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
- vcpu->arch.pfault_token = kvm_run->s.regs.pft;
- vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
- vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
+ riccb = (struct runtime_instr_cb *) &run->s.regs.riccb;
+ gscb = (struct gs_cb *) &run->s.regs.gscb;
+ vcpu->arch.sie_block->gpsw.mask = run->psw_mask;
+ vcpu->arch.sie_block->gpsw.addr = run->psw_addr;
+ if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
+ vcpu->arch.sie_block->todpr = run->s.regs.todpr;
+ vcpu->arch.sie_block->pp = run->s.regs.pp;
+ vcpu->arch.sie_block->gbea = run->s.regs.gbea;
+ }
+ if (run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
+ vcpu->arch.pfault_token = run->s.regs.pft;
+ vcpu->arch.pfault_select = run->s.regs.pfs;
+ vcpu->arch.pfault_compare = run->s.regs.pfc;
if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
kvm_clear_async_pf_completion_queue(vcpu);
}
@@ -4198,7 +4199,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
* If userspace sets the riccb (e.g. after migration) to a valid state,
* we should enable RI here instead of doing the lazy enablement.
*/
- if ((kvm_run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
+ if ((run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
test_kvm_facility(vcpu->kvm, 64) &&
riccb->v &&
!(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
@@ -4209,7 +4210,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
* If userspace sets the gscb (e.g. after migration) to non-zero,
* we should enable GS here instead of doing the lazy enablement.
*/
- if ((kvm_run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
+ if ((run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
test_kvm_facility(vcpu->kvm, 133) &&
gscb->gssm &&
!vcpu->arch.gs_enabled) {
@@ -4218,10 +4219,10 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
vcpu->arch.gs_enabled = 1;
}
- if ((kvm_run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
+ if ((run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
test_kvm_facility(vcpu->kvm, 82)) {
vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
- vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 0;
+ vcpu->arch.sie_block->fpf |= run->s.regs.bpbc ? FPF_BPBC : 0;
}
if (MACHINE_HAS_GS) {
preempt_disable();
@@ -4232,45 +4233,47 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
}
if (vcpu->arch.gs_enabled) {
current->thread.gs_cb = (struct gs_cb *)
- &vcpu->run->s.regs.gscb;
+ &run->s.regs.gscb;
restore_gs_cb(current->thread.gs_cb);
}
preempt_enable();
}
- /* SIE will load etoken directly from SDNX and therefore kvm_run */
+ /* SIE will load etoken directly from SDNX and therefore run */
}

-static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
+static void sync_regs(struct kvm_vcpu *vcpu)
{
- if (kvm_run->kvm_dirty_regs & KVM_SYNC_PREFIX)
- kvm_s390_set_prefix(vcpu, kvm_run->s.regs.prefix);
- if (kvm_run->kvm_dirty_regs & KVM_SYNC_CRS) {
- memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
+ struct kvm_run *run = vcpu->run;
+
+ if (run->kvm_dirty_regs & KVM_SYNC_PREFIX)
+ kvm_s390_set_prefix(vcpu, run->s.regs.prefix);
+ if (run->kvm_dirty_regs & KVM_SYNC_CRS) {
+ memcpy(&vcpu->arch.sie_block->gcr, &run->s.regs.crs, 128);
/* some control register changes require a tlb flush */
kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
}
- if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
- kvm_s390_set_cpu_timer(vcpu, kvm_run->s.regs.cputm);
- vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
+ if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
+ kvm_s390_set_cpu_timer(vcpu, run->s.regs.cputm);
+ vcpu->arch.sie_block->ckc = run->s.regs.ckc;
}
save_access_regs(vcpu->arch.host_acrs);
- restore_access_regs(vcpu->run->s.regs.acrs);
+ restore_access_regs(run->s.regs.acrs);
/* save host (userspace) fprs/vrs */
save_fpu_regs();
vcpu->arch.host_fpregs.fpc = current->thread.fpu.fpc;
vcpu->arch.host_fpregs.regs = current->thread.fpu.regs;
if (MACHINE_HAS_VX)
- current->thread.fpu.regs = vcpu->run->s.regs.vrs;
+ current->thread.fpu.regs = run->s.regs.vrs;
else
- current->thread.fpu.regs = vcpu->run->s.regs.fprs;
- current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
+ current->thread.fpu.regs = run->s.regs.fprs;
+ current->thread.fpu.fpc = run->s.regs.fpc;
if (test_fp_ctl(current->thread.fpu.fpc))
/* User space provided an invalid FPC, let's clear it */
current->thread.fpu.fpc = 0;

/* Sync fmt2 only data */
if (likely(!kvm_s390_pv_cpu_is_protected(vcpu))) {
- sync_regs_fmt2(vcpu, kvm_run);
+ sync_regs_fmt2(vcpu);
} else {
/*
* In several places we have to modify our internal view to
@@ -4282,19 +4285,21 @@ static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
* do only accept the condition code from userspace.
*/
vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
- vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
+ vcpu->arch.sie_block->gpsw.mask |= run->psw_mask &
PSW_MASK_CC;
}

- kvm_run->kvm_dirty_regs = 0;
+ run->kvm_dirty_regs = 0;
}

-static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
+static void store_regs_fmt2(struct kvm_vcpu *vcpu)
{
- kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
- kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
- kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
- kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
+ struct kvm_run *run = vcpu->run;
+
+ run->s.regs.todpr = vcpu->arch.sie_block->todpr;
+ run->s.regs.pp = vcpu->arch.sie_block->pp;
+ run->s.regs.gbea = vcpu->arch.sie_block->gbea;
+ run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
if (MACHINE_HAS_GS) {
__ctl_set_bit(2, 4);
if (vcpu->arch.gs_enabled)
@@ -4310,39 +4315,41 @@ static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
/* SIE will save etoken directly into SDNX and therefore kvm_run */
}

-static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
+static void store_regs(struct kvm_vcpu *vcpu)
{
- kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
- kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
- kvm_run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
- memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
- kvm_run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
- kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
- kvm_run->s.regs.pft = vcpu->arch.pfault_token;
- kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
- kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
- save_access_regs(vcpu->run->s.regs.acrs);
+ struct kvm_run *run = vcpu->run;
+
+ run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
+ run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
+ run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
+ memcpy(&run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
+ run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
+ run->s.regs.ckc = vcpu->arch.sie_block->ckc;
+ run->s.regs.pft = vcpu->arch.pfault_token;
+ run->s.regs.pfs = vcpu->arch.pfault_select;
+ run->s.regs.pfc = vcpu->arch.pfault_compare;
+ save_access_regs(run->s.regs.acrs);
restore_access_regs(vcpu->arch.host_acrs);
/* Save guest register state */
save_fpu_regs();
- vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
+ run->s.regs.fpc = current->thread.fpu.fpc;
/* Restore will be done lazily at return */
current->thread.fpu.fpc = vcpu->arch.host_fpregs.fpc;
current->thread.fpu.regs = vcpu->arch.host_fpregs.regs;
if (likely(!kvm_s390_pv_cpu_is_protected(vcpu)))
- store_regs_fmt2(vcpu, kvm_run);
+ store_regs_fmt2(vcpu);
}

int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
{
- struct kvm_run *kvm_run = vcpu->run;
+ struct kvm_run *run = vcpu->run;
int rc;

- if (kvm_run->immediate_exit)
+ if (run->immediate_exit)
return -EINTR;

- if (kvm_run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
- kvm_run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
+ if (run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
+ run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
return -EINVAL;

vcpu_load(vcpu);
@@ -4368,14 +4375,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
goto out;
}

- sync_regs(vcpu, kvm_run);
+ sync_regs(vcpu);
enable_cpu_timer_accounting(vcpu);

might_fault();
rc = __vcpu_run(vcpu);

if (signal_pending(current) && !rc) {
- kvm_run->exit_reason = KVM_EXIT_INTR;
+ run->exit_reason = KVM_EXIT_INTR;
rc = -EINTR;
}

@@ -4390,7 +4397,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
}

disable_cpu_timer_accounting(vcpu);
- store_regs(vcpu, kvm_run);
+ store_regs(vcpu);

kvm_sigset_deactivate(vcpu);

--
2.17.1


2020-04-20 07:11:48

by Christian Borntraeger

[permalink] [raw]
Subject: Re: [PATCH 1/7] KVM: s390: clean up redundant 'kvm_run' parameters



On 19.04.20 09:51, Tianjia Zhang wrote:
> In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu'
> structure. Earlier than historical reasons, many kvm-related function
> parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time.
> This patch does a unified cleanup of these remaining redundant parameters.
>
> Signed-off-by: Tianjia Zhang <[email protected]>
> ---
> arch/s390/kvm/kvm-s390.c | 127 +++++++++++++++++++++------------------
> 1 file changed, 67 insertions(+), 60 deletions(-)
>
> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
> index 443af3ead739..cf420d013ba3 100644
> --- a/arch/s390/kvm/kvm-s390.c
> +++ b/arch/s390/kvm/kvm-s390.c
> @@ -4173,24 +4173,25 @@ static int __vcpu_run(struct kvm_vcpu *vcpu)
> return rc;
> }
>
> -static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
> {
> + struct kvm_run *run = vcpu->run;

Please use kvm_run as variable name. This makes all of the changes below go away.


> struct runtime_instr_cb *riccb;
> struct gs_cb *gscb;
>
> - riccb = (struct runtime_instr_cb *) &kvm_run->s.regs.riccb;
> - gscb = (struct gs_cb *) &kvm_run->s.regs.gscb;
> - vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
> - vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> - vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
> - vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
> - vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
> - }
> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
> - vcpu->arch.pfault_token = kvm_run->s.regs.pft;
> - vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
> - vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
> + riccb = (struct runtime_instr_cb *) &run->s.regs.riccb;
> + gscb = (struct gs_cb *) &run->s.regs.gscb;
> + vcpu->arch.sie_block->gpsw.mask = run->psw_mask;
> + vcpu->arch.sie_block->gpsw.addr = run->psw_addr;
> + if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> + vcpu->arch.sie_block->todpr = run->s.regs.todpr;
> + vcpu->arch.sie_block->pp = run->s.regs.pp;
> + vcpu->arch.sie_block->gbea = run->s.regs.gbea;
> + }
> + if (run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
> + vcpu->arch.pfault_token = run->s.regs.pft;
> + vcpu->arch.pfault_select = run->s.regs.pfs;
> + vcpu->arch.pfault_compare = run->s.regs.pfc;
> if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
> kvm_clear_async_pf_completion_queue(vcpu);
> }
> @@ -4198,7 +4199,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> * If userspace sets the riccb (e.g. after migration) to a valid state,
> * we should enable RI here instead of doing the lazy enablement.
> */
> - if ((kvm_run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
> + if ((run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
> test_kvm_facility(vcpu->kvm, 64) &&
> riccb->v &&
> !(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
> @@ -4209,7 +4210,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> * If userspace sets the gscb (e.g. after migration) to non-zero,
> * we should enable GS here instead of doing the lazy enablement.
> */
> - if ((kvm_run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
> + if ((run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
> test_kvm_facility(vcpu->kvm, 133) &&
> gscb->gssm &&
> !vcpu->arch.gs_enabled) {
> @@ -4218,10 +4219,10 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
> vcpu->arch.gs_enabled = 1;
> }
> - if ((kvm_run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
> + if ((run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
> test_kvm_facility(vcpu->kvm, 82)) {
> vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
> - vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 0;
> + vcpu->arch.sie_block->fpf |= run->s.regs.bpbc ? FPF_BPBC : 0;
> }
> if (MACHINE_HAS_GS) {
> preempt_disable();
> @@ -4232,45 +4233,47 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> }
> if (vcpu->arch.gs_enabled) {
> current->thread.gs_cb = (struct gs_cb *)
> - &vcpu->run->s.regs.gscb;
> + &run->s.regs.gscb;
> restore_gs_cb(current->thread.gs_cb);
> }
> preempt_enable();
> }
> - /* SIE will load etoken directly from SDNX and therefore kvm_run */
> + /* SIE will load etoken directly from SDNX and therefore run */
> }
>
> -static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void sync_regs(struct kvm_vcpu *vcpu)
> {
> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_PREFIX)
> - kvm_s390_set_prefix(vcpu, kvm_run->s.regs.prefix);
> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_CRS) {
> - memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
> + struct kvm_run *run = vcpu->run;
> +
> + if (run->kvm_dirty_regs & KVM_SYNC_PREFIX)
> + kvm_s390_set_prefix(vcpu, run->s.regs.prefix);
> + if (run->kvm_dirty_regs & KVM_SYNC_CRS) {
> + memcpy(&vcpu->arch.sie_block->gcr, &run->s.regs.crs, 128);
> /* some control register changes require a tlb flush */
> kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
> }
> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> - kvm_s390_set_cpu_timer(vcpu, kvm_run->s.regs.cputm);
> - vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
> + if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> + kvm_s390_set_cpu_timer(vcpu, run->s.regs.cputm);
> + vcpu->arch.sie_block->ckc = run->s.regs.ckc;
> }
> save_access_regs(vcpu->arch.host_acrs);
> - restore_access_regs(vcpu->run->s.regs.acrs);
> + restore_access_regs(run->s.regs.acrs);
> /* save host (userspace) fprs/vrs */
> save_fpu_regs();
> vcpu->arch.host_fpregs.fpc = current->thread.fpu.fpc;
> vcpu->arch.host_fpregs.regs = current->thread.fpu.regs;
> if (MACHINE_HAS_VX)
> - current->thread.fpu.regs = vcpu->run->s.regs.vrs;
> + current->thread.fpu.regs = run->s.regs.vrs;
> else
> - current->thread.fpu.regs = vcpu->run->s.regs.fprs;
> - current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
> + current->thread.fpu.regs = run->s.regs.fprs;
> + current->thread.fpu.fpc = run->s.regs.fpc;
> if (test_fp_ctl(current->thread.fpu.fpc))
> /* User space provided an invalid FPC, let's clear it */
> current->thread.fpu.fpc = 0;
>
> /* Sync fmt2 only data */
> if (likely(!kvm_s390_pv_cpu_is_protected(vcpu))) {
> - sync_regs_fmt2(vcpu, kvm_run);
> + sync_regs_fmt2(vcpu);
> } else {
> /*
> * In several places we have to modify our internal view to
> @@ -4282,19 +4285,21 @@ static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> * do only accept the condition code from userspace.
> */
> vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
> - vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
> + vcpu->arch.sie_block->gpsw.mask |= run->psw_mask &
> PSW_MASK_CC;
> }
>
> - kvm_run->kvm_dirty_regs = 0;
> + run->kvm_dirty_regs = 0;
> }
>
> -static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void store_regs_fmt2(struct kvm_vcpu *vcpu)
> {
> - kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
> - kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
> - kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
> - kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
> + struct kvm_run *run = vcpu->run;
> +
> + run->s.regs.todpr = vcpu->arch.sie_block->todpr;
> + run->s.regs.pp = vcpu->arch.sie_block->pp;
> + run->s.regs.gbea = vcpu->arch.sie_block->gbea;
> + run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
> if (MACHINE_HAS_GS) {
> __ctl_set_bit(2, 4);
> if (vcpu->arch.gs_enabled)
> @@ -4310,39 +4315,41 @@ static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> /* SIE will save etoken directly into SDNX and therefore kvm_run */
> }
>
> -static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void store_regs(struct kvm_vcpu *vcpu)
> {
> - kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
> - kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
> - kvm_run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
> - memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
> - kvm_run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
> - kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
> - kvm_run->s.regs.pft = vcpu->arch.pfault_token;
> - kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
> - kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
> - save_access_regs(vcpu->run->s.regs.acrs);
> + struct kvm_run *run = vcpu->run;
> +
> + run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
> + run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
> + run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
> + memcpy(&run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
> + run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
> + run->s.regs.ckc = vcpu->arch.sie_block->ckc;
> + run->s.regs.pft = vcpu->arch.pfault_token;
> + run->s.regs.pfs = vcpu->arch.pfault_select;
> + run->s.regs.pfc = vcpu->arch.pfault_compare;
> + save_access_regs(run->s.regs.acrs);
> restore_access_regs(vcpu->arch.host_acrs);
> /* Save guest register state */
> save_fpu_regs();
> - vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
> + run->s.regs.fpc = current->thread.fpu.fpc;
> /* Restore will be done lazily at return */
> current->thread.fpu.fpc = vcpu->arch.host_fpregs.fpc;
> current->thread.fpu.regs = vcpu->arch.host_fpregs.regs;
> if (likely(!kvm_s390_pv_cpu_is_protected(vcpu)))
> - store_regs_fmt2(vcpu, kvm_run);
> + store_regs_fmt2(vcpu);
> }
>
> int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
> {
> - struct kvm_run *kvm_run = vcpu->run;
> + struct kvm_run *run = vcpu->run;
> int rc;
>
> - if (kvm_run->immediate_exit)
> + if (run->immediate_exit)
> return -EINTR;
>
> - if (kvm_run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
> - kvm_run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
> + if (run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
> + run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
> return -EINVAL;
>
> vcpu_load(vcpu);
> @@ -4368,14 +4375,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
> goto out;
> }
>
> - sync_regs(vcpu, kvm_run);
> + sync_regs(vcpu);
> enable_cpu_timer_accounting(vcpu);
>
> might_fault();
> rc = __vcpu_run(vcpu);
>
> if (signal_pending(current) && !rc) {
> - kvm_run->exit_reason = KVM_EXIT_INTR;
> + run->exit_reason = KVM_EXIT_INTR;
> rc = -EINTR;
> }
>
> @@ -4390,7 +4397,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
> }
>
> disable_cpu_timer_accounting(vcpu);
> - store_regs(vcpu, kvm_run);
> + store_regs(vcpu);
>
> kvm_sigset_deactivate(vcpu);
>
>

2020-04-20 07:19:33

by Tianjia Zhang

[permalink] [raw]
Subject: Re: [PATCH 1/7] KVM: s390: clean up redundant 'kvm_run' parameters



On 2020/4/20 15:07, Christian Borntraeger wrote:
>
>
> On 19.04.20 09:51, Tianjia Zhang wrote:
>> In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu'
>> structure. Earlier than historical reasons, many kvm-related function
>> parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time.
>> This patch does a unified cleanup of these remaining redundant parameters.
>>
>> Signed-off-by: Tianjia Zhang <[email protected]>
>> ---
>> arch/s390/kvm/kvm-s390.c | 127 +++++++++++++++++++++------------------
>> 1 file changed, 67 insertions(+), 60 deletions(-)
>>
>> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
>> index 443af3ead739..cf420d013ba3 100644
>> --- a/arch/s390/kvm/kvm-s390.c
>> +++ b/arch/s390/kvm/kvm-s390.c
>> @@ -4173,24 +4173,25 @@ static int __vcpu_run(struct kvm_vcpu *vcpu)
>> return rc;
>> }
>>
>> -static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
>> {
>> + struct kvm_run *run = vcpu->run;
>
> Please use kvm_run as variable name. This makes all of the changes below go away.
>

It's OK, I will fix it in v2 patch.

Thanks,
Tianjia

>
>> struct runtime_instr_cb *riccb;
>> struct gs_cb *gscb;
>>
>> - riccb = (struct runtime_instr_cb *) &kvm_run->s.regs.riccb;
>> - gscb = (struct gs_cb *) &kvm_run->s.regs.gscb;
>> - vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
>> - vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> - vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
>> - vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
>> - vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
>> - }
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
>> - vcpu->arch.pfault_token = kvm_run->s.regs.pft;
>> - vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
>> - vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
>> + riccb = (struct runtime_instr_cb *) &run->s.regs.riccb;
>> + gscb = (struct gs_cb *) &run->s.regs.gscb;
>> + vcpu->arch.sie_block->gpsw.mask = run->psw_mask;
>> + vcpu->arch.sie_block->gpsw.addr = run->psw_addr;
>> + if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> + vcpu->arch.sie_block->todpr = run->s.regs.todpr;
>> + vcpu->arch.sie_block->pp = run->s.regs.pp;
>> + vcpu->arch.sie_block->gbea = run->s.regs.gbea;
>> + }
>> + if (run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
>> + vcpu->arch.pfault_token = run->s.regs.pft;
>> + vcpu->arch.pfault_select = run->s.regs.pfs;
>> + vcpu->arch.pfault_compare = run->s.regs.pfc;
>> if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
>> kvm_clear_async_pf_completion_queue(vcpu);
>> }
>> @@ -4198,7 +4199,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> * If userspace sets the riccb (e.g. after migration) to a valid state,
>> * we should enable RI here instead of doing the lazy enablement.
>> */
>> - if ((kvm_run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>> + if ((run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>> test_kvm_facility(vcpu->kvm, 64) &&
>> riccb->v &&
>> !(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
>> @@ -4209,7 +4210,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> * If userspace sets the gscb (e.g. after migration) to non-zero,
>> * we should enable GS here instead of doing the lazy enablement.
>> */
>> - if ((kvm_run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>> + if ((run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>> test_kvm_facility(vcpu->kvm, 133) &&
>> gscb->gssm &&
>> !vcpu->arch.gs_enabled) {
>> @@ -4218,10 +4219,10 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
>> vcpu->arch.gs_enabled = 1;
>> }
>> - if ((kvm_run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>> + if ((run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>> test_kvm_facility(vcpu->kvm, 82)) {
>> vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
>> - vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 0;
>> + vcpu->arch.sie_block->fpf |= run->s.regs.bpbc ? FPF_BPBC : 0;
>> }
>> if (MACHINE_HAS_GS) {
>> preempt_disable();
>> @@ -4232,45 +4233,47 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> }
>> if (vcpu->arch.gs_enabled) {
>> current->thread.gs_cb = (struct gs_cb *)
>> - &vcpu->run->s.regs.gscb;
>> + &run->s.regs.gscb;
>> restore_gs_cb(current->thread.gs_cb);
>> }
>> preempt_enable();
>> }
>> - /* SIE will load etoken directly from SDNX and therefore kvm_run */
>> + /* SIE will load etoken directly from SDNX and therefore run */
>> }
>>
>> -static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void sync_regs(struct kvm_vcpu *vcpu)
>> {
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_PREFIX)
>> - kvm_s390_set_prefix(vcpu, kvm_run->s.regs.prefix);
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_CRS) {
>> - memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
>> + struct kvm_run *run = vcpu->run;
>> +
>> + if (run->kvm_dirty_regs & KVM_SYNC_PREFIX)
>> + kvm_s390_set_prefix(vcpu, run->s.regs.prefix);
>> + if (run->kvm_dirty_regs & KVM_SYNC_CRS) {
>> + memcpy(&vcpu->arch.sie_block->gcr, &run->s.regs.crs, 128);
>> /* some control register changes require a tlb flush */
>> kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
>> }
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> - kvm_s390_set_cpu_timer(vcpu, kvm_run->s.regs.cputm);
>> - vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
>> + if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> + kvm_s390_set_cpu_timer(vcpu, run->s.regs.cputm);
>> + vcpu->arch.sie_block->ckc = run->s.regs.ckc;
>> }
>> save_access_regs(vcpu->arch.host_acrs);
>> - restore_access_regs(vcpu->run->s.regs.acrs);
>> + restore_access_regs(run->s.regs.acrs);
>> /* save host (userspace) fprs/vrs */
>> save_fpu_regs();
>> vcpu->arch.host_fpregs.fpc = current->thread.fpu.fpc;
>> vcpu->arch.host_fpregs.regs = current->thread.fpu.regs;
>> if (MACHINE_HAS_VX)
>> - current->thread.fpu.regs = vcpu->run->s.regs.vrs;
>> + current->thread.fpu.regs = run->s.regs.vrs;
>> else
>> - current->thread.fpu.regs = vcpu->run->s.regs.fprs;
>> - current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
>> + current->thread.fpu.regs = run->s.regs.fprs;
>> + current->thread.fpu.fpc = run->s.regs.fpc;
>> if (test_fp_ctl(current->thread.fpu.fpc))
>> /* User space provided an invalid FPC, let's clear it */
>> current->thread.fpu.fpc = 0;
>>
>> /* Sync fmt2 only data */
>> if (likely(!kvm_s390_pv_cpu_is_protected(vcpu))) {
>> - sync_regs_fmt2(vcpu, kvm_run);
>> + sync_regs_fmt2(vcpu);
>> } else {
>> /*
>> * In several places we have to modify our internal view to
>> @@ -4282,19 +4285,21 @@ static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> * do only accept the condition code from userspace.
>> */
>> vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
>> - vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
>> + vcpu->arch.sie_block->gpsw.mask |= run->psw_mask &
>> PSW_MASK_CC;
>> }
>>
>> - kvm_run->kvm_dirty_regs = 0;
>> + run->kvm_dirty_regs = 0;
>> }
>>
>> -static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void store_regs_fmt2(struct kvm_vcpu *vcpu)
>> {
>> - kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
>> - kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
>> - kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
>> - kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>> + struct kvm_run *run = vcpu->run;
>> +
>> + run->s.regs.todpr = vcpu->arch.sie_block->todpr;
>> + run->s.regs.pp = vcpu->arch.sie_block->pp;
>> + run->s.regs.gbea = vcpu->arch.sie_block->gbea;
>> + run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>> if (MACHINE_HAS_GS) {
>> __ctl_set_bit(2, 4);
>> if (vcpu->arch.gs_enabled)
>> @@ -4310,39 +4315,41 @@ static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> /* SIE will save etoken directly into SDNX and therefore kvm_run */
>> }
>>
>> -static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void store_regs(struct kvm_vcpu *vcpu)
>> {
>> - kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
>> - kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
>> - kvm_run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
>> - memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
>> - kvm_run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
>> - kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
>> - kvm_run->s.regs.pft = vcpu->arch.pfault_token;
>> - kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
>> - kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
>> - save_access_regs(vcpu->run->s.regs.acrs);
>> + struct kvm_run *run = vcpu->run;
>> +
>> + run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
>> + run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
>> + run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
>> + memcpy(&run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
>> + run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
>> + run->s.regs.ckc = vcpu->arch.sie_block->ckc;
>> + run->s.regs.pft = vcpu->arch.pfault_token;
>> + run->s.regs.pfs = vcpu->arch.pfault_select;
>> + run->s.regs.pfc = vcpu->arch.pfault_compare;
>> + save_access_regs(run->s.regs.acrs);
>> restore_access_regs(vcpu->arch.host_acrs);
>> /* Save guest register state */
>> save_fpu_regs();
>> - vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
>> + run->s.regs.fpc = current->thread.fpu.fpc;
>> /* Restore will be done lazily at return */
>> current->thread.fpu.fpc = vcpu->arch.host_fpregs.fpc;
>> current->thread.fpu.regs = vcpu->arch.host_fpregs.regs;
>> if (likely(!kvm_s390_pv_cpu_is_protected(vcpu)))
>> - store_regs_fmt2(vcpu, kvm_run);
>> + store_regs_fmt2(vcpu);
>> }
>>
>> int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>> {
>> - struct kvm_run *kvm_run = vcpu->run;
>> + struct kvm_run *run = vcpu->run;
>> int rc;
>>
>> - if (kvm_run->immediate_exit)
>> + if (run->immediate_exit)
>> return -EINTR;
>>
>> - if (kvm_run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
>> - kvm_run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>> + if (run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
>> + run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>> return -EINVAL;
>>
>> vcpu_load(vcpu);
>> @@ -4368,14 +4375,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>> goto out;
>> }
>>
>> - sync_regs(vcpu, kvm_run);
>> + sync_regs(vcpu);
>> enable_cpu_timer_accounting(vcpu);
>>
>> might_fault();
>> rc = __vcpu_run(vcpu);
>>
>> if (signal_pending(current) && !rc) {
>> - kvm_run->exit_reason = KVM_EXIT_INTR;
>> + run->exit_reason = KVM_EXIT_INTR;
>> rc = -EINTR;
>> }
>>
>> @@ -4390,7 +4397,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>> }
>>
>> disable_cpu_timer_accounting(vcpu);
>> - store_regs(vcpu, kvm_run);
>> + store_regs(vcpu);
>>
>> kvm_sigset_deactivate(vcpu);
>>
>>

2020-04-20 08:27:32

by Janosch Frank

[permalink] [raw]
Subject: Re: [PATCH 1/7] KVM: s390: clean up redundant 'kvm_run' parameters

On 4/20/20 9:07 AM, Christian Borntraeger wrote:
>
>
> On 19.04.20 09:51, Tianjia Zhang wrote:
>> In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu'
>> structure. Earlier than historical reasons, many kvm-related function
>> parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time.
>> This patch does a unified cleanup of these remaining redundant parameters.
>>
>> Signed-off-by: Tianjia Zhang <[email protected]>
>> ---
>> arch/s390/kvm/kvm-s390.c | 127 +++++++++++++++++++++------------------
>> 1 file changed, 67 insertions(+), 60 deletions(-)
>>
>> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
>> index 443af3ead739..cf420d013ba3 100644
>> --- a/arch/s390/kvm/kvm-s390.c
>> +++ b/arch/s390/kvm/kvm-s390.c
>> @@ -4173,24 +4173,25 @@ static int __vcpu_run(struct kvm_vcpu *vcpu)
>> return rc;
>> }
>>
>> -static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
>> {
>> + struct kvm_run *run = vcpu->run;
>
> Please use kvm_run as variable name. This makes all of the changes below go away.

Definitely

@Tianjia
This patch produces way to much churn to be picked up as is.
I do not want patches that touch so much for small cosmetic changes.

>
>
>> struct runtime_instr_cb *riccb;
>> struct gs_cb *gscb;
>>
>> - riccb = (struct runtime_instr_cb *) &kvm_run->s.regs.riccb;
>> - gscb = (struct gs_cb *) &kvm_run->s.regs.gscb;
>> - vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
>> - vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> - vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
>> - vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
>> - vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
>> - }
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
>> - vcpu->arch.pfault_token = kvm_run->s.regs.pft;
>> - vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
>> - vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
>> + riccb = (struct runtime_instr_cb *) &run->s.regs.riccb;
>> + gscb = (struct gs_cb *) &run->s.regs.gscb;
>> + vcpu->arch.sie_block->gpsw.mask = run->psw_mask;
>> + vcpu->arch.sie_block->gpsw.addr = run->psw_addr;
>> + if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> + vcpu->arch.sie_block->todpr = run->s.regs.todpr;
>> + vcpu->arch.sie_block->pp = run->s.regs.pp;
>> + vcpu->arch.sie_block->gbea = run->s.regs.gbea;
>> + }
>> + if (run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
>> + vcpu->arch.pfault_token = run->s.regs.pft;
>> + vcpu->arch.pfault_select = run->s.regs.pfs;
>> + vcpu->arch.pfault_compare = run->s.regs.pfc;
>> if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
>> kvm_clear_async_pf_completion_queue(vcpu);
>> }
>> @@ -4198,7 +4199,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> * If userspace sets the riccb (e.g. after migration) to a valid state,
>> * we should enable RI here instead of doing the lazy enablement.
>> */
>> - if ((kvm_run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>> + if ((run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>> test_kvm_facility(vcpu->kvm, 64) &&
>> riccb->v &&
>> !(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
>> @@ -4209,7 +4210,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> * If userspace sets the gscb (e.g. after migration) to non-zero,
>> * we should enable GS here instead of doing the lazy enablement.
>> */
>> - if ((kvm_run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>> + if ((run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>> test_kvm_facility(vcpu->kvm, 133) &&
>> gscb->gssm &&
>> !vcpu->arch.gs_enabled) {
>> @@ -4218,10 +4219,10 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
>> vcpu->arch.gs_enabled = 1;
>> }
>> - if ((kvm_run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>> + if ((run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>> test_kvm_facility(vcpu->kvm, 82)) {
>> vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
>> - vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 0;
>> + vcpu->arch.sie_block->fpf |= run->s.regs.bpbc ? FPF_BPBC : 0;
>> }
>> if (MACHINE_HAS_GS) {
>> preempt_disable();
>> @@ -4232,45 +4233,47 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> }
>> if (vcpu->arch.gs_enabled) {
>> current->thread.gs_cb = (struct gs_cb *)
>> - &vcpu->run->s.regs.gscb;
>> + &run->s.regs.gscb;
>> restore_gs_cb(current->thread.gs_cb);
>> }
>> preempt_enable();
>> }
>> - /* SIE will load etoken directly from SDNX and therefore kvm_run */
>> + /* SIE will load etoken directly from SDNX and therefore run */
>> }
>>
>> -static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void sync_regs(struct kvm_vcpu *vcpu)
>> {
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_PREFIX)
>> - kvm_s390_set_prefix(vcpu, kvm_run->s.regs.prefix);
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_CRS) {
>> - memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
>> + struct kvm_run *run = vcpu->run;
>> +
>> + if (run->kvm_dirty_regs & KVM_SYNC_PREFIX)
>> + kvm_s390_set_prefix(vcpu, run->s.regs.prefix);
>> + if (run->kvm_dirty_regs & KVM_SYNC_CRS) {
>> + memcpy(&vcpu->arch.sie_block->gcr, &run->s.regs.crs, 128);
>> /* some control register changes require a tlb flush */
>> kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
>> }
>> - if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> - kvm_s390_set_cpu_timer(vcpu, kvm_run->s.regs.cputm);
>> - vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
>> + if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
>> + kvm_s390_set_cpu_timer(vcpu, run->s.regs.cputm);
>> + vcpu->arch.sie_block->ckc = run->s.regs.ckc;
>> }
>> save_access_regs(vcpu->arch.host_acrs);
>> - restore_access_regs(vcpu->run->s.regs.acrs);
>> + restore_access_regs(run->s.regs.acrs);
>> /* save host (userspace) fprs/vrs */
>> save_fpu_regs();
>> vcpu->arch.host_fpregs.fpc = current->thread.fpu.fpc;
>> vcpu->arch.host_fpregs.regs = current->thread.fpu.regs;
>> if (MACHINE_HAS_VX)
>> - current->thread.fpu.regs = vcpu->run->s.regs.vrs;
>> + current->thread.fpu.regs = run->s.regs.vrs;
>> else
>> - current->thread.fpu.regs = vcpu->run->s.regs.fprs;
>> - current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
>> + current->thread.fpu.regs = run->s.regs.fprs;
>> + current->thread.fpu.fpc = run->s.regs.fpc;
>> if (test_fp_ctl(current->thread.fpu.fpc))
>> /* User space provided an invalid FPC, let's clear it */
>> current->thread.fpu.fpc = 0;
>>
>> /* Sync fmt2 only data */
>> if (likely(!kvm_s390_pv_cpu_is_protected(vcpu))) {
>> - sync_regs_fmt2(vcpu, kvm_run);
>> + sync_regs_fmt2(vcpu);
>> } else {
>> /*
>> * In several places we have to modify our internal view to
>> @@ -4282,19 +4285,21 @@ static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> * do only accept the condition code from userspace.
>> */
>> vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
>> - vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
>> + vcpu->arch.sie_block->gpsw.mask |= run->psw_mask &
>> PSW_MASK_CC;
>> }
>>
>> - kvm_run->kvm_dirty_regs = 0;
>> + run->kvm_dirty_regs = 0;
>> }
>>
>> -static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void store_regs_fmt2(struct kvm_vcpu *vcpu)
>> {
>> - kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
>> - kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
>> - kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
>> - kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>> + struct kvm_run *run = vcpu->run;
>> +
>> + run->s.regs.todpr = vcpu->arch.sie_block->todpr;
>> + run->s.regs.pp = vcpu->arch.sie_block->pp;
>> + run->s.regs.gbea = vcpu->arch.sie_block->gbea;
>> + run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>> if (MACHINE_HAS_GS) {
>> __ctl_set_bit(2, 4);
>> if (vcpu->arch.gs_enabled)
>> @@ -4310,39 +4315,41 @@ static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> /* SIE will save etoken directly into SDNX and therefore kvm_run */
>> }
>>
>> -static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
>> +static void store_regs(struct kvm_vcpu *vcpu)
>> {
>> - kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
>> - kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
>> - kvm_run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
>> - memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
>> - kvm_run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
>> - kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
>> - kvm_run->s.regs.pft = vcpu->arch.pfault_token;
>> - kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
>> - kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
>> - save_access_regs(vcpu->run->s.regs.acrs);
>> + struct kvm_run *run = vcpu->run;
>> +
>> + run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
>> + run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
>> + run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
>> + memcpy(&run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
>> + run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
>> + run->s.regs.ckc = vcpu->arch.sie_block->ckc;
>> + run->s.regs.pft = vcpu->arch.pfault_token;
>> + run->s.regs.pfs = vcpu->arch.pfault_select;
>> + run->s.regs.pfc = vcpu->arch.pfault_compare;
>> + save_access_regs(run->s.regs.acrs);
>> restore_access_regs(vcpu->arch.host_acrs);
>> /* Save guest register state */
>> save_fpu_regs();
>> - vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
>> + run->s.regs.fpc = current->thread.fpu.fpc;
>> /* Restore will be done lazily at return */
>> current->thread.fpu.fpc = vcpu->arch.host_fpregs.fpc;
>> current->thread.fpu.regs = vcpu->arch.host_fpregs.regs;
>> if (likely(!kvm_s390_pv_cpu_is_protected(vcpu)))
>> - store_regs_fmt2(vcpu, kvm_run);
>> + store_regs_fmt2(vcpu);
>> }
>>
>> int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>> {
>> - struct kvm_run *kvm_run = vcpu->run;
>> + struct kvm_run *run = vcpu->run;
>> int rc;
>>
>> - if (kvm_run->immediate_exit)
>> + if (run->immediate_exit)
>> return -EINTR;
>>
>> - if (kvm_run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
>> - kvm_run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>> + if (run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
>> + run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>> return -EINVAL;
>>
>> vcpu_load(vcpu);
>> @@ -4368,14 +4375,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>> goto out;
>> }
>>
>> - sync_regs(vcpu, kvm_run);
>> + sync_regs(vcpu);
>> enable_cpu_timer_accounting(vcpu);
>>
>> might_fault();
>> rc = __vcpu_run(vcpu);
>>
>> if (signal_pending(current) && !rc) {
>> - kvm_run->exit_reason = KVM_EXIT_INTR;
>> + run->exit_reason = KVM_EXIT_INTR;
>> rc = -EINTR;
>> }
>>
>> @@ -4390,7 +4397,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>> }
>>
>> disable_cpu_timer_accounting(vcpu);
>> - store_regs(vcpu, kvm_run);
>> + store_regs(vcpu);
>>
>> kvm_sigset_deactivate(vcpu);
>>
>>



Attachments:
signature.asc (849.00 B)
OpenPGP digital signature