On 19.04.20 09:51, Tianjia Zhang wrote:
> In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu'
> structure. Earlier than historical reasons, many kvm-related function
> parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time.
> This patch does a unified cleanup of these remaining redundant parameters.
> 
> Signed-off-by: Tianjia Zhang <tianjia.zh...@linux.alibaba.com>
> ---
>  arch/s390/kvm/kvm-s390.c | 127 +++++++++++++++++++++------------------
>  1 file changed, 67 insertions(+), 60 deletions(-)
> 
> diff --git a/arch/s390/kvm/kvm-s390.c b/arch/s390/kvm/kvm-s390.c
> index 443af3ead739..cf420d013ba3 100644
> --- a/arch/s390/kvm/kvm-s390.c
> +++ b/arch/s390/kvm/kvm-s390.c
> @@ -4173,24 +4173,25 @@ static int __vcpu_run(struct kvm_vcpu *vcpu)
>       return rc;
>  }
>  
> -static void sync_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void sync_regs_fmt2(struct kvm_vcpu *vcpu)
>  {
> +     struct kvm_run *run = vcpu->run;

Please use kvm_run as variable name. This makes all of the changes below go 
away.


>       struct runtime_instr_cb *riccb;
>       struct gs_cb *gscb;
>  
> -     riccb = (struct runtime_instr_cb *) &kvm_run->s.regs.riccb;
> -     gscb = (struct gs_cb *) &kvm_run->s.regs.gscb;
> -     vcpu->arch.sie_block->gpsw.mask = kvm_run->psw_mask;
> -     vcpu->arch.sie_block->gpsw.addr = kvm_run->psw_addr;
> -     if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> -             vcpu->arch.sie_block->todpr = kvm_run->s.regs.todpr;
> -             vcpu->arch.sie_block->pp = kvm_run->s.regs.pp;
> -             vcpu->arch.sie_block->gbea = kvm_run->s.regs.gbea;
> -     }
> -     if (kvm_run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
> -             vcpu->arch.pfault_token = kvm_run->s.regs.pft;
> -             vcpu->arch.pfault_select = kvm_run->s.regs.pfs;
> -             vcpu->arch.pfault_compare = kvm_run->s.regs.pfc;
> +     riccb = (struct runtime_instr_cb *) &run->s.regs.riccb;
> +     gscb = (struct gs_cb *) &run->s.regs.gscb;
> +     vcpu->arch.sie_block->gpsw.mask = run->psw_mask;
> +     vcpu->arch.sie_block->gpsw.addr = run->psw_addr;
> +     if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> +             vcpu->arch.sie_block->todpr = run->s.regs.todpr;
> +             vcpu->arch.sie_block->pp = run->s.regs.pp;
> +             vcpu->arch.sie_block->gbea = run->s.regs.gbea;
> +     }
> +     if (run->kvm_dirty_regs & KVM_SYNC_PFAULT) {
> +             vcpu->arch.pfault_token = run->s.regs.pft;
> +             vcpu->arch.pfault_select = run->s.regs.pfs;
> +             vcpu->arch.pfault_compare = run->s.regs.pfc;
>               if (vcpu->arch.pfault_token == KVM_S390_PFAULT_TOKEN_INVALID)
>                       kvm_clear_async_pf_completion_queue(vcpu);
>       }
> @@ -4198,7 +4199,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, 
> struct kvm_run *kvm_run)
>        * If userspace sets the riccb (e.g. after migration) to a valid state,
>        * we should enable RI here instead of doing the lazy enablement.
>        */
> -     if ((kvm_run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
> +     if ((run->kvm_dirty_regs & KVM_SYNC_RICCB) &&
>           test_kvm_facility(vcpu->kvm, 64) &&
>           riccb->v &&
>           !(vcpu->arch.sie_block->ecb3 & ECB3_RI)) {
> @@ -4209,7 +4210,7 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, 
> struct kvm_run *kvm_run)
>        * If userspace sets the gscb (e.g. after migration) to non-zero,
>        * we should enable GS here instead of doing the lazy enablement.
>        */
> -     if ((kvm_run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
> +     if ((run->kvm_dirty_regs & KVM_SYNC_GSCB) &&
>           test_kvm_facility(vcpu->kvm, 133) &&
>           gscb->gssm &&
>           !vcpu->arch.gs_enabled) {
> @@ -4218,10 +4219,10 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, 
> struct kvm_run *kvm_run)
>               vcpu->arch.sie_block->ecd |= ECD_HOSTREGMGMT;
>               vcpu->arch.gs_enabled = 1;
>       }
> -     if ((kvm_run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
> +     if ((run->kvm_dirty_regs & KVM_SYNC_BPBC) &&
>           test_kvm_facility(vcpu->kvm, 82)) {
>               vcpu->arch.sie_block->fpf &= ~FPF_BPBC;
> -             vcpu->arch.sie_block->fpf |= kvm_run->s.regs.bpbc ? FPF_BPBC : 
> 0;
> +             vcpu->arch.sie_block->fpf |= run->s.regs.bpbc ? FPF_BPBC : 0;
>       }
>       if (MACHINE_HAS_GS) {
>               preempt_disable();
> @@ -4232,45 +4233,47 @@ static void sync_regs_fmt2(struct kvm_vcpu *vcpu, 
> struct kvm_run *kvm_run)
>               }
>               if (vcpu->arch.gs_enabled) {
>                       current->thread.gs_cb = (struct gs_cb *)
> -                                             &vcpu->run->s.regs.gscb;
> +                                             &run->s.regs.gscb;
>                       restore_gs_cb(current->thread.gs_cb);
>               }
>               preempt_enable();
>       }
> -     /* SIE will load etoken directly from SDNX and therefore kvm_run */
> +     /* SIE will load etoken directly from SDNX and therefore run */
>  }
>  
> -static void sync_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void sync_regs(struct kvm_vcpu *vcpu)
>  {
> -     if (kvm_run->kvm_dirty_regs & KVM_SYNC_PREFIX)
> -             kvm_s390_set_prefix(vcpu, kvm_run->s.regs.prefix);
> -     if (kvm_run->kvm_dirty_regs & KVM_SYNC_CRS) {
> -             memcpy(&vcpu->arch.sie_block->gcr, &kvm_run->s.regs.crs, 128);
> +     struct kvm_run *run = vcpu->run;
> +
> +     if (run->kvm_dirty_regs & KVM_SYNC_PREFIX)
> +             kvm_s390_set_prefix(vcpu, run->s.regs.prefix);
> +     if (run->kvm_dirty_regs & KVM_SYNC_CRS) {
> +             memcpy(&vcpu->arch.sie_block->gcr, &run->s.regs.crs, 128);
>               /* some control register changes require a tlb flush */
>               kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
>       }
> -     if (kvm_run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> -             kvm_s390_set_cpu_timer(vcpu, kvm_run->s.regs.cputm);
> -             vcpu->arch.sie_block->ckc = kvm_run->s.regs.ckc;
> +     if (run->kvm_dirty_regs & KVM_SYNC_ARCH0) {
> +             kvm_s390_set_cpu_timer(vcpu, run->s.regs.cputm);
> +             vcpu->arch.sie_block->ckc = run->s.regs.ckc;
>       }
>       save_access_regs(vcpu->arch.host_acrs);
> -     restore_access_regs(vcpu->run->s.regs.acrs);
> +     restore_access_regs(run->s.regs.acrs);
>       /* save host (userspace) fprs/vrs */
>       save_fpu_regs();
>       vcpu->arch.host_fpregs.fpc = current->thread.fpu.fpc;
>       vcpu->arch.host_fpregs.regs = current->thread.fpu.regs;
>       if (MACHINE_HAS_VX)
> -             current->thread.fpu.regs = vcpu->run->s.regs.vrs;
> +             current->thread.fpu.regs = run->s.regs.vrs;
>       else
> -             current->thread.fpu.regs = vcpu->run->s.regs.fprs;
> -     current->thread.fpu.fpc = vcpu->run->s.regs.fpc;
> +             current->thread.fpu.regs = run->s.regs.fprs;
> +     current->thread.fpu.fpc = run->s.regs.fpc;
>       if (test_fp_ctl(current->thread.fpu.fpc))
>               /* User space provided an invalid FPC, let's clear it */
>               current->thread.fpu.fpc = 0;
>  
>       /* Sync fmt2 only data */
>       if (likely(!kvm_s390_pv_cpu_is_protected(vcpu))) {
> -             sync_regs_fmt2(vcpu, kvm_run);
> +             sync_regs_fmt2(vcpu);
>       } else {
>               /*
>                * In several places we have to modify our internal view to
> @@ -4282,19 +4285,21 @@ static void sync_regs(struct kvm_vcpu *vcpu, struct 
> kvm_run *kvm_run)
>                * do only accept the condition code from userspace.
>                */
>               vcpu->arch.sie_block->gpsw.mask &= ~PSW_MASK_CC;
> -             vcpu->arch.sie_block->gpsw.mask |= kvm_run->psw_mask &
> +             vcpu->arch.sie_block->gpsw.mask |= run->psw_mask &
>                                                  PSW_MASK_CC;
>       }
>  
> -     kvm_run->kvm_dirty_regs = 0;
> +     run->kvm_dirty_regs = 0;
>  }
>  
> -static void store_regs_fmt2(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void store_regs_fmt2(struct kvm_vcpu *vcpu)
>  {
> -     kvm_run->s.regs.todpr = vcpu->arch.sie_block->todpr;
> -     kvm_run->s.regs.pp = vcpu->arch.sie_block->pp;
> -     kvm_run->s.regs.gbea = vcpu->arch.sie_block->gbea;
> -     kvm_run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == 
> FPF_BPBC;
> +     struct kvm_run *run = vcpu->run;
> +
> +     run->s.regs.todpr = vcpu->arch.sie_block->todpr;
> +     run->s.regs.pp = vcpu->arch.sie_block->pp;
> +     run->s.regs.gbea = vcpu->arch.sie_block->gbea;
> +     run->s.regs.bpbc = (vcpu->arch.sie_block->fpf & FPF_BPBC) == FPF_BPBC;
>       if (MACHINE_HAS_GS) {
>               __ctl_set_bit(2, 4);
>               if (vcpu->arch.gs_enabled)
> @@ -4310,39 +4315,41 @@ static void store_regs_fmt2(struct kvm_vcpu *vcpu, 
> struct kvm_run *kvm_run)
>       /* SIE will save etoken directly into SDNX and therefore kvm_run */
>  }
>  
> -static void store_regs(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
> +static void store_regs(struct kvm_vcpu *vcpu)
>  {
> -     kvm_run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
> -     kvm_run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
> -     kvm_run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
> -     memcpy(&kvm_run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
> -     kvm_run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
> -     kvm_run->s.regs.ckc = vcpu->arch.sie_block->ckc;
> -     kvm_run->s.regs.pft = vcpu->arch.pfault_token;
> -     kvm_run->s.regs.pfs = vcpu->arch.pfault_select;
> -     kvm_run->s.regs.pfc = vcpu->arch.pfault_compare;
> -     save_access_regs(vcpu->run->s.regs.acrs);
> +     struct kvm_run *run = vcpu->run;
> +
> +     run->psw_mask = vcpu->arch.sie_block->gpsw.mask;
> +     run->psw_addr = vcpu->arch.sie_block->gpsw.addr;
> +     run->s.regs.prefix = kvm_s390_get_prefix(vcpu);
> +     memcpy(&run->s.regs.crs, &vcpu->arch.sie_block->gcr, 128);
> +     run->s.regs.cputm = kvm_s390_get_cpu_timer(vcpu);
> +     run->s.regs.ckc = vcpu->arch.sie_block->ckc;
> +     run->s.regs.pft = vcpu->arch.pfault_token;
> +     run->s.regs.pfs = vcpu->arch.pfault_select;
> +     run->s.regs.pfc = vcpu->arch.pfault_compare;
> +     save_access_regs(run->s.regs.acrs);
>       restore_access_regs(vcpu->arch.host_acrs);
>       /* Save guest register state */
>       save_fpu_regs();
> -     vcpu->run->s.regs.fpc = current->thread.fpu.fpc;
> +     run->s.regs.fpc = current->thread.fpu.fpc;
>       /* Restore will be done lazily at return */
>       current->thread.fpu.fpc = vcpu->arch.host_fpregs.fpc;
>       current->thread.fpu.regs = vcpu->arch.host_fpregs.regs;
>       if (likely(!kvm_s390_pv_cpu_is_protected(vcpu)))
> -             store_regs_fmt2(vcpu, kvm_run);
> +             store_regs_fmt2(vcpu);
>  }
>  
>  int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>  {
> -     struct kvm_run *kvm_run = vcpu->run;
> +     struct kvm_run *run = vcpu->run;
>       int rc;
>  
> -     if (kvm_run->immediate_exit)
> +     if (run->immediate_exit)
>               return -EINTR;
>  
> -     if (kvm_run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
> -         kvm_run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
> +     if (run->kvm_valid_regs & ~KVM_SYNC_S390_VALID_FIELDS ||
> +         run->kvm_dirty_regs & ~KVM_SYNC_S390_VALID_FIELDS)
>               return -EINVAL;
>  
>       vcpu_load(vcpu);
> @@ -4368,14 +4375,14 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>               goto out;
>       }
>  
> -     sync_regs(vcpu, kvm_run);
> +     sync_regs(vcpu);
>       enable_cpu_timer_accounting(vcpu);
>  
>       might_fault();
>       rc = __vcpu_run(vcpu);
>  
>       if (signal_pending(current) && !rc) {
> -             kvm_run->exit_reason = KVM_EXIT_INTR;
> +             run->exit_reason = KVM_EXIT_INTR;
>               rc = -EINTR;
>       }
>  
> @@ -4390,7 +4397,7 @@ int kvm_arch_vcpu_ioctl_run(struct kvm_vcpu *vcpu)
>       }
>  
>       disable_cpu_timer_accounting(vcpu);
> -     store_regs(vcpu, kvm_run);
> +     store_regs(vcpu);
>  
>       kvm_sigset_deactivate(vcpu);
>  
> 

Reply via email to