On Mon Aug 7, 2023 at 11:45 AM AEST, Jordan Niethe wrote:
> More "wrapper" style accessor generating macros will be introduced for
> the nestedv2 guest support. Rename the existing macros with more
> descriptive names now so there is a consistent naming convention.
>
> Signed-off-by: Jordan Niethe <jniet...@gmail.com>

> ---
> v3:
>   - New to series
> ---
>  arch/powerpc/include/asm/kvm_ppc.h | 60 +++++++++++++++---------------
>  1 file changed, 30 insertions(+), 30 deletions(-)
>
> diff --git a/arch/powerpc/include/asm/kvm_ppc.h 
> b/arch/powerpc/include/asm/kvm_ppc.h
> index d16d80ad2ae4..b66084a81dd0 100644
> --- a/arch/powerpc/include/asm/kvm_ppc.h
> +++ b/arch/powerpc/include/asm/kvm_ppc.h
> @@ -927,19 +927,19 @@ static inline bool kvmppc_shared_big_endian(struct 
> kvm_vcpu *vcpu)
>  #endif
>  }
>  
> -#define SPRNG_WRAPPER_GET(reg, bookehv_spr)                          \
> +#define KVMPPC_BOOKE_HV_SPRNG_ACESSOR_GET(reg, bookehv_spr)          \
>  static inline ulong kvmppc_get_##reg(struct kvm_vcpu *vcpu)          \
>  {                                                                    \
>       return mfspr(bookehv_spr);                                      \
>  }                                                                    \
>  
> -#define SPRNG_WRAPPER_SET(reg, bookehv_spr)                          \
> +#define KVMPPC_BOOKE_HV_SPRNG_ACESSOR_SET(reg, bookehv_spr)          \
>  static inline void kvmppc_set_##reg(struct kvm_vcpu *vcpu, ulong val)        
> \
>  {                                                                    \
>       mtspr(bookehv_spr, val);                                                
> \
>  }                                                                    \
>  
> -#define SHARED_WRAPPER_GET(reg, size)                                        
> \
> +#define KVMPPC_VCPU_SHARED_REGS_ACESSOR_GET(reg, size)                       
> \
>  static inline u##size kvmppc_get_##reg(struct kvm_vcpu *vcpu)                
> \
>  {                                                                    \
>       if (kvmppc_shared_big_endian(vcpu))                             \
> @@ -948,7 +948,7 @@ static inline u##size kvmppc_get_##reg(struct kvm_vcpu 
> *vcpu)             \
>              return le##size##_to_cpu(vcpu->arch.shared->reg);        \
>  }                                                                    \
>  
> -#define SHARED_WRAPPER_SET(reg, size)                                        
> \
> +#define KVMPPC_VCPU_SHARED_REGS_ACESSOR_SET(reg, size)                       
> \
>  static inline void kvmppc_set_##reg(struct kvm_vcpu *vcpu, u##size val)      
> \
>  {                                                                    \
>       if (kvmppc_shared_big_endian(vcpu))                             \
> @@ -957,36 +957,36 @@ static inline void kvmppc_set_##reg(struct kvm_vcpu 
> *vcpu, u##size val) \
>              vcpu->arch.shared->reg = cpu_to_le##size(val);           \
>  }                                                                    \
>  
> -#define SHARED_WRAPPER(reg, size)                                    \
> -     SHARED_WRAPPER_GET(reg, size)                                   \
> -     SHARED_WRAPPER_SET(reg, size)                                   \
> +#define KVMPPC_VCPU_SHARED_REGS_ACESSOR(reg, size)                           
>         \
> +     KVMPPC_VCPU_SHARED_REGS_ACESSOR_GET(reg, size)                          
>         \
> +     KVMPPC_VCPU_SHARED_REGS_ACESSOR_SET(reg, size)                          
>         \
>  
> -#define SPRNG_WRAPPER(reg, bookehv_spr)                                      
> \
> -     SPRNG_WRAPPER_GET(reg, bookehv_spr)                             \
> -     SPRNG_WRAPPER_SET(reg, bookehv_spr)                             \
> +#define KVMPPC_BOOKE_HV_SPRNG_ACESSOR(reg, bookehv_spr)                      
>                 \
> +     KVMPPC_BOOKE_HV_SPRNG_ACESSOR_GET(reg, bookehv_spr)                     
>         \
> +     KVMPPC_BOOKE_HV_SPRNG_ACESSOR_SET(reg, bookehv_spr)                     
>         \
>  
>  #ifdef CONFIG_KVM_BOOKE_HV
>  
> -#define SHARED_SPRNG_WRAPPER(reg, size, bookehv_spr)                 \
> -     SPRNG_WRAPPER(reg, bookehv_spr)                                 \
> +#define KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(reg, size, 
> bookehv_spr)   \
> +     KVMPPC_BOOKE_HV_SPRNG_ACESSOR(reg, bookehv_spr)                 \
>  
>  #else
>  
> -#define SHARED_SPRNG_WRAPPER(reg, size, bookehv_spr)                 \
> -     SHARED_WRAPPER(reg, size)                                       \
> +#define KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(reg, size, 
> bookehv_spr)   \
> +     KVMPPC_VCPU_SHARED_REGS_ACESSOR(reg, size)                      \

Not the greatest name I've ever seen :D Hard to be concice and
consistent though, this is an odd one.

Reviewed-by: Nicholas Piggin <npig...@gmail.com>

>  
>  #endif
>  
> -SHARED_WRAPPER(critical, 64)
> -SHARED_SPRNG_WRAPPER(sprg0, 64, SPRN_GSPRG0)
> -SHARED_SPRNG_WRAPPER(sprg1, 64, SPRN_GSPRG1)
> -SHARED_SPRNG_WRAPPER(sprg2, 64, SPRN_GSPRG2)
> -SHARED_SPRNG_WRAPPER(sprg3, 64, SPRN_GSPRG3)
> -SHARED_SPRNG_WRAPPER(srr0, 64, SPRN_GSRR0)
> -SHARED_SPRNG_WRAPPER(srr1, 64, SPRN_GSRR1)
> -SHARED_SPRNG_WRAPPER(dar, 64, SPRN_GDEAR)
> -SHARED_SPRNG_WRAPPER(esr, 64, SPRN_GESR)
> -SHARED_WRAPPER_GET(msr, 64)
> +KVMPPC_VCPU_SHARED_REGS_ACESSOR(critical, 64)
> +KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(sprg0, 64, SPRN_GSPRG0)
> +KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(sprg1, 64, SPRN_GSPRG1)
> +KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(sprg2, 64, SPRN_GSPRG2)
> +KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(sprg3, 64, SPRN_GSPRG3)
> +KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(srr0, 64, SPRN_GSRR0)
> +KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(srr1, 64, SPRN_GSRR1)
> +KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(dar, 64, SPRN_GDEAR)
> +KVMPPC_BOOKE_HV_SPRNG_OR_VCPU_SHARED_REGS_ACCESSOR(esr, 64, SPRN_GESR)
> +KVMPPC_VCPU_SHARED_REGS_ACESSOR_GET(msr, 64)
>  static inline void kvmppc_set_msr_fast(struct kvm_vcpu *vcpu, u64 val)
>  {
>       if (kvmppc_shared_big_endian(vcpu))
> @@ -994,12 +994,12 @@ static inline void kvmppc_set_msr_fast(struct kvm_vcpu 
> *vcpu, u64 val)
>       else
>              vcpu->arch.shared->msr = cpu_to_le64(val);
>  }
> -SHARED_WRAPPER(dsisr, 32)
> -SHARED_WRAPPER(int_pending, 32)
> -SHARED_WRAPPER(sprg4, 64)
> -SHARED_WRAPPER(sprg5, 64)
> -SHARED_WRAPPER(sprg6, 64)
> -SHARED_WRAPPER(sprg7, 64)
> +KVMPPC_VCPU_SHARED_REGS_ACESSOR(dsisr, 32)
> +KVMPPC_VCPU_SHARED_REGS_ACESSOR(int_pending, 32)
> +KVMPPC_VCPU_SHARED_REGS_ACESSOR(sprg4, 64)
> +KVMPPC_VCPU_SHARED_REGS_ACESSOR(sprg5, 64)
> +KVMPPC_VCPU_SHARED_REGS_ACESSOR(sprg6, 64)
> +KVMPPC_VCPU_SHARED_REGS_ACESSOR(sprg7, 64)
>  
>  static inline u32 kvmppc_get_sr(struct kvm_vcpu *vcpu, int nr)
>  {

Reply via email to