On Thu, 2017-01-12 at 20:07 +1100, Paul Mackerras wrote:
> This adapts our implementations of the MMU notifier callbacks
> (unmap_hva, unmap_hva_range, age_hva, test_age_hva, set_spte_hva)
> to call radix functions when the guest is using radix.  These
> implementations are much simpler than for HPT guests because we
> have only one PTE to deal with, so we don't need to traverse
> rmap chains.
> 
> Signed-off-by: Paul Mackerras <pau...@ozlabs.org>
> ---
>  arch/powerpc/include/asm/kvm_book3s.h  |  6 ++++
>  arch/powerpc/kvm/book3s_64_mmu_hv.c    | 64 +++++++++++++++++++++++-
> ----------
>  arch/powerpc/kvm/book3s_64_mmu_radix.c | 54
> ++++++++++++++++++++++++++++
>  3 files changed, 103 insertions(+), 21 deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/kvm_book3s.h
> b/arch/powerpc/include/asm/kvm_book3s.h
> index ff5cd5c..952cc4b 100644
> --- a/arch/powerpc/include/asm/kvm_book3s.h
> +++ b/arch/powerpc/include/asm/kvm_book3s.h
> @@ -192,6 +192,12 @@ extern int kvmppc_mmu_radix_xlate(struct
> kvm_vcpu *vcpu, gva_t eaddr,
>  extern void kvmppc_free_radix(struct kvm *kvm);
>  extern int kvmppc_radix_init(void);
>  extern void kvmppc_radix_exit(void);
> +extern int kvm_unmap_radix(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
> +                     unsigned long gfn);
> +extern int kvm_age_radix(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
> +                     unsigned long gfn);
> +extern int kvm_test_age_radix(struct kvm *kvm, struct
> kvm_memory_slot *memslot,
> +                     unsigned long gfn);
>  
>  /* XXX remove this export when load_last_inst() is generic */
>  extern int kvmppc_ld(struct kvm_vcpu *vcpu, ulong *eaddr, int size,
> void *ptr, bool data);
> diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c
> b/arch/powerpc/kvm/book3s_64_mmu_hv.c
> index 57690c2..fbb3de4 100644
> --- a/arch/powerpc/kvm/book3s_64_mmu_hv.c
> +++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c
> @@ -701,12 +701,13 @@ static void kvmppc_rmap_reset(struct kvm *kvm)
>       srcu_read_unlock(&kvm->srcu, srcu_idx);
>  }
>  
> +typedef int (*hva_handler_fn)(struct kvm *kvm, struct
> kvm_memory_slot *memslot,
> +                           unsigned long gfn);
> +
>  static int kvm_handle_hva_range(struct kvm *kvm,
>                               unsigned long start,
>                               unsigned long end,
> -                             int (*handler)(struct kvm *kvm,
> -                                            unsigned long *rmapp,
> -                                            unsigned long gfn))
> +                             hva_handler_fn handler)
>  {
>       int ret;
>       int retval = 0;
> @@ -731,9 +732,7 @@ static int kvm_handle_hva_range(struct kvm *kvm,
>               gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE -
> 1, memslot);
>  
>               for (; gfn < gfn_end; ++gfn) {
> -                     gfn_t gfn_offset = gfn - memslot->base_gfn;
> -
> -                     ret = handler(kvm, &memslot-
> >arch.rmap[gfn_offset], gfn);
> +                     ret = handler(kvm, memslot, gfn);
>                       retval |= ret;
>               }
>       }
> @@ -742,20 +741,21 @@ static int kvm_handle_hva_range(struct kvm
> *kvm,
>  }
>  
>  static int kvm_handle_hva(struct kvm *kvm, unsigned long hva,
> -                       int (*handler)(struct kvm *kvm, unsigned
> long *rmapp,
> -                                      unsigned long gfn))
> +                       hva_handler_fn handler)
>  {
>       return kvm_handle_hva_range(kvm, hva, hva + 1, handler);
>  }
>  
> -static int kvm_unmap_rmapp(struct kvm *kvm, unsigned long *rmapp,
> +static int kvm_unmap_rmapp(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
>                          unsigned long gfn)
>  {
>       struct revmap_entry *rev = kvm->arch.revmap;
>       unsigned long h, i, j;
>       __be64 *hptep;
>       unsigned long ptel, psize, rcbits;
> +     unsigned long *rmapp;
>  
> +     rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
>       for (;;) {
>               lock_rmap(rmapp);
>               if (!(*rmapp & KVMPPC_RMAP_PRESENT)) {
> @@ -816,26 +816,36 @@ static int kvm_unmap_rmapp(struct kvm *kvm,
> unsigned long *rmapp,
>  
>  int kvm_unmap_hva_hv(struct kvm *kvm, unsigned long hva)
>  {
> -     kvm_handle_hva(kvm, hva, kvm_unmap_rmapp);
> +     hva_handler_fn handler;
> +
> +     handler = kvm->arch.radix ? kvm_unmap_radix : 
kvm_is_radix() for consistency?
> kvm_unmap_rmapp;
> +     kvm_handle_hva(kvm, hva, handler);
>       return 0;
>  }
>  
>  int kvm_unmap_hva_range_hv(struct kvm *kvm, unsigned long start,
> unsigned long end)
>  {
> -     kvm_handle_hva_range(kvm, start, end, kvm_unmap_rmapp);
> +     hva_handler_fn handler;
> +
> +     handler = kvm->arch.radix ? kvm_unmap_radix : 
ditto
> kvm_unmap_rmapp;
> +     kvm_handle_hva_range(kvm, start, end, handler);
>       return 0;
>  }
>  
>  void kvmppc_core_flush_memslot_hv(struct kvm *kvm,
>                                 struct kvm_memory_slot *memslot)
>  {
> -     unsigned long *rmapp;
>       unsigned long gfn;
>       unsigned long n;
> +     unsigned long *rmapp;
>  
> -     rmapp = memslot->arch.rmap;
>       gfn = memslot->base_gfn;
> -     for (n = memslot->npages; n; --n) {
> +     rmapp = memslot->arch.rmap;
> +     for (n = memslot->npages; n; --n, ++gfn) {
> +             if (kvm->arch.radix) {
ditto
> +                     kvm_unmap_radix(kvm, memslot, gfn);
> +                     continue;
> +             }
>               /*
>                * Testing the present bit without locking is OK
> because
>                * the memslot has been marked invalid already, and
> hence
> @@ -843,20 +853,21 @@ void kvmppc_core_flush_memslot_hv(struct kvm
> *kvm,
>                * thus the present bit can't go from 0 to 1.
>                */
>               if (*rmapp & KVMPPC_RMAP_PRESENT)
> -                     kvm_unmap_rmapp(kvm, rmapp, gfn);
> +                     kvm_unmap_rmapp(kvm, memslot, gfn);
>               ++rmapp;
> -             ++gfn;
>       }
>  }
>  
> -static int kvm_age_rmapp(struct kvm *kvm, unsigned long *rmapp,
> +static int kvm_age_rmapp(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
>                        unsigned long gfn)
>  {
>       struct revmap_entry *rev = kvm->arch.revmap;
>       unsigned long head, i, j;
>       __be64 *hptep;
>       int ret = 0;
> +     unsigned long *rmapp;
>  
> +     rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
>   retry:
>       lock_rmap(rmapp);
>       if (*rmapp & KVMPPC_RMAP_REFERENCED) {
> @@ -904,17 +915,22 @@ static int kvm_age_rmapp(struct kvm *kvm,
> unsigned long *rmapp,
>  
>  int kvm_age_hva_hv(struct kvm *kvm, unsigned long start, unsigned
> long end)
>  {
> -     return kvm_handle_hva_range(kvm, start, end, kvm_age_rmapp);
> +     hva_handler_fn handler;
> +
> +     handler = kvm->arch.radix ? kvm_age_radix : kvm_age_rmapp;
ditto
> +     return kvm_handle_hva_range(kvm, start, end, handler);
>  }
>  
> -static int kvm_test_age_rmapp(struct kvm *kvm, unsigned long *rmapp,
> +static int kvm_test_age_rmapp(struct kvm *kvm, struct
> kvm_memory_slot *memslot,
>                             unsigned long gfn)
>  {
>       struct revmap_entry *rev = kvm->arch.revmap;
>       unsigned long head, i, j;
>       unsigned long *hp;
>       int ret = 1;
> +     unsigned long *rmapp;
>  
> +     rmapp = &memslot->arch.rmap[gfn - memslot->base_gfn];
>       if (*rmapp & KVMPPC_RMAP_REFERENCED)
>               return 1;
>  
> @@ -940,12 +956,18 @@ static int kvm_test_age_rmapp(struct kvm *kvm,
> unsigned long *rmapp,
>  
>  int kvm_test_age_hva_hv(struct kvm *kvm, unsigned long hva)
>  {
> -     return kvm_handle_hva(kvm, hva, kvm_test_age_rmapp);
> +     hva_handler_fn handler;
> +
> +     handler = kvm->arch.radix ? kvm_test_age_radix : 
ditto
> kvm_test_age_rmapp;
> +     return kvm_handle_hva(kvm, hva, handler);
>  }
>  
>  void kvm_set_spte_hva_hv(struct kvm *kvm, unsigned long hva, pte_t
> pte)
>  {
> -     kvm_handle_hva(kvm, hva, kvm_unmap_rmapp);
> +     hva_handler_fn handler;
> +
> +     handler = kvm->arch.radix ? kvm_unmap_radix : 
ditto
> kvm_unmap_rmapp;
> +     kvm_handle_hva(kvm, hva, handler);
>  }
>  
>  static int vcpus_running(struct kvm *kvm)
> diff --git a/arch/powerpc/kvm/book3s_64_mmu_radix.c
> b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> index 865ea9b..69cabad 100644
> --- a/arch/powerpc/kvm/book3s_64_mmu_radix.c
> +++ b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> @@ -463,6 +463,60 @@ int kvmppc_book3s_radix_page_fault(struct
> kvm_run *run, struct kvm_vcpu *vcpu,
>       return ret;
>  }
>  
> +/* Called with kvm->lock held */
> +int kvm_unmap_radix(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
> +                 unsigned long gfn)
> +{
> +     pte_t *ptep;
> +     unsigned long gpa = gfn << PAGE_SHIFT;
> +     unsigned int shift;
> +
> +     ptep = __find_linux_pte_or_hugepte(kvm->arch.pgtable, gpa,
> +                                        NULL, &shift);
> +     if (ptep && pte_present(*ptep)) {
> +             kvmppc_radix_update_pte(kvm, ptep, _PAGE_PRESENT, 0,
> +                                     gpa, shift);
> +             kvmppc_radix_tlbie_page(kvm, gpa, shift);
> +     }
> +     return 0;                               
> +}
> +
> +/* Called with kvm->lock held */
> +int kvm_age_radix(struct kvm *kvm, struct kvm_memory_slot *memslot,
> +               unsigned long gfn)
> +{
> +     pte_t *ptep;
> +     unsigned long gpa = gfn << PAGE_SHIFT;
> +     unsigned int shift;
> +     int ref = 0;
> +
> +     ptep = __find_linux_pte_or_hugepte(kvm->arch.pgtable, gpa,
> +                                        NULL, &shift);
> +     if (ptep && pte_present(*ptep) && pte_young(*ptep)) {
> +             kvmppc_radix_update_pte(kvm, ptep, _PAGE_ACCESSED,
> 0,
> +                                     gpa, shift);
> +             /* XXX need to flush tlb here? */
> +             ref = 1;
> +     }
> +     return ref;
> +}
> +
> +/* Called with kvm->lock held */
> +int kvm_test_age_radix(struct kvm *kvm, struct kvm_memory_slot
> *memslot,
> +                    unsigned long gfn)
> +{
> +     pte_t *ptep;
> +     unsigned long gpa = gfn << PAGE_SHIFT;
> +     unsigned int shift;
> +     int ref = 0;
> +
> +     ptep = __find_linux_pte_or_hugepte(kvm->arch.pgtable, gpa,
> +                                        NULL, &shift);
> +     if (ptep && pte_present(*ptep) && pte_young(*ptep))
> +             ref = 1;
> +     return ref;
> +}
> +
>  void kvmppc_free_radix(struct kvm *kvm)
>  {
>       unsigned long ig, iu, im;

Reply via email to