On Fri, Jun 08, 2018 at 03:46:30PM +1000, Alexey Kardashevskiy wrote:
> We are going to reuse multilevel TCE code for the userspace copy of
> the TCE table and since it is big endian, let's make the copy big endian
> too.
> 
> Signed-off-by: Alexey Kardashevskiy <a...@ozlabs.ru>

Reviewed-by: David Gibson <da...@gibson.dropbear.id.au>

> ---
>  arch/powerpc/include/asm/iommu.h    |  2 +-
>  arch/powerpc/kvm/book3s_64_vio.c    | 11 ++++++-----
>  arch/powerpc/kvm/book3s_64_vio_hv.c | 10 +++++-----
>  drivers/vfio/vfio_iommu_spapr_tce.c | 19 +++++++++----------
>  4 files changed, 21 insertions(+), 21 deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/iommu.h 
> b/arch/powerpc/include/asm/iommu.h
> index 20febe0..803ac70 100644
> --- a/arch/powerpc/include/asm/iommu.h
> +++ b/arch/powerpc/include/asm/iommu.h
> @@ -117,7 +117,7 @@ struct iommu_table {
>       unsigned long *it_map;       /* A simple allocation bitmap for now */
>       unsigned long  it_page_shift;/* table iommu page size */
>       struct list_head it_group_list;/* List of iommu_table_group_link */
> -     unsigned long *it_userspace; /* userspace view of the table */
> +     __be64 *it_userspace; /* userspace view of the table */
>       struct iommu_table_ops *it_ops;
>       struct kref    it_kref;
>  };
> diff --git a/arch/powerpc/kvm/book3s_64_vio.c 
> b/arch/powerpc/kvm/book3s_64_vio.c
> index 80ead38..1dbca4b 100644
> --- a/arch/powerpc/kvm/book3s_64_vio.c
> +++ b/arch/powerpc/kvm/book3s_64_vio.c
> @@ -378,19 +378,19 @@ static long kvmppc_tce_iommu_mapped_dec(struct kvm *kvm,
>  {
>       struct mm_iommu_table_group_mem_t *mem = NULL;
>       const unsigned long pgsize = 1ULL << tbl->it_page_shift;
> -     unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
> +     __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
>  
>       if (!pua)
>               /* it_userspace allocation might be delayed */
>               return H_TOO_HARD;
>  
> -     mem = mm_iommu_lookup(kvm->mm, *pua, pgsize);
> +     mem = mm_iommu_lookup(kvm->mm, be64_to_cpu(*pua), pgsize);
>       if (!mem)
>               return H_TOO_HARD;
>  
>       mm_iommu_mapped_dec(mem);
>  
> -     *pua = 0;
> +     *pua = cpu_to_be64(0);
>  
>       return H_SUCCESS;
>  }
> @@ -437,7 +437,8 @@ long kvmppc_tce_iommu_do_map(struct kvm *kvm, struct 
> iommu_table *tbl,
>               enum dma_data_direction dir)
>  {
>       long ret;
> -     unsigned long hpa, *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
> +     unsigned long hpa;
> +     __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
>       struct mm_iommu_table_group_mem_t *mem;
>  
>       if (!pua)
> @@ -464,7 +465,7 @@ long kvmppc_tce_iommu_do_map(struct kvm *kvm, struct 
> iommu_table *tbl,
>       if (dir != DMA_NONE)
>               kvmppc_tce_iommu_mapped_dec(kvm, tbl, entry);
>  
> -     *pua = ua;
> +     *pua = cpu_to_be64(ua);
>  
>       return 0;
>  }
> diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c 
> b/arch/powerpc/kvm/book3s_64_vio_hv.c
> index 635f3ca..18109f3 100644
> --- a/arch/powerpc/kvm/book3s_64_vio_hv.c
> +++ b/arch/powerpc/kvm/book3s_64_vio_hv.c
> @@ -200,7 +200,7 @@ static long kvmppc_rm_tce_iommu_mapped_dec(struct kvm 
> *kvm,
>  {
>       struct mm_iommu_table_group_mem_t *mem = NULL;
>       const unsigned long pgsize = 1ULL << tbl->it_page_shift;
> -     unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
> +     __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
>  
>       if (!pua)
>               /* it_userspace allocation might be delayed */
> @@ -210,13 +210,13 @@ static long kvmppc_rm_tce_iommu_mapped_dec(struct kvm 
> *kvm,
>       if (WARN_ON_ONCE_RM(!pua))
>               return H_HARDWARE;
>  
> -     mem = mm_iommu_lookup_rm(kvm->mm, *pua, pgsize);
> +     mem = mm_iommu_lookup_rm(kvm->mm, be64_to_cpu(*pua), pgsize);
>       if (!mem)
>               return H_TOO_HARD;
>  
>       mm_iommu_mapped_dec(mem);
>  
> -     *pua = 0;
> +     *pua = cpu_to_be64(0);
>  
>       return H_SUCCESS;
>  }
> @@ -268,7 +268,7 @@ static long kvmppc_rm_tce_iommu_do_map(struct kvm *kvm, 
> struct iommu_table *tbl,
>  {
>       long ret;
>       unsigned long hpa = 0;
> -     unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
> +     __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
>       struct mm_iommu_table_group_mem_t *mem;
>  
>       if (!pua)
> @@ -302,7 +302,7 @@ static long kvmppc_rm_tce_iommu_do_map(struct kvm *kvm, 
> struct iommu_table *tbl,
>       if (dir != DMA_NONE)
>               kvmppc_rm_tce_iommu_mapped_dec(kvm, tbl, entry);
>  
> -     *pua = ua;
> +     *pua = cpu_to_be64(ua);
>  
>       return 0;
>  }
> diff --git a/drivers/vfio/vfio_iommu_spapr_tce.c 
> b/drivers/vfio/vfio_iommu_spapr_tce.c
> index 47071f3..81f48114 100644
> --- a/drivers/vfio/vfio_iommu_spapr_tce.c
> +++ b/drivers/vfio/vfio_iommu_spapr_tce.c
> @@ -231,7 +231,7 @@ static long tce_iommu_userspace_view_alloc(struct 
> iommu_table *tbl,
>               decrement_locked_vm(mm, cb >> PAGE_SHIFT);
>               return -ENOMEM;
>       }
> -     tbl->it_userspace = uas;
> +     tbl->it_userspace = (__be64 *) uas;
>  
>       return 0;
>  }
> @@ -494,20 +494,20 @@ static void tce_iommu_unuse_page_v2(struct 
> tce_container *container,
>       struct mm_iommu_table_group_mem_t *mem = NULL;
>       int ret;
>       unsigned long hpa = 0;
> -     unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
> +     __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry);
>  
>       if (!pua)
>               return;
>  
> -     ret = tce_iommu_prereg_ua_to_hpa(container, *pua, IOMMU_PAGE_SIZE(tbl),
> -                     &hpa, &mem);
> +     ret = tce_iommu_prereg_ua_to_hpa(container, be64_to_cpu(*pua),
> +                     IOMMU_PAGE_SIZE(tbl), &hpa, &mem);
>       if (ret)
> -             pr_debug("%s: tce %lx at #%lx was not cached, ret=%d\n",
> -                             __func__, *pua, entry, ret);
> +             pr_debug("%s: tce %llx at #%lx was not cached, ret=%d\n",
> +                             __func__, be64_to_cpu(*pua), entry, ret);
>       if (mem)
>               mm_iommu_mapped_dec(mem);
>  
> -     *pua = 0;
> +     *pua = cpu_to_be64(0);
>  }
>  
>  static int tce_iommu_clear(struct tce_container *container,
> @@ -616,8 +616,7 @@ static long tce_iommu_build_v2(struct tce_container 
> *container,
>  
>       for (i = 0; i < pages; ++i) {
>               struct mm_iommu_table_group_mem_t *mem = NULL;
> -             unsigned long *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl,
> -                             entry + i);
> +             __be64 *pua = IOMMU_TABLE_USERSPACE_ENTRY(tbl, entry + i);
>  
>               ret = tce_iommu_prereg_ua_to_hpa(container,
>                               tce, IOMMU_PAGE_SIZE(tbl), &hpa, &mem);
> @@ -650,7 +649,7 @@ static long tce_iommu_build_v2(struct tce_container 
> *container,
>               if (dirtmp != DMA_NONE)
>                       tce_iommu_unuse_page_v2(container, tbl, entry + i);
>  
> -             *pua = tce;
> +             *pua = cpu_to_be64(tce);
>  
>               tce += IOMMU_PAGE_SIZE(tbl);
>       }

-- 
David Gibson                    | I'll have my music baroque, and my code
david AT gibson.dropbear.id.au  | minimalist, thank you.  NOT _the_ _other_
                                | _way_ _around_!
http://www.ozlabs.org/~dgibson

Attachment: signature.asc
Description: PGP signature

Reply via email to