On 27/03/16 19:23, Aneesh Kumar K.V wrote:
> This split _PAGE_RW bit to _PAGE_READ and _PAGE_WRITE. It also remove
> the dependency on _PAGE_USER for implying read only. Few things to note
> here is that, we have read implied with write and execute permission.
> Hence we should always find _PAGE_READ set on hash pte fault.
>
> We still can't switch PROT_NONE to !(_PAGE_RWX). Auto numa do depend
> on marking a prot none pte _PAGE_WRITE. (For more details look at
> b191f9b106ea "mm: numa: preserve PTE write permissions across a NUMA hinting 
> fault")
>
> Signed-off-by: Aneesh Kumar K.V <aneesh.ku...@linux.vnet.ibm.com>
> ---
>  arch/powerpc/include/asm/book3s/64/hash-64k.h |  4 +--
>  arch/powerpc/include/asm/book3s/64/hash.h     | 35 
> ++++++++++++++++-----------
>  arch/powerpc/include/asm/pte-common.h         |  5 ++++
>  arch/powerpc/mm/hash64_4k.c                   |  2 +-
>  arch/powerpc/mm/hash64_64k.c                  |  4 +--
>  arch/powerpc/mm/hash_utils_64.c               |  9 ++++---
>  arch/powerpc/mm/hugepage-hash64.c             |  2 +-
>  arch/powerpc/mm/hugetlbpage-hash64.c          |  2 +-
>  arch/powerpc/mm/hugetlbpage.c                 |  4 +--
>  arch/powerpc/mm/pgtable.c                     |  4 +--
>  arch/powerpc/mm/pgtable_64.c                  |  5 ++--
>  arch/powerpc/platforms/cell/spu_base.c        |  2 +-
>  arch/powerpc/platforms/cell/spufs/fault.c     |  4 +--
>  drivers/misc/cxl/fault.c                      |  4 +--
>  14 files changed, 49 insertions(+), 37 deletions(-)
>
> diff --git a/arch/powerpc/include/asm/book3s/64/hash-64k.h 
> b/arch/powerpc/include/asm/book3s/64/hash-64k.h
> index 0a7956a80a08..279ded72f1db 100644
> --- a/arch/powerpc/include/asm/book3s/64/hash-64k.h
> +++ b/arch/powerpc/include/asm/book3s/64/hash-64k.h
> @@ -291,10 +291,10 @@ static inline void pmdp_set_wrprotect(struct mm_struct 
> *mm, unsigned long addr,
>                                     pmd_t *pmdp)
>  {
>  
> -     if ((pmd_val(*pmdp) & _PAGE_RW) == 0)
> +     if ((pmd_val(*pmdp) & _PAGE_WRITE) == 0)
>               return;
>  
> -     pmd_hugepage_update(mm, addr, pmdp, _PAGE_RW, 0);
> +     pmd_hugepage_update(mm, addr, pmdp, _PAGE_WRITE, 0);
>  }
>  
>  #endif /*  CONFIG_TRANSPARENT_HUGEPAGE */
> diff --git a/arch/powerpc/include/asm/book3s/64/hash.h 
> b/arch/powerpc/include/asm/book3s/64/hash.h
> index 2113de051824..f092d83fa623 100644
> --- a/arch/powerpc/include/asm/book3s/64/hash.h
> +++ b/arch/powerpc/include/asm/book3s/64/hash.h
> @@ -16,8 +16,10 @@
>  #define _PAGE_BIT_SWAP_TYPE  0
>  
>  #define _PAGE_EXEC           0x00001 /* execute permission */
> -#define _PAGE_RW             0x00002 /* read & write access allowed */
> +#define _PAGE_WRITE          0x00002 /* write access allowed */
>  #define _PAGE_READ           0x00004 /* read access allowed */
> +#define _PAGE_RW             (_PAGE_READ | _PAGE_WRITE)
> +#define _PAGE_RWX            (_PAGE_READ | _PAGE_WRITE | _PAGE_EXEC)
>  #define _PAGE_USER           0x00008 /* page may be accessed by userspace */
>  #define _PAGE_GUARDED                0x00010 /* G: guarded (side-effect) 
> page */
>  /* M (memory coherence) is always set in the HPTE, so we don't need it here 
> */
> @@ -147,8 +149,8 @@
>   */
>  #define PAGE_PROT_BITS       (_PAGE_GUARDED | _PAGE_COHERENT | 
> _PAGE_NO_CACHE | \
>                        _PAGE_WRITETHRU | _PAGE_4K_PFN | \
> -                      _PAGE_USER | _PAGE_ACCESSED |  \
> -                      _PAGE_RW |  _PAGE_DIRTY | _PAGE_EXEC | \
> +                      _PAGE_USER | _PAGE_ACCESSED |  _PAGE_READ |\
> +                      _PAGE_WRITE |  _PAGE_DIRTY | _PAGE_EXEC | \
>                        _PAGE_SOFT_DIRTY)
>  /*
>   * We define 2 sets of base prot bits, one for basic pages (ie,
> @@ -173,10 +175,12 @@
>  #define PAGE_SHARED  __pgprot(_PAGE_BASE | _PAGE_USER | _PAGE_RW)
>  #define PAGE_SHARED_X        __pgprot(_PAGE_BASE | _PAGE_USER | _PAGE_RW | \
>                                _PAGE_EXEC)
> -#define PAGE_COPY    __pgprot(_PAGE_BASE | _PAGE_USER )
> -#define PAGE_COPY_X  __pgprot(_PAGE_BASE | _PAGE_USER | _PAGE_EXEC)
> -#define PAGE_READONLY        __pgprot(_PAGE_BASE | _PAGE_USER )
> -#define PAGE_READONLY_X      __pgprot(_PAGE_BASE | _PAGE_USER | _PAGE_EXEC)
> +#define PAGE_COPY    __pgprot(_PAGE_BASE | _PAGE_USER | _PAGE_READ)
> +#define PAGE_COPY_X  __pgprot(_PAGE_BASE | _PAGE_USER | _PAGE_READ| \
> +                              _PAGE_EXEC)
> +#define PAGE_READONLY        __pgprot(_PAGE_BASE | _PAGE_USER | _PAGE_READ)
> +#define PAGE_READONLY_X      __pgprot(_PAGE_BASE | _PAGE_USER | _PAGE_READ| \
> +                              _PAGE_EXEC)
>  
>  #define __P000       PAGE_NONE
>  #define __P001       PAGE_READONLY
> @@ -300,19 +304,19 @@ static inline void ptep_set_wrprotect(struct mm_struct 
> *mm, unsigned long addr,
>                                     pte_t *ptep)
>  {
>  
> -     if ((pte_val(*ptep) & _PAGE_RW) == 0)
> +     if ((pte_val(*ptep) & _PAGE_WRITE) == 0)
>               return;
>  
> -     pte_update(mm, addr, ptep, _PAGE_RW, 0, 0);
> +     pte_update(mm, addr, ptep, _PAGE_WRITE, 0, 0);
>  }
>  
>  static inline void huge_ptep_set_wrprotect(struct mm_struct *mm,
>                                          unsigned long addr, pte_t *ptep)
>  {
> -     if ((pte_val(*ptep) & _PAGE_RW) == 0)
> +     if ((pte_val(*ptep) & _PAGE_WRITE) == 0)
>               return;
>  
> -     pte_update(mm, addr, ptep, _PAGE_RW, 0, 1);
> +     pte_update(mm, addr, ptep, _PAGE_WRITE, 0, 1);
>  }
>  
>  /*
> @@ -352,7 +356,7 @@ static inline void pte_clear(struct mm_struct *mm, 
> unsigned long addr,
>  static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry)
>  {
>       unsigned long bits = pte_val(entry) &
> -             (_PAGE_DIRTY | _PAGE_ACCESSED | _PAGE_RW | _PAGE_EXEC |
> +             (_PAGE_DIRTY | _PAGE_ACCESSED | _PAGE_READ | _PAGE_WRITE | 
> _PAGE_EXEC |
>                _PAGE_SOFT_DIRTY);
>  
>       unsigned long old, tmp;
> @@ -386,7 +390,7 @@ static inline unsigned long pgd_page_vaddr(pgd_t pgd)
>  
>  
>  /* Generic accessors to PTE bits */
> -static inline int pte_write(pte_t pte)               { return 
> !!(pte_val(pte) & _PAGE_RW);}
> +static inline int pte_write(pte_t pte)               { return 
> !!(pte_val(pte) & _PAGE_WRITE);}
>  static inline int pte_dirty(pte_t pte)               { return 
> !!(pte_val(pte) & _PAGE_DIRTY); }
>  static inline int pte_young(pte_t pte)               { return 
> !!(pte_val(pte) & _PAGE_ACCESSED); }
>  static inline int pte_special(pte_t pte)     { return !!(pte_val(pte) & 
> _PAGE_SPECIAL); }
> @@ -447,7 +451,7 @@ static inline unsigned long pte_pfn(pte_t pte)
>  /* Generic modifiers for PTE bits */
>  static inline pte_t pte_wrprotect(pte_t pte)
>  {
> -     return __pte(pte_val(pte) & ~_PAGE_RW);
> +     return __pte(pte_val(pte) & ~_PAGE_WRITE);
>  }
>  
>  static inline pte_t pte_mkclean(pte_t pte)
> @@ -462,6 +466,9 @@ static inline pte_t pte_mkold(pte_t pte)
>  
>  static inline pte_t pte_mkwrite(pte_t pte)
>  {
> +     /*
> +      * write implies read, hence set both
> +      */
>       return __pte(pte_val(pte) | _PAGE_RW);
>  }
>  
> diff --git a/arch/powerpc/include/asm/pte-common.h 
> b/arch/powerpc/include/asm/pte-common.h
> index 1ec67b043065..9f5dea58b0db 100644
> --- a/arch/powerpc/include/asm/pte-common.h
> +++ b/arch/powerpc/include/asm/pte-common.h
> @@ -198,3 +198,8 @@ extern unsigned long bad_call_to_PMD_PAGE_SIZE(void);
>  /* Advertise support for _PAGE_SPECIAL */
>  #define __HAVE_ARCH_PTE_SPECIAL
>  
> +#ifndef _PAGE_READ
> +/* if not defined, we should not find _PAGE_WRITE too */
> +#define _PAGE_READ 0
> +#define _PAGE_WRITE _PAGE_RW
> +#endif
> diff --git a/arch/powerpc/mm/hash64_4k.c b/arch/powerpc/mm/hash64_4k.c
> index 71abd4c44c27..7ebac279d38e 100644
> --- a/arch/powerpc/mm/hash64_4k.c
> +++ b/arch/powerpc/mm/hash64_4k.c
> @@ -46,7 +46,7 @@ int __hash_page_4K(unsigned long ea, unsigned long access, 
> unsigned long vsid,
>                * also add _PAGE_COMBO
>                */
>               new_pte = old_pte | _PAGE_BUSY | _PAGE_ACCESSED;
> -             if (access & _PAGE_RW)
> +             if (access & _PAGE_WRITE)
>                       new_pte |= _PAGE_DIRTY;
>  
>               opte = cpu_to_be64(old_pte);
> diff --git a/arch/powerpc/mm/hash64_64k.c b/arch/powerpc/mm/hash64_64k.c
> index 6f9b3c34a5c0..83ac9f658733 100644
> --- a/arch/powerpc/mm/hash64_64k.c
> +++ b/arch/powerpc/mm/hash64_64k.c
> @@ -78,7 +78,7 @@ int __hash_page_4K(unsigned long ea, unsigned long access, 
> unsigned long vsid,
>                * also add _PAGE_COMBO
>                */
>               new_pte = old_pte | _PAGE_BUSY | _PAGE_ACCESSED | _PAGE_COMBO;
> -             if (access & _PAGE_RW)
> +             if (access & _PAGE_WRITE)
>                       new_pte |= _PAGE_DIRTY;
>  
>               opte = cpu_to_be64(old_pte);
> @@ -255,7 +255,7 @@ int __hash_page_64K(unsigned long ea, unsigned long 
> access,
>                * a write access.
>                */
>               new_pte = old_pte | _PAGE_BUSY | _PAGE_ACCESSED;
> -             if (access & _PAGE_RW)
> +             if (access & _PAGE_WRITE)
>                       new_pte |= _PAGE_DIRTY;
>               opte = cpu_to_be64(old_pte);
>               npte = cpu_to_be64(new_pte);
> diff --git a/arch/powerpc/mm/hash_utils_64.c b/arch/powerpc/mm/hash_utils_64.c
> index 90dd9280894f..ea23403b3fc0 100644
> --- a/arch/powerpc/mm/hash_utils_64.c
> +++ b/arch/powerpc/mm/hash_utils_64.c
> @@ -175,8 +175,9 @@ unsigned long htab_convert_pte_flags(unsigned long 
> pteflags)
>        * or PP=0x3 for read-only (including writeable but clean pages).
>        */
>       if (pteflags & _PAGE_USER) {
> -             rflags |= 0x2;
> -             if (!((pteflags & _PAGE_RW) && (pteflags & _PAGE_DIRTY)))
> +             if (pteflags & _PAGE_RWX)
Should this be pteflags & _PAGE_RW?
> +                     rflags |= 0x2;
> +             if (!((pteflags & _PAGE_WRITE) && (pteflags & _PAGE_DIRTY)))
>                       rflags |= 0x1;
if pteflags == _PAGE_USER | _PAGE_WRITE | _PAGE_DIRTY, what is rflags set to?



>       }
>       /*
> @@ -1205,7 +1206,7 @@ EXPORT_SYMBOL_GPL(hash_page);
>  int __hash_page(unsigned long ea, unsigned long msr, unsigned long trap,
>               unsigned long dsisr)
>  {
> -     unsigned long access = _PAGE_PRESENT;
> +     unsigned long access = _PAGE_PRESENT | _PAGE_READ;
>       unsigned long flags = 0;
>       struct mm_struct *mm = current->mm;
>  
> @@ -1216,7 +1217,7 @@ int __hash_page(unsigned long ea, unsigned long msr, 
> unsigned long trap,
>               flags |= HPTE_NOHPTE_UPDATE;
>  
>       if (dsisr & DSISR_ISSTORE)
> -             access |= _PAGE_RW;
> +             access |= _PAGE_WRITE;
>       /*
>        * We need to set the _PAGE_USER bit if MSR_PR is set or if we are
>        * accessing a userspace segment (even from the kernel). We assume
> diff --git a/arch/powerpc/mm/hugepage-hash64.c 
> b/arch/powerpc/mm/hugepage-hash64.c
> index 98891139c044..39342638a498 100644
> --- a/arch/powerpc/mm/hugepage-hash64.c
> +++ b/arch/powerpc/mm/hugepage-hash64.c
> @@ -48,7 +48,7 @@ int __hash_page_thp(unsigned long ea, unsigned long access, 
> unsigned long vsid,
>                * a write access
>                */
>               new_pmd = old_pmd | _PAGE_BUSY | _PAGE_ACCESSED;
> -             if (access & _PAGE_RW)
> +             if (access & _PAGE_WRITE)
>                       new_pmd |= _PAGE_DIRTY;
>               opmd = cpu_to_be64(old_pmd);
>               npmd = cpu_to_be64(new_pmd);
> diff --git a/arch/powerpc/mm/hugetlbpage-hash64.c 
> b/arch/powerpc/mm/hugetlbpage-hash64.c
> index 5bcb28606158..e6e54a04bd32 100644
> --- a/arch/powerpc/mm/hugetlbpage-hash64.c
> +++ b/arch/powerpc/mm/hugetlbpage-hash64.c
> @@ -56,7 +56,7 @@ int __hash_page_huge(unsigned long ea, unsigned long 
> access, unsigned long vsid,
>               /* Try to lock the PTE, add ACCESSED and DIRTY if it was
>                * a write access */
>               new_pte = old_pte | _PAGE_BUSY | _PAGE_ACCESSED;
> -             if (access & _PAGE_RW)
> +             if (access & _PAGE_WRITE)
>                       new_pte |= _PAGE_DIRTY;
>               opte = cpu_to_be64(old_pte);
>               npte = cpu_to_be64(new_pte);
> diff --git a/arch/powerpc/mm/hugetlbpage.c b/arch/powerpc/mm/hugetlbpage.c
> index 6dd272b6196f..6e52e722d3f2 100644
> --- a/arch/powerpc/mm/hugetlbpage.c
> +++ b/arch/powerpc/mm/hugetlbpage.c
> @@ -1003,9 +1003,9 @@ int gup_hugepte(pte_t *ptep, unsigned long sz, unsigned 
> long addr,
>               end = pte_end;
>  
>       pte = READ_ONCE(*ptep);
> -     mask = _PAGE_PRESENT | _PAGE_USER;
> +     mask = _PAGE_PRESENT | _PAGE_USER | _PAGE_READ;
>       if (write)
> -             mask |= _PAGE_RW;
> +             mask |= _PAGE_WRITE;
>  
>       if ((pte_val(pte) & mask) != mask)
>               return 0;
> diff --git a/arch/powerpc/mm/pgtable.c b/arch/powerpc/mm/pgtable.c
> index 83dfd7925c72..98b5c03e344d 100644
> --- a/arch/powerpc/mm/pgtable.c
> +++ b/arch/powerpc/mm/pgtable.c
> @@ -177,8 +177,8 @@ void set_pte_at(struct mm_struct *mm, unsigned long addr, 
> pte_t *ptep,
>        * _PAGE_PRESENT, but we can be sure that it is not in hpte.
>        * Hence we can use set_pte_at for them.
>        */
> -     VM_WARN_ON((pte_val(*ptep) & (_PAGE_PRESENT | _PAGE_USER)) ==
> -             (_PAGE_PRESENT | _PAGE_USER));
> +     VM_WARN_ON(pte_present(*ptep) && !pte_protnone(*ptep));
> +
What was wrong with the previous check? The compiler will optimize it, but the 
new
check uses two pte_val() calls on *ptep
>       /*
>        * Add the pte bit when tryint set a pte
>        */
> diff --git a/arch/powerpc/mm/pgtable_64.c b/arch/powerpc/mm/pgtable_64.c
> index aa742aa35b64..00d8d985bba3 100644
> --- a/arch/powerpc/mm/pgtable_64.c
> +++ b/arch/powerpc/mm/pgtable_64.c
> @@ -277,7 +277,7 @@ void __iomem * ioremap_prot(phys_addr_t addr, unsigned 
> long size,
>       void *caller = __builtin_return_address(0);
>  
>       /* writeable implies dirty for kernel addresses */
> -     if (flags & _PAGE_RW)
> +     if (flags & _PAGE_WRITE)
>               flags |= _PAGE_DIRTY;
>  
>       /* we don't want to let _PAGE_USER and _PAGE_EXEC leak out */
> @@ -681,8 +681,7 @@ void set_pmd_at(struct mm_struct *mm, unsigned long addr,
>               pmd_t *pmdp, pmd_t pmd)
>  {
>  #ifdef CONFIG_DEBUG_VM
> -     WARN_ON((pmd_val(*pmdp) & (_PAGE_PRESENT | _PAGE_USER)) ==
> -             (_PAGE_PRESENT | _PAGE_USER));
> +     WARN_ON(pte_present(pmd_pte(*pmdp)) && !pte_protnone(pmd_pte(*pmdp)));
Same as above, plus I think we can move these to VM_WARN_ON just to be 
consistent

snip...
Balbir
_______________________________________________
Linuxppc-dev mailing list
Linuxppc-dev@lists.ozlabs.org
https://lists.ozlabs.org/listinfo/linuxppc-dev

Reply via email to