On Wed, 2016-08-24 at 15:03 +0530, Aneesh Kumar K.V wrote: > POWER9 DD1 requires pte to be marked invalid (V=0) before updating > it with the new value. This makes this distinction for the different > revisions. > > Signed-off-by: Aneesh Kumar K.V <aneesh.ku...@linux.vnet.ibm.com>
Acked-by: Michael Neuling <mi...@neuling.org> > --- > arch/powerpc/include/asm/book3s/32/pgtable.h | 3 +- > arch/powerpc/include/asm/book3s/64/pgtable.h | 5 +- > arch/powerpc/include/asm/book3s/64/radix.h | 75 > ++++++++++++++++++++++------ > arch/powerpc/include/asm/nohash/32/pgtable.h | 3 +- > arch/powerpc/include/asm/nohash/64/pgtable.h | 3 +- > arch/powerpc/mm/pgtable-book3s64.c | 2 +- > arch/powerpc/mm/pgtable.c | 2 +- > 7 files changed, 71 insertions(+), 22 deletions(-) > > diff --git a/arch/powerpc/include/asm/book3s/32/pgtable.h > b/arch/powerpc/include/asm/book3s/32/pgtable.h > index 38b33dcfcc9d..6b8b2d57fdc8 100644 > --- a/arch/powerpc/include/asm/book3s/32/pgtable.h > +++ b/arch/powerpc/include/asm/book3s/32/pgtable.h > @@ -223,7 +223,8 @@ static inline void huge_ptep_set_wrprotect(struct > mm_struct *mm, > } > > > -static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry) > +static inline void __ptep_set_access_flags(struct mm_struct *mm, > + pte_t *ptep, pte_t entry) > { > unsigned long set = pte_val(entry) & > (_PAGE_DIRTY | _PAGE_ACCESSED | _PAGE_RW | _PAGE_EXEC); > diff --git a/arch/powerpc/include/asm/book3s/64/pgtable.h > b/arch/powerpc/include/asm/book3s/64/pgtable.h > index 263bf39ced40..8ec8be9495ba 100644 > --- a/arch/powerpc/include/asm/book3s/64/pgtable.h > +++ b/arch/powerpc/include/asm/book3s/64/pgtable.h > @@ -565,10 +565,11 @@ static inline bool check_pte_access(unsigned long > access, unsigned long ptev) > * Generic functions with hash/radix callbacks > */ > > -static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry) > +static inline void __ptep_set_access_flags(struct mm_struct *mm, > + pte_t *ptep, pte_t entry) > { > if (radix_enabled()) > - return radix__ptep_set_access_flags(ptep, entry); > + return radix__ptep_set_access_flags(mm, ptep, entry); > return hash__ptep_set_access_flags(ptep, entry); > } > > diff --git a/arch/powerpc/include/asm/book3s/64/radix.h > b/arch/powerpc/include/asm/book3s/64/radix.h > index a2fe8fbfbd3d..2a46dea8e1b1 100644 > --- a/arch/powerpc/include/asm/book3s/64/radix.h > +++ b/arch/powerpc/include/asm/book3s/64/radix.h > @@ -11,6 +11,11 @@ > #include > #endif > > +#ifndef __ASSEMBLY__ > +#include > +#include > +#endif > + > /* An empty PTE can still have a R or C writeback */ > #define RADIX_PTE_NONE_MASK (_PAGE_DIRTY | _PAGE_ACCESSED) > > @@ -105,11 +110,8 @@ > #define RADIX_PUD_TABLE_SIZE (sizeof(pud_t) << RADIX_PUD_INDEX_SIZE) > #define RADIX_PGD_TABLE_SIZE (sizeof(pgd_t) << RADIX_PGD_INDEX_SIZE) > > -static inline unsigned long radix__pte_update(struct mm_struct *mm, > - unsigned long addr, > - pte_t *ptep, unsigned long clr, > - unsigned long set, > - int huge) > +static inline unsigned long __radix_pte_update(pte_t *ptep, unsigned long > clr, > + unsigned long set) > { > pte_t pte; > unsigned long old_pte, new_pte; > @@ -121,9 +123,39 @@ static inline unsigned long radix__pte_update(struct > mm_struct *mm, > > } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); > > - /* We already do a sync in cmpxchg, is ptesync needed ?*/ > + return old_pte; > +} > + > + > +static inline unsigned long radix__pte_update(struct mm_struct *mm, > + unsigned long addr, > + pte_t *ptep, unsigned long clr, > + unsigned long set, > + int huge) > +{ > + unsigned long old_pte; > + > + if (cpu_has_feature(CPU_FTR_POWER9_DD1)) { > + > + unsigned long new_pte; > + > + old_pte = __radix_pte_update(ptep, ~0, 0); > + asm volatile("ptesync" : : : "memory"); > + /* > + * new value of pte > + */ > + new_pte = (old_pte | set) & ~clr; > + > + /* > + * For now let's do heavy pid flush > + * radix__flush_tlb_page_psize(mm, addr, mmu_virtual_psize); > + */ > + radix__flush_tlb_mm(mm); > + > + __radix_pte_update(ptep, 0, new_pte); > + } else > + old_pte = __radix_pte_update(ptep, clr, set); > asm volatile("ptesync" : : : "memory"); > - /* huge pages use the old page table lock */ > if (!huge) > assert_pte_locked(mm, addr); > > @@ -134,20 +166,33 @@ static inline unsigned long radix__pte_update(struct > mm_struct *mm, > * Set the dirty and/or accessed bits atomically in a linux PTE, this > * function doesn't need to invalidate tlb. > */ > -static inline void radix__ptep_set_access_flags(pte_t *ptep, pte_t entry) > +static inline void radix__ptep_set_access_flags(struct mm_struct *mm, > + pte_t *ptep, pte_t entry) > { > - pte_t pte; > - unsigned long old_pte, new_pte; > + > unsigned long set = pte_val(entry) & (_PAGE_DIRTY | _PAGE_ACCESSED | > _PAGE_RW | _PAGE_EXEC); > - do { > - pte = READ_ONCE(*ptep); > - old_pte = pte_val(pte); > + > + if (cpu_has_feature(CPU_FTR_POWER9_DD1)) { > + > + unsigned long old_pte, new_pte; > + > + old_pte = __radix_pte_update(ptep, ~0, 0); > + asm volatile("ptesync" : : : "memory"); > + /* > + * new value of pte > + */ > new_pte = old_pte | set; > > - } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); > + /* > + * For now let's do heavy pid flush > + * radix__flush_tlb_page_psize(mm, addr, mmu_virtual_psize); > + */ > + radix__flush_tlb_mm(mm); > > - /* We already do a sync in cmpxchg, is ptesync needed ?*/ > + __radix_pte_update(ptep, 0, new_pte); > + } else > + __radix_pte_update(ptep, 0, set); > asm volatile("ptesync" : : : "memory"); > } > > diff --git a/arch/powerpc/include/asm/nohash/32/pgtable.h > b/arch/powerpc/include/asm/nohash/32/pgtable.h > index 780847597514..c219ef7be53b 100644 > --- a/arch/powerpc/include/asm/nohash/32/pgtable.h > +++ b/arch/powerpc/include/asm/nohash/32/pgtable.h > @@ -267,7 +267,8 @@ static inline void huge_ptep_set_wrprotect(struct > mm_struct *mm, > } > > > -static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry) > +static inline void __ptep_set_access_flags(struct mm_struct *mm, > + pte_t *ptep, pte_t entry) > { > unsigned long set = pte_val(entry) & > (_PAGE_DIRTY | _PAGE_ACCESSED | _PAGE_RW | _PAGE_EXEC); > diff --git a/arch/powerpc/include/asm/nohash/64/pgtable.h > b/arch/powerpc/include/asm/nohash/64/pgtable.h > index d4d808cf905e..653a1838469d 100644 > --- a/arch/powerpc/include/asm/nohash/64/pgtable.h > +++ b/arch/powerpc/include/asm/nohash/64/pgtable.h > @@ -300,7 +300,8 @@ static inline void pte_clear(struct mm_struct *mm, > unsigned long addr, > /* Set the dirty and/or accessed bits atomically in a linux PTE, this > * function doesn't need to flush the hash entry > */ > -static inline void __ptep_set_access_flags(pte_t *ptep, pte_t entry) > +static inline void __ptep_set_access_flags(struct mm_struct *mm, > + pte_t *ptep, pte_t entry) > { > unsigned long bits = pte_val(entry) & > (_PAGE_DIRTY | _PAGE_ACCESSED | _PAGE_RW | _PAGE_EXEC); > diff --git a/arch/powerpc/mm/pgtable-book3s64.c > b/arch/powerpc/mm/pgtable-book3s64.c > index 34079302cc17..7328886bca4c 100644 > --- a/arch/powerpc/mm/pgtable-book3s64.c > +++ b/arch/powerpc/mm/pgtable-book3s64.c > @@ -35,7 +35,7 @@ int pmdp_set_access_flags(struct vm_area_struct *vma, > unsigned long address, > #endif > changed = !pmd_same(*(pmdp), entry); > if (changed) { > - __ptep_set_access_flags(pmdp_ptep(pmdp), pmd_pte(entry)); > + __ptep_set_access_flags(vma->vm_mm, pmdp_ptep(pmdp), > pmd_pte(entry)); > flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE); > } > return changed; > diff --git a/arch/powerpc/mm/pgtable.c b/arch/powerpc/mm/pgtable.c > index 0b6fb244d0a1..911fdfb63ec1 100644 > --- a/arch/powerpc/mm/pgtable.c > +++ b/arch/powerpc/mm/pgtable.c > @@ -224,7 +224,7 @@ int ptep_set_access_flags(struct vm_area_struct *vma, > unsigned long address, > if (changed) { > if (!is_vm_hugetlb_page(vma)) > assert_pte_locked(vma->vm_mm, address); > - __ptep_set_access_flags(ptep, entry); > + __ptep_set_access_flags(vma->vm_mm, ptep, entry); > flush_tlb_page(vma, address); > } > return changed;