Nicholas Piggin <npig...@gmail.com> writes: > According to the architecture, the process table entry cache must be > flushed with tlbie RIC=2. > > Currently the process table entry is set to invalid right before the > PID is returned to the allocator, with no invalidation. This works on > existing implementations that are known to not cache the process table > entry for any except the current PIDR. > > It is architecturally correct and cleaner to invalidate with RIC=2 > after clearing the process table entry and before the PID is returned > to the allocator. This can be done in arch_exit_mmap that runs before > the final flush, and to ensure the final flush (fullmm) is always a > RIC=2 variant. >
Reviewed-by: Aneesh Kumar K.V <aneesh.ku...@linux.vnet.ibm.com> > Signed-off-by: Nicholas Piggin <npig...@gmail.com> > --- > arch/powerpc/include/asm/mmu_context.h | 4 ++++ > arch/powerpc/mm/mmu_context_book3s64.c | 25 ++++++++++++++++++++----- > arch/powerpc/mm/tlb-radix.c | 6 +++++- > 3 files changed, 29 insertions(+), 6 deletions(-) > > diff --git a/arch/powerpc/include/asm/mmu_context.h > b/arch/powerpc/include/asm/mmu_context.h > index a0d7145d6cd2..20eae6f76247 100644 > --- a/arch/powerpc/include/asm/mmu_context.h > +++ b/arch/powerpc/include/asm/mmu_context.h > @@ -164,9 +164,13 @@ static inline void arch_dup_mmap(struct mm_struct *oldmm, > { > } > > +#ifndef CONFIG_PPC_BOOK3S_64 > static inline void arch_exit_mmap(struct mm_struct *mm) > { > } > +#else > +extern void arch_exit_mmap(struct mm_struct *mm); > +#endif > > static inline void arch_unmap(struct mm_struct *mm, > struct vm_area_struct *vma, > diff --git a/arch/powerpc/mm/mmu_context_book3s64.c > b/arch/powerpc/mm/mmu_context_book3s64.c > index 05e15386d4cb..6d724dab27c2 100644 > --- a/arch/powerpc/mm/mmu_context_book3s64.c > +++ b/arch/powerpc/mm/mmu_context_book3s64.c > @@ -216,19 +216,34 @@ void destroy_context(struct mm_struct *mm) > #ifdef CONFIG_SPAPR_TCE_IOMMU > WARN_ON_ONCE(!list_empty(&mm->context.iommu_group_mem_list)); > #endif > + if (radix_enabled()) > + WARN_ON(process_tb[mm->context.id].prtb0 != 0); > + else > + subpage_prot_free(mm); > + destroy_pagetable_page(mm); > + __destroy_context(mm->context.id); > + mm->context.id = MMU_NO_CONTEXT; > +} > + > +void arch_exit_mmap(struct mm_struct *mm) > +{ > if (radix_enabled()) { > /* > * Radix doesn't have a valid bit in the process table > * entries. However we know that at least P9 implementation > * will avoid caching an entry with an invalid RTS field, > * and 0 is invalid. So this will do. > + * > + * This runs before the "fullmm" tlb flush in exit_mmap, > + * which does a RIC=2 tlbie to clear the process table > + * entry. See the "fullmm" comments in tlb-radix.c. > + * > + * No barrier required here after the store because > + * this process will do the invalidate, which starts with > + * ptesync. > */ > process_tb[mm->context.id].prtb0 = 0; > - } else > - subpage_prot_free(mm); > - destroy_pagetable_page(mm); > - __destroy_context(mm->context.id); > - mm->context.id = MMU_NO_CONTEXT; > + } > } > > #ifdef CONFIG_PPC_RADIX_MMU > diff --git a/arch/powerpc/mm/tlb-radix.c b/arch/powerpc/mm/tlb-radix.c > index 18170dc264aa..0c8464653aa3 100644 > --- a/arch/powerpc/mm/tlb-radix.c > +++ b/arch/powerpc/mm/tlb-radix.c > @@ -297,10 +297,14 @@ void radix__tlb_flush(struct mmu_gather *tlb) > psize = radix_get_mmu_psize(page_size); > /* > * if page size is not something we understand, do a full mm flush > + * > + * A "fullmm" flush must always do a flush_all_mm (RIC=2) flush > + * that flushes the process table entry cache upon process teardown. > + * See the comment for radix in arch_exit_mmap(). > */ > if (psize != -1 && !tlb->fullmm && !tlb->need_flush_all) > radix__flush_tlb_range_psize(mm, tlb->start, tlb->end, psize); > - else if (tlb->need_flush_all) { > + else if (tlb->fullmm || tlb->need_flush_all) { > tlb->need_flush_all = 0; > radix__flush_all_mm(mm); > } else > -- > 2.13.3