On Wed, 2017-07-19 at 12:29 +1000, Balbir Singh wrote: > > > diff --git a/arch/powerpc/kvm/book3s_hv_rmhandlers.S > > b/arch/powerpc/kvm/book3s_hv_rmhandlers.S > > index 6ea4b53..4fb3581b 100644 > > --- a/arch/powerpc/kvm/book3s_hv_rmhandlers.S > > +++ b/arch/powerpc/kvm/book3s_hv_rmhandlers.S > > @@ -1522,6 +1522,18 @@ END_FTR_SECTION_IFCLR(CPU_FTR_ARCH_207S) > > std r6, VCPU_BESCR(r9) > > stw r7, VCPU_GUEST_PID(r9) > > std r8, VCPU_WORT(r9) > > + > > + /* Handle the case where the guest used an illegal PID */ > > + LOAD_REG_ADDR(r4, mmu_base_pid) > > + lwz r3, 0(r4) > > + cmpw cr0,r7,r3 > > + blt 1f > > So the boundary is [1..(1<<mmu_pid_bits-1)]? Do we flush the tlb > for pid 0 always?
What do you mean ? mmu_base_pid is going to be 1 << (mmu_pid_bits-1), we flush the TLB if the guest PID r7 is not lower than the boundary. Ntoe: this is patch v1, look at v2. > > + > > + /* Illegal PID, flush the TLB */ > > + bl radix_flush_pid > > + ld r9, HSTATE_KVM_VCPU(r13) > > +1: > > + > > BEGIN_FTR_SECTION > > mfspr r5, SPRN_TCSCR > > mfspr r6, SPRN_ACOP > > diff --git a/arch/powerpc/mm/mmu_context_book3s64.c > > b/arch/powerpc/mm/mmu_context_book3s64.c > > index abed1fe..183a67b 100644 > > --- a/arch/powerpc/mm/mmu_context_book3s64.c > > +++ b/arch/powerpc/mm/mmu_context_book3s64.c > > @@ -25,6 +25,10 @@ > > #include <asm/mmu_context.h> > > #include <asm/pgalloc.h> > > > > +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE > > +#include <asm/kvm_book3s_asm.h> > > +#endif > > + > > #include "icswx.h" > > > > static DEFINE_SPINLOCK(mmu_context_lock); > > @@ -126,9 +130,10 @@ static int hash__init_new_context(struct mm_struct *mm) > > static int radix__init_new_context(struct mm_struct *mm) > > { > > unsigned long rts_field; > > - int index; > > + int index, max_id; > > > > - index = alloc_context_id(1, PRTB_ENTRIES - 1); > > + max_id = (1 << mmu_pid_bits) - 1; > > + index = alloc_context_id(mmu_base_pid, max_id); > > if (index < 0) > > return index; > > > > @@ -247,8 +252,40 @@ void destroy_context(struct mm_struct *mm) > > } > > > > #ifdef CONFIG_PPC_RADIX_MMU > > -void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct > > *next) > > +void radix__switch_mmu_context(struct mm_struct *prev, struct mm_struct > > *next, > > + bool new_on_cpu) > > { > > + /* > > + * If this context hasn't run on that CPU before and KVM is > > + * around, there's a slim chance that the guest on another > > + * CPU just brought in obsolete translation into the TLB of > > + * this CPU due to a bad prefetch using the guest PID on > > + * the way into the hypervisor. > > + * > > + * We work around this here. If KVM is possible, we check if > > + * any sibling thread is in KVM. If it is, the window may exist > > + * and thus we flush that PID from the core. > > + * > > + * A potential future improvement would be to mark which PIDs > > + * have never been used on the system and avoid it if the PID > > + * is new and the process has no other cpumask bit set. > > Also due to the pid split, the chances of the context bringing in > TLB entries is low, but a bad guest could bring in stale entries > > > + */ > > +#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE > > + if (cpu_has_feature(CPU_FTR_HVMODE) && new_on_cpu) { > > + int cpu = smp_processor_id(); > > + int sib = cpu_first_thread_sibling(cpu); > > + bool flush = false; > > + > > + for (; sib <= cpu_last_thread_sibling(cpu) && !flush; sib++) { > > + if (sib == cpu) > > + continue; > > + if (paca[sib].kvm_hstate.kvm_vcpu) > > + flush = true; > > + } > > + if (flush) > > + radix__local_flush_all_mm(next); > > + } > > +#endif /* CONFIG_KVM_BOOK3S_HV_POSSIBLE */ > >