Nicholas Piggin <npig...@gmail.com> writes: > When an interrupt is taken, the SRR registers are set to return to > where it left off. Unless they are modified in the meantime, or the > return address or MSR are modified, there is no need to reload these > registers when returning from interrupt. > > Introduce per-CPU flags that track the validity of SRR and HSRR > registers, clear them when returning from interrupt, using the registers > for something else (e.g., OPAL calls), or adjusting return address or MSR. > > This improves the performance of interrupt returns. > > XXX: may not need to invalidate both hsrr and srr all the time > > Signed-off-by: Nicholas Piggin <npig...@gmail.com> > ---
I needed something like below to get 32-bit building. cheers diff --git a/arch/powerpc/include/asm/ptrace.h b/arch/powerpc/include/asm/ptrace.h index 6d6237e0cbd7..7f9bbd19db10 100644 --- a/arch/powerpc/include/asm/ptrace.h +++ b/arch/powerpc/include/asm/ptrace.h @@ -153,15 +153,21 @@ static inline void regs_set_return_value(struct pt_regs *regs, unsigned long rc) regs->gpr[3] = rc; } -static inline void regs_set_return_ip(struct pt_regs *regs, unsigned long ip) +static inline void invalidate_srrs(void) { - regs->nip = ip; #ifdef CONFIG_PPC_BOOK3S_64 + // XXX: We may not need to invalidate both hsrr and srr all the time local_paca->hsrr_valid = 0; local_paca->srr_valid = 0; #endif } +static inline void regs_set_return_ip(struct pt_regs *regs, unsigned long ip) +{ + regs->nip = ip; + invalidate_srrs(); +} + static inline void regs_add_return_ip(struct pt_regs *regs, long offset) { regs_set_return_ip(regs, regs->nip + offset); diff --git a/arch/powerpc/kernel/process.c b/arch/powerpc/kernel/process.c index 200b4805f999..82623b57e2d6 100644 --- a/arch/powerpc/kernel/process.c +++ b/arch/powerpc/kernel/process.c @@ -98,8 +98,7 @@ static void check_if_tm_restore_required(struct task_struct *tsk) !test_thread_flag(TIF_RESTORE_TM)) { tsk->thread.ckpt_regs.msr = tsk->thread.regs->msr; set_thread_flag(TIF_RESTORE_TM); - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + invalidate_srrs(); } } @@ -164,8 +163,7 @@ static void __giveup_fpu(struct task_struct *tsk) if (cpu_has_feature(CPU_FTR_VSX)) msr &= ~MSR_VSX; tsk->thread.regs->msr = msr; - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + invalidate_srrs(); } void giveup_fpu(struct task_struct *tsk) @@ -249,8 +247,7 @@ static void __giveup_altivec(struct task_struct *tsk) if (cpu_has_feature(CPU_FTR_VSX)) msr &= ~MSR_VSX; tsk->thread.regs->msr = msr; - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + invalidate_srrs(); } void giveup_altivec(struct task_struct *tsk) @@ -566,8 +563,7 @@ void notrace restore_math(struct pt_regs *regs) msr_check_and_clear(new_msr); regs->msr |= new_msr | fpexc_mode; - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + invalidate_srrs(); } } #endif /* CONFIG_PPC_BOOK3S_64 */ @@ -1293,8 +1289,7 @@ struct task_struct *__switch_to(struct task_struct *prev, atomic_read(¤t->mm->context.vas_windows))) asm volatile(PPC_CP_ABORT); } - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + invalidate_srrs(); #endif /* CONFIG_PPC_BOOK3S_64 */ return last; @@ -1884,8 +1879,7 @@ void start_thread(struct pt_regs *regs, unsigned long start, unsigned long sp) current->thread.load_tm = 0; #endif /* CONFIG_PPC_TRANSACTIONAL_MEM */ - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + invalidate_srrs(); } EXPORT_SYMBOL(start_thread); @@ -1936,8 +1930,7 @@ int set_fpexc_mode(struct task_struct *tsk, unsigned int val) if (regs != NULL && (regs->msr & MSR_FP) != 0) { regs->msr = (regs->msr & ~(MSR_FE0|MSR_FE1)) | tsk->thread.fpexc_mode; - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + invalidate_srrs(); } return 0; } @@ -1990,8 +1983,7 @@ int set_endian(struct task_struct *tsk, unsigned int val) else return -EINVAL; - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + invalidate_srrs(); return 0; } diff --git a/arch/powerpc/kernel/syscalls.c b/arch/powerpc/kernel/syscalls.c index 4cb38afa28a8..9d1d6070a516 100644 --- a/arch/powerpc/kernel/syscalls.c +++ b/arch/powerpc/kernel/syscalls.c @@ -115,8 +115,8 @@ SYSCALL_DEFINE0(switch_endian) struct thread_info *ti; current->thread.regs->msr ^= MSR_LE; - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + + invalidate_srrs(); /* * Set TIF_RESTOREALL so that r3 isn't clobbered on return to diff --git a/arch/powerpc/lib/sstep.c b/arch/powerpc/lib/sstep.c index 96505d4bba1c..2b94bf21d6ae 100644 --- a/arch/powerpc/lib/sstep.c +++ b/arch/powerpc/lib/sstep.c @@ -3480,8 +3480,7 @@ int emulate_step(struct pt_regs *regs, struct ppc_inst instr) unsigned long val; unsigned long ea; - local_paca->hsrr_valid = 0; - local_paca->srr_valid = 0; + invalidate_srrs(); r = analyse_instr(&op, regs, instr); if (r < 0)