This simplifies spin unlock code and mmio primitives. This may not be the best use of an r14 bit, but it was a simple first proof of concept after the per-cpu data_offset, and so it can stay until we get low on bits. --- arch/powerpc/include/asm/io.h | 11 ++++------ arch/powerpc/include/asm/paca.h | 44 ++++++++++++++++++++++++++++++++++++- arch/powerpc/include/asm/spinlock.h | 21 ++++++++++-------- arch/powerpc/xmon/xmon.c | 1 - 4 files changed, 59 insertions(+), 18 deletions(-)
diff --git a/arch/powerpc/include/asm/io.h b/arch/powerpc/include/asm/io.h index 422f99cf9924..c817f3a83fcc 100644 --- a/arch/powerpc/include/asm/io.h +++ b/arch/powerpc/include/asm/io.h @@ -104,8 +104,8 @@ extern bool isa_io_special; * */ -#ifdef CONFIG_PPC64 -#define IO_SET_SYNC_FLAG() do { local_paca->io_sync = 1; } while(0) +#if defined(CONFIG_PPC64) && defined(CONFIG_SMP) +#define IO_SET_SYNC_FLAG() do { r14_set_bits(R14_BIT_IO_SYNC); } while(0) #else #define IO_SET_SYNC_FLAG() #endif @@ -673,11 +673,8 @@ static inline void name at \ */ static inline void mmiowb(void) { - unsigned long tmp; - - __asm__ __volatile__("sync; li %0,0; stb %0,%1(13)" - : "=&r" (tmp) : "i" (offsetof(struct paca_struct, io_sync)) - : "memory"); + __asm__ __volatile__("sync" : : : "memory"); + r14_clear_bits(R14_BIT_IO_SYNC); } #endif /* !CONFIG_PPC32 */ diff --git a/arch/powerpc/include/asm/paca.h b/arch/powerpc/include/asm/paca.h index 4dd4ac69e84f..408fa079e00d 100644 --- a/arch/powerpc/include/asm/paca.h +++ b/arch/powerpc/include/asm/paca.h @@ -35,12 +35,55 @@ register struct paca_struct *local_paca asm("r13"); #ifdef CONFIG_PPC_BOOK3S + +#define R14_BIT_IO_SYNC 0x0001 + /* * The top 32-bits of r14 is used as the per-cpu offset, shifted by PAGE_SHIFT. * The per-cpu could be moved completely to vmalloc space if we had large * vmalloc page mapping? (no, must access it in real mode). */ register u64 local_r14 asm("r14"); + +/* + * r14 should not be modified by C code, because we can not guarantee it + * will be done with non-atomic (vs interrupts) read-modify-write sequences. + * All updates must be of the form `op r14,r14,xxx` or similar (i.e., atomic + * updates). + * + * Make asm statements have r14 for input and output so that the compiler + * does not re-order it with respect to other r14 manipulations. + */ +static inline void r14_set_bits(unsigned long mask) +{ + if (__builtin_constant_p(mask)) + asm volatile("ori %0,%0,%2\n" + : "=r" (local_r14) + : "0" (local_r14), "i" (mask)); + else + asm volatile("or %0,%0,%2\n" + : "=r" (local_r14) + : "0" (local_r14), "r" (mask)); +} + +static inline void r14_flip_bits(unsigned long mask) +{ + if (__builtin_constant_p(mask)) + asm volatile("xori %0,%0,%2\n" + : "=r" (local_r14) + : "0" (local_r14), "i" (mask)); + else + asm volatile("xor %0,%0,%2\n" + : "=r" (local_r14) + : "0" (local_r14), "r" (mask)); +} + +static inline void r14_clear_bits(unsigned long mask) +{ + asm volatile("andc %0,%0,%2\n" + : "=r" (local_r14) + : "0" (local_r14), "r" (mask)); +} #endif #if defined(CONFIG_DEBUG_PREEMPT) && defined(CONFIG_SMP) @@ -169,7 +212,6 @@ struct paca_struct { u16 trap_save; /* Used when bad stack is encountered */ u8 irq_soft_mask; /* mask for irq soft masking */ u8 irq_happened; /* irq happened while soft-disabled */ - u8 io_sync; /* writel() needs spin_unlock sync */ u8 irq_work_pending; /* IRQ_WORK interrupt while soft-disable */ u8 nap_state_lost; /* NV GPR values lost in power7_idle */ u64 sprg_vdso; /* Saved user-visible sprg */ diff --git a/arch/powerpc/include/asm/spinlock.h b/arch/powerpc/include/asm/spinlock.h index b9ebc3085fb7..182bb9304c79 100644 --- a/arch/powerpc/include/asm/spinlock.h +++ b/arch/powerpc/include/asm/spinlock.h @@ -40,16 +40,9 @@ #endif #if defined(CONFIG_PPC64) && defined(CONFIG_SMP) -#define CLEAR_IO_SYNC (get_paca()->io_sync = 0) -#define SYNC_IO do { \ - if (unlikely(get_paca()->io_sync)) { \ - mb(); \ - get_paca()->io_sync = 0; \ - } \ - } while (0) +#define CLEAR_IO_SYNC do { r14_clear_bits(R14_BIT_IO_SYNC); } while(0) #else #define CLEAR_IO_SYNC -#define SYNC_IO #endif #ifdef CONFIG_PPC_PSERIES @@ -165,9 +158,19 @@ void arch_spin_lock_flags(arch_spinlock_t *lock, unsigned long flags) static inline void arch_spin_unlock(arch_spinlock_t *lock) { - SYNC_IO; +#if defined(CONFIG_PPC64) && defined(CONFIG_SMP) + bool io_sync = local_r14 & R14_BIT_IO_SYNC; + if (unlikely(io_sync)) { + mb(); + CLEAR_IO_SYNC; + } else { + __asm__ __volatile__("# arch_spin_unlock\n\t" + PPC_RELEASE_BARRIER: : :"memory"); + } +#else __asm__ __volatile__("# arch_spin_unlock\n\t" PPC_RELEASE_BARRIER: : :"memory"); +#endif lock->slock = 0; } diff --git a/arch/powerpc/xmon/xmon.c b/arch/powerpc/xmon/xmon.c index a53454f61d09..40f0d02ae92d 100644 --- a/arch/powerpc/xmon/xmon.c +++ b/arch/powerpc/xmon/xmon.c @@ -2393,7 +2393,6 @@ static void dump_one_paca(int cpu) DUMP(p, trap_save, "x"); DUMP(p, irq_soft_mask, "x"); DUMP(p, irq_happened, "x"); - DUMP(p, io_sync, "x"); DUMP(p, irq_work_pending, "x"); DUMP(p, nap_state_lost, "x"); DUMP(p, sprg_vdso, "llx"); -- 2.15.0