From: Andi Kleen <a...@linux.intel.com> Very straight forward. Use the non-adaptive elision wrappers for bit spinlocks. This is useful because they perform very poorly under contention.
The functions are a bit on the big side for inlining now, but I kept them inline for now. Signed-off-by: Andi Kleen <a...@linux.intel.com> --- arch/x86/kernel/rtm-locks.c | 5 +++-- include/linux/bit_spinlock.h | 16 ++++++++++++++++ 2 files changed, 19 insertions(+), 2 deletions(-) diff --git a/arch/x86/kernel/rtm-locks.c b/arch/x86/kernel/rtm-locks.c index bc3275a..40a0e7d 100644 --- a/arch/x86/kernel/rtm-locks.c +++ b/arch/x86/kernel/rtm-locks.c @@ -464,14 +464,15 @@ void init_rtm_spinlocks(void) static_key_slow_inc(&rwlock_elision); static_key_slow_inc(&mutex_elision); + bitlock_elision = true; } __read_mostly struct elision_config mutex_elision_config = DEFAULT_ELISION_CONFIG; TUNE_ELISION_CONFIG(mutex, mutex_elision_config); -__read_mostly bool rwsem_elision = true; -module_param(rwsem_elision, bool, 0644); +__read_mostly bool bitlock_elision; +module_param(bitlock_elision, bool, 0644); module_param_cb(lock_el_skip, ¶m_ops_percpu_uint, &lock_el_skip, 0644); diff --git a/include/linux/bit_spinlock.h b/include/linux/bit_spinlock.h index 3b5bafc..2954b86 100644 --- a/include/linux/bit_spinlock.h +++ b/include/linux/bit_spinlock.h @@ -5,6 +5,9 @@ #include <linux/preempt.h> #include <linux/atomic.h> #include <linux/bug.h> +#include <linux/elide.h> + +extern bool bitlock_elision; /* * bit-based spin_lock() @@ -14,6 +17,9 @@ */ static inline void bit_spin_lock(int bitnum, unsigned long *addr) { + if (elide_lock(bitlock_elision, test_bit(bitnum, addr) == 0)) + return; + /* * Assuming the lock is uncontended, this never enters * the body of the outer loop. If it is contended, then @@ -39,6 +45,9 @@ static inline void bit_spin_lock(int bitnum, unsigned long *addr) */ static inline int bit_spin_trylock(int bitnum, unsigned long *addr) { + if (elide_lock(bitlock_elision, test_bit(bitnum, addr) == 0)) + return 1; + preempt_disable(); #if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK) if (unlikely(test_and_set_bit_lock(bitnum, addr))) { @@ -55,6 +64,9 @@ static inline int bit_spin_trylock(int bitnum, unsigned long *addr) */ static inline void bit_spin_unlock(int bitnum, unsigned long *addr) { + if (elide_unlock(test_bit(bitnum, addr) == 0)) + return; + #ifdef CONFIG_DEBUG_SPINLOCK BUG_ON(!test_bit(bitnum, addr)); #endif @@ -72,6 +84,9 @@ static inline void bit_spin_unlock(int bitnum, unsigned long *addr) */ static inline void __bit_spin_unlock(int bitnum, unsigned long *addr) { + if (elide_unlock(test_bit(bitnum, addr) == 0)) + return; + #ifdef CONFIG_DEBUG_SPINLOCK BUG_ON(!test_bit(bitnum, addr)); #endif @@ -87,6 +102,7 @@ static inline void __bit_spin_unlock(int bitnum, unsigned long *addr) */ static inline int bit_spin_is_locked(int bitnum, unsigned long *addr) { + elide_abort(); #if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK) return test_bit(bitnum, addr); #elif defined CONFIG_PREEMPT_COUNT -- 1.7.7.6 -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/