From: Andi Kleen <a...@linux.intel.com>

Very straight forward. Use the non-adaptive elision wrappers for
bit spinlocks. This is useful because they perform very poorly
under contention.

The functions are a bit on the big side for inlining now, but
I kept them inline for now.

Signed-off-by: Andi Kleen <a...@linux.intel.com>
---
 arch/x86/kernel/rtm-locks.c  |    5 +++--
 include/linux/bit_spinlock.h |   16 ++++++++++++++++
 2 files changed, 19 insertions(+), 2 deletions(-)

diff --git a/arch/x86/kernel/rtm-locks.c b/arch/x86/kernel/rtm-locks.c
index bc3275a..40a0e7d 100644
--- a/arch/x86/kernel/rtm-locks.c
+++ b/arch/x86/kernel/rtm-locks.c
@@ -464,14 +464,15 @@ void init_rtm_spinlocks(void)
 
        static_key_slow_inc(&rwlock_elision);
        static_key_slow_inc(&mutex_elision);
+       bitlock_elision = true;
 }
 
 __read_mostly struct elision_config mutex_elision_config =
        DEFAULT_ELISION_CONFIG;
 TUNE_ELISION_CONFIG(mutex, mutex_elision_config);
 
-__read_mostly bool rwsem_elision = true;
-module_param(rwsem_elision, bool, 0644);
+__read_mostly bool bitlock_elision;
+module_param(bitlock_elision, bool, 0644);
 
 module_param_cb(lock_el_skip, &param_ops_percpu_uint, &lock_el_skip,
                0644);
diff --git a/include/linux/bit_spinlock.h b/include/linux/bit_spinlock.h
index 3b5bafc..2954b86 100644
--- a/include/linux/bit_spinlock.h
+++ b/include/linux/bit_spinlock.h
@@ -5,6 +5,9 @@
 #include <linux/preempt.h>
 #include <linux/atomic.h>
 #include <linux/bug.h>
+#include <linux/elide.h>
+
+extern bool bitlock_elision;
 
 /*
  *  bit-based spin_lock()
@@ -14,6 +17,9 @@
  */
 static inline void bit_spin_lock(int bitnum, unsigned long *addr)
 {
+       if (elide_lock(bitlock_elision, test_bit(bitnum, addr) == 0))
+               return;
+
        /*
         * Assuming the lock is uncontended, this never enters
         * the body of the outer loop. If it is contended, then
@@ -39,6 +45,9 @@ static inline void bit_spin_lock(int bitnum, unsigned long 
*addr)
  */
 static inline int bit_spin_trylock(int bitnum, unsigned long *addr)
 {
+       if (elide_lock(bitlock_elision, test_bit(bitnum, addr) == 0))
+               return 1;
+
        preempt_disable();
 #if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK)
        if (unlikely(test_and_set_bit_lock(bitnum, addr))) {
@@ -55,6 +64,9 @@ static inline int bit_spin_trylock(int bitnum, unsigned long 
*addr)
  */
 static inline void bit_spin_unlock(int bitnum, unsigned long *addr)
 {
+       if (elide_unlock(test_bit(bitnum, addr) == 0))
+               return;
+
 #ifdef CONFIG_DEBUG_SPINLOCK
        BUG_ON(!test_bit(bitnum, addr));
 #endif
@@ -72,6 +84,9 @@ static inline void bit_spin_unlock(int bitnum, unsigned long 
*addr)
  */
 static inline void __bit_spin_unlock(int bitnum, unsigned long *addr)
 {
+       if (elide_unlock(test_bit(bitnum, addr) == 0))
+               return;
+
 #ifdef CONFIG_DEBUG_SPINLOCK
        BUG_ON(!test_bit(bitnum, addr));
 #endif
@@ -87,6 +102,7 @@ static inline void __bit_spin_unlock(int bitnum, unsigned 
long *addr)
  */
 static inline int bit_spin_is_locked(int bitnum, unsigned long *addr)
 {
+       elide_abort();
 #if defined(CONFIG_SMP) || defined(CONFIG_DEBUG_SPINLOCK)
        return test_bit(bitnum, addr);
 #elif defined CONFIG_PREEMPT_COUNT
-- 
1.7.7.6

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to