From: Andi Kleen <a...@linux.intel.com>

Add the elision adaption state to struct mutex and use elide_adapt()
This allows mutexes that do not elide to automatically disable
elision on themselves for some time. This means we have a fail-safe
for mutexes that do not elide well and do not need to annotate
very mutex.

Signed-off-by: Andi Kleen <a...@linux.intel.com>
---
 arch/x86/include/asm/mutex.h |   20 +++++++++++++++-----
 arch/x86/kernel/rtm-locks.c  |   10 ++++++++--
 include/linux/mutex.h        |    3 +++
 3 files changed, 26 insertions(+), 7 deletions(-)

diff --git a/arch/x86/include/asm/mutex.h b/arch/x86/include/asm/mutex.h
index 7f585e3..bb9cf9d 100644
--- a/arch/x86/include/asm/mutex.h
+++ b/arch/x86/include/asm/mutex.h
@@ -10,8 +10,10 @@
 #define  ARCH_HAS_MUTEX_AND_OWN 1
 
 #include <linux/elide.h>
+#include <linux/jump_label.h>
 
-extern bool mutex_elision;
+extern struct static_key mutex_elision;
+extern struct elision_config mutex_elision_config;
 
 /*
  * Try speculation first and only do the normal locking and owner setting
@@ -21,8 +23,10 @@ extern bool mutex_elision;
 #define mutex_free(l) (atomic_read(&(l)->count) == 1)
 
 #define __mutex_fastpath_lock_and_own(l, s) ({                         \
-                       if (!elide_lock(mutex_elision,                  \
-                                       mutex_free(l))) {               \
+                       if (!elide_lock_adapt(mutex_elision,            \
+                                       mutex_free(l),                  \
+                                       &l->elision_adapt,              \
+                                       &mutex_elision_config)) {       \
                                __mutex_fastpath_lock(&(l)->count, s);  \
                                mutex_set_owner(l);                     \
                        }                                               \
@@ -37,7 +41,10 @@ extern bool mutex_elision;
 
 #define __mutex_fastpath_lock_retval_and_own(l, s) ({                  \
                        int ret = 0;                                    \
-                       if (!elide_lock(mutex_elision, mutex_free(l))) { \
+                       if (!elide_lock_adapt(mutex_elision,            \
+                                             mutex_free(l),            \
+                                             &l->elision_adapt,        \
+                                             &mutex_elision_config)) { \
                                ret = __mutex_fastpath_lock_retval(&(l)->count, 
s); \
                                if (!ret)                               \
                                        mutex_set_owner(l);             \
@@ -46,7 +53,10 @@ extern bool mutex_elision;
 
 #define __mutex_fastpath_trylock_and_own(l, s) ({                      \
                        int ret = 1;                                    \
-                       if (!elide_lock(mutex_elision, mutex_free(l))) {\
+                       if (!elide_lock_adapt(mutex_elision,            \
+                                             mutex_free(l),            \
+                                             &l->elision_adapt,        \
+                                             &mutex_elision_config)) { \
                                ret = __mutex_fastpath_trylock(&(l)->count, s); 
\
                                if (ret)                                \
                                        mutex_set_owner(l);             \
diff --git a/arch/x86/kernel/rtm-locks.c b/arch/x86/kernel/rtm-locks.c
index 8d4763d..a313a81 100644
--- a/arch/x86/kernel/rtm-locks.c
+++ b/arch/x86/kernel/rtm-locks.c
@@ -416,6 +416,9 @@ static unsigned rtm_patch(u8 type, u16 clobbers, void *ibuf,
        }
 }
 
+struct static_key mutex_elision = STATIC_KEY_INIT_FALSE;
+module_param(mutex_elision, static_key, 0644);
+
 void init_rtm_spinlocks(void)
 {
        if (!boot_cpu_has(X86_FEATURE_RTM))
@@ -441,10 +444,13 @@ void init_rtm_spinlocks(void)
        pv_irq_ops.irq_enable = PV_CALLEE_SAVE(rtm_irq_enable);
        pv_irq_ops.restore_fl = PV_CALLEE_SAVE(rtm_restore_fl);
        pv_init_ops.patch = rtm_patch;
+
+       static_key_slow_inc(&mutex_elision);
 }
 
-__read_mostly bool mutex_elision = true;
-module_param(mutex_elision, bool, 0644);
+__read_mostly struct elision_config mutex_elision_config =
+       DEFAULT_ELISION_CONFIG;
+TUNE_ELISION_CONFIG(mutex, mutex_elision_config);
 
 __read_mostly bool rwsem_elision = true;
 module_param(rwsem_elision, bool, 0644);
diff --git a/include/linux/mutex.h b/include/linux/mutex.h
index 0574095..7ae9a08 100644
--- a/include/linux/mutex.h
+++ b/include/linux/mutex.h
@@ -54,6 +54,9 @@ struct mutex {
 #if defined(CONFIG_DEBUG_MUTEXES) || defined(CONFIG_SMP)
        struct task_struct      *owner;
 #endif
+#ifdef CONFIG_RTM_LOCKS
+       short                   elision_adapt;
+#endif
 #ifdef CONFIG_DEBUG_MUTEXES
        const char              *name;
        void                    *magic;
-- 
1.7.7.6

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to