On Thu, 13 Sep 2012 20:15:26 -0700
Vikram Mulukutla <mark...@codeaurora.org> wrote:

> The logic in do_raw_spin_lock attempts to acquire a spinlock
> by invoking arch_spin_trylock in a loop with a delay between
> each attempt. Now consider the following situation in a 2
> CPU system:
> 
> 1. CPU-0 continually acquires and releases a spinlock in a
>    tight loop; it stays in this loop until some condition X
>    is satisfied. X can only be satisfied by another CPU.
> 
> 2. CPU-1 tries to acquire the same spinlock, in an attempt
>    to satisfy the aforementioned condition X. However, it
>    never sees the unlocked value of the lock because the
>    debug spinlock code uses trylock instead of just lock;
>    it checks at all the wrong moments - whenever CPU-0 has
>    locked the lock.
> 
> Now in the absence of debug spinlocks, the architecture specific
> spinlock code can correctly allow CPU-1 to wait in a "queue"
> (e.g., ticket spinlocks), ensuring that it acquires the lock at
> some point. However, with the debug spinlock code, livelock
> can easily occur due to the use of try_lock, which obviously
> cannot put the CPU in that "queue". This queueing mechanism is
> implemented in both x86 and ARM spinlock code.
> 
> Note that the situation mentioned above is not hypothetical.
> A real problem was encountered where CPU-0 was running
> hrtimer_cancel with interrupts disabled, and CPU-1 was attempting
> to run the hrtimer that CPU-0 was trying to cancel.

I'm surprised.  Yes, I guess it will happen if both CPUs have local
interrupts disabled.  And perhaps serialisation of the locked operation
helps.

> ...
>
> --- a/lib/spinlock_debug.c
> +++ b/lib/spinlock_debug.c
> @@ -107,23 +107,27 @@ static void __spin_lock_debug(raw_spinlock_t *lock)
>  {
>       u64 i;
>       u64 loops = loops_per_jiffy * HZ;
> -     int print_once = 1;
>  
> -     for (;;) {
> -             for (i = 0; i < loops; i++) {
> -                     if (arch_spin_trylock(&lock->raw_lock))
> -                             return;
> -                     __delay(1);
> -             }
> -             /* lockup suspected: */
> -             if (print_once) {
> -                     print_once = 0;
> -                     spin_dump(lock, "lockup suspected");
> +     for (i = 0; i < loops; i++) {
> +             if (arch_spin_trylock(&lock->raw_lock))
> +                     return;
> +             __delay(1);
> +     }
> +     /* lockup suspected: */
> +     spin_dump(lock, "lockup suspected");
>  #ifdef CONFIG_SMP
> -                     trigger_all_cpu_backtrace();
> +     trigger_all_cpu_backtrace();
>  #endif
> -             }
> -     }
> +
> +     /*
> +      * In case the trylock above was causing a livelock, give the lower
> +      * level arch specific lock code a chance to acquire the lock. We have
> +      * already printed a warning/backtrace at this point. The non-debug arch
> +      * specific code might actually succeed in acquiring the lock. If it is
> +      * not successful, the end-result is the same - there is no forward
> +      * progress.
> +      */
> +     arch_spin_lock(&lock->raw_lock);
>  }

The change looks reasonable to me.  I suggest we disambiguate that
comment a bit:

--- 
a/lib/spinlock_debug.c~lib-spinlock_debug-avoid-livelock-in-do_raw_spin_lock-fix
+++ a/lib/spinlock_debug.c
@@ -120,7 +120,7 @@ static void __spin_lock_debug(raw_spinlo
 #endif
 
        /*
-        * In case the trylock above was causing a livelock, give the lower
+        * The trylock above was causing a livelock.  Give the lower
         * level arch specific lock code a chance to acquire the lock. We have
         * already printed a warning/backtrace at this point. The non-debug arch
         * specific code might actually succeed in acquiring the lock. If it is
_

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to