On Sun, May 08, 2016 at 09:56:10PM -0700, Davidlohr Bueso wrote:
> Readers that are awoken will expect a nil ->task indicating
> that a wakeup has occurred. There is a mismatch between the
> smp_mb() and its documentation, in that the serialization is
> done between reading the task and the nil store. Furthermore,
> in addition to having the overlapping of loads and stores to
> waiter->task guaranteed to be ordered within that CPU, both
> wake_up_process() originally and now wake_q_add() already
> imply barriers upon successful calls, which serves the comment.
> 
> Just atomically do a xchg() and simplify the whole thing. We can
> use relaxed semantics as before mentioned in addition to the
> barrier provided by wake_q_add(), delaying there is no risk in
> reordering with the actual wakeup.

> @@ -190,24 +189,18 @@ __rwsem_mark_wake(struct rw_semaphore *sem,
>       next = sem->wait_list.next;
>       loop = woken;
>       do {
> +             struct task_struct *tsk;
> +
>               waiter = list_entry(next, struct rwsem_waiter, list);
>               next = waiter->list.next;
> -             tsk = waiter->task;
> -             /*
> -              * Make sure we do not wakeup the next reader before
> -              * setting the nil condition to grant the next reader;
> -              * otherwise we could miss the wakeup on the other
> -              * side and end up sleeping again. See the pairing
> -              * in rwsem_down_read_failed().
> -              */
> -             smp_mb();
> -             waiter->task = NULL;
> +
> +             tsk = xchg_relaxed(&waiter->task, NULL);
>               wake_q_add(wake_q, tsk);

Not a great fan of this patch; it again doesn't fix the race, and
smp_store_release() is a cheaper option on x86.

Reply via email to