On 04/18/2015 11:40 AM, Peter Zijlstra wrote:
On Fri, Apr 17, 2015 at 10:03:18PM -0400, Waiman Long wrote:
@@ -478,7 +515,28 @@ struct rw_semaphore *rwsem_wake(struct rw_semaphore *sem)
  {
        unsigned long flags;

-       raw_spin_lock_irqsave(&sem->wait_lock, flags);
+       /*
+        * If a spinner is present, it is not necessary to do the wakeup.
+        * Try to do wakeup only if the trylock succeeds to minimize
+        * spinlock contention which may introduce too much delay in the
+        * unlock operation.
+        *
+        * In case the spinning writer is just going to break out of the
+        * waiting loop, it will still do a trylock in
+        * rwsem_down_write_failed() before sleeping.
+        * IOW, if rwsem_has_spinner() is true, it will guarantee at least
+        * one trylock attempt on the rwsem.
successful trylock? I think we're having 'issues' on if failed trylocks
(and cmpxchg) imply full barriers.

+        *
+        *    spinning writer
+        *    ---------------
+        * [S]   osq_unlock()
+        *       MB
+        * [RmW] rwsem_try_write_lock()
+        */
Ordering comes in pairs, this is incomplete.

I am sorry that I am a bit sloppy here. I have just sent out an updated patch to remedy this. I have added a smp_mb__after_atomic() to ensure proper memory ordering. However, I am not so sure if this primitive or just a simple smp_rmb() will be more expensive in other non-x86 architectures.

Cheers,
Longman
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to