Figured this out last night:

On Jul 31, 2014, at 3:21 PM, Jarno Rajahalme <jrajaha...@nicira.com> wrote:
(snip)
> +#define atomic_compare_exchange__(DST, EXP, SRC, RES, CLOB)           \
> +    asm volatile("lock; cmpxchg %3,%1 ; "                             \
> +                 "      sete    %0      "                             \
> +                 "# atomic_compare_exchange__"                        \
> +                 : "=q" (RES),           /* 0 */                      \
> +                   "+m" (*DST),          /* 1 */                      \
> +                   "+a" (EXP)            /* 2 */                      \
> +                 : "r" (SRC)             /* 3 */                      \
> +                 : CLOB, "cc")
> +
> +/* All memory models are valid for read-modify-write operations.
> + *
> + * Valid memory_models for the read operation of the current value in

“memory models”

> + * the failure case are the same as for atomic read, but can not be
> + * stronger than the success memory model. */
> +#define atomic_compare_exchange_strong_explicit(DST, EXP, SRC, ORDER, 
> ORD_FAIL) \
> +    ({                                                              \
> +        typeof(DST) dst__ = (DST);                                  \
> +        typeof(DST) expp__ = (EXP);                                 \
> +        typeof(*DST) src__ = (SRC);                                 \
> +        typeof(*DST) exp__ = *expp__;                               \
> +        uint8_t res__;                                              \
> +                                                                    \
> +        if (ORDER > memory_order_consume) {                         \
> +            atomic_compare_exchange__(dst__, exp__, src__, res__,   \
> +                                      "memory");                    \
> +        } else {                                                    \
> +            atomic_compare_exchange__(dst__, exp__, src__, res__,   \
> +                                      "cc");                        \
> +        }                                                           \
> +        if (!res__) {                                               \
> +            *expp__ = exp__;                                        \
> +            atomic_compiler_barrier(ORD_FAIL);                      \

This barrier is not needed, as the barrier of the atomic_compare_exchange__ is 
always strong enough. Also the assignment here is not an atomic operation, the 
atomic load operation was performed by atomic_compare_exchange__, loading the 
value of *dst__ to exp__ in case of failure. Here we just move the value from 
one local variable to another; the compiler should have the right to optimize 
this in any way it likes.

Just to make this clear, this is the breakdown of the different allowed 
combinations:

ORDER > memory_order_consume: atomic_compare_exchange__ already implements a 
full CPU & compiler barrier
ORDER == memory_order_acquire: atomic_compare_exchange__ already implements the 
strongest possible barrier allowed for ORD_FAIL (memory_order_acquire)
ORDER == memory_order_consume: atomic_compare_exchange__ already implements the 
strongest possible barrier allowed for ORD_FAIL (memory_order_consume)
ORDER == memory_order_relaxed: atomic_compare_exchange__ already implements the 
strongest possible barrier allowed for ORD_FAIL (memory_order_relaxed)


> +        }                                                           \
> +        (bool)res__;                                                \
> +    })

  Jarno

_______________________________________________
dev mailing list
dev@openvswitch.org
http://openvswitch.org/mailman/listinfo/dev

Reply via email to