On Wed, Jun 07, 2023 at 02:53:52PM -0400, Olivier Dion wrote:
> Introducing the CMM memory model with the following new primitives:
> 
>   - uatomic_load(addr, memory_order)
> 
>   - uatomic_store(addr, value, memory_order)
>   - uatomic_and_mo(addr, mask, memory_order)
>   - uatomic_or_mo(addr, mask, memory_order)
>   - uatomic_add_mo(addr, value, memory_order)
>   - uatomic_sub_mo(addr, value, memory_order)
>   - uatomic_inc_mo(addr, memory_order)
>   - uatomic_dec_mo(addr, memory_order)
> 
>   - uatomic_add_return_mo(addr, value, memory_order)
>   - uatomic_sub_return_mo(addr, value, memory_order)
> 
>   - uatomic_xchg_mo(addr, value, memory_order)
> 
>   - uatomic_cmpxchg_mo(addr, old, new,
>                        memory_order_success,
>                        memory_order_failure)
> 
> The CMM memory model reflects the C11 memory model with an additional
> CMM_SEQ_CST_FENCE memory order. The memory order can be selected through
> the enum cmm_memorder.
> 
> * With Atomic Builtins
> 
> If configured with atomic builtins, the correspondence between the CMM
> memory model and the C11 memory model is a one to one at the exception
> of the CMM_SEQ_CST_FENCE memory order which implies the memory order
> CMM_SEQ_CST and a thread fence after the operation.
> 
> * Without Atomic Builtins
> 
> However, if not configured with atomic builtins, the following stipulate
> the memory model.
> 
> For load operations with uatomic_load(), the memory orders CMM_RELAXED,
> CMM_CONSUME, CMM_ACQUIRE, CMM_SEQ_CST and CMM_SEQ_CST_FENCE are
> allowed. A barrier may be inserted before and after the load from memory
> depending on the memory order:
> 
>   - CMM_RELAXED: No barrier
>   - CMM_CONSUME: Memory barrier after read
>   - CMM_ACQUIRE: Memory barrier after read
>   - CMM_SEQ_CST: Memory barriers before and after read
>   - CMM_SEQ_CST_FENCE: Memory barriers before and after read
> 
> For store operations with uatomic_store(), the memory orders
> CMM_RELAXED, CMM_RELEASE, CMM_SEQ_CST and CMM_SEQ_CST_FENCE are
> allowed. A barrier may be inserted before and after the store to memory
> depending on the memory order:
> 
>   - CMM_RELAXED: No barrier
>   - CMM_RELEASE: Memory barrier before operation
>   - CMM_SEQ_CST: Memory barriers before and after operation
>   - CMM_SEQ_CST_FENCE: Memory barriers before and after operation
> 
> For load/store operations with uatomic_and_mo(), uatomic_or_mo(),
> uatomic_add_mo(), uatomic_sub_mo(), uatomic_inc_mo(), uatomic_dec_mo(),
> uatomic_add_return_mo() and uatomic_sub_return_mo(), all memory orders
> are allowed. A barrier may be inserted before and after the operation
> depending on the memory order:
> 
>   - CMM_RELAXED: No barrier
>   - CMM_ACQUIRE: Memory barrier after operation
>   - CMM_CONSUME: Memory barrier after operation
>   - CMM_RELEASE: Memory barrier before operation
>   - CMM_ACQ_REL: Memory barriers before and after operation
>   - CMM_SEQ_CST: Memory barriers before and after operation
>   - CMM_SEQ_CST_FENCE: Memory barriers before and after operation
> 
> For the exchange operation uatomic_xchg_mo(), any memory order is
> valid. A barrier may be inserted before and after the exchange to memory
> depending on the memory order:
> 
>   - CMM_RELAXED: No barrier
>   - CMM_ACQUIRE: Memory barrier after operation
>   - CMM_CONSUME: Memory barrier after operation
>   - CMM_RELEASE: Memory barrier before operation
>   - CMM_ACQ_REL: Memory barriers before and after operation
>   - CMM_SEQ_CST: Memory barriers before and after operation
>   - CMM_SEQ_CST_FENCE: Memory barriers before and after operation
> 
> For the compare exchange operation uatomic_cmpxchg_mo(), the success
> memory order can be anything while the failure memory order cannot be
> CMM_RELEASE nor CMM_ACQ_REL and cannot be stronger than the success
> memory order. A barrier may be inserted before and after the store to
> memory depending on the memory orders:
> 
>  Success memory order:
> 
>   - CMM_RELAXED: No barrier
>   - CMM_ACQUIRE: Memory barrier after operation
>   - CMM_CONSUME: Memory barrier after operation
>   - CMM_RELEASE: Memory barrier before operation
>   - CMM_ACQ_REL: Memory barriers before and after operation
>   - CMM_SEQ_CST: Memory barriers before and after operation
>   - CMM_SEQ_CST_FENCE: Memory barriers before and after operation
> 
>   Barriers after the operations are only emitted if the compare exchange
>   succeed.
> 
>  Failure memory order:
>   - CMM_RELAXED: No barrier
>   - CMM_ACQUIRE: Memory barrier after operation
>   - CMM_CONSUME: Memory barrier after operation
>   - CMM_SEQ_CST: Memory barriers before and after operation
>   - CMM_SEQ_CST_FENCE: Memory barriers before and after operation
> 
>   Barriers after the operations are only emitted if the compare exchange
>   failed.  Barriers before the operation are never emitted by this
>   memory order.
> 
> Change-Id: I213ba19c84e82a63083f00143a3142ffbdab1d52
> Co-authored-by: Mathieu Desnoyers <mathieu.desnoy...@efficios.com>
> Signed-off-by: Olivier Dion <od...@efficios.com>
> ---
>  doc/uatomic-api.md                      |   3 +-
>  include/Makefile.am                     |   2 +
>  include/urcu/static/pointer.h           |  40 ++--
>  include/urcu/uatomic.h                  |  63 ++++++-
>  include/urcu/uatomic/builtins-generic.h | 170 +++++++++++++++++
>  include/urcu/uatomic/builtins.h         |  79 ++++++++
>  include/urcu/uatomic/generic.h          | 234 ++++++++++++++++++++++++
>  src/urcu-pointer.c                      |   9 +-
>  8 files changed, 565 insertions(+), 35 deletions(-)
>  create mode 100644 include/urcu/uatomic/builtins-generic.h
>  create mode 100644 include/urcu/uatomic/builtins.h
> 
> diff --git a/doc/uatomic-api.md b/doc/uatomic-api.md
> index 0962399..7341ee8 100644
> --- a/doc/uatomic-api.md
> +++ b/doc/uatomic-api.md
> @@ -52,7 +52,8 @@ An atomic read-modify-write operation that performs this
>  sequence of operations atomically: check if `addr` contains `old`.
>  If true, then replace the content of `addr` by `new`. Return the
>  value previously contained by `addr`. This function implies a full
> -memory barrier before and after the atomic operation.
> +memory barrier before and after the atomic operation. The second memory
> +barrier is only emitted if the operation succeeded.
>  
>  
>  ```c
> diff --git a/include/Makefile.am b/include/Makefile.am
> index ba1fe60..b20e56d 100644
> --- a/include/Makefile.am
> +++ b/include/Makefile.am
> @@ -63,6 +63,8 @@ nobase_include_HEADERS = \
>       urcu/uatomic/alpha.h \
>       urcu/uatomic_arch.h \
>       urcu/uatomic/arm.h \
> +     urcu/uatomic/builtins.h \
> +     urcu/uatomic/builtins-generic.h \
>       urcu/uatomic/gcc.h \
>       urcu/uatomic/generic.h \
>       urcu/uatomic.h \
> diff --git a/include/urcu/static/pointer.h b/include/urcu/static/pointer.h
> index 9e46a57..9da8657 100644
> --- a/include/urcu/static/pointer.h
> +++ b/include/urcu/static/pointer.h
> @@ -96,23 +96,8 @@ extern "C" {
>   * -Wincompatible-pointer-types errors.  Using the statement expression
>   * makes it an rvalue and gets rid of the const-ness.
>   */
> -#ifdef __URCU_DEREFERENCE_USE_ATOMIC_CONSUME
> -# define _rcu_dereference(p) __extension__ ({                                
>                 \
> -                             __typeof__(__extension__ ({                     
>         \
> -                                     __typeof__(p) __attribute__((unused)) 
> _________p0 = { 0 }; \
> -                                     _________p0;                            
>         \
> -                             })) _________p1;                                
>         \
> -                             __atomic_load(&(p), &_________p1, 
> __ATOMIC_CONSUME);    \

There is talk of getting rid of memory_order_consume.  But for the moment,
it is what there is.  Another alternative is to use a volatile load,
similar to old-style CMM_LOAD_SHARED() or in-kernel READ_ONCE().

                                                Thanx, Paul

> -                             (_________p1);                                  
>         \
> -                     })
> -#else
> -# define _rcu_dereference(p) __extension__ ({                                
>                 \
> -                             __typeof__(p) _________p1 = CMM_LOAD_SHARED(p); 
>         \
> -                             cmm_smp_read_barrier_depends();                 
>         \
> -                             (_________p1);                                  
>         \
> -                     })
> -#endif
> -
> +# define _rcu_dereference(p)                 \
> +     uatomic_load(&(p), CMM_CONSUME)
>  /**
>   * _rcu_cmpxchg_pointer - same as rcu_assign_pointer, but tests if the 
> pointer
>   * is as expected by "old". If succeeds, returns the previous pointer to the
> @@ -131,8 +116,9 @@ extern "C" {
>       ({                                                              \
>               __typeof__(*p) _________pold = (old);                   \
>               __typeof__(*p) _________pnew = (_new);                  \
> -             uatomic_cmpxchg(p, _________pold, _________pnew);       \
> -     })
> +             uatomic_cmpxchg_mo(p, _________pold, _________pnew,     \
> +                                CMM_SEQ_CST, CMM_SEQ_CST);           \
> +     });
>  
>  /**
>   * _rcu_xchg_pointer - same as rcu_assign_pointer, but returns the previous
> @@ -149,17 +135,17 @@ extern "C" {
>       __extension__                                   \
>       ({                                              \
>               __typeof__(*p) _________pv = (v);       \
> -             uatomic_xchg(p, _________pv);           \
> +             uatomic_xchg_mo(p, _________pv,         \
> +                             CMM_SEQ_CST);           \
>       })
>  
>  
> -#define _rcu_set_pointer(p, v)                               \
> -     do {                                            \
> -             __typeof__(*p) _________pv = (v);       \
> -             if (!__builtin_constant_p(v) ||         \
> -                 ((v) != NULL))                      \
> -                     cmm_wmb();                              \
> -             uatomic_set(p, _________pv);            \
> +#define _rcu_set_pointer(p, v)                                               
> \
> +     do {                                                            \
> +             __typeof__(*p) _________pv = (v);                       \
> +             uatomic_store(p, _________pv,                           \
> +                     __builtin_constant_p(v) && (v) == NULL ?        \
> +                     CMM_RELAXED : CMM_RELEASE);                     \
>       } while (0)
>  
>  /**
> diff --git a/include/urcu/uatomic.h b/include/urcu/uatomic.h
> index 2fb5fd4..be857e1 100644
> --- a/include/urcu/uatomic.h
> +++ b/include/urcu/uatomic.h
> @@ -21,9 +21,70 @@
>  #ifndef _URCU_UATOMIC_H
>  #define _URCU_UATOMIC_H
>  
> +#include <assert.h>
> +
>  #include <urcu/arch.h>
> +#include <urcu/config.h>
>  
> -#if defined(URCU_ARCH_X86)
> +enum cmm_memorder {
> +     CMM_RELAXED = 0,
> +     CMM_CONSUME = 1,
> +     CMM_ACQUIRE = 2,
> +     CMM_RELEASE = 3,
> +     CMM_ACQ_REL = 4,
> +     CMM_SEQ_CST = 5,
> +     CMM_SEQ_CST_FENCE = 6,
> +};
> +
> +#ifdef CONFIG_RCU_USE_ATOMIC_BUILTINS
> +
> +/*
> + * Make sure that CMM_SEQ_CST_FENCE is not equivalent to other memory orders.
> + */
> +# ifdef static_assert
> +static_assert(CMM_RELAXED == __ATOMIC_RELAXED, "");
> +static_assert(CMM_CONSUME == __ATOMIC_CONSUME, "");
> +static_assert(CMM_ACQUIRE == __ATOMIC_ACQUIRE, "");
> +static_assert(CMM_RELEASE == __ATOMIC_RELEASE, "");
> +static_assert(CMM_ACQ_REL == __ATOMIC_ACQ_REL, "");
> +static_assert(CMM_SEQ_CST == __ATOMIC_SEQ_CST, "");
> +# endif
> +
> +/*
> + * This is not part of the public API. It it used internally to implement the
> + * CMM_SEQ_CST_FENCE memory order.
> + *
> + * NOTE: Using switch here instead of if statement to avoid -Wduplicated-cond
> + * warning when memory order is conditionally determined.
> + */
> +static inline void cmm_seq_cst_fence_after_atomic(enum cmm_memorder mo)
> +{
> +     switch (mo) {
> +     case CMM_SEQ_CST_FENCE:
> +             cmm_smp_mb();
> +             break;
> +     default:
> +             break;
> +     }
> +}
> +
> +#endif
> +
> +/*
> + * This is not part of the public API. It is used internally to convert from 
> the
> + * CMM memory model to the C11 memory model.
> + */
> +static inline int cmm_to_c11(int mo)
> +{
> +     if (mo == CMM_SEQ_CST_FENCE) {
> +             return CMM_SEQ_CST;
> +     }
> +     return mo;
> +}
> +
> +#if defined(CONFIG_RCU_USE_ATOMIC_BUILTINS)
> +#include <urcu/uatomic/builtins.h>
> +#elif defined(URCU_ARCH_X86)
>  #include <urcu/uatomic/x86.h>
>  #elif defined(URCU_ARCH_PPC)
>  #include <urcu/uatomic/ppc.h>
> diff --git a/include/urcu/uatomic/builtins-generic.h 
> b/include/urcu/uatomic/builtins-generic.h
> new file mode 100644
> index 0000000..673e888
> --- /dev/null
> +++ b/include/urcu/uatomic/builtins-generic.h
> @@ -0,0 +1,170 @@
> +/*
> + * urcu/uatomic/builtins-generic.h
> + *
> + * Copyright (c) 2023 Olivier Dion <od...@efficios.com>
> + *
> + * This library is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU Lesser General Public
> + * License as published by the Free Software Foundation; either
> + * version 2.1 of the License, or (at your option) any later version.
> + *
> + * This library is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> + * Lesser General Public License for more details.
> + *
> + * You should have received a copy of the GNU Lesser General Public
> + * License along with this library; if not, write to the Free Software
> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 
> USA
> + */
> +
> +#ifndef _URCU_UATOMIC_BUILTINS_GENERIC_H
> +#define _URCU_UATOMIC_BUILTINS_GENERIC_H
> +
> +#include <urcu/system.h>
> +
> +#define uatomic_store(addr, v, mo)                           \
> +     __extension__                                           \
> +     ({                                                      \
> +             __atomic_store_n(addr, v, cmm_to_c11(mo));      \
> +             cmm_seq_cst_fence_after_atomic(mo);             \
> +     })
> +
> +#define uatomic_set(addr, v)                 \
> +     uatomic_store(addr, v, CMM_RELAXED)
> +
> +#define uatomic_load(addr, mo)                                               
> \
> +     __extension__                                                   \
> +     ({                                                              \
> +             __typeof__(*(addr)) _value = __atomic_load_n(addr,      \
> +                                                     cmm_to_c11(mo)); \
> +             cmm_seq_cst_fence_after_atomic(mo);                     \
> +                                                                     \
> +             _value;                                                 \
> +     })
> +
> +#define uatomic_read(addr)                   \
> +     uatomic_load(addr, CMM_RELAXED)
> +
> +#define uatomic_cmpxchg_mo(addr, old, new, mos, mof)                 \
> +     __extension__                                                   \
> +     ({                                                              \
> +             __typeof__(*(addr)) _old = (__typeof__(*(addr)))old;    \
> +                                                                     \
> +             if (__atomic_compare_exchange_n(addr, &_old, new, 0,    \
> +                                                     cmm_to_c11(mos), \
> +                                                     cmm_to_c11(mof))) { \
> +                     cmm_seq_cst_fence_after_atomic(mos);            \
> +             } else {                                                \
> +                     cmm_seq_cst_fence_after_atomic(mof);            \
> +             }                                                       \
> +             _old;                                                   \
> +     })
> +
> +#define uatomic_cmpxchg(addr, old, new)                                      
> \
> +     uatomic_cmpxchg_mo(addr, old, new, CMM_SEQ_CST_FENCE, CMM_RELAXED)
> +
> +#define uatomic_xchg_mo(addr, v, mo)                                 \
> +     __extension__                                                   \
> +     ({                                                              \
> +             __typeof__((*addr)) _old = __atomic_exchange_n(addr, v, \
> +                                                     cmm_to_c11(mo)); \
> +             cmm_seq_cst_fence_after_atomic(mo);                     \
> +             _old;                                                   \
> +     })
> +
> +#define uatomic_xchg(addr, v)                                                
> \
> +     uatomic_xchg_mo(addr, v, CMM_SEQ_CST_FENCE)
> +
> +#define uatomic_add_return_mo(addr, v, mo)                           \
> +     __extension__                                                   \
> +     ({                                                              \
> +             __typeof__(*(addr)) _old = __atomic_add_fetch(addr, v,  \
> +                                                     cmm_to_c11(mo)); \
> +             cmm_seq_cst_fence_after_atomic(mo);                     \
> +             _old;                                                   \
> +     })
> +
> +#define uatomic_add_return(addr, v)                                  \
> +     uatomic_add_return_mo(addr, v, CMM_SEQ_CST_FENCE)
> +
> +#define uatomic_sub_return_mo(addr, v, mo)                           \
> +     __extension__                                                   \
> +     ({                                                              \
> +             __typeof__(*(addr)) _old = __atomic_sub_fetch(addr, v,  \
> +                                                     cmm_to_c11(mo)); \
> +             cmm_seq_cst_fence_after_atomic(mo);                     \
> +             _old;                                                   \
> +     })
> +
> +#define uatomic_sub_return(addr, v)                                  \
> +     uatomic_sub_return_mo(addr, v, CMM_SEQ_CST_FENCE)
> +
> +#define uatomic_and_mo(addr, mask, mo)                                       
> \
> +     __extension__                                                   \
> +     ({                                                              \
> +             __typeof__(*(addr)) _old = __atomic_and_fetch(addr, mask, \
> +                                                     cmm_to_c11(mo)); \
> +             cmm_seq_cst_fence_after_atomic(mo);                     \
> +             _old;                                                   \
> +     })
> +
> +#define uatomic_and(addr, mask)                              \
> +     (void) uatomic_and_mo(addr, mask, CMM_SEQ_CST)
> +
> +#define uatomic_or_mo(addr, mask, mo)                                        
> \
> +     __extension__                                                   \
> +     ({                                                              \
> +             __typeof__(*(addr)) _old = __atomic_or_fetch(addr, mask, \
> +                                                     cmm_to_c11(mo)); \
> +             cmm_seq_cst_fence_after_atomic(mo);                     \
> +             _old;                                                   \
> +     })
> +
> +
> +#define uatomic_or(addr, mask)                               \
> +     (void) uatomic_or_mo(addr, mask, CMM_RELAXED)
> +
> +#define uatomic_add_mo(addr, v, mo)                  \
> +     (void) uatomic_add_return_mo(addr, v, mo)
> +
> +#define uatomic_add(addr, v)                         \
> +     (void) uatomic_add_mo(addr, v, CMM_RELAXED)
> +
> +#define uatomic_sub_mo(addr, v, mo)                  \
> +     (void) uatomic_sub_return_mo(addr, v, mo)
> +
> +#define uatomic_sub(addr, v)                         \
> +     (void) uatomic_sub_mo(addr, v, CMM_RELAXED)
> +
> +#define uatomic_inc_mo(addr, mo)             \
> +     (void) uatomic_add_mo(addr, 1, mo)
> +
> +#define uatomic_inc(addr)                            \
> +     (void) uatomic_inc_mo(addr, CMM_RELAXED)
> +
> +#define uatomic_dec_mo(addr, mo)             \
> +     (void) uatomic_sub_mo(addr, 1, mo)
> +
> +#define uatomic_dec(addr)                            \
> +     (void) uatomic_dec_mo(addr, CMM_RELAXED)
> +
> +#define cmm_smp_mb__before_uatomic_and() cmm_smp_mb()
> +#define cmm_smp_mb__after_uatomic_and()  cmm_smp_mb()
> +
> +#define cmm_smp_mb__before_uatomic_or() cmm_smp_mb()
> +#define cmm_smp_mb__after_uatomic_or()  cmm_smp_mb()
> +
> +#define cmm_smp_mb__before_uatomic_add() cmm_smp_mb()
> +#define cmm_smp_mb__after_uatomic_add()  cmm_smp_mb()
> +
> +#define cmm_smp_mb__before_uatomic_sub() cmm_smp_mb()
> +#define cmm_smp_mb__after_uatomic_sub()  cmm_smp_mb()
> +
> +#define cmm_smp_mb__before_uatomic_inc() cmm_smp_mb()
> +#define cmm_smp_mb__after_uatomic_inc() cmm_smp_mb()
> +
> +#define cmm_smp_mb__before_uatomic_dec() cmm_smp_mb()
> +#define cmm_smp_mb__after_uatomic_dec() cmm_smp_mb()
> +
> +#endif /* _URCU_UATOMIC_BUILTINS_X86_H */
> diff --git a/include/urcu/uatomic/builtins.h b/include/urcu/uatomic/builtins.h
> new file mode 100644
> index 0000000..82e98f8
> --- /dev/null
> +++ b/include/urcu/uatomic/builtins.h
> @@ -0,0 +1,79 @@
> +/*
> + * urcu/uatomic/builtins.h
> + *
> + * Copyright (c) 2023 Olivier Dion <od...@efficios.com>
> + *
> + * This library is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU Lesser General Public
> + * License as published by the Free Software Foundation; either
> + * version 2.1 of the License, or (at your option) any later version.
> + *
> + * This library is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> + * Lesser General Public License for more details.
> + *
> + * You should have received a copy of the GNU Lesser General Public
> + * License along with this library; if not, write to the Free Software
> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 
> USA
> + */
> +
> +#ifndef _URCU_UATOMIC_BUILTINS_H
> +#define _URCU_UATOMIC_BUILTINS_H
> +
> +#include <urcu/arch.h>
> +
> +#if defined(__has_builtin)
> +# if !__has_builtin(__atomic_store_n)
> +#  error "Toolchain does not support __atomic_store_n."
> +# endif
> +# if !__has_builtin(__atomic_load_n)
> +#  error "Toolchain does not support __atomic_load_n."
> +# endif
> +# if !__has_builtin(__atomic_exchange_n)
> +#  error "Toolchain does not support __atomic_exchange_n."
> +# endif
> +# if !__has_builtin(__atomic_compare_exchange_n)
> +#  error "Toolchain does not support __atomic_compare_exchange_n."
> +# endif
> +# if !__has_builtin(__atomic_add_fetch)
> +#  error "Toolchain does not support __atomic_add_fetch."
> +# endif
> +# if !__has_builtin(__atomic_sub_fetch)
> +#  error "Toolchain does not support __atomic_sub_fetch."
> +# endif
> +# if !__has_builtin(__atomic_or_fetch)
> +#  error "Toolchain does not support __atomic_or_fetch."
> +# endif
> +# if !__has_builtin(__atomic_thread_fence)
> +#  error "Toolchain does not support __atomic_thread_fence."
> +# endif
> +# if !__has_builtin(__atomic_signal_fence)
> +#  error "Toolchain does not support __atomic_signal_fence."
> +# endif
> +#elif defined(__GNUC__)
> +# define GCC_VERSION (__GNUC__       * 10000 + \
> +                    __GNUC_MINOR__ * 100   + \
> +                    __GNUC_PATCHLEVEL__)
> +# if  GCC_VERSION < 40700
> +#  error "GCC version is too old. Version must be 4.7 or greater"
> +# endif
> +# undef  GCC_VERSION
> +#else
> +# error "Toolchain is not supported."
> +#endif
> +
> +#if defined(__GNUC__)
> +# define UATOMIC_HAS_ATOMIC_BYTE  __GCC_ATOMIC_CHAR_LOCK_FREE
> +# define UATOMIC_HAS_ATOMIC_SHORT __GCC_ATOMIC_SHORT_LOCK_FREE
> +#elif defined(__clang__)
> +# define UATOMIC_HAS_ATOMIC_BYTE  __CLANG_ATOMIC_CHAR_LOCK_FREE
> +# define UATOMIC_HAS_ATOMIC_SHORT __CLANG_ATOMIC_SHORT_LOCK_FREE
> +#else
> +/* #  define UATOMIC_HAS_ATOMIC_BYTE  */
> +/* #  define UATOMIC_HAS_ATOMIC_SHORT */
> +#endif
> +
> +#include <urcu/uatomic/builtins-generic.h>
> +
> +#endif       /* _URCU_UATOMIC_BUILTINS_H */
> diff --git a/include/urcu/uatomic/generic.h b/include/urcu/uatomic/generic.h
> index e31a19b..6b9c153 100644
> --- a/include/urcu/uatomic/generic.h
> +++ b/include/urcu/uatomic/generic.h
> @@ -33,10 +33,244 @@ extern "C" {
>  #define uatomic_set(addr, v) ((void) CMM_STORE_SHARED(*(addr), (v)))
>  #endif
>  
> +extern void abort(void);
> +
> +#define uatomic_load_store_return_op(op, addr, v, mo)                        
> \
> +     __extension__                                                   \
> +     ({                                                              \
> +                                                                     \
> +             switch (mo) {                                           \
> +             case CMM_ACQUIRE:                                       \
> +             case CMM_CONSUME:                                       \
> +             case CMM_RELAXED:                                       \
> +                     break;                                          \
> +             case CMM_RELEASE:                                       \
> +             case CMM_ACQ_REL:                                       \
> +             case CMM_SEQ_CST:                                       \
> +             case CMM_SEQ_CST_FENCE:                                 \
> +                     cmm_smp_mb();                                   \
> +                     break;                                          \
> +             default:                                                \
> +                     abort();                                        \
> +             }                                                       \
> +                                                                     \
> +             __typeof__((*addr)) _value = op(addr, v);               \
> +                                                                     \
> +             switch (mo) {                                           \
> +             case CMM_CONSUME:                                       \
> +                     cmm_smp_read_barrier_depends();                 \
> +                     break;                                          \
> +             case CMM_ACQUIRE:                                       \
> +             case CMM_ACQ_REL:                                       \
> +             case CMM_SEQ_CST:                                       \
> +             case CMM_SEQ_CST_FENCE:                                 \
> +                     cmm_smp_mb();                                   \
> +                     break;                                          \
> +             case CMM_RELAXED:                                       \
> +             case CMM_RELEASE:                                       \
> +                     break;                                          \
> +             default:                                                \
> +                     abort();                                        \
> +             }                                                       \
> +             _value;                                                 \
> +     })
> +
> +#define uatomic_load_store_op(op, addr, v, mo)                               
> \
> +     ({                                                              \
> +             switch (mo) {                                           \
> +             case CMM_ACQUIRE:                                       \
> +             case CMM_CONSUME:                                       \
> +             case CMM_RELAXED:                                       \
> +                     break;                                          \
> +             case CMM_RELEASE:                                       \
> +             case CMM_ACQ_REL:                                       \
> +             case CMM_SEQ_CST:                                       \
> +             case CMM_SEQ_CST_FENCE:                                 \
> +                     cmm_smp_mb();                                   \
> +                     break;                                          \
> +             default:                                                \
> +                     abort();                                        \
> +             }                                                       \
> +                                                                     \
> +             op(addr, v);                                            \
> +                                                                     \
> +             switch (mo) {                                           \
> +             case CMM_CONSUME:                                       \
> +                     cmm_smp_read_barrier_depends();                 \
> +                     break;                                          \
> +             case CMM_ACQUIRE:                                       \
> +             case CMM_ACQ_REL:                                       \
> +             case CMM_SEQ_CST:                                       \
> +             case CMM_SEQ_CST_FENCE:                                 \
> +                     cmm_smp_mb();                                   \
> +                     break;                                          \
> +             case CMM_RELAXED:                                       \
> +             case CMM_RELEASE:                                       \
> +                     break;                                          \
> +             default:                                                \
> +                     abort();                                        \
> +             }                                                       \
> +     })
> +
> +#define uatomic_store(addr, v, mo)                   \
> +     ({                                              \
> +             switch (mo) {                           \
> +             case CMM_RELAXED:                       \
> +                     break;                          \
> +             case CMM_RELEASE:                       \
> +             case CMM_SEQ_CST:                       \
> +             case CMM_SEQ_CST_FENCE:                 \
> +                     cmm_smp_mb();                   \
> +                     break;                          \
> +             default:                                \
> +                     abort();                        \
> +             }                                       \
> +                                                     \
> +             uatomic_set(addr, v);                   \
> +                                                     \
> +             switch (mo) {                           \
> +             case CMM_RELAXED:                       \
> +             case CMM_RELEASE:                       \
> +                     break;                          \
> +             case CMM_SEQ_CST:                       \
> +             case CMM_SEQ_CST_FENCE:                 \
> +                     cmm_smp_mb();                   \
> +                     break;                          \
> +             default:                                \
> +                     abort();                        \
> +             }                                       \
> +     })
> +
> +#define uatomic_and_mo(addr, v, mo)                          \
> +     uatomic_load_store_op(uatomic_and, addr, v, mo)
> +
> +#define uatomic_or_mo(addr, v, mo)                           \
> +     uatomic_load_store_op(uatomic_or, addr, v, mo)
> +
> +#define uatomic_add_mo(addr, v, mo)                          \
> +     uatomic_load_store_op(uatomic_add, addr, v, mo)
> +
> +#define uatomic_sub_mo(addr, v, mo)                          \
> +     uatomic_load_store_op(uatomic_sub, addr, v, mo)
> +
> +#define uatomic_inc_mo(addr, mo)                             \
> +     uatomic_load_store_op(uatomic_add, addr, 1, mo)
> +
> +#define uatomic_dec_mo(addr, mo)                             \
> +     uatomic_load_store_op(uatomic_add, addr, -1, mo)
> +/*
> + * NOTE: We can not just do switch (_value == (old) ? mos : mof) otherwise 
> the
> + * compiler emit a -Wduplicated-cond warning.
> + */
> +#define uatomic_cmpxchg_mo(addr, old, new, mos, mof)                 \
> +     __extension__                                                   \
> +     ({                                                              \
> +             switch (mos) {                                          \
> +             case CMM_ACQUIRE:                                       \
> +             case CMM_CONSUME:                                       \
> +             case CMM_RELAXED:                                       \
> +                     break;                                          \
> +             case CMM_RELEASE:                                       \
> +             case CMM_ACQ_REL:                                       \
> +             case CMM_SEQ_CST:                                       \
> +             case CMM_SEQ_CST_FENCE:                                 \
> +                     cmm_smp_mb();                                   \
> +                     break;                                          \
> +             default:                                                \
> +                     abort();                                        \
> +             }                                                       \
> +                                                                     \
> +             __typeof__(*(addr)) _value = uatomic_cmpxchg(addr, old, \
> +                                                     new);           \
> +                                                                     \
> +             if (_value == (old)) {                                  \
> +                     switch (mos) {                                  \
> +                     case CMM_CONSUME:                               \
> +                             cmm_smp_read_barrier_depends();         \
> +                             break;                                  \
> +                     case CMM_ACQUIRE:                               \
> +                     case CMM_ACQ_REL:                               \
> +                     case CMM_SEQ_CST:                               \
> +                     case CMM_SEQ_CST_FENCE:                         \
> +                             cmm_smp_mb();                           \
> +                             break;                                  \
> +                     case CMM_RELAXED:                               \
> +                     case CMM_RELEASE:                               \
> +                             break;                                  \
> +                     default:                                        \
> +                             abort();                                \
> +                     }                                               \
> +             } else {                                                \
> +                     switch (mof) {                                  \
> +                     case CMM_CONSUME:                               \
> +                             cmm_smp_read_barrier_depends();         \
> +                             break;                                  \
> +                     case CMM_ACQUIRE:                               \
> +                     case CMM_ACQ_REL:                               \
> +                     case CMM_SEQ_CST:                               \
> +                     case CMM_SEQ_CST_FENCE:                         \
> +                             cmm_smp_mb();                           \
> +                             break;                                  \
> +                     case CMM_RELAXED:                               \
> +                     case CMM_RELEASE:                               \
> +                             break;                                  \
> +                     default:                                        \
> +                             abort();                                \
> +                     }                                               \
> +             }                                                       \
> +             _value;                                                 \
> +     })
> +
> +#define uatomic_xchg_mo(addr, v, mo)                         \
> +     uatomic_load_store_return_op(uatomic_xchg, addr, v, mo)
> +
> +#define uatomic_add_return_mo(addr, v, mo)                           \
> +     uatomic_load_store_return_op(uatomic_add_return, addr, v)
> +
> +#define uatomic_sub_return_mo(addr, v, mo)                           \
> +     uatomic_load_store_return_op(uatomic_sub_return, addr, v)
> +
> +
>  #ifndef uatomic_read
>  #define uatomic_read(addr)   CMM_LOAD_SHARED(*(addr))
>  #endif
>  
> +#define uatomic_load(addr, mo)                                               
> \
> +     __extension__                                                   \
> +     ({                                                              \
> +             switch (mo) {                                           \
> +             case CMM_ACQUIRE:                                       \
> +             case CMM_CONSUME:                                       \
> +             case CMM_RELAXED:                                       \
> +                     break;                                          \
> +             case CMM_SEQ_CST:                                       \
> +             case CMM_SEQ_CST_FENCE:                                 \
> +                     cmm_smp_mb();                                   \
> +                     break;                                          \
> +             default:                                                \
> +                     abort();                                        \
> +             }                                                       \
> +                                                                     \
> +             __typeof__(*(addr)) _rcu_value = uatomic_read(addr);    \
> +                                                                     \
> +             switch (mo) {                                           \
> +             case CMM_RELAXED:                                       \
> +                     break;                                          \
> +             case CMM_CONSUME:                                       \
> +                     cmm_smp_read_barrier_depends();                 \
> +                     break;                                          \
> +             case CMM_ACQUIRE:                                       \
> +             case CMM_SEQ_CST:                                       \
> +             case CMM_SEQ_CST_FENCE:                                 \
> +                     cmm_smp_mb();                                   \
> +                     break;                                          \
> +             default:                                                \
> +                     abort();                                        \
> +             }                                                       \
> +                                                                     \
> +             _rcu_value;                                             \
> +     })
> +
>  #if !defined __OPTIMIZE__  || defined UATOMIC_NO_LINK_ERROR
>  #ifdef ILLEGAL_INSTR
>  static inline __attribute__((always_inline))
> diff --git a/src/urcu-pointer.c b/src/urcu-pointer.c
> index d0854ac..cea8aeb 100644
> --- a/src/urcu-pointer.c
> +++ b/src/urcu-pointer.c
> @@ -39,19 +39,16 @@ void *rcu_dereference_sym(void *p)
>  
>  void *rcu_set_pointer_sym(void **p, void *v)
>  {
> -     cmm_wmb();
> -     uatomic_set(p, v);
> +     uatomic_store(p, v, CMM_RELEASE);
>       return v;
>  }
>  
>  void *rcu_xchg_pointer_sym(void **p, void *v)
>  {
> -     cmm_wmb();
> -     return uatomic_xchg(p, v);
> +     return uatomic_xchg_mo(p, v, CMM_SEQ_CST);
>  }
>  
>  void *rcu_cmpxchg_pointer_sym(void **p, void *old, void *_new)
>  {
> -     cmm_wmb();
> -     return uatomic_cmpxchg(p, old, _new);
> +     return uatomic_cmpxchg_mo(p, old, _new, CMM_SEQ_CST, CMM_SEQ_CST);
>  }
> -- 
> 2.40.1
> 
_______________________________________________
lttng-dev mailing list
lttng-dev@lists.lttng.org
https://lists.lttng.org/cgi-bin/mailman/listinfo/lttng-dev

Reply via email to