'__sync' builtins are deprecated, enable '__atomic' builtins for generic atomic operations.
Signed-off-by: Phil Yang <phil.y...@arm.com> Reviewed-by: Gavin Hu <gavin...@arm.com> Tested-by: Phil Yang <phil.y...@arm.com> --- lib/librte_eal/common/include/generic/rte_atomic.h | 80 ++++++++++++++++++++++ 1 file changed, 80 insertions(+) diff --git a/lib/librte_eal/common/include/generic/rte_atomic.h b/lib/librte_eal/common/include/generic/rte_atomic.h index b99ba46..260cdf3 100644 --- a/lib/librte_eal/common/include/generic/rte_atomic.h +++ b/lib/librte_eal/common/include/generic/rte_atomic.h @@ -186,7 +186,12 @@ rte_atomic16_cmpset(volatile uint16_t *dst, uint16_t exp, uint16_t src); static inline int rte_atomic16_cmpset(volatile uint16_t *dst, uint16_t exp, uint16_t src) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_bool_compare_and_swap(dst, exp, src); +#else + return __atomic_compare_exchange(dst, &exp, &src, 0, __ATOMIC_ACQUIRE, + __ATOMIC_ACQUIRE) ? 1 : 0; +#endif } #endif @@ -283,7 +288,11 @@ rte_atomic16_set(rte_atomic16_t *v, int16_t new_value) static inline void rte_atomic16_add(rte_atomic16_t *v, int16_t inc) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) __sync_fetch_and_add(&v->cnt, inc); +#else + __atomic_fetch_add(&v->cnt, inc, __ATOMIC_ACQUIRE); +#endif } /** @@ -297,7 +306,11 @@ rte_atomic16_add(rte_atomic16_t *v, int16_t inc) static inline void rte_atomic16_sub(rte_atomic16_t *v, int16_t dec) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) __sync_fetch_and_sub(&v->cnt, dec); +#else + __atomic_fetch_sub(&v->cnt, dec, __ATOMIC_ACQUIRE); +#endif } /** @@ -350,7 +363,11 @@ rte_atomic16_dec(rte_atomic16_t *v) static inline int16_t rte_atomic16_add_return(rte_atomic16_t *v, int16_t inc) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_add_and_fetch(&v->cnt, inc); +#else + return __atomic_add_fetch(&v->cnt, inc, __ATOMIC_ACQUIRE); +#endif } /** @@ -370,7 +387,11 @@ rte_atomic16_add_return(rte_atomic16_t *v, int16_t inc) static inline int16_t rte_atomic16_sub_return(rte_atomic16_t *v, int16_t dec) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_sub_and_fetch(&v->cnt, dec); +#else + return __atomic_sub_fetch(&v->cnt, dec, __ATOMIC_ACQUIRE); +#endif } /** @@ -389,7 +410,11 @@ static inline int rte_atomic16_inc_and_test(rte_atomic16_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic16_inc_and_test(rte_atomic16_t *v) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_add_and_fetch(&v->cnt, 1) == 0; +#else + return __atomic_add_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE) == 0; +#endif } #endif @@ -409,7 +434,11 @@ static inline int rte_atomic16_dec_and_test(rte_atomic16_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic16_dec_and_test(rte_atomic16_t *v) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_sub_and_fetch(&v->cnt, 1) == 0; +#else + return __atomic_sub_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE) == 0; +#endif } #endif @@ -469,7 +498,13 @@ rte_atomic32_cmpset(volatile uint32_t *dst, uint32_t exp, uint32_t src); static inline int rte_atomic32_cmpset(volatile uint32_t *dst, uint32_t exp, uint32_t src) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_bool_compare_and_swap(dst, exp, src); +#else + return __atomic_compare_exchange(dst, &exp, &src, 0, __ATOMIC_ACQUIRE, + __ATOMIC_ACQUIRE) ? 1 : 0; +#endif + } #endif @@ -566,7 +601,11 @@ rte_atomic32_set(rte_atomic32_t *v, int32_t new_value) static inline void rte_atomic32_add(rte_atomic32_t *v, int32_t inc) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) __sync_fetch_and_add(&v->cnt, inc); +#else + __atomic_fetch_add(&v->cnt, inc, __ATOMIC_ACQUIRE); +#endif } /** @@ -580,7 +619,11 @@ rte_atomic32_add(rte_atomic32_t *v, int32_t inc) static inline void rte_atomic32_sub(rte_atomic32_t *v, int32_t dec) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) __sync_fetch_and_sub(&v->cnt, dec); +#else + __atomic_fetch_sub(&v->cnt, dec, __ATOMIC_ACQUIRE); +#endif } /** @@ -633,7 +676,11 @@ rte_atomic32_dec(rte_atomic32_t *v) static inline int32_t rte_atomic32_add_return(rte_atomic32_t *v, int32_t inc) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_add_and_fetch(&v->cnt, inc); +#else + return __atomic_add_fetch(&v->cnt, inc, __ATOMIC_ACQUIRE); +#endif } /** @@ -653,7 +700,11 @@ rte_atomic32_add_return(rte_atomic32_t *v, int32_t inc) static inline int32_t rte_atomic32_sub_return(rte_atomic32_t *v, int32_t dec) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_sub_and_fetch(&v->cnt, dec); +#else + return __atomic_sub_fetch(&v->cnt, dec, __ATOMIC_ACQUIRE); +#endif } /** @@ -672,7 +723,11 @@ static inline int rte_atomic32_inc_and_test(rte_atomic32_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic32_inc_and_test(rte_atomic32_t *v) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_add_and_fetch(&v->cnt, 1) == 0; +#else + return __atomic_add_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE) == 0; +#endif } #endif @@ -692,7 +747,11 @@ static inline int rte_atomic32_dec_and_test(rte_atomic32_t *v); #ifdef RTE_FORCE_INTRINSICS static inline int rte_atomic32_dec_and_test(rte_atomic32_t *v) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_sub_and_fetch(&v->cnt, 1) == 0; +#else + return __atomic_sub_fetch(&v->cnt, 1, __ATOMIC_ACQUIRE) == 0; +#endif } #endif @@ -751,7 +810,12 @@ rte_atomic64_cmpset(volatile uint64_t *dst, uint64_t exp, uint64_t src); static inline int rte_atomic64_cmpset(volatile uint64_t *dst, uint64_t exp, uint64_t src) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_bool_compare_and_swap(dst, exp, src); +#else + return __atomic_compare_exchange(dst, &exp, &src, 0, __ATOMIC_ACQUIRE, + __ATOMIC_ACQUIRE) ? 1 : 0; +#endif } #endif @@ -902,7 +966,11 @@ rte_atomic64_add(rte_atomic64_t *v, int64_t inc); static inline void rte_atomic64_add(rte_atomic64_t *v, int64_t inc) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) __sync_fetch_and_add(&v->cnt, inc); +#else + __atomic_fetch_add(&v->cnt, inc, __ATOMIC_ACQUIRE); +#endif } #endif @@ -921,7 +989,11 @@ rte_atomic64_sub(rte_atomic64_t *v, int64_t dec); static inline void rte_atomic64_sub(rte_atomic64_t *v, int64_t dec) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) __sync_fetch_and_sub(&v->cnt, dec); +#else + __atomic_fetch_sub(&v->cnt, dec, __ATOMIC_ACQUIRE); +#endif } #endif @@ -979,7 +1051,11 @@ rte_atomic64_add_return(rte_atomic64_t *v, int64_t inc); static inline int64_t rte_atomic64_add_return(rte_atomic64_t *v, int64_t inc) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_add_and_fetch(&v->cnt, inc); +#else + return __atomic_add_fetch(&v->cnt, inc, __ATOMIC_ACQUIRE); +#endif } #endif @@ -1003,7 +1079,11 @@ rte_atomic64_sub_return(rte_atomic64_t *v, int64_t dec); static inline int64_t rte_atomic64_sub_return(rte_atomic64_t *v, int64_t dec) { +#if defined(RTE_TOOLCHAIN_GCC) && (GCC_VERSION < 70100) return __sync_sub_and_fetch(&v->cnt, dec); +#else + return __atomic_sub_fetch(&v->cnt, dec, __ATOMIC_ACQUIRE); +#endif } #endif -- 2.7.4