> -----Original Message----- > From: Morten Brørup <m...@smartsharesystems.com> > Sent: Tuesday, May 12, 2020 7:18 PM > To: Phil Yang <phil.y...@arm.com>; tho...@monjalon.net; dev@dpdk.org > Cc: bruce.richard...@intel.com; ferruh.yi...@intel.com; > hemant.agra...@nxp.com; Honnappa Nagarahalli > <honnappa.nagaraha...@arm.com>; jer...@marvell.com; > ktray...@redhat.com; konstantin.anan...@intel.com; > maxime.coque...@redhat.com; olivier.m...@6wind.com; > step...@networkplumber.org; mattias.ronnb...@ericsson.com; > harry.van.haa...@intel.com; erik.g.carri...@intel.com; nd <n...@arm.com>; > David Christensen <d...@linux.vnet.ibm.com>; nd <n...@arm.com> > Subject: RE: [PATCH v4 4/4] eal/atomic: add wrapper for c11 atomics > > > From: Phil Yang [mailto:phil.y...@arm.com] > > Sent: Tuesday, May 12, 2020 10:03 AM > > > > Wraps up compiler c11 atomic built-ins with explicit memory ordering > > parameter. > > > > Signed-off-by: Phil Yang <phil.y...@arm.com> > > --- > > lib/librte_eal/include/generic/rte_atomic_c11.h | 139 > > ++++++++++++++++++++++++ > > lib/librte_eal/include/meson.build | 1 + > > 2 files changed, 140 insertions(+) > > create mode 100644 lib/librte_eal/include/generic/rte_atomic_c11.h > > > > diff --git a/lib/librte_eal/include/generic/rte_atomic_c11.h > > b/lib/librte_eal/include/generic/rte_atomic_c11.h > > new file mode 100644 > > index 0000000..20490f4 > > --- /dev/null > > +++ b/lib/librte_eal/include/generic/rte_atomic_c11.h > > @@ -0,0 +1,139 @@ > > +/* SPDX-License-Identifier: BSD-3-Clause > > + * Copyright(c) 2020 Arm Limited > > + */ > > + > > +#ifndef _RTE_ATOMIC_C11_H_ > > +#define _RTE_ATOMIC_C11_H_ > > + > > +#include <rte_common.h> > > + > > +/** > > + * @file > > + * c11 atomic operations > > + * > > + * This file wraps up compiler (GCC) c11 atomic built-ins. > > + * https://gcc.gnu.org/onlinedocs/gcc/_005f_005fatomic-Builtins.html > > + */ > > + > > +#define memory_order_relaxed __ATOMIC_RELAXED > > +#define memory_order_consume __ATOMIC_CONSUME > > +#define memory_order_acquire __ATOMIC_ACQUIRE > > +#define memory_order_release __ATOMIC_RELEASE > > +#define memory_order_acq_rel __ATOMIC_ACQ_REL > > +#define memory_order_seq_cst __ATOMIC_SEQ_CST > > Why redefine these instead of using the original names? > > If we need to redefine them, they should be upper case and RTE_ prefixed.
Agreed, we don't need to redefine them. I was trying to align with the stdatomic library. I will remove them in the next version. > > > + > > +/* Generic atomic load. > > + * It returns the contents of *PTR. > > + * > > + * The valid memory order variants are: > > + * memory_order_relaxed > > + * memory_order_consume > > + * memory_order_acquire > > + * memory_order_seq_cst > > + */ > > +#define rte_atomic_load(PTR, MO) \ > > + (__extension__ ({ \ > > + typeof(PTR) _ptr = (PTR); \ > > + typeof(*_ptr) _ret; \ > > + __atomic_load(_ptr, &_ret, (MO)); \ > > + _ret; \ > > + })) > > + > > +/* Generic atomic store. > > + * It stores the value of VAL into *PTR. > > + * > > + * The valid memory order variants are: > > + * memory_order_relaxed > > + * memory_order_release > > + * memory_order_seq_cst > > + */ > > +#define rte_atomic_store(PTR, VAL, MO) \ > > + (__extension__ ({ \ > > + typeof(PTR) _ptr = (PTR); \ > > + typeof(*_ptr) _val = (VAL); \ > > + __atomic_store(_ptr, &_val, (MO)); \ > > + })) > > + > > +/* Generic atomic exchange. > > + * It stores the value of VAL into *PTR. > > + * It returns the original value of *PTR. > > + * > > + * The valid memory order variants are: > > + * memory_order_relaxed > > + * memory_order_acquire > > + * memory_order_release > > + * memory_order_acq_rel > > + * memory_order_seq_cst > > + */ > > +#define rte_atomic_exchange(PTR, VAL, MO) \ > > + (__extension__ ({ \ > > + typeof(PTR) _ptr = (PTR); \ > > + typeof(*_ptr) _val = (VAL); \ > > + typeof(*_ptr) _ret; \ > > + __atomic_exchange(_ptr, &_val, &_ret, (MO)); \ > > + _ret; \ > > + })) > > + > > +/* Generic atomic compare and exchange. > > + * It compares the contents of *PTR with the contents of *EXP. > > + * If equal, the operation is a read-modify-write operation that > > + * writes DES into *PTR. > > + * If they are not equal, the operation is a read and the current > > + * contents of *PTR are written into *EXP. > > + * > > + * The weak compare_exchange may fail spuriously and the strong > > + * variation will never fails spuriously. > > "will never fails spuriously" -> "will never fail" / "never fails". Thanks, I will fix it in the next version. > > And I suggest that you elaborate what "fail" means here, > i.e. what exactly can happen when it fails. Yes. That would be better. I will update it in the new version. Fail spuriously means the compare exchange operation acts as *PTR != *EXP and return false even if they are equal. > > > + * > > + * If DES is written into *PTR then true is returned and memory is > > + * affected according to the memory order specified by SUC_MO. > > + * There are no restrictions on what memory order can be used here. > > + * > > + * Otherwise, false is returned and memory is affected according to > > + * FAIL_MO. This memory order cannot be memory_order_release nor > > + * memory_order_acq_rel. It also cannot be a stronger order than that > > + * specified by SUC_MO. > > + */ > > +#define rte_atomic_compare_exchange_weak(PTR, EXP, DES, SUC_MO, > > FAIL_MO) \ > > + (__extension__ ({ \ > > + typeof(PTR) _ptr = (PTR); \ > > + typeof(*_ptr) _des = (DES); \ > > + __atomic_compare_exchange(_ptr, (EXP), &_des, 1, \ > > + (SUC_MO), (FAIL_MO)); > \ > > + })) > > + > > +#define rte_atomic_compare_exchange_strong(PTR, EXP, DES, SUC_MO, > > FAIL_MO) \ > > + (__extension__ ({ \ > > + typeof(PTR) _ptr = (PTR); \ > > + typeof(*_ptr) _des = (DES); \ > > + __atomic_compare_exchange(_ptr, (EXP), &_des, 0, \ > > + (SUC_MO), (FAIL_MO)); > \ > > + })) > > + > > +#define rte_atomic_fetch_add(PTR, VAL, MO) \ > > + __atomic_fetch_add((PTR), (VAL), (MO)) > > +#define rte_atomic_fetch_sub(PTR, VAL, MO) \ > > + __atomic_fetch_sub((PTR), (VAL), (MO)) > > +#define rte_atomic_fetch_or(PTR, VAL, MO) \ > > + __atomic_fetch_or((PTR), (VAL), (MO)) > > +#define rte_atomic_fetch_xor(PTR, VAL, MO) \ > > + __atomic_fetch_xor((PTR), (VAL), (MO)) > > +#define rte_atomic_fetch_and(PTR, VAL, MO) \ > > + __atomic_fetch_and((PTR), (VAL), (MO)) > > + > > +#define rte_atomic_add_fetch(PTR, VAL, MO) \ > > + __atomic_add_fetch((PTR), (VAL), (MO)) > > +#define rte_atomic_sub_fetch(PTR, VAL, MO) \ > > + __atomic_sub_fetch((PTR), (VAL), (MO)) > > +#define rte_atomic_or_fetch(PTR, VAL, MO) \ > > + __atomic_or_fetch((PTR), (VAL), (MO)) > > +#define rte_atomic_xor_fetch(PTR, VAL, MO) \ > > + __atomic_xor_fetch((PTR), (VAL), (MO)) > > +#define rte_atomic_and_fetch(PTR, VAL, MO) \ > > + __atomic_and_fetch((PTR), (VAL), (MO)) > > + > > +/* Synchronization fence between threads based on > > + * the specified memory order. > > + */ > > +#define rte_atomic_thread_fence(MO) __atomic_thread_fence((MO)) > > + > > +#endif /* _RTE_ATOMIC_C11_H_ */ > > diff --git a/lib/librte_eal/include/meson.build > > b/lib/librte_eal/include/meson.build > > index bc73ec2..dac1aac 100644 > > --- a/lib/librte_eal/include/meson.build > > +++ b/lib/librte_eal/include/meson.build > > @@ -51,6 +51,7 @@ headers += files( > > # special case install the generic headers, since they go in a subdir > > generic_headers = files( > > 'generic/rte_atomic.h', > > + 'generic/rte_atomic_c11.h', > > 'generic/rte_byteorder.h', > > 'generic/rte_cpuflags.h', > > 'generic/rte_cycles.h', > > -- > > 2.7.4 > > > > Thumbs up for the good function documentation. :-) Thank you for your comments. Thanks, Phil > > > Med venlig hilsen / kind regards > - Morten Brørup > >