Replace the use of gcc builtin __atomic_xxx intrinsics with
corresponding rte_atomic_xxx optional stdatomic API
Signed-off-by: Tyler Retzlaff <roret...@linux.microsoft.com>
---
lib/stack/rte_stack.h | 2 +-
lib/stack/rte_stack_lf_c11.h | 24 ++++++++++++------------
lib/stack/rte_stack_lf_generic.h | 18 +++++++++---------
3 files changed, 22 insertions(+), 22 deletions(-)
diff --git a/lib/stack/rte_stack.h b/lib/stack/rte_stack.h
index 921d29a..a379300 100644
--- a/lib/stack/rte_stack.h
+++ b/lib/stack/rte_stack.h
@@ -44,7 +44,7 @@ struct rte_stack_lf_list {
/** List head */
struct rte_stack_lf_head head __rte_aligned(16);
/** List len */
- uint64_t len;
+ RTE_ATOMIC(uint64_t) len;
};
/* Structure containing two lock-free LIFO lists: the stack itself and a list
diff --git a/lib/stack/rte_stack_lf_c11.h b/lib/stack/rte_stack_lf_c11.h
index 687a6f6..9cb6998 100644
--- a/lib/stack/rte_stack_lf_c11.h
+++ b/lib/stack/rte_stack_lf_c11.h
@@ -26,8 +26,8 @@
* elements. If the mempool is near-empty to the point that this is a
* concern, the user should consider increasing the mempool size.
*/
- return (unsigned int)__atomic_load_n(&s->stack_lf.used.len,
- __ATOMIC_RELAXED);
+ return (unsigned int)rte_atomic_load_explicit(&s->stack_lf.used.len,
+ rte_memory_order_relaxed);
}
static __rte_always_inline void
@@ -59,14 +59,14 @@
(rte_int128_t *)&list->head,
(rte_int128_t *)&old_head,
(rte_int128_t *)&new_head,
- 1, __ATOMIC_RELEASE,
- __ATOMIC_RELAXED);
+ 1, rte_memory_order_release,
+ rte_memory_order_relaxed);
} while (success == 0);
/* Ensure the stack modifications are not reordered with respect
* to the LIFO len update.
*/
- __atomic_fetch_add(&list->len, num, __ATOMIC_RELEASE);
+ rte_atomic_fetch_add_explicit(&list->len, num,
rte_memory_order_release);
}
static __rte_always_inline struct rte_stack_lf_elem *
@@ -80,7 +80,7 @@
int success;
/* Reserve num elements, if available */
- len = __atomic_load_n(&list->len, __ATOMIC_RELAXED);
+ len = rte_atomic_load_explicit(&list->len, rte_memory_order_relaxed);
while (1) {
/* Does the list contain enough elements? */
@@ -88,10 +88,10 @@
return NULL;
/* len is updated on failure */
- if (__atomic_compare_exchange_n(&list->len,
+ if (rte_atomic_compare_exchange_weak_explicit(&list->len,
&len, len - num,
- 1, __ATOMIC_ACQUIRE,
- __ATOMIC_RELAXED))
+ rte_memory_order_acquire,
+ rte_memory_order_relaxed))
break;
}
@@ -110,7 +110,7 @@
* elements are properly ordered with respect to the head
* pointer read.
*/
- __atomic_thread_fence(__ATOMIC_ACQUIRE);
+ __atomic_thread_fence(rte_memory_order_acquire);
rte_prefetch0(old_head.top);
@@ -159,8 +159,8 @@
(rte_int128_t *)&list->head,
(rte_int128_t *)&old_head,
(rte_int128_t *)&new_head,
- 0, __ATOMIC_RELAXED,
- __ATOMIC_RELAXED);
+ 0, rte_memory_order_relaxed,
+ rte_memory_order_relaxed);
} while (success == 0);
return old_head.top;
diff --git a/lib/stack/rte_stack_lf_generic.h b/lib/stack/rte_stack_lf_generic.h
index 39f7ff3..cc69e4d 100644
--- a/lib/stack/rte_stack_lf_generic.h
+++ b/lib/stack/rte_stack_lf_generic.h
@@ -27,7 +27,7 @@
* concern, the user should consider increasing the mempool size.
*/
/* NOTE: review for potential ordering optimization */
- return __atomic_load_n(&s->stack_lf.used.len, __ATOMIC_SEQ_CST);
+ return rte_atomic_load_explicit(&s->stack_lf.used.len,
rte_memory_order_seq_cst);
}
static __rte_always_inline void
@@ -64,11 +64,11 @@
(rte_int128_t *)&list->head,
(rte_int128_t *)&old_head,
(rte_int128_t *)&new_head,
- 1, __ATOMIC_RELEASE,
- __ATOMIC_RELAXED);
+ 1, rte_memory_order_release,
+ rte_memory_order_relaxed);
} while (success == 0);
/* NOTE: review for potential ordering optimization */
- __atomic_fetch_add(&list->len, num, __ATOMIC_SEQ_CST);
+ rte_atomic_fetch_add_explicit(&list->len, num,
rte_memory_order_seq_cst);
}
static __rte_always_inline struct rte_stack_lf_elem *
@@ -83,15 +83,15 @@
/* Reserve num elements, if available */
while (1) {
/* NOTE: review for potential ordering optimization */
- uint64_t len = __atomic_load_n(&list->len, __ATOMIC_SEQ_CST);
+ uint64_t len = rte_atomic_load_explicit(&list->len,
rte_memory_order_seq_cst);
/* Does the list contain enough elements? */
if (unlikely(len < num))
return NULL;
/* NOTE: review for potential ordering optimization */
- if (__atomic_compare_exchange_n(&list->len, &len, len - num,
- 0, __ATOMIC_SEQ_CST, __ATOMIC_SEQ_CST))
+ if (rte_atomic_compare_exchange_strong_explicit(&list->len,
&len, len - num,
+ rte_memory_order_seq_cst,
rte_memory_order_seq_cst))
break;
}
@@ -143,8 +143,8 @@
(rte_int128_t *)&list->head,
(rte_int128_t *)&old_head,
(rte_int128_t *)&new_head,
- 1, __ATOMIC_RELEASE,
- __ATOMIC_RELAXED);
+ 1, rte_memory_order_release,
+ rte_memory_order_relaxed);
} while (success == 0);
return old_head.top;