Use rte_atomic_thread_fence instead of directly using
__atomic_thread_fence builtin gcc intrinsic

Update rte_mcslock.h to use rte_atomic_thread_fence instead of
directly using internal __rte_atomic_thread_fence

Signed-off-by: Tyler Retzlaff <roret...@linux.microsoft.com>
Acked-by: Morten Brørup <m...@smartsharesystems.com>
---
 lib/eal/common/eal_common_trace.c | 2 +-
 lib/eal/include/rte_mcslock.h     | 4 ++--
 2 files changed, 3 insertions(+), 3 deletions(-)

diff --git a/lib/eal/common/eal_common_trace.c 
b/lib/eal/common/eal_common_trace.c
index 6ad87fc..918f49b 100644
--- a/lib/eal/common/eal_common_trace.c
+++ b/lib/eal/common/eal_common_trace.c
@@ -526,7 +526,7 @@ rte_trace_mode rte_trace_mode_get(void)
 
        /* Add the trace point at tail */
        STAILQ_INSERT_TAIL(&tp_list, tp, next);
-       __atomic_thread_fence(rte_memory_order_release);
+       rte_atomic_thread_fence(rte_memory_order_release);
 
        /* All Good !!! */
        return 0;
diff --git a/lib/eal/include/rte_mcslock.h b/lib/eal/include/rte_mcslock.h
index 2ca967f..0aeb1a0 100644
--- a/lib/eal/include/rte_mcslock.h
+++ b/lib/eal/include/rte_mcslock.h
@@ -83,7 +83,7 @@
         * store to prev->next. Otherwise it will cause a deadlock. Need a
         * store-load barrier.
         */
-       __rte_atomic_thread_fence(rte_memory_order_acq_rel);
+       rte_atomic_thread_fence(rte_memory_order_acq_rel);
        /* If the lock has already been acquired, it first atomically
         * places the node at the end of the queue and then proceeds
         * to spin on me->locked until the previous lock holder resets
@@ -117,7 +117,7 @@
                 * while-loop first. This has the potential to cause a
                 * deadlock. Need a load barrier.
                 */
-               __rte_atomic_thread_fence(rte_memory_order_acquire);
+               rte_atomic_thread_fence(rte_memory_order_acquire);
                /* More nodes added to the queue by other CPUs.
                 * Wait until the next pointer is set.
                 */
-- 
1.8.3.1

Reply via email to