Convert spinlock (and as a consequence seqlock) to the clang
capability annotations.

Signed-off-by: David Marchand <david.march...@redhat.com>
Acked-by: Mattias Rönnnblom <mattias.ronnb...@ericsson.com>
---
 drivers/bus/dpaa/base/qbman/qman.c     |  4 ++--
 drivers/net/fm10k/fm10k_ethdev.c       |  4 ++--
 lib/eal/include/generic/rte_spinlock.h | 14 +++++++-------
 lib/eal/include/rte_eal_memconfig.h    |  4 ++--
 lib/eal/include/rte_seqlock.h          |  4 ++--
 lib/graph/graph_private.h              |  4 ++--
 lib/vhost/vdpa.c                       |  2 +-
 7 files changed, 18 insertions(+), 18 deletions(-)

diff --git a/drivers/bus/dpaa/base/qbman/qman.c 
b/drivers/bus/dpaa/base/qbman/qman.c
index f92b25343a..11fabcaff5 100644
--- a/drivers/bus/dpaa/base/qbman/qman.c
+++ b/drivers/bus/dpaa/base/qbman/qman.c
@@ -45,7 +45,7 @@ static inline int fq_isset(struct qman_fq *fq, u32 mask)
 }
 
 static inline void fq_lock(struct qman_fq *fq)
-       __rte_exclusive_lock_function(&fq->fqlock)
+       __rte_acquire_capability(&fq->fqlock)
        __rte_no_thread_safety_analysis
 {
        if (fq_isset(fq, QMAN_FQ_FLAG_LOCKED))
@@ -53,7 +53,7 @@ static inline void fq_lock(struct qman_fq *fq)
 }
 
 static inline void fq_unlock(struct qman_fq *fq)
-        __rte_unlock_function(&fq->fqlock)
+       __rte_release_capability(&fq->fqlock)
        __rte_no_thread_safety_analysis
 {
        if (fq_isset(fq, QMAN_FQ_FLAG_LOCKED))
diff --git a/drivers/net/fm10k/fm10k_ethdev.c b/drivers/net/fm10k/fm10k_ethdev.c
index 7b490bea17..747042d621 100644
--- a/drivers/net/fm10k/fm10k_ethdev.c
+++ b/drivers/net/fm10k/fm10k_ethdev.c
@@ -116,7 +116,7 @@ fm10k_mbx_initlock(struct fm10k_hw *hw)
 
 static void
 fm10k_mbx_lock(struct fm10k_hw *hw)
-       __rte_exclusive_lock_function(FM10K_DEV_PRIVATE_TO_MBXLOCK(hw->back))
+       __rte_acquire_capability(FM10K_DEV_PRIVATE_TO_MBXLOCK(hw->back))
 {
        while (!rte_spinlock_trylock(FM10K_DEV_PRIVATE_TO_MBXLOCK(hw->back)))
                rte_delay_us(FM10K_MBXLOCK_DELAY_US);
@@ -124,7 +124,7 @@ fm10k_mbx_lock(struct fm10k_hw *hw)
 
 static void
 fm10k_mbx_unlock(struct fm10k_hw *hw)
-       __rte_unlock_function(FM10K_DEV_PRIVATE_TO_MBXLOCK(hw->back))
+       __rte_release_capability(FM10K_DEV_PRIVATE_TO_MBXLOCK(hw->back))
 {
        rte_spinlock_unlock(FM10K_DEV_PRIVATE_TO_MBXLOCK(hw->back));
 }
diff --git a/lib/eal/include/generic/rte_spinlock.h 
b/lib/eal/include/generic/rte_spinlock.h
index c2980601b2..c907d4e45c 100644
--- a/lib/eal/include/generic/rte_spinlock.h
+++ b/lib/eal/include/generic/rte_spinlock.h
@@ -32,7 +32,7 @@ extern "C" {
 /**
  * The rte_spinlock_t type.
  */
-typedef struct __rte_lockable {
+typedef struct __rte_capability("spinlock") {
        volatile RTE_ATOMIC(int) locked; /**< lock status 0 = unlocked, 1 = 
locked */
 } rte_spinlock_t;
 
@@ -61,7 +61,7 @@ rte_spinlock_init(rte_spinlock_t *sl)
  */
 static inline void
 rte_spinlock_lock(rte_spinlock_t *sl)
-       __rte_exclusive_lock_function(sl);
+       __rte_acquire_capability(sl);
 
 #ifdef RTE_FORCE_INTRINSICS
 static inline void
@@ -87,7 +87,7 @@ rte_spinlock_lock(rte_spinlock_t *sl)
  */
 static inline void
 rte_spinlock_unlock(rte_spinlock_t *sl)
-       __rte_unlock_function(sl);
+       __rte_release_capability(sl);
 
 #ifdef RTE_FORCE_INTRINSICS
 static inline void
@@ -109,7 +109,7 @@ rte_spinlock_unlock(rte_spinlock_t *sl)
 __rte_warn_unused_result
 static inline int
 rte_spinlock_trylock(rte_spinlock_t *sl)
-       __rte_exclusive_trylock_function(1, sl);
+       __rte_try_acquire_capability(true, sl);
 
 #ifdef RTE_FORCE_INTRINSICS
 static inline int
@@ -158,7 +158,7 @@ static inline int rte_tm_supported(void);
  */
 static inline void
 rte_spinlock_lock_tm(rte_spinlock_t *sl)
-       __rte_exclusive_lock_function(sl);
+       __rte_acquire_capability(sl);
 
 /**
  * Commit hardware memory transaction or release the spinlock if
@@ -169,7 +169,7 @@ rte_spinlock_lock_tm(rte_spinlock_t *sl)
  */
 static inline void
 rte_spinlock_unlock_tm(rte_spinlock_t *sl)
-       __rte_unlock_function(sl);
+       __rte_release_capability(sl);
 
 /**
  * Try to execute critical section in a hardware memory transaction,
@@ -190,7 +190,7 @@ rte_spinlock_unlock_tm(rte_spinlock_t *sl)
 __rte_warn_unused_result
 static inline int
 rte_spinlock_trylock_tm(rte_spinlock_t *sl)
-       __rte_exclusive_trylock_function(1, sl);
+       __rte_try_acquire_capability(true, sl);
 
 /**
  * The rte_spinlock_recursive_t type.
diff --git a/lib/eal/include/rte_eal_memconfig.h 
b/lib/eal/include/rte_eal_memconfig.h
index 0b1d0d4ff0..55d78de334 100644
--- a/lib/eal/include/rte_eal_memconfig.h
+++ b/lib/eal/include/rte_eal_memconfig.h
@@ -132,14 +132,14 @@ rte_mcfg_mempool_write_unlock(void)
  */
 void
 rte_mcfg_timer_lock(void)
-       __rte_exclusive_lock_function(rte_mcfg_timer_get_lock());
+       __rte_acquire_capability(rte_mcfg_timer_get_lock());
 
 /**
  * Unlock the internal EAL Timer Library lock for exclusive access.
  */
 void
 rte_mcfg_timer_unlock(void)
-       __rte_unlock_function(rte_mcfg_timer_get_lock());
+       __rte_release_capability(rte_mcfg_timer_get_lock());
 
 /**
  * If true, pages are put in single files (per memseg list),
diff --git a/lib/eal/include/rte_seqlock.h b/lib/eal/include/rte_seqlock.h
index e0e94900d1..670a680443 100644
--- a/lib/eal/include/rte_seqlock.h
+++ b/lib/eal/include/rte_seqlock.h
@@ -197,7 +197,7 @@ rte_seqlock_read_retry(const rte_seqlock_t *seqlock, 
uint32_t begin_sn)
  */
 static inline void
 rte_seqlock_write_lock(rte_seqlock_t *seqlock)
-       __rte_exclusive_lock_function(&seqlock->lock)
+       __rte_acquire_capability(&seqlock->lock)
 {
        /* To synchronize with other writers. */
        rte_spinlock_lock(&seqlock->lock);
@@ -219,7 +219,7 @@ rte_seqlock_write_lock(rte_seqlock_t *seqlock)
  */
 static inline void
 rte_seqlock_write_unlock(rte_seqlock_t *seqlock)
-       __rte_unlock_function(&seqlock->lock)
+       __rte_release_capability(&seqlock->lock)
 {
        rte_seqcount_write_end(&seqlock->count);
 
diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h
index da48d73587..813dd78b9d 100644
--- a/lib/graph/graph_private.h
+++ b/lib/graph/graph_private.h
@@ -221,7 +221,7 @@ graph_spinlock_get(void);
  * Take a lock on the graph internal spin lock.
  */
 void graph_spinlock_lock(void)
-       __rte_exclusive_lock_function(graph_spinlock_get());
+       __rte_acquire_capability(graph_spinlock_get());
 
 /**
  * @internal
@@ -229,7 +229,7 @@ void graph_spinlock_lock(void)
  * Release a lock on the graph internal spin lock.
  */
 void graph_spinlock_unlock(void)
-       __rte_unlock_function(graph_spinlock_get());
+       __rte_release_capability(graph_spinlock_get());
 
 /* Graph operations */
 /**
diff --git a/lib/vhost/vdpa.c b/lib/vhost/vdpa.c
index 8abb073675..329dd572b7 100644
--- a/lib/vhost/vdpa.c
+++ b/lib/vhost/vdpa.c
@@ -32,7 +32,7 @@ static struct vdpa_device_list * const vdpa_device_list
 
 static struct rte_vdpa_device *
 __vdpa_find_device_by_name(const char *name)
-       __rte_exclusive_locks_required(&vdpa_device_list_lock)
+       __rte_requires_capability(&vdpa_device_list_lock)
 {
        struct rte_vdpa_device *dev, *ret = NULL;
 
-- 
2.47.0

Reply via email to