In order to get more accurate the cntvct_el0 reading, SW must invoke isb and arch_counter_enforce_ordering.
Reference of linux kernel: https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/ linux.git/tree/arch/arm64/include/asm/arch_timer.h?h=v5.5#n220 Fixes: ccad39ea0712 ("eal/arm: add cpu cycle operations for ARMv8") Cc: sta...@dpdk.org Reviewed-by: Jerin Jacob <jerinjac...@gmail.com> Reviewed-by: Gavin Hu <gavin...@arm.com> Signed-off-by: Haifeng Lin <haifeng....@huawei.com> --- .../common/include/arch/arm/rte_cycles_64.h | 26 +++++++++++++++++-- 1 file changed, 24 insertions(+), 2 deletions(-) diff --git a/lib/librte_eal/common/include/arch/arm/rte_cycles_64.h b/lib/librte_eal/common/include/arch/arm/rte_cycles_64.h index 68e7c7338..6b0df5b0a 100644 --- a/lib/librte_eal/common/include/arch/arm/rte_cycles_64.h +++ b/lib/librte_eal/common/include/arch/arm/rte_cycles_64.h @@ -59,11 +59,33 @@ rte_rdtsc(void) } #endif +static inline void +isb(void) +{ + asm volatile("isb" : : : "memory"); +} + +static inline void +__rte_arm64_cntvct_el0_enforce_ordering(uint64_t val) +{ + uint64_t tmp; + + asm volatile( + " eor %0, %1, %1\n" + " add %0, sp, %0\n" + " ldr xzr, [%0]" + : "=r" (tmp) : "r" (val)); +} + static inline uint64_t rte_rdtsc_precise(void) { - rte_mb(); - return rte_rdtsc(); + uint64_t tsc; + + isb(); + tsc = rte_rdtsc(); + __rte_arm64_cntvct_el0_enforce_ordering(tsc); + return tsc; } static inline uint64_t -- 2.24.1.windows.2