On Fri, Oct 11, 2024 at 12:42:01PM +0000, Konstantin Ananyev wrote: > > > > Rather than re-querying the HW each time a CPU flag is requested, we can > > just save the return value in the flags array. This should speed up > > repeated querying of CPU flags, and provides a workaround for a reported > > issue where errors are seen with constant querying of the AVX-512 CPU > > flag from a non-AVX VM. > > > > Bugzilla Id: 1501 > > > > Signed-off-by: Bruce Richardson <bruce.richard...@intel.com> > > --- > > lib/eal/x86/rte_cpuflags.c | 20 +++++++++++++++----- > > 1 file changed, 15 insertions(+), 5 deletions(-) > > > > diff --git a/lib/eal/x86/rte_cpuflags.c b/lib/eal/x86/rte_cpuflags.c > > index 26163ab746..62e782fb4b 100644 > > --- a/lib/eal/x86/rte_cpuflags.c > > +++ b/lib/eal/x86/rte_cpuflags.c > > @@ -8,6 +8,7 @@ > > #include <errno.h> > > #include <stdint.h> > > #include <string.h> > > +#include <stdbool.h> > > > > #include "rte_cpuid.h" > > > > @@ -21,12 +22,14 @@ struct feature_entry { > > uint32_t bit; /**< cpuid register bit */ > > #define CPU_FLAG_NAME_MAX_LEN 64 > > char name[CPU_FLAG_NAME_MAX_LEN]; /**< String for printing */ > > + bool has_value; > > + bool value; > > }; > > > > #define FEAT_DEF(name, leaf, subleaf, reg, bit) \ > > [RTE_CPUFLAG_##name] = {leaf, subleaf, reg, bit, #name }, > > > > -const struct feature_entry rte_cpu_feature_table[] = { > > +struct feature_entry rte_cpu_feature_table[] = { > > FEAT_DEF(SSE3, 0x00000001, 0, RTE_REG_ECX, 0) > > FEAT_DEF(PCLMULQDQ, 0x00000001, 0, RTE_REG_ECX, 1) > > FEAT_DEF(DTES64, 0x00000001, 0, RTE_REG_ECX, 2) > > @@ -147,7 +150,7 @@ const struct feature_entry rte_cpu_feature_table[] = { > > int > > rte_cpu_get_flag_enabled(enum rte_cpu_flag_t feature) > > { > > - const struct feature_entry *feat; > > + struct feature_entry *feat; > > cpuid_registers_t regs; > > unsigned int maxleaf; > > > > @@ -156,6 +159,8 @@ rte_cpu_get_flag_enabled(enum rte_cpu_flag_t feature) > > return -ENOENT; > > > > feat = &rte_cpu_feature_table[feature]; > > + if (feat->has_value) > > + return feat->value; > > > > if (!feat->leaf) > > /* This entry in the table wasn't filled out! */ > > @@ -163,8 +168,10 @@ rte_cpu_get_flag_enabled(enum rte_cpu_flag_t feature) > > > > maxleaf = __get_cpuid_max(feat->leaf & 0x80000000, NULL); > > > > - if (maxleaf < feat->leaf) > > - return 0; > > + if (maxleaf < feat->leaf) { > > + feat->value = 0; > > + goto out; > > + } > > > > #ifdef RTE_TOOLCHAIN_MSVC > > __cpuidex(regs, feat->leaf, feat->subleaf); > > @@ -175,7 +182,10 @@ rte_cpu_get_flag_enabled(enum rte_cpu_flag_t feature) > > #endif > > > > /* check if the feature is enabled */ > > - return (regs[feat->reg] >> feat->bit) & 1; > > + feat->value = (regs[feat->reg] >> feat->bit) & 1; > > +out: > > + feat->has_value = true; > > + return feat->value; > > If that function can be called by 2 (or more) threads simultaneously, > then In theory 'feat->has_value = true;' can be reordered with > ' feat->value = (regs[feat->reg] >> feat->bit) & 1;' (by cpu or complier) > and some thread(s) can get wrong feat->value. > The probability of such collision is really low, but still seems not > impossible. >
Well since this code is x86-specific the externally visible store ordering will match the instruction store ordering. Therefore, I think a compiler barrier is all that is necessary before feat->has_value assignment, correct? /Bruce