PING for review. Stephen, the discussion took quite a few turns, but didn't seem to reach a better solution. If you don't object to this simple patch, could you please also ack/review it, so it can be applied.
> From: Mattias Rönnblom [mailto:hof...@lysator.liu.se] > Sent: Monday, 4 September 2023 13.57 > > On 2023-09-04 11:26, Morten Brørup wrote: > > The per-lcore random state is frequently updated by their individual > > lcores, so add a cache guard to prevent CPU cache thrashing. > > > > "to prevent false sharing in case the CPU employs a next-N-lines (or > similar) hardware prefetcher" > > In my world, cache trashing and cache line contention are two different > things. You are right, Mattias. I didn't think give the description much thought, and simply used "cache trashing" in a broad, general sense. I think most readers will get the point anyway. Or they could take a look at the description provided for the RTE_CACHE_GUARD itself. :-) > > Other than that, > Acked-by: Mattias Rönnblom <mattias.ronnb...@ericsson.com> > > > Depends-on: series-29415 ("clarify purpose of empty cache lines") > > > > Signed-off-by: Morten Brørup <m...@smartsharesystems.com> > > --- > > lib/eal/common/rte_random.c | 1 + > > 1 file changed, 1 insertion(+) > > > > diff --git a/lib/eal/common/rte_random.c b/lib/eal/common/rte_random.c > > index 565f2401ce..3df0c7004a 100644 > > --- a/lib/eal/common/rte_random.c > > +++ b/lib/eal/common/rte_random.c > > @@ -18,6 +18,7 @@ struct rte_rand_state { > > uint64_t z3; > > uint64_t z4; > > uint64_t z5; > > + RTE_CACHE_GUARD; > > } __rte_cache_aligned; > > > > /* One instance each for every lcore id-equipped thread, and one