> From: Morten Brørup [mailto:m...@smartsharesystems.com] > Sent: Wednesday, 2 February 2022 11.34 > > This patch fixes the rte_mempool_do_generic_put() caching algorithm, > which was fundamentally wrong, causing multiple performance issues when > flushing. >
[...] Olivier, Will you please consider this patch [1] and the other one [2]. The primary bug here is this: When a mempool cache becomes full (i.e. exceeds the "flush threshold"), and is flushed to the backing ring, it is still full afterwards; but it should be empty afterwards. It is not flushed entirely, only the elements exceeding "size" are flushed. E.g. pipelined applications having ingress threads and egress threads running on different lcores are affected by this bug. I don't think the real performance impact is very big, but these algorithm level bugs really annoy me. I'm still wondering how the patch introducing the mempool cache flush threshold could pass internal code review with so many bugs. [1] https://patchwork.dpdk.org/project/dpdk/patch/20220202103354.79832-1...@smartsharesystems.com/ [2] https://patchwork.dpdk.org/project/dpdk/patch/20220202081426.77975-1...@smartsharesystems.com/ -Morten > Signed-off-by: Morten Brørup <m...@smartsharesystems.com> > --- > lib/mempool/rte_mempool.h | 34 ++++++++++++++++++++++------------ > 1 file changed, 22 insertions(+), 12 deletions(-) > > diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h > index 1e7a3c1527..e7e09e48fc 100644 > --- a/lib/mempool/rte_mempool.h > +++ b/lib/mempool/rte_mempool.h > @@ -1344,31 +1344,41 @@ rte_mempool_do_generic_put(struct rte_mempool > *mp, void * const *obj_table, > if (unlikely(cache == NULL || n > RTE_MEMPOOL_CACHE_MAX_SIZE)) > goto ring_enqueue; > > - cache_objs = &cache->objs[cache->len]; > + /* If the request itself is too big for the cache */ > + if (unlikely(n > cache->flushthresh)) > + goto ring_enqueue; > > /* > * The cache follows the following algorithm > - * 1. Add the objects to the cache > - * 2. Anything greater than the cache min value (if it crosses > the > - * cache flush threshold) is flushed to the ring. In the code, "the cache min value" is actually "the cache size". This indicates an intention to do something more. Perhaps the patch introducing the "flush threshold" was committed while still incomplete, and just never got completed? > + * 1. If the objects cannot be added to the cache without > + * crossing the flush threshold, flush the cache to the ring. > + * 2. Add the objects to the cache. > */ > > - /* Add elements back into the cache */ > - rte_memcpy(&cache_objs[0], obj_table, sizeof(void *) * n); > + if (cache->len + n <= cache->flushthresh) { > + cache_objs = &cache->objs[cache->len]; > > - cache->len += n; > + cache->len += n; > + } else { > + cache_objs = &cache->objs[0]; > > - if (cache->len >= cache->flushthresh) { > - rte_mempool_ops_enqueue_bulk(mp, &cache->objs[cache->size], > - cache->len - cache->size); > - cache->len = cache->size; > +#ifdef RTE_LIBRTE_MEMPOOL_DEBUG > + if (rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache- > >len) < 0) > + rte_panic("cannot put objects in mempool\n"); > +#else > + rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len); > +#endif > + cache->len = n; > } > > + /* Add the objects to the cache. */ > + rte_memcpy(cache_objs, obj_table, sizeof(void *) * n); > + > return; > > ring_enqueue: > > - /* push remaining objects in ring */ > + /* Put the objects into the ring */ > #ifdef RTE_LIBRTE_MEMPOOL_DEBUG > if (rte_mempool_ops_enqueue_bulk(mp, obj_table, n) < 0) > rte_panic("cannot put objects in mempool\n"); > -- > 2.17.1