> From: Bruce Richardson [mailto:bruce.richard...@intel.com]
> Sent: Thursday, 7 April 2022 11.14
> 
> On Thu, Apr 07, 2022 at 11:04:53AM +0200, Morten Brørup wrote:
> > > From: Morten Brørup [mailto:m...@smartsharesystems.com]
> > > Sent: Wednesday, 2 February 2022 11.34
> > >
> > > This patch fixes the rte_mempool_do_generic_put() caching
> algorithm,
> > > which was fundamentally wrong, causing multiple performance issues
> when
> > > flushing.
> > >
> >
> > [...]
> >
> > Olivier,
> >
> > Will you please consider this patch [1] and the other one [2].
> >
> > The primary bug here is this: When a mempool cache becomes full (i.e.
> exceeds the "flush threshold"), and is flushed to the backing ring, it
> is still full afterwards; but it should be empty afterwards. It is not
> flushed entirely, only the elements exceeding "size" are flushed.
> >
> 
> I don't believe it should be flushed entirely, there should always be
> some
> elements left so that even after flush we can still allocate an
> additional
> burst. We want to avoid the situation where a flush of all elements is
> immediately followed by a refill of new elements. However, we can flush
> to
> maybe size/2, and improve things. In short, this not emptying is by
> design
> rather than a bug, though we can look to tweak the behaviour.
> 

I initially agreed with you about flushing to size/2.

However, I did think further about it when I wrote the patch, and came to this 
conclusion: If an application thread repeatedly puts objects into the mempool, 
and does it so often that the cache overflows (i.e. reaches the flush 
threshold) and needs to be flushed, it is far more likely that the application 
thread will continue doing that, rather than start getting objects from the 
mempool. This speaks for flushing the cache entirely.

Both solutions are better than flushing to size, so if there is a preference 
for keeping some objects in the cache after flushing, I can update the patch 
accordingly.

> > E.g. pipelined applications having ingress threads and egress threads
> running on different lcores are affected by this bug.
> >
> If we are looking at improvements for pipelined applications, I think a
> bigger win would be to change the default mempool from ring-based to
> stack-based. For apps using a run-to-completion model, they should run
> out
> of cache and should therefore be largely unaffected by such a change.
> 
> > I don't think the real performance impact is very big, but these
> algorithm level bugs really annoy me.
> >
> > I'm still wondering how the patch introducing the mempool cache flush
> threshold could pass internal code review with so many bugs.
> >
> > [1]
> https://patchwork.dpdk.org/project/dpdk/patch/20220202103354.79832-1-
> m...@smartsharesystems.com/
> > [2]
> https://patchwork.dpdk.org/project/dpdk/patch/20220202081426.77975-1-
> m...@smartsharesystems.com/
> >
> > -Morten
> >
> > > Signed-off-by: Morten Brørup <m...@smartsharesystems.com>
> > > ---
> > >  lib/mempool/rte_mempool.h | 34 ++++++++++++++++++++++------------
> > >  1 file changed, 22 insertions(+), 12 deletions(-)
> > >
> > > diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
> > > index 1e7a3c1527..e7e09e48fc 100644
> > > --- a/lib/mempool/rte_mempool.h
> > > +++ b/lib/mempool/rte_mempool.h
> > > @@ -1344,31 +1344,41 @@ rte_mempool_do_generic_put(struct
> rte_mempool
> > > *mp, void * const *obj_table,
> > >   if (unlikely(cache == NULL || n > RTE_MEMPOOL_CACHE_MAX_SIZE))
> > >           goto ring_enqueue;
> > >
> > > - cache_objs = &cache->objs[cache->len];
> > > + /* If the request itself is too big for the cache */
> > > + if (unlikely(n > cache->flushthresh))
> > > +         goto ring_enqueue;
> > >
> > >   /*
> > >    * The cache follows the following algorithm
> > > -  *   1. Add the objects to the cache
> > > -  *   2. Anything greater than the cache min value (if it crosses
> > > the
> > > -  *   cache flush threshold) is flushed to the ring.
> >
> > In the code, "the cache min value" is actually "the cache size". This
> indicates an intention to do something more. Perhaps the patch
> introducing the "flush threshold" was committed while still incomplete,
> and just never got completed?
> >
> > > +  *   1. If the objects cannot be added to the cache without
> > > +  *   crossing the flush threshold, flush the cache to the ring.
> > > +  *   2. Add the objects to the cache.
> > >    */
> > >
> > > - /* Add elements back into the cache */
> > > - rte_memcpy(&cache_objs[0], obj_table, sizeof(void *) * n);
> > > + if (cache->len + n <= cache->flushthresh) {
> > > +         cache_objs = &cache->objs[cache->len];
> > >
> > > - cache->len += n;
> > > +         cache->len += n;
> > > + } else {
> > > +         cache_objs = &cache->objs[0];
> > >
> > > - if (cache->len >= cache->flushthresh) {
> > > -         rte_mempool_ops_enqueue_bulk(mp, &cache->objs[cache->size],
> > > -                         cache->len - cache->size);
> > > -         cache->len = cache->size;
> > > +#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> > > +         if (rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache-
> > > >len) < 0)
> > > +                 rte_panic("cannot put objects in mempool\n");
> > > +#else
> > > +         rte_mempool_ops_enqueue_bulk(mp, cache_objs, cache->len);
> > > +#endif
> > > +         cache->len = n;
> > >   }
> > >
> > > + /* Add the objects to the cache. */
> > > + rte_memcpy(cache_objs, obj_table, sizeof(void *) * n);
> > > +
> > >   return;
> > >
> > >  ring_enqueue:
> > >
> > > - /* push remaining objects in ring */
> > > + /* Put the objects into the ring */
> > >  #ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> > >   if (rte_mempool_ops_enqueue_bulk(mp, obj_table, n) < 0)
> > >           rte_panic("cannot put objects in mempool\n");
> > > --
> > > 2.17.1
> >

Reply via email to