On Sat, Jan 08, 2022 at 12:00:17PM +0100, Morten Brørup wrote:
> > From: Bruce Richardson [mailto:bruce.richard...@intel.com]
> > Sent: Friday, 7 January 2022 16.12
> > 
> > On Tue, Dec 28, 2021 at 03:28:45PM +0100, Morten Brørup wrote:
> > > Hi mempool maintainers and DPDK team.
> > >
> > > Does anyone know the reason or history why
> > CACHE_FLUSHTHRESH_MULTIPLIER was chosen to be 1.5? I think it is
> > counterintuitive.
> > >
> > > The mempool cache flush threshold was introduced in DPDK version 1.3;
> > it was not in DPDK version 1.2. The copyright notice for rte_mempool.c
> > says year 2012.
> > >
> > >
> > > Here is my analysis:
> > >
> > > With the multiplier of 1.5, a mempool cache is allowed to be filled
> > up to 50 % above than its target size before its excess entries are
> > flushed to the mempool (thereby reducing the cache length to the target
> > size).
> > >
> > > In the opposite direction, a mempool cache is allowed to be drained
> > completely, i.e. up to 100 % below its target size.
> > >
> > > My instinct tells me that it would be more natural to let a mempool
> > cache go the same amount above and below its target size, i.e. using a
> > flush multiplier of 2 instead of 1.5.
> > >
> > > Also, the cache should be allowed to fill up to and including the
> > flush threshold, so it is flushed when the threshold is exceeded,
> > instead of when it is reached.
> > >
> > > Here is a simplified example:
> > >
> > > Imagine a cache target size of 32, corresponding to a typical packet
> > burst. With a flush threshold of 2 (and len > threshold instead of len
> > >= threshold), the cache could hold 1 +/-1 packet bursts. With the
> > current multiplier it can only hold [0 .. 1.5[ packet bursts, not
> > really providing a lot of elasticity.
> > >
> > Hi Morten,
> > 
> > Interesting to see this being looked at again. The original idea of
> > adding
> > in some extra room above the requested value was to avoid the worst-
> > case
> > scenario of a pool oscillating between full and empty repeatedly due to
> > the
> > addition/removal of perhaps a single packet. As for why 1.5 was chosen
> > as
> > the value, I don't recall any particular reason for it myself. The main
> > objective was to have a separate flush and size values so that we could
> > go
> > a bit above full, and when flushing, not emptying the entire cache down
> > to
> > zero.
> 
> Thanks for providing the historical background for this feature, Bruce.
> 
> > 
> > In terms of the behavioural points you make above, I wonder if symmetry
> > is
> > actually necessary or desirable in this case. After all, the ideal case
> > is
> > probably to keep the mempool neither full nor empty, so that both
> > allocations or frees can be done without having to go to the underlying
> > shared data structure. To accomodate this, the mempool will only flush
> > when
> > the number of elements is greater than size * 1.5, and then it only
> > flushes
> > elements down to size, ensuring that allocations can still take place.
> > On allocation, new buffers are taken when we don't have enough in the
> > cache
> > to fullfil the request, and then the cache is filled up to size, not to
> > the
> > flush threshold.
> 
> I agree with the ideal case.
> 
> However, it looks like the addition of the flush threshold also changed the 
> "size" parameter to effectively become "desired length" instead. This 
> interpretation is also supported by the flush algorithm, which doesn't flush 
> below the "size", but to the "size". So based on interpretation, I was 
> wondering why it is not symmetrical around the "desired length", but allowed 
> to go 100 % below and only 50 % above.
> 
> > 
> > Now, for the scenario you describe - where the mempool cache size is
> > set to
> > be the same as the burst size, this scheme probably does break down, in
> > that we don't really have any burst elasticity. However, I would query
> > if
> > that is a configuration that is used, since to the user it should
> > appear
> > correctly to provide no elasticity. Looking at testpmd, and our example
> > apps, the standard there is a burst size of 32 and mempool cache of
> > ~256.
> > In OVS code, netdev-dpdk.c seems to initialize the mempool  with cache
> > size
> > of RTE_MEMPOOL_CACHE_MAX_SIZE (through define MP_CACHE_SZ). In all
> > these
> > cases, I think the 1.5 threshold should work just fine for us.
> 
> My example was for demonstration only, and hopefully not being used by any 
> applications.
> 
> The simplified example was intended to demonstrate the theoretical effect of 
> the unbalance in using the 1.5 threshold. It will be similar with a cache 
> size of 256 objects: You will be allowed to go 8 bursts (calculation: 256 
> objects / 32 objects per burst) below the cache "size" without retrieving 
> objects from the backing pool, but only 4 bursts above the cache "size" 
> before flushing objects to the backing pool.
> 
> > That said,
> > if you want to bump it up to 2x, I can't say I'd object strongly as it
> > should be harmless, I think.
> 
> From an academic viewpoint, 2x seems much better to me than 1.5x. And based 
> on your background information from the introduction of the flush threshold, 
> there was no academic reason why 1.5x was chosen - this is certainly valuable 
> feedback.
> 
> I will consider providing a patch. Worst case it probably doesn't do any 
> harm. And if anyone is really interested, they can look at the mempool debug 
> stats to see the difference it makes for their specific application, e.g. the 
> put_common_pool_bulk/put_bulk and get_common_pool_bulk/get_success_bulk 
> ratios should decrease.

Sure, thanks.

/Bruce

Reply via email to