On Sun, Oct 09, 2022 at 04:44:08PM +0200, Morten Brørup wrote: > > From: Andrew Rybchenko [mailto:andrew.rybche...@oktetlabs.ru] > > Sent: Sunday, 9 October 2022 15.38 > > To: Olivier Matz > > Cc: dev@dpdk.org; Morten Brørup; Bruce Richardson > > Subject: [PATCH v6 4/4] mempool: flush cache completely on overflow > > > > The cache was still full after flushing. In the opposite direction, > > i.e. when getting objects from the cache, the cache is refilled to full > > level when it crosses the low watermark (which happens to be zero). > > Similarly, the cache should be flushed to empty level when it crosses > > the high watermark (which happens to be 1.5 x the size of the cache). > > The existing flushing behaviour was suboptimal for real applications, > > because crossing the low or high watermark typically happens when the > > application is in a state where the number of put/get events are out of > > balance, e.g. when absorbing a burst of packets into a QoS queue > > (getting more mbufs from the mempool), or when a burst of packets is > > trickling out from the QoS queue (putting the mbufs back into the > > mempool). > > Now, the mempool cache is completely flushed when crossing the flush > > threshold, so only the newly put (hot) objects remain in the mempool > > cache afterwards. > > > > This bug degraded performance caused by too frequent flushing. > > > > Consider this application scenario: > > > > Either, an lcore thread in the application is in a state of balance, > > where it uses the mempool cache within its flush/refill boundaries; in > > this situation, the flush method is less important, and this fix is > > irrelevant. > > > > Or, an lcore thread in the application is out of balance (either > > permanently or temporarily), and mostly gets or puts objects from/to > > the > > mempool. If it mostly puts objects, not flushing all of the objects > > will > > cause more frequent flushing. This is the scenario addressed by this > > fix. E.g.: > > > > Cache size=256, flushthresh=384 (1.5x size), initial len=256; > > application burst len=32. > > > > If there are "size" objects in the cache after flushing, the cache is > > flushed at every 4th burst. > > > > If the cache is flushed completely, the cache is only flushed at every > > 16th burst. > > > > As you can see, this bug caused the cache to be flushed 4x too > > frequently in this example. > > > > And when/if the application thread breaks its pattern of continuously > > putting objects, and suddenly starts to get objects instead, it will > > either get objects already in the cache, or the get() function will > > refill the cache. > > > > The concept of not flushing the cache completely was probably based on > > an assumption that it is more likely for an application's lcore thread > > to get() after flushing than to put() after flushing. > > I strongly disagree with this assumption! If an application thread is > > continuously putting so much that it overflows the cache, it is much > > more likely to keep putting than it is to start getting. If in doubt, > > consider how CPU branch predictors work: When the application has done > > something many times consecutively, the branch predictor will expect > > the > > application to do the same again, rather than suddenly do something > > else. > > > > Signed-off-by: Morten Brørup <m...@smartsharesystems.com> > > Signed-off-by: Andrew Rybchenko <andrew.rybche...@oktetlabs.ru> > > --- > > Reviewed-by: Morten Brørup <m...@smartsharesystems.com> >
Acked-by: Olivier Matz <olivier.m...@6wind.com>