> From: Konstantin Ananyev [mailto:konstantin.anan...@huawei.com]
> Sent: Thursday, 22 December 2022 16.57
> 
> > Zero-copy access to mempool caches is beneficial for PMD performance,
> and
> > must be provided by the mempool library to fix [Bug 1052] without a
> > performance regression.
> 
> LGTM in general, thank you for working on it.
> Few comments below.
> 
> >
> > [Bug 1052]: https://bugs.dpdk.org/show_bug.cgi?id=1052
> >
> > v2:
> > * Fix checkpatch warnings.
> > * Fix missing registration of trace points.
> > * The functions are inline, so they don't go into the map file.
> > v1 changes from the RFC:
> > * Removed run-time parameter checks. (Honnappa)
> >   This is a hot fast path function; requiring correct application
> >   behaviour, i.e. function parameters must be valid.
> > * Added RTE_ASSERT for parameters instead.
> 
> RTE_ASSERT(n <= RTE_MEMPOOL_CACHE_MAX_SIZE);
> I think it is too excessive.
> Just:
> if (n <= RTE_MEMPOOL_CACHE_MAX_SIZE) return NULL;
> seems much more convenient for the users here and
> more close to other mempool/ring API behavior.
> In terms of performance - I don’t think one extra comparison here
> would really count.

The insignificant performance degradation seems like a good tradeoff for making 
the function more generic.
I will update the function documentation and place the run-time check here, so 
both trace and stats reflect what happened:

        RTE_ASSERT(cache != NULL);
        RTE_ASSERT(mp != NULL);
-       RTE_ASSERT(n <= RTE_MEMPOOL_CACHE_MAX_SIZE);

        rte_mempool_trace_cache_zc_put_bulk(cache, mp, n);
+
+       if (unlikely(n > RTE_MEMPOOL_CACHE_MAX_SIZE)) {
+               rte_errno = -ENOSPC; // Or EINVAL?
+               return NULL;
+       }

        /* Increment stats now, adding in mempool always succeeds. */

I will probably also be able to come up with solution for zc_get_bulk(), so 
both trace and stats make sense if called with n > RTE_MEMPOOL_CACHE_MAX_SIZE.

> 
> I also think would be really good to add:
> add zc_(get|put)_bulk_start(),  zc_(get|put)_bulk_finish().
> Where _start would check/fill the cache and return the pointer,
> while _finsih would updathe cache->len.
> Similar to what we have for rte_ring _peek_ API.
> That would allow to extend this API usage - let say inside PMDs
> it could be used not only for MBUF_FAST_FREE case,  but for generic
> TX code path (one that have to call rte_mbuf_prefree()) also.

I don't see a use case for zc_get_start()/_finish().

And since the mempool cache is a stack, it would *require* that the application 
reads the array in reverse order. In such case, the function should not return 
a pointer to the array of objects, but a pointer to the top of the stack.

So I prefer to stick with the single-function zero-copy get, i.e. without 
start/finish.


I do agree with you about the use case for zc_put_start()/_finish().

Unlike the ring, there is no need for locking with the mempool cache, so we can 
implement something much simpler:

Instead of requiring calling both zc_put_start() and _finish() for every 
zero-copy burst, we could add a zc_put_rewind() function, only to be called if 
some number of objects were not added anyway:

/* FIXME: Function documentation here. */
__rte_experimental
static __rte_always_inline void
rte_mempool_cache_zc_put_rewind(struct rte_mempool_cache *cache,
                unsigned int n)
{
        RTE_ASSERT(cache != NULL);
        RTE_ASSERT(n <= cache->len);

        rte_mempool_trace_cache_zc_put_rewind(cache, n);

        /* Rewind stats. */
        RTE_MEMPOOL_CACHE_STAT_ADD(cache, put_objs, -n);

        cache->len -= n;
}

I have a strong preference for _rewind() over _start() and _finish(), because 
in the full burst case, it only touches the rte_mempool_cache structure once, 
whereas splitting it up into _start() and _finish() touches the 
rte_mempool_cache structure both before and after copying the array of objects.

What do you think?

I am open for other names than _rewind(), so feel free to speak up if you have 
a better name.


> 
> >   Code for this is only generated if built with RTE_ENABLE_ASSERT.
> > * Removed fallback when 'cache' parameter is not set. (Honnappa)
> > * Chose the simple get function; i.e. do not move the existing
> objects in
> >   the cache to the top of the new stack, just leave them at the
> bottom.
> > * Renamed the functions. Other suggestions are welcome, of course. ;-
> )
> > * Updated the function descriptions.
> > * Added the functions to trace_fp and version.map.
> 
> Would be great to add some test-cases in app/test to cover this new
> API.

Good point. I will look at it.

BTW: Akshitha already has zc_put_bulk working in the i40e PMD.

Reply via email to