Use rte_memcpy to improve the pointer array copy. This optimization method
has already been applied to __mempool_generic_put() [1], this patch applies
it to __mempool_generic_get(). Slight performance gain can be observed in
testpmd txonly test.

[1] 863bfb47449 ("mempool: optimize copy in cache")

Signed-off-by: Xiao Wang <xiao.w.w...@intel.com>
---
 lib/librte_mempool/rte_mempool.h | 7 +------
 1 file changed, 1 insertion(+), 6 deletions(-)

diff --git a/lib/librte_mempool/rte_mempool.h b/lib/librte_mempool/rte_mempool.h
index 8053f7a04..975da8d22 100644
--- a/lib/librte_mempool/rte_mempool.h
+++ b/lib/librte_mempool/rte_mempool.h
@@ -1344,15 +1344,11 @@ __mempool_generic_get(struct rte_mempool *mp, void 
**obj_table,
                      unsigned int n, struct rte_mempool_cache *cache)
 {
        int ret;
-       uint32_t index, len;
-       void **cache_objs;
 
        /* No cache provided or cannot be satisfied from cache */
        if (unlikely(cache == NULL || n >= cache->size))
                goto ring_dequeue;
 
-       cache_objs = cache->objs;
-
        /* Can this be satisfied from the cache? */
        if (cache->len < n) {
                /* No. Backfill the cache first, and then fill from it */
@@ -1375,8 +1371,7 @@ __mempool_generic_get(struct rte_mempool *mp, void 
**obj_table,
        }
 
        /* Now fill in the response ... */
-       for (index = 0, len = cache->len - 1; index < n; ++index, len--, 
obj_table++)
-               *obj_table = cache_objs[len];
+       rte_memcpy(obj_table, &cache->objs[cache->len - n], sizeof(void *) * n);
 
        cache->len -= n;
 
-- 
2.15.1

Reply via email to