<snip> > > int n = txq->tx_rs_thresh; > > int32_t i = 0, j = 0; > > const int32_t k = RTE_ALIGN_FLOOR(n, RTE_I40E_TX_MAX_FREE_BUF_SZ); > > const int32_t m = n % RTE_I40E_TX_MAX_FREE_BUF_SZ; struct rte_mbuf > > *free[RTE_I40E_TX_MAX_FREE_BUF_SZ]; > > > > For FAST_FREE_MODE: > > > > if (k) { > > for (j = 0; j != k - RTE_I40E_TX_MAX_FREE_BUF_SZ; > > j += RTE_I40E_TX_MAX_FREE_BUF_SZ) { > > for (i = 0; i <RTE_I40E_TX_MAX_FREE_BUF_SZ; ++i, ++txep) { > > free[i] = txep->mbuf; > > txep->mbuf = NULL; > > } > > rte_mempool_put_bulk(free[0]->pool, (void **)free, > > RTE_I40E_TX_MAX_FREE_BUF_SZ); > > } > > } > > > > if (m) { > > for (i = 0; i < m; ++i, ++txep) { > > free[i] = txep->mbuf; > > txep->mbuf = NULL; > > } > > } > > rte_mempool_put_bulk(free[0]->pool, (void **)free, m); }
> Seems no logical problem, but the code looks heavy due to for loops. > Did you run performance with this change when tx_rs_thresh > > RTE_I40E_TX_MAX_FREE_BUF_SZ? Sorry for my late rely. It takes me some time to do the test for this path and following is my test results: First, I come up with another way to solve this bug and compare it with "loop"(size of 'free' is 64). That is set the size of 'free' as a large constant. We know: tx_rs_thresh < ring_desc_size < I40E_MAX_RING_DESC(4096), so we can directly define as: struct rte_mbuf *free[RTE_I40E_TX_MAX_FREE_BUF_SZ]; [1]Test Config: MRR Test: two porst & bi-directional flows & one core RX API: i40e_recv_pkts_bulk_alloc TX API: i40e_xmit_pkts_simple ring_descs_size: 1024 Ring_I40E_TX_MAX_FREE_SZ: 64 [2]Scheme: tx_rs_thresh = I40E_DEFAULT_TX_RSBIT_THRESH tx_free_thresh = I40E_DEFAULT_TX_FREE_THRESH tx_rs_thresh <= tx_free_thresh < nb_tx_desc So we change the value of 'tx_rs_thresh' by adjust I40E_DEFAULT_TX_RSBIT_THRESH [3]Test Results (performance improve): In X86: tx_rs_thresh/ tx_free_thresh 32/32 256/256 512/512 1.mempool_put(base) 0 0 0 2.mempool_put_bulk:loop +4.7% +5.6% +7.0% 3.mempool_put_bulk:large size for free +3.8% +2.3% -2.0% (free[I40E_MAX_RING_DESC]) In Arm: N1SDP: tx_rs_thresh/ tx_free_thresh 32/32 256/256 512/512 1.mempool_put(base) 0 0 0 2.mempool_put_bulk:loop +7.9% +9.1% +2.9% 3.mempool_put_bulk:large size for free +7.1% +8.7% +3.4% (free[I40E_MAX_RING_DESC]) Thunderx2: tx_rs_thresh/ tx_free_thresh 32/32 256/256 512/512 1.mempool_put(base) 0 0 0 2.mempool_put_bulk:loop +7.6% +10.5% +7.6% 3.mempool_put_bulk:large size for free +1.7% +18.4% +10.2% (free[I40E_MAX_RING_DESC]) As a result, I feel maybe 'loop' is better and it seems not very heavy according to the test. What about your views and look forward to your reply. Thanks a lot.