On 2/3/2018 3:11 AM, Mallesh Koujalagi wrote: > After bulk allocation and freeing of multiple mbufs increase more than ~2% > throughput on single core. > > Signed-off-by: Mallesh Koujalagi <malleshx.koujal...@intel.com> > --- > drivers/net/null/rte_eth_null.c | 16 +++++++--------- > 1 file changed, 7 insertions(+), 9 deletions(-) > > diff --git a/drivers/net/null/rte_eth_null.c b/drivers/net/null/rte_eth_null.c > index 9385ffd..247ede0 100644 > --- a/drivers/net/null/rte_eth_null.c > +++ b/drivers/net/null/rte_eth_null.c > @@ -130,10 +130,11 @@ eth_null_copy_rx(void *q, struct rte_mbuf **bufs, > uint16_t nb_bufs) > return 0; > > packet_size = h->internals->packet_size; > + > + if (rte_pktmbuf_alloc_bulk(h->mb_pool, bufs, nb_bufs) != 0) > + return 0; > + > for (i = 0; i < nb_bufs; i++) { > - bufs[i] = rte_pktmbuf_alloc(h->mb_pool); > - if (!bufs[i]) > - break; > rte_memcpy(rte_pktmbuf_mtod(bufs[i], void *), h->dummy_packet, > packet_size); > bufs[i]->data_len = (uint16_t)packet_size; > @@ -149,18 +150,15 @@ eth_null_copy_rx(void *q, struct rte_mbuf **bufs, > uint16_t nb_bufs) > static uint16_t > eth_null_tx(void *q, struct rte_mbuf **bufs, uint16_t nb_bufs) > { > - int i; > struct null_queue *h = q; > > if ((q == NULL) || (bufs == NULL)) > return 0; > > - for (i = 0; i < nb_bufs; i++) > - rte_pktmbuf_free(bufs[i]); > + rte_mempool_put_bulk(bufs[0]->pool, (void **)bufs, nb_bufs);
Is it guarantied that all mbufs will be from same mempool? > + rte_atomic64_add(&h->tx_pkts, nb_bufs); > > - rte_atomic64_add(&(h->tx_pkts), i); > - > - return i; > + return nb_bufs; > } > > static uint16_t >