On Tue, 24 Jun 2014 23:32:15 +0100 Bruce Richardson <bruce.richardson at intel.com> wrote:
> > +static void > +free_unsent_pkts(struct rte_mbuf **pkts, uint16_t unsent, > + void *userdata __rte_unused) > +{ > + unsigned i; > + for (i = 0; i < unsent; i++) > + rte_pktmbuf_free(pkts[i]); > +} > + This should be moved into mbuf layer, and there it could be optimized to do a rte_mempool_mp_put_bulk. This would speed up operations becuase it would mean a single ring operation per set rather per mbuf segment. Of course, the optimization would have to handle the refcnt issues.