On Mon, May 24, 2021 at 6:36 PM <pbhagavat...@marvell.com> wrote: > > From: Pavan Nikhilesh <pbhagavat...@marvell.com> > > Add Rx event vector fastpath to convert HW defined metadata into > rte_mbuf and rte_event_vector. > > Signed-off-by: Pavan Nikhilesh <pbhagavat...@marvell.com> > --- > drivers/event/cnxk/cn10k_worker.h | 50 +++++++++++++++++++++++++++++++ > 1 file changed, 50 insertions(+) > > diff --git a/drivers/event/cnxk/cn10k_worker.h > b/drivers/event/cnxk/cn10k_worker.h > index b3f71202ad..8c2cd72873 100644 > --- a/drivers/event/cnxk/cn10k_worker.h > +++ b/drivers/event/cnxk/cn10k_worker.h > @@ -5,6 +5,8 @@ > #ifndef __CN10K_WORKER_H__ > #define __CN10K_WORKER_H__ > > +#include <rte_vect.h> > + > #include "cnxk_ethdev.h" > #include "cnxk_eventdev.h" > #include "cnxk_worker.h" > @@ -101,6 +103,44 @@ cn10k_wqe_to_mbuf(uint64_t wqe, const uint64_t mbuf, > uint8_t port_id, > mbuf_init.value, flags); > } > > +static __rte_always_inline void > +cn10k_process_vwqe(uintptr_t vwqe, uint16_t port_id, const uint32_t flags, > + void *lookup_mem) > +{ > + union mbuf_initializer mbuf_init = { > + .fields = {.data_off = RTE_PKTMBUF_HEADROOM, > + .refcnt = 1, > + .nb_segs = 1, > + .port = port_id}, > + };
If there is performance improvement move mbuf_init population to slow path and store it in HWS software structure or so. > + struct rte_event_vector *vec; > + uint16_t nb_mbufs, non_vec; > + uint64_t **wqe; > + > + vec = (struct rte_event_vector *)vwqe; > + wqe = vec->u64s; > + > + nb_mbufs = RTE_ALIGN_FLOOR(vec->nb_elem, NIX_DESCS_PER_LOOP); > + nb_mbufs = cn10k_nix_recv_pkts_vector(&mbuf_init.value, vec->mbufs, > + nb_mbufs, flags | NIX_RX_VWQE_F, > + lookup_mem); > + wqe += nb_mbufs; > + non_vec = vec->nb_elem - nb_mbufs; > + > + while (non_vec) { > + struct nix_cqe_hdr_s *cqe = (struct nix_cqe_hdr_s *)wqe[0]; > + struct rte_mbuf *mbuf; > + > + mbuf = (struct rte_mbuf *)((char *)cqe - > + sizeof(struct rte_mbuf)); > + cn10k_nix_cqe_to_mbuf(cqe, cqe->tag, mbuf, lookup_mem, > + mbuf_init.value, flags); > + wqe[0] = (uint64_t *)mbuf; > + non_vec--; > + wqe++; > + } > +} > + > static __rte_always_inline uint16_t > cn10k_sso_hws_get_work(struct cn10k_sso_hws *ws, struct rte_event *ev, > const uint32_t flags, void *lookup_mem) > @@ -141,6 +181,16 @@ cn10k_sso_hws_get_work(struct cn10k_sso_hws *ws, struct > rte_event *ev, > gw.u64[0] & 0xFFFFF, flags, > lookup_mem); > gw.u64[1] = mbuf; > + } else if (CNXK_EVENT_TYPE_FROM_TAG(gw.u64[0]) == > + RTE_EVENT_TYPE_ETHDEV_VECTOR) { > + uint8_t port = CNXK_SUB_EVENT_FROM_TAG(gw.u64[0]); > + __uint128_t vwqe_hdr = *(__uint128_t *)gw.u64[1]; > + > + vwqe_hdr = ((vwqe_hdr >> 64) & 0xFFF) | BIT_ULL(31) | > + ((vwqe_hdr & 0xFFFF) << 48) | > + ((uint64_t)port << 32); > + *(uint64_t *)gw.u64[1] = (uint64_t)vwqe_hdr; > + cn10k_process_vwqe(gw.u64[1], port, flags, > lookup_mem); > } > } > > -- > 2.17.1 >