Hi Pavan,

Thanks for the review and feedback. I will send next version with suggested 
changes.

Thanks,
Amit Shukla

<snip>

> > diff --git a/app/test-eventdev/test_perf_common.c b/app/test-
> > eventdev/test_perf_common.c index 5e0255cfeb..509d3f9232 100644
> > --- a/app/test-eventdev/test_perf_common.c
> > +++ b/app/test-eventdev/test_perf_common.c
> > @@ -559,6 +559,84 @@ crypto_adapter_enq_op_fwd(struct prod_data
> *p)
> >                    __func__, rte_lcore_id(), alloc_failures);  }
> >
> > +static inline void
> > +dma_adapter_enq_op_new(struct prod_data *p) {
> > +   struct test_perf *t = p->t;
> > +   const uint32_t nb_flows = t->nb_flows;
> > +   const uint64_t nb_pkts = t->nb_pkts;
> > +   struct rte_event_dma_adapter_op *op;
> > +   struct evt_options *opt = t->opt;
> > +   uint32_t flow_counter = 0;
> > +   uint64_t count = 0;
> > +
> > +   if (opt->verbose_level > 1)
> > +           printf("%s(): lcore %d queue %d dma_dev_id %u
> > dma_dev_vhcan_id %u\n",
> > +                  __func__, rte_lcore_id(), p->queue_id, p->da.dma_dev_id,
> > +                  p->da.vchan_id);
> > +
> > +   while (count < nb_pkts && t->done == false) {
> > +           op = p->da.dma_op[flow_counter++ % nb_flows];
> > +           while (rte_dma_copy_sg(op->dma_dev_id, op->vchan, op-
> > >src_seg,
> > +                                  op->dst_seg, op->nb_src, op->nb_dst,
> > +                                  op->flags) < 0 && t->done == false)
> 
> Is op type new really sending events to eventdev? If not, you can remove this
> test from the app for now and add it when dmadev supports enqueueing ops
> similar to cryptodev.
> 
> You can set --dma_adptr_mode   default to FORWARD and say NEW is not
> supported.
> 
> > +                   rte_pause();
> > +
> > +           count++;
> > +   }
> > +}
> > +
> > +static inline void
> > +dma_adapter_enq_op_fwd(struct prod_data *p) {
> > +   struct test_perf *t = p->t;
> > +   const uint32_t nb_flows = t->nb_flows;
> > +   const uint64_t nb_pkts = t->nb_pkts;
> > +   struct rte_event_dma_adapter_op *op;
> > +   const uint8_t dev_id = p->dev_id;
> > +   struct evt_options *opt = t->opt;
> > +   const uint8_t port = p->port_id;
> > +   uint32_t flow_counter = 0;
> > +   struct rte_event ev;
> > +   uint64_t count = 0;
> > +
> > +   if (opt->verbose_level > 1)
> > +           printf("%s(): lcore %d port %d queue %d dma_dev_id %u
> > dma_dev_vchan_id %u\n",
> > +                  __func__, rte_lcore_id(), port, p->queue_id,
> > +                  p->da.dma_dev_id, p->da.vchan_id);
> > +
> > +   ev.event = 0;
> > +   ev.op = RTE_EVENT_OP_NEW;
> > +   ev.queue_id = p->queue_id;
> > +   ev.sched_type = RTE_SCHED_TYPE_ATOMIC;
> > +   ev.event_type = RTE_EVENT_TYPE_CPU;
> > +
> > +   while (count < nb_pkts && t->done == false) {
> > +           op = p->da.dma_op[flow_counter++ % nb_flows];
> > +           ev.event_ptr = op;
> > +
> > +           while (rte_event_dma_adapter_enqueue(dev_id, port, &ev, 1)
> > != 1 &&
> > +                                                t->done == false)
> > +                   rte_pause();
> > +
> > +           count++;
> > +   }
> > +}
> > +
> > +static inline int
> > +perf_event_dma_producer(void *arg)
> > +{
> > +   struct prod_data *p = arg;
> > +   struct evt_options *opt = p->t->opt;
> > +
> > +   if (opt->dma_adptr_mode == RTE_EVENT_DMA_ADAPTER_OP_NEW)
> > +           dma_adapter_enq_op_new(p);
> > +   else
> > +           dma_adapter_enq_op_fwd(p);
> > +
> > +   return 0;
> > +}
> > +
> >  static inline int
> >  perf_event_crypto_producer(void *arg)  { @@ -841,7 +919,9 @@
> > perf_producer_wrapper(void *arg)
> >                     return perf_event_crypto_producer_burst(arg);
> >             else
> >                     return perf_event_crypto_producer(arg);
> > -   }
> > +   } else if (t->opt->prod_type == EVT_PROD_TYPE_EVENT_DMA_ADPTR)
> > +           return perf_event_dma_producer(arg);
> > +
> >     return 0;
> >  }
> >
> > @@ -952,7 +1032,9 @@ perf_launch_lcores(struct evt_test *test, struct
> > evt_options *opt,
> >                                 opt->prod_type ==
> >
> > EVT_PROD_TYPE_EVENT_TIMER_ADPTR ||
> >                                 opt->prod_type ==
> > -
> > EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR) {
> > +
> > EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR ||
> > +                               opt->prod_type ==
> > +
> > EVT_PROD_TYPE_EVENT_DMA_ADPTR) {
> >                                     t->done = true;
> >                                     break;
> >                             }
> > @@ -962,7 +1044,8 @@ perf_launch_lcores(struct evt_test *test, struct
> > evt_options *opt,
> >             if (new_cycles - dead_lock_cycles > dead_lock_sample &&
> >                 (opt->prod_type == EVT_PROD_TYPE_SYNT ||
> >                  opt->prod_type == EVT_PROD_TYPE_EVENT_TIMER_ADPTR
> > ||
> > -                opt->prod_type ==
> > EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR)) {
> > +                opt->prod_type ==
> > EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR ||
> > +                opt->prod_type ==
> > EVT_PROD_TYPE_EVENT_DMA_ADPTR)) {
> >                     remaining = t->outstand_pkts - processed_pkts(t);
> >                     if (dead_lock_remaining == remaining) {
> >                             rte_event_dev_dump(opt->dev_id, stdout);
> @@ -1162,6 +1245,39 @@
> > perf_event_crypto_adapter_setup(struct
> > test_perf *t, struct prod_data *p)
> >     return ret;
> >  }
> >
> > +static int
> > +perf_event_dma_adapter_setup(struct test_perf *t, struct prod_data
> > +*p) {
> > +   struct evt_options *opt = t->opt;
> > +   struct rte_event event;
> > +   uint32_t cap;
> > +   int ret;
> > +
> > +   ret = rte_event_dma_adapter_caps_get(p->dev_id, p-
> > >da.dma_dev_id, &cap);
> > +   if (ret) {
> > +           evt_err("Failed to get dma adapter capabilities");
> > +           return ret;
> > +   }
> > +
> > +   if (((opt->dma_adptr_mode ==
> > RTE_EVENT_DMA_ADAPTER_OP_NEW) &&
> > +        !(cap &
> > RTE_EVENT_DMA_ADAPTER_CAP_INTERNAL_PORT_OP_NEW)) ||
> > +       ((opt->dma_adptr_mode ==
> > RTE_EVENT_DMA_ADAPTER_OP_FORWARD) &&
> > +        !(cap &
> > RTE_EVENT_DMA_ADAPTER_CAP_INTERNAL_PORT_OP_FWD))) {
> > +           evt_err("dma adapter %s mode unsupported\n",
> > +                   opt->dma_adptr_mode ? "OP_FORWARD" :
> > "OP_NEW");
> > +           return -ENOTSUP;
> > +   }
> > +
> > +   if (cap &
> > RTE_EVENT_DMA_ADAPTER_CAP_INTERNAL_PORT_VCHAN_EV_BIND)
> > +           ret = rte_event_dma_adapter_vchan_add(TEST_PERF_DA_ID,
> > p->da.dma_dev_id,
> > +                                                 p->da.vchan_id, &event);
> > +   else
> > +           ret = rte_event_dma_adapter_vchan_add(TEST_PERF_DA_ID,
> > p->da.dma_dev_id,
> > +                                                 p->da.vchan_id, NULL);
> > +
> > +   return ret;
> > +}
> > +
> >  static void *
> >  cryptodev_sym_sess_create(struct prod_data *p, struct test_perf *t)
> > { @@ -1399,6 +1515,77 @@ perf_event_dev_port_setup(struct evt_test
> > *test, struct evt_options *opt,
> >                     }
> >
> >                     qp_id++;
> > +                   prod++;
> > +           }
> > +   }  else if (opt->prod_type == EVT_PROD_TYPE_EVENT_DMA_ADPTR) {
> > +           struct rte_event_port_conf conf = *port_conf;
> > +           struct rte_event_dma_adapter_op *op;
> > +           struct rte_mempool *pool = t->pool;
> > +           uint8_t dma_dev_id = 0;
> > +           uint16_t vchan_id = 0;
> > +
> > +           ret = rte_event_dma_adapter_create(TEST_PERF_DA_ID, opt-
> > >dev_id, &conf, 0);
> > +           if (ret) {
> > +                   evt_err("Failed to create dma adapter");
> > +                   return ret;
> > +           }
> > +
> > +           prod = 0;
> > +           for (; port < perf_nb_event_ports(opt); port++) {
> > +                   struct prod_data *p = &t->prod[port];
> > +                   struct rte_event *response_info;
> > +                   uint32_t flow_id;
> > +
> > +                   p->dev_id = opt->dev_id;
> > +                   p->port_id = port;
> > +                   p->queue_id = prod * stride;
> > +                   p->da.dma_dev_id = dma_dev_id;
> > +                   p->da.vchan_id = vchan_id;
> > +                   p->da.dma_op = rte_zmalloc_socket(NULL, sizeof(void
> > *) * t->nb_flows,
> > +                                   RTE_CACHE_LINE_SIZE, opt-
> > >socket_id);
> > +
> > +                   p->t = t;
> > +
> > +                   ret = perf_event_dma_adapter_setup(t, p);
> > +                   if (ret)
> > +                           return ret;
> > +
> > +                   for (flow_id = 0; flow_id < t->nb_flows; flow_id++) {
> > +                           rte_mempool_get(t->da_op_pool, (void
> > **)&op);
> > +
> > +                           op->src_seg = rte_malloc(NULL, sizeof(struct
> > rte_dma_sge), 0);
> > +                           op->dst_seg = rte_malloc(NULL, sizeof(struct
> > rte_dma_sge), 0);
> > +
> > +                           op->src_seg->addr =
> > rte_pktmbuf_iova(rte_pktmbuf_alloc(pool));
> > +                           op->dst_seg->addr =
> > rte_pktmbuf_iova(rte_pktmbuf_alloc(pool));
> > +                           op->src_seg->length = 1024;
> > +                           op->dst_seg->length = 1024;
> > +                           op->nb_src = 1;
> > +                           op->nb_dst = 1;
> > +                           op->flags = RTE_DMA_OP_FLAG_SUBMIT;
> > +                           op->op_mp = t->da_op_pool;
> > +                           op->dma_dev_id = dma_dev_id;
> > +                           op->vchan = vchan_id;
> > +
> > +                           response_info = (struct rte_event *)((uint8_t
> > *)op +
> > +                                            sizeof(struct
> > rte_event_dma_adapter_op));
> > +                           response_info->queue_id = p->queue_id;
> > +                           response_info->sched_type =
> > RTE_SCHED_TYPE_ATOMIC;
> > +                           response_info->flow_id = flow_id;
> > +
> > +                           p->da.dma_op[flow_id] = op;
> > +                   }
> > +
> > +                   conf.event_port_cfg |=
> > +                           RTE_EVENT_PORT_CFG_HINT_PRODUCER |
> > +                           RTE_EVENT_PORT_CFG_HINT_CONSUMER;
> > +
> > +                   ret = rte_event_port_setup(opt->dev_id, port, &conf);
> > +                   if (ret) {
> > +                           evt_err("failed to setup port %d", port);
> > +                           return ret;
> > +                   }
> > +
> >                     prod++;
> >             }
> >     } else {
> > @@ -1463,7 +1650,8 @@ perf_opt_check(struct evt_options *opt,
> uint64_t
> > nb_queues)
> >
> >     if (opt->prod_type == EVT_PROD_TYPE_SYNT ||
> >         opt->prod_type == EVT_PROD_TYPE_EVENT_TIMER_ADPTR ||
> > -       opt->prod_type == EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR) {
> > +       opt->prod_type == EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR ||
> > +       opt->prod_type == EVT_PROD_TYPE_EVENT_DMA_ADPTR) {
> >             /* Validate producer lcores */
> >             if (evt_lcores_has_overlap(opt->plcores,
> >                                     rte_get_main_lcore())) {
> > @@ -1855,6 +2043,103 @@ perf_cryptodev_destroy(struct evt_test *test,
> > struct evt_options *opt)
> >     rte_mempool_free(t->ca_vector_pool);
> >  }
> >
> > +int
> > +perf_dmadev_setup(struct evt_test *test, struct evt_options *opt) {
> > +   const struct rte_dma_conf conf = { .nb_vchans = 1};
> > +   const struct rte_dma_vchan_conf qconf = {
> > +                   .direction = RTE_DMA_DIR_MEM_TO_MEM,
> > +                   .nb_desc = 1024,
> > +   };
> > +   struct test_perf *t = evt_test_priv(test);
> > +   uint8_t dma_dev_count, dma_dev_id;
> > +   unsigned int elt_size;
> > +   int ret;
> > +
> > +   if (opt->prod_type != EVT_PROD_TYPE_EVENT_DMA_ADPTR)
> > +           return 0;
> > +
> > +   dma_dev_count = rte_dma_count_avail();
> > +   if (dma_dev_count == 0) {
> > +           evt_err("No dma devices available\n");
> > +           return -ENODEV;
> > +   }
> > +
> > +   elt_size = sizeof(struct rte_event_dma_adapter_op) + sizeof(struct
> > rte_event);
> > +   t->da_op_pool = rte_mempool_create("dma_op_pool", opt->pool_sz,
> > elt_size, 256,
> > +                                      0, NULL, NULL, NULL, NULL,
> > rte_socket_id(), 0);
> > +   if (t->da_op_pool == NULL) {
> > +           evt_err("Failed to create dma op pool");
> > +           return -ENOMEM;
> > +   }
> > +
> > +   for (dma_dev_id = 0; dma_dev_id < dma_dev_count; dma_dev_id++) {
> 
> Since only dma_dev_id 0 is used, we can skip configuring the rest.
> 
> > +           int vchan_id;
> > +
> > +           ret = rte_dma_configure(dma_dev_id, &conf);
> > +           if (ret) {
> > +                   evt_err("Failed to configure dma dev (%u)",
> > dma_dev_id);
> > +                   goto err;
> > +           }
> > +
> > +           for (vchan_id = 0; vchan_id < conf.nb_vchans; vchan_id++) {
> > +                   ret = rte_dma_vchan_setup(dma_dev_id, vchan_id,
> > &qconf);
> > +                   if (ret) {
> > +                           evt_err("Failed to setup vchan on dma dev
> > %u\n",
> > +                                   dma_dev_id);
> > +                           goto err;
> > +                   }
> > +           }
> > +   }
> > +
> > +   return 0;
> > +err:
> > +   for (dma_dev_id = 0; dma_dev_id < dma_dev_count; dma_dev_id++)
> > +           rte_dma_close(dma_dev_id);
> > +
> > +   rte_mempool_free(t->da_op_pool);
> > +
> > +   return ret;
> > +}
> > +
> > +void
> > +perf_dmadev_destroy(struct evt_test *test, struct evt_options *opt) {
> > +   uint8_t dma_dev_id, dma_dev_count = rte_dma_count_avail();
> > +   struct test_perf *t = evt_test_priv(test);
> > +   uint16_t port;
> > +
> > +   if (opt->prod_type != EVT_PROD_TYPE_EVENT_DMA_ADPTR)
> > +           return;
> > +
> > +   for (port = t->nb_workers; port < perf_nb_event_ports(opt); port++) {
> > +           struct prod_data *p = &t->prod[port];
> > +           struct rte_event_dma_adapter_op *op;
> > +           uint32_t flow_id;
> > +
> > +           for (flow_id = 0; flow_id < t->nb_flows; flow_id++) {
> > +                   op = p->da.dma_op[flow_id];
> > +
> > +                   rte_pktmbuf_free((struct rte_mbuf *)(uintptr_t)op-
> > >src_seg->addr);
> > +                   rte_pktmbuf_free((struct rte_mbuf *)(uintptr_t)op-
> > >dst_seg->addr);
> > +                   rte_free(op->src_seg);
> > +                   rte_free(op->dst_seg);
> > +                   rte_mempool_put(op->op_mp, op);
> > +           }
> > +
> > +           rte_event_dma_adapter_vchan_del(TEST_PERF_DA_ID, p-
> > >da.dma_dev_id, p->da.vchan_id);
> > +   }
> > +
> > +   rte_event_dma_adapter_free(TEST_PERF_DA_ID);
> > +
> > +   for (dma_dev_id = 0; dma_dev_id < dma_dev_count; dma_dev_id++) {
> 
> Same as above.
> 
> > +           rte_dma_stop(dma_dev_id);
> > +           rte_dma_close(dma_dev_id);
> > +   }
> > +
> > +   rte_mempool_free(t->da_op_pool);
> > +}
> > +
> >  int
> >  perf_mempool_setup(struct evt_test *test, struct evt_options *opt)  {
> > diff --git a/app/test-eventdev/test_perf_common.h b/app/test-
> > eventdev/test_perf_common.h index faedd471c6..2b4f572b7f 100644
> > --- a/app/test-eventdev/test_perf_common.h
> > +++ b/app/test-eventdev/test_perf_common.h
> > @@ -27,6 +27,7 @@
> >  #include "evt_test.h"
> >
> >  #define TEST_PERF_CA_ID 0
> > +#define TEST_PERF_DA_ID 0
> >
> >  struct test_perf;
> >
> > @@ -43,11 +44,19 @@ struct crypto_adptr_data {
> >     uint16_t cdev_qp_id;
> >     void **crypto_sess;
> >  };
> > +
> > +struct dma_adptr_data {
> > +   uint8_t dma_dev_id;
> > +   uint16_t vchan_id;
> > +   void **dma_op;
> > +};
> > +
> >  struct prod_data {
> >     uint8_t dev_id;
> >     uint8_t port_id;
> >     uint8_t queue_id;
> >     struct crypto_adptr_data ca;
> > +   struct dma_adptr_data da;
> >     struct test_perf *t;
> >  } __rte_cache_aligned;
> >
> > @@ -72,6 +81,7 @@ struct test_perf {
> >     struct rte_mempool *ca_sess_pool;
> >     struct rte_mempool *ca_asym_sess_pool;
> >     struct rte_mempool *ca_vector_pool;
> > +   struct rte_mempool *da_op_pool;
> >  } __rte_cache_aligned;
> >
> >  struct perf_elt {
> > @@ -95,9 +105,8 @@ struct perf_elt {
> >     const uint8_t port = w->port_id;\
> >     const uint8_t prod_timer_type = \
> >             opt->prod_type == EVT_PROD_TYPE_EVENT_TIMER_ADPTR;\
> > -   const uint8_t prod_crypto_type = \
> > -           opt->prod_type ==
> > EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR;\
> >     uint8_t *const sched_type_list = &t->sched_type_list[0];\
> > +   const enum evt_prod_type prod_type = opt->prod_type;\
> >     struct rte_mempool *const pool = t->pool;\
> >     const uint8_t nb_stages = t->opt->nb_stages;\
> >     const uint8_t laststage = nb_stages - 1;\ @@ -206,9 +215,9 @@
> > perf_handle_crypto_vector_ev(struct rte_event *ev, struct perf_elt
> > **pe,  }
> >
> >  static __rte_always_inline int
> > -perf_process_last_stage(struct rte_mempool *const pool, uint8_t
> > prod_crypto_type,
> > -           struct rte_event *const ev, struct worker_data *const w,
> > -           void *bufs[], int const buf_sz, uint8_t count)
> > +perf_process_last_stage(struct rte_mempool *const pool, enum
> > evt_prod_type prod_type,
> > +                   struct rte_event *const ev, struct worker_data *const
> > w,
> > +                   void *bufs[], int const buf_sz, uint8_t count)
> >  {
> >     void *to_free_in_bulk;
> >
> > @@ -219,7 +228,7 @@ perf_process_last_stage(struct rte_mempool
> *const
> > pool, uint8_t prod_crypto_type
> >     rte_atomic_thread_fence(__ATOMIC_RELEASE);
> >     w->processed_pkts++;
> >
> > -   if (prod_crypto_type) {
> > +   if (prod_type == EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR) {
> >             struct rte_crypto_op *op = ev->event_ptr;
> >             struct rte_mbuf *m;
> >
> > @@ -234,6 +243,8 @@ perf_process_last_stage(struct rte_mempool
> *const
> > pool, uint8_t prod_crypto_type
> >                     to_free_in_bulk = op->asym->modex.result.data;
> >             }
> >             rte_crypto_op_free(op);
> > +   } else if (prod_type == EVT_PROD_TYPE_EVENT_DMA_ADPTR) {
> > +           return count;
> >     } else {
> >             to_free_in_bulk = ev->event_ptr;
> >     }
> > @@ -248,9 +259,9 @@ perf_process_last_stage(struct rte_mempool
> *const
> > pool, uint8_t prod_crypto_type  }
> >
> >  static __rte_always_inline uint8_t
> > -perf_process_last_stage_latency(struct rte_mempool *const pool,
> > uint8_t prod_crypto_type,
> > -           struct rte_event *const ev, struct worker_data *const w,
> > -           void *bufs[], int const buf_sz, uint8_t count)
> > +perf_process_last_stage_latency(struct rte_mempool *const pool, enum
> > evt_prod_type prod_type,
> > +                           struct rte_event *const ev, struct worker_data
> > *const w,
> > +                           void *bufs[], int const buf_sz, uint8_t count)
> >  {
> >     uint64_t latency;
> >     struct perf_elt *pe;
> > @@ -262,7 +273,7 @@ perf_process_last_stage_latency(struct
> rte_mempool
> > *const pool, uint8_t prod_cry
> >     rte_atomic_thread_fence(__ATOMIC_RELEASE);
> >     w->processed_pkts++;
> >
> > -   if (prod_crypto_type) {
> > +   if (prod_type == EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR) {
> >             struct rte_crypto_op *op = ev->event_ptr;
> >             struct rte_mbuf *m;
> >
> > @@ -280,6 +291,8 @@ perf_process_last_stage_latency(struct
> rte_mempool
> > *const pool, uint8_t prod_cry
> >                     to_free_in_bulk = op->asym->modex.result.data;
> >             }
> >             rte_crypto_op_free(op);
> > +   } else if (prod_type == EVT_PROD_TYPE_EVENT_DMA_ADPTR) {
> > +           return count;
> >     } else {
> >             pe = ev->event_ptr;
> >             to_free_in_bulk = pe;
> > @@ -346,6 +359,7 @@ int perf_opt_check(struct evt_options *opt,
> > uint64_t nb_queues);  int perf_test_setup(struct evt_test *test,
> > struct evt_options *opt);  int perf_ethdev_setup(struct evt_test
> > *test, struct evt_options *opt);  int perf_cryptodev_setup(struct
> > evt_test *test, struct evt_options *opt);
> > +int perf_dmadev_setup(struct evt_test *test, struct evt_options
> > +*opt);
> >  int perf_mempool_setup(struct evt_test *test, struct evt_options
> > *opt);  int perf_event_dev_port_setup(struct evt_test *test, struct
> evt_options *opt,
> >                             uint8_t stride, uint8_t nb_queues, @@ -
> 357,6 +371,7 @@ void
> > perf_opt_dump(struct evt_options *opt, uint8_t nb_queues);  void
> > perf_test_destroy(struct evt_test *test, struct evt_options *opt);
> > void perf_eventdev_destroy(struct evt_test *test, struct evt_options
> > *opt);  void perf_cryptodev_destroy(struct evt_test *test, struct
> > evt_options *opt);
> > +void perf_dmadev_destroy(struct evt_test *test, struct evt_options
> > +*opt);
> >  void perf_ethdev_destroy(struct evt_test *test, struct evt_options
> > *opt);  void perf_ethdev_rx_stop(struct evt_test *test, struct
> > evt_options *opt);  void perf_mempool_destroy(struct evt_test *test,
> > struct evt_options *opt); diff --git
> > a/app/test-eventdev/test_perf_queue.c b/app/test-
> > eventdev/test_perf_queue.c index 2399cfb69b..8b6b85c1ad 100644
> > --- a/app/test-eventdev/test_perf_queue.c
> > +++ b/app/test-eventdev/test_perf_queue.c
> > @@ -50,7 +50,8 @@ perf_queue_worker(void *arg, const int
> > enable_fwd_latency)
> >                     continue;
> >             }
> >
> > -           if (prod_crypto_type && (ev.event_type ==
> > RTE_EVENT_TYPE_CRYPTODEV)) {
> > +           if ((prod_type == EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR)
> > &&
> > +               (ev.event_type == RTE_EVENT_TYPE_CRYPTODEV)) {
> >                     if (perf_handle_crypto_ev(&ev, &pe,
> > enable_fwd_latency))
> >                             continue;
> >             } else {
> > @@ -65,10 +66,10 @@ perf_queue_worker(void *arg, const int
> > enable_fwd_latency)
> >             /* last stage in pipeline */
> >             if (unlikely(stage == laststage)) {
> >                     if (enable_fwd_latency)
> > -                           cnt = perf_process_last_stage_latency(pool,
> > prod_crypto_type,
> > +                           cnt = perf_process_last_stage_latency(pool,
> > prod_type,
> >                                     &ev, w, bufs, sz, cnt);
> >                     else
> > -                           cnt = perf_process_last_stage(pool,
> > prod_crypto_type,
> > +                           cnt = perf_process_last_stage(pool,
> > prod_type,
> >                                     &ev, w, bufs, sz, cnt);
> >             } else {
> >                     fwd_event(&ev, sched_type_list, nb_stages); @@ -
> 101,7 +102,8 @@
> > perf_queue_worker_burst(void *arg, const int
> > enable_fwd_latency)
> >             }
> >
> >             for (i = 0; i < nb_rx; i++) {
> > -                   if (prod_crypto_type && (ev[i].event_type ==
> > RTE_EVENT_TYPE_CRYPTODEV)) {
> > +                   if ((prod_type ==
> > EVT_PROD_TYPE_EVENT_CRYPTO_ADPTR) &&
> > +                       (ev[i].event_type ==
> > RTE_EVENT_TYPE_CRYPTODEV)) {
> >                             if (perf_handle_crypto_ev(&ev[i], &pe,
> > enable_fwd_latency))
> >                                     continue;
> >                     }
> > @@ -118,9 +120,9 @@ perf_queue_worker_burst(void *arg, const int
> > enable_fwd_latency)
> >                     if (unlikely(stage == laststage)) {
> >                             if (enable_fwd_latency)
> >                                     cnt =
> > perf_process_last_stage_latency(pool,
> > -                                           prod_crypto_type, &ev[i], w,
> > bufs, sz, cnt);
> > +                                           prod_type, &ev[i], w, bufs, sz,
> > cnt);
> >                             else
> > -                                   cnt = perf_process_last_stage(pool,
> > prod_crypto_type,
> > +                                   cnt = perf_process_last_stage(pool,
> > prod_type,
> >                                             &ev[i], w, bufs, sz, cnt);
> >
> >                             ev[i].op = RTE_EVENT_OP_RELEASE;
> > @@ -151,7 +153,7 @@ perf_queue_worker_vector(void *arg, const int
> > enable_fwd_latency)
> >
> >     RTE_SET_USED(sz);
> >     RTE_SET_USED(cnt);
> > -   RTE_SET_USED(prod_crypto_type);
> > +   RTE_SET_USED(prod_type);
> >
> >     while (t->done == false) {
> >             deq = rte_event_dequeue_burst(dev, port, &ev, 1, 0); @@ -
> 346,6
> > +348,18 @@ perf_queue_eventdev_setup(struct evt_test *test, struct
> > evt_options *opt)
> >                             return ret;
> >                     }
> >             }
> > +   } else if (opt->prod_type == EVT_PROD_TYPE_EVENT_DMA_ADPTR) {
> > +           uint8_t dma_dev_id, dma_dev_count;
> > +
> > +           dma_dev_count = rte_dma_count_avail();
> > +           for (dma_dev_id = 0; dma_dev_id < dma_dev_count;
> > dma_dev_id++) {
> 
> Since we only use the 1st dma_dev_id in fp, we should skip this loop.
> 
> > +                   ret = rte_dma_start(dma_dev_id);
> > +                   if (ret) {
> > +                           evt_err("Failed to start dmadev %u",
> > +                                   dma_dev_id);
> > +                           return ret;
> > +                   }
> > +           }
> >     }
> >
> >     return 0;
> > @@ -389,6 +403,7 @@ static const struct evt_test_ops perf_queue =  {
> >     .mempool_setup      = perf_mempool_setup,
> >     .ethdev_setup       = perf_ethdev_setup,
> >     .cryptodev_setup    = perf_cryptodev_setup,
> > +   .dmadev_setup       = perf_dmadev_setup,
> >     .ethdev_rx_stop     = perf_ethdev_rx_stop,
> >     .eventdev_setup     = perf_queue_eventdev_setup,
> >     .launch_lcores      = perf_queue_launch_lcores,
> > @@ -396,6 +411,7 @@ static const struct evt_test_ops perf_queue =  {
> >     .mempool_destroy    = perf_mempool_destroy,
> >     .ethdev_destroy     = perf_ethdev_destroy,
> >     .cryptodev_destroy  = perf_cryptodev_destroy,
> > +   .dmadev_destroy     = perf_dmadev_destroy,
> >     .test_result        = perf_test_result,
> >     .test_destroy       = perf_test_destroy,
> >  };
> > diff --git a/doc/guides/tools/testeventdev.rst
> > b/doc/guides/tools/testeventdev.rst
> > index 3fcc2c9894..a29afe2cb2 100644
> > --- a/doc/guides/tools/testeventdev.rst
> > +++ b/doc/guides/tools/testeventdev.rst
> 
> Please Add example command
> 
> > @@ -124,6 +124,10 @@ The following are the application command-line
> > options:
> >
> >          Use crypto device as producer.
> >
> > +* ``--prod_type_dmadev``
> > +
> > +        Use dma device as producer.
> > +
> >  * ``--timer_tick_nsec``
> >
> >         Used to dictate number of nano seconds between bucket
> > traversal of the @@ -157,6 +161,11 @@ The following are the
> > application command-line
> > options:
> >          Set crypto adapter mode. Use 0 for OP_NEW (default) and 1 for
> >          OP_FORWARD mode.
> >
> > +* ``--dma_adptr_mode``
> > +
> > +        Set dma adapter mode. Use 0 for OP_NEW (default) and 1 for
> > +        OP_FORWARD mode.
> > +
> >  * ``--crypto_op_type``
> >
> >          Set crypto operation type. Use 0 for symmetric crypto ops
> > (default) @@ -459,6 +468,7 @@ Supported application command line
> > options are
> > following::
> >          --prod_type_timerdev_burst
> >          --prod_type_timerdev
> >          --prod_type_cryptodev
> > +        --prod_type_dmadev
> >          --prod_enq_burst_sz
> >          --timer_tick_nsec
> >          --max_tmo_nsec
> > @@ -467,6 +477,7 @@ Supported application command line options are
> > following::
> >          --nb_timer_adptrs
> >          --deq_tmo_nsec
> >          --crypto_adptr_mode
> > +        --dma_adptr_mode
> >
> >  Example
> >  ^^^^^^^
> > @@ -570,6 +581,7 @@ Supported application command line options are
> > following::
> >          --prod_type_timerdev_burst
> >          --prod_type_timerdev
> >          --prod_type_cryptodev
> > +        --prod_type_dmadev
> >          --timer_tick_nsec
> >          --max_tmo_nsec
> >          --expiry_nsec
> > @@ -577,6 +589,7 @@ Supported application command line options are
> > following::
> >          --nb_timer_adptrs
> >          --deq_tmo_nsec
> >          --crypto_adptr_mode
> > +        --dma_adptr_mode
> >
> >  Example
> >  ^^^^^^^
> > --
> > 2.34.1

Reply via email to