-----Original Message-----
> Date: Thu, 1 Jun 2017 21:04:15 +0000
> From: "Eads, Gage" <gage.e...@intel.com>
> To: Jerin Jacob <jerin.ja...@caviumnetworks.com>, "dev@dpdk.org"
>  <dev@dpdk.org>
> CC: "Richardson, Bruce" <bruce.richard...@intel.com>, "Van Haaren, Harry"
>  <harry.van.haa...@intel.com>, "hemant.agra...@nxp.com"
>  <hemant.agra...@nxp.com>, "nipun.gu...@nxp.com" <nipun.gu...@nxp.com>,
>  "Vangati, Narender" <narender.vang...@intel.com>, "Rao, Nikhil"
>  <nikhil....@intel.com>, "gprathyu...@caviumnetworks.com"
>  <gprathyu...@caviumnetworks.com>
> Subject: RE: [dpdk-dev] [PATCH 25/33] app/testeventdev: perf queue: add
>  worker functions
> 
> 
> 
> >  -----Original Message-----
> >  From: Jerin Jacob [mailto:jerin.ja...@caviumnetworks.com]
> >  Sent: Sunday, May 28, 2017 2:59 PM
> >  To: dev@dpdk.org
> >  Cc: Richardson, Bruce <bruce.richard...@intel.com>; Van Haaren, Harry
> >  <harry.van.haa...@intel.com>; hemant.agra...@nxp.com; Eads, Gage
> >  <gage.e...@intel.com>; nipun.gu...@nxp.com; Vangati, Narender
> >  <narender.vang...@intel.com>; Rao, Nikhil <nikhil....@intel.com>;
> >  gprathyu...@caviumnetworks.com; Jerin Jacob
> >  <jerin.ja...@caviumnetworks.com>
> >  Subject: [dpdk-dev] [PATCH 25/33] app/testeventdev: perf queue: add worker
> >  functions
> >  
> >  Signed-off-by: Jerin Jacob <jerin.ja...@caviumnetworks.com>
> >  ---
> >   app/test-eventdev/test_perf_common.h |  60 +++++++++++++++  app/test-
> >  eventdev/test_perf_queue.c  | 137 +++++++++++++++++++++++++++++++++++
> >   2 files changed, 197 insertions(+)
> >  
> >  diff --git a/app/test-eventdev/test_perf_common.h b/app/test-
> >  eventdev/test_perf_common.h
> >  index f8246953a..9888e5078 100644
> >  --- a/app/test-eventdev/test_perf_common.h
> >  +++ b/app/test-eventdev/test_perf_common.h
> >  @@ -86,6 +86,66 @@ struct perf_elt {
> >     uint64_t timestamp;
> >   } __rte_cache_aligned;
> >  
> >  +#define BURST_SIZE 16
> >  +
> >  +#define PERF_WORKER_INIT\
> >  +  struct worker_data *w  = arg;\
> >  +  struct test_perf *t = w->t;\
> >  +  struct evt_options *opt = t->opt;\
> >  +  const uint8_t dev = w->dev_id;\
> >  +  const uint8_t port = w->port_id;\
> >  +  uint8_t *const sched_type_list = &t->sched_type_list[0];\
> >  +  struct rte_mempool *const pool = t->pool;\
> >  +  const uint8_t nb_stages = t->opt->nb_stages;\
> >  +  const uint8_t laststage = nb_stages - 1;\
> >  +  uint8_t cnt = 0;\
> >  +  void *bufs[16] __rte_cache_aligned;\
> >  +  int const sz = RTE_DIM(bufs);\
> >  +  if (opt->verbose_level > 1)\
> >  +          printf("%s(): lcore %d dev_id %d port=%d\n", __func__,\
> >  +                          rte_lcore_id(), dev, port)
> >  +
> >  +static inline __attribute__((always_inline)) int
> >  +perf_process_last_stage(struct rte_mempool *const pool,
> >  +          struct rte_event *const ev, struct worker_data *const w,
> >  +          void *bufs[], int const buf_sz, uint8_t count) {
> >  +  bufs[count++] = ev->event_ptr;
> >  +  w->processed_pkts++;
> >  +  rte_smp_wmb();
> >  +
> >  +  if (unlikely(count == buf_sz)) {
> >  +          count = 0;
> >  +          rte_mempool_put_bulk(pool, bufs, buf_sz);
> >  +  }
> >  +  return count;
> >  +}
> >  +
> >  +static inline __attribute__((always_inline)) uint8_t
> >  +perf_process_last_stage_latency(struct rte_mempool *const pool,
> >  +          struct rte_event *const ev, struct worker_data *const w,
> >  +          void *bufs[], int const buf_sz, uint8_t count) {
> >  +  uint64_t latency;
> >  +  struct perf_elt *const m = ev->event_ptr;
> >  +
> >  +  bufs[count++] = ev->event_ptr;
> >  +  w->processed_pkts++;
> >  +
> >  +  if (unlikely(count == buf_sz)) {
> >  +          count = 0;
> >  +          latency = rte_get_timer_cycles() - m->timestamp;
> >  +          rte_mempool_put_bulk(pool, bufs, buf_sz);
> >  +  } else {
> >  +          latency = rte_get_timer_cycles() - m->timestamp;
> >  +  }
> >  +
> >  +  w->latency += latency;
> >  +  rte_smp_wmb();
> >  +  return count;
> >  +}
> 
> What purpose does the store barrier serve in these two functions?

The master core(!worker core) reads w->latency and
w->processed_pkts periodically from all workers.


Reply via email to