Re: [dpdk-dev] [EXT] [PATCH] event/cnxk: fix clang warning on Arm

2021-06-18 Thread Pavan Nikhilesh Bhagavatula



>-Original Message-
>From: Ruifeng Wang 
>Sent: Thursday, June 10, 2021 12:55 PM
>To: Pavan Nikhilesh Bhagavatula ; Shijith
>Thotton 
>Cc: dev@dpdk.org; n...@arm.com; Ruifeng Wang
>
>Subject: [EXT] [PATCH] event/cnxk: fix clang warning on Arm
>
>External Email
>
>--
>Build with Clang-10 has warning:
>drivers/event/cnxk/cnxk_tim_worker.h:372:23: warning: value size
>does not match register size specified by the constraint and modifier [-
>Wasm-operand-widths]
> : [rem] "=&r"(rem)
>   ^
>drivers/event/cnxk/cnxk_tim_worker.h:365:17: note: use constraint
>modifier "w"
> "  ldxr %[rem], [%[crem]]  \n"
> ^~
> %w[rem]
>
>Changed variable type to match register size, which placates clang.
>
>Fixes: 300b796262a1 ("event/cnxk: add timer arm routine")
>Cc: pbhagavat...@marvell.com
>
>Signed-off-by: Ruifeng Wang 

LGTM, thanks.

Acked-by: Pavan Nikhilesh 

>---
> drivers/event/cnxk/cnxk_tim_worker.h | 2 +-
> 1 file changed, 1 insertion(+), 1 deletion(-)
>
>diff --git a/drivers/event/cnxk/cnxk_tim_worker.h
>b/drivers/event/cnxk/cnxk_tim_worker.h
>index 7caeb1a8fb..78e36ffafe 100644
>--- a/drivers/event/cnxk/cnxk_tim_worker.h
>+++ b/drivers/event/cnxk/cnxk_tim_worker.h
>@@ -320,7 +320,7 @@ cnxk_tim_add_entry_mp(struct cnxk_tim_ring
>*const tim_ring,
>   struct cnxk_tim_ent *chunk;
>   struct cnxk_tim_bkt *bkt;
>   uint64_t lock_sema;
>-  int16_t rem;
>+  int64_t rem;
>
> __retry:
>   cnxk_tim_get_target_bucket(tim_ring, rel_bkt, &bkt,
>&mirr_bkt);
>--
>2.25.1



Re: [dpdk-dev] [PATCH v4 1/7] vhost: fix missing memory table NUMA realloc

2021-06-18 Thread Maxime Coquelin



On 6/18/21 6:34 AM, Xia, Chenbo wrote:
> Hi Maxime,
> 
>> -Original Message-
>> From: Maxime Coquelin 
>> Sent: Thursday, June 17, 2021 11:38 PM
>> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo 
>> 
>> Cc: Maxime Coquelin ; sta...@dpdk.org
>> Subject: [PATCH v4 1/7] vhost: fix missing memory table NUMA realloc
>>
>> When the guest allocates virtqueues on a different NUMA node
>> than the one the Vhost metadata are allocated, both the Vhost
>> device struct and the virtqueues struct are reallocated.
>>
>> However, reallocating the Vhost memory table was missing, which
>> likely causes iat least one cross-NUMA accesses for every burst
>> of packets.
> 
> 'at least' ?

yes.

>>
>> This patch reallocates this table on the same NUMA node as the
>> other metadata.
>>
>> Fixes: 552e8fd3d2b4 ("vhost: simplify memory regions handling")
>> Cc: sta...@dpdk.org
>>
>> Reported-by: David Marchand 
>> Signed-off-by: Maxime Coquelin 
>> ---
>>  lib/vhost/vhost_user.c | 15 +++
>>  1 file changed, 15 insertions(+)
>>
>> diff --git a/lib/vhost/vhost_user.c b/lib/vhost/vhost_user.c
>> index 8f0eba6412..031e3bfa2f 100644
>> --- a/lib/vhost/vhost_user.c
>> +++ b/lib/vhost/vhost_user.c
>> @@ -557,6 +557,9 @@ numa_realloc(struct virtio_net *dev, int index)
> 
> As we are realloc more things now, the comment above 'numa_realloc(XXX)'
> should also be changed like:
> 
> Reallocate related data structure to make them on the same numa node as
> the memory of vring descriptor.

Agree, I'll put this:

"
Reallocate virtio_dev, vhost_virtqueue and related data structures to
make them on the same numa node as the memory of vring descriptor.
"

Thanks,
Maxime

> Thanks,
> Chenbo
> 
>>  goto out;
>>  }
>>  if (oldnode != newnode) {
>> +struct rte_vhost_memory *old_mem;
>> +ssize_t mem_size;
>> +
>>  VHOST_LOG_CONFIG(INFO,
>>  "reallocate dev from %d to %d node\n",
>>  oldnode, newnode);
>> @@ -568,6 +571,18 @@ numa_realloc(struct virtio_net *dev, int index)
>>
>>  memcpy(dev, old_dev, sizeof(*dev));
>>  rte_free(old_dev);
>> +
>> +mem_size = sizeof(struct rte_vhost_memory) +
>> +sizeof(struct rte_vhost_mem_region) * 
>> dev->mem->nregions;
>> +old_mem = dev->mem;
>> +dev->mem = rte_malloc_socket(NULL, mem_size, 0, newnode);
>> +if (!dev->mem) {
>> +dev->mem = old_mem;
>> +goto out;
>> +}
>> +
>> +memcpy(dev->mem, old_mem, mem_size);
>> +rte_free(old_mem);
>>  }
>>
>>  out:
>> --
>> 2.31.1
> 



Re: [dpdk-dev] [RFC v2] porting AddressSanitizer feature to DPDK

2021-06-18 Thread Lin, Xueqin
> -Original Message-
> From: Jerin Jacob 
> Sent: Wednesday, June 16, 2021 7:35 PM
> To: Lin, Xueqin 
> Cc: Peng, ZhihongX ; Burakov, Anatoly
> ; Ananyev, Konstantin
> ; Stephen Hemminger
> ; dpdk-dev 
> Subject: Re: [dpdk-dev] [RFC v2] porting AddressSanitizer feature to DPDK
> 
> On Wed, Jun 16, 2021 at 2:43 PM Lin, Xueqin  wrote:
> >
> > > -Original Message-
> > > From: Jerin Jacob 
> > > Sent: Tuesday, June 15, 2021 4:40 PM
> > > To: Peng, ZhihongX 
> > > Cc: Burakov, Anatoly ; Ananyev,
> > > Konstantin ; Stephen Hemminger
> > > ; dpdk-dev ; Lin,
> Xueqin
> > > 
> > > Subject: Re: [dpdk-dev] [RFC v2] porting AddressSanitizer feature to
> > > DPDK
> > >
> > > On Tue, Jun 15, 2021 at 1:46 PM  wrote:
> > > >
> > > > From: Zhihong Peng 
> > > >
> > > > AddressSanitizer (ASan) is a google memory error detect standard tool.
> > > > It could help to detect use-after-free and
> > > > {heap,stack,global}-buffer overflow bugs in C/C++ programs, print
> > > > detailed error information when error happens, large improve debug
> efficiency.
> > > >
> > > > By referring to its implementation algorithm
> > > > (https://github.com/google/sanitizers/wiki/AddressSanitizerAlgorit
> > > > hm), ported heap-buffer-overflow and use-after-freefunctions to
> > > > dpdk.
> > > >
> > > > Here is an example of heap-buffer-overflow bug:
> > > > ..
> > > > char *p = rte_zmalloc(NULL, 7, 0);
> > > > p[7] = 'a';
> > > > ..
> > > >
> > > > Here is an example of use-after-free bug:
> > > > ..
> > > > char *p = rte_zmalloc(NULL, 7, 0);
> > > > rte_free(p);
> > > > *p = 'a';
> > > > ..
> > > >
> > > > If you want to use this feature,
> > > > you need to use the following compilation options:
> > > > -Db_lundef=false -Db_sanitize=address
> > >
> > > # Thanks for this patch. It is a useful item.
> > >
> > > # Subject could be changed
> > > from:
> > > porting AddressSanitizer feature to DPDK to
> > > eal: support for  AddressSanitizer
> > > or so
> >
> > Thanks for your positive feedback and review.
> > Good point, we will update the title in next version.
> >
> > >
> > > # Could you add a section in the documentation for Sanitizers to
> > > document the build time option and other points that users need to know.
> >
> > Make sense to add build option and key points to document, will add
> > this part in doc
> folder.
> >
> > > We can add other sanitizers such as UBSan etc in the future here
> > WIP to research other sanitizer tool.
> 
> UBsan is a good candate.
> Some old DPDK patch for the same:
> http://patches.dpdk.org/project/dpdk/patch/1573832013-18946-1-git-send-
> email-hka...@marvell.com/
> 

Thanks for share, we will research the tool next. 

> > >
> > > # Add a UT test case to make sure it is working in app/test or so.
> >
> > This tool could help to detect memory issue, need to change bad code to
> check if working.
> 
> It is better to have a  UT to test things are working. You could add it in
> app/test/test_address_sanity.c. Tests can be such that
> - Skip if not complied with Sanity enabled
> - Pass if the code detects the known bad code. You can have test cases with
> pubic rte_ API that internally exercise the verify your implementation related
> to new asan_* APIs.

Thanks Jacob for your suggestion, we will add unit test for Asan tool later. 

> 
> 
> > Suggest listing demo code and tool capture information for user to try if
> tool works, also add this part into doc.
> >
> > >
> > > # Also, Please update the release note for this feature.
> > Sure, we can update the release note if code merge.
> 
> Probably you can send v1 version next i.e change the RFC status to get
> merged.

Sure, we will send v1 patch if no obvious objection for that, hope patch could 
receive some ACKs and could success to merge, thanks.  

> 
> 
> >


Re: [dpdk-dev] [PATCH v4 4/7] vhost: fix NUMA reallocation with multiqueue

2021-06-18 Thread Maxime Coquelin



On 6/18/21 6:34 AM, Xia, Chenbo wrote:
> Hi Maxime,
> 
>> -Original Message-
>> From: Maxime Coquelin 
>> Sent: Thursday, June 17, 2021 11:38 PM
>> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo 
>> 
>> Cc: Maxime Coquelin ; sta...@dpdk.org
>> Subject: [PATCH v4 4/7] vhost: fix NUMA reallocation with multiqueue
>>
>> Since the Vhost-user device initialization has been reworked,
>> enabling the application to start using the device as soon as
>> the first queue pair is ready, NUMA reallocation no more
>> happened on queue pairs other than the first one since
>> numa_realloc() was returning early if the device was running.
>>
>> This patch fixes this issue by only preventing the device
>> metadata to be allocated if the device is running. For the
>> virtqueues, a vring state change notification is sent to
>> notify the application of its disablement. Since the callback
>> is supposed to be blocking, it is safe to reallocate it
>> afterwards.
> 
> Is there a corner case? Numa_realloc may happen during vhost-user msg
> set_vring_addr/kick, set_mem_table and iotlb msg. And iotlb msg will
> not take vq access lock. It could happen when numa_realloc happens on
> iotlb msg and app accesses vq in the meantime?

I think we are safe wrt to numa_realloc(), because the app's
.vring_state_changed() callback is only returning when it is no more
processing the rings.


> Thanks,
> Chenbo
> 
>>
>> Fixes: d0fcc38f5fa4 ("vhost: improve device readiness notifications")
>> Cc: sta...@dpdk.org
>>
>> Signed-off-by: Maxime Coquelin 
>> ---
>>  lib/vhost/vhost_user.c | 11 ---
>>  1 file changed, 8 insertions(+), 3 deletions(-)
>>
>> diff --git a/lib/vhost/vhost_user.c b/lib/vhost/vhost_user.c
>> index 0e9e26ebe0..6e7b327ef8 100644
>> --- a/lib/vhost/vhost_user.c
>> +++ b/lib/vhost/vhost_user.c
>> @@ -488,9 +488,6 @@ numa_realloc(struct virtio_net *dev, int index)
>>  struct batch_copy_elem *new_batch_copy_elems;
>>  int ret;
>>
>> -if (dev->flags & VIRTIO_DEV_RUNNING)
>> -return dev;
>> -
>>  old_dev = dev;
>>  vq = old_vq = dev->virtqueue[index];
>>
>> @@ -506,6 +503,11 @@ numa_realloc(struct virtio_net *dev, int index)
>>  return dev;
>>  }
>>  if (oldnode != newnode) {
>> +if (vq->ready) {
>> +vq->ready = false;
>> +vhost_user_notify_queue_state(dev, index, 0);
>> +}
>> +
>>  VHOST_LOG_CONFIG(INFO,
>>  "reallocate vq from %d to %d node\n", oldnode, newnode);
>>  vq = rte_malloc_socket(NULL, sizeof(*vq), 0, newnode);
>> @@ -558,6 +560,9 @@ numa_realloc(struct virtio_net *dev, int index)
>>  rte_free(old_vq);
>>  }
>>
>> +if (dev->flags & VIRTIO_DEV_RUNNING)
>> +goto out;
>> +
>>  /* check if we need to reallocate dev */
>>  ret = get_mempolicy(&oldnode, NULL, 0, old_dev,
>>  MPOL_F_NODE | MPOL_F_ADDR);
>> --
>> 2.31.1
> 



Re: [dpdk-dev] [PATCH v4 5/7] vhost: improve NUMA reallocation

2021-06-18 Thread Maxime Coquelin



On 6/18/21 6:34 AM, Xia, Chenbo wrote:
> Hi Maxime,
> 
>> -Original Message-
>> From: Maxime Coquelin 
>> Sent: Thursday, June 17, 2021 11:38 PM
>> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo 
>> 
>> Cc: Maxime Coquelin 
>> Subject: [PATCH v4 5/7] vhost: improve NUMA reallocation
>>
>> This patch improves the numa_realloc() function by making use
>> of rte_realloc_socket(), which takes care of the memory copy
>> and freeing of the old data.
>>
>> Suggested-by: David Marchand 
>> Signed-off-by: Maxime Coquelin 
>> ---
>>  lib/vhost/vhost_user.c | 195 ++---
>>  1 file changed, 86 insertions(+), 109 deletions(-)
>>
>> diff --git a/lib/vhost/vhost_user.c b/lib/vhost/vhost_user.c
>> index 6e7b327ef8..0590ef6d14 100644
>> --- a/lib/vhost/vhost_user.c
>> +++ b/lib/vhost/vhost_user.c
>> @@ -480,144 +480,121 @@ vhost_user_set_vring_num(struct virtio_net **pdev,
>>  static struct virtio_net*
>>  numa_realloc(struct virtio_net *dev, int index)
>>  {
>> -int oldnode, newnode;
>> +int node, dev_node;
>>  struct virtio_net *old_dev;
>> -struct vhost_virtqueue *old_vq, *vq;
>> -struct vring_used_elem *new_shadow_used_split;
>> -struct vring_used_elem_packed *new_shadow_used_packed;
>> -struct batch_copy_elem *new_batch_copy_elems;
>> +struct vhost_virtqueue *vq;
>> +struct batch_copy_elem *bce;
>> +struct guest_page *gp;
>> +struct rte_vhost_memory *mem;
>> +size_t mem_size;
>>  int ret;
>>
>>  old_dev = dev;
>> -vq = old_vq = dev->virtqueue[index];
>> -
>> -ret = get_mempolicy(&newnode, NULL, 0, old_vq->desc,
>> -MPOL_F_NODE | MPOL_F_ADDR);
>> +vq = dev->virtqueue[index];
>>
>> -/* check if we need to reallocate vq */
>> -ret |= get_mempolicy(&oldnode, NULL, 0, old_vq,
>> - MPOL_F_NODE | MPOL_F_ADDR);
>> +ret = get_mempolicy(&node, NULL, 0, vq->desc, MPOL_F_NODE | 
>> MPOL_F_ADDR);
>>  if (ret) {
>> -VHOST_LOG_CONFIG(ERR,
>> -"Unable to get vq numa information.\n");
>> +VHOST_LOG_CONFIG(ERR, "Unable to get virtqueue %d numa
>> information.\n", index);
>>  return dev;
>>  }
>> -if (oldnode != newnode) {
>> -if (vq->ready) {
>> -vq->ready = false;
>> -vhost_user_notify_queue_state(dev, index, 0);
>> -}
>>
>> -VHOST_LOG_CONFIG(INFO,
>> -"reallocate vq from %d to %d node\n", oldnode, newnode);
>> -vq = rte_malloc_socket(NULL, sizeof(*vq), 0, newnode);
>> -if (!vq)
>> -return dev;
>> +if (vq->ready) {
>> +vq->ready = false;
>> +vhost_user_notify_queue_state(dev, index, 0);
>> +}
>>
>> -memcpy(vq, old_vq, sizeof(*vq));
>> +vq = rte_realloc_socket(vq, sizeof(*vq), 0, node);
>> +if (!vq) {
>> +VHOST_LOG_CONFIG(ERR, "Failed to realloc virtqueue %d on
>> node %d\n",
>> +index, node);
>> +return dev;
>> +}
>>
>> -if (vq_is_packed(dev)) {
>> -new_shadow_used_packed = rte_malloc_socket(NULL,
>> -vq->size *
>> -sizeof(struct vring_used_elem_packed),
>> -RTE_CACHE_LINE_SIZE,
>> -newnode);
>> -if (new_shadow_used_packed) {
>> -rte_free(vq->shadow_used_packed);
>> -vq->shadow_used_packed = new_shadow_used_packed;
>> -}
>> -} else {
>> -new_shadow_used_split = rte_malloc_socket(NULL,
>> -vq->size *
>> -sizeof(struct vring_used_elem),
>> -RTE_CACHE_LINE_SIZE,
>> -newnode);
>> -if (new_shadow_used_split) {
>> -rte_free(vq->shadow_used_split);
>> -vq->shadow_used_split = new_shadow_used_split;
>> -}
>> -}
>> +if (vq != dev->virtqueue[index]) {
>> +VHOST_LOG_CONFIG(INFO, "reallocated virtqueue on node %d\n", 
>> node);
>> +dev->virtqueue[index] = vq;
>> +vhost_user_iotlb_init(dev, index);
>> +}
>> +
>> +if (vq_is_packed(dev)) {
>> +struct vring_used_elem_packed *sup;
>>
>> -new_batch_copy_elems = rte_malloc_socket(NULL,
>> -vq->size * sizeof(struct batch_copy_elem),
>> -RTE_CACHE_LINE_SIZE,
>> -newnode);
>> -if (new_batch_copy_elems) {
>> -rte_free(vq->batch_copy_elems);
>> -vq->batch_copy_elems = new_batch_copy_elems;
>>

[dpdk-dev] [PATCH v2 1/4] examples/vhost: refactor vhost enqueue and dequeue datapaths.

2021-06-18 Thread Wenwu Ma
Previously, by judging the flag, we call different enqueue/dequeue
functions in data path.

Now, we use an ops that was initialized when Vhost was created,
so that we can call ops directly in Vhost data path without any more
flag judgment.

Signed-off-by: Wenwu Ma 
---
 examples/vhost/main.c   | 112 
 examples/vhost/main.h   |  33 +--
 examples/vhost/virtio_net.c |  16 +-
 3 files changed, 105 insertions(+), 56 deletions(-)

diff --git a/examples/vhost/main.c b/examples/vhost/main.c
index d2179eadb9..aebdc3a566 100644
--- a/examples/vhost/main.c
+++ b/examples/vhost/main.c
@@ -106,6 +106,8 @@ static uint32_t burst_rx_retry_num = BURST_RX_RETRIES;
 static char *socket_files;
 static int nb_sockets;
 
+static struct vhost_queue_ops vdev_queue_ops[MAX_VHOST_DEVICE];
+
 /* empty vmdq configuration structure. Filled in programatically */
 static struct rte_eth_conf vmdq_conf_default = {
.rxmode = {
@@ -885,27 +887,8 @@ drain_vhost(struct vhost_dev *vdev)
uint16_t nr_xmit = vhost_txbuff[buff_idx]->len;
struct rte_mbuf **m = vhost_txbuff[buff_idx]->m_table;
 
-   if (builtin_net_driver) {
-   ret = vs_enqueue_pkts(vdev, VIRTIO_RXQ, m, nr_xmit);
-   } else if (async_vhost_driver) {
-   uint32_t cpu_cpl_nr = 0;
-   uint16_t enqueue_fail = 0;
-   struct rte_mbuf *m_cpu_cpl[nr_xmit];
-
-   complete_async_pkts(vdev);
-   ret = rte_vhost_submit_enqueue_burst(vdev->vid, VIRTIO_RXQ,
-   m, nr_xmit, m_cpu_cpl, &cpu_cpl_nr);
-
-   if (cpu_cpl_nr)
-   free_pkts(m_cpu_cpl, cpu_cpl_nr);
-
-   enqueue_fail = nr_xmit - ret;
-   if (enqueue_fail)
-   free_pkts(&m[ret], nr_xmit - ret);
-   } else {
-   ret = rte_vhost_enqueue_burst(vdev->vid, VIRTIO_RXQ,
-   m, nr_xmit);
-   }
+   ret = vdev_queue_ops[vdev->vid].enqueue_pkt_burst(vdev,
+   VIRTIO_RXQ, m, nr_xmit);
 
if (enable_stats) {
__atomic_add_fetch(&vdev->stats.rx_total_atomic, nr_xmit,
@@ -1184,6 +1167,36 @@ drain_mbuf_table(struct mbuf_table *tx_q)
}
 }
 
+uint16_t
+async_enqueue_pkts(struct vhost_dev *vdev, uint16_t queue_id,
+   struct rte_mbuf **pkts, uint32_t rx_count)
+{
+   uint16_t enqueue_count;
+   uint32_t cpu_cpl_nr = 0;
+   uint16_t enqueue_fail = 0;
+   struct rte_mbuf *m_cpu_cpl[MAX_PKT_BURST];
+
+   complete_async_pkts(vdev);
+   enqueue_count = rte_vhost_submit_enqueue_burst(vdev->vid,
+   queue_id, pkts, rx_count,
+   m_cpu_cpl, &cpu_cpl_nr);
+   if (cpu_cpl_nr)
+   free_pkts(m_cpu_cpl, cpu_cpl_nr);
+
+   enqueue_fail = rx_count - enqueue_count;
+   if (enqueue_fail)
+   free_pkts(&pkts[enqueue_count], enqueue_fail);
+
+   return enqueue_count;
+}
+
+uint16_t
+sync_enqueue_pkts(struct vhost_dev *vdev, uint16_t queue_id,
+   struct rte_mbuf **pkts, uint32_t rx_count)
+{
+   return rte_vhost_enqueue_burst(vdev->vid, queue_id, pkts, rx_count);
+}
+
 static __rte_always_inline void
 drain_eth_rx(struct vhost_dev *vdev)
 {
@@ -1214,29 +1227,8 @@ drain_eth_rx(struct vhost_dev *vdev)
}
}
 
-   if (builtin_net_driver) {
-   enqueue_count = vs_enqueue_pkts(vdev, VIRTIO_RXQ,
-   pkts, rx_count);
-   } else if (async_vhost_driver) {
-   uint32_t cpu_cpl_nr = 0;
-   uint16_t enqueue_fail = 0;
-   struct rte_mbuf *m_cpu_cpl[MAX_PKT_BURST];
-
-   complete_async_pkts(vdev);
-   enqueue_count = rte_vhost_submit_enqueue_burst(vdev->vid,
-   VIRTIO_RXQ, pkts, rx_count,
-   m_cpu_cpl, &cpu_cpl_nr);
-   if (cpu_cpl_nr)
-   free_pkts(m_cpu_cpl, cpu_cpl_nr);
-
-   enqueue_fail = rx_count - enqueue_count;
-   if (enqueue_fail)
-   free_pkts(&pkts[enqueue_count], enqueue_fail);
-
-   } else {
-   enqueue_count = rte_vhost_enqueue_burst(vdev->vid, VIRTIO_RXQ,
-   pkts, rx_count);
-   }
+   enqueue_count = vdev_queue_ops[vdev->vid].enqueue_pkt_burst(vdev,
+   VIRTIO_RXQ, pkts, rx_count);
 
if (enable_stats) {
__atomic_add_fetch(&vdev->stats.rx_total_atomic, rx_count,
@@ -1249,6 +1241,14 @@ drain_eth_rx(struct vhost_dev *vdev)
free_pkts(pkts, rx_count);
 }
 
+uint16_t sync_dequeue_pkts(struct vhost_dev *dev, uint16_t queue_id,
+   struct rte_mempool *mbuf_pool,
+   

[dpdk-dev] [PATCH v2 0/4] vhost: support async dequeue for split ring

2021-06-18 Thread Wenwu Ma
This patch implements asynchronous dequeue data path for split ring.
A new asynchronous dequeue function is introduced. With this function,
the application can try to receive packets from the guest with
offloading large copies to the DMA engine, thus saving precious CPU
cycles.

v2:
- Refactor vhost datapath as preliminary patch for this series.
- The change of using new API in examples/vhost is put into a
  dedicated patch.
- Check queue_id value before using it.
- Async dequeue performance enhancement. 160% performance improvement
  for v2 vs. v1.
- Async dequeue API name change from rte_vhost_try_dequeue_burst to
  rte_vhost_async_try_dequeue_burst.
- The completed package updates the used ring directly.

Wenwu Ma (3):
  examples/vhost: refactor vhost enqueue and dequeue datapaths.
  examples/vhost: use a new API to query remaining ring space
  examples/vhost: support vhost async dequeue data path

Yuan Wang (1):
  vhost: support async dequeue for split ring

 doc/guides/prog_guide/vhost_lib.rst |  10 +
 doc/guides/sample_app_ug/vhost.rst  |   9 +-
 examples/vhost/ioat.c   |  40 +-
 examples/vhost/ioat.h   |   4 +
 examples/vhost/main.c   | 225 +++
 examples/vhost/main.h   |  38 +-
 examples/vhost/virtio_net.c |  16 +-
 lib/vhost/rte_vhost_async.h |  44 ++-
 lib/vhost/version.map   |   3 +
 lib/vhost/virtio_net.c  | 553 
 10 files changed, 839 insertions(+), 103 deletions(-)

-- 
2.25.1



[dpdk-dev] [PATCH v2 2/4] examples/vhost: use a new API to query remaining ring space

2021-06-18 Thread Wenwu Ma
A new API for querying the remaining descriptor ring capacity
is available, so we use the new one instead of the old one.

Signed-off-by: Wenwu Ma 
---
 examples/vhost/ioat.c | 6 +-
 1 file changed, 1 insertion(+), 5 deletions(-)

diff --git a/examples/vhost/ioat.c b/examples/vhost/ioat.c
index 2a2c2d7202..bf4e033bdb 100644
--- a/examples/vhost/ioat.c
+++ b/examples/vhost/ioat.c
@@ -17,7 +17,6 @@ struct packet_tracker {
unsigned short next_read;
unsigned short next_write;
unsigned short last_remain;
-   unsigned short ioat_space;
 };
 
 struct packet_tracker cb_tracker[MAX_VHOST_DEVICE];
@@ -113,7 +112,6 @@ open_ioat(const char *value)
goto out;
}
rte_rawdev_start(dev_id);
-   cb_tracker[dev_id].ioat_space = IOAT_RING_SIZE - 1;
dma_info->nr++;
i++;
}
@@ -140,7 +138,7 @@ ioat_transfer_data_cb(int vid, uint16_t queue_id,
src = descs[i_desc].src;
dst = descs[i_desc].dst;
i_seg = 0;
-   if (cb_tracker[dev_id].ioat_space < src->nr_segs)
+   if (rte_ioat_burst_capacity(dev_id) < src->nr_segs)
break;
while (i_seg < src->nr_segs) {
rte_ioat_enqueue_copy(dev_id,
@@ -155,7 +153,6 @@ ioat_transfer_data_cb(int vid, uint16_t queue_id,
}
write &= mask;
cb_tracker[dev_id].size_track[write] = src->nr_segs;
-   cb_tracker[dev_id].ioat_space -= src->nr_segs;
write++;
}
} else {
@@ -194,7 +191,6 @@ ioat_check_completed_copies_cb(int vid, uint16_t queue_id,
if (n_seg == 0)
return 0;
 
-   cb_tracker[dev_id].ioat_space += n_seg;
n_seg += cb_tracker[dev_id].last_remain;
 
read = cb_tracker[dev_id].next_read;
-- 
2.25.1



[dpdk-dev] [PATCH v2 3/4] vhost: support async dequeue for split ring

2021-06-18 Thread Wenwu Ma
From: Yuan Wang 

This patch implements asynchronous dequeue data path for split ring.
A new asynchronous dequeue function is introduced. With this function,
the application can try to receive packets from the guest with
offloading large copies to the DMA engine, thus saving precious CPU
cycles.

Signed-off-by: Yuan Wang 
Signed-off-by: Jiayu Hu 
Signed-off-by: Wenwu Ma 
---
 doc/guides/prog_guide/vhost_lib.rst |  10 +
 lib/vhost/rte_vhost_async.h |  44 ++-
 lib/vhost/version.map   |   3 +
 lib/vhost/virtio_net.c  | 553 
 4 files changed, 607 insertions(+), 3 deletions(-)

diff --git a/doc/guides/prog_guide/vhost_lib.rst 
b/doc/guides/prog_guide/vhost_lib.rst
index d18fb98910..05c42c9b11 100644
--- a/doc/guides/prog_guide/vhost_lib.rst
+++ b/doc/guides/prog_guide/vhost_lib.rst
@@ -281,6 +281,16 @@ The following is an overview of some key Vhost API 
functions:
   Poll enqueue completion status from async data path. Completed packets
   are returned to applications through ``pkts``.
 
+* ``rte_vhost_async_try_dequeue_burst(vid, queue_id, mbuf_pool, pkts, count, 
nr_inflight)``
+
+  Try to receive packets from the guest with offloading large packets
+  to the DMA engine. Successfully dequeued packets are transfer
+  completed and returned in ``pkts``. But there may be other packets
+  that are sent from the guest but being transferred by the DMA engine,
+  called in-flight packets. This function will return in-flight packets
+  only after the DMA engine finishes transferring. The amount of
+  in-flight packets by now is returned in ``nr_inflight``.
+
 Vhost-user Implementations
 --
 
diff --git a/lib/vhost/rte_vhost_async.h b/lib/vhost/rte_vhost_async.h
index 6faa31f5ad..58019408f1 100644
--- a/lib/vhost/rte_vhost_async.h
+++ b/lib/vhost/rte_vhost_async.h
@@ -84,13 +84,21 @@ struct rte_vhost_async_channel_ops {
 };
 
 /**
- * inflight async packet information
+ * in-flight async packet information
  */
+struct async_nethdr {
+   struct virtio_net_hdr hdr;
+   bool valid;
+};
+
 struct async_inflight_info {
struct rte_mbuf *mbuf;
-   uint16_t descs; /* num of descs inflight */
+   union {
+   uint16_t descs; /* num of descs in-flight */
+   struct async_nethdr nethdr;
+   };
uint16_t nr_buffers; /* num of buffers inflight for packed ring */
-};
+} __rte_cache_aligned;
 
 /**
  *  dma channel feature bit definition
@@ -193,4 +201,34 @@ __rte_experimental
 uint16_t rte_vhost_poll_enqueue_completed(int vid, uint16_t queue_id,
struct rte_mbuf **pkts, uint16_t count);
 
+/**
+ * This function tries to receive packets from the guest with offloading
+ * large copies to the DMA engine. Successfully dequeued packets are
+ * transfer completed, either by the CPU or the DMA engine, and they are
+ * returned in "pkts". There may be other packets that are sent from
+ * the guest but being transferred by the DMA engine, called in-flight
+ * packets. The amount of in-flight packets by now is returned in
+ * "nr_inflight". This function will return in-flight packets only after
+ * the DMA engine finishes transferring.
+ *
+ * @param vid
+ *  id of vhost device to dequeue data
+ * @param queue_id
+ *  queue id to dequeue data
+ * @param pkts
+ *  blank array to keep successfully dequeued packets
+ * @param count
+ *  size of the packet array
+ * @param nr_inflight
+ *  the amount of in-flight packets by now. If error occurred, its
+ *  value is set to -1.
+ * @return
+ *  num of successfully dequeued packets
+ */
+__rte_experimental
+uint16_t
+rte_vhost_async_try_dequeue_burst(int vid, uint16_t queue_id,
+   struct rte_mempool *mbuf_pool, struct rte_mbuf **pkts, uint16_t count,
+   int *nr_inflight);
+
 #endif /* _RTE_VHOST_ASYNC_H_ */
diff --git a/lib/vhost/version.map b/lib/vhost/version.map
index 9103a23cd4..a320f889cd 100644
--- a/lib/vhost/version.map
+++ b/lib/vhost/version.map
@@ -79,4 +79,7 @@ EXPERIMENTAL {
 
# added in 21.05
rte_vhost_get_negotiated_protocol_features;
+
+   # added in 21.08
+   rte_vhost_async_try_dequeue_burst;
 };
diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index 8da8a86a10..28e54df804 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -3165,3 +3165,556 @@ rte_vhost_dequeue_burst(int vid, uint16_t queue_id,
 
return count;
 }
+
+static __rte_always_inline int
+async_desc_to_mbuf(struct virtio_net *dev, struct vhost_virtqueue *vq,
+ struct buf_vector *buf_vec, uint16_t nr_vec,
+ struct rte_mbuf *m, struct rte_mempool *mbuf_pool,
+ struct iovec *src_iovec, struct iovec *dst_iovec,
+ struct rte_vhost_iov_iter *src_it,
+ struct rte_vhost_iov_iter *dst_it,
+ struct async_nethdr *nethdr,
+ bool legacy_ol_flags)
+{
+   uint64_t buf_addr;
+   uint32_t tlen = 

[dpdk-dev] [PATCH v2 4/4] examples/vhost: support vhost async dequeue data path

2021-06-18 Thread Wenwu Ma
This patch is to add vhost async dequeue data-path in vhost sample.
vswitch can leverage IOAT to accelerate vhost async dequeue data-path.

Signed-off-by: Wenwu Ma 
---
 doc/guides/sample_app_ug/vhost.rst |   9 +-
 examples/vhost/ioat.c  |  34 +--
 examples/vhost/ioat.h  |   4 +
 examples/vhost/main.c  | 141 -
 examples/vhost/main.h  |   5 +
 5 files changed, 140 insertions(+), 53 deletions(-)

diff --git a/doc/guides/sample_app_ug/vhost.rst 
b/doc/guides/sample_app_ug/vhost.rst
index 9afde9c7f5..63dcf181e1 100644
--- a/doc/guides/sample_app_ug/vhost.rst
+++ b/doc/guides/sample_app_ug/vhost.rst
@@ -169,9 +169,12 @@ demonstrates how to use the async vhost APIs. It's used in 
combination with dmas
 **--dmas**
 This parameter is used to specify the assigned DMA device of a vhost device.
 Async vhost-user net driver will be used if --dmas is set. For example
---dmas [txd0@00:04.0,txd1@00:04.1] means use DMA channel 00:04.0 for vhost
-device 0 enqueue operation and use DMA channel 00:04.1 for vhost device 1
-enqueue operation.
+--dmas [txd0@00:04.0,txd1@00:04.1,rxd0@00:04.2,rxd1@00:04.3] means use
+DMA channel 00:04.0/00:04.2 for vhost device 0 enqueue/dequeue operation
+and use DMA channel 00:04.1/00:04.3 for vhost device 1 enqueue/dequeue
+operation. The index of the device corresponds to the socket file in order,
+that means vhost device 0 is created through the first socket file, vhost
+device 1 is created through the second socket file, and so on.
 
 Common Issues
 -
diff --git a/examples/vhost/ioat.c b/examples/vhost/ioat.c
index bf4e033bdb..179ae87deb 100644
--- a/examples/vhost/ioat.c
+++ b/examples/vhost/ioat.c
@@ -21,6 +21,8 @@ struct packet_tracker {
 
 struct packet_tracker cb_tracker[MAX_VHOST_DEVICE];
 
+int vid2txd[MAX_VHOST_DEVICE];
+
 int
 open_ioat(const char *value)
 {
@@ -60,6 +62,8 @@ open_ioat(const char *value)
goto out;
}
while (i < args_nr) {
+   char *txd, *rxd;
+   bool is_txd;
char *arg_temp = dma_arg[i];
uint8_t sub_nr;
sub_nr = rte_strsplit(arg_temp, strlen(arg_temp), ptrs, 2, '@');
@@ -68,10 +72,20 @@ open_ioat(const char *value)
goto out;
}
 
-   start = strstr(ptrs[0], "txd");
-   if (start == NULL) {
+   int async_flag;
+   txd = strstr(ptrs[0], "txd");
+   rxd = strstr(ptrs[0], "rxd");
+   if (txd == NULL && rxd == NULL) {
ret = -1;
goto out;
+   } else if (txd) {
+   is_txd = true;
+   start = txd;
+   async_flag = ASYNC_RX_VHOST;
+   } else {
+   is_txd = false;
+   start = rxd;
+   async_flag = ASYNC_TX_VHOST;
}
 
start += 3;
@@ -81,7 +95,8 @@ open_ioat(const char *value)
goto out;
}
 
-   vring_id = 0 + VIRTIO_RXQ;
+   vring_id = is_txd ? VIRTIO_RXQ : VIRTIO_TXQ;
+
if (rte_pci_addr_parse(ptrs[1],
&(dma_info + vid)->dmas[vring_id].addr) < 0) {
ret = -1;
@@ -105,6 +120,7 @@ open_ioat(const char *value)
 
(dma_info + vid)->dmas[vring_id].dev_id = dev_id;
(dma_info + vid)->dmas[vring_id].is_valid = true;
+   (dma_info + vid)->async_flag |= async_flag;
config.ring_size = IOAT_RING_SIZE;
config.hdls_disable = true;
if (rte_rawdev_configure(dev_id, &info, sizeof(config)) < 0) {
@@ -126,13 +142,16 @@ ioat_transfer_data_cb(int vid, uint16_t queue_id,
struct rte_vhost_async_status *opaque_data, uint16_t count)
 {
uint32_t i_desc;
-   uint16_t dev_id = dma_bind[vid].dmas[queue_id * 2 + VIRTIO_RXQ].dev_id;
struct rte_vhost_iov_iter *src = NULL;
struct rte_vhost_iov_iter *dst = NULL;
unsigned long i_seg;
unsigned short mask = MAX_ENQUEUED_SIZE - 1;
-   unsigned short write = cb_tracker[dev_id].next_write;
 
+   if (queue_id >= MAX_RING_COUNT)
+   return -1;
+
+   uint16_t dev_id = dma_bind[vid2txd[vid]].dmas[queue_id].dev_id;
+   unsigned short write = cb_tracker[dev_id].next_write;
if (!opaque_data) {
for (i_desc = 0; i_desc < count; i_desc++) {
src = descs[i_desc].src;
@@ -170,7 +189,7 @@ ioat_check_completed_copies_cb(int vid, uint16_t queue_id,
struct rte_vhost_async_status *opaque_data,
uint16_t max_packets)
 {
-   if (!opaque_data) {
+   if (!opaque_data && (queue_id < MAX_RING_COUNT)) {
uintptr_t dump[255];
int n_seg;
   

Re: [dpdk-dev] [PATCH v4 4/7] vhost: fix NUMA reallocation with multiqueue

2021-06-18 Thread Xia, Chenbo
Hi Maxime,

> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, June 18, 2021 4:01 PM
> To: Xia, Chenbo ; dev@dpdk.org;
> david.march...@redhat.com
> Cc: sta...@dpdk.org
> Subject: Re: [PATCH v4 4/7] vhost: fix NUMA reallocation with multiqueue
> 
> 
> 
> On 6/18/21 6:34 AM, Xia, Chenbo wrote:
> > Hi Maxime,
> >
> >> -Original Message-
> >> From: Maxime Coquelin 
> >> Sent: Thursday, June 17, 2021 11:38 PM
> >> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> 
> >> Cc: Maxime Coquelin ; sta...@dpdk.org
> >> Subject: [PATCH v4 4/7] vhost: fix NUMA reallocation with multiqueue
> >>
> >> Since the Vhost-user device initialization has been reworked,
> >> enabling the application to start using the device as soon as
> >> the first queue pair is ready, NUMA reallocation no more
> >> happened on queue pairs other than the first one since
> >> numa_realloc() was returning early if the device was running.
> >>
> >> This patch fixes this issue by only preventing the device
> >> metadata to be allocated if the device is running. For the
> >> virtqueues, a vring state change notification is sent to
> >> notify the application of its disablement. Since the callback
> >> is supposed to be blocking, it is safe to reallocate it
> >> afterwards.
> >
> > Is there a corner case? Numa_realloc may happen during vhost-user msg
> > set_vring_addr/kick, set_mem_table and iotlb msg. And iotlb msg will
> > not take vq access lock. It could happen when numa_realloc happens on
> > iotlb msg and app accesses vq in the meantime?
> 
> I think we are safe wrt to numa_realloc(), because the app's
> .vring_state_changed() callback is only returning when it is no more
> processing the rings.

Yes, I think it should be. But in this iotlb msg case (take vhost pmd for 
example),
can't vhost pmd still access vq since vq access lock is not took? Do I miss 
something?

Thanks,
Chenbo

> 
> 
> > Thanks,
> > Chenbo
> >
> >>
> >> Fixes: d0fcc38f5fa4 ("vhost: improve device readiness notifications")
> >> Cc: sta...@dpdk.org
> >>
> >> Signed-off-by: Maxime Coquelin 
> >> ---
> >>  lib/vhost/vhost_user.c | 11 ---
> >>  1 file changed, 8 insertions(+), 3 deletions(-)
> >>
> >> diff --git a/lib/vhost/vhost_user.c b/lib/vhost/vhost_user.c
> >> index 0e9e26ebe0..6e7b327ef8 100644
> >> --- a/lib/vhost/vhost_user.c
> >> +++ b/lib/vhost/vhost_user.c
> >> @@ -488,9 +488,6 @@ numa_realloc(struct virtio_net *dev, int index)
> >>struct batch_copy_elem *new_batch_copy_elems;
> >>int ret;
> >>
> >> -  if (dev->flags & VIRTIO_DEV_RUNNING)
> >> -  return dev;
> >> -
> >>old_dev = dev;
> >>vq = old_vq = dev->virtqueue[index];
> >>
> >> @@ -506,6 +503,11 @@ numa_realloc(struct virtio_net *dev, int index)
> >>return dev;
> >>}
> >>if (oldnode != newnode) {
> >> +  if (vq->ready) {
> >> +  vq->ready = false;
> >> +  vhost_user_notify_queue_state(dev, index, 0);
> >> +  }
> >> +
> >>VHOST_LOG_CONFIG(INFO,
> >>"reallocate vq from %d to %d node\n", oldnode, newnode);
> >>vq = rte_malloc_socket(NULL, sizeof(*vq), 0, newnode);
> >> @@ -558,6 +560,9 @@ numa_realloc(struct virtio_net *dev, int index)
> >>rte_free(old_vq);
> >>}
> >>
> >> +  if (dev->flags & VIRTIO_DEV_RUNNING)
> >> +  goto out;
> >> +
> >>/* check if we need to reallocate dev */
> >>ret = get_mempolicy(&oldnode, NULL, 0, old_dev,
> >>MPOL_F_NODE | MPOL_F_ADDR);
> >> --
> >> 2.31.1
> >



[dpdk-dev] [PATCH] net/memif: fix abstract socket addr_len

2021-06-18 Thread Nathan Skrzypczak
This fixes using abstract sockets with memifs.
we were not passing the exact addr_len, which
requires zeroing the remaining sun_path and
doesn't appear well in other utilities (e.g.
lsof -U)

Signed-off-by: Nathan Skrzypczak 
---
 drivers/net/memif/memif_socket.c | 13 ++---
 1 file changed, 10 insertions(+), 3 deletions(-)

diff --git a/drivers/net/memif/memif_socket.c b/drivers/net/memif/memif_socket.c
index 5b373738e6..7498a2f58d 100644
--- a/drivers/net/memif/memif_socket.c
+++ b/drivers/net/memif/memif_socket.c
@@ -866,6 +866,7 @@ memif_socket_create(char *key, uint8_t listener, bool 
is_abstract)
 {
struct memif_socket *sock;
struct sockaddr_un un = { 0 };
+   uint32_t sunlen;
int sockfd;
int ret;
int on = 1;
@@ -890,7 +891,10 @@ memif_socket_create(char *key, uint8_t listener, bool 
is_abstract)
/* abstract address */
un.sun_path[0] = '\0';
strlcpy(un.sun_path + 1, sock->filename, 
MEMIF_SOCKET_UN_SIZE - 1);
+   sunlen = 1 + strlen(sock->filename) +
+sizeof(un.sun_family);
} else {
+   sunlen = sizeof(un);
strlcpy(un.sun_path, sock->filename, 
MEMIF_SOCKET_UN_SIZE);
}
 
@@ -899,7 +903,7 @@ memif_socket_create(char *key, uint8_t listener, bool 
is_abstract)
if (ret < 0)
goto error;
 
-   ret = bind(sockfd, (struct sockaddr *)&un, sizeof(un));
+   ret = bind(sockfd, (struct sockaddr *)&un, sunlen);
if (ret < 0)
goto error;
 
@@ -1061,6 +1065,7 @@ memif_connect_client(struct rte_eth_dev *dev)
 {
int sockfd;
int ret;
+   uint32_t sunlen;
struct sockaddr_un sun = { 0 };
struct pmd_internals *pmd = dev->data->dev_private;
 
@@ -1075,16 +1080,18 @@ memif_connect_client(struct rte_eth_dev *dev)
}
 
sun.sun_family = AF_UNIX;
+   sunlen = sizeof(struct sockaddr_un);
if (pmd->flags & ETH_MEMIF_FLAG_SOCKET_ABSTRACT) {
/* abstract address */
sun.sun_path[0] = '\0';
strlcpy(sun.sun_path + 1,  pmd->socket_filename, 
MEMIF_SOCKET_UN_SIZE - 1);
+   sunlen = strlen(pmd->socket_filename) + 1 +
+sizeof(sun.sun_family);
} else {
strlcpy(sun.sun_path,  pmd->socket_filename, 
MEMIF_SOCKET_UN_SIZE);
}
 
-   ret = connect(sockfd, (struct sockaddr *)&sun,
- sizeof(struct sockaddr_un));
+   ret = connect(sockfd, (struct sockaddr *)&sun, sunlen);
if (ret < 0) {
MIF_LOG(ERR, "Failed to connect socket: %s.", 
pmd->socket_filename);
goto error;
-- 
2.30.2



Re: [dpdk-dev] [PATCH] Add Armv7 support

2021-06-18 Thread Supriya Shekhar Mane
Hi Thomas,

Here are the meson/ninja build results, we see no warnings here:

supriya@supriya:~/dpdk-meson-openwrt/dpdk-21.05$ meson armv7-build
--cross-file config/arm/arm_armv7a_linux_gcc
The Meson build system
Version: 0.58.0
Source dir: /home/supriya/dpdk-meson-openwrt/dpdk-21.05
Build dir: /home/supriya/dpdk-meson-openwrt/dpdk-21.05/armv7-build
Build type: cross build
Program cat found: YES (/bin/cat)
Project name: DPDK
Project version: 21.05.0
C compiler for the host machine: arm-none-linux-gnueabihf-gcc (gcc
10.2.1 "arm-none-linux-gnueabihf-gcc (GNU Toolchain for the A-profile
Architecture 10.2-2020.11 (arm-10.16)) 10.2.1 20201103")
C linker for the host machine: arm-none-linux-gnueabihf-gcc ld.bfd 10.2-2020
C compiler for the build machine: ccache cc (gcc 7.5.0 "cc (Ubuntu
7.5.0-3ubuntu1~18.04) 7.5.0")
C linker for the build machine: cc ld.bfd 2.30
Build machine cpu family: x86_64
Build machine cpu: x86_64
Host machine cpu family: arm
Host machine cpu: armv7-a
Target machine cpu family: arm
Target machine cpu: armv7-a
Program pkg-config found: YES (/usr/bin/pkg-config)
Program check-symbols.sh found: YES
(/home/supriya/dpdk-meson-openwrt/dpdk-21.05/buildtools/check-symbols.sh)
Program options-ibverbs-static.sh found: YES
(/home/supriya/dpdk-meson-openwrt/dpdk-21.05/buildtools/options-ibverbs-static.sh)
Program binutils-avx512-check.sh found: YES
(/home/supriya/dpdk-meson-openwrt/dpdk-21.05/buildtools/binutils-avx512-check.sh)
Program python3 found: YES (/usr/bin/python3)
Program cat found: YES (/bin/cat)
Program ../buildtools/symlink-drivers-solibs.sh found: YES (/bin/sh
/home/supriya/dpdk-meson-openwrt/dpdk-21.05/config/../buildtools/symlink-drivers-solibs.sh)
Checking for size of "void *" : 4
Checking for size of "void *" : 4
Library m found: YES
Library numa found: NO
Library libfdt found: NO
Library libexecinfo found: NO
Found pkg-config: /usr/bin/arm-linux-gnueabihf-pkg-config (0.29.1)
Run-time dependency libbsd found: NO (tried pkgconfig)
Run-time dependency libpcap found: NO (tried pkgconfig)
Library pcap found: NO
Compiler for C supports arguments -Wextra: YES
config/meson.build:245: WARNING: Consider using the built-in
warning_level option instead of using "-Wextra".
Compiler for C supports arguments -Wcast-qual: YES
Compiler for C supports arguments -Wdeprecated: YES
Compiler for C supports arguments -Wformat: YES
Compiler for C supports arguments -Wformat-nonliteral: YES
Compiler for C supports arguments -Wformat-security: YES
Compiler for C supports arguments -Wmissing-declarations: YES
Compiler for C supports arguments -Wmissing-prototypes: YES
Compiler for C supports arguments -Wnested-externs: YES
Compiler for C supports arguments -Wold-style-definition: YES
Compiler for C supports arguments -Wpointer-arith: YES
Compiler for C supports arguments -Wsign-compare: YES
Compiler for C supports arguments -Wstrict-prototypes: YES
Compiler for C supports arguments -Wundef: YES
Compiler for C supports arguments -Wwrite-strings: YES
Compiler for C supports arguments -Wno-address-of-packed-member: YES
Compiler for C supports arguments -Wno-packed-not-aligned: YES
Compiler for C supports arguments -Wno-missing-field-initializers: YES
Compiler for C supports arguments -Wno-zero-length-bounds: YES
Compiler for C supports arguments -Wno-pointer-to-int-cast: YES
Message: Using machine args: ['-march=armv7-a', '-mfpu=neon']
Fetching value of define "__ARM_NEON" : 1
Fetching value of define "__ARM_FEATURE_CRC32" :
Fetching value of define "__ARM_FEATURE_CRYPTO" :
Compiler for C supports arguments -Wno-format-truncation: YES
Checking for function "getentropy" : YES
Run-time dependency jansson found: NO (tried pkgconfig)
Has header "linux/userfaultfd.h" : YES
Run-time dependency libelf found: NO (tried pkgconfig)
Compiler for C supports arguments -Wno-format-truncation: YES (cached)
Compiler for C supports arguments -Wno-cast-qual: YES
Compiler for C supports arguments -Wno-pointer-arith: YES
Compiler for C supports arguments -Wno-pointer-to-int-cast: YES (cached)
Run-time dependency libmusdk found: NO (tried pkgconfig)
Compiler for C supports arguments -Wno-cast-qual: YES (cached)
Compiler for C supports arguments -Wno-pointer-arith: YES (cached)
Compiler for C supports arguments -std=c11: YES
Compiler for C supports arguments -Wno-strict-prototypes: YES
Compiler for C supports arguments -D_BSD_SOURCE: YES
Compiler for C supports arguments -D_DEFAULT_SOURCE: YES
Compiler for C supports arguments -D_XOPEN_SOURCE=600: YES
Run-time dependency libmlx5 found: NO (tried pkgconfig)
Library mlx5 found: NO
Run-time dependency libcrypto found: NO (tried pkgconfig)
Compiler for C supports arguments -Wdisabled-optimization: YES
Compiler for C supports arguments -Waggregate-return: YES
Compiler for C supports arguments -Wbad-function-cast: YES
Compiler for C supports arguments -Wno-sign-compare: YES
Compiler for C supports arguments -Wno-unused-parameter: YES
Compiler for C supports arguments -Wno-unused-variable: 

[dpdk-dev] Regarding Armv7 support in DPDK-21.05

2021-06-18 Thread Supriya Shekhar Mane
Hi Thomas,

The checklist shows a functional test error(as shown in attachment),
but the error is in X86,
how to reproduce and fix this issue?

Regards,
Supriya Mane

-- 
Disclaimer:This message is intended only for the designated recipient(s). 
It may contain confidential or proprietary information and may be subject 
to other confidentiality protections. If you are not a designated 
recipient, you may not review, copy or distribute this message. Please 
notify the sender by e-mail and delete this message. GlobalEdge does not 
accept any liability for virus infected mails.



Re: [dpdk-dev] [PATCH v3] app/testpmd: send failure logs to stderr

2021-06-18 Thread Li, Xiaoyun
> -Original Message-
> From: Andrew Rybchenko 
> Sent: Thursday, June 17, 2021 22:20
> To: Li, Xiaoyun ; Ori Kam 
> Cc: dev@dpdk.org; Richardson, Bruce ; Yigit,
> Ferruh ; Singh, Aman Deep
> 
> Subject: [PATCH v3] app/testpmd: send failure logs to stderr
> 
> Running with stdout suppressed or redirected for further processing
> is very confusing in the case of errors. Fix it by logging errors and
> warnings to stderr.
> 
> Since lines with log messages are touched anyway concatenate split
> format strings to make it easier to search using grep.
> 
> Fix indent of format string arguments.
> 
> Signed-off-by: Andrew Rybchenko 
> ---
> v3:
>  - fix typo in the patch description
>  - fix applicable coding style issues reported by patchwork
> 
> v2:
>  - switch from printf() to fpritnf(stderr, ...) in more cases
>  - do not inherit acks from the previous version since the patch is
>much bigger
>  - fix style in few cases (TAB vs spaces, missing space separtor etc)
>  - still don't use TESTPMD_LOG() since the patch does not add new logs.
>Also switching to TESTPMD_LOG() will add "testpmd: " prefix to log
>messages and it is a real change and could be a pain for automation.
> 
>  app/test-pmd/bpf_cmd.c |   6 +-
>  app/test-pmd/cmdline.c | 957 ++---
>  app/test-pmd/cmdline_flow.c|  20 +-
>  app/test-pmd/cmdline_mtr.c |   8 +-
>  app/test-pmd/cmdline_tm.c  |  33 +-
>  app/test-pmd/config.c  | 455 ++--
>  app/test-pmd/csumonly.c|   5 +-
>  app/test-pmd/parameters.c  |  21 +-
>  app/test-pmd/testpmd.c | 298 
>  app/test-pmd/util.c|  19 +-
>  doc/guides/rel_notes/release_21_08.rst |   5 +
>  11 files changed, 1004 insertions(+), 823 deletions(-)

Acked-by: Xiaoyun Li 


Re: [dpdk-dev] Regarding Armv7 support in DPDK-21.05

2021-06-18 Thread David Marchand
Hello Supriya,

I waived 2/3 mails from you in the ml moderation list.
Please register to the dev@ mailing list if you wish to contribute to
the project.


On Fri, Jun 18, 2021 at 10:27 AM Supriya Shekhar Mane
 wrote:
>
> Hi Thomas,
>
> The checklist shows a functional test error(as shown in attachment),
> but the error is in X86,
> how to reproduce and fix this issue?

In this specific case, those failures come from CI issues itself.
You can ignore them.


>
> Regards,
> Supriya Mane
>
> --
> Disclaimer:This message is intended only for the designated recipient(s).
> It may contain confidential or proprietary information and may be subject
> to other confidentiality protections. If you are not a designated
> recipient, you may not review, copy or distribute this message. Please
> notify the sender by e-mail and delete this message. GlobalEdge does not
> accept any liability for virus infected mails.

This trailer does not make sense on a public mailing list.
Please ask your IT to remove it.


Thanks!

-- 
David Marchand



[dpdk-dev] [PATCH] net/i40e: fix L2 payload RSS mask input set

2021-06-18 Thread Alvin Zhang
Allow VLAN tag being added to L2 payload packet type RSS input set.

Fixes: ef4c16fd9148 ("net/i40e: refactor RSS flow")
Cc: sta...@dpdk.org

Signed-off-by: Alvin Zhang 
---
 drivers/net/i40e/i40e_hash.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/drivers/net/i40e/i40e_hash.c b/drivers/net/i40e/i40e_hash.c
index b1cb24f..722edc9 100644
--- a/drivers/net/i40e/i40e_hash.c
+++ b/drivers/net/i40e/i40e_hash.c
@@ -201,11 +201,11 @@ struct i40e_hash_match_pattern {
 #define I40E_HASH_MAP_CUS_PATTERN(pattern, rss_mask, cus_pctype) { \
pattern, rss_mask, true, cus_pctype }
 
-#define I40E_HASH_L2_RSS_MASK  (ETH_RSS_ETH | ETH_RSS_L2_SRC_ONLY | \
+#define I40E_HASH_L2_RSS_MASK  (ETH_RSS_VLAN | ETH_RSS_ETH | \
+   ETH_RSS_L2_SRC_ONLY | \
ETH_RSS_L2_DST_ONLY)
 
 #define I40E_HASH_L23_RSS_MASK (I40E_HASH_L2_RSS_MASK | \
-   ETH_RSS_VLAN | \
ETH_RSS_L3_SRC_ONLY | \
ETH_RSS_L3_DST_ONLY)
 
-- 
1.8.3.1



Re: [dpdk-dev] [PATCH v4 4/7] vhost: fix NUMA reallocation with multiqueue

2021-06-18 Thread Maxime Coquelin



On 6/18/21 10:21 AM, Xia, Chenbo wrote:
> Hi Maxime,
> 
>> -Original Message-
>> From: Maxime Coquelin 
>> Sent: Friday, June 18, 2021 4:01 PM
>> To: Xia, Chenbo ; dev@dpdk.org;
>> david.march...@redhat.com
>> Cc: sta...@dpdk.org
>> Subject: Re: [PATCH v4 4/7] vhost: fix NUMA reallocation with multiqueue
>>
>>
>>
>> On 6/18/21 6:34 AM, Xia, Chenbo wrote:
>>> Hi Maxime,
>>>
 -Original Message-
 From: Maxime Coquelin 
 Sent: Thursday, June 17, 2021 11:38 PM
 To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
>> 
 Cc: Maxime Coquelin ; sta...@dpdk.org
 Subject: [PATCH v4 4/7] vhost: fix NUMA reallocation with multiqueue

 Since the Vhost-user device initialization has been reworked,
 enabling the application to start using the device as soon as
 the first queue pair is ready, NUMA reallocation no more
 happened on queue pairs other than the first one since
 numa_realloc() was returning early if the device was running.

 This patch fixes this issue by only preventing the device
 metadata to be allocated if the device is running. For the
 virtqueues, a vring state change notification is sent to
 notify the application of its disablement. Since the callback
 is supposed to be blocking, it is safe to reallocate it
 afterwards.
>>>
>>> Is there a corner case? Numa_realloc may happen during vhost-user msg
>>> set_vring_addr/kick, set_mem_table and iotlb msg. And iotlb msg will
>>> not take vq access lock. It could happen when numa_realloc happens on
>>> iotlb msg and app accesses vq in the meantime?
>>
>> I think we are safe wrt to numa_realloc(), because the app's
>> .vring_state_changed() callback is only returning when it is no more
>> processing the rings.
> 
> Yes, I think it should be. But in this iotlb msg case (take vhost pmd for 
> example),
> can't vhost pmd still access vq since vq access lock is not took? Do I miss 
> something?

Vhost PMD sends RTE_ETH_EVENT_QUEUE_STATE, and my assumption was that
the application would stop processing the rings when handling this
event and only return from the callback when it's one, but this seems
that's not done at least in testpmd. So we may not rely on that after
all :/.

We cannot rely on the VQ's access lock since the goal of numa_realloc is
to reallocate the vhost_virtqueue itself which contains the acces_lock.
Relying on it would cause a use after free.

Maybe the safest thing to do is to just skip the reallocation if
vq->ready == true.

Having vq->ready == true means we already received all the vrings info
from QEMU, which means the driver has already initialized the device.

It should not change runtime behavior compared to this patch since it
would not reallocate anyway.

What do you think?

> Thanks,
> Chenbo
> 
>>
>>
>>> Thanks,
>>> Chenbo
>>>

 Fixes: d0fcc38f5fa4 ("vhost: improve device readiness notifications")
 Cc: sta...@dpdk.org

 Signed-off-by: Maxime Coquelin 
 ---
  lib/vhost/vhost_user.c | 11 ---
  1 file changed, 8 insertions(+), 3 deletions(-)

 diff --git a/lib/vhost/vhost_user.c b/lib/vhost/vhost_user.c
 index 0e9e26ebe0..6e7b327ef8 100644
 --- a/lib/vhost/vhost_user.c
 +++ b/lib/vhost/vhost_user.c
 @@ -488,9 +488,6 @@ numa_realloc(struct virtio_net *dev, int index)
struct batch_copy_elem *new_batch_copy_elems;
int ret;

 -  if (dev->flags & VIRTIO_DEV_RUNNING)
 -  return dev;
 -
old_dev = dev;
vq = old_vq = dev->virtqueue[index];

 @@ -506,6 +503,11 @@ numa_realloc(struct virtio_net *dev, int index)
return dev;
}
if (oldnode != newnode) {
 +  if (vq->ready) {
 +  vq->ready = false;
 +  vhost_user_notify_queue_state(dev, index, 0);
 +  }
 +
VHOST_LOG_CONFIG(INFO,
"reallocate vq from %d to %d node\n", oldnode, newnode);
vq = rte_malloc_socket(NULL, sizeof(*vq), 0, newnode);
 @@ -558,6 +560,9 @@ numa_realloc(struct virtio_net *dev, int index)
rte_free(old_vq);
}

 +  if (dev->flags & VIRTIO_DEV_RUNNING)
 +  goto out;
 +
/* check if we need to reallocate dev */
ret = get_mempolicy(&oldnode, NULL, 0, old_dev,
MPOL_F_NODE | MPOL_F_ADDR);
 --
 2.31.1
>>>
> 



Re: [dpdk-dev] [PATCH V6] config/arm: add Qualcomm Centriq 2400 part number

2021-06-18 Thread Thomas Monjalon
18/06/2021 04:09, Ruifeng Wang:
> From: Thierry Herbelot 
> >  'part_number_config': {
> > -'0xc00': {'machine_args':  ['-march=armv8-a+crc']}
> > +'0x800': {'machine_args':  ['-march=armv8-a+crc']},
> > +'0xc00': {'machine_args':  ['-march=armv8-a+crc']},
> Nit, redundant comma at the end of the line.

What is redundant?





Re: [dpdk-dev] [RFC PATCH] dmadev: introduce DMA device library

2021-06-18 Thread fengchengwen
On 2021/6/17 22:18, Bruce Richardson wrote:
> On Thu, Jun 17, 2021 at 12:02:00PM +0100, Bruce Richardson wrote:
>> On Thu, Jun 17, 2021 at 05:48:05PM +0800, fengchengwen wrote:
>>> On 2021/6/17 1:31, Bruce Richardson wrote:
 On Wed, Jun 16, 2021 at 05:41:45PM +0800, fengchengwen wrote:
> On 2021/6/16 0:38, Bruce Richardson wrote:
>> On Tue, Jun 15, 2021 at 09:22:07PM +0800, Chengwen Feng wrote:
>>> This patch introduces 'dmadevice' which is a generic type of DMA
>>> device.
>>>
>>> The APIs of dmadev library exposes some generic operations which can
>>> enable configuration and I/O with the DMA devices.
>>>
>>> Signed-off-by: Chengwen Feng 
>>> ---
>> Thanks for sending this.
>>
>> Of most interest to me right now are the key data-plane APIs. While we 
>> are
>> still in the prototyping phase, below is a draft of what we are thinking
>> for the key enqueue/perform_ops/completed_ops APIs.
>>
>> Some key differences I note in below vs your original RFC:
>> * Use of void pointers rather than iova addresses. While using iova's 
>> makes
>>   sense in the general case when using hardware, in that it can work with
>>   both physical addresses and virtual addresses, if we change the APIs 
>> to use
>>   void pointers instead it will still work for DPDK in VA mode, while at 
>> the
>>   same time allow use of software fallbacks in error cases, and also a 
>> stub
>>   driver than uses memcpy in the background. Finally, using iova's makes 
>> the
>>   APIs a lot more awkward to use with anything but mbufs or similar 
>> buffers
>>   where we already have a pre-computed physical address.
>
> The iova is an hint to application, and widely used in DPDK.
> If switch to void, how to pass the address (iova or just va ?)
> this may introduce implementation dependencies here.
>
> Or always pass the va, and the driver performs address translation, and 
> this
> translation may cost too much cpu I think.
>

 On the latter point, about driver doing address translation I would agree.
 However, we probably need more discussion about the use of iova vs just
 virtual addresses. My thinking on this is that if we specify the API using
 iovas it will severely hurt usability of the API, since it forces the user
 to take more inefficient codepaths in a large number of cases. Given a
 pointer to the middle of an mbuf, one cannot just pass that straight as an
 iova but must instead do a translation into offset from mbuf pointer and
 then readd the offset to the mbuf base address.

 My preference therefore is to require the use of an IOMMU when using a
 dmadev, so that it can be a much closer analog of memcpy. Once an iommu is
 present, DPDK will run in VA mode, allowing virtual addresses to our
 hugepage memory to be sent directly to hardware. Also, when using
 dmadevs on top of an in-kernel driver, that kernel driver may do all iommu
 management for the app, removing further the restrictions on what memory
 can be addressed by hardware.
>>>
>>> Some DMA devices many don't support IOMMU or IOMMU bypass default, so 
>>> driver may
>>> should call rte_mem_virt2phy() do the address translate, but the 
>>> rte_mem_virt2phy()
>>> cost too many CPU cycles.
>>>
>>> If the API defined as iova, it will work fine in:
>>> 1) If DMA don't support IOMMU or IOMMU bypass, then start application with
>>>--iova-mode=pa
>>> 2) If DMA support IOMMU, --iova-mode=pa/va work both fine
>>>
>>
>> I suppose if we keep the iova as the datatype, we can just cast "void *"
>> pointers to that in the case that virtual addresses can be used directly. I
>> believe your RFC included a capability query API - "uses void * as iova" 
>> should probably be one of those capabilities, and that would resolve this.
>> If DPDK is in iova=va mode because of the presence of an iommu, all drivers
>> could report this capability too.
>>

>> * Use of id values rather than user-provided handles. Allowing the 
>> user/app
>>   to manage the amount of data stored per operation is a better 
>> solution, I
>>   feel than proscribing a certain about of in-driver tracking. Some apps 
>> may
>>   not care about anything other than a job being completed, while other 
>> apps
>>   may have significant metadata to be tracked. Taking the user-context
>>   handles out of the API also makes the driver code simpler.
>
> The user-provided handle was mainly used to simply application 
> implementation,
> It provides the ability to quickly locate contexts.
>
> The "use of id values" seem like the dma_cookie of Linux DMA engine 
> framework,
> user will get a unique dma_cookie after calling dmaengine_submit(), and 
> then
> could use it to call dma_async_is_tx_complete() to get comple

Re: [dpdk-dev] [PATCH V6] config/arm: add Qualcomm Centriq 2400 part number

2021-06-18 Thread Thierry Herbelot

On 6/18/21 10:51 AM, Thomas Monjalon wrote:

18/06/2021 04:09, Ruifeng Wang:

From: Thierry Herbelot 

  'part_number_config': {
-'0xc00': {'machine_args':  ['-march=armv8-a+crc']}
+'0x800': {'machine_args':  ['-march=armv8-a+crc']},
+'0xc00': {'machine_args':  ['-march=armv8-a+crc']},

Nit, redundant comma at the end of the line.


What is redundant?


The comma at the end of the second line is not necessary.

Thierry

--
Thierry Herbelot
Senior Software Engineer
Tel: +33 1 39 30 92 61
http://www.6wind.com/

Follow us:
https://www.linkedin.com/company/6wind/
https://twitter.com/6WINDsoftware
https://www.youtube.com/user/6windsoftware


Re: [dpdk-dev] [PATCH V6] config/arm: add Qualcomm Centriq 2400 part number

2021-06-18 Thread Thomas Monjalon
18/06/2021 10:53, Thierry Herbelot:
> On 6/18/21 10:51 AM, Thomas Monjalon wrote:
> > 18/06/2021 04:09, Ruifeng Wang:
> >> From: Thierry Herbelot 
> >>>   'part_number_config': {
> >>> -'0xc00': {'machine_args':  ['-march=armv8-a+crc']}
> >>> +'0x800': {'machine_args':  ['-march=armv8-a+crc']},
> >>> +'0xc00': {'machine_args':  ['-march=armv8-a+crc']},
> >> Nit, redundant comma at the end of the line.
> > 
> > What is redundant?
> 
> The comma at the end of the second line is not necessary.

It is a good practice to have comma like other lines,
so no need to update this line when adding more.




Re: [dpdk-dev] [RFC v2] porting AddressSanitizer feature to DPDK

2021-06-18 Thread David Marchand
On Fri, Jun 18, 2021 at 9:49 AM Lin, Xueqin  wrote:
> > > Suggest listing demo code and tool capture information for user to try if
> > tool works, also add this part into doc.
> > >
> > > >
> > > > # Also, Please update the release note for this feature.
> > > Sure, we can update the release note if code merge.
> >
> > Probably you can send v1 version next i.e change the RFC status to get
> > merged.
>
> Sure, we will send v1 patch if no obvious objection for that, hope patch 
> could receive some ACKs and could success to merge, thanks.

How did you test this work?

UNH recently started testing with ASAN and it reveals leaks just in
the unit test.

Merging these annotations will help catch more issues.
But users will hit the current issues that we must fix first.


-- 
David Marchand



Re: [dpdk-dev] [PATCH] parray: introduce internal API for dynamic arrays

2021-06-18 Thread Morten Brørup
> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Ananyev,
> Konstantin
> Sent: Thursday, 17 June 2021 19.06
> 
> I think it would be better to change rx_pkt_burst() to accept port_id
> and queue_id, instead of void *.

Current:

typedef uint16_t (*eth_rx_burst_t)(void *rxq,
   struct rte_mbuf **rx_pkts,
   uint16_t nb_pkts);

> I.E:
> typedef uint16_t (*eth_rx_burst_t)(uint16_t port_id,
> uint16_t queue_id,
> struct rte_mbuf **rx_pkts,
> uint16_t nb_pkts);
> 
> And we can do actual de-referencing of private rxq data inside the
> actual rx function.

Good idea, if it can be done without a performance cost.

The X64 calling convention allows up to 4 parameters passed as registers, so 
the added parameter should not be a problem.


Another thing:

I just noticed that struct rte_eth_dev_data has "void **rx_queues;" (and 
similarly for tx_queues).

That should be "void *rx_queues[RTE_MAX_QUEUES_PER_PORT];", like in all the 
other ethdev structures.

The same structure even has "uint8_t rx_queue_state[RTE_MAX_QUEUES_PER_PORT];", 
so it's following two different conventions.




Re: [dpdk-dev] [RFC PATCH] dmadev: introduce DMA device library

2021-06-18 Thread Bruce Richardson
On Fri, Jun 18, 2021 at 04:52:00PM +0800, fengchengwen wrote:
> On 2021/6/17 22:18, Bruce Richardson wrote:
> > On Thu, Jun 17, 2021 at 12:02:00PM +0100, Bruce Richardson wrote:
> >> On Thu, Jun 17, 2021 at 05:48:05PM +0800, fengchengwen wrote:
> >>> On 2021/6/17 1:31, Bruce Richardson wrote:
>  On Wed, Jun 16, 2021 at 05:41:45PM +0800, fengchengwen wrote:
> > On 2021/6/16 0:38, Bruce Richardson wrote:
> >> On Tue, Jun 15, 2021 at 09:22:07PM +0800, Chengwen Feng wrote:
> >>> This patch introduces 'dmadevice' which is a generic type of DMA
> >>> device.
> >>>
> >>> The APIs of dmadev library exposes some generic operations which
> >>> can enable configuration and I/O with the DMA devices.
> >>>
> >>> Signed-off-by: Chengwen Feng  ---
> >> Thanks for sending this.
> >>
> >> Of most interest to me right now are the key data-plane APIs.
> >> While we are still in the prototyping phase, below is a draft of
> >> what we are thinking for the key enqueue/perform_ops/completed_ops
> >> APIs.
> >>
> >> Some key differences I note in below vs your original RFC: * Use
> >> of void pointers rather than iova addresses. While using iova's
> >> makes sense in the general case when using hardware, in that it
> >> can work with both physical addresses and virtual addresses, if we
> >> change the APIs to use void pointers instead it will still work
> >> for DPDK in VA mode, while at the same time allow use of software
> >> fallbacks in error cases, and also a stub driver than uses memcpy
> >> in the background. Finally, using iova's makes the APIs a lot more
> >> awkward to use with anything but mbufs or similar buffers where we
> >> already have a pre-computed physical address.
> >
> > The iova is an hint to application, and widely used in DPDK.  If
> > switch to void, how to pass the address (iova or just va ?) this
> > may introduce implementation dependencies here.
> >
> > Or always pass the va, and the driver performs address translation,
> > and this translation may cost too much cpu I think.
> >
> 
>  On the latter point, about driver doing address translation I would
>  agree.  However, we probably need more discussion about the use of
>  iova vs just virtual addresses. My thinking on this is that if we
>  specify the API using iovas it will severely hurt usability of the
>  API, since it forces the user to take more inefficient codepaths in
>  a large number of cases. Given a pointer to the middle of an mbuf,
>  one cannot just pass that straight as an iova but must instead do a
>  translation into offset from mbuf pointer and then readd the offset
>  to the mbuf base address.
> 
>  My preference therefore is to require the use of an IOMMU when using
>  a dmadev, so that it can be a much closer analog of memcpy. Once an
>  iommu is present, DPDK will run in VA mode, allowing virtual
>  addresses to our hugepage memory to be sent directly to hardware.
>  Also, when using dmadevs on top of an in-kernel driver, that kernel
>  driver may do all iommu management for the app, removing further the
>  restrictions on what memory can be addressed by hardware.
> >>>
> >>> Some DMA devices many don't support IOMMU or IOMMU bypass default, so
> >>> driver may should call rte_mem_virt2phy() do the address translate,
> >>> but the rte_mem_virt2phy() cost too many CPU cycles.
> >>>
> >>> If the API defined as iova, it will work fine in: 1) If DMA don't
> >>> support IOMMU or IOMMU bypass, then start application with
> >>> --iova-mode=pa 2) If DMA support IOMMU, --iova-mode=pa/va work both
> >>> fine
> >>>
> >>
> >> I suppose if we keep the iova as the datatype, we can just cast "void
> >> *" pointers to that in the case that virtual addresses can be used
> >> directly. I believe your RFC included a capability query API - "uses
> >> void * as iova" should probably be one of those capabilities, and that
> >> would resolve this.  If DPDK is in iova=va mode because of the
> >> presence of an iommu, all drivers could report this capability too.
> >>
> 
> >> * Use of id values rather than user-provided handles. Allowing the
> >> user/app to manage the amount of data stored per operation is a
> >> better solution, I feel than proscribing a certain about of
> >> in-driver tracking. Some apps may not care about anything other
> >> than a job being completed, while other apps may have significant
> >> metadata to be tracked. Taking the user-context handles out of the
> >> API also makes the driver code simpler.
> >
> > The user-provided handle was mainly used to simply application
> > implementation, It provides the ability to quickly locate contexts.
> >
> > The "use of id values" seem like the dma_cookie of Linux DMA engine
> > framework, user will get a unique dma_cookie aft

Re: [dpdk-dev] [RFC PATCH] dmadev: introduce DMA device library

2021-06-18 Thread fengchengwen
On 2021/6/18 13:52, Jerin Jacob wrote:
> On Thu, Jun 17, 2021 at 2:46 PM Bruce Richardson
>  wrote:
>>
>> On Wed, Jun 16, 2021 at 08:07:26PM +0530, Jerin Jacob wrote:
>>> On Wed, Jun 16, 2021 at 3:47 PM fengchengwen  
>>> wrote:

 On 2021/6/16 15:09, Morten Brørup wrote:
>> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Bruce Richardson
>> Sent: Tuesday, 15 June 2021 18.39
>>
>> On Tue, Jun 15, 2021 at 09:22:07PM +0800, Chengwen Feng wrote:
>>> This patch introduces 'dmadevice' which is a generic type of DMA
>>> device.
>>>
>>> The APIs of dmadev library exposes some generic operations which can
>>> enable configuration and I/O with the DMA devices.
>>>
>>> Signed-off-by: Chengwen Feng 
>>> ---
>> Thanks for sending this.
>>
>> Of most interest to me right now are the key data-plane APIs. While we
>> are
>> still in the prototyping phase, below is a draft of what we are
>> thinking
>> for the key enqueue/perform_ops/completed_ops APIs.
>>
>> Some key differences I note in below vs your original RFC:
>> * Use of void pointers rather than iova addresses. While using iova's
>> makes
>>   sense in the general case when using hardware, in that it can work
>> with
>>   both physical addresses and virtual addresses, if we change the APIs
>> to use
>>   void pointers instead it will still work for DPDK in VA mode, while
>> at the
>>   same time allow use of software fallbacks in error cases, and also a
>> stub
>>   driver than uses memcpy in the background. Finally, using iova's
>> makes the
>>   APIs a lot more awkward to use with anything but mbufs or similar
>> buffers
>>   where we already have a pre-computed physical address.
>> * Use of id values rather than user-provided handles. Allowing the
>> user/app
>>   to manage the amount of data stored per operation is a better
>> solution, I
>>   feel than proscribing a certain about of in-driver tracking. Some
>> apps may
>>   not care about anything other than a job being completed, while other
>> apps
>>   may have significant metadata to be tracked. Taking the user-context
>>   handles out of the API also makes the driver code simpler.
>> * I've kept a single combined API for completions, which differs from
>> the
>>   separate error handling completion API you propose. I need to give
>> the
>>   two function approach a bit of thought, but likely both could work.
>> If we
>>   (likely) never expect failed ops, then the specifics of error
>> handling
>>   should not matter that much.
>>
>> For the rest, the control / setup APIs are likely to be rather
>> uncontroversial, I suspect. However, I think that rather than xstats
>> APIs,
>> the library should first provide a set of standardized stats like
>> ethdev
>> does. If driver-specific stats are needed, we can add xstats later to
>> the
>> API.
>>
>> Appreciate your further thoughts on this, thanks.
>>
>> Regards,
>> /Bruce
>
> I generally agree with Bruce's points above.
>
> I would like to share a couple of ideas for further discussion:
>>>
>>>
>>> I believe some of the other requirements and comments for generic DMA will 
>>> be
>>>
>>> 1) Support for the _channel_, Each channel may have different
>>> capabilities and functionalities.
>>> Typical cases are, each channel have separate source and destination
>>> devices like
>>> DMA between PCIe EP to Host memory, Host memory to Host memory, PCIe
>>> EP to PCIe EP.
>>> So we need some notion of the channel in the specification.
>>>
>>
>> Can you share a bit more detail on what constitutes a channel in this case?
>> Is it equivalent to a device queue (which we are flattening to individual
>> devices in this API), or to a specific configuration on a queue?
> 
> It not a queue. It is one of the attributes for transfer.
> I.e in the same queue, for a given transfer it can specify the
> different "source" and "destination" device.
> Like CPU to Sound card, CPU to network card etc.
> 
> 
>>
>>> 2) I assume current data plane APIs are not thread-safe. Is it right?
>>>
>> Yes.
>>
>>>
>>> 3) Cookie scheme outlined earlier looks good to me. Instead of having
>>> generic dequeue() API
>>>
>>> 4) Can split the rte_dmadev_enqueue_copy(uint16_t dev_id, void * src,
>>> void * dst, unsigned int length);
>>> to two stage API like, Where one will be used in fastpath and other
>>> one will use used in slowpath.
>>>
>>> - slowpath API will for take channel and take other attributes for transfer
>>>
>>> Example syantx will be:
>>>
>>> struct rte_dmadev_desc {
>>>channel id;
>>>ops ; // copy, xor, fill etc
>>>   other arguments specific to dma transfer // it can be set
>>> based on capability.
>>>
>>> };
>>>
>>> rte_dmadev_desc_t rte_dmadev_preprare(

Re: [dpdk-dev] [RFC PATCH] dmadev: introduce DMA device library

2021-06-18 Thread Bruce Richardson
On Fri, Jun 18, 2021 at 11:22:28AM +0530, Jerin Jacob wrote:
> On Thu, Jun 17, 2021 at 2:46 PM Bruce Richardson
>  wrote:
> >
> > On Wed, Jun 16, 2021 at 08:07:26PM +0530, Jerin Jacob wrote:
> > > On Wed, Jun 16, 2021 at 3:47 PM fengchengwen  
> > > wrote:
> > > >
> > > > On 2021/6/16 15:09, Morten Brørup wrote:
> > > > >> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Bruce Richardson
> > > > >> Sent: Tuesday, 15 June 2021 18.39
> > > > >>
> > > > >> On Tue, Jun 15, 2021 at 09:22:07PM +0800, Chengwen Feng wrote:
> > > > >>> This patch introduces 'dmadevice' which is a generic type of DMA
> > > > >>> device.
> > > > >>>
> > > > >>> The APIs of dmadev library exposes some generic operations which can
> > > > >>> enable configuration and I/O with the DMA devices.
> > > > >>>
> > > > >>> Signed-off-by: Chengwen Feng 
> > > > >>> ---
> > > > >> Thanks for sending this.
> > > > >>
> > > > >> Of most interest to me right now are the key data-plane APIs. While 
> > > > >> we
> > > > >> are
> > > > >> still in the prototyping phase, below is a draft of what we are
> > > > >> thinking
> > > > >> for the key enqueue/perform_ops/completed_ops APIs.
> > > > >>
> > > > >> Some key differences I note in below vs your original RFC:
> > > > >> * Use of void pointers rather than iova addresses. While using iova's
> > > > >> makes
> > > > >>   sense in the general case when using hardware, in that it can work
> > > > >> with
> > > > >>   both physical addresses and virtual addresses, if we change the 
> > > > >> APIs
> > > > >> to use
> > > > >>   void pointers instead it will still work for DPDK in VA mode, while
> > > > >> at the
> > > > >>   same time allow use of software fallbacks in error cases, and also 
> > > > >> a
> > > > >> stub
> > > > >>   driver than uses memcpy in the background. Finally, using iova's
> > > > >> makes the
> > > > >>   APIs a lot more awkward to use with anything but mbufs or similar
> > > > >> buffers
> > > > >>   where we already have a pre-computed physical address.
> > > > >> * Use of id values rather than user-provided handles. Allowing the
> > > > >> user/app
> > > > >>   to manage the amount of data stored per operation is a better
> > > > >> solution, I
> > > > >>   feel than proscribing a certain about of in-driver tracking. Some
> > > > >> apps may
> > > > >>   not care about anything other than a job being completed, while 
> > > > >> other
> > > > >> apps
> > > > >>   may have significant metadata to be tracked. Taking the 
> > > > >> user-context
> > > > >>   handles out of the API also makes the driver code simpler.
> > > > >> * I've kept a single combined API for completions, which differs from
> > > > >> the
> > > > >>   separate error handling completion API you propose. I need to give
> > > > >> the
> > > > >>   two function approach a bit of thought, but likely both could work.
> > > > >> If we
> > > > >>   (likely) never expect failed ops, then the specifics of error
> > > > >> handling
> > > > >>   should not matter that much.
> > > > >>
> > > > >> For the rest, the control / setup APIs are likely to be rather
> > > > >> uncontroversial, I suspect. However, I think that rather than xstats
> > > > >> APIs,
> > > > >> the library should first provide a set of standardized stats like
> > > > >> ethdev
> > > > >> does. If driver-specific stats are needed, we can add xstats later to
> > > > >> the
> > > > >> API.
> > > > >>
> > > > >> Appreciate your further thoughts on this, thanks.
> > > > >>
> > > > >> Regards,
> > > > >> /Bruce
> > > > >
> > > > > I generally agree with Bruce's points above.
> > > > >
> > > > > I would like to share a couple of ideas for further discussion:
> > >
> > >
> > > I believe some of the other requirements and comments for generic DMA 
> > > will be
> > >
> > > 1) Support for the _channel_, Each channel may have different
> > > capabilities and functionalities.
> > > Typical cases are, each channel have separate source and destination
> > > devices like
> > > DMA between PCIe EP to Host memory, Host memory to Host memory, PCIe
> > > EP to PCIe EP.
> > > So we need some notion of the channel in the specification.
> > >
> >
> > Can you share a bit more detail on what constitutes a channel in this case?
> > Is it equivalent to a device queue (which we are flattening to individual
> > devices in this API), or to a specific configuration on a queue?
> 
> It not a queue. It is one of the attributes for transfer.
> I.e in the same queue, for a given transfer it can specify the
> different "source" and "destination" device.
> Like CPU to Sound card, CPU to network card etc.
>
Ok. Thanks for clarifying. Do you think it's best given as a
device-specific parameter to the various functions, and NULL for hardware
that doesn't need it?
 
> 
> >
> > > 2) I assume current data plane APIs are not thread-safe. Is it right?
> > >
> > Yes.
> >
> > >
> > > 3) Cookie scheme outlined earlier looks good to me. Instead of having
> > > generic dequeue() API
> > >
> > 

Re: [dpdk-dev] [RFC PATCH] dmadev: introduce DMA device library

2021-06-18 Thread Bruce Richardson
On Fri, Jun 18, 2021 at 10:46:08AM +0530, Jerin Jacob wrote:
> On Thu, Jun 17, 2021 at 1:30 PM Bruce Richardson
>  wrote:
> >
> > On Thu, Jun 17, 2021 at 01:12:22PM +0530, Jerin Jacob wrote:
> > > On Thu, Jun 17, 2021 at 12:43 AM Bruce Richardson
> > >  wrote:
> > > >
> > > > On Wed, Jun 16, 2021 at 11:38:08PM +0530, Jerin Jacob wrote:
> > > > > On Wed, Jun 16, 2021 at 11:01 PM Bruce Richardson
> > > > >  wrote:
> > > > > >
> > > > > > On Wed, Jun 16, 2021 at 05:41:45PM +0800, fengchengwen wrote:
> > > > > > > On 2021/6/16 0:38, Bruce Richardson wrote:
> > > > > > > > On Tue, Jun 15, 2021 at 09:22:07PM +0800, Chengwen Feng wrote:
> > > > > > > >> This patch introduces 'dmadevice' which is a generic type of 
> > > > > > > >> DMA
> > > > > > > >> device.
> > > > > > > >>
> > > > > > > >> The APIs of dmadev library exposes some generic operations 
> > > > > > > >> which can
> > > > > > > >> enable configuration and I/O with the DMA devices.
> > > > > > > >>
> > > > > > > >> Signed-off-by: Chengwen Feng 
> > > > > > > >> ---
> > > > > > > > Thanks for sending this.
> > > > > > > >
> > > > > > > > Of most interest to me right now are the key data-plane APIs. 
> > > > > > > > While we are
> > > > > > > > still in the prototyping phase, below is a draft of what we are 
> > > > > > > > thinking
> > > > > > > > for the key enqueue/perform_ops/completed_ops APIs.
> > > > > > > >
> > > > > > > > Some key differences I note in below vs your original RFC:
> > > > > > > > * Use of void pointers rather than iova addresses. While using 
> > > > > > > > iova's makes
> > > > > > > >   sense in the general case when using hardware, in that it can 
> > > > > > > > work with
> > > > > > > >   both physical addresses and virtual addresses, if we change 
> > > > > > > > the APIs to use
> > > > > > > >   void pointers instead it will still work for DPDK in VA mode, 
> > > > > > > > while at the
> > > > > > > >   same time allow use of software fallbacks in error cases, and 
> > > > > > > > also a stub
> > > > > > > >   driver than uses memcpy in the background. Finally, using 
> > > > > > > > iova's makes the
> > > > > > > >   APIs a lot more awkward to use with anything but mbufs or 
> > > > > > > > similar buffers
> > > > > > > >   where we already have a pre-computed physical address.
> > > > > > >
> > > > > > > The iova is an hint to application, and widely used in DPDK.
> > > > > > > If switch to void, how to pass the address (iova or just va ?)
> > > > > > > this may introduce implementation dependencies here.
> > > > > > >
> > > > > > > Or always pass the va, and the driver performs address 
> > > > > > > translation, and this
> > > > > > > translation may cost too much cpu I think.
> > > > > > >
> > > > > >
> > > > > > On the latter point, about driver doing address translation I would 
> > > > > > agree.
> > > > > > However, we probably need more discussion about the use of iova vs 
> > > > > > just
> > > > > > virtual addresses. My thinking on this is that if we specify the 
> > > > > > API using
> > > > > > iovas it will severely hurt usability of the API, since it forces 
> > > > > > the user
> > > > > > to take more inefficient codepaths in a large number of cases. 
> > > > > > Given a
> > > > > > pointer to the middle of an mbuf, one cannot just pass that 
> > > > > > straight as an
> > > > > > iova but must instead do a translation into offset from mbuf 
> > > > > > pointer and
> > > > > > then readd the offset to the mbuf base address.
> > > > > >
> > > > > > My preference therefore is to require the use of an IOMMU when 
> > > > > > using a
> > > > > > dmadev, so that it can be a much closer analog of memcpy. Once an 
> > > > > > iommu is
> > > > > > present, DPDK will run in VA mode, allowing virtual addresses to our
> > > > > > hugepage memory to be sent directly to hardware. Also, when using
> > > > > > dmadevs on top of an in-kernel driver, that kernel driver may do 
> > > > > > all iommu
> > > > > > management for the app, removing further the restrictions on what 
> > > > > > memory
> > > > > > can be addressed by hardware.
> > > > >
> > > > >
> > > > > One issue of keeping void * is that memory can come from stack or 
> > > > > heap .
> > > > > which HW can not really operate it on.
> > > >
> > > > when kernel driver is managing the IOMMU all process memory can be 
> > > > worked
> > > > on, not just hugepage memory, so using iova is wrong in these cases.
> > >
> > > But not for stack and heap memory. Right?
> > >
> > Yes, even stack and heap can be accessed.
> 
> The HW device cannot as that memory is NOT mapped to IOMMU. It will
> result in the transaction
> fault.
>

Not if the kernel driver rather than DPDK is managing the IOMMU:
https://www.kernel.org/doc/html/latest/x86/sva.html
"Shared Virtual Addressing (SVA) allows the processor and device to use the
same virtual addresses avoiding the need for software to translate virtual
addresses to physical addresses. SVA is what PCIe calls Shared Virtual
Mem

Re: [dpdk-dev] [PATCH] parray: introduce internal API for dynamic arrays

2021-06-18 Thread Ferruh Yigit
On 6/17/2021 5:55 PM, Morten Brørup wrote:
>> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Ferruh Yigit
>> Sent: Thursday, 17 June 2021 18.13
>>
>> On 6/17/2021 4:17 PM, Morten Brørup wrote:
 From: Ananyev, Konstantin [mailto:konstantin.anan...@intel.com]
 Sent: Thursday, 17 June 2021 16.59


 14/06/2021 15:15, Bruce Richardson:
> On Mon, Jun 14, 2021 at 02:22:42PM +0200, Morten Brørup wrote:
>>> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Thomas
 Monjalon
>>> Sent: Monday, 14 June 2021 12.59
>>>
>>> Performance of access in a fixed-size array is very good
>>> because of cache locality
>>> and because there is a single pointer to dereference.
>>> The only drawback is the lack of flexibility:
>>> the size of such an array cannot be increase at runtime.
>>>
>>> An approach to this problem is to allocate the array at
 runtime,
>>> being as efficient as static arrays, but still limited to a
 maximum.
>>>
>>> That's why the API rte_parray is introduced,
>>> allowing to declare an array of pointer which can be resized
>>> dynamically
>>> and automatically at runtime while keeping a good read
 performance.
>>>
>>> After resize, the previous array is kept until the next
>> resize
>>> to avoid crashs during a read without any lock.
>>>
>>> Each element is a pointer to a memory chunk dynamically
 allocated.
>>> This is not good for cache locality but it allows to keep the
 same
>>> memory per element, no matter how the array is resized.
>>> Cache locality could be improved with mempools.
>>> The other drawback is having to dereference one more pointer
>>> to read an element.
>>>
>>> There is not much locks, so the API is for internal use only.
>>> This API may be used to completely remove some compilation-
 time
>>> maximums.
>>
>> I get the purpose and overall intention of this library.
>>
>> I probably already mentioned that I prefer "embedded style
 programming" with fixed size arrays, rather than runtime
 configurability.
>>> It's
 my personal opinion, and the DPDK Tech Board clearly prefers
 reducing the amount of compile time configurability, so there is no
>> way
> for
 me to stop this progress, and I do not intend to oppose to this
 library. :-)
>>
>> This library is likely to become a core library of DPDK, so I
 think it is important getting it right. Could you please mention a
>> few
>>> examples
 where you think this internal library should be used, and where
 it should not be used. Then it is easier to discuss if the border
>> line
> between
 control path and data plane is correct. E.g. this library is not
 intended to be used for dynamically sized packet queues that grow
>> and
> shrink
>>> in
 the fast path.
>>
>> If the library becomes a core DPDK library, it should probably
 be public instead of internal. E.g. if the library is used to make
 RTE_MAX_ETHPORTS dynamic instead of compile time fixed, then
>> some
 applications might also need dynamically sized arrays for their
 application specific per-port runtime data, and this library
 could serve that purpose too.
>>
>
> Thanks Thomas for starting this discussion and Morten for
 follow-up.
>
> My thinking is as follows, and I'm particularly keeping in mind
 the cases
> of e.g. RTE_MAX_ETHPORTS, as a leading candidate here.
>
> While I dislike the hard-coded limits in DPDK, I'm also not
 convinced that
> we should switch away from the flat arrays or that we need
>> fully
 dynamic
> arrays that grow/shrink at runtime for ethdevs. I would suggest
 a half-way
> house here, where we keep the ethdevs as an array, but one
 allocated/sized
> at runtime rather than statically. This would allow us to have
>> a
> compile-time default value, but, for use cases that need it,
 allow use of a
> flag e.g.  "max-ethdevs" to change the size of the parameter
 given to the
> malloc call for the array.  This max limit could then be
 provided to apps
> too if they want to match any array sizes. [Alternatively those
 apps could
> check the provided size and error out if the size has been
 increased beyond
> what the app is designed to use?]. There would be no extra
 dereferences per
> rx/tx burst call in this scenario so performance should be the
 same as
> before (potentially better if array is in hugepage memory, I
 suppose).

 I think we need some

Re: [dpdk-dev] [PATCH] parray: introduce internal API for dynamic arrays

2021-06-18 Thread Ferruh Yigit
On 6/17/2021 6:05 PM, Ananyev, Konstantin wrote:
> 
> 
>> On 6/17/2021 4:17 PM, Morten Brørup wrote:
 From: Ananyev, Konstantin [mailto:konstantin.anan...@intel.com]
 Sent: Thursday, 17 June 2021 16.59


 14/06/2021 15:15, Bruce Richardson:
> On Mon, Jun 14, 2021 at 02:22:42PM +0200, Morten Brørup wrote:
>>> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Thomas
 Monjalon
>>> Sent: Monday, 14 June 2021 12.59
>>>
>>> Performance of access in a fixed-size array is very good
>>> because of cache locality
>>> and because there is a single pointer to dereference.
>>> The only drawback is the lack of flexibility:
>>> the size of such an array cannot be increase at runtime.
>>>
>>> An approach to this problem is to allocate the array at
 runtime,
>>> being as efficient as static arrays, but still limited to a
 maximum.
>>>
>>> That's why the API rte_parray is introduced,
>>> allowing to declare an array of pointer which can be resized
>>> dynamically
>>> and automatically at runtime while keeping a good read
 performance.
>>>
>>> After resize, the previous array is kept until the next resize
>>> to avoid crashs during a read without any lock.
>>>
>>> Each element is a pointer to a memory chunk dynamically
 allocated.
>>> This is not good for cache locality but it allows to keep the
 same
>>> memory per element, no matter how the array is resized.
>>> Cache locality could be improved with mempools.
>>> The other drawback is having to dereference one more pointer
>>> to read an element.
>>>
>>> There is not much locks, so the API is for internal use only.
>>> This API may be used to completely remove some compilation-
 time
>>> maximums.
>>
>> I get the purpose and overall intention of this library.
>>
>> I probably already mentioned that I prefer "embedded style
 programming" with fixed size arrays, rather than runtime
 configurability.
>>> It's
 my personal opinion, and the DPDK Tech Board clearly prefers
 reducing the amount of compile time configurability, so there is no way
> for
 me to stop this progress, and I do not intend to oppose to this
 library. :-)
>>
>> This library is likely to become a core library of DPDK, so I
 think it is important getting it right. Could you please mention a few
>>> examples
 where you think this internal library should be used, and where
 it should not be used. Then it is easier to discuss if the border line
> between
 control path and data plane is correct. E.g. this library is not
 intended to be used for dynamically sized packet queues that grow and
> shrink
>>> in
 the fast path.
>>
>> If the library becomes a core DPDK library, it should probably
 be public instead of internal. E.g. if the library is used to make
 RTE_MAX_ETHPORTS dynamic instead of compile time fixed, then some
 applications might also need dynamically sized arrays for their
 application specific per-port runtime data, and this library
 could serve that purpose too.
>>
>
> Thanks Thomas for starting this discussion and Morten for
 follow-up.
>
> My thinking is as follows, and I'm particularly keeping in mind
 the cases
> of e.g. RTE_MAX_ETHPORTS, as a leading candidate here.
>
> While I dislike the hard-coded limits in DPDK, I'm also not
 convinced that
> we should switch away from the flat arrays or that we need fully
 dynamic
> arrays that grow/shrink at runtime for ethdevs. I would suggest
 a half-way
> house here, where we keep the ethdevs as an array, but one
 allocated/sized
> at runtime rather than statically. This would allow us to have a
> compile-time default value, but, for use cases that need it,
 allow use of a
> flag e.g.  "max-ethdevs" to change the size of the parameter
 given to the
> malloc call for the array.  This max limit could then be
 provided to apps
> too if they want to match any array sizes. [Alternatively those
 apps could
> check the provided size and error out if the size has been
 increased beyond
> what the app is designed to use?]. There would be no extra
 dereferences per
> rx/tx burst call in this scenario so performance should be the
 same as
> before (potentially better if array is in hugepage memory, I
 suppose).

 I think we need some benchmarks to decide what is the best
 tradeoff.
 I spent time on this implementation, but sorry I won't have

[dpdk-dev] [PATCH v3 00/62] Marvell CNXK Ethdev Driver

2021-06-18 Thread Nithin Dabilpuram
This patchset adds support for Marvell CN106XX SoC based on 'common/cnxk'
driver. In future, CN9K a.k.a octeontx2 will also be supported by same
driver when code is ready and 'net/octeontx2' will be deprecated.

Harman Kalra (1):
  common/cnxk: allocate lmt region in userspace

Jerin Jacob (7):
  common/cnxk: fix batch alloc completion poll logic
  net/cnxk: add Rx support for cn9k
  net/cnxk: add Rx vector version for cn9k
  net/cnxk: add Tx support for cn9k
  net/cnxk: add Rx support for cn10k
  net/cnxk: add Rx vector version for cn10k
  net/cnxk: add Tx support for cn10k

Kiran Kumar K (2):
  net/cnxk: add support to configure npc
  net/cnxk: add initial version of rte flow support

Nithin Dabilpuram (17):
  net/cnxk: add build infra and common probe
  net/cnxk: add platform specific probe and remove
  net/cnxk: add common devargs parsing function
  net/cnxk: add common dev infos get support
  net/cnxk: add device configuration operation
  net/cnxk: add link status update support
  net/cnxk: add Rx queue setup and release
  net/cnxk: add Tx queue setup and release
  net/cnxk: add packet type support
  net/cnxk: add queue start and stop support
  net/cnxk: add Rx multi-segmented version for cn9k
  net/cnxk: add Tx multi-segment version for cn9k
  net/cnxk: add Tx vector version for cn9k
  net/cnxk: add Rx multi-segment version for cn10k
  net/cnxk: add Tx multi-segment version for cn10k
  net/cnxk: add Tx vector version for cn10k
  net/cnxk: add device start and stop operations

Satha Rao (8):
  common/cnxk: add support to lock NIX RQ contexts
  common/cnxk: add provision to enable RED on RQ
  net/cnxk: add port/queue stats
  net/cnxk: add xstats apis
  net/cnxk: add rxq/txq info get operations
  net/cnxk: add ethdev firmware version get
  net/cnxk: add get register operation
  net/cnxk: added RETA and RSS hash operations

Satheesh Paul (7):
  common/cnxk: add support to dump flow entries
  common/cnxk: support for mark and flag flow actions
  common/cnxk: support for VLAN push and pop flow actions
  common/cnxk: fix flow create on CN98xx
  net/cnxk: add flow ops get operation
  net/cnxk: support for RSS in rte flow
  net/cnxk: add marking and VLAN tagging support

Sunil Kumar Kori (20):
  net/cnxk: add MAC address set ops
  net/cnxk: add MTU set device operation
  net/cnxk: add promiscuous mode enable and disable
  net/cnxk: add DMAC filter support
  net/cnxk: add all multicast enable/disable ethops
  net/cnxk: add Rx/Tx burst mode get ops
  net/cnxk: add flow ctrl set/get ops
  net/cnxk: add link up/down operations
  net/cnxk: add EEPROM module info get operations
  net/cnxk: add Rx queue interrupt enable/disable ops
  net/cnxk: add validation API for mempool ops
  net/cnxk: add device close and reset operations
  net/cnxk: add pending Tx mbuf cleanup operation
  net/cnxk: register callback to get PTP status
  net/cnxk: add base PTP timesync support
  net/cnxk: add timesync enable/disable operations
  net/cnxk: add Rx/Tx timestamp read operations
  net/cnxk: add time read/write/adjust operations
  net/cnxk: add read clock operation
  net/cnxk: add multicast filter support

--

v3:
- Updated release notes
- Removed RTE_ETH_DEV_AUTOFILL_QUEUE_XSTATS flag and add support for queue
  stats in xstats
- Fixed issue with LSO format indices
- Removed mbox sync changes patch from this series
- Fixed documentation issues
- Removed repetitive code in fast path SIMD
- Optimize cn10k LMTST logic
- Make rte_flow_create implementation specific
  to handle VLAN Stripping and MARK actions/offloads
- Use rte_atomic_thread_fence() instead of rte_rmb()
- Handle other comments from Jerin.
- Merged rte flow dump API patch to flow ops get patch
- Added marking and vlan tagging support.
- Fixed some checkpatch and git check log issues.

v2:
- Fixed issue with flow validate and flow create for 98xx
- Fixed issue batch alloc logic
- Fix lmtline allocation to be cached
- Sync Inline IPSec Rx mbox with kernel
- Add support for mark and flag flow actions
- Add reta key and hash update ops
- Added PTP and multicast filter support
 
 MAINTAINERS |5 +-
 doc/guides/nics/cnxk.rst|  232 +
 doc/guides/nics/features/cnxk.ini   |   90 ++
 doc/guides/nics/features/cnxk_vec.ini   |   44 +
 doc/guides/nics/features/cnxk_vf.ini|   40 +
 doc/guides/nics/index.rst   |1 +
 doc/guides/platform/cnxk.rst|3 +
 doc/guides/rel_notes/release_21_08.rst  |5 +
 drivers/common/cnxk/hw/npc.h|2 +
 drivers/common/cnxk/meson.build |1 +
 drivers/common/cnxk/roc_api.h   |2 +
 drivers/common/cnxk/roc_dev.c   |   98 +-
 drivers/common/cnxk/roc_dev_priv.h  |1 +
 drivers/common/cnxk/roc_mbox.h  |3 +
 drivers/common/cnxk/roc_model.h |6 +
 drivers/common/cnxk/roc_nix.h   |   39 +-
 drivers/common/cnxk/roc_nix_queue.c |   52 +
 drivers/common/cnxk/roc_nix_rss.c   |   51 +-
 driver

[dpdk-dev] [PATCH v3 01/62] common/cnxk: add support to lock NIX RQ contexts

2021-06-18 Thread Nithin Dabilpuram
From: Satha Rao 

This patch will consider device argument to lock rss table
in NIX.

This patch also adds few misc fixes such as disabling NIX Tx
vlan insertion conf in SMQ, enabling SSO in NIX Tx SQ
for Tx completions and TM related stats API.

Signed-off-by: Satha Rao 
---
 drivers/common/cnxk/roc_nix.h  | 31 ++--
 drivers/common/cnxk/roc_nix_queue.c|  2 +
 drivers/common/cnxk/roc_nix_rss.c  | 51 ++--
 drivers/common/cnxk/roc_nix_tm_utils.c | 86 +-
 drivers/common/cnxk/roc_platform.h |  2 +
 drivers/common/cnxk/version.map|  1 +
 6 files changed, 163 insertions(+), 10 deletions(-)

diff --git a/drivers/common/cnxk/roc_nix.h b/drivers/common/cnxk/roc_nix.h
index b39f461..6d9ac10 100644
--- a/drivers/common/cnxk/roc_nix.h
+++ b/drivers/common/cnxk/roc_nix.h
@@ -85,10 +85,11 @@ struct roc_nix_eeprom_info {
 #define ROC_NIX_LF_RX_CFG_LEN_OL3 BIT_ULL(41)
 
 /* Group 0 will be used for RSS, 1 -7 will be used for npc_flow RSS action*/
-#define ROC_NIX_RSS_GROUP_DEFAULT 0
-#define ROC_NIX_RSS_GRPS 8
-#define ROC_NIX_RSS_RETA_MAX ROC_NIX_RSS_RETA_SZ_256
-#define ROC_NIX_RSS_KEY_LEN  48 /* 352 Bits */
+#define ROC_NIX_RSS_GROUP_DEFAULT0
+#define ROC_NIX_RSS_GRPS8
+#define ROC_NIX_RSS_RETA_MAXROC_NIX_RSS_RETA_SZ_256
+#define ROC_NIX_RSS_KEY_LEN 48 /* 352 Bits */
+#define ROC_NIX_RSS_MCAM_IDX_DEFAULT (-1)
 
 #define ROC_NIX_DEFAULT_HW_FRS 1514
 
@@ -184,6 +185,7 @@ struct roc_nix_sq {
enum roc_nix_sq_max_sqe_sz max_sqe_sz;
uint32_t nb_desc;
uint16_t qid;
+   bool sso_ena;
/* End of Input parameters */
uint16_t sqes_per_sqb_log2;
struct roc_nix *roc_nix;
@@ -241,6 +243,8 @@ struct roc_nix {
uint16_t max_sqb_count;
enum roc_nix_rss_reta_sz reta_sz;
bool enable_loop;
+   bool hw_vlan_ins;
+   uint8_t lock_rx_ctx;
/* End of input parameters */
/* LMT line base for "Per Core Tx LMT line" mode*/
uintptr_t lmt_base;
@@ -371,6 +375,22 @@ struct roc_nix_tm_shaper_profile {
void (*free_fn)(void *profile);
 };
 
+enum roc_nix_tm_node_stats_type {
+   ROC_NIX_TM_NODE_PKTS_DROPPED,
+   ROC_NIX_TM_NODE_BYTES_DROPPED,
+   ROC_NIX_TM_NODE_GREEN_PKTS,
+   ROC_NIX_TM_NODE_GREEN_BYTES,
+   ROC_NIX_TM_NODE_YELLOW_PKTS,
+   ROC_NIX_TM_NODE_YELLOW_BYTES,
+   ROC_NIX_TM_NODE_RED_PKTS,
+   ROC_NIX_TM_NODE_RED_BYTES,
+   ROC_NIX_TM_NODE_STATS_MAX,
+};
+
+struct roc_nix_tm_node_stats {
+   uint64_t stats[ROC_NIX_TM_NODE_STATS_MAX];
+};
+
 int __roc_api roc_nix_tm_node_add(struct roc_nix *roc_nix,
  struct roc_nix_tm_node *roc_node);
 int __roc_api roc_nix_tm_node_delete(struct roc_nix *roc_nix, uint32_t node_id,
@@ -408,6 +428,9 @@ roc_nix_tm_shaper_profile_get(struct roc_nix *roc_nix, 
uint32_t profile_id);
 struct roc_nix_tm_shaper_profile *__roc_api roc_nix_tm_shaper_profile_next(
struct roc_nix *roc_nix, struct roc_nix_tm_shaper_profile *__prev);
 
+int __roc_api roc_nix_tm_node_stats_get(struct roc_nix *roc_nix,
+   uint32_t node_id, bool clear,
+   struct roc_nix_tm_node_stats *stats);
 /*
  * TM ratelimit tree API.
  */
diff --git a/drivers/common/cnxk/roc_nix_queue.c 
b/drivers/common/cnxk/roc_nix_queue.c
index fbf7efa..1c62aa2 100644
--- a/drivers/common/cnxk/roc_nix_queue.c
+++ b/drivers/common/cnxk/roc_nix_queue.c
@@ -582,6 +582,7 @@ sq_cn9k_init(struct nix *nix, struct roc_nix_sq *sq, 
uint32_t rr_quantum,
aq->sq.default_chan = nix->tx_chan_base;
aq->sq.sqe_stype = NIX_STYPE_STF;
aq->sq.ena = 1;
+   aq->sq.sso_ena = !!sq->sso_ena;
if (aq->sq.max_sqe_size == NIX_MAXSQESZ_W8)
aq->sq.sqe_stype = NIX_STYPE_STP;
aq->sq.sqb_aura = roc_npa_aura_handle_to_aura(sq->aura_handle);
@@ -679,6 +680,7 @@ sq_init(struct nix *nix, struct roc_nix_sq *sq, uint32_t 
rr_quantum,
aq->sq.default_chan = nix->tx_chan_base;
aq->sq.sqe_stype = NIX_STYPE_STF;
aq->sq.ena = 1;
+   aq->sq.sso_ena = !!sq->sso_ena;
if (aq->sq.max_sqe_size == NIX_MAXSQESZ_W8)
aq->sq.sqe_stype = NIX_STYPE_STP;
aq->sq.sqb_aura = roc_npa_aura_handle_to_aura(sq->aura_handle);
diff --git a/drivers/common/cnxk/roc_nix_rss.c 
b/drivers/common/cnxk/roc_nix_rss.c
index 2d7b84a..7de69aa 100644
--- a/drivers/common/cnxk/roc_nix_rss.c
+++ b/drivers/common/cnxk/roc_nix_rss.c
@@ -52,7 +52,7 @@ roc_nix_rss_key_get(struct roc_nix *roc_nix, uint8_t 
key[ROC_NIX_RSS_KEY_LEN])
 
 static int
 nix_cn9k_rss_reta_set(struct nix *nix, uint8_t group,
- uint16_t reta[ROC_NIX_RSS_RETA_MAX])
+ uint16_t reta[ROC_NIX_RSS_RETA_MAX], uint8_t lock_rx_ctx)
 {
struct mbox *mbox = (&nix->dev)->mbox;
struct nix_aq_enq_req *req;
@@ -77,6 +77,27 @@

[dpdk-dev] [PATCH v3 02/62] common/cnxk: fix batch alloc completion poll logic

2021-06-18 Thread Nithin Dabilpuram
From: Jerin Jacob 

The instruction generation was not correct due to
fact that volatile suppose to use with ccode variable
as well.

Change the logic to use gcc atomic builtin to
simplify and avoid explicit volatile from the code.

Fixes: 81af26789316 ("common/cnxk: support NPA batch alloc/free")

Signed-off-by: Jerin Jacob 
Signed-off-by: Ashwin Sekhar T K 
---
 drivers/common/cnxk/roc_npa.c |  2 +-
 drivers/common/cnxk/roc_npa.h | 30 +++---
 2 files changed, 16 insertions(+), 16 deletions(-)

diff --git a/drivers/common/cnxk/roc_npa.c b/drivers/common/cnxk/roc_npa.c
index f1e03b7..5ba6e81 100644
--- a/drivers/common/cnxk/roc_npa.c
+++ b/drivers/common/cnxk/roc_npa.c
@@ -236,7 +236,7 @@ npa_aura_pool_pair_alloc(struct npa_lf *lf, const uint32_t 
block_size,
 
/* Block size should be cache line aligned and in range of 128B-128KB */
if (block_size % ROC_ALIGN || block_size < 128 ||
-   block_size > 128 * 1024)
+   block_size > ROC_NPA_MAX_BLOCK_SZ)
return NPA_ERR_INVALID_BLOCK_SZ;
 
pos = 0;
diff --git a/drivers/common/cnxk/roc_npa.h b/drivers/common/cnxk/roc_npa.h
index 89f5c6f..59d6223 100644
--- a/drivers/common/cnxk/roc_npa.h
+++ b/drivers/common/cnxk/roc_npa.h
@@ -8,6 +8,7 @@
 #define ROC_AURA_ID_MASK   (BIT_ULL(16) - 1)
 #define ROC_AURA_OP_LIMIT_MASK (BIT_ULL(36) - 1)
 
+#define ROC_NPA_MAX_BLOCK_SZ  (128 * 1024)
 #define ROC_CN10K_NPA_BATCH_ALLOC_MAX_PTRS 512
 #define ROC_CN10K_NPA_BATCH_FREE_MAX_PTRS  15
 
@@ -219,6 +220,17 @@ roc_npa_aura_batch_alloc_issue(uint64_t aura_handle, 
uint64_t *buf,
return 0;
 }
 
+static inline void
+roc_npa_batch_alloc_wait(uint64_t *cache_line)
+{
+   /* Batch alloc status code is updated in bits [5:6] of the first word
+* of the 128 byte cache line.
+*/
+   while (((__atomic_load_n(cache_line, __ATOMIC_RELAXED) >> 5) & 0x3) ==
+  ALLOC_CCODE_INVAL)
+   ;
+}
+
 static inline unsigned int
 roc_npa_aura_batch_alloc_count(uint64_t *aligned_buf, unsigned int num)
 {
@@ -231,17 +243,10 @@ roc_npa_aura_batch_alloc_count(uint64_t *aligned_buf, 
unsigned int num)
/* Check each ROC cache line one by one */
for (i = 0; i < num; i += (ROC_ALIGN >> 3)) {
struct npa_batch_alloc_status_s *status;
-   int ccode;
 
status = (struct npa_batch_alloc_status_s *)&aligned_buf[i];
 
-   /* Status is updated in first 7 bits of each 128 byte cache
-* line. Wait until the status gets updated.
-*/
-   do {
-   ccode = (volatile int)status->ccode;
-   } while (ccode == ALLOC_CCODE_INVAL);
-
+   roc_npa_batch_alloc_wait(&aligned_buf[i]);
count += status->count;
}
 
@@ -261,16 +266,11 @@ roc_npa_aura_batch_alloc_extract(uint64_t *buf, uint64_t 
*aligned_buf,
/* Check each ROC cache line one by one */
for (i = 0; i < num; i += (ROC_ALIGN >> 3)) {
struct npa_batch_alloc_status_s *status;
-   int line_count, ccode;
+   int line_count;
 
status = (struct npa_batch_alloc_status_s *)&aligned_buf[i];
 
-   /* Status is updated in first 7 bits of each 128 byte cache
-* line. Wait until the status gets updated.
-*/
-   do {
-   ccode = (volatile int)status->ccode;
-   } while (ccode == ALLOC_CCODE_INVAL);
+   roc_npa_batch_alloc_wait(&aligned_buf[i]);
 
line_count = status->count;
 
-- 
2.8.4



[dpdk-dev] [PATCH v3 03/62] common/cnxk: add support to dump flow entries

2021-06-18 Thread Nithin Dabilpuram
From: Satheesh Paul 

Add NPC support API to dump created flow entries.

Signed-off-by: Satheesh Paul 
---
 drivers/common/cnxk/hw/npc.h|   2 +
 drivers/common/cnxk/meson.build |   1 +
 drivers/common/cnxk/roc_npc.c   |  20 ++
 drivers/common/cnxk/roc_npc.h   |  12 +-
 drivers/common/cnxk/roc_npc_mcam_dump.c | 611 
 drivers/common/cnxk/roc_npc_priv.h  |   2 +-
 drivers/common/cnxk/roc_npc_utils.c |   4 +
 drivers/common/cnxk/version.map |   2 +
 8 files changed, 652 insertions(+), 2 deletions(-)
 create mode 100644 drivers/common/cnxk/roc_npc_mcam_dump.c

diff --git a/drivers/common/cnxk/hw/npc.h b/drivers/common/cnxk/hw/npc.h
index e0f06bf..68c5037 100644
--- a/drivers/common/cnxk/hw/npc.h
+++ b/drivers/common/cnxk/hw/npc.h
@@ -193,6 +193,7 @@ enum npc_kpu_lb_ltype {
NPC_LT_LB_EXDSA,
NPC_LT_LB_EXDSA_VLAN,
NPC_LT_LB_FDSA,
+   NPC_LT_LB_VLAN_EXDSA,
NPC_LT_LB_CUSTOM0 = 0xE,
NPC_LT_LB_CUSTOM1 = 0xF,
 };
@@ -208,6 +209,7 @@ enum npc_kpu_lc_ltype {
NPC_LT_LC_MPLS,
NPC_LT_LC_NSH,
NPC_LT_LC_FCOE,
+   NPC_LT_LC_NGIO,
NPC_LT_LC_CUSTOM0 = 0xE,
NPC_LT_LC_CUSTOM1 = 0xF,
 };
diff --git a/drivers/common/cnxk/meson.build b/drivers/common/cnxk/meson.build
index 178bce7..e7ab79f 100644
--- a/drivers/common/cnxk/meson.build
+++ b/drivers/common/cnxk/meson.build
@@ -37,6 +37,7 @@ sources = files(
 'roc_npa_irq.c',
 'roc_npc.c',
 'roc_npc_mcam.c',
+'roc_npc_mcam_dump.c',
 'roc_npc_parse.c',
 'roc_npc_utils.c',
 'roc_platform.c',
diff --git a/drivers/common/cnxk/roc_npc.c b/drivers/common/cnxk/roc_npc.c
index abaef77..81c7fd9 100644
--- a/drivers/common/cnxk/roc_npc.c
+++ b/drivers/common/cnxk/roc_npc.c
@@ -870,3 +870,23 @@ roc_npc_flow_destroy(struct roc_npc *roc_npc, struct 
roc_npc_flow *flow)
plt_free(flow);
return 0;
 }
+
+void
+roc_npc_flow_dump(FILE *file, struct roc_npc *roc_npc)
+{
+   struct npc *npc = roc_npc_to_npc_priv(roc_npc);
+   struct roc_npc_flow *flow_iter;
+   struct npc_flow_list *list;
+   uint32_t max_prio, i;
+
+   max_prio = npc->flow_max_priority;
+
+   for (i = 0; i < max_prio; i++) {
+   list = &npc->flow_list[i];
+
+   /* List in ascending order of mcam entries */
+   TAILQ_FOREACH(flow_iter, list, next) {
+   roc_npc_flow_mcam_dump(file, roc_npc, flow_iter);
+   }
+   }
+}
diff --git a/drivers/common/cnxk/roc_npc.h b/drivers/common/cnxk/roc_npc.h
index 223c4ba..115bcd5 100644
--- a/drivers/common/cnxk/roc_npc.h
+++ b/drivers/common/cnxk/roc_npc.h
@@ -90,6 +90,11 @@ struct roc_npc_attr {
uint32_t reserved : 30; /**< Reserved, must be zero. */
 };
 
+struct roc_npc_flow_dump_data {
+   uint8_t lid;
+   uint16_t ltype;
+};
+
 struct roc_npc_flow {
uint8_t nix_intf;
uint8_t enable;
@@ -102,6 +107,9 @@ struct roc_npc_flow {
uint64_t mcam_mask[ROC_NPC_MAX_MCAM_WIDTH_DWORDS];
uint64_t npc_action;
uint64_t vtag_action;
+#define ROC_NPC_MAX_FLOW_PATTERNS 32
+   struct roc_npc_flow_dump_data dump_data[ROC_NPC_MAX_FLOW_PATTERNS];
+   uint16_t num_patterns;
 
TAILQ_ENTRY(roc_npc_flow) next;
 };
@@ -185,5 +193,7 @@ int __roc_api roc_npc_mcam_clear_counter(struct roc_npc 
*roc_npc,
 uint32_t ctr_id);
 
 int __roc_api roc_npc_mcam_free_all_resources(struct roc_npc *roc_npc);
-
+void __roc_api roc_npc_flow_dump(FILE *file, struct roc_npc *roc_npc);
+void __roc_api roc_npc_flow_mcam_dump(FILE *file, struct roc_npc *roc_npc,
+ struct roc_npc_flow *mcam);
 #endif /* _ROC_NPC_H_ */
diff --git a/drivers/common/cnxk/roc_npc_mcam_dump.c 
b/drivers/common/cnxk/roc_npc_mcam_dump.c
new file mode 100644
index 000..19b4901
--- /dev/null
+++ b/drivers/common/cnxk/roc_npc_mcam_dump.c
@@ -0,0 +1,611 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+
+#include "roc_api.h"
+#include "roc_priv.h"
+
+#define NPC_MAX_FIELD_NAME_SIZE   80
+#define NPC_RX_ACTIONOP_MASK  GENMASK(3, 0)
+#define NPC_RX_ACTION_PFFUNC_MASK  GENMASK(19, 4)
+#define NPC_RX_ACTION_INDEX_MASK   GENMASK(39, 20)
+#define NPC_RX_ACTION_MATCH_MASK   GENMASK(55, 40)
+#define NPC_RX_ACTION_FLOWKEY_MASK GENMASK(60, 56)
+
+#define NPC_TX_ACTION_INDEX_MASK GENMASK(31, 12)
+#define NPC_TX_ACTION_MATCH_MASK GENMASK(47, 32)
+
+#define NIX_RX_VTAGACT_VTAG0_RELPTR_MASK GENMASK(7, 0)
+#define NIX_RX_VTAGACT_VTAG0_LID_MASK   GENMASK(10, 8)
+#define NIX_RX_VTAGACT_VTAG0_TYPE_MASK  GENMASK(14, 12)
+#define NIX_RX_VTAGACT_VTAG0_VALID_MASK BIT_ULL(15)
+
+#define NIX_RX_VTAGACT_VTAG1_RELPTR_MASK GENMASK(39, 32)
+#define NIX_RX_VTAGACT_VTAG1_LID_MASK   GENMASK(42, 40)
+#define NIX_RX_VTAGACT_VTAG1_TYPE_MASK  GENMASK(46, 44)
+#define NIX_RX_VTAGACT

[dpdk-dev] [PATCH v3 04/62] common/cnxk: support for mark and flag flow actions

2021-06-18 Thread Nithin Dabilpuram
From: Satheesh Paul 

Add roc API to get mark action.

Signed-off-by: Satheesh Paul 
---
 drivers/common/cnxk/roc_npc.c   | 17 +
 drivers/common/cnxk/roc_npc.h   |  3 +++
 drivers/common/cnxk/version.map |  2 ++
 3 files changed, 22 insertions(+)

diff --git a/drivers/common/cnxk/roc_npc.c b/drivers/common/cnxk/roc_npc.c
index 81c7fd9..e6a5036 100644
--- a/drivers/common/cnxk/roc_npc.c
+++ b/drivers/common/cnxk/roc_npc.c
@@ -757,6 +757,23 @@ npc_rss_action_program(struct roc_npc *roc_npc,
return 0;
 }
 
+int
+roc_npc_mark_actions_get(struct roc_npc *roc_npc)
+{
+   struct npc *npc = roc_npc_to_npc_priv(roc_npc);
+
+   return npc->mark_actions;
+}
+
+int
+roc_npc_mark_actions_sub_return(struct roc_npc *roc_npc, uint32_t count)
+{
+   struct npc *npc = roc_npc_to_npc_priv(roc_npc);
+
+   npc->mark_actions -= count;
+   return npc->mark_actions;
+}
+
 struct roc_npc_flow *
 roc_npc_flow_create(struct roc_npc *roc_npc, const struct roc_npc_attr *attr,
const struct roc_npc_item_info pattern[],
diff --git a/drivers/common/cnxk/roc_npc.h b/drivers/common/cnxk/roc_npc.h
index 115bcd5..cf6f732 100644
--- a/drivers/common/cnxk/roc_npc.h
+++ b/drivers/common/cnxk/roc_npc.h
@@ -196,4 +196,7 @@ int __roc_api roc_npc_mcam_free_all_resources(struct 
roc_npc *roc_npc);
 void __roc_api roc_npc_flow_dump(FILE *file, struct roc_npc *roc_npc);
 void __roc_api roc_npc_flow_mcam_dump(FILE *file, struct roc_npc *roc_npc,
  struct roc_npc_flow *mcam);
+int __roc_api roc_npc_mark_actions_get(struct roc_npc *roc_npc);
+int __roc_api roc_npc_mark_actions_sub_return(struct roc_npc *roc_npc,
+ uint32_t count);
 #endif /* _ROC_NPC_H_ */
diff --git a/drivers/common/cnxk/version.map b/drivers/common/cnxk/version.map
index a11ba4d..554459b 100644
--- a/drivers/common/cnxk/version.map
+++ b/drivers/common/cnxk/version.map
@@ -165,6 +165,8 @@ INTERNAL {
roc_npc_flow_parse;
roc_npc_get_low_priority_mcam;
roc_npc_init;
+   roc_npc_mark_actions_get;
+   roc_npc_mark_actions_sub_return;
roc_npc_mcam_alloc_entries;
roc_npc_mcam_alloc_entry;
roc_npc_mcam_clear_counter;
-- 
2.8.4



[dpdk-dev] [PATCH v3 05/62] common/cnxk: allocate lmt region in userspace

2021-06-18 Thread Nithin Dabilpuram
From: Harman Kalra 

As per the new LMTST design, userspace shall allocate lmt region,
setup the DMA translation and share the IOVA with kernel via MBOX.
Kernel will convert this IOVA to physical memory and update the
LMT table entry with the same.
With this new design also shared mode (i.e. all pci funcs sharing
the LMT region allocated by primary/base pci func) is intact.

Signed-off-by: Harman Kalra 
---
 drivers/common/cnxk/roc_api.h  |  2 +
 drivers/common/cnxk/roc_dev.c  | 98 ++
 drivers/common/cnxk/roc_dev_priv.h |  1 +
 drivers/common/cnxk/roc_mbox.h |  3 ++
 drivers/common/cnxk/roc_platform.h | 11 +
 5 files changed, 63 insertions(+), 52 deletions(-)

diff --git a/drivers/common/cnxk/roc_api.h b/drivers/common/cnxk/roc_api.h
index 67f5d13..32e383c 100644
--- a/drivers/common/cnxk/roc_api.h
+++ b/drivers/common/cnxk/roc_api.h
@@ -24,6 +24,8 @@
 /* Platform definition */
 #include "roc_platform.h"
 
+#define ROC_LMT_LINE_SZ128
+#define ROC_NUM_LMT_LINES  2048
 #define ROC_LMT_LINES_PER_CORE_LOG2 5
 #define ROC_LMT_LINE_SIZE_LOG2 7
 #define ROC_LMT_BASE_PER_CORE_LOG2 
\
diff --git a/drivers/common/cnxk/roc_dev.c b/drivers/common/cnxk/roc_dev.c
index a39acc9..adff779 100644
--- a/drivers/common/cnxk/roc_dev.c
+++ b/drivers/common/cnxk/roc_dev.c
@@ -915,43 +915,30 @@ dev_vf_mbase_put(struct plt_pci_device *pci_dev, 
uintptr_t vf_mbase)
mbox_mem_unmap((void *)vf_mbase, MBOX_SIZE * pci_dev->max_vfs);
 }
 
-static uint16_t
-dev_pf_total_vfs(struct plt_pci_device *pci_dev)
-{
-   uint16_t total_vfs = 0;
-   int sriov_pos, rc;
-
-   sriov_pos =
-   plt_pci_find_ext_capability(pci_dev, ROC_PCI_EXT_CAP_ID_SRIOV);
-   if (sriov_pos <= 0) {
-   plt_warn("Unable to find SRIOV cap, rc=%d", sriov_pos);
-   return 0;
-   }
-
-   rc = plt_pci_read_config(pci_dev, &total_vfs, 2,
-sriov_pos + ROC_PCI_SRIOV_TOTAL_VF);
-   if (rc < 0) {
-   plt_warn("Unable to read SRIOV cap, rc=%d", rc);
-   return 0;
-   }
-
-   return total_vfs;
-}
-
 static int
-dev_setup_shared_lmt_region(struct mbox *mbox)
+dev_setup_shared_lmt_region(struct mbox *mbox, bool valid_iova, uint64_t iova)
 {
struct lmtst_tbl_setup_req *req;
 
req = mbox_alloc_msg_lmtst_tbl_setup(mbox);
-   req->pcifunc = idev_lmt_pffunc_get();
+   /* This pcifunc is defined with primary pcifunc whose LMT address
+* will be shared. If call contains valid IOVA, following pcifunc
+* field is of no use.
+*/
+   req->pcifunc = valid_iova ? 0 : idev_lmt_pffunc_get();
+   req->use_local_lmt_region = valid_iova;
+   req->lmt_iova = iova;
 
return mbox_process(mbox);
 }
 
+/* Total no of lines * size of each lmtline */
+#define LMT_REGION_SIZE (ROC_NUM_LMT_LINES * ROC_LMT_LINE_SZ)
 static int
-dev_lmt_setup(struct plt_pci_device *pci_dev, struct dev *dev)
+dev_lmt_setup(struct dev *dev)
 {
+   char name[PLT_MEMZONE_NAMESIZE];
+   const struct plt_memzone *mz;
struct idev_cfg *idev;
int rc;
 
@@ -965,8 +952,11 @@ dev_lmt_setup(struct plt_pci_device *pci_dev, struct dev 
*dev)
/* Set common lmt region from second pf_func onwards. */
if (!dev->disable_shared_lmt && idev_lmt_pffunc_get() &&
dev->pf_func != idev_lmt_pffunc_get()) {
-   rc = dev_setup_shared_lmt_region(dev->mbox);
+   rc = dev_setup_shared_lmt_region(dev->mbox, false, 0);
if (!rc) {
+   /* On success, updating lmt base of secondary pf_funcs
+* with primary pf_func's lmt base.
+*/
dev->lmt_base = roc_idev_lmt_base_addr_get();
return rc;
}
@@ -975,34 +965,30 @@ dev_lmt_setup(struct plt_pci_device *pci_dev, struct dev 
*dev)
dev->pf_func, rc);
}
 
-   if (dev_is_vf(dev)) {
-   /* VF BAR4 should always be sufficient enough to
-* hold LMT lines.
-*/
-   if (pci_dev->mem_resource[4].len <
-   (RVU_LMT_LINE_MAX * RVU_LMT_SZ)) {
-   plt_err("Not enough bar4 space for lmt lines");
-   return -EFAULT;
-   }
+   /* Allocating memory for LMT region */
+   sprintf(name, "LMT_MAP%x", dev->pf_func);
 
-   dev->lmt_base = dev->bar4;
-   } else {
-   uint64_t bar4_mbox_sz = MBOX_SIZE;
-
-   /* PF BAR4 should always be sufficient enough to
-* hold PF-AF MBOX + PF-VF MBOX + LMT lines.
-*/
-   if (pci_dev->mem_resource[4].len <
-   (bar4_mbox_sz + (RVU_LMT_LINE_MAX * RVU_LMT_SZ))) {
-   plt_err("Not enough b

[dpdk-dev] [PATCH v3 06/62] common/cnxk: add provision to enable RED on RQ

2021-06-18 Thread Nithin Dabilpuram
From: Satha Rao 

Send RED pass/drop levels based on rq configurations to kernel.
Fixed the aura and pool shift value calculation.

Signed-off-by: Satha Rao 
---
 drivers/common/cnxk/roc_nix.h   |  8 ++
 drivers/common/cnxk/roc_nix_queue.c | 50 +
 drivers/common/cnxk/roc_npa.c   |  8 --
 drivers/common/cnxk/roc_npa.h   |  5 
 4 files changed, 69 insertions(+), 2 deletions(-)

diff --git a/drivers/common/cnxk/roc_nix.h b/drivers/common/cnxk/roc_nix.h
index 6d9ac10..bb69027 100644
--- a/drivers/common/cnxk/roc_nix.h
+++ b/drivers/common/cnxk/roc_nix.h
@@ -161,6 +161,14 @@ struct roc_nix_rq {
uint32_t vwqe_max_sz_exp;
uint64_t vwqe_wait_tmo;
uint64_t vwqe_aura_handle;
+   /* Average LPB aura level drop threshold for RED */
+   uint8_t red_drop;
+   /* Average LPB aura level pass threshold for RED */
+   uint8_t red_pass;
+   /* Average SPB aura level drop threshold for RED */
+   uint8_t spb_red_drop;
+   /* Average SPB aura level pass threshold for RED */
+   uint8_t spb_red_pass;
/* End of Input parameters */
struct roc_nix *roc_nix;
 };
diff --git a/drivers/common/cnxk/roc_nix_queue.c 
b/drivers/common/cnxk/roc_nix_queue.c
index 1c62aa2..0604e7a 100644
--- a/drivers/common/cnxk/roc_nix_queue.c
+++ b/drivers/common/cnxk/roc_nix_queue.c
@@ -119,6 +119,15 @@ rq_cn9k_cfg(struct nix *nix, struct roc_nix_rq *rq, bool 
cfg, bool ena)
aq->rq.qint_idx = rq->qid % nix->qints;
aq->rq.xqe_drop_ena = 1;
 
+   /* If RED enabled, then fill enable for all cases */
+   if (rq->red_pass && (rq->red_pass >= rq->red_drop)) {
+   aq->rq.spb_aura_pass = rq->spb_red_pass;
+   aq->rq.lpb_aura_pass = rq->red_pass;
+
+   aq->rq.spb_aura_drop = rq->spb_red_drop;
+   aq->rq.lpb_aura_drop = rq->red_drop;
+   }
+
if (cfg) {
if (rq->sso_ena) {
/* SSO mode */
@@ -155,6 +164,14 @@ rq_cn9k_cfg(struct nix *nix, struct roc_nix_rq *rq, bool 
cfg, bool ena)
aq->rq_mask.rq_int_ena = ~aq->rq_mask.rq_int_ena;
aq->rq_mask.qint_idx = ~aq->rq_mask.qint_idx;
aq->rq_mask.xqe_drop_ena = ~aq->rq_mask.xqe_drop_ena;
+
+   if (rq->red_pass && (rq->red_pass >= rq->red_drop)) {
+   aq->rq_mask.spb_aura_pass = ~aq->rq_mask.spb_aura_pass;
+   aq->rq_mask.lpb_aura_pass = ~aq->rq_mask.lpb_aura_pass;
+
+   aq->rq_mask.spb_aura_drop = ~aq->rq_mask.spb_aura_drop;
+   aq->rq_mask.lpb_aura_drop = ~aq->rq_mask.lpb_aura_drop;
+   }
}
 
return 0;
@@ -244,6 +261,23 @@ rq_cfg(struct nix *nix, struct roc_nix_rq *rq, bool cfg, 
bool ena)
aq->rq.qint_idx = rq->qid % nix->qints;
aq->rq.xqe_drop_ena = 1;
 
+   /* If RED enabled, then fill enable for all cases */
+   if (rq->red_pass && (rq->red_pass >= rq->red_drop)) {
+   aq->rq.spb_pool_pass = rq->red_pass;
+   aq->rq.spb_aura_pass = rq->red_pass;
+   aq->rq.lpb_pool_pass = rq->red_pass;
+   aq->rq.lpb_aura_pass = rq->red_pass;
+   aq->rq.wqe_pool_pass = rq->red_pass;
+   aq->rq.xqe_pass = rq->red_pass;
+
+   aq->rq.spb_pool_drop = rq->red_drop;
+   aq->rq.spb_aura_drop = rq->red_drop;
+   aq->rq.lpb_pool_drop = rq->red_drop;
+   aq->rq.lpb_aura_drop = rq->red_drop;
+   aq->rq.wqe_pool_drop = rq->red_drop;
+   aq->rq.xqe_drop = rq->red_drop;
+   }
+
if (cfg) {
if (rq->sso_ena) {
/* SSO mode */
@@ -296,6 +330,22 @@ rq_cfg(struct nix *nix, struct roc_nix_rq *rq, bool cfg, 
bool ena)
aq->rq_mask.rq_int_ena = ~aq->rq_mask.rq_int_ena;
aq->rq_mask.qint_idx = ~aq->rq_mask.qint_idx;
aq->rq_mask.xqe_drop_ena = ~aq->rq_mask.xqe_drop_ena;
+
+   if (rq->red_pass && (rq->red_pass >= rq->red_drop)) {
+   aq->rq_mask.spb_pool_pass = ~aq->rq_mask.spb_pool_pass;
+   aq->rq_mask.spb_aura_pass = ~aq->rq_mask.spb_aura_pass;
+   aq->rq_mask.lpb_pool_pass = ~aq->rq_mask.lpb_pool_pass;
+   aq->rq_mask.lpb_aura_pass = ~aq->rq_mask.lpb_aura_pass;
+   aq->rq_mask.wqe_pool_pass = ~aq->rq_mask.wqe_pool_pass;
+   aq->rq_mask.xqe_pass = ~aq->rq_mask.xqe_pass;
+
+   aq->rq_mask.spb_pool_drop = ~aq->rq_mask.spb_pool_drop;
+   aq->rq_mask.spb_aura_drop = ~aq->rq_mask.spb_aura_drop;
+   aq->rq_mask.lpb_pool_drop = ~aq->rq_mask.lpb_pool_drop;
+   aq->rq_mask.lpb_aura_drop = ~aq->rq_mask.lpb_aura_drop;
+   aq->rq_mask.wqe_pool_drop = ~aq->rq_mask.wqe_poo

[dpdk-dev] [PATCH v3 07/62] common/cnxk: support for VLAN push and pop flow actions

2021-06-18 Thread Nithin Dabilpuram
From: Satheesh Paul 

Add roc API to configure VLAN tag addition and removal.

Signed-off-by: Satheesh Paul 
---
 drivers/common/cnxk/roc_npc.c  | 255 +++--
 drivers/common/cnxk/roc_npc.h  |  24 
 drivers/common/cnxk/roc_npc_mcam.c |   2 +-
 drivers/common/cnxk/roc_npc_priv.h |   1 +
 drivers/common/cnxk/version.map|   2 +
 5 files changed, 269 insertions(+), 15 deletions(-)

diff --git a/drivers/common/cnxk/roc_npc.c b/drivers/common/cnxk/roc_npc.c
index e6a5036..bb55f3d 100644
--- a/drivers/common/cnxk/roc_npc.c
+++ b/drivers/common/cnxk/roc_npc.c
@@ -6,6 +6,23 @@
 #include "roc_priv.h"
 
 int
+roc_npc_vtag_actions_get(struct roc_npc *roc_npc)
+{
+   struct npc *npc = roc_npc_to_npc_priv(roc_npc);
+
+   return npc->vtag_actions;
+}
+
+int
+roc_npc_vtag_actions_sub_return(struct roc_npc *roc_npc, uint32_t count)
+{
+   struct npc *npc = roc_npc_to_npc_priv(roc_npc);
+
+   npc->vtag_actions -= count;
+   return npc->vtag_actions;
+}
+
+int
 roc_npc_mcam_free_counter(struct roc_npc *roc_npc, uint16_t ctr_id)
 {
struct npc *npc = roc_npc_to_npc_priv(roc_npc);
@@ -330,6 +347,7 @@ npc_parse_actions(struct npc *npc, const struct 
roc_npc_attr *attr,
const struct roc_npc_action_mark *act_mark;
const struct roc_npc_action_queue *act_q;
const struct roc_npc_action_vf *vf_act;
+   bool vlan_insert_action = false;
int sel_act, req_act = 0;
uint16_t pf_func, vf_id;
int errcode = 0;
@@ -417,25 +435,69 @@ npc_parse_actions(struct npc *npc, const struct 
roc_npc_attr *attr,
req_act |= ROC_NPC_ACTION_TYPE_SEC;
rq = 0;
break;
+   case ROC_NPC_ACTION_TYPE_VLAN_STRIP:
+   req_act |= ROC_NPC_ACTION_TYPE_VLAN_STRIP;
+   break;
+   case ROC_NPC_ACTION_TYPE_VLAN_INSERT:
+   req_act |= ROC_NPC_ACTION_TYPE_VLAN_INSERT;
+   break;
+   case ROC_NPC_ACTION_TYPE_VLAN_ETHTYPE_INSERT:
+   req_act |= ROC_NPC_ACTION_TYPE_VLAN_ETHTYPE_INSERT;
+   break;
+   case ROC_NPC_ACTION_TYPE_VLAN_PCP_INSERT:
+   req_act |= ROC_NPC_ACTION_TYPE_VLAN_PCP_INSERT;
+   break;
default:
errcode = NPC_ERR_ACTION_NOTSUP;
goto err_exit;
}
}
 
+   if (req_act & (ROC_NPC_ACTION_TYPE_VLAN_INSERT |
+  ROC_NPC_ACTION_TYPE_VLAN_ETHTYPE_INSERT |
+  ROC_NPC_ACTION_TYPE_VLAN_PCP_INSERT))
+   vlan_insert_action = true;
+
+   if ((req_act & (ROC_NPC_ACTION_TYPE_VLAN_INSERT |
+   ROC_NPC_ACTION_TYPE_VLAN_ETHTYPE_INSERT |
+   ROC_NPC_ACTION_TYPE_VLAN_PCP_INSERT)) ==
+   ROC_NPC_ACTION_TYPE_VLAN_PCP_INSERT) {
+   plt_err("PCP insert action can't be supported alone");
+   errcode = NPC_ERR_ACTION_NOTSUP;
+   goto err_exit;
+   }
+
+   /* Both STRIP and INSERT actions are not supported */
+   if (vlan_insert_action && (req_act & ROC_NPC_ACTION_TYPE_VLAN_STRIP)) {
+   errcode = NPC_ERR_ACTION_NOTSUP;
+   goto err_exit;
+   }
+
/* Check if actions specified are compatible */
if (attr->egress) {
-   /* Only DROP/COUNT is supported */
-   if (!(req_act & ROC_NPC_ACTION_TYPE_DROP)) {
+   if (req_act & ROC_NPC_ACTION_TYPE_VLAN_STRIP) {
+   plt_err("VLAN pop action is not supported on Egress");
errcode = NPC_ERR_ACTION_NOTSUP;
goto err_exit;
-   } else if (req_act & ~(ROC_NPC_ACTION_TYPE_DROP |
-  ROC_NPC_ACTION_TYPE_COUNT)) {
+   }
+
+   if (req_act & ROC_NPC_ACTION_TYPE_DROP) {
+   flow->npc_action = NIX_TX_ACTIONOP_DROP;
+   } else if ((req_act & ROC_NPC_ACTION_TYPE_COUNT) ||
+  vlan_insert_action) {
+   flow->npc_action = NIX_TX_ACTIONOP_UCAST_DEFAULT;
+   } else {
+   plt_err("Unsupported action for egress");
errcode = NPC_ERR_ACTION_NOTSUP;
goto err_exit;
}
-   flow->npc_action = NIX_TX_ACTIONOP_DROP;
+
goto set_pf_func;
+   } else {
+   if (vlan_insert_action) {
+   errcode = NPC_ERR_ACTION_NOTSUP;
+   goto err_exit;
+   }
}
 
/* We have already verified the attr, this is ingress.
@@ -463,6 +525,13 @@ npc_parse_actions(struct npc *npc, const struct 
roc_npc_attr *attr,
goto err_exit;
}
 
+   if (req_act & ROC_NPC_

[dpdk-dev] [PATCH v3 08/62] common/cnxk: fix flow create on CN98xx

2021-06-18 Thread Nithin Dabilpuram
From: Satheesh Paul 

CN96xx and CN98xx have 4096 and 16384 MCAM entries respectively.
Aligning the code with the same numbers.

Fixes: a07f7ced436d ("common/cnxk: add NPC init and fini")

Signed-off-by: Satheesh Paul 
---
 drivers/common/cnxk/roc_model.h | 6 ++
 drivers/common/cnxk/roc_npc.c   | 2 +-
 2 files changed, 7 insertions(+), 1 deletion(-)

diff --git a/drivers/common/cnxk/roc_model.h b/drivers/common/cnxk/roc_model.h
index fb774ac..d6ef459 100644
--- a/drivers/common/cnxk/roc_model.h
+++ b/drivers/common/cnxk/roc_model.h
@@ -88,6 +88,12 @@ roc_model_is_cn10k(void)
 }
 
 static inline uint64_t
+roc_model_is_cn98xx(void)
+{
+   return (roc_model->flag & ROC_MODEL_CN98xx_A0);
+}
+
+static inline uint64_t
 roc_model_is_cn96_A0(void)
 {
return roc_model->flag & ROC_MODEL_CN96xx_A0;
diff --git a/drivers/common/cnxk/roc_npc.c b/drivers/common/cnxk/roc_npc.c
index bb55f3d..8a76823 100644
--- a/drivers/common/cnxk/roc_npc.c
+++ b/drivers/common/cnxk/roc_npc.c
@@ -118,7 +118,7 @@ npc_mcam_tot_entries(void)
/* FIXME: change to reading in AF from NPC_AF_CONST1/2
 * MCAM_BANK_DEPTH(_EXT) * MCAM_BANKS
 */
-   if (roc_model_is_cn10k())
+   if (roc_model_is_cn10k() || roc_model_is_cn98xx())
return 16 * 1024; /* MCAM_BANKS = 4, BANK_DEPTH_EXT = 4096 */
else
return 4 * 1024; /* MCAM_BANKS = 4, BANK_DEPTH_EXT = 1024 */
-- 
2.8.4



[dpdk-dev] [PATCH v3 09/62] net/cnxk: add build infra and common probe

2021-06-18 Thread Nithin Dabilpuram
Add build infrastructure and common probe and remove for cnxk driver
which is used by both CN10K and CN9K SoC.

Signed-off-by: Nithin Dabilpuram 
---
 MAINTAINERS|   5 +-
 doc/guides/nics/cnxk.rst   |  29 +
 doc/guides/nics/features/cnxk.ini  |   9 ++
 doc/guides/nics/features/cnxk_vec.ini  |   9 ++
 doc/guides/nics/features/cnxk_vf.ini   |   9 ++
 doc/guides/nics/index.rst  |   1 +
 doc/guides/platform/cnxk.rst   |   3 +
 doc/guides/rel_notes/release_21_08.rst |   5 +
 drivers/net/cnxk/cnxk_ethdev.c | 218 +
 drivers/net/cnxk/cnxk_ethdev.h |  57 +
 drivers/net/cnxk/meson.build   |  14 +++
 drivers/net/cnxk/version.map   |   3 +
 drivers/net/meson.build|   1 +
 13 files changed, 362 insertions(+), 1 deletion(-)
 create mode 100644 doc/guides/nics/cnxk.rst
 create mode 100644 doc/guides/nics/features/cnxk.ini
 create mode 100644 doc/guides/nics/features/cnxk_vec.ini
 create mode 100644 doc/guides/nics/features/cnxk_vf.ini
 create mode 100644 drivers/net/cnxk/cnxk_ethdev.c
 create mode 100644 drivers/net/cnxk/cnxk_ethdev.h
 create mode 100644 drivers/net/cnxk/meson.build
 create mode 100644 drivers/net/cnxk/version.map

diff --git a/MAINTAINERS b/MAINTAINERS
index 5877a16..b39a1c2 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -745,8 +745,11 @@ M: Kiran Kumar K 
 M: Sunil Kumar Kori 
 M: Satha Rao 
 T: git://dpdk.org/next/dpdk-next-net-mrvl
-F: drivers/common/cnxk/
+F: doc/guides/nics/cnxk.rst
+F: doc/guides/nics/features/cnxk*.ini
 F: doc/guides/platform/cnxk.rst
+F: drivers/common/cnxk/
+F: drivers/net/cnxk/
 
 Marvell mvpp2
 M: Liron Himi 
diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
new file mode 100644
index 000..ca21842
--- /dev/null
+++ b/doc/guides/nics/cnxk.rst
@@ -0,0 +1,29 @@
+..  SPDX-License-Identifier: BSD-3-Clause
+Copyright(C) 2021 Marvell.
+
+CNXK Poll Mode driver
+=
+
+The CNXK ETHDEV PMD (**librte_net_cnxk**) provides poll mode ethdev driver
+support for the inbuilt network device found in **Marvell OCTEON CN9K/CN10K**
+SoC family as well as for their virtual functions (VF) in SR-IOV context.
+
+More information can be found at `Marvell Official Website
+`_.
+
+Features
+
+
+Features of the CNXK Ethdev PMD are:
+
+Prerequisites
+-
+
+See :doc:`../platform/cnxk` for setup information.
+
+
+Driver compilation and testing
+--
+
+Refer to the document :ref:`compiling and testing a PMD for a NIC 
`
+for details.
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
new file mode 100644
index 000..2c23464
--- /dev/null
+++ b/doc/guides/nics/features/cnxk.ini
@@ -0,0 +1,9 @@
+;
+; Supported features of the 'cnxk' network poll mode driver.
+;
+; Refer to default.ini for the full list of available PMD features.
+;
+[Features]
+Linux= Y
+ARMv8= Y
+Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
new file mode 100644
index 000..de78516
--- /dev/null
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -0,0 +1,9 @@
+;
+; Supported features of the 'cnxk_vec' network poll mode driver.
+;
+; Refer to default.ini for the full list of available PMD features.
+;
+[Features]
+Linux= Y
+ARMv8= Y
+Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
new file mode 100644
index 000..9c96351
--- /dev/null
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -0,0 +1,9 @@
+;
+; Supported features of the 'cnxk_vf' network poll mode driver.
+;
+; Refer to default.ini for the full list of available PMD features.
+;
+[Features]
+Linux= Y
+ARMv8= Y
+Usage doc= Y
diff --git a/doc/guides/nics/index.rst b/doc/guides/nics/index.rst
index 799697c..c1a04d9 100644
--- a/doc/guides/nics/index.rst
+++ b/doc/guides/nics/index.rst
@@ -19,6 +19,7 @@ Network Interface Controller Drivers
 axgbe
 bnx2x
 bnxt
+cnxk
 cxgbe
 dpaa
 dpaa2
diff --git a/doc/guides/platform/cnxk.rst b/doc/guides/platform/cnxk.rst
index cebb3d0..b506c11 100644
--- a/doc/guides/platform/cnxk.rst
+++ b/doc/guides/platform/cnxk.rst
@@ -142,6 +142,9 @@ HW Offload Drivers
 
 This section lists dataplane H/W block(s) available in cnxk SoC.
 
+#. **Ethdev Driver**
+   See :doc:`../nics/cnxk` for NIX Ethdev driver information.
+
 #. **Mempool Driver**
See :doc:`../mempool/cnxk` for NPA mempool driver information.
 
diff --git a/doc/guides/rel_notes/release_21_08.rst 
b/doc/guides/rel_notes/release_21_08.rst
index a6ecfdf..31e49e1 100644
--- a/doc/guides/rel_notes/release_21_08.rst
+++ b/doc/guides/rel_notes/release_21_08.rst
@@ -55,6 +55,11 @@ New Features
   

[dpdk-dev] [PATCH v3 10/62] net/cnxk: add platform specific probe and remove

2021-06-18 Thread Nithin Dabilpuram
Add platform specific probe and remove callbacks for CN9K
and CN10K which use common probe and remove functions.
Register ethdev driver for CN9K and CN10K.

Signed-off-by: Nithin Dabilpuram 
---
 drivers/net/cnxk/cn10k_ethdev.c | 64 
 drivers/net/cnxk/cn10k_ethdev.h |  9 +
 drivers/net/cnxk/cn9k_ethdev.c  | 82 +
 drivers/net/cnxk/cn9k_ethdev.h  |  9 +
 drivers/net/cnxk/cnxk_ethdev.c  | 42 +
 drivers/net/cnxk/cnxk_ethdev.h  | 19 ++
 drivers/net/cnxk/meson.build|  5 +++
 7 files changed, 230 insertions(+)
 create mode 100644 drivers/net/cnxk/cn10k_ethdev.c
 create mode 100644 drivers/net/cnxk/cn10k_ethdev.h
 create mode 100644 drivers/net/cnxk/cn9k_ethdev.c
 create mode 100644 drivers/net/cnxk/cn9k_ethdev.h

diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
new file mode 100644
index 000..ff8ce31
--- /dev/null
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -0,0 +1,64 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+#include "cn10k_ethdev.h"
+
+static int
+cn10k_nix_remove(struct rte_pci_device *pci_dev)
+{
+   return cnxk_nix_remove(pci_dev);
+}
+
+static int
+cn10k_nix_probe(struct rte_pci_driver *pci_drv, struct rte_pci_device *pci_dev)
+{
+   struct rte_eth_dev *eth_dev;
+   int rc;
+
+   if (RTE_CACHE_LINE_SIZE != 64) {
+   plt_err("Driver not compiled for CN10K");
+   return -EFAULT;
+   }
+
+   rc = roc_plt_init();
+   if (rc) {
+   plt_err("Failed to initialize platform model, rc=%d", rc);
+   return rc;
+   }
+
+   /* Common probe */
+   rc = cnxk_nix_probe(pci_drv, pci_dev);
+   if (rc)
+   return rc;
+
+   if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
+   eth_dev = rte_eth_dev_allocated(pci_dev->device.name);
+   if (!eth_dev)
+   return -ENOENT;
+   }
+   return 0;
+}
+
+static const struct rte_pci_id cn10k_pci_nix_map[] = {
+   CNXK_PCI_ID(PCI_SUBSYSTEM_DEVID_CN10KA, PCI_DEVID_CNXK_RVU_PF),
+   CNXK_PCI_ID(PCI_SUBSYSTEM_DEVID_CN10KAS, PCI_DEVID_CNXK_RVU_PF),
+   CNXK_PCI_ID(PCI_SUBSYSTEM_DEVID_CN10KA, PCI_DEVID_CNXK_RVU_VF),
+   CNXK_PCI_ID(PCI_SUBSYSTEM_DEVID_CN10KAS, PCI_DEVID_CNXK_RVU_VF),
+   CNXK_PCI_ID(PCI_SUBSYSTEM_DEVID_CN10KA, PCI_DEVID_CNXK_RVU_AF_VF),
+   CNXK_PCI_ID(PCI_SUBSYSTEM_DEVID_CN10KAS, PCI_DEVID_CNXK_RVU_AF_VF),
+   {
+   .vendor_id = 0,
+   },
+};
+
+static struct rte_pci_driver cn10k_pci_nix = {
+   .id_table = cn10k_pci_nix_map,
+   .drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_NEED_IOVA_AS_VA |
+RTE_PCI_DRV_INTR_LSC,
+   .probe = cn10k_nix_probe,
+   .remove = cn10k_nix_remove,
+};
+
+RTE_PMD_REGISTER_PCI(net_cn10k, cn10k_pci_nix);
+RTE_PMD_REGISTER_PCI_TABLE(net_cn10k, cn10k_pci_nix_map);
+RTE_PMD_REGISTER_KMOD_DEP(net_cn10k, "vfio-pci");
diff --git a/drivers/net/cnxk/cn10k_ethdev.h b/drivers/net/cnxk/cn10k_ethdev.h
new file mode 100644
index 000..1bf4a65
--- /dev/null
+++ b/drivers/net/cnxk/cn10k_ethdev.h
@@ -0,0 +1,9 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+#ifndef __CN10K_ETHDEV_H__
+#define __CN10K_ETHDEV_H__
+
+#include 
+
+#endif /* __CN10K_ETHDEV_H__ */
diff --git a/drivers/net/cnxk/cn9k_ethdev.c b/drivers/net/cnxk/cn9k_ethdev.c
new file mode 100644
index 000..701dc12
--- /dev/null
+++ b/drivers/net/cnxk/cn9k_ethdev.c
@@ -0,0 +1,82 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+#include "cn9k_ethdev.h"
+
+static int
+cn9k_nix_remove(struct rte_pci_device *pci_dev)
+{
+   return cnxk_nix_remove(pci_dev);
+}
+
+static int
+cn9k_nix_probe(struct rte_pci_driver *pci_drv, struct rte_pci_device *pci_dev)
+{
+   struct rte_eth_dev *eth_dev;
+   struct cnxk_eth_dev *dev;
+   int rc;
+
+   if (RTE_CACHE_LINE_SIZE != 128) {
+   plt_err("Driver not compiled for CN9K");
+   return -EFAULT;
+   }
+
+   rc = roc_plt_init();
+   if (rc) {
+   plt_err("Failed to initialize platform model, rc=%d", rc);
+   return rc;
+   }
+
+   /* Common probe */
+   rc = cnxk_nix_probe(pci_drv, pci_dev);
+   if (rc)
+   return rc;
+
+   /* Find eth dev allocated */
+   eth_dev = rte_eth_dev_allocated(pci_dev->device.name);
+   if (!eth_dev)
+   return -ENOENT;
+
+   dev = cnxk_eth_pmd_priv(eth_dev);
+   /* Update capabilities already set for TSO.
+* TSO not supported for earlier chip revisions
+*/
+   if (roc_model_is_cn96_A0() || roc_model_is_cn95_A0())
+   dev->tx_offload_capa &= ~(DEV_TX_OFFLOAD_TCP_TSO |
+ DEV_TX_OFFLOAD_VXLAN_TNL_TSO |
+ DEV_TX_OF

[dpdk-dev] [PATCH v3 11/62] net/cnxk: add common devargs parsing function

2021-06-18 Thread Nithin Dabilpuram
Add various devargs parsing command line arguments
parsing functions supported by CN9K and CN10K.

Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/cnxk.rst   |  94 +++
 drivers/net/cnxk/cnxk_ethdev.c |   7 ++
 drivers/net/cnxk/cnxk_ethdev.h |   9 ++
 drivers/net/cnxk/cnxk_ethdev_devargs.c | 166 +
 drivers/net/cnxk/meson.build   |   3 +-
 5 files changed, 278 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cnxk_ethdev_devargs.c

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index ca21842..6652e17 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -27,3 +27,97 @@ Driver compilation and testing
 
 Refer to the document :ref:`compiling and testing a PMD for a NIC 
`
 for details.
+
+Runtime Config Options
+--
+
+- ``Rx&Tx scalar mode enable`` (default ``0``)
+
+   PMD supports both scalar and vector mode, it may be selected at runtime
+   using ``scalar_enable`` ``devargs`` parameter.
+
+- ``RSS reta size`` (default ``64``)
+
+   RSS redirection table size may be configured during runtime using 
``reta_size``
+   ``devargs`` parameter.
+
+   For example::
+
+  -a 0002:02:00.0,reta_size=256
+
+   With the above configuration, reta table of size 256 is populated.
+
+- ``Flow priority levels`` (default ``3``)
+
+   RTE Flow priority levels can be configured during runtime using
+   ``flow_max_priority`` ``devargs`` parameter.
+
+   For example::
+
+  -a 0002:02:00.0,flow_max_priority=10
+
+   With the above configuration, priority level was set to 10 (0-9). Max
+   priority level supported is 32.
+
+- ``Reserve Flow entries`` (default ``8``)
+
+   RTE flow entries can be pre allocated and the size of pre allocation can be
+   selected runtime using ``flow_prealloc_size`` ``devargs`` parameter.
+
+   For example::
+
+  -a 0002:02:00.0,flow_prealloc_size=4
+
+   With the above configuration, pre alloc size was set to 4. Max pre alloc
+   size supported is 32.
+
+- ``Max SQB buffer count`` (default ``512``)
+
+   Send queue descriptor buffer count may be limited during runtime using
+   ``max_sqb_count`` ``devargs`` parameter.
+
+   For example::
+
+  -a 0002:02:00.0,max_sqb_count=64
+
+   With the above configuration, each send queue's descriptor buffer count is
+   limited to a maximum of 64 buffers.
+
+- ``Switch header enable`` (default ``none``)
+
+   A port can be configured to a specific switch header type by using
+   ``switch_header`` ``devargs`` parameter.
+
+   For example::
+
+  -a 0002:02:00.0,switch_header="higig2"
+
+   With the above configuration, higig2 will be enabled on that port and the
+   traffic on this port should be higig2 traffic only. Supported switch header
+   types are "higig2", "dsa", "chlen90b" and "chlen24b".
+
+- ``RSS tag as XOR`` (default ``0``)
+
+   The HW gives two options to configure the RSS adder i.e
+
+   * ``rss_adder<7:0> = flow_tag<7:0> ^ flow_tag<15:8> ^ flow_tag<23:16> ^ 
flow_tag<31:24>``
+
+   * ``rss_adder<7:0> = flow_tag<7:0>``
+
+   Latter one aligns with standard NIC behavior vs former one is a legacy
+   RSS adder scheme used in OCTEON TX2 products.
+
+   By default, the driver runs in the latter mode.
+   Setting this flag to 1 to select the legacy mode.
+
+   For example to select the legacy mode(RSS tag adder as XOR)::
+
+  -a 0002:02:00.0,tag_as_xor=1
+
+
+
+.. note::
+
+   Above devarg parameters are configurable per device, user needs to pass the
+   parameters to all the PCIe devices if application requires to configure on
+   all the ethdev ports.
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 526c19b..109fd35 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -57,6 +57,13 @@ cnxk_eth_dev_init(struct rte_eth_dev *eth_dev)
pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
rte_eth_copy_pci_info(eth_dev, pci_dev);
 
+   /* Parse devargs string */
+   rc = cnxk_ethdev_parse_devargs(eth_dev->device->devargs, dev);
+   if (rc) {
+   plt_err("Failed to parse devargs rc=%d", rc);
+   goto error;
+   }
+
/* Initialize base roc nix */
nix->pci_dev = pci_dev;
rc = roc_nix_dev_init(nix);
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index ba2bfcd..97e3a15 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -9,11 +9,15 @@
 
 #include 
 #include 
+#include 
 
 #include "roc_api.h"
 
 #define CNXK_ETH_DEV_PMD_VERSION "1.0"
 
+/* Max supported SQB count */
+#define CNXK_NIX_TX_MAX_SQB 512
+
 #define CNXK_NIX_TX_OFFLOAD_CAPA   
\
(DEV_TX_OFFLOAD_MBUF_FAST_FREE | DEV_TX_OFFLOAD_MT_LOCKFREE |  \
 DEV_TX_OFFLOAD_VLAN_INSERT | DEV_TX_OFFLOAD_QINQ_INSERT | \
@@ -38,6 +42,7 @@ struct cnxk_eth_dev {
uint

[dpdk-dev] [PATCH v3 12/62] net/cnxk: add common dev infos get support

2021-06-18 Thread Nithin Dabilpuram
Add support to retrieve dev infos get for CN9K and CN10K.

Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/cnxk.rst  |  3 ++
 doc/guides/nics/features/cnxk.ini |  4 ++
 doc/guides/nics/features/cnxk_vec.ini |  4 ++
 doc/guides/nics/features/cnxk_vf.ini  |  3 ++
 drivers/net/cnxk/cnxk_ethdev.c|  4 +-
 drivers/net/cnxk/cnxk_ethdev.h| 33 
 drivers/net/cnxk/cnxk_ethdev_ops.c| 71 +++
 drivers/net/cnxk/meson.build  |  1 +
 8 files changed, 122 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cnxk_ethdev_ops.c

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 6652e17..6bd410b 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -16,6 +16,9 @@ Features
 
 Features of the CNXK Ethdev PMD are:
 
+- SR-IOV VF
+- Lock-free Tx queue
+
 Prerequisites
 -
 
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 2c23464..b426340 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -4,6 +4,10 @@
 ; Refer to default.ini for the full list of available PMD features.
 ;
 [Features]
+Speed capabilities   = Y
+Lock-free Tx queue   = Y
+SR-IOV   = Y
+Multiprocess aware   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index de78516..292ac1e 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -4,6 +4,10 @@
 ; Refer to default.ini for the full list of available PMD features.
 ;
 [Features]
+Speed capabilities   = Y
+Lock-free Tx queue   = Y
+SR-IOV   = Y
+Multiprocess aware   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 9c96351..bc2eb8a 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -4,6 +4,9 @@
 ; Refer to default.ini for the full list of available PMD features.
 ;
 [Features]
+Speed capabilities   = Y
+Lock-free Tx queue   = Y
+Multiprocess aware   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 109fd35..066e01c 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -38,7 +38,9 @@ nix_get_speed_capa(struct cnxk_eth_dev *dev)
 }
 
 /* CNXK platform independent eth dev ops */
-struct eth_dev_ops cnxk_eth_dev_ops;
+struct eth_dev_ops cnxk_eth_dev_ops = {
+   .dev_infos_get = cnxk_nix_info_get,
+};
 
 static int
 cnxk_eth_dev_init(struct rte_eth_dev *eth_dev)
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 97e3a15..8d9a7e0 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -15,9 +15,40 @@
 
 #define CNXK_ETH_DEV_PMD_VERSION "1.0"
 
+/* VLAN tag inserted by NIX_TX_VTAG_ACTION.
+ * In Tx space is always reserved for this in FRS.
+ */
+#define CNXK_NIX_MAX_VTAG_INS 2
+#define CNXK_NIX_MAX_VTAG_ACT_SIZE (4 * CNXK_NIX_MAX_VTAG_INS)
+
+/* ETH_HLEN+ETH_FCS+2*VLAN_HLEN */
+#define CNXK_NIX_L2_OVERHEAD (RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + 8)
+
+#define CNXK_NIX_RX_MIN_DESC   16
+#define CNXK_NIX_RX_MIN_DESC_ALIGN  16
+#define CNXK_NIX_RX_NB_SEG_MAX 6
+#define CNXK_NIX_RX_DEFAULT_RING_SZ 4096
 /* Max supported SQB count */
 #define CNXK_NIX_TX_MAX_SQB 512
 
+/* If PTP is enabled additional SEND MEM DESC is required which
+ * takes 2 words, hence max 7 iova address are possible
+ */
+#if defined(RTE_LIBRTE_IEEE1588)
+#define CNXK_NIX_TX_NB_SEG_MAX 7
+#else
+#define CNXK_NIX_TX_NB_SEG_MAX 9
+#endif
+
+#define CNXK_NIX_RSS_L3_L4_SRC_DST 
\
+   (ETH_RSS_L3_SRC_ONLY | ETH_RSS_L3_DST_ONLY | ETH_RSS_L4_SRC_ONLY | \
+ETH_RSS_L4_DST_ONLY)
+
+#define CNXK_NIX_RSS_OFFLOAD   
\
+   (ETH_RSS_PORT | ETH_RSS_IP | ETH_RSS_UDP | ETH_RSS_TCP |   \
+ETH_RSS_SCTP | ETH_RSS_TUNNEL | ETH_RSS_L2_PAYLOAD |  \
+CNXK_NIX_RSS_L3_L4_SRC_DST | ETH_RSS_LEVEL_MASK | ETH_RSS_C_VLAN)
+
 #define CNXK_NIX_TX_OFFLOAD_CAPA   
\
(DEV_TX_OFFLOAD_MBUF_FAST_FREE | DEV_TX_OFFLOAD_MT_LOCKFREE |  \
 DEV_TX_OFFLOAD_VLAN_INSERT | DEV_TX_OFFLOAD_QINQ_INSERT | \
@@ -77,6 +108,8 @@ extern struct eth_dev_ops cnxk_eth_dev_ops;
 int cnxk_nix_probe(struct rte_pci_driver *pci_drv,
   struct rte_pci_device *pci_dev);
 int cnxk_nix_remove(struct rte_pci_device *pci_dev);
+int cnxk_nix_info_get(struct rte_eth_dev *eth_dev,
+ struct rte_eth_dev_info *dev_info);
 
 /* Devargs */
 int cnxk_ethdev_par

[dpdk-dev] [PATCH v3 13/62] net/cnxk: add device configuration operation

2021-06-18 Thread Nithin Dabilpuram
Add device configuration op for CN9K and CN10K. Most of the
device configuration is common between two platforms except for
some supported offloads.

Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/cnxk.rst  |   2 +
 doc/guides/nics/features/cnxk.ini |   2 +
 doc/guides/nics/features/cnxk_vec.ini |   2 +
 doc/guides/nics/features/cnxk_vf.ini  |   2 +
 drivers/net/cnxk/cn10k_ethdev.c   |  34 +++
 drivers/net/cnxk/cn9k_ethdev.c|  45 +++
 drivers/net/cnxk/cnxk_ethdev.c| 556 ++
 drivers/net/cnxk/cnxk_ethdev.h|  85 ++
 8 files changed, 728 insertions(+)

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 6bd410b..0c2ea89 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -18,6 +18,8 @@ Features of the CNXK Ethdev PMD are:
 
 - SR-IOV VF
 - Lock-free Tx queue
+- Multiple queues for TX and RX
+- Receiver Side Scaling (RSS)
 
 Prerequisites
 -
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index b426340..96dba2a 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -8,6 +8,8 @@ Speed capabilities   = Y
 Lock-free Tx queue   = Y
 SR-IOV   = Y
 Multiprocess aware   = Y
+RSS hash = Y
+Inner RSS= Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 292ac1e..616991c 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -8,6 +8,8 @@ Speed capabilities   = Y
 Lock-free Tx queue   = Y
 SR-IOV   = Y
 Multiprocess aware   = Y
+RSS hash = Y
+Inner RSS= Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index bc2eb8a..a0bd2f1 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -7,6 +7,8 @@
 Speed capabilities   = Y
 Lock-free Tx queue   = Y
 Multiprocess aware   = Y
+RSS hash = Y
+Inner RSS= Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index ff8ce31..d971bbd 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -4,6 +4,38 @@
 #include "cn10k_ethdev.h"
 
 static int
+cn10k_nix_configure(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   int rc;
+
+   /* Common nix configure */
+   rc = cnxk_nix_configure(eth_dev);
+   if (rc)
+   return rc;
+
+   plt_nix_dbg("Configured port%d platform specific rx_offload_flags=%x"
+   " tx_offload_flags=0x%x",
+   eth_dev->data->port_id, dev->rx_offload_flags,
+   dev->tx_offload_flags);
+   return 0;
+}
+
+/* Update platform specific eth dev ops */
+static void
+nix_eth_dev_ops_override(void)
+{
+   static int init_once;
+
+   if (init_once)
+   return;
+   init_once = 1;
+
+   /* Update platform specific ops */
+   cnxk_eth_dev_ops.dev_configure = cn10k_nix_configure;
+}
+
+static int
 cn10k_nix_remove(struct rte_pci_device *pci_dev)
 {
return cnxk_nix_remove(pci_dev);
@@ -26,6 +58,8 @@ cn10k_nix_probe(struct rte_pci_driver *pci_drv, struct 
rte_pci_device *pci_dev)
return rc;
}
 
+   nix_eth_dev_ops_override();
+
/* Common probe */
rc = cnxk_nix_probe(pci_drv, pci_dev);
if (rc)
diff --git a/drivers/net/cnxk/cn9k_ethdev.c b/drivers/net/cnxk/cn9k_ethdev.c
index 701dc12..7f3e910 100644
--- a/drivers/net/cnxk/cn9k_ethdev.c
+++ b/drivers/net/cnxk/cn9k_ethdev.c
@@ -4,6 +4,49 @@
 #include "cn9k_ethdev.h"
 
 static int
+cn9k_nix_configure(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct rte_eth_conf *conf = ð_dev->data->dev_conf;
+   struct rte_eth_txmode *txmode = &conf->txmode;
+   int rc;
+
+   /* Platform specific checks */
+   if ((roc_model_is_cn96_A0() || roc_model_is_cn95_A0()) &&
+   (txmode->offloads & DEV_TX_OFFLOAD_SCTP_CKSUM) &&
+   ((txmode->offloads & DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM) ||
+(txmode->offloads & DEV_TX_OFFLOAD_OUTER_UDP_CKSUM))) {
+   plt_err("Outer IP and SCTP checksum unsupported");
+   return -EINVAL;
+   }
+
+   /* Common nix configure */
+   rc = cnxk_nix_configure(eth_dev);
+   if (rc)
+   return rc;
+
+   plt_nix_dbg("Configured port%d platform specific rx_offload_flags=%x"
+   " tx_offload_flags=0x%x",
+   eth_dev->data->port_id, dev->rx_offload_flags,
+   dev->tx_of

[dpdk-dev] [PATCH v3 14/62] net/cnxk: add link status update support

2021-06-18 Thread Nithin Dabilpuram
Add link status update callback to get current
link status.

Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/cnxk.rst  |   1 +
 doc/guides/nics/features/cnxk.ini |   2 +
 doc/guides/nics/features/cnxk_vec.ini |   2 +
 doc/guides/nics/features/cnxk_vf.ini  |   2 +
 drivers/net/cnxk/cnxk_ethdev.c|   7 +++
 drivers/net/cnxk/cnxk_ethdev.h|   8 +++
 drivers/net/cnxk/cnxk_link.c  | 102 ++
 drivers/net/cnxk/meson.build  |   3 +-
 8 files changed, 126 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cnxk_link.c

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 0c2ea89..7bf6cf5 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -20,6 +20,7 @@ Features of the CNXK Ethdev PMD are:
 - Lock-free Tx queue
 - Multiple queues for TX and RX
 - Receiver Side Scaling (RSS)
+- Link state information
 
 Prerequisites
 -
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 96dba2a..affbbd9 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -8,6 +8,8 @@ Speed capabilities   = Y
 Lock-free Tx queue   = Y
 SR-IOV   = Y
 Multiprocess aware   = Y
+Link status  = Y
+Link status event= Y
 RSS hash = Y
 Inner RSS= Y
 Linux= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 616991c..836cc9f 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -8,6 +8,8 @@ Speed capabilities   = Y
 Lock-free Tx queue   = Y
 SR-IOV   = Y
 Multiprocess aware   = Y
+Link status  = Y
+Link status event= Y
 RSS hash = Y
 Inner RSS= Y
 Linux= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index a0bd2f1..29bb24f 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -7,6 +7,8 @@
 Speed capabilities   = Y
 Lock-free Tx queue   = Y
 Multiprocess aware   = Y
+Link status  = Y
+Link status event= Y
 RSS hash = Y
 Inner RSS= Y
 Linux= Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 14e28fe..002afdc 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -589,6 +589,7 @@ cnxk_nix_configure(struct rte_eth_dev *eth_dev)
 /* CNXK platform independent eth dev ops */
 struct eth_dev_ops cnxk_eth_dev_ops = {
.dev_infos_get = cnxk_nix_info_get,
+   .link_update = cnxk_nix_link_update,
 };
 
 static int
@@ -623,6 +624,9 @@ cnxk_eth_dev_init(struct rte_eth_dev *eth_dev)
goto error;
}
 
+   /* Register up msg callbacks */
+   roc_nix_mac_link_cb_register(nix, cnxk_eth_dev_link_status_cb);
+
dev->eth_dev = eth_dev;
dev->configured = 0;
 
@@ -711,6 +715,9 @@ cnxk_eth_dev_uninit(struct rte_eth_dev *eth_dev, bool 
mbox_close)
 
roc_nix_npc_rx_ena_dis(nix, false);
 
+   /* Disable link status events */
+   roc_nix_mac_link_event_start_stop(nix, false);
+
/* Free up SQs */
for (i = 0; i < eth_dev->data->nb_tx_queues; i++) {
dev_ops->tx_queue_release(eth_dev->data->tx_queues[i]);
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 291f5f9..daa87af 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -15,6 +15,9 @@
 
 #define CNXK_ETH_DEV_PMD_VERSION "1.0"
 
+/* Used for struct cnxk_eth_dev::flags */
+#define CNXK_LINK_CFG_IN_PROGRESS_F BIT_ULL(0)
+
 /* VLAN tag inserted by NIX_TX_VTAG_ACTION.
  * In Tx space is always reserved for this in FRS.
  */
@@ -196,6 +199,11 @@ int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
 uint32_t cnxk_rss_ethdev_to_nix(struct cnxk_eth_dev *dev, uint64_t ethdev_rss,
uint8_t rss_level);
 
+/* Link */
+void cnxk_eth_dev_link_status_cb(struct roc_nix *nix,
+struct roc_nix_link_info *link);
+int cnxk_nix_link_update(struct rte_eth_dev *eth_dev, int wait_to_complete);
+
 /* Devargs */
 int cnxk_ethdev_parse_devargs(struct rte_devargs *devargs,
  struct cnxk_eth_dev *dev);
diff --git a/drivers/net/cnxk/cnxk_link.c b/drivers/net/cnxk/cnxk_link.c
new file mode 100644
index 000..b0273e7
--- /dev/null
+++ b/drivers/net/cnxk/cnxk_link.c
@@ -0,0 +1,102 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+
+#include "cnxk_ethdev.h"
+
+static inline int
+nix_wait_for_link_cfg(struct cnxk_eth_dev *dev)
+{
+   uint16_t wait = 1000;
+
+   do {
+   rte_atomic_thread_fence(__ATOMIC_ACQUIRE);
+   if (!(dev->flags & CNXK_LINK_CFG_IN_PROGRESS_F))
+   break;
+   wait--;
+ 

[dpdk-dev] [PATCH v3 15/62] net/cnxk: add Rx queue setup and release

2021-06-18 Thread Nithin Dabilpuram
Add Rx queue setup and release op for CN9K and CN10K
SoC. Release is completely common while setup is platform
dependent due to fast path Rx queue structure variation.
Fastpath is platform dependent partly due to core cacheline
size difference.

Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/features/cnxk.ini |   1 +
 doc/guides/nics/features/cnxk_vec.ini |   1 +
 doc/guides/nics/features/cnxk_vf.ini  |   1 +
 drivers/net/cnxk/cn10k_ethdev.c   |  44 +
 drivers/net/cnxk/cn10k_ethdev.h   |  14 +++
 drivers/net/cnxk/cn9k_ethdev.c|  44 +
 drivers/net/cnxk/cn9k_ethdev.h|  14 +++
 drivers/net/cnxk/cnxk_ethdev.c| 172 ++
 drivers/net/cnxk/cnxk_ethdev.h|   9 ++
 9 files changed, 300 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index affbbd9..a9d2b03 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -10,6 +10,7 @@ SR-IOV   = Y
 Multiprocess aware   = Y
 Link status  = Y
 Link status event= Y
+Runtime Rx queue setup = Y
 RSS hash = Y
 Inner RSS= Y
 Linux= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 836cc9f..6a8ca1f 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -10,6 +10,7 @@ SR-IOV   = Y
 Multiprocess aware   = Y
 Link status  = Y
 Link status event= Y
+Runtime Rx queue setup = Y
 RSS hash = Y
 Inner RSS= Y
 Linux= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 29bb24f..f761638 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -9,6 +9,7 @@ Lock-free Tx queue   = Y
 Multiprocess aware   = Y
 Link status  = Y
 Link status event= Y
+Runtime Rx queue setup = Y
 RSS hash = Y
 Inner RSS= Y
 Linux= Y
diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index d971bbd..b87c4e5 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -4,6 +4,49 @@
 #include "cn10k_ethdev.h"
 
 static int
+cn10k_nix_rx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
+uint16_t nb_desc, unsigned int socket,
+const struct rte_eth_rxconf *rx_conf,
+struct rte_mempool *mp)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct cn10k_eth_rxq *rxq;
+   struct roc_nix_rq *rq;
+   struct roc_nix_cq *cq;
+   int rc;
+
+   RTE_SET_USED(socket);
+
+   /* CQ Errata needs min 4K ring */
+   if (dev->cq_min_4k && nb_desc < 4096)
+   nb_desc = 4096;
+
+   /* Common Rx queue setup */
+   rc = cnxk_nix_rx_queue_setup(eth_dev, qid, nb_desc,
+sizeof(struct cn10k_eth_rxq), rx_conf, mp);
+   if (rc)
+   return rc;
+
+   rq = &dev->rqs[qid];
+   cq = &dev->cqs[qid];
+
+   /* Update fast path queue */
+   rxq = eth_dev->data->rx_queues[qid];
+   rxq->rq = qid;
+   rxq->desc = (uintptr_t)cq->desc_base;
+   rxq->cq_door = cq->door;
+   rxq->cq_status = cq->status;
+   rxq->wdata = cq->wdata;
+   rxq->head = cq->head;
+   rxq->qmask = cq->qmask;
+
+   /* Data offset from data to start of mbuf is first_skip */
+   rxq->data_off = rq->first_skip;
+   rxq->mbuf_initializer = cnxk_nix_rxq_mbuf_setup(dev);
+   return 0;
+}
+
+static int
 cn10k_nix_configure(struct rte_eth_dev *eth_dev)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
@@ -33,6 +76,7 @@ nix_eth_dev_ops_override(void)
 
/* Update platform specific ops */
cnxk_eth_dev_ops.dev_configure = cn10k_nix_configure;
+   cnxk_eth_dev_ops.rx_queue_setup = cn10k_nix_rx_queue_setup;
 }
 
 static int
diff --git a/drivers/net/cnxk/cn10k_ethdev.h b/drivers/net/cnxk/cn10k_ethdev.h
index 1bf4a65..08e11bb 100644
--- a/drivers/net/cnxk/cn10k_ethdev.h
+++ b/drivers/net/cnxk/cn10k_ethdev.h
@@ -6,4 +6,18 @@
 
 #include 
 
+struct cn10k_eth_rxq {
+   uint64_t mbuf_initializer;
+   uintptr_t desc;
+   void *lookup_mem;
+   uintptr_t cq_door;
+   uint64_t wdata;
+   int64_t *cq_status;
+   uint32_t head;
+   uint32_t qmask;
+   uint32_t available;
+   uint16_t data_off;
+   uint16_t rq;
+} __plt_cache_aligned;
+
 #endif /* __CN10K_ETHDEV_H__ */
diff --git a/drivers/net/cnxk/cn9k_ethdev.c b/drivers/net/cnxk/cn9k_ethdev.c
index 7f3e910..ed9a813 100644
--- a/drivers/net/cnxk/cn9k_ethdev.c
+++ b/drivers/net/cnxk/cn9k_ethdev.c
@@ -4,6 +4,49 @@
 #include "cn9k_ethdev.h"
 
 static int
+cn9k_nix_rx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
+   uint16_t nb_

[dpdk-dev] [PATCH v3 16/62] net/cnxk: add Tx queue setup and release

2021-06-18 Thread Nithin Dabilpuram
aDD tx queue setup and release for CN9K and CN10K.
Release is common while setup is platform dependent due
to differences in fast path Tx queue structures.

Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 doc/guides/nics/features/cnxk_vf.ini  |  1 +
 drivers/net/cnxk/cn10k_ethdev.c   | 72 +
 drivers/net/cnxk/cn10k_ethdev.h   | 13 +
 drivers/net/cnxk/cn10k_tx.h   | 13 +
 drivers/net/cnxk/cn9k_ethdev.c| 70 +
 drivers/net/cnxk/cn9k_ethdev.h| 11 
 drivers/net/cnxk/cn9k_tx.h| 13 +
 drivers/net/cnxk/cnxk_ethdev.c| 98 +++
 drivers/net/cnxk/cnxk_ethdev.h|  3 ++
 11 files changed, 296 insertions(+)
 create mode 100644 drivers/net/cnxk/cn10k_tx.h
 create mode 100644 drivers/net/cnxk/cn9k_tx.h

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index a9d2b03..462d7c4 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -11,6 +11,7 @@ Multiprocess aware   = Y
 Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
+Runtime Tx queue setup = Y
 RSS hash = Y
 Inner RSS= Y
 Linux= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 6a8ca1f..09e0d3a 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -11,6 +11,7 @@ Multiprocess aware   = Y
 Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
+Runtime Tx queue setup = Y
 RSS hash = Y
 Inner RSS= Y
 Linux= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index f761638..4a93a35 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -10,6 +10,7 @@ Multiprocess aware   = Y
 Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
+Runtime Tx queue setup = Y
 RSS hash = Y
 Inner RSS= Y
 Linux= Y
diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index b87c4e5..454c8ca 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -2,6 +2,77 @@
  * Copyright(C) 2021 Marvell.
  */
 #include "cn10k_ethdev.h"
+#include "cn10k_tx.h"
+
+static void
+nix_form_default_desc(struct cnxk_eth_dev *dev, struct cn10k_eth_txq *txq,
+ uint16_t qid)
+{
+   struct nix_send_ext_s *send_hdr_ext;
+   union nix_send_hdr_w0_u send_hdr_w0;
+   union nix_send_sg_s sg_w0;
+
+   RTE_SET_USED(dev);
+
+   /* Initialize the fields based on basic single segment packet */
+   memset(&txq->cmd, 0, sizeof(txq->cmd));
+   send_hdr_w0.u = 0;
+   sg_w0.u = 0;
+
+   if (dev->tx_offload_flags & NIX_TX_NEED_EXT_HDR) {
+   /* 2(HDR) + 2(EXT_HDR) + 1(SG) + 1(IOVA) = 6/2 - 1 = 2 */
+   send_hdr_w0.sizem1 = 2;
+
+   send_hdr_ext = (struct nix_send_ext_s *)&txq->cmd[0];
+   send_hdr_ext->w0.subdc = NIX_SUBDC_EXT;
+   } else {
+   /* 2(HDR) + 1(SG) + 1(IOVA) = 4/2 - 1 = 1 */
+   send_hdr_w0.sizem1 = 1;
+   }
+
+   send_hdr_w0.sq = qid;
+   sg_w0.subdc = NIX_SUBDC_SG;
+   sg_w0.segs = 1;
+   sg_w0.ld_type = NIX_SENDLDTYPE_LDD;
+
+   txq->send_hdr_w0 = send_hdr_w0.u;
+   txq->sg_w0 = sg_w0.u;
+
+   rte_wmb();
+}
+
+static int
+cn10k_nix_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
+uint16_t nb_desc, unsigned int socket,
+const struct rte_eth_txconf *tx_conf)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct cn10k_eth_txq *txq;
+   struct roc_nix_sq *sq;
+   int rc;
+
+   RTE_SET_USED(socket);
+
+   /* Common Tx queue setup */
+   rc = cnxk_nix_tx_queue_setup(eth_dev, qid, nb_desc,
+sizeof(struct cn10k_eth_txq), tx_conf);
+   if (rc)
+   return rc;
+
+   sq = &dev->sqs[qid];
+   /* Update fast path queue */
+   txq = eth_dev->data->tx_queues[qid];
+   txq->fc_mem = sq->fc;
+   /* Store lmt base in tx queue for easy access */
+   txq->lmt_base = dev->nix.lmt_base;
+   txq->io_addr = sq->io_addr;
+   txq->nb_sqb_bufs_adj = sq->nb_sqb_bufs_adj;
+   txq->sqes_per_sqb_log2 = sq->sqes_per_sqb_log2;
+
+   nix_form_default_desc(dev, txq, qid);
+   txq->lso_tun_fmt = dev->lso_tun_fmt;
+   return 0;
+}
 
 static int
 cn10k_nix_rx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
@@ -76,6 +147,7 @@ nix_eth_dev_ops_override(void)
 
/* Update platform specific ops */
cnxk_eth_dev_ops.dev_configure = cn10k_nix_configure;
+   cnxk_

[dpdk-dev] [PATCH v3 17/62] net/cnxk: add packet type support

2021-06-18 Thread Nithin Dabilpuram
Add support for packet type lookup on Rx to translate HW
specific types to  RTE_PTYPE_* defines

Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/cnxk.rst  |   1 +
 doc/guides/nics/features/cnxk.ini |   1 +
 doc/guides/nics/features/cnxk_vec.ini |   1 +
 doc/guides/nics/features/cnxk_vf.ini  |   1 +
 drivers/net/cnxk/cn10k_ethdev.c   |  21 +++
 drivers/net/cnxk/cn10k_rx.h   |  11 ++
 drivers/net/cnxk/cn9k_ethdev.c|  21 +++
 drivers/net/cnxk/cn9k_rx.h|  12 ++
 drivers/net/cnxk/cnxk_ethdev.c|   2 +
 drivers/net/cnxk/cnxk_ethdev.h|  14 ++
 drivers/net/cnxk/cnxk_lookup.c| 326 ++
 drivers/net/cnxk/meson.build  |   3 +-
 12 files changed, 413 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cn10k_rx.h
 create mode 100644 drivers/net/cnxk/cn9k_rx.h
 create mode 100644 drivers/net/cnxk/cnxk_lookup.c

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 7bf6cf5..8bc85c0 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -16,6 +16,7 @@ Features
 
 Features of the CNXK Ethdev PMD are:
 
+- Packet type information
 - SR-IOV VF
 - Lock-free Tx queue
 - Multiple queues for TX and RX
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 462d7c4..503582c 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -14,6 +14,7 @@ Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
 RSS hash = Y
 Inner RSS= Y
+Packet type parsing  = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 09e0d3a..9ad225a 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -14,6 +14,7 @@ Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
 RSS hash = Y
 Inner RSS= Y
+Packet type parsing  = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 4a93a35..8c93ba7 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -13,6 +13,7 @@ Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
 RSS hash = Y
 Inner RSS= Y
+Packet type parsing  = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index 454c8ca..f79d03c 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -2,8 +2,25 @@
  * Copyright(C) 2021 Marvell.
  */
 #include "cn10k_ethdev.h"
+#include "cn10k_rx.h"
 #include "cn10k_tx.h"
 
+static int
+cn10k_nix_ptypes_set(struct rte_eth_dev *eth_dev, uint32_t ptype_mask)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   if (ptype_mask) {
+   dev->rx_offload_flags |= NIX_RX_OFFLOAD_PTYPE_F;
+   dev->ptype_disable = 0;
+   } else {
+   dev->rx_offload_flags &= ~NIX_RX_OFFLOAD_PTYPE_F;
+   dev->ptype_disable = 1;
+   }
+
+   return 0;
+}
+
 static void
 nix_form_default_desc(struct cnxk_eth_dev *dev, struct cn10k_eth_txq *txq,
  uint16_t qid)
@@ -114,6 +131,9 @@ cn10k_nix_rx_queue_setup(struct rte_eth_dev *eth_dev, 
uint16_t qid,
/* Data offset from data to start of mbuf is first_skip */
rxq->data_off = rq->first_skip;
rxq->mbuf_initializer = cnxk_nix_rxq_mbuf_setup(dev);
+
+   /* Lookup mem */
+   rxq->lookup_mem = cnxk_nix_fastpath_lookup_mem_get();
return 0;
 }
 
@@ -149,6 +169,7 @@ nix_eth_dev_ops_override(void)
cnxk_eth_dev_ops.dev_configure = cn10k_nix_configure;
cnxk_eth_dev_ops.tx_queue_setup = cn10k_nix_tx_queue_setup;
cnxk_eth_dev_ops.rx_queue_setup = cn10k_nix_rx_queue_setup;
+   cnxk_eth_dev_ops.dev_ptypes_set = cn10k_nix_ptypes_set;
 }
 
 static int
diff --git a/drivers/net/cnxk/cn10k_rx.h b/drivers/net/cnxk/cn10k_rx.h
new file mode 100644
index 000..d3d1661
--- /dev/null
+++ b/drivers/net/cnxk/cn10k_rx.h
@@ -0,0 +1,11 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+#ifndef __CN10K_RX_H__
+#define __CN10K_RX_H__
+
+#include 
+
+#define NIX_RX_OFFLOAD_PTYPE_F  BIT(1)
+
+#endif /* __CN10K_RX_H__ */
diff --git a/drivers/net/cnxk/cn9k_ethdev.c b/drivers/net/cnxk/cn9k_ethdev.c
index a8ac2c5..b5a4bd7 100644
--- a/drivers/net/cnxk/cn9k_ethdev.c
+++ b/drivers/net/cnxk/cn9k_ethdev.c
@@ -2,8 +2,25 @@
  * Copyright(C) 2021 Marvell.
  */
 #include "cn9k_ethdev.h"
+#include "cn9k_rx.h"
 #include "cn9k_tx.h"
 
+static int
+cn9k_nix_ptypes_set(struct rte_eth_dev *eth_dev, uint32_t ptype_mask)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   if

[dpdk-dev] [PATCH v3 18/62] net/cnxk: add queue start and stop support

2021-06-18 Thread Nithin Dabilpuram
Add Rx/Tx queue start and stop callbacks for
CN9K and CN10K.

Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 doc/guides/nics/features/cnxk_vf.ini  |  1 +
 drivers/net/cnxk/cn10k_ethdev.c   | 16 ++
 drivers/net/cnxk/cn9k_ethdev.c| 16 ++
 drivers/net/cnxk/cnxk_ethdev.c| 92 +++
 drivers/net/cnxk/cnxk_ethdev.h|  1 +
 7 files changed, 128 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 503582c..712f8d5 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -12,6 +12,7 @@ Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
+Queue start/stop = Y
 RSS hash = Y
 Inner RSS= Y
 Packet type parsing  = Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 9ad225a..82f2af0 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -12,6 +12,7 @@ Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
+Queue start/stop = Y
 RSS hash = Y
 Inner RSS= Y
 Packet type parsing  = Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 8c93ba7..61fed11 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -11,6 +11,7 @@ Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
+Queue start/stop = Y
 RSS hash = Y
 Inner RSS= Y
 Packet type parsing  = Y
diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index f79d03c..d70ab00 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -138,6 +138,21 @@ cn10k_nix_rx_queue_setup(struct rte_eth_dev *eth_dev, 
uint16_t qid,
 }
 
 static int
+cn10k_nix_tx_queue_stop(struct rte_eth_dev *eth_dev, uint16_t qidx)
+{
+   struct cn10k_eth_txq *txq = eth_dev->data->tx_queues[qidx];
+   int rc;
+
+   rc = cnxk_nix_tx_queue_stop(eth_dev, qidx);
+   if (rc)
+   return rc;
+
+   /* Clear fc cache pkts to trigger worker stop */
+   txq->fc_cache_pkts = 0;
+   return 0;
+}
+
+static int
 cn10k_nix_configure(struct rte_eth_dev *eth_dev)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
@@ -169,6 +184,7 @@ nix_eth_dev_ops_override(void)
cnxk_eth_dev_ops.dev_configure = cn10k_nix_configure;
cnxk_eth_dev_ops.tx_queue_setup = cn10k_nix_tx_queue_setup;
cnxk_eth_dev_ops.rx_queue_setup = cn10k_nix_rx_queue_setup;
+   cnxk_eth_dev_ops.tx_queue_stop = cn10k_nix_tx_queue_stop;
cnxk_eth_dev_ops.dev_ptypes_set = cn10k_nix_ptypes_set;
 }
 
diff --git a/drivers/net/cnxk/cn9k_ethdev.c b/drivers/net/cnxk/cn9k_ethdev.c
index b5a4bd7..1dab096 100644
--- a/drivers/net/cnxk/cn9k_ethdev.c
+++ b/drivers/net/cnxk/cn9k_ethdev.c
@@ -136,6 +136,21 @@ cn9k_nix_rx_queue_setup(struct rte_eth_dev *eth_dev, 
uint16_t qid,
 }
 
 static int
+cn9k_nix_tx_queue_stop(struct rte_eth_dev *eth_dev, uint16_t qidx)
+{
+   struct cn9k_eth_txq *txq = eth_dev->data->tx_queues[qidx];
+   int rc;
+
+   rc = cnxk_nix_tx_queue_stop(eth_dev, qidx);
+   if (rc)
+   return rc;
+
+   /* Clear fc cache pkts to trigger worker stop */
+   txq->fc_cache_pkts = 0;
+   return 0;
+}
+
+static int
 cn9k_nix_configure(struct rte_eth_dev *eth_dev)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
@@ -178,6 +193,7 @@ nix_eth_dev_ops_override(void)
cnxk_eth_dev_ops.dev_configure = cn9k_nix_configure;
cnxk_eth_dev_ops.tx_queue_setup = cn9k_nix_tx_queue_setup;
cnxk_eth_dev_ops.rx_queue_setup = cn9k_nix_rx_queue_setup;
+   cnxk_eth_dev_ops.tx_queue_stop = cn9k_nix_tx_queue_stop;
cnxk_eth_dev_ops.dev_ptypes_set = cn9k_nix_ptypes_set;
 }
 
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 6a0bd50..98e0ba0 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -854,12 +854,104 @@ cnxk_nix_configure(struct rte_eth_dev *eth_dev)
return rc;
 }
 
+static int
+cnxk_nix_tx_queue_start(struct rte_eth_dev *eth_dev, uint16_t qid)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct rte_eth_dev_data *data = eth_dev->data;
+   struct roc_nix_sq *sq = &dev->sqs[qid];
+   int rc = -EINVAL;
+
+   if (data->tx_queue_state[qid] == RTE_ETH_QUEUE_STATE_STARTED)
+   return 0;
+
+   rc = roc_nix_tm_sq_aura_fc(sq, true);
+   if (rc) {
+   plt_err("Failed to enable sq aura fc, txq=%u, rc=%d", qid, rc);
+   goto done;
+   }
+
+   data->tx_queue_state[qid] = 

[dpdk-dev] [PATCH v3 19/62] net/cnxk: add Rx support for cn9k

2021-06-18 Thread Nithin Dabilpuram
From: Jerin Jacob 

Add Rx burst scalar version for CN9K.

Signed-off-by: Jerin Jacob 
---
 drivers/net/cnxk/cn9k_ethdev.h |   3 +
 drivers/net/cnxk/cn9k_rx.c |  46 
 drivers/net/cnxk/cn9k_rx.h | 237 +
 drivers/net/cnxk/cnxk_ethdev.h |   3 +
 drivers/net/cnxk/meson.build   |   3 +-
 5 files changed, 291 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cn9k_rx.c

diff --git a/drivers/net/cnxk/cn9k_ethdev.h b/drivers/net/cnxk/cn9k_ethdev.h
index de635fa..bd41d3c 100644
--- a/drivers/net/cnxk/cn9k_ethdev.h
+++ b/drivers/net/cnxk/cn9k_ethdev.h
@@ -31,4 +31,7 @@ struct cn9k_eth_rxq {
uint16_t rq;
 } __plt_cache_aligned;
 
+/* Rx and Tx routines */
+void cn9k_eth_set_rx_function(struct rte_eth_dev *eth_dev);
+
 #endif /* __CN9K_ETHDEV_H__ */
diff --git a/drivers/net/cnxk/cn9k_rx.c b/drivers/net/cnxk/cn9k_rx.c
new file mode 100644
index 000..a4297f9
--- /dev/null
+++ b/drivers/net/cnxk/cn9k_rx.c
@@ -0,0 +1,46 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+
+#include "cn9k_ethdev.h"
+#include "cn9k_rx.h"
+
+#define R(name, f3, f2, f1, f0, flags)\
+   uint16_t __rte_noinline __rte_hot cn9k_nix_recv_pkts_##name(   \
+   void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t pkts)  \
+   {  \
+   return cn9k_nix_recv_pkts(rx_queue, rx_pkts, pkts, (flags));   \
+   }
+
+NIX_RX_FASTPATH_MODES
+#undef R
+
+static inline void
+pick_rx_func(struct rte_eth_dev *eth_dev,
+const eth_rx_burst_t rx_burst[2][2][2][2])
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   /* [MARK] [CKSUM] [PTYPE] [RSS] */
+   eth_dev->rx_pkt_burst = rx_burst
+   [!!(dev->rx_offload_flags & NIX_RX_OFFLOAD_MARK_UPDATE_F)]
+   [!!(dev->rx_offload_flags & NIX_RX_OFFLOAD_CHECKSUM_F)]
+   [!!(dev->rx_offload_flags & NIX_RX_OFFLOAD_PTYPE_F)]
+   [!!(dev->rx_offload_flags & NIX_RX_OFFLOAD_RSS_F)];
+}
+
+void
+cn9k_eth_set_rx_function(struct rte_eth_dev *eth_dev)
+{
+   const eth_rx_burst_t nix_eth_rx_burst[2][2][2][2] = {
+#define R(name, f3, f2, f1, f0, flags) \
+   [f3][f2][f1][f0] = cn9k_nix_recv_pkts_##name,
+
+   NIX_RX_FASTPATH_MODES
+#undef R
+   };
+
+   pick_rx_func(eth_dev, nix_eth_rx_burst);
+
+   rte_mb();
+}
diff --git a/drivers/net/cnxk/cn9k_rx.h b/drivers/net/cnxk/cn9k_rx.h
index 95a1e69..92f3c7c 100644
--- a/drivers/net/cnxk/cn9k_rx.h
+++ b/drivers/net/cnxk/cn9k_rx.h
@@ -7,6 +7,243 @@
 
 #include 
 
+#define NIX_RX_OFFLOAD_NONE (0)
+#define NIX_RX_OFFLOAD_RSS_FBIT(0)
 #define NIX_RX_OFFLOAD_PTYPE_F  BIT(1)
+#define NIX_RX_OFFLOAD_CHECKSUM_FBIT(2)
+#define NIX_RX_OFFLOAD_MARK_UPDATE_F BIT(3)
+
+/* Flags to control cqe_to_mbuf conversion function.
+ * Defining it from backwards to denote its been
+ * not used as offload flags to pick function
+ */
+#define NIX_RX_MULTI_SEG_F BIT(15)
+
+#define CNXK_NIX_CQ_ENTRY_SZ 128
+#define NIX_DESCS_PER_LOOP   4
+#define CQE_CAST(x) ((struct nix_cqe_hdr_s *)(x))
+#define CQE_SZ(x)   ((x) * CNXK_NIX_CQ_ENTRY_SZ)
+
+union mbuf_initializer {
+   struct {
+   uint16_t data_off;
+   uint16_t refcnt;
+   uint16_t nb_segs;
+   uint16_t port;
+   } fields;
+   uint64_t value;
+};
+
+static __rte_always_inline uint64_t
+nix_clear_data_off(uint64_t oldval)
+{
+   union mbuf_initializer mbuf_init = {.value = oldval};
+
+   mbuf_init.fields.data_off = 0;
+   return mbuf_init.value;
+}
+
+static __rte_always_inline struct rte_mbuf *
+nix_get_mbuf_from_cqe(void *cq, const uint64_t data_off)
+{
+   rte_iova_t buff;
+
+   /* Skip CQE, NIX_RX_PARSE_S and SG HDR(9 DWORDs) and peek buff addr */
+   buff = *((rte_iova_t *)((uint64_t *)cq + 9));
+   return (struct rte_mbuf *)(buff - data_off);
+}
+
+static __rte_always_inline uint32_t
+nix_ptype_get(const void *const lookup_mem, const uint64_t in)
+{
+   const uint16_t *const ptype = lookup_mem;
+   const uint16_t lh_lg_lf = (in & 0xFFF0) >> 52;
+   const uint16_t tu_l2 = ptype[(in & 0x0000) >> 36];
+   const uint16_t il4_tu = ptype[PTYPE_NON_TUNNEL_ARRAY_SZ + lh_lg_lf];
+
+   return (il4_tu << PTYPE_NON_TUNNEL_WIDTH) | tu_l2;
+}
+
+static __rte_always_inline uint32_t
+nix_rx_olflags_get(const void *const lookup_mem, const uint64_t in)
+{
+   const uint32_t *const ol_flags =
+   (const uint32_t *)((const uint8_t *)lookup_mem +
+  PTYPE_ARRAY_SZ);
+
+   return ol_flags[(in & 0xfff0) >> 20];
+}
+
+static inline uint64_t
+nix_update_match_id(const uint16_t match_id, uint64_t ol_flags,
+   struct rte_mbuf *mbuf)
+{
+ 

[dpdk-dev] [PATCH v3 20/62] net/cnxk: add Rx multi-segmented version for cn9k

2021-06-18 Thread Nithin Dabilpuram
Add Rx burst multi-segmented version for CN9K.

Signed-off-by: Nithin Dabilpuram 
Signed-off-by: Pavan Nikhilesh 
---
 drivers/net/cnxk/cn9k_rx.c  | 17 
 drivers/net/cnxk/cn9k_rx.h  | 60 ++---
 drivers/net/cnxk/cn9k_rx_mseg.c | 17 
 drivers/net/cnxk/cnxk_ethdev.h  |  3 +++
 drivers/net/cnxk/meson.build|  3 ++-
 5 files changed, 96 insertions(+), 4 deletions(-)
 create mode 100644 drivers/net/cnxk/cn9k_rx_mseg.c

diff --git a/drivers/net/cnxk/cn9k_rx.c b/drivers/net/cnxk/cn9k_rx.c
index a4297f9..87a62c9 100644
--- a/drivers/net/cnxk/cn9k_rx.c
+++ b/drivers/net/cnxk/cn9k_rx.c
@@ -32,6 +32,8 @@ pick_rx_func(struct rte_eth_dev *eth_dev,
 void
 cn9k_eth_set_rx_function(struct rte_eth_dev *eth_dev)
 {
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
const eth_rx_burst_t nix_eth_rx_burst[2][2][2][2] = {
 #define R(name, f3, f2, f1, f0, flags) \
[f3][f2][f1][f0] = cn9k_nix_recv_pkts_##name,
@@ -40,7 +42,22 @@ cn9k_eth_set_rx_function(struct rte_eth_dev *eth_dev)
 #undef R
};
 
+   const eth_rx_burst_t nix_eth_rx_burst_mseg[2][2][2][2] = {
+#define R(name, f3, f2, f1, f0, flags) \
+   [f3][f2][f1][f0] = cn9k_nix_recv_pkts_mseg_##name,
+
+   NIX_RX_FASTPATH_MODES
+#undef R
+   };
+
pick_rx_func(eth_dev, nix_eth_rx_burst);
 
+   if (dev->rx_offloads & DEV_RX_OFFLOAD_SCATTER)
+   pick_rx_func(eth_dev, nix_eth_rx_burst_mseg);
+
+   /* Copy multi seg version with no offload for tear down sequence */
+   if (rte_eal_process_type() == RTE_PROC_PRIMARY)
+   dev->rx_pkt_burst_no_offload =
+   nix_eth_rx_burst_mseg[0][0][0][0];
rte_mb();
 }
diff --git a/drivers/net/cnxk/cn9k_rx.h b/drivers/net/cnxk/cn9k_rx.h
index 92f3c7c..49f80ce 100644
--- a/drivers/net/cnxk/cn9k_rx.h
+++ b/drivers/net/cnxk/cn9k_rx.h
@@ -104,6 +104,53 @@ nix_update_match_id(const uint16_t match_id, uint64_t 
ol_flags,
 }
 
 static __rte_always_inline void
+nix_cqe_xtract_mseg(const union nix_rx_parse_u *rx, struct rte_mbuf *mbuf,
+   uint64_t rearm)
+{
+   const rte_iova_t *iova_list;
+   struct rte_mbuf *head;
+   const rte_iova_t *eol;
+   uint8_t nb_segs;
+   uint64_t sg;
+
+   sg = *(const uint64_t *)(rx + 1);
+   nb_segs = (sg >> 48) & 0x3;
+   mbuf->nb_segs = nb_segs;
+   mbuf->data_len = sg & 0x;
+   sg = sg >> 16;
+
+   eol = ((const rte_iova_t *)(rx + 1) +
+  ((rx->cn9k.desc_sizem1 + 1) << 1));
+   /* Skip SG_S and first IOVA*/
+   iova_list = ((const rte_iova_t *)(rx + 1)) + 2;
+   nb_segs--;
+
+   rearm = rearm & ~0x;
+
+   head = mbuf;
+   while (nb_segs) {
+   mbuf->next = ((struct rte_mbuf *)*iova_list) - 1;
+   mbuf = mbuf->next;
+
+   __mempool_check_cookies(mbuf->pool, (void **)&mbuf, 1, 1);
+
+   mbuf->data_len = sg & 0x;
+   sg = sg >> 16;
+   *(uint64_t *)(&mbuf->rearm_data) = rearm;
+   nb_segs--;
+   iova_list++;
+
+   if (!nb_segs && (iova_list + 1 < eol)) {
+   sg = *(const uint64_t *)(iova_list);
+   nb_segs = (sg >> 48) & 0x3;
+   head->nb_segs += nb_segs;
+   iova_list = (const rte_iova_t *)(iova_list + 1);
+   }
+   }
+   mbuf->next = NULL;
+}
+
+static __rte_always_inline void
 cn9k_nix_cqe_to_mbuf(const struct nix_cqe_hdr_s *cq, const uint32_t tag,
 struct rte_mbuf *mbuf, const void *lookup_mem,
 const uint64_t val, const uint16_t flag)
@@ -138,8 +185,12 @@ cn9k_nix_cqe_to_mbuf(const struct nix_cqe_hdr_s *cq, const 
uint32_t tag,
*(uint64_t *)(&mbuf->rearm_data) = val;
mbuf->pkt_len = len;
 
-   mbuf->data_len = len;
-   mbuf->next = NULL;
+   if (flag & NIX_RX_MULTI_SEG_F) {
+   nix_cqe_xtract_mseg(rx, mbuf, val);
+   } else {
+   mbuf->data_len = len;
+   mbuf->next = NULL;
+   }
 }
 
 static inline uint16_t
@@ -239,8 +290,11 @@ R(mark_cksum_rss,  1, 1, 0, 1, MARK_F | CKSUM_F | 
RSS_F)  \
 R(mark_cksum_ptype,1, 1, 1, 0, MARK_F | CKSUM_F | PTYPE_F)\
 R(mark_cksum_ptype_rss,1, 1, 1, 1, MARK_F | CKSUM_F | PTYPE_F 
| RSS_F)
 
-#define R(name, f3, f2, f1, f0, flags)\
+#define R(name, f3, f2, f1, f0, flags) 
\
uint16_t __rte_noinline __rte_hot cn9k_nix_recv_pkts_##name(   \
+   void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t pkts); \
+  \
+   uint16_t __rte_noinline __rte_hot cn9k_nix_recv_pkts_mseg_##name(  \
  

[dpdk-dev] [PATCH v3 21/62] net/cnxk: add Rx vector version for cn9k

2021-06-18 Thread Nithin Dabilpuram
From: Jerin Jacob 

Add Rx burst vector version for CN9K.

Signed-off-by: Jerin Jacob 
Signed-off-by: Nithin Dabilpuram 
---
 drivers/net/cnxk/cn9k_rx.c |  13 ++-
 drivers/net/cnxk/cn9k_rx.h | 221 +
 drivers/net/cnxk/cn9k_rx_vec.c |  17 
 drivers/net/cnxk/meson.build   |  11 +-
 4 files changed, 260 insertions(+), 2 deletions(-)
 create mode 100644 drivers/net/cnxk/cn9k_rx_vec.c

diff --git a/drivers/net/cnxk/cn9k_rx.c b/drivers/net/cnxk/cn9k_rx.c
index 87a62c9..01eb21f 100644
--- a/drivers/net/cnxk/cn9k_rx.c
+++ b/drivers/net/cnxk/cn9k_rx.c
@@ -50,7 +50,18 @@ cn9k_eth_set_rx_function(struct rte_eth_dev *eth_dev)
 #undef R
};
 
-   pick_rx_func(eth_dev, nix_eth_rx_burst);
+   const eth_rx_burst_t nix_eth_rx_vec_burst[2][2][2][2] = {
+#define R(name, f3, f2, f1, f0, flags) \
+   [f3][f2][f1][f0] = cn9k_nix_recv_pkts_vec_##name,
+
+   NIX_RX_FASTPATH_MODES
+#undef R
+   };
+
+   if (dev->scalar_ena)
+   pick_rx_func(eth_dev, nix_eth_rx_burst);
+   else
+   pick_rx_func(eth_dev, nix_eth_rx_vec_burst);
 
if (dev->rx_offloads & DEV_RX_OFFLOAD_SCATTER)
pick_rx_func(eth_dev, nix_eth_rx_burst_mseg);
diff --git a/drivers/net/cnxk/cn9k_rx.h b/drivers/net/cnxk/cn9k_rx.h
index 49f80ce..bc04f5c 100644
--- a/drivers/net/cnxk/cn9k_rx.h
+++ b/drivers/net/cnxk/cn9k_rx.h
@@ -6,6 +6,7 @@
 #define __CN9K_RX_H__
 
 #include 
+#include 
 
 #define NIX_RX_OFFLOAD_NONE (0)
 #define NIX_RX_OFFLOAD_RSS_FBIT(0)
@@ -266,6 +267,223 @@ cn9k_nix_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts, uint16_t pkts,
return nb_pkts;
 }
 
+#if defined(RTE_ARCH_ARM64)
+
+static __rte_always_inline uint16_t
+cn9k_nix_recv_pkts_vector(void *rx_queue, struct rte_mbuf **rx_pkts,
+ uint16_t pkts, const uint16_t flags)
+{
+   struct cn9k_eth_rxq *rxq = rx_queue;
+   uint16_t packets = 0;
+   uint64x2_t cq0_w8, cq1_w8, cq2_w8, cq3_w8, mbuf01, mbuf23;
+   const uint64_t mbuf_initializer = rxq->mbuf_initializer;
+   const uint64x2_t data_off = vdupq_n_u64(rxq->data_off);
+   uint64_t ol_flags0, ol_flags1, ol_flags2, ol_flags3;
+   uint64x2_t rearm0 = vdupq_n_u64(mbuf_initializer);
+   uint64x2_t rearm1 = vdupq_n_u64(mbuf_initializer);
+   uint64x2_t rearm2 = vdupq_n_u64(mbuf_initializer);
+   uint64x2_t rearm3 = vdupq_n_u64(mbuf_initializer);
+   struct rte_mbuf *mbuf0, *mbuf1, *mbuf2, *mbuf3;
+   const uint16_t *lookup_mem = rxq->lookup_mem;
+   const uint32_t qmask = rxq->qmask;
+   const uint64_t wdata = rxq->wdata;
+   const uintptr_t desc = rxq->desc;
+   uint8x16_t f0, f1, f2, f3;
+   uint32_t head = rxq->head;
+   uint16_t pkts_left;
+
+   pkts = nix_rx_nb_pkts(rxq, wdata, pkts, qmask);
+   pkts_left = pkts & (NIX_DESCS_PER_LOOP - 1);
+
+   /* Packets has to be floor-aligned to NIX_DESCS_PER_LOOP */
+   pkts = RTE_ALIGN_FLOOR(pkts, NIX_DESCS_PER_LOOP);
+
+   while (packets < pkts) {
+   /* Exit loop if head is about to wrap and become unaligned */
+   if (((head + NIX_DESCS_PER_LOOP - 1) & qmask) <
+   NIX_DESCS_PER_LOOP) {
+   pkts_left += (pkts - packets);
+   break;
+   }
+
+   const uintptr_t cq0 = desc + CQE_SZ(head);
+
+   /* Prefetch N desc ahead */
+   rte_prefetch_non_temporal((void *)(cq0 + CQE_SZ(8)));
+   rte_prefetch_non_temporal((void *)(cq0 + CQE_SZ(9)));
+   rte_prefetch_non_temporal((void *)(cq0 + CQE_SZ(10)));
+   rte_prefetch_non_temporal((void *)(cq0 + CQE_SZ(11)));
+
+   /* Get NIX_RX_SG_S for size and buffer pointer */
+   cq0_w8 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(0) + 64));
+   cq1_w8 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(1) + 64));
+   cq2_w8 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(2) + 64));
+   cq3_w8 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(3) + 64));
+
+   /* Extract mbuf from NIX_RX_SG_S */
+   mbuf01 = vzip2q_u64(cq0_w8, cq1_w8);
+   mbuf23 = vzip2q_u64(cq2_w8, cq3_w8);
+   mbuf01 = vqsubq_u64(mbuf01, data_off);
+   mbuf23 = vqsubq_u64(mbuf23, data_off);
+
+   /* Move mbufs to scalar registers for future use */
+   mbuf0 = (struct rte_mbuf *)vgetq_lane_u64(mbuf01, 0);
+   mbuf1 = (struct rte_mbuf *)vgetq_lane_u64(mbuf01, 1);
+   mbuf2 = (struct rte_mbuf *)vgetq_lane_u64(mbuf23, 0);
+   mbuf3 = (struct rte_mbuf *)vgetq_lane_u64(mbuf23, 1);
+
+   /* Mask to get packet len from NIX_RX_SG_S */
+   const uint8x16_t shuf_msk = {
+   0xFF, 0xFF, /* pkt_type set as unknown */
+   0xFF, 0xFF, /* pkt_type set as unkn

[dpdk-dev] [PATCH v3 22/62] net/cnxk: add Tx support for cn9k

2021-06-18 Thread Nithin Dabilpuram
From: Jerin Jacob 

Add Tx burst scalar version for CN9K.

Signed-off-by: Jerin Jacob 
Signed-off-by: Nithin Dabilpuram 
Signed-off-by: Pavan Nikhilesh 
Signed-off-by: Harman Kalra 
---
 drivers/net/cnxk/cn9k_ethdev.h |   1 +
 drivers/net/cnxk/cn9k_tx.c |  53 ++
 drivers/net/cnxk/cn9k_tx.h | 419 +
 drivers/net/cnxk/cnxk_ethdev.h |  71 +++
 drivers/net/cnxk/meson.build   |   3 +-
 5 files changed, 546 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cn9k_tx.c

diff --git a/drivers/net/cnxk/cn9k_ethdev.h b/drivers/net/cnxk/cn9k_ethdev.h
index bd41d3c..b92f3fc 100644
--- a/drivers/net/cnxk/cn9k_ethdev.h
+++ b/drivers/net/cnxk/cn9k_ethdev.h
@@ -33,5 +33,6 @@ struct cn9k_eth_rxq {
 
 /* Rx and Tx routines */
 void cn9k_eth_set_rx_function(struct rte_eth_dev *eth_dev);
+void cn9k_eth_set_tx_function(struct rte_eth_dev *eth_dev);
 
 #endif /* __CN9K_ETHDEV_H__ */
diff --git a/drivers/net/cnxk/cn9k_tx.c b/drivers/net/cnxk/cn9k_tx.c
new file mode 100644
index 000..a0b022a
--- /dev/null
+++ b/drivers/net/cnxk/cn9k_tx.c
@@ -0,0 +1,53 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+
+#include "cn9k_ethdev.h"
+#include "cn9k_tx.h"
+
+#define T(name, f4, f3, f2, f1, f0, sz, flags)\
+   uint16_t __rte_noinline __rte_hot cn9k_nix_xmit_pkts_##name(   \
+   void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t pkts)  \
+   {  \
+   uint64_t cmd[sz];  \
+  \
+   /* For TSO inner checksum is a must */ \
+   if (((flags) & NIX_TX_OFFLOAD_TSO_F) &&\
+   !((flags) & NIX_TX_OFFLOAD_L3_L4_CSUM_F))  \
+   return 0;  \
+   return cn9k_nix_xmit_pkts(tx_queue, tx_pkts, pkts, cmd, flags);\
+   }
+
+NIX_TX_FASTPATH_MODES
+#undef T
+
+static inline void
+pick_tx_func(struct rte_eth_dev *eth_dev,
+const eth_tx_burst_t tx_burst[2][2][2][2][2])
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   /* [TSO] [NOFF] [VLAN] [OL3_OL4_CSUM] [IL3_IL4_CSUM] */
+   eth_dev->tx_pkt_burst = tx_burst
+   [!!(dev->tx_offload_flags & NIX_TX_OFFLOAD_TSO_F)]
+   [!!(dev->tx_offload_flags & NIX_TX_OFFLOAD_MBUF_NOFF_F)]
+   [!!(dev->tx_offload_flags & NIX_TX_OFFLOAD_VLAN_QINQ_F)]
+   [!!(dev->tx_offload_flags & NIX_TX_OFFLOAD_OL3_OL4_CSUM_F)]
+   [!!(dev->tx_offload_flags & NIX_TX_OFFLOAD_L3_L4_CSUM_F)];
+}
+
+void
+cn9k_eth_set_tx_function(struct rte_eth_dev *eth_dev)
+{
+   const eth_tx_burst_t nix_eth_tx_burst[2][2][2][2][2] = {
+#define T(name, f4, f3, f2, f1, f0, sz, flags)\
+   [f4][f3][f2][f1][f0] = cn9k_nix_xmit_pkts_##name,
+
+   NIX_TX_FASTPATH_MODES
+#undef T
+   };
+
+   pick_tx_func(eth_dev, nix_eth_tx_burst);
+
+   rte_mb();
+}
diff --git a/drivers/net/cnxk/cn9k_tx.h b/drivers/net/cnxk/cn9k_tx.h
index bb6379b..7acecc6 100644
--- a/drivers/net/cnxk/cn9k_tx.h
+++ b/drivers/net/cnxk/cn9k_tx.h
@@ -4,10 +4,429 @@
 #ifndef __CN9K_TX_H__
 #define __CN9K_TX_H__
 
+#define NIX_TX_OFFLOAD_NONE  (0)
+#define NIX_TX_OFFLOAD_L3_L4_CSUM_F   BIT(0)
+#define NIX_TX_OFFLOAD_OL3_OL4_CSUM_F BIT(1)
 #define NIX_TX_OFFLOAD_VLAN_QINQ_FBIT(2)
+#define NIX_TX_OFFLOAD_MBUF_NOFF_FBIT(3)
 #define NIX_TX_OFFLOAD_TSO_F BIT(4)
 
+/* Flags to control xmit_prepare function.
+ * Defining it from backwards to denote its been
+ * not used as offload flags to pick function
+ */
+#define NIX_TX_MULTI_SEG_F BIT(15)
+
+#define NIX_TX_NEED_SEND_HDR_W1
\
+   (NIX_TX_OFFLOAD_L3_L4_CSUM_F | NIX_TX_OFFLOAD_OL3_OL4_CSUM_F | \
+NIX_TX_OFFLOAD_VLAN_QINQ_F | NIX_TX_OFFLOAD_TSO_F)
+
 #define NIX_TX_NEED_EXT_HDR
\
(NIX_TX_OFFLOAD_VLAN_QINQ_F | NIX_TX_OFFLOAD_TSO_F)
 
+#define NIX_XMIT_FC_OR_RETURN(txq, pkts)   
\
+   do {   \
+   /* Cached value is low, Update the fc_cache_pkts */\
+   if (unlikely((txq)->fc_cache_pkts < (pkts))) { \
+   /* Multiply with sqe_per_sqb to express in pkts */ \
+   (txq)->fc_cache_pkts = \
+   ((txq)->nb_sqb_bufs_adj - *(txq)->fc_mem)  \
+   << (txq)->sqes_per_sqb_log2;   \
+   /* Check it

[dpdk-dev] [PATCH v3 23/62] net/cnxk: add Tx multi-segment version for cn9k

2021-06-18 Thread Nithin Dabilpuram
Add Tx burst multi-segment version for CN9K.

Signed-off-by: Nithin Dabilpuram 
Signed-off-by: Pavan Nikhilesh 
---
 drivers/net/cnxk/cn9k_tx.c  |  14 
 drivers/net/cnxk/cn9k_tx.h  | 150 
 drivers/net/cnxk/cn9k_tx_mseg.c |  25 +++
 drivers/net/cnxk/cnxk_ethdev.h  |   4 ++
 drivers/net/cnxk/meson.build|   3 +-
 5 files changed, 195 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cn9k_tx_mseg.c

diff --git a/drivers/net/cnxk/cn9k_tx.c b/drivers/net/cnxk/cn9k_tx.c
index a0b022a..8f1d5f5 100644
--- a/drivers/net/cnxk/cn9k_tx.c
+++ b/drivers/net/cnxk/cn9k_tx.c
@@ -21,6 +21,7 @@
 NIX_TX_FASTPATH_MODES
 #undef T
 
+
 static inline void
 pick_tx_func(struct rte_eth_dev *eth_dev,
 const eth_tx_burst_t tx_burst[2][2][2][2][2])
@@ -39,6 +40,8 @@ pick_tx_func(struct rte_eth_dev *eth_dev,
 void
 cn9k_eth_set_tx_function(struct rte_eth_dev *eth_dev)
 {
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
const eth_tx_burst_t nix_eth_tx_burst[2][2][2][2][2] = {
 #define T(name, f4, f3, f2, f1, f0, sz, flags)\
[f4][f3][f2][f1][f0] = cn9k_nix_xmit_pkts_##name,
@@ -47,7 +50,18 @@ cn9k_eth_set_tx_function(struct rte_eth_dev *eth_dev)
 #undef T
};
 
+   const eth_tx_burst_t nix_eth_tx_burst_mseg[2][2][2][2][2] = {
+#define T(name, f4, f3, f2, f1, f0, sz, flags)\
+   [f4][f3][f2][f1][f0] = cn9k_nix_xmit_pkts_mseg_##name,
+
+   NIX_TX_FASTPATH_MODES
+#undef T
+   };
+
pick_tx_func(eth_dev, nix_eth_tx_burst);
 
+   if (dev->tx_offloads & DEV_TX_OFFLOAD_MULTI_SEGS)
+   pick_tx_func(eth_dev, nix_eth_tx_burst_mseg);
+
rte_mb();
 }
diff --git a/drivers/net/cnxk/cn9k_tx.h b/drivers/net/cnxk/cn9k_tx.h
index 7acecc6..d9aa406 100644
--- a/drivers/net/cnxk/cn9k_tx.h
+++ b/drivers/net/cnxk/cn9k_tx.h
@@ -311,6 +311,111 @@ cn9k_nix_xmit_submit_lmt_release(const rte_iova_t io_addr)
 }
 
 static __rte_always_inline uint16_t
+cn9k_nix_prepare_mseg(struct rte_mbuf *m, uint64_t *cmd, const uint16_t flags)
+{
+   struct nix_send_hdr_s *send_hdr;
+   union nix_send_sg_s *sg;
+   struct rte_mbuf *m_next;
+   uint64_t *slist, sg_u;
+   uint64_t nb_segs;
+   uint64_t segdw;
+   uint8_t off, i;
+
+   send_hdr = (struct nix_send_hdr_s *)cmd;
+   send_hdr->w0.total = m->pkt_len;
+   send_hdr->w0.aura = roc_npa_aura_handle_to_aura(m->pool->pool_id);
+
+   if (flags & NIX_TX_NEED_EXT_HDR)
+   off = 2;
+   else
+   off = 0;
+
+   sg = (union nix_send_sg_s *)&cmd[2 + off];
+   /* Clear sg->u header before use */
+   sg->u &= 0xFC00;
+   sg_u = sg->u;
+   slist = &cmd[3 + off];
+
+   i = 0;
+   nb_segs = m->nb_segs;
+
+   /* Fill mbuf segments */
+   do {
+   m_next = m->next;
+   sg_u = sg_u | ((uint64_t)m->data_len << (i << 4));
+   *slist = rte_mbuf_data_iova(m);
+   /* Set invert df if buffer is not to be freed by H/W */
+   if (flags & NIX_TX_OFFLOAD_MBUF_NOFF_F) {
+   sg_u |= (cnxk_nix_prefree_seg(m) << (i + 55));
+   /* Commit changes to mbuf */
+   rte_io_wmb();
+   }
+   /* Mark mempool object as "put" since it is freed by NIX */
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+   if (!(sg_u & (1ULL << (i + 55
+   __mempool_check_cookies(m->pool, (void **)&m, 1, 0);
+   rte_io_wmb();
+#endif
+   slist++;
+   i++;
+   nb_segs--;
+   if (i > 2 && nb_segs) {
+   i = 0;
+   /* Next SG subdesc */
+   *(uint64_t *)slist = sg_u & 0xFC00;
+   sg->u = sg_u;
+   sg->segs = 3;
+   sg = (union nix_send_sg_s *)slist;
+   sg_u = sg->u;
+   slist++;
+   }
+   m = m_next;
+   } while (nb_segs);
+
+   sg->u = sg_u;
+   sg->segs = i;
+   segdw = (uint64_t *)slist - (uint64_t *)&cmd[2 + off];
+   /* Roundup extra dwords to multiple of 2 */
+   segdw = (segdw >> 1) + (segdw & 0x1);
+   /* Default dwords */
+   segdw += (off >> 1) + 1;
+   send_hdr->w0.sizem1 = segdw - 1;
+
+   return segdw;
+}
+
+static __rte_always_inline void
+cn9k_nix_xmit_mseg_prep_lmt(uint64_t *cmd, void *lmt_addr, uint16_t segdw)
+{
+   roc_lmt_mov_seg(lmt_addr, (const void *)cmd, segdw);
+}
+
+static __rte_always_inline void
+cn9k_nix_xmit_mseg_one(uint64_t *cmd, void *lmt_addr, rte_iova_t io_addr,
+  uint16_t segdw)
+{
+   uint64_t lmt_status;
+
+   do {
+   roc_lmt_mov_seg(lmt_addr, (const void *)cmd, segdw);
+   lmt_status = r

[dpdk-dev] [PATCH v3 24/62] net/cnxk: add Tx vector version for cn9k

2021-06-18 Thread Nithin Dabilpuram
Add Tx burst vector version for CN9K.

Signed-off-by: Nithin Dabilpuram 
Signed-off-by: Pavan Nikhilesh 
---
 drivers/net/cnxk/cn9k_tx.c |  16 +-
 drivers/net/cnxk/cn9k_tx.h | 743 +
 drivers/net/cnxk/cn9k_tx_vec.c |  25 ++
 drivers/net/cnxk/meson.build   |   3 +-
 4 files changed, 784 insertions(+), 3 deletions(-)
 create mode 100644 drivers/net/cnxk/cn9k_tx_vec.c

diff --git a/drivers/net/cnxk/cn9k_tx.c b/drivers/net/cnxk/cn9k_tx.c
index 8f1d5f5..2ff9720 100644
--- a/drivers/net/cnxk/cn9k_tx.c
+++ b/drivers/net/cnxk/cn9k_tx.c
@@ -21,7 +21,6 @@
 NIX_TX_FASTPATH_MODES
 #undef T
 
-
 static inline void
 pick_tx_func(struct rte_eth_dev *eth_dev,
 const eth_tx_burst_t tx_burst[2][2][2][2][2])
@@ -58,7 +57,20 @@ cn9k_eth_set_tx_function(struct rte_eth_dev *eth_dev)
 #undef T
};
 
-   pick_tx_func(eth_dev, nix_eth_tx_burst);
+   const eth_tx_burst_t nix_eth_tx_vec_burst[2][2][2][2][2] = {
+#define T(name, f4, f3, f2, f1, f0, sz, flags)\
+   [f4][f3][f2][f1][f0] = cn9k_nix_xmit_pkts_vec_##name,
+
+   NIX_TX_FASTPATH_MODES
+#undef T
+   };
+
+   if (dev->scalar_ena ||
+   (dev->tx_offload_flags &
+(NIX_TX_OFFLOAD_VLAN_QINQ_F | NIX_TX_OFFLOAD_TSO_F)))
+   pick_tx_func(eth_dev, nix_eth_tx_burst);
+   else
+   pick_tx_func(eth_dev, nix_eth_tx_vec_burst);
 
if (dev->tx_offloads & DEV_TX_OFFLOAD_MULTI_SEGS)
pick_tx_func(eth_dev, nix_eth_tx_burst_mseg);
diff --git a/drivers/net/cnxk/cn9k_tx.h b/drivers/net/cnxk/cn9k_tx.h
index d9aa406..7b0d536 100644
--- a/drivers/net/cnxk/cn9k_tx.h
+++ b/drivers/net/cnxk/cn9k_tx.h
@@ -4,6 +4,8 @@
 #ifndef __CN9K_TX_H__
 #define __CN9K_TX_H__
 
+#include 
+
 #define NIX_TX_OFFLOAD_NONE  (0)
 #define NIX_TX_OFFLOAD_L3_L4_CSUM_F   BIT(0)
 #define NIX_TX_OFFLOAD_OL3_OL4_CSUM_F BIT(1)
@@ -495,6 +497,744 @@ cn9k_nix_xmit_pkts_mseg(void *tx_queue, struct rte_mbuf 
**tx_pkts,
return pkts;
 }
 
+#if defined(RTE_ARCH_ARM64)
+
+#define NIX_DESCS_PER_LOOP 4
+static __rte_always_inline uint16_t
+cn9k_nix_xmit_pkts_vector(void *tx_queue, struct rte_mbuf **tx_pkts,
+ uint16_t pkts, uint64_t *cmd, const uint16_t flags)
+{
+   uint64x2_t dataoff_iova0, dataoff_iova1, dataoff_iova2, dataoff_iova3;
+   uint64x2_t len_olflags0, len_olflags1, len_olflags2, len_olflags3;
+   uint64x2_t cmd0[NIX_DESCS_PER_LOOP], cmd1[NIX_DESCS_PER_LOOP];
+   uint64_t *mbuf0, *mbuf1, *mbuf2, *mbuf3;
+   uint64x2_t senddesc01_w0, senddesc23_w0;
+   uint64x2_t senddesc01_w1, senddesc23_w1;
+   uint64x2_t sgdesc01_w0, sgdesc23_w0;
+   uint64x2_t sgdesc01_w1, sgdesc23_w1;
+   struct cn9k_eth_txq *txq = tx_queue;
+   uint64_t *lmt_addr = txq->lmt_addr;
+   rte_iova_t io_addr = txq->io_addr;
+   uint64x2_t ltypes01, ltypes23;
+   uint64x2_t xtmp128, ytmp128;
+   uint64x2_t xmask01, xmask23;
+   uint64_t lmt_status, i;
+   uint16_t pkts_left;
+
+   NIX_XMIT_FC_OR_RETURN(txq, pkts);
+
+   pkts_left = pkts & (NIX_DESCS_PER_LOOP - 1);
+   pkts = RTE_ALIGN_FLOOR(pkts, NIX_DESCS_PER_LOOP);
+
+   /* Reduce the cached count */
+   txq->fc_cache_pkts -= pkts;
+
+   /* Lets commit any changes in the packet here as no further changes
+* to the packet will be done unless no fast free is enabled.
+*/
+   if (!(flags & NIX_TX_OFFLOAD_MBUF_NOFF_F))
+   rte_io_wmb();
+
+   senddesc01_w0 = vld1q_dup_u64(&txq->cmd[0]);
+   senddesc23_w0 = senddesc01_w0;
+   senddesc01_w1 = vdupq_n_u64(0);
+   senddesc23_w1 = senddesc01_w1;
+   sgdesc01_w0 = vld1q_dup_u64(&txq->cmd[2]);
+   sgdesc23_w0 = sgdesc01_w0;
+
+   for (i = 0; i < pkts; i += NIX_DESCS_PER_LOOP) {
+   /* Clear lower 32bit of SEND_HDR_W0 and SEND_SG_W0 */
+   senddesc01_w0 =
+   vbicq_u64(senddesc01_w0, vdupq_n_u64(0x));
+   sgdesc01_w0 = vbicq_u64(sgdesc01_w0, vdupq_n_u64(0x));
+
+   senddesc23_w0 = senddesc01_w0;
+   sgdesc23_w0 = sgdesc01_w0;
+
+   /* Move mbufs to iova */
+   mbuf0 = (uint64_t *)tx_pkts[0];
+   mbuf1 = (uint64_t *)tx_pkts[1];
+   mbuf2 = (uint64_t *)tx_pkts[2];
+   mbuf3 = (uint64_t *)tx_pkts[3];
+
+   mbuf0 = (uint64_t *)((uintptr_t)mbuf0 +
+offsetof(struct rte_mbuf, buf_iova));
+   mbuf1 = (uint64_t *)((uintptr_t)mbuf1 +
+offsetof(struct rte_mbuf, buf_iova));
+   mbuf2 = (uint64_t *)((uintptr_t)mbuf2 +
+offsetof(struct rte_mbuf, buf_iova));
+   mbuf3 = (uint64_t *)((uintptr_t)mbuf3 +
+offsetof(struct rte_mbuf, buf_iova));
+   /*
+

[dpdk-dev] [PATCH v3 25/62] net/cnxk: add Rx support for cn10k

2021-06-18 Thread Nithin Dabilpuram
From: Jerin Jacob 

Add Rx burst support for CN10K SoC.

Signed-off-by: Jerin Jacob 
Signed-off-by: Nithin Dabilpuram 
Signed-off-by: Pavan Nikhilesh 
Signed-off-by: Harman Kalra 
---
 drivers/net/cnxk/cn10k_ethdev.h |   3 +
 drivers/net/cnxk/cn10k_rx.c |  45 
 drivers/net/cnxk/cn10k_rx.h | 236 
 drivers/net/cnxk/meson.build|   3 +-
 4 files changed, 286 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cn10k_rx.c

diff --git a/drivers/net/cnxk/cn10k_ethdev.h b/drivers/net/cnxk/cn10k_ethdev.h
index 18deb95..596985f 100644
--- a/drivers/net/cnxk/cn10k_ethdev.h
+++ b/drivers/net/cnxk/cn10k_ethdev.h
@@ -33,4 +33,7 @@ struct cn10k_eth_rxq {
uint16_t rq;
 } __plt_cache_aligned;
 
+/* Rx and Tx routines */
+void cn10k_eth_set_rx_function(struct rte_eth_dev *eth_dev);
+
 #endif /* __CN10K_ETHDEV_H__ */
diff --git a/drivers/net/cnxk/cn10k_rx.c b/drivers/net/cnxk/cn10k_rx.c
new file mode 100644
index 000..8b422d0
--- /dev/null
+++ b/drivers/net/cnxk/cn10k_rx.c
@@ -0,0 +1,45 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+
+#include "cn10k_ethdev.h"
+#include "cn10k_rx.h"
+
+#define R(name, f3, f2, f1, f0, flags)\
+   uint16_t __rte_noinline __rte_hot cn10k_nix_recv_pkts_##name(  \
+   void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t pkts)  \
+   {  \
+   return cn10k_nix_recv_pkts(rx_queue, rx_pkts, pkts, (flags));  \
+   }
+
+NIX_RX_FASTPATH_MODES
+#undef R
+
+static inline void
+pick_rx_func(struct rte_eth_dev *eth_dev,
+const eth_rx_burst_t rx_burst[2][2][2][2])
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   /* [MARK] [CKSUM] [PTYPE] [RSS] */
+   eth_dev->rx_pkt_burst = rx_burst
+   [!!(dev->rx_offload_flags & NIX_RX_OFFLOAD_MARK_UPDATE_F)]
+   [!!(dev->rx_offload_flags & NIX_RX_OFFLOAD_CHECKSUM_F)]
+   [!!(dev->rx_offload_flags & NIX_RX_OFFLOAD_PTYPE_F)]
+   [!!(dev->rx_offload_flags & NIX_RX_OFFLOAD_RSS_F)];
+}
+
+void
+cn10k_eth_set_rx_function(struct rte_eth_dev *eth_dev)
+{
+   const eth_rx_burst_t nix_eth_rx_burst[2][2][2][2] = {
+#define R(name, f3, f2, f1, f0, flags)   \
+   [f3][f2][f1][f0] = cn10k_nix_recv_pkts_##name,
+
+   NIX_RX_FASTPATH_MODES
+#undef R
+   };
+
+   pick_rx_func(eth_dev, nix_eth_rx_burst);
+   rte_mb();
+}
diff --git a/drivers/net/cnxk/cn10k_rx.h b/drivers/net/cnxk/cn10k_rx.h
index d3d1661..01c9d29 100644
--- a/drivers/net/cnxk/cn10k_rx.h
+++ b/drivers/net/cnxk/cn10k_rx.h
@@ -6,6 +6,242 @@
 
 #include 
 
+#define NIX_RX_OFFLOAD_NONE (0)
+#define NIX_RX_OFFLOAD_RSS_FBIT(0)
 #define NIX_RX_OFFLOAD_PTYPE_F  BIT(1)
+#define NIX_RX_OFFLOAD_CHECKSUM_FBIT(2)
+#define NIX_RX_OFFLOAD_MARK_UPDATE_F BIT(3)
+
+/* Flags to control cqe_to_mbuf conversion function.
+ * Defining it from backwards to denote its been
+ * not used as offload flags to pick function
+ */
+#define NIX_RX_MULTI_SEG_F BIT(15)
+
+#define CNXK_NIX_CQ_ENTRY_SZ 128
+#define NIX_DESCS_PER_LOOP   4
+#define CQE_CAST(x) ((struct nix_cqe_hdr_s *)(x))
+#define CQE_SZ(x)   ((x) * CNXK_NIX_CQ_ENTRY_SZ)
+
+union mbuf_initializer {
+   struct {
+   uint16_t data_off;
+   uint16_t refcnt;
+   uint16_t nb_segs;
+   uint16_t port;
+   } fields;
+   uint64_t value;
+};
+
+static __rte_always_inline uint64_t
+nix_clear_data_off(uint64_t oldval)
+{
+   union mbuf_initializer mbuf_init = {.value = oldval};
+
+   mbuf_init.fields.data_off = 0;
+   return mbuf_init.value;
+}
+
+static __rte_always_inline struct rte_mbuf *
+nix_get_mbuf_from_cqe(void *cq, const uint64_t data_off)
+{
+   rte_iova_t buff;
+
+   /* Skip CQE, NIX_RX_PARSE_S and SG HDR(9 DWORDs) and peek buff addr */
+   buff = *((rte_iova_t *)((uint64_t *)cq + 9));
+   return (struct rte_mbuf *)(buff - data_off);
+}
+
+static __rte_always_inline uint32_t
+nix_ptype_get(const void *const lookup_mem, const uint64_t in)
+{
+   const uint16_t *const ptype = lookup_mem;
+   const uint16_t lh_lg_lf = (in & 0xFFF0) >> 52;
+   const uint16_t tu_l2 = ptype[(in & 0x0000) >> 36];
+   const uint16_t il4_tu = ptype[PTYPE_NON_TUNNEL_ARRAY_SZ + lh_lg_lf];
+
+   return (il4_tu << PTYPE_NON_TUNNEL_WIDTH) | tu_l2;
+}
+
+static __rte_always_inline uint32_t
+nix_rx_olflags_get(const void *const lookup_mem, const uint64_t in)
+{
+   const uint32_t *const ol_flags =
+   (const uint32_t *)((const uint8_t *)lookup_mem +
+  PTYPE_ARRAY_SZ);
+
+   return ol_flags[(in & 0xfff0) >> 20];
+}
+
+static inline uint64_t
+nix_update_match_id(const uint16_t ma

[dpdk-dev] [PATCH v3 26/62] net/cnxk: add Rx multi-segment version for cn10k

2021-06-18 Thread Nithin Dabilpuram
Add Rx burst multi-segment version for CN10K.

Signed-off-by: Nithin Dabilpuram 
Signed-off-by: Pavan Nikhilesh 
---
 doc/guides/nics/cnxk.rst  |  2 ++
 doc/guides/nics/features/cnxk.ini |  2 ++
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 doc/guides/nics/features/cnxk_vf.ini  |  2 ++
 drivers/net/cnxk/cn10k_rx.c   | 20 +++-
 drivers/net/cnxk/cn10k_rx.h   | 57 +--
 drivers/net/cnxk/cn10k_rx_mseg.c  | 17 +++
 drivers/net/cnxk/meson.build  |  3 +-
 8 files changed, 100 insertions(+), 4 deletions(-)
 create mode 100644 drivers/net/cnxk/cn10k_rx_mseg.c

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 8bc85c0..fd7f2dd 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -17,11 +17,13 @@ Features
 Features of the CNXK Ethdev PMD are:
 
 - Packet type information
+- Jumbo frames
 - SR-IOV VF
 - Lock-free Tx queue
 - Multiple queues for TX and RX
 - Receiver Side Scaling (RSS)
 - Link state information
+- Scatter-Gather IO support
 
 Prerequisites
 -
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 712f8d5..23564b7 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -15,6 +15,8 @@ Runtime Tx queue setup = Y
 Queue start/stop = Y
 RSS hash = Y
 Inner RSS= Y
+Jumbo frame  = Y
+Scattered Rx = Y
 Packet type parsing  = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 82f2af0..421048d 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -15,6 +15,7 @@ Runtime Tx queue setup = Y
 Queue start/stop = Y
 RSS hash = Y
 Inner RSS= Y
+Jumbo frame  = Y
 Packet type parsing  = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 61fed11..e901fa2 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -14,6 +14,8 @@ Runtime Tx queue setup = Y
 Queue start/stop = Y
 RSS hash = Y
 Inner RSS= Y
+Jumbo frame  = Y
+Scattered Rx = Y
 Packet type parsing  = Y
 Linux= Y
 ARMv8= Y
diff --git a/drivers/net/cnxk/cn10k_rx.c b/drivers/net/cnxk/cn10k_rx.c
index 8b422d0..ce2cfee 100644
--- a/drivers/net/cnxk/cn10k_rx.c
+++ b/drivers/net/cnxk/cn10k_rx.c
@@ -10,7 +10,7 @@
void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t pkts)  \
{  \
return cn10k_nix_recv_pkts(rx_queue, rx_pkts, pkts, (flags));  \
-   }
+   }  \
 
 NIX_RX_FASTPATH_MODES
 #undef R
@@ -32,6 +32,8 @@ pick_rx_func(struct rte_eth_dev *eth_dev,
 void
 cn10k_eth_set_rx_function(struct rte_eth_dev *eth_dev)
 {
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
const eth_rx_burst_t nix_eth_rx_burst[2][2][2][2] = {
 #define R(name, f3, f2, f1, f0, flags)   \
[f3][f2][f1][f0] = cn10k_nix_recv_pkts_##name,
@@ -40,6 +42,22 @@ cn10k_eth_set_rx_function(struct rte_eth_dev *eth_dev)
 #undef R
};
 
+   const eth_rx_burst_t nix_eth_rx_burst_mseg[2][2][2][2] = {
+#define R(name, f3, f2, f1, f0, flags)   \
+   [f3][f2][f1][f0] = cn10k_nix_recv_pkts_mseg_##name,
+
+   NIX_RX_FASTPATH_MODES
+#undef R
+   };
+
pick_rx_func(eth_dev, nix_eth_rx_burst);
+
+   if (dev->rx_offloads & DEV_RX_OFFLOAD_SCATTER)
+   pick_rx_func(eth_dev, nix_eth_rx_burst_mseg);
+
+   /* Copy multi seg version with no offload for tear down sequence */
+   if (rte_eal_process_type() == RTE_PROC_PRIMARY)
+   dev->rx_pkt_burst_no_offload =
+   nix_eth_rx_burst_mseg[0][0][0][0];
rte_mb();
 }
diff --git a/drivers/net/cnxk/cn10k_rx.h b/drivers/net/cnxk/cn10k_rx.h
index 01c9d29..c667c9a 100644
--- a/drivers/net/cnxk/cn10k_rx.h
+++ b/drivers/net/cnxk/cn10k_rx.h
@@ -103,6 +103,52 @@ nix_update_match_id(const uint16_t match_id, uint64_t 
ol_flags,
 }
 
 static __rte_always_inline void
+nix_cqe_xtract_mseg(const union nix_rx_parse_u *rx, struct rte_mbuf *mbuf,
+   uint64_t rearm)
+{
+   const rte_iova_t *iova_list;
+   struct rte_mbuf *head;
+   const rte_iova_t *eol;
+   uint8_t nb_segs;
+   uint64_t sg;
+
+   sg = *(const uint64_t *)(rx + 1);
+   nb_segs = (sg >> 48) & 0x3;
+   mbuf->nb_segs = nb_segs;
+   mbuf->data_len = sg & 0x;
+   sg = sg >> 16;
+
+   eol = ((const rte_iova_t *)(rx + 1) + ((rx->desc_sizem1 + 1) << 1));
+ 

[dpdk-dev] [PATCH v3 27/62] net/cnxk: add Rx vector version for cn10k

2021-06-18 Thread Nithin Dabilpuram
From: Jerin Jacob 

Add Rx burst vector version for CN10K.

Signed-off-by: Jerin Jacob 
Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/cnxk.rst|   1 +
 drivers/net/cnxk/cn10k_rx.c |  13 ++-
 drivers/net/cnxk/cn10k_rx.h | 222 
 drivers/net/cnxk/cn10k_rx_vec.c |  19 
 drivers/net/cnxk/meson.build|   3 +-
 5 files changed, 256 insertions(+), 2 deletions(-)
 create mode 100644 drivers/net/cnxk/cn10k_rx_vec.c

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index fd7f2dd..481bc7e 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -24,6 +24,7 @@ Features of the CNXK Ethdev PMD are:
 - Receiver Side Scaling (RSS)
 - Link state information
 - Scatter-Gather IO support
+- Vector Poll mode driver
 
 Prerequisites
 -
diff --git a/drivers/net/cnxk/cn10k_rx.c b/drivers/net/cnxk/cn10k_rx.c
index ce2cfee..0598111 100644
--- a/drivers/net/cnxk/cn10k_rx.c
+++ b/drivers/net/cnxk/cn10k_rx.c
@@ -50,7 +50,18 @@ cn10k_eth_set_rx_function(struct rte_eth_dev *eth_dev)
 #undef R
};
 
-   pick_rx_func(eth_dev, nix_eth_rx_burst);
+   const eth_rx_burst_t nix_eth_rx_vec_burst[2][2][2][2] = {
+#define R(name, f3, f2, f1, f0, flags)   \
+   [f3][f2][f1][f0] = cn10k_nix_recv_pkts_vec_##name,
+
+   NIX_RX_FASTPATH_MODES
+#undef R
+   };
+
+   if (dev->scalar_ena)
+   pick_rx_func(eth_dev, nix_eth_rx_burst);
+   else
+   pick_rx_func(eth_dev, nix_eth_rx_vec_burst);
 
if (dev->rx_offloads & DEV_RX_OFFLOAD_SCATTER)
pick_rx_func(eth_dev, nix_eth_rx_burst_mseg);
diff --git a/drivers/net/cnxk/cn10k_rx.h b/drivers/net/cnxk/cn10k_rx.h
index c667c9a..7bb9dd8 100644
--- a/drivers/net/cnxk/cn10k_rx.h
+++ b/drivers/net/cnxk/cn10k_rx.h
@@ -5,6 +5,7 @@
 #define __CN10K_RX_H__
 
 #include 
+#include 
 
 #define NIX_RX_OFFLOAD_NONE (0)
 #define NIX_RX_OFFLOAD_RSS_FBIT(0)
@@ -263,6 +264,224 @@ cn10k_nix_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts, uint16_t pkts,
return nb_pkts;
 }
 
+#if defined(RTE_ARCH_ARM64)
+
+static __rte_always_inline uint16_t
+cn10k_nix_recv_pkts_vector(void *rx_queue, struct rte_mbuf **rx_pkts,
+  uint16_t pkts, const uint16_t flags)
+{
+   struct cn10k_eth_rxq *rxq = rx_queue;
+   uint16_t packets = 0;
+   uint64x2_t cq0_w8, cq1_w8, cq2_w8, cq3_w8, mbuf01, mbuf23;
+   const uint64_t mbuf_initializer = rxq->mbuf_initializer;
+   const uint64x2_t data_off = vdupq_n_u64(rxq->data_off);
+   uint64_t ol_flags0, ol_flags1, ol_flags2, ol_flags3;
+   uint64x2_t rearm0 = vdupq_n_u64(mbuf_initializer);
+   uint64x2_t rearm1 = vdupq_n_u64(mbuf_initializer);
+   uint64x2_t rearm2 = vdupq_n_u64(mbuf_initializer);
+   uint64x2_t rearm3 = vdupq_n_u64(mbuf_initializer);
+   struct rte_mbuf *mbuf0, *mbuf1, *mbuf2, *mbuf3;
+   const uint16_t *lookup_mem = rxq->lookup_mem;
+   const uint32_t qmask = rxq->qmask;
+   const uint64_t wdata = rxq->wdata;
+   const uintptr_t desc = rxq->desc;
+   uint8x16_t f0, f1, f2, f3;
+   uint32_t head = rxq->head;
+   uint16_t pkts_left;
+
+   pkts = nix_rx_nb_pkts(rxq, wdata, pkts, qmask);
+   pkts_left = pkts & (NIX_DESCS_PER_LOOP - 1);
+
+   /* Packets has to be floor-aligned to NIX_DESCS_PER_LOOP */
+   pkts = RTE_ALIGN_FLOOR(pkts, NIX_DESCS_PER_LOOP);
+
+   while (packets < pkts) {
+   /* Exit loop if head is about to wrap and become unaligned */
+   if (((head + NIX_DESCS_PER_LOOP - 1) & qmask) <
+   NIX_DESCS_PER_LOOP) {
+   pkts_left += (pkts - packets);
+   break;
+   }
+
+   const uintptr_t cq0 = desc + CQE_SZ(head);
+
+   /* Prefetch N desc ahead */
+   rte_prefetch_non_temporal((void *)(cq0 + CQE_SZ(8)));
+   rte_prefetch_non_temporal((void *)(cq0 + CQE_SZ(9)));
+   rte_prefetch_non_temporal((void *)(cq0 + CQE_SZ(10)));
+   rte_prefetch_non_temporal((void *)(cq0 + CQE_SZ(11)));
+
+   /* Get NIX_RX_SG_S for size and buffer pointer */
+   cq0_w8 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(0) + 64));
+   cq1_w8 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(1) + 64));
+   cq2_w8 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(2) + 64));
+   cq3_w8 = vld1q_u64((uint64_t *)(cq0 + CQE_SZ(3) + 64));
+
+   /* Extract mbuf from NIX_RX_SG_S */
+   mbuf01 = vzip2q_u64(cq0_w8, cq1_w8);
+   mbuf23 = vzip2q_u64(cq2_w8, cq3_w8);
+   mbuf01 = vqsubq_u64(mbuf01, data_off);
+   mbuf23 = vqsubq_u64(mbuf23, data_off);
+
+   /* Move mbufs to scalar registers for future use */
+   mbuf0 = (struct rte_mbuf *)vgetq_lane_u64(mbuf01, 0);
+   mbuf1

[dpdk-dev] [PATCH v3 28/62] net/cnxk: add Tx support for cn10k

2021-06-18 Thread Nithin Dabilpuram
From: Jerin Jacob 

Add Tx burst scalar version for CN10K.

Signed-off-by: Jerin Jacob 
Signed-off-by: Nithin Dabilpuram 
Signed-off-by: Pavan Nikhilesh 
Signed-off-by: Harman Kalra 
---
 doc/guides/nics/cnxk.rst  |   1 +
 doc/guides/nics/features/cnxk.ini |   7 +
 doc/guides/nics/features/cnxk_vec.ini |   6 +
 doc/guides/nics/features/cnxk_vf.ini  |   7 +
 drivers/net/cnxk/cn10k_ethdev.h   |   1 +
 drivers/net/cnxk/cn10k_tx.c   |  54 
 drivers/net/cnxk/cn10k_tx.h   | 491 ++
 drivers/net/cnxk/meson.build  |   7 +-
 8 files changed, 571 insertions(+), 3 deletions(-)
 create mode 100644 drivers/net/cnxk/cn10k_tx.c

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 481bc7e..17da141 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -22,6 +22,7 @@ Features of the CNXK Ethdev PMD are:
 - Lock-free Tx queue
 - Multiple queues for TX and RX
 - Receiver Side Scaling (RSS)
+- Inner and Outer Checksum offload
 - Link state information
 - Scatter-Gather IO support
 - Vector Poll mode driver
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 23564b7..02be26b 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -12,11 +12,18 @@ Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
+Fast mbuf free   = Y
+Free Tx mbuf on demand = Y
 Queue start/stop = Y
+TSO  = Y
 RSS hash = Y
 Inner RSS= Y
 Jumbo frame  = Y
 Scattered Rx = Y
+L3 checksum offload  = Y
+L4 checksum offload  = Y
+Inner L3 checksum= Y
+Inner L4 checksum= Y
 Packet type parsing  = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 421048d..8c63853 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -12,10 +12,16 @@ Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
+Fast mbuf free   = Y
+Free Tx mbuf on demand = Y
 Queue start/stop = Y
 RSS hash = Y
 Inner RSS= Y
 Jumbo frame  = Y
+L3 checksum offload  = Y
+L4 checksum offload  = Y
+Inner L3 checksum= Y
+Inner L4 checksum= Y
 Packet type parsing  = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index e901fa2..a1bd49b 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -11,11 +11,18 @@ Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
+Fast mbuf free   = Y
+Free Tx mbuf on demand = Y
 Queue start/stop = Y
+TSO  = Y
 RSS hash = Y
 Inner RSS= Y
 Jumbo frame  = Y
 Scattered Rx = Y
+L3 checksum offload  = Y
+L4 checksum offload  = Y
+Inner L3 checksum= Y
+Inner L4 checksum= Y
 Packet type parsing  = Y
 Linux= Y
 ARMv8= Y
diff --git a/drivers/net/cnxk/cn10k_ethdev.h b/drivers/net/cnxk/cn10k_ethdev.h
index 596985f..d39ca31 100644
--- a/drivers/net/cnxk/cn10k_ethdev.h
+++ b/drivers/net/cnxk/cn10k_ethdev.h
@@ -35,5 +35,6 @@ struct cn10k_eth_rxq {
 
 /* Rx and Tx routines */
 void cn10k_eth_set_rx_function(struct rte_eth_dev *eth_dev);
+void cn10k_eth_set_tx_function(struct rte_eth_dev *eth_dev);
 
 #endif /* __CN10K_ETHDEV_H__ */
diff --git a/drivers/net/cnxk/cn10k_tx.c b/drivers/net/cnxk/cn10k_tx.c
new file mode 100644
index 000..13c605f
--- /dev/null
+++ b/drivers/net/cnxk/cn10k_tx.c
@@ -0,0 +1,54 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+
+#include "cn10k_ethdev.h"
+#include "cn10k_tx.h"
+
+#define T(name, f4, f3, f2, f1, f0, sz, flags)\
+   uint16_t __rte_noinline __rte_hot cn10k_nix_xmit_pkts_##name(  \
+   void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t pkts)  \
+   {  \
+   uint64_t cmd[sz];  \
+  \
+   /* For TSO inner checksum is a must */ \
+   if (((flags) & NIX_TX_OFFLOAD_TSO_F) &&\
+   !((flags) & NIX_TX_OFFLOAD_L3_L4_CSUM_F))  \
+   return 0;  \
+   return cn10k_nix_xmit_pkts(tx_queue, tx_pkts, pkts, cmd,   \
+  flags); \
+   }
+
+NIX_TX_FASTPATH_MODES
+#undef T
+
+static inline void
+pi

[dpdk-dev] [PATCH v3 29/62] net/cnxk: add Tx multi-segment version for cn10k

2021-06-18 Thread Nithin Dabilpuram
Add Tx burst multi-segment version for CN10K.

Signed-off-by: Nithin Dabilpuram 
Signed-off-by: Pavan Nikhilesh 
---
 drivers/net/cnxk/cn10k_tx.c  |  18 -
 drivers/net/cnxk/cn10k_tx.h  | 171 +++
 drivers/net/cnxk/cn10k_tx_mseg.c |  25 ++
 drivers/net/cnxk/meson.build |   3 +-
 4 files changed, 215 insertions(+), 2 deletions(-)
 create mode 100644 drivers/net/cnxk/cn10k_tx_mseg.c

diff --git a/drivers/net/cnxk/cn10k_tx.c b/drivers/net/cnxk/cn10k_tx.c
index 13c605f..9803002 100644
--- a/drivers/net/cnxk/cn10k_tx.c
+++ b/drivers/net/cnxk/cn10k_tx.c
@@ -40,6 +40,8 @@ pick_tx_func(struct rte_eth_dev *eth_dev,
 void
 cn10k_eth_set_tx_function(struct rte_eth_dev *eth_dev)
 {
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
const eth_tx_burst_t nix_eth_tx_burst[2][2][2][2][2] = {
 #define T(name, f4, f3, f2, f1, f0, sz, flags) \
[f4][f3][f2][f1][f0] = cn10k_nix_xmit_pkts_##name,
@@ -48,7 +50,21 @@ cn10k_eth_set_tx_function(struct rte_eth_dev *eth_dev)
 #undef T
};
 
-   pick_tx_func(eth_dev, nix_eth_tx_burst);
+   const eth_tx_burst_t nix_eth_tx_burst_mseg[2][2][2][2][2] = {
+#define T(name, f4, f3, f2, f1, f0, sz, flags) \
+   [f4][f3][f2][f1][f0] = cn10k_nix_xmit_pkts_mseg_##name,
+
+   NIX_TX_FASTPATH_MODES
+#undef T
+   };
+
+   if (dev->scalar_ena ||
+   (dev->tx_offload_flags &
+(NIX_TX_OFFLOAD_VLAN_QINQ_F | NIX_TX_OFFLOAD_TSO_F)))
+   pick_tx_func(eth_dev, nix_eth_tx_burst);
+
+   if (dev->tx_offloads & DEV_TX_OFFLOAD_MULTI_SEGS)
+   pick_tx_func(eth_dev, nix_eth_tx_burst_mseg);
 
rte_mb();
 }
diff --git a/drivers/net/cnxk/cn10k_tx.h b/drivers/net/cnxk/cn10k_tx.h
index c54fbfe..63e9848 100644
--- a/drivers/net/cnxk/cn10k_tx.h
+++ b/drivers/net/cnxk/cn10k_tx.h
@@ -339,6 +339,77 @@ cn10k_nix_xmit_prepare(struct rte_mbuf *m, uint64_t *cmd, 
uintptr_t lmt_addr,
 }
 
 static __rte_always_inline uint16_t
+cn10k_nix_prepare_mseg(struct rte_mbuf *m, uint64_t *cmd, const uint16_t flags)
+{
+   struct nix_send_hdr_s *send_hdr;
+   union nix_send_sg_s *sg;
+   struct rte_mbuf *m_next;
+   uint64_t *slist, sg_u;
+   uint64_t nb_segs;
+   uint64_t segdw;
+   uint8_t off, i;
+
+   send_hdr = (struct nix_send_hdr_s *)cmd;
+   send_hdr->w0.total = m->pkt_len;
+   send_hdr->w0.aura = roc_npa_aura_handle_to_aura(m->pool->pool_id);
+
+   if (flags & NIX_TX_NEED_EXT_HDR)
+   off = 2;
+   else
+   off = 0;
+
+   sg = (union nix_send_sg_s *)&cmd[2 + off];
+   /* Clear sg->u header before use */
+   sg->u &= 0xFC00;
+   sg_u = sg->u;
+   slist = &cmd[3 + off];
+
+   i = 0;
+   nb_segs = m->nb_segs;
+
+   /* Fill mbuf segments */
+   do {
+   m_next = m->next;
+   sg_u = sg_u | ((uint64_t)m->data_len << (i << 4));
+   *slist = rte_mbuf_data_iova(m);
+   /* Set invert df if buffer is not to be freed by H/W */
+   if (flags & NIX_TX_OFFLOAD_MBUF_NOFF_F)
+   sg_u |= (cnxk_nix_prefree_seg(m) << (i + 55));
+   /* Mark mempool object as "put" since it is freed by NIX
+*/
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+   if (!(sg_u & (1ULL << (i + 55
+   __mempool_check_cookies(m->pool, (void **)&m, 1, 0);
+#endif
+   slist++;
+   i++;
+   nb_segs--;
+   if (i > 2 && nb_segs) {
+   i = 0;
+   /* Next SG subdesc */
+   *(uint64_t *)slist = sg_u & 0xFC00;
+   sg->u = sg_u;
+   sg->segs = 3;
+   sg = (union nix_send_sg_s *)slist;
+   sg_u = sg->u;
+   slist++;
+   }
+   m = m_next;
+   } while (nb_segs);
+
+   sg->u = sg_u;
+   sg->segs = i;
+   segdw = (uint64_t *)slist - (uint64_t *)&cmd[2 + off];
+   /* Roundup extra dwords to multiple of 2 */
+   segdw = (segdw >> 1) + (segdw & 0x1);
+   /* Default dwords */
+   segdw += (off >> 1) + 1;
+   send_hdr->w0.sizem1 = segdw - 1;
+
+   return segdw;
+}
+
+static __rte_always_inline uint16_t
 cn10k_nix_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t pkts,
uint64_t *cmd, const uint16_t flags)
 {
@@ -421,6 +492,103 @@ cn10k_nix_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts, uint16_t pkts,
return pkts;
 }
 
+static __rte_always_inline uint16_t
+cn10k_nix_xmit_pkts_mseg(void *tx_queue, struct rte_mbuf **tx_pkts,
+uint16_t pkts, uint64_t *cmd, const uint16_t flags)
+{
+   struct cn10k_eth_txq *txq = tx_queue;
+   uintptr_t pa0, pa1, lmt_addr = txq->lmt

[dpdk-dev] [PATCH v3 31/62] net/cnxk: add device start and stop operations

2021-06-18 Thread Nithin Dabilpuram
Add device start and stop operation callbacks for
CN9K and CN10K. Device stop is common for both platforms
while device start as some platform dependent portion where
the platform specific offload flags are recomputed and
the right Rx/Tx burst function is chosen.

Signed-off-by: Nithin Dabilpuram 
---
 doc/guides/nics/cnxk.rst|  84 ++
 drivers/net/cnxk/cn10k_ethdev.c | 124 +++
 drivers/net/cnxk/cn9k_ethdev.c  | 127 
 drivers/net/cnxk/cnxk_ethdev.c  |  90 
 drivers/net/cnxk/cnxk_ethdev.h  |   2 +
 drivers/net/cnxk/cnxk_link.c|  11 
 6 files changed, 438 insertions(+)

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 17da141..15911ee 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -39,6 +39,58 @@ Driver compilation and testing
 Refer to the document :ref:`compiling and testing a PMD for a NIC 
`
 for details.
 
+#. Running testpmd:
+
+   Follow instructions available in the document
+   :ref:`compiling and testing a PMD for a NIC `
+   to run testpmd.
+
+   Example output:
+
+   .. code-block:: console
+
+  .//app/dpdk-testpmd -c 0xc -a 0002:02:00.0 -- --portmask=0x1 
--nb-cores=1 --port-topology=loop --rxq=1 --txq=1
+  EAL: Detected 4 lcore(s)
+  EAL: Detected 1 NUMA nodes
+  EAL: Multi-process socket /var/run/dpdk/rte/mp_socket
+  EAL: Selected IOVA mode 'VA'
+  EAL: No available hugepages reported in hugepages-16777216kB
+  EAL: No available hugepages reported in hugepages-2048kB
+  EAL: Probing VFIO support...
+  EAL: VFIO support initialized
+  EAL:   using IOMMU type 1 (Type 1)
+  [ 2003.202721] vfio-pci 0002:02:00.0: vfio_cap_init: hiding cap 0x14@0x98
+  EAL: Probe PCI driver: net_cn10k (177d:a063) device: 0002:02:00.0 
(socket 0)
+  PMD: RoC Model: cn10k
+  EAL: No legacy callbacks, legacy socket not created
+  testpmd: create a new mbuf pool : n=155456, size=2176, 
socket=0
+  testpmd: preferred mempool ops selected: cn10k_mempool_ops
+  Configuring Port 0 (socket 0)
+  PMD: Port 0: Link Up - speed 25000 Mbps - full-duplex
+
+  Port 0: link state change event
+  Port 0: 96:D4:99:72:A5:BF
+  Checking link statuses...
+  Done
+  No commandline core given, start packet forwarding
+  io packet forwarding - ports=1 - cores=1 - streams=1 - NUMA support 
enabled, MP allocation mode: native
+  Logical Core 3 (socket 0) forwards packets on 1 streams:
+RX P=0/Q=0 (socket 0) -> TX P=0/Q=0 (socket 0) peer=02:00:00:00:00:00
+
+io packet forwarding packets/burst=32
+nb forwarding cores=1 - nb forwarding ports=1
+port 0: RX queue number: 1 Tx queue number: 1
+  Rx offloads=0x0 Tx offloads=0x1
+  RX queue: 0
+RX desc=4096 - RX free threshold=0
+RX threshold registers: pthresh=0 hthresh=0  wthresh=0
+RX Offloads=0x0
+  TX queue: 0
+TX desc=512 - TX free threshold=0
+TX threshold registers: pthresh=0 hthresh=0  wthresh=0
+TX offloads=0x0 - TX RS bit threshold=0
+  Press enter to exit
+
 Runtime Config Options
 --
 
@@ -132,3 +184,35 @@ Runtime Config Options
Above devarg parameters are configurable per device, user needs to pass the
parameters to all the PCIe devices if application requires to configure on
all the ethdev ports.
+
+Limitations
+---
+
+``mempool_cnxk`` external mempool handler dependency
+~
+
+The OCTEON CN9K/CN10K SoC family NIC has inbuilt HW assisted external mempool 
manager.
+``net_cnxk`` pmd only works with ``mempool_cnxk`` mempool handler
+as it is performance wise most effective way for packet allocation and Tx 
buffer
+recycling on OCTEON TX2 SoC platform.
+
+CRC stripping
+~
+
+The OCTEON CN9K/CN10K SoC family NICs strip the CRC for every packet being 
received by
+the host interface irrespective of the offload configuration.
+
+Debugging Options
+-
+
+.. _table_cnxk_ethdev_debug_options:
+
+.. table:: cnxk ethdev debug options
+
+   +---++---+
+   | # | Component  | EAL log command   |
+   +===++===+
+   | 1 | NIX| --log-level='pmd\.net.cnxk,8' |
+   +---++---+
+   | 2 | NPC| --log-level='pmd\.net.cnxk\.flow,8'   |
+   +---++---+
diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index d70ab00..5ff36bb 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.

[dpdk-dev] [PATCH v3 30/62] net/cnxk: add Tx vector version for cn10k

2021-06-18 Thread Nithin Dabilpuram
Add Tx burst vector version for CN10K.

Signed-off-by: Nithin Dabilpuram 
Signed-off-by: Pavan Nikhilesh 
---
 drivers/net/cnxk/cn10k_tx.c |  10 +
 drivers/net/cnxk/cn10k_tx.h | 815 
 drivers/net/cnxk/cn10k_tx_vec.c |  25 ++
 drivers/net/cnxk/meson.build|   3 +-
 4 files changed, 852 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cn10k_tx_vec.c

diff --git a/drivers/net/cnxk/cn10k_tx.c b/drivers/net/cnxk/cn10k_tx.c
index 9803002..e6eb101 100644
--- a/drivers/net/cnxk/cn10k_tx.c
+++ b/drivers/net/cnxk/cn10k_tx.c
@@ -58,10 +58,20 @@ cn10k_eth_set_tx_function(struct rte_eth_dev *eth_dev)
 #undef T
};
 
+   const eth_tx_burst_t nix_eth_tx_vec_burst[2][2][2][2][2] = {
+#define T(name, f4, f3, f2, f1, f0, sz, flags) \
+   [f4][f3][f2][f1][f0] = cn10k_nix_xmit_pkts_vec_##name,
+
+   NIX_TX_FASTPATH_MODES
+#undef T
+   };
+
if (dev->scalar_ena ||
(dev->tx_offload_flags &
 (NIX_TX_OFFLOAD_VLAN_QINQ_F | NIX_TX_OFFLOAD_TSO_F)))
pick_tx_func(eth_dev, nix_eth_tx_burst);
+   else
+   pick_tx_func(eth_dev, nix_eth_tx_vec_burst);
 
if (dev->tx_offloads & DEV_TX_OFFLOAD_MULTI_SEGS)
pick_tx_func(eth_dev, nix_eth_tx_burst_mseg);
diff --git a/drivers/net/cnxk/cn10k_tx.h b/drivers/net/cnxk/cn10k_tx.h
index 63e9848..b74df10 100644
--- a/drivers/net/cnxk/cn10k_tx.h
+++ b/drivers/net/cnxk/cn10k_tx.h
@@ -4,6 +4,8 @@
 #ifndef __CN10K_TX_H__
 #define __CN10K_TX_H__
 
+#include 
+
 #define NIX_TX_OFFLOAD_NONE  (0)
 #define NIX_TX_OFFLOAD_L3_L4_CSUM_F   BIT(0)
 #define NIX_TX_OFFLOAD_OL3_OL4_CSUM_F BIT(1)
@@ -38,6 +40,9 @@
}  \
} while (0)
 
+#define LMT_OFF(lmt_addr, lmt_num, offset) 
\
+   (void *)((lmt_addr) + ((lmt_num) << ROC_LMT_LINE_SIZE_LOG2) + (offset))
+
 /* Function to determine no of tx subdesc required in case ext
  * sub desc is enabled.
  */
@@ -48,6 +53,14 @@ cn10k_nix_tx_ext_subs(const uint16_t flags)
(NIX_TX_OFFLOAD_VLAN_QINQ_F | NIX_TX_OFFLOAD_TSO_F)) ? 1 : 0;
 }
 
+static __rte_always_inline uint8_t
+cn10k_nix_pkts_per_vec_brst(const uint16_t flags)
+{
+   RTE_SET_USED(flags);
+   /* We can pack up to 4 packets per LMTLINE if there are no offloads. */
+   return 4 << ROC_LMT_LINES_PER_CORE_LOG2;
+}
+
 static __rte_always_inline uint64_t
 cn10k_nix_tx_steor_data(const uint16_t flags)
 {
@@ -76,6 +89,35 @@ cn10k_nix_tx_steor_data(const uint16_t flags)
return data;
 }
 
+static __rte_always_inline uint64_t
+cn10k_nix_tx_steor_vec_data(const uint16_t flags)
+{
+   const uint64_t dw_m1 = 0x7;
+   uint64_t data;
+
+   RTE_SET_USED(flags);
+   /* This will be moved to addr area */
+   data = dw_m1;
+   /* 15 vector sizes for single seg */
+   data |= dw_m1 << 19;
+   data |= dw_m1 << 22;
+   data |= dw_m1 << 25;
+   data |= dw_m1 << 28;
+   data |= dw_m1 << 31;
+   data |= dw_m1 << 34;
+   data |= dw_m1 << 37;
+   data |= dw_m1 << 40;
+   data |= dw_m1 << 43;
+   data |= dw_m1 << 46;
+   data |= dw_m1 << 49;
+   data |= dw_m1 << 52;
+   data |= dw_m1 << 55;
+   data |= dw_m1 << 58;
+   data |= dw_m1 << 61;
+
+   return data;
+}
+
 static __rte_always_inline void
 cn10k_nix_tx_skeleton(const struct cn10k_eth_txq *txq, uint64_t *cmd,
  const uint16_t flags)
@@ -589,6 +631,776 @@ cn10k_nix_xmit_pkts_mseg(void *tx_queue, struct rte_mbuf 
**tx_pkts,
return pkts;
 }
 
+#if defined(RTE_ARCH_ARM64)
+
+#define NIX_DESCS_PER_LOOP 4
+static __rte_always_inline uint16_t
+cn10k_nix_xmit_pkts_vector(void *tx_queue, struct rte_mbuf **tx_pkts,
+  uint16_t pkts, uint64_t *cmd, const uint16_t flags)
+{
+   uint64x2_t dataoff_iova0, dataoff_iova1, dataoff_iova2, dataoff_iova3;
+   uint64x2_t len_olflags0, len_olflags1, len_olflags2, len_olflags3;
+   uint64x2_t cmd0[NIX_DESCS_PER_LOOP], cmd1[NIX_DESCS_PER_LOOP];
+   uint64_t *mbuf0, *mbuf1, *mbuf2, *mbuf3, data, pa;
+   uint64x2_t senddesc01_w0, senddesc23_w0;
+   uint64x2_t senddesc01_w1, senddesc23_w1;
+   uint16_t left, scalar, burst, i, lmt_id;
+   uint64x2_t sgdesc01_w0, sgdesc23_w0;
+   uint64x2_t sgdesc01_w1, sgdesc23_w1;
+   struct cn10k_eth_txq *txq = tx_queue;
+   uintptr_t laddr = txq->lmt_base;
+   rte_iova_t io_addr = txq->io_addr;
+   uint64x2_t ltypes01, ltypes23;
+   uint64x2_t xtmp128, ytmp128;
+   uint64x2_t xmask01, xmask23;
+   uint8_t lnum;
+
+   NIX_XMIT_FC_OR_RETURN(txq, pkts);
+
+   scalar = pkts & (NIX_DESCS_PER_LOOP - 1);
+   pkts = RTE_ALIGN_FLOOR(pkts, NIX_DESCS_PER_LOOP);
+
+   /* Reduce the cached count */
+   txq->fc_cache_pkts -= pkts;
+
+   senddesc01_w0 = vl

[dpdk-dev] [PATCH v3 32/62] net/cnxk: add MAC address set ops

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Default mac address set operation is implemented for
cn9k and cn10k platforms.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cnxk_ethdev.c |  1 +
 drivers/net/cnxk/cnxk_ethdev.h |  2 ++
 drivers/net/cnxk/cnxk_ethdev_ops.c | 29 +
 3 files changed, 32 insertions(+)

diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index beb07ae..8908016 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1034,6 +1034,7 @@ cnxk_nix_dev_start(struct rte_eth_dev *eth_dev)
 
 /* CNXK platform independent eth dev ops */
 struct eth_dev_ops cnxk_eth_dev_ops = {
+   .mac_addr_set = cnxk_nix_mac_addr_set,
.dev_infos_get = cnxk_nix_info_get,
.link_update = cnxk_nix_link_update,
.tx_queue_release = cnxk_nix_tx_queue_release,
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 50c75e1..a5380a5 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -218,6 +218,8 @@ extern struct eth_dev_ops cnxk_eth_dev_ops;
 int cnxk_nix_probe(struct rte_pci_driver *pci_drv,
   struct rte_pci_device *pci_dev);
 int cnxk_nix_remove(struct rte_pci_device *pci_dev);
+int cnxk_nix_mac_addr_set(struct rte_eth_dev *eth_dev,
+ struct rte_ether_addr *addr);
 int cnxk_nix_info_get(struct rte_eth_dev *eth_dev,
  struct rte_eth_dev_info *dev_info);
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 4a45956..87cf4ee 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -69,3 +69,32 @@ cnxk_nix_info_get(struct rte_eth_dev *eth_dev, struct 
rte_eth_dev_info *devinfo)
RTE_ETH_DEV_CAPA_RUNTIME_TX_QUEUE_SETUP;
return 0;
 }
+
+int
+cnxk_nix_mac_addr_set(struct rte_eth_dev *eth_dev, struct rte_ether_addr *addr)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   int rc;
+
+   /* Update mac address at NPC */
+   rc = roc_nix_npc_mac_addr_set(nix, addr->addr_bytes);
+   if (rc)
+   goto exit;
+
+   /* Update mac address at CGX for PFs only */
+   if (!roc_nix_is_vf_or_sdp(nix)) {
+   rc = roc_nix_mac_addr_set(nix, addr->addr_bytes);
+   if (rc) {
+   /* Rollback to previous mac address */
+   roc_nix_npc_mac_addr_set(nix, dev->mac_addr);
+   goto exit;
+   }
+   }
+
+   /* Update mac address to cnxk ethernet device */
+   rte_memcpy(dev->mac_addr, addr->addr_bytes, RTE_ETHER_ADDR_LEN);
+
+exit:
+   return rc;
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 33/62] net/cnxk: add MTU set device operation

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

This Patch implements mtu set dev op for cn9k and cn10k platforms.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/cnxk.rst  |  1 +
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 doc/guides/nics/features/cnxk_vf.ini  |  1 +
 drivers/net/cnxk/cnxk_ethdev.c| 51 +++
 drivers/net/cnxk/cnxk_ethdev.h|  5 ++-
 drivers/net/cnxk/cnxk_ethdev_ops.c| 77 ++-
 7 files changed, 135 insertions(+), 2 deletions(-)

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 15911ee..d34d3fa 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -24,6 +24,7 @@ Features of the CNXK Ethdev PMD are:
 - Receiver Side Scaling (RSS)
 - Inner and Outer Checksum offload
 - Link state information
+- MTU update
 - Scatter-Gather IO support
 - Vector Poll mode driver
 
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 02be26b..6fef725 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -15,6 +15,7 @@ Runtime Tx queue setup = Y
 Fast mbuf free   = Y
 Free Tx mbuf on demand = Y
 Queue start/stop = Y
+MTU update   = Y
 TSO  = Y
 RSS hash = Y
 Inner RSS= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 8c63853..79cb1e2 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -15,6 +15,7 @@ Runtime Tx queue setup = Y
 Fast mbuf free   = Y
 Free Tx mbuf on demand = Y
 Queue start/stop = Y
+MTU update   = Y
 RSS hash = Y
 Inner RSS= Y
 Jumbo frame  = Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index a1bd49b..5cc9f3f 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -14,6 +14,7 @@ Runtime Tx queue setup = Y
 Fast mbuf free   = Y
 Free Tx mbuf on demand = Y
 Queue start/stop = Y
+MTU update   = Y
 TSO  = Y
 RSS hash = Y
 Inner RSS= Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 8908016..8c5a1f6 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -37,6 +37,50 @@ nix_get_speed_capa(struct cnxk_eth_dev *dev)
return speed_capa;
 }
 
+static void
+nix_enable_mseg_on_jumbo(struct cnxk_eth_rxq_sp *rxq)
+{
+   struct rte_pktmbuf_pool_private *mbp_priv;
+   struct rte_eth_dev *eth_dev;
+   struct cnxk_eth_dev *dev;
+   uint32_t buffsz;
+
+   dev = rxq->dev;
+   eth_dev = dev->eth_dev;
+
+   /* Get rx buffer size */
+   mbp_priv = rte_mempool_get_priv(rxq->qconf.mp);
+   buffsz = mbp_priv->mbuf_data_room_size - RTE_PKTMBUF_HEADROOM;
+
+   if (eth_dev->data->dev_conf.rxmode.max_rx_pkt_len > buffsz) {
+   dev->rx_offloads |= DEV_RX_OFFLOAD_SCATTER;
+   dev->tx_offloads |= DEV_TX_OFFLOAD_MULTI_SEGS;
+   }
+}
+
+static int
+nix_recalc_mtu(struct rte_eth_dev *eth_dev)
+{
+   struct rte_eth_dev_data *data = eth_dev->data;
+   struct cnxk_eth_rxq_sp *rxq;
+   uint16_t mtu;
+   int rc;
+
+   rxq = ((struct cnxk_eth_rxq_sp *)data->rx_queues[0]) - 1;
+   /* Setup scatter mode if needed by jumbo */
+   nix_enable_mseg_on_jumbo(rxq);
+
+   /* Setup MTU based on max_rx_pkt_len */
+   mtu = data->dev_conf.rxmode.max_rx_pkt_len - CNXK_NIX_L2_OVERHEAD +
+   CNXK_NIX_MAX_VTAG_ACT_SIZE;
+
+   rc = cnxk_nix_mtu_set(eth_dev, mtu);
+   if (rc)
+   plt_err("Failed to set default MTU size, rc=%d", rc);
+
+   return rc;
+}
+
 uint64_t
 cnxk_nix_rxq_mbuf_setup(struct cnxk_eth_dev *dev)
 {
@@ -990,6 +1034,12 @@ cnxk_nix_dev_start(struct rte_eth_dev *eth_dev)
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
int rc, i;
 
+   if (eth_dev->data->nb_rx_queues != 0) {
+   rc = nix_recalc_mtu(eth_dev);
+   if (rc)
+   return rc;
+   }
+
/* Start rx queues */
for (i = 0; i < eth_dev->data->nb_rx_queues; i++) {
rc = cnxk_nix_rx_queue_start(eth_dev, i);
@@ -1034,6 +1084,7 @@ cnxk_nix_dev_start(struct rte_eth_dev *eth_dev)
 
 /* CNXK platform independent eth dev ops */
 struct eth_dev_ops cnxk_eth_dev_ops = {
+   .mtu_set = cnxk_nix_mtu_set,
.mac_addr_set = cnxk_nix_mac_addr_set,
.dev_infos_get = cnxk_nix_info_get,
.link_update = cnxk_nix_link_update,
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index a5380a5..c216dd5 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -28,7 +28,9 @@
 #define CNXK_NIX_MAX_VTAG_ACT_SIZE (4 * CNXK_NIX_MAX_VTAG_INS)
 
 /* ETH_HLEN+ETH

[dpdk-dev] [PATCH v3 34/62] net/cnxk: add promiscuous mode enable and disable

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Add device operations to enable and disable promisc mode
for cn9k and cn10k.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/cnxk.rst  |  1 +
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 drivers/net/cnxk/cnxk_ethdev.c|  2 ++
 drivers/net/cnxk/cnxk_ethdev.h|  2 ++
 drivers/net/cnxk/cnxk_ethdev_ops.c| 56 +++
 6 files changed, 63 insertions(+)

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index d34d3fa..87ceda5 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -17,6 +17,7 @@ Features
 Features of the CNXK Ethdev PMD are:
 
 - Packet type information
+- Promiscuous mode
 - Jumbo frames
 - SR-IOV VF
 - Lock-free Tx queue
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 6fef725..9b2e163 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -17,6 +17,7 @@ Free Tx mbuf on demand = Y
 Queue start/stop = Y
 MTU update   = Y
 TSO  = Y
+Promiscuous mode = Y
 RSS hash = Y
 Inner RSS= Y
 Jumbo frame  = Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 79cb1e2..31471e0 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -16,6 +16,7 @@ Fast mbuf free   = Y
 Free Tx mbuf on demand = Y
 Queue start/stop = Y
 MTU update   = Y
+Promiscuous mode = Y
 RSS hash = Y
 Inner RSS= Y
 Jumbo frame  = Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 8c5a1f6..d3dbe54 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1095,6 +1095,8 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.rx_queue_start = cnxk_nix_rx_queue_start,
.rx_queue_stop = cnxk_nix_rx_queue_stop,
.dev_supported_ptypes_get = cnxk_nix_supported_ptypes_get,
+   .promiscuous_enable = cnxk_nix_promisc_enable,
+   .promiscuous_disable = cnxk_nix_promisc_disable,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index c216dd5..2fce20e 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -223,6 +223,8 @@ int cnxk_nix_remove(struct rte_pci_device *pci_dev);
 int cnxk_nix_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu);
 int cnxk_nix_mac_addr_set(struct rte_eth_dev *eth_dev,
  struct rte_ether_addr *addr);
+int cnxk_nix_promisc_enable(struct rte_eth_dev *eth_dev);
+int cnxk_nix_promisc_disable(struct rte_eth_dev *eth_dev);
 int cnxk_nix_info_get(struct rte_eth_dev *eth_dev,
  struct rte_eth_dev_info *dev_info);
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 21b55c4..6feb3a9 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -173,3 +173,59 @@ cnxk_nix_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu)
 exit:
return rc;
 }
+
+int
+cnxk_nix_promisc_enable(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   int rc = 0;
+
+   if (roc_nix_is_vf_or_sdp(nix))
+   return rc;
+
+   rc = roc_nix_npc_promisc_ena_dis(nix, true);
+   if (rc) {
+   plt_err("Failed to setup promisc mode in npc, rc=%d(%s)", rc,
+   roc_error_msg_get(rc));
+   return rc;
+   }
+
+   rc = roc_nix_mac_promisc_mode_enable(nix, true);
+   if (rc) {
+   plt_err("Failed to setup promisc mode in mac, rc=%d(%s)", rc,
+   roc_error_msg_get(rc));
+   roc_nix_npc_promisc_ena_dis(nix, false);
+   return rc;
+   }
+
+   return 0;
+}
+
+int
+cnxk_nix_promisc_disable(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   int rc = 0;
+
+   if (roc_nix_is_vf_or_sdp(nix))
+   return rc;
+
+   rc = roc_nix_npc_promisc_ena_dis(nix, false);
+   if (rc < 0) {
+   plt_err("Failed to setup promisc mode in npc, rc=%d(%s)", rc,
+   roc_error_msg_get(rc));
+   return rc;
+   }
+
+   rc = roc_nix_mac_promisc_mode_enable(nix, false);
+   if (rc) {
+   plt_err("Failed to setup promisc mode in mac, rc=%d(%s)", rc,
+   roc_error_msg_get(rc));
+   roc_nix_npc_promisc_ena_dis(nix, true);
+   return rc;
+   }
+
+   return 0;
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 35/62] net/cnxk: add DMAC filter support

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

DMAC filter support is added for cn9k and cn10k platforms.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/cnxk.rst  |  1 +
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 drivers/net/cnxk/cnxk_ethdev.c|  2 ++
 drivers/net/cnxk/cnxk_ethdev.h|  5 
 drivers/net/cnxk/cnxk_ethdev_ops.c| 44 ---
 6 files changed, 51 insertions(+), 3 deletions(-)

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 87ceda5..af7141f 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -23,6 +23,7 @@ Features of the CNXK Ethdev PMD are:
 - Lock-free Tx queue
 - Multiple queues for TX and RX
 - Receiver Side Scaling (RSS)
+- MAC filtering
 - Inner and Outer Checksum offload
 - Link state information
 - MTU update
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 9b2e163..20d4d12 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -18,6 +18,7 @@ Queue start/stop = Y
 MTU update   = Y
 TSO  = Y
 Promiscuous mode = Y
+Unicast MAC filter   = Y
 RSS hash = Y
 Inner RSS= Y
 Jumbo frame  = Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 31471e0..e1de8ab 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -17,6 +17,7 @@ Free Tx mbuf on demand = Y
 Queue start/stop = Y
 MTU update   = Y
 Promiscuous mode = Y
+Unicast MAC filter   = Y
 RSS hash = Y
 Inner RSS= Y
 Jumbo frame  = Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index d3dbe54..a813c08 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1085,6 +1085,8 @@ cnxk_nix_dev_start(struct rte_eth_dev *eth_dev)
 /* CNXK platform independent eth dev ops */
 struct eth_dev_ops cnxk_eth_dev_ops = {
.mtu_set = cnxk_nix_mtu_set,
+   .mac_addr_add = cnxk_nix_mac_addr_add,
+   .mac_addr_remove = cnxk_nix_mac_addr_del,
.mac_addr_set = cnxk_nix_mac_addr_set,
.dev_infos_get = cnxk_nix_info_get,
.link_update = cnxk_nix_link_update,
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 2fce20e..ab94b99 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -139,6 +139,7 @@ struct cnxk_eth_dev {
 
/* Max macfilter entries */
uint8_t max_mac_entries;
+   bool dmac_filter_enable;
 
uint16_t flags;
uint8_t ptype_disable;
@@ -221,6 +222,10 @@ int cnxk_nix_probe(struct rte_pci_driver *pci_drv,
   struct rte_pci_device *pci_dev);
 int cnxk_nix_remove(struct rte_pci_device *pci_dev);
 int cnxk_nix_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu);
+int cnxk_nix_mac_addr_add(struct rte_eth_dev *eth_dev,
+ struct rte_ether_addr *addr, uint32_t index,
+ uint32_t pool);
+void cnxk_nix_mac_addr_del(struct rte_eth_dev *eth_dev, uint32_t index);
 int cnxk_nix_mac_addr_set(struct rte_eth_dev *eth_dev,
  struct rte_ether_addr *addr);
 int cnxk_nix_promisc_enable(struct rte_eth_dev *eth_dev);
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 6feb3a9..fc60576 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -101,6 +101,43 @@ cnxk_nix_mac_addr_set(struct rte_eth_dev *eth_dev, struct 
rte_ether_addr *addr)
 }
 
 int
+cnxk_nix_mac_addr_add(struct rte_eth_dev *eth_dev, struct rte_ether_addr *addr,
+ uint32_t index, uint32_t pool)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   int rc;
+
+   PLT_SET_USED(index);
+   PLT_SET_USED(pool);
+
+   rc = roc_nix_mac_addr_add(nix, addr->addr_bytes);
+   if (rc < 0) {
+   plt_err("Failed to add mac address, rc=%d", rc);
+   return rc;
+   }
+
+   /* Enable promiscuous mode at NIX level */
+   roc_nix_npc_promisc_ena_dis(nix, true);
+   dev->dmac_filter_enable = true;
+   eth_dev->data->promiscuous = false;
+
+   return 0;
+}
+
+void
+cnxk_nix_mac_addr_del(struct rte_eth_dev *eth_dev, uint32_t index)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   int rc;
+
+   rc = roc_nix_mac_addr_del(nix, index);
+   if (rc)
+   plt_err("Failed to delete mac address, rc=%d", rc);
+}
+
+int
 cnxk_nix_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu)
 {
uint32_t old_frame_size, frame_size = mtu + CNXK_NIX_L2_OVERHEAD;
@@ -212,8 +249,8 @@ cnxk_nix_promisc_disable(struct rte_eth_dev *eth_dev)
if (roc_nix_is_vf_or_sdp(nix))

[dpdk-dev] [PATCH v3 36/62] net/cnxk: add all multicast enable/disable ethops

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

L2 multicast packets can be allowed or blocked. Patch implements
corresponding ethops.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 drivers/net/cnxk/cnxk_ethdev.c|  2 ++
 drivers/net/cnxk/cnxk_ethdev.h|  2 ++
 drivers/net/cnxk/cnxk_ethdev_ops.c| 17 +
 5 files changed, 23 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 20d4d12..b41af2d 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -18,6 +18,7 @@ Queue start/stop = Y
 MTU update   = Y
 TSO  = Y
 Promiscuous mode = Y
+Allmulticast mode= Y
 Unicast MAC filter   = Y
 RSS hash = Y
 Inner RSS= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index e1de8ab..7fe8018 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -17,6 +17,7 @@ Free Tx mbuf on demand = Y
 Queue start/stop = Y
 MTU update   = Y
 Promiscuous mode = Y
+Allmulticast mode= Y
 Unicast MAC filter   = Y
 RSS hash = Y
 Inner RSS= Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index a813c08..f3d5a9d 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1099,6 +1099,8 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.dev_supported_ptypes_get = cnxk_nix_supported_ptypes_get,
.promiscuous_enable = cnxk_nix_promisc_enable,
.promiscuous_disable = cnxk_nix_promisc_disable,
+   .allmulticast_enable = cnxk_nix_allmulticast_enable,
+   .allmulticast_disable = cnxk_nix_allmulticast_disable,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index ab94b99..70bc374 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -230,6 +230,8 @@ int cnxk_nix_mac_addr_set(struct rte_eth_dev *eth_dev,
  struct rte_ether_addr *addr);
 int cnxk_nix_promisc_enable(struct rte_eth_dev *eth_dev);
 int cnxk_nix_promisc_disable(struct rte_eth_dev *eth_dev);
+int cnxk_nix_allmulticast_enable(struct rte_eth_dev *eth_dev);
+int cnxk_nix_allmulticast_disable(struct rte_eth_dev *eth_dev);
 int cnxk_nix_info_get(struct rte_eth_dev *eth_dev,
  struct rte_eth_dev_info *dev_info);
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index fc60576..61ecbab 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -267,3 +267,20 @@ cnxk_nix_promisc_disable(struct rte_eth_dev *eth_dev)
dev->dmac_filter_enable = false;
return 0;
 }
+
+int
+cnxk_nix_allmulticast_enable(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   return roc_nix_npc_mcast_config(&dev->nix, true, false);
+}
+
+int
+cnxk_nix_allmulticast_disable(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   return roc_nix_npc_mcast_config(&dev->nix, false,
+   eth_dev->data->promiscuous);
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 37/62] net/cnxk: add Rx/Tx burst mode get ops

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch implements ethdev operations to get Rx and Tx burst
mode.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/features/cnxk.ini |   1 +
 doc/guides/nics/features/cnxk_vec.ini |   1 +
 doc/guides/nics/features/cnxk_vf.ini  |   1 +
 drivers/net/cnxk/cnxk_ethdev.c|   2 +
 drivers/net/cnxk/cnxk_ethdev.h|   4 ++
 drivers/net/cnxk/cnxk_ethdev_ops.c| 127 ++
 6 files changed, 136 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index b41af2d..298f167 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -12,6 +12,7 @@ Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
+Burst mode info  = Y
 Fast mbuf free   = Y
 Free Tx mbuf on demand = Y
 Queue start/stop = Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 7fe8018..a673cc1 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -12,6 +12,7 @@ Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
+Burst mode info  = Y
 Fast mbuf free   = Y
 Free Tx mbuf on demand = Y
 Queue start/stop = Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 5cc9f3f..335d082 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -11,6 +11,7 @@ Link status  = Y
 Link status event= Y
 Runtime Rx queue setup = Y
 Runtime Tx queue setup = Y
+Burst mode info  = Y
 Fast mbuf free   = Y
 Free Tx mbuf on demand = Y
 Queue start/stop = Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index f3d5a9d..4ec0dfb 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1101,6 +1101,8 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.promiscuous_disable = cnxk_nix_promisc_disable,
.allmulticast_enable = cnxk_nix_allmulticast_enable,
.allmulticast_disable = cnxk_nix_allmulticast_disable,
+   .rx_burst_mode_get = cnxk_nix_rx_burst_mode_get,
+   .tx_burst_mode_get = cnxk_nix_tx_burst_mode_get,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 70bc374..aea0005 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -234,6 +234,10 @@ int cnxk_nix_allmulticast_enable(struct rte_eth_dev 
*eth_dev);
 int cnxk_nix_allmulticast_disable(struct rte_eth_dev *eth_dev);
 int cnxk_nix_info_get(struct rte_eth_dev *eth_dev,
  struct rte_eth_dev_info *dev_info);
+int cnxk_nix_rx_burst_mode_get(struct rte_eth_dev *eth_dev, uint16_t queue_id,
+  struct rte_eth_burst_mode *mode);
+int cnxk_nix_tx_burst_mode_get(struct rte_eth_dev *eth_dev, uint16_t queue_id,
+  struct rte_eth_burst_mode *mode);
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
 int cnxk_nix_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
uint16_t nb_desc, uint16_t fp_tx_q_sz,
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 61ecbab..7ae961a 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -72,6 +72,133 @@ cnxk_nix_info_get(struct rte_eth_dev *eth_dev, struct 
rte_eth_dev_info *devinfo)
 }
 
 int
+cnxk_nix_rx_burst_mode_get(struct rte_eth_dev *eth_dev, uint16_t queue_id,
+  struct rte_eth_burst_mode *mode)
+{
+   ssize_t bytes = 0, str_size = RTE_ETH_BURST_MODE_INFO_SIZE, rc;
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   const struct burst_info {
+   uint64_t flags;
+   const char *output;
+   } rx_offload_map[] = {
+   {DEV_RX_OFFLOAD_VLAN_STRIP, " VLAN Strip,"},
+   {DEV_RX_OFFLOAD_IPV4_CKSUM, " Inner IPv4 Checksum,"},
+   {DEV_RX_OFFLOAD_UDP_CKSUM, " UDP Checksum,"},
+   {DEV_RX_OFFLOAD_TCP_CKSUM, " TCP Checksum,"},
+   {DEV_RX_OFFLOAD_TCP_LRO, " TCP LRO,"},
+   {DEV_RX_OFFLOAD_QINQ_STRIP, " QinQ VLAN Strip,"},
+   {DEV_RX_OFFLOAD_OUTER_IPV4_CKSUM, " Outer IPv4 Checksum,"},
+   {DEV_RX_OFFLOAD_MACSEC_STRIP, " MACsec Strip,"},
+   {DEV_RX_OFFLOAD_HEADER_SPLIT, " Header Split,"},
+   {DEV_RX_OFFLOAD_VLAN_FILTER, " VLAN Filter,"},
+   {DEV_RX_OFFLOAD_VLAN_EXTEND, " VLAN Extend,"},
+   {DEV_RX_OFFLOAD_JUMBO_FRAME, " Jumbo Frame,"},
+   {DEV_RX_OFFLOAD_SCATTER, " Scattered,"},
+   {DEV_RX_OFFLOAD_TIMESTAMP, " Timestamp,"},
+   {DEV_RX_OFFLOAD_SECURITY, " Security,"},
+   {DEV_RX_OFFLOAD_KEEP_CRC, " Keep CRC,"},
+   {D

[dpdk-dev] [PATCH v3 38/62] net/cnxk: add flow ctrl set/get ops

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch implements set and get operations for flow control.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/cnxk.rst  |   1 +
 doc/guides/nics/features/cnxk.ini |   1 +
 doc/guides/nics/features/cnxk_vec.ini |   1 +
 drivers/net/cnxk/cnxk_ethdev.c|  74 +
 drivers/net/cnxk/cnxk_ethdev.h|  13 +
 drivers/net/cnxk/cnxk_ethdev_ops.c| 101 ++
 6 files changed, 191 insertions(+)

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index af7141f..88ca313 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -26,6 +26,7 @@ Features of the CNXK Ethdev PMD are:
 - MAC filtering
 - Inner and Outer Checksum offload
 - Link state information
+- Link flow control
 - MTU update
 - Scatter-Gather IO support
 - Vector Poll mode driver
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 298f167..afd0f01 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -23,6 +23,7 @@ Allmulticast mode= Y
 Unicast MAC filter   = Y
 RSS hash = Y
 Inner RSS= Y
+Flow control = Y
 Jumbo frame  = Y
 Scattered Rx = Y
 L3 checksum offload  = Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index a673cc1..4bd11ce 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -22,6 +22,7 @@ Allmulticast mode= Y
 Unicast MAC filter   = Y
 RSS hash = Y
 Inner RSS= Y
+Flow control = Y
 Jumbo frame  = Y
 L3 checksum offload  = Y
 L4 checksum offload  = Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 4ec0dfb..4fe9c1c 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -81,6 +81,55 @@ nix_recalc_mtu(struct rte_eth_dev *eth_dev)
return rc;
 }
 
+static int
+nix_init_flow_ctrl_config(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct cnxk_fc_cfg *fc = &dev->fc_cfg;
+   struct rte_eth_fc_conf fc_conf = {0};
+   int rc;
+
+   /* Both Rx & Tx flow ctrl get enabled(RTE_FC_FULL) in HW
+* by AF driver, update those info in PMD structure.
+*/
+   rc = cnxk_nix_flow_ctrl_get(eth_dev, &fc_conf);
+   if (rc)
+   goto exit;
+
+   fc->mode = fc_conf.mode;
+   fc->rx_pause = (fc_conf.mode == RTE_FC_FULL) ||
+   (fc_conf.mode == RTE_FC_RX_PAUSE);
+   fc->tx_pause = (fc_conf.mode == RTE_FC_FULL) ||
+   (fc_conf.mode == RTE_FC_TX_PAUSE);
+
+exit:
+   return rc;
+}
+
+static int
+nix_update_flow_ctrl_config(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct cnxk_fc_cfg *fc = &dev->fc_cfg;
+   struct rte_eth_fc_conf fc_cfg = {0};
+
+   if (roc_nix_is_vf_or_sdp(&dev->nix))
+   return 0;
+
+   fc_cfg.mode = fc->mode;
+
+   /* To avoid Link credit deadlock on Ax, disable Tx FC if it's enabled */
+   if (roc_model_is_cn96_Ax() &&
+   (fc_cfg.mode == RTE_FC_FULL || fc_cfg.mode == RTE_FC_RX_PAUSE)) {
+   fc_cfg.mode =
+   (fc_cfg.mode == RTE_FC_FULL ||
+   fc_cfg.mode == RTE_FC_TX_PAUSE) ?
+   RTE_FC_TX_PAUSE : RTE_FC_NONE;
+   }
+
+   return cnxk_nix_flow_ctrl_set(eth_dev, &fc_cfg);
+}
+
 uint64_t
 cnxk_nix_rxq_mbuf_setup(struct cnxk_eth_dev *dev)
 {
@@ -674,6 +723,7 @@ cnxk_nix_configure(struct rte_eth_dev *eth_dev)
struct rte_eth_rxmode *rxmode = &conf->rxmode;
struct rte_eth_txmode *txmode = &conf->txmode;
char ea_fmt[RTE_ETHER_ADDR_FMT_SIZE];
+   struct roc_nix_fc_cfg fc_cfg = {0};
struct roc_nix *nix = &dev->nix;
struct rte_ether_addr *ea;
uint8_t nb_rxq, nb_txq;
@@ -855,6 +905,21 @@ cnxk_nix_configure(struct rte_eth_dev *eth_dev)
goto cq_fini;
}
 
+   /* Init flow control configuration */
+   fc_cfg.cq_cfg_valid = false;
+   fc_cfg.rxchan_cfg.enable = true;
+   rc = roc_nix_fc_config_set(nix, &fc_cfg);
+   if (rc) {
+   plt_err("Failed to initialize flow control rc=%d", rc);
+   goto cq_fini;
+   }
+
+   /* Update flow control configuration to PMD */
+   rc = nix_init_flow_ctrl_config(eth_dev);
+   if (rc) {
+   plt_err("Failed to initialize flow control rc=%d", rc);
+   goto cq_fini;
+   }
/*
 * Restore queue config when reconfigure followed by
 * reconfigure and no queue configure invoked from application case.
@@ -1054,6 +1119,13 @@ cnxk_nix_dev_start(struct rte_eth_dev *eth_dev)
return rc;
}
 
+   /* Update Flow con

[dpdk-dev] [PATCH v3 39/62] net/cnxk: add link up/down operations

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch implements link up/down ethdev operations for
cn9k and cn10k platform.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cnxk_ethdev.c |  4 +++-
 drivers/net/cnxk/cnxk_ethdev.h |  4 
 drivers/net/cnxk/cnxk_ethdev_ops.c | 47 ++
 3 files changed, 54 insertions(+), 1 deletion(-)

diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 4fe9c1c..c6072c1 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -963,7 +963,7 @@ cnxk_nix_configure(struct rte_eth_dev *eth_dev)
return rc;
 }
 
-static int
+int
 cnxk_nix_tx_queue_start(struct rte_eth_dev *eth_dev, uint16_t qid)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
@@ -1177,6 +1177,8 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.tx_burst_mode_get = cnxk_nix_tx_burst_mode_get,
.flow_ctrl_get = cnxk_nix_flow_ctrl_get,
.flow_ctrl_set = cnxk_nix_flow_ctrl_set,
+   .dev_set_link_up = cnxk_nix_set_link_up,
+   .dev_set_link_down = cnxk_nix_set_link_down,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index e788a42..5e982f9 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -251,6 +251,9 @@ int cnxk_nix_flow_ctrl_set(struct rte_eth_dev *eth_dev,
   struct rte_eth_fc_conf *fc_conf);
 int cnxk_nix_flow_ctrl_get(struct rte_eth_dev *eth_dev,
   struct rte_eth_fc_conf *fc_conf);
+int cnxk_nix_set_link_up(struct rte_eth_dev *eth_dev);
+int cnxk_nix_set_link_down(struct rte_eth_dev *eth_dev);
+
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
 int cnxk_nix_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
uint16_t nb_desc, uint16_t fp_tx_q_sz,
@@ -259,6 +262,7 @@ int cnxk_nix_rx_queue_setup(struct rte_eth_dev *eth_dev, 
uint16_t qid,
uint16_t nb_desc, uint16_t fp_rx_q_sz,
const struct rte_eth_rxconf *rx_conf,
struct rte_mempool *mp);
+int cnxk_nix_tx_queue_start(struct rte_eth_dev *eth_dev, uint16_t qid);
 int cnxk_nix_tx_queue_stop(struct rte_eth_dev *eth_dev, uint16_t qid);
 int cnxk_nix_dev_start(struct rte_eth_dev *eth_dev);
 
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 420c928..b58d21e 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -512,3 +512,50 @@ cnxk_nix_allmulticast_disable(struct rte_eth_dev *eth_dev)
return roc_nix_npc_mcast_config(&dev->nix, false,
eth_dev->data->promiscuous);
 }
+
+int
+cnxk_nix_set_link_up(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   int rc, i;
+
+   if (roc_nix_is_vf_or_sdp(nix))
+   return -ENOTSUP;
+
+   rc = roc_nix_mac_link_state_set(nix, true);
+   if (rc)
+   goto exit;
+
+   /* Start tx queues  */
+   for (i = 0; i < eth_dev->data->nb_tx_queues; i++) {
+   rc = cnxk_nix_tx_queue_start(eth_dev, i);
+   if (rc)
+   goto exit;
+   }
+
+exit:
+   return rc;
+}
+
+int
+cnxk_nix_set_link_down(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   int rc, i;
+
+   if (roc_nix_is_vf_or_sdp(nix))
+   return -ENOTSUP;
+
+   /* Stop tx queues  */
+   for (i = 0; i < eth_dev->data->nb_tx_queues; i++) {
+   rc = cnxk_nix_tx_queue_stop(eth_dev, i);
+   if (rc)
+   goto exit;
+   }
+
+   rc = roc_nix_mac_link_state_set(nix, false);
+exit:
+   return rc;
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 40/62] net/cnxk: add EEPROM module info get operations

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch implements eeprom module info get ethops for cn9k and
cn10k platforms.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 doc/guides/nics/features/cnxk_vf.ini  |  1 +
 drivers/net/cnxk/cnxk_ethdev.c|  2 ++
 drivers/net/cnxk/cnxk_ethdev.h|  4 
 drivers/net/cnxk/cnxk_ethdev_ops.c| 39 +++
 6 files changed, 48 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index afd0f01..b1e8641 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -31,6 +31,7 @@ L4 checksum offload  = Y
 Inner L3 checksum= Y
 Inner L4 checksum= Y
 Packet type parsing  = Y
+Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 4bd11ce..0f99634 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -29,6 +29,7 @@ L4 checksum offload  = Y
 Inner L3 checksum= Y
 Inner L4 checksum= Y
 Packet type parsing  = Y
+Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 335d082..cecced9 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -26,6 +26,7 @@ L4 checksum offload  = Y
 Inner L3 checksum= Y
 Inner L4 checksum= Y
 Packet type parsing  = Y
+Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index c6072c1..2e34867 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1179,6 +1179,8 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.flow_ctrl_set = cnxk_nix_flow_ctrl_set,
.dev_set_link_up = cnxk_nix_set_link_up,
.dev_set_link_down = cnxk_nix_set_link_down,
+   .get_module_info = cnxk_nix_get_module_info,
+   .get_module_eeprom = cnxk_nix_get_module_eeprom,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 5e982f9..083af29 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -253,6 +253,10 @@ int cnxk_nix_flow_ctrl_get(struct rte_eth_dev *eth_dev,
   struct rte_eth_fc_conf *fc_conf);
 int cnxk_nix_set_link_up(struct rte_eth_dev *eth_dev);
 int cnxk_nix_set_link_down(struct rte_eth_dev *eth_dev);
+int cnxk_nix_get_module_info(struct rte_eth_dev *eth_dev,
+struct rte_eth_dev_module_info *modinfo);
+int cnxk_nix_get_module_eeprom(struct rte_eth_dev *eth_dev,
+  struct rte_dev_eeprom_info *info);
 
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
 int cnxk_nix_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index b58d21e..6e83a63 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -559,3 +559,42 @@ cnxk_nix_set_link_down(struct rte_eth_dev *eth_dev)
 exit:
return rc;
 }
+
+int
+cnxk_nix_get_module_info(struct rte_eth_dev *eth_dev,
+struct rte_eth_dev_module_info *modinfo)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix_eeprom_info eeprom_info = {0};
+   struct roc_nix *nix = &dev->nix;
+   int rc;
+
+   rc = roc_nix_eeprom_info_get(nix, &eeprom_info);
+   if (rc)
+   return rc;
+
+   modinfo->type = eeprom_info.sff_id;
+   modinfo->eeprom_len = ROC_NIX_EEPROM_SIZE;
+   return 0;
+}
+
+int
+cnxk_nix_get_module_eeprom(struct rte_eth_dev *eth_dev,
+  struct rte_dev_eeprom_info *info)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix_eeprom_info eeprom_info = {0};
+   struct roc_nix *nix = &dev->nix;
+   int rc = -EINVAL;
+
+   if (!info->data || !info->length ||
+   (info->offset + info->length > ROC_NIX_EEPROM_SIZE))
+   return rc;
+
+   rc = roc_nix_eeprom_info_get(nix, &eeprom_info);
+   if (rc)
+   return rc;
+
+   rte_memcpy(info->data, eeprom_info.buf + info->offset, info->length);
+   return 0;
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 41/62] net/cnxk: add Rx queue interrupt enable/disable ops

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Application may choose to enable/disable interrupts on Rx queues
so that application can select its processing if no packets are
available on queues for a longer period.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/cnxk.rst  |  1 +
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 doc/guides/nics/features/cnxk_vf.ini  |  1 +
 drivers/net/cnxk/cnxk_ethdev.c|  2 ++
 drivers/net/cnxk/cnxk_ethdev.h|  4 
 drivers/net/cnxk/cnxk_ethdev_ops.c| 19 +++
 7 files changed, 29 insertions(+)

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 88ca313..6678486 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -30,6 +30,7 @@ Features of the CNXK Ethdev PMD are:
 - MTU update
 - Scatter-Gather IO support
 - Vector Poll mode driver
+- Support Rx interrupt
 
 Prerequisites
 -
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index b1e8641..e5669f5 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -5,6 +5,7 @@
 ;
 [Features]
 Speed capabilities   = Y
+Rx interrupt = Y
 Lock-free Tx queue   = Y
 SR-IOV   = Y
 Multiprocess aware   = Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 0f99634..dff0c9b 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -5,6 +5,7 @@
 ;
 [Features]
 Speed capabilities   = Y
+Rx interrupt = Y
 Lock-free Tx queue   = Y
 SR-IOV   = Y
 Multiprocess aware   = Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index cecced9..b950d2f 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -5,6 +5,7 @@
 ;
 [Features]
 Speed capabilities   = Y
+Rx interrupt = Y
 Lock-free Tx queue   = Y
 Multiprocess aware   = Y
 Link status  = Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 2e34867..fd8d1fb 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1181,6 +1181,8 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.dev_set_link_down = cnxk_nix_set_link_down,
.get_module_info = cnxk_nix_get_module_info,
.get_module_eeprom = cnxk_nix_get_module_eeprom,
+   .rx_queue_intr_enable = cnxk_nix_rx_queue_intr_enable,
+   .rx_queue_intr_disable = cnxk_nix_rx_queue_intr_disable,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 083af29..a01b72a 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -257,6 +257,10 @@ int cnxk_nix_get_module_info(struct rte_eth_dev *eth_dev,
 struct rte_eth_dev_module_info *modinfo);
 int cnxk_nix_get_module_eeprom(struct rte_eth_dev *eth_dev,
   struct rte_dev_eeprom_info *info);
+int cnxk_nix_rx_queue_intr_enable(struct rte_eth_dev *eth_dev,
+ uint16_t rx_queue_id);
+int cnxk_nix_rx_queue_intr_disable(struct rte_eth_dev *eth_dev,
+  uint16_t rx_queue_id);
 
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
 int cnxk_nix_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 6e83a63..6399090 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -598,3 +598,22 @@ cnxk_nix_get_module_eeprom(struct rte_eth_dev *eth_dev,
rte_memcpy(info->data, eeprom_info.buf + info->offset, info->length);
return 0;
 }
+
+int
+cnxk_nix_rx_queue_intr_enable(struct rte_eth_dev *eth_dev, uint16_t 
rx_queue_id)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   roc_nix_rx_queue_intr_enable(&dev->nix, rx_queue_id);
+   return 0;
+}
+
+int
+cnxk_nix_rx_queue_intr_disable(struct rte_eth_dev *eth_dev,
+  uint16_t rx_queue_id)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   roc_nix_rx_queue_intr_disable(&dev->nix, rx_queue_id);
+   return 0;
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 42/62] net/cnxk: add validation API for mempool ops

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

cn9k and cn10k supports platform specific mempool ops.
This patch implements API to validate whether given mempool
ops is supported or not.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cnxk_ethdev.c |  1 +
 drivers/net/cnxk/cnxk_ethdev.h |  1 +
 drivers/net/cnxk/cnxk_ethdev_ops.c | 11 +++
 3 files changed, 13 insertions(+)

diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index fd8d1fb..627ac32 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1183,6 +1183,7 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.get_module_eeprom = cnxk_nix_get_module_eeprom,
.rx_queue_intr_enable = cnxk_nix_rx_queue_intr_enable,
.rx_queue_intr_disable = cnxk_nix_rx_queue_intr_disable,
+   .pool_ops_supported = cnxk_nix_pool_ops_supported,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index a01b72a..e6dac95 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -261,6 +261,7 @@ int cnxk_nix_rx_queue_intr_enable(struct rte_eth_dev 
*eth_dev,
  uint16_t rx_queue_id);
 int cnxk_nix_rx_queue_intr_disable(struct rte_eth_dev *eth_dev,
   uint16_t rx_queue_id);
+int cnxk_nix_pool_ops_supported(struct rte_eth_dev *eth_dev, const char *pool);
 
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
 int cnxk_nix_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 6399090..ddfaffa 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -617,3 +617,14 @@ cnxk_nix_rx_queue_intr_disable(struct rte_eth_dev *eth_dev,
roc_nix_rx_queue_intr_disable(&dev->nix, rx_queue_id);
return 0;
 }
+
+int
+cnxk_nix_pool_ops_supported(struct rte_eth_dev *eth_dev, const char *pool)
+{
+   RTE_SET_USED(eth_dev);
+
+   if (!strcmp(pool, rte_mbuf_platform_mempool_ops()))
+   return 0;
+
+   return -ENOTSUP;
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 43/62] net/cnxk: add port/queue stats

2021-06-18 Thread Nithin Dabilpuram
From: Satha Rao 

This patch implements regular port statistics and queue mapping set
api to get queue statistics

Signed-off-by: Satha Rao 
---
 doc/guides/nics/cnxk.rst  |  1 +
 doc/guides/nics/features/cnxk.ini |  2 +
 doc/guides/nics/features/cnxk_vec.ini |  2 +
 doc/guides/nics/features/cnxk_vf.ini  |  2 +
 drivers/net/cnxk/cnxk_ethdev.c|  3 ++
 drivers/net/cnxk/cnxk_ethdev.h|  8 
 drivers/net/cnxk/cnxk_stats.c | 85 +++
 drivers/net/cnxk/meson.build  |  3 +-
 8 files changed, 105 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/cnxk/cnxk_stats.c

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index 6678486..e875e55 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -25,6 +25,7 @@ Features of the CNXK Ethdev PMD are:
 - Receiver Side Scaling (RSS)
 - MAC filtering
 - Inner and Outer Checksum offload
+- Port hardware statistics
 - Link state information
 - Link flow control
 - MTU update
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index e5669f5..40952a9 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -32,6 +32,8 @@ L4 checksum offload  = Y
 Inner L3 checksum= Y
 Inner L4 checksum= Y
 Packet type parsing  = Y
+Basic stats  = Y
+Stats per queue  = Y
 Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index dff0c9b..32035bb 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -30,6 +30,8 @@ L4 checksum offload  = Y
 Inner L3 checksum= Y
 Inner L4 checksum= Y
 Packet type parsing  = Y
+Basic stats  = Y
+Stats per queue  = Y
 Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index b950d2f..8060a68 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -27,6 +27,8 @@ L4 checksum offload  = Y
 Inner L3 checksum= Y
 Inner L4 checksum= Y
 Packet type parsing  = Y
+Basic stats  = Y
+Stats per queue  = Y
 Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 627ac32..ecc7f13 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1184,6 +1184,9 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.rx_queue_intr_enable = cnxk_nix_rx_queue_intr_enable,
.rx_queue_intr_disable = cnxk_nix_rx_queue_intr_disable,
.pool_ops_supported = cnxk_nix_pool_ops_supported,
+   .queue_stats_mapping_set = cnxk_nix_queue_stats_mapping,
+   .stats_get = cnxk_nix_stats_get,
+   .stats_reset = cnxk_nix_stats_reset,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index e6dac95..860cfe1 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -191,6 +191,10 @@ struct cnxk_eth_dev {
 
/* LSO Tunnel format indices */
uint64_t lso_tun_fmt;
+
+   /* Per queue statistics counters */
+   uint32_t txq_stat_map[RTE_ETHDEV_QUEUE_STAT_CNTRS];
+   uint32_t rxq_stat_map[RTE_ETHDEV_QUEUE_STAT_CNTRS];
 };
 
 struct cnxk_eth_rxq_sp {
@@ -286,6 +290,10 @@ void cnxk_nix_toggle_flag_link_cfg(struct cnxk_eth_dev 
*dev, bool set);
 void cnxk_eth_dev_link_status_cb(struct roc_nix *nix,
 struct roc_nix_link_info *link);
 int cnxk_nix_link_update(struct rte_eth_dev *eth_dev, int wait_to_complete);
+int cnxk_nix_queue_stats_mapping(struct rte_eth_dev *dev, uint16_t queue_id,
+uint8_t stat_idx, uint8_t is_rx);
+int cnxk_nix_stats_reset(struct rte_eth_dev *dev);
+int cnxk_nix_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats);
 
 /* Lookup configuration */
 const uint32_t *cnxk_nix_supported_ptypes_get(struct rte_eth_dev *eth_dev);
diff --git a/drivers/net/cnxk/cnxk_stats.c b/drivers/net/cnxk/cnxk_stats.c
new file mode 100644
index 000..24bff0b
--- /dev/null
+++ b/drivers/net/cnxk/cnxk_stats.c
@@ -0,0 +1,85 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2021 Marvell.
+ */
+
+#include "cnxk_ethdev.h"
+
+int
+cnxk_nix_stats_get(struct rte_eth_dev *eth_dev, struct rte_eth_stats *stats)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   struct roc_nix_stats nix_stats;
+   int rc = 0, i;
+
+   rc = roc_nix_stats_get(nix, &nix_stats);
+   if (rc)
+   goto exit;
+
+   stats->opackets = nix_stats.tx_ucast;
+   stats->opackets += nix_stats.tx_mcast;
+   stats->opackets += nix_stats.tx_bcast;
+   stats->oerrors = nix_stats.tx_drop;
+

[dpdk-dev] [PATCH v3 44/62] net/cnxk: add xstats apis

2021-06-18 Thread Nithin Dabilpuram
From: Satha Rao 

Initial implementation of xstats operations.

Signed-off-by: Satha Rao 
---
 doc/guides/nics/features/cnxk.ini |   1 +
 doc/guides/nics/features/cnxk_vec.ini |   1 +
 doc/guides/nics/features/cnxk_vf.ini  |   1 +
 drivers/net/cnxk/cnxk_ethdev.c|   5 +
 drivers/net/cnxk/cnxk_ethdev.h|  11 ++
 drivers/net/cnxk/cnxk_stats.c | 235 ++
 6 files changed, 254 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 40952a9..192c15a 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -34,6 +34,7 @@ Inner L4 checksum= Y
 Packet type parsing  = Y
 Basic stats  = Y
 Stats per queue  = Y
+Extended stats   = Y
 Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 32035bb..e990480 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -32,6 +32,7 @@ Inner L4 checksum= Y
 Packet type parsing  = Y
 Basic stats  = Y
 Stats per queue  = Y
+Extended stats   = Y
 Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 8060a68..3a4417c 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -29,6 +29,7 @@ Inner L4 checksum= Y
 Packet type parsing  = Y
 Basic stats  = Y
 Stats per queue  = Y
+Extended stats   = Y
 Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index ecc7f13..27b92f9 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1187,6 +1187,11 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.queue_stats_mapping_set = cnxk_nix_queue_stats_mapping,
.stats_get = cnxk_nix_stats_get,
.stats_reset = cnxk_nix_stats_reset,
+   .xstats_get = cnxk_nix_xstats_get,
+   .xstats_get_names = cnxk_nix_xstats_get_names,
+   .xstats_reset = cnxk_nix_xstats_reset,
+   .xstats_get_by_id = cnxk_nix_xstats_get_by_id,
+   .xstats_get_names_by_id = cnxk_nix_xstats_get_names_by_id,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 860cfe1..5a0dc13 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -294,6 +294,17 @@ int cnxk_nix_queue_stats_mapping(struct rte_eth_dev *dev, 
uint16_t queue_id,
 uint8_t stat_idx, uint8_t is_rx);
 int cnxk_nix_stats_reset(struct rte_eth_dev *dev);
 int cnxk_nix_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats);
+int cnxk_nix_xstats_get(struct rte_eth_dev *eth_dev,
+   struct rte_eth_xstat *xstats, unsigned int n);
+int cnxk_nix_xstats_get_names(struct rte_eth_dev *eth_dev,
+ struct rte_eth_xstat_name *xstats_names,
+ unsigned int limit);
+int cnxk_nix_xstats_get_names_by_id(struct rte_eth_dev *eth_dev,
+   struct rte_eth_xstat_name *xstats_names,
+   const uint64_t *ids, unsigned int limit);
+int cnxk_nix_xstats_get_by_id(struct rte_eth_dev *eth_dev, const uint64_t *ids,
+ uint64_t *values, unsigned int n);
+int cnxk_nix_xstats_reset(struct rte_eth_dev *eth_dev);
 
 /* Lookup configuration */
 const uint32_t *cnxk_nix_supported_ptypes_get(struct rte_eth_dev *eth_dev);
diff --git a/drivers/net/cnxk/cnxk_stats.c b/drivers/net/cnxk/cnxk_stats.c
index 24bff0b..19bab21 100644
--- a/drivers/net/cnxk/cnxk_stats.c
+++ b/drivers/net/cnxk/cnxk_stats.c
@@ -4,6 +4,9 @@
 
 #include "cnxk_ethdev.h"
 
+#define CNXK_NB_RXQ_STATS 5
+#define CNXK_NB_TXQ_STATS 4
+
 int
 cnxk_nix_stats_get(struct rte_eth_dev *eth_dev, struct rte_eth_stats *stats)
 {
@@ -83,3 +86,235 @@ cnxk_nix_queue_stats_mapping(struct rte_eth_dev *eth_dev, 
uint16_t queue_id,
 
return 0;
 }
+
+int
+cnxk_nix_xstats_get(struct rte_eth_dev *eth_dev, struct rte_eth_xstat *xstats,
+   unsigned int n)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix_xstat roc_xstats[n];
+   struct roc_nix *nix = &dev->nix;
+   int roc_size, q, idx = 0, size;
+
+   roc_size = roc_nix_xstats_get(nix, roc_xstats, n);
+
+   if (roc_size < 0)
+   return roc_size;
+
+   /* Per Queue statistics also returned as part of xstats */
+   size = roc_size + (dev->nb_rxq * CNXK_NB_RXQ_STATS) +
+  (dev->nb_txq * CNXK_NB_TXQ_STATS);
+
+   /* If requested array do not have space then return with count */
+   if (size > (int)n || xstats == NULL)
+   return size;
+
+   for 

Re: [dpdk-dev] [PATCH] parray: introduce internal API for dynamic arrays

2021-06-18 Thread Ananyev, Konstantin

> >
> > 14/06/2021 15:15, Bruce Richardson:
> >> On Mon, Jun 14, 2021 at 02:22:42PM +0200, Morten Brørup wrote:
>  From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Thomas Monjalon
>  Sent: Monday, 14 June 2021 12.59
> 
>  Performance of access in a fixed-size array is very good
>  because of cache locality
>  and because there is a single pointer to dereference.
>  The only drawback is the lack of flexibility:
>  the size of such an array cannot be increase at runtime.
> 
>  An approach to this problem is to allocate the array at runtime,
>  being as efficient as static arrays, but still limited to a maximum.
> 
>  That's why the API rte_parray is introduced,
>  allowing to declare an array of pointer which can be resized
>  dynamically
>  and automatically at runtime while keeping a good read performance.
> 
>  After resize, the previous array is kept until the next resize
>  to avoid crashs during a read without any lock.
> 
>  Each element is a pointer to a memory chunk dynamically allocated.
>  This is not good for cache locality but it allows to keep the same
>  memory per element, no matter how the array is resized.
>  Cache locality could be improved with mempools.
>  The other drawback is having to dereference one more pointer
>  to read an element.
> 
>  There is not much locks, so the API is for internal use only.
>  This API may be used to completely remove some compilation-time
>  maximums.
> >>>
> >>> I get the purpose and overall intention of this library.
> >>>
> >>> I probably already mentioned that I prefer "embedded style 
> >>> programming" with fixed size arrays, rather than runtime
> configurability.
>  It's
> > my personal opinion, and the DPDK Tech Board clearly prefers reducing 
> > the amount of compile time configurability, so there is no
> way
> >> for
> > me to stop this progress, and I do not intend to oppose to this 
> > library. :-)
> >>>
> >>> This library is likely to become a core library of DPDK, so I think 
> >>> it is important getting it right. Could you please mention a few
>  examples
> > where you think this internal library should be used, and where it 
> > should not be used. Then it is easier to discuss if the border line
> >> between
> > control path and data plane is correct. E.g. this library is not 
> > intended to be used for dynamically sized packet queues that grow and
> >> shrink
>  in
> > the fast path.
> >>>
> >>> If the library becomes a core DPDK library, it should probably be 
> >>> public instead of internal. E.g. if the library is used to make
> > RTE_MAX_ETHPORTS dynamic instead of compile time fixed, then some 
> > applications might also need dynamically sized arrays for
> their
> > application specific per-port runtime data, and this library could 
> > serve that purpose too.
> >>>
> >>
> >> Thanks Thomas for starting this discussion and Morten for follow-up.
> >>
> >> My thinking is as follows, and I'm particularly keeping in mind the 
> >> cases
> >> of e.g. RTE_MAX_ETHPORTS, as a leading candidate here.
> >>
> >> While I dislike the hard-coded limits in DPDK, I'm also not convinced 
> >> that
> >> we should switch away from the flat arrays or that we need fully 
> >> dynamic
> >> arrays that grow/shrink at runtime for ethdevs. I would suggest a 
> >> half-way
> >> house here, where we keep the ethdevs as an array, but one 
> >> allocated/sized
> >> at runtime rather than statically. This would allow us to have a
> >> compile-time default value, but, for use cases that need it, allow use 
> >> of a
> >> flag e.g.  "max-ethdevs" to change the size of the parameter given to 
> >> the
> >> malloc call for the array.  This max limit could then be provided to 
> >> apps
> >> too if they want to match any array sizes. [Alternatively those apps 
> >> could
> >> check the provided size and error out if the size has been increased 
> >> beyond
> >> what the app is designed to use?]. There would be no extra 
> >> dereferences per
> >> rx/tx burst call in this scenario so performance should be the same as
> >> before (potentially better if array is in hugepage memory, I suppose).
> >
> > I think we need some benchmarks to decide what is the best tradeoff.
> > I spent time on this implementation, but sorry I won't have time for 
> > benchmarks.
> > Volunteers?
> 
>  I had only a quick look at your approach so far.
>  But from what I can read, in MT environment your suggestion will require
>  extra synchronization for each read-write access

[dpdk-dev] [PATCH v3 45/62] net/cnxk: add rxq/txq info get operations

2021-06-18 Thread Nithin Dabilpuram
From: Satha Rao 

Initial apis to get default queue information.

Signed-off-by: Satha Rao 
---
 drivers/net/cnxk/cnxk_ethdev.c |  2 ++
 drivers/net/cnxk/cnxk_ethdev.h |  4 
 drivers/net/cnxk/cnxk_ethdev_ops.c | 30 ++
 3 files changed, 36 insertions(+)

diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 27b92f9..288e069 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1192,6 +1192,8 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.xstats_reset = cnxk_nix_xstats_reset,
.xstats_get_by_id = cnxk_nix_xstats_get_by_id,
.xstats_get_names_by_id = cnxk_nix_xstats_get_names_by_id,
+   .rxq_info_get = cnxk_nix_rxq_info_get,
+   .txq_info_get = cnxk_nix_txq_info_get,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 5a0dc13..32977eb 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -305,6 +305,10 @@ int cnxk_nix_xstats_get_names_by_id(struct rte_eth_dev 
*eth_dev,
 int cnxk_nix_xstats_get_by_id(struct rte_eth_dev *eth_dev, const uint64_t *ids,
  uint64_t *values, unsigned int n);
 int cnxk_nix_xstats_reset(struct rte_eth_dev *eth_dev);
+void cnxk_nix_rxq_info_get(struct rte_eth_dev *eth_dev, uint16_t qid,
+  struct rte_eth_rxq_info *qinfo);
+void cnxk_nix_txq_info_get(struct rte_eth_dev *eth_dev, uint16_t qid,
+  struct rte_eth_txq_info *qinfo);
 
 /* Lookup configuration */
 const uint32_t *cnxk_nix_supported_ptypes_get(struct rte_eth_dev *eth_dev);
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index ddfaffa..55c26c4 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -628,3 +628,33 @@ cnxk_nix_pool_ops_supported(struct rte_eth_dev *eth_dev, 
const char *pool)
 
return -ENOTSUP;
 }
+
+void
+cnxk_nix_rxq_info_get(struct rte_eth_dev *eth_dev, uint16_t qid,
+ struct rte_eth_rxq_info *qinfo)
+{
+   void *rxq = eth_dev->data->rx_queues[qid];
+   struct cnxk_eth_rxq_sp *rxq_sp = cnxk_eth_rxq_to_sp(rxq);
+
+   memset(qinfo, 0, sizeof(*qinfo));
+
+   qinfo->mp = rxq_sp->qconf.mp;
+   qinfo->scattered_rx = eth_dev->data->scattered_rx;
+   qinfo->nb_desc = rxq_sp->qconf.nb_desc;
+
+   memcpy(&qinfo->conf, &rxq_sp->qconf.conf.rx, sizeof(qinfo->conf));
+}
+
+void
+cnxk_nix_txq_info_get(struct rte_eth_dev *eth_dev, uint16_t qid,
+ struct rte_eth_txq_info *qinfo)
+{
+   void *txq = eth_dev->data->tx_queues[qid];
+   struct cnxk_eth_txq_sp *txq_sp = cnxk_eth_txq_to_sp(txq);
+
+   memset(qinfo, 0, sizeof(*qinfo));
+
+   qinfo->nb_desc = txq_sp->qconf.nb_desc;
+
+   memcpy(&qinfo->conf, &txq_sp->qconf.conf.tx, sizeof(qinfo->conf));
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 46/62] net/cnxk: add device close and reset operations

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch implements device close and reset operations for cn9k
and cn10k platforms.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cnxk_ethdev.c | 35 ---
 1 file changed, 28 insertions(+), 7 deletions(-)

diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 288e069..3929e7e 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1154,6 +1154,9 @@ cnxk_nix_dev_start(struct rte_eth_dev *eth_dev)
return rc;
 }
 
+static int cnxk_nix_dev_reset(struct rte_eth_dev *eth_dev);
+static int cnxk_nix_dev_close(struct rte_eth_dev *eth_dev);
+
 /* CNXK platform independent eth dev ops */
 struct eth_dev_ops cnxk_eth_dev_ops = {
.mtu_set = cnxk_nix_mtu_set,
@@ -1165,6 +1168,8 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.tx_queue_release = cnxk_nix_tx_queue_release,
.rx_queue_release = cnxk_nix_rx_queue_release,
.dev_stop = cnxk_nix_dev_stop,
+   .dev_close = cnxk_nix_dev_close,
+   .dev_reset = cnxk_nix_dev_reset,
.tx_queue_start = cnxk_nix_tx_queue_start,
.rx_queue_start = cnxk_nix_rx_queue_start,
.rx_queue_stop = cnxk_nix_rx_queue_stop,
@@ -1304,7 +1309,7 @@ cnxk_eth_dev_init(struct rte_eth_dev *eth_dev)
 }
 
 static int
-cnxk_eth_dev_uninit(struct rte_eth_dev *eth_dev, bool mbox_close)
+cnxk_eth_dev_uninit(struct rte_eth_dev *eth_dev, bool reset)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
const struct eth_dev_ops *dev_ops = eth_dev->dev_ops;
@@ -1358,14 +1363,11 @@ cnxk_eth_dev_uninit(struct rte_eth_dev *eth_dev, bool 
mbox_close)
rte_free(eth_dev->data->mac_addrs);
eth_dev->data->mac_addrs = NULL;
 
-   /* Check if mbox close is needed */
-   if (!mbox_close)
-   return 0;
-
rc = roc_nix_dev_fini(nix);
/* Can be freed later by PMD if NPA LF is in use */
if (rc == -EAGAIN) {
-   eth_dev->data->dev_private = NULL;
+   if (!reset)
+   eth_dev->data->dev_private = NULL;
return 0;
} else if (rc) {
plt_err("Failed in nix dev fini, rc=%d", rc);
@@ -1374,6 +1376,25 @@ cnxk_eth_dev_uninit(struct rte_eth_dev *eth_dev, bool 
mbox_close)
return rc;
 }
 
+static int
+cnxk_nix_dev_close(struct rte_eth_dev *eth_dev)
+{
+   cnxk_eth_dev_uninit(eth_dev, false);
+   return 0;
+}
+
+static int
+cnxk_nix_dev_reset(struct rte_eth_dev *eth_dev)
+{
+   int rc;
+
+   rc = cnxk_eth_dev_uninit(eth_dev, true);
+   if (rc)
+   return rc;
+
+   return cnxk_eth_dev_init(eth_dev);
+}
+
 int
 cnxk_nix_remove(struct rte_pci_device *pci_dev)
 {
@@ -1384,7 +1405,7 @@ cnxk_nix_remove(struct rte_pci_device *pci_dev)
eth_dev = rte_eth_dev_allocated(pci_dev->device.name);
if (eth_dev) {
/* Cleanup eth dev */
-   rc = cnxk_eth_dev_uninit(eth_dev, true);
+   rc = cnxk_eth_dev_uninit(eth_dev, false);
if (rc)
return rc;
 
-- 
2.8.4



[dpdk-dev] [PATCH v3 47/62] net/cnxk: add pending Tx mbuf cleanup operation

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Once mbufs are transmitted, mbufs are freed by H/W. No mbufs are
accumalated as a pending mbuf.
Hence operation is NOP for cnxk platform.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cnxk_ethdev.c |  1 +
 drivers/net/cnxk/cnxk_ethdev.h |  1 +
 drivers/net/cnxk/cnxk_ethdev_ops.c | 10 ++
 3 files changed, 12 insertions(+)

diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 3929e7e..f2df0ba 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1199,6 +1199,7 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.xstats_get_names_by_id = cnxk_nix_xstats_get_names_by_id,
.rxq_info_get = cnxk_nix_rxq_info_get,
.txq_info_get = cnxk_nix_txq_info_get,
+   .tx_done_cleanup = cnxk_nix_tx_done_cleanup,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 32977eb..1c4403c 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -266,6 +266,7 @@ int cnxk_nix_rx_queue_intr_enable(struct rte_eth_dev 
*eth_dev,
 int cnxk_nix_rx_queue_intr_disable(struct rte_eth_dev *eth_dev,
   uint16_t rx_queue_id);
 int cnxk_nix_pool_ops_supported(struct rte_eth_dev *eth_dev, const char *pool);
+int cnxk_nix_tx_done_cleanup(void *txq, uint32_t free_cnt);
 
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
 int cnxk_nix_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 55c26c4..ce94eb6 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -658,3 +658,13 @@ cnxk_nix_txq_info_get(struct rte_eth_dev *eth_dev, 
uint16_t qid,
 
memcpy(&qinfo->conf, &txq_sp->qconf.conf.tx, sizeof(qinfo->conf));
 }
+
+/* It is a NOP for cnxk as HW frees the buffer on xmit */
+int
+cnxk_nix_tx_done_cleanup(void *txq, uint32_t free_cnt)
+{
+   RTE_SET_USED(txq);
+   RTE_SET_USED(free_cnt);
+
+   return 0;
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 48/62] net/cnxk: add support to configure npc

2021-06-18 Thread Nithin Dabilpuram
From: Kiran Kumar K 

Adding support to configure NPC on device initialization. This involves
reading the MKEX and initializing the necessary data.

Signed-off-by: Kiran Kumar K 
---
 drivers/net/cnxk/cnxk_ethdev.c | 25 ++---
 drivers/net/cnxk/cnxk_ethdev.h |  3 +++
 drivers/net/cnxk/cnxk_ethdev_devargs.c |  3 +++
 3 files changed, 28 insertions(+), 3 deletions(-)

diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index f2df0ba..d1a7f2e 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -8,7 +8,8 @@ nix_get_rx_offload_capa(struct cnxk_eth_dev *dev)
 {
uint64_t capa = CNXK_NIX_RX_OFFLOAD_CAPA;
 
-   if (roc_nix_is_vf_or_sdp(&dev->nix))
+   if (roc_nix_is_vf_or_sdp(&dev->nix) ||
+   dev->npc.switch_header_type == ROC_PRIV_FLAGS_HIGIG)
capa &= ~DEV_RX_OFFLOAD_TIMESTAMP;
 
return capa;
@@ -120,6 +121,7 @@ nix_update_flow_ctrl_config(struct rte_eth_dev *eth_dev)
 
/* To avoid Link credit deadlock on Ax, disable Tx FC if it's enabled */
if (roc_model_is_cn96_Ax() &&
+   dev->npc.switch_header_type != ROC_PRIV_FLAGS_HIGIG &&
(fc_cfg.mode == RTE_FC_FULL || fc_cfg.mode == RTE_FC_RX_PAUSE)) {
fc_cfg.mode =
(fc_cfg.mode == RTE_FC_FULL ||
@@ -419,8 +421,10 @@ cnxk_rss_ethdev_to_nix(struct cnxk_eth_dev *dev, uint64_t 
ethdev_rss,
 
dev->ethdev_rss_hf = ethdev_rss;
 
-   if (ethdev_rss & ETH_RSS_L2_PAYLOAD)
+   if (ethdev_rss & ETH_RSS_L2_PAYLOAD &&
+   dev->npc.switch_header_type == ROC_PRIV_FLAGS_LEN_90B) {
flowkey_cfg |= FLOW_KEY_TYPE_CH_LEN_90B;
+   }
 
if (ethdev_rss & ETH_RSS_C_VLAN)
flowkey_cfg |= FLOW_KEY_TYPE_VLAN;
@@ -837,11 +841,18 @@ cnxk_nix_configure(struct rte_eth_dev *eth_dev)
roc_nix_err_intr_ena_dis(nix, true);
roc_nix_ras_intr_ena_dis(nix, true);
 
-   if (nix->rx_ptp_ena) {
+   if (nix->rx_ptp_ena &&
+   dev->npc.switch_header_type == ROC_PRIV_FLAGS_HIGIG) {
plt_err("Both PTP and switch header enabled");
goto free_nix_lf;
}
 
+   rc = roc_nix_switch_hdr_set(nix, dev->npc.switch_header_type);
+   if (rc) {
+   plt_err("Failed to enable switch type nix_lf rc=%d", rc);
+   goto free_nix_lf;
+   }
+
/* Setup LSO if needed */
rc = nix_lso_fmt_setup(dev);
if (rc) {
@@ -1293,6 +1304,11 @@ cnxk_eth_dev_init(struct rte_eth_dev *eth_dev)
dev->speed_capa = nix_get_speed_capa(dev);
 
/* Initialize roc npc */
+   dev->npc.roc_nix = nix;
+   rc = roc_npc_init(&dev->npc);
+   if (rc)
+   goto free_mac_addrs;
+
plt_nix_dbg("Port=%d pf=%d vf=%d ver=%s hwcap=0x%" PRIx64
" rxoffload_capa=0x%" PRIx64 " txoffload_capa=0x%" PRIx64,
eth_dev->data->port_id, roc_nix_get_pf(nix),
@@ -1326,6 +1342,9 @@ cnxk_eth_dev_uninit(struct rte_eth_dev *eth_dev, bool 
reset)
 
roc_nix_npc_rx_ena_dis(nix, false);
 
+   /* Disable and free rte_flow entries */
+   roc_npc_fini(&dev->npc);
+
/* Disable link status events */
roc_nix_mac_link_event_start_stop(nix, false);
 
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 1c4403c..4879ef9 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -133,6 +133,9 @@ struct cnxk_eth_dev {
/* ROC NIX */
struct roc_nix nix;
 
+   /* ROC NPC */
+   struct roc_npc npc;
+
/* ROC RQs, SQs and CQs */
struct roc_nix_rq *rqs;
struct roc_nix_sq *sqs;
diff --git a/drivers/net/cnxk/cnxk_ethdev_devargs.c 
b/drivers/net/cnxk/cnxk_ethdev_devargs.c
index 4af2803..7fd06eb 100644
--- a/drivers/net/cnxk/cnxk_ethdev_devargs.c
+++ b/drivers/net/cnxk/cnxk_ethdev_devargs.c
@@ -150,6 +150,9 @@ cnxk_ethdev_parse_devargs(struct rte_devargs *devargs, 
struct cnxk_eth_dev *dev)
dev->nix.rss_tag_as_xor = !!rss_tag_as_xor;
dev->nix.max_sqb_count = sqb_count;
dev->nix.reta_sz = reta_sz;
+   dev->npc.flow_prealloc_size = flow_prealloc_size;
+   dev->npc.flow_max_priority = flow_max_priority;
+   dev->npc.switch_header_type = switch_header_type;
return 0;
 
 exit:
-- 
2.8.4



[dpdk-dev] [PATCH v3 49/62] net/cnxk: add initial version of rte flow support

2021-06-18 Thread Nithin Dabilpuram
From: Kiran Kumar K 

Adding initial version of rte_flow support for cnxk family device.
Supported rte_flow ops are flow_validate, flow_create, flow_crstroy,
flow_flush, flow_query, flow_isolate.

Signed-off-by: Kiran Kumar K 
---
 doc/guides/nics/cnxk.rst  |   7 +
 doc/guides/nics/features/cnxk.ini |  36 +
 drivers/net/cnxk/cn10k_ethdev.c   |  16 ++
 drivers/net/cnxk/cn10k_rte_flow.c |  30 
 drivers/net/cnxk/cn10k_rte_flow.h |  17 ++
 drivers/net/cnxk/cn9k_ethdev.c|  16 ++
 drivers/net/cnxk/cn9k_rte_flow.c  |  30 
 drivers/net/cnxk/cn9k_rte_flow.h  |  17 ++
 drivers/net/cnxk/cnxk_ethdev.h|   3 +
 drivers/net/cnxk/cnxk_rte_flow.c  | 330 ++
 drivers/net/cnxk/cnxk_rte_flow.h  |  27 
 drivers/net/cnxk/meson.build  |   3 +
 12 files changed, 532 insertions(+)
 create mode 100644 drivers/net/cnxk/cn10k_rte_flow.c
 create mode 100644 drivers/net/cnxk/cn10k_rte_flow.h
 create mode 100644 drivers/net/cnxk/cn9k_rte_flow.c
 create mode 100644 drivers/net/cnxk/cn9k_rte_flow.h
 create mode 100644 drivers/net/cnxk/cnxk_rte_flow.c
 create mode 100644 drivers/net/cnxk/cnxk_rte_flow.h

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index e875e55..cc71b22 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -24,6 +24,7 @@ Features of the CNXK Ethdev PMD are:
 - Multiple queues for TX and RX
 - Receiver Side Scaling (RSS)
 - MAC filtering
+- Generic flow API
 - Inner and Outer Checksum offload
 - Port hardware statistics
 - Link state information
@@ -208,6 +209,12 @@ CRC stripping
 The OCTEON CN9K/CN10K SoC family NICs strip the CRC for every packet being 
received by
 the host interface irrespective of the offload configuration.
 
+RTE Flow GRE support
+
+
+- ``RTE_FLOW_ITEM_TYPE_GRE_KEY`` works only when checksum and routing
+  bits in the GRE header are equal to 0.
+
 Debugging Options
 -
 
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 192c15a..e2da5dd 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -39,3 +39,39 @@ Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
+
+[rte_flow items]
+any  = Y
+arp_eth_ipv4 = Y
+esp  = Y
+eth  = Y
+e_tag= Y
+geneve   = Y
+gre  = Y
+gre_key  = Y
+gtpc = Y
+gtpu = Y
+higig2   = Y
+icmp = Y
+ipv4 = Y
+ipv6 = Y
+ipv6_ext = Y
+mpls = Y
+nvgre= Y
+raw  = Y
+sctp = Y
+tcp  = Y
+udp  = Y
+vlan = Y
+vxlan= Y
+vxlan_gpe= Y
+
+[rte_flow actions]
+count= Y
+drop = Y
+flag = Y
+pf   = Y
+port_id  = Y
+queue= Y
+security = Y
+vf   = Y
diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index 5ff36bb..0396ff6 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -2,6 +2,7 @@
  * Copyright(C) 2021 Marvell.
  */
 #include "cn10k_ethdev.h"
+#include "cn10k_rte_flow.h"
 #include "cn10k_rx.h"
 #include "cn10k_tx.h"
 
@@ -308,6 +309,20 @@ nix_eth_dev_ops_override(void)
cnxk_eth_dev_ops.dev_ptypes_set = cn10k_nix_ptypes_set;
 }
 
+static void
+npc_flow_ops_override(void)
+{
+   static int init_once;
+
+   if (init_once)
+   return;
+   init_once = 1;
+
+   /* Update platform specific ops */
+   cnxk_flow_ops.create = cn10k_flow_create;
+   cnxk_flow_ops.destroy = cn10k_flow_destroy;
+}
+
 static int
 cn10k_nix_remove(struct rte_pci_device *pci_dev)
 {
@@ -332,6 +347,7 @@ cn10k_nix_probe(struct rte_pci_driver *pci_drv, struct 
rte_pci_device *pci_dev)
}
 
nix_eth_dev_ops_override();
+   npc_flow_ops_override();
 
/* Common probe */
rc = cnxk_nix_probe(pci_drv, pci_dev);
diff --git a/drivers/net/cnxk/cn10k_rte_flow.c 
b/drivers/net/cnxk/cn10k_rte_flow.c
new file mode 100644
index 000..65893cc
--- /dev/null
+++ b/drivers/net/cnxk/cn10k_rte_flow.c
@@ -0,0 +1,30 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2020 Marvell.
+ */
+#include 
+#include "cn10k_rte_flow.h"
+#include "cn10k_ethdev.h"
+
+struct rte_flow *
+cn10k_flow_create(struct rte_eth_dev *eth_dev, const struct rte_flow_attr 
*attr,
+ const struct rte_flow_item pattern[],
+ const struct rte_flow_action actions[],
+ struct rte_flow_error *error)
+{
+   struct roc_npc_flow *flow;
+
+   flow = cnxk_flow_create(eth_dev, attr, pattern, actions, error);
+   if (!flow)
+   return NULL;
+

[dpdk-dev] [PATCH v3 50/62] net/cnxk: add flow ops get operation

2021-06-18 Thread Nithin Dabilpuram
From: Satheesh Paul 

This patch adds flow ops get operation to enable rte_flow_ops.A

This patch also add support for flow dev dump API.
Every flow rule added will be dumped in the below format.

MCAM Index:1881
Interface :NIX-RX (0)
Priority  :1
NPC RX Action:0X404001
ActionOp:NIX_RX_ACTIONOP_UCAST (1)
PF_FUNC: 0X400
RQ Index:0X004
Match Id:
Flow Key Alg:0
NPC RX VTAG Action:0X008100
VTAG0:relptr:0
lid:0X1
type:0
Patterns:
NPC_PARSE_NIBBLE_CHAN:000
NPC_PARSE_NIBBLE_LA_LTYPE:LA_ETHER
NPC_PARSE_NIBBLE_LB_LTYPE:NONE
NPC_PARSE_NIBBLE_LC_LTYPE:LC_IP
NPC_PARSE_NIBBLE_LD_LTYPE:LD_TCP
NPC_PARSE_NIBBLE_LE_LTYPE:NONE
LA_ETHER, hdr offset:0, len:0X6, key offset:0X8,\
Data:0X4AE124FC7FFF, Mask:0X
LA_ETHER, hdr offset:0XC, len:0X2, key offset:0X4, Data:0XCA5A,\
Mask:0X
LC_IP, hdr offset:0XC, len:0X8, key offset:0X10,\
Data:0X0A010103, Mask:0X
LD_TCP, hdr offset:0, len:0X4, key offset:0X18, Data:0X0345,\
Mask:0X
MCAM Raw Data :
DW0 :CA5A01202000
DW0_Mask:0FF0F000
DW1 :4AE124FC7FFF
DW1_Mask:
DW2 :0A010103
DW2_Mask:
DW3 :0345
DW3_Mask:
DW4 :
DW4_Mask:
DW5 :
DW5_Mask:
DW6 :
DW6_Mask:

Signed-off-by: Satheesh Paul 
---
 drivers/common/cnxk/roc_npc.c  |  2 ++
 drivers/net/cnxk/cnxk_ethdev.c |  3 +++
 drivers/net/cnxk/cnxk_ethdev.h |  3 ++-
 drivers/net/cnxk/cnxk_ethdev_ops.c | 10 ++
 drivers/net/cnxk/cnxk_rte_flow.c   | 28 
 5 files changed, 45 insertions(+), 1 deletion(-)

diff --git a/drivers/common/cnxk/roc_npc.c b/drivers/common/cnxk/roc_npc.c
index 8a76823..aff4eef 100644
--- a/drivers/common/cnxk/roc_npc.c
+++ b/drivers/common/cnxk/roc_npc.c
@@ -1009,6 +1009,8 @@ roc_npc_flow_create(struct roc_npc *roc_npc, const struct 
roc_npc_attr *attr,
struct npc_flow_list *list;
int rc;
 
+   npc->channel = roc_npc->channel;
+
flow = plt_zmalloc(sizeof(*flow), 0);
if (flow == NULL) {
*errcode = NPC_ERR_NO_MEM;
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index d1a7f2e..a330875 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -807,6 +807,8 @@ cnxk_nix_configure(struct rte_eth_dev *eth_dev)
goto fail_configure;
}
 
+   dev->npc.channel = roc_nix_get_base_chan(nix);
+
nb_rxq = data->nb_rx_queues;
nb_txq = data->nb_tx_queues;
rc = -ENOMEM;
@@ -1211,6 +1213,7 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.rxq_info_get = cnxk_nix_rxq_info_get,
.txq_info_get = cnxk_nix_txq_info_get,
.tx_done_cleanup = cnxk_nix_tx_done_cleanup,
+   .flow_ops_get = cnxk_nix_flow_ops_get,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index b7869b4..07c87ea 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -273,7 +273,8 @@ int cnxk_nix_rx_queue_intr_disable(struct rte_eth_dev 
*eth_dev,
   uint16_t rx_queue_id);
 int cnxk_nix_pool_ops_supported(struct rte_eth_dev *eth_dev, const char *pool);
 int cnxk_nix_tx_done_cleanup(void *txq, uint32_t free_cnt);
-
+int cnxk_nix_flow_ops_get(struct rte_eth_dev *eth_dev,
+ const struct rte_flow_ops **ops);
 int cnxk_nix_configure(struct rte_eth_dev *eth_dev);
 int cnxk_nix_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t qid,
uint16_t nb_desc, uint16_t fp_tx_q_sz,
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index ce94eb6..0b62911 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -300,6 +300,16 @@ cnxk_nix_flow_ctrl_set(struct rte_eth_dev *eth_dev,
 }
 
 int
+cnxk_nix_flow_ops_get(struct rte_eth_dev *eth_dev,
+ const struct rte_flow_ops **ops)
+{
+   RTE_SET_USED(eth_dev);
+
+   *ops = &cnxk_flow_ops;
+   return 0;
+}
+
+int
 cnxk_nix_mac_addr_set(struct rte_eth_dev *eth_dev, struct rte_ether_addr *addr)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
diff --git a/drivers/net/cnxk/cnxk_rte_flow.c b/drivers/net/cnxk/cnxk_rte_flow.c
index 1695d4f..2dee19e 100644
--- a/drivers/net/cnxk/cnxk_rte_flow.c
+++ b/drivers/net/cnxk/cnxk_rte_flow.c
@@ -322,9 +322,37 @@ cnxk_flow_isolate(struct rte_eth_dev *eth_dev __rte_unused,
return -rte_errno;
 }
 
+static int
+cnxk_flow_dev_dump(struct rte_et

[dpdk-dev] [PATCH v3 51/62] net/cnxk: add ethdev firmware version get

2021-06-18 Thread Nithin Dabilpuram
From: Satha Rao 

Add callback to get ethdev firmware version.

Signed-off-by: Satha Rao 
---
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 doc/guides/nics/features/cnxk_vf.ini  |  1 +
 drivers/net/cnxk/cnxk_ethdev.c|  1 +
 drivers/net/cnxk/cnxk_ethdev.h|  2 ++
 drivers/net/cnxk/cnxk_ethdev_ops.c| 19 +++
 6 files changed, 25 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index e2da5dd..d43b1c4 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -35,6 +35,7 @@ Packet type parsing  = Y
 Basic stats  = Y
 Stats per queue  = Y
 Extended stats   = Y
+FW version   = Y
 Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index e990480..6b9ce2d 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -33,6 +33,7 @@ Packet type parsing  = Y
 Basic stats  = Y
 Stats per queue  = Y
 Extended stats   = Y
+FW version   = Y
 Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 3a4417c..5629d07 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -30,6 +30,7 @@ Packet type parsing  = Y
 Basic stats  = Y
 Stats per queue  = Y
 Extended stats   = Y
+FW version   = Y
 Module EEPROM dump   = Y
 Linux= Y
 ARMv8= Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index a330875..284c8f4 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1210,6 +1210,7 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.xstats_reset = cnxk_nix_xstats_reset,
.xstats_get_by_id = cnxk_nix_xstats_get_by_id,
.xstats_get_names_by_id = cnxk_nix_xstats_get_names_by_id,
+   .fw_version_get = cnxk_nix_fw_version_get,
.rxq_info_get = cnxk_nix_rxq_info_get,
.txq_info_get = cnxk_nix_txq_info_get,
.tx_done_cleanup = cnxk_nix_tx_done_cleanup,
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 07c87ea..a25be8a 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -313,6 +313,8 @@ int cnxk_nix_xstats_get_names_by_id(struct rte_eth_dev 
*eth_dev,
 int cnxk_nix_xstats_get_by_id(struct rte_eth_dev *eth_dev, const uint64_t *ids,
  uint64_t *values, unsigned int n);
 int cnxk_nix_xstats_reset(struct rte_eth_dev *eth_dev);
+int cnxk_nix_fw_version_get(struct rte_eth_dev *eth_dev, char *fw_version,
+   size_t fw_size);
 void cnxk_nix_rxq_info_get(struct rte_eth_dev *eth_dev, uint16_t qid,
   struct rte_eth_rxq_info *qinfo);
 void cnxk_nix_txq_info_get(struct rte_eth_dev *eth_dev, uint16_t qid,
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 0b62911..4e863e2 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -639,6 +639,25 @@ cnxk_nix_pool_ops_supported(struct rte_eth_dev *eth_dev, 
const char *pool)
return -ENOTSUP;
 }
 
+int
+cnxk_nix_fw_version_get(struct rte_eth_dev *eth_dev, char *fw_version,
+   size_t fw_size)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   const char *str = roc_npc_profile_name_get(&dev->npc);
+   uint32_t size = strlen(str) + 1;
+
+   if (fw_size > size)
+   fw_size = size;
+
+   rte_strlcpy(fw_version, str, fw_size);
+
+   if (fw_size < size)
+   return size;
+
+   return 0;
+}
+
 void
 cnxk_nix_rxq_info_get(struct rte_eth_dev *eth_dev, uint16_t qid,
  struct rte_eth_rxq_info *qinfo)
-- 
2.8.4



[dpdk-dev] [PATCH v3 52/62] net/cnxk: add get register operation

2021-06-18 Thread Nithin Dabilpuram
From: Satha Rao 

With this patch implemented api to dump platform registers for
debug purposes.

Signed-off-by: Satha Rao 
---
 doc/guides/nics/cnxk.rst  |  1 +
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 doc/guides/nics/features/cnxk_vf.ini  |  1 +
 drivers/net/cnxk/cnxk_ethdev.c|  1 +
 drivers/net/cnxk/cnxk_ethdev.h|  4 
 drivers/net/cnxk/cnxk_ethdev_ops.c| 25 +
 7 files changed, 34 insertions(+)

diff --git a/doc/guides/nics/cnxk.rst b/doc/guides/nics/cnxk.rst
index cc71b22..e561321 100644
--- a/doc/guides/nics/cnxk.rst
+++ b/doc/guides/nics/cnxk.rst
@@ -32,6 +32,7 @@ Features of the CNXK Ethdev PMD are:
 - MTU update
 - Scatter-Gather IO support
 - Vector Poll mode driver
+- Debug utilities - Context dump and error interrupt support
 - Support Rx interrupt
 
 Prerequisites
diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index d43b1c4..246caf9 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -37,6 +37,7 @@ Stats per queue  = Y
 Extended stats   = Y
 FW version   = Y
 Module EEPROM dump   = Y
+Registers dump   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 6b9ce2d..65ee8ba 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -35,6 +35,7 @@ Stats per queue  = Y
 Extended stats   = Y
 FW version   = Y
 Module EEPROM dump   = Y
+Registers dump   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 5629d07..00bde9b 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -32,6 +32,7 @@ Stats per queue  = Y
 Extended stats   = Y
 FW version   = Y
 Module EEPROM dump   = Y
+Registers dump   = Y
 Linux= Y
 ARMv8= Y
 Usage doc= Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 284c8f4..cb0dd45 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1215,6 +1215,7 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.txq_info_get = cnxk_nix_txq_info_get,
.tx_done_cleanup = cnxk_nix_tx_done_cleanup,
.flow_ops_get = cnxk_nix_flow_ops_get,
+   .get_reg = cnxk_nix_dev_get_reg,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index a25be8a..d51a6d7 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -328,6 +328,10 @@ void *cnxk_nix_fastpath_lookup_mem_get(void);
 int cnxk_ethdev_parse_devargs(struct rte_devargs *devargs,
  struct cnxk_eth_dev *dev);
 
+/* Debug */
+int cnxk_nix_dev_get_reg(struct rte_eth_dev *eth_dev,
+struct rte_dev_reg_info *regs);
+
 /* Inlines */
 static __rte_always_inline uint64_t
 cnxk_pktmbuf_detach(struct rte_mbuf *m)
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 4e863e2..27d5dc0 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -697,3 +697,28 @@ cnxk_nix_tx_done_cleanup(void *txq, uint32_t free_cnt)
 
return 0;
 }
+
+int
+cnxk_nix_dev_get_reg(struct rte_eth_dev *eth_dev, struct rte_dev_reg_info 
*regs)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   uint64_t *data = regs->data;
+   int rc = -ENOTSUP;
+
+   if (data == NULL) {
+   rc = roc_nix_lf_get_reg_count(nix);
+   if (rc > 0) {
+   regs->length = rc;
+   regs->width = 8;
+   rc = 0;
+   }
+   return rc;
+   }
+
+   if (!regs->length ||
+   regs->length == (uint32_t)roc_nix_lf_get_reg_count(nix))
+   return roc_nix_lf_reg_dump(nix, data);
+
+   return rc;
+}
-- 
2.8.4



[dpdk-dev] [PATCH v3 53/62] net/cnxk: support for RSS in rte flow

2021-06-18 Thread Nithin Dabilpuram
From: Satheesh Paul 

Added support for RSS action in rte flow code based on ROC.

Signed-off-by: Satheesh Paul 
---
 doc/guides/nics/features/cnxk.ini |  1 +
 drivers/net/cnxk/cnxk_rte_flow.c  | 74 ++-
 2 files changed, 67 insertions(+), 8 deletions(-)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 246caf9..eba4107 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -75,5 +75,6 @@ flag = Y
 pf   = Y
 port_id  = Y
 queue= Y
+rss  = Y
 security = Y
 vf   = Y
diff --git a/drivers/net/cnxk/cnxk_rte_flow.c b/drivers/net/cnxk/cnxk_rte_flow.c
index 2dee19e..c61618a 100644
--- a/drivers/net/cnxk/cnxk_rte_flow.c
+++ b/drivers/net/cnxk/cnxk_rte_flow.c
@@ -56,14 +56,64 @@ const struct cnxk_rte_flow_term_info term[] = {
 };
 
 static int
-cnxk_map_actions(struct rte_eth_dev *eth_dev,
+npc_rss_action_validate(struct rte_eth_dev *eth_dev,
+   const struct rte_flow_attr *attr,
+   const struct rte_flow_action *act)
+{
+   const struct rte_flow_action_rss *rss;
+
+   rss = (const struct rte_flow_action_rss *)act->conf;
+
+   if (attr->egress) {
+   plt_err("No support of RSS in egress");
+   return -EINVAL;
+   }
+
+   if (eth_dev->data->dev_conf.rxmode.mq_mode != ETH_MQ_RX_RSS) {
+   plt_err("multi-queue mode is disabled");
+   return -ENOTSUP;
+   }
+
+   if (!rss || !rss->queue_num) {
+   plt_err("no valid queues");
+   return -EINVAL;
+   }
+
+   if (rss->func != RTE_ETH_HASH_FUNCTION_DEFAULT) {
+   plt_err("non-default RSS hash functions are not supported");
+   return -ENOTSUP;
+   }
+
+   if (rss->key_len && rss->key_len > ROC_NIX_RSS_KEY_LEN) {
+   plt_err("RSS hash key too large");
+   return -ENOTSUP;
+   }
+
+   return 0;
+}
+
+static void
+npc_rss_flowkey_get(struct cnxk_eth_dev *eth_dev,
+   const struct roc_npc_action *rss_action,
+   uint32_t *flowkey_cfg)
+{
+   const struct roc_npc_action_rss *rss;
+
+   rss = (const struct roc_npc_action_rss *)rss_action->conf;
+
+   *flowkey_cfg = cnxk_rss_ethdev_to_nix(eth_dev, rss->types, rss->level);
+}
+
+static int
+cnxk_map_actions(struct rte_eth_dev *eth_dev, const struct rte_flow_attr *attr,
 const struct rte_flow_action actions[],
-struct roc_npc_action in_actions[])
+struct roc_npc_action in_actions[], uint32_t *flowkey_cfg)
 {
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
const struct rte_flow_action_count *act_count;
const struct rte_flow_action_queue *act_q;
+   int i = 0, rc = 0;
int rq;
-   int i = 0;
 
for (; actions->type != RTE_FLOW_ACTION_TYPE_END; actions++) {
switch (actions->type) {
@@ -117,7 +167,12 @@ cnxk_map_actions(struct rte_eth_dev *eth_dev,
break;
 
case RTE_FLOW_ACTION_TYPE_RSS:
+   rc = npc_rss_action_validate(eth_dev, attr, actions);
+   if (rc)
+   goto err_exit;
in_actions[i].type = ROC_NPC_ACTION_TYPE_RSS;
+   in_actions[i].conf = actions->conf;
+   npc_rss_flowkey_get(dev, &in_actions[i], flowkey_cfg);
break;
 
case RTE_FLOW_ACTION_TYPE_SECURITY:
@@ -144,7 +199,7 @@ cnxk_map_flow_data(struct rte_eth_dev *eth_dev,
   const struct rte_flow_action actions[],
   struct roc_npc_attr *in_attr,
   struct roc_npc_item_info in_pattern[],
-  struct roc_npc_action in_actions[])
+  struct roc_npc_action in_actions[], uint32_t *flowkey_cfg)
 {
int i = 0;
 
@@ -163,7 +218,8 @@ cnxk_map_flow_data(struct rte_eth_dev *eth_dev,
}
in_pattern[i].type = ROC_NPC_ITEM_TYPE_END;
 
-   return cnxk_map_actions(eth_dev, actions, in_actions);
+   return cnxk_map_actions(eth_dev, attr, actions, in_actions,
+   flowkey_cfg);
 }
 
 static int
@@ -179,12 +235,13 @@ cnxk_flow_validate(struct rte_eth_dev *eth_dev,
struct roc_npc *npc = &dev->npc;
struct roc_npc_attr in_attr;
struct roc_npc_flow flow;
+   uint32_t flowkey_cfg = 0;
int rc;
 
memset(&flow, 0, sizeof(flow));
 
rc = cnxk_map_flow_data(eth_dev, attr, pattern, actions, &in_attr,
-   in_pattern, in_actions);
+   in_pattern, in_actions, &flowkey_cfg);
if (rc) {
rte_flow_error_set(error, 0, RTE_FLOW_ERROR_TYPE_ACTION_NUM,
 

[dpdk-dev] [PATCH v3 54/62] net/cnxk: register callback to get PTP status

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Once PTP status is changed at H/W i.e. enable/disable then
it is propagated to user via registered callback.

So corresponding callback is registered to get PTP status.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cn10k_ethdev.c| 87 --
 drivers/net/cnxk/cn10k_rx.h|  1 +
 drivers/net/cnxk/cn9k_ethdev.c | 73 
 drivers/net/cnxk/cn9k_rx.h |  1 +
 drivers/net/cnxk/cnxk_ethdev.c |  2 +-
 drivers/net/cnxk/cnxk_ethdev.h |  5 +++
 drivers/net/cnxk/cnxk_ethdev_ops.c |  2 +
 7 files changed, 166 insertions(+), 5 deletions(-)

diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index 0396ff6..bddb7fb 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -268,6 +268,76 @@ cn10k_nix_configure(struct rte_eth_dev *eth_dev)
return 0;
 }
 
+/* Function to enable ptp config for VFs */
+static void
+nix_ptp_enable_vf(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   if (nix_recalc_mtu(eth_dev))
+   plt_err("Failed to set MTU size for ptp");
+
+   dev->scalar_ena = true;
+   dev->rx_offload_flags |= NIX_RX_OFFLOAD_TSTAMP_F;
+
+   /* Setting up the function pointers as per new offload flags */
+   cn10k_eth_set_rx_function(eth_dev);
+   cn10k_eth_set_tx_function(eth_dev);
+}
+
+static uint16_t
+nix_ptp_vf_burst(void *queue, struct rte_mbuf **mbufs, uint16_t pkts)
+{
+   struct cn10k_eth_rxq *rxq = queue;
+   struct cnxk_eth_rxq_sp *rxq_sp;
+   struct rte_eth_dev *eth_dev;
+
+   RTE_SET_USED(mbufs);
+   RTE_SET_USED(pkts);
+
+   rxq_sp = cnxk_eth_rxq_to_sp(rxq);
+   eth_dev = rxq_sp->dev->eth_dev;
+   nix_ptp_enable_vf(eth_dev);
+
+   return 0;
+}
+
+static int
+cn10k_nix_ptp_info_update_cb(struct roc_nix *nix, bool ptp_en)
+{
+   struct cnxk_eth_dev *dev = (struct cnxk_eth_dev *)nix;
+   struct rte_eth_dev *eth_dev;
+   struct cn10k_eth_rxq *rxq;
+   int i;
+
+   if (!dev)
+   return -EINVAL;
+
+   eth_dev = dev->eth_dev;
+   if (!eth_dev)
+   return -EINVAL;
+
+   dev->ptp_en = ptp_en;
+
+   for (i = 0; i < eth_dev->data->nb_rx_queues; i++) {
+   rxq = eth_dev->data->rx_queues[i];
+   rxq->mbuf_initializer = cnxk_nix_rxq_mbuf_setup(dev);
+   }
+
+   if (roc_nix_is_vf_or_sdp(nix) && !(roc_nix_is_sdp(nix)) &&
+   !(roc_nix_is_lbk(nix))) {
+   /* In case of VF, setting of MTU cannot be done directly in this
+* function as this is running as part of MBOX request(PF->VF)
+* and MTU setting also requires MBOX message to be
+* sent(VF->PF)
+*/
+   eth_dev->rx_pkt_burst = nix_ptp_vf_burst;
+   rte_mb();
+   }
+
+   return 0;
+}
+
 static int
 cn10k_nix_dev_start(struct rte_eth_dev *eth_dev)
 {
@@ -333,6 +403,7 @@ static int
 cn10k_nix_probe(struct rte_pci_driver *pci_drv, struct rte_pci_device *pci_dev)
 {
struct rte_eth_dev *eth_dev;
+   struct cnxk_eth_dev *dev;
int rc;
 
if (RTE_CACHE_LINE_SIZE != 64) {
@@ -354,15 +425,23 @@ cn10k_nix_probe(struct rte_pci_driver *pci_drv, struct 
rte_pci_device *pci_dev)
if (rc)
return rc;
 
+   /* Find eth dev allocated */
+   eth_dev = rte_eth_dev_allocated(pci_dev->device.name);
+   if (!eth_dev)
+   return -ENOENT;
+
if (rte_eal_process_type() != RTE_PROC_PRIMARY) {
-   eth_dev = rte_eth_dev_allocated(pci_dev->device.name);
-   if (!eth_dev)
-   return -ENOENT;
-
/* Setup callbacks for secondary process */
cn10k_eth_set_tx_function(eth_dev);
cn10k_eth_set_rx_function(eth_dev);
+   return 0;
}
+
+   dev = cnxk_eth_pmd_priv(eth_dev);
+
+   /* Register up msg callbacks for PTP information */
+   roc_nix_ptp_info_cb_register(&dev->nix, cn10k_nix_ptp_info_update_cb);
+
return 0;
 }
 
diff --git a/drivers/net/cnxk/cn10k_rx.h b/drivers/net/cnxk/cn10k_rx.h
index 7bb9dd8..29ee0ac 100644
--- a/drivers/net/cnxk/cn10k_rx.h
+++ b/drivers/net/cnxk/cn10k_rx.h
@@ -12,6 +12,7 @@
 #define NIX_RX_OFFLOAD_PTYPE_F  BIT(1)
 #define NIX_RX_OFFLOAD_CHECKSUM_FBIT(2)
 #define NIX_RX_OFFLOAD_MARK_UPDATE_F BIT(3)
+#define NIX_RX_OFFLOAD_TSTAMP_F BIT(4)
 
 /* Flags to control cqe_to_mbuf conversion function.
  * Defining it from backwards to denote its been
diff --git a/drivers/net/cnxk/cn9k_ethdev.c b/drivers/net/cnxk/cn9k_ethdev.c
index 5cc0db9..fc3054d 100644
--- a/drivers/net/cnxk/cn9k_ethdev.c
+++ b/drivers/net/cnxk/cn9k_ethdev.c
@@ -277,6 +277,76 @@ cn9k_nix_configure(struct rte_eth_dev *eth_dev)
return 0;
 }
 
+/* Function to enable ptp config for VFs *

[dpdk-dev] [PATCH v3 55/62] net/cnxk: add base PTP timesync support

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Base PTP timesync support is added for cn9k and cn10k platforms.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cn10k_ethdev.c  |  31 +
 drivers/net/cnxk/cn10k_ethdev.h  |   1 +
 drivers/net/cnxk/cn10k_rx.c  |  32 ++---
 drivers/net/cnxk/cn10k_rx.h  |  61 +++---
 drivers/net/cnxk/cn10k_rx_mseg.c |   2 +-
 drivers/net/cnxk/cn10k_rx_vec.c  |   5 +-
 drivers/net/cnxk/cn10k_tx.c  |  28 +++--
 drivers/net/cnxk/cn10k_tx.h  | 193 +++--
 drivers/net/cnxk/cn10k_tx_mseg.c |   2 +-
 drivers/net/cnxk/cn10k_tx_vec.c  |   3 +-
 drivers/net/cnxk/cn9k_ethdev.c   |  34 +-
 drivers/net/cnxk/cn9k_ethdev.h   |   1 +
 drivers/net/cnxk/cn9k_rx.c   |  32 ++---
 drivers/net/cnxk/cn9k_rx.h   |  61 +++---
 drivers/net/cnxk/cn9k_rx_mseg.c  |   2 +-
 drivers/net/cnxk/cn9k_rx_vec.c   |   5 +-
 drivers/net/cnxk/cn9k_tx.c   |  28 +++--
 drivers/net/cnxk/cn9k_tx.h   | 253 ---
 drivers/net/cnxk/cn9k_tx_mseg.c  |   2 +-
 drivers/net/cnxk/cn9k_tx_vec.c   |   3 +-
 drivers/net/cnxk/cnxk_ethdev.c   |  36 +-
 drivers/net/cnxk/cnxk_ethdev.h   |  64 +-
 drivers/net/cnxk/cnxk_ptp.c  | 169 ++
 drivers/net/cnxk/meson.build |   1 +
 24 files changed, 834 insertions(+), 215 deletions(-)
 create mode 100644 drivers/net/cnxk/cnxk_ptp.c

diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index bddb7fb..5e0de13 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -30,6 +30,9 @@ nix_rx_offload_flags(struct rte_eth_dev *eth_dev)
if (dev->rx_offloads & DEV_RX_OFFLOAD_SCATTER)
flags |= NIX_RX_MULTI_SEG_F;
 
+   if ((dev->rx_offloads & DEV_RX_OFFLOAD_TIMESTAMP))
+   flags |= NIX_RX_OFFLOAD_TSTAMP_F;
+
if (!dev->ptype_disable)
flags |= NIX_RX_OFFLOAD_PTYPE_F;
 
@@ -95,6 +98,9 @@ nix_tx_offload_flags(struct rte_eth_dev *eth_dev)
flags |= (NIX_TX_OFFLOAD_TSO_F | NIX_TX_OFFLOAD_OL3_OL4_CSUM_F |
  NIX_TX_OFFLOAD_L3_L4_CSUM_F);
 
+   if ((dev->rx_offloads & DEV_RX_OFFLOAD_TIMESTAMP))
+   flags |= NIX_TX_OFFLOAD_TSTAMP_F;
+
return flags;
 }
 
@@ -121,6 +127,7 @@ nix_form_default_desc(struct cnxk_eth_dev *dev, struct 
cn10k_eth_txq *txq,
 {
struct nix_send_ext_s *send_hdr_ext;
union nix_send_hdr_w0_u send_hdr_w0;
+   struct nix_send_mem_s *send_mem;
union nix_send_sg_s sg_w0;
 
RTE_SET_USED(dev);
@@ -136,6 +143,22 @@ nix_form_default_desc(struct cnxk_eth_dev *dev, struct 
cn10k_eth_txq *txq,
 
send_hdr_ext = (struct nix_send_ext_s *)&txq->cmd[0];
send_hdr_ext->w0.subdc = NIX_SUBDC_EXT;
+   if (dev->tx_offload_flags & NIX_TX_OFFLOAD_TSTAMP_F) {
+   /* Default: one seg packet would have:
+* 2(HDR) + 2(EXT) + 1(SG) + 1(IOVA) + 2(MEM)
+* => 8/2 - 1 = 3
+*/
+   send_hdr_w0.sizem1 = 3;
+   send_hdr_ext->w0.tstmp = 1;
+
+   /* To calculate the offset for send_mem,
+* send_hdr->w0.sizem1 * 2
+*/
+   send_mem = (struct nix_send_mem_s *)(txq->cmd + 2);
+   send_mem->w0.subdc = NIX_SUBDC_MEM;
+   send_mem->w0.alg = NIX_SENDMEMALG_SETTSTMP;
+   send_mem->addr = dev->tstamp.tx_tstamp_iova;
+   }
} else {
/* 2(HDR) + 1(SG) + 1(IOVA) = 4/2 - 1 = 1 */
send_hdr_w0.sizem1 = 1;
@@ -221,6 +244,7 @@ cn10k_nix_rx_queue_setup(struct rte_eth_dev *eth_dev, 
uint16_t qid,
rxq->wdata = cq->wdata;
rxq->head = cq->head;
rxq->qmask = cq->qmask;
+   rxq->tstamp = &dev->tstamp;
 
/* Data offset from data to start of mbuf is first_skip */
rxq->data_off = rq->first_skip;
@@ -342,6 +366,7 @@ static int
 cn10k_nix_dev_start(struct rte_eth_dev *eth_dev)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
int rc;
 
/* Common eth dev start */
@@ -349,6 +374,12 @@ cn10k_nix_dev_start(struct rte_eth_dev *eth_dev)
if (rc)
return rc;
 
+   /* Update VF about data off shifted by 8 bytes if PTP already
+* enabled in PF owning this VF
+*/
+   if (dev->ptp_en && (!roc_nix_is_pf(nix) && (!roc_nix_is_sdp(nix
+   nix_ptp_enable_vf(eth_dev);
+
/* Setting up the rx[tx]_offload_flags due to change
 * in rx[tx]_offloads.
 */
diff --git a/drivers/net/cnxk/cn10k_ethdev.h b/drivers/net/cnxk/cn10k_ethdev.h
index d39ca31..8b6e0f2 100644
--- a/drivers/net/cnxk/cn10k_ethdev.h
+++ b/drivers/net/cnxk/cn10k_ethdev.h
@@ -31,6 +31,7 @@ struct cn10k_eth_rxq {
  

[dpdk-dev] [PATCH v3 56/62] net/cnxk: add timesync enable/disable operations

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch implements timesync enable/disable operations for
cn9k and cn10k platforms.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cn10k_ethdev.c | 50 +
 drivers/net/cnxk/cn9k_ethdev.c  | 50 +
 2 files changed, 100 insertions(+)

diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index 5e0de13..b079edb 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -363,6 +363,54 @@ cn10k_nix_ptp_info_update_cb(struct roc_nix *nix, bool 
ptp_en)
 }
 
 static int
+cn10k_nix_timesync_enable(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   int i, rc;
+
+   rc = cnxk_nix_timesync_enable(eth_dev);
+   if (rc)
+   return rc;
+
+   dev->rx_offload_flags |= NIX_RX_OFFLOAD_TSTAMP_F;
+   dev->tx_offload_flags |= NIX_TX_OFFLOAD_TSTAMP_F;
+
+   for (i = 0; i < eth_dev->data->nb_tx_queues; i++)
+   nix_form_default_desc(dev, eth_dev->data->tx_queues[i], i);
+
+   /* Setting up the rx[tx]_offload_flags due to change
+* in rx[tx]_offloads.
+*/
+   cn10k_eth_set_rx_function(eth_dev);
+   cn10k_eth_set_tx_function(eth_dev);
+   return 0;
+}
+
+static int
+cn10k_nix_timesync_disable(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   int i, rc;
+
+   rc = cnxk_nix_timesync_disable(eth_dev);
+   if (rc)
+   return rc;
+
+   dev->rx_offload_flags &= ~NIX_RX_OFFLOAD_TSTAMP_F;
+   dev->tx_offload_flags &= ~NIX_TX_OFFLOAD_TSTAMP_F;
+
+   for (i = 0; i < eth_dev->data->nb_tx_queues; i++)
+   nix_form_default_desc(dev, eth_dev->data->tx_queues[i], i);
+
+   /* Setting up the rx[tx]_offload_flags due to change
+* in rx[tx]_offloads.
+*/
+   cn10k_eth_set_rx_function(eth_dev);
+   cn10k_eth_set_tx_function(eth_dev);
+   return 0;
+}
+
+static int
 cn10k_nix_dev_start(struct rte_eth_dev *eth_dev)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
@@ -408,6 +456,8 @@ nix_eth_dev_ops_override(void)
cnxk_eth_dev_ops.tx_queue_stop = cn10k_nix_tx_queue_stop;
cnxk_eth_dev_ops.dev_start = cn10k_nix_dev_start;
cnxk_eth_dev_ops.dev_ptypes_set = cn10k_nix_ptypes_set;
+   cnxk_eth_dev_ops.timesync_enable = cn10k_nix_timesync_enable;
+   cnxk_eth_dev_ops.timesync_disable = cn10k_nix_timesync_disable;
 }
 
 static void
diff --git a/drivers/net/cnxk/cn9k_ethdev.c b/drivers/net/cnxk/cn9k_ethdev.c
index 924b747..107a540 100644
--- a/drivers/net/cnxk/cn9k_ethdev.c
+++ b/drivers/net/cnxk/cn9k_ethdev.c
@@ -371,6 +371,54 @@ cn9k_nix_ptp_info_update_cb(struct roc_nix *nix, bool 
ptp_en)
 }
 
 static int
+cn9k_nix_timesync_enable(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   int i, rc;
+
+   rc = cnxk_nix_timesync_enable(eth_dev);
+   if (rc)
+   return rc;
+
+   dev->rx_offload_flags |= NIX_RX_OFFLOAD_TSTAMP_F;
+   dev->tx_offload_flags |= NIX_TX_OFFLOAD_TSTAMP_F;
+
+   for (i = 0; i < eth_dev->data->nb_tx_queues; i++)
+   nix_form_default_desc(dev, eth_dev->data->tx_queues[i], i);
+
+   /* Setting up the rx[tx]_offload_flags due to change
+* in rx[tx]_offloads.
+*/
+   cn9k_eth_set_rx_function(eth_dev);
+   cn9k_eth_set_tx_function(eth_dev);
+   return 0;
+}
+
+static int
+cn9k_nix_timesync_disable(struct rte_eth_dev *eth_dev)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   int i, rc;
+
+   rc = cnxk_nix_timesync_disable(eth_dev);
+   if (rc)
+   return rc;
+
+   dev->rx_offload_flags &= ~NIX_RX_OFFLOAD_TSTAMP_F;
+   dev->tx_offload_flags &= ~NIX_TX_OFFLOAD_TSTAMP_F;
+
+   for (i = 0; i < eth_dev->data->nb_tx_queues; i++)
+   nix_form_default_desc(dev, eth_dev->data->tx_queues[i], i);
+
+   /* Setting up the rx[tx]_offload_flags due to change
+* in rx[tx]_offloads.
+*/
+   cn9k_eth_set_rx_function(eth_dev);
+   cn9k_eth_set_tx_function(eth_dev);
+   return 0;
+}
+
+static int
 cn9k_nix_dev_start(struct rte_eth_dev *eth_dev)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
@@ -416,6 +464,8 @@ nix_eth_dev_ops_override(void)
cnxk_eth_dev_ops.tx_queue_stop = cn9k_nix_tx_queue_stop;
cnxk_eth_dev_ops.dev_start = cn9k_nix_dev_start;
cnxk_eth_dev_ops.dev_ptypes_set = cn9k_nix_ptypes_set;
+   cnxk_eth_dev_ops.timesync_enable = cn9k_nix_timesync_enable;
+   cnxk_eth_dev_ops.timesync_disable = cn9k_nix_timesync_disable;
 }
 
 static void
-- 
2.8.4



[dpdk-dev] [PATCH v3 57/62] net/cnxk: add Rx/Tx timestamp read operations

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch implements Rx/Tx timestamp read operations for cn9k
and cn10k platforms.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cnxk_ethdev.c |  2 ++
 drivers/net/cnxk/cnxk_ethdev.h |  5 +
 drivers/net/cnxk/cnxk_ptp.c| 38 ++
 3 files changed, 45 insertions(+)

diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 71d567a..0bef942 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1248,6 +1248,8 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.tx_done_cleanup = cnxk_nix_tx_done_cleanup,
.flow_ops_get = cnxk_nix_flow_ops_get,
.get_reg = cnxk_nix_dev_get_reg,
+   .timesync_read_rx_timestamp = cnxk_nix_timesync_read_rx_timestamp,
+   .timesync_read_tx_timestamp = cnxk_nix_timesync_read_tx_timestamp,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index de6d533..76df84a 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -312,6 +312,11 @@ int cnxk_nix_tx_queue_stop(struct rte_eth_dev *eth_dev, 
uint16_t qid);
 int cnxk_nix_dev_start(struct rte_eth_dev *eth_dev);
 int cnxk_nix_timesync_enable(struct rte_eth_dev *eth_dev);
 int cnxk_nix_timesync_disable(struct rte_eth_dev *eth_dev);
+int cnxk_nix_timesync_read_rx_timestamp(struct rte_eth_dev *eth_dev,
+   struct timespec *timestamp,
+   uint32_t flags);
+int cnxk_nix_timesync_read_tx_timestamp(struct rte_eth_dev *eth_dev,
+   struct timespec *timestamp);
 int cnxk_nix_tsc_convert(struct cnxk_eth_dev *dev);
 
 uint64_t cnxk_nix_rxq_mbuf_setup(struct cnxk_eth_dev *dev);
diff --git a/drivers/net/cnxk/cnxk_ptp.c b/drivers/net/cnxk/cnxk_ptp.c
index fc317965..7b00f87 100644
--- a/drivers/net/cnxk/cnxk_ptp.c
+++ b/drivers/net/cnxk/cnxk_ptp.c
@@ -56,6 +56,44 @@ cnxk_nix_tsc_convert(struct cnxk_eth_dev *dev)
 }
 
 int
+cnxk_nix_timesync_read_rx_timestamp(struct rte_eth_dev *eth_dev,
+   struct timespec *timestamp, uint32_t flags)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct cnxk_timesync_info *tstamp = &dev->tstamp;
+   uint64_t ns;
+
+   PLT_SET_USED(flags);
+
+   if (!tstamp->rx_ready)
+   return -EINVAL;
+
+   ns = rte_timecounter_update(&dev->rx_tstamp_tc, tstamp->rx_tstamp);
+   *timestamp = rte_ns_to_timespec(ns);
+   tstamp->rx_ready = 0;
+   return 0;
+}
+
+int
+cnxk_nix_timesync_read_tx_timestamp(struct rte_eth_dev *eth_dev,
+   struct timespec *timestamp)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct cnxk_timesync_info *tstamp = &dev->tstamp;
+   uint64_t ns;
+
+   if (*tstamp->tx_tstamp == 0)
+   return -EINVAL;
+
+   ns = rte_timecounter_update(&dev->tx_tstamp_tc, *tstamp->tx_tstamp);
+   *timestamp = rte_ns_to_timespec(ns);
+   *tstamp->tx_tstamp = 0;
+   rte_wmb();
+
+   return 0;
+}
+
+int
 cnxk_nix_timesync_enable(struct rte_eth_dev *eth_dev)
 {
struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
-- 
2.8.4



[dpdk-dev] [PATCH v3 58/62] net/cnxk: add time read/write/adjust operations

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch implements read/write/adjust time operations for
cn9k and cn10k platforms.

Signed-off-by: Sunil Kumar Kori 
---
 drivers/net/cnxk/cnxk_ethdev.c |  3 ++
 drivers/net/cnxk/cnxk_ethdev.h |  5 
 drivers/net/cnxk/cnxk_ptp.c| 63 ++
 3 files changed, 71 insertions(+)

diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index 0bef942..cac8ddf 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1250,6 +1250,9 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.get_reg = cnxk_nix_dev_get_reg,
.timesync_read_rx_timestamp = cnxk_nix_timesync_read_rx_timestamp,
.timesync_read_tx_timestamp = cnxk_nix_timesync_read_tx_timestamp,
+   .timesync_read_time = cnxk_nix_timesync_read_time,
+   .timesync_write_time = cnxk_nix_timesync_write_time,
+   .timesync_adjust_time = cnxk_nix_timesync_adjust_time,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 76df84a..4214365 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -317,6 +317,11 @@ int cnxk_nix_timesync_read_rx_timestamp(struct rte_eth_dev 
*eth_dev,
uint32_t flags);
 int cnxk_nix_timesync_read_tx_timestamp(struct rte_eth_dev *eth_dev,
struct timespec *timestamp);
+int cnxk_nix_timesync_read_time(struct rte_eth_dev *eth_dev,
+   struct timespec *ts);
+int cnxk_nix_timesync_write_time(struct rte_eth_dev *eth_dev,
+const struct timespec *ts);
+int cnxk_nix_timesync_adjust_time(struct rte_eth_dev *eth_dev, int64_t delta);
 int cnxk_nix_tsc_convert(struct cnxk_eth_dev *dev);
 
 uint64_t cnxk_nix_rxq_mbuf_setup(struct cnxk_eth_dev *dev);
diff --git a/drivers/net/cnxk/cnxk_ptp.c b/drivers/net/cnxk/cnxk_ptp.c
index 7b00f87..52f6eb1 100644
--- a/drivers/net/cnxk/cnxk_ptp.c
+++ b/drivers/net/cnxk/cnxk_ptp.c
@@ -56,6 +56,69 @@ cnxk_nix_tsc_convert(struct cnxk_eth_dev *dev)
 }
 
 int
+cnxk_nix_timesync_read_time(struct rte_eth_dev *eth_dev, struct timespec *ts)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   uint64_t clock, ns;
+   int rc;
+
+   rc = roc_nix_ptp_clock_read(nix, &clock, NULL, false);
+   if (rc)
+   return rc;
+
+   ns = rte_timecounter_update(&dev->systime_tc, clock);
+   *ts = rte_ns_to_timespec(ns);
+   return 0;
+}
+
+int
+cnxk_nix_timesync_write_time(struct rte_eth_dev *eth_dev,
+const struct timespec *ts)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   uint64_t ns;
+
+   ns = rte_timespec_to_ns(ts);
+   /* Set the time counters to a new value. */
+   dev->systime_tc.nsec = ns;
+   dev->rx_tstamp_tc.nsec = ns;
+   dev->tx_tstamp_tc.nsec = ns;
+
+   return 0;
+}
+
+int
+cnxk_nix_timesync_adjust_time(struct rte_eth_dev *eth_dev, int64_t delta)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   struct roc_nix *nix = &dev->nix;
+   int rc;
+
+   /* Adjust the frequent to make tics increments in 10^9 tics per sec */
+   if (delta < ROC_NIX_PTP_FREQ_ADJUST &&
+   delta > -ROC_NIX_PTP_FREQ_ADJUST) {
+   rc = roc_nix_ptp_sync_time_adjust(nix, delta);
+   if (rc)
+   return rc;
+
+   /* Since the frequency of PTP comp register is tuned, delta and
+* freq mult calculation for deriving PTP_HI from timestamp
+* counter should be done again.
+*/
+   rc = cnxk_nix_tsc_convert(dev);
+   if (rc)
+   plt_err("Failed to calculate delta and freq mult");
+   }
+
+   dev->systime_tc.nsec += delta;
+   dev->rx_tstamp_tc.nsec += delta;
+   dev->tx_tstamp_tc.nsec += delta;
+
+   return 0;
+}
+
+int
 cnxk_nix_timesync_read_rx_timestamp(struct rte_eth_dev *eth_dev,
struct timespec *timestamp, uint32_t flags)
 {
-- 
2.8.4



[dpdk-dev] [PATCH v3 59/62] net/cnxk: add read clock operation

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch implements read raw clock operation for cn9k and
cn10k.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/features/cnxk.ini |  2 ++
 drivers/net/cnxk/cnxk_ethdev.c|  1 +
 drivers/net/cnxk/cnxk_ethdev.h|  1 +
 drivers/net/cnxk/cnxk_ptp.c   | 17 +
 4 files changed, 21 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index eba4107..5874531 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -32,6 +32,8 @@ L4 checksum offload  = Y
 Inner L3 checksum= Y
 Inner L4 checksum= Y
 Packet type parsing  = Y
+Timesync = Y
+Timestamp offload= Y
 Basic stats  = Y
 Stats per queue  = Y
 Extended stats   = Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index cac8ddf..e1373cf 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1253,6 +1253,7 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.timesync_read_time = cnxk_nix_timesync_read_time,
.timesync_write_time = cnxk_nix_timesync_write_time,
.timesync_adjust_time = cnxk_nix_timesync_adjust_time,
+   .read_clock = cnxk_nix_read_clock,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index 4214365..fa6f16f 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -323,6 +323,7 @@ int cnxk_nix_timesync_write_time(struct rte_eth_dev 
*eth_dev,
 const struct timespec *ts);
 int cnxk_nix_timesync_adjust_time(struct rte_eth_dev *eth_dev, int64_t delta);
 int cnxk_nix_tsc_convert(struct cnxk_eth_dev *dev);
+int cnxk_nix_read_clock(struct rte_eth_dev *eth_dev, uint64_t *clock);
 
 uint64_t cnxk_nix_rxq_mbuf_setup(struct cnxk_eth_dev *dev);
 
diff --git a/drivers/net/cnxk/cnxk_ptp.c b/drivers/net/cnxk/cnxk_ptp.c
index 52f6eb1..449489f 100644
--- a/drivers/net/cnxk/cnxk_ptp.c
+++ b/drivers/net/cnxk/cnxk_ptp.c
@@ -4,6 +4,23 @@
 
 #include "cnxk_ethdev.h"
 
+int
+cnxk_nix_read_clock(struct rte_eth_dev *eth_dev, uint64_t *clock)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+
+   /* This API returns the raw PTP HI clock value. Since LFs do not
+* have direct access to PTP registers and it requires mbox msg
+* to AF for this value. In fastpath reading this value for every
+* packet (which involes mbox call) becomes very expensive, hence
+* we should be able to derive PTP HI clock value from tsc by
+* using freq_mult and clk_delta calculated during configure stage.
+*/
+   *clock = (rte_get_tsc_cycles() + dev->clk_delta) * dev->clk_freq_mult;
+
+   return 0;
+}
+
 /* This function calculates two parameters "clk_freq_mult" and
  * "clk_delta" which is useful in deriving PTP HI clock from
  * timestamp counter (tsc) value.
-- 
2.8.4



[dpdk-dev] [PATCH v3 60/62] net/cnxk: added RETA and RSS hash operations

2021-06-18 Thread Nithin Dabilpuram
From: Satha Rao 

This patch will implement RETA and RSS hash apis. Also added
device argument to lock rx context.

Signed-off-by: Satha Rao 
---
 doc/guides/nics/features/cnxk.ini  |   2 +
 doc/guides/nics/features/cnxk_vec.ini  |   2 +
 doc/guides/nics/features/cnxk_vf.ini   |   2 +
 drivers/net/cnxk/cnxk_ethdev.c |   4 ++
 drivers/net/cnxk/cnxk_ethdev.h |  10 +++
 drivers/net/cnxk/cnxk_ethdev_devargs.c |   4 ++
 drivers/net/cnxk/cnxk_ethdev_ops.c | 121 +
 7 files changed, 145 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 5874531..9945af9 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -23,6 +23,8 @@ Promiscuous mode = Y
 Allmulticast mode= Y
 Unicast MAC filter   = Y
 RSS hash = Y
+RSS key update   = Y
+RSS reta update  = Y
 Inner RSS= Y
 Flow control = Y
 Jumbo frame  = Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index 65ee8ba..b4b1169 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -22,6 +22,8 @@ Promiscuous mode = Y
 Allmulticast mode= Y
 Unicast MAC filter   = Y
 RSS hash = Y
+RSS key update   = Y
+RSS reta update  = Y
 Inner RSS= Y
 Flow control = Y
 Jumbo frame  = Y
diff --git a/doc/guides/nics/features/cnxk_vf.ini 
b/doc/guides/nics/features/cnxk_vf.ini
index 00bde9b..fbc50a8 100644
--- a/doc/guides/nics/features/cnxk_vf.ini
+++ b/doc/guides/nics/features/cnxk_vf.ini
@@ -19,6 +19,8 @@ Queue start/stop = Y
 MTU update   = Y
 TSO  = Y
 RSS hash = Y
+RSS key update   = Y
+RSS reta update  = Y
 Inner RSS= Y
 Jumbo frame  = Y
 Scattered Rx = Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index e1373cf..f49c9e6 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1254,6 +1254,10 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.timesync_write_time = cnxk_nix_timesync_write_time,
.timesync_adjust_time = cnxk_nix_timesync_adjust_time,
.read_clock = cnxk_nix_read_clock,
+   .reta_update = cnxk_nix_reta_update,
+   .reta_query = cnxk_nix_reta_query,
+   .rss_hash_update = cnxk_nix_rss_hash_update,
+   .rss_hash_conf_get = cnxk_nix_rss_hash_conf_get,
 };
 
 static int
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index fa6f16f..cd08e3a 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -330,6 +330,16 @@ uint64_t cnxk_nix_rxq_mbuf_setup(struct cnxk_eth_dev *dev);
 /* RSS */
 uint32_t cnxk_rss_ethdev_to_nix(struct cnxk_eth_dev *dev, uint64_t ethdev_rss,
uint8_t rss_level);
+int cnxk_nix_reta_update(struct rte_eth_dev *eth_dev,
+struct rte_eth_rss_reta_entry64 *reta_conf,
+uint16_t reta_size);
+int cnxk_nix_reta_query(struct rte_eth_dev *eth_dev,
+   struct rte_eth_rss_reta_entry64 *reta_conf,
+   uint16_t reta_size);
+int cnxk_nix_rss_hash_update(struct rte_eth_dev *eth_dev,
+struct rte_eth_rss_conf *rss_conf);
+int cnxk_nix_rss_hash_conf_get(struct rte_eth_dev *eth_dev,
+  struct rte_eth_rss_conf *rss_conf);
 
 /* Link */
 void cnxk_nix_toggle_flag_link_cfg(struct cnxk_eth_dev *dev, bool set);
diff --git a/drivers/net/cnxk/cnxk_ethdev_devargs.c 
b/drivers/net/cnxk/cnxk_ethdev_devargs.c
index 7fd06eb..c76b628 100644
--- a/drivers/net/cnxk/cnxk_ethdev_devargs.c
+++ b/drivers/net/cnxk/cnxk_ethdev_devargs.c
@@ -109,6 +109,7 @@ parse_switch_header_type(const char *key, const char 
*value, void *extra_args)
 #define CNXK_FLOW_MAX_PRIORITY "flow_max_priority"
 #define CNXK_SWITCH_HEADER_TYPE "switch_header"
 #define CNXK_RSS_TAG_AS_XOR"tag_as_xor"
+#define CNXK_LOCK_RX_CTX   "lock_rx_ctx"
 
 int
 cnxk_ethdev_parse_devargs(struct rte_devargs *devargs, struct cnxk_eth_dev 
*dev)
@@ -120,6 +121,7 @@ cnxk_ethdev_parse_devargs(struct rte_devargs *devargs, 
struct cnxk_eth_dev *dev)
uint16_t flow_max_priority = 3;
uint16_t rss_tag_as_xor = 0;
uint16_t scalar_enable = 0;
+   uint8_t lock_rx_ctx = 0;
struct rte_kvargs *kvlist;
 
if (devargs == NULL)
@@ -143,6 +145,7 @@ cnxk_ethdev_parse_devargs(struct rte_devargs *devargs, 
struct cnxk_eth_dev *dev)
   &parse_switch_header_type, &switch_header_type);
rte_kvargs_process(kvlist, CNXK_RSS_TAG_AS_XOR, &parse_flag,
   &rss_tag_as_xor);
+   rte_kvargs_process(kvlist, CNXK_LOCK_RX_CTX, &parse_flag, &lock_rx_ctx);
rte_kvargs_free(kvlist);
 
 null_devargs:
@@ -150,6 +153,7 @@ cnxk_ethdev_parse_de

[dpdk-dev] [PATCH v3 61/62] net/cnxk: add multicast filter support

2021-06-18 Thread Nithin Dabilpuram
From: Sunil Kumar Kori 

Patch adds multicast filter support for cn9k and cn10k platforms.

CGX DMAC filter table(32 entries) is divided among all LMACs
connected to it i.e. if CGX has 4 LMACs then each LMAC can have
up to 8 filters. If CGX has 1 LMAC then it can have up to 32
filters.

Above mentioned filter table is used to install unicast and multicast
DMAC address filters. Unicast filters are installed via
rte_eth_dev_mac_addr_add API while multicast filters are installed
via rte_eth_dev_set_mc_addr_list API.

So in total, supported MAC filters are equal to DMAC filters plus
mcast filters.

Signed-off-by: Sunil Kumar Kori 
---
 doc/guides/nics/features/cnxk.ini |  1 +
 doc/guides/nics/features/cnxk_vec.ini |  1 +
 drivers/net/cnxk/cnxk_ethdev.c|  2 ++
 drivers/net/cnxk/cnxk_ethdev.h|  4 +++
 drivers/net/cnxk/cnxk_ethdev_ops.c| 63 +++
 5 files changed, 71 insertions(+)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index 9945af9..c686ceb 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -22,6 +22,7 @@ TSO  = Y
 Promiscuous mode = Y
 Allmulticast mode= Y
 Unicast MAC filter   = Y
+Multicast MAC filter = Y
 RSS hash = Y
 RSS key update   = Y
 RSS reta update  = Y
diff --git a/doc/guides/nics/features/cnxk_vec.ini 
b/doc/guides/nics/features/cnxk_vec.ini
index b4b1169..4c3f78e 100644
--- a/doc/guides/nics/features/cnxk_vec.ini
+++ b/doc/guides/nics/features/cnxk_vec.ini
@@ -21,6 +21,7 @@ MTU update   = Y
 Promiscuous mode = Y
 Allmulticast mode= Y
 Unicast MAC filter   = Y
+Multicast MAC filter = Y
 RSS hash = Y
 RSS key update   = Y
 RSS reta update  = Y
diff --git a/drivers/net/cnxk/cnxk_ethdev.c b/drivers/net/cnxk/cnxk_ethdev.c
index f49c9e6..1edea35 100644
--- a/drivers/net/cnxk/cnxk_ethdev.c
+++ b/drivers/net/cnxk/cnxk_ethdev.c
@@ -1258,6 +1258,7 @@ struct eth_dev_ops cnxk_eth_dev_ops = {
.reta_query = cnxk_nix_reta_query,
.rss_hash_update = cnxk_nix_rss_hash_update,
.rss_hash_conf_get = cnxk_nix_rss_hash_conf_get,
+   .set_mc_addr_list = cnxk_nix_mc_addr_list_configure,
 };
 
 static int
@@ -1322,6 +1323,7 @@ cnxk_eth_dev_init(struct rte_eth_dev *eth_dev)
}
 
dev->max_mac_entries = max_entries;
+   dev->dmac_filter_count = 1;
 
/* Get mac address */
rc = roc_nix_npc_mac_addr_get(nix, dev->mac_addr);
diff --git a/drivers/net/cnxk/cnxk_ethdev.h b/drivers/net/cnxk/cnxk_ethdev.h
index cd08e3a..c07e988 100644
--- a/drivers/net/cnxk/cnxk_ethdev.h
+++ b/drivers/net/cnxk/cnxk_ethdev.h
@@ -162,6 +162,7 @@ struct cnxk_eth_dev {
uint8_t configured;
 
/* Max macfilter entries */
+   uint8_t dmac_filter_count;
uint8_t max_mac_entries;
bool dmac_filter_enable;
 
@@ -265,6 +266,9 @@ int cnxk_nix_probe(struct rte_pci_driver *pci_drv,
   struct rte_pci_device *pci_dev);
 int cnxk_nix_remove(struct rte_pci_device *pci_dev);
 int cnxk_nix_mtu_set(struct rte_eth_dev *eth_dev, uint16_t mtu);
+int cnxk_nix_mc_addr_list_configure(struct rte_eth_dev *eth_dev,
+   struct rte_ether_addr *mc_addr_set,
+   uint32_t nb_mc_addr);
 int cnxk_nix_mac_addr_add(struct rte_eth_dev *eth_dev,
  struct rte_ether_addr *addr, uint32_t index,
  uint32_t pool);
diff --git a/drivers/net/cnxk/cnxk_ethdev_ops.c 
b/drivers/net/cnxk/cnxk_ethdev_ops.c
index 28b49ad..3be678a 100644
--- a/drivers/net/cnxk/cnxk_ethdev_ops.c
+++ b/drivers/net/cnxk/cnxk_ethdev_ops.c
@@ -359,6 +359,7 @@ cnxk_nix_mac_addr_add(struct rte_eth_dev *eth_dev, struct 
rte_ether_addr *addr,
roc_nix_npc_promisc_ena_dis(nix, true);
dev->dmac_filter_enable = true;
eth_dev->data->promiscuous = false;
+   dev->dmac_filter_count++;
 
return 0;
 }
@@ -373,6 +374,8 @@ cnxk_nix_mac_addr_del(struct rte_eth_dev *eth_dev, uint32_t 
index)
rc = roc_nix_mac_addr_del(nix, index);
if (rc)
plt_err("Failed to delete mac address, rc=%d", rc);
+
+   dev->dmac_filter_count--;
 }
 
 int
@@ -845,3 +848,63 @@ cnxk_nix_rss_hash_conf_get(struct rte_eth_dev *eth_dev,
 
return 0;
 }
+
+int
+cnxk_nix_mc_addr_list_configure(struct rte_eth_dev *eth_dev,
+   struct rte_ether_addr *mc_addr_set,
+   uint32_t nb_mc_addr)
+{
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   const struct rte_ether_addr null_mac_addr = {0};
+   struct rte_eth_dev_data *data = eth_dev->data;
+   struct roc_nix *nix = &dev->nix;
+   int rc, index;
+   uint32_t i;
+
+   /* All configured multicast filters should be flushed first */
+   for (i = 0; i < dev->max_mac_entries; i++) {
+   if (rte_is_multicast_ether_ad

[dpdk-dev] [PATCH v3 62/62] net/cnxk: add marking and VLAN tagging support

2021-06-18 Thread Nithin Dabilpuram
From: Satheesh Paul 

This patch adds support for mark, flag, VLAN pop and
push flow actions.

Signed-off-by: Satheesh Paul 
---
 doc/guides/nics/features/cnxk.ini |   5 ++
 drivers/net/cnxk/cn10k_rte_flow.c |  42 +
 drivers/net/cnxk/cn10k_rx.c   |  27 +++---
 drivers/net/cnxk/cn10k_rx.h   | 175 ++
 drivers/net/cnxk/cn10k_rx_mseg.c  |   2 +-
 drivers/net/cnxk/cn10k_rx_vec.c   |   2 +-
 drivers/net/cnxk/cn9k_rte_flow.c  |  42 +
 drivers/net/cnxk/cn9k_rx.c|  27 +++---
 drivers/net/cnxk/cn9k_rx.h| 175 ++
 drivers/net/cnxk/cn9k_rx_mseg.c   |   2 +-
 drivers/net/cnxk/cn9k_rx_vec.c|   2 +-
 drivers/net/cnxk/cnxk_ethdev.c|   1 +
 drivers/net/cnxk/cnxk_ethdev.h|   3 +-
 drivers/net/cnxk/cnxk_rte_flow.c  |  17 
 14 files changed, 423 insertions(+), 99 deletions(-)

diff --git a/doc/guides/nics/features/cnxk.ini 
b/doc/guides/nics/features/cnxk.ini
index c686ceb..337ad24 100644
--- a/doc/guides/nics/features/cnxk.ini
+++ b/doc/guides/nics/features/cnxk.ini
@@ -77,6 +77,11 @@ vxlan_gpe= Y
 count= Y
 drop = Y
 flag = Y
+mark = Y
+of_pop_vlan  = Y
+of_push_vlan = Y
+of_set_vlan_pcp  = Y
+of_set_vlan_vid  = Y
 pf   = Y
 port_id  = Y
 queue= Y
diff --git a/drivers/net/cnxk/cn10k_rte_flow.c 
b/drivers/net/cnxk/cn10k_rte_flow.c
index 65893cc..b04de6a 100644
--- a/drivers/net/cnxk/cn10k_rte_flow.c
+++ b/drivers/net/cnxk/cn10k_rte_flow.c
@@ -4,6 +4,7 @@
 #include 
 #include "cn10k_rte_flow.h"
 #include "cn10k_ethdev.h"
+#include "cn10k_rx.h"
 
 struct rte_flow *
 cn10k_flow_create(struct rte_eth_dev *eth_dev, const struct rte_flow_attr 
*attr,
@@ -11,12 +12,29 @@ cn10k_flow_create(struct rte_eth_dev *eth_dev, const struct 
rte_flow_attr *attr,
  const struct rte_flow_action actions[],
  struct rte_flow_error *error)
 {
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   int mark_actions = 0, vtag_actions = 0;
+   struct roc_npc *npc = &dev->npc;
struct roc_npc_flow *flow;
 
flow = cnxk_flow_create(eth_dev, attr, pattern, actions, error);
if (!flow)
return NULL;
 
+   mark_actions = roc_npc_mark_actions_get(npc);
+
+   if (mark_actions) {
+   dev->rx_offload_flags |= NIX_RX_OFFLOAD_MARK_UPDATE_F;
+   cn10k_eth_set_rx_function(eth_dev);
+   }
+
+   vtag_actions = roc_npc_vtag_actions_get(npc);
+
+   if (vtag_actions) {
+   dev->rx_offload_flags |= NIX_RX_OFFLOAD_VLAN_STRIP_F;
+   cn10k_eth_set_rx_function(eth_dev);
+   }
+
return (struct rte_flow *)flow;
 }
 
@@ -25,6 +43,30 @@ cn10k_flow_destroy(struct rte_eth_dev *eth_dev, struct 
rte_flow *rte_flow,
   struct rte_flow_error *error)
 {
struct roc_npc_flow *flow = (struct roc_npc_flow *)rte_flow;
+   struct cnxk_eth_dev *dev = cnxk_eth_pmd_priv(eth_dev);
+   int mark_actions = 0, vtag_actions = 0;
+   struct roc_npc *npc = &dev->npc;
+
+   mark_actions = roc_npc_mark_actions_get(npc);
+   if (mark_actions) {
+   mark_actions = roc_npc_mark_actions_sub_return(npc, 1);
+   if (mark_actions == 0) {
+   dev->rx_offload_flags &= ~NIX_RX_OFFLOAD_MARK_UPDATE_F;
+   cn10k_eth_set_rx_function(eth_dev);
+   }
+   }
+
+   vtag_actions = roc_npc_vtag_actions_get(npc);
+   if (vtag_actions) {
+   if (flow->nix_intf == ROC_NPC_INTF_RX) {
+   vtag_actions = roc_npc_vtag_actions_sub_return(npc, 1);
+   if (vtag_actions == 0) {
+   dev->rx_offload_flags &=
+   ~NIX_RX_OFFLOAD_VLAN_STRIP_F;
+   cn10k_eth_set_rx_function(eth_dev);
+   }
+   }
+   }
 
return cnxk_flow_destroy(eth_dev, flow, error);
 }
diff --git a/drivers/net/cnxk/cn10k_rx.c b/drivers/net/cnxk/cn10k_rx.c
index c9744e2..5c956c0 100644
--- a/drivers/net/cnxk/cn10k_rx.c
+++ b/drivers/net/cnxk/cn10k_rx.c
@@ -5,7 +5,7 @@
 #include "cn10k_ethdev.h"
 #include "cn10k_rx.h"
 
-#define R(name, f4, f3, f2, f1, f0, flags)\
+#define R(name, f5, f4, f3, f2, f1, f0, flags)\
uint16_t __rte_noinline __rte_hot cn10k_nix_recv_pkts_##name(  \
void *rx_queue, struct rte_mbuf **rx_pkts, uint16_t pkts)  \
{  \
@@ -17,12 +17,13 @@ NIX_RX_FASTPATH_MODES
 
 static inline void
 pick_rx_func(struct rte_eth_dev *eth_dev,
-const eth_rx_burst_t rx_burst[2][2][2][2][2])
+const eth_rx_burst_t rx_burst[2][2][2

Re: [dpdk-dev] [PATCH] parray: introduce internal API for dynamic arrays

2021-06-18 Thread Ferruh Yigit
On 6/18/2021 10:14 AM, Morten Brørup wrote:
>> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Ananyev,
>> Konstantin
>> Sent: Thursday, 17 June 2021 19.06
>>
>> I think it would be better to change rx_pkt_burst() to accept port_id
>> and queue_id, instead of void *.
> 
> Current:
> 
> typedef uint16_t (*eth_rx_burst_t)(void *rxq,
>  struct rte_mbuf **rx_pkts,
>  uint16_t nb_pkts);
> 
>> I.E:
>> typedef uint16_t (*eth_rx_burst_t)(uint16_t port_id,
>> uint16_t queue_id,
>> struct rte_mbuf **rx_pkts,
>> uint16_t nb_pkts);
>>
>> And we can do actual de-referencing of private rxq data inside the
>> actual rx function.
> 
> Good idea, if it can be done without a performance cost.
> 
> The X64 calling convention allows up to 4 parameters passed as registers, so 
> the added parameter should not be a problem.
> 
> 
> Another thing:
> 
> I just noticed that struct rte_eth_dev_data has "void **rx_queues;" (and 
> similarly for tx_queues).
> 
> That should be "void *rx_queues[RTE_MAX_QUEUES_PER_PORT];", like in all the 
> other ethdev structures.
> 

Why have a fixed size rx_queues array? It is already allocated dynamically in
'rte_eth_dev_configure()' based on actual Rx/Tx queue number.

We are already trying to get rid of compile time fixed array sizes, so I think
better to keep it as it is.

Also this will increase the strcut size.

> The same structure even has "uint8_t 
> rx_queue_state[RTE_MAX_QUEUES_PER_PORT];", so it's following two different 
> conventions.
> 

I wonder if we should should switch these to dynamic allocation too.



Re: [dpdk-dev] [PATCH] parray: introduce internal API for dynamic arrays

2021-06-18 Thread Ferruh Yigit
On 6/18/2021 11:41 AM, Ananyev, Konstantin wrote:
> 
>>>
>>> 14/06/2021 15:15, Bruce Richardson:
 On Mon, Jun 14, 2021 at 02:22:42PM +0200, Morten Brørup wrote:
>> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Thomas Monjalon
>> Sent: Monday, 14 June 2021 12.59
>>
>> Performance of access in a fixed-size array is very good
>> because of cache locality
>> and because there is a single pointer to dereference.
>> The only drawback is the lack of flexibility:
>> the size of such an array cannot be increase at runtime.
>>
>> An approach to this problem is to allocate the array at runtime,
>> being as efficient as static arrays, but still limited to a maximum.
>>
>> That's why the API rte_parray is introduced,
>> allowing to declare an array of pointer which can be resized
>> dynamically
>> and automatically at runtime while keeping a good read performance.
>>
>> After resize, the previous array is kept until the next resize
>> to avoid crashs during a read without any lock.
>>
>> Each element is a pointer to a memory chunk dynamically allocated.
>> This is not good for cache locality but it allows to keep the same
>> memory per element, no matter how the array is resized.
>> Cache locality could be improved with mempools.
>> The other drawback is having to dereference one more pointer
>> to read an element.
>>
>> There is not much locks, so the API is for internal use only.
>> This API may be used to completely remove some compilation-time
>> maximums.
>
> I get the purpose and overall intention of this library.
>
> I probably already mentioned that I prefer "embedded style 
> programming" with fixed size arrays, rather than runtime
>> configurability.
>> It's
>>> my personal opinion, and the DPDK Tech Board clearly prefers reducing 
>>> the amount of compile time configurability, so there is no
>> way
 for
>>> me to stop this progress, and I do not intend to oppose to this 
>>> library. :-)
>
> This library is likely to become a core library of DPDK, so I think 
> it is important getting it right. Could you please mention a few
>> examples
>>> where you think this internal library should be used, and where it 
>>> should not be used. Then it is easier to discuss if the border line
 between
>>> control path and data plane is correct. E.g. this library is not 
>>> intended to be used for dynamically sized packet queues that grow and
 shrink
>> in
>>> the fast path.
>
> If the library becomes a core DPDK library, it should probably be 
> public instead of internal. E.g. if the library is used to make
>>> RTE_MAX_ETHPORTS dynamic instead of compile time fixed, then some 
>>> applications might also need dynamically sized arrays for
>> their
>>> application specific per-port runtime data, and this library could 
>>> serve that purpose too.
>

 Thanks Thomas for starting this discussion and Morten for follow-up.

 My thinking is as follows, and I'm particularly keeping in mind the 
 cases
 of e.g. RTE_MAX_ETHPORTS, as a leading candidate here.

 While I dislike the hard-coded limits in DPDK, I'm also not convinced 
 that
 we should switch away from the flat arrays or that we need fully 
 dynamic
 arrays that grow/shrink at runtime for ethdevs. I would suggest a 
 half-way
 house here, where we keep the ethdevs as an array, but one 
 allocated/sized
 at runtime rather than statically. This would allow us to have a
 compile-time default value, but, for use cases that need it, allow use 
 of a
 flag e.g.  "max-ethdevs" to change the size of the parameter given to 
 the
 malloc call for the array.  This max limit could then be provided to 
 apps
 too if they want to match any array sizes. [Alternatively those apps 
 could
 check the provided size and error out if the size has been increased 
 beyond
 what the app is designed to use?]. There would be no extra 
 dereferences per
 rx/tx burst call in this scenario so performance should be the same as
 before (potentially better if array is in hugepage memory, I suppose).
>>>
>>> I think we need some benchmarks to decide what is the best tradeoff.
>>> I spent time on this implementation, but sorry I won't have time for 
>>> benchmarks.
>>> Volunteers?
>>
>> I had only a quick look at your approach so far.
>> But from what I can read, in MT environment your suggestion will require

[dpdk-dev] [Bug 728] [dpdk-19.11.9-rc2]When compiling dpdk with make or meson under SUSE Linux Enterprise Server 15 SP2, the compilation fails

2021-06-18 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=728

jiang,yu (yux.ji...@intel.com) changed:

   What|Removed |Added

 CC||yux.ji...@intel.com
 Status|UNCONFIRMED |RESOLVED
 Resolution|--- |FIXED

--- Comment #2 from jiang,yu (yux.ji...@intel.com) ---
Test passed on LTS19.11.9-rc3
commit e5f56f22a82a1df49c224a8af13e769ff4b04cd8
Author: Christian Ehrhardt 
Date:   Thu Jun 17 08:25:20 2021 +0200

version: 19.11.9-rc3

Signed-off-by: Christian Ehrhardt 

-- 
You are receiving this mail because:
You are the assignee for the bug.

[dpdk-dev] [Bug 732] [dpdk-19.11.9-rc2] DPDK build failed on Win10

2021-06-18 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=732

jiang,yu (yux.ji...@intel.com) changed:

   What|Removed |Added

 Status|CONFIRMED   |RESOLVED
 Resolution|--- |FIXED

--- Comment #3 from jiang,yu (yux.ji...@intel.com) ---
Test passed on LTS19.11.9-rc3
commit e5f56f22a82a1df49c224a8af13e769ff4b04cd8
Author: Christian Ehrhardt 
Date:   Thu Jun 17 08:25:20 2021 +0200

version: 19.11.9-rc3

Signed-off-by: Christian Ehrhardt 

-- 
You are receiving this mail because:
You are the assignee for the bug.

Re: [dpdk-dev] [PATCH] parray: introduce internal API for dynamic arrays

2021-06-18 Thread Morten Brørup
> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Ferruh Yigit
> Sent: Friday, 18 June 2021 12.47
> 
> On 6/18/2021 10:14 AM, Morten Brørup wrote:
> > Another thing:
> >
> > I just noticed that struct rte_eth_dev_data has "void **rx_queues;"
> (and similarly for tx_queues).
> >
> > That should be "void *rx_queues[RTE_MAX_QUEUES_PER_PORT];", like in
> all the other ethdev structures.
> >
> 
> Why have a fixed size rx_queues array? It is already allocated
> dynamically in
> 'rte_eth_dev_configure()' based on actual Rx/Tx queue number.

For performance reasons, I guess.

> 
> We are already trying to get rid of compile time fixed array sizes, so
> I think
> better to keep it as it is.
> 
> Also this will increase the strcut size.
> 
> > The same structure even has "uint8_t
> rx_queue_state[RTE_MAX_QUEUES_PER_PORT];", so it's following two
> different conventions.
> >
> 
> I wonder if we should should switch these to dynamic allocation too.
> 

I agree that we should generally stick with one or the other, either fixed or 
dynamically allocated arrays, for consistency reasons. However, sometimes 
performance trumps consistency. :-)

If we change more per-queue arrays to dynamic allocation, perhaps it would be 
beneficial gathering these fields into one per-queue struct, so the struct 
rte_eth_dev_data only has one array instead of multiple arrays. It depends on 
how they are being used. (Also, maybe there should be one array in each 
direction, so it is two arrays, not just one.)




[dpdk-dev] [Bug 744] [dpdk-19.11.9-rc3] make build failed on Fedora34 with GCC11

2021-06-18 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=744

Bug ID: 744
   Summary: [dpdk-19.11.9-rc3] make build failed on Fedora34 with
GCC11
   Product: DPDK
   Version: 19.11
  Hardware: All
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: core
  Assignee: dev@dpdk.org
  Reporter: yux.ji...@intel.com
  Target Milestone: ---

[DPDK version]:
e5f56f22a8 (HEAD, tag: v19.11.9-rc3, origin/19.11) version: 19.11.9-rc3

[OS version]:
Fedora 34 (Server Edition)/5.12.10-300.fc34.x86_64
gcc version 11.1.1 20210531 (Red Hat 11.1.1-3) (GCC)


[Test Setup]:
export RTE_TARGET=x86_64-native-linuxapp-gcc
export RTE_SDK=`pwd`
rm -rf x86_64-native-linuxapp-gcc
rm -rf ./app/test/test_resource_c.res.o
rm -rf ./app/test/test_resource_tar.res.o
rm -rf ./app/test/test_pci_sysfs.res.o
make -j 20 install T=x86_64-native-linuxapp-gcc  MAKE_PAUSE=n

Should support to build 19.11.9 with GCC11 or not?

[Show the output from the previous commands]:
  PMDINFO hn_ethdev.o.pmd.c
  CC hn_ethdev.o.pmd.o
  LD hn_ethdev.o
  CC hn_rxtx.o
In file included from /root/jaccy/dpdk/drivers/net/ice/base/ice_type.h:43,
 from /root/jaccy/dpdk/drivers/net/ice/base/ice_common.h:8,
 from /root/jaccy/dpdk/drivers/net/ice/base/ice_switch.h:8,
 from /root/jaccy/dpdk/drivers/net/ice/base/ice_switch.c:5:
In function ‘ice_is_bit_set_internal’,
inlined from ‘ice_is_bit_set’ at
/root/jaccy/dpdk/drivers/net/ice/base/ice_bitops.h:85:9,
inlined from ‘ice_get_recp_to_prof_map’ at
/root/jaccy/dpdk/drivers/net/ice/base/ice_switch.c:654:8,
inlined from ‘ice_get_recp_frm_fw’ at
/root/jaccy/dpdk/drivers/net/ice/base/ice_switch.c:520:3,
inlined from ‘ice_find_recp’ at
/root/jaccy/dpdk/drivers/net/ice/base/ice_switch.c:4695:8:
/root/jaccy/dpdk/drivers/net/ice/base/ice_bitops.h:30:19: error: ‘r_bitmap’ may
be used uninitialized [-Werror=maybe-uninitialized]
   30 | return !!(*bitmap & BIT(nr));
  |   ^~~
/root/jaccy/dpdk/drivers/net/ice/base/ice_switch.c: In function
‘ice_find_recp’:
/root/jaccy/dpdk/drivers/net/ice/base/ice_switch.c:641:28: note: ‘r_bitmap’
declared here
  641 | ice_declare_bitmap(r_bitmap, ICE_MAX_NUM_RECIPES);
  |^~~~
/root/jaccy/dpdk/drivers/net/ice/base/ice_bitops.h:26:22: note: in definition
of macro ‘ice_declare_bitmap’
   26 | ice_bitmap_t A[BITS_TO_CHUNKS(sz)]
  |  ^
  CC ixgbe_phy.o
  PMDINFO igb_ethdev.o.pmd.c
  CC igb_ethdev.o.pmd.o
  CC ixgbe_api.o
  LD igb_ethdev.o
  CC igb_pf.o
  CC ixgbe_vf.o
  CC bnxt_ethdev.o
  CC lio_23xx_vf.o
  CC bnxt_filter.o
  CC clip_tbl.o
  CC nfp_mutex.o
  PMDINFO rte_eth_null.o.pmd.c
  CC rte_eth_null.o.pmd.o
  CC igb_flow.o
  CC em_ethdev.o
  CC em_rxtx.o
  LD rte_eth_null.o
  AR librte_pmd_null.a
  CC base/vnic_rq.o
  CC hns3_stats.o
  INSTALL-LIB librte_pmd_null.a
  CC hns3_regs.o
  CC hns3_dcb.o
  CC hns3_mp.o
  CC hinic_pmd_ethdev.o
  CC lio_mbox.o
  CC ixgbe_hv_vf.o
  CC ice_flex_pipe.o
  CC nfp_resource.o
  CC mps_tcam.o
  CC l2t.o
  CC ixgbe_dcb.o
  CC enic_rxtx_vec_avx2.o
  PMDINFO em_ethdev.o.pmd.c
  CC em_ethdev.o.pmd.o
  LD em_ethdev.o
  CC t4vf_hw.o
  CC ixgbe_dcb_82599.o
  CC hn_rndis.o
  CC ixgbe_dcb_82598.o
  CC ixgbe_mbx.o
  CC ixgbe_rxtx.o
cc1: all warnings being treated as errors
  AR librte_pmd_lio.a
make[6]: *** [/root/jaccy/dpdk/mk/internal/rte.compile-pre.mk:116:
ice_switch.o] Error 1
make[6]: *** Waiting for unfinished jobs
  CC nfp_crc.o
  CC nfp_mip.o
  INSTALL-LIB librte_pmd_lio.a
  CC ixgbe_ethdev.o
  CC bnxt_flow.o
  CC bnxt_hwrm.o
  CC bnxt_ring.o
  CC bnxt_rxq.o
  CC bnxt_rxr.o
  CC hinic_pmd_rx.o
  CC bnxt_stats.o
  CC bnxt_txq.o
  AR librte_pmd_e1000.a
  PMDINFO hinic_pmd_ethdev.o.pmd.c
  CC hinic_pmd_ethdev.o.pmd.o
  INSTALL-LIB librte_pmd_e1000.a
  AR librte_pmd_hns3.a
  CC bnxt_txr.o
  LD hinic_pmd_ethdev.o
  INSTALL-LIB librte_pmd_hns3.a
  CC bnxt_vnic.o
  CC bnxt_irq.o
  AR librte_pmd_enic.a
  INSTALL-LIB librte_pmd_enic.a
  CC nfp_nffw.o
  PMDINFO bnxt_ethdev.o.pmd.c
  CC bnxt_ethdev.o.pmd.o
  CC hn_nvs.o
  LD bnxt_ethdev.o
  CC hn_vf.o
  CC bnxt_util.o
  AR librte_pmd_cxgbe.a
  INSTALL-LIB librte_pmd_cxgbe.a
== Build drivers/net/octeontx
== Build drivers/net/octeontx2
  CC rte_pmd_bnxt.o
  CC bnxt_rxtx_vec_sse.o
== Build drivers/net/qede
  CC ixgbe_fdir.o
  CC nfp_hwinfo.o
  CC hinic_pmd_tx.o
  CC nfp_rtsym.o
  CC nfp_nsp.o
  CC i40e_rxtx_vec_sse.o
  CC i40e_ethdev_vf.o
  CC ixgbe_pf.o
== Build drivers/net/ring
  AR librte_pmd_netvsc.a
  CC otx2_rx.o
  CC octeontx_rxtx.o
  INSTALL-LIB librte_pmd_netvsc.a
== Build drivers/net/sfc
  CC nfp_nsp_cmds.o
  CC nfp_nsp_eth.o
  CC ecore_dev.o
make[5]: *** [/root/jaccy/dpdk/mk/rte.subdir.mk:37: ice] Error 2
make[5]: *** Waiting for unfinished jobs
  CC ixgbe_flow.o
  SYMLINK-FILE include/rte_eth_ring.h
  CC rte_eth

  1   2   >