Re: [dpdk-dev] [PATCH v1 08/14] vhost: improve IO vector logic

2021-10-26 Thread Hu, Jiayu
Hi Maxime,

> -Original Message-
> From: Maxime Coquelin 
> Sent: Monday, October 25, 2021 6:03 PM
> To: Hu, Jiayu ; dev@dpdk.org; Xia, Chenbo
> ; Wang, YuanX ; Ma,
> WenwuX ; Richardson, Bruce
> ; Mcnamara, John
> ; david.march...@redhat.com
> Subject: Re: [PATCH v1 08/14] vhost: improve IO vector logic
> 
> Hi Jiayu,
> 
> On 10/25/21 09:22, Hu, Jiayu wrote:
> > Hi Maxime,
> >
> >> -Original Message-
> >> From: Maxime Coquelin 
> >> Sent: Monday, October 18, 2021 9:02 PM
> >> To: dev@dpdk.org; Xia, Chenbo ; Hu, Jiayu
> >> ; Wang, YuanX ; Ma,
> WenwuX
> >> ; Richardson, Bruce
> >> ; Mcnamara, John
> >> ; david.march...@redhat.com
> >> Cc: Maxime Coquelin 
> >> Subject: [PATCH v1 08/14] vhost: improve IO vector logic
> >>
> >> IO vectors and their iterators arrays were part of the async metadata
> >> but not their indexes.
> >>
> >> In order to makes this more consistent, the patch adds the indexes to
> >> the async metadata. Doing that, we can avoid triggering DMA transfer
> >> within the loop as it IO vector index overflow is now prevented in
> >> the
> >> async_mbuf_to_desc() function.
> >>
> >> Note that previous detection mechanism was broken since the overflow
> >> already happened when detected, so OOB memory access would already
> >> have happened.
> >>
> >> With this changes done, virtio_dev_rx_async_submit_split()
> >> and virtio_dev_rx_async_submit_packed() can be further simplified.
> >>
> >> Signed-off-by: Maxime Coquelin 
> >> ---
> >>   lib/vhost/vhost.h  |   2 +
> >>   lib/vhost/virtio_net.c | 291 ++---
> >>   2 files changed, 131 insertions(+), 162 deletions(-)
> >>
> >> diff --git a/lib/vhost/vhost.h b/lib/vhost/vhost.h index
> >> dae9a1ac2d..812d4c55a5 100644
> >> --- a/lib/vhost/vhost.h
> >> +++ b/lib/vhost/vhost.h
> >> @@ -134,6 +134,8 @@ struct vhost_async {
> >>
> >>struct rte_vhost_iov_iter iov_iter[VHOST_MAX_ASYNC_IT];
> >>struct rte_vhost_iovec iovec[VHOST_MAX_ASYNC_VEC];
> >> +  uint16_t iter_idx;
> >> +  uint16_t iovec_idx;
> >>
> >>/* data transfer status */
> >>struct async_inflight_info *pkts_info; diff --git
> >> a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c index
> >> ae7dded979..c80823a8de 100644
> >> --- a/lib/vhost/virtio_net.c
> >> +++ b/lib/vhost/virtio_net.c
> >> @@ -924,33 +924,86 @@ copy_mbuf_to_desc(struct virtio_net *dev,
> >> struct vhost_virtqueue *vq,
> >>return error;
> >>   }
> >>
> >> +static __rte_always_inline int
> >> +async_iter_initialize(struct vhost_async *async) {
> >> +  struct rte_vhost_iov_iter *iter;
> >> +
> >> +  if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
> >> +  VHOST_LOG_DATA(ERR, "no more async iovec available\n");
> >> +  return -1;
> >> +  }
> >> +
> >> +  iter = async->iov_iter + async->iter_idx;
> >> +  iter->iov = async->iovec + async->iovec_idx;
> >> +  iter->nr_segs = 0;
> >> +
> >> +  return 0;
> >> +}
> >> +
> >> +static __rte_always_inline int
> >> +async_iter_add_iovec(struct vhost_async *async, void *src, void
> >> +*dst, size_t len) {
> >> +  struct rte_vhost_iov_iter *iter;
> >> +  struct rte_vhost_iovec *iovec;
> >> +
> >> +  if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
> >> +  VHOST_LOG_DATA(ERR, "no more async iovec available\n");
> >> +  return -1;
> >> +  }
> >
> > For large packets, like 64KB in iperf test, async_iter_add_iovec()
> > frequently reports the log above, as we run out of iovecs. I think
> > it's better to change the log from ERR to DEBUG.
> 
> I think it is better to keep it as an error, we want to see it if it happens
> without having the user to enable debug.
> 
> But maybe we can only print it once, not to flood the logs.

OK.

> 
> > In addition, the size of iovec is too small. For burst 32 and 64KB
> > pkts, it's easy to run out of iovecs and we will drop the pkts to
> > enqueue if it happens, which hurts performance. Enlarging the array is
> > a choice to mitigate the issue, but another solution is to reallocate
> > iovec once we run out of it. How do you think?
> 
> I would prefer we enlarge the array, reallocating the array when the issue
> happens sounds like over-engineering to me.
> 
> Any idea what size it should be based on your experiments?

2048 is enough for iperf and 64KB pkts.

Thanks,
Jiayu
> 
> Thanks,
> Maxime
> 
> > Thanks,
> > Jiayu
> >> +
> >> +  iter = async->iov_iter + async->iter_idx;
> >> +  iovec = async->iovec + async->iovec_idx;
> >> +
> >> +  iovec->src_addr = src;
> >> +  iovec->dst_addr = dst;
> >> +  iovec->len = len;
> >> +
> >> +  iter->nr_segs++;
> >> +  async->iovec_idx++;
> >> +
> >> +  return 0;
> >> +}
> >



Re: [dpdk-dev] [PATCH 1/2] ethdev: fix log level of Tx and Rx dummy functions

2021-10-26 Thread Bing Zhao
Hi David,

> -Original Message-
> From: David Marchand 
> Sent: Monday, October 25, 2021 9:32 PM
> To: NBU-Contact-Thomas Monjalon 
> Cc: Ananyev, Konstantin ; Bing Zhao
> ; Yigit, Ferruh ;
> andrew.rybche...@oktetlabs.ru; dev@dpdk.org
> Subject: Re: [PATCH 1/2] ethdev: fix log level of Tx and Rx dummy
> functions
> 
> External email: Use caution opening links or attachments
> 
> 
> On Mon, Oct 25, 2021 at 3:27 PM Thomas Monjalon 
> wrote:
> > > > > There is a concern about getting efficient log report,
> > > > > especially when looking at CI issues.
> > > >
> > > > +1.
> > > > The current solution with logs is a real pain.
> > >
> > > Are you guys talking about problems with
> > > app/test/sample_packet_forward.* David reported?
> > > Or some extra problems arise?
> >
> > The problem will arise each time an app is misbehaving.
> > That's going to be a recurring problem in the CI.
> >
> 
> One thing that could be done is compiling with asserts in CI, and
> let default build not have those asserts.
> 
> Otherwise, logging once should be enough (I have a patch for this
> latter idea).

If you already have a patch to log once, I will suppress this patch. Thanks

> 
> 
> --
> David Marchand

BR. Bing



Re: [dpdk-dev] [PATCH v1 08/14] vhost: improve IO vector logic

2021-10-26 Thread Maxime Coquelin




On 10/26/21 09:07, Hu, Jiayu wrote:

Hi Maxime,


-Original Message-
From: Maxime Coquelin 
Sent: Monday, October 25, 2021 6:03 PM
To: Hu, Jiayu ; dev@dpdk.org; Xia, Chenbo
; Wang, YuanX ; Ma,
WenwuX ; Richardson, Bruce
; Mcnamara, John
; david.march...@redhat.com
Subject: Re: [PATCH v1 08/14] vhost: improve IO vector logic

Hi Jiayu,

On 10/25/21 09:22, Hu, Jiayu wrote:

Hi Maxime,


-Original Message-
From: Maxime Coquelin 
Sent: Monday, October 18, 2021 9:02 PM
To: dev@dpdk.org; Xia, Chenbo ; Hu, Jiayu
; Wang, YuanX ; Ma,

WenwuX

; Richardson, Bruce
; Mcnamara, John
; david.march...@redhat.com
Cc: Maxime Coquelin 
Subject: [PATCH v1 08/14] vhost: improve IO vector logic

IO vectors and their iterators arrays were part of the async metadata
but not their indexes.

In order to makes this more consistent, the patch adds the indexes to
the async metadata. Doing that, we can avoid triggering DMA transfer
within the loop as it IO vector index overflow is now prevented in
the
async_mbuf_to_desc() function.

Note that previous detection mechanism was broken since the overflow
already happened when detected, so OOB memory access would already
have happened.

With this changes done, virtio_dev_rx_async_submit_split()
and virtio_dev_rx_async_submit_packed() can be further simplified.

Signed-off-by: Maxime Coquelin 
---
   lib/vhost/vhost.h  |   2 +
   lib/vhost/virtio_net.c | 291 ++---
   2 files changed, 131 insertions(+), 162 deletions(-)

diff --git a/lib/vhost/vhost.h b/lib/vhost/vhost.h index
dae9a1ac2d..812d4c55a5 100644
--- a/lib/vhost/vhost.h
+++ b/lib/vhost/vhost.h
@@ -134,6 +134,8 @@ struct vhost_async {

struct rte_vhost_iov_iter iov_iter[VHOST_MAX_ASYNC_IT];
struct rte_vhost_iovec iovec[VHOST_MAX_ASYNC_VEC];
+   uint16_t iter_idx;
+   uint16_t iovec_idx;

/* data transfer status */
struct async_inflight_info *pkts_info; diff --git
a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c index
ae7dded979..c80823a8de 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -924,33 +924,86 @@ copy_mbuf_to_desc(struct virtio_net *dev,
struct vhost_virtqueue *vq,
return error;
   }

+static __rte_always_inline int
+async_iter_initialize(struct vhost_async *async) {
+   struct rte_vhost_iov_iter *iter;
+
+   if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
+   VHOST_LOG_DATA(ERR, "no more async iovec available\n");
+   return -1;
+   }
+
+   iter = async->iov_iter + async->iter_idx;
+   iter->iov = async->iovec + async->iovec_idx;
+   iter->nr_segs = 0;
+
+   return 0;
+}
+
+static __rte_always_inline int
+async_iter_add_iovec(struct vhost_async *async, void *src, void
+*dst, size_t len) {
+   struct rte_vhost_iov_iter *iter;
+   struct rte_vhost_iovec *iovec;
+
+   if (unlikely(async->iovec_idx >= VHOST_MAX_ASYNC_VEC)) {
+   VHOST_LOG_DATA(ERR, "no more async iovec available\n");
+   return -1;
+   }


For large packets, like 64KB in iperf test, async_iter_add_iovec()
frequently reports the log above, as we run out of iovecs. I think
it's better to change the log from ERR to DEBUG.


I think it is better to keep it as an error, we want to see it if it happens
without having the user to enable debug.

But maybe we can only print it once, not to flood the logs.


OK.




In addition, the size of iovec is too small. For burst 32 and 64KB
pkts, it's easy to run out of iovecs and we will drop the pkts to
enqueue if it happens, which hurts performance. Enlarging the array is
a choice to mitigate the issue, but another solution is to reallocate
iovec once we run out of it. How do you think?


I would prefer we enlarge the array, reallocating the array when the issue
happens sounds like over-engineering to me.

Any idea what size it should be based on your experiments?


2048 is enough for iperf and 64KB pkts.


Thanks for the insight, I will change to 2048 in next revision.

Maxime



Thanks,
Jiayu


Thanks,
Maxime


Thanks,
Jiayu

+
+   iter = async->iov_iter + async->iter_idx;
+   iovec = async->iovec + async->iovec_idx;
+
+   iovec->src_addr = src;
+   iovec->dst_addr = dst;
+   iovec->len = len;
+
+   iter->nr_segs++;
+   async->iovec_idx++;
+
+   return 0;
+}








[dpdk-dev] [PATCH 1/2] eventdev/eth_rx: add queue stats get API

2021-10-26 Thread Naga Harish K S V
This patch adds new api ``rte_event_eth_rx_adapter_queue_stats_get``
to retrieve queue stats. The queue stats are in the format
``struct rte_event_eth_rx_adapter_queue_stats``.

The adapter stats_get and stats_reset apis are also updated to
handle queue level event buffer use case.

Signed-off-by: Naga Harish K S V 
---
 lib/eventdev/rte_event_eth_rx_adapter.c | 195 ++--
 lib/eventdev/rte_event_eth_rx_adapter.h |  43 ++
 lib/eventdev/version.map|   1 +
 3 files changed, 196 insertions(+), 43 deletions(-)

diff --git a/lib/eventdev/rte_event_eth_rx_adapter.c 
b/lib/eventdev/rte_event_eth_rx_adapter.c
index a175c61551..1d63252c95 100644
--- a/lib/eventdev/rte_event_eth_rx_adapter.c
+++ b/lib/eventdev/rte_event_eth_rx_adapter.c
@@ -245,6 +245,7 @@ struct eth_rx_queue_info {
uint64_t event;
struct eth_rx_vector_data vector_data;
struct eth_event_enqueue_buffer *event_buf;
+   struct rte_event_eth_rx_adapter_stats *stats;
 };
 
 static struct event_eth_rx_adapter **event_eth_rx_adapter;
@@ -268,14 +269,18 @@ rxa_validate_id(uint8_t id)
 
 static inline struct eth_event_enqueue_buffer *
 rxa_event_buf_get(struct event_eth_rx_adapter *rx_adapter, uint16_t eth_dev_id,
- uint16_t rx_queue_id)
+ uint16_t rx_queue_id,
+ struct rte_event_eth_rx_adapter_stats **stats)
 {
if (rx_adapter->use_queue_event_buf) {
struct eth_device_info *dev_info =
&rx_adapter->eth_devices[eth_dev_id];
+   *stats = dev_info->rx_queue[rx_queue_id].stats;
return dev_info->rx_queue[rx_queue_id].event_buf;
-   } else
+   } else {
+   *stats = &rx_adapter->stats;
return &rx_adapter->event_enqueue_buffer;
+   }
 }
 
 #define RTE_EVENT_ETH_RX_ADAPTER_ID_VALID_OR_ERR_RET(id, retval) do { \
@@ -766,9 +771,9 @@ rxa_enq_block_end_ts(struct event_eth_rx_adapter 
*rx_adapter,
 /* Enqueue buffered events to event device */
 static inline uint16_t
 rxa_flush_event_buffer(struct event_eth_rx_adapter *rx_adapter,
-  struct eth_event_enqueue_buffer *buf)
+  struct eth_event_enqueue_buffer *buf,
+  struct rte_event_eth_rx_adapter_stats *stats)
 {
-   struct rte_event_eth_rx_adapter_stats *stats = &rx_adapter->stats;
uint16_t count = buf->last ? buf->last - buf->head : buf->count;
 
if (!count)
@@ -883,7 +888,8 @@ rxa_create_event_vector(struct event_eth_rx_adapter 
*rx_adapter,
 static inline void
 rxa_buffer_mbufs(struct event_eth_rx_adapter *rx_adapter, uint16_t eth_dev_id,
 uint16_t rx_queue_id, struct rte_mbuf **mbufs, uint16_t num,
-struct eth_event_enqueue_buffer *buf)
+struct eth_event_enqueue_buffer *buf,
+struct rte_event_eth_rx_adapter_stats *stats)
 {
uint32_t i;
struct eth_device_info *dev_info =
@@ -954,7 +960,7 @@ rxa_buffer_mbufs(struct event_eth_rx_adapter *rx_adapter, 
uint16_t eth_dev_id,
else
num = nb_cb;
if (dropped)
-   rx_adapter->stats.rx_dropped += dropped;
+   stats->rx_dropped += dropped;
}
 
buf->count += num;
@@ -985,11 +991,10 @@ rxa_pkt_buf_available(struct eth_event_enqueue_buffer 
*buf)
 static inline uint32_t
 rxa_eth_rx(struct event_eth_rx_adapter *rx_adapter, uint16_t port_id,
   uint16_t queue_id, uint32_t rx_count, uint32_t max_rx,
-  int *rxq_empty, struct eth_event_enqueue_buffer *buf)
+  int *rxq_empty, struct eth_event_enqueue_buffer *buf,
+  struct rte_event_eth_rx_adapter_stats *stats)
 {
struct rte_mbuf *mbufs[BATCH_SIZE];
-   struct rte_event_eth_rx_adapter_stats *stats =
-   &rx_adapter->stats;
uint16_t n;
uint32_t nb_rx = 0;
 
@@ -1000,7 +1005,7 @@ rxa_eth_rx(struct event_eth_rx_adapter *rx_adapter, 
uint16_t port_id,
 */
while (rxa_pkt_buf_available(buf)) {
if (buf->count >= BATCH_SIZE)
-   rxa_flush_event_buffer(rx_adapter, buf);
+   rxa_flush_event_buffer(rx_adapter, buf, stats);
 
stats->rx_poll_count++;
n = rte_eth_rx_burst(port_id, queue_id, mbufs, BATCH_SIZE);
@@ -1009,14 +1014,17 @@ rxa_eth_rx(struct event_eth_rx_adapter *rx_adapter, 
uint16_t port_id,
*rxq_empty = 1;
break;
}
-   rxa_buffer_mbufs(rx_adapter, port_id, queue_id, mbufs, n, buf);
+   rxa_buffer_mbufs(rx_adapter, port_id, queue_id, mbufs, n, buf,
+stats);
nb_rx += n;
if (rx_count + nb_rx > max_rx)
break;
}
 
if (buf->count > 0)
-   rxa_flush_ev

[dpdk-dev] [PATCH 2/2] eventdev/eth_rx: support telemetry

2021-10-26 Thread Naga Harish K S V
Added telemetry support for rxa_queue_stats to get rx queue
stats

Signed-off-by: Naga Harish K S V 
---
 lib/eventdev/rte_event_eth_rx_adapter.c | 67 +
 1 file changed, 67 insertions(+)

diff --git a/lib/eventdev/rte_event_eth_rx_adapter.c 
b/lib/eventdev/rte_event_eth_rx_adapter.c
index 1d63252c95..2e0b316213 100644
--- a/lib/eventdev/rte_event_eth_rx_adapter.c
+++ b/lib/eventdev/rte_event_eth_rx_adapter.c
@@ -3272,6 +3272,69 @@ handle_rxa_get_queue_conf(const char *cmd __rte_unused,
return 0;
 }
 
+static int
+handle_rxa_get_queue_stats(const char *cmd __rte_unused,
+  const char *params,
+  struct rte_tel_data *d)
+{
+   uint8_t rx_adapter_id;
+   uint16_t rx_queue_id;
+   int eth_dev_id;
+   char *token, *l_params;
+   struct rte_event_eth_rx_adapter_queue_stats q_stats;
+
+   if (params == NULL || strlen(params) == 0 || !isdigit(*params))
+   return -1;
+
+   /* Get Rx adapter ID from parameter string */
+   l_params = strdup(params);
+   token = strtok(l_params, ",");
+   rx_adapter_id = strtoul(token, NULL, 10);
+   RTE_EVENT_ETH_RX_ADAPTER_ID_VALID_OR_ERR_RET(rx_adapter_id, -EINVAL);
+
+   token = strtok(NULL, ",");
+   if (token == NULL || strlen(token) == 0 || !isdigit(*token))
+   return -1;
+
+   /* Get device ID from parameter string */
+   eth_dev_id = strtoul(token, NULL, 10);
+   RTE_EVENTDEV_VALID_DEVID_OR_ERR_RET(eth_dev_id, -EINVAL);
+
+   token = strtok(NULL, ",");
+   if (token == NULL || strlen(token) == 0 || !isdigit(*token))
+   return -1;
+
+   /* Get Rx queue ID from parameter string */
+   rx_queue_id = strtoul(token, NULL, 10);
+   if (rx_queue_id >= rte_eth_devices[eth_dev_id].data->nb_rx_queues) {
+   RTE_EDEV_LOG_ERR("Invalid rx queue_id %u", rx_queue_id);
+   return -EINVAL;
+   }
+
+   token = strtok(NULL, "\0");
+   if (token != NULL)
+   RTE_EDEV_LOG_ERR("Extra parameters passed to eventdev"
+" telemetry command, igrnoring");
+
+   if (rte_event_eth_rx_adapter_queue_stats_get(rx_adapter_id, eth_dev_id,
+   rx_queue_id, &q_stats)) {
+   RTE_EDEV_LOG_ERR("Failed to get Rx adapter queue stats");
+   return -1;
+   }
+
+   rte_tel_data_start_dict(d);
+   rte_tel_data_add_dict_u64(d, "rx_adapter_id", rx_adapter_id);
+   rte_tel_data_add_dict_u64(d, "eth_dev_id", eth_dev_id);
+   rte_tel_data_add_dict_u64(d, "rx_queue_id", rx_queue_id);
+   RXA_ADD_DICT(q_stats, rx_event_buf_count);
+   RXA_ADD_DICT(q_stats, rx_event_buf_size);
+   RXA_ADD_DICT(q_stats, rx_poll_count);
+   RXA_ADD_DICT(q_stats, rx_packets);
+   RXA_ADD_DICT(q_stats, rx_dropped);
+
+   return 0;
+}
+
 RTE_INIT(rxa_init_telemetry)
 {
rte_telemetry_register_cmd("/eventdev/rxa_stats",
@@ -3285,4 +3348,8 @@ RTE_INIT(rxa_init_telemetry)
rte_telemetry_register_cmd("/eventdev/rxa_queue_conf",
handle_rxa_get_queue_conf,
"Returns Rx queue config. Parameter: rxa_id, dev_id, queue_id");
+
+   rte_telemetry_register_cmd("/eventdev/rxa_queue_stats",
+   handle_rxa_get_queue_stats,
+   "Returns Rx queue stats. Parameter: rxa_id, dev_id, queue_id");
 }
-- 
2.25.1



Re: [dpdk-dev] [PATCH v2] doc: propose correction rte_{bsf, fls} inline functions type use

2021-10-26 Thread Morten Brørup
> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Thomas Monjalon
> Sent: Monday, 25 October 2021 21.14
> 
> 15/03/2021 20:34, Tyler Retzlaff:
> > The proposal has resulted from request to review [1] the following
> > functions where there appeared to be inconsistency in return type
> > or parameter type selections for the following inline functions.
> >
> > rte_bsf32()
> > rte_bsf32_safe()
> > rte_bsf64()
> > rte_bsf64_safe()
> > rte_fls_u32()
> > rte_fls_u64()
> > rte_log2_u32()
> > rte_log2_u64()
> >
> > [1] http://mails.dpdk.org/archives/dev/2021-March/201590.html
> >
> > Signed-off-by: Tyler Retzlaff 
> > ---
> > --- a/doc/guides/rel_notes/deprecation.rst
> > +++ b/doc/guides/rel_notes/deprecation.rst
> > +* eal: Fix inline function return and parameter types for
> rte_{bsf,fls}
> > +  inline functions to be consistent.
> > +  Change ``rte_bsf32_safe`` parameter ``v`` from ``uint64_t`` to
> ``uint32_t``.
> > +  Change ``rte_bsf64`` return type to  ``uint32_t`` instead of
> ``int``.
> > +  Change ``rte_fls_u32`` return type to ``uint32_t`` instead of
> ``int``.
> > +  Change ``rte_fls_u64`` return type to ``uint32_t`` instead of
> ``int``.
> 
> It seems we completely forgot this.
> How critical is it?

Not updating has near zero effect on bug probability: Incorrectly returning 
signed int instead of unsigned int is extremely unlikely to cause problems.

Updating has near zero performance improvement: The unnecessary expansion of a 
parameter value from 32 to 64 bits is cheap.

The update's only tangible benefit is API consistency. :-)

-Morten


[dpdk-dev] [PATCH] net/mlx5: enhancement for flow dump value

2021-10-26 Thread Haifei Luo
Multiple rules could use the same encap_decap/modify_hdr/counter action.
The flow dump data could be duplicated.

To avoid redundancy, flow dump value is based on the actions' pointer
instead of previous rules' pointer.

For counter, the data is stored in cmng of priv->sh.
For encap_decap/modify_hdr, the data stored in encaps_decaps/modify_cmds.
Traverse the fields and get action's pointer and information.

Formats are same for information in the dump except "id" stands for
actions' pointer:
Counter: rec_type,id,hits,bytes
Modify_hdr:  rec_type,id,actions_number,actions
Encap_decap: rec_type,id,buf

Signed-off-by: Haifei Luo 
Acked-by: Viacheslav Ovsiienko 
---
 drivers/net/mlx5/mlx5.h |   2 +-
 drivers/net/mlx5/mlx5_flow.c| 161 ++--
 drivers/net/mlx5/mlx5_flow.h|   5 ++
 drivers/net/mlx5/mlx5_flow_dv.c |  44 ++-
 4 files changed, 188 insertions(+), 24 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index 5da5cea..5fad077 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -1685,7 +1685,7 @@ int mlx5_counter_query(struct rte_eth_dev *dev, uint32_t 
cnt,
 int mlx5_flow_dev_dump(struct rte_eth_dev *dev, struct rte_flow *flow,
FILE *file, struct rte_flow_error *error);
 int save_dump_file(const unsigned char *data, uint32_t size,
-   uint32_t type, uint32_t id, void *arg, FILE *file);
+   uint32_t type, uint64_t id, void *arg, FILE *file);
 int mlx5_flow_query_counter(struct rte_eth_dev *dev, struct rte_flow *flow,
struct rte_flow_query_count *count, struct rte_flow_error *error);
 #ifdef HAVE_IBV_FLOW_DV_SUPPORT
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 5d19ef1..e746e20 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -8173,7 +8173,7 @@ struct mlx5_flow_workspace*
 
 int
 save_dump_file(const uint8_t *data, uint32_t size,
-   uint32_t type, uint32_t id, void *arg, FILE *file)
+   uint32_t type, uint64_t id, void *arg, FILE *file)
 {
char line[BUF_SIZE];
uint32_t out = 0;
@@ -8185,17 +8185,18 @@ struct mlx5_flow_workspace*
switch (type) {
case DR_DUMP_REC_TYPE_PMD_MODIFY_HDR:
actions_num = *(uint32_t *)(arg);
-   out += snprintf(line + out, BUF_SIZE - out, "%d,0x%x,%d,",
+   out += snprintf(line + out, BUF_SIZE - out, "%d,0x%" PRIx64 
",%d,",
type, id, actions_num);
break;
case DR_DUMP_REC_TYPE_PMD_PKT_REFORMAT:
-   out += snprintf(line + out, BUF_SIZE - out, "%d,0x%x,",
+   out += snprintf(line + out, BUF_SIZE - out, "%d,0x%" PRIx64 ",",
type, id);
break;
case DR_DUMP_REC_TYPE_PMD_COUNTER:
count = (struct rte_flow_query_count *)arg;
-   fprintf(file, "%d,0x%x,%" PRIu64 ",%" PRIu64 "\n", type,
-   id, count->hits, count->bytes);
+   fprintf(file,
+   "%d,0x%" PRIx64 ",%" PRIu64 ",%" PRIu64 "\n",
+   type, id, count->hits, count->bytes);
return 0;
default:
return -1;
@@ -8269,30 +8270,34 @@ struct mlx5_flow_workspace*
uint32_t actions_num;
const uint8_t *data;
size_t size;
-   uint32_t id;
+   uint64_t id;
uint32_t type;
+   void *action = NULL;
 
if (!flow) {
return rte_flow_error_set(error, ENOENT,
-   RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
-   NULL,
-   "invalid flow handle");
+   RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
+   NULL,
+   "invalid flow handle");
}
handle_idx = flow->dev_handles;
while (handle_idx) {
dh = mlx5_ipool_get(priv->sh->ipool
-   [MLX5_IPOOL_MLX5_FLOW], handle_idx);
+   [MLX5_IPOOL_MLX5_FLOW], handle_idx);
if (!dh)
continue;
handle_idx = dh->next.next;
-   id = (uint32_t)(uintptr_t)dh->drv_flow;
 
/* query counter */
type = DR_DUMP_REC_TYPE_PMD_COUNTER;
-   if (!mlx5_flow_query_counter(dev, flow, &count, error))
-   save_dump_file(NULL, 0, type,
-   id, (void *)&count, file);
-
+   flow_dv_query_count_ptr(dev, flow->counter,
+   &action, error);
+   if (action) {
+   id = (uint64_t)(uintptr_t)action;
+   if (!mlx5_flow_query_counter(dev, flow, &count, error))
+   save_dump_file(NULL, 0, type,
+  

[dpdk-dev] [PATCH v5 0/5] add new definitions for wait scheme

2021-10-26 Thread Feifei Wang
Add new definitions for wait scheme, and apply this new definitions into
lib to replace rte_pause.

v2:
1. use macro to create new wait scheme (Stephen)

v3:
1. delete unnecessary bug fix in bpf (Konstantin)

v4:
1. put size into the macro body (Konstantin)
2. replace assert with BUILD_BUG_ON (Stephen)
3. delete unnecessary compiler barrier for bpf (Konstantin)

v5:
1. 'size' is not the parameter (Konstantin)
2. put () around macro parameters (Konstantin)
3. fix some original typo issue (Jerin)
4. swap 'rte_wait_event' parameter location (Jerin)
4. add new macro '__LOAD_EXC'
5. delete 'undef' to prevent compilation warning
 
Feifei Wang (5):
  eal: add new definitions for wait scheme
  eal: use wait event for read pflock
  eal: use wait event scheme for mcslock
  lib/bpf: use wait event scheme for Rx/Tx iteration
  lib/distributor: use wait event scheme

 lib/bpf/bpf_pkt.c|  11 +-
 lib/distributor/rte_distributor_single.c |  10 +-
 lib/eal/arm/include/rte_pause_64.h   | 135 +--
 lib/eal/include/generic/rte_mcslock.h|   9 +-
 lib/eal/include/generic/rte_pause.h  |  27 +
 lib/eal/include/generic/rte_pflock.h |   4 +-
 6 files changed, 121 insertions(+), 75 deletions(-)

-- 
2.25.1



[dpdk-dev] [PATCH v5 1/5] eal: add new definitions for wait scheme

2021-10-26 Thread Feifei Wang
Introduce macros as generic interface for address monitoring.
For different size, encapsulate '__LOAD_EXC_16', '__LOAD_EXC_32'
and '__LOAD_EXC_64' into a new macro '__LOAD_EXC'.

Furthermore, to prevent compilation warning in arm:
--
'warning: implicit declaration of function ...'
--
Delete 'undef' constructions for '__LOAD_EXC_xx', '__SEVL' and '__WFE'.

This is because original macros are undefine at the end of the file.
If new macro 'rte_wait_event' calls them in other files, they will be
seen as 'not defined'.

Signed-off-by: Feifei Wang 
Reviewed-by: Ruifeng Wang 
---
 lib/eal/arm/include/rte_pause_64.h  | 135 
 lib/eal/include/generic/rte_pause.h |  27 ++
 2 files changed, 105 insertions(+), 57 deletions(-)

diff --git a/lib/eal/arm/include/rte_pause_64.h 
b/lib/eal/arm/include/rte_pause_64.h
index e87d10b8cc..1fea0dec63 100644
--- a/lib/eal/arm/include/rte_pause_64.h
+++ b/lib/eal/arm/include/rte_pause_64.h
@@ -31,20 +31,12 @@ static inline void rte_pause(void)
 /* Put processor into low power WFE(Wait For Event) state. */
 #define __WFE() { asm volatile("wfe" : : : "memory"); }
 
-static __rte_always_inline void
-rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t expected,
-   int memorder)
-{
-   uint16_t value;
-
-   assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED);
-
-   /*
-* Atomic exclusive load from addr, it returns the 16-bit content of
-* *addr while making it 'monitored',when it is written by someone
-* else, the 'monitored' state is cleared and a event is generated
-* implicitly to exit WFE.
-*/
+/*
+ * Atomic exclusive load from addr, it returns the 16-bit content of
+ * *addr while making it 'monitored', when it is written by someone
+ * else, the 'monitored' state is cleared and an event is generated
+ * implicitly to exit WFE.
+ */
 #define __LOAD_EXC_16(src, dst, memorder) {   \
if (memorder == __ATOMIC_RELAXED) {   \
asm volatile("ldxrh %w[tmp], [%x[addr]]"  \
@@ -58,6 +50,62 @@ rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t 
expected,
: "memory");  \
} }
 
+/*
+ * Atomic exclusive load from addr, it returns the 32-bit content of
+ * *addr while making it 'monitored', when it is written by someone
+ * else, the 'monitored' state is cleared and an event is generated
+ * implicitly to exit WFE.
+ */
+#define __LOAD_EXC_32(src, dst, memorder) {  \
+   if (memorder == __ATOMIC_RELAXED) {  \
+   asm volatile("ldxr %w[tmp], [%x[addr]]"  \
+   : [tmp] "=&r" (dst)  \
+   : [addr] "r"(src)\
+   : "memory"); \
+   } else { \
+   asm volatile("ldaxr %w[tmp], [%x[addr]]" \
+   : [tmp] "=&r" (dst)  \
+   : [addr] "r"(src)\
+   : "memory"); \
+   } }
+
+/*
+ * Atomic exclusive load from addr, it returns the 64-bit content of
+ * *addr while making it 'monitored', when it is written by someone
+ * else, the 'monitored' state is cleared and an event is generated
+ * implicitly to exit WFE.
+ */
+#define __LOAD_EXC_64(src, dst, memorder) {  \
+   if (memorder == __ATOMIC_RELAXED) {  \
+   asm volatile("ldxr %x[tmp], [%x[addr]]"  \
+   : [tmp] "=&r" (dst)  \
+   : [addr] "r"(src)\
+   : "memory"); \
+   } else { \
+   asm volatile("ldaxr %x[tmp], [%x[addr]]" \
+   : [tmp] "=&r" (dst)  \
+   : [addr] "r"(src)\
+   : "memory"); \
+   } }
+
+#define __LOAD_EXC(src, dst, memorder, size) {  \
+   assert(size == 16 || size == 32 || size == 64); \
+   if (size == 16) \
+   __LOAD_EXC_16(src, dst, memorder)   \
+   else if (size == 32)\
+   __LOAD_EXC_32(src, dst, memorder)   \
+   else if (size == 64)\
+   __LOAD_EXC_64(src, dst, memorder)   \
+}
+
+static __rte_always_inline void
+rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t expected,
+   int memorder)
+{
+   uint16_t value;
+
+   assert(memorder == __ATOMIC_ACQUIRE || memorder == __ATOMIC_RELAXED);
+
__LOAD_EXC_16(addr, value, memorder)
if (value != expected) {
__SEVL()
@@ -66,7 +114,6 @@ rte_wait_

[dpdk-dev] [PATCH v5 2/5] eal: use wait event for read pflock

2021-10-26 Thread Feifei Wang
Instead of polling for read pflock update, use wait event scheme for
this case.

Signed-off-by: Feifei Wang 
Reviewed-by: Ruifeng Wang 
---
 lib/eal/include/generic/rte_pflock.h | 4 +---
 1 file changed, 1 insertion(+), 3 deletions(-)

diff --git a/lib/eal/include/generic/rte_pflock.h 
b/lib/eal/include/generic/rte_pflock.h
index e57c179ef2..7573b036bf 100644
--- a/lib/eal/include/generic/rte_pflock.h
+++ b/lib/eal/include/generic/rte_pflock.h
@@ -121,9 +121,7 @@ rte_pflock_read_lock(rte_pflock_t *pf)
return;
 
/* Wait for current write phase to complete. */
-   while ((__atomic_load_n(&pf->rd.in, __ATOMIC_ACQUIRE)
-   & RTE_PFLOCK_WBITS) == w)
-   rte_pause();
+   rte_wait_event(&pf->rd.in, RTE_PFLOCK_WBITS, ==, w, __ATOMIC_ACQUIRE);
 }
 
 /**
-- 
2.25.1



[dpdk-dev] [PATCH v5 3/5] eal: use wait event scheme for mcslock

2021-10-26 Thread Feifei Wang
Instead of polling for mcslock to be updated, use wait event scheme
for this case.

Signed-off-by: Feifei Wang 
Reviewed-by: Ruifeng Wang 
---
 lib/eal/include/generic/rte_mcslock.h | 9 +++--
 1 file changed, 7 insertions(+), 2 deletions(-)

diff --git a/lib/eal/include/generic/rte_mcslock.h 
b/lib/eal/include/generic/rte_mcslock.h
index 34f33c64a5..806a2b2c7e 100644
--- a/lib/eal/include/generic/rte_mcslock.h
+++ b/lib/eal/include/generic/rte_mcslock.h
@@ -116,8 +116,13 @@ rte_mcslock_unlock(rte_mcslock_t **msl, rte_mcslock_t *me)
/* More nodes added to the queue by other CPUs.
 * Wait until the next pointer is set.
 */
-   while (__atomic_load_n(&me->next, __ATOMIC_RELAXED) == NULL)
-   rte_pause();
+#ifdef RTE_ARCH_32
+   rte_wait_event((uint32_t *)&me->next, UINT32_MAX, ==, 0,
+   __ATOMIC_RELAXED);
+#else
+   rte_wait_event((uint64_t *)&me->next, UINT64_MAX, ==, 0,
+   __ATOMIC_RELAXED);
+#endif
}
 
/* Pass lock to next waiter. */
-- 
2.25.1



[dpdk-dev] [PATCH v5 4/5] lib/bpf: use wait event scheme for Rx/Tx iteration

2021-10-26 Thread Feifei Wang
Instead of polling for cbi->use to be updated, use wait event scheme.

Furthermore, delete 'const' for 'bpf_eth_cbi_wait'. This is because of
a compilation error:
---
../lib/eal/include/rte_common.h:36:13: error: read-only variable ‘value’
used as ‘asm’ output
   36 | #define asm __asm__
  | ^~~

../lib/eal/arm/include/rte_pause_64.h:66:3: note: in expansion of macro
‘asm’
   66 |   asm volatile("ldaxr %w[tmp], [%x[addr]]" \
  |   ^~~

../lib/eal/arm/include/rte_pause_64.h:96:3: note: in expansion of macro
‘__LOAD_EXC_32’
   96 |   __LOAD_EXC_32((src), dst, memorder) \
  |   ^

../lib/eal/arm/include/rte_pause_64.h:167:4: note: in expansion of macro
‘__LOAD_EXC’
  167 |__LOAD_EXC((addr), value, memorder, size) \
  |^~

../lib/bpf/bpf_pkt.c:125:3: note: in expansion of macro ‘rte_wait_event’
  125 |   rte_wait_event(&cbi->use, UINT32_MAX, ==, puse,
---

Signed-off-by: Feifei Wang 
Reviewed-by: Ruifeng Wang 
---
 lib/bpf/bpf_pkt.c | 11 ---
 1 file changed, 4 insertions(+), 7 deletions(-)

diff --git a/lib/bpf/bpf_pkt.c b/lib/bpf/bpf_pkt.c
index 6e8248f0d6..213d44a75a 100644
--- a/lib/bpf/bpf_pkt.c
+++ b/lib/bpf/bpf_pkt.c
@@ -111,9 +111,9 @@ bpf_eth_cbi_unuse(struct bpf_eth_cbi *cbi)
  * Waits till datapath finished using given callback.
  */
 static void
-bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
+bpf_eth_cbi_wait(struct bpf_eth_cbi *cbi)
 {
-   uint32_t nuse, puse;
+   uint32_t puse;
 
/* make sure all previous loads and stores are completed */
rte_smp_mb();
@@ -122,11 +122,8 @@ bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
 
/* in use, busy wait till current RX/TX iteration is finished */
if ((puse & BPF_ETH_CBI_INUSE) != 0) {
-   do {
-   rte_pause();
-   rte_compiler_barrier();
-   nuse = cbi->use;
-   } while (nuse == puse);
+   rte_wait_event(&cbi->use, UINT32_MAX, ==, puse,
+   __ATOMIC_RELAXED);
}
 }
 
-- 
2.25.1



[dpdk-dev] [PATCH v5 5/5] lib/distributor: use wait event scheme

2021-10-26 Thread Feifei Wang
Instead of polling for bufptr64 to be updated, use
wait event for this case.

Signed-off-by: Feifei Wang 
Reviewed-by: Ruifeng Wang 
---
 lib/distributor/rte_distributor_single.c | 10 --
 1 file changed, 4 insertions(+), 6 deletions(-)

diff --git a/lib/distributor/rte_distributor_single.c 
b/lib/distributor/rte_distributor_single.c
index f4725b1d0b..d52b24a453 100644
--- a/lib/distributor/rte_distributor_single.c
+++ b/lib/distributor/rte_distributor_single.c
@@ -33,9 +33,8 @@ rte_distributor_request_pkt_single(struct 
rte_distributor_single *d,
union rte_distributor_buffer_single *buf = &d->bufs[worker_id];
int64_t req = (((int64_t)(uintptr_t)oldpkt) << RTE_DISTRIB_FLAG_BITS)
| RTE_DISTRIB_GET_BUF;
-   while (unlikely(__atomic_load_n(&buf->bufptr64, __ATOMIC_RELAXED)
-   & RTE_DISTRIB_FLAGS_MASK))
-   rte_pause();
+   rte_wait_event(&buf->bufptr64, RTE_DISTRIB_FLAGS_MASK,
+   !=, 0, __ATOMIC_RELAXED);
 
/* Sync with distributor on GET_BUF flag. */
__atomic_store_n(&(buf->bufptr64), req, __ATOMIC_RELEASE);
@@ -74,9 +73,8 @@ rte_distributor_return_pkt_single(struct 
rte_distributor_single *d,
union rte_distributor_buffer_single *buf = &d->bufs[worker_id];
uint64_t req = (((int64_t)(uintptr_t)oldpkt) << RTE_DISTRIB_FLAG_BITS)
| RTE_DISTRIB_RETURN_BUF;
-   while (unlikely(__atomic_load_n(&buf->bufptr64, __ATOMIC_RELAXED)
-   & RTE_DISTRIB_FLAGS_MASK))
-   rte_pause();
+   rte_wait_event(&buf->bufptr64, RTE_DISTRIB_FLAGS_MASK,
+   !=, 0, __ATOMIC_RELAXED);
 
/* Sync with distributor on RETURN_BUF flag. */
__atomic_store_n(&(buf->bufptr64), req, __ATOMIC_RELEASE);
-- 
2.25.1



[dpdk-dev] 回复: [PATCH v5 1/5] eal: add new definitions for wait scheme

2021-10-26 Thread Feifei Wang
> -邮件原件-
> 发件人: Feifei Wang 
> 发送时间: Tuesday, October 26, 2021 4:02 PM
> 收件人: Ruifeng Wang 
> 抄送: dev@dpdk.org; nd ; Feifei Wang
> 
> 主题: [PATCH v5 1/5] eal: add new definitions for wait scheme
> 
> Introduce macros as generic interface for address monitoring.
> For different size, encapsulate '__LOAD_EXC_16', '__LOAD_EXC_32'
> and '__LOAD_EXC_64' into a new macro '__LOAD_EXC'.
> 
> Furthermore, to prevent compilation warning in arm:
> --
> 'warning: implicit declaration of function ...'
> --
> Delete 'undef' constructions for '__LOAD_EXC_xx', '__SEVL' and '__WFE'.
> 
> This is because original macros are undefine at the end of the file.
> If new macro 'rte_wait_event' calls them in other files, they will be seen as
> 'not defined'.
> 
> Signed-off-by: Feifei Wang 
> Reviewed-by: Ruifeng Wang 
> ---
>  lib/eal/arm/include/rte_pause_64.h  | 135 
> lib/eal/include/generic/rte_pause.h |  27 ++
>  2 files changed, 105 insertions(+), 57 deletions(-)
> 
> diff --git a/lib/eal/arm/include/rte_pause_64.h
> b/lib/eal/arm/include/rte_pause_64.h
> index e87d10b8cc..1fea0dec63 100644
> --- a/lib/eal/arm/include/rte_pause_64.h
> +++ b/lib/eal/arm/include/rte_pause_64.h
> @@ -31,20 +31,12 @@ static inline void rte_pause(void)
>  /* Put processor into low power WFE(Wait For Event) state. */  #define
> __WFE() { asm volatile("wfe" : : : "memory"); }
> 
> -static __rte_always_inline void
> -rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t expected,
> - int memorder)
> -{
> - uint16_t value;
> -
> - assert(memorder == __ATOMIC_ACQUIRE || memorder ==
> __ATOMIC_RELAXED);
> -
> - /*
> -  * Atomic exclusive load from addr, it returns the 16-bit content of
> -  * *addr while making it 'monitored',when it is written by someone
> -  * else, the 'monitored' state is cleared and a event is generated
> -  * implicitly to exit WFE.
> -  */
> +/*
> + * Atomic exclusive load from addr, it returns the 16-bit content of
> + * *addr while making it 'monitored', when it is written by someone
> + * else, the 'monitored' state is cleared and an event is generated
> + * implicitly to exit WFE.
> + */
>  #define __LOAD_EXC_16(src, dst, memorder) {   \
>   if (memorder == __ATOMIC_RELAXED) {   \
>   asm volatile("ldxrh %w[tmp], [%x[addr]]"  \ @@ -58,6 +50,62
> @@ rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t expected,
>   : "memory");  \
>   } }
> 
> +/*
> + * Atomic exclusive load from addr, it returns the 32-bit content of
> + * *addr while making it 'monitored', when it is written by someone
> + * else, the 'monitored' state is cleared and an event is generated
> + * implicitly to exit WFE.
> + */
> +#define __LOAD_EXC_32(src, dst, memorder) {  \
> + if (memorder == __ATOMIC_RELAXED) {  \
> + asm volatile("ldxr %w[tmp], [%x[addr]]"  \
> + : [tmp] "=&r" (dst)  \
> + : [addr] "r"(src)\
> + : "memory"); \
> + } else { \
> + asm volatile("ldaxr %w[tmp], [%x[addr]]" \
> + : [tmp] "=&r" (dst)  \
> + : [addr] "r"(src)\
> + : "memory"); \
> + } }
> +
> +/*
> + * Atomic exclusive load from addr, it returns the 64-bit content of
> + * *addr while making it 'monitored', when it is written by someone
> + * else, the 'monitored' state is cleared and an event is generated
> + * implicitly to exit WFE.
> + */
> +#define __LOAD_EXC_64(src, dst, memorder) {  \
> + if (memorder == __ATOMIC_RELAXED) {  \
> + asm volatile("ldxr %x[tmp], [%x[addr]]"  \
> + : [tmp] "=&r" (dst)  \
> + : [addr] "r"(src)\
> + : "memory"); \
> + } else { \
> + asm volatile("ldaxr %x[tmp], [%x[addr]]" \
> + : [tmp] "=&r" (dst)  \
> + : [addr] "r"(src)\
> + : "memory"); \
> + } }
> +
> +#define __LOAD_EXC(src, dst, memorder, size) {  \
> + assert(size == 16 || size == 32 || size == 64); \
> + if (size == 16) \
> + __LOAD_EXC_16(src, dst, memorder)   \
> + else if (size == 32)\
> + __LOAD_EXC_32(src, dst, memorder)   \
> + else if (size == 64)\
> + __LOAD_EXC_64(src, dst, memorder)   \
> +}
> +
> +static __rte_always_inline void
> +rte_wa

[dpdk-dev] [Bug 837] [build]lib/ethdev build failed on Ubuntu20.04.3 with gcc 10.3.0.

2021-10-26 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=837

Bug ID: 837
   Summary: [build]lib/ethdev build failed on Ubuntu20.04.3 with
gcc 10.3.0.
   Product: DPDK
   Version: 21.11
  Hardware: x86
OS: Linux
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: meson
  Assignee: dev@dpdk.org
  Reporter: longfengx.li...@intel.com
  Target Milestone: ---

1.DPDK version:
Now build failed issue branch dpdk-next-net/dpdk-next-net-intel
bad commit d7a55a39bce30d3bf7e9e7de3f25b1f603b63926 (HEAD -> main)
Author: Ferruh Yigit 
Date:   Fri Oct 22 12:03:12 2021 +0100

ethdev: add namespace

Add 'RTE_ETH' namespace to all enums & macros in a backward compatible
way. The macros for backward compatibility can be removed in next LTS.
Also updated some struct names to have 'rte_eth' prefix.

All internal components switched to using new names.

Syntax fixed on lines that this patch touches.

Signed-off-by: Ferruh Yigit 
Acked-by: Tyler Retzlaff 
Acked-by: Andrew Rybchenko 
Acked-by: Ajit Khaparde 
Acked-by: Jerin Jacob 
Acked-by: Wisam Jaddo 
Acked-by: Rosen Xu 
Acked-by: Chenbo Xia 
Acked-by: Hemant Agrawal 
Acked-by: Somnath Kotur 


2.OS version:
  Ubuntu 20.04.3 LTS /kernel 5.8.0-48-generic
  gcc version:gcc (Ubuntu 10.3.0-1ubuntu1~20.04) 10.3.0

3.Test Setup:

build cmd:
CC=gcc meson --werror -Denable_kmods=True -Dlibdir=lib -Denable_docs=true
x86_64-native-linuxapp-doc
ninja -C x86_64-native-linuxapp-doc/


4.build failed output info:
root@dpdk-xuemin4x-ub2004-doc:~/dpdk-next-net# ninja -C
x86_64-native-linuxapp-doc/
ninja: Entering directory `x86_64-native-linuxapp-doc/'
[2668/2697] Generating rte_kni with a custom command.
make: Entering directory '/usr/src/linux-headers-5.8.0-48-generic'
  CC [M] 
/root/dpdk-next-net/x86_64-native-linuxapp-doc/kernel/linux/kni/kni_net.o
  CC [M] 
/root/dpdk-next-net/x86_64-native-linuxapp-doc/kernel/linux/kni/kni_misc.o
  LD [M] 
/root/dpdk-next-net/x86_64-native-linuxapp-doc/kernel/linux/kni/rte_kni.o
  MODPOST
/root/dpdk-next-net/x86_64-native-linuxapp-doc/kernel/linux/kni/Module.symvers
  CC [M] 
/root/dpdk-next-net/x86_64-native-linuxapp-doc/kernel/linux/kni/rte_kni.mod.o
  LD [M] 
/root/dpdk-next-net/x86_64-native-linuxapp-doc/kernel/linux/kni/rte_kni.ko
make: Leaving directory '/usr/src/linux-headers-5.8.0-48-generic'
[2671/2697] Generating doxygen with a custom command.
FAILED: doc/api/html
/root/dpdk-next-net/doc/api/generate_doxygen.sh doc/api/doxy-api.conf
doc/api/html /root/dpdk-next-net/doc/api/doxy-html-custom.sh
/root/dpdk-next-net/lib/ethdev/rte_ethdev.h:397: error: Member
RTE_ETH_MQ_RX_RSS_FLAG (macro definition) of file rte_ethdev.h is not
documented. (warning treated as error, aborting now)
[2676/2697] Compiling C object
'drivers/a715181@@tmp_rte_event_octeontx2@sta/event_octeontx2_otx2_worker_dual.c.o'.
ninja: build stopped: subcommand failed.

-- 
You are receiving this mail because:
You are the assignee for the bug.

[dpdk-dev] 回复: [PATCH v5 4/5] lib/bpf: use wait event scheme for Rx/Tx iteration

2021-10-26 Thread Feifei Wang


> -邮件原件-
> 发件人: Feifei Wang 
> 发送时间: Tuesday, October 26, 2021 4:02 PM
> 收件人: Konstantin Ananyev 
> 抄送: dev@dpdk.org; nd ; Feifei Wang
> ; Ruifeng Wang 
> 主题: [PATCH v5 4/5] lib/bpf: use wait event scheme for Rx/Tx iteration
> 
> Instead of polling for cbi->use to be updated, use wait event scheme.
> 
> Furthermore, delete 'const' for 'bpf_eth_cbi_wait'. This is because of a
> compilation error:
> ---
> ../lib/eal/include/rte_common.h:36:13: error: read-only variable ‘value’
> used as ‘asm’ output
>36 | #define asm __asm__
>   | ^~~
> 
> ../lib/eal/arm/include/rte_pause_64.h:66:3: note: in expansion of macro ‘asm’
>66 |   asm volatile("ldaxr %w[tmp], [%x[addr]]" \
>   |   ^~~
> 
> ../lib/eal/arm/include/rte_pause_64.h:96:3: note: in expansion of macro
> ‘__LOAD_EXC_32’
>96 |   __LOAD_EXC_32((src), dst, memorder) \
>   |   ^
> 
> ../lib/eal/arm/include/rte_pause_64.h:167:4: note: in expansion of macro
> ‘__LOAD_EXC’
>   167 |__LOAD_EXC((addr), value, memorder, size) \
>   |^~
> 
> ../lib/bpf/bpf_pkt.c:125:3: note: in expansion of macro ‘rte_wait_event’
>   125 |   rte_wait_event(&cbi->use, UINT32_MAX, ==, puse,
> ---
> 
> Signed-off-by: Feifei Wang 
> Reviewed-by: Ruifeng Wang 
> ---
>  lib/bpf/bpf_pkt.c | 11 ---
>  1 file changed, 4 insertions(+), 7 deletions(-)
> 
> diff --git a/lib/bpf/bpf_pkt.c b/lib/bpf/bpf_pkt.c index
> 6e8248f0d6..213d44a75a 100644
> --- a/lib/bpf/bpf_pkt.c
> +++ b/lib/bpf/bpf_pkt.c
> @@ -111,9 +111,9 @@ bpf_eth_cbi_unuse(struct bpf_eth_cbi *cbi)
>   * Waits till datapath finished using given callback.
>   */
>  static void
> -bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
> +bpf_eth_cbi_wait(struct bpf_eth_cbi *cbi)

Hi, Konstantin

For this bpf patch, I delete 'const' through this is contrary to what we
discussed earlier. This is because if  we keep 'constant' here and use 
'rte_wait_event'
new macro, compiler will report error. And earlier the arm version cannot be 
compiled
due to I forgot enable "wfe" config in the meson file, so this issue can not 
happen before.

>  {
> - uint32_t nuse, puse;
> + uint32_t puse;
> 
>   /* make sure all previous loads and stores are completed */
>   rte_smp_mb();
> @@ -122,11 +122,8 @@ bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
> 
>   /* in use, busy wait till current RX/TX iteration is finished */
>   if ((puse & BPF_ETH_CBI_INUSE) != 0) {
> - do {
> - rte_pause();
> - rte_compiler_barrier();
> - nuse = cbi->use;
> - } while (nuse == puse);
> + rte_wait_event(&cbi->use, UINT32_MAX, ==, puse,
> + __ATOMIC_RELAXED);
>   }
>  }
> 
> --
> 2.25.1



[dpdk-dev] [PATCH v3] net/ice: simplify the use of DCF device reset

2021-10-26 Thread dapengx . yu
From: Dapeng Yu 

After DCF is reset by PF, the DCF device un-initialization cannot
function normally since the resource is already invalidated. So
reset DCF twice is necessary, the first simplified reset
re-initializes the AdminQ of DCF, only then second reset can clean
the filters successfully.

This patch detects the reset flag, which is set by PF on DCF reset,
if the flag is true, do DCF reset twice automatically.

Fixes: 1a86f4dbdf42 ("net/ice: support DCF device reset")
Cc: sta...@dpdk.org

Signed-off-by: Dapeng Yu 
---
V2:
* Ignore the returned error of dev_uninit when DCF is reset by PF
V3:
* Add a reset function to re-initialize AdminQ resource
* Add a function to check the reset flag
---
 drivers/net/ice/ice_dcf.c|  2 ++
 drivers/net/ice/ice_dcf_ethdev.c | 32 
 2 files changed, 34 insertions(+)

diff --git a/drivers/net/ice/ice_dcf.c b/drivers/net/ice/ice_dcf.c
index 084f7a53db..366ff0a907 100644
--- a/drivers/net/ice/ice_dcf.c
+++ b/drivers/net/ice/ice_dcf.c
@@ -593,6 +593,8 @@ ice_dcf_init_hw(struct rte_eth_dev *eth_dev, struct 
ice_dcf_hw *hw)
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
int ret, size;
 
+   hw->resetting = false;
+
hw->avf.hw_addr = pci_dev->mem_resource[0].addr;
hw->avf.back = hw;
 
diff --git a/drivers/net/ice/ice_dcf_ethdev.c b/drivers/net/ice/ice_dcf_ethdev.c
index 7c71a48010..f47219ee17 100644
--- a/drivers/net/ice/ice_dcf_ethdev.c
+++ b/drivers/net/ice/ice_dcf_ethdev.c
@@ -1025,11 +1025,43 @@ ice_dcf_tm_ops_get(struct rte_eth_dev *dev __rte_unused,
return 0;
 }
 
+static inline void
+ice_dcf_simple_reset(struct rte_eth_dev *eth_dev, struct ice_dcf_hw *hw)
+{
+   ice_dcf_uninit_hw(eth_dev, hw);
+   ice_dcf_init_hw(eth_dev, hw);
+}
+
+/* Check if reset has been triggered by PF */
+static inline bool
+dcf_is_reset(struct rte_eth_dev *dev)
+{
+   struct ice_dcf_adapter *ad = dev->data->dev_private;
+   struct iavf_hw *hw = &ad->real_hw.avf;
+
+   return !(IAVF_READ_REG(hw, IAVF_VF_ARQLEN1) &
+IAVF_VF_ARQLEN1_ARQENABLE_MASK);
+}
+
 static int
 ice_dcf_dev_reset(struct rte_eth_dev *dev)
 {
+   struct ice_dcf_adapter *ad = dev->data->dev_private;
+   struct ice_dcf_hw *hw = &ad->real_hw;
int ret;
 
+   if (dcf_is_reset(dev)) {
+   if (!ad->real_hw.resetting)
+   ad->real_hw.resetting = true;
+   PMD_DRV_LOG(ERR, "The DCF has been reset by PF");
+
+   /*
+* Do the simplified reset to make DCF get AdminQ resource.
+* Then the next uninit/init can clean filters successfully.
+*/
+   ice_dcf_simple_reset(dev, hw);
+   }
+
ret = ice_dcf_dev_uninit(dev);
if (ret)
return ret;
-- 
2.27.0



Re: [dpdk-dev] [EXT] [PATCH v5] cryptodev: add telemetry callbacks

2021-10-26 Thread Troy, Rebecca
> -Original Message-
> From: Akhil Goyal 
> Sent: Monday 25 October 2021 05:32
> To: Troy, Rebecca ; dev@dpdk.org
> Cc: Power, Ciara ; Zhang, Roy Fan
> ; Doherty, Declan 
> Subject: RE: [EXT] [PATCH v5] cryptodev: add telemetry callbacks
> 
> > ---
> > v5:
> >  - Added missing telemetry dependency to meson.build.
> > v4:
> >   - Corrected doc heading underline and link.
> >   - Replaced remaining -1 return values with -EINVAL.
> 
> 
> > +
> > +static int
> > +cryptodev_handle_dev_list(const char *cmd __rte_unused,
> > +   const char *params __rte_unused,
> > +   struct rte_tel_data *d)
> > +{
> > +   int dev_id;
> > +
> > +   if (rte_cryptodev_count() < 1)
> > +   return -1;
> 
> EINVAL missed here.
> 

Good catch, thanks! Will fix now.


Re: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library

2021-10-26 Thread Liu, Yu Y
Hi Thomas,

Would you merge this patch as the series is acked by Cristian as below?
https://patchwork.dpdk.org/project/dpdk/cover/20211019081902.3514841-1-wojciechx.liguzin...@intel.com/
 

Thanks & Regards,
Yu Liu

-Original Message-
From: dev  On Behalf Of Liguzinski, WojciechX
Sent: Monday, October 25, 2021 7:32 PM
To: dev@dpdk.org; Singh, Jasvinder ; Dumitrescu, 
Cristian 
Cc: Ajmera, Megha 
Subject: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library

DPDK sched library is equipped with mechanism that secures it from the 
bufferbloat problem which is a situation when excess buffers in the network 
cause high latency and latency variation. Currently, it supports RED for active 
queue management. However, more advanced queue management is required to 
address this problem and provide desirable quality of service to users.

This solution (RFC) proposes usage of new algorithm called "PIE" (Proportional 
Integral controller Enhanced) that can effectively and directly control queuing 
latency to address the bufferbloat problem.

The implementation of mentioned functionality includes modification of existing 
and adding a new set of data structures to the library, adding PIE related APIs.
This affects structures in public API/ABI. That is why deprecation notice is 
going to be prepared and sent.

Liguzinski, WojciechX (5):
  sched: add PIE based congestion management
  example/qos_sched: add PIE support
  example/ip_pipeline: add PIE support
  doc/guides/prog_guide: added PIE
  app/test: add tests for PIE

 app/test/meson.build |4 +
 app/test/test_pie.c  | 1065 ++
 config/rte_config.h  |1 -
 doc/guides/prog_guide/glossary.rst   |3 +
 doc/guides/prog_guide/qos_framework.rst  |   64 +-
 doc/guides/prog_guide/traffic_management.rst |   13 +-
 drivers/net/softnic/rte_eth_softnic_tm.c |6 +-
 examples/ip_pipeline/tmgr.c  |  142 +--
 examples/qos_sched/cfg_file.c|  127 ++-
 examples/qos_sched/cfg_file.h|5 +
 examples/qos_sched/init.c|   27 +-
 examples/qos_sched/main.h|3 +
 examples/qos_sched/profile.cfg   |  196 ++--
 lib/sched/meson.build|3 +-
 lib/sched/rte_pie.c  |   86 ++
 lib/sched/rte_pie.h  |  398 +++
 lib/sched/rte_sched.c|  241 ++--
 lib/sched/rte_sched.h|   63 +-
 lib/sched/version.map|4 +
 19 files changed, 2172 insertions(+), 279 deletions(-)  create mode 100644 
app/test/test_pie.c  create mode 100644 lib/sched/rte_pie.c  create mode 100644 
lib/sched/rte_pie.h

--
2.25.1

Series-acked-by: Cristian Dumitrescu 


[dpdk-dev] [Bug 837] [build]lib/ethdev build failed on Ubuntu20.04.3 with gcc 10.3.0.

2021-10-26 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=837

Thomas Monjalon (ad...@dpdk.org) changed:

   What|Removed |Added

 Status|UNCONFIRMED |RESOLVED
 CC||ad...@dpdk.org
 Resolution|--- |FIXED

--- Comment #1 from Thomas Monjalon (ad...@dpdk.org) ---
It has been fixed when pulling next-net in main.

-- 
You are receiving this mail because:
You are the assignee for the bug.

Re: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library

2021-10-26 Thread Thomas Monjalon
26/10/2021 10:24, Liu, Yu Y:
> Hi Thomas,
> 
> Would you merge this patch as the series is acked by Cristian as below?
> https://patchwork.dpdk.org/project/dpdk/cover/20211019081902.3514841-1-wojciechx.liguzin...@intel.com/

I didn't see any email from Cristian.
It seems you just added this ack silently at the bottom of the cover letter.

1/ an email from Cristian is far better
2/ when integrating ack, it must be done in patches, not cover letter


> 
> Thanks & Regards,
> Yu Liu
> 
> -Original Message-
> From: dev  On Behalf Of Liguzinski, WojciechX
> Sent: Monday, October 25, 2021 7:32 PM
> To: dev@dpdk.org; Singh, Jasvinder ; Dumitrescu, 
> Cristian 
> Cc: Ajmera, Megha 
> Subject: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library
> 
> DPDK sched library is equipped with mechanism that secures it from the 
> bufferbloat problem which is a situation when excess buffers in the network 
> cause high latency and latency variation. Currently, it supports RED for 
> active queue management. However, more advanced queue management is required 
> to address this problem and provide desirable quality of service to users.
> 
> This solution (RFC) proposes usage of new algorithm called "PIE" 
> (Proportional Integral controller Enhanced) that can effectively and directly 
> control queuing latency to address the bufferbloat problem.
> 
> The implementation of mentioned functionality includes modification of 
> existing and adding a new set of data structures to the library, adding PIE 
> related APIs.
> This affects structures in public API/ABI. That is why deprecation notice is 
> going to be prepared and sent.
> 
> Liguzinski, WojciechX (5):
>   sched: add PIE based congestion management
>   example/qos_sched: add PIE support
>   example/ip_pipeline: add PIE support
>   doc/guides/prog_guide: added PIE
>   app/test: add tests for PIE
> 
>  app/test/meson.build |4 +
>  app/test/test_pie.c  | 1065 ++
>  config/rte_config.h  |1 -
>  doc/guides/prog_guide/glossary.rst   |3 +
>  doc/guides/prog_guide/qos_framework.rst  |   64 +-
>  doc/guides/prog_guide/traffic_management.rst |   13 +-
>  drivers/net/softnic/rte_eth_softnic_tm.c |6 +-
>  examples/ip_pipeline/tmgr.c  |  142 +--
>  examples/qos_sched/cfg_file.c|  127 ++-
>  examples/qos_sched/cfg_file.h|5 +
>  examples/qos_sched/init.c|   27 +-
>  examples/qos_sched/main.h|3 +
>  examples/qos_sched/profile.cfg   |  196 ++--
>  lib/sched/meson.build|3 +-
>  lib/sched/rte_pie.c  |   86 ++
>  lib/sched/rte_pie.h  |  398 +++
>  lib/sched/rte_sched.c|  241 ++--
>  lib/sched/rte_sched.h|   63 +-
>  lib/sched/version.map|4 +
>  19 files changed, 2172 insertions(+), 279 deletions(-)  create mode 100644 
> app/test/test_pie.c  create mode 100644 lib/sched/rte_pie.c  create mode 
> 100644 lib/sched/rte_pie.h
> 
> --
> 2.25.1
> 
> Series-acked-by: Cristian Dumitrescu 
> 







Re: [dpdk-dev] [PATCH 1/4] common/cnxk: add DPI DMA support

2021-10-26 Thread Jerin Jacob
On Tue, Oct 26, 2021 at 9:43 AM Radha Mohan Chintakuntla
 wrote:
>
> Add base support as ROC(Rest of Chip) API which will be used by PMD
> dmadev driver.
>
> This patch adds routines to init, fini, configure the DPI DMA device
> found in Marvell's CN9k or CN10k SoC familes.

families

>
> Signed-off-by: Radha Mohan Chintakuntla 
> ---
>  drivers/common/cnxk/hw/dpi.h   | 136 
>  drivers/common/cnxk/meson.build|   1 +
>  drivers/common/cnxk/roc_api.h  |   4 +
>  drivers/common/cnxk/roc_dpi.c  | 193 +
>  drivers/common/cnxk/roc_dpi.h  |  44 +++
>  drivers/common/cnxk/roc_dpi_priv.h |  40 ++
>  drivers/common/cnxk/roc_platform.h |   1 +
>  drivers/common/cnxk/roc_priv.h |   3 +
>  drivers/common/cnxk/version.map|   5 +
>  9 files changed, 427 insertions(+)
>  create mode 100644 drivers/common/cnxk/hw/dpi.h
>  create mode 100644 drivers/common/cnxk/roc_dpi.c
>  create mode 100644 drivers/common/cnxk/roc_dpi.h
>  create mode 100644 drivers/common/cnxk/roc_dpi_priv.h
>
> diff --git a/drivers/common/cnxk/hw/dpi.h b/drivers/common/cnxk/hw/dpi.h
> new file mode 100644
> index 00..aa1e66aa11
> --- /dev/null
> +++ b/drivers/common/cnxk/hw/dpi.h
> @@ -0,0 +1,136 @@
> +/* SPDX-License-Identifier: BSD-3-Clause
> + * Copyright(C) 2021 Marvell.
> + */
> +/**
> + * DPI device HW definitions.
> + */
> +#ifndef __DEV_DPI_HW_H__
> +#define __DEV_DPI_HW_H__
> +
> +#include 
> +
> +/** @cond __INTERNAL_DOCUMENTATION__ */

This is not required.

> +
> +/* DPI VF register offsets from VF_BAR0 */
> +#define DPI_VDMA_EN   (0x0)
> +#define DPI_VDMA_REQQ_CTL  (0x8)
> +#define DPI_VDMA_DBELL(0x10)
> +#define DPI_VDMA_SADDR(0x18)
> +#define DPI_VDMA_COUNTS   (0x20)
> +#define DPI_VDMA_NADDR(0x28)
> +#define DPI_VDMA_IWBUSY   (0x30)
> +#define DPI_VDMA_CNT  (0x38)
> +#define DPI_VF_INT(0x100)
> +#define DPI_VF_INT_W1S(0x108)
> +#define DPI_VF_INT_ENA_W1C (0x110)
> +#define DPI_VF_INT_ENA_W1S (0x118)
> +
> +/**
> + * Enumeration dpi_hdr_xtype_e
> + *
> + * DPI Transfer Type Enumeration
> + * Enumerates the pointer type in DPI_DMA_INSTR_HDR_S[XTYPE].
> + */
> +#define DPI_XTYPE_OUTBOUND (0)
> +#define DPI_XTYPE_INBOUND  (1)
> +#define DPI_XTYPE_INTERNAL_ONLY (2)
> +#define DPI_XTYPE_EXTERNAL_ONLY (3)
> +#define DPI_HDR_XTYPE_MASK 0x3
> +#define DPI_HDR_PT_MASK0x3
> +#define DPI_HDR_TT_MASK0x3
> +#define DPI_HDR_GRP_MASK   0x3FF
> +#define DPI_HDR_FUNC_MASK  0x
> +
> +/* Big endian data bit position in DMA local pointer */
> +#define DPI_LPTR_BED_BIT_POS (60)
> +
> +#define DPI_MIN_CMD_SIZE 8
> +#define DPI_MAX_CMD_SIZE 64
> +
> +/**
> + * Structure dpi_instr_hdr_s for CN9K
> + *
> + * DPI DMA Instruction Header Format
> + */
> +union dpi_instr_hdr_s {
> +   uint64_t u[4];
> +   struct dpi_dma_instr_hdr_s_s {
> +   uint64_t tag : 32;
> +   uint64_t tt : 2;
> +   uint64_t grp : 10;
> +   uint64_t reserved_44_47 : 4;
> +   uint64_t nfst : 4;
> +   uint64_t reserved_52_53 : 2;
> +   uint64_t nlst : 4;
> +   uint64_t reserved_58_63 : 6;
> +   /* Word 0 - End */
> +   uint64_t aura : 20;
> +   uint64_t func : 16;
> +   uint64_t pt : 2;
> +   uint64_t reserved_102 : 1;
> +   uint64_t pvfe : 1;
> +   uint64_t fl : 1;
> +   uint64_t ii : 1;
> +   uint64_t fi : 1;
> +   uint64_t ca : 1;
> +   uint64_t csel : 1;
> +   uint64_t reserved_109_111 : 3;
> +   uint64_t xtype : 2;
> +   uint64_t reserved_114_119 : 6;
> +   uint64_t fport : 2;
> +   uint64_t reserved_122_123 : 2;
> +   uint64_t lport : 2;
> +   uint64_t reserved_126_127 : 2;
> +   /* Word 1 - End */
> +   uint64_t ptr : 64;
> +   /* Word 2 - End */
> +   uint64_t reserved_192_255 : 64;
> +   /* Word 3 - End */
> +   } s;
> +};
> +
> +/**
> + * Structure dpi_cn10k_instr_hdr_s for CN10K
> + *
> + * DPI DMA Instruction Header Format
> + */
> +union dpi_cn10k_instr_hdr_s {
> +   uint64_t u[4];
> +   struct dpi_cn10k_dma_instr_hdr_s_s {
> +   uint64_t nfst : 4;
> +   uint64_t reserved_4_5 : 2;
> +   uint64_t nlst : 4;
> +   uint64_t reserved_10_11 : 2;
> +   uint64_t pvfe : 1;
> +   uint64_t reserved_13 : 1;
> +   uint64_t func : 16;
> +   uint64_t aura : 20;
> +   uint64_t xtype : 2;
> +   uint64_t reserved_52_53 : 2;
> +   uint64_t pt : 2;
> +   uint64_t fport : 2;
> +   uint64_t reserved_58_59 : 2;
> +   uint64_t lport : 2;
> +   uint64_t r

Re: [dpdk-dev] [PATCH 2/4] dma/cnxk: create and initialize dmadev on pci probe

2021-10-26 Thread Jerin Jacob
On Tue, Oct 26, 2021 at 9:43 AM Radha Mohan Chintakuntla
 wrote:
>
> This patch creates and initializes a dmadev device on pci probe.
>
> Signed-off-by: Radha Mohan Chintakuntla 
> ---
>  MAINTAINERS|   7 +-
>  doc/guides/dmadevs/cnxk.rst|  53 +++
>  doc/guides/dmadevs/index.rst   |   1 +
>  drivers/dma/cnxk/cnxk_dmadev.c | 119 +
>  drivers/dma/cnxk/cnxk_dmadev.h |  11 +++
>  drivers/dma/cnxk/meson.build   |   7 ++
>  drivers/dma/meson.build|   1 +
>  7 files changed, 198 insertions(+), 1 deletion(-)
>  create mode 100644 doc/guides/dmadevs/cnxk.rst
>  create mode 100644 drivers/dma/cnxk/cnxk_dmadev.c
>  create mode 100644 drivers/dma/cnxk/cnxk_dmadev.h
>  create mode 100644 drivers/dma/cnxk/meson.build
>
> diff --git a/MAINTAINERS b/MAINTAINERS
> index be2c9b6815..cdc2d98a6b 100644
> --- a/MAINTAINERS
> +++ b/MAINTAINERS
> @@ -1186,7 +1186,6 @@ F: drivers/compress/zlib/
>  F: doc/guides/compressdevs/zlib.rst
>  F: doc/guides/compressdevs/features/zlib.ini
>
> -

Unwanted line deletion.

>  DMAdev Drivers
>  --
>
> @@ -1202,6 +1201,12 @@ M: Conor Walsh 
>  F: drivers/dma/ioat/
>  F: doc/guides/dmadevs/ioat.rst
>
> +Marvell CNXK DPI DMA
> +M: Radha Mohan Chintakuntla 
> +M: Veerasenareddy Burru 
> +F: drivers/dma/cnxk/
> +F: doc/guides/dmadevs/cnxk.rst
> +
>
>  RegEx Drivers
>  -
> diff --git a/doc/guides/dmadevs/cnxk.rst b/doc/guides/dmadevs/cnxk.rst
> new file mode 100644
> index 00..8ae7c1f8cd
> --- /dev/null
> +++ b/doc/guides/dmadevs/cnxk.rst
> @@ -0,0 +1,53 @@
> +..  SPDX-License-Identifier: BSD-3-Clause
> +Copyright(c) 2021 Marvell International Ltd.
> +
> +.. include:: 
> +


Please link top-level doc/guides/platform/cnxk.rst documentation file
to this file.


> +CNXK DMA Device Driver
> +==
> +
> +The ``cnxk`` dmadev driver provides a poll-mode driver (PMD) for Marvell DPI 
> DMA
> +Hardware Accelerator block found in OCTEONTX2 and OCTEONTX3 family of SoCs. 
> Each
> +DMA queue is exposed as a VF function when SRIOV is enabled.
> +
> +The block supports following modes of DMA transfers
> +
> +#. Internal - DMA within SoC DRAM to DRAM
> +
> +#. Inbound  - Host DRAM to SoC DRAM when SoC is in PCIe Endpoint
> +
> +#. Outbound - SoC DRAM to Host DRAM when SoC is in PCIe Endpoint
> +
> +Device Setup
> +-
> +The ``dpdk-devbind.py`` script, included with DPDK, can be used to show the
> +presence of supported hardware. Running ``dpdk-devbind.py --status-dev dma``
> +will show all the CNXK DMA devices.
> +
> +Devices using VFIO drivers
> +~~~
> +
> +The HW devices to be used will need to be bound to a user-space IO driver 
> for use.
> +The ``dpdk-devbind.py`` script can be used to view the state of the devices
> +and to bind them to a suitable DPDK-supported driver, such as ``vfio-pci``.
> +For example::
> +
> + $ dpdk-devbind.py -b vfio-pci :05:00.1
> +
> +Device Probing and Initialization
> +~~
> +
> +To use the devices from an application, the dmadev API can be used.
> +CNXK DMA device configuration requirements:
> +
> +* Only one ``vchan`` is supported per device.
> +* CNXK DMA devices do not support silent mode.
> +
> +Once configured, the device can then be made ready for use by calling the
> + ``rte_dma_start()`` API.
> +
> +Performing Data Copies
> +~~~
> +
> +Refer to the :ref:`Enqueue / Dequeue APIs ` section 
> of the dmadev library
> +documentation for details on operation enqueue and submission API usage.
> diff --git a/doc/guides/dmadevs/index.rst b/doc/guides/dmadevs/index.rst
> index 20476039a5..227fa00c68 100644
> --- a/doc/guides/dmadevs/index.rst
> +++ b/doc/guides/dmadevs/index.rst
> @@ -11,5 +11,6 @@ an application through DMA API.
> :maxdepth: 2
> :numbered:
>
> +   cnxk
> idxd
> ioat
> diff --git a/drivers/dma/cnxk/cnxk_dmadev.c b/drivers/dma/cnxk/cnxk_dmadev.c
> new file mode 100644
> index 00..620766743d
> --- /dev/null
> +++ b/drivers/dma/cnxk/cnxk_dmadev.c
> @@ -0,0 +1,119 @@
> +/* SPDX-License-Identifier: BSD-3-Clause
> + * Copyright (C) 2021 Marvell International Ltd.
> + */
> +
> +#include 
> +#include 
> +
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +
> +#include 
> +#include 
> +
> +static int
> +cnxk_dmadev_probe(struct rte_pci_driver *pci_drv __rte_unused,
> + struct rte_pci_device *pci_dev)
> +{
> +   struct cnxk_dpi_vf_s *dpivf = NULL;
> +   char name[RTE_DEV_NAME_MAX_LEN];
> +   struct rte_dma_dev *dmadev;
> +   struct roc_dpi *rdpi = NULL;
> +   int rc;
> +
> +   if (!pci_dev->mem_resource[0].addr)
> +   return -ENODEV;
> +
> +   rc = roc_plt_init();
> +   if (rc) {
> +   plt_err("Failed to initialize platform model, rc=%d", rc);
> +   return rc;
> +   }
> +

Re: [dpdk-dev] [PATCH] doc: document a limitation for a meter with RSS action

2021-10-26 Thread Thomas Monjalon
26/10/2021 04:08, Li Zhang:
> A meter policy with RSS/Queue action is not supported
> when dv_xmeta_en != 0.

You could say why. HW limitation? driver design?

>- Policy actions of RSS for green and yellow should have the same 
> configuration except queues.
> +  - Policy with RSS/queue action is not supported when dv_xmeta_en != 0.

dv_xmeta_en should be enclosed with ticks: ``dv_xmeta_en``
!= 0 should be said in english: "is not zero" or "is enabled".




Re: [dpdk-dev] [PATCH 4/4] dma/cnxk: add copy_sg function

2021-10-26 Thread Jerin Jacob
On Tue, Oct 26, 2021 at 9:43 AM Radha Mohan Chintakuntla
 wrote:
>
> Add the copy_sg function that will do the multiple DMA transfers of
> different sizes and different source/destination as well.

> +static int
> +cnxk_dmadev_copy_sg(void *dev_private, uint16_t vchan,
> +   const struct rte_dma_sge *src,
> +   const struct rte_dma_sge *dst,
> +   uint16_t nb_src, uint16_t nb_dst, uint64_t flags)
> +{
> +   uint64_t cmd[DPI_MAX_CMD_SIZE] = {0};
> +   union dpi_instr_hdr_s *header = (union dpi_instr_hdr_s *)&cmd[0];
> +   struct cnxk_dpi_vf_s *dpivf = dev_private;
> +   const struct rte_dma_sge *fptr, *lptr;
> +   struct cnxk_dpi_compl_s *comp_ptr;
> +   int num_words = 0;
> +   int i, rc;
> +
> +   RTE_SET_USED(vchan);
> +
> +   header->s.xtype = dpivf->conf.direction;
> +   header->s.pt = DPI_HDR_PT_ZBW_CA;
> +   header->s.grp = 0;
> +   header->s.tag = 0;
> +   header->s.tt = 0;
> +   header->s.func = 0;
> +   comp_ptr = dpivf->conf.c_desc.compl_ptr[dpivf->conf.c_desc.tail];
> +   comp_ptr->cdata = DPI_REQ_CDATA;
> +   header->s.ptr = (uint64_t)comp_ptr;
> +   STRM_INC(dpivf->conf.c_desc);
> +
> +   /* pvfs should be set for inbound and outbound only */
> +   if (header->s.xtype <= 1)
> +   header->s.pvfe = 1;
> +   num_words += 4;


# Please change the logic to populate the static items based on
configure/channel setup
in slowpath and update only per transfer-specific items to have better
performance.


# Also make sure test application and example application passes this
patch series.


[dpdk-dev] [PATCH v4] net/ice: simplify the use of DCF device reset

2021-10-26 Thread dapengx . yu
From: Dapeng Yu 

After DCF is reset by PF, the DCF device un-initialization cannot
function normally since the resource is already invalidated. So
reset DCF twice is necessary, the first simplified reset
re-initializes the AdminQ of DCF, only then second reset can clean
the filters successfully.

This patch detects the reset flag, which is set by PF on DCF reset,
if the flag is true, do DCF reset twice automatically.

Fixes: 1a86f4dbdf42 ("net/ice: support DCF device reset")
Cc: sta...@dpdk.org

Signed-off-by: Dapeng Yu 
---
V2:
* Ignore the returned error of dev_uninit when DCF is reset by PF
V3:
* Add a reset function to re-initialize AdminQ resource
* Add a function to check the reset flag
V4:
* Remove redundant reset flag setting
---
 drivers/net/ice/ice_dcf.c|  2 ++
 drivers/net/ice/ice_dcf_ethdev.c | 33 +++-
 2 files changed, 34 insertions(+), 1 deletion(-)

diff --git a/drivers/net/ice/ice_dcf.c b/drivers/net/ice/ice_dcf.c
index 084f7a53db..366ff0a907 100644
--- a/drivers/net/ice/ice_dcf.c
+++ b/drivers/net/ice/ice_dcf.c
@@ -593,6 +593,8 @@ ice_dcf_init_hw(struct rte_eth_dev *eth_dev, struct 
ice_dcf_hw *hw)
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
int ret, size;
 
+   hw->resetting = false;
+
hw->avf.hw_addr = pci_dev->mem_resource[0].addr;
hw->avf.back = hw;
 
diff --git a/drivers/net/ice/ice_dcf_ethdev.c b/drivers/net/ice/ice_dcf_ethdev.c
index 7c71a48010..171b58a748 100644
--- a/drivers/net/ice/ice_dcf_ethdev.c
+++ b/drivers/net/ice/ice_dcf_ethdev.c
@@ -1025,11 +1025,43 @@ ice_dcf_tm_ops_get(struct rte_eth_dev *dev __rte_unused,
return 0;
 }
 
+static inline void
+ice_dcf_simple_reset(struct rte_eth_dev *eth_dev, struct ice_dcf_hw *hw)
+{
+   ice_dcf_uninit_hw(eth_dev, hw);
+   ice_dcf_init_hw(eth_dev, hw);
+}
+
+/* Check if reset has been triggered by PF */
+static inline bool
+dcf_is_reset(struct rte_eth_dev *dev)
+{
+   struct ice_dcf_adapter *ad = dev->data->dev_private;
+   struct iavf_hw *hw = &ad->real_hw.avf;
+
+   return !(IAVF_READ_REG(hw, IAVF_VF_ARQLEN1) &
+IAVF_VF_ARQLEN1_ARQENABLE_MASK);
+}
+
 static int
 ice_dcf_dev_reset(struct rte_eth_dev *dev)
 {
+   struct ice_dcf_adapter *ad = dev->data->dev_private;
+   struct ice_dcf_hw *hw = &ad->real_hw;
int ret;
 
+   if (dcf_is_reset(dev)) {
+   if (!ad->real_hw.resetting)
+   ad->real_hw.resetting = true;
+   PMD_DRV_LOG(ERR, "The DCF has been reset by PF");
+
+   /*
+* Do the simplified reset to make DCF get AdminQ resource.
+* Then the next uninit/init can clean filters successfully.
+*/
+   ice_dcf_simple_reset(dev, hw);
+   }
+
ret = ice_dcf_dev_uninit(dev);
if (ret)
return ret;
@@ -1072,7 +1104,6 @@ ice_dcf_dev_init(struct rte_eth_dev *eth_dev)
 {
struct ice_dcf_adapter *adapter = eth_dev->data->dev_private;
 
-   adapter->real_hw.resetting = false;
eth_dev->dev_ops = &ice_dcf_eth_dev_ops;
eth_dev->rx_pkt_burst = ice_dcf_recv_pkts;
eth_dev->tx_pkt_burst = ice_dcf_xmit_pkts;
-- 
2.27.0



Re: [dpdk-dev] [PATCH v6 01/12] lib: build libraries that some tests depend on

2021-10-26 Thread Thomas Monjalon
26/10/2021 02:47, Jie Zhou:
> On Mon, Oct 25, 2021 at 05:38:42PM +0200, Thomas Monjalon wrote:
> > 14/10/2021 18:21, Jie Zhou:
> > > Enable building subset of libraries that tests depend on for Windows
> > > 
> > > Signed-off-by: Jie Zhou 
> > > ---
> > >  lib/meson.build | 2 ++
> > >  1 file changed, 2 insertions(+)
> > > 
> > > diff --git a/lib/meson.build b/lib/meson.build
> > > index b2ba7258d8..bd6c27deef 100644
> > > --- a/lib/meson.build
> > > +++ b/lib/meson.build
> > > @@ -82,9 +82,11 @@ if is_windows
> > >  'bitratestats',
> > >  'cryptodev',
> > >  'cfgfile',
> > > +'efd',
> > >  'gro',
> > >  'gso',
> > >  'latencystats',
> > > +'lpm',
> > >  'pdump',
> > >  'stack',
> > >  'security',
> > > 
> > 
> > It needs to be rebased after the recent changes to disable libs on Windows.
> > But instead of such patch, I would prefer one patch per lib,
> > with Tested-by tags showing that the lib is tested and working on Windows.
> > Also you need to Cc the maintainer of the lib being enabled.
> >
> Thanks Thomas. Will rebase in V7, and will separately enabling efd, lpm, and 
> their corresponding unit tests in two patches after this one. 

I think it enabling can be done in separate patches out of the series.
There is a chance to merge such patch before the big series.




Re: [dpdk-dev] [PATCH 3/4] dma/cnxk: add dma channel operations

2021-10-26 Thread Jerin Jacob
On Tue, Oct 26, 2021 at 9:43 AM Radha Mohan Chintakuntla
 wrote:
>
> Add functions for the dmadev vchan setup and DMA operations.
>
> Signed-off-by: Radha Mohan Chintakuntla 
> ---
>  drivers/dma/cnxk/cnxk_dmadev.c | 322 +
>  drivers/dma/cnxk/cnxk_dmadev.h |  53 ++
>  drivers/dma/cnxk/version.map   |   3 +
>  3 files changed, 378 insertions(+)
>  create mode 100644 drivers/dma/cnxk/version.map
>
> diff --git a/drivers/dma/cnxk/cnxk_dmadev.c b/drivers/dma/cnxk/cnxk_dmadev.c
> index 620766743d..8434579aa2 100644
> --- a/drivers/dma/cnxk/cnxk_dmadev.c
> +++ b/drivers/dma/cnxk/cnxk_dmadev.c
> @@ -18,6 +18,322 @@
>  #include 
>  #include 
>
> +static int
> +cnxk_dmadev_info_get(const struct rte_dma_dev *dev,
> +struct rte_dma_info *dev_info, uint32_t size)
> +{
> +   RTE_SET_USED(dev);
> +   RTE_SET_USED(size);
> +
> +   dev_info->max_vchans = 1;
> +   dev_info->nb_vchans = 1;
> +   dev_info->dev_capa = RTE_DMA_CAPA_MEM_TO_MEM |
> +   RTE_DMA_CAPA_MEM_TO_DEV | RTE_DMA_CAPA_DEV_TO_MEM |
> +   RTE_DMA_CAPA_OPS_COPY;
> +   dev_info->max_desc = DPI_MAX_DESC;
> +   dev_info->min_desc = 1;
> +   dev_info->max_sges = DPI_MAX_POINTER;
> +
> +   return 0;
> +}
> +
> +static int
> +cnxk_dmadev_configure(struct rte_dma_dev *dev,
> + const struct rte_dma_conf *conf, uint32_t conf_sz)
> +{
> +   struct cnxk_dpi_vf_s *dpivf = NULL;
> +   int rc = 0;
> +
> +   RTE_SET_USED(conf);
> +   RTE_SET_USED(conf);
> +   RTE_SET_USED(conf_sz);
> +   RTE_SET_USED(conf_sz);
> +   dpivf = dev->fp_obj->dev_private;
> +   rc = roc_dpi_queue_configure(&dpivf->rdpi);
> +   if (rc < 0)
> +   plt_err("DMA queue configure failed err = %d", rc);
> +
> +   return rc;
> +}
> +
> +static int
> +cnxk_dmadev_vchan_setup(struct rte_dma_dev *dev, uint16_t vchan,
> +   const struct rte_dma_vchan_conf *conf,
> +   uint32_t conf_sz)
> +{
> +   struct cnxk_dpi_vf_s *dpivf = dev->fp_obj->dev_private;
> +   struct cnxk_dpi_compl_s *comp_data;
> +   int i;
> +
> +   RTE_SET_USED(vchan);
> +   RTE_SET_USED(conf_sz);
> +
> +   switch (conf->direction) {
> +   case RTE_DMA_DIR_DEV_TO_MEM:
> +   dpivf->conf.direction = DPI_XTYPE_INBOUND;
> +   dpivf->conf.src_port = conf->src_port.pcie.coreid;
> +   dpivf->conf.dst_port = 0;
> +   break;
> +   case RTE_DMA_DIR_MEM_TO_DEV:
> +   dpivf->conf.direction = DPI_XTYPE_OUTBOUND;
> +   dpivf->conf.src_port = 0;
> +   dpivf->conf.dst_port = conf->dst_port.pcie.coreid;
> +   break;
> +   case RTE_DMA_DIR_MEM_TO_MEM:
> +   dpivf->conf.direction = DPI_XTYPE_INTERNAL_ONLY;
> +   dpivf->conf.src_port = 0;
> +   dpivf->conf.dst_port = 0;
> +   break;
> +   case RTE_DMA_DIR_DEV_TO_DEV:
> +   dpivf->conf.direction = DPI_XTYPE_EXTERNAL_ONLY;
> +   dpivf->conf.src_port = conf->src_port.pcie.coreid;
> +   dpivf->conf.dst_port = conf->src_port.pcie.coreid;
> +   };
> +
> +   for (i = 0; i < conf->nb_desc; i++) {
> +   comp_data = rte_zmalloc(NULL, sizeof(*comp_data), 0);
> +   dpivf->conf.c_desc.compl_ptr[i] = comp_data;
> +   };
> +   dpivf->conf.c_desc.max_cnt = DPI_MAX_DESC;
> +   dpivf->conf.c_desc.head = 0;
> +   dpivf->conf.c_desc.tail = 0;
> +
> +   return 0;
> +}
> +
> +static int
> +cnxk_dmadev_start(struct rte_dma_dev *dev)
> +{
> +   struct cnxk_dpi_vf_s *dpivf = dev->fp_obj->dev_private;
> +
> +   roc_dpi_queue_start(&dpivf->rdpi);
> +
> +   return 0;
> +}
> +
> +static int
> +cnxk_dmadev_stop(struct rte_dma_dev *dev)
> +{
> +   struct cnxk_dpi_vf_s *dpivf = dev->fp_obj->dev_private;
> +
> +   roc_dpi_queue_stop(&dpivf->rdpi);
> +
> +   return 0;
> +}
> +
> +static int
> +cnxk_dmadev_close(struct rte_dma_dev *dev)
> +{
> +   struct cnxk_dpi_vf_s *dpivf = dev->fp_obj->dev_private;
> +
> +   roc_dpi_queue_stop(&dpivf->rdpi);
> +   roc_dpi_dev_fini(&dpivf->rdpi);
> +
> +   return 0;
> +}
> +
> +static inline int
> +__dpi_queue_write(struct roc_dpi *dpi, uint64_t *cmds, int cmd_count)
> +{
> +   uint64_t *ptr = dpi->chunk_base;
> +
> +   if ((cmd_count < DPI_MIN_CMD_SIZE) || (cmd_count > DPI_MAX_CMD_SIZE) 
> ||
> +   cmds == NULL)
> +   return -EINVAL;
> +
> +   /*
> +* Normally there is plenty of room in the current buffer for the
> +* command
> +*/
> +   if (dpi->chunk_head + cmd_count < dpi->pool_size_m1) {
> +   ptr += dpi->chunk_head;
> +   dpi->chunk_head += cmd_count;
> +   while (cmd_count--)
> +   *ptr++ = *cmds++;
> +   } else {
> +   int count;
> +  

[dpdk-dev] [PATCH] net/ice: change RTE log level to DEBUG

2021-10-26 Thread dapengx . yu
From: Dapeng Yu 

When VF is reset in high frequency, the AdminQ command may fail, but
the failure can be recovered actually. So the error messages for getting
VF resource, getting VSI map, and re-directing flows should be
suppressed, in order to avoid causing concern for DCF users.

Fixes: 7564d5509611 ("net/ice: add DCF hardware initialization")
Cc: sta...@dpdk.org

Signed-off-by: Dapeng Yu 
---
 drivers/net/ice/ice_dcf.c   | 8 
 drivers/net/ice/ice_generic_flow.c  | 2 +-
 drivers/net/ice/ice_switch_filter.c | 4 ++--
 3 files changed, 7 insertions(+), 7 deletions(-)

diff --git a/drivers/net/ice/ice_dcf.c b/drivers/net/ice/ice_dcf.c
index 084f7a53db..9a53ba3538 100644
--- a/drivers/net/ice/ice_dcf.c
+++ b/drivers/net/ice/ice_dcf.c
@@ -241,7 +241,7 @@ ice_dcf_get_vf_resource(struct ice_dcf_hw *hw)
err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_GET_VF_RESOURCES,
  (uint8_t *)&caps, sizeof(caps));
if (err) {
-   PMD_DRV_LOG(ERR, "Failed to send msg OP_GET_VF_RESOURCE");
+   PMD_DRV_LOG(DEBUG, "Failed to send msg OP_GET_VF_RESOURCE");
return err;
}
 
@@ -249,7 +249,7 @@ ice_dcf_get_vf_resource(struct ice_dcf_hw *hw)
  (uint8_t *)hw->vf_res,
  ICE_DCF_VF_RES_BUF_SZ, NULL);
if (err) {
-   PMD_DRV_LOG(ERR, "Failed to get response of 
OP_GET_VF_RESOURCE");
+   PMD_DRV_LOG(DEBUG, "Failed to get response of 
OP_GET_VF_RESOURCE");
return -1;
}
 
@@ -283,7 +283,7 @@ ice_dcf_get_vf_vsi_map(struct ice_dcf_hw *hw)
err = ice_dcf_send_cmd_req_no_irq(hw, VIRTCHNL_OP_DCF_GET_VSI_MAP,
  NULL, 0);
if (err) {
-   PMD_DRV_LOG(ERR, "Failed to send msg OP_DCF_GET_VSI_MAP");
+   PMD_DRV_LOG(DEBUG, "Failed to send msg OP_DCF_GET_VSI_MAP");
return err;
}
 
@@ -291,7 +291,7 @@ ice_dcf_get_vf_vsi_map(struct ice_dcf_hw *hw)
  hw->arq_buf, ICE_DCF_AQ_BUF_SZ,
  &len);
if (err) {
-   PMD_DRV_LOG(ERR, "Failed to get response of 
OP_DCF_GET_VSI_MAP");
+   PMD_DRV_LOG(DEBUG, "Failed to get response of 
OP_DCF_GET_VSI_MAP");
return err;
}
 
diff --git a/drivers/net/ice/ice_generic_flow.c 
b/drivers/net/ice/ice_generic_flow.c
index 02f854666a..4aca24de89 100644
--- a/drivers/net/ice/ice_generic_flow.c
+++ b/drivers/net/ice/ice_generic_flow.c
@@ -2551,7 +2551,7 @@ ice_flow_redirect(struct ice_adapter *ad,
continue;
ret = p_flow->engine->redirect(ad, p_flow, rd);
if (ret) {
-   PMD_DRV_LOG(ERR, "Failed to redirect flows");
+   PMD_DRV_LOG(DEBUG, "Failed to redirect flows");
break;
}
}
diff --git a/drivers/net/ice/ice_switch_filter.c 
b/drivers/net/ice/ice_switch_filter.c
index 6b0c1bff1e..b4f3bbafff 100644
--- a/drivers/net/ice/ice_switch_filter.c
+++ b/drivers/net/ice/ice_switch_filter.c
@@ -1943,7 +1943,7 @@ ice_switch_redirect(struct ice_adapter *ad,
ret = ice_rem_adv_rule(hw, list_itr->lkups,
   lkups_cnt, &rinfo);
if (ret) {
-   PMD_DRV_LOG(ERR, "Failed to delete the old rule %d",
+   PMD_DRV_LOG(DEBUG, "Failed to delete the old rule %d",
rdata->rule_id);
ret = -EINVAL;
goto out;
@@ -1956,7 +1956,7 @@ ice_switch_redirect(struct ice_adapter *ad,
ret = ice_add_adv_rule(hw, lkups_dp, lkups_cnt,
   &rinfo, rdata);
if (ret) {
-   PMD_DRV_LOG(ERR, "Failed to replay the rule");
+   PMD_DRV_LOG(DEBUG, "Failed to replay the rule");
ret = -EINVAL;
}
 
-- 
2.27.0



Re: [dpdk-dev] [PATCH] vhost: fix async DMA map

2021-10-26 Thread Ding, Xuan
Hi Maxime,

>-Original Message-
>From: Maxime Coquelin 
>Sent: Tuesday, October 26, 2021 2:53 PM
>To: Ding, Xuan ; dev@dpdk.org;
>david.march...@redhat.com; Xia, Chenbo 
>Cc: Burakov, Anatoly 
>Subject: Re: [PATCH] vhost: fix async DMA map
>
>
>
>On 10/26/21 04:07, Ding, Xuan wrote:
>> Hi Maxime,
>>
>>> -Original Message-
>>> From: Maxime Coquelin 
>>> Sent: Tuesday, October 26, 2021 4:47 AM
>>> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
>>> ; Ding, Xuan 
>>> Subject: Re: [PATCH] vhost: fix async DMA map
>>>
>>> Hi Xuan,
>>>
>>> On 10/25/21 22:33, Maxime Coquelin wrote:
 This patch fixes possible NULL-pointer dereferencing
 reported by Coverity and also fixes NUMA reallocation
 of the async DMA map.

 Fixes: 7c61fa08b716 ("vhost: enable IOMMU for async vhost")

 Coverity issue: 373655

 Signed-off-by: Maxime Coquelin 
 ---
lib/vhost/vhost_user.c | 45 +++---
1 file changed, 20 insertions(+), 25 deletions(-)

>>>
>>> I posted this patch to fix the issue reported by Coverity and also other
>>> issue on NUMA realloc that I found at the same time. But I wonder
>>> whether all this async_map_status is needed.
>>
>> Thanks for your fix! I can help to review and test the patch later.
>>
>> I add the async_map_status in v2 for compatibility. Some DMA device,
>> like DSA, may use kernel idxd driver only. If there is no device bound to
>> DPDK vfio and kernel vfio module is modprobed to ensure
>rte_vfio_is_enabled() is true,
>> we will unavoidably do DMA map/unmap and it will fail.
>>
>> Therefore, the dma_map_status here is used to filter this situation by
>preventing
>> unnecessary DMA unmap.
>
>Ok, then I think we can just remove the async DMA map.
>
>>>
>>> Indeed, if the only place where we DMA map is in
>>> vhost_user_mmap_region(). If it fails, the error is propagated, the mem
>>> table are freed and NACK is replied to the master. IOW, the device will
>>> be in an unusable state.
>>
>> I agree with you, this is the place I consider right to do DMA map
>> because we also do SW mapping here, any suggestions?
>
>No suggestion, I was just explaining that at the only place where
>DMA map were done, mapping errors were properly handled and propagated.

What about just setting async_copy to false, and allow switching to sync path.

>
>>>
>>> Removing the async DMA map will simplify a lot the code, do you agree to
>>> remove it or there is something I missed?
>>
>> See above. Indeed, it adds a lot of code. But we can't know the driver for
>> each device in vhost lib, or we can only restrict the user to bind some
>devices
>> to DPDK vfio if async logic needed.
>
>I would think we don't care if DMA unmap fails, we can just do the same
>as what you do for DMA map, i.e. just ignore the error.

Get your idea, we can do the same as DMA map, and in this way dma_map_status 
flag can be removed.

>
>Thanks to this discussion, I have now more concerns on how it works. I
>think we have a problem here in case of DMA device hotplug, that device
>could miss the necessary map entries from Vhost if no VFIO devices were
>attached at VHST_USER_SET_MEM_TABLE time. How would you handle that
>case?

DMA device is uncore, so I don't see the  hotplug issue here.
I will have another patch containing compatibility with sync path, and 
async_map_status flag will be removed.
Hope to get your insights.

Thanks,
Xuan

>
>Regards,
>Maxime
>
>>>
>>> Thanks,
>>> Maxime
>>



Re: [dpdk-dev] [PATCH v4] net/ice: simplify the use of DCF device reset

2021-10-26 Thread Zhang, Qi Z



> -Original Message-
> From: Yu, DapengX 
> Sent: Tuesday, October 26, 2021 4:44 PM
> To: Yang, Qiming ; Zhang, Qi Z
> 
> Cc: dev@dpdk.org; Yu, DapengX ; sta...@dpdk.org
> Subject: [PATCH v4] net/ice: simplify the use of DCF device reset
> 
> From: Dapeng Yu 
> 
> After DCF is reset by PF, the DCF device un-initialization cannot function
> normally since the resource is already invalidated. So reset DCF twice is
> necessary, the first simplified reset re-initializes the AdminQ of DCF, only 
> then
> second reset can clean the filters successfully.
> 
> This patch detects the reset flag, which is set by PF on DCF reset, if the 
> flag is
> true, do DCF reset twice automatically.
> 
> Fixes: 1a86f4dbdf42 ("net/ice: support DCF device reset")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Dapeng Yu 
> ---
> V2:
> * Ignore the returned error of dev_uninit when DCF is reset by PF
> V3:
> * Add a reset function to re-initialize AdminQ resource
> * Add a function to check the reset flag
> V4:
> * Remove redundant reset flag setting
> ---
>  drivers/net/ice/ice_dcf.c|  2 ++
>  drivers/net/ice/ice_dcf_ethdev.c | 33 +++-
>  2 files changed, 34 insertions(+), 1 deletion(-)
> 
> diff --git a/drivers/net/ice/ice_dcf.c b/drivers/net/ice/ice_dcf.c index
> 084f7a53db..366ff0a907 100644
> --- a/drivers/net/ice/ice_dcf.c
> +++ b/drivers/net/ice/ice_dcf.c
> @@ -593,6 +593,8 @@ ice_dcf_init_hw(struct rte_eth_dev *eth_dev, struct
> ice_dcf_hw *hw)
>   struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
>   int ret, size;
> 
> + hw->resetting = false;
> +
>   hw->avf.hw_addr = pci_dev->mem_resource[0].addr;
>   hw->avf.back = hw;
> 
> diff --git a/drivers/net/ice/ice_dcf_ethdev.c 
> b/drivers/net/ice/ice_dcf_ethdev.c
> index 7c71a48010..171b58a748 100644
> --- a/drivers/net/ice/ice_dcf_ethdev.c
> +++ b/drivers/net/ice/ice_dcf_ethdev.c
> @@ -1025,11 +1025,43 @@ ice_dcf_tm_ops_get(struct rte_eth_dev *dev
> __rte_unused,
>   return 0;
>  }
> 
> +static inline void
> +ice_dcf_simple_reset(struct rte_eth_dev *eth_dev, struct ice_dcf_hw  

Better to rename the function to ice_dcf_reset_hw 

> +*hw) {
> + ice_dcf_uninit_hw(eth_dev, hw);
> + ice_dcf_init_hw(eth_dev, hw);
> +}
> +
> +/* Check if reset has been triggered by PF */ static inline bool
> +dcf_is_reset(struct rte_eth_dev *dev) {
> + struct ice_dcf_adapter *ad = dev->data->dev_private;
> + struct iavf_hw *hw = &ad->real_hw.avf;
> +
> + return !(IAVF_READ_REG(hw, IAVF_VF_ARQLEN1) &
> +  IAVF_VF_ARQLEN1_ARQENABLE_MASK);
> +}
> +
>  static int
>  ice_dcf_dev_reset(struct rte_eth_dev *dev)  {
> + struct ice_dcf_adapter *ad = dev->data->dev_private;
> + struct ice_dcf_hw *hw = &ad->real_hw;
>   int ret;
> 
> + if (dcf_is_reset(dev)) {
> + if (!ad->real_hw.resetting)
> + ad->real_hw.resetting = true;
> + PMD_DRV_LOG(ERR, "The DCF has been reset by PF");
> +
> + /*
> +  * Do the simplified reset to make DCF get AdminQ resource.
> +  * Then the next uninit/init can clean filters successfully.
> +  */ 

Can reword as below:

Simply reset hw to trigger an additional DCF enable/disable cycle which help to 
workaround
the issue that kernel driver may not clean up resource during previous reset.

> + ice_dcf_simple_reset(dev, hw);
> + }
> +
>   ret = ice_dcf_dev_uninit(dev);
>   if (ret)
>   return ret;
> @@ -1072,7 +1104,6 @@ ice_dcf_dev_init(struct rte_eth_dev *eth_dev)  {
>   struct ice_dcf_adapter *adapter = eth_dev->data->dev_private;
> 
> - adapter->real_hw.resetting = false;
>   eth_dev->dev_ops = &ice_dcf_eth_dev_ops;
>   eth_dev->rx_pkt_burst = ice_dcf_recv_pkts;
>   eth_dev->tx_pkt_burst = ice_dcf_xmit_pkts;
> --
> 2.27.0



[dpdk-dev] [PATCH v2 00/19] fixes and enhancements to Truflow

2021-10-26 Thread Venkat Duvvuru
This patch set adds enhancements and fixes to Truflow feature.

Enhancements include:
* Scaling numbers on Thor
* Inner IP header support for GRE tunnel flows
* Enable wildcard match for ingress flows
* Add clear on read for flow stats on Thor
* Add nat support for dest IP and port combination
* Remove 2-slice WC support
* Add support for socket redirect feature
* Add new API TruFlow get SRAM resources
* Remove accumulation of stats devargs argument
* Add TruFlow and AFM SRAM partitioning support

V2:
* Compilation fixes
* Remove 2-slice WC support
* Add support for socket redirect feature
* Add new API TruFlow get SRAM resources
* Remove accumulation of stats devargs argument
* Add TruFlow and AFM SRAM partitioning support

Farah Smith (1):
  net/bnxt: add clear on read support

Jay Ding (4):
  net/bnxt: get TruFlow version
  net/bnxt: add new API TruFlow get SRAM resources
  net/bnxt: add TruFlow and AFM SRAM partitioning support
  net/bnxt: add Tx TruFlow table config for p4

Kishore Padmanabha (11):
  net/bnxt: add NAT support for dest IP and port combination
  net/bnxt: add support for multi root capability
  net/bnxt: fix the out of boundary issue in hash list
  net/bnxt: add capability option for socket redirect
  net/bnxt: enable wildcard match for ingress flows
  net/bnxt: support inner IP header for GRE tunnel flows
  net/bnxt: remove accumulation of stats devargs argument
  net/bnxt: updated the log messages
  net/bnxt: add support for socket redirect feature
  net/bnxt: delete the VF pair before VF representor alloc
  net/bnxt: check for mismatch of control and physical port

Mike Baucom (1):
  net/bnxt: remove 2-slice WC entries for scale

Shahaji Bhosle (2):
  net/bnxt: increase flow scale for Thor
  net/bnxt: fix clang compiler warnings

 drivers/net/bnxt/bnxt.h   |   25 +-
 drivers/net/bnxt/bnxt_cpr.c   |2 +-
 drivers/net/bnxt/bnxt_ethdev.c|  274 +-
 drivers/net/bnxt/bnxt_hwrm.c  |   36 +
 drivers/net/bnxt/bnxt_hwrm.h  |1 +
 drivers/net/bnxt/bnxt_reps.c  |9 +-
 drivers/net/bnxt/hsi_struct_def_dpdk.h| 2970 +++--
 drivers/net/bnxt/tf_core/tf_core.c|  197 +
 drivers/net/bnxt/tf_core/tf_core.h|  202 +
 drivers/net/bnxt/tf_core/tf_device.c  |   11 +-
 drivers/net/bnxt/tf_core/tf_device.h  |  102 +
 drivers/net/bnxt/tf_core/tf_device_p4.c   |  178 +-
 drivers/net/bnxt/tf_core/tf_device_p4.h   |  137 +-
 drivers/net/bnxt/tf_core/tf_device_p58.c  |  421 +-
 drivers/net/bnxt/tf_core/tf_device_p58.h  |  205 +-
 drivers/net/bnxt/tf_core/tf_msg.c |   87 +-
 drivers/net/bnxt/tf_core/tf_msg.h |   31 +-
 drivers/net/bnxt/tf_core/tf_rm.c  |3 +-
 drivers/net/bnxt/tf_core/tf_sram_mgr.h|   10 -
 drivers/net/bnxt/tf_core/tf_tbl.c |8 +-
 drivers/net/bnxt/tf_core/tf_tbl_sram.c|   13 +-
 drivers/net/bnxt/tf_ulp/bnxt_ulp.c|   58 +-
 drivers/net/bnxt/tf_ulp/bnxt_ulp.h|9 +-
 drivers/net/bnxt/tf_ulp/bnxt_ulp_flow.c   |   29 +-
 .../generic_templates/ulp_template_db_act.c   |  376 +-
 .../generic_templates/ulp_template_db_class.c | 1986 -
 .../generic_templates/ulp_template_db_enum.h  |   91 +-
 .../generic_templates/ulp_template_db_tbl.c   | 3946 +
 .../ulp_template_db_thor_act.c|  150 +-
 .../ulp_template_db_thor_class.c  |  909 ++--
 .../ulp_template_db_wh_plus_act.c |  336 +-
 .../ulp_template_db_wh_plus_class.c   |  222 +-
 drivers/net/bnxt/tf_ulp/ulp_def_rules.c   |8 +-
 drivers/net/bnxt/tf_ulp/ulp_fc_mgr.c  |   12 +-
 drivers/net/bnxt/tf_ulp/ulp_flow_db.c |2 +-
 drivers/net/bnxt/tf_ulp/ulp_gen_hash.c|   20 +-
 drivers/net/bnxt/tf_ulp/ulp_mark_mgr.c|2 +-
 drivers/net/bnxt/tf_ulp/ulp_port_db.c |   23 +
 drivers/net/bnxt/tf_ulp/ulp_port_db.h |   13 +
 drivers/net/bnxt/tf_ulp/ulp_rte_parser.c  |   32 +-
 drivers/net/bnxt/tf_ulp/ulp_template_struct.h |2 +-
 41 files changed, 9254 insertions(+), 3894 deletions(-)

-- 
2.17.1



[dpdk-dev] [PATCH v2 01/19] net/bnxt: add NAT support for dest IP and port combination

2021-10-26 Thread Venkat Duvvuru
From: Kishore Padmanabha 

* Added support for NAT action for the destination IP and port
  combination for the Thor platform. This is not supported for
  Whitney platform.
* Consolidated the encapsulation and NAT entries for scaling flows
  with NAT actions.

Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Shahaji Bhosle 
Reviewed-by: Michael Baucom 
Reviewed-by: Randy Schacher 
---
 .../generic_templates/ulp_template_db_act.c   | 376 +++---
 .../generic_templates/ulp_template_db_enum.h  |  18 +-
 .../generic_templates/ulp_template_db_tbl.c   |  14 +-
 .../ulp_template_db_thor_class.c  |   2 +-
 .../ulp_template_db_wh_plus_act.c |  96 +++--
 5 files changed, 317 insertions(+), 189 deletions(-)

diff --git a/drivers/net/bnxt/tf_ulp/generic_templates/ulp_template_db_act.c 
b/drivers/net/bnxt/tf_ulp/generic_templates/ulp_template_db_act.c
index 0da6070d7d..ce878d8e02 100644
--- a/drivers/net/bnxt/tf_ulp/generic_templates/ulp_template_db_act.c
+++ b/drivers/net/bnxt/tf_ulp/generic_templates/ulp_template_db_act.c
@@ -3,7 +3,7 @@
  * All rights reserved.
  */
 
-/* date: Mon May 17 15:30:41 2021 */
+/* date: Wed Aug 25 14:37:06 2021 */
 
 #include "ulp_template_db_enum.h"
 #include "ulp_template_db_field.h"
@@ -47,59 +47,67 @@ uint16_t ulp_act_sig_tbl[BNXT_ULP_ACT_SIG_TBL_MAX_SZ] = {
[BNXT_ULP_ACT_HID_04bc] = 30,
[BNXT_ULP_ACT_HID_00a9] = 31,
[BNXT_ULP_ACT_HID_020f] = 32,
-   [BNXT_ULP_ACT_HID_04a9] = 33,
-   [BNXT_ULP_ACT_HID_01fc] = 34,
-   [BNXT_ULP_ACT_HID_04be] = 35,
-   [BNXT_ULP_ACT_HID_00ab] = 36,
-   [BNXT_ULP_ACT_HID_0211] = 37,
-   [BNXT_ULP_ACT_HID_04ab] = 38,
-   [BNXT_ULP_ACT_HID_01fe] = 39,
-   [BNXT_ULP_ACT_HID_0667] = 40,
-   [BNXT_ULP_ACT_HID_0254] = 41,
-   [BNXT_ULP_ACT_HID_03ba] = 42,
-   [BNXT_ULP_ACT_HID_0654] = 43,
-   [BNXT_ULP_ACT_HID_03a7] = 44,
-   [BNXT_ULP_ACT_HID_0669] = 45,
-   [BNXT_ULP_ACT_HID_0256] = 46,
-   [BNXT_ULP_ACT_HID_03bc] = 47,
-   [BNXT_ULP_ACT_HID_0656] = 48,
-   [BNXT_ULP_ACT_HID_03a9] = 49,
-   [BNXT_ULP_ACT_HID_021b] = 50,
-   [BNXT_ULP_ACT_HID_021c] = 51,
-   [BNXT_ULP_ACT_HID_021e] = 52,
-   [BNXT_ULP_ACT_HID_063f] = 53,
-   [BNXT_ULP_ACT_HID_0510] = 54,
-   [BNXT_ULP_ACT_HID_03c6] = 55,
-   [BNXT_ULP_ACT_HID_0082] = 56,
-   [BNXT_ULP_ACT_HID_06bb] = 57,
-   [BNXT_ULP_ACT_HID_021d] = 58,
-   [BNXT_ULP_ACT_HID_0641] = 59,
-   [BNXT_ULP_ACT_HID_0512] = 60,
-   [BNXT_ULP_ACT_HID_03c8] = 61,
-   [BNXT_ULP_ACT_HID_0084] = 62,
-   [BNXT_ULP_ACT_HID_06bd] = 63,
-   [BNXT_ULP_ACT_HID_06d7] = 64,
-   [BNXT_ULP_ACT_HID_02c4] = 65,
-   [BNXT_ULP_ACT_HID_042a] = 66,
-   [BNXT_ULP_ACT_HID_06c4] = 67,
-   [BNXT_ULP_ACT_HID_0417] = 68,
-   [BNXT_ULP_ACT_HID_06d9] = 69,
-   [BNXT_ULP_ACT_HID_02c6] = 70,
-   [BNXT_ULP_ACT_HID_042c] = 71,
-   [BNXT_ULP_ACT_HID_06c6] = 72,
-   [BNXT_ULP_ACT_HID_0419] = 73,
-   [BNXT_ULP_ACT_HID_0119] = 74,
-   [BNXT_ULP_ACT_HID_046f] = 75,
-   [BNXT_ULP_ACT_HID_05d5] = 76,
-   [BNXT_ULP_ACT_HID_0106] = 77,
-   [BNXT_ULP_ACT_HID_05c2] = 78,
-   [BNXT_ULP_ACT_HID_011b] = 79,
-   [BNXT_ULP_ACT_HID_0471] = 80,
-   [BNXT_ULP_ACT_HID_05d7] = 81,
-   [BNXT_ULP_ACT_HID_0108] = 82,
-   [BNXT_ULP_ACT_HID_05c4] = 83,
-   [BNXT_ULP_ACT_HID_00a2] = 84,
-   [BNXT_ULP_ACT_HID_00a4] = 85
+   [BNXT_ULP_ACT_HID_0153] = 33,
+   [BNXT_ULP_ACT_HID_04a9] = 34,
+   [BNXT_ULP_ACT_HID_01fc] = 35,
+   [BNXT_ULP_ACT_HID_04be] = 36,
+   [BNXT_ULP_ACT_HID_00ab] = 37,
+   [BNXT_ULP_ACT_HID_0211] = 38,
+   [BNXT_ULP_ACT_HID_0155] = 39,
+   [BNXT_ULP_ACT_HID_04ab] = 40,
+   [BNXT_ULP_ACT_HID_01fe] = 41,
+   [BNXT_ULP_ACT_HID_0667] = 42,
+   [BNXT_ULP_ACT_HID_0254] = 43,
+   [BNXT_ULP_ACT_HID_03ba] = 44,
+   [BNXT_ULP_ACT_HID_02fe] = 45,
+   [BNXT_ULP_ACT_HID_0654] = 46,
+   [BNXT_ULP_ACT_HID_03a7] = 47,
+   [BNXT_ULP_ACT_HID_0669] = 48,
+   [BNXT_ULP_ACT_HID_0256] = 49,
+   [BNXT_ULP_ACT_HID_03bc] = 50,
+   [BNXT_ULP_ACT_HID_0300] = 51,
+   [BNXT_ULP_ACT_HID_0656] = 52,
+   [BNXT_ULP_ACT_HID_03a9] = 53,
+   [BNXT_ULP_ACT_HID_021b] = 54,
+   [BNXT_ULP_ACT_HID_021c] = 55,
+   [BNXT_ULP_ACT_HID_021e] = 56,
+   [BNXT_ULP_ACT_HID_063f] = 57,
+   [BNXT_ULP_ACT_HID_0510] = 58,
+   [BNXT_ULP_ACT_HID_03c6] = 59,
+   [BNXT_ULP_ACT_HID_0082] = 60,
+   [BNXT_ULP_ACT_HID_06bb] = 61,
+   [BNXT_ULP_ACT_HID_021d] = 62,
+   [BNXT_ULP_ACT_HID_0641] = 63,
+   [BNXT_ULP_ACT_HID_0512] = 64,
+   [BNXT_ULP_ACT_HID_03c8] = 65,
+   [BNXT_ULP_ACT_HID_0084] = 66,
+   [BNXT_ULP_ACT_HID_06bd] = 67,
+   [BNXT_ULP_ACT_HID_06d7] = 68,
+   [BNXT_ULP_ACT_HID_02c4] = 69,
+   [BNXT_ULP_ACT_HID_042a] = 70,
+   [BNXT_ULP_ACT_HID_036e] = 71,
+   [BNXT_ULP_ACT_H

[dpdk-dev] [PATCH] bus/pci: fix selection of default device NUMA node

2021-10-26 Thread Houssem Bouhlel
There can be dev binding issue when no hugepages
are allocated for socket 0.
To avoid this, set device numa node value based on
the first lcore instead of 0.

Fixes: 831dba47bd36 ("bus/vmbus: add Hyper-V virtual bus support")
Cc: sta...@dpdk.org

Signed-off-by: Houssem Bouhlel 
Signed-off-by: Olivier Matz 
---
 drivers/bus/pci/pci_common.c | 4 +++-
 1 file changed, 3 insertions(+), 1 deletion(-)

diff --git a/drivers/bus/pci/pci_common.c b/drivers/bus/pci/pci_common.c
index f8fff2c98ebf..c70ab2373c79 100644
--- a/drivers/bus/pci/pci_common.c
+++ b/drivers/bus/pci/pci_common.c
@@ -166,6 +166,7 @@ rte_pci_probe_one_driver(struct rte_pci_driver *dr,
 struct rte_pci_device *dev)
 {
int ret;
+   unsigned int socket_id;
bool already_probed;
struct rte_pci_addr *loc;
 
@@ -194,7 +195,8 @@ rte_pci_probe_one_driver(struct rte_pci_driver *dr,
if (rte_socket_count() > 1)
RTE_LOG(INFO, EAL, "Device %s is not NUMA-aware, 
defaulting socket to 0\n",
dev->name);
-   dev->device.numa_node = 0;
+   socket_id = rte_lcore_to_socket_id(rte_get_next_lcore(-1, 0, 
0));
+   dev->device.numa_node = socket_id;
}
 
already_probed = rte_dev_is_probed(&dev->device);
-- 
2.30.2



[dpdk-dev] [PATCH v2 09/19] net/bnxt: increase flow scale for Thor

2021-10-26 Thread Venkat Duvvuru
From: Shahaji Bhosle 

* Updated defines and data types to allow 256 VFRs.
* Increased the encap record cache to support 256 to 4K entries. So
  VxLAN connections can be scaled to 4K entries.

Signed-off-by: Shahaji Bhosle 
Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Kishore Padmanabha 
Reviewed-by: Randy Schacher 
---
 drivers/net/bnxt/bnxt.h|  6 +-
 drivers/net/bnxt/bnxt_cpr.c|  2 +-
 drivers/net/bnxt/bnxt_ethdev.c | 18 +-
 drivers/net/bnxt/bnxt_reps.c   |  3 +--
 drivers/net/bnxt/tf_ulp/bnxt_ulp.h |  2 +-
 .../generic_templates/ulp_template_db_tbl.c|  8 
 drivers/net/bnxt/tf_ulp/ulp_def_rules.c|  6 +++---
 7 files changed, 24 insertions(+), 21 deletions(-)

diff --git a/drivers/net/bnxt/bnxt.h b/drivers/net/bnxt/bnxt.h
index c65e360446..5c064fd119 100644
--- a/drivers/net/bnxt/bnxt.h
+++ b/drivers/net/bnxt/bnxt.h
@@ -240,7 +240,11 @@ struct bnxt_parent_info {
 struct bnxt_pf_info {
 #define BNXT_FIRST_PF_FID  1
 #define BNXT_MAX_VFS(bp)   ((bp)->pf->max_vfs)
-#define BNXT_MAX_VF_REPS   64
+#define BNXT_MAX_VF_REPS_WH 64
+#define BNXT_MAX_VF_REPS_TH 256
+#define BNXT_MAX_VF_REPS(bp) \
+   (BNXT_CHIP_P5(bp) ? BNXT_MAX_VF_REPS_TH : \
+   BNXT_MAX_VF_REPS_WH)
 #define BNXT_TOTAL_VFS(bp) ((bp)->pf->total_vfs)
 #define BNXT_FIRST_VF_FID  128
 #define BNXT_PF_RINGS_USED(bp) bnxt_get_num_queues(bp)
diff --git a/drivers/net/bnxt/bnxt_cpr.c b/drivers/net/bnxt/bnxt_cpr.c
index 63ff02a198..6bb70d516e 100644
--- a/drivers/net/bnxt/bnxt_cpr.c
+++ b/drivers/net/bnxt/bnxt_cpr.c
@@ -74,7 +74,7 @@ bnxt_process_default_vnic_change(struct bnxt *bp,
BNXT_DEFAULT_VNIC_CHANGE_VF_ID_SFT;
PMD_DRV_LOG(INFO, "async event received vf_id 0x%x\n", vf_fid);
 
-   for (vf_id = 0; vf_id < BNXT_MAX_VF_REPS; vf_id++) {
+   for (vf_id = 0; vf_id < BNXT_MAX_VF_REPS(bp); vf_id++) {
eth_dev = bp->rep_info[vf_id].vfr_eth_dev;
if (!eth_dev)
continue;
diff --git a/drivers/net/bnxt/bnxt_ethdev.c b/drivers/net/bnxt/bnxt_ethdev.c
index bf034db336..8c72ab8fd1 100644
--- a/drivers/net/bnxt/bnxt_ethdev.c
+++ b/drivers/net/bnxt/bnxt_ethdev.c
@@ -1179,7 +1179,7 @@ void bnxt_print_link_info(struct rte_eth_dev *eth_dev)
struct rte_eth_link *link = ð_dev->data->dev_link;
 
if (link->link_status)
-   PMD_DRV_LOG(INFO, "Port %d Link Up - speed %u Mbps - %s\n",
+   PMD_DRV_LOG(DEBUG, "Port %d Link Up - speed %u Mbps - %s\n",
eth_dev->data->port_id,
(uint32_t)link->link_speed,
(link->link_duplex == ETH_LINK_FULL_DUPLEX) ?
@@ -6034,7 +6034,7 @@ static int bnxt_init_rep_info(struct bnxt *bp)
return 0;
 
bp->rep_info = rte_zmalloc("bnxt_rep_info",
-  sizeof(bp->rep_info[0]) * BNXT_MAX_VF_REPS,
+  sizeof(bp->rep_info[0]) * 
BNXT_MAX_VF_REPS(bp),
   0);
if (!bp->rep_info) {
PMD_DRV_LOG(ERR, "Failed to alloc memory for rep info\n");
@@ -6076,7 +6076,9 @@ static int bnxt_rep_port_probe(struct rte_pci_device 
*pci_dev,
 {
struct rte_eth_dev *vf_rep_eth_dev;
char name[RTE_ETH_NAME_MAX_LEN];
-   struct bnxt *backing_bp;
+   struct bnxt *backing_bp = backing_eth_dev->data->dev_private;
+   uint16_t max_vf_reps = BNXT_MAX_VF_REPS(backing_bp);
+
uint16_t num_rep;
int i, ret = 0;
struct rte_kvargs *kvlist = NULL;
@@ -6089,9 +6091,9 @@ static int bnxt_rep_port_probe(struct rte_pci_device 
*pci_dev,
return -ENOTSUP;
}
num_rep = eth_da->nb_representor_ports;
-   if (num_rep > BNXT_MAX_VF_REPS) {
+   if (num_rep > max_vf_reps) {
PMD_DRV_LOG(ERR, "nb_representor_ports = %d > %d MAX VF REPS\n",
-   num_rep, BNXT_MAX_VF_REPS);
+   num_rep, max_vf_reps);
return -EINVAL;
}
 
@@ -6102,8 +6104,6 @@ static int bnxt_rep_port_probe(struct rte_pci_device 
*pci_dev,
return -EINVAL;
}
 
-   backing_bp = backing_eth_dev->data->dev_private;
-
if (!(BNXT_PF(backing_bp) || BNXT_VF_IS_TRUSTED(backing_bp))) {
PMD_DRV_LOG(ERR,
"Not a PF or trusted VF. No Representor support\n");
@@ -6123,9 +6123,9 @@ static int bnxt_rep_port_probe(struct rte_pci_device 
*pci_dev,
.parent_dev = backing_eth_dev
};
 
-   if (representor.vf_id >= BNXT_MAX_VF_REPS) {
+   if (representor.vf_id >= max_vf_reps) {
PMD_DRV_LOG(ERR, "VF-Rep id %d >= %d MAX VF ID\n",
-

[dpdk-dev] [PATCH v2 03/19] net/bnxt: fix the out of boundary issue in hash list

2021-10-26 Thread Venkat Duvvuru
From: Kishore Padmanabha 

The number of hash bucket list calculation is fixed and added
check to limit the out of boundary condition

Fixes: 0001cc58d362 ("net/bnxt: support generic hash table")

Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Michael Baucom 
Reviewed-by: Randy Schacher 
---
 drivers/net/bnxt/tf_ulp/ulp_gen_hash.c | 20 ++--
 1 file changed, 14 insertions(+), 6 deletions(-)

diff --git a/drivers/net/bnxt/tf_ulp/ulp_gen_hash.c 
b/drivers/net/bnxt/tf_ulp/ulp_gen_hash.c
index 3c6e7fe924..84c83de35c 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_gen_hash.c
+++ b/drivers/net/bnxt/tf_ulp/ulp_gen_hash.c
@@ -16,20 +16,21 @@ int32_t ulp_bit_alloc_list_alloc(struct bit_alloc_list 
*blist,
 {
uint64_t bentry;
uint32_t idx = 0, jdx = 0;
+   uint32_t bsize_64 = blist->bsize / ULP_64B_IN_BYTES;
 
/* Iterate all numbers that have all 1's */
do {
bentry = blist->bdata[idx++];
-   } while (bentry == -1UL && idx < blist->bsize);
+   } while (bentry == -1UL && idx <= bsize_64);
 
-   if (idx < blist->bsize) {
+   if (idx <= bsize_64) {
if (bentry)
jdx = __builtin_clzl(~bentry);
*index = ((idx - 1) * ULP_INDEX_BITMAP_SIZE) + jdx;
ULP_INDEX_BITMAP_SET(blist->bdata[(idx - 1)], jdx);
return 0;
}
-   jdx = (uint32_t)(blist->bsize * ULP_INDEX_BITMAP_SIZE);
+   jdx = (uint32_t)(bsize_64 * ULP_INDEX_BITMAP_SIZE);
BNXT_TF_DBG(ERR, "bit allocator is full reached max:%x\n", jdx);
return -1;
 }
@@ -39,9 +40,10 @@ int32_t ulp_bit_alloc_list_dealloc(struct bit_alloc_list 
*blist,
   uint32_t index)
 {
uint32_t idx = 0, jdx;
+   uint32_t bsize_64 = blist->bsize / ULP_64B_IN_BYTES;
 
idx = index / ULP_INDEX_BITMAP_SIZE;
-   if (idx >= blist->bsize) {
+   if (idx >= bsize_64) {
BNXT_TF_DBG(ERR, "invalid bit index %x:%x\n", idx,
blist->bsize);
return -EINVAL;
@@ -127,7 +129,8 @@ ulp_gen_hash_tbl_list_init(struct ulp_hash_create_params 
*cparams,
hash_tbl->hash_mask = size - 1;
 
/* allocate the memory for the bit allocator */
-   size = (cparams->num_key_entries / sizeof(uint64_t)) + 1;
+   size = (cparams->num_key_entries / sizeof(uint64_t));
+   size = ULP_BYTE_ROUND_OFF_8(size);
hash_tbl->bit_list.bsize = size;
hash_tbl->bit_list.bdata = rte_zmalloc("Generic hash bit alloc", size,
   ULP_BUFFER_ALIGN_64_BYTE);
@@ -311,7 +314,12 @@ ulp_gen_hash_tbl_list_add(struct ulp_gen_hash_tbl 
*hash_tbl,
BNXT_TF_DBG(ERR, "Error in bit list alloc\n");
return -ENOMEM;
}
-
+   if (key_index > hash_tbl->num_key_entries) {
+   BNXT_TF_DBG(ERR, "reached max size %u:%u\n", key_index,
+   hash_tbl->num_key_entries);
+   ulp_bit_alloc_list_dealloc(&hash_tbl->bit_list, key_index);
+   return -ENOMEM;
+   }
/* Update the hash entry */
ULP_HASH_BUCKET_MARK_INUSE(bucket, (uint16_t)key_index);
 
-- 
2.17.1



[dpdk-dev] [PATCH v2 02/19] net/bnxt: add support for multi root capability

2021-10-26 Thread Venkat Duvvuru
From: Kishore Padmanabha 

Update driver to read the multi root capability and ignore
pci address check while creating ulp session when multi root
capability is enabled in the hardware. DPDK HSI version updated
from 1.10.1.70 to 1.10.2.54.

Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Michael Baucom 
Reviewed-by: Ajit Khaparde 
Reviewed-by: Randy Schacher 
---
 drivers/net/bnxt/bnxt.h|3 +
 drivers/net/bnxt/bnxt_hwrm.c   |8 +
 drivers/net/bnxt/hsi_struct_def_dpdk.h | 2682 +---
 drivers/net/bnxt/tf_ulp/bnxt_ulp.c |   10 +-
 4 files changed, 2407 insertions(+), 296 deletions(-)

diff --git a/drivers/net/bnxt/bnxt.h b/drivers/net/bnxt/bnxt.h
index 6743cf92b0..e3e38ffa19 100644
--- a/drivers/net/bnxt/bnxt.h
+++ b/drivers/net/bnxt/bnxt.h
@@ -723,6 +723,9 @@ struct bnxt {
uint16_tchip_num;
 #define CHIP_NUM_58818 0xd818
 #define BNXT_CHIP_SR2(bp)  ((bp)->chip_num == CHIP_NUM_58818)
+#defineBNXT_FLAGS2_MULTIROOT_ENBIT(4)
+#defineBNXT_MULTIROOT_EN(bp)   \
+   ((bp)->flags2 & BNXT_FLAGS2_MULTIROOT_EN)
 
uint32_tfw_cap;
 #define BNXT_FW_CAP_HOT_RESET  BIT(0)
diff --git a/drivers/net/bnxt/bnxt_hwrm.c b/drivers/net/bnxt/bnxt_hwrm.c
index 181e607d7b..c7041143a3 100644
--- a/drivers/net/bnxt/bnxt_hwrm.c
+++ b/drivers/net/bnxt/bnxt_hwrm.c
@@ -3394,6 +3394,7 @@ int bnxt_hwrm_parent_pf_qcfg(struct bnxt *bp)
 {
struct hwrm_func_qcfg_input req = {0};
struct hwrm_func_qcfg_output *resp = bp->hwrm_cmd_resp_addr;
+   uint16_t flags;
int rc;
 
if (!BNXT_VF_IS_TRUSTED(bp))
@@ -3417,6 +3418,13 @@ int bnxt_hwrm_parent_pf_qcfg(struct bnxt *bp)
bp->parent->fid = rte_le_to_cpu_16(resp->fid);
bp->parent->port_id = rte_le_to_cpu_16(resp->port_id);
 
+   flags = rte_le_to_cpu_16(resp->flags);
+   /* check for the multi-root support */
+   if (flags & HWRM_FUNC_QCFG_OUTPUT_FLAGS_MULTI_ROOT) {
+   bp->flags2 |= BNXT_FLAGS2_MULTIROOT_EN;
+   PMD_DRV_LOG(DEBUG, "PF enabled with multi root capability\n");
+   }
+
HWRM_UNLOCK();
 
return 0;
diff --git a/drivers/net/bnxt/hsi_struct_def_dpdk.h 
b/drivers/net/bnxt/hsi_struct_def_dpdk.h
index 4d7efb19f4..2a1e6ab97e 100644
--- a/drivers/net/bnxt/hsi_struct_def_dpdk.h
+++ b/drivers/net/bnxt/hsi_struct_def_dpdk.h
@@ -657,6 +657,8 @@ struct cmd_nums {
#define HWRM_FUNC_PTP_EXT_CFG UINT32_C(0x1a0)
/* PTP - Query extended PTP configuration. */
#define HWRM_FUNC_PTP_EXT_QCFGUINT32_C(0x1a1)
+   /* The command is used to allocate KTLS crypto key contexts. */
+   #define HWRM_FUNC_KEY_CTX_ALLOC   UINT32_C(0x1a2)
/* Experimental */
#define HWRM_SELFTEST_QLIST   UINT32_C(0x200)
/* Experimental */
@@ -1056,8 +1058,8 @@ struct hwrm_err_output {
 #define HWRM_VERSION_MINOR 10
 #define HWRM_VERSION_UPDATE 2
 /* non-zero means beta version */
-#define HWRM_VERSION_RSVD 44
-#define HWRM_VERSION_STR "1.10.2.44"
+#define HWRM_VERSION_RSVD 54
+#define HWRM_VERSION_STR "1.10.2.54"
 
 /
  * hwrm_ver_get *
@@ -1357,6 +1359,12 @@ struct hwrm_ver_get_output {
 */
#define HWRM_VER_GET_OUTPUT_DEV_CAPS_CFG_CFA_TRUFLOW_SUPPORTED \
UINT32_C(0x4000)
+   /*
+* If set to 1, then firmware supports secure boot.
+* If set to 0, then firmware doesn't support secure boot.
+*/
+   #define HWRM_VER_GET_OUTPUT_DEV_CAPS_CFG_SECURE_BOOT_CAPABLE \
+   UINT32_C(0x8000)
/*
 * This field represents the major version of RoCE firmware.
 * A change in major version represents a major release.
@@ -8283,8 +8291,14 @@ struct hwrm_async_event_cmpl_reset_notify {
/* Fast reset */
#define 
HWRM_ASYNC_EVENT_CMPL_RESET_NOTIFY_EVENT_DATA1_REASON_CODE_FAST_RESET \
(UINT32_C(0x4) << 8)
+   /*
+* Reset was a result of a firmware activation. That is, the
+* fw_activation flag was set in a FW_RESET operation.
+*/
+   #define 
HWRM_ASYNC_EVENT_CMPL_RESET_NOTIFY_EVENT_DATA1_REASON_CODE_FW_ACTIVATION \
+   (UINT32_C(0x5) << 8)
#define HWRM_ASYNC_EVENT_CMPL_RESET_NOTIFY_EVENT_DATA1_REASON_CODE_LAST 
\
-   
HWRM_ASYNC_EVENT_CMPL_RESET_NOTIFY_EVENT_DATA1_REASON_CODE_FAST_RESET
+   
HWRM_ASYNC_EVENT_CMPL_RESET_NOTIFY_EVENT_DATA1_REASON_CODE_FW_ACTIVATION
/*
 * Minimum time before driver should attempt access - units 100ms ticks.
 * Range 0-65535
@@ -10244,8 +10258,21 @@ struct hwrm_async_event_cmpl_error_report_base {
 */
#define 
HWRM_ASYNC_EVENT_CMPL_ERROR_REPORT_BASE_EVENT_DATA1_ERROR_TYPE_INVALID_SIGNAL \
UINT32_C(0x2)
+   /*
+* There wa

[dpdk-dev] [PATCH v2 05/19] net/bnxt: add capability option for socket redirect

2021-10-26 Thread Venkat Duvvuru
From: Kishore Padmanabha 

Added support for socket redirect feature capability so applications
can enable or disable this feature. This patch contains the template
changes.

Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Michael Baucom 
Reviewed-by: Randy Schacher 
---
 drivers/net/bnxt/bnxt.h   |  13 --
 drivers/net/bnxt/bnxt_ethdev.c| 203 --
 drivers/net/bnxt/tf_ulp/bnxt_ulp.c|  11 +-
 drivers/net/bnxt/tf_ulp/bnxt_ulp.h|   6 +-
 drivers/net/bnxt/tf_ulp/bnxt_ulp_flow.c   |  10 +-
 .../generic_templates/ulp_template_db_enum.h  |   8 +-
 .../generic_templates/ulp_template_db_tbl.c   |   5 +-
 7 files changed, 30 insertions(+), 226 deletions(-)

diff --git a/drivers/net/bnxt/bnxt.h b/drivers/net/bnxt/bnxt.h
index e3e38ffa19..c65e360446 100644
--- a/drivers/net/bnxt/bnxt.h
+++ b/drivers/net/bnxt/bnxt.h
@@ -1054,19 +1054,6 @@ int32_t
 bnxt_ulp_create_vfr_default_rules(struct rte_eth_dev *vfr_ethdev);
 int32_t
 bnxt_ulp_delete_vfr_default_rules(struct bnxt_representor *vfr);
-void bnxt_get_iface_mac(uint16_t port, enum bnxt_ulp_intf_type type,
-   uint8_t *mac, uint8_t *parent_mac);
-uint16_t bnxt_get_vnic_id(uint16_t port, enum bnxt_ulp_intf_type type);
-uint16_t bnxt_get_parent_vnic_id(uint16_t port, enum bnxt_ulp_intf_type type);
-struct bnxt *bnxt_get_bp(uint16_t port);
-uint16_t bnxt_get_svif(uint16_t port_id, bool func_svif,
-  enum bnxt_ulp_intf_type type);
-uint16_t bnxt_get_fw_func_id(uint16_t port, enum bnxt_ulp_intf_type type);
-uint16_t bnxt_get_parif(uint16_t port, enum bnxt_ulp_intf_type type);
-uint16_t bnxt_get_phy_port_id(uint16_t port);
-uint16_t bnxt_get_vport(uint16_t port);
-enum bnxt_ulp_intf_type
-bnxt_get_interface_type(uint16_t port);
 int bnxt_rep_dev_start_op(struct rte_eth_dev *eth_dev);
 
 void bnxt_cancel_fc_thread(struct bnxt *bp);
diff --git a/drivers/net/bnxt/bnxt_ethdev.c b/drivers/net/bnxt/bnxt_ethdev.c
index f385723a9f..bf034db336 100644
--- a/drivers/net/bnxt/bnxt_ethdev.c
+++ b/drivers/net/bnxt/bnxt_ethdev.c
@@ -5056,209 +5056,6 @@ static void bnxt_config_vf_req_fwd(struct bnxt *bp)
BNXT_HWRM_CMD_TO_FORWARD(HWRM_OEM_CMD);
 }
 
-struct bnxt *
-bnxt_get_bp(uint16_t port)
-{
-   struct bnxt *bp;
-   struct rte_eth_dev *dev;
-
-   if (!rte_eth_dev_is_valid_port(port)) {
-   PMD_DRV_LOG(ERR, "Invalid port %d\n", port);
-   return NULL;
-   }
-
-   dev = &rte_eth_devices[port];
-   if (!is_bnxt_supported(dev)) {
-   PMD_DRV_LOG(ERR, "Device %d not supported\n", port);
-   return NULL;
-   }
-
-   bp = (struct bnxt *)dev->data->dev_private;
-   if (!BNXT_TRUFLOW_EN(bp)) {
-   PMD_DRV_LOG(ERR, "TRUFLOW not enabled\n");
-   return NULL;
-   }
-
-   return bp;
-}
-
-uint16_t
-bnxt_get_svif(uint16_t port_id, bool func_svif,
- enum bnxt_ulp_intf_type type)
-{
-   struct rte_eth_dev *eth_dev;
-   struct bnxt *bp;
-
-   eth_dev = &rte_eth_devices[port_id];
-   if (BNXT_ETH_DEV_IS_REPRESENTOR(eth_dev)) {
-   struct bnxt_representor *vfr = eth_dev->data->dev_private;
-   if (!vfr)
-   return 0;
-
-   if (type == BNXT_ULP_INTF_TYPE_VF_REP)
-   return vfr->svif;
-
-   eth_dev = vfr->parent_dev;
-   }
-
-   bp = eth_dev->data->dev_private;
-
-   return func_svif ? bp->func_svif : bp->port_svif;
-}
-
-void
-bnxt_get_iface_mac(uint16_t port, enum bnxt_ulp_intf_type type,
-  uint8_t *mac, uint8_t *parent_mac)
-{
-   struct rte_eth_dev *eth_dev;
-   struct bnxt *bp;
-
-   if (type != BNXT_ULP_INTF_TYPE_TRUSTED_VF &&
-   type != BNXT_ULP_INTF_TYPE_PF)
-   return;
-
-   eth_dev = &rte_eth_devices[port];
-   bp = eth_dev->data->dev_private;
-   memcpy(mac, bp->mac_addr, RTE_ETHER_ADDR_LEN);
-
-   if (type == BNXT_ULP_INTF_TYPE_TRUSTED_VF)
-   memcpy(parent_mac, bp->parent->mac_addr, RTE_ETHER_ADDR_LEN);
-}
-
-uint16_t
-bnxt_get_parent_vnic_id(uint16_t port, enum bnxt_ulp_intf_type type)
-{
-   struct rte_eth_dev *eth_dev;
-   struct bnxt *bp;
-
-   if (type != BNXT_ULP_INTF_TYPE_TRUSTED_VF)
-   return 0;
-
-   eth_dev = &rte_eth_devices[port];
-   bp = eth_dev->data->dev_private;
-
-   return bp->parent->vnic;
-}
-uint16_t
-bnxt_get_vnic_id(uint16_t port, enum bnxt_ulp_intf_type type)
-{
-   struct rte_eth_dev *eth_dev;
-   struct bnxt_vnic_info *vnic;
-   struct bnxt *bp;
-
-   eth_dev = &rte_eth_devices[port];
-   if (BNXT_ETH_DEV_IS_REPRESENTOR(eth_dev)) {
-   struct bnxt_representor *vfr = eth_dev->data->dev_private;
-   if (!vfr)
-   return 0;
-
-   if (type == BNXT_ULP_INTF_TYPE_VF_REP)
-   r

[dpdk-dev] [PATCH v2 04/19] net/bnxt: add clear on read support

2021-10-26 Thread Venkat Duvvuru
From: Farah Smith 

Add clear on read stats support for Thor. Currently, the
flow stats are not cleared after they are read from the FW.
This patch adds support for clear on read. Since clear on
read support is added for flow stats in Thor, the flow
accumulation is enabled on Thor as well.

Signed-off-by: Farah Smith 
Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Randy Schacher 
Reviewed-by: Shahaji Bhosle 
---
 drivers/net/bnxt/hsi_struct_def_dpdk.h | 34 --
 drivers/net/bnxt/tf_core/tf_msg.c  | 24 +++---
 drivers/net/bnxt/tf_core/tf_msg.h  |  6 +++--
 drivers/net/bnxt/tf_core/tf_tbl.c  |  6 +++--
 drivers/net/bnxt/tf_core/tf_tbl_sram.c | 13 --
 drivers/net/bnxt/tf_ulp/bnxt_ulp.c |  3 ++-
 6 files changed, 67 insertions(+), 19 deletions(-)

diff --git a/drivers/net/bnxt/hsi_struct_def_dpdk.h 
b/drivers/net/bnxt/hsi_struct_def_dpdk.h
index 2a1e6ab97e..77981b5cdb 100644
--- a/drivers/net/bnxt/hsi_struct_def_dpdk.h
+++ b/drivers/net/bnxt/hsi_struct_def_dpdk.h
@@ -1058,8 +1058,8 @@ struct hwrm_err_output {
 #define HWRM_VERSION_MINOR 10
 #define HWRM_VERSION_UPDATE 2
 /* non-zero means beta version */
-#define HWRM_VERSION_RSVD 54
-#define HWRM_VERSION_STR "1.10.2.54"
+#define HWRM_VERSION_RSVD 55
+#define HWRM_VERSION_STR "1.10.2.55"
 
 /
  * hwrm_ver_get *
@@ -46253,13 +46253,22 @@ struct hwrm_tf_tbl_type_get_input {
/* Control flags. */
uint16_tflags;
/* Indicates the flow direction. */
-   #define HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR UINT32_C(0x1)
+   #define HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR \
+   UINT32_C(0x1)
/* If this bit set to 0, then it indicates rx flow. */
-   #define HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR_RXUINT32_C(0x0)
+   #define HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR_RX \
+   UINT32_C(0x0)
/* If this bit is set to 1, then it indicates tx flow. */
-   #define HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR_TXUINT32_C(0x1)
+   #define HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR_TX \
+   UINT32_C(0x1)
#define HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR_LAST \
HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR_TX
+   /*
+* When set use the special access register access to clear
+* the table entry on read.
+*/
+   #define HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_CLEAR_ON_READ \
+   UINT32_C(0x2)
/* unused. */
uint8_t unused0[2];
/*
@@ -48489,13 +48498,22 @@ struct hwrm_tf_tbl_type_bulk_get_input {
/* Control flags. */
uint16_tflags;
/* Indicates the flow direction. */
-   #define HWRM_TF_TBL_TYPE_BULK_GET_INPUT_FLAGS_DIR UINT32_C(0x1)
+   #define HWRM_TF_TBL_TYPE_BULK_GET_INPUT_FLAGS_DIR \
+   UINT32_C(0x1)
/* If this bit set to 0, then it indicates rx flow. */
-   #define HWRM_TF_TBL_TYPE_BULK_GET_INPUT_FLAGS_DIR_RXUINT32_C(0x0)
+   #define HWRM_TF_TBL_TYPE_BULK_GET_INPUT_FLAGS_DIR_RX \
+   UINT32_C(0x0)
/* If this bit is set to 1, then it indicates tx flow. */
-   #define HWRM_TF_TBL_TYPE_BULK_GET_INPUT_FLAGS_DIR_TXUINT32_C(0x1)
+   #define HWRM_TF_TBL_TYPE_BULK_GET_INPUT_FLAGS_DIR_TX \
+   UINT32_C(0x1)
#define HWRM_TF_TBL_TYPE_BULK_GET_INPUT_FLAGS_DIR_LAST \
HWRM_TF_TBL_TYPE_BULK_GET_INPUT_FLAGS_DIR_TX
+   /*
+* When set use the special access register access to clear
+* the table entries on read.
+*/
+   #define HWRM_TF_TBL_TYPE_BULK_GET_INPUT_FLAGS_CLEAR_ON_READ \
+   UINT32_C(0x2)
/* unused. */
uint8_t unused0[2];
/*
diff --git a/drivers/net/bnxt/tf_core/tf_msg.c 
b/drivers/net/bnxt/tf_core/tf_msg.c
index 0fbb2fe837..ea6e2af7ce 100644
--- a/drivers/net/bnxt/tf_core/tf_msg.c
+++ b/drivers/net/bnxt/tf_core/tf_msg.c
@@ -1851,7 +1851,8 @@ tf_msg_get_tbl_entry(struct tf *tfp,
 uint16_t hcapi_type,
 uint16_t size,
 uint8_t *data,
-uint32_t index)
+uint32_t index,
+bool clear_on_read)
 {
int rc;
struct hwrm_tf_tbl_type_get_input req = { 0 };
@@ -1860,6 +1861,7 @@ tf_msg_get_tbl_entry(struct tf *tfp,
uint8_t fw_session_id;
struct tf_dev_info *dev;
struct tf_session *tfs;
+   uint32_t flags = 0;
 
/* Retrieve the session information */
rc = tf_session_get_session_internal(tfp, &tfs);
@@ -1889,10 +1891,16 @@ tf_msg_get_tbl_entry(struct tf *tfp,
strerror(-rc));
return rc;
}
+   flags = (dir == TF_DIR_TX ?
+HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR_TX :
+HWRM_TF_TBL_TYPE_GET_INPUT_FLAGS_DIR_RX);
+
+   if (clear_on_read)
+   flags |= HWRM_TF_TBL_TYPE_G

[dpdk-dev] [PATCH v2 10/19] net/bnxt: remove accumulation of stats devargs argument

2021-10-26 Thread Venkat Duvvuru
From: Kishore Padmanabha 

The accumulation of flow counters is not determined by the
application device arguments instead it is dictated by the platform
capabilities whether to do software based accumulation or not.

Change-Id: I7275a0ed6ab089358e0a95f43d6291ec65ebe030
Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Ajit Khaparde 
Reviewed-by: Michael Baucom 
---
 drivers/net/bnxt/bnxt.h  |  3 --
 drivers/net/bnxt/bnxt_ethdev.c   | 53 
 drivers/net/bnxt/tf_ulp/bnxt_ulp.c   |  8 -
 drivers/net/bnxt/tf_ulp/bnxt_ulp.h   |  1 -
 drivers/net/bnxt/tf_ulp/ulp_fc_mgr.c | 12 +++
 5 files changed, 4 insertions(+), 73 deletions(-)

diff --git a/drivers/net/bnxt/bnxt.h b/drivers/net/bnxt/bnxt.h
index 5c064fd119..5cc4a5afae 100644
--- a/drivers/net/bnxt/bnxt.h
+++ b/drivers/net/bnxt/bnxt.h
@@ -718,11 +718,8 @@ struct bnxt {
uint32_tflags2;
 #define BNXT_FLAGS2_PTP_TIMESYNC_ENABLED   BIT(0)
 #define BNXT_FLAGS2_PTP_ALARM_SCHEDULEDBIT(1)
-#defineBNXT_FLAGS2_ACCUM_STATS_EN  BIT(2)
 #define BNXT_P5_PTP_TIMESYNC_ENABLED(bp)   \
((bp)->flags2 & BNXT_FLAGS2_PTP_TIMESYNC_ENABLED)
-#defineBNXT_ACCUM_STATS_EN(bp) \
-   ((bp)->flags2 & BNXT_FLAGS2_ACCUM_STATS_EN)
 
uint16_tchip_num;
 #define CHIP_NUM_58818 0xd818
diff --git a/drivers/net/bnxt/bnxt_ethdev.c b/drivers/net/bnxt/bnxt_ethdev.c
index 8c72ab8fd1..070f69d8e9 100644
--- a/drivers/net/bnxt/bnxt_ethdev.c
+++ b/drivers/net/bnxt/bnxt_ethdev.c
@@ -87,7 +87,6 @@ static const struct rte_pci_id bnxt_pci_id_map[] = {
{ .vendor_id = 0, /* sentinel */ },
 };
 
-#defineBNXT_DEVARG_ACCUM_STATS "accum-stats"
 #define BNXT_DEVARG_FLOW_XSTAT "flow-xstat"
 #define BNXT_DEVARG_MAX_NUM_KFLOWS  "max-num-kflows"
 #define BNXT_DEVARG_REPRESENTOR"representor"
@@ -101,7 +100,6 @@ static const struct rte_pci_id bnxt_pci_id_map[] = {
 
 static const char *const bnxt_dev_args[] = {
BNXT_DEVARG_REPRESENTOR,
-   BNXT_DEVARG_ACCUM_STATS,
BNXT_DEVARG_FLOW_XSTAT,
BNXT_DEVARG_MAX_NUM_KFLOWS,
BNXT_DEVARG_REP_BASED_PF,
@@ -114,12 +112,6 @@ static const char *const bnxt_dev_args[] = {
NULL
 };
 
-/*
- * accum-stats == false to disable flow counter accumulation
- * accum-stats == true to enable flow counter accumulation
- */
-#defineBNXT_DEVARG_ACCUM_STATS_INVALID(accum_stats)((accum_stats) 
> 1)
-
 /*
  * app-id = an non-negative 8-bit number
  */
@@ -5299,45 +5291,6 @@ static int bnxt_init_resources(struct bnxt *bp, bool 
reconfig_dev)
return 0;
 }
 
-static int
-bnxt_parse_devarg_accum_stats(__rte_unused const char *key,
- const char *value, void *opaque_arg)
-{
-   struct bnxt *bp = opaque_arg;
-   unsigned long accum_stats;
-   char *end = NULL;
-
-   if (!value || !opaque_arg) {
-   PMD_DRV_LOG(ERR,
-   "Invalid parameter passed to accum-stats 
devargs.\n");
-   return -EINVAL;
-   }
-
-   accum_stats = strtoul(value, &end, 10);
-   if (end == NULL || *end != '\0' ||
-   (accum_stats == ULONG_MAX && errno == ERANGE)) {
-   PMD_DRV_LOG(ERR,
-   "Invalid parameter passed to accum-stats 
devargs.\n");
-   return -EINVAL;
-   }
-
-   if (BNXT_DEVARG_ACCUM_STATS_INVALID(accum_stats)) {
-   PMD_DRV_LOG(ERR,
-   "Invalid value passed to accum-stats devargs.\n");
-   return -EINVAL;
-   }
-
-   if (accum_stats) {
-   bp->flags2 |= BNXT_FLAGS2_ACCUM_STATS_EN;
-   PMD_DRV_LOG(INFO, "Host-based accum-stats feature enabled.\n");
-   } else {
-   bp->flags2 &= ~BNXT_FLAGS2_ACCUM_STATS_EN;
-   PMD_DRV_LOG(INFO, "Host-based accum-stats feature disabled.\n");
-   }
-
-   return 0;
-}
-
 static int
 bnxt_parse_devarg_flow_xstat(__rte_unused const char *key,
 const char *value, void *opaque_arg)
@@ -5690,12 +5643,6 @@ bnxt_parse_dev_args(struct bnxt *bp, struct rte_devargs 
*devargs)
if (ret)
goto err;
 
-   /*
-* Handler for "accum-stats" devarg.
-* Invoked as for ex: "-a :00:0d.0,accum-stats=1"
-*/
-   rte_kvargs_process(kvlist, BNXT_DEVARG_ACCUM_STATS,
-  bnxt_parse_devarg_accum_stats, bp);
/*
 * Handler for "max_num_kflows" devarg.
 * Invoked as for ex: "-a 000:00:0d.0,max_num_kflows=32"
diff --git a/drivers/net/bnxt/tf_ulp/bnxt_ulp.c 
b/drivers/net/bnxt/tf_ulp/bnxt_ulp.c
index 2ac1a8625f..7deacd1f3e 100644
--- a/drivers/net/bnxt/tf_ulp/bnxt_ulp.c
+++ b/drivers/net/bnxt/tf_ulp/bnxt_ulp.c
@@ -1490,14 +1490,6 @@ bnxt_ulp_port_init(struct bnxt *bp)
goto jump_to_error;
}
 
-

[dpdk-dev] [PATCH v2 08/19] net/bnxt: get TruFlow version

2021-10-26 Thread Venkat Duvvuru
From: Jay Ding 

Implement tf_get_version that returns TruFlow version
numbers and CFA resources capbilities.

Signed-off-by: Jay Ding 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Farah Smith 
Reviewed-by: Randy Schacher 
Reviewed-by: Peter Spreadborough 
---
 drivers/net/bnxt/hsi_struct_def_dpdk.h   | 15 ++--
 drivers/net/bnxt/tf_core/tf_core.c   | 29 
 drivers/net/bnxt/tf_core/tf_core.h   | 75 
 drivers/net/bnxt/tf_core/tf_device.h | 50 ++
 drivers/net/bnxt/tf_core/tf_device_p4.c  | 65 +-
 drivers/net/bnxt/tf_core/tf_device_p4.h  | 79 +
 drivers/net/bnxt/tf_core/tf_device_p58.c | 65 +-
 drivers/net/bnxt/tf_core/tf_device_p58.h | 87 
 drivers/net/bnxt/tf_core/tf_msg.c| 35 ++
 drivers/net/bnxt/tf_core/tf_msg.h| 19 ++
 10 files changed, 513 insertions(+), 6 deletions(-)

diff --git a/drivers/net/bnxt/hsi_struct_def_dpdk.h 
b/drivers/net/bnxt/hsi_struct_def_dpdk.h
index 77981b5cdb..6d04e39e48 100644
--- a/drivers/net/bnxt/hsi_struct_def_dpdk.h
+++ b/drivers/net/bnxt/hsi_struct_def_dpdk.h
@@ -1058,8 +1058,8 @@ struct hwrm_err_output {
 #define HWRM_VERSION_MINOR 10
 #define HWRM_VERSION_UPDATE 2
 /* non-zero means beta version */
-#define HWRM_VERSION_RSVD 55
-#define HWRM_VERSION_STR "1.10.2.55"
+#define HWRM_VERSION_RSVD 58
+#define HWRM_VERSION_STR "1.10.2.58"
 
 /
  * hwrm_ver_get *
@@ -45178,7 +45178,7 @@ struct hwrm_tf_version_get_input {
uint64_tresp_addr;
 } __rte_packed;
 
-/* hwrm_tf_version_get_output (size:128b/16B) */
+/* hwrm_tf_version_get_output (size:256b/32B) */
 struct hwrm_tf_version_get_output {
/* The specific error status for the command. */
uint16_terror_code;
@@ -45195,7 +45195,14 @@ struct hwrm_tf_version_get_output {
/* Version Update number. */
uint8_t update;
/* unused. */
-   uint8_t unused0[4];
+   uint8_t unused0[5];
+   /*
+* This field is used to indicate device's capabilities and
+* configurations.
+*/
+   uint64_tdev_caps_cfg;
+   /* unused. */
+   uint8_t unused1[7];
/*
 * This field is used in Output records to indicate that the output
 * is completely written to RAM. This field should be read as '1'
diff --git a/drivers/net/bnxt/tf_core/tf_core.c 
b/drivers/net/bnxt/tf_core/tf_core.c
index 936102c804..86dfec0eb4 100644
--- a/drivers/net/bnxt/tf_core/tf_core.c
+++ b/drivers/net/bnxt/tf_core/tf_core.c
@@ -1802,3 +1802,32 @@ int tf_get_session_info(struct tf *tfp,
 
return 0;
 }
+
+int tf_get_version(struct tf *tfp,
+  struct tf_get_version_parms *parms)
+{
+   int rc;
+   struct tf_dev_info dev;
+
+   TF_CHECK_PARMS2(tfp, parms);
+
+   /* This function can be called before open session, filter
+* out any non-supported device types on the Core side.
+*/
+   if (parms->device_type != TF_DEVICE_TYPE_WH &&
+   parms->device_type != TF_DEVICE_TYPE_THOR &&
+   parms->device_type != TF_DEVICE_TYPE_SR) {
+   TFP_DRV_LOG(ERR,
+   "Unsupported device type %d\n",
+   parms->device_type);
+   return -ENOTSUP;
+   }
+
+   tf_dev_bind_ops(parms->device_type, &dev);
+
+   rc = tf_msg_get_version(parms->bp, &dev, parms);
+   if (rc)
+   return rc;
+
+   return 0;
+}
diff --git a/drivers/net/bnxt/tf_core/tf_core.h 
b/drivers/net/bnxt/tf_core/tf_core.h
index fb02c2b161..ba9881c69d 100644
--- a/drivers/net/bnxt/tf_core/tf_core.h
+++ b/drivers/net/bnxt/tf_core/tf_core.h
@@ -2363,4 +2363,79 @@ struct tf_get_if_tbl_entry_parms {
 int tf_get_if_tbl_entry(struct tf *tfp,
struct tf_get_if_tbl_entry_parms *parms);
 
+/**
+ * tf_get_version parameters definition.
+ */
+struct tf_get_version_parms {
+   /**
+* [in] device type
+*
+* Device type for the session.
+*/
+   enum tf_device_type device_type;
+
+   /**
+* [in] bp
+* The pointer to the parent bp struct. This is only used for HWRM
+* message passing within the portability layer. The type is struct
+* bnxt.
+*/
+   void *bp;
+
+   /* [out] major
+*
+* Version Major number.
+*/
+   uint8_t major;
+
+   /* [out] minor
+*
+* Version Minor number.
+*/
+   uint8_t minor;
+
+   /* [out] update
+*
+* Version Update number.
+*/
+   uint8_t update;
+
+   /**
+* [out] dev_ident_caps
+*
+* fw available identifier resource list
+*/
+   uint32_t dev_ident_caps;
+
+   /**
+* [out] dev_tbl_caps
+*
+* fw available table resource list
+*/
+   uint32_t dev_tbl_caps;
+
+   /**
+* [o

[dpdk-dev] [PATCH v2 07/19] net/bnxt: support inner IP header for GRE tunnel flows

2021-10-26 Thread Venkat Duvvuru
From: Kishore Padmanabha 

This change allows adding IP header matches for GRE flows that
does not specify outer IP header in the flow match pattern.

Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Michael Baucom 
Reviewed-by: Ajit Khaparde 
Reviewed-by: Randy Schacher 
---
 drivers/net/bnxt/tf_ulp/ulp_rte_parser.c  | 7 ---
 drivers/net/bnxt/tf_ulp/ulp_template_struct.h | 2 +-
 2 files changed, 5 insertions(+), 4 deletions(-)

diff --git a/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c 
b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c
index 40da953f06..605c29223c 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c
+++ b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c
@@ -1119,7 +1119,8 @@ ulp_rte_ipv4_hdr_handler(const struct rte_flow_item *item,
 
/* Set the ipv4 header bitmap and computed l3 header bitmaps */
if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV4) ||
-   ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV6)) {
+   ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV6) ||
+   ULP_COMP_FLD_IDX_RD(params, BNXT_ULP_CF_IDX_L3_TUN)) {
ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_IPV4);
ULP_COMP_FLD_IDX_WR(params, BNXT_ULP_CF_IDX_I_L3, 1);
inner_flag = 1;
@@ -1245,7 +1246,8 @@ ulp_rte_ipv6_hdr_handler(const struct rte_flow_item *item,
 
/* Set the ipv6 header bitmap and computed l3 header bitmaps */
if (ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV4) ||
-   ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV6)) {
+   ULP_BITMAP_ISSET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_O_IPV6) ||
+   ULP_COMP_FLD_IDX_RD(params, BNXT_ULP_CF_IDX_L3_TUN)) {
ULP_BITMAP_SET(hdr_bitmap->bits, BNXT_ULP_HDR_BIT_I_IPV6);
ULP_COMP_FLD_IDX_WR(params, BNXT_ULP_CF_IDX_I_L3, 1);
inner_flag = 1;
@@ -2127,7 +2129,6 @@ ulp_rte_vxlan_decap_act_handler(const struct 
rte_flow_action *action_item
   BNXT_ULP_ACT_BIT_VXLAN_DECAP);
/* Update computational field with tunnel decap info */
ULP_COMP_FLD_IDX_WR(params, BNXT_ULP_CF_IDX_L3_TUN_DECAP, 1);
-   ULP_COMP_FLD_IDX_WR(params, BNXT_ULP_CF_IDX_L3_TUN, 1);
return BNXT_TF_RC_SUCCESS;
 }
 
diff --git a/drivers/net/bnxt/tf_ulp/ulp_template_struct.h 
b/drivers/net/bnxt/tf_ulp/ulp_template_struct.h
index d3bfb8c12d..7d1bc06a3e 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_template_struct.h
+++ b/drivers/net/bnxt/tf_ulp/ulp_template_struct.h
@@ -27,7 +27,7 @@
 #define BNXT_ULP_PROTO_HDR_UDP_NUM 4
 #define BNXT_ULP_PROTO_HDR_TCP_NUM 9
 #define BNXT_ULP_PROTO_HDR_VXLAN_NUM   4
-#define BNXT_ULP_PROTO_HDR_GRE_NUM 6
+#define BNXT_ULP_PROTO_HDR_GRE_NUM 2
 #define BNXT_ULP_PROTO_HDR_ICMP_NUM5
 #define BNXT_ULP_PROTO_HDR_MAX 128
 #define BNXT_ULP_PROTO_HDR_ENCAP_MAX   64
-- 
2.17.1



[dpdk-dev] [PATCH v2 12/19] net/bnxt: updated the log messages

2021-10-26 Thread Venkat Duvvuru
From: Kishore Padmanabha 

Some of the error level log messages are made debug level messages.
When Truflow is not enabled then Truflow init error messages are
moved to debug level instead.

Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Shahaji Bhosle 
Reviewed-by: Michael Baucom 
Reviewed-by: Ajit Khaparde 
---
 drivers/net/bnxt/tf_ulp/bnxt_ulp.c  | 24 
 drivers/net/bnxt/tf_ulp/ulp_def_rules.c |  2 +-
 drivers/net/bnxt/tf_ulp/ulp_flow_db.c   |  2 +-
 drivers/net/bnxt/tf_ulp/ulp_mark_mgr.c  |  2 +-
 4 files changed, 15 insertions(+), 15 deletions(-)

diff --git a/drivers/net/bnxt/tf_ulp/bnxt_ulp.c 
b/drivers/net/bnxt/tf_ulp/bnxt_ulp.c
index 7deacd1f3e..21f71d6445 100644
--- a/drivers/net/bnxt/tf_ulp/bnxt_ulp.c
+++ b/drivers/net/bnxt/tf_ulp/bnxt_ulp.c
@@ -1396,16 +1396,16 @@ bnxt_ulp_port_init(struct bnxt *bp)
uint32_t ulp_flags;
int32_t rc = 0;
 
-   if (!BNXT_PF(bp) && !BNXT_VF_IS_TRUSTED(bp)) {
-   BNXT_TF_DBG(ERR,
-   "Skip ulp init for port: %d, not a TVF or PF\n",
+   if (!BNXT_TRUFLOW_EN(bp)) {
+   BNXT_TF_DBG(DEBUG,
+   "Skip ulp init for port: %d, TF is not enabled\n",
bp->eth_dev->data->port_id);
return rc;
}
 
-   if (!BNXT_TRUFLOW_EN(bp)) {
-   BNXT_TF_DBG(ERR,
-   "Skip ulp init for port: %d, truflow is not 
enabled\n",
+   if (!BNXT_PF(bp) && !BNXT_VF_IS_TRUSTED(bp)) {
+   BNXT_TF_DBG(DEBUG,
+   "Skip ulp init for port: %d, not a TVF or PF\n",
bp->eth_dev->data->port_id);
return rc;
}
@@ -1520,16 +1520,16 @@ bnxt_ulp_port_deinit(struct bnxt *bp)
struct rte_pci_device *pci_dev;
struct rte_pci_addr *pci_addr;
 
-   if (!BNXT_PF(bp) && !BNXT_VF_IS_TRUSTED(bp)) {
-   BNXT_TF_DBG(ERR,
-   "Skip ULP deinit port:%d, not a TVF or PF\n",
+   if (!BNXT_TRUFLOW_EN(bp)) {
+   BNXT_TF_DBG(DEBUG,
+   "Skip ULP deinit for port:%d, TF is not enabled\n",
bp->eth_dev->data->port_id);
return;
}
 
-   if (!BNXT_TRUFLOW_EN(bp)) {
-   BNXT_TF_DBG(ERR,
-   "Skip ULP deinit for port:%d, truflow is not 
enabled\n",
+   if (!BNXT_PF(bp) && !BNXT_VF_IS_TRUSTED(bp)) {
+   BNXT_TF_DBG(DEBUG,
+   "Skip ULP deinit port:%d, not a TVF or PF\n",
bp->eth_dev->data->port_id);
return;
}
diff --git a/drivers/net/bnxt/tf_ulp/ulp_def_rules.c 
b/drivers/net/bnxt/tf_ulp/ulp_def_rules.c
index 01233c0f5e..8790d7ac0d 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_def_rules.c
+++ b/drivers/net/bnxt/tf_ulp/ulp_def_rules.c
@@ -323,7 +323,7 @@ ulp_default_flow_create(struct rte_eth_dev *eth_dev,
ulp_ctx = bnxt_ulp_eth_dev_ptr2_cntxt_get(eth_dev);
if (!ulp_ctx) {
BNXT_TF_DBG(ERR,
-   "ULP context is not initialized. Failed to create 
dflt flow.\n");
+   "ULP is not init'ed. Fail to create dflt flow.\n");
return -EINVAL;
}
 
diff --git a/drivers/net/bnxt/tf_ulp/ulp_flow_db.c 
b/drivers/net/bnxt/tf_ulp/ulp_flow_db.c
index 79dc869e64..9968311c44 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_flow_db.c
+++ b/drivers/net/bnxt/tf_ulp/ulp_flow_db.c
@@ -470,7 +470,7 @@ ulp_flow_db_init(struct bnxt_ulp_context *ulp_ctxt)
}
 
/* All good so return. */
-   BNXT_TF_DBG(INFO, "FlowDB initialized with %d flows.\n",
+   BNXT_TF_DBG(DEBUG, "FlowDB initialized with %d flows.\n",
flow_tbl->num_flows);
return 0;
 error_free:
diff --git a/drivers/net/bnxt/tf_ulp/ulp_mark_mgr.c 
b/drivers/net/bnxt/tf_ulp/ulp_mark_mgr.c
index 271520e1d3..9dffaef73b 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_mark_mgr.c
+++ b/drivers/net/bnxt/tf_ulp/ulp_mark_mgr.c
@@ -116,7 +116,7 @@ ulp_mark_db_init(struct bnxt_ulp_context *ctxt)
mark_tbl->gfid_mask = (mark_tbl->gfid_num_entries / 2) - 1;
mark_tbl->gfid_type_bit = (mark_tbl->gfid_num_entries / 2);
 
-   BNXT_TF_DBG(DEBUG, "GFID Max = 0x%08x\nGFID MASK = 0x%08x\n",
+   BNXT_TF_DBG(DEBUG, "GFID Max = 0x%08x GFID MASK = 0x%08x\n",
mark_tbl->gfid_num_entries - 1,
mark_tbl->gfid_mask);
 
-- 
2.17.1



[dpdk-dev] [PATCH v2 11/19] net/bnxt: fix clang compiler warnings

2021-10-26 Thread Venkat Duvvuru
From: Shahaji Bhosle 

Typecast flow_item type, action_item type and the ENUMs to uint32_t
before comparing.

Fixes: 53a0d4f7663 ("net/bnxt: support flow API item parsing")

Signed-off-by: Shahaji Bhosle 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Ajit Khaparde 
---
 drivers/net/bnxt/tf_ulp/ulp_rte_parser.c | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git a/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c 
b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c
index 605c29223c..d21c088d59 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c
+++ b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c
@@ -137,10 +137,10 @@ bnxt_ulp_rte_parser_hdr_parse(const struct rte_flow_item 
pattern[],
 
/* Parse all the items in the pattern */
while (item && item->type != RTE_FLOW_ITEM_TYPE_END) {
-   if (item->type >= (uint32_t)
+   if (item->type >= (typeof(item->type))
BNXT_RTE_FLOW_ITEM_TYPE_END) {
if (item->type >=
-   (uint32_t)BNXT_RTE_FLOW_ITEM_TYPE_LAST)
+   (typeof(item->type))BNXT_RTE_FLOW_ITEM_TYPE_LAST)
goto hdr_parser_error;
/* get the header information */
hdr_info = &ulp_vendor_hdr_info[item->type -
@@ -186,9 +186,9 @@ bnxt_ulp_rte_parser_act_parse(const struct rte_flow_action 
actions[],
/* Parse all the items in the pattern */
while (action_item && action_item->type != RTE_FLOW_ACTION_TYPE_END) {
if (action_item->type >=
-   (uint32_t)BNXT_RTE_FLOW_ACTION_TYPE_END) {
+   (typeof(action_item->type))BNXT_RTE_FLOW_ACTION_TYPE_END) {
if (action_item->type >=
-   (uint32_t)BNXT_RTE_FLOW_ACTION_TYPE_LAST)
+   
(typeof(action_item->type))BNXT_RTE_FLOW_ACTION_TYPE_LAST)
goto act_parser_error;
/* get the header information from bnxt actinfo table */
hdr_info = &ulp_vendor_act_info[action_item->type -
-- 
2.17.1



[dpdk-dev] [PATCH v2 17/19] net/bnxt: add Tx TruFlow table config for p4

2021-10-26 Thread Venkat Duvvuru
From: Jay Ding 

Add TX direction TruFlow table type config to be
compatible with other devices. For P4, the TX cfg
is duplicated from RX.

Signed-off-by: Jay Ding 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Farah Smith 
Reviewed-by: Randy Schacher 
---
 drivers/net/bnxt/tf_core/tf_device.c|   4 +-
 drivers/net/bnxt/tf_core/tf_device_p4.c | 107 
 drivers/net/bnxt/tf_core/tf_device_p4.h |  58 +
 3 files changed, 111 insertions(+), 58 deletions(-)

diff --git a/drivers/net/bnxt/tf_core/tf_device.c 
b/drivers/net/bnxt/tf_core/tf_device.c
index 40db546604..4c416270b6 100644
--- a/drivers/net/bnxt/tf_core/tf_device.c
+++ b/drivers/net/bnxt/tf_core/tf_device.c
@@ -131,11 +131,11 @@ tf_dev_bind_p4(struct tf *tfp,
}
 
rsv_cnt = tf_dev_reservation_check(TF_TBL_TYPE_MAX,
-  tf_tbl_p4,
+  tf_tbl_p4[TF_DIR_RX],
   (uint16_t *)resources->tbl_cnt);
if (rsv_cnt) {
tbl_cfg.num_elements = TF_TBL_TYPE_MAX;
-   tbl_cfg.cfg = tf_tbl_p4;
+   tbl_cfg.cfg = tf_tbl_p4[TF_DIR_RX];
tbl_cfg.resources = resources;
rc = tf_tbl_bind(tfp, &tbl_cfg);
if (rc) {
diff --git a/drivers/net/bnxt/tf_core/tf_device_p4.c 
b/drivers/net/bnxt/tf_core/tf_device_p4.c
index 244bd08914..a6a59b8a07 100644
--- a/drivers/net/bnxt/tf_core/tf_device_p4.c
+++ b/drivers/net/bnxt/tf_core/tf_device_p4.c
@@ -59,6 +59,113 @@ const char *tf_resource_str_p4[CFA_RESOURCE_TYPE_P4_LAST + 
1] = {
[CFA_RESOURCE_TYPE_P4_TBL_SCOPE] = "tb_scope",
 };
 
+struct tf_rm_element_cfg tf_tbl_p4[TF_DIR_MAX][TF_TBL_TYPE_MAX] = {
+   [TF_DIR_RX][TF_TBL_TYPE_FULL_ACT_RECORD] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_FULL_ACTION,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_MCAST_GROUPS] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_MCG,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_ACT_ENCAP_8B] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_ENCAP_8B,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_ACT_ENCAP_16B] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_ENCAP_16B,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_ACT_ENCAP_64B] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_ENCAP_64B,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_ACT_SP_SMAC] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_SP_MAC,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_ACT_SP_SMAC_IPV4] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_SP_MAC_IPV4,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_ACT_SP_SMAC_IPV6] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_SP_MAC_IPV6,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_ACT_STATS_64] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_COUNTER_64B,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_ACT_MODIFY_IPV4] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_NAT_IPV4,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_METER_PROF] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_METER_PROF,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_METER_INST] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_METER,
+   0, 0
+   },
+   [TF_DIR_RX][TF_TBL_TYPE_MIRROR_CONFIG] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_MIRROR,
+   0, 0
+   },
+   [TF_DIR_TX][TF_TBL_TYPE_FULL_ACT_RECORD] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_FULL_ACTION,
+   0, 0
+   },
+   [TF_DIR_TX][TF_TBL_TYPE_MCAST_GROUPS] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_MCG,
+   0, 0
+   },
+   [TF_DIR_TX][TF_TBL_TYPE_ACT_ENCAP_8B] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_ENCAP_8B,
+   0, 0
+   },
+   [TF_DIR_TX][TF_TBL_TYPE_ACT_ENCAP_16B] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_ENCAP_16B,
+   0, 0
+   },
+   [TF_DIR_TX][TF_TBL_TYPE_ACT_ENCAP_64B] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_ENCAP_64B,
+   0, 0
+   },
+   [TF_DIR_TX][TF_TBL_TYPE_ACT_SP_SMAC] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_SP_MAC,
+   0, 0
+   },
+   [TF_DIR_TX][TF_TBL_TYPE_ACT_SP_SMAC_IPV4] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_SP_MAC_IPV4,
+   0, 0
+   },
+   [TF_DIR_TX][TF_TBL_TYPE_ACT_SP_SMAC_IPV6] = {
+   TF_RM_ELEM_CFG_HCAPI_BA, CFA_RESOURCE_TYPE_P4_SP_MAC_IPV6,
+ 

[dpdk-dev] [PATCH v2 14/19] net/bnxt: delete the VF pair before VF representor alloc

2021-10-26 Thread Venkat Duvvuru
From: Kishore Padmanabha 

When the VF representor interface is created, the VF pair relationship
is established between the VF and it is representor. If the pair
already exists then the pair needs to be deleted before allocation.
This could happen if the application is abruptly killed and restarted.
If the deletion of an existing VF rep is not done then hw pipeline is not
cleaned and a new allocation shall leave the hw in inconsistent state.

Signed-off-by: Kishore Padmanabha 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Ajit Khaparde 
Reviewed-by: Shahaji Bhosle 
---
 drivers/net/bnxt/bnxt_hwrm.c | 28 
 drivers/net/bnxt/bnxt_hwrm.h |  1 +
 drivers/net/bnxt/bnxt_reps.c |  6 ++
 3 files changed, 35 insertions(+)

diff --git a/drivers/net/bnxt/bnxt_hwrm.c b/drivers/net/bnxt/bnxt_hwrm.c
index c7041143a3..f8c27fe316 100644
--- a/drivers/net/bnxt/bnxt_hwrm.c
+++ b/drivers/net/bnxt/bnxt_hwrm.c
@@ -6025,6 +6025,34 @@ int bnxt_hwrm_first_vf_id_query(struct bnxt *bp, 
uint16_t fid,
return rc;
 }
 
+int bnxt_hwrm_cfa_pair_exists(struct bnxt *bp, struct bnxt_representor *rep_bp)
+{
+   struct hwrm_cfa_pair_info_output *resp = bp->hwrm_cmd_resp_addr;
+   struct hwrm_cfa_pair_info_input req = {0};
+   int rc = 0;
+
+   if (!(BNXT_PF(bp) || BNXT_VF_IS_TRUSTED(bp))) {
+   PMD_DRV_LOG(DEBUG,
+   "Not a PF or trusted VF. Command not supported\n");
+   return 0;
+   }
+
+   HWRM_PREP(&req, HWRM_CFA_PAIR_INFO, BNXT_USE_CHIMP_MB);
+   snprintf(req.pair_name, sizeof(req.pair_name), "%svfr%d",
+bp->eth_dev->data->name, rep_bp->vf_id);
+   req.flags =
+   rte_cpu_to_le_32(HWRM_CFA_PAIR_INFO_INPUT_FLAGS_LOOKUP_TYPE);
+
+   rc = bnxt_hwrm_send_message(bp, &req, sizeof(req), BNXT_USE_CHIMP_MB);
+   HWRM_CHECK_RESULT();
+   if (rc == HWRM_ERR_CODE_SUCCESS && strlen(resp->pair_name)) {
+   HWRM_UNLOCK();
+   return !rc;
+   }
+   HWRM_UNLOCK();
+   return rc;
+}
+
 int bnxt_hwrm_cfa_pair_alloc(struct bnxt *bp, struct bnxt_representor *rep_bp)
 {
struct hwrm_cfa_pair_alloc_output *resp = bp->hwrm_cmd_resp_addr;
diff --git a/drivers/net/bnxt/bnxt_hwrm.h b/drivers/net/bnxt/bnxt_hwrm.h
index 6dc23b93ac..33e6ba99d5 100644
--- a/drivers/net/bnxt/bnxt_hwrm.h
+++ b/drivers/net/bnxt/bnxt_hwrm.h
@@ -294,6 +294,7 @@ int bnxt_clear_one_vnic_filter(struct bnxt *bp,
 void bnxt_free_vf_info(struct bnxt *bp);
 int bnxt_hwrm_first_vf_id_query(struct bnxt *bp, uint16_t fid,
uint16_t *first_vf_id);
+int bnxt_hwrm_cfa_pair_exists(struct bnxt *bp, struct bnxt_representor 
*rep_bp);
 int bnxt_hwrm_cfa_pair_alloc(struct bnxt *bp, struct bnxt_representor *rep);
 int bnxt_hwrm_cfa_pair_free(struct bnxt *bp, struct bnxt_representor *rep);
 int bnxt_hwrm_cfa_adv_flow_mgmt_qcaps(struct bnxt *bp);
diff --git a/drivers/net/bnxt/bnxt_reps.c b/drivers/net/bnxt/bnxt_reps.c
index 988f5a33a7..111d226a59 100644
--- a/drivers/net/bnxt/bnxt_reps.c
+++ b/drivers/net/bnxt/bnxt_reps.c
@@ -315,6 +315,12 @@ static int bnxt_tf_vfr_alloc(struct rte_eth_dev 
*vfr_ethdev)
BNXT_TF_DBG(ERR, "Invalid arguments\n");
return 0;
}
+   /* update the port id so you can backtrack to ethdev */
+   vfr->dpdk_port_id = vfr_ethdev->data->port_id;
+
+   /* If pair is present, then delete the pair */
+   if (bnxt_hwrm_cfa_pair_exists(parent_bp, vfr))
+   (void)bnxt_hwrm_cfa_pair_free(parent_bp, vfr);
 
/* Update the ULP portdata base with the new VFR interface */
rc = ulp_port_db_dev_port_intf_update(parent_bp->ulp_ctx, vfr_ethdev);
-- 
2.17.1



[dpdk-dev] [PATCH v2 18/19] net/bnxt: remove 2-slice WC entries for scale

2021-10-26 Thread Venkat Duvvuru
From: Mike Baucom 

The type-5 WC IPv6 flows were removed in order to increase the scale for
app-id=3.  The app no longer supports 2-slice WC entries.

Signed-off-by: Mike Baucom 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Kishore Padmanabha 
Reviewed-by: Randy Schacher 
---
 drivers/net/bnxt/tf_ulp/bnxt_ulp.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/drivers/net/bnxt/tf_ulp/bnxt_ulp.c 
b/drivers/net/bnxt/tf_ulp/bnxt_ulp.c
index 14be079eaa..d40d726c58 100644
--- a/drivers/net/bnxt/tf_ulp/bnxt_ulp.c
+++ b/drivers/net/bnxt/tf_ulp/bnxt_ulp.c
@@ -461,7 +461,7 @@ ulp_ctx_shared_session_open(struct bnxt *bp,
 
parms.shadow_copy = true;
parms.bp = bp;
-   if (app_id == 0 || app_id == 3)
+   if (app_id == 0)
parms.wc_num_slices = TF_WC_TCAM_2_SLICE_PER_ROW;
else
parms.wc_num_slices = TF_WC_TCAM_1_SLICE_PER_ROW;
@@ -583,7 +583,7 @@ ulp_ctx_session_open(struct bnxt *bp,
return rc;
 
params.bp = bp;
-   if (app_id == 0 || app_id == 3)
+   if (app_id == 0)
params.wc_num_slices = TF_WC_TCAM_2_SLICE_PER_ROW;
else
params.wc_num_slices = TF_WC_TCAM_1_SLICE_PER_ROW;
-- 
2.17.1



[dpdk-dev] [PATCH v2 16/19] net/bnxt: add TruFlow and AFM SRAM partitioning support

2021-10-26 Thread Venkat Duvvuru
From: Jay Ding 

Implement set/get_sram_policy which support both rx/tx
direction truflow type the specific SRAM bank.

Signed-off-by: Jay Ding 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Randy Schacher 
Reviewed-by: Farah Smith 
---
 drivers/net/bnxt/tf_core/tf_core.c   |  82 ++
 drivers/net/bnxt/tf_core/tf_core.h   |  66 -
 drivers/net/bnxt/tf_core/tf_device.c |   7 +-
 drivers/net/bnxt/tf_core/tf_device.h |  34 ++-
 drivers/net/bnxt/tf_core/tf_device_p4.c  |   8 +-
 drivers/net/bnxt/tf_core/tf_device_p58.c | 311 ++-
 drivers/net/bnxt/tf_core/tf_device_p58.h | 118 +
 drivers/net/bnxt/tf_core/tf_tbl.c|   2 +-
 8 files changed, 503 insertions(+), 125 deletions(-)

diff --git a/drivers/net/bnxt/tf_core/tf_core.c 
b/drivers/net/bnxt/tf_core/tf_core.c
index 346d220c87..90ff93946b 100644
--- a/drivers/net/bnxt/tf_core/tf_core.c
+++ b/drivers/net/bnxt/tf_core/tf_core.c
@@ -1917,3 +1917,85 @@ int tf_query_sram_resources(struct tf *tfp,
 
return 0;
 }
+
+int tf_set_sram_policy(struct tf *tfp,
+  struct tf_set_sram_policy_parms *parms)
+{
+   int rc = 0;
+   struct tf_dev_info dev;
+
+   TF_CHECK_PARMS2(tfp, parms);
+
+   /* This function can be called before open session, filter
+* out any non-supported device types on the Core side.
+*/
+   if (parms->device_type != TF_DEVICE_TYPE_THOR) {
+   TFP_DRV_LOG(ERR,
+   "Unsupported device type %d\n",
+   parms->device_type);
+   return -ENOTSUP;
+   }
+
+   tf_dev_bind_ops(parms->device_type, &dev);
+
+   if (dev.ops->tf_dev_set_sram_policy == NULL) {
+   rc = -EOPNOTSUPP;
+   TFP_DRV_LOG(ERR,
+   "%s: Operation not supported, rc:%s\n",
+   tf_dir_2_str(parms->dir),
+   strerror(-rc));
+   return rc;
+   }
+
+   rc = dev.ops->tf_dev_set_sram_policy(parms->dir, parms->bank_id);
+   if (rc) {
+   TFP_DRV_LOG(ERR,
+   "%s: SRAM policy set failed, rc:%s\n",
+   tf_dir_2_str(parms->dir),
+   strerror(-rc));
+   return rc;
+   }
+
+   return rc;
+}
+
+int tf_get_sram_policy(struct tf *tfp,
+  struct tf_get_sram_policy_parms *parms)
+{
+   int rc = 0;
+   struct tf_dev_info dev;
+
+   TF_CHECK_PARMS2(tfp, parms);
+
+   /* This function can be called before open session, filter
+* out any non-supported device types on the Core side.
+*/
+   if (parms->device_type != TF_DEVICE_TYPE_THOR) {
+   TFP_DRV_LOG(ERR,
+   "Unsupported device type %d\n",
+   parms->device_type);
+   return -ENOTSUP;
+   }
+
+   tf_dev_bind_ops(parms->device_type, &dev);
+
+   if (dev.ops->tf_dev_get_sram_policy == NULL) {
+   rc = -EOPNOTSUPP;
+   TFP_DRV_LOG(ERR,
+   "%s: Operation not supported, rc:%s\n",
+   tf_dir_2_str(parms->dir),
+   strerror(-rc));
+   return rc;
+   }
+
+   rc = dev.ops->tf_dev_get_sram_policy(parms->dir, parms->bank_id);
+   if (rc) {
+   TFP_DRV_LOG(ERR,
+   "%s: SRAM policy get failed, rc:%s\n",
+   tf_dir_2_str(parms->dir),
+   strerror(-rc));
+   return rc;
+   }
+
+   return rc;
+}
diff --git a/drivers/net/bnxt/tf_core/tf_core.h 
b/drivers/net/bnxt/tf_core/tf_core.h
index 078fd278a1..b2886355fa 100644
--- a/drivers/net/bnxt/tf_core/tf_core.h
+++ b/drivers/net/bnxt/tf_core/tf_core.h
@@ -2455,7 +2455,7 @@ int tf_get_version(struct tf *tfp,
  */
 struct tf_query_sram_resources_parms {
/**
-* [in] device type
+* [in] Device type
 *
 * Device type for the session.
 */
@@ -2501,4 +2501,68 @@ struct tf_query_sram_resources_parms {
 int tf_query_sram_resources(struct tf *tfp,
struct tf_query_sram_resources_parms *parms);
 
+/**
+ * tf_set_sram_policy parameter definition
+ */
+struct tf_set_sram_policy_parms {
+   /**
+* [in] Device type
+*
+* Device type for the session.
+*/
+   enum tf_device_type device_type;
+
+   /**
+* [in] Receive or transmit direction
+*/
+   enum tf_dir dir;
+
+   /**
+* [in] Array of Bank id for each truflow tbl type
+*/
+   uint8_t *bank_id;
+};
+
+/**
+ * Set SRAM policy
+ *
+ * Used to assign SRAM bank index to all truflow table type.
+ *
+ * Returns success or failure code.
+ */
+int tf_set_sram_policy(struct tf *tfp,
+  struct tf_set_sram_policy_parms *parms);
+
+/**
+ * tf_get_sram_po

[dpdk-dev] [PATCH v2 15/19] net/bnxt: add new API TruFlow get SRAM resources

2021-10-26 Thread Venkat Duvvuru
From: Jay Ding 

Implement tf_get_sram_resources to return SRAM
partition information, including bank count and
SRAM profile number.

Signed-off-by: Jay Ding 
Signed-off-by: Venkat Duvvuru 
Reviewed-by: Farah Smith 
Reviewed-by: Randy Schacher 
---
 drivers/net/bnxt/hsi_struct_def_dpdk.h   | 11 ++-
 drivers/net/bnxt/tf_core/tf_core.c   | 86 
 drivers/net/bnxt/tf_core/tf_core.h   | 63 +
 drivers/net/bnxt/tf_core/tf_device.h | 20 ++
 drivers/net/bnxt/tf_core/tf_device_p4.c  |  6 +-
 drivers/net/bnxt/tf_core/tf_device_p58.c | 53 ++-
 drivers/net/bnxt/tf_core/tf_msg.c| 28 ++--
 drivers/net/bnxt/tf_core/tf_msg.h|  6 +-
 drivers/net/bnxt/tf_core/tf_rm.c |  3 +-
 drivers/net/bnxt/tf_core/tf_sram_mgr.h   | 10 ---
 10 files changed, 245 insertions(+), 41 deletions(-)

diff --git a/drivers/net/bnxt/hsi_struct_def_dpdk.h 
b/drivers/net/bnxt/hsi_struct_def_dpdk.h
index 6d04e39e48..1a46f34249 100644
--- a/drivers/net/bnxt/hsi_struct_def_dpdk.h
+++ b/drivers/net/bnxt/hsi_struct_def_dpdk.h
@@ -1058,8 +1058,8 @@ struct hwrm_err_output {
 #define HWRM_VERSION_MINOR 10
 #define HWRM_VERSION_UPDATE 2
 /* non-zero means beta version */
-#define HWRM_VERSION_RSVD 58
-#define HWRM_VERSION_STR "1.10.2.58"
+#define HWRM_VERSION_RSVD 64
+#define HWRM_VERSION_STR "1.10.2.64"
 
 /
  * hwrm_ver_get *
@@ -45817,8 +45817,13 @@ struct hwrm_tf_session_resc_qcaps_output {
 * qcaps_size.
 */
uint16_tsize;
+   /*
+* SRAM profile number that sets the partition of SRAM memory
+* between TF and AFM within the 4 internal memory banks (Thor).
+*/
+   uint8_t sram_profile;
/* unused. */
-   uint16_tunused0;
+   uint8_t unused0;
/* unused. */
uint8_t unused1[7];
/*
diff --git a/drivers/net/bnxt/tf_core/tf_core.c 
b/drivers/net/bnxt/tf_core/tf_core.c
index 86dfec0eb4..346d220c87 100644
--- a/drivers/net/bnxt/tf_core/tf_core.c
+++ b/drivers/net/bnxt/tf_core/tf_core.c
@@ -1831,3 +1831,89 @@ int tf_get_version(struct tf *tfp,
 
return 0;
 }
+
+int tf_query_sram_resources(struct tf *tfp,
+   struct tf_query_sram_resources_parms *parms)
+{
+   int rc;
+   struct tf_dev_info dev;
+   uint16_t max_types;
+   struct tfp_calloc_parms cparms;
+   struct tf_rm_resc_req_entry *query;
+   enum tf_rm_resc_resv_strategy resv_strategy;
+
+   TF_CHECK_PARMS2(tfp, parms);
+
+   /* This function can be called before open session, filter
+* out any non-supported device types on the Core side.
+*/
+   if (parms->device_type != TF_DEVICE_TYPE_THOR) {
+   TFP_DRV_LOG(ERR,
+   "Unsupported device type %d\n",
+   parms->device_type);
+   return -ENOTSUP;
+   }
+
+   tf_dev_bind_ops(parms->device_type, &dev);
+
+   if (dev.ops->tf_dev_get_max_types == NULL) {
+   rc = -EOPNOTSUPP;
+   TFP_DRV_LOG(ERR,
+   "%s: Operation not supported, rc:%s\n",
+   tf_dir_2_str(parms->dir),
+   strerror(-rc));
+   return -EOPNOTSUPP;
+   }
+
+   /* Need device max number of elements for the RM QCAPS */
+   rc = dev.ops->tf_dev_get_max_types(tfp, &max_types);
+   if (rc) {
+   TFP_DRV_LOG(ERR,
+   "Get SRAM resc info failed, rc:%s\n",
+   strerror(-rc));
+   return rc;
+   }
+
+   /* Allocate memory for RM QCAPS request */
+   cparms.nitems = max_types;
+   cparms.size = sizeof(struct tf_rm_resc_req_entry);
+   cparms.alignment = 0;
+   rc = tfp_calloc(&cparms);
+   if (rc)
+   return rc;
+
+   query = (struct tf_rm_resc_req_entry *)cparms.mem_va;
+   tfp->bp = parms->bp;
+
+   /* Get Firmware Capabilities */
+   rc = tf_msg_session_resc_qcaps(tfp,
+  &dev,
+  parms->dir,
+  max_types,
+  query,
+  &resv_strategy,
+  &parms->sram_profile);
+   if (rc)
+   return rc;
+
+   if (dev.ops->tf_dev_get_sram_resources == NULL) {
+   rc = -EOPNOTSUPP;
+   TFP_DRV_LOG(ERR,
+   "%s: Operation not supported, rc:%s\n",
+   tf_dir_2_str(parms->dir),
+   strerror(-rc));
+   return -EOPNOTSUPP;
+   }
+
+   rc = dev.ops->tf_dev_get_sram_resources((void *)query,
+   parms->bank_resc_count,
+   &parms->dynamic_sram_capable);
+   if (rc) {
+   TFP_DRV_LOG(ERR,
+  

[dpdk-dev] [PATCH v2 19/19] net/bnxt: check for mismatch of control and physical port

2021-10-26 Thread Venkat Duvvuru
From: Kishore Padmanabha 

During the parsing of the ingress port ignore for a flow, added
check to match the control port and the physical port that is configured
to be ignored. If they do not match then the configuration to setup the
svif ignore shall fail.

Signed-off-by: Kishore Padmanabha 
Reviewed-by: Michael Baucom 
Reviewed-by: Shahaji Bhosle 
---
 drivers/net/bnxt/tf_ulp/ulp_port_db.c| 23 +++
 drivers/net/bnxt/tf_ulp/ulp_port_db.h| 13 +
 drivers/net/bnxt/tf_ulp/ulp_rte_parser.c | 12 
 3 files changed, 48 insertions(+)

diff --git a/drivers/net/bnxt/tf_ulp/ulp_port_db.c 
b/drivers/net/bnxt/tf_ulp/ulp_port_db.c
index 5e7c1d1c17..f8ffb567b5 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_port_db.c
+++ b/drivers/net/bnxt/tf_ulp/ulp_port_db.c
@@ -679,3 +679,26 @@ ulp_port_db_parent_vnic_get(struct bnxt_ulp_context 
*ulp_ctxt,
}
return -EINVAL;
 }
+
+/*
+ * Api to get the phy port for a given port id.
+ *
+ * ulp_ctxt [in] Ptr to ulp context
+ * port_id [in] device port id
+ * phy_port [out] phy_port of the dpdk port_id
+ *
+ * Returns 0 on success or negative number on failure.
+ */
+int32_t
+ulp_port_db_phy_port_get(struct bnxt_ulp_context *ulp_ctxt,
+uint32_t port_id, uint16_t *phy_port)
+{
+   struct ulp_func_if_info *info;
+
+   info = ulp_port_db_func_if_info_get(ulp_ctxt, port_id);
+   if (info) {
+   *phy_port = info->phy_port_id;
+   return 0;
+   }
+   return -EINVAL;
+}
diff --git a/drivers/net/bnxt/tf_ulp/ulp_port_db.h 
b/drivers/net/bnxt/tf_ulp/ulp_port_db.h
index 740c186e12..b112f1a216 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_port_db.h
+++ b/drivers/net/bnxt/tf_ulp/ulp_port_db.h
@@ -314,4 +314,17 @@ int32_t
 ulp_port_db_parent_vnic_get(struct bnxt_ulp_context *ulp_ctxt,
uint32_t port_id, uint8_t **vnic);
 
+/*
+ * Api to get the phy port for a given port id.
+ *
+ * ulp_ctxt [in] Ptr to ulp context
+ * port_id [in] device port id
+ * phy_port [out] phy_port of the dpdk port_id
+ *
+ * Returns 0 on success or negative number on failure.
+ */
+int32_t
+ulp_port_db_phy_port_get(struct bnxt_ulp_context *ulp_ctxt,
+uint32_t port_id, uint16_t *phy_port);
+
 #endif /* _ULP_PORT_DB_H_ */
diff --git a/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c 
b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c
index 2ec3279239..f4274dd634 100644
--- a/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c
+++ b/drivers/net/bnxt/tf_ulp/ulp_rte_parser.c
@@ -686,6 +686,18 @@ ulp_rte_phy_port_hdr_handler(const struct rte_flow_item 
*item,
ULP_COMP_FLD_IDX_WR(params, BNXT_ULP_CF_IDX_SVIF_FLAG,
rte_be_to_cpu_16(svif));
if (!mask) {
+   uint32_t port_id = 0;
+   uint16_t phy_port = 0;
+
+   /* Validate the control port */
+   port_id = ULP_COMP_FLD_IDX_RD(params,
+ BNXT_ULP_CF_IDX_DEV_PORT_ID);
+   if (ulp_port_db_phy_port_get(params->ulp_ctx,
+port_id, &phy_port) ||
+   (uint16_t)port_spec->index != phy_port) {
+   BNXT_TF_DBG(ERR, "Mismatch of control and phy_port\n");
+   return BNXT_TF_RC_PARSE_ERR;
+   }
ULP_BITMAP_SET(params->hdr_bitmap.bits,
   BNXT_ULP_HDR_BIT_SVIF_IGNORE);
memset(hdr_field->mask, 0xFF, sizeof(mask));
-- 
2.17.1



[dpdk-dev] [PATCH] net/bnxt: fix RSS action parser

2021-10-26 Thread Ajit Khaparde
Minor fixes are needed in the RTE_FLOW RSS action parser.
1. Update the comment in the parser to indicate rss level 1 implies
RSS on outer header.
2. RSS action will not be supported if level is > 1.
3. RSS action will not be supported if user or application specifies
MARK or COUNT action.
4. If RSS types is not specified i.e., is 0, the best effort RSS should
use IPV4 and IPV6 headers. Currently we are considering only IPV4.

Fixes: fe0bab7eb34e ("net/bnxt: enhance support for RSS action")

Signed-off-by: Ajit Khaparde 
Acked-by: Kalesh AP 
Acked-by: Somnath Kotur 
---
 drivers/net/bnxt/bnxt_flow.c | 13 -
 1 file changed, 4 insertions(+), 9 deletions(-)

diff --git a/drivers/net/bnxt/bnxt_flow.c b/drivers/net/bnxt/bnxt_flow.c
index ced697a739..9526a8fc5a 100644
--- a/drivers/net/bnxt/bnxt_flow.c
+++ b/drivers/net/bnxt/bnxt_flow.c
@@ -1084,10 +1084,6 @@ bnxt_validate_rss_action(const struct rte_flow_action 
actions[])
break;
case RTE_FLOW_ACTION_TYPE_RSS:
break;
-   case RTE_FLOW_ACTION_TYPE_MARK:
-   break;
-   case RTE_FLOW_ACTION_TYPE_COUNT:
-   break;
default:
return -ENOTSUP;
}
@@ -1151,11 +1147,10 @@ bnxt_vnic_rss_cfg_update(struct bnxt *bp,
}
 
/* Currently RSS hash on inner and outer headers are supported.
-* 0 => Default setting
-* 1 => Inner
-* 2 => Outer
+* 0 => Default (innermost RSS) setting
+* 1 => Outermost
 */
-   if (rss->level > 2) {
+   if (rss->level > 1) {
rte_flow_error_set(error,
   ENOTSUP,
   RTE_FLOW_ERROR_TYPE_ACTION,
@@ -1177,7 +1172,7 @@ bnxt_vnic_rss_cfg_update(struct bnxt *bp,
}
 
/* If RSS types is 0, use a best effort configuration */
-   types = rss->types ? rss->types : RTE_ETH_RSS_IPV4;
+   types = rss->types ? rss->types : ETH_RSS_IPV4 | ETH_RSS_IPV6;
 
hash_type = bnxt_rte_to_hwrm_hash_types(types);
 
-- 
2.30.1 (Apple Git-130)



[dpdk-dev] [PATCH] net/bnxt: refactor Rx ring cleanup for representors

2021-10-26 Thread Ajit Khaparde
Rx ring for representors does not use aggregation rings for Rx.
Instead they use simple software buffers for handling Rx packets.
So there is no need to use the same cleanup routine as done by
the non-representor code path.

Signed-off-by: Ajit Khaparde 
Acked-by: Kalesh AP 
Acked-by: Somnath Kotur 
---
 drivers/net/bnxt/bnxt_reps.c | 22 +-
 1 file changed, 21 insertions(+), 1 deletion(-)

diff --git a/drivers/net/bnxt/bnxt_reps.c b/drivers/net/bnxt/bnxt_reps.c
index 1c07db3ca9..92beea3558 100644
--- a/drivers/net/bnxt/bnxt_reps.c
+++ b/drivers/net/bnxt/bnxt_reps.c
@@ -386,6 +386,26 @@ static int bnxt_vfr_alloc(struct rte_eth_dev *vfr_ethdev)
return rc;
 }
 
+static void bnxt_vfr_rx_queue_release_mbufs(struct bnxt_rx_queue *rxq)
+{
+   struct rte_mbuf **sw_ring;
+   unsigned int i;
+
+   if (!rxq || !rxq->rx_ring)
+   return;
+
+   sw_ring = rxq->rx_ring->rx_buf_ring;
+   if (sw_ring) {
+   for (i = 0; i < rxq->rx_ring->rx_ring_struct->ring_size; i++) {
+   if (sw_ring[i]) {
+   if (sw_ring[i] != &rxq->fake_mbuf)
+   rte_pktmbuf_free_seg(sw_ring[i]);
+   sw_ring[i] = NULL;
+   }
+   }
+   }
+}
+
 static void bnxt_rep_free_rx_mbufs(struct bnxt_representor *rep_bp)
 {
struct bnxt_rx_queue *rxq;
@@ -393,7 +413,7 @@ static void bnxt_rep_free_rx_mbufs(struct bnxt_representor 
*rep_bp)
 
for (i = 0; i < rep_bp->rx_nr_rings; i++) {
rxq = rep_bp->rx_queues[i];
-   bnxt_rx_queue_release_mbufs(rxq);
+   bnxt_vfr_rx_queue_release_mbufs(rxq);
}
 }
 
-- 
2.30.1 (Apple Git-130)



Re: [dpdk-dev] [PATCH] bus/pci: fix selection of default device NUMA node

2021-10-26 Thread Olivier Matz
On Tue, Oct 26, 2021 at 11:06:10AM +0200, Houssem Bouhlel wrote:
> There can be dev binding issue when no hugepages
> are allocated for socket 0.
> To avoid this, set device numa node value based on
> the first lcore instead of 0.
> 
> Fixes: 831dba47bd36 ("bus/vmbus: add Hyper-V virtual bus support")

Sorry, the Fixes line is wrong. This is the correct one:
Fixes: 8a04cb612589 ("pci: set default numa node for broken systems")

> Cc: sta...@dpdk.org
> 
> Signed-off-by: Houssem Bouhlel 
> Signed-off-by: Olivier Matz 
> ---
>  drivers/bus/pci/pci_common.c | 4 +++-
>  1 file changed, 3 insertions(+), 1 deletion(-)
> 
> diff --git a/drivers/bus/pci/pci_common.c b/drivers/bus/pci/pci_common.c
> index f8fff2c98ebf..c70ab2373c79 100644
> --- a/drivers/bus/pci/pci_common.c
> +++ b/drivers/bus/pci/pci_common.c
> @@ -166,6 +166,7 @@ rte_pci_probe_one_driver(struct rte_pci_driver *dr,
>struct rte_pci_device *dev)
>  {
>   int ret;
> + unsigned int socket_id;
>   bool already_probed;
>   struct rte_pci_addr *loc;
>  
> @@ -194,7 +195,8 @@ rte_pci_probe_one_driver(struct rte_pci_driver *dr,
>   if (rte_socket_count() > 1)
>   RTE_LOG(INFO, EAL, "Device %s is not NUMA-aware, 
> defaulting socket to 0\n",
>   dev->name);

One more comment (sorry, I should have done it before you send the mail):
We should move this log below, and use the socket_id instead of 0.

> - dev->device.numa_node = 0;
> + socket_id = rte_lcore_to_socket_id(rte_get_next_lcore(-1, 0, 
> 0));
> + dev->device.numa_node = socket_id;
>   }
>  
>   already_probed = rte_dev_is_probed(&dev->device);
> -- 
> 2.30.2
> 


Re: [dpdk-dev] [PATCH v5 4/5] lib/bpf: use wait event scheme for Rx/Tx iteration

2021-10-26 Thread Ananyev, Konstantin
Hi Feifei,

> > Instead of polling for cbi->use to be updated, use wait event scheme.
> >
> > Furthermore, delete 'const' for 'bpf_eth_cbi_wait'. This is because of a
> > compilation error:
> > ---
> > ../lib/eal/include/rte_common.h:36:13: error: read-only variable ‘value’
> > used as ‘asm’ output
> >36 | #define asm __asm__
> >   | ^~~
> >
> > ../lib/eal/arm/include/rte_pause_64.h:66:3: note: in expansion of macro 
> > ‘asm’
> >66 |   asm volatile("ldaxr %w[tmp], [%x[addr]]" \
> >   |   ^~~
> >
> > ../lib/eal/arm/include/rte_pause_64.h:96:3: note: in expansion of macro
> > ‘__LOAD_EXC_32’
> >96 |   __LOAD_EXC_32((src), dst, memorder) \
> >   |   ^
> >
> > ../lib/eal/arm/include/rte_pause_64.h:167:4: note: in expansion of macro
> > ‘__LOAD_EXC’
> >   167 |__LOAD_EXC((addr), value, memorder, size) \
> >   |^~
> >
> > ../lib/bpf/bpf_pkt.c:125:3: note: in expansion of macro ‘rte_wait_event’
> >   125 |   rte_wait_event(&cbi->use, UINT32_MAX, ==, puse,
> > ---
> >
> > Signed-off-by: Feifei Wang 
> > Reviewed-by: Ruifeng Wang 
> > ---
> >  lib/bpf/bpf_pkt.c | 11 ---
> >  1 file changed, 4 insertions(+), 7 deletions(-)
> >
> > diff --git a/lib/bpf/bpf_pkt.c b/lib/bpf/bpf_pkt.c index
> > 6e8248f0d6..213d44a75a 100644
> > --- a/lib/bpf/bpf_pkt.c
> > +++ b/lib/bpf/bpf_pkt.c
> > @@ -111,9 +111,9 @@ bpf_eth_cbi_unuse(struct bpf_eth_cbi *cbi)
> >   * Waits till datapath finished using given callback.
> >   */
> >  static void
> > -bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
> > +bpf_eth_cbi_wait(struct bpf_eth_cbi *cbi)  
> 
> Hi, Konstantin
> 
> For this bpf patch, I delete 'const' through this is contrary to what we
> discussed earlier. This is because if  we keep 'constant' here and use 
> 'rte_wait_event'
> new macro, compiler will report error. And earlier the arm version cannot be 
> compiled
> due to I forgot enable "wfe" config in the meson file, so this issue can not 
> happen before.


Honestly, I don't understand why we have to remove perfectly valid 'const' 
qualifier here.
If this macro can't be used with pointers to const (still don't understand why),
then let's just not use this macro here.
Strictly speaking I don't see much benefit here from it.

> 
> >  {
> > -   uint32_t nuse, puse;
> > +   uint32_t puse;
> >
> > /* make sure all previous loads and stores are completed */
> > rte_smp_mb();
> > @@ -122,11 +122,8 @@ bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
> >
> > /* in use, busy wait till current RX/TX iteration is finished */
> > if ((puse & BPF_ETH_CBI_INUSE) != 0) {
> > -   do {
> > -   rte_pause();
> > -   rte_compiler_barrier();
> > -   nuse = cbi->use;
> > -   } while (nuse == puse);
> > +   rte_wait_event(&cbi->use, UINT32_MAX, ==, puse,
> > +   __ATOMIC_RELAXED);
> > }
> >  }
> >
> > --
> > 2.25.1



Re: [dpdk-dev] [PATCH v5 1/5] eal: add new definitions for wait scheme

2021-10-26 Thread Ananyev, Konstantin


> >
> > Introduce macros as generic interface for address monitoring.
> > For different size, encapsulate '__LOAD_EXC_16', '__LOAD_EXC_32'
> > and '__LOAD_EXC_64' into a new macro '__LOAD_EXC'.
> >
> > Furthermore, to prevent compilation warning in arm:
> > --
> > 'warning: implicit declaration of function ...'
> > --
> > Delete 'undef' constructions for '__LOAD_EXC_xx', '__SEVL' and '__WFE'.
> >
> > This is because original macros are undefine at the end of the file.
> > If new macro 'rte_wait_event' calls them in other files, they will be seen 
> > as
> > 'not defined'.
> >
> > Signed-off-by: Feifei Wang 
> > Reviewed-by: Ruifeng Wang 
> > ---
> >  lib/eal/arm/include/rte_pause_64.h  | 135 
> > lib/eal/include/generic/rte_pause.h |  27 ++
> >  2 files changed, 105 insertions(+), 57 deletions(-)
> >
> > diff --git a/lib/eal/arm/include/rte_pause_64.h
> > b/lib/eal/arm/include/rte_pause_64.h
> > index e87d10b8cc..1fea0dec63 100644
> > --- a/lib/eal/arm/include/rte_pause_64.h
> > +++ b/lib/eal/arm/include/rte_pause_64.h
> > @@ -31,20 +31,12 @@ static inline void rte_pause(void)
> >  /* Put processor into low power WFE(Wait For Event) state. */  #define
> > __WFE() { asm volatile("wfe" : : : "memory"); }
> >
> > -static __rte_always_inline void
> > -rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t expected,
> > -   int memorder)
> > -{
> > -   uint16_t value;
> > -
> > -   assert(memorder == __ATOMIC_ACQUIRE || memorder ==
> > __ATOMIC_RELAXED);
> > -
> > -   /*
> > -* Atomic exclusive load from addr, it returns the 16-bit content of
> > -* *addr while making it 'monitored',when it is written by someone
> > -* else, the 'monitored' state is cleared and a event is generated
> > -* implicitly to exit WFE.
> > -*/
> > +/*
> > + * Atomic exclusive load from addr, it returns the 16-bit content of
> > + * *addr while making it 'monitored', when it is written by someone
> > + * else, the 'monitored' state is cleared and an event is generated
> > + * implicitly to exit WFE.
> > + */
> >  #define __LOAD_EXC_16(src, dst, memorder) {   \
> > if (memorder == __ATOMIC_RELAXED) {   \
> > asm volatile("ldxrh %w[tmp], [%x[addr]]"  \ @@ -58,6 +50,62
> > @@ rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t expected,
> > : "memory");  \
> > } }
> >
> > +/*
> > + * Atomic exclusive load from addr, it returns the 32-bit content of
> > + * *addr while making it 'monitored', when it is written by someone
> > + * else, the 'monitored' state is cleared and an event is generated
> > + * implicitly to exit WFE.
> > + */
> > +#define __LOAD_EXC_32(src, dst, memorder) {  \
> > +   if (memorder == __ATOMIC_RELAXED) {  \
> > +   asm volatile("ldxr %w[tmp], [%x[addr]]"  \
> > +   : [tmp] "=&r" (dst)  \
> > +   : [addr] "r"(src)\
> > +   : "memory"); \
> > +   } else { \
> > +   asm volatile("ldaxr %w[tmp], [%x[addr]]" \
> > +   : [tmp] "=&r" (dst)  \
> > +   : [addr] "r"(src)\
> > +   : "memory"); \
> > +   } }
> > +
> > +/*
> > + * Atomic exclusive load from addr, it returns the 64-bit content of
> > + * *addr while making it 'monitored', when it is written by someone
> > + * else, the 'monitored' state is cleared and an event is generated
> > + * implicitly to exit WFE.
> > + */
> > +#define __LOAD_EXC_64(src, dst, memorder) {  \
> > +   if (memorder == __ATOMIC_RELAXED) {  \
> > +   asm volatile("ldxr %x[tmp], [%x[addr]]"  \
> > +   : [tmp] "=&r" (dst)  \
> > +   : [addr] "r"(src)\
> > +   : "memory"); \
> > +   } else { \
> > +   asm volatile("ldaxr %x[tmp], [%x[addr]]" \
> > +   : [tmp] "=&r" (dst)  \
> > +   : [addr] "r"(src)\
> > +   : "memory"); \
> > +   } }
> > +
> > +#define __LOAD_EXC(src, dst, memorder, size) {  \
> > +   assert(size == 16 || size == 32 || size == 64); \
> > +   if (size == 16) \
> > +   __LOAD_EXC_16(src, dst, memorder)   \
> > +   else if (size == 32)\
> > +   __LOAD_EXC_32(src, dst, memorder)   \
> > +   else if (size == 64)\
> > +   __LOAD_EXC_64(src, dst, memorder)   \
> > +}
> > +
> > +static __rte_always_inline void
> > +rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t expected,
> > + 

[dpdk-dev] [Bug 838] [ubsan] bnxt: left shift cannot be represented in bnxt_hwrm_ver_get()

2021-10-26 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=838

Bug ID: 838
   Summary: [ubsan] bnxt: left shift cannot be represented in
bnxt_hwrm_ver_get()
   Product: DPDK
   Version: unspecified
  Hardware: All
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: ethdev
  Assignee: dev@dpdk.org
  Reporter: burner-em...@caramail.com
  Target Milestone: ---

Hello,

An issue found by UBSan:

DPDK/drivers/net/bnxt/bnxt_hwrm.c:X:Y: runtime error: left shift of 216 by 24
places cannot be represented in type 'int'.

The offending code is in bnxt_hwrm_ver_get():

bp->fw_ver = (resp->hwrm_fw_maj_8b << 24) |   <--- HERE
 (resp->hwrm_fw_min_8b << 16) |
 (resp->hwrm_fw_bld_8b << 8) |
 resp->hwrm_fw_rsvd_8b;

The left shifts here are of type 'int', and the first one cannot be
represented. This is an undefined behavior.

The first shift should be explicitly cast to uint32_t, ie:

bp->fw_ver = ((uint32_t)resp->hwrm_fw_maj_8b << 24) |
 (resp->hwrm_fw_min_8b << 16) |
 (resp->hwrm_fw_bld_8b << 8) |
 resp->hwrm_fw_rsvd_8b;

I have tested this patch, and could confirm with UBSan that there is no
undefined behavior anymore.

Thanks!

-- 
You are receiving this mail because:
You are the assignee for the bug.

Re: [dpdk-dev] [PATCH] vhost: fix async DMA map

2021-10-26 Thread Maxime Coquelin




On 10/26/21 10:49, Ding, Xuan wrote:

Hi Maxime,


-Original Message-
From: Maxime Coquelin 
Sent: Tuesday, October 26, 2021 2:53 PM
To: Ding, Xuan ; dev@dpdk.org;
david.march...@redhat.com; Xia, Chenbo 
Cc: Burakov, Anatoly 
Subject: Re: [PATCH] vhost: fix async DMA map



On 10/26/21 04:07, Ding, Xuan wrote:

Hi Maxime,


-Original Message-
From: Maxime Coquelin 
Sent: Tuesday, October 26, 2021 4:47 AM
To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
; Ding, Xuan 
Subject: Re: [PATCH] vhost: fix async DMA map

Hi Xuan,

On 10/25/21 22:33, Maxime Coquelin wrote:

This patch fixes possible NULL-pointer dereferencing
reported by Coverity and also fixes NUMA reallocation
of the async DMA map.

Fixes: 7c61fa08b716 ("vhost: enable IOMMU for async vhost")

Coverity issue: 373655

Signed-off-by: Maxime Coquelin 
---
lib/vhost/vhost_user.c | 45 +++---
1 file changed, 20 insertions(+), 25 deletions(-)



I posted this patch to fix the issue reported by Coverity and also other
issue on NUMA realloc that I found at the same time. But I wonder
whether all this async_map_status is needed.


Thanks for your fix! I can help to review and test the patch later.

I add the async_map_status in v2 for compatibility. Some DMA device,
like DSA, may use kernel idxd driver only. If there is no device bound to
DPDK vfio and kernel vfio module is modprobed to ensure

rte_vfio_is_enabled() is true,

we will unavoidably do DMA map/unmap and it will fail.

Therefore, the dma_map_status here is used to filter this situation by

preventing

unnecessary DMA unmap.


Ok, then I think we can just remove the async DMA map.



Indeed, if the only place where we DMA map is in
vhost_user_mmap_region(). If it fails, the error is propagated, the mem
table are freed and NACK is replied to the master. IOW, the device will
be in an unusable state.


I agree with you, this is the place I consider right to do DMA map
because we also do SW mapping here, any suggestions?


No suggestion, I was just explaining that at the only place where
DMA map were done, mapping errors were properly handled and propagated.


What about just setting async_copy to false, and allow switching to sync path.





Removing the async DMA map will simplify a lot the code, do you agree to
remove it or there is something I missed?


See above. Indeed, it adds a lot of code. But we can't know the driver for
each device in vhost lib, or we can only restrict the user to bind some

devices

to DPDK vfio if async logic needed.


I would think we don't care if DMA unmap fails, we can just do the same
as what you do for DMA map, i.e. just ignore the error.


Get your idea, we can do the same as DMA map, and in this way dma_map_status 
flag can be removed.



Thanks to this discussion, I have now more concerns on how it works. I
think we have a problem here in case of DMA device hotplug, that device
could miss the necessary map entries from Vhost if no VFIO devices were
attached at VHST_USER_SET_MEM_TABLE time. How would you handle that
case?


DMA device is uncore, so I don't see the  hotplug issue here.


I'm not sure what 'uncore' is, I suppose you mean your device cannot be
physically added/removed to the system.

I was not clear enough in my question. I meant that for example, the
application is started and the Vhost port is created. Then, the DMA
device is bound to VFIO, and probed by the DPDK application. Finally,
the application register the DMA device as an async channel in Vhost.

I think it will not work as the SET_MEM_TABLE will already have
happened, so the guest memory won't be mapped in the VFIO container.

Do you have the same understanding?


I will have another patch containing compatibility with sync path, and 
async_map_status flag will be removed.
Hope to get your insights.


What do you mean by compatibility with sync path?

Thanks for taking care of the async map status removal.

Maxime

Thanks,
Xuan



Regards,
Maxime



Thanks,
Maxime








Re: [dpdk-dev] [PATCH v3] doc: document a limitation for a meter with RSS action

2021-10-26 Thread Thomas Monjalon
26/10/2021 11:31, Li Zhang:
> A meter policy with RSS/Queue action is not supported
> when dv_xmeta_en is not zero.

It is still not explaining why.

> 
> Fixes: 51ec04d ("net/mlx5: connect meter policy to created flows")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Li Zhang 
> Acked-by: Matan Azrad 
> ---
> --- a/doc/guides/nics/mlx5.rst
> +++ b/doc/guides/nics/mlx5.rst
>- Policy actions of RSS for green and yellow should have the same 
> configuration except queues.
> +  - Policy with RSS/queue action is not supported when dv_xmeta_en is not 
> zero.

It is still not enclosed like this: ``dv_xmeta_en``

Don't you think "enabled" is better than "not zero"?





[dpdk-dev] [PATCH v5] net/ice: simplify the use of DCF device reset

2021-10-26 Thread dapengx . yu
From: Dapeng Yu 

After DCF is reset by PF, the DCF device un-initialization cannot
function normally since the kernel may not clean up resource.

This patch detects the reset flag, which is set by PF on DCF reset,
if the flag is true, reset hw to trigger an additional DCF
enable/disable cycle which helps to work around the issue that kernel
driver may not clean up resource during previous reset.

Fixes: 1a86f4dbdf42 ("net/ice: support DCF device reset")
Cc: sta...@dpdk.org

Signed-off-by: Dapeng Yu 
---
V2:
* Ignore the returned error of dev_uninit when DCF is reset by PF
V3:
* Add a reset function to re-initialize AdminQ resource
* Add a function to check the reset flag
V4:
* Remove redundant reset flag setting
V5:
* Rename function and modify comments
---
 drivers/net/ice/ice_dcf.c|  2 ++
 drivers/net/ice/ice_dcf_ethdev.c | 34 +++-
 2 files changed, 35 insertions(+), 1 deletion(-)

diff --git a/drivers/net/ice/ice_dcf.c b/drivers/net/ice/ice_dcf.c
index 084f7a53db..366ff0a907 100644
--- a/drivers/net/ice/ice_dcf.c
+++ b/drivers/net/ice/ice_dcf.c
@@ -593,6 +593,8 @@ ice_dcf_init_hw(struct rte_eth_dev *eth_dev, struct 
ice_dcf_hw *hw)
struct rte_pci_device *pci_dev = RTE_ETH_DEV_TO_PCI(eth_dev);
int ret, size;
 
+   hw->resetting = false;
+
hw->avf.hw_addr = pci_dev->mem_resource[0].addr;
hw->avf.back = hw;
 
diff --git a/drivers/net/ice/ice_dcf_ethdev.c b/drivers/net/ice/ice_dcf_ethdev.c
index 7c71a48010..4d9484e994 100644
--- a/drivers/net/ice/ice_dcf_ethdev.c
+++ b/drivers/net/ice/ice_dcf_ethdev.c
@@ -1025,11 +1025,44 @@ ice_dcf_tm_ops_get(struct rte_eth_dev *dev __rte_unused,
return 0;
 }
 
+static inline void
+ice_dcf_reset_hw(struct rte_eth_dev *eth_dev, struct ice_dcf_hw *hw)
+{
+   ice_dcf_uninit_hw(eth_dev, hw);
+   ice_dcf_init_hw(eth_dev, hw);
+}
+
+/* Check if reset has been triggered by PF */
+static inline bool
+ice_dcf_is_reset(struct rte_eth_dev *dev)
+{
+   struct ice_dcf_adapter *ad = dev->data->dev_private;
+   struct iavf_hw *hw = &ad->real_hw.avf;
+
+   return !(IAVF_READ_REG(hw, IAVF_VF_ARQLEN1) &
+IAVF_VF_ARQLEN1_ARQENABLE_MASK);
+}
+
 static int
 ice_dcf_dev_reset(struct rte_eth_dev *dev)
 {
+   struct ice_dcf_adapter *ad = dev->data->dev_private;
+   struct ice_dcf_hw *hw = &ad->real_hw;
int ret;
 
+   if (ice_dcf_is_reset(dev)) {
+   if (!ad->real_hw.resetting)
+   ad->real_hw.resetting = true;
+   PMD_DRV_LOG(ERR, "The DCF has been reset by PF");
+
+   /*
+* Simply reset hw to trigger an additional DCF enable/disable
+* cycle which help to workaround the issue that kernel driver
+* may not clean up resource during previous reset.
+*/
+   ice_dcf_reset_hw(dev, hw);
+   }
+
ret = ice_dcf_dev_uninit(dev);
if (ret)
return ret;
@@ -1072,7 +1105,6 @@ ice_dcf_dev_init(struct rte_eth_dev *eth_dev)
 {
struct ice_dcf_adapter *adapter = eth_dev->data->dev_private;
 
-   adapter->real_hw.resetting = false;
eth_dev->dev_ops = &ice_dcf_eth_dev_ops;
eth_dev->rx_pkt_burst = ice_dcf_recv_pkts;
eth_dev->tx_pkt_burst = ice_dcf_xmit_pkts;
-- 
2.27.0



[dpdk-dev] Please help to see this DPDK problem//Reply: DPDK 20.11 - i40evf: No response for 14

2021-10-26 Thread liaobiting
BUG: DPDK-20.11- i40evf: No response for 14

Hi:
I am using Intel SR-IOV XL710 VF with DPDK v20.11 to create a dpdkbond but 
failed.
 [cid:image001.png@01D7CA75.76DF9EB0]
 As the picture shows, when dpdk bond start and slave link up, it 
triggers lsc callback function which registers in the eal thread, and then in 
the activate_slave process,
i40evf send msg 14 to pf to config promisc, but there is no response received. 
Because function "i40evf_handle_aq_msg" in which vf receives response from pf
is in the same thread of eal with function "_i40evf_execute_vf_cmd" in which vf 
sends msg to pf. Thus, when eal thread goes in function _i40evf_execute_vf_cmd,
it can't handle msg from pf at the same time, then it will report "No response 
for 14" in dpdk log.

And there are the error logs in dpdk:
2021-10-11T11:03:32.812516+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1: set mtu: 9058 .
2021-10-11T11:03:32.822434+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1: dev configure succeed.
2021-10-11T11:03:33.022426+08:00|info|ovs-vswitchd[1721043]|EAL: LSC CALLBACK: 
slave :87:02.0 links up
2021-10-11T11:03:33.022455+08:00|info|ovs-vswitchd[1721043]|PMD: Slave 1: 
lacp_rate has been set to slow.
2021-10-11T11:03:35.034445+08:00|warning|ovs-vswitchd[1721043]|_i40evf_execute_vf_cmd():
 No response for 14
2021-10-11T11:03:35.034628+08:00|err|ovs-vswitchd[1721043]|i40evf_config_promisc():
 fail to execute command CONFIG_PROMISCUOUS_MODE
2021-10-11T11:03:35.034657+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1: enable allmulticast.
2021-10-11T11:03:35.034680+08:00|err|ovs-vswitchd[1721043]|bond_mode_8023ad_register_lacp_mac(1250)
 - failed to enable allmulti mode for port 1: Resource temporarily unavailable
2021-10-11T11:03:35.034702+08:00|debug|ovs-vswitchd[1721043]|bond_mode_8023ad_register_lacp_mac(1266)
 - forced promiscuous for port 1
2021-10-11T11:03:35.034725+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1 (net_bonding_trunk1): Configuring and applying resources successed.
2021-10-11T11:03:35.034746+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
slave 1 (net_bonding_trunk1) activate.
2021-10-11T11:03:35.034771+08:00|debug|ovs-vswitchd[1721043]| 0 [Port 1: 
rx_machine] -> INITIALIZE
2021-10-11T11:03:35.034794+08:00|debug|ovs-vswitchd[1721043]| 0 [Port 1: 
mux_machine] -> DETACHED
2021-10-11T11:03:35.084534+08:00|err|ovs-vswitchd[1721043]|i40evf_handle_aq_msg():
 command mismatch,expect 8, get 14
2021-10-11T11:03:35.134545+08:00|debug|ovs-vswitchd[1721043]|   100 [Port 1: 
mux_machine] DETACHED -> WAITING
2021-10-11T11:03:35.544648+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1: start dev.

How can I solve this problem ?
Thakns for help.
Regards,
Liao


[dpdk-dev] [dpdk-users] is i40evf support promisc? // DPDK 20.11 - i40evf: No response for 14

2021-10-26 Thread liaobiting
Hi:
Please help to see this DPDK problem. And I want to know whether i40e vf 
support promisc or not. Thanks a lot.

From: liaobiting 
Subject: Please help to see this DPDK problem//Reply: DPDK 20.11 - i40evf: No 
response for 14

BUG: DPDK-20.11- i40evf: No response for 14

Hi:
I am using Intel SR-IOV XL710 VF with DPDK v20.11 to create a dpdkbond but 
failed.
 [cid:image001.png@01D7CA79.73F04A20]
 As the picture shows, when dpdk bond start and slave link up, it 
triggers lsc callback function which registers in the eal thread, and then in 
the activate_slave process,
i40evf send msg 14 to pf to config promisc, but there is no response received. 
Because function "i40evf_handle_aq_msg" in which vf receives response from pf
is in the same thread of eal with function "_i40evf_execute_vf_cmd" in which vf 
sends msg to pf. Thus, when eal thread goes in function _i40evf_execute_vf_cmd,
it can't handle msg from pf at the same time, then it will report "No response 
for 14" in dpdk log.

And there are the error logs in dpdk:
2021-10-11T11:03:32.812516+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1: set mtu: 9058 .
2021-10-11T11:03:32.822434+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1: dev configure succeed.
2021-10-11T11:03:33.022426+08:00|info|ovs-vswitchd[1721043]|EAL: LSC CALLBACK: 
slave :87:02.0 links up
2021-10-11T11:03:33.022455+08:00|info|ovs-vswitchd[1721043]|PMD: Slave 1: 
lacp_rate has been set to slow.
2021-10-11T11:03:35.034445+08:00|warning|ovs-vswitchd[1721043]|_i40evf_execute_vf_cmd():
 No response for 14
2021-10-11T11:03:35.034628+08:00|err|ovs-vswitchd[1721043]|i40evf_config_promisc():
 fail to execute command CONFIG_PROMISCUOUS_MODE
2021-10-11T11:03:35.034657+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1: enable allmulticast.
2021-10-11T11:03:35.034680+08:00|err|ovs-vswitchd[1721043]|bond_mode_8023ad_register_lacp_mac(1250)
 - failed to enable allmulti mode for port 1: Resource temporarily unavailable
2021-10-11T11:03:35.034702+08:00|debug|ovs-vswitchd[1721043]|bond_mode_8023ad_register_lacp_mac(1266)
 - forced promiscuous for port 1
2021-10-11T11:03:35.034725+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1 (net_bonding_trunk1): Configuring and applying resources successed.
2021-10-11T11:03:35.034746+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
slave 1 (net_bonding_trunk1) activate.
2021-10-11T11:03:35.034771+08:00|debug|ovs-vswitchd[1721043]| 0 [Port 1: 
rx_machine] -> INITIALIZE
2021-10-11T11:03:35.034794+08:00|debug|ovs-vswitchd[1721043]| 0 [Port 1: 
mux_machine] -> DETACHED
2021-10-11T11:03:35.084534+08:00|err|ovs-vswitchd[1721043]|i40evf_handle_aq_msg():
 command mismatch,expect 8, get 14
2021-10-11T11:03:35.134545+08:00|debug|ovs-vswitchd[1721043]|   100 [Port 1: 
mux_machine] DETACHED -> WAITING
2021-10-11T11:03:35.544648+08:00|info|ovs-vswitchd[1721043]|EAL: [eth_dev_ops] 
Slave 1: start dev.

How can I solve this problem ?
Thakns for help.
Regards,
Liao


Re: [dpdk-dev] [PATCH v5 1/5] eal: add new definitions for wait scheme

2021-10-26 Thread Ananyev, Konstantin

> > > Introduce macros as generic interface for address monitoring.
> > > For different size, encapsulate '__LOAD_EXC_16', '__LOAD_EXC_32'
> > > and '__LOAD_EXC_64' into a new macro '__LOAD_EXC'.
> > >
> > > Furthermore, to prevent compilation warning in arm:
> > > --
> > > 'warning: implicit declaration of function ...'
> > > --
> > > Delete 'undef' constructions for '__LOAD_EXC_xx', '__SEVL' and '__WFE'.
> > >
> > > This is because original macros are undefine at the end of the file.
> > > If new macro 'rte_wait_event' calls them in other files, they will be 
> > > seen as
> > > 'not defined'.
> > >
> > > Signed-off-by: Feifei Wang 
> > > Reviewed-by: Ruifeng Wang 
> > > ---
> > >  lib/eal/arm/include/rte_pause_64.h  | 135 
> > > lib/eal/include/generic/rte_pause.h |  27 ++
> > >  2 files changed, 105 insertions(+), 57 deletions(-)
> > >
> > > diff --git a/lib/eal/arm/include/rte_pause_64.h
> > > b/lib/eal/arm/include/rte_pause_64.h
> > > index e87d10b8cc..1fea0dec63 100644
> > > --- a/lib/eal/arm/include/rte_pause_64.h
> > > +++ b/lib/eal/arm/include/rte_pause_64.h
> > > @@ -31,20 +31,12 @@ static inline void rte_pause(void)
> > >  /* Put processor into low power WFE(Wait For Event) state. */  #define
> > > __WFE() { asm volatile("wfe" : : : "memory"); }
> > >
> > > -static __rte_always_inline void
> > > -rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t expected,
> > > - int memorder)
> > > -{
> > > - uint16_t value;
> > > -
> > > - assert(memorder == __ATOMIC_ACQUIRE || memorder ==
> > > __ATOMIC_RELAXED);
> > > -
> > > - /*
> > > -  * Atomic exclusive load from addr, it returns the 16-bit content of
> > > -  * *addr while making it 'monitored',when it is written by someone
> > > -  * else, the 'monitored' state is cleared and a event is generated
> > > -  * implicitly to exit WFE.
> > > -  */
> > > +/*
> > > + * Atomic exclusive load from addr, it returns the 16-bit content of
> > > + * *addr while making it 'monitored', when it is written by someone
> > > + * else, the 'monitored' state is cleared and an event is generated
> > > + * implicitly to exit WFE.
> > > + */
> > >  #define __LOAD_EXC_16(src, dst, memorder) {   \
> > >   if (memorder == __ATOMIC_RELAXED) {   \
> > >   asm volatile("ldxrh %w[tmp], [%x[addr]]"  \ @@ -58,6 +50,62
> > > @@ rte_wait_until_equal_16(volatile uint16_t *addr, uint16_t expected,
> > >   : "memory");  \
> > >   } }
> > >
> > > +/*
> > > + * Atomic exclusive load from addr, it returns the 32-bit content of
> > > + * *addr while making it 'monitored', when it is written by someone
> > > + * else, the 'monitored' state is cleared and an event is generated
> > > + * implicitly to exit WFE.
> > > + */
> > > +#define __LOAD_EXC_32(src, dst, memorder) {  \
> > > + if (memorder == __ATOMIC_RELAXED) {  \
> > > + asm volatile("ldxr %w[tmp], [%x[addr]]"  \
> > > + : [tmp] "=&r" (dst)  \
> > > + : [addr] "r"(src)\
> > > + : "memory"); \
> > > + } else { \
> > > + asm volatile("ldaxr %w[tmp], [%x[addr]]" \
> > > + : [tmp] "=&r" (dst)  \
> > > + : [addr] "r"(src)\
> > > + : "memory"); \
> > > + } }
> > > +
> > > +/*
> > > + * Atomic exclusive load from addr, it returns the 64-bit content of
> > > + * *addr while making it 'monitored', when it is written by someone
> > > + * else, the 'monitored' state is cleared and an event is generated
> > > + * implicitly to exit WFE.
> > > + */
> > > +#define __LOAD_EXC_64(src, dst, memorder) {  \
> > > + if (memorder == __ATOMIC_RELAXED) {  \
> > > + asm volatile("ldxr %x[tmp], [%x[addr]]"  \
> > > + : [tmp] "=&r" (dst)  \
> > > + : [addr] "r"(src)\
> > > + : "memory"); \
> > > + } else { \
> > > + asm volatile("ldaxr %x[tmp], [%x[addr]]" \
> > > + : [tmp] "=&r" (dst)  \
> > > + : [addr] "r"(src)\
> > > + : "memory"); \
> > > + } }
> > > +
> > > +#define __LOAD_EXC(src, dst, memorder, size) {  \
> > > + assert(size == 16 || size == 32 || size == 64); \
> > > + if (size == 16) \
> > > + __LOAD_EXC_16(src, dst, memorder)   \
> > > + else if (size == 32)\
> > > + __LOAD_EXC_32(src, dst, memorder)   \
> > > + else if (size == 64)\
> > > + __LOAD_EXC_64(src, dst, memorder)   \
> > > +}

Re: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library

2021-10-26 Thread Dumitrescu, Cristian



> -Original Message-
> From: Thomas Monjalon 
> Sent: Tuesday, October 26, 2021 9:33 AM
> To: Liguzinski, WojciechX ; Singh, Jasvinder
> ; Dumitrescu, Cristian
> ; Liu, Yu Y 
> Cc: dev@dpdk.org; Ajmera, Megha ; Liu, Yu Y
> ; david.march...@redhat.com
> Subject: Re: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library
>
> 26/10/2021 10:24, Liu, Yu Y:
> > Hi Thomas,
> >
> > Would you merge this patch as the series is acked by Cristian as below?
> >
> https://patchwork.dpdk.org/project/dpdk/cover/20211019081902.3514841-
> 1-wojciechx.liguzin...@intel.com/
>
> I didn't see any email from Cristian.
> It seems you just added this ack silently at the bottom of the cover letter.
>
> 1/ an email from Cristian is far better
> 2/ when integrating ack, it must be done in patches, not cover letter
>

Hi Thomas,

I did ack this set in a previous version (V15) by replying with 
"Series-acked-by" on the cover letter email, which does not show in patchwork. 
Is there a better way to do this?

It would be good to have Jasvinder's ack as well on this series, as he is 
looking into some other aspects of the sched library.

Regards,
Cristian
>
> >
> > Thanks & Regards,
> > Yu Liu
> >
> > -Original Message-
> > From: dev  On Behalf Of Liguzinski, WojciechX
> > Sent: Monday, October 25, 2021 7:32 PM
> > To: dev@dpdk.org; Singh, Jasvinder ;
> Dumitrescu, Cristian 
> > Cc: Ajmera, Megha 
> > Subject: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library
> >
> > DPDK sched library is equipped with mechanism that secures it from the
> bufferbloat problem which is a situation when excess buffers in the network
> cause high latency and latency variation. Currently, it supports RED for 
> active
> queue management. However, more advanced queue management is
> required to address this problem and provide desirable quality of service to
> users.
> >
> > This solution (RFC) proposes usage of new algorithm called "PIE"
> (Proportional Integral controller Enhanced) that can effectively and directly
> control queuing latency to address the bufferbloat problem.
> >
> > The implementation of mentioned functionality includes modification of
> existing and adding a new set of data structures to the library, adding PIE
> related APIs.
> > This affects structures in public API/ABI. That is why deprecation notice is
> going to be prepared and sent.
> >
> > Liguzinski, WojciechX (5):
> >   sched: add PIE based congestion management
> >   example/qos_sched: add PIE support
> >   example/ip_pipeline: add PIE support
> >   doc/guides/prog_guide: added PIE
> >   app/test: add tests for PIE
> >
> >  app/test/meson.build |4 +
> >  app/test/test_pie.c  | 1065 ++
> >  config/rte_config.h  |1 -
> >  doc/guides/prog_guide/glossary.rst   |3 +
> >  doc/guides/prog_guide/qos_framework.rst  |   64 +-
> >  doc/guides/prog_guide/traffic_management.rst |   13 +-
> >  drivers/net/softnic/rte_eth_softnic_tm.c |6 +-
> >  examples/ip_pipeline/tmgr.c  |  142 +--
> >  examples/qos_sched/cfg_file.c|  127 ++-
> >  examples/qos_sched/cfg_file.h|5 +
> >  examples/qos_sched/init.c|   27 +-
> >  examples/qos_sched/main.h|3 +
> >  examples/qos_sched/profile.cfg   |  196 ++--
> >  lib/sched/meson.build|3 +-
> >  lib/sched/rte_pie.c  |   86 ++
> >  lib/sched/rte_pie.h  |  398 +++
> >  lib/sched/rte_sched.c|  241 ++--
> >  lib/sched/rte_sched.h|   63 +-
> >  lib/sched/version.map|4 +
> >  19 files changed, 2172 insertions(+), 279 deletions(-)  create mode 100644
> app/test/test_pie.c  create mode 100644 lib/sched/rte_pie.c  create mode
> 100644 lib/sched/rte_pie.h
> >
> > --
> > 2.25.1
> >
> > Series-acked-by: Cristian Dumitrescu 
> >
>
>
>
>



Re: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library

2021-10-26 Thread Thomas Monjalon
26/10/2021 12:02, Dumitrescu, Cristian:
> From: Thomas Monjalon 
> > 26/10/2021 10:24, Liu, Yu Y:
> > > Hi Thomas,
> > >
> > > Would you merge this patch as the series is acked by Cristian as below?
> > >
> > https://patchwork.dpdk.org/project/dpdk/cover/20211019081902.3514841-
> > 1-wojciechx.liguzin...@intel.com/
> > 
> > I didn't see any email from Cristian.
> > It seems you just added this ack silently at the bottom of the cover letter.
> > 
> > 1/ an email from Cristian is far better
> > 2/ when integrating ack, it must be done in patches, not cover letter
> > 
> 
> Hi Thomas,
> 
> I did ack this set in a previous version (V15) by replying with 
> "Series-acked-by" on the cover letter email, which does not show in 
> patchwork. Is there a better way to do this?

No you did the right thing (I missed this email on v15).
But v16 did not show your ack.
And v17 added it only in the cover letter instead of reporting it in all 
patches.


> It would be good to have Jasvinder's ack as well on this series, as he is 
> looking into some other aspects of the sched library.

Yes




Re: [dpdk-dev] [PATCH v4] doc: document a limitation for a meter with RSS action

2021-10-26 Thread Thomas Monjalon
26/10/2021 12:10, Li Zhang:
> A meter policy with RSS/Queue action is not supported
> when dv_xmeta_en enabled.
> 
> When dv_xmeta_en enabled in legacy creating flow,
> it will split into two flows
> (one set_tag with jump flow and one RSS/queue action flow).
> For meter policy as termination table,
> it cannot split flow and
> cannot support dv_xmeta_en enabled.

Thank you for the explanation.

> Fixes: 51ec04d ("net/mlx5: connect meter policy to created flows")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Li Zhang 
> Acked-by: Matan Azrad 
> ---
>  doc/guides/nics/mlx5.rst | 1 +
>  1 file changed, 1 insertion(+)
> 
> diff --git a/doc/guides/nics/mlx5.rst b/doc/guides/nics/mlx5.rst
> index 47709d93b3..2f46ed0263 100644
> --- a/doc/guides/nics/mlx5.rst
> +++ b/doc/guides/nics/mlx5.rst
> @@ -435,6 +435,7 @@ Limitations
>   - yellow: QUEUE, RSS, PORT_ID, REPRESENTED_PORT, JUMP, DROP, MARK and 
> SET_TAG.
>   - RED: must be DROP.
>- Policy actions of RSS for green and yellow should have the same 
> configuration except queues.
> +  - Policy with RSS/queue action is not supported when dv_xmeta_en enabled.

Did you understand my comment about the formatting?
You must add quotes like that: ``dv_xmeta_en``
Look at other occurences in the doc.






Re: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library

2021-10-26 Thread Liguzinski, WojciechX
Hi,

V16 - My bad, probably I haven't copied it correctly when preparing cover letter
V17 - I understood Cristian's comment as to copy the Series ACK to next 
versions of patches, and not to "split it" for each one. If that was the 
correct way I had no knowledge about it.

Wojtek

-Original Message-
From: Thomas Monjalon  
Sent: Tuesday, October 26, 2021 12:10 PM
To: Liguzinski, WojciechX ; Singh, Jasvinder 
; Liu, Yu Y ; Singh, Jasvinder 
; Dumitrescu, Cristian 

Cc: dev@dpdk.org; Ajmera, Megha ; Liu, Yu Y 
; david.march...@redhat.com
Subject: Re: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library

26/10/2021 12:02, Dumitrescu, Cristian:
> From: Thomas Monjalon 
> > 26/10/2021 10:24, Liu, Yu Y:
> > > Hi Thomas,
> > >
> > > Would you merge this patch as the series is acked by Cristian as below?
> > >
> > https://patchwork.dpdk.org/project/dpdk/cover/20211019081902.3514841
> > -
> > 1-wojciechx.liguzin...@intel.com/
> > 
> > I didn't see any email from Cristian.
> > It seems you just added this ack silently at the bottom of the cover letter.
> > 
> > 1/ an email from Cristian is far better 2/ when integrating ack, it 
> > must be done in patches, not cover letter
> > 
> 
> Hi Thomas,
> 
> I did ack this set in a previous version (V15) by replying with 
> "Series-acked-by" on the cover letter email, which does not show in 
> patchwork. Is there a better way to do this?

No you did the right thing (I missed this email on v15).
But v16 did not show your ack.
And v17 added it only in the cover letter instead of reporting it in all 
patches.


> It would be good to have Jasvinder's ack as well on this series, as he is 
> looking into some other aspects of the sched library.

Yes




Re: [dpdk-dev] [PATCH v2] ethdev: fix one MAC address occupies two index in mac addrs

2021-10-26 Thread Ferruh Yigit

On 10/22/2021 3:04 AM, lihuisong (C) wrote:


在 2021/10/21 16:30, Ferruh Yigit 写道:

On 10/21/2021 3:05 AM, lihuisong (C) wrote:


在 2021/10/21 0:32, Ferruh Yigit 写道:

On 10/20/2021 11:15 AM, Kevin Traynor wrote:

On 20/10/2021 08:41, Ferruh Yigit wrote:

On 10/20/2021 7:49 AM, lihuisong (C) wrote:

Hi Ferruh

在 2021/10/20 1:45, Ferruh Yigit 写道:

On 10/11/2021 10:28 AM, Min Hu (Connor) wrote:

From: Huisong Li 

The dev->data->mac_addrs[0] will be changed to a new MAC address when
applications modify the default MAC address by
rte_eth_dev_default_mac_addr_set() API. However, If the new default
MAC address has been added as a non-default MAC address by
rte_eth_dev_mac_addr_add() API, the rte_eth_dev_default_mac_addr_set()
API doesn't remove it from dev->data->mac_addrs[]. As a result, one MAC
address occupies two index capacities in dev->data->mac_addrs[].



Hi Connor,

I see the problem, but can you please clarify what is the impact to the end 
user?

If application does as following:
   rte_eth_dev_mac_addr_add(MAC1);
   rte_eth_dev_mac_addr_add(MAC2);
   rte_eth_dev_mac_addr_add(MAC3);
   rte_eth_dev_default_mac_addr_set(MAC2);

The 'dev->data->mac_addrs[]' will have: "MAC2,MAC2,MAC3" which has 'MAC2' 
duplicated.

Will this cause any problem for the application to receive the packets
with 'MAC2' address?
Or is the only problem one extra space used in 'dev->data->mac_addrs[]'
without any other impact to the application?

I think it's just a waste of space.


True, it is a waste. But if there is no other visible user impact, we can
handle the issue with lower priority and clarifying the impact in commit log
helps to others.




This patch adds the logic of removing MAC addresses for this scenario.

Fixes: 854d8ad4ef68 ("ethdev: add default mac address modifier")
Cc: sta...@dpdk.org

Signed-off-by: Huisong Li 
Signed-off-by: Min Hu (Connor) 
---
v2:
* fixed commit log.
---
   lib/ethdev/rte_ethdev.c | 15 +++
   1 file changed, 15 insertions(+)

diff --git a/lib/ethdev/rte_ethdev.c b/lib/ethdev/rte_ethdev.c
index 028907bc4b..7faff17d9a 100644
--- a/lib/ethdev/rte_ethdev.c
+++ b/lib/ethdev/rte_ethdev.c
@@ -4340,6 +4340,7 @@ int
   rte_eth_dev_default_mac_addr_set(uint16_t port_id, struct rte_ether_addr 
*addr)
   {
   struct rte_eth_dev *dev;
+    int index;
   int ret;
     RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
@@ -4361,6 +4362,20 @@ rte_eth_dev_default_mac_addr_set(uint16_t port_id, 
struct rte_ether_addr *addr)
   if (ret < 0)
   return ret;
   +    /*
+ * If the address has been added as a non-default MAC address by
+ * rte_eth_dev_mac_addr_add API, it should be removed from
+ * dev->data->mac_addrs[].
+ */
+    index = eth_dev_get_mac_addr_index(port_id, addr);
+    if (index > 0) {
+    /* remove address in NIC data structure */
+    rte_ether_addr_copy(&null_mac_addr,
+ &dev->data->mac_addrs[index]);
+    /* reset pool bitmap */
+    dev->data->mac_pool_sel[index] = 0;
+    }
+


Here only 'dev->data->mac_addrs[]' array is updated and it assumes
driver removes similar duplication itself, but I am not sure if this is
valid for all drivers.

If driver is not removing the duplicate in the HW configuration, the driver
config and 'dev->data->mac_addrs[]' will diverge, which is not good.

The same MAC address does not occupy two HW entries, which is also a
waste for HW. After all, HW entry resources are also limited.
The PMD should also take this into account.
So, I think, we don't have to think about it here.


I am not sure all PMD take this into account, I briefly checked the ixgbe
code and I am not sure if it handles this.

Also it is possible to think that this responsibility is pushed to the
application, like application should remove a MAC address before setting
it as default MAC...



Yes, the API view is more important than saving one entry in an array. From API 
perspective with this patch,

set_default(MAC1)
add(MAC2)
add(MAC3)
add(MAC4)
default=MAC1, Filters=MAC2, MAC3, MAC4

set_default(MAC2)
default=MAC2, Filters= MAC3, MAC4

set_default(MAC3)
default=MAC3, Filters= MAC4

set_default(MAC4)
default=MAC4, Filters=

set_default(MAC5)
default=MAC5, Filters=

Did I get it right? If so, it seems wrong to silently remove the filters. In 
which case, it would be easier to just not remove them in the first place 
(current behaviour).



Yep, this is the updated behavior. And agree it looks wrong when you
show like this. (btw, this is only ethdev record of MAC filters, what
is updated in this patch, HW still may be keeping all filters.)


Whether HW saves all filters depends on the implementation of the set_default()

in the driver. According to the implementation of this API of all PMDs, some 
drivers

will first remove the old default MAC in HW and then add the new one when 
calling

the set_default(). I am not sure if the HW that didn't do this would remove the 
old

default MAC. If not, we may need to standardize this 

Re: [dpdk-dev] [PATCH v18 0/5] Add PIE support for HQoS library

2021-10-26 Thread Thomas Monjalon
26/10/2021 12:20, Liguzinski, WojciechX:
> Hi,
> 
> V16 - My bad, probably I haven't copied it correctly when preparing cover 
> letter
> V17 - I understood Cristian's comment as to copy the Series ACK to next 
> versions of patches, and not to "split it" for each one. If that was the 
> correct way I had no knowledge about it.

Yes you had to reproduce it in each patch. Otherwise who would do it
to make it appear in patchwork and in the git history when merged?
I understand you did not have that knowledge.
For future, I hope the Intel team will better track features patches
of newcomers so they don't miss something knew by others.

John, I know knowledge sharing is not an easy task, we always have to improve :)



> From: Thomas Monjalon  
> 26/10/2021 12:02, Dumitrescu, Cristian:
> > From: Thomas Monjalon 
> > > 26/10/2021 10:24, Liu, Yu Y:
> > > > Hi Thomas,
> > > >
> > > > Would you merge this patch as the series is acked by Cristian as below?
> > > >
> > > https://patchwork.dpdk.org/project/dpdk/cover/20211019081902.3514841
> > > -
> > > 1-wojciechx.liguzin...@intel.com/
> > > 
> > > I didn't see any email from Cristian.
> > > It seems you just added this ack silently at the bottom of the cover 
> > > letter.
> > > 
> > > 1/ an email from Cristian is far better 2/ when integrating ack, it 
> > > must be done in patches, not cover letter
> > > 
> > 
> > Hi Thomas,
> > 
> > I did ack this set in a previous version (V15) by replying with 
> > "Series-acked-by" on the cover letter email, which does not show in 
> > patchwork. Is there a better way to do this?
> 
> No you did the right thing (I missed this email on v15).
> But v16 did not show your ack.
> And v17 added it only in the cover letter instead of reporting it in all 
> patches.
> 
> 
> > It would be good to have Jasvinder's ack as well on this series, as he is 
> > looking into some other aspects of the sched library.
> 
> Yes
> 
> 
> 







Re: [dpdk-dev] [PATCH] vhost: fix async DMA map

2021-10-26 Thread Ding, Xuan
Hi,

>-Original Message-
>From: Maxime Coquelin 
>Sent: Tuesday, October 26, 2021 5:49 PM
>To: Ding, Xuan ; dev@dpdk.org;
>david.march...@redhat.com; Xia, Chenbo 
>Cc: Burakov, Anatoly 
>Subject: Re: [PATCH] vhost: fix async DMA map
>
>
>
>On 10/26/21 10:49, Ding, Xuan wrote:
>> Hi Maxime,
>>
>>> -Original Message-
>>> From: Maxime Coquelin 
>>> Sent: Tuesday, October 26, 2021 2:53 PM
>>> To: Ding, Xuan ; dev@dpdk.org;
>>> david.march...@redhat.com; Xia, Chenbo 
>>> Cc: Burakov, Anatoly 
>>> Subject: Re: [PATCH] vhost: fix async DMA map
>>>
>>>
>>>
>>> On 10/26/21 04:07, Ding, Xuan wrote:
 Hi Maxime,

> -Original Message-
> From: Maxime Coquelin 
> Sent: Tuesday, October 26, 2021 4:47 AM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; Ding, Xuan 
> Subject: Re: [PATCH] vhost: fix async DMA map
>
> Hi Xuan,
>
> On 10/25/21 22:33, Maxime Coquelin wrote:
>> This patch fixes possible NULL-pointer dereferencing
>> reported by Coverity and also fixes NUMA reallocation
>> of the async DMA map.
>>
>> Fixes: 7c61fa08b716 ("vhost: enable IOMMU for async vhost")
>>
>> Coverity issue: 373655
>>
>> Signed-off-by: Maxime Coquelin 
>> ---
>> lib/vhost/vhost_user.c | 45 
>> +++---
>> 1 file changed, 20 insertions(+), 25 deletions(-)
>>
>
> I posted this patch to fix the issue reported by Coverity and also other
> issue on NUMA realloc that I found at the same time. But I wonder
> whether all this async_map_status is needed.

 Thanks for your fix! I can help to review and test the patch later.

 I add the async_map_status in v2 for compatibility. Some DMA device,
 like DSA, may use kernel idxd driver only. If there is no device bound to
 DPDK vfio and kernel vfio module is modprobed to ensure
>>> rte_vfio_is_enabled() is true,
 we will unavoidably do DMA map/unmap and it will fail.

 Therefore, the dma_map_status here is used to filter this situation by
>>> preventing
 unnecessary DMA unmap.
>>>
>>> Ok, then I think we can just remove the async DMA map.
>>>
>
> Indeed, if the only place where we DMA map is in
> vhost_user_mmap_region(). If it fails, the error is propagated, the mem
> table are freed and NACK is replied to the master. IOW, the device will
> be in an unusable state.

 I agree with you, this is the place I consider right to do DMA map
 because we also do SW mapping here, any suggestions?
>>>
>>> No suggestion, I was just explaining that at the only place where
>>> DMA map were done, mapping errors were properly handled and
>propagated.
>>
>> What about just setting async_copy to false, and allow switching to sync
>path.
>>
>>>
>
> Removing the async DMA map will simplify a lot the code, do you agree
>to
> remove it or there is something I missed?

 See above. Indeed, it adds a lot of code. But we can't know the driver for
 each device in vhost lib, or we can only restrict the user to bind some
>>> devices
 to DPDK vfio if async logic needed.
>>>
>>> I would think we don't care if DMA unmap fails, we can just do the same
>>> as what you do for DMA map, i.e. just ignore the error.
>>
>> Get your idea, we can do the same as DMA map, and in this way
>dma_map_status flag can be removed.
>>
>>>
>>> Thanks to this discussion, I have now more concerns on how it works. I
>>> think we have a problem here in case of DMA device hotplug, that device
>>> could miss the necessary map entries from Vhost if no VFIO devices were
>>> attached at VHST_USER_SET_MEM_TABLE time. How would you handle
>that
>>> case?
>>
>> DMA device is uncore, so I don't see the  hotplug issue here.
>
>I'm not sure what 'uncore' is, I suppose you mean your device cannot be
>physically added/removed to the system.

Yes, we are at the same understanding.

>
>I was not clear enough in my question. I meant that for example, the
>application is started and the Vhost port is created. Then, the DMA
>device is bound to VFIO, and probed by the DPDK application. Finally,
>the application register the DMA device as an async channel in Vhost.
>
>I think it will not work as the SET_MEM_TABLE will already have
>happened, so the guest memory won't be mapped in the VFIO container.

Assuming the DMA device supports hotplug, this situation was not actually took 
into consideration, and maybe the handling should be added in app, with an 
event callback.

>
>Do you have the same understanding?
>
>> I will have another patch containing compatibility with sync path, and
>async_map_status flag will be removed.
>> Hope to get your insights.
>
>What do you mean by compatibility with sync path?

I mean whatever DMA map succeeds or fails, we return 0 so as not to prevent
SET_EM_TABLE.

>
>Thanks for taking care of the async map status removal.

Found a tricky point for removing 

[dpdk-dev] [PATCH 1/2] eventdev/eth_rx: add queue stats get API

2021-10-26 Thread Naga Harish K S V
This patch adds new api ``rte_event_eth_rx_adapter_queue_stats_get``
to retrieve queue stats. The queue stats are in the format
``struct rte_event_eth_rx_adapter_queue_stats``.

The adapter stats_get and stats_reset apis are also updated to
handle queue level event buffer use case.

Signed-off-by: Naga Harish K S V 
---
 lib/eventdev/rte_event_eth_rx_adapter.c | 201 +++-
 lib/eventdev/rte_event_eth_rx_adapter.h |  43 +
 lib/eventdev/version.map|   1 +
 3 files changed, 202 insertions(+), 43 deletions(-)

diff --git a/lib/eventdev/rte_event_eth_rx_adapter.c 
b/lib/eventdev/rte_event_eth_rx_adapter.c
index a175c61551..a485c89ffe 100644
--- a/lib/eventdev/rte_event_eth_rx_adapter.c
+++ b/lib/eventdev/rte_event_eth_rx_adapter.c
@@ -245,6 +245,7 @@ struct eth_rx_queue_info {
uint64_t event;
struct eth_rx_vector_data vector_data;
struct eth_event_enqueue_buffer *event_buf;
+   struct rte_event_eth_rx_adapter_stats *stats;
 };
 
 static struct event_eth_rx_adapter **event_eth_rx_adapter;
@@ -268,14 +269,18 @@ rxa_validate_id(uint8_t id)
 
 static inline struct eth_event_enqueue_buffer *
 rxa_event_buf_get(struct event_eth_rx_adapter *rx_adapter, uint16_t eth_dev_id,
- uint16_t rx_queue_id)
+ uint16_t rx_queue_id,
+ struct rte_event_eth_rx_adapter_stats **stats)
 {
if (rx_adapter->use_queue_event_buf) {
struct eth_device_info *dev_info =
&rx_adapter->eth_devices[eth_dev_id];
+   *stats = dev_info->rx_queue[rx_queue_id].stats;
return dev_info->rx_queue[rx_queue_id].event_buf;
-   } else
+   } else {
+   *stats = &rx_adapter->stats;
return &rx_adapter->event_enqueue_buffer;
+   }
 }
 
 #define RTE_EVENT_ETH_RX_ADAPTER_ID_VALID_OR_ERR_RET(id, retval) do { \
@@ -766,9 +771,9 @@ rxa_enq_block_end_ts(struct event_eth_rx_adapter 
*rx_adapter,
 /* Enqueue buffered events to event device */
 static inline uint16_t
 rxa_flush_event_buffer(struct event_eth_rx_adapter *rx_adapter,
-  struct eth_event_enqueue_buffer *buf)
+  struct eth_event_enqueue_buffer *buf,
+  struct rte_event_eth_rx_adapter_stats *stats)
 {
-   struct rte_event_eth_rx_adapter_stats *stats = &rx_adapter->stats;
uint16_t count = buf->last ? buf->last - buf->head : buf->count;
 
if (!count)
@@ -883,7 +888,8 @@ rxa_create_event_vector(struct event_eth_rx_adapter 
*rx_adapter,
 static inline void
 rxa_buffer_mbufs(struct event_eth_rx_adapter *rx_adapter, uint16_t eth_dev_id,
 uint16_t rx_queue_id, struct rte_mbuf **mbufs, uint16_t num,
-struct eth_event_enqueue_buffer *buf)
+struct eth_event_enqueue_buffer *buf,
+struct rte_event_eth_rx_adapter_stats *stats)
 {
uint32_t i;
struct eth_device_info *dev_info =
@@ -954,7 +960,7 @@ rxa_buffer_mbufs(struct event_eth_rx_adapter *rx_adapter, 
uint16_t eth_dev_id,
else
num = nb_cb;
if (dropped)
-   rx_adapter->stats.rx_dropped += dropped;
+   stats->rx_dropped += dropped;
}
 
buf->count += num;
@@ -985,11 +991,10 @@ rxa_pkt_buf_available(struct eth_event_enqueue_buffer 
*buf)
 static inline uint32_t
 rxa_eth_rx(struct event_eth_rx_adapter *rx_adapter, uint16_t port_id,
   uint16_t queue_id, uint32_t rx_count, uint32_t max_rx,
-  int *rxq_empty, struct eth_event_enqueue_buffer *buf)
+  int *rxq_empty, struct eth_event_enqueue_buffer *buf,
+  struct rte_event_eth_rx_adapter_stats *stats)
 {
struct rte_mbuf *mbufs[BATCH_SIZE];
-   struct rte_event_eth_rx_adapter_stats *stats =
-   &rx_adapter->stats;
uint16_t n;
uint32_t nb_rx = 0;
 
@@ -1000,7 +1005,7 @@ rxa_eth_rx(struct event_eth_rx_adapter *rx_adapter, 
uint16_t port_id,
 */
while (rxa_pkt_buf_available(buf)) {
if (buf->count >= BATCH_SIZE)
-   rxa_flush_event_buffer(rx_adapter, buf);
+   rxa_flush_event_buffer(rx_adapter, buf, stats);
 
stats->rx_poll_count++;
n = rte_eth_rx_burst(port_id, queue_id, mbufs, BATCH_SIZE);
@@ -1009,14 +1014,17 @@ rxa_eth_rx(struct event_eth_rx_adapter *rx_adapter, 
uint16_t port_id,
*rxq_empty = 1;
break;
}
-   rxa_buffer_mbufs(rx_adapter, port_id, queue_id, mbufs, n, buf);
+   rxa_buffer_mbufs(rx_adapter, port_id, queue_id, mbufs, n, buf,
+stats);
nb_rx += n;
if (rx_count + nb_rx > max_rx)
break;
}
 
if (buf->count > 0)
-   rxa_flush_eve

[dpdk-dev] [PATCH 2/2] eventdev/eth_rx: support telemetry

2021-10-26 Thread Naga Harish K S V
Added telemetry support for rxa_queue_stats to get rx queue
stats

Signed-off-by: Naga Harish K S V 
---
 lib/eventdev/rte_event_eth_rx_adapter.c | 67 +
 1 file changed, 67 insertions(+)

diff --git a/lib/eventdev/rte_event_eth_rx_adapter.c 
b/lib/eventdev/rte_event_eth_rx_adapter.c
index a485c89ffe..3e28ee5379 100644
--- a/lib/eventdev/rte_event_eth_rx_adapter.c
+++ b/lib/eventdev/rte_event_eth_rx_adapter.c
@@ -3278,6 +3278,69 @@ handle_rxa_get_queue_conf(const char *cmd __rte_unused,
return 0;
 }
 
+static int
+handle_rxa_get_queue_stats(const char *cmd __rte_unused,
+  const char *params,
+  struct rte_tel_data *d)
+{
+   uint8_t rx_adapter_id;
+   uint16_t rx_queue_id;
+   int eth_dev_id;
+   char *token, *l_params;
+   struct rte_event_eth_rx_adapter_queue_stats q_stats;
+
+   if (params == NULL || strlen(params) == 0 || !isdigit(*params))
+   return -1;
+
+   /* Get Rx adapter ID from parameter string */
+   l_params = strdup(params);
+   token = strtok(l_params, ",");
+   rx_adapter_id = strtoul(token, NULL, 10);
+   RTE_EVENT_ETH_RX_ADAPTER_ID_VALID_OR_ERR_RET(rx_adapter_id, -EINVAL);
+
+   token = strtok(NULL, ",");
+   if (token == NULL || strlen(token) == 0 || !isdigit(*token))
+   return -1;
+
+   /* Get device ID from parameter string */
+   eth_dev_id = strtoul(token, NULL, 10);
+   RTE_EVENTDEV_VALID_DEVID_OR_ERR_RET(eth_dev_id, -EINVAL);
+
+   token = strtok(NULL, ",");
+   if (token == NULL || strlen(token) == 0 || !isdigit(*token))
+   return -1;
+
+   /* Get Rx queue ID from parameter string */
+   rx_queue_id = strtoul(token, NULL, 10);
+   if (rx_queue_id >= rte_eth_devices[eth_dev_id].data->nb_rx_queues) {
+   RTE_EDEV_LOG_ERR("Invalid rx queue_id %u", rx_queue_id);
+   return -EINVAL;
+   }
+
+   token = strtok(NULL, "\0");
+   if (token != NULL)
+   RTE_EDEV_LOG_ERR("Extra parameters passed to eventdev"
+" telemetry command, igrnoring");
+
+   if (rte_event_eth_rx_adapter_queue_stats_get(rx_adapter_id, eth_dev_id,
+   rx_queue_id, &q_stats)) {
+   RTE_EDEV_LOG_ERR("Failed to get Rx adapter queue stats");
+   return -1;
+   }
+
+   rte_tel_data_start_dict(d);
+   rte_tel_data_add_dict_u64(d, "rx_adapter_id", rx_adapter_id);
+   rte_tel_data_add_dict_u64(d, "eth_dev_id", eth_dev_id);
+   rte_tel_data_add_dict_u64(d, "rx_queue_id", rx_queue_id);
+   RXA_ADD_DICT(q_stats, rx_event_buf_count);
+   RXA_ADD_DICT(q_stats, rx_event_buf_size);
+   RXA_ADD_DICT(q_stats, rx_poll_count);
+   RXA_ADD_DICT(q_stats, rx_packets);
+   RXA_ADD_DICT(q_stats, rx_dropped);
+
+   return 0;
+}
+
 RTE_INIT(rxa_init_telemetry)
 {
rte_telemetry_register_cmd("/eventdev/rxa_stats",
@@ -3291,4 +3354,8 @@ RTE_INIT(rxa_init_telemetry)
rte_telemetry_register_cmd("/eventdev/rxa_queue_conf",
handle_rxa_get_queue_conf,
"Returns Rx queue config. Parameter: rxa_id, dev_id, queue_id");
+
+   rte_telemetry_register_cmd("/eventdev/rxa_queue_stats",
+   handle_rxa_get_queue_stats,
+   "Returns Rx queue stats. Parameter: rxa_id, dev_id, queue_id");
 }
-- 
2.25.1



[dpdk-dev] [PATCH v11 0/7] iavf: add iAVF IPsec inline crypto support

2021-10-26 Thread Radu Nicolau
Add support for inline crypto for IPsec, for ESP transport and
tunnel over IPv4 and IPv6, as well as supporting the offload for
ESP over UDP, and inconjunction with TSO for UDP and TCP flows.

Radu Nicolau (7):
  common/iavf: add iAVF IPsec inline crypto support
  net/iavf: rework tx path
  net/iavf: add support for asynchronous virt channel messages
  net/iavf: add iAVF IPsec inline crypto support
  net/iavf: add xstats support for inline IPsec crypto
  net/iavf: add watchdog for VFLR
  net/iavf: update doc with inline crypto support

 doc/guides/nics/features/iavf.ini |2 +
 doc/guides/nics/intel_vf.rst  |   10 +
 doc/guides/rel_notes/release_21_11.rst|1 +
 drivers/common/iavf/iavf_type.h   |1 +
 drivers/common/iavf/virtchnl.h|   17 +-
 drivers/common/iavf/virtchnl_inline_ipsec.h   |  553 +
 drivers/net/iavf/iavf.h   |   52 +-
 drivers/net/iavf/iavf_ethdev.c|  219 +-
 drivers/net/iavf/iavf_generic_flow.c  |   15 +
 drivers/net/iavf/iavf_generic_flow.h  |2 +
 drivers/net/iavf/iavf_ipsec_crypto.c  | 1894 +
 drivers/net/iavf/iavf_ipsec_crypto.h  |  160 ++
 .../net/iavf/iavf_ipsec_crypto_capabilities.h |  383 
 drivers/net/iavf/iavf_rxtx.c  |  710 --
 drivers/net/iavf/iavf_rxtx.h  |  220 +-
 drivers/net/iavf/iavf_rxtx_vec_sse.c  |   10 +-
 drivers/net/iavf/iavf_vchnl.c |  167 +-
 drivers/net/iavf/meson.build  |3 +-
 drivers/net/iavf/rte_pmd_iavf.h   |1 +
 drivers/net/iavf/version.map  |3 +
 20 files changed, 4101 insertions(+), 322 deletions(-)
 create mode 100644 drivers/common/iavf/virtchnl_inline_ipsec.h
 create mode 100644 drivers/net/iavf/iavf_ipsec_crypto.c
 create mode 100644 drivers/net/iavf/iavf_ipsec_crypto.h
 create mode 100644 drivers/net/iavf/iavf_ipsec_crypto_capabilities.h

-- 
v2: small updates and fixes in the flow related section
v3: split the huge patch and address feedback
v4: small changes due to dependencies changes
v5: updated the watchdow patch
v6: rebased and updated the common section
v7: fixed TSO issue and disabled watchdog by default
v8: rebased to next-net-intel and added doc updates
v9: fixed IV len for AEAD and GMAC
v10: removed blank lines at EOF
v11: rebased patchset
 
2.25.1



[dpdk-dev] [PATCH v11 1/7] common/iavf: add iAVF IPsec inline crypto support

2021-10-26 Thread Radu Nicolau
Add support for inline crypto for IPsec.

Signed-off-by: Declan Doherty 
Signed-off-by: Abhijit Sinha 
Signed-off-by: Radu Nicolau 
---
 drivers/common/iavf/iavf_type.h |   1 +
 drivers/common/iavf/virtchnl.h  |  17 +-
 drivers/common/iavf/virtchnl_inline_ipsec.h | 553 
 3 files changed, 569 insertions(+), 2 deletions(-)
 create mode 100644 drivers/common/iavf/virtchnl_inline_ipsec.h

diff --git a/drivers/common/iavf/iavf_type.h b/drivers/common/iavf/iavf_type.h
index 73dfb47e70..51267ca3b3 100644
--- a/drivers/common/iavf/iavf_type.h
+++ b/drivers/common/iavf/iavf_type.h
@@ -723,6 +723,7 @@ enum iavf_tx_desc_dtype_value {
IAVF_TX_DESC_DTYPE_NOP  = 0x1, /* same as Context desc */
IAVF_TX_DESC_DTYPE_CONTEXT  = 0x1,
IAVF_TX_DESC_DTYPE_FCOE_CTX = 0x2,
+   IAVF_TX_DESC_DTYPE_IPSEC= 0x3,
IAVF_TX_DESC_DTYPE_FILTER_PROG  = 0x8,
IAVF_TX_DESC_DTYPE_DDP_CTX  = 0x9,
IAVF_TX_DESC_DTYPE_FLEX_DATA= 0xB,
diff --git a/drivers/common/iavf/virtchnl.h b/drivers/common/iavf/virtchnl.h
index 067f715945..269578f7c0 100644
--- a/drivers/common/iavf/virtchnl.h
+++ b/drivers/common/iavf/virtchnl.h
@@ -38,6 +38,8 @@
  * value in current and future projects
  */
 
+#include "virtchnl_inline_ipsec.h"
+
 /* Error Codes */
 enum virtchnl_status_code {
VIRTCHNL_STATUS_SUCCESS = 0,
@@ -133,7 +135,8 @@ enum virtchnl_ops {
VIRTCHNL_OP_DISABLE_CHANNELS = 31,
VIRTCHNL_OP_ADD_CLOUD_FILTER = 32,
VIRTCHNL_OP_DEL_CLOUD_FILTER = 33,
-   /* opcodes 34, 35, 36, and 37 are reserved */
+   VIRTCHNL_OP_INLINE_IPSEC_CRYPTO = 34,
+   /* opcodes 35 and 36 are reserved */
VIRTCHNL_OP_DCF_CONFIG_BW = 37,
VIRTCHNL_OP_DCF_VLAN_OFFLOAD = 38,
VIRTCHNL_OP_DCF_CMD_DESC = 39,
@@ -225,6 +228,8 @@ static inline const char *virtchnl_op_str(enum virtchnl_ops 
v_opcode)
return "VIRTCHNL_OP_ADD_CLOUD_FILTER";
case VIRTCHNL_OP_DEL_CLOUD_FILTER:
return "VIRTCHNL_OP_DEL_CLOUD_FILTER";
+   case VIRTCHNL_OP_INLINE_IPSEC_CRYPTO:
+   return "VIRTCHNL_OP_INLINE_IPSEC_CRYPTO";
case VIRTCHNL_OP_DCF_CMD_DESC:
return "VIRTCHNL_OP_DCF_CMD_DESC";
case VIRTCHNL_OP_DCF_CMD_BUFF:
@@ -385,7 +390,7 @@ VIRTCHNL_CHECK_STRUCT_LEN(16, virtchnl_vsi_resource);
 #define VIRTCHNL_VF_OFFLOAD_REQ_QUEUES BIT(6)
 /* used to negotiate communicating link speeds in Mbps */
 #define VIRTCHNL_VF_CAP_ADV_LINK_SPEED BIT(7)
-   /* BIT(8) is reserved */
+#define VIRTCHNL_VF_OFFLOAD_INLINE_IPSEC_CRYPTOBIT(8)
 #define VIRTCHNL_VF_LARGE_NUM_QPAIRS   BIT(9)
 #define VIRTCHNL_VF_OFFLOAD_CRCBIT(10)
 #define VIRTCHNL_VF_OFFLOAD_VLAN_V2BIT(15)
@@ -2291,6 +2296,14 @@ virtchnl_vc_validate_vf_msg(struct virtchnl_version_info 
*ver, u32 v_opcode,
  sizeof(struct virtchnl_queue_vector);
}
break;
+
+   case VIRTCHNL_OP_INLINE_IPSEC_CRYPTO:
+   {
+   struct inline_ipsec_msg *iim = (struct inline_ipsec_msg *)msg;
+   valid_len =
+   virtchnl_inline_ipsec_val_msg_len(iim->ipsec_opcode);
+   break;
+   }
/* These are always errors coming from the VF. */
case VIRTCHNL_OP_EVENT:
case VIRTCHNL_OP_UNKNOWN:
diff --git a/drivers/common/iavf/virtchnl_inline_ipsec.h 
b/drivers/common/iavf/virtchnl_inline_ipsec.h
new file mode 100644
index 00..1e9134501e
--- /dev/null
+++ b/drivers/common/iavf/virtchnl_inline_ipsec.h
@@ -0,0 +1,553 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2001-2021 Intel Corporation
+ */
+
+#ifndef _VIRTCHNL_INLINE_IPSEC_H_
+#define _VIRTCHNL_INLINE_IPSEC_H_
+
+#define VIRTCHNL_IPSEC_MAX_CRYPTO_CAP_NUM  3
+#define VIRTCHNL_IPSEC_MAX_ALGO_CAP_NUM16
+#define VIRTCHNL_IPSEC_MAX_TX_DESC_NUM 128
+#define VIRTCHNL_IPSEC_MAX_CRYPTO_ITEM_NUMBER  2
+#define VIRTCHNL_IPSEC_MAX_KEY_LEN 128
+#define VIRTCHNL_IPSEC_MAX_SA_DESTROY_NUM  8
+#define VIRTCHNL_IPSEC_SA_DESTROY  0
+#define VIRTCHNL_IPSEC_BROADCAST_VFID  0x
+#define VIRTCHNL_IPSEC_INVALID_REQ_ID  0x
+#define VIRTCHNL_IPSEC_INVALID_SA_CFG_RESP 0x
+#define VIRTCHNL_IPSEC_INVALID_SP_CFG_RESP 0x
+
+/* crypto type */
+#define VIRTCHNL_AUTH  1
+#define VIRTCHNL_CIPHER2
+#define VIRTCHNL_AEAD  3
+
+/* caps enabled */
+#define VIRTCHNL_IPSEC_ESN_ENA BIT(0)
+#define VIRTCHNL_IPSEC_UDP_ENCAP_ENA   BIT(1)
+#define VIRTCHNL_IPSEC_SA_INDEX_SW_ENA BIT(2)
+#define VIRTCHNL_IPSEC_AUDIT_ENA   BIT(3)
+#define VIRTCHNL_IPSEC_BYTE_LIMIT_ENA  BIT(4)
+#define VIRTCHNL_IPSEC_DROP_ON_AUTH_FAIL_ENA   BIT(5)
+#define VIRTCHNL_IPSEC_ARW_CHECK_ENA  

[dpdk-dev] [PATCH v11 2/7] net/iavf: rework tx path

2021-10-26 Thread Radu Nicolau
Rework the TX path and TX descriptor usage in order to
allow for better use of oflload flags and to facilitate enabling of
inline crypto offload feature.

Signed-off-by: Declan Doherty 
Signed-off-by: Abhijit Sinha 
Signed-off-by: Radu Nicolau 
Acked-by: Jingjing Wu 
---
 drivers/net/iavf/iavf_rxtx.c | 538 ---
 drivers/net/iavf/iavf_rxtx.h | 117 +-
 drivers/net/iavf/iavf_rxtx_vec_sse.c |  10 +-
 3 files changed, 431 insertions(+), 234 deletions(-)

diff --git a/drivers/net/iavf/iavf_rxtx.c b/drivers/net/iavf/iavf_rxtx.c
index ac4db117f5..dbf71747c0 100644
--- a/drivers/net/iavf/iavf_rxtx.c
+++ b/drivers/net/iavf/iavf_rxtx.c
@@ -1054,27 +1054,31 @@ iavf_rxd_to_vlan_tci(struct rte_mbuf *mb, volatile 
union iavf_rx_desc *rxdp)
 
 static inline void
 iavf_flex_rxd_to_vlan_tci(struct rte_mbuf *mb,
- volatile union iavf_rx_flex_desc *rxdp,
- uint8_t rx_flags)
+ volatile union iavf_rx_flex_desc *rxdp)
 {
-   uint16_t vlan_tci = 0;
-
-   if (rx_flags & IAVF_RX_FLAGS_VLAN_TAG_LOC_L2TAG1 &&
-   rte_le_to_cpu_64(rxdp->wb.status_error0) &
-   (1 << IAVF_RX_FLEX_DESC_STATUS0_L2TAG1P_S))
-   vlan_tci = rte_le_to_cpu_16(rxdp->wb.l2tag1);
+   if (rte_le_to_cpu_64(rxdp->wb.status_error0) &
+   (1 << IAVF_RX_FLEX_DESC_STATUS0_L2TAG1P_S)) {
+   mb->ol_flags |= PKT_RX_VLAN | PKT_RX_VLAN_STRIPPED;
+   mb->vlan_tci =
+   rte_le_to_cpu_16(rxdp->wb.l2tag1);
+   } else {
+   mb->vlan_tci = 0;
+   }
 
 #ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC
-   if (rx_flags & IAVF_RX_FLAGS_VLAN_TAG_LOC_L2TAG2_2 &&
-   rte_le_to_cpu_16(rxdp->wb.status_error1) &
-   (1 << IAVF_RX_FLEX_DESC_STATUS1_L2TAG2P_S))
-   vlan_tci = rte_le_to_cpu_16(rxdp->wb.l2tag2_2nd);
-#endif
-
-   if (vlan_tci) {
-   mb->ol_flags |= PKT_RX_VLAN | PKT_RX_VLAN_STRIPPED;
-   mb->vlan_tci = vlan_tci;
+   if (rte_le_to_cpu_16(rxdp->wb.status_error1) &
+   (1 << IAVF_RX_FLEX_DESC_STATUS1_L2TAG2P_S)) {
+   mb->ol_flags |= PKT_RX_QINQ_STRIPPED | PKT_RX_QINQ |
+   PKT_RX_VLAN_STRIPPED | PKT_RX_VLAN;
+   mb->vlan_tci_outer = mb->vlan_tci;
+   mb->vlan_tci = rte_le_to_cpu_16(rxdp->wb.l2tag2_2nd);
+   PMD_RX_LOG(DEBUG, "Descriptor l2tag2_1: %u, l2tag2_2: %u",
+  rte_le_to_cpu_16(rxdp->wb.l2tag2_1st),
+  rte_le_to_cpu_16(rxdp->wb.l2tag2_2nd));
+   } else {
+   mb->vlan_tci_outer = 0;
}
+#endif
 }
 
 /* Translate the rx descriptor status and error fields to pkt flags */
@@ -1394,7 +1398,7 @@ iavf_recv_pkts_flex_rxd(void *rx_queue,
rxm->ol_flags = 0;
rxm->packet_type = ptype_tbl[IAVF_RX_FLEX_DESC_PTYPE_M &
rte_le_to_cpu_16(rxd.wb.ptype_flex_flags0)];
-   iavf_flex_rxd_to_vlan_tci(rxm, &rxd, rxq->rx_flags);
+   iavf_flex_rxd_to_vlan_tci(rxm, &rxd);
rxq->rxd_to_pkt_fields(rxq, rxm, &rxd);
pkt_flags = iavf_flex_rxd_error_to_pkt_flags(rx_stat_err0);
rxm->ol_flags |= pkt_flags;
@@ -1536,7 +1540,7 @@ iavf_recv_scattered_pkts_flex_rxd(void *rx_queue, struct 
rte_mbuf **rx_pkts,
first_seg->ol_flags = 0;
first_seg->packet_type = ptype_tbl[IAVF_RX_FLEX_DESC_PTYPE_M &
rte_le_to_cpu_16(rxd.wb.ptype_flex_flags0)];
-   iavf_flex_rxd_to_vlan_tci(first_seg, &rxd, rxq->rx_flags);
+   iavf_flex_rxd_to_vlan_tci(first_seg, &rxd);
rxq->rxd_to_pkt_fields(rxq, first_seg, &rxd);
pkt_flags = iavf_flex_rxd_error_to_pkt_flags(rx_stat_err0);
 
@@ -1774,7 +1778,7 @@ iavf_rx_scan_hw_ring_flex_rxd(struct iavf_rx_queue *rxq)
 
mb->packet_type = ptype_tbl[IAVF_RX_FLEX_DESC_PTYPE_M &
rte_le_to_cpu_16(rxdp[j].wb.ptype_flex_flags0)];
-   iavf_flex_rxd_to_vlan_tci(mb, &rxdp[j], rxq->rx_flags);
+   iavf_flex_rxd_to_vlan_tci(mb, &rxdp[j]);
rxq->rxd_to_pkt_fields(rxq, mb, &rxdp[j]);
stat_err0 = rte_le_to_cpu_16(rxdp[j].wb.status_error0);
pkt_flags = iavf_flex_rxd_error_to_pkt_flags(stat_err0);
@@ -2068,190 +2072,302 @@ iavf_xmit_cleanup(struct iavf_tx_queue *txq)
return 0;
 }
 
-/* Check if the context descriptor is needed for TX offloading */
+
+
+static inline void
+iavf_fill_ctx_desc_cmd_field(volatile uint64_t *field, struct rte_mbuf *m)
+{
+   uint64_t cmd = 0;
+
+   /* TSO enabled */
+   if (m->ol_flags & (PKT_TX_TCP_SEG | PKT_TX_UDP_SEG))
+   cmd = IAVF_TX_CTX_DESC_TSO << IAVF_TXD_DATA_QW1_CMD_SHIFT;
+
+   /* Time Sync - Currently not suppo

[dpdk-dev] [PATCH v11 3/7] net/iavf: add support for asynchronous virt channel messages

2021-10-26 Thread Radu Nicolau
Add support for asynchronous virtual channel messages, specifically for
inline IPsec messages.

Signed-off-by: Declan Doherty 
Signed-off-by: Abhijit Sinha 
Signed-off-by: Radu Nicolau 
Acked-by: Jingjing Wu 
---
 drivers/net/iavf/iavf.h   |  16 
 drivers/net/iavf/iavf_vchnl.c | 138 +-
 2 files changed, 101 insertions(+), 53 deletions(-)

diff --git a/drivers/net/iavf/iavf.h b/drivers/net/iavf/iavf.h
index 12f541f539..efc90f9072 100644
--- a/drivers/net/iavf/iavf.h
+++ b/drivers/net/iavf/iavf.h
@@ -193,6 +193,7 @@ struct iavf_info {
uint64_t supported_rxdid;
uint8_t *proto_xtr; /* proto xtr type for all queues */
volatile enum virtchnl_ops pend_cmd; /* pending command not finished */
+   rte_atomic32_t pend_cmd_count;
int cmd_retval; /* return value of the cmd response from PF */
uint8_t *aq_resp; /* buffer to store the adminq response from PF */
 
@@ -345,9 +346,24 @@ _atomic_set_cmd(struct iavf_info *vf, enum virtchnl_ops 
ops)
if (!ret)
PMD_DRV_LOG(ERR, "There is incomplete cmd %d", vf->pend_cmd);
 
+   rte_atomic32_set(&vf->pend_cmd_count, 1);
+
return !ret;
 }
 
+/* Check there is pending cmd in execution. If none, set new command. */
+static inline int
+_atomic_set_async_response_cmd(struct iavf_info *vf, enum virtchnl_ops ops)
+{
+   int ret = rte_atomic32_cmpset(&vf->pend_cmd, VIRTCHNL_OP_UNKNOWN, ops);
+
+   if (!ret)
+   PMD_DRV_LOG(ERR, "There is incomplete cmd %d", vf->pend_cmd);
+
+   rte_atomic32_set(&vf->pend_cmd_count, 2);
+
+   return !ret;
+}
 int iavf_check_api_version(struct iavf_adapter *adapter);
 int iavf_get_vf_resource(struct iavf_adapter *adapter);
 void iavf_handle_virtchnl_msg(struct rte_eth_dev *dev);
diff --git a/drivers/net/iavf/iavf_vchnl.c b/drivers/net/iavf/iavf_vchnl.c
index 0f4dd21d44..da4654957a 100644
--- a/drivers/net/iavf/iavf_vchnl.c
+++ b/drivers/net/iavf/iavf_vchnl.c
@@ -24,8 +24,8 @@
 #include "iavf.h"
 #include "iavf_rxtx.h"
 
-#define MAX_TRY_TIMES 200
-#define ASQ_DELAY_MS  10
+#define MAX_TRY_TIMES 2000
+#define ASQ_DELAY_MS  1
 
 static uint32_t
 iavf_convert_link_speed(enum virtchnl_link_speed virt_link_speed)
@@ -143,7 +143,8 @@ iavf_read_msg_from_pf(struct iavf_adapter *adapter, 
uint16_t buf_len,
 }
 
 static int
-iavf_execute_vf_cmd(struct iavf_adapter *adapter, struct iavf_cmd_info *args)
+iavf_execute_vf_cmd(struct iavf_adapter *adapter, struct iavf_cmd_info *args,
+   int async)
 {
struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(adapter);
@@ -155,8 +156,14 @@ iavf_execute_vf_cmd(struct iavf_adapter *adapter, struct 
iavf_cmd_info *args)
if (vf->vf_reset)
return -EIO;
 
-   if (_atomic_set_cmd(vf, args->ops))
-   return -1;
+
+   if (async) {
+   if (_atomic_set_async_response_cmd(vf, args->ops))
+   return -1;
+   } else {
+   if (_atomic_set_cmd(vf, args->ops))
+   return -1;
+   }
 
ret = iavf_aq_send_msg_to_pf(hw, args->ops, IAVF_SUCCESS,
args->in_args, args->in_args_size, NULL);
@@ -252,9 +259,11 @@ static void
 iavf_handle_pf_event_msg(struct rte_eth_dev *dev, uint8_t *msg,
uint16_t msglen)
 {
+   struct iavf_adapter *adapter =
+   IAVF_DEV_PRIVATE_TO_ADAPTER(dev->data->dev_private);
+   struct iavf_info *vf = &adapter->vf;
struct virtchnl_pf_event *pf_msg =
(struct virtchnl_pf_event *)msg;
-   struct iavf_info *vf = IAVF_DEV_PRIVATE_TO_VF(dev->data->dev_private);
 
if (msglen < sizeof(struct virtchnl_pf_event)) {
PMD_DRV_LOG(DEBUG, "Error event");
@@ -330,18 +339,40 @@ iavf_handle_virtchnl_msg(struct rte_eth_dev *dev)
case iavf_aqc_opc_send_msg_to_vf:
if (msg_opc == VIRTCHNL_OP_EVENT) {
iavf_handle_pf_event_msg(dev, info.msg_buf,
-   info.msg_len);
+   info.msg_len);
} else {
+   /* check for inline IPsec events */
+   struct inline_ipsec_msg *imsg =
+   (struct inline_ipsec_msg *)info.msg_buf;
+   struct rte_eth_event_ipsec_desc desc;
+   if (msg_opc == VIRTCHNL_OP_INLINE_IPSEC_CRYPTO
+   && imsg->ipsec_opcode ==
+   INLINE_IPSEC_OP_EVENT) {
+   struct virtchnl_ipsec_event *ev =
+   imsg->ipsec_data.event;
+   desc.subtype =
+ 

[dpdk-dev] [PATCH v11 4/7] net/iavf: add iAVF IPsec inline crypto support

2021-10-26 Thread Radu Nicolau
Add support for inline crypto for IPsec, for ESP transport and
tunnel over IPv4 and IPv6, as well as supporting the offload for
ESP over UDP, and inconjunction with TSO for UDP and TCP flows.
Implement support for rte_security packet metadata

Add definition for IPsec descriptors, extend support for offload
in data and context descriptor to support

Add support to virtual channel mailbox for IPsec Crypto request
operations. IPsec Crypto requests receive an initial acknowledgment
from phsyical function driver of receipt of request and then an
asynchronous response with success/failure of request including any
response data.

Add enhanced descriptor debugging

Refactor of scalar tx burst function to support integration of offload

Signed-off-by: Declan Doherty 
Signed-off-by: Abhijit Sinha 
Signed-off-by: Radu Nicolau 
Reviewed-by: Jingjing Wu 
---
 drivers/net/iavf/iavf.h   |   10 +
 drivers/net/iavf/iavf_ethdev.c|   41 +-
 drivers/net/iavf/iavf_generic_flow.c  |   15 +
 drivers/net/iavf/iavf_generic_flow.h  |2 +
 drivers/net/iavf/iavf_ipsec_crypto.c  | 1894 +
 drivers/net/iavf/iavf_ipsec_crypto.h  |  160 ++
 .../net/iavf/iavf_ipsec_crypto_capabilities.h |  383 
 drivers/net/iavf/iavf_rxtx.c  |  202 +-
 drivers/net/iavf/iavf_rxtx.h  |  115 +-
 drivers/net/iavf/iavf_vchnl.c |   29 +
 drivers/net/iavf/meson.build  |3 +-
 drivers/net/iavf/rte_pmd_iavf.h   |1 +
 drivers/net/iavf/version.map  |3 +
 13 files changed, 2826 insertions(+), 32 deletions(-)
 create mode 100644 drivers/net/iavf/iavf_ipsec_crypto.c
 create mode 100644 drivers/net/iavf/iavf_ipsec_crypto.h
 create mode 100644 drivers/net/iavf/iavf_ipsec_crypto_capabilities.h

diff --git a/drivers/net/iavf/iavf.h b/drivers/net/iavf/iavf.h
index efc90f9072..6df31a649e 100644
--- a/drivers/net/iavf/iavf.h
+++ b/drivers/net/iavf/iavf.h
@@ -221,6 +221,7 @@ struct iavf_info {
rte_spinlock_t flow_ops_lock;
struct iavf_parser_list rss_parser_list;
struct iavf_parser_list dist_parser_list;
+   struct iavf_parser_list ipsec_crypto_parser_list;
 
struct iavf_fdir_info fdir; /* flow director info */
/* indicate large VF support enabled or not */
@@ -245,6 +246,7 @@ enum iavf_proto_xtr_type {
IAVF_PROTO_XTR_IPV6_FLOW,
IAVF_PROTO_XTR_TCP,
IAVF_PROTO_XTR_IP_OFFSET,
+   IAVF_PROTO_XTR_IPSEC_CRYPTO_SAID,
IAVF_PROTO_XTR_MAX,
 };
 
@@ -256,11 +258,14 @@ struct iavf_devargs {
uint8_t proto_xtr[IAVF_MAX_QUEUE_NUM];
 };
 
+struct iavf_security_ctx;
+
 /* Structure to store private data for each VF instance. */
 struct iavf_adapter {
struct iavf_hw hw;
struct rte_eth_dev_data *dev_data;
struct iavf_info vf;
+   struct iavf_security_ctx *security_ctx;
 
bool rx_bulk_alloc_allowed;
/* For vector PMD */
@@ -279,6 +284,8 @@ struct iavf_adapter {
(&((struct iavf_adapter *)adapter)->vf)
 #define IAVF_DEV_PRIVATE_TO_HW(adapter) \
(&((struct iavf_adapter *)adapter)->hw)
+#define IAVF_DEV_PRIVATE_TO_IAVF_SECURITY_CTX(adapter) \
+   (((struct iavf_adapter *)adapter)->security_ctx)
 
 /* IAVF_VSI_TO */
 #define IAVF_VSI_TO_HW(vsi) \
@@ -421,5 +428,8 @@ int iavf_set_q_tc_map(struct rte_eth_dev *dev,
uint16_t size);
 void iavf_tm_conf_init(struct rte_eth_dev *dev);
 void iavf_tm_conf_uninit(struct rte_eth_dev *dev);
+int iavf_ipsec_crypto_request(struct iavf_adapter *adapter,
+   uint8_t *msg, size_t msg_len,
+   uint8_t *resp_msg, size_t resp_msg_len);
 extern const struct rte_tm_ops iavf_tm_ops;
 #endif /* _IAVF_ETHDEV_H_ */
diff --git a/drivers/net/iavf/iavf_ethdev.c b/drivers/net/iavf/iavf_ethdev.c
index b2b413c247..9ab42b6452 100644
--- a/drivers/net/iavf/iavf_ethdev.c
+++ b/drivers/net/iavf/iavf_ethdev.c
@@ -30,6 +30,7 @@
 #include "iavf_rxtx.h"
 #include "iavf_generic_flow.h"
 #include "rte_pmd_iavf.h"
+#include "iavf_ipsec_crypto.h"
 
 /* devargs */
 #define IAVF_PROTO_XTR_ARG "proto_xtr"
@@ -71,6 +72,11 @@ static struct iavf_proto_xtr_ol iavf_proto_xtr_params[] = {
[IAVF_PROTO_XTR_IP_OFFSET] = {
.param = { .name = "intel_pmd_dynflag_proto_xtr_ip_offset" },
.ol_flag = &rte_pmd_ifd_dynflag_proto_xtr_ip_offset_mask },
+   [IAVF_PROTO_XTR_IPSEC_CRYPTO_SAID] = {
+   .param = {
+   .name = "intel_pmd_dynflag_proto_xtr_ipsec_crypto_said" },
+   .ol_flag =
+   &rte_pmd_ifd_dynflag_proto_xtr_ipsec_crypto_said_mask },
 };
 
 static int iavf_dev_configure(struct rte_eth_dev *dev);
@@ -924,6 +930,9 @@ iavf_dev_stop(struct rte_eth_dev *dev)
iavf_add_del_mc_addr_list(adapter, vf->mc_addrs, vf->mc_addrs_num,
  false);
 
+   /* free iAVF security device contex

[dpdk-dev] [PATCH v11 5/7] net/iavf: add xstats support for inline IPsec crypto

2021-10-26 Thread Radu Nicolau
Add per queue counters for maintaining statistics for inline IPsec
crypto offload, which can be retrieved through the
rte_security_session_stats_get() with more detailed errors through the
rte_ethdev xstats.

Signed-off-by: Declan Doherty 
Signed-off-by: Radu Nicolau 
Acked-by: Jingjing Wu 
---
 drivers/net/iavf/iavf.h| 21 -
 drivers/net/iavf/iavf_ethdev.c | 84 --
 drivers/net/iavf/iavf_rxtx.h   | 12 -
 3 files changed, 89 insertions(+), 28 deletions(-)

diff --git a/drivers/net/iavf/iavf.h b/drivers/net/iavf/iavf.h
index 6df31a649e..f314373ab0 100644
--- a/drivers/net/iavf/iavf.h
+++ b/drivers/net/iavf/iavf.h
@@ -96,6 +96,25 @@ struct iavf_adapter;
 struct iavf_rx_queue;
 struct iavf_tx_queue;
 
+
+struct iavf_ipsec_crypto_stats {
+   uint64_t icount;
+   uint64_t ibytes;
+   struct {
+   uint64_t count;
+   uint64_t sad_miss;
+   uint64_t not_processed;
+   uint64_t icv_check;
+   uint64_t ipsec_length;
+   uint64_t misc;
+   } ierrors;
+};
+
+struct iavf_eth_xstats {
+   struct virtchnl_eth_stats eth_stats;
+   struct iavf_ipsec_crypto_stats ips_stats;
+};
+
 /* Structure that defines a VSI, associated with a adapter. */
 struct iavf_vsi {
struct iavf_adapter *adapter; /* Backreference to associated adapter */
@@ -105,7 +124,7 @@ struct iavf_vsi {
uint16_t max_macaddrs;   /* Maximum number of MAC addresses */
uint16_t base_vector;
uint16_t msix_intr;  /* The MSIX interrupt binds to VSI */
-   struct virtchnl_eth_stats eth_stats_offset;
+   struct iavf_eth_xstats eth_stats_offset;
 };
 
 struct rte_flow;
diff --git a/drivers/net/iavf/iavf_ethdev.c b/drivers/net/iavf/iavf_ethdev.c
index 9ab42b6452..e1e6f49dec 100644
--- a/drivers/net/iavf/iavf_ethdev.c
+++ b/drivers/net/iavf/iavf_ethdev.c
@@ -90,6 +90,7 @@ static const uint32_t *iavf_dev_supported_ptypes_get(struct 
rte_eth_dev *dev);
 static int iavf_dev_stats_get(struct rte_eth_dev *dev,
 struct rte_eth_stats *stats);
 static int iavf_dev_stats_reset(struct rte_eth_dev *dev);
+static int iavf_dev_xstats_reset(struct rte_eth_dev *dev);
 static int iavf_dev_xstats_get(struct rte_eth_dev *dev,
 struct rte_eth_xstat *xstats, unsigned int n);
 static int iavf_dev_xstats_get_names(struct rte_eth_dev *dev,
@@ -145,21 +146,37 @@ struct rte_iavf_xstats_name_off {
unsigned int offset;
 };
 
+#define _OFF_OF(a) offsetof(struct iavf_eth_xstats, a)
 static const struct rte_iavf_xstats_name_off rte_iavf_stats_strings[] = {
-   {"rx_bytes", offsetof(struct iavf_eth_stats, rx_bytes)},
-   {"rx_unicast_packets", offsetof(struct iavf_eth_stats, rx_unicast)},
-   {"rx_multicast_packets", offsetof(struct iavf_eth_stats, rx_multicast)},
-   {"rx_broadcast_packets", offsetof(struct iavf_eth_stats, rx_broadcast)},
-   {"rx_dropped_packets", offsetof(struct iavf_eth_stats, rx_discards)},
+   {"rx_bytes", _OFF_OF(eth_stats.rx_bytes)},
+   {"rx_unicast_packets", _OFF_OF(eth_stats.rx_unicast)},
+   {"rx_multicast_packets", _OFF_OF(eth_stats.rx_multicast)},
+   {"rx_broadcast_packets", _OFF_OF(eth_stats.rx_broadcast)},
+   {"rx_dropped_packets", _OFF_OF(eth_stats.rx_discards)},
{"rx_unknown_protocol_packets", offsetof(struct iavf_eth_stats,
rx_unknown_protocol)},
-   {"tx_bytes", offsetof(struct iavf_eth_stats, tx_bytes)},
-   {"tx_unicast_packets", offsetof(struct iavf_eth_stats, tx_unicast)},
-   {"tx_multicast_packets", offsetof(struct iavf_eth_stats, tx_multicast)},
-   {"tx_broadcast_packets", offsetof(struct iavf_eth_stats, tx_broadcast)},
-   {"tx_dropped_packets", offsetof(struct iavf_eth_stats, tx_discards)},
-   {"tx_error_packets", offsetof(struct iavf_eth_stats, tx_errors)},
+   {"tx_bytes", _OFF_OF(eth_stats.tx_bytes)},
+   {"tx_unicast_packets", _OFF_OF(eth_stats.tx_unicast)},
+   {"tx_multicast_packets", _OFF_OF(eth_stats.tx_multicast)},
+   {"tx_broadcast_packets", _OFF_OF(eth_stats.tx_broadcast)},
+   {"tx_dropped_packets", _OFF_OF(eth_stats.tx_discards)},
+   {"tx_error_packets", _OFF_OF(eth_stats.tx_errors)},
+
+   {"inline_ipsec_crypto_ipackets", _OFF_OF(ips_stats.icount)},
+   {"inline_ipsec_crypto_ibytes", _OFF_OF(ips_stats.ibytes)},
+   {"inline_ipsec_crypto_ierrors", _OFF_OF(ips_stats.ierrors.count)},
+   {"inline_ipsec_crypto_ierrors_sad_lookup",
+   _OFF_OF(ips_stats.ierrors.sad_miss)},
+   {"inline_ipsec_crypto_ierrors_not_processed",
+   _OFF_OF(ips_stats.ierrors.not_processed)},
+   {"inline_ipsec_crypto_ierrors_icv_fail",
+   _OFF_OF(ips_stats.ierrors.icv_check)},
+   {"inline_ipsec_crypto_ierrors_length",
+   _OFF_OF(ips_stats.ierrors.ipsec_length)},
+   {"inline_ipsec_crypto

[dpdk-dev] [PATCH v11 6/7] net/iavf: add watchdog for VFLR

2021-10-26 Thread Radu Nicolau
Add watchdog to iAVF PMD which support monitoring the VFLR register. If
the device is not already in reset then if a VF reset in progress is
detected then notfiy user through callback and set into reset state.
If the device is already in reset then poll for completion of reset.

The watchdog is disabled by default, to enable it set
IAVF_DEV_WATCHDOG_PERIOD to a non zero value (microseconds)

Signed-off-by: Declan Doherty 
Signed-off-by: Radu Nicolau 
Acked-by: Jingjing Wu 
---
 drivers/net/iavf/iavf.h|  5 ++
 drivers/net/iavf/iavf_ethdev.c | 94 ++
 2 files changed, 99 insertions(+)

diff --git a/drivers/net/iavf/iavf.h b/drivers/net/iavf/iavf.h
index f314373ab0..40c8045de1 100644
--- a/drivers/net/iavf/iavf.h
+++ b/drivers/net/iavf/iavf.h
@@ -31,6 +31,8 @@
 
 #define IAVF_NUM_MACADDR_MAX  64
 
+#define IAVF_DEV_WATCHDOG_PERIOD 0
+
 #define IAVF_DEFAULT_RX_PTHRESH  8
 #define IAVF_DEFAULT_RX_HTHRESH  8
 #define IAVF_DEFAULT_RX_WTHRESH  0
@@ -216,6 +218,9 @@ struct iavf_info {
int cmd_retval; /* return value of the cmd response from PF */
uint8_t *aq_resp; /* buffer to store the adminq response from PF */
 
+   /** iAVF watchdog enable */
+   bool watchdog_enabled;
+
/* Event from pf */
bool dev_closed;
bool link_up;
diff --git a/drivers/net/iavf/iavf_ethdev.c b/drivers/net/iavf/iavf_ethdev.c
index e1e6f49dec..e2897441aa 100644
--- a/drivers/net/iavf/iavf_ethdev.c
+++ b/drivers/net/iavf/iavf_ethdev.c
@@ -25,6 +25,7 @@
 #include 
 #include 
 #include 
+#include 
 
 #include "iavf.h"
 #include "iavf_rxtx.h"
@@ -240,6 +241,91 @@ iavf_tm_ops_get(struct rte_eth_dev *dev __rte_unused,
return 0;
 }
 
+__rte_unused
+static int
+iavf_vfr_inprogress(struct iavf_hw *hw)
+{
+   int inprogress = 0;
+
+   if ((IAVF_READ_REG(hw, IAVF_VFGEN_RSTAT) &
+   IAVF_VFGEN_RSTAT_VFR_STATE_MASK) ==
+   VIRTCHNL_VFR_INPROGRESS)
+   inprogress = 1;
+
+   if (inprogress)
+   PMD_DRV_LOG(INFO, "Watchdog detected VFR in progress");
+
+   return inprogress;
+}
+
+__rte_unused
+static void
+iavf_dev_watchdog(void *cb_arg)
+{
+   struct iavf_adapter *adapter = cb_arg;
+   struct iavf_hw *hw = IAVF_DEV_PRIVATE_TO_HW(adapter);
+   int vfr_inprogress = 0, rc = 0;
+
+   /* check if watchdog has been disabled since last call */
+   if (!adapter->vf.watchdog_enabled)
+   return;
+
+   /* If in reset then poll vfr_inprogress register for completion */
+   if (adapter->vf.vf_reset) {
+   vfr_inprogress = iavf_vfr_inprogress(hw);
+
+   if (!vfr_inprogress) {
+   PMD_DRV_LOG(INFO, "VF \"%s\" reset has completed",
+   adapter->vf.eth_dev->data->name);
+   adapter->vf.vf_reset = false;
+   }
+   /* If not in reset then poll vfr_inprogress register for VFLR event */
+   } else {
+   vfr_inprogress = iavf_vfr_inprogress(hw);
+
+   if (vfr_inprogress) {
+   PMD_DRV_LOG(INFO,
+   "VF \"%s\" reset event detected by watchdog",
+   adapter->vf.eth_dev->data->name);
+
+   /* enter reset state with VFLR event */
+   adapter->vf.vf_reset = true;
+
+   rte_eth_dev_callback_process(adapter->vf.eth_dev,
+   RTE_ETH_EVENT_INTR_RESET, NULL);
+   }
+   }
+
+   /* re-alarm watchdog */
+   rc = rte_eal_alarm_set(IAVF_DEV_WATCHDOG_PERIOD,
+   &iavf_dev_watchdog, cb_arg);
+
+   if (rc)
+   PMD_DRV_LOG(ERR, "Failed \"%s\" to reset device watchdog alarm",
+   adapter->vf.eth_dev->data->name);
+}
+
+static void
+iavf_dev_watchdog_enable(struct iavf_adapter *adapter __rte_unused)
+{
+#if (IAVF_DEV_WATCHDOG_PERIOD > 0)
+   PMD_DRV_LOG(INFO, "Enabling device watchdog");
+   adapter->vf.watchdog_enabled = true;
+   if (rte_eal_alarm_set(IAVF_DEV_WATCHDOG_PERIOD,
+   &iavf_dev_watchdog, (void *)adapter))
+   PMD_DRV_LOG(ERR, "Failed to enabled device watchdog");
+#endif
+}
+
+static void
+iavf_dev_watchdog_disable(struct iavf_adapter *adapter __rte_unused)
+{
+#if (IAVF_DEV_WATCHDOG_PERIOD > 0)
+   PMD_DRV_LOG(INFO, "Disabling device watchdog");
+   adapter->vf.watchdog_enabled = false;
+#endif
+}
+
 static int
 iavf_set_mc_addr_list(struct rte_eth_dev *dev,
struct rte_ether_addr *mc_addrs,
@@ -2466,6 +2552,11 @@ iavf_dev_init(struct rte_eth_dev *eth_dev)
 
iavf_default_rss_disable(adapter);
 
+
+   /* Start device watchdog */
+   iavf_dev_watchdog_enable(adapter);
+
+
return 0;
 
 flow_init_err:
@@ -2549,6 +2640,9 @@ iavf_dev_close(struct rte_eth_dev *dev)
if (vf->vf_reset && !rte_pci_

[dpdk-dev] [PATCH v11 7/7] net/iavf: update doc with inline crypto support

2021-10-26 Thread Radu Nicolau
Update the PMD doc, feature matrix and release notes with the
new inline crypto feature.

Signed-off-by: Radu Nicolau 
---
 doc/guides/nics/features/iavf.ini  |  2 ++
 doc/guides/nics/intel_vf.rst   | 10 ++
 doc/guides/rel_notes/release_21_11.rst |  1 +
 3 files changed, 13 insertions(+)

diff --git a/doc/guides/nics/features/iavf.ini 
b/doc/guides/nics/features/iavf.ini
index dd3519e1e2..01f514239e 100644
--- a/doc/guides/nics/features/iavf.ini
+++ b/doc/guides/nics/features/iavf.ini
@@ -27,6 +27,7 @@ L4 checksum offload  = P
 Packet type parsing  = Y
 Rx descriptor status = Y
 Tx descriptor status = Y
+Inline crypto= Y
 Basic stats  = Y
 Multiprocess aware   = Y
 FreeBSD  = Y
@@ -65,3 +66,4 @@ mark = Y
 passthru = Y
 queue= Y
 rss  = Y
+security = Y
diff --git a/doc/guides/nics/intel_vf.rst b/doc/guides/nics/intel_vf.rst
index a1e236ad75..fd235e1463 100644
--- a/doc/guides/nics/intel_vf.rst
+++ b/doc/guides/nics/intel_vf.rst
@@ -633,3 +633,13 @@ Windows Support
 
 *   To load NetUIO driver, follow the steps mentioned in `dpdk-kmods repository
 `_.
+
+
+Inline IPsec Support
+
+
+*   IAVF PMD supports inline crypto processing depending on the underlying
+hardware crypto capabilities. IPsec Security Gateway Sample Application
+supports inline IPsec processing for IAVF PMD. For more details see the
+IPsec Security Gateway Sample Application and Security library
+documentation.
diff --git a/doc/guides/rel_notes/release_21_11.rst 
b/doc/guides/rel_notes/release_21_11.rst
index b327c2bfca..6c0cb55f17 100644
--- a/doc/guides/rel_notes/release_21_11.rst
+++ b/doc/guides/rel_notes/release_21_11.rst
@@ -152,6 +152,7 @@ New Features
   * Added Intel iavf support on Windows.
   * Added IPv4 and L4 (TCP/UDP/SCTP) checksum hash support in RSS flow.
   * Added PPPoL2TPv2oUDP RSS hash based on inner IP address and TCP/UDP port.
+  * Added Intel iavf inline crypto support.
 
 * **Updated Intel ice driver.**
 
-- 
2.25.1



Re: [dpdk-dev] [PATCH] vhost: fix async DMA map

2021-10-26 Thread Burakov, Anatoly

On 26-Oct-21 11:27 AM, Ding, Xuan wrote:

Hi,


-Original Message-
From: Maxime Coquelin 
Sent: Tuesday, October 26, 2021 5:49 PM
To: Ding, Xuan ; dev@dpdk.org;
david.march...@redhat.com; Xia, Chenbo 
Cc: Burakov, Anatoly 
Subject: Re: [PATCH] vhost: fix async DMA map



On 10/26/21 10:49, Ding, Xuan wrote:

Hi Maxime,


-Original Message-
From: Maxime Coquelin 
Sent: Tuesday, October 26, 2021 2:53 PM
To: Ding, Xuan ; dev@dpdk.org;
david.march...@redhat.com; Xia, Chenbo 
Cc: Burakov, Anatoly 
Subject: Re: [PATCH] vhost: fix async DMA map



On 10/26/21 04:07, Ding, Xuan wrote:

Hi Maxime,


-Original Message-
From: Maxime Coquelin 
Sent: Tuesday, October 26, 2021 4:47 AM
To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
; Ding, Xuan 
Subject: Re: [PATCH] vhost: fix async DMA map

Hi Xuan,

On 10/25/21 22:33, Maxime Coquelin wrote:

This patch fixes possible NULL-pointer dereferencing
reported by Coverity and also fixes NUMA reallocation
of the async DMA map.

Fixes: 7c61fa08b716 ("vhost: enable IOMMU for async vhost")

Coverity issue: 373655

Signed-off-by: Maxime Coquelin 
---
 lib/vhost/vhost_user.c | 45 +++---
 1 file changed, 20 insertions(+), 25 deletions(-)



I posted this patch to fix the issue reported by Coverity and also other
issue on NUMA realloc that I found at the same time. But I wonder
whether all this async_map_status is needed.


Thanks for your fix! I can help to review and test the patch later.

I add the async_map_status in v2 for compatibility. Some DMA device,
like DSA, may use kernel idxd driver only. If there is no device bound to
DPDK vfio and kernel vfio module is modprobed to ensure

rte_vfio_is_enabled() is true,

we will unavoidably do DMA map/unmap and it will fail.

Therefore, the dma_map_status here is used to filter this situation by

preventing

unnecessary DMA unmap.


Ok, then I think we can just remove the async DMA map.



Indeed, if the only place where we DMA map is in
vhost_user_mmap_region(). If it fails, the error is propagated, the mem
table are freed and NACK is replied to the master. IOW, the device will
be in an unusable state.


I agree with you, this is the place I consider right to do DMA map
because we also do SW mapping here, any suggestions?


No suggestion, I was just explaining that at the only place where
DMA map were done, mapping errors were properly handled and

propagated.


What about just setting async_copy to false, and allow switching to sync

path.






Removing the async DMA map will simplify a lot the code, do you agree

to

remove it or there is something I missed?


See above. Indeed, it adds a lot of code. But we can't know the driver for
each device in vhost lib, or we can only restrict the user to bind some

devices

to DPDK vfio if async logic needed.


I would think we don't care if DMA unmap fails, we can just do the same
as what you do for DMA map, i.e. just ignore the error.


Get your idea, we can do the same as DMA map, and in this way

dma_map_status flag can be removed.




Thanks to this discussion, I have now more concerns on how it works. I
think we have a problem here in case of DMA device hotplug, that device
could miss the necessary map entries from Vhost if no VFIO devices were
attached at VHST_USER_SET_MEM_TABLE time. How would you handle

that

case?


DMA device is uncore, so I don't see the  hotplug issue here.


I'm not sure what 'uncore' is, I suppose you mean your device cannot be
physically added/removed to the system.


Yes, we are at the same understanding.



I was not clear enough in my question. I meant that for example, the
application is started and the Vhost port is created. Then, the DMA
device is bound to VFIO, and probed by the DPDK application. Finally,
the application register the DMA device as an async channel in Vhost.

I think it will not work as the SET_MEM_TABLE will already have
happened, so the guest memory won't be mapped in the VFIO container.


Assuming the DMA device supports hotplug, this situation was not actually took 
into consideration, and maybe the handling should be added in app, with an 
event callback.



Do you have the same understanding?


I will have another patch containing compatibility with sync path, and

async_map_status flag will be removed.

Hope to get your insights.


What do you mean by compatibility with sync path?


I mean whatever DMA map succeeds or fails, we return 0 so as not to prevent
SET_EM_TABLE.



Thanks for taking care of the async map status removal.


Found a tricky point for removing dma_map_status flag, currently DMA unmap API 
will directly access the vfio_config,
And if without dma_map_status, we cannot get rte_errno at this moment.


This is a bug in VFIO API. We always assume VFIO type is valid, but it 
is only valid whenever 1) VFIO support is detected at startup, and 2) we 
assign at least one device to VFIO. If VFIO was detected but no devices 
were assigned, the 

[dpdk-dev] [PATCH] test: fix forwarding packets through not-ready port

2021-10-26 Thread Konstantin Ananyev
(bitratestats_autotest|latencystats_autotest|pdump_autotest) tests
generate a log of error messages like that:

test_packet_forward() line 104: Error sending packet to port 0
Send pkts Failed

These tests use of app/test/sample_packet_forward.* code.
This code creates a portid from a ring, but doesn't properly
configure/start it.
The fix adds code to configure/start given port before usage.

Fixes: 7a0935239b9e ("ethdev: make fast-path functions to use new flat array")
Fixes: a52966cd48fd ("test: add helpers using ring PMD Rx/Tx")
Cc: sta...@dpdk.org

Reported-by: David Marchand 
Signed-off-by: Konstantin Ananyev 
---
 app/test/sample_packet_forward.c | 29 +
 app/test/sample_packet_forward.h |  3 +++
 app/test/test_bitratestats.c | 12 +++-
 app/test/test_latencystats.c | 12 +++-
 app/test/test_pdump.c| 12 ++--
 5 files changed, 64 insertions(+), 4 deletions(-)

diff --git a/app/test/sample_packet_forward.c b/app/test/sample_packet_forward.c
index 61384b3d9b..aa897274d8 100644
--- a/app/test/sample_packet_forward.c
+++ b/app/test/sample_packet_forward.c
@@ -15,6 +15,35 @@
 
 #include "sample_packet_forward.h"
 
+/*
+ * heper function: configure and start test device
+ */
+int
+test_dev_start(uint16_t port, struct rte_mempool *mp)
+{
+   int32_t rc;
+   struct rte_eth_conf pconf;
+
+   memset(&pconf, 0, sizeof(pconf));
+
+   rc =  rte_eth_dev_configure(port, NUM_QUEUES, NUM_QUEUES, &pconf);
+   if (rc != 0)
+   return rc;
+
+   rc = rte_eth_rx_queue_setup(port, 0, RING_SIZE, SOCKET_ID_ANY,
+   NULL, mp);
+   if (rc != 0)
+   return rc;
+
+   rc = rte_eth_tx_queue_setup(port, 0, RING_SIZE, SOCKET_ID_ANY,
+   NULL);
+   if (rc != 0)
+   return rc;
+
+   rc = rte_eth_dev_start(port);
+   return rc;
+}
+
 /* Sample test to create virtual rings and tx,rx portid from rings */
 int
 test_ring_setup(struct rte_ring **ring, uint16_t *portid)
diff --git a/app/test/sample_packet_forward.h b/app/test/sample_packet_forward.h
index 6789217de3..af0b1d9924 100644
--- a/app/test/sample_packet_forward.h
+++ b/app/test/sample_packet_forward.h
@@ -21,6 +21,9 @@ struct rte_ring;
 /* Sample test to create virtual rings and tx,rx portid from rings */
 int test_ring_setup(struct rte_ring **ring, uint16_t *portid);
 
+/* configure and start device created by test_ring_setup */
+int test_dev_start(uint16_t port, struct rte_mempool *mp);
+
 /* Sample test to free the virtual rings */
 void test_ring_free(struct rte_ring *rxtx);
 
diff --git a/app/test/test_bitratestats.c b/app/test/test_bitratestats.c
index f4a92c9be6..1ff540f4c4 100644
--- a/app/test/test_bitratestats.c
+++ b/app/test/test_bitratestats.c
@@ -11,6 +11,7 @@
 #include 
 #include 
 #include 
+#include 
 
 #include "sample_packet_forward.h"
 #include "test.h"
@@ -159,12 +160,21 @@ test_bit_packet_forward(void)
printf("allocate mbuf pool Failed\n");
return TEST_FAILED;
}
+   ret = test_dev_start(portid, mp);
+   if (ret < 0) {
+   printf("test_dev_start(%hu, %p) failed, error code: %d\n",
+   portid, mp, ret);
+   return TEST_FAILED;
+   }
+
ret = test_packet_forward(pbuf, portid, QUEUE_ID);
if (ret < 0)
printf("send pkts Failed\n");
+
+   rte_eth_dev_stop(portid);
test_put_mbuf_to_pool(mp, pbuf);
 
-   return TEST_SUCCESS;
+   return (ret >= 0) ? TEST_SUCCESS : TEST_FAILED;
 }
 
 static int
diff --git a/app/test/test_latencystats.c b/app/test/test_latencystats.c
index 724acbc315..db06c7d5c7 100644
--- a/app/test/test_latencystats.c
+++ b/app/test/test_latencystats.c
@@ -6,6 +6,7 @@
 #include 
 #include 
 
+#include 
 #include 
 #include "rte_lcore.h"
 #include "rte_metrics.h"
@@ -158,12 +159,21 @@ static int test_latency_packet_forward(void)
printf("allocate mbuf pool Failed\n");
return TEST_FAILED;
}
+   ret = test_dev_start(portid, mp);
+   if (ret < 0) {
+   printf("test_dev_start(%hu, %p) failed, error code: %d\n",
+   portid, mp, ret);
+   return TEST_FAILED;
+   }
+
ret = test_packet_forward(pbuf, portid, QUEUE_ID);
if (ret < 0)
printf("send pkts Failed\n");
+
+   rte_eth_dev_stop(portid);
test_put_mbuf_to_pool(mp, pbuf);
 
-   return TEST_SUCCESS;
+   return (ret >= 0) ? TEST_SUCCESS : TEST_FAILED;
 }
 
 static struct
diff --git a/app/test/test_pdump.c b/app/test/test_pdump.c
index b49fcfb3f1..ea03056b47 100644
--- a/app/test/test_pdump.c
+++ b/app/test/test_pdump.c
@@ -147,11 +147,19 @@ send_pkts(void *empty)
ret = test_get_mbuf_from_pool(&mp, pbuf, poolname);
if (ret < 0)
printf("get_mbuf_from_pool failed\n");
-   do {
+
+   ret = test_dev_start(portid, mp);
+ 

Re: [dpdk-dev] [PATCH v4 0/6] Flow entites behavior on port restart

2021-10-26 Thread Ferruh Yigit

On 10/21/2021 7:34 AM, Dmitry Kozlyuk wrote:

It is unspecified whether flow rules and indirect actions are kept
when a port is stopped, possibly reconfigured, and started again.
Vendors approach the topic differently, e.g. mlx5 and i40e PMD
disagree in whether flow rules can be kept, and mlx5 PMD would keep
indirect actions. In the end, applications are greatly affected
by whatever contract there is and need to know it.

Applications may wish to restart the port to reconfigure it,
e.g. switch offloads or even modify queues.
Keeping rte_flow entities enables application improvements:
1. Since keeping the rules across restart comes with the ability
to create rules before the device is started. This allows
to have all the rules created at the moment of start,
so that there is no time frame when traffic is coming already,
but the rules are not yet created (restored).
2. When a rule or an indirect action has some associated state,
such as a counter, application saves the need to keep
additional state in order to cope with information loss
if such an entity would be destroyed.

It is proposed to advertise capabilities of keeping flow rules
and indirect actions (as a special case of shared object)
using a combination of ethdev info and rte_flow calls.
Then a bug is fixed in mlx5 PMD that prevented indirect RSS action
from being kept, and the driver starts advertising the new capability.

Prior discussions:
1) http://inbox.dpdk.org/dev/20210727073121.895620-1-dkozl...@nvidia.com/
2) http://inbox.dpdk.org/dev/20210901085516.3647814-1-dkozl...@nvidia.com/

v4:  1. Fix rebase conflicts (CI).
  2. State rule behavior when a port is not started or stopped (Ori).
  3. Improve wording on rule features, add examples (Andrew).
  4. State that rules/actions that cannot be kept while other can be
 must be destroyed by the application (Andrew/Ori).
  5. Add rationale to the cover letter (Andrew).

Dmitry Kozlyuk (6):
   ethdev: add capability to keep flow rules on restart
   ethdev: add capability to keep shared objects on restart
   net: advertise no support for keeping flow rules
   net/mlx5: discover max flow priority using DevX
   net/mlx5: create drop queue using DevX
   net/mlx5: preserve indirect actions on restart



Requesting review from PMD maintainers.

Since this patch tries to define behavior on keeping/flushing flow rules
after port stop/start/configure, better to get more feedback from various
vendors, please review/comment on patch so that we can get it for -rc2.

If there is no comment the patch can go in as it is for -rc2.

Thanks,
ferruh




Re: [dpdk-dev] [PATCH v4 0/8] port ioatfwd app to dmadev

2021-10-26 Thread Kevin Laatz

On 26/10/2021 01:56, fengchengwen wrote:

Hi Kevin,

We test whole patch set and found it should add one judgement:
the ring_size should be less than or equal to MBUF_RING_SIZE.
If ring_size greater than MBUF_RING_SIZE, the tracking DMA bufs
may be overwrited when the DMA copy is not in time.

Thanks.



Thanks for testing, Chengwen. I'll include this check in the v5.

/Kevin



[dpdk-dev] [PATCH] pdump: fix uninit not freeing statistics memzone

2021-10-26 Thread Konstantin Ananyev
rte_pdump_init() always allocates new memzone for pdump_stats.
Though rte_pdump_uninit() never frees it.
So the following combination will always fail:
rte_pdump_init(); rte_pdump_uninit(); rte_pdump_init();
The issue was caught by pdump_autotest UT.
While first test run successful, any consecutive runs
of this test-case will fail.
Fix the issue by calling rte_memzone_free() for statistics memzone.

Fixes: 10f726efe26c ("pdump: support pcapng and filtering")

Signed-off-by: Konstantin Ananyev 
---
 lib/pdump/rte_pdump.c | 7 +++
 1 file changed, 7 insertions(+)

diff --git a/lib/pdump/rte_pdump.c b/lib/pdump/rte_pdump.c
index 71602685d5..a708935861 100644
--- a/lib/pdump/rte_pdump.c
+++ b/lib/pdump/rte_pdump.c
@@ -74,6 +74,7 @@ static const char MZ_RTE_PDUMP_STATS[] = "rte_pdump_stats";
 static struct {
struct rte_pdump_stats rx[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT];
struct rte_pdump_stats tx[RTE_MAX_ETHPORTS][RTE_MAX_QUEUES_PER_PORT];
+   const struct rte_memzone *mz;
 } *pdump_stats;
 
 /* Create a clone of mbuf to be placed into ring. */
@@ -429,6 +430,7 @@ rte_pdump_init(void)
return -1;
}
pdump_stats = mz->addr;
+   pdump_stats->mz = mz;
 
ret = rte_mp_action_register(PDUMP_MP, pdump_server);
if (ret && rte_errno != ENOTSUP)
@@ -441,6 +443,11 @@ rte_pdump_uninit(void)
 {
rte_mp_action_unregister(PDUMP_MP);
 
+   if (pdump_stats != NULL) {
+   rte_memzone_free(pdump_stats->mz);
+   pdump_stats = NULL;
+   }
+
return 0;
 }
 
-- 
2.26.3



[dpdk-dev] [PATCH v6] cryptodev: add telemetry callbacks

2021-10-26 Thread Rebecca Troy
The cryptodev library now registers commands with telemetry, and
implements the corresponding callback functions. These commands
allow a list of cryptodevs to be queried, as well as info and stats
for the corresponding cryptodev.

An example usage can be seen below:

Connecting to /var/run/dpdk/rte/dpdk_telemetry.v2
{"version": "DPDK 21.11.0-rc0", "pid": 1135019, "max_output_len": 16384}
--> /
{"/": ["/", "/cryptodev/info", "/cryptodev/list", "/cryptodev/stats", ...]}
--> /cryptodev/list
{"/cryptodev/list": [0,1,2,3]}
--> /cryptodev/info,0
{"/cryptodev/info": {"device_name": ":1c:01.0_qat_sym", \
 "max_nb_queue_pairs": 2}}
--> /cryptodev/stats,0
{"/cryptodev/stats": {"enqueued_count": 0, "dequeued_count": 0, \
"enqueue_err_count": 0, "dequeue_err_count": 0}}

Signed-off-by: Rebecca Troy 
Acked-by: Ciara Power 

---
v6:
 - Replaced missed -1 return value
v5:
 - Added missing telemetry dependency to meson.build.
v4:
  - Corrected doc heading underline and link.
  - Replaced remaining -1 return values with -EINVAL.
v3:
  - Added missing version tag to patch.
v2:
  - Added documentation and release notes.
  - Changed the /cryptodev/list command to list the devices as an
array of IDs, rather than as names and IDs.
  - Added the /cryptodev/info command as described above.
---
 doc/guides/prog_guide/cryptodev_lib.rst | 28 
 doc/guides/rel_notes/release_21_11.rst  |  5 ++
 lib/cryptodev/meson.build   |  2 +-
 lib/cryptodev/rte_cryptodev.c   | 92 +
 4 files changed, 126 insertions(+), 1 deletion(-)

diff --git a/doc/guides/prog_guide/cryptodev_lib.rst 
b/doc/guides/prog_guide/cryptodev_lib.rst
index 9b1cf8d49f..25663e552e 100644
--- a/doc/guides/prog_guide/cryptodev_lib.rst
+++ b/doc/guides/prog_guide/cryptodev_lib.rst
@@ -1282,3 +1282,31 @@ Asymmetric Crypto Device API
 
 The cryptodev Library API is described in the
 `DPDK API Reference `_
+
+
+Device Statistics
+-
+
+The Cryptodev library has support for displaying Crypto device information
+through the Telemetry interface. Telemetry commands that can be used
+are shown below.
+
+#. Get the list of available Crypto devices by ID::
+
+ --> /cryptodev/list
+ {"/cryptodev/list": [0, 1, 2, 3]}
+
+#. Get general information from a Crypto device::
+
+ --> /cryptodev/info,0
+ {"/cryptodev/info": {"device_name": ":1c:01.0_qat_sym",
+ "max_nb_queue_pairs": 2}}
+
+#. Get the statistics for a particular Crypto device::
+
+ --> /cryptodev/stats,0
+ {"/cryptodev/stats": {"enqueued_count": 0, "dequeued_count": 0,
+ "enqueue_err_count": 0, "dequeue_err_count": 0}}
+
+For more information on how to use the Telemetry interface, see
+the :doc:`../howto/telemetry`.
diff --git a/doc/guides/rel_notes/release_21_11.rst 
b/doc/guides/rel_notes/release_21_11.rst
index 1ccac87b73..ee1e557309 100644
--- a/doc/guides/rel_notes/release_21_11.rst
+++ b/doc/guides/rel_notes/release_21_11.rst
@@ -269,6 +269,11 @@ New Features
   * Added support for SA telemetry.
   * Added support for setting a non default starting ESN value.
 
+* **Added Telemetry callbacks to Cryptodev library.**
+
+  Added Telemetry callback functions which allow a list of Crypto devices,
+  stats for a Crypto device, and other device information to be queried.
+
 * **Added multi-process support for testpmd.**
 
   Added command-line options to specify total number of processes and
diff --git a/lib/cryptodev/meson.build b/lib/cryptodev/meson.build
index 289b66ab76..19de3073bb 100644
--- a/lib/cryptodev/meson.build
+++ b/lib/cryptodev/meson.build
@@ -21,4 +21,4 @@ driver_sdk_headers += files(
 'cryptodev_pmd.h',
 )
 
-deps += ['kvargs', 'mbuf', 'rcu']
+deps += ['kvargs', 'mbuf', 'rcu', 'telemetry']
diff --git a/lib/cryptodev/rte_cryptodev.c b/lib/cryptodev/rte_cryptodev.c
index 305e013ebb..3b01a4bfbc 100644
--- a/lib/cryptodev/rte_cryptodev.c
+++ b/lib/cryptodev/rte_cryptodev.c
@@ -35,6 +35,7 @@
 #include 
 #include 
 #include 
+#include 
 
 #include "rte_crypto.h"
 #include "rte_cryptodev.h"
@@ -2430,3 +2431,94 @@ RTE_INIT(cryptodev_init_fp_ops)
for (i = 0; i != RTE_DIM(rte_crypto_fp_ops); i++)
cryptodev_fp_ops_reset(rte_crypto_fp_ops + i);
 }
+
+static int
+cryptodev_handle_dev_list(const char *cmd __rte_unused,
+   const char *params __rte_unused,
+   struct rte_tel_data *d)
+{
+   int dev_id;
+
+   if (rte_cryptodev_count() < 1)
+   return -EINVAL;
+
+   rte_tel_data_start_array(d, RTE_TEL_INT_VAL);
+   for (dev_id = 0; dev_id < RTE_CRYPTO_MAX_DEVS; dev_id++)
+   if (rte_cryptodev_is_valid_dev(dev_id))
+   rte_tel_data_add_array_int(d, dev_id);
+
+   return 0;
+}
+
+static int
+cryptodev_handle_dev_info(const char *cmd __rte_unused,
+   const char *params, struct rte_tel_data *d)
+{
+   struct rte_cryptodev_info cryptodev_info;
+ 

[dpdk-dev] [PATCH v4 0/4] enable protocol agnostic flow offloading in FDIR

2021-10-26 Thread Junfeng Guo
Protocol agnostic flow offloading in Flow Director is enabled by this
patch set based on the Parser Library using existing rte_flow raw API

[PATCH v4 1/4] net/ice/base: add method to disable FDIR SWAP option.
[PATCH v4 2/4] net/ice/base: add function to set HW profile for raw flow.
[PATCH v4 3/4] app/testpmd: update Max RAW pattern size to 512.
[PATCH v4 4/4] net/ice: enable protocol agnostic flow offloading in FDIR.

Junfeng Guo (4):
  net/ice/base: add method to disable FDIR SWAP option
  net/ice/base: add function to set HW profile for raw flow
  app/testpmd: update Max RAW pattern size to 512
  net/ice: enable protocol agnostic flow offloading in FDIR

* v4:
Added list to check inputset conflict.

* v3:
Added necessary base code for raw flow in FDIR.

* v2:
Enabled vxlan port add for raw flow and updated commit message

 app/test-pmd/cmdline_flow.c|   2 +-
 doc/guides/rel_notes/release_21_11.rst |   1 +
 drivers/net/ice/base/ice_flex_pipe.c   |  99 +-
 drivers/net/ice/base/ice_flex_pipe.h   |   7 +-
 drivers/net/ice/base/ice_flow.c|  87 -
 drivers/net/ice/base/ice_flow.h|   4 +
 drivers/net/ice/ice_ethdev.h   |  17 ++
 drivers/net/ice/ice_fdir_filter.c  | 249 +
 drivers/net/ice/ice_generic_flow.c |   7 +
 drivers/net/ice/ice_generic_flow.h |   3 +
 10 files changed, 471 insertions(+), 5 deletions(-)

-- 
2.25.1



[dpdk-dev] [PATCH v4 1/4] net/ice/base: add method to disable FDIR SWAP option

2021-10-26 Thread Junfeng Guo
The SWAP Flag in the FDIR Programming Descriptor doesn't work, thus
add a method to disable the FDIR SWAP option by setting the swap and
inset register set with certain values. The boolean fd_swap is used
to enable/disable the SWAP option.

Signed-off-by: Junfeng Guo 
---
 drivers/net/ice/base/ice_flex_pipe.c | 44 ++--
 drivers/net/ice/base/ice_flex_pipe.h |  3 +-
 drivers/net/ice/base/ice_flow.c  |  2 +-
 3 files changed, 45 insertions(+), 4 deletions(-)

diff --git a/drivers/net/ice/base/ice_flex_pipe.c 
b/drivers/net/ice/base/ice_flex_pipe.c
index f35d59f4f5..06a233990f 100644
--- a/drivers/net/ice/base/ice_flex_pipe.c
+++ b/drivers/net/ice/base/ice_flex_pipe.c
@@ -4952,6 +4952,43 @@ ice_add_prof_attrib(struct ice_prof_map *prof, u8 ptg, 
u16 ptype,
return ICE_SUCCESS;
 }
 
+/**
+ * ice_disable_fd_swap - set register appropriately to disable FD swap
+ * @hw: pointer to the HW struct
+ * @prof_id: profile ID
+ */
+void ice_disable_fd_swap(struct ice_hw *hw, u16 prof_id)
+{
+   u8 swap_val = ICE_SWAP_VALID;
+   u8 i;
+   /* Since the SWAP Flag in the Programming Desc doesn't work,
+* here add method to disable the SWAP Option via setting
+* certain SWAP and INSET register set.
+*/
+   for (i = 0; i < hw->blk[ICE_BLK_FD].es.fvw / 4; i++) {
+   u32 raw_swap = 0;
+   u32 raw_in = 0;
+   u8 j;
+
+   for (j = 0; j < 4; j++) {
+   raw_swap |= (swap_val++) << (j * BITS_PER_BYTE);
+   raw_in |= ICE_INSET_DFLT << (j * BITS_PER_BYTE);
+   }
+
+   /* write the FDIR swap register set */
+   wr32(hw, GLQF_FDSWAP(prof_id, i), raw_swap);
+
+   ice_debug(hw, ICE_DBG_INIT, "swap wr(%d, %d): %x = %08x\n",
+   prof_id, i, GLQF_FDSWAP(prof_id, i), raw_swap);
+
+   /* write the FDIR inset register set */
+   wr32(hw, GLQF_FDINSET(prof_id, i), raw_in);
+
+   ice_debug(hw, ICE_DBG_INIT, "inset wr(%d, %d): %x = %08x\n",
+   prof_id, i, GLQF_FDINSET(prof_id, i), raw_in);
+   }
+}
+
 /**
  * ice_add_prof - add profile
  * @hw: pointer to the HW struct
@@ -4962,6 +4999,7 @@ ice_add_prof_attrib(struct ice_prof_map *prof, u8 ptg, 
u16 ptype,
  * @attr_cnt: number of elements in attrib array
  * @es: extraction sequence (length of array is determined by the block)
  * @masks: mask for extraction sequence
+ * @fd_swap: enable/disable FDIR paired src/dst fields swap option
  *
  * This function registers a profile, which matches a set of PTYPES with a
  * particular extraction sequence. While the hardware profile is allocated
@@ -4971,7 +5009,7 @@ ice_add_prof_attrib(struct ice_prof_map *prof, u8 ptg, 
u16 ptype,
 enum ice_status
 ice_add_prof(struct ice_hw *hw, enum ice_block blk, u64 id, u8 ptypes[],
 const struct ice_ptype_attributes *attr, u16 attr_cnt,
-struct ice_fv_word *es, u16 *masks)
+struct ice_fv_word *es, u16 *masks, bool fd_swap)
 {
u32 bytes = DIVIDE_AND_ROUND_UP(ICE_FLOW_PTYPE_MAX, BITS_PER_BYTE);
ice_declare_bitmap(ptgs_used, ICE_XLT1_CNT);
@@ -4991,7 +5029,7 @@ ice_add_prof(struct ice_hw *hw, enum ice_block blk, u64 
id, u8 ptypes[],
status = ice_alloc_prof_id(hw, blk, &prof_id);
if (status)
goto err_ice_add_prof;
-   if (blk == ICE_BLK_FD) {
+   if (blk == ICE_BLK_FD && fd_swap) {
/* For Flow Director block, the extraction sequence may
 * need to be altered in the case where there are paired
 * fields that have no match. This is necessary because
@@ -5002,6 +5040,8 @@ ice_add_prof(struct ice_hw *hw, enum ice_block blk, u64 
id, u8 ptypes[],
status = ice_update_fd_swap(hw, prof_id, es);
if (status)
goto err_ice_add_prof;
+   } else if (blk == ICE_BLK_FD) {
+   ice_disable_fd_swap(hw, prof_id);
}
status = ice_update_prof_masking(hw, blk, prof_id, masks);
if (status)
diff --git a/drivers/net/ice/base/ice_flex_pipe.h 
b/drivers/net/ice/base/ice_flex_pipe.h
index 9733c4b214..dd332312dd 100644
--- a/drivers/net/ice/base/ice_flex_pipe.h
+++ b/drivers/net/ice/base/ice_flex_pipe.h
@@ -61,10 +61,11 @@ bool ice_hw_ptype_ena(struct ice_hw *hw, u16 ptype);
 /* XLT2/VSI group functions */
 enum ice_status
 ice_vsig_find_vsi(struct ice_hw *hw, enum ice_block blk, u16 vsi, u16 *vsig);
+void ice_disable_fd_swap(struct ice_hw *hw, u16 prof_id);
 enum ice_status
 ice_add_prof(struct ice_hw *hw, enum ice_block blk, u64 id, u8 ptypes[],
 const struct ice_ptype_attributes *attr, u16 attr_cnt,
-struct ice_fv_word *es, u16 *masks);
+struct ice_fv_

[dpdk-dev] [PATCH v4 2/4] net/ice/base: add function to set HW profile for raw flow

2021-10-26 Thread Junfeng Guo
Based on the parser library, we can directly set HW profile and
associate the main/ctrl vsi.

Signed-off-by: Junfeng Guo 
---
 drivers/net/ice/base/ice_flex_pipe.c | 55 ++
 drivers/net/ice/base/ice_flex_pipe.h |  4 ++
 drivers/net/ice/base/ice_flow.c  | 85 
 drivers/net/ice/base/ice_flow.h  |  4 ++
 4 files changed, 148 insertions(+)

diff --git a/drivers/net/ice/base/ice_flex_pipe.c 
b/drivers/net/ice/base/ice_flex_pipe.c
index 06a233990f..030655f3f0 100644
--- a/drivers/net/ice/base/ice_flex_pipe.c
+++ b/drivers/net/ice/base/ice_flex_pipe.c
@@ -6365,3 +6365,58 @@ ice_rem_prof_id_flow(struct ice_hw *hw, enum ice_block 
blk, u16 vsi, u64 hdl)
 
return status;
 }
+
+/**
+ * ice_flow_assoc_hw_prof - add profile id flow for main/ctrl VSI flow entry
+ * @hw: pointer to the HW struct
+ * @blk: HW block
+ * @dest_vsi_handle: dest VSI handle
+ * @fdir_vsi_handle: fdir programming VSI handle
+ * @id: profile id (handle)
+ * @fv_found: found fv in fdir fv list
+ *
+ * Calling this function will update the hardware tables to enable the
+ * profile indicated by the ID parameter for the VSIs specified in the VSI
+ * array. Once successfully called, the flow will be enabled.
+ */
+enum ice_status
+ice_flow_assoc_hw_prof(struct ice_hw *hw, enum ice_block blk,
+  u16 dest_vsi_handle, u16 fdir_vsi_handle, int id,
+  bool fv_found)
+{
+   enum ice_status status = ICE_SUCCESS;
+   u16 vsi_num;
+
+   vsi_num = ice_get_hw_vsi_num(hw, dest_vsi_handle);
+   if (!fv_found) {
+   status = ice_add_prof_id_flow(hw, blk, vsi_num, id);
+   if (status) {
+   ice_debug(hw, ICE_DBG_FLOW, "HW profile add failed for 
main VSI flow entry, %d\n",
+ status);
+   goto err_add_prof;
+   }
+   }
+
+   if (blk != ICE_BLK_FD)
+   return status;
+
+   vsi_num = ice_get_hw_vsi_num(hw, fdir_vsi_handle);
+   if (!fv_found) {
+   status = ice_add_prof_id_flow(hw, blk, vsi_num, id);
+   if (status) {
+   ice_debug(hw, ICE_DBG_FLOW, "HW profile add failed for 
ctrl VSI flow entry, %d\n",
+ status);
+   goto err_add_entry;
+   }
+   }
+
+   return status;
+
+err_add_entry:
+   vsi_num = ice_get_hw_vsi_num(hw, dest_vsi_handle);
+   ice_rem_prof_id_flow(hw, blk, vsi_num, id);
+err_add_prof:
+   ice_flow_rem_prof(hw, blk, id);
+
+   return status;
+}
diff --git a/drivers/net/ice/base/ice_flex_pipe.h 
b/drivers/net/ice/base/ice_flex_pipe.h
index dd332312dd..3621de18dc 100644
--- a/drivers/net/ice/base/ice_flex_pipe.h
+++ b/drivers/net/ice/base/ice_flex_pipe.h
@@ -76,6 +76,10 @@ enum ice_status
 ice_add_prof_id_flow(struct ice_hw *hw, enum ice_block blk, u16 vsi, u64 hdl);
 enum ice_status
 ice_rem_prof_id_flow(struct ice_hw *hw, enum ice_block blk, u16 vsi, u64 hdl);
+enum ice_status
+ice_flow_assoc_hw_prof(struct ice_hw *hw, enum ice_block blk,
+  u16 dest_vsi_handle, u16 fdir_vsi_handle, int id,
+  bool fv_found);
 enum ice_status ice_init_pkg(struct ice_hw *hw, u8 *buff, u32 len);
 enum ice_status
 ice_copy_and_init_pkg(struct ice_hw *hw, const u8 *buf, u32 len);
diff --git a/drivers/net/ice/base/ice_flow.c b/drivers/net/ice/base/ice_flow.c
index 77b6b130c1..43a72b0882 100644
--- a/drivers/net/ice/base/ice_flow.c
+++ b/drivers/net/ice/base/ice_flow.c
@@ -2524,6 +2524,91 @@ ice_flow_disassoc_prof(struct ice_hw *hw, enum ice_block 
blk,
return status;
 }
 
+#define FLAG_GTP_EH_PDU_LINK   BIT_ULL(13)
+#define FLAG_GTP_EH_PDUBIT_ULL(14)
+
+#define FLAG_GTPU_MSK  \
+   (FLAG_GTP_EH_PDU | FLAG_GTP_EH_PDU_LINK)
+#define FLAG_GTPU_DW   \
+   (FLAG_GTP_EH_PDU | FLAG_GTP_EH_PDU_LINK)
+#define FLAG_GTPU_UP   \
+   (FLAG_GTP_EH_PDU)
+/**
+ * ice_flow_set_hw_prof - Set HW flow profile based on the parsed profile info
+ * @hw: pointer to the HW struct
+ * @dest_vsi_handle: dest VSI handle
+ * @fdir_vsi_handle: fdir programming VSI handle
+ * @prof: stores parsed profile info from raw flow
+ * @blk: classification stage
+ * @fv_found: found fv in fdir fv list
+ */
+enum ice_status
+ice_flow_set_hw_prof(struct ice_hw *hw, u16 dest_vsi_handle,
+u16 fdir_vsi_handle, struct ice_parser_profile *prof,
+enum ice_block blk, bool fv_found)
+{
+   int id = ice_find_first_bit(prof->ptypes, UINT16_MAX);
+   struct ice_flow_prof_params *params;
+   u8 fv_words = hw->blk[blk].es.fvw;
+   enum ice_status status;
+   u16 vsi_num;
+   int i, idx;
+
+   params = (struct ice_flow_prof_params *)ice_malloc(hw, sizeof(*params));
+   if (!params)
+   return ICE_ERR_NO_MEMORY;
+
+   for (i = 0; i < ICE_MAX_FV_WORDS; i++) {
+   params->es[i].pro

[dpdk-dev] [PATCH v4 3/4] app/testpmd: update Max RAW pattern size to 512

2021-10-26 Thread Junfeng Guo
Update max size for pattern in struct rte_flow_item_raw to enable
protocol agnostic flow offloading.

Signed-off-by: Junfeng Guo 
---
 app/test-pmd/cmdline_flow.c | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/app/test-pmd/cmdline_flow.c b/app/test-pmd/cmdline_flow.c
index d8218771fb..ef24710bd8 100644
--- a/app/test-pmd/cmdline_flow.c
+++ b/app/test-pmd/cmdline_flow.c
@@ -495,7 +495,7 @@ enum index {
 };
 
 /** Maximum size for pattern in struct rte_flow_item_raw. */
-#define ITEM_RAW_PATTERN_SIZE 40
+#define ITEM_RAW_PATTERN_SIZE 512
 
 /** Maximum size for GENEVE option data pattern in bytes. */
 #define ITEM_GENEVE_OPT_DATA_SIZE 124
-- 
2.25.1



[dpdk-dev] [PATCH v4 4/4] net/ice: enable protocol agnostic flow offloading in FDIR

2021-10-26 Thread Junfeng Guo
Protocol agnostic flow offloading in Flow Director is enabled by this
patch based on the Parser Library, using existing rte_flow raw API.

Note that the raw flow requires:
1. byte string of raw target packet bits.
2. byte string of mask of target packet.

Here is an example:
FDIR matching ipv4 dst addr with 1.2.3.4 and redirect to queue 3:

flow create 0 ingress pattern raw \
pattern spec \
080045144000401001020304 \
pattern mask \
 \
/ end actions queue index 3 / mark id 3 / end

Note that mask of some key bits (e.g., 0x0800 to indicate ipv4 proto)
is optional in our cases. To avoid redundancy, we just omit the mask
of 0x0800 (with 0x) in the mask byte string example. The prefix
'0x' for the spec and mask byte (hex) strings are also omitted here.

Signed-off-by: Junfeng Guo 
---
 doc/guides/rel_notes/release_21_11.rst |   1 +
 drivers/net/ice/ice_ethdev.h   |  17 ++
 drivers/net/ice/ice_fdir_filter.c  | 249 +
 drivers/net/ice/ice_generic_flow.c |   7 +
 drivers/net/ice/ice_generic_flow.h |   3 +
 5 files changed, 277 insertions(+)

diff --git a/doc/guides/rel_notes/release_21_11.rst 
b/doc/guides/rel_notes/release_21_11.rst
index b327c2bfca..f7f45b9fa5 100644
--- a/doc/guides/rel_notes/release_21_11.rst
+++ b/doc/guides/rel_notes/release_21_11.rst
@@ -155,6 +155,7 @@ New Features
 
 * **Updated Intel ice driver.**
 
+  * Added protocol agnostic flow offloading support in Flow Director.
   * Added 1PPS out support by a devargs.
   * Added IPv4 and L4 (TCP/UDP/SCTP) checksum hash support in RSS flow.
   * Added DEV_RX_OFFLOAD_TIMESTAMP support.
diff --git a/drivers/net/ice/ice_ethdev.h b/drivers/net/ice/ice_ethdev.h
index 599e0028f7..6d93fca033 100644
--- a/drivers/net/ice/ice_ethdev.h
+++ b/drivers/net/ice/ice_ethdev.h
@@ -318,6 +318,11 @@ struct ice_fdir_filter_conf {
uint64_t input_set_o; /* used for non-tunnel or tunnel outer fields */
uint64_t input_set_i; /* only for tunnel inner fields */
uint32_t mark_flag;
+
+   struct ice_parser_profile *prof;
+   const u8 *pkt_buf;
+   bool parser_ena;
+   u8 pkt_len;
 };
 
 #define ICE_MAX_FDIR_FILTER_NUM(1024 * 16)
@@ -487,6 +492,17 @@ struct ice_devargs {
uint8_t pps_out_ena;
 };
 
+/**
+ * Structure to store fdir fv entry.
+ */
+struct ice_fdir_prof_info {
+   struct LIST_ENTRY_TYPE l_entry;
+
+   struct ice_parser_profile prof;
+   u16 ptype;
+   u64 fdir_actived_cnt;
+};
+
 /**
  * Structure to store private data for each PF/VF instance.
  */
@@ -509,6 +525,7 @@ struct ice_adapter {
struct rte_timecounter rx_tstamp_tc;
struct rte_timecounter tx_tstamp_tc;
bool ptp_ena;
+   struct LIST_HEAD_TYPE fdir_prof_list;
 #ifdef RTE_ARCH_X86
bool rx_use_avx2;
bool rx_use_avx512;
diff --git a/drivers/net/ice/ice_fdir_filter.c 
b/drivers/net/ice/ice_fdir_filter.c
index bd627e3aa8..81a07f191a 100644
--- a/drivers/net/ice/ice_fdir_filter.c
+++ b/drivers/net/ice/ice_fdir_filter.c
@@ -107,6 +107,7 @@
ICE_INSET_NAT_T_ESP_SPI)
 
 static struct ice_pattern_match_item ice_fdir_pattern_list[] = {
+   {pattern_raw,   ICE_INSET_NONE, 
ICE_INSET_NONE, ICE_INSET_NONE},
{pattern_ethertype, ICE_FDIR_INSET_ETH, 
ICE_INSET_NONE, ICE_INSET_NONE},
{pattern_eth_ipv4,  
ICE_FDIR_INSET_ETH_IPV4,ICE_INSET_NONE, ICE_INSET_NONE},
{pattern_eth_ipv4_udp,  
ICE_FDIR_INSET_ETH_IPV4_UDP,ICE_INSET_NONE, ICE_INSET_NONE},
@@ -1158,6 +1159,8 @@ ice_fdir_init(struct ice_adapter *ad)
if (ret)
return ret;
 
+   INIT_LIST_HEAD(&ad->fdir_prof_list);
+
parser = &ice_fdir_parser;
 
return ice_register_parser(parser, ad);
@@ -1188,6 +1191,24 @@ ice_fdir_is_tunnel_profile(enum ice_fdir_tunnel_type 
tunnel_type)
return 0;
 }
 
+static int
+ice_fdir_add_del_raw(struct ice_pf *pf,
+struct ice_fdir_filter_conf *filter,
+bool add)
+{
+   struct ice_hw *hw = ICE_PF_TO_HW(pf);
+
+   unsigned char *pkt = (unsigned char *)pf->fdir.prg_pkt;
+   rte_memcpy(pkt, filter->pkt_buf, filter->pkt_len);
+
+   struct ice_fltr_desc desc;
+   memset(&desc, 0, sizeof(desc));
+   filter->input.comp_report = ICE_FXD_FLTR_QW0_COMP_REPORT_SW;
+   ice_fdir_get_prgm_desc(hw, &filter->input, &desc, add);
+
+   return ice_fdir_programming(pf, &desc);
+}
+
 static int
 ice_fdir_add_del_filter(struct ice_pf *pf,
struct ice_fdir_filter_conf *filter,
@@ -1303,6 +1324,91 @@ ice_fdir_create_filter(struct ice_adapter *ad,
struct ice_fdir_fltr_pattern key;
   

[dpdk-dev] [Bug 839] pdump: any subsequent runs of pdump_autotest fail

2021-10-26 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=839

Bug ID: 839
   Summary: pdump: any subsequent runs of pdump_autotest fail
   Product: DPDK
   Version: 21.11
  Hardware: All
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: other
  Assignee: dev@dpdk.org
  Reporter: konstantin.anan...@intel.com
  Target Milestone: ---

rte_pdump_init() always allocates new memzone for pdump_stats.
Though rte_pdump_uninit() never frees it.
So the following combination will always fail:
rte_pdump_init(); rte_pdump_uninit(); rte_pdump_init();
The issue was caught by pdump_autotest UT.
While first test run successful, any consecutive runs
of this test-case will fail like that:
RTE>>pdump_autotest
IN PRIMARY PROCESS
rte_pdump_init(): cannot allocate pdump statistics
rte_pdump_init failed
...

I submitted fix to free pdum_stats memzone at rte_pdump_uninit():
https://patches.dpdk.org/project/dpdk/patch/20211026115301.5456-1-konstantin.anan...@intel.com/

If you think it is not a proper way (instead we should keep memzone forever, or
so),
please feel free to supersede my patch with newer one.

-- 
You are receiving this mail because:
You are the assignee for the bug.

[dpdk-dev] [PATCH] mempool/cnxk: fix max pools argument parsing

2021-10-26 Thread Volodymyr Fialko
roc_idev_npa_maxpools_set expects max_pools original value, not the aura

Fixes: 0a50a5aad299 ("mempool/cnxk: add device probe/remove")
Cc: sta...@dpdk.org

Signed-off-by: Volodymyr Fialko 
Reviewed-by: Jerin Jacob 
---
 drivers/mempool/cnxk/cnxk_mempool.c | 20 ++--
 1 file changed, 10 insertions(+), 10 deletions(-)

diff --git a/drivers/mempool/cnxk/cnxk_mempool.c 
b/drivers/mempool/cnxk/cnxk_mempool.c
index dc36be54f6..828bf3fc36 100644
--- a/drivers/mempool/cnxk/cnxk_mempool.c
+++ b/drivers/mempool/cnxk/cnxk_mempool.c
@@ -31,25 +31,25 @@ npa_aura_size_to_u32(uint8_t val)
 }
 
 static int
-parse_max_pools(const char *key, const char *value, void *extra_args)
+parse_max_pools_handler(const char *key, const char *value, void *extra_args)
 {
RTE_SET_USED(key);
uint32_t val;
 
-   val = atoi(value);
+   val = rte_align32pow2(atoi(value));
if (val < npa_aura_size_to_u32(NPA_AURA_SZ_128))
val = 128;
if (val > npa_aura_size_to_u32(NPA_AURA_SZ_1M))
val = BIT_ULL(20);
 
-   *(uint8_t *)extra_args = rte_log2_u32(val) - 6;
+   *(uint32_t *)extra_args = val;
return 0;
 }
 
-static inline uint8_t
-parse_aura_size(struct rte_devargs *devargs)
+static inline uint32_t
+parse_max_pools(struct rte_devargs *devargs)
 {
-   uint8_t aura_sz = NPA_AURA_SZ_128;
+   uint32_t max_pools = npa_aura_size_to_u32(NPA_AURA_SZ_128);
struct rte_kvargs *kvlist;
 
if (devargs == NULL)
@@ -58,11 +58,11 @@ parse_aura_size(struct rte_devargs *devargs)
if (kvlist == NULL)
goto exit;
 
-   rte_kvargs_process(kvlist, CNXK_NPA_MAX_POOLS_PARAM, &parse_max_pools,
-  &aura_sz);
+   rte_kvargs_process(kvlist, CNXK_NPA_MAX_POOLS_PARAM,
+  &parse_max_pools_handler, &max_pools);
rte_kvargs_free(kvlist);
 exit:
-   return aura_sz;
+   return max_pools;
 }
 
 static inline char *
@@ -92,7 +92,7 @@ npa_init(struct rte_pci_device *pci_dev)
dev = mz->addr;
dev->pci_dev = pci_dev;
 
-   roc_idev_npa_maxpools_set(parse_aura_size(pci_dev->device.devargs));
+   roc_idev_npa_maxpools_set(parse_max_pools(pci_dev->device.devargs));
rc = roc_npa_dev_init(dev);
if (rc)
goto mz_free;
-- 
2.25.1



Re: [dpdk-dev] [PATCH v5] doc: document a limitation for a meter with RSS action

2021-10-26 Thread Thomas Monjalon
The title should mention "mlx5 guide" as the scope.

26/10/2021 13:53, Li Zhang:
> A meter policy with RSS/Queue action is not supported
> when ``dv_xmeta_en`` enabled.

The quotes are RST syntax.
Raslan, while merging, you can remove the quotes from the commit log.

> 
> When ``dv_xmeta_en`` enabled in legacy creating flow,
> it will split into two flows
> (one set_tag with jump flow and one RSS/queue action flow).
> For meter policy as termination table,
> it cannot split flow and
> cannot support when ``dv_xmeta_en`` enabled.
> 
> Fixes: 51ec04d ("net/mlx5: connect meter policy to created flows")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Li Zhang 
> Acked-by: Matan Azrad 
> ---
>  doc/guides/nics/mlx5.rst | 1 +
>  1 file changed, 1 insertion(+)
> 
> diff --git a/doc/guides/nics/mlx5.rst b/doc/guides/nics/mlx5.rst
> index 47709d93b3..cd28c04c1e 100644
> --- a/doc/guides/nics/mlx5.rst
> +++ b/doc/guides/nics/mlx5.rst
> @@ -435,6 +435,7 @@ Limitations
>   - yellow: QUEUE, RSS, PORT_ID, REPRESENTED_PORT, JUMP, DROP, MARK and 
> SET_TAG.
>   - RED: must be DROP.
>- Policy actions of RSS for green and yellow should have the same 
> configuration except queues.
> +  - Policy with RSS/queue action is not supported when ``dv_xmeta_en`` 
> enabled.
>- meter profile packet mode is supported.
>- meter profiles of RFC2697, RFC2698 and RFC4115 are supported.





Re: [dpdk-dev] [PATCH v11 0/7] iavf: add iAVF IPsec inline crypto support

2021-10-26 Thread Zhang, Qi Z



> -Original Message-
> From: Nicolau, Radu 
> Sent: Tuesday, October 26, 2021 6:38 PM
> Cc: dev@dpdk.org; Doherty, Declan ; Sinha,
> Abhijit ; Wu, Jingjing ;
> Zhang, Qi Z ; Xing, Beilei ;
> Richardson, Bruce ; Ananyev, Konstantin
> ; Nicolau, Radu 
> Subject: [PATCH v11 0/7] iavf: add iAVF IPsec inline crypto support
> 
> Add support for inline crypto for IPsec, for ESP transport and tunnel over 
> IPv4
> and IPv6, as well as supporting the offload for ESP over UDP, and
> inconjunction with TSO for UDP and TCP flows.
> 
> Radu Nicolau (7):
>   common/iavf: add iAVF IPsec inline crypto support
>   net/iavf: rework tx path
>   net/iavf: add support for asynchronous virt channel messages
>   net/iavf: add iAVF IPsec inline crypto support
>   net/iavf: add xstats support for inline IPsec crypto
>   net/iavf: add watchdog for VFLR
>   net/iavf: update doc with inline crypto support
> 
>  doc/guides/nics/features/iavf.ini |2 +
>  doc/guides/nics/intel_vf.rst  |   10 +
>  doc/guides/rel_notes/release_21_11.rst|1 +
>  drivers/common/iavf/iavf_type.h   |1 +
>  drivers/common/iavf/virtchnl.h|   17 +-
>  drivers/common/iavf/virtchnl_inline_ipsec.h   |  553 +
>  drivers/net/iavf/iavf.h   |   52 +-
>  drivers/net/iavf/iavf_ethdev.c|  219 +-
>  drivers/net/iavf/iavf_generic_flow.c  |   15 +
>  drivers/net/iavf/iavf_generic_flow.h  |2 +
>  drivers/net/iavf/iavf_ipsec_crypto.c  | 1894 +
>  drivers/net/iavf/iavf_ipsec_crypto.h  |  160 ++
>  .../net/iavf/iavf_ipsec_crypto_capabilities.h |  383 
>  drivers/net/iavf/iavf_rxtx.c  |  710 --
>  drivers/net/iavf/iavf_rxtx.h  |  220 +-
>  drivers/net/iavf/iavf_rxtx_vec_sse.c  |   10 +-
>  drivers/net/iavf/iavf_vchnl.c |  167 +-
>  drivers/net/iavf/meson.build  |3 +-
>  drivers/net/iavf/rte_pmd_iavf.h   |1 +
>  drivers/net/iavf/version.map  |3 +
>  20 files changed, 4101 insertions(+), 322 deletions(-)  create mode 100644
> drivers/common/iavf/virtchnl_inline_ipsec.h
>  create mode 100644 drivers/net/iavf/iavf_ipsec_crypto.c
>  create mode 100644 drivers/net/iavf/iavf_ipsec_crypto.h
>  create mode 100644 drivers/net/iavf/iavf_ipsec_crypto_capabilities.h
> 
> --
> v2: small updates and fixes in the flow related section
> v3: split the huge patch and address feedback
> v4: small changes due to dependencies changes
> v5: updated the watchdow patch
> v6: rebased and updated the common section
> v7: fixed TSO issue and disabled watchdog by default
> v8: rebased to next-net-intel and added doc updates
> v9: fixed IV len for AEAD and GMAC
> v10: removed blank lines at EOF
> v11: rebased patchset
> 
> 2.25.1

For patch 1/7 and patch 7/7

Acked-by: Qi Zhang 

Applied to dpdk-next-net-intel.

Thanks
Qi



Re: [dpdk-dev] [PATCH 1/2] ethdev: fix log level of Tx and Rx dummy functions

2021-10-26 Thread Ananyev, Konstantin

> > > > > > > There is a concern about getting efficient log report,
> > > > > > > especially when looking at CI issues.
> > > > > >
> > > > > > +1.
> > > > > > The current solution with logs is a real pain.
> > > > >
> > > > > Are you guys talking about problems with
> > > > > app/test/sample_packet_forward.* David reported?
> > > > > Or some extra problems arise?
> > > >
> > > > The problem will arise each time an app is misbehaving.
> > > > That's going to be a recurring problem in the CI.
> >
> > It is still not clear to me why it is going to be a recurring one?
> > Ok, right now we have some test-cases that are misbehaving unintentionally.
> > So we need to fix them.
> > I admit that it might be a pain, but it still looks like a one time job to 
> > me.
> > With new test-cases we should be able to catch such misbehaving at patch
> > submission stage (by checking then logs).
> > I guess there might be some test-cases that misbehave intentionally -
> > some negative test-cases for error-condition checking etc.
> > But for them error message in the log and error return value seems like a
> > right thing, no? Again I expect such test-cases do erroneous rx/tx_burst
> > just few times (not dozens or hundreds) so they shouldn't pollute log too 
> > much.
> > So, what I am missing here?
> 
> You don't miss anything, but as you said above, we are going to catch
> some issues at patch submission stage.
> And we want this stage to be easy to catch.
> Having megabytes of log does not help to check in the CI.
> 
> > > One thing that could be done is compiling with asserts in CI, and let
> > > default build not have those asserts.
> >
> > Agree, log+assert seems like a good alternative to panic() for me.
> >
> > > Otherwise, logging once should be enough (I have a patch for this latter 
> > > idea).
> >
> > I understand the intention, but I am a bit sceptical about that one:
> > it is quite often people don’t pay much attention to single log message.
> 
> Not a good argument in my opinion.
> One error == one log.
> We are not going to flood all error logs to make sure devs pay attention :)

Well, that error could come from different sources (rx/tx, different ports, 
etc.).
But yes, healthy CI is important.
So if suppressing subsequent messages will help it anyhow, I wouldn't object.
Few thoughts though:
we probably need to make it more informative (and scary :)) then now:
bump log-level, print current lcore id and dump current call-stack.
Another thought - might be worth to make it logging once per lcore
(instead of global logging once).
 



[dpdk-dev] [PATCH v5 0/2] support socket direct mode bonding

2021-10-26 Thread Rongwei Liu
In socket direct mode, it's possible to bind any two (maybe four
in the future) PCIe devices with IDs like :xx:xx.x and 
:yy:yy.y. Bonding member interfaces are unnecessary to have
the same PCIe domain/bus/device ID anymore.

Doesn't need to backport to DPDK 20.11

v2: fix ci warnings.
v3: add description in release_21_11.rst.
v4: add description in mlx5.rst.
v5: rebase on top of master-net-mlx

Rongwei Liu (2):
  common/mlx5: support pcie device guid query
  net/mlx5: support socket direct mode bonding

 doc/guides/nics/mlx5.rst   |  4 ++
 doc/guides/rel_notes/release_21_11.rst |  4 ++
 drivers/common/mlx5/linux/mlx5_common_os.c | 40 
 drivers/common/mlx5/linux/mlx5_common_os.h | 19 ++
 drivers/net/mlx5/linux/mlx5_os.c   | 43 +-
 5 files changed, 101 insertions(+), 9 deletions(-)

-- 
2.27.0



[dpdk-dev] [PATCH v5 1/2] common/mlx5: support pcie device guid query

2021-10-26 Thread Rongwei Liu
sysfs entry "phys_switch_id" holds each PCIe device'
guid.

The devices which reside in the same physical NIC should
have the same guid.

Signed-off-by: Rongwei Liu 
Acked-by: Viacheslav Ovsiienko 
---
 drivers/common/mlx5/linux/mlx5_common_os.c | 40 ++
 drivers/common/mlx5/linux/mlx5_common_os.h | 19 ++
 2 files changed, 59 insertions(+)

diff --git a/drivers/common/mlx5/linux/mlx5_common_os.c 
b/drivers/common/mlx5/linux/mlx5_common_os.c
index 8db3fe790a..b516564b79 100644
--- a/drivers/common/mlx5/linux/mlx5_common_os.c
+++ b/drivers/common/mlx5/linux/mlx5_common_os.c
@@ -2,6 +2,7 @@
  * Copyright 2020 Mellanox Technologies, Ltd
  */
 
+#include 
 #include 
 #include 
 #include 
@@ -704,3 +705,42 @@ mlx5_os_open_device(struct mlx5_common_device *cdev, 
uint32_t classes)
DRV_LOG(ERR, "Failed to open IB device \"%s\".", ibv->name);
return -rte_errno;
 }
+int
+mlx5_get_device_guid(const struct rte_pci_addr *dev, uint8_t *guid, size_t len)
+{
+   char tmp[512];
+   char cur_ifname[IF_NAMESIZE + 1];
+   FILE *id_file;
+   DIR *dir;
+   struct dirent *ptr;
+   int ret;
+
+   if (guid == NULL || len < sizeof(u_int64_t) + 1)
+   return -1;
+   memset(guid, 0, len);
+   snprintf(tmp, sizeof(tmp), "/sys/bus/pci/devices/%04x:%02x:%02x.%x/net",
+   dev->domain, dev->bus, dev->devid, dev->function);
+   dir = opendir(tmp);
+   if (dir == NULL)
+   return -1;
+   /* Traverse to identify PF interface */
+   do {
+   ptr = readdir(dir);
+   if (ptr == NULL || ptr->d_type != DT_DIR) {
+   closedir(dir);
+   return -1;
+   }
+   } while (strchr(ptr->d_name, '.') || strchr(ptr->d_name, '_') ||
+strchr(ptr->d_name, 'v'));
+   snprintf(cur_ifname, sizeof(cur_ifname), "%s", ptr->d_name);
+   closedir(dir);
+   snprintf(tmp + strlen(tmp), sizeof(tmp) - strlen(tmp),
+   "/%s/phys_switch_id", cur_ifname);
+   /* Older OFED like 5.3 doesn't support read */
+   id_file = fopen(tmp, "r");
+   if (!id_file)
+   return 0;
+   ret = fscanf(id_file, "%16s", guid);
+   fclose(id_file);
+   return ret;
+}
diff --git a/drivers/common/mlx5/linux/mlx5_common_os.h 
b/drivers/common/mlx5/linux/mlx5_common_os.h
index c2957f91ec..83066e752d 100644
--- a/drivers/common/mlx5/linux/mlx5_common_os.h
+++ b/drivers/common/mlx5/linux/mlx5_common_os.h
@@ -284,4 +284,23 @@ mlx5_os_free(void *addr)
 void
 mlx5_set_context_attr(struct rte_device *dev, struct ibv_context *ctx);
 
+/**
+ * This is used to query system_image_guid as describing in PRM.
+ *
+ * @param dev[in]
+ *  Pointer to a device instance as PCIe id.
+ * @param guid[out]
+ *  Pointer to the buffer to hold device guid.
+ *  Guid is uint64_t and corresponding to 17 bytes string.
+ * @param len[in]
+ *  Guid buffer length, 17 bytes at least.
+ *
+ * @return
+ *  -1 if internal failure.
+ *  0 if OFED doesn't support.
+ *  >0 if success.
+ */
+int
+mlx5_get_device_guid(const struct rte_pci_addr *dev, uint8_t *guid, size_t 
len);
+
 #endif /* RTE_PMD_MLX5_COMMON_OS_H_ */
-- 
2.27.0



[dpdk-dev] [PATCH v5 2/2] net/mlx5: support socket direct mode bonding

2021-10-26 Thread Rongwei Liu
In socket direct mode, it's possible to bind any two (maybe four
in future) PCIe devices with IDs like :xx:xx.x and
:yy:yy.y. Bonding member interfaces are unnecessary to have
the same PCIe domain/bus/device ID anymore,

Kernel driver uses "system_image_guid" to identify if devices can
be bound together or not. Sysfs "phys_switch_id" is used to get
"system_image_guid" of each network interface.

OFED 5.4+ is required to support "phys_switch_id".

Signed-off-by: Rongwei Liu 
Acked-by: Viacheslav Ovsiienko 
---
 doc/guides/nics/mlx5.rst   |  4 +++
 doc/guides/rel_notes/release_21_11.rst |  4 +++
 drivers/net/mlx5/linux/mlx5_os.c   | 43 --
 3 files changed, 42 insertions(+), 9 deletions(-)

diff --git a/doc/guides/nics/mlx5.rst b/doc/guides/nics/mlx5.rst
index 47709d93b3..45f44c97d7 100644
--- a/doc/guides/nics/mlx5.rst
+++ b/doc/guides/nics/mlx5.rst
@@ -468,6 +468,10 @@ Limitations
 
   - TXQ affinity subjects to HW hash once enabled.
 
+- Bonding under socket direct mode
+
+  - Needs OFED 5.4+.
+
 Statistics
 --
 
diff --git a/doc/guides/rel_notes/release_21_11.rst 
b/doc/guides/rel_notes/release_21_11.rst
index 1ccac87b73..2f46b27709 100644
--- a/doc/guides/rel_notes/release_21_11.rst
+++ b/doc/guides/rel_notes/release_21_11.rst
@@ -217,6 +217,10 @@ New Features
   * Added PDCP short MAC-I support.
   * Added raw vector datapath API support.
 
+* **Updated Mellanox mlx5 driver.**
+
+  * Added socket direct mode bonding support.
+
 * **Updated NXP dpaa2_sec crypto PMD.**
 
   * Added PDCP short MAC-I support.
diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 72bbb665cf..3deae861d5 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -1898,6 +1898,8 @@ mlx5_device_bond_pci_match(const char *ibdev_name,
FILE *bond_file = NULL, *file;
int pf = -1;
int ret;
+   uint8_t cur_guid[32] = {0};
+   uint8_t guid[32] = {0};
 
/*
 * Try to get master device name. If something goes wrong suppose
@@ -1911,6 +1913,8 @@ mlx5_device_bond_pci_match(const char *ibdev_name,
np = mlx5_nl_portnum(nl_rdma, ibdev_name);
if (!np)
return -1;
+   if (mlx5_get_device_guid(pci_dev, cur_guid, sizeof(cur_guid)) < 0)
+   return -1;
/*
 * The master device might not be on the predefined port(not on port
 * index 1, it is not guaranteed), we have to scan all Infiniband
@@ -1938,6 +1942,7 @@ mlx5_device_bond_pci_match(const char *ibdev_name,
char tmp_str[IF_NAMESIZE + 32];
struct rte_pci_addr pci_addr;
struct mlx5_switch_info info;
+   int ret;
 
/* Process slave interface names in the loop. */
snprintf(tmp_str, sizeof(tmp_str),
@@ -1969,15 +1974,6 @@ mlx5_device_bond_pci_match(const char *ibdev_name,
tmp_str);
break;
}
-   /* Match PCI address, allows BDF0+pfx or BDFx+pfx. */
-   if (pci_dev->domain == pci_addr.domain &&
-   pci_dev->bus == pci_addr.bus &&
-   pci_dev->devid == pci_addr.devid &&
-   ((pci_dev->function == 0 &&
- pci_dev->function + owner == pci_addr.function) ||
-(pci_dev->function == owner &&
- pci_addr.function == owner)))
-   pf = info.port_name;
/* Get ifindex. */
snprintf(tmp_str, sizeof(tmp_str),
 "/sys/class/net/%s/ifindex", ifname);
@@ -1994,6 +1990,30 @@ mlx5_device_bond_pci_match(const char *ibdev_name,
bond_info->ports[info.port_name].pci_addr = pci_addr;
bond_info->ports[info.port_name].ifindex = ifindex;
bond_info->n_port++;
+   /*
+* Under socket direct mode, bonding will use
+* system_image_guid as identification.
+* After OFED 5.4, guid is readable (ret >= 0) under sysfs.
+* All bonding members should have the same guid even if driver
+* is using PCIe BDF.
+*/
+   ret = mlx5_get_device_guid(&pci_addr, guid, sizeof(guid));
+   if (ret < 0)
+   break;
+   else if (ret > 0) {
+   if (!memcmp(guid, cur_guid, sizeof(guid)) &&
+   owner == info.port_name &&
+   (owner != 0 || (owner == 0 &&
+   !rte_pci_addr_cmp(pci_dev, &pci_addr
+   pf = info.port_name;
+   } else if (pci_dev->domain == pci_addr.domain &&
+   pci_dev->bus == pci_addr.bus &&
+   pci_dev->devid == pci_addr.devid &&
+   ((pci_dev->function == 0 &&
+

[dpdk-dev] [v4] cryptodev: add telemetry endpoint for cryptodev capabilities

2021-10-26 Thread Gowrishankar Muthukrishnan
Add telemetry endpoint for cryptodev capabilities.

Signed-off-by: Gowrishankar Muthukrishnan 
---
Depends-on: patch-19601 ("cryptodev: add telemetry callbacks")

v4:
 - minor cleanup
 
---
 lib/cryptodev/rte_cryptodev.c | 61 +++
 1 file changed, 61 insertions(+)

diff --git a/lib/cryptodev/rte_cryptodev.c b/lib/cryptodev/rte_cryptodev.c
index f9e241c60a..e01f651cf9 100644
--- a/lib/cryptodev/rte_cryptodev.c
+++ b/lib/cryptodev/rte_cryptodev.c
@@ -2503,6 +2503,64 @@ cryptodev_handle_dev_stats(const char *cmd __rte_unused,
return 0;
 }
 
+#define CRYPTO_CAPS_SZ \
+   (RTE_ALIGN_CEIL(sizeof(struct rte_cryptodev_capabilities), \
+   sizeof(uint64_t)) /\
+sizeof(uint64_t))
+
+static int
+crypto_caps_array(struct rte_tel_data *d,
+ const struct rte_cryptodev_capabilities *capabilities)
+{
+   const struct rte_cryptodev_capabilities *dev_caps;
+   uint64_t caps_val[CRYPTO_CAPS_SZ];
+   unsigned int i = 0, j;
+
+   rte_tel_data_start_array(d, RTE_TEL_U64_VAL);
+
+   while ((dev_caps = &capabilities[i++])->op !=
+  RTE_CRYPTO_OP_TYPE_UNDEFINED) {
+   memset(&caps_val, 0, sizeof(caps_val[0]));
+   rte_memcpy(caps_val, dev_caps, sizeof(capabilities[0]));
+   for (j = 0; j < CRYPTO_CAPS_SZ; j++)
+   rte_tel_data_add_array_u64(d, caps_val[j]);
+   }
+
+   return i;
+}
+
+static int
+cryptodev_handle_dev_caps(const char *cmd __rte_unused, const char *params,
+ struct rte_tel_data *d)
+{
+   struct rte_cryptodev_info dev_info;
+   struct rte_tel_data *crypto_caps;
+   int crypto_caps_n;
+   char *end_param;
+   int dev_id;
+
+   if (!params || strlen(params) == 0 || !isdigit(*params))
+   return -EINVAL;
+
+   dev_id = strtoul(params, &end_param, 0);
+   if (*end_param != '\0')
+   CDEV_LOG_ERR("Extra parameters passed to command, ignoring");
+   if (!rte_cryptodev_is_valid_dev(dev_id))
+   return -EINVAL;
+
+   rte_tel_data_start_dict(d);
+   crypto_caps = rte_tel_data_alloc();
+   if (!crypto_caps)
+   return -ENOMEM;
+
+   rte_cryptodev_info_get(dev_id, &dev_info);
+   crypto_caps_n = crypto_caps_array(crypto_caps, dev_info.capabilities);
+   rte_tel_data_add_dict_container(d, "crypto_caps", crypto_caps, 0);
+   rte_tel_data_add_dict_int(d, "crypto_caps_n", crypto_caps_n);
+
+   return 0;
+}
+
 RTE_INIT(cryptodev_init_fp_ops)
 {
uint32_t i;
@@ -2522,4 +2580,7 @@ RTE_INIT(cryptodev_init_telemetry)
rte_telemetry_register_cmd("/cryptodev/stats",
cryptodev_handle_dev_stats,
"Returns the stats for a cryptodev. Parameters: int 
dev_id");
+   rte_telemetry_register_cmd("/cryptodev/caps",
+   cryptodev_handle_dev_caps,
+   "Returns the capabilities for a cryptodev. Parameters: 
int dev_id");
 }
-- 
2.25.1



Re: [dpdk-dev] [PATCH v5 4/5] lib/bpf: use wait event scheme for Rx/Tx iteration

2021-10-26 Thread Ananyev, Konstantin

> Hi Feifei,
> 
> > > Instead of polling for cbi->use to be updated, use wait event scheme.
> > >
> > > Furthermore, delete 'const' for 'bpf_eth_cbi_wait'. This is because of a
> > > compilation error:
> > > ---
> > > ../lib/eal/include/rte_common.h:36:13: error: read-only variable ‘value’
> > > used as ‘asm’ output
> > >36 | #define asm __asm__
> > >   | ^~~
> > >
> > > ../lib/eal/arm/include/rte_pause_64.h:66:3: note: in expansion of macro 
> > > ‘asm’
> > >66 |   asm volatile("ldaxr %w[tmp], [%x[addr]]" \
> > >   |   ^~~
> > >
> > > ../lib/eal/arm/include/rte_pause_64.h:96:3: note: in expansion of macro
> > > ‘__LOAD_EXC_32’
> > >96 |   __LOAD_EXC_32((src), dst, memorder) \
> > >   |   ^
> > >
> > > ../lib/eal/arm/include/rte_pause_64.h:167:4: note: in expansion of macro
> > > ‘__LOAD_EXC’
> > >   167 |__LOAD_EXC((addr), value, memorder, size) \
> > >   |^~
> > >
> > > ../lib/bpf/bpf_pkt.c:125:3: note: in expansion of macro ‘rte_wait_event’
> > >   125 |   rte_wait_event(&cbi->use, UINT32_MAX, ==, puse,
> > > ---
> > >
> > > Signed-off-by: Feifei Wang 
> > > Reviewed-by: Ruifeng Wang 
> > > ---
> > >  lib/bpf/bpf_pkt.c | 11 ---
> > >  1 file changed, 4 insertions(+), 7 deletions(-)
> > >
> > > diff --git a/lib/bpf/bpf_pkt.c b/lib/bpf/bpf_pkt.c index
> > > 6e8248f0d6..213d44a75a 100644
> > > --- a/lib/bpf/bpf_pkt.c
> > > +++ b/lib/bpf/bpf_pkt.c
> > > @@ -111,9 +111,9 @@ bpf_eth_cbi_unuse(struct bpf_eth_cbi *cbi)
> > >   * Waits till datapath finished using given callback.
> > >   */
> > >  static void
> > > -bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
> > > +bpf_eth_cbi_wait(struct bpf_eth_cbi *cbi)
> >
> > Hi, Konstantin
> >
> > For this bpf patch, I delete 'const' through this is contrary to what we
> > discussed earlier. This is because if  we keep 'constant' here and use 
> > 'rte_wait_event'
> > new macro, compiler will report error. And earlier the arm version cannot 
> > be compiled
> > due to I forgot enable "wfe" config in the meson file, so this issue can 
> > not happen before.
> 
> 
> Honestly, I don't understand why we have to remove perfectly valid 'const' 
> qualifier here.
> If this macro can't be used with pointers to const (still don't understand 
> why),
> then let's just not use this macro here.
> Strictly speaking I don't see much benefit here from it.
> 
> >
> > >  {
> > > - uint32_t nuse, puse;
> > > + uint32_t puse;
> > >
> > >   /* make sure all previous loads and stores are completed */
> > >   rte_smp_mb();
> > > @@ -122,11 +122,8 @@ bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
> > >
> > >   /* in use, busy wait till current RX/TX iteration is finished */
> > >   if ((puse & BPF_ETH_CBI_INUSE) != 0) {
> > > - do {
> > > - rte_pause();
> > > - rte_compiler_barrier();
> > > - nuse = cbi->use;
> > > - } while (nuse == puse);
> > > + rte_wait_event(&cbi->use, UINT32_MAX, ==, puse,
> > > + __ATOMIC_RELAXED);

After another thought, if we do type conversion at macro invocation time:

bpf_eth_cbi_wait(const struct bpf_eth_cbi *cbi)
{
  ...
  rte_wait_event((uint32_t *)&cbi->use, UINT32_MAX, ==, puse, __ATOMIC_RELAXED);

would that help?


> > >   }
> > >  }
> > >
> > > --
> > > 2.25.1



  1   2   3   >