RE: [PATCH 1/3] net/mlx5/hws: support jump FDB Rx

2025-02-26 Thread Suanming Mou



> -Original Message-
> From: Bing Zhao 
> Sent: Wednesday, February 26, 2025 6:29 PM
> To: Suanming Mou ; Dariusz Sosnowski
> ; Slava Ovsiienko ; Ori
> Kam ; Matan Azrad 
> Cc: dev@dpdk.org; Raslan Darawsheh ; Alex Vesker
> 
> Subject: RE: [PATCH 1/3] net/mlx5/hws: support jump FDB Rx
> 
> Hi Suanming,
> 
> PSB
> 
> > -Original Message-
> > From: Suanming Mou 
> > Sent: Tuesday, February 25, 2025 8:45 AM
> > To: Dariusz Sosnowski ; Slava Ovsiienko
> > ; Bing Zhao ; Ori Kam
> > ; Matan Azrad 
> > Cc: dev@dpdk.org; Raslan Darawsheh ; Alex Vesker
> > 
> > Subject: [PATCH 1/3] net/mlx5/hws: support jump FDB Rx
> >
> > Before FW introduced the JUMP_FDB_RX action feature, jump from FDB Tx
> > to Rx is not allowed. JUMP_FDB_RX feature introduces the internal
> > loopback for Tx case and allow the REG C0 C1 B be preserved as well.
> >
> > This commit adds the JUMP_FDB_RX cap bit check and use JUMP_FDB_RX
> > instead of FT for dest table FDB Rx case.
> >
> > Signed-off-by: Suanming Mou 
> > Signed-off-by: Alex Vesker 
> > ---
> >  drivers/common/mlx5/mlx5_prm.h|  9 -
> >  drivers/net/mlx5/hws/mlx5dr_action.c  | 26
> +-
> > drivers/net/mlx5/hws/mlx5dr_action.h  |  4 
> >  drivers/net/mlx5/hws/mlx5dr_cmd.c |  9 +
> >  drivers/net/mlx5/hws/mlx5dr_cmd.h |  2 ++
> >  drivers/net/mlx5/hws/mlx5dr_context.c | 17 +
> > drivers/net/mlx5/hws/mlx5dr_context.h |  2 ++
> >  7 files changed, 63 insertions(+), 6 deletions(-)
> >
[snip]
> MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX)
> > &&
> > +   action->dest_tbl.type == MLX5DR_TABLE_TYPE_FDB_RX)
> > +   attr->action_type =
> > MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FLOW_TABLE_FDB_RX;
> > +   else
> > +   attr->action_type =
> MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FT;
> 
> Fix the long line warnings, same as other places.
> Also mlx5dr_context_cap_stc() seems to contain some calculation inside. If
> the context and the CAP will not be changed during the runtime on a specific
> port. Can this CAP BIT checking be done only once and saved for later usage?
> Since the code is used in the flow insertion.WDYT?

Ack for the line. But mlx5dr_context_cap_stc() is planned as get other caps 
based on the bit as well. And meanwhile that function is called by action 
creation, the jump action creation is done inside table creation, it is not in 
the flow insertion routine, so we are OK here.

> 
> > +
> > +   break;
[snip]


[PATCH] app/test-pmd: enable hw checksum offload for txonly

2025-02-26 Thread Nicolas PLANEL
This patch fix txonly forward mode, as it was lacking of MBUF
offload flags for IPv4 and UDP when the hardware support it.

Signed-off-by: Nicolas PLANEL 
---
 app/test-pmd/txonly.c | 4 
 1 file changed, 4 insertions(+)

diff --git a/app/test-pmd/txonly.c b/app/test-pmd/txonly.c
index 59d821a22d6f..bdcf6ea66008 100644
--- a/app/test-pmd/txonly.c
+++ b/app/test-pmd/txonly.c
@@ -334,6 +334,10 @@ pkt_burst_transmit(struct fwd_stream *fs)
ol_flags |= RTE_MBUF_F_TX_QINQ;
if (tx_offloads & RTE_ETH_TX_OFFLOAD_MACSEC_INSERT)
ol_flags |= RTE_MBUF_F_TX_MACSEC;
+   if (tx_offloads & RTE_ETH_TX_OFFLOAD_IPV4_CKSUM)
+   ol_flags |= RTE_MBUF_F_TX_IP_CKSUM;
+   if (tx_offloads & RTE_ETH_TX_OFFLOAD_UDP_CKSUM)
+   ol_flags |= RTE_MBUF_F_TX_UDP_CKSUM;
 
/*
 * Initialize Ethernet header.
-- 
2.43.0



RE: [PATCH 3/3] net/mlx5: allow FDB RSS

2025-02-26 Thread Bing Zhao
Hi



> -Original Message-
> From: Suanming Mou 
> Sent: Wednesday, February 26, 2025 9:43 PM
> To: Bing Zhao ; Dariusz Sosnowski
> ; Slava Ovsiienko ; Ori Kam
> ; Matan Azrad 
> Cc: dev@dpdk.org; Raslan Darawsheh 
> Subject: RE: [PATCH 3/3] net/mlx5: allow FDB RSS
> 
> 
> 
> > -Original Message-
> > From: Bing Zhao 
> > Sent: Wednesday, February 26, 2025 6:50 PM
> > To: Suanming Mou ; Dariusz Sosnowski
> > ; Slava Ovsiienko ; Ori
> > Kam ; Matan Azrad 
> > Cc: dev@dpdk.org; Raslan Darawsheh 
> > Subject: RE: [PATCH 3/3] net/mlx5: allow FDB RSS
> >
> >
> >
> > > -Original Message-
> > > From: Suanming Mou 
> > > Sent: Tuesday, February 25, 2025 8:45 AM
> > > To: Dariusz Sosnowski ; Slava Ovsiienko
> > > ; Bing Zhao ; Ori Kam
> > > ; Matan Azrad 
> > > Cc: dev@dpdk.org; Raslan Darawsheh 
> > > Subject: [PATCH 3/3] net/mlx5: allow FDB RSS
> > >
> > > RSS can be used in FDB Rx rules when JUMP_FDB_RX action to allow
> > > jump from FDB Tx to FDB Rx table.
> > >
> > > Different with NIC RSS, FDB RSS will not do the internal implicit
> > > metadata copy.
> > >
> > > This commit enables the FDB RSS if JUMP_FDB_RX is supported.
> > >
> > > Signed-off-by: Suanming Mou 
> > > ---
> > >  drivers/net/mlx5/mlx5_flow.c|  4 ++--
> > >  drivers/net/mlx5/mlx5_flow_dv.c | 32
> > > +--- drivers/net/mlx5/mlx5_flow_hw.c |
> > > 19 ++-
> > >  3 files changed, 33 insertions(+), 22 deletions(-)
> > >
> > > diff --git a/drivers/net/mlx5/mlx5_flow.c
> > > b/drivers/net/mlx5/mlx5_flow.c index 3fbe89a9d4..9c6a4f39fb 100644
> > > --- a/drivers/net/mlx5/mlx5_flow.c
> > > +++ b/drivers/net/mlx5/mlx5_flow.c
> > > @@ -1977,9 +1977,9 @@ mlx5_flow_validate_action_mark(struct
> > > rte_eth_dev *dev,
> > > RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
> > > "mark action not supported for "
> > > "egress");
> > > - if (attr->transfer && mlx5_hws_active(dev))
> > > + if (attr->transfer && !mlx5_hws_active(dev))
> > >   return rte_flow_error_set(error, ENOTSUP,
> > > -   RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
> > > +   RTE_FLOW_ERROR_TYPE_ATTR_TRANSFER, 
> > > NULL,
> > > "non-template mark action not 
> > > supported
> for
> > transfer");
> > >   return 0;
> > >  }
> > > diff --git a/drivers/net/mlx5/mlx5_flow_dv.c
> > > b/drivers/net/mlx5/mlx5_flow_dv.c index 633c41e358..61d3101ce8
> > 100644
> > > --- a/drivers/net/mlx5/mlx5_flow_dv.c
> > > +++ b/drivers/net/mlx5/mlx5_flow_dv.c
> > > @@ -8939,21 +8939,23 @@ flow_dv_validate(struct rte_eth_dev *dev,
> > > const struct rte_flow_attr *attr,
> > > RTE_FLOW_ERROR_TYPE_ACTION,
> > > NULL,
> > > "unsupported action MARK");
> > > - if (action_flags & MLX5_FLOW_ACTION_QUEUE)
> > > - return rte_flow_error_set(error, ENOTSUP,
> > > -   RTE_FLOW_ERROR_TYPE_ACTION,
> > > -   NULL,
> > > -   "unsupported action QUEUE");
> > > - if (action_flags & MLX5_FLOW_ACTION_RSS)
> > > - return rte_flow_error_set(error, ENOTSUP,
> > > -   RTE_FLOW_ERROR_TYPE_ACTION,
> > > -   NULL,
> > > -   "unsupported action RSS");
> > > - if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
> > > - return rte_flow_error_set(error, EINVAL,
> > > -   RTE_FLOW_ERROR_TYPE_ACTION,
> > > -   actions,
> > > -   "no fate action is found");
> > > + if (!priv->jump_fdb_rx_en) {
> > > + if (action_flags & MLX5_FLOW_ACTION_QUEUE)
> > > + return rte_flow_error_set(error, ENOTSUP,
> > > +   
> > > RTE_FLOW_ERROR_TYPE_ACTION,
> > > +   NULL,
> > > +   "unsupported action 
> > > QUEUE");
> > > + if (action_flags & MLX5_FLOW_ACTION_RSS)
> > > + return rte_flow_error_set(error, ENOTSUP,
> > > +   
> > > RTE_FLOW_ERROR_TYPE_ACTION,
> > > +   NULL,
> > > +   "unsupported action 
> > > RSS");
> > > + if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
> > > + return rte_flow_error_set(error, EINVAL,
> > > + 

RE: [PATCH 3/3] net/mlx5: allow FDB RSS

2025-02-26 Thread Suanming Mou



> -Original Message-
> From: Bing Zhao 
> Sent: Wednesday, February 26, 2025 6:50 PM
> To: Suanming Mou ; Dariusz Sosnowski
> ; Slava Ovsiienko ; Ori
> Kam ; Matan Azrad 
> Cc: dev@dpdk.org; Raslan Darawsheh 
> Subject: RE: [PATCH 3/3] net/mlx5: allow FDB RSS
> 
> 
> 
> > -Original Message-
> > From: Suanming Mou 
> > Sent: Tuesday, February 25, 2025 8:45 AM
> > To: Dariusz Sosnowski ; Slava Ovsiienko
> > ; Bing Zhao ; Ori Kam
> > ; Matan Azrad 
> > Cc: dev@dpdk.org; Raslan Darawsheh 
> > Subject: [PATCH 3/3] net/mlx5: allow FDB RSS
> >
> > RSS can be used in FDB Rx rules when JUMP_FDB_RX action to allow jump
> > from FDB Tx to FDB Rx table.
> >
> > Different with NIC RSS, FDB RSS will not do the internal implicit
> > metadata copy.
> >
> > This commit enables the FDB RSS if JUMP_FDB_RX is supported.
> >
> > Signed-off-by: Suanming Mou 
> > ---
> >  drivers/net/mlx5/mlx5_flow.c|  4 ++--
> >  drivers/net/mlx5/mlx5_flow_dv.c | 32 +---
> > drivers/net/mlx5/mlx5_flow_hw.c | 19 ++-
> >  3 files changed, 33 insertions(+), 22 deletions(-)
> >
> > diff --git a/drivers/net/mlx5/mlx5_flow.c
> > b/drivers/net/mlx5/mlx5_flow.c index 3fbe89a9d4..9c6a4f39fb 100644
> > --- a/drivers/net/mlx5/mlx5_flow.c
> > +++ b/drivers/net/mlx5/mlx5_flow.c
> > @@ -1977,9 +1977,9 @@ mlx5_flow_validate_action_mark(struct
> > rte_eth_dev *dev,
> >   RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
> >   "mark action not supported for "
> >   "egress");
> > -   if (attr->transfer && mlx5_hws_active(dev))
> > +   if (attr->transfer && !mlx5_hws_active(dev))
> > return rte_flow_error_set(error, ENOTSUP,
> > - RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
> > + RTE_FLOW_ERROR_TYPE_ATTR_TRANSFER, 
> > NULL,
> >   "non-template mark action not 
> > supported for
> transfer");
> > return 0;
> >  }
> > diff --git a/drivers/net/mlx5/mlx5_flow_dv.c
> > b/drivers/net/mlx5/mlx5_flow_dv.c index 633c41e358..61d3101ce8
> 100644
> > --- a/drivers/net/mlx5/mlx5_flow_dv.c
> > +++ b/drivers/net/mlx5/mlx5_flow_dv.c
> > @@ -8939,21 +8939,23 @@ flow_dv_validate(struct rte_eth_dev *dev,
> > const struct rte_flow_attr *attr,
> >   RTE_FLOW_ERROR_TYPE_ACTION,
> >   NULL,
> >   "unsupported action MARK");
> > -   if (action_flags & MLX5_FLOW_ACTION_QUEUE)
> > -   return rte_flow_error_set(error, ENOTSUP,
> > - RTE_FLOW_ERROR_TYPE_ACTION,
> > - NULL,
> > - "unsupported action QUEUE");
> > -   if (action_flags & MLX5_FLOW_ACTION_RSS)
> > -   return rte_flow_error_set(error, ENOTSUP,
> > - RTE_FLOW_ERROR_TYPE_ACTION,
> > - NULL,
> > - "unsupported action RSS");
> > -   if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
> > -   return rte_flow_error_set(error, EINVAL,
> > - RTE_FLOW_ERROR_TYPE_ACTION,
> > - actions,
> > - "no fate action is found");
> > +   if (!priv->jump_fdb_rx_en) {
> > +   if (action_flags & MLX5_FLOW_ACTION_QUEUE)
> > +   return rte_flow_error_set(error, ENOTSUP,
> > + 
> > RTE_FLOW_ERROR_TYPE_ACTION,
> > + NULL,
> > + "unsupported action 
> > QUEUE");
> > +   if (action_flags & MLX5_FLOW_ACTION_RSS)
> > +   return rte_flow_error_set(error, ENOTSUP,
> > + 
> > RTE_FLOW_ERROR_TYPE_ACTION,
> > + NULL,
> > + "unsupported action 
> > RSS");
> > +   if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
> > +   return rte_flow_error_set(error, EINVAL,
> > + 
> > RTE_FLOW_ERROR_TYPE_ACTION,
> > + actions,
> > + "no fate action is 
> > found");
> > +   }
> > } else {
> > if (!(action_flags & MLX5_FLOW_FATE_ACTIONS) && attr->ingress)
> > return rte_flow_error_

RE: [PATCH 2/3] net/mlx5: add jump FDB Rx flag

2025-02-26 Thread Suanming Mou



> -Original Message-
> From: Bing Zhao 
> Sent: Wednesday, February 26, 2025 6:38 PM
> To: Suanming Mou ; Dariusz Sosnowski
> ; Slava Ovsiienko ; Ori
> Kam ; Matan Azrad 
> Cc: dev@dpdk.org; Raslan Darawsheh 
> Subject: RE: [PATCH 2/3] net/mlx5: add jump FDB Rx flag
> 
> Hi
> 
> > -Original Message-
> > From: Suanming Mou 
> > Sent: Tuesday, February 25, 2025 8:45 AM
> > To: Dariusz Sosnowski ; Slava Ovsiienko
> > ; Bing Zhao ; Ori Kam
> > ; Matan Azrad 
> > Cc: dev@dpdk.org; Raslan Darawsheh 
> > Subject: [PATCH 2/3] net/mlx5: add jump FDB Rx flag
> >
> > When jump FDB Rx is supported, flow will be able to jump from FDB Tx
> > to FDB Rx, in that case the dest action in FDB Rx table should support
> > FDB Tx as well.
> >
> > Signed-off-by: Suanming Mou 
> > ---
> >  drivers/common/mlx5/mlx5_devx_cmds.c | 8 
> > drivers/common/mlx5/mlx5_devx_cmds.h | 1 +
> >  drivers/net/mlx5/linux/mlx5_os.c | 1 +
> >  drivers/net/mlx5/mlx5.h  | 1 +
> >  drivers/net/mlx5/mlx5_flow_hw.c  | 8 +++-
> >  5 files changed, 18 insertions(+), 1 deletion(-)
> >
> > diff --git a/drivers/common/mlx5/mlx5_devx_cmds.c
> > b/drivers/common/mlx5/mlx5_devx_cmds.c
> > index f504b29f31..eb8553e8ad 100644
> > --- a/drivers/common/mlx5/mlx5_devx_cmds.c
> > +++ b/drivers/common/mlx5/mlx5_devx_cmds.c
> > @@ -924,6 +924,7 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
> > uint32_t out[MLX5_ST_SZ_DW(query_hca_cap_out)] = {0};
> > bool hca_cap_2_sup;
> > uint64_t general_obj_types_supported = 0;
> > +   uint64_t stc_action_type_127_64;
> > void *hcattr;
> > int rc, i;
> >
> > @@ -1352,6 +1353,13 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
> > attr->fdb_unified_en = MLX5_GET(wqe_based_flow_table_cap,
> > hcattr,
> > fdb_unified_en);
> > +   stc_action_type_127_64 =
> MLX5_GET64(wqe_based_flow_table_cap,
> > +   hcattr,
> > +   stc_action_type_127_64);
> > +   if (stc_action_type_127_64 &
> > +  (1 <<
> > (MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX -
> > + MLX5_IFC_STC_ACTION_TYPE_BIT_64_INDEX)))
> > +   attr->jump_fdb_rx_en = true;
> 
> I assume integer value 1 is enough, since it is a bit field, no need to use 
> bool
> type.
Will change it to 1.

> 
> > }
> > /* Query HCA attribute for ROCE. */
> > if (attr->roce) {
> > diff --git a/drivers/common/mlx5/mlx5_devx_cmds.h
> > b/drivers/common/mlx5/mlx5_devx_cmds.h
> > index 8de4210fb2..6c726a0d46 100644
> > --- a/drivers/common/mlx5/mlx5_devx_cmds.h
> > +++ b/drivers/common/mlx5/mlx5_devx_cmds.h
> > @@ -326,6 +326,7 @@ struct mlx5_hca_attr {
> > uint32_t lag_rx_port_affinity:1;
> > uint32_t wqe_based_flow_table_sup:1;
> > uint32_t fdb_unified_en:1;
> > +   uint32_t jump_fdb_rx_en:1;
> > uint8_t max_header_modify_pattern_length;
> > uint64_t system_image_guid;
> > uint32_t log_max_conn_track_offload:5; diff --git
> > a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
> > index 9410211e3b..4e64026137 100644
> > --- a/drivers/net/mlx5/linux/mlx5_os.c
> > +++ b/drivers/net/mlx5/linux/mlx5_os.c
> > @@ -1718,6 +1718,7 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
> > priv->unified_fdb_en = !!priv->master;
> > DRV_LOG(DEBUG, "port %u: unified FDB %s enabled.",
> > eth_dev->data->port_id, priv->unified_fdb_en ? "is" :
> > "isn't");
> > +   priv->jump_fdb_rx_en = sh->cdev-
> > >config.hca_attr.jump_fdb_rx_en;
> 
> Can we add some INFO or DEBUG level log to indicate this to the user?
Sure, and I think here we have another improvement as if unified_fdb is not 
enabled, jump_fdb_rx_en does not make any sense. Will update.

> 
> > if (priv->sh->config.dv_esw_en) {
> > uint32_t usable_bits;
> > uint32_t required_bits;
> > diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h index
> > f73f6e63ff..545ba48b3c 100644
> > --- a/drivers/net/mlx5/mlx5.h
> > +++ b/drivers/net/mlx5/mlx5.h
> > @@ -1987,6 +1987,7 @@ struct mlx5_priv {
> > uint32_t num_lag_ports:4; /* Number of ports can be bonded. */
> > uint32_t tunnel_enabled:1; /* If tunnel offloading is enabled on
> > rxqs. */
> > uint32_t unified_fdb_en:1; /* Unified FDB flag per port. */
> > +   uint32_t jump_fdb_rx_en:1; /* Jump from FDB Tx to FDB Rx flag per
> > +port. */
> > uint16_t domain_id; /* Switch domain identifier. */
> > uint16_t vport_id; /* Associated VF vport index (if any). */
> > uint32_t vport_meta_tag; /* Used for vport index match ove VF LAG.
> > */ diff --git a/drivers/net/mlx5/mlx5_flow_hw.c
> > b/drivers/net/mlx5/mlx5_flow_hw.c index f0888dbf0e..83f55ed3e8 100644
> > --- a/drivers/net/mlx5/mlx5_flow_hw.c
> > +++ b/drivers/net/mlx

[PATCH] mempool: micro optimizations

2025-02-26 Thread Morten Brørup
The comparisons lcore_id < RTE_MAX_LCORE and lcore_id != LCORE_ID_ANY are
equivalent, but the latter compiles to fewer bytes of code space.
Similarly for lcore_id >= RTE_MAX_LCORE and lcore_id == LCORE_ID_ANY.

The rte_mempool_get_ops() function is also used in the fast path, so
RTE_VERIFY() was replaced by RTE_ASSERT().

Compilers implicitly consider comparisons of variable == 0 likely, so
unlikely() was added to the check for no mempool cache (mp->cache_size ==
0) in the rte_mempool_default_cache() function.

The rte_mempool_do_generic_put() function for adding objects to a mempool
was refactored as follows:
- The comparison for the request itself being too big, which is considered
  unlikely, was moved down and out of the code path where the cache has
  sufficient room for the added objects, which is considered the most
  likely code path.
- Added __rte_assume() about the cache length, size and threshold, for
  compiler optimization when "n" is compile time constant.
- Added __rte_assume() about "ret" being zero, so other functions using
  the value returned by this function can be potentially optimized by the
  compiler; especially when it merges multiple sequential code paths of
  inlined code depending on the return value being either zero or
  negative.
- The refactored source code (with comments) made the separate comment
  describing the cache flush/add algorithm superfluous, so it was removed.

A few more likely()/unlikely() were added.

A few comments were improved for readability.

Some assertions, RTE_ASSERT(), were added. Most importantly to assert that
the return values of the mempool drivers' enqueue and dequeue operations
are API compliant, i.e. 0 (for success) or negative (for failure), and
never positive.

Signed-off-by: Morten Brørup 
---
 lib/mempool/rte_mempool.h | 67 ++-
 1 file changed, 38 insertions(+), 29 deletions(-)

diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index c495cc012f..aedc100964 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -334,7 +334,7 @@ struct __rte_cache_aligned rte_mempool {
 #ifdef RTE_LIBRTE_MEMPOOL_STATS
 #define RTE_MEMPOOL_STAT_ADD(mp, name, n) do { 
 \
unsigned int __lcore_id = rte_lcore_id();   
\
-   if (likely(__lcore_id < RTE_MAX_LCORE)) 
\
+   if (likely(__lcore_id != LCORE_ID_ANY)) 
\
(mp)->stats[__lcore_id].name += (n);
\
else
\

rte_atomic_fetch_add_explicit(&((mp)->stats[RTE_MAX_LCORE].name),  \
@@ -751,7 +751,7 @@ extern struct rte_mempool_ops_table rte_mempool_ops_table;
 static inline struct rte_mempool_ops *
 rte_mempool_get_ops(int ops_index)
 {
-   RTE_VERIFY((ops_index >= 0) && (ops_index < RTE_MEMPOOL_MAX_OPS_IDX));
+   RTE_ASSERT((ops_index >= 0) && (ops_index < RTE_MEMPOOL_MAX_OPS_IDX));
 
return &rte_mempool_ops_table.ops[ops_index];
 }
@@ -791,7 +791,8 @@ rte_mempool_ops_dequeue_bulk(struct rte_mempool *mp,
rte_mempool_trace_ops_dequeue_bulk(mp, obj_table, n);
ops = rte_mempool_get_ops(mp->ops_index);
ret = ops->dequeue(mp, obj_table, n);
-   if (ret == 0) {
+   RTE_ASSERT(ret <= 0);
+   if (likely(ret == 0)) {
RTE_MEMPOOL_STAT_ADD(mp, get_common_pool_bulk, 1);
RTE_MEMPOOL_STAT_ADD(mp, get_common_pool_objs, n);
}
@@ -816,11 +817,14 @@ rte_mempool_ops_dequeue_contig_blocks(struct rte_mempool 
*mp,
void **first_obj_table, unsigned int n)
 {
struct rte_mempool_ops *ops;
+   int ret;
 
ops = rte_mempool_get_ops(mp->ops_index);
RTE_ASSERT(ops->dequeue_contig_blocks != NULL);
rte_mempool_trace_ops_dequeue_contig_blocks(mp, first_obj_table, n);
-   return ops->dequeue_contig_blocks(mp, first_obj_table, n);
+   ret = ops->dequeue_contig_blocks(mp, first_obj_table, n);
+   RTE_ASSERT(ret <= 0);
+   return ret;
 }
 
 /**
@@ -848,6 +852,7 @@ rte_mempool_ops_enqueue_bulk(struct rte_mempool *mp, void * 
const *obj_table,
rte_mempool_trace_ops_enqueue_bulk(mp, obj_table, n);
ops = rte_mempool_get_ops(mp->ops_index);
ret = ops->enqueue(mp, obj_table, n);
+   RTE_ASSERT(ret <= 0);
 #ifdef RTE_LIBRTE_MEMPOOL_DEBUG
if (unlikely(ret < 0))
RTE_MEMPOOL_LOG(CRIT, "cannot enqueue %u objects to mempool %s",
@@ -1333,10 +1338,10 @@ rte_mempool_cache_free(struct rte_mempool_cache *cache);
 static __rte_always_inline struct rte_mempool_cache *
 rte_mempool_default_cache(struct rte_mempool *mp, unsigned lcore_id)
 {
-   if (mp->cache_size == 0)
+   if (unlikely(mp->cache_size == 0))
return NULL;
 
-   if (lcore_id >= RTE_MAX_LCORE)
+   if (unlikely(lcore_id == 

Re: [PATCH v3 0/3] net/mlx5: support jump FDB Rx

2025-02-26 Thread Raslan Darawsheh
Hi,

From: Dariusz Sosnowski 
Sent: Wednesday, February 26, 2025 6:04 PM
To: Dariusz Sosnowski; Slava Ovsiienko; Bing Zhao; Ori Kam; Suanming Mou; Matan 
Azrad
Cc: dev@dpdk.org; Raslan Darawsheh
Subject: [PATCH v3 0/3] net/mlx5: support jump FDB Rx

Before FW introduced the JUMP_FDB_RX action feature, jump from
FDB Tx to Rx is not allowed. JUMP_FDB_RX feature introduces
the internal loopback for Tx case and allow the REG C0 C1 B be
preserved as well.

This series adds the jump from FDB Tx to FDB Rx support and allows
RSS action in FDB domain rule when jump FDB Rx is supported.

v3:
- Rebased on v2 of mlx5 Unified FDB patches (series 34723)

Suanming Mou (3):
  net/mlx5/hws: support jump FDB Rx
  net/mlx5: add jump FDB Rx flag
  net/mlx5: allow FDB RSS

 doc/guides/nics/mlx5.rst  |  4 
 drivers/common/mlx5/mlx5_devx_cmds.c  |  8 +++
 drivers/common/mlx5/mlx5_devx_cmds.h  |  1 +
 drivers/common/mlx5/mlx5_prm.h|  9 +++-
 drivers/net/mlx5/hws/mlx5dr_action.c  | 26 +-
 drivers/net/mlx5/hws/mlx5dr_action.h  |  4 
 drivers/net/mlx5/hws/mlx5dr_cmd.c |  9 
 drivers/net/mlx5/hws/mlx5dr_cmd.h |  2 ++
 drivers/net/mlx5/hws/mlx5dr_context.c | 17 ++
 drivers/net/mlx5/hws/mlx5dr_context.h |  2 ++
 drivers/net/mlx5/linux/mlx5_os.c  |  9 ++--
 drivers/net/mlx5/mlx5.h   |  1 +
 drivers/net/mlx5/mlx5_flow.c  |  4 ++--
 drivers/net/mlx5/mlx5_flow_dv.c   | 32 ++-
 drivers/net/mlx5/mlx5_flow_hw.c   | 28 ++-
 15 files changed, 125 insertions(+), 31 deletions(-)

--
2.39.5

Series applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh




Re: [PATCH] mempool: micro optimizations

2025-02-26 Thread Bruce Richardson
On Wed, Feb 26, 2025 at 03:59:22PM +, Morten Brørup wrote:
> The comparisons lcore_id < RTE_MAX_LCORE and lcore_id != LCORE_ID_ANY are
> equivalent, but the latter compiles to fewer bytes of code space.
> Similarly for lcore_id >= RTE_MAX_LCORE and lcore_id == LCORE_ID_ANY.
> 
> The rte_mempool_get_ops() function is also used in the fast path, so
> RTE_VERIFY() was replaced by RTE_ASSERT().
> 
> Compilers implicitly consider comparisons of variable == 0 likely, so
> unlikely() was added to the check for no mempool cache (mp->cache_size ==
> 0) in the rte_mempool_default_cache() function.
> 
> The rte_mempool_do_generic_put() function for adding objects to a mempool
> was refactored as follows:
> - The comparison for the request itself being too big, which is considered
>   unlikely, was moved down and out of the code path where the cache has
>   sufficient room for the added objects, which is considered the most
>   likely code path.
> - Added __rte_assume() about the cache length, size and threshold, for
>   compiler optimization when "n" is compile time constant.
> - Added __rte_assume() about "ret" being zero, so other functions using
>   the value returned by this function can be potentially optimized by the
>   compiler; especially when it merges multiple sequential code paths of
>   inlined code depending on the return value being either zero or
>   negative.
> - The refactored source code (with comments) made the separate comment
>   describing the cache flush/add algorithm superfluous, so it was removed.
> 
> A few more likely()/unlikely() were added.
> 
> A few comments were improved for readability.
> 
> Some assertions, RTE_ASSERT(), were added. Most importantly to assert that
> the return values of the mempool drivers' enqueue and dequeue operations
> are API compliant, i.e. 0 (for success) or negative (for failure), and
> never positive.
> 
> Signed-off-by: Morten Brørup 
> ---
>  lib/mempool/rte_mempool.h | 67 ++-
>  1 file changed, 38 insertions(+), 29 deletions(-)
> 
Is there any measurable performance change with these modifications?

/Bruce


RE: [EXTERNAL] Re: [v5 1/5] vhost: skip crypto op fetch before vring init

2025-02-26 Thread Gowrishankar Muthukrishnan
Hi Maxime,
> >
> > +   if (unlikely(vq == NULL)) {
> > +   VC_LOG_ERR("Invalid virtqueue %u", qid);
> > +   return 0;
> > +   }
> > +
> > +   if (unlikely(vq->avail == NULL)) {
> > +   VC_LOG_DBG("Virtqueue ring not yet initialized %u", qid);
> > +   return 0;
> > +   }
> > +
> 
> I think the proper fix to:
> 1. acquire the access_lock as reader
> 2. check whether vq is enabled
> 3. acquire the iotlb lock as a reader
> 4. check vq's access_ok
> 

Ack. Except for vq->enabled, next version of this patch would carry these 
checks.
Current virtio crypto for vhost_user does not request SET_VRING_ENABLE, so this 
flag would not be set.

Thanks,
Gowrishankar

> You can have a look at rte_vhost_dequeue_burst for example.
> 
> > avail_idx = *((volatile uint16_t *)&vq->avail->idx);
> > start_idx = vq->last_used_idx;
> > count = avail_idx - start_idx;



[v6 4/5] vhost: support asymmetric RSA crypto ops

2025-02-26 Thread Gowrishankar Muthukrishnan
Support asymmetric RSA crypto operations in vhost-user.

Signed-off-by: Gowrishankar Muthukrishnan 
Acked-by: Akhil Goyal 
---
 doc/guides/rel_notes/release_25_03.rst |   3 +
 lib/vhost/vhost_crypto.c   | 486 +++--
 lib/vhost/virtio_crypto.h  |  67 
 3 files changed, 521 insertions(+), 35 deletions(-)

diff --git a/doc/guides/rel_notes/release_25_03.rst 
b/doc/guides/rel_notes/release_25_03.rst
index 8867a4bd74..087a407337 100644
--- a/doc/guides/rel_notes/release_25_03.rst
+++ b/doc/guides/rel_notes/release_25_03.rst
@@ -151,6 +151,9 @@ New Features
 
   See the :doc:`../compressdevs/zsda` guide for more details on the new driver.
 
+* **Updated vhost library.**
+
+  Updated vhost library to support RSA crypto operations.
 
 Removed Items
 -
diff --git a/lib/vhost/vhost_crypto.c b/lib/vhost/vhost_crypto.c
index c014c85797..c55db0108b 100644
--- a/lib/vhost/vhost_crypto.c
+++ b/lib/vhost/vhost_crypto.c
@@ -55,6 +55,14 @@ RTE_LOG_REGISTER_SUFFIX(vhost_crypto_logtype, crypto, INFO);
  */
 #define vhost_crypto_desc vring_desc
 
+struct vhost_crypto_session {
+   union {
+   struct rte_cryptodev_asym_session *asym;
+   struct rte_cryptodev_sym_session *sym;
+   };
+   enum rte_crypto_op_type type;
+};
+
 static int
 cipher_algo_transform(uint32_t virtio_cipher_algo,
enum rte_crypto_cipher_algorithm *algo)
@@ -207,8 +215,10 @@ struct __rte_cache_aligned vhost_crypto {
 
uint64_t last_session_id;
 
-   uint64_t cache_session_id;
-   struct rte_cryptodev_sym_session *cache_session;
+   uint64_t cache_sym_session_id;
+   struct rte_cryptodev_sym_session *cache_sym_session;
+   uint64_t cache_asym_session_id;
+   struct rte_cryptodev_asym_session *cache_asym_session;
/** socket id for the device */
int socket_id;
 
@@ -335,10 +345,11 @@ transform_chain_param(struct rte_crypto_sym_xform *xforms,
 }
 
 static void
-vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
+vhost_crypto_create_sym_sess(struct vhost_crypto *vcrypto,
VhostUserCryptoSessionParam *sess_param)
 {
struct rte_crypto_sym_xform xform1 = {0}, xform2 = {0};
+   struct vhost_crypto_session *vhost_session;
struct rte_cryptodev_sym_session *session;
int ret;
 
@@ -385,42 +396,277 @@ vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
return;
}
 
-   /* insert hash to map */
-   if (rte_hash_add_key_data(vcrypto->session_map,
-   &vcrypto->last_session_id, session) < 0) {
+   vhost_session = rte_zmalloc(NULL, sizeof(*vhost_session), 0);
+   if (vhost_session == NULL) {
+   VC_LOG_ERR("Failed to alloc session memory");
+   goto error_exit;
+   }
+
+   vhost_session->type = RTE_CRYPTO_OP_TYPE_SYMMETRIC;
+   vhost_session->sym = session;
+
+   /* insert session to map */
+   if ((rte_hash_add_key_data(vcrypto->session_map,
+   &vcrypto->last_session_id, vhost_session) < 0)) {
VC_LOG_ERR("Failed to insert session to hash table");
+   goto error_exit;
+   }
+
+   VC_LOG_INFO("Session %"PRIu64" created for vdev %i.",
+   vcrypto->last_session_id, vcrypto->dev->vid);
+
+   sess_param->session_id = vcrypto->last_session_id;
+   vcrypto->last_session_id++;
+   return;
+
+error_exit:
+   if (rte_cryptodev_sym_session_free(vcrypto->cid, session) < 0)
+   VC_LOG_ERR("Failed to free session");
+
+   sess_param->session_id = -VIRTIO_CRYPTO_ERR;
+   rte_free(vhost_session);
+}
+
+static int
+tlv_decode(uint8_t *tlv, uint8_t type, uint8_t **data, size_t *data_len)
+{
+   size_t tlen = -EINVAL, len;
+
+   if (tlv[0] != type)
+   return -EINVAL;
 
-   if (rte_cryptodev_sym_session_free(vcrypto->cid, session) < 0)
-   VC_LOG_ERR("Failed to free session");
+   if (tlv[1] == 0x82) {
+   len = (tlv[2] << 8) | tlv[3];
+   *data = &tlv[4];
+   tlen = len + 4;
+   } else if (tlv[1] == 0x81) {
+   len = tlv[2];
+   *data = &tlv[3];
+   tlen = len + 3;
+   } else {
+   len = tlv[1];
+   *data = &tlv[2];
+   tlen = len + 2;
+   }
+
+   *data_len = len;
+   return tlen;
+}
+
+static int
+virtio_crypto_asym_rsa_der_to_xform(uint8_t *der, size_t der_len,
+   struct rte_crypto_asym_xform *xform)
+{
+   uint8_t *n = NULL, *e = NULL, *d = NULL, *p = NULL, *q = NULL, *dp = 
NULL,
+   *dq = NULL, *qinv = NULL, *v = NULL, *tlv;
+   size_t nlen, elen, dlen, plen, qlen, dplen, dqlen, qinvlen, vlen;
+   int len;
+
+   RTE_SET_USED(der_len);
+
+   if (der[0] != 0x30)
+   return -EINVAL;
+
+   if (der[1] == 0x82)
+   tlv = &der[4];
+  

[v6 5/5] examples/vhost_crypto: support asymmetric crypto

2025-02-26 Thread Gowrishankar Muthukrishnan
Support asymmetric crypto operations.

Signed-off-by: Gowrishankar Muthukrishnan 
Acked-by: Akhil Goyal 
---
 doc/guides/sample_app_ug/vhost_crypto.rst |  5 +++
 examples/vhost_crypto/main.c  | 50 +--
 2 files changed, 43 insertions(+), 12 deletions(-)

diff --git a/doc/guides/sample_app_ug/vhost_crypto.rst 
b/doc/guides/sample_app_ug/vhost_crypto.rst
index 7ae7addac4..b00f2bf3ae 100644
--- a/doc/guides/sample_app_ug/vhost_crypto.rst
+++ b/doc/guides/sample_app_ug/vhost_crypto.rst
@@ -33,6 +33,7 @@ Start the vhost_crypto example
--socket-file lcore,PATH
[--zero-copy]
[--guest-polling]
+   [--asymmetric-crypto]
 
 where,
 
@@ -54,6 +55,10 @@ where,
   guest works in polling mode, thus will NOT notify the guest completion of
   processing.
 
+* asymmetric-crypto: the presence of this item means the application
+  can handle the asymmetric crypto requests. When this option is used,
+  symmetric crypto requests can not be handled by the application.
+
 The application requires that crypto devices capable of performing
 the specified crypto operation are available on application initialization.
 This means that HW crypto device/s must be bound to a DPDK driver or
diff --git a/examples/vhost_crypto/main.c b/examples/vhost_crypto/main.c
index b1fe4120b9..8bdfc40c4b 100644
--- a/examples/vhost_crypto/main.c
+++ b/examples/vhost_crypto/main.c
@@ -59,6 +59,7 @@ struct vhost_crypto_options {
uint32_t nb_los;
uint32_t zero_copy;
uint32_t guest_polling;
+   bool asymmetric_crypto;
 } options;
 
 enum {
@@ -70,6 +71,8 @@ enum {
OPT_ZERO_COPY_NUM,
 #define OPT_POLLING "guest-polling"
OPT_POLLING_NUM,
+#define OPT_ASYM"asymmetric-crypto"
+   OPT_ASYM_NUM,
 };
 
 #define NB_SOCKET_FIELDS   (2)
@@ -202,9 +205,10 @@ vhost_crypto_usage(const char *prgname)
"  --%s ,SOCKET-FILE-PATH\n"
"  --%s (lcore,cdev_id,queue_id)[,(lcore,cdev_id,queue_id)]\n"
"  --%s: zero copy\n"
-   "  --%s: guest polling\n",
+   "  --%s: guest polling\n"
+   "  --%s: asymmetric crypto\n",
prgname, OPT_SOCKET_FILE, OPT_CONFIG,
-   OPT_ZERO_COPY, OPT_POLLING);
+   OPT_ZERO_COPY, OPT_POLLING, OPT_ASYM);
 }
 
 static int
@@ -223,6 +227,8 @@ vhost_crypto_parse_args(int argc, char **argv)
NULL, OPT_ZERO_COPY_NUM},
{OPT_POLLING, no_argument,
NULL, OPT_POLLING_NUM},
+   {OPT_ASYM, no_argument,
+   NULL, OPT_ASYM_NUM},
{NULL, 0, 0, 0}
};
 
@@ -262,6 +268,10 @@ vhost_crypto_parse_args(int argc, char **argv)
options.guest_polling = 1;
break;
 
+   case OPT_ASYM_NUM:
+   options.asymmetric_crypto = true;
+   break;
+
default:
vhost_crypto_usage(prgname);
return -EINVAL;
@@ -376,6 +386,7 @@ vhost_crypto_worker(void *arg)
int callfds[VIRTIO_CRYPTO_MAX_NUM_BURST_VQS];
uint32_t lcore_id = rte_lcore_id();
uint32_t burst_size = MAX_PKT_BURST;
+   enum rte_crypto_op_type cop_type;
uint32_t i, j, k;
uint32_t to_fetch, fetched;
 
@@ -383,9 +394,13 @@ vhost_crypto_worker(void *arg)
 
RTE_LOG(INFO, USER1, "Processing on Core %u started\n", lcore_id);
 
+   cop_type = RTE_CRYPTO_OP_TYPE_SYMMETRIC;
+   if (options.asymmetric_crypto)
+   cop_type = RTE_CRYPTO_OP_TYPE_ASYMMETRIC;
+
for (i = 0; i < NB_VIRTIO_QUEUES; i++) {
if (rte_crypto_op_bulk_alloc(info->cop_pool,
-   RTE_CRYPTO_OP_TYPE_SYMMETRIC, ops[i],
+   cop_type, ops[i],
burst_size) < burst_size) {
RTE_LOG(ERR, USER1, "Failed to alloc cops\n");
ret = -1;
@@ -411,12 +426,11 @@ vhost_crypto_worker(void *arg)
fetched);
if (unlikely(rte_crypto_op_bulk_alloc(
info->cop_pool,
-   RTE_CRYPTO_OP_TYPE_SYMMETRIC,
+   cop_type,
ops[j], fetched) < fetched)) {
RTE_LOG(ERR, USER1, "Failed realloc\n");
return -1;
}
-
fetched = rte_cryptodev_dequeue_burst(
info->cid, info->qid,
ops_deq[j], RTE_MIN(burst_size,
@@ -477,6 +491,7 @@ m

[v5 4/6] crypto/virtio: add vDPA backend

2025-02-26 Thread Gowrishankar Muthukrishnan
Add vDPA backend to virtio_user crypto.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 drivers/crypto/virtio/meson.build |   7 +
 drivers/crypto/virtio/virtio_cryptodev.c  |  57 +-
 drivers/crypto/virtio/virtio_cryptodev.h  |   3 +
 drivers/crypto/virtio/virtio_logs.h   |   6 +-
 drivers/crypto/virtio/virtio_pci.h|   7 +
 drivers/crypto/virtio/virtio_ring.h   |   6 -
 drivers/crypto/virtio/virtio_user/vhost.h |  90 +++
 .../crypto/virtio/virtio_user/vhost_vdpa.c| 710 +
 .../virtio/virtio_user/virtio_user_dev.c  | 749 ++
 .../virtio/virtio_user/virtio_user_dev.h  |  85 ++
 drivers/crypto/virtio/virtio_user_cryptodev.c | 575 ++
 11 files changed, 2265 insertions(+), 30 deletions(-)
 create mode 100644 drivers/crypto/virtio/virtio_user/vhost.h
 create mode 100644 drivers/crypto/virtio/virtio_user/vhost_vdpa.c
 create mode 100644 drivers/crypto/virtio/virtio_user/virtio_user_dev.c
 create mode 100644 drivers/crypto/virtio/virtio_user/virtio_user_dev.h
 create mode 100644 drivers/crypto/virtio/virtio_user_cryptodev.c

diff --git a/drivers/crypto/virtio/meson.build 
b/drivers/crypto/virtio/meson.build
index d2c3b3ad07..3763e86746 100644
--- a/drivers/crypto/virtio/meson.build
+++ b/drivers/crypto/virtio/meson.build
@@ -16,3 +16,10 @@ sources = files(
 'virtio_rxtx.c',
 'virtqueue.c',
 )
+
+if is_linux
+sources += files('virtio_user_cryptodev.c',
+'virtio_user/vhost_vdpa.c',
+'virtio_user/virtio_user_dev.c')
+deps += ['bus_vdev']
+endif
diff --git a/drivers/crypto/virtio/virtio_cryptodev.c 
b/drivers/crypto/virtio/virtio_cryptodev.c
index 92fea557ab..bc737f1e68 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.c
+++ b/drivers/crypto/virtio/virtio_cryptodev.c
@@ -544,24 +544,12 @@ virtio_crypto_init_device(struct rte_cryptodev *cryptodev,
return 0;
 }
 
-/*
- * This function is based on probe() function
- * It returns 0 on success.
- */
-static int
-crypto_virtio_create(const char *name, struct rte_pci_device *pci_dev,
-   struct rte_cryptodev_pmd_init_params *init_params)
+int
+crypto_virtio_dev_init(struct rte_cryptodev *cryptodev, uint64_t features,
+   struct rte_pci_device *pci_dev)
 {
-   struct rte_cryptodev *cryptodev;
struct virtio_crypto_hw *hw;
 
-   PMD_INIT_FUNC_TRACE();
-
-   cryptodev = rte_cryptodev_pmd_create(name, &pci_dev->device,
-   init_params);
-   if (cryptodev == NULL)
-   return -ENODEV;
-
cryptodev->driver_id = cryptodev_virtio_driver_id;
cryptodev->dev_ops = &virtio_crypto_dev_ops;
 
@@ -578,16 +566,41 @@ crypto_virtio_create(const char *name, struct 
rte_pci_device *pci_dev,
hw->dev_id = cryptodev->data->dev_id;
hw->virtio_dev_capabilities = virtio_capabilities;
 
-   VIRTIO_CRYPTO_INIT_LOG_DBG("dev %d vendorID=0x%x deviceID=0x%x",
-   cryptodev->data->dev_id, pci_dev->id.vendor_id,
-   pci_dev->id.device_id);
+   if (pci_dev) {
+   /* pci device init */
+   VIRTIO_CRYPTO_INIT_LOG_DBG("dev %d vendorID=0x%x deviceID=0x%x",
+   cryptodev->data->dev_id, pci_dev->id.vendor_id,
+   pci_dev->id.device_id);
 
-   /* pci device init */
-   if (vtpci_cryptodev_init(pci_dev, hw))
+   if (vtpci_cryptodev_init(pci_dev, hw))
+   return -1;
+   }
+
+   if (virtio_crypto_init_device(cryptodev, features) < 0)
return -1;
 
-   if (virtio_crypto_init_device(cryptodev,
-   VIRTIO_CRYPTO_PMD_GUEST_FEATURES) < 0)
+   return 0;
+}
+
+/*
+ * This function is based on probe() function
+ * It returns 0 on success.
+ */
+static int
+crypto_virtio_create(const char *name, struct rte_pci_device *pci_dev,
+   struct rte_cryptodev_pmd_init_params *init_params)
+{
+   struct rte_cryptodev *cryptodev;
+
+   PMD_INIT_FUNC_TRACE();
+
+   cryptodev = rte_cryptodev_pmd_create(name, &pci_dev->device,
+   init_params);
+   if (cryptodev == NULL)
+   return -ENODEV;
+
+   if (crypto_virtio_dev_init(cryptodev, VIRTIO_CRYPTO_PMD_GUEST_FEATURES,
+   pci_dev) < 0)
return -1;
 
rte_cryptodev_pmd_probing_finish(cryptodev);
diff --git a/drivers/crypto/virtio/virtio_cryptodev.h 
b/drivers/crypto/virtio/virtio_cryptodev.h
index f8498246e2..fad73d54a8 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.h
+++ b/drivers/crypto/virtio/virtio_cryptodev.h
@@ -76,4 +76,7 @@ uint16_t virtio_crypto_pkt_rx_burst(void *tx_queue,
struct rte_crypto_op **tx_pkts,
uint16_t nb_pkts);
 
+int crypto_virtio_dev_init(struct rte_cryptodev *cryptodev, uint64_t features,
+   struct rte_pci_device *pci_dev);
+
 #endif /* _VI

[v5 6/6] test/crypto: add tests for virtio user PMD

2025-02-26 Thread Gowrishankar Muthukrishnan
Reuse virtio_crypto tests for testing virtio_crypto_user PMD.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 app/test/test_cryptodev.c  |  7 +++
 app/test/test_cryptodev.h  |  1 +
 app/test/test_cryptodev_asym.c | 15 +++
 3 files changed, 23 insertions(+)

diff --git a/app/test/test_cryptodev.c b/app/test/test_cryptodev.c
index 441ecc6ad5..60aacdc155 100644
--- a/app/test/test_cryptodev.c
+++ b/app/test/test_cryptodev.c
@@ -19737,6 +19737,12 @@ test_cryptodev_virtio(void)
return run_cryptodev_testsuite(RTE_STR(CRYPTODEV_NAME_VIRTIO_PMD));
 }
 
+static int
+test_cryptodev_virtio_user(void)
+{
+   return run_cryptodev_testsuite(RTE_STR(CRYPTODEV_NAME_VIRTIO_USER_PMD));
+}
+
 static int
 test_cryptodev_aesni_mb(void)
 {
@@ -20074,6 +20080,7 @@ REGISTER_DRIVER_TEST(cryptodev_dpaa_sec_autotest, 
test_cryptodev_dpaa_sec);
 REGISTER_DRIVER_TEST(cryptodev_ccp_autotest, test_cryptodev_ccp);
 REGISTER_DRIVER_TEST(cryptodev_uadk_autotest, test_cryptodev_uadk);
 REGISTER_DRIVER_TEST(cryptodev_virtio_autotest, test_cryptodev_virtio);
+REGISTER_DRIVER_TEST(cryptodev_virtio_user_autotest, 
test_cryptodev_virtio_user);
 REGISTER_DRIVER_TEST(cryptodev_octeontx_autotest, test_cryptodev_octeontx);
 REGISTER_DRIVER_TEST(cryptodev_caam_jr_autotest, test_cryptodev_caam_jr);
 REGISTER_DRIVER_TEST(cryptodev_nitrox_autotest, test_cryptodev_nitrox);
diff --git a/app/test/test_cryptodev.h b/app/test/test_cryptodev.h
index bb54a33d62..f6c7478f19 100644
--- a/app/test/test_cryptodev.h
+++ b/app/test/test_cryptodev.h
@@ -64,6 +64,7 @@
 #define CRYPTODEV_NAME_MVSAM_PMD   crypto_mvsam
 #define CRYPTODEV_NAME_CCP_PMD crypto_ccp
 #define CRYPTODEV_NAME_VIRTIO_PMD  crypto_virtio
+#define CRYPTODEV_NAME_VIRTIO_USER_PMD crypto_virtio_user
 #define CRYPTODEV_NAME_OCTEONTX_SYM_PMDcrypto_octeontx
 #define CRYPTODEV_NAME_CAAM_JR_PMD crypto_caam_jr
 #define CRYPTODEV_NAME_NITROX_PMD  crypto_nitrox_sym
diff --git a/app/test/test_cryptodev_asym.c b/app/test/test_cryptodev_asym.c
index ac47be724f..a98e3dc824 100644
--- a/app/test/test_cryptodev_asym.c
+++ b/app/test/test_cryptodev_asym.c
@@ -4118,9 +4118,24 @@ test_cryptodev_virtio_asym(void)
return unit_test_suite_runner(&cryptodev_virtio_asym_testsuite);
 }
 
+static int
+test_cryptodev_virtio_user_asym(void)
+{
+   gbl_driver_id = rte_cryptodev_driver_id_get(
+   RTE_STR(CRYPTODEV_NAME_VIRTIO_USER_PMD));
+   if (gbl_driver_id == -1) {
+   RTE_LOG(ERR, USER1, "virtio user PMD must be loaded.\n");
+   return TEST_FAILED;
+   }
+
+   /* Use test suite registered for crypto_virtio_user PMD */
+   return unit_test_suite_runner(&cryptodev_virtio_asym_testsuite);
+}
+
 REGISTER_DRIVER_TEST(cryptodev_openssl_asym_autotest, 
test_cryptodev_openssl_asym);
 REGISTER_DRIVER_TEST(cryptodev_qat_asym_autotest, test_cryptodev_qat_asym);
 REGISTER_DRIVER_TEST(cryptodev_octeontx_asym_autotest, 
test_cryptodev_octeontx_asym);
 REGISTER_DRIVER_TEST(cryptodev_cn9k_asym_autotest, test_cryptodev_cn9k_asym);
 REGISTER_DRIVER_TEST(cryptodev_cn10k_asym_autotest, test_cryptodev_cn10k_asym);
 REGISTER_DRIVER_TEST(cryptodev_virtio_asym_autotest, 
test_cryptodev_virtio_asym);
+REGISTER_DRIVER_TEST(cryptodev_virtio_user_asym_autotest, 
test_cryptodev_virtio_user_asym);
-- 
2.25.1



[v5 5/6] test/crypto: add asymmetric tests for virtio PMD

2025-02-26 Thread Gowrishankar Muthukrishnan
Add asymmetric tests for Virtio PMD.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 app/test/test_cryptodev_asym.c | 28 
 1 file changed, 28 insertions(+)

diff --git a/app/test/test_cryptodev_asym.c b/app/test/test_cryptodev_asym.c
index 9b5f3c545e..ac47be724f 100644
--- a/app/test/test_cryptodev_asym.c
+++ b/app/test/test_cryptodev_asym.c
@@ -4023,6 +4023,19 @@ static struct unit_test_suite 
cryptodev_octeontx_asym_testsuite  = {
}
 };
 
+static struct unit_test_suite cryptodev_virtio_asym_testsuite  = {
+   .suite_name = "Crypto Device VIRTIO ASYM Unit Test Suite",
+   .setup = testsuite_setup,
+   .teardown = testsuite_teardown,
+   .unit_test_cases = {
+   TEST_CASE_ST(ut_setup_asym, ut_teardown_asym, test_capability),
+   TEST_CASE_ST(ut_setup_asym, ut_teardown_asym,
+   test_rsa_sign_verify_crt),
+   TEST_CASE_ST(ut_setup_asym, ut_teardown_asym, 
test_rsa_enc_dec_crt),
+   TEST_CASES_END() /**< NULL terminate unit test array */
+   }
+};
+
 static int
 test_cryptodev_openssl_asym(void)
 {
@@ -4091,8 +4104,23 @@ test_cryptodev_cn10k_asym(void)
return unit_test_suite_runner(&cryptodev_octeontx_asym_testsuite);
 }
 
+static int
+test_cryptodev_virtio_asym(void)
+{
+   gbl_driver_id = rte_cryptodev_driver_id_get(
+   RTE_STR(CRYPTODEV_NAME_VIRTIO_PMD));
+   if (gbl_driver_id == -1) {
+   RTE_LOG(ERR, USER1, "virtio PMD must be loaded.\n");
+   return TEST_FAILED;
+   }
+
+   /* Use test suite registered for crypto_virtio PMD */
+   return unit_test_suite_runner(&cryptodev_virtio_asym_testsuite);
+}
+
 REGISTER_DRIVER_TEST(cryptodev_openssl_asym_autotest, 
test_cryptodev_openssl_asym);
 REGISTER_DRIVER_TEST(cryptodev_qat_asym_autotest, test_cryptodev_qat_asym);
 REGISTER_DRIVER_TEST(cryptodev_octeontx_asym_autotest, 
test_cryptodev_octeontx_asym);
 REGISTER_DRIVER_TEST(cryptodev_cn9k_asym_autotest, test_cryptodev_cn9k_asym);
 REGISTER_DRIVER_TEST(cryptodev_cn10k_asym_autotest, test_cryptodev_cn10k_asym);
+REGISTER_DRIVER_TEST(cryptodev_virtio_asym_autotest, 
test_cryptodev_virtio_asym);
-- 
2.25.1



[v5 2/6] crypto/virtio: refactor queue operations

2025-02-26 Thread Gowrishankar Muthukrishnan
Move existing control queue operations into a common place
that would be shared with other virtio type of devices.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 drivers/crypto/virtio/meson.build  |   1 +
 drivers/crypto/virtio/virtio_crypto_algs.h |   2 +-
 drivers/crypto/virtio/virtio_cryptodev.c   | 563 -
 drivers/crypto/virtio/virtio_cvq.c | 129 +
 drivers/crypto/virtio/virtio_cvq.h |  33 ++
 drivers/crypto/virtio/virtio_pci.h |   6 +-
 drivers/crypto/virtio/virtio_ring.h|  12 +-
 drivers/crypto/virtio/virtio_rxtx.c|  44 +-
 drivers/crypto/virtio/virtio_rxtx.h|  13 +
 drivers/crypto/virtio/virtqueue.c  | 191 ++-
 drivers/crypto/virtio/virtqueue.h  |  89 +++-
 11 files changed, 692 insertions(+), 391 deletions(-)
 create mode 100644 drivers/crypto/virtio/virtio_cvq.c
 create mode 100644 drivers/crypto/virtio/virtio_cvq.h
 create mode 100644 drivers/crypto/virtio/virtio_rxtx.h

diff --git a/drivers/crypto/virtio/meson.build 
b/drivers/crypto/virtio/meson.build
index 45533c9b89..d2c3b3ad07 100644
--- a/drivers/crypto/virtio/meson.build
+++ b/drivers/crypto/virtio/meson.build
@@ -11,6 +11,7 @@ includes += include_directories('../../../lib/vhost')
 deps += 'bus_pci'
 sources = files(
 'virtio_cryptodev.c',
+'virtio_cvq.c',
 'virtio_pci.c',
 'virtio_rxtx.c',
 'virtqueue.c',
diff --git a/drivers/crypto/virtio/virtio_crypto_algs.h 
b/drivers/crypto/virtio/virtio_crypto_algs.h
index 4c44af3733..3824017ca5 100644
--- a/drivers/crypto/virtio/virtio_crypto_algs.h
+++ b/drivers/crypto/virtio/virtio_crypto_algs.h
@@ -22,7 +22,7 @@ struct virtio_crypto_session {
phys_addr_t phys_addr;
} aad;
 
-   struct virtio_crypto_op_ctrl_req ctrl;
+   struct virtio_pmd_ctrl ctrl;
 };
 
 #endif /* _VIRTIO_CRYPTO_ALGS_H_ */
diff --git a/drivers/crypto/virtio/virtio_cryptodev.c 
b/drivers/crypto/virtio/virtio_cryptodev.c
index 6a264bc24a..6bb76ff15e 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.c
+++ b/drivers/crypto/virtio/virtio_cryptodev.c
@@ -64,211 +64,6 @@ static const struct rte_cryptodev_capabilities 
virtio_capabilities[] = {
 
 uint8_t cryptodev_virtio_driver_id;
 
-#define NUM_ENTRY_SYM_CREATE_SESSION 4
-
-static int
-virtio_crypto_send_command(struct virtqueue *vq,
-   struct virtio_crypto_op_ctrl_req *ctrl, uint8_t *cipher_key,
-   uint8_t *auth_key, struct virtio_crypto_session *session)
-{
-   uint8_t idx = 0;
-   uint8_t needed = 1;
-   uint32_t head = 0;
-   uint32_t len_cipher_key = 0;
-   uint32_t len_auth_key = 0;
-   uint32_t len_ctrl_req = sizeof(struct virtio_crypto_op_ctrl_req);
-   uint32_t len_session_input = sizeof(struct virtio_crypto_session_input);
-   uint32_t len_total = 0;
-   uint32_t input_offset = 0;
-   void *virt_addr_started = NULL;
-   phys_addr_t phys_addr_started;
-   struct vring_desc *desc;
-   uint32_t desc_offset;
-   struct virtio_crypto_session_input *input;
-   int ret;
-
-   PMD_INIT_FUNC_TRACE();
-
-   if (session == NULL) {
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("session is NULL.");
-   return -EINVAL;
-   }
-   /* cipher only is supported, it is available if auth_key is NULL */
-   if (!cipher_key) {
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("cipher key is NULL.");
-   return -EINVAL;
-   }
-
-   head = vq->vq_desc_head_idx;
-   VIRTIO_CRYPTO_INIT_LOG_DBG("vq->vq_desc_head_idx = %d, vq = %p",
-   head, vq);
-
-   if (vq->vq_free_cnt < needed) {
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("Not enough entry");
-   return -ENOSPC;
-   }
-
-   /* calculate the length of cipher key */
-   if (session->ctrl.header.algo == VIRTIO_CRYPTO_SERVICE_CIPHER) {
-   switch (ctrl->u.sym_create_session.op_type) {
-   case VIRTIO_CRYPTO_SYM_OP_CIPHER:
-   len_cipher_key = 
ctrl->u.sym_create_session.u.cipher.para.keylen;
-   break;
-   case VIRTIO_CRYPTO_SYM_OP_ALGORITHM_CHAINING:
-   len_cipher_key =
-   
ctrl->u.sym_create_session.u.chain.para.cipher_param.keylen;
-   break;
-   default:
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("invalid op type");
-   return -EINVAL;
-   }
-   } else if (session->ctrl.header.algo == VIRTIO_CRYPTO_AKCIPHER_RSA) {
-   len_cipher_key = ctrl->u.akcipher_create_session.para.keylen;
-   } else {
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("Invalid crypto service for 
cipher key");
-   return -EINVAL;
-   }
-
-   /* calculate the length of auth key */
-   if (auth_key) {
-   len_auth_key =
-   ctrl->u.sym_

[v6 2/5] vhost: update vhost_user crypto session parameters

2025-02-26 Thread Gowrishankar Muthukrishnan
As per requirements on vhost_user spec, session id should be
located at the end of session parameter.

Update VhostUserCryptoSessionParam structure to support newer QEMU
versions (v9). Due to additional parameters added in QEMU,
received payload from QEMU would be larger than existing payload.
Hence, it would break parsing vhost_user messages.

This patch addresses both of the above problems.

Signed-off-by: Gowrishankar Muthukrishnan 
Acked-by: Akhil Goyal 
---
 lib/vhost/vhost_crypto.c | 12 ++--
 lib/vhost/vhost_user.h   | 33 +
 2 files changed, 35 insertions(+), 10 deletions(-)

diff --git a/lib/vhost/vhost_crypto.c b/lib/vhost/vhost_crypto.c
index d3d13eff07..c014c85797 100644
--- a/lib/vhost/vhost_crypto.c
+++ b/lib/vhost/vhost_crypto.c
@@ -238,7 +238,7 @@ struct vhost_crypto_data_req {
 
 static int
 transform_cipher_param(struct rte_crypto_sym_xform *xform,
-   VhostUserCryptoSessionParam *param)
+   VhostUserCryptoSymSessionParam *param)
 {
int ret;
 
@@ -274,7 +274,7 @@ transform_cipher_param(struct rte_crypto_sym_xform *xform,
 
 static int
 transform_chain_param(struct rte_crypto_sym_xform *xforms,
-   VhostUserCryptoSessionParam *param)
+   VhostUserCryptoSymSessionParam *param)
 {
struct rte_crypto_sym_xform *xform_cipher, *xform_auth;
int ret;
@@ -342,10 +342,10 @@ vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
struct rte_cryptodev_sym_session *session;
int ret;
 
-   switch (sess_param->op_type) {
+   switch (sess_param->u.sym_sess.op_type) {
case VIRTIO_CRYPTO_SYM_OP_NONE:
case VIRTIO_CRYPTO_SYM_OP_CIPHER:
-   ret = transform_cipher_param(&xform1, sess_param);
+   ret = transform_cipher_param(&xform1, &sess_param->u.sym_sess);
if (unlikely(ret)) {
VC_LOG_ERR("Error transform session msg (%i)", ret);
sess_param->session_id = ret;
@@ -353,7 +353,7 @@ vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
}
break;
case VIRTIO_CRYPTO_SYM_OP_ALGORITHM_CHAINING:
-   if (unlikely(sess_param->hash_mode !=
+   if (unlikely(sess_param->u.sym_sess.hash_mode !=
VIRTIO_CRYPTO_SYM_HASH_MODE_AUTH)) {
sess_param->session_id = -VIRTIO_CRYPTO_NOTSUPP;
VC_LOG_ERR("Error transform session message (%i)",
@@ -363,7 +363,7 @@ vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
 
xform1.next = &xform2;
 
-   ret = transform_chain_param(&xform1, sess_param);
+   ret = transform_chain_param(&xform1, &sess_param->u.sym_sess);
if (unlikely(ret)) {
VC_LOG_ERR("Error transform session message (%i)", ret);
sess_param->session_id = ret;
diff --git a/lib/vhost/vhost_user.h b/lib/vhost/vhost_user.h
index 9a905ee5f4..ef486545ba 100644
--- a/lib/vhost/vhost_user.h
+++ b/lib/vhost/vhost_user.h
@@ -99,11 +99,10 @@ typedef struct VhostUserLog {
 /* Comply with Cryptodev-Linux */
 #define VHOST_USER_CRYPTO_MAX_HMAC_KEY_LENGTH  512
 #define VHOST_USER_CRYPTO_MAX_CIPHER_KEY_LENGTH64
+#define VHOST_USER_CRYPTO_MAX_KEY_LENGTH   1024
 
 /* Same structure as vhost-user backend session info */
-typedef struct VhostUserCryptoSessionParam {
-   int64_t session_id;
-   uint32_t op_code;
+typedef struct VhostUserCryptoSymSessionParam {
uint32_t cipher_algo;
uint32_t cipher_key_len;
uint32_t hash_algo;
@@ -114,10 +113,36 @@ typedef struct VhostUserCryptoSessionParam {
uint8_t dir;
uint8_t hash_mode;
uint8_t chaining_dir;
-   uint8_t *ciphe_key;
+   uint8_t *cipher_key;
uint8_t *auth_key;
uint8_t cipher_key_buf[VHOST_USER_CRYPTO_MAX_CIPHER_KEY_LENGTH];
uint8_t auth_key_buf[VHOST_USER_CRYPTO_MAX_HMAC_KEY_LENGTH];
+} VhostUserCryptoSymSessionParam;
+
+
+typedef struct VhostUserCryptoAsymRsaParam {
+   uint32_t padding_algo;
+   uint32_t hash_algo;
+} VhostUserCryptoAsymRsaParam;
+
+typedef struct VhostUserCryptoAsymSessionParam {
+   uint32_t algo;
+   uint32_t key_type;
+   uint32_t key_len;
+   uint8_t *key;
+   union {
+   VhostUserCryptoAsymRsaParam rsa;
+   } u;
+   uint8_t key_buf[VHOST_USER_CRYPTO_MAX_KEY_LENGTH];
+} VhostUserCryptoAsymSessionParam;
+
+typedef struct VhostUserCryptoSessionParam {
+   uint32_t op_code;
+   union {
+   VhostUserCryptoSymSessionParam sym_sess;
+   VhostUserCryptoAsymSessionParam asym_sess;
+   } u;
+   int64_t session_id;
 } VhostUserCryptoSessionParam;
 
 typedef struct VhostUserVringArea {
-- 
2.25.1



[v6 1/5] vhost: skip crypto op fetch before vring init

2025-02-26 Thread Gowrishankar Muthukrishnan
Until virtio avail ring is initialized (by VHOST_USER_SET_VRING_ADDR),
worker thread should not try to fetch crypto op, which would lead to
memory fault.

Fixes: 939066d96563 ("vhost/crypto: add public function implementation")
Cc: sta...@dpdk.org

Signed-off-by: Gowrishankar Muthukrishnan 
Acked-by: Akhil Goyal 
---
v6:
 - added lock checks.
---
 lib/vhost/vhost_crypto.c | 20 
 1 file changed, 20 insertions(+)

diff --git a/lib/vhost/vhost_crypto.c b/lib/vhost/vhost_crypto.c
index 3dc41a3bd5..d3d13eff07 100644
--- a/lib/vhost/vhost_crypto.c
+++ b/lib/vhost/vhost_crypto.c
@@ -8,6 +8,7 @@
 #include 
 #include 
 
+#include "iotlb.h"
 #include "rte_vhost_crypto.h"
 #include "vhost.h"
 #include "vhost_user.h"
@@ -1580,7 +1581,26 @@ rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
 
vq = dev->virtqueue[qid];
 
+   if (unlikely(vq == NULL)) {
+   VC_LOG_ERR("Invalid virtqueue %u", qid);
+   return 0;
+   }
+
+   if (unlikely(rte_rwlock_read_trylock(&vq->access_lock) != 0))
+   return 0;
+
+   vhost_user_iotlb_rd_lock(vq);
+   if (unlikely(!vq->access_ok)) {
+   VC_LOG_DBG("Virtqueue %u vrings not yet initialized", qid);
+   vhost_user_iotlb_rd_unlock(vq);
+   rte_rwlock_read_unlock(&vq->access_lock);
+   return 0;
+   }
+
avail_idx = *((volatile uint16_t *)&vq->avail->idx);
+   vhost_user_iotlb_rd_unlock(vq);
+   rte_rwlock_read_unlock(&vq->access_lock);
+
start_idx = vq->last_used_idx;
count = avail_idx - start_idx;
count = RTE_MIN(count, VHOST_CRYPTO_MAX_BURST_SIZE);
-- 
2.25.1



[v6 3/5] examples/vhost_crypto: fix user callbacks

2025-02-26 Thread Gowrishankar Muthukrishnan
In order to handle new vhost user connection, use new_connection
and destroy_connection callbacks.

Fixes: f5188211c721 ("examples/vhost_crypto: add sample application")
Cc: sta...@dpdk.org

Signed-off-by: Gowrishankar Muthukrishnan 
Acked-by: Akhil Goyal 
---
 examples/vhost_crypto/main.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/examples/vhost_crypto/main.c b/examples/vhost_crypto/main.c
index 558c09a60f..b1fe4120b9 100644
--- a/examples/vhost_crypto/main.c
+++ b/examples/vhost_crypto/main.c
@@ -362,8 +362,8 @@ destroy_device(int vid)
 }
 
 static const struct rte_vhost_device_ops virtio_crypto_device_ops = {
-   .new_device =  new_device,
-   .destroy_device = destroy_device,
+   .new_connection =  new_device,
+   .destroy_connection = destroy_device,
 };
 
 static int
-- 
2.25.1



[v6 0/5] vhost: add RSA support

2025-02-26 Thread Gowrishankar Muthukrishnan
This patch series supports asymmetric RSA in vhost crypto library.
It also includes changes to improve vhost crypto library:
 * support newer QEMU versions.
 * fix broken vhost_crypto example application.
 * stabilize crypto fastpath operations.

Gowrishankar Muthukrishnan (5):
  vhost: skip crypto op fetch before vring init
  vhost: update vhost_user crypto session parameters
  examples/vhost_crypto: fix user callbacks
  vhost: support asymmetric RSA crypto ops
  examples/vhost_crypto: support asymmetric crypto

 doc/guides/rel_notes/release_25_03.rst|   3 +
 doc/guides/sample_app_ug/vhost_crypto.rst |   5 +
 examples/vhost_crypto/main.c  |  54 ++-
 lib/vhost/vhost_crypto.c  | 518 --
 lib/vhost/vhost_user.h|  33 +-
 lib/vhost/virtio_crypto.h |  67 +++
 6 files changed, 621 insertions(+), 59 deletions(-)

-- 
2.25.1



[v5 0/6] crypto/virtio: enhancements for RSA and vDPA

2025-02-26 Thread Gowrishankar Muthukrishnan
This patch series enhances virtio crypto PMD to:
 * support RSA
 * support packed virtio ring
 * support vDPA backend

Depends-on: patch-152109 ("vhost: support asymmetric RSA crypto ops")

Gowrishankar Muthukrishnan (6):
  crypto/virtio: add asymmetric RSA support
  crypto/virtio: refactor queue operations
  crypto/virtio: add packed ring support
  crypto/virtio: add vDPA backend
  test/crypto: add asymmetric tests for virtio PMD
  test/crypto: add tests for virtio user PMD

 app/test/test_cryptodev.c |7 +
 app/test/test_cryptodev.h |1 +
 app/test/test_cryptodev_asym.c|   43 +
 doc/guides/cryptodevs/features/virtio.ini |4 +
 doc/guides/rel_notes/release_25_03.rst|4 +
 drivers/crypto/virtio/meson.build |8 +
 drivers/crypto/virtio/virtio_crypto_algs.h|2 +-
 .../virtio/virtio_crypto_capabilities.h   |   19 +
 drivers/crypto/virtio/virtio_cryptodev.c  | 1060 +++--
 drivers/crypto/virtio/virtio_cryptodev.h  |   18 +-
 drivers/crypto/virtio/virtio_cvq.c|  228 
 drivers/crypto/virtio/virtio_cvq.h|   33 +
 drivers/crypto/virtio/virtio_logs.h   |6 +-
 drivers/crypto/virtio/virtio_pci.h|   38 +-
 drivers/crypto/virtio/virtio_ring.h   |   65 +-
 drivers/crypto/virtio/virtio_rxtx.c   |  725 ++-
 drivers/crypto/virtio/virtio_rxtx.h   |   13 +
 drivers/crypto/virtio/virtio_user/vhost.h |   90 ++
 .../crypto/virtio/virtio_user/vhost_vdpa.c|  710 +++
 .../virtio/virtio_user/virtio_user_dev.c  |  749 
 .../virtio/virtio_user/virtio_user_dev.h  |   85 ++
 drivers/crypto/virtio/virtio_user_cryptodev.c |  575 +
 drivers/crypto/virtio/virtqueue.c |  229 +++-
 drivers/crypto/virtio/virtqueue.h |  221 +++-
 lib/cryptodev/cryptodev_pmd.h |6 +
 25 files changed, 4447 insertions(+), 492 deletions(-)
 create mode 100644 drivers/crypto/virtio/virtio_cvq.c
 create mode 100644 drivers/crypto/virtio/virtio_cvq.h
 create mode 100644 drivers/crypto/virtio/virtio_rxtx.h
 create mode 100644 drivers/crypto/virtio/virtio_user/vhost.h
 create mode 100644 drivers/crypto/virtio/virtio_user/vhost_vdpa.c
 create mode 100644 drivers/crypto/virtio/virtio_user/virtio_user_dev.c
 create mode 100644 drivers/crypto/virtio/virtio_user/virtio_user_dev.h
 create mode 100644 drivers/crypto/virtio/virtio_user_cryptodev.c

-- 
2.25.1



[v5 1/6] crypto/virtio: add asymmetric RSA support

2025-02-26 Thread Gowrishankar Muthukrishnan
Asymmetric RSA operations (SIGN, VERIFY, ENCRYPT and DECRYPT) are
supported in virtio PMD.

Signed-off-by: Gowrishankar Muthukrishnan 
---
Depends-on: patch-152109 ("vhost: support asymmetric RSA crypto ops")

v5:
 - fixed asym enqueue xmit func for correct head idx.
 - release notes updated.
---
 doc/guides/cryptodevs/features/virtio.ini |   4 +
 doc/guides/rel_notes/release_25_03.rst|   4 +
 .../virtio/virtio_crypto_capabilities.h   |  19 +
 drivers/crypto/virtio/virtio_cryptodev.c  | 347 +++---
 drivers/crypto/virtio/virtio_cryptodev.h  |   2 +
 drivers/crypto/virtio/virtio_rxtx.c   | 247 +++--
 lib/cryptodev/cryptodev_pmd.h |   6 +
 7 files changed, 551 insertions(+), 78 deletions(-)

diff --git a/doc/guides/cryptodevs/features/virtio.ini 
b/doc/guides/cryptodevs/features/virtio.ini
index 8f22b6210b..f720def6b4 100644
--- a/doc/guides/cryptodevs/features/virtio.ini
+++ b/doc/guides/cryptodevs/features/virtio.ini
@@ -4,8 +4,11 @@
 ;
 [Features]
 Symmetric crypto   = Y
+Asymmetric crypto  = Y
 Sym operation chaining = Y
 OOP LB  In LB  Out = Y
+RSA PRIV OP KEY EXP= Y
+RSA PRIV OP KEY QT = Y
 
 ;
 ; Supported crypto algorithms of the 'virtio' crypto driver.
@@ -30,6 +33,7 @@ SHA1 HMAC  = Y
 ; Supported Asymmetric algorithms of the 'virtio' crypto driver.
 ;
 [Asymmetric]
+RSA = Y
 
 ;
 ; Supported Operating systems of the 'virtio' crypto driver.
diff --git a/doc/guides/rel_notes/release_25_03.rst 
b/doc/guides/rel_notes/release_25_03.rst
index 087a407337..5199cb6f31 100644
--- a/doc/guides/rel_notes/release_25_03.rst
+++ b/doc/guides/rel_notes/release_25_03.rst
@@ -155,6 +155,10 @@ New Features
 
   Updated vhost library to support RSA crypto operations.
 
+* **Updated virtio crypto driver.**
+
+  Updated virtio crypto driver to support RSA crypto operations.
+
 Removed Items
 -
 
diff --git a/drivers/crypto/virtio/virtio_crypto_capabilities.h 
b/drivers/crypto/virtio/virtio_crypto_capabilities.h
index 03c30deefd..1b26ff6720 100644
--- a/drivers/crypto/virtio/virtio_crypto_capabilities.h
+++ b/drivers/crypto/virtio/virtio_crypto_capabilities.h
@@ -48,4 +48,23 @@
}, }\
}
 
+#define VIRTIO_ASYM_CAPABILITIES   \
+   {   /* RSA */   \
+   .op = RTE_CRYPTO_OP_TYPE_ASYMMETRIC,\
+   {.asym = {  \
+   .xform_capa = { \
+   .xform_type = RTE_CRYPTO_ASYM_XFORM_RSA,\
+   .op_types = ((1 << RTE_CRYPTO_ASYM_OP_SIGN) |   \
+   (1 << RTE_CRYPTO_ASYM_OP_VERIFY) |  \
+   (1 << RTE_CRYPTO_ASYM_OP_ENCRYPT) | \
+   (1 << RTE_CRYPTO_ASYM_OP_DECRYPT)), \
+   {.modlen = {\
+   .min = 1,   \
+   .max = 1024,\
+   .increment = 1  \
+   }, }\
+   }   \
+}, }   \
+   }
+
 #endif /* _VIRTIO_CRYPTO_CAPABILITIES_H_ */
diff --git a/drivers/crypto/virtio/virtio_cryptodev.c 
b/drivers/crypto/virtio/virtio_cryptodev.c
index 793f50059f..6a264bc24a 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.c
+++ b/drivers/crypto/virtio/virtio_cryptodev.c
@@ -41,6 +41,11 @@ static void virtio_crypto_sym_clear_session(struct 
rte_cryptodev *dev,
 static int virtio_crypto_sym_configure_session(struct rte_cryptodev *dev,
struct rte_crypto_sym_xform *xform,
struct rte_cryptodev_sym_session *session);
+static void virtio_crypto_asym_clear_session(struct rte_cryptodev *dev,
+   struct rte_cryptodev_asym_session *sess);
+static int virtio_crypto_asym_configure_session(struct rte_cryptodev *dev,
+   struct rte_crypto_asym_xform *xform,
+   struct rte_cryptodev_asym_session *session);
 
 /*
  * The set of PCI devices this driver supports
@@ -53,6 +58,7 @@ static const struct rte_pci_id pci_id_virtio_crypto_map[] = {
 
 static const struct rte_cryptodev_capabilities virtio_capabilities[] = {
VIRTIO_SYM_CAPABILITIES,
+   VIRTIO_ASYM_CAPABILITIES,
RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST()
 };
 
@@ -103,22 +109,24 @@ virtio_crypto_send_command(struct virtqueue *vq,
}
 
/* calculate the length of cipher key */
-   if (cipher_key) {
+   if (session->ctrl.header.algo == VIRT

[PATCH v11 1/2] net/af_xdp: Fix use after free in af_xdp_tx_zc

2025-02-26 Thread Ariel Otilibili
tx_bytes is computed after both legs are tested. This might
produce a use after memory free.

The computation is now moved into each leg.

Bugzilla ID: 1440
Fixes: d8a210774e1d ("net/af_xdp: support unaligned umem chunks")
Signed-off-by: Ariel Otilibili 
Acked-by: Stephen Hemminger 
---
 .mailmap| 2 +-
 drivers/net/af_xdp/rte_eth_af_xdp.c | 4 ++--
 2 files changed, 3 insertions(+), 3 deletions(-)

diff --git a/.mailmap b/.mailmap
index a03d3cfb591b..ea68d6180ccc 100644
--- a/.mailmap
+++ b/.mailmap
@@ -135,7 +135,7 @@ Anupam Kapoor 
 Apeksha Gupta 
 Archana Muniganti  
 Archit Pandey 
-Ariel Otilibili  
+Ariel Otilibili  
 Arkadiusz Kubalewski 
 Arkadiusz Kusztal 
 Arnaud Fiorini 
diff --git a/drivers/net/af_xdp/rte_eth_af_xdp.c 
b/drivers/net/af_xdp/rte_eth_af_xdp.c
index 814398ba4b44..092bcb73aa0a 100644
--- a/drivers/net/af_xdp/rte_eth_af_xdp.c
+++ b/drivers/net/af_xdp/rte_eth_af_xdp.c
@@ -574,6 +574,7 @@ af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, uint16_t 
nb_pkts)
umem->mb_pool->header_size;
offset = offset << XSK_UNALIGNED_BUF_OFFSET_SHIFT;
desc->addr = addr | offset;
+   tx_bytes += desc->len;
count++;
} else {
struct rte_mbuf *local_mbuf =
@@ -601,11 +602,10 @@ af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, 
uint16_t nb_pkts)
desc->addr = addr | offset;
rte_memcpy(pkt, rte_pktmbuf_mtod(mbuf, void *),
desc->len);
+   tx_bytes += desc->len;
rte_pktmbuf_free(mbuf);
count++;
}
-
-   tx_bytes += mbuf->pkt_len;
}
 
 out:
-- 
2.30.2



[v5 3/6] crypto/virtio: add packed ring support

2025-02-26 Thread Gowrishankar Muthukrishnan
Add packed ring support.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 drivers/crypto/virtio/virtio_cryptodev.c | 125 +++
 drivers/crypto/virtio/virtio_cryptodev.h |  13 +-
 drivers/crypto/virtio/virtio_cvq.c   | 103 +-
 drivers/crypto/virtio/virtio_pci.h   |  25 ++
 drivers/crypto/virtio/virtio_ring.h  |  59 ++-
 drivers/crypto/virtio/virtio_rxtx.c  | 444 ++-
 drivers/crypto/virtio/virtqueue.c|  50 ++-
 drivers/crypto/virtio/virtqueue.h| 134 ++-
 8 files changed, 924 insertions(+), 29 deletions(-)

diff --git a/drivers/crypto/virtio/virtio_cryptodev.c 
b/drivers/crypto/virtio/virtio_cryptodev.c
index 6bb76ff15e..92fea557ab 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.c
+++ b/drivers/crypto/virtio/virtio_cryptodev.c
@@ -869,6 +869,125 @@ virtio_crypto_clear_session(
rte_free(ctrl);
 }
 
+static void
+virtio_crypto_clear_session_packed(
+   struct rte_cryptodev *dev,
+   struct virtio_crypto_op_ctrl_req *ctrl)
+{
+   struct virtio_crypto_hw *hw;
+   struct virtqueue *vq;
+   struct vring_packed_desc *desc;
+   uint8_t *status;
+   uint8_t needed = 1;
+   uint32_t head;
+   uint64_t malloc_phys_addr;
+   uint8_t len_inhdr = sizeof(struct virtio_crypto_inhdr);
+   uint32_t len_op_ctrl_req = sizeof(struct virtio_crypto_op_ctrl_req);
+   uint64_t session_id = ctrl->u.destroy_session.session_id;
+   uint16_t flags;
+   uint8_t nb_descs = 0;
+
+   hw = dev->data->dev_private;
+   vq = virtcrypto_cq_to_vq(hw->cvq);
+   head = vq->vq_avail_idx;
+   flags = vq->vq_packed.cached_flags;
+
+   VIRTIO_CRYPTO_SESSION_LOG_INFO("vq->vq_desc_head_idx = %d, "
+   "vq = %p", vq->vq_desc_head_idx, vq);
+
+   if (vq->vq_free_cnt < needed) {
+   VIRTIO_CRYPTO_SESSION_LOG_ERR(
+   "vq->vq_free_cnt = %d is less than %d, "
+   "not enough", vq->vq_free_cnt, needed);
+   return;
+   }
+
+   malloc_phys_addr = rte_malloc_virt2iova(ctrl);
+
+   /* status part */
+   status = &(((struct virtio_crypto_inhdr *)
+   ((uint8_t *)ctrl + len_op_ctrl_req))->status);
+   *status = VIRTIO_CRYPTO_ERR;
+
+   /* indirect desc vring part */
+   desc = vq->vq_packed.ring.desc;
+
+   /* ctrl request part */
+   desc[head].addr = malloc_phys_addr;
+   desc[head].len = len_op_ctrl_req;
+   desc[head].flags = VRING_DESC_F_NEXT | vq->vq_packed.cached_flags;
+   vq->vq_free_cnt--;
+   nb_descs++;
+   if (++vq->vq_avail_idx >= vq->vq_nentries) {
+   vq->vq_avail_idx -= vq->vq_nentries;
+   vq->vq_packed.cached_flags ^= VRING_PACKED_DESC_F_AVAIL_USED;
+   }
+
+   /* status part */
+   desc[vq->vq_avail_idx].addr = malloc_phys_addr + len_op_ctrl_req;
+   desc[vq->vq_avail_idx].len = len_inhdr;
+   desc[vq->vq_avail_idx].flags = VRING_DESC_F_WRITE;
+   vq->vq_free_cnt--;
+   nb_descs++;
+   if (++vq->vq_avail_idx >= vq->vq_nentries) {
+   vq->vq_avail_idx -= vq->vq_nentries;
+   vq->vq_packed.cached_flags ^= VRING_PACKED_DESC_F_AVAIL_USED;
+   }
+
+   virtqueue_store_flags_packed(&desc[head], VRING_DESC_F_NEXT | flags,
+   vq->hw->weak_barriers);
+
+   virtio_wmb(vq->hw->weak_barriers);
+   virtqueue_notify(vq);
+
+   /* wait for used desc in virtqueue
+* desc_is_used has a load-acquire or rte_io_rmb inside
+*/
+   rte_rmb();
+   while (!desc_is_used(&desc[head], vq)) {
+   rte_rmb();
+   usleep(100);
+   }
+
+   /* now get used descriptors */
+   vq->vq_free_cnt += nb_descs;
+   vq->vq_used_cons_idx += nb_descs;
+   if (vq->vq_used_cons_idx >= vq->vq_nentries) {
+   vq->vq_used_cons_idx -= vq->vq_nentries;
+   vq->vq_packed.used_wrap_counter ^= 1;
+   }
+
+   PMD_INIT_LOG(DEBUG, "vq->vq_free_cnt=%d "
+   "vq->vq_queue_idx=%d "
+   "vq->vq_avail_idx=%d "
+   "vq->vq_used_cons_idx=%d "
+   "vq->vq_packed.cached_flags=0x%x "
+   "vq->vq_packed.used_wrap_counter=%d",
+   vq->vq_free_cnt,
+   vq->vq_queue_index,
+   vq->vq_avail_idx,
+   vq->vq_used_cons_idx,
+   vq->vq_packed.cached_flags,
+   vq->vq_packed.used_wrap_counter);
+
+   if (*status != VIRTIO_CRYPTO_OK) {
+   VIRTIO_CRYPTO_SESSION_LOG_ERR("Close session failed "
+   "status=%"PRIu32", session_id=%"PRIu64"",
+   *status, session_id);
+   rte_free(ctrl);
+   return;
+   }
+
+   VIRTIO_CRYPTO_INIT_LOG_DBG("vq->vq_free_cnt=%d "
+  

[PATCH v11 2/2] net/af_xdp: Refactor af_xdp_tx_zc

2025-02-26 Thread Ariel Otilibili
Both legs of the loop share the same logic: the common parts are about
reserving and filling both address and length into the description.

This is moved into reserve_and_fill().

Bugzilla ID: 1440
Suggested-by: Maryam Tahhan 
Signed-off-by: Ariel Otilibili 
Acked-by: Maryam Tahhan 
---
 drivers/net/af_xdp/rte_eth_af_xdp.c | 75 -
 1 file changed, 42 insertions(+), 33 deletions(-)

diff --git a/drivers/net/af_xdp/rte_eth_af_xdp.c 
b/drivers/net/af_xdp/rte_eth_af_xdp.c
index 092bcb73aa0a..013f30dfc3bf 100644
--- a/drivers/net/af_xdp/rte_eth_af_xdp.c
+++ b/drivers/net/af_xdp/rte_eth_af_xdp.c
@@ -536,21 +536,49 @@ kick_tx(struct pkt_tx_queue *txq, struct xsk_ring_cons 
*cq)
}
 }
 
+static inline struct xdp_desc *
+reserve_and_fill(struct pkt_tx_queue *txq, struct rte_mbuf *mbuf,
+struct xsk_umem_info *umem, void **pkt_ptr)
+{
+   struct xdp_desc *desc = NULL;
+   uint64_t addr, offset;
+   uint32_t idx_tx;
+
+   if (!xsk_ring_prod__reserve(&txq->tx, 1, &idx_tx))
+   goto out;
+
+   desc = xsk_ring_prod__tx_desc(&txq->tx, idx_tx);
+   desc->len = mbuf->pkt_len;
+
+   addr = (uint64_t)mbuf - (uint64_t)umem->buffer
+   - umem->mb_pool->header_size;
+   offset = rte_pktmbuf_mtod(mbuf, uint64_t) - (uint64_t)mbuf
+   + umem->mb_pool->header_size;
+
+   if (pkt_ptr)
+   *pkt_ptr = xsk_umem__get_data(umem->buffer, addr + offset);
+
+   offset = offset << XSK_UNALIGNED_BUF_OFFSET_SHIFT;
+   desc->addr = addr | offset;
+
+out:
+   return desc;
+}
+
 #if defined(XDP_UMEM_UNALIGNED_CHUNK_FLAG)
 static uint16_t
 af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
 {
struct pkt_tx_queue *txq = queue;
struct xsk_umem_info *umem = txq->umem;
-   struct rte_mbuf *mbuf;
+   struct rte_mbuf *mbuf, *local_mbuf = NULL;
unsigned long tx_bytes = 0;
int i;
-   uint32_t idx_tx;
uint16_t count = 0;
struct xdp_desc *desc;
-   uint64_t addr, offset;
struct xsk_ring_cons *cq = &txq->pair->cq;
uint32_t free_thresh = cq->size >> 1;
+   void *pkt;
 
if (xsk_cons_nb_avail(cq, free_thresh) >= free_thresh)
pull_umem_cq(umem, XSK_RING_CONS__DEFAULT_NUM_DESCS, cq);
@@ -559,51 +587,32 @@ af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, 
uint16_t nb_pkts)
mbuf = bufs[i];
 
if (mbuf->pool == umem->mb_pool) {
-   if (!xsk_ring_prod__reserve(&txq->tx, 1, &idx_tx)) {
+ desc = reserve_and_fill(txq, mbuf, umem, NULL);
+ if (!desc) {
kick_tx(txq, cq);
-   if (!xsk_ring_prod__reserve(&txq->tx, 1,
-   &idx_tx))
+   desc = reserve_and_fill(txq, mbuf, umem, NULL);
+   if (!desc)
goto out;
}
-   desc = xsk_ring_prod__tx_desc(&txq->tx, idx_tx);
-   desc->len = mbuf->pkt_len;
-   addr = (uint64_t)mbuf - (uint64_t)umem->buffer -
-   umem->mb_pool->header_size;
-   offset = rte_pktmbuf_mtod(mbuf, uint64_t) -
-   (uint64_t)mbuf +
-   umem->mb_pool->header_size;
-   offset = offset << XSK_UNALIGNED_BUF_OFFSET_SHIFT;
-   desc->addr = addr | offset;
+
tx_bytes += desc->len;
count++;
} else {
-   struct rte_mbuf *local_mbuf =
-   rte_pktmbuf_alloc(umem->mb_pool);
-   void *pkt;
-
-   if (local_mbuf == NULL)
+   local_mbuf = rte_pktmbuf_alloc(umem->mb_pool);
+   if (!local_mbuf)
goto out;
 
-   if (!xsk_ring_prod__reserve(&txq->tx, 1, &idx_tx)) {
+   desc = reserve_and_fill(txq, local_mbuf, umem, &pkt);
+   if (!desc) {
rte_pktmbuf_free(local_mbuf);
goto out;
}
 
-   desc = xsk_ring_prod__tx_desc(&txq->tx, idx_tx);
desc->len = mbuf->pkt_len;
-
-   addr = (uint64_t)local_mbuf - (uint64_t)umem->buffer -
-   umem->mb_pool->header_size;
-   offset = rte_pktmbuf_mtod(local_mbuf, uint64_t) -
-   (uint64_t)local_mbuf +
-   umem->mb_pool->header_size;
-   pkt = xsk_umem__get_data(umem

Re: [PATCH] common/mlx5: add ConnectX-8 device ID

2025-02-26 Thread Raslan Darawsheh
Hi

From: Raslan Darawsheh 
Sent: Sunday, February 23, 2025 9:16 AM
To: NBU-Contact-Thomas Monjalon (EXTERNAL)
Cc: dev@dpdk.org
Subject: [PATCH] common/mlx5: add ConnectX-8 device ID

This adds the ConnectX-8 device id to the list of
supported Nvidia devices that run the MLX5 PMDs.

Signed-off-by: Raslan Darawsheh 

Patch applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh


Re: [PATCH v2 1/1] pcapng: warn if NULL is passed to rte_pcapng_close

2025-02-26 Thread Ariel Otilibili
Hello Patrick,

On Tue, Feb 25, 2025 at 9:42 PM Patrick Robb  wrote:

> Recheck-request: iol-marvell-Functional
>
> Putting in a recheck as I believe the fail CI reported is false.
>
>>
>>
Thanks for the heads up. This patch did supersede version 2,
https://patches.dpdk.org/project/dpdk/patch/20250224164031.587829-2-ariel.otilib...@6wind.com/

Regards,
Ariel


[DPDK/other Bug 1665] __rte_trace_mem_get causing out of bounds write

2025-02-26 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1665

Bug ID: 1665
   Summary: __rte_trace_mem_get causing out of bounds write
   Product: DPDK
   Version: 24.11
  Hardware: All
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: other
  Assignee: dev@dpdk.org
  Reporter: oleksan...@interfacemasters.com
  Target Milestone: ---

When almost out of trace memory, __rte_trace_mem_get can write out of bounds.

It happens in my case if I have trace events of sizes that are not aligned to
__RTE_TRACE_EVENT_HEADER_SZ. like 27,33 etc.

I suspect that the issue is with the incorrect bounds check in
__rte_trace_mem_get.

>uint32_t offset = trace->offset;
>if (unlikely((offset + sz) >= trace->len)) { // assume condition is false,
>and offset is not aligned
>...}
>offset = RTE_ALIGN_CEIL(offset, __RTE_TRACE_EVENT_HEADER_SZ); // after
>offset alignment offset + size might be bigger than trace->len
>void *mem = RTE_PTR_ADD(&trace->mem[0], offset); // returning memory chunk
>that is smaller than requested size


For example:
offset = 21, len = 32, size = 9 -> offset + size is smaller than len
align offset to __RTE_TRACE_EVENT_HEADER_SZ -> offset = 24
offset + size is bigger than len.

-- 
You are receiving this mail because:
You are the assignee for the bug.

Re: [PATCH] Skip vfio in the scenario of non-privileged mode

2025-02-26 Thread Stephen Hemminger
On Wed, 22 Jan 2025 16:15:03 +0800
Yang Ming  wrote:

> On 2025/1/18 00:47, Stephen Hemminger wrote:
> > Caution: This is an external email. Please be very careful when clicking 
> > links or opening attachments. See http://nok.it/nsb for additional 
> > information.
> >
> > On Fri, 17 Jan 2025 15:28:47 +0800
> > Yang Ming  wrote:
> >  
> >> DPDK detect vfio container according the existence of vfio
> >> module. But for container with non-privileged mode, there is
> >> possibility that no VFIO_DIR(/dev/vfio) mapping from host to
> >> container when host have both Intel NIC and Mellanox NIC but
> >> this conntainer only allocate VFs from Mellanox NIC.
> >> In this case, vfio kernel module has already been loaded from
> >> the host.
> >> This scenario will cause the error log occurs in DPDK primary
> >> process as below:
> >> 'EAL:   cannot open VFIO container, error 2 (No such file or
> >> directory)'
> >> 'EAL: VFIO support could not be initialized'
> >> Because `rte_vfio_enable()` call `rte_vfio_get_container_fd()`
> >> to execute `vfio_container_fd = open(VFIO_CONTAINER_PATH,
> >> O_RDWR);` but VFIO_CONTAINER_PATH(/dev/vfio/vfio) doesn't exist
> >> in this container.
> >> This scenario will also lead to the delay of DPDK secondary
> >> process because `default_vfio_cfg->vfio_enabled = 0` and
> >> `default_vfio_cfg->vfio_container_fd = -1`, socket error will
> >> be set in DPDK primary process when it sync this info to
> >> the secondary process.
> >> This patch use to skip this kind of useless detection for this
> >> scenario.
> >>
> >> Signed-off-by: Yang Ming 
> >> ---
> >>   lib/eal/linux/eal_vfio.c | 11 +++
> >>   1 file changed, 11 insertions(+)
> >>
> >> diff --git a/lib/eal/linux/eal_vfio.c b/lib/eal/linux/eal_vfio.c
> >> index 7132e24cba..1679d29263 100644
> >> --- a/lib/eal/linux/eal_vfio.c
> >> +++ b/lib/eal/linux/eal_vfio.c
> >> @@ -7,6 +7,7 @@
> >>   #include 
> >>   #include 
> >>   #include 
> >> +#include 
> >>   
> >>   #include 
> >>   #include 
> >> @@ -1083,6 +1084,7 @@ rte_vfio_enable(const char *modname)
> >>/* initialize group list */
> >>int i, j;
> >>int vfio_available;
> >> +  DIR *dir;
> >>const struct internal_config *internal_conf =
> >>eal_get_internal_configuration();
> >>   
> >> @@ -1119,6 +1121,15 @@ rte_vfio_enable(const char *modname)
> >>return 0;
> >>}
> >>   
> >> +  /* return 0 if VFIO directory not exist for container with 
> >> non-privileged mode */
> >> +  dir = opendir(VFIO_DIR);
> >> +  if (dir == NULL) {
> >> +  EAL_LOG(DEBUG,
> >> +  "VFIO directory not exist, skipping VFIO support...");
> >> +  return 0;
> >> +  }
> >> +  closedir(dir);  
> > You need to test the non-container cases.
> > If vfio is loaded /dev/vfio is a character device (not a directory)
> >
> > Also looks suspicious that VFIO_DIR is defined but never used currently.
> >  
> Hi Stephen,
> For non-container test, /dev/vfio/vfio will be character device, not 
> /dev/vfio.
> Here is the command result on my testing environment with Intel NIC.
> 
> [root@computer-1 testuser]# ls -l /dev/vfio
> total 0
> crw-rw-rw-. 1 root root 10, 196 Jan 22 01:50 vfio
> [root@computer-1 testuser]# dpdk-devbind.py -b vfio-pci :04:10.2
> [root@computer-1 testuser]# ls -l /dev/vfio
> total 0
> crw---. 1 root root 239,   0 Jan 22 01:52 59
> crw-rw-rw-. 1 root root  10, 196 Jan 22 01:50 vfio
> [root@computer-1 testuser]# dpdk-devbind.py -b ixgbevf :04:10.2
> [root@computer-1 testuser]# ls -l /dev/vfio
> total 0
> crw-rw-rw-. 1 root root 10, 196 Jan 22 01:50 vfio
> 
> Can you confirm your test scenario?
> 
> 

When vfio-pci is loaded but no device bound:
$ ls -l /dev/vfio
total 0
crw-rw-rw- 1 root root 10, 196 Feb 26 05:39 vfio

After binding device
$ ls -l /dev/vfio
total 0
crw--- 1 root root 511,   0 Feb 26 05:42 15
crw-rw-rw- 1 root root  10, 196 Feb 26 05:39 vfio

So testing for /dev/vfio is good indication that module is loaded.
Not sure what I was thinking earlier.






Re: [PATCH] net/mlx5/hws: change error flow on matcher disconnect

2025-02-26 Thread Raslan Darawsheh
Hi,

From: Gregory Etelson 
Sent: Sunday, February 9, 2025 3:48 PM
To: dev@dpdk.org
Cc: Gregory Etelson; Maayan Kashani; Raslan Darawsheh; Itamar Gozlan
Subject: [PATCH] net/mlx5/hws: change error flow on matcher disconnect

From: Itamar Gozlan 

Currently, when a firmware (FW) failure occurs during matcher error flow,
the system attempts to reconnect the matcher, leading to a segmentation
fault. This happens because the matcher is freed but remains in the list.

Example scenario:
Given matchers M1->M2->M3, if a FW failure occurs:
1. System tries to destroy M1 and fails
2. M1 remains in the list but is freed
3. When destroying M2, it attempts to remove itself and create M1->M3
4. This results in a segmentation fault as M1 is already freed

Signed-off-by: Itamar Gozlan 

Patch applied to next-net-mlx,


Kindest regards,
Raslan Darawsheh



Re: [PATCH] net/mlx5: fix packet hardware ptype translation

2025-02-26 Thread Raslan Darawsheh
Hi,

From: Gregory Etelson 
Sent: Sunday, February 9, 2025 4:01 PM
To: dev@dpdk.org
Cc: Gregory Etelson; Maayan Kashani; Raslan Darawsheh; Slava Ovsiienko; 
sta...@dpdk.org
Subject: [PATCH] net/mlx5: fix packet hardware ptype translation

From: Viacheslav Ovsiienko 

Since ConnectX-8 NIC the hardware recognizes L4 protocol in the
IP fragments if the IP offset is zero and reports the L4 protocol
type in CQE fields. This case was not supported in the ptype translation
table cause not reporting the ptype in the mbufs.

Fixes: ea16068c0064 ("net/mlx5: fix L4 packet type support")
Cc: sta...@dpdk.org

Signed-off-by: Viacheslav Ovsiienko 


Patch applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh


Re: [PATCH 1/2] net/mlx5: add DROP destination to mirror clone actions

2025-02-26 Thread Raslan Darawsheh
Hi,

From: Gregory Etelson 
Sent: Thursday, February 13, 2025 10:32 AM
To: dev@dpdk.org
Cc: Gregory Etelson; Maayan Kashani; Raslan Darawsheh; Tamar Mashiah; 
sta...@dpdk.org
Subject: [PATCH 1/2] net/mlx5: add DROP destination to mirror clone actions

From: Tamar Mashiah 

Add DROP destination to mirror clone actions.

cc: sta...@dpdk.org

Signed-off-by: Tamar Mashiah 


Series applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh


Re: [PATCH] net/mlx5: support ESP in non-template RSS expansion

2025-02-26 Thread Raslan Darawsheh
Hi,

From: Gregory Etelson 
Sent: Monday, February 17, 2025 9:13 AM
To: dev@dpdk.org
Cc: Gregory Etelson; Maayan Kashani; Raslan Darawsheh
Subject: [PATCH] net/mlx5: support ESP in non-template RSS expansion

Add support for the ESP protocol in non-template RSS expansion.

Signed-off-by: Gregory Etelson 


Patch applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh


[PATCH v2 2/3] net/mlx5: add jump FDB Rx flag

2025-02-26 Thread Suanming Mou
When jump FDB Rx is supported, flow will be able to jump
from FDB Tx to FDB Rx, in that case the dest action in FDB
Rx table should support FDB Tx as well.

Signed-off-by: Suanming Mou 
---

v2: fix bool and update log info.

---
 drivers/common/mlx5/mlx5_devx_cmds.c | 8 
 drivers/common/mlx5/mlx5_devx_cmds.h | 1 +
 drivers/net/mlx5/linux/mlx5_os.c | 9 +++--
 drivers/net/mlx5/mlx5.h  | 1 +
 drivers/net/mlx5/mlx5_flow_hw.c  | 8 +++-
 5 files changed, 24 insertions(+), 3 deletions(-)

diff --git a/drivers/common/mlx5/mlx5_devx_cmds.c 
b/drivers/common/mlx5/mlx5_devx_cmds.c
index f504b29f31..440820fd4f 100644
--- a/drivers/common/mlx5/mlx5_devx_cmds.c
+++ b/drivers/common/mlx5/mlx5_devx_cmds.c
@@ -924,6 +924,7 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
uint32_t out[MLX5_ST_SZ_DW(query_hca_cap_out)] = {0};
bool hca_cap_2_sup;
uint64_t general_obj_types_supported = 0;
+   uint64_t stc_action_type_127_64;
void *hcattr;
int rc, i;
 
@@ -1352,6 +1353,13 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
attr->fdb_unified_en = MLX5_GET(wqe_based_flow_table_cap,
hcattr,
fdb_unified_en);
+   stc_action_type_127_64 = MLX5_GET64(wqe_based_flow_table_cap,
+   hcattr,
+   stc_action_type_127_64);
+   if (stc_action_type_127_64 &
+  (1 << 
(MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX -
+ MLX5_IFC_STC_ACTION_TYPE_BIT_64_INDEX)))
+   attr->jump_fdb_rx_en = 1;
}
/* Query HCA attribute for ROCE. */
if (attr->roce) {
diff --git a/drivers/common/mlx5/mlx5_devx_cmds.h 
b/drivers/common/mlx5/mlx5_devx_cmds.h
index 8de4210fb2..6c726a0d46 100644
--- a/drivers/common/mlx5/mlx5_devx_cmds.h
+++ b/drivers/common/mlx5/mlx5_devx_cmds.h
@@ -326,6 +326,7 @@ struct mlx5_hca_attr {
uint32_t lag_rx_port_affinity:1;
uint32_t wqe_based_flow_table_sup:1;
uint32_t fdb_unified_en:1;
+   uint32_t jump_fdb_rx_en:1;
uint8_t max_header_modify_pattern_length;
uint64_t system_image_guid;
uint32_t log_max_conn_track_offload:5;
diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 9410211e3b..77f17f1a14 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -1716,8 +1716,13 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
 *   4. all representors in HWS
 */
priv->unified_fdb_en = !!priv->master;
-   DRV_LOG(DEBUG, "port %u: unified FDB %s enabled.",
-   eth_dev->data->port_id, priv->unified_fdb_en ? "is" : 
"isn't");
+   /* Jump FDB Rx works only with unified FDB enabled. */
+   if (priv->unified_fdb_en)
+   priv->jump_fdb_rx_en = 
sh->cdev->config.hca_attr.jump_fdb_rx_en;
+   DRV_LOG(DEBUG, "port %u: unified FDB %s enabled, jump_fdb_rx %s 
enabled.",
+   eth_dev->data->port_id,
+   priv->unified_fdb_en ? "is" : "isn't",
+   priv->jump_fdb_rx_en ? "is" : "isn't");
if (priv->sh->config.dv_esw_en) {
uint32_t usable_bits;
uint32_t required_bits;
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index f73f6e63ff..545ba48b3c 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -1987,6 +1987,7 @@ struct mlx5_priv {
uint32_t num_lag_ports:4; /* Number of ports can be bonded. */
uint32_t tunnel_enabled:1; /* If tunnel offloading is enabled on rxqs. 
*/
uint32_t unified_fdb_en:1; /* Unified FDB flag per port. */
+   uint32_t jump_fdb_rx_en:1; /* Jump from FDB Tx to FDB Rx flag per port. 
*/
uint16_t domain_id; /* Switch domain identifier. */
uint16_t vport_id; /* Associated VF vport index (if any). */
uint32_t vport_meta_tag; /* Used for vport index match ove VF LAG. */
diff --git a/drivers/net/mlx5/mlx5_flow_hw.c b/drivers/net/mlx5/mlx5_flow_hw.c
index 07d36bfb8f..cdb3dd6420 100644
--- a/drivers/net/mlx5/mlx5_flow_hw.c
+++ b/drivers/net/mlx5/mlx5_flow_hw.c
@@ -9350,6 +9350,7 @@ flow_hw_grp_create_cb(void *tool_ctx, void *cb_ctx)
struct mlx5_flow_group *grp_data;
struct mlx5dr_table *tbl = NULL;
struct mlx5dr_action *jump;
+   uint32_t hws_flags;
uint32_t idx = 0;
MKSTR(matcher_name, "%s_%s_%u_%u_matcher_list",
  attr->transfer ? "FDB" : "NIC", attr->egress ? "egress" : 
"ingress",
@@ -9370,10 +9371,15 @@ flow_hw_grp_create_cb(void *tool_ctx, void *cb_ctx)
goto error;
grp_data->tbl = tbl;
if (attr->group) {
+   hws_flags =

[PATCH v2 3/3] net/mlx5: allow FDB RSS

2025-02-26 Thread Suanming Mou
RSS can be used in FDB Rx rules when JUMP_FDB_RX action to allow
jump from FDB Tx to FDB Rx table.

Different with NIC RSS, FDB RSS will not do the internal implicit
metadata copy.

This commit enables the FDB RSS if JUMP_FDB_RX is supported.

Signed-off-by: Suanming Mou 
---

v2: update limitation.

---
 doc/guides/nics/mlx5.rst|  4 
 drivers/net/mlx5/mlx5_flow.c|  4 ++--
 drivers/net/mlx5/mlx5_flow_dv.c | 32 +---
 drivers/net/mlx5/mlx5_flow_hw.c | 20 +++-
 4 files changed, 38 insertions(+), 22 deletions(-)

diff --git a/doc/guides/nics/mlx5.rst b/doc/guides/nics/mlx5.rst
index d40878e336..0371a397b6 100644
--- a/doc/guides/nics/mlx5.rst
+++ b/doc/guides/nics/mlx5.rst
@@ -968,6 +968,10 @@ Limitations
   The flow engine of a process cannot move from active to standby mode
   if preceding active application rules are still present and vice versa.
 
+- Since only when unified FDB is enabled, there will be FDB Rx/Tx, jump FDB Rx
+  is valid only when unfied is enabled. And in unified FDB mode, the tag and 
RSS
+  actions are only allowed in FDB Rx domain.
+
 
 Statistics
 --
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 69d7621d0b..f8b3e504b3 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -1977,9 +1977,9 @@ mlx5_flow_validate_action_mark(struct rte_eth_dev *dev,
  RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
  "mark action not supported for "
  "egress");
-   if (attr->transfer && mlx5_hws_active(dev))
+   if (attr->transfer && !mlx5_hws_active(dev))
return rte_flow_error_set(error, ENOTSUP,
- RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
+ RTE_FLOW_ERROR_TYPE_ATTR_TRANSFER, 
NULL,
  "non-template mark action not 
supported for transfer");
return 0;
 }
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 633c41e358..61d3101ce8 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -8939,21 +8939,23 @@ flow_dv_validate(struct rte_eth_dev *dev, const struct 
rte_flow_attr *attr,
  RTE_FLOW_ERROR_TYPE_ACTION,
  NULL,
  "unsupported action MARK");
-   if (action_flags & MLX5_FLOW_ACTION_QUEUE)
-   return rte_flow_error_set(error, ENOTSUP,
- RTE_FLOW_ERROR_TYPE_ACTION,
- NULL,
- "unsupported action QUEUE");
-   if (action_flags & MLX5_FLOW_ACTION_RSS)
-   return rte_flow_error_set(error, ENOTSUP,
- RTE_FLOW_ERROR_TYPE_ACTION,
- NULL,
- "unsupported action RSS");
-   if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
-   return rte_flow_error_set(error, EINVAL,
- RTE_FLOW_ERROR_TYPE_ACTION,
- actions,
- "no fate action is found");
+   if (!priv->jump_fdb_rx_en) {
+   if (action_flags & MLX5_FLOW_ACTION_QUEUE)
+   return rte_flow_error_set(error, ENOTSUP,
+ 
RTE_FLOW_ERROR_TYPE_ACTION,
+ NULL,
+ "unsupported action 
QUEUE");
+   if (action_flags & MLX5_FLOW_ACTION_RSS)
+   return rte_flow_error_set(error, ENOTSUP,
+ 
RTE_FLOW_ERROR_TYPE_ACTION,
+ NULL,
+ "unsupported action 
RSS");
+   if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
+   return rte_flow_error_set(error, EINVAL,
+ 
RTE_FLOW_ERROR_TYPE_ACTION,
+ actions,
+ "no fate action is 
found");
+   }
} else {
if (!(action_flags & MLX5_FLOW_FATE_ACTIONS) && attr->ingress)
return rte_flow_error_set(error, EINVAL,
diff --git a/drivers/net/mlx5/

[PATCH v2 1/3] net/mlx5/hws: support jump FDB Rx

2025-02-26 Thread Suanming Mou
Before FW introduced the JUMP_FDB_RX action feature, jump from
FDB Tx to Rx is not allowed. JUMP_FDB_RX feature introduces
the internal loopback for Tx case and allow the REG C0 C1 B be
preserved as well.

This commit adds the JUMP_FDB_RX cap bit check and use
JUMP_FDB_RX instead of FT for dest table FDB Rx case.

Signed-off-by: Suanming Mou 
Signed-off-by: Alex Vesker 
---

v2: fix line lengh.

---
 drivers/common/mlx5/mlx5_prm.h|  9 -
 drivers/net/mlx5/hws/mlx5dr_action.c  | 26 +-
 drivers/net/mlx5/hws/mlx5dr_action.h  |  4 
 drivers/net/mlx5/hws/mlx5dr_cmd.c |  9 +
 drivers/net/mlx5/hws/mlx5dr_cmd.h |  2 ++
 drivers/net/mlx5/hws/mlx5dr_context.c | 17 +
 drivers/net/mlx5/hws/mlx5dr_context.h |  2 ++
 7 files changed, 63 insertions(+), 6 deletions(-)

diff --git a/drivers/common/mlx5/mlx5_prm.h b/drivers/common/mlx5/mlx5_prm.h
index 3fc3b0cd2a..84e3347794 100644
--- a/drivers/common/mlx5/mlx5_prm.h
+++ b/drivers/common/mlx5/mlx5_prm.h
@@ -2466,7 +2466,8 @@ struct mlx5_ifc_wqe_based_flow_table_cap_bits {
u8 reserved_at_60[0x8];
u8 max_header_modify_pattern_length[0x8];
u8 ste_format[0x10];
-   u8 stc_action_type[0x80];
+   u8 stc_action_type_63_0[0x40];
+   u8 stc_action_type_127_64[0x40];
u8 header_insert_type[0x10];
u8 header_remove_type[0x10];
u8 trivial_match_definer[0x20];
@@ -3543,6 +3544,11 @@ enum mlx5_ifc_rtc_reparse_mode {
MLX5_IFC_RTC_REPARSE_BY_STC = 0x2,
 };
 
+enum mlx5_ifc_stc_action_type_bit_index {
+   MLX5_IFC_STC_ACTION_TYPE_BIT_64_INDEX = 64,
+   MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX = 71,
+};
+
 #define MLX5_IFC_RTC_LINEAR_LOOKUP_TBL_LOG_MAX 16
 
 struct mlx5_ifc_rtc_bits {
@@ -3621,6 +3627,7 @@ enum mlx5_ifc_stc_action_type {
MLX5_IFC_STC_ACTION_TYPE_ALLOW = 0x84,
MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_VPORT = 0x85,
MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_UPLINK = 0x86,
+   MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FLOW_TABLE_FDB_RX = 0x87,
 };
 
 enum mlx5_ifc_stc_reparse_mode {
diff --git a/drivers/net/mlx5/hws/mlx5dr_action.c 
b/drivers/net/mlx5/hws/mlx5dr_action.c
index b9452a3ebc..e21db5b327 100644
--- a/drivers/net/mlx5/hws/mlx5dr_action.c
+++ b/drivers/net/mlx5/hws/mlx5dr_action.c
@@ -803,6 +803,9 @@ int mlx5dr_action_root_build_attr(struct mlx5dr_rule_action 
rule_actions[],
 
switch (action->type) {
case MLX5DR_ACTION_TYP_TBL:
+   attr[i].type = MLX5DV_FLOW_ACTION_DEST_DEVX;
+   attr[i].obj = action->dest_tbl.devx_obj->obj;
+   break;
case MLX5DR_ACTION_TYP_TIR:
attr[i].type = MLX5DV_FLOW_ACTION_DEST_DEVX;
attr[i].obj = action->devx_obj;
@@ -1097,6 +1100,17 @@ static void mlx5dr_action_fill_stc_attr(struct 
mlx5dr_action *action,
}
break;
case MLX5DR_ACTION_TYP_TBL:
+   attr->action_offset = MLX5DR_ACTION_OFFSET_HIT;
+   attr->dest_table_id = obj->id;
+   /* Only for unified FDB Rx case */
+   if (mlx5dr_context_cap_stc(action->ctx,
+   MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX) 
&&
+   action->dest_tbl.type == MLX5DR_TABLE_TYPE_FDB_RX)
+   attr->action_type = 
MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FLOW_TABLE_FDB_RX;
+   else
+   attr->action_type = MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FT;
+
+   break;
case MLX5DR_ACTION_TYP_DEST_ARRAY:
attr->action_type = MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FT;
attr->action_offset = MLX5DR_ACTION_OFFSET_HIT;
@@ -1419,17 +1433,19 @@ mlx5dr_action_create_dest_table(struct mlx5dr_context 
*ctx,
if (!action)
return NULL;
 
+   action->dest_tbl.type = tbl->type;
+
if (mlx5dr_action_is_root_flags(flags)) {
if (mlx5dr_context_shared_gvmi_used(ctx))
-   action->devx_obj = tbl->local_ft->obj;
+   action->dest_tbl.devx_obj = tbl->local_ft;
else
-   action->devx_obj = tbl->ft->obj;
+   action->dest_tbl.devx_obj = tbl->ft;
} else {
+   action->dest_tbl.devx_obj = tbl->ft;
+
ret = mlx5dr_action_create_stcs(action, tbl->ft);
if (ret)
goto free_action;
-
-   action->devx_dest.devx_obj = tbl->ft;
}
 
return action;
@@ -2526,7 +2542,7 @@ mlx5dr_action_create_dest_array(struct mlx5dr_context 
*ctx,
case MLX5DR_ACTION_TYP_TBL:
dest_list[i].destination_type =
MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE;
-   dest_list[i].destination_id = 
dests[i].

[PATCH v2 0/3] net/mlx5: support jump FDB Rx

2025-02-26 Thread Suanming Mou
Before FW introduced the JUMP_FDB_RX action feature, jump from
FDB Tx to Rx is not allowed. JUMP_FDB_RX feature introduces
the internal loopback for Tx case and allow the REG C0 C1 B be
preserved as well.

This series adds the jump from FDB Tx to FDB Rx support and allows
RSS action in FDB domain rule when jump FDB Rx is supported.


Suanming Mou (3):
  net/mlx5/hws: support jump FDB Rx
  net/mlx5: add jump FDB Rx flag
  net/mlx5: allow FDB RSS

 doc/guides/nics/mlx5.rst  |  4 
 drivers/common/mlx5/mlx5_devx_cmds.c  |  8 +++
 drivers/common/mlx5/mlx5_devx_cmds.h  |  1 +
 drivers/common/mlx5/mlx5_prm.h|  9 +++-
 drivers/net/mlx5/hws/mlx5dr_action.c  | 26 +-
 drivers/net/mlx5/hws/mlx5dr_action.h  |  4 
 drivers/net/mlx5/hws/mlx5dr_cmd.c |  9 
 drivers/net/mlx5/hws/mlx5dr_cmd.h |  2 ++
 drivers/net/mlx5/hws/mlx5dr_context.c | 17 ++
 drivers/net/mlx5/hws/mlx5dr_context.h |  2 ++
 drivers/net/mlx5/linux/mlx5_os.c  |  9 ++--
 drivers/net/mlx5/mlx5.h   |  1 +
 drivers/net/mlx5/mlx5_flow.c  |  4 ++--
 drivers/net/mlx5/mlx5_flow_dv.c   | 32 ++-
 drivers/net/mlx5/mlx5_flow_hw.c   | 28 ++-
 15 files changed, 125 insertions(+), 31 deletions(-)

-- 
2.34.1



Re: [PATCH V2 0/4] mlx5 unified fdb

2025-02-26 Thread Raslan Darawsheh
Hi,

From: Ophir Munk 
Sent: Wednesday, February 26, 2025 10:38 AM
To: Dariusz Sosnowski; Slava Ovsiienko; Bing Zhao; Ori Kam; Suanming Mou; Matan 
Azrad
Cc: dev@dpdk.org; Raslan Darawsheh
Subject: [PATCH V2 0/4] mlx5 unified fdb

V1: initial version
V2: rebase + determine unified fdb based on FW capability

Bing Zhao (1):
  net/mlx5: remove unneeded FDB flag on representor action

Ophir Munk (3):
  common/mlx5: support FDB unified capability query
  net/mlx5: support FDB unified domain
  net/mlx5/hws: allow different types in miss validation

 drivers/common/mlx5/mlx5_devx_cmds.c |   3 +
 drivers/common/mlx5/mlx5_devx_cmds.h |   1 +
 drivers/net/mlx5/hws/mlx5dr_table.c  |  18 +-
 drivers/net/mlx5/linux/mlx5_os.c |  11 +
 drivers/net/mlx5/mlx5.h  |   8 +
 drivers/net/mlx5/mlx5_flow_hw.c  | 521 ---
 drivers/net/mlx5/mlx5_flow_meter.c   |  16 +-
 drivers/net/mlx5/mlx5_flow_quota.c   |   9 +-
 drivers/net/mlx5/mlx5_hws_cnt.c  |   9 +-
 9 files changed, 429 insertions(+), 167 deletions(-)

--
2.8.4


rebased squashed second and third patch together,
series applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh



[PATCH v3 0/3] net/mlx5: support jump FDB Rx

2025-02-26 Thread Dariusz Sosnowski
Before FW introduced the JUMP_FDB_RX action feature, jump from
FDB Tx to Rx is not allowed. JUMP_FDB_RX feature introduces
the internal loopback for Tx case and allow the REG C0 C1 B be
preserved as well.

This series adds the jump from FDB Tx to FDB Rx support and allows
RSS action in FDB domain rule when jump FDB Rx is supported.

v3:
- Rebased on v2 of mlx5 Unified FDB patches (series 34723)

Suanming Mou (3):
  net/mlx5/hws: support jump FDB Rx
  net/mlx5: add jump FDB Rx flag
  net/mlx5: allow FDB RSS

 doc/guides/nics/mlx5.rst  |  4 
 drivers/common/mlx5/mlx5_devx_cmds.c  |  8 +++
 drivers/common/mlx5/mlx5_devx_cmds.h  |  1 +
 drivers/common/mlx5/mlx5_prm.h|  9 +++-
 drivers/net/mlx5/hws/mlx5dr_action.c  | 26 +-
 drivers/net/mlx5/hws/mlx5dr_action.h  |  4 
 drivers/net/mlx5/hws/mlx5dr_cmd.c |  9 
 drivers/net/mlx5/hws/mlx5dr_cmd.h |  2 ++
 drivers/net/mlx5/hws/mlx5dr_context.c | 17 ++
 drivers/net/mlx5/hws/mlx5dr_context.h |  2 ++
 drivers/net/mlx5/linux/mlx5_os.c  |  9 ++--
 drivers/net/mlx5/mlx5.h   |  1 +
 drivers/net/mlx5/mlx5_flow.c  |  4 ++--
 drivers/net/mlx5/mlx5_flow_dv.c   | 32 ++-
 drivers/net/mlx5/mlx5_flow_hw.c   | 28 ++-
 15 files changed, 125 insertions(+), 31 deletions(-)

--
2.39.5



[PATCH v3 3/3] net/mlx5: allow FDB RSS

2025-02-26 Thread Dariusz Sosnowski
From: Suanming Mou 

RSS can be used in FDB Rx rules when JUMP_FDB_RX action to allow
jump from FDB Tx to FDB Rx table.

Different with NIC RSS, FDB RSS will not do the internal implicit
metadata copy.

This commit enables the FDB RSS if JUMP_FDB_RX is supported.

Signed-off-by: Suanming Mou 
Acked-by: Dariusz Sosnowski 
---
 doc/guides/nics/mlx5.rst|  4 
 drivers/net/mlx5/mlx5_flow.c|  4 ++--
 drivers/net/mlx5/mlx5_flow_dv.c | 32 +---
 drivers/net/mlx5/mlx5_flow_hw.c | 20 +++-
 4 files changed, 38 insertions(+), 22 deletions(-)

diff --git a/doc/guides/nics/mlx5.rst b/doc/guides/nics/mlx5.rst
index def40fc20a..190c58b54d 100644
--- a/doc/guides/nics/mlx5.rst
+++ b/doc/guides/nics/mlx5.rst
@@ -970,6 +970,10 @@ Limitations
   The flow engine of a process cannot move from active to standby mode
   if preceding active application rules are still present and vice versa.
 
+- Since only when unified FDB is enabled, there will be FDB Rx/Tx, jump FDB Rx
+  is valid only when unfied is enabled. And in unified FDB mode, the tag and 
RSS
+  actions are only allowed in FDB Rx domain.
+
 
 Statistics
 --
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 69d7621d0b..f8b3e504b3 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -1977,9 +1977,9 @@ mlx5_flow_validate_action_mark(struct rte_eth_dev *dev,
  RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
  "mark action not supported for "
  "egress");
-   if (attr->transfer && mlx5_hws_active(dev))
+   if (attr->transfer && !mlx5_hws_active(dev))
return rte_flow_error_set(error, ENOTSUP,
- RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
+ RTE_FLOW_ERROR_TYPE_ATTR_TRANSFER, 
NULL,
  "non-template mark action not 
supported for transfer");
return 0;
 }
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 633c41e358..61d3101ce8 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -8939,21 +8939,23 @@ flow_dv_validate(struct rte_eth_dev *dev, const struct 
rte_flow_attr *attr,
  RTE_FLOW_ERROR_TYPE_ACTION,
  NULL,
  "unsupported action MARK");
-   if (action_flags & MLX5_FLOW_ACTION_QUEUE)
-   return rte_flow_error_set(error, ENOTSUP,
- RTE_FLOW_ERROR_TYPE_ACTION,
- NULL,
- "unsupported action QUEUE");
-   if (action_flags & MLX5_FLOW_ACTION_RSS)
-   return rte_flow_error_set(error, ENOTSUP,
- RTE_FLOW_ERROR_TYPE_ACTION,
- NULL,
- "unsupported action RSS");
-   if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
-   return rte_flow_error_set(error, EINVAL,
- RTE_FLOW_ERROR_TYPE_ACTION,
- actions,
- "no fate action is found");
+   if (!priv->jump_fdb_rx_en) {
+   if (action_flags & MLX5_FLOW_ACTION_QUEUE)
+   return rte_flow_error_set(error, ENOTSUP,
+ 
RTE_FLOW_ERROR_TYPE_ACTION,
+ NULL,
+ "unsupported action 
QUEUE");
+   if (action_flags & MLX5_FLOW_ACTION_RSS)
+   return rte_flow_error_set(error, ENOTSUP,
+ 
RTE_FLOW_ERROR_TYPE_ACTION,
+ NULL,
+ "unsupported action 
RSS");
+   if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
+   return rte_flow_error_set(error, EINVAL,
+ 
RTE_FLOW_ERROR_TYPE_ACTION,
+ actions,
+ "no fate action is 
found");
+   }
} else {
if (!(action_flags & MLX5_FLOW_FATE_ACTIONS) && attr->ingress)
return rte_flow_error_set(error, EINVAL,
diff --gi

[PATCH v3 2/3] net/mlx5: add jump FDB Rx flag

2025-02-26 Thread Dariusz Sosnowski
From: Suanming Mou 

When jump FDB Rx is supported, flow will be able to jump
from FDB Tx to FDB Rx, in that case the dest action in FDB
Rx table should support FDB Tx as well.

Signed-off-by: Suanming Mou 
Acked-by: Dariusz Sosnowski 
---
 drivers/common/mlx5/mlx5_devx_cmds.c | 8 
 drivers/common/mlx5/mlx5_devx_cmds.h | 1 +
 drivers/net/mlx5/linux/mlx5_os.c | 9 +++--
 drivers/net/mlx5/mlx5.h  | 1 +
 drivers/net/mlx5/mlx5_flow_hw.c  | 8 +++-
 5 files changed, 24 insertions(+), 3 deletions(-)

diff --git a/drivers/common/mlx5/mlx5_devx_cmds.c 
b/drivers/common/mlx5/mlx5_devx_cmds.c
index f504b29f31..440820fd4f 100644
--- a/drivers/common/mlx5/mlx5_devx_cmds.c
+++ b/drivers/common/mlx5/mlx5_devx_cmds.c
@@ -924,6 +924,7 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
uint32_t out[MLX5_ST_SZ_DW(query_hca_cap_out)] = {0};
bool hca_cap_2_sup;
uint64_t general_obj_types_supported = 0;
+   uint64_t stc_action_type_127_64;
void *hcattr;
int rc, i;
 
@@ -1352,6 +1353,13 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
attr->fdb_unified_en = MLX5_GET(wqe_based_flow_table_cap,
hcattr,
fdb_unified_en);
+   stc_action_type_127_64 = MLX5_GET64(wqe_based_flow_table_cap,
+   hcattr,
+   stc_action_type_127_64);
+   if (stc_action_type_127_64 &
+  (1 << 
(MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX -
+ MLX5_IFC_STC_ACTION_TYPE_BIT_64_INDEX)))
+   attr->jump_fdb_rx_en = 1;
}
/* Query HCA attribute for ROCE. */
if (attr->roce) {
diff --git a/drivers/common/mlx5/mlx5_devx_cmds.h 
b/drivers/common/mlx5/mlx5_devx_cmds.h
index 8de4210fb2..6c726a0d46 100644
--- a/drivers/common/mlx5/mlx5_devx_cmds.h
+++ b/drivers/common/mlx5/mlx5_devx_cmds.h
@@ -326,6 +326,7 @@ struct mlx5_hca_attr {
uint32_t lag_rx_port_affinity:1;
uint32_t wqe_based_flow_table_sup:1;
uint32_t fdb_unified_en:1;
+   uint32_t jump_fdb_rx_en:1;
uint8_t max_header_modify_pattern_length;
uint64_t system_image_guid;
uint32_t log_max_conn_track_offload:5;
diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 7f13cb2c7d..573e846ed2 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -1716,8 +1716,13 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
 *   4. all representors in HWS
 */
priv->unified_fdb_en = !!priv->master && 
sh->cdev->config.hca_attr.fdb_unified_en;
-   DRV_LOG(DEBUG, "port %u: unified FDB %s enabled.",
-   eth_dev->data->port_id, priv->unified_fdb_en ? "is" : 
"isn't");
+   /* Jump FDB Rx works only with unified FDB enabled. */
+   if (priv->unified_fdb_en)
+   priv->jump_fdb_rx_en = 
sh->cdev->config.hca_attr.jump_fdb_rx_en;
+   DRV_LOG(DEBUG, "port %u: unified FDB %s enabled, jump_fdb_rx %s 
enabled.",
+   eth_dev->data->port_id,
+   priv->unified_fdb_en ? "is" : "isn't",
+   priv->jump_fdb_rx_en ? "is" : "isn't");
if (priv->sh->config.dv_esw_en) {
uint32_t usable_bits;
uint32_t required_bits;
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index f73f6e63ff..545ba48b3c 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -1987,6 +1987,7 @@ struct mlx5_priv {
uint32_t num_lag_ports:4; /* Number of ports can be bonded. */
uint32_t tunnel_enabled:1; /* If tunnel offloading is enabled on rxqs. 
*/
uint32_t unified_fdb_en:1; /* Unified FDB flag per port. */
+   uint32_t jump_fdb_rx_en:1; /* Jump from FDB Tx to FDB Rx flag per port. 
*/
uint16_t domain_id; /* Switch domain identifier. */
uint16_t vport_id; /* Associated VF vport index (if any). */
uint32_t vport_meta_tag; /* Used for vport index match ove VF LAG. */
diff --git a/drivers/net/mlx5/mlx5_flow_hw.c b/drivers/net/mlx5/mlx5_flow_hw.c
index 983af38c96..8909f5b506 100644
--- a/drivers/net/mlx5/mlx5_flow_hw.c
+++ b/drivers/net/mlx5/mlx5_flow_hw.c
@@ -9350,6 +9350,7 @@ flow_hw_grp_create_cb(void *tool_ctx, void *cb_ctx)
struct mlx5_flow_group *grp_data;
struct mlx5dr_table *tbl = NULL;
struct mlx5dr_action *jump;
+   uint32_t hws_flags;
uint32_t idx = 0;
MKSTR(matcher_name, "%s_%s_%u_%u_matcher_list",
  attr->transfer ? "FDB" : "NIC", attr->egress ? "egress" : 
"ingress",
@@ -9370,10 +9371,15 @@ flow_hw_grp_create_cb(void *tool_ctx, void *cb_ctx)
goto error;
grp_data->tbl = tbl;

[PATCH v3 1/3] net/mlx5/hws: support jump FDB Rx

2025-02-26 Thread Dariusz Sosnowski
From: Suanming Mou 

Before FW introduced the JUMP_FDB_RX action feature, jump from
FDB Tx to Rx is not allowed. JUMP_FDB_RX feature introduces
the internal loopback for Tx case and allow the REG C0 C1 B be
preserved as well.

This commit adds the JUMP_FDB_RX cap bit check and use
JUMP_FDB_RX instead of FT for dest table FDB Rx case.

Signed-off-by: Suanming Mou 
Signed-off-by: Alex Vesker 
Acked-by: Dariusz Sosnowski 
---
 drivers/common/mlx5/mlx5_prm.h|  9 -
 drivers/net/mlx5/hws/mlx5dr_action.c  | 26 +-
 drivers/net/mlx5/hws/mlx5dr_action.h  |  4 
 drivers/net/mlx5/hws/mlx5dr_cmd.c |  9 +
 drivers/net/mlx5/hws/mlx5dr_cmd.h |  2 ++
 drivers/net/mlx5/hws/mlx5dr_context.c | 17 +
 drivers/net/mlx5/hws/mlx5dr_context.h |  2 ++
 7 files changed, 63 insertions(+), 6 deletions(-)

diff --git a/drivers/common/mlx5/mlx5_prm.h b/drivers/common/mlx5/mlx5_prm.h
index 0885f8757f..742c274a85 100644
--- a/drivers/common/mlx5/mlx5_prm.h
+++ b/drivers/common/mlx5/mlx5_prm.h
@@ -2466,7 +2466,8 @@ struct mlx5_ifc_wqe_based_flow_table_cap_bits {
u8 reserved_at_60[0x8];
u8 max_header_modify_pattern_length[0x8];
u8 ste_format[0x10];
-   u8 stc_action_type[0x80];
+   u8 stc_action_type_63_0[0x40];
+   u8 stc_action_type_127_64[0x40];
u8 header_insert_type[0x10];
u8 header_remove_type[0x10];
u8 trivial_match_definer[0x20];
@@ -3543,6 +3544,11 @@ enum mlx5_ifc_rtc_reparse_mode {
MLX5_IFC_RTC_REPARSE_BY_STC = 0x2,
 };
 
+enum mlx5_ifc_stc_action_type_bit_index {
+   MLX5_IFC_STC_ACTION_TYPE_BIT_64_INDEX = 64,
+   MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX = 71,
+};
+
 #define MLX5_IFC_RTC_LINEAR_LOOKUP_TBL_LOG_MAX 16
 
 struct mlx5_ifc_rtc_bits {
@@ -3621,6 +3627,7 @@ enum mlx5_ifc_stc_action_type {
MLX5_IFC_STC_ACTION_TYPE_ALLOW = 0x84,
MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_VPORT = 0x85,
MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_UPLINK = 0x86,
+   MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FLOW_TABLE_FDB_RX = 0x87,
 };
 
 enum mlx5_ifc_stc_reparse_mode {
diff --git a/drivers/net/mlx5/hws/mlx5dr_action.c 
b/drivers/net/mlx5/hws/mlx5dr_action.c
index 200a4995d0..c3e67f8a0b 100644
--- a/drivers/net/mlx5/hws/mlx5dr_action.c
+++ b/drivers/net/mlx5/hws/mlx5dr_action.c
@@ -803,6 +803,9 @@ int mlx5dr_action_root_build_attr(struct mlx5dr_rule_action 
rule_actions[],
 
switch (action->type) {
case MLX5DR_ACTION_TYP_TBL:
+   attr[i].type = MLX5DV_FLOW_ACTION_DEST_DEVX;
+   attr[i].obj = action->dest_tbl.devx_obj->obj;
+   break;
case MLX5DR_ACTION_TYP_TIR:
attr[i].type = MLX5DV_FLOW_ACTION_DEST_DEVX;
attr[i].obj = action->devx_obj;
@@ -1097,6 +1100,17 @@ static void mlx5dr_action_fill_stc_attr(struct 
mlx5dr_action *action,
}
break;
case MLX5DR_ACTION_TYP_TBL:
+   attr->action_offset = MLX5DR_ACTION_OFFSET_HIT;
+   attr->dest_table_id = obj->id;
+   /* Only for unified FDB Rx case */
+   if (mlx5dr_context_cap_stc(action->ctx,
+   MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX) 
&&
+   action->dest_tbl.type == MLX5DR_TABLE_TYPE_FDB_RX)
+   attr->action_type = 
MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FLOW_TABLE_FDB_RX;
+   else
+   attr->action_type = MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FT;
+
+   break;
case MLX5DR_ACTION_TYP_DEST_ARRAY:
attr->action_type = MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FT;
attr->action_offset = MLX5DR_ACTION_OFFSET_HIT;
@@ -1419,17 +1433,19 @@ mlx5dr_action_create_dest_table(struct mlx5dr_context 
*ctx,
if (!action)
return NULL;
 
+   action->dest_tbl.type = tbl->type;
+
if (mlx5dr_action_is_root_flags(flags)) {
if (mlx5dr_context_shared_gvmi_used(ctx))
-   action->devx_obj = tbl->local_ft->obj;
+   action->dest_tbl.devx_obj = tbl->local_ft;
else
-   action->devx_obj = tbl->ft->obj;
+   action->dest_tbl.devx_obj = tbl->ft;
} else {
+   action->dest_tbl.devx_obj = tbl->ft;
+
ret = mlx5dr_action_create_stcs(action, tbl->ft);
if (ret)
goto free_action;
-
-   action->devx_dest.devx_obj = tbl->ft;
}
 
return action;
@@ -2526,7 +2542,7 @@ mlx5dr_action_create_dest_array(struct mlx5dr_context 
*ctx,
case MLX5DR_ACTION_TYP_TBL:
dest_list[i].destination_type =
MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE;
-   dest_list[i].des

RE: release candidate 25.03-rc1

2025-02-26 Thread Xu, HailinX
> -Original Message-
> From: Thomas Monjalon 
> Sent: Thursday, February 13, 2025 12:50 AM
> To: annou...@dpdk.org
> Subject: release candidate 25.03-rc1
> 
> A new DPDK release candidate is ready for testing:
>   https://git.dpdk.org/dpdk/tag/?id=v25.03-rc1
> 
> There are 478 new patches in this snapshot.
> 
> Release notes:
>   https://doc.dpdk.org/guides/rel_notes/release_25_03.html
> 
> Highlights of 25.03-rc1:
>   - Staged-Ordered ring (SORING)
>   - mbuf raw bulk functions
>   - some Intel drivers merged together
>   - Yunsilicon xsc networking driver
>   - ZTE Storage Data Accelerator (ZSDA) driver
>   - Intel compiler icc support replaced by icx
>   - more function annotations for analyzers
>   - more MSVC compatibility
> 
> Please test and report issues on bugs.dpdk.org.
> 
> DPDK 25.03-rc2 is expected in more than two weeks (end of February).
> 
> Thank you everyone
> 
Update the test status for Intel part. dpdk25.03-rc1 all test is done. found 
six new issues.

New issues:
1. [dpdk25.03] nvgre/nvgre_ipv4_checksum_offload: core dumped when quit testpmd 
   -> has fix patch
2. [dpdk-25.03] ice_dcf_acl_filter: the rule that supports both switch and ACL 
will be prioritized for switch-> Intel dev is under investigating
3. [dpdk-25.03rc1][Ubuntu24.04.1] 
ipsec_gw_cryptodev_func/qat_3des_cbc_ipv4_transport: buffer overflow detected, 
Aborted (core dumped)-> has fix patch
4. Bug 1663 - [dpdk-25.03] 
flexible_rxd/check_IPv4_IPv6_TCP_fields_in_RXD_on_specific_queues: port config 
rxq/txq failed after attach port-> not fix yet
5. [dpdk-25.03] [NNT] pmd_bonded_8023ad: port start failed after set bonding 
lacp-> has fix patch
6. [dpdk-25.03]port_control/test_pf_start_stop_reset_close: Failed to manually 
reset pf-> has fix patch

# Basic Intel(R) NIC testing
* Build or compile:  
 *Build: cover the build test combination with latest GCC/Clang version and the 
popular OS revision such as Ubuntu24.10, Ubuntu24.04.1, Fedora41, RHEL10, 
RHEL9.4, RHEL9.5-RT, FreeBSD14.2, SUSE15.6, OpenAnolis8.9, AzureLinux3.0 etc.
  - All test passed.
 *Compile: cover the CFLAGES(O0/O1/O2/O3) with popular OS such as Ubuntu24.04.1 
and RHEL9.4.
  - All test passed with latest dpdk.
* PF/VF(i40e, ixgbe): test scenarios including 
PF/VF-RTE_FLOW/TSO/Jumboframe/checksum offload/VLAN/VXLAN, etc. 
- All test case is done. found the 5 issue.
* PF/VF(ice): test scenarios including Switch features/Package Management/Flow 
Director/Advanced Tx/Advanced RSS/ACL/DCF/Flexible Descriptor, etc.
- Execution rate is done. found the 1,2,4,6 issues.
* PF/VF(apfi/cpfi): test scenarios including 
cpfi_rte_flow/TSO/MTU/Jumboframe/checksum offload, etc.
- Execution rate is done. No new issue is found.
* Intel NIC single core/NIC performance: test scenarios including PF/VF single 
core performance test, RFC2544 Zero packet loss performance test, etc.
- Execution rate is done. No new issue is found.
* Power and IPsec: 
 * Power: test scenarios including bi-direction/Telemetry/Empty Poll 
Lib/Priority Base Frequency, etc. 
- Execution rate is done. No new issue is found.
 * IPsec: test scenarios including ipsec/ipsec-gw/ipsec library basic test - 
QAT&SW/FIB library, etc.
- Execution rate is done. No new issue is found. 
# Basic cryptodev and virtio testing
* Virtio: both function and performance test are covered. Such as 
PVP/Virtio_loopback/virtio-user loopback/virtio-net VM2VM perf testing/VMAWARE 
ESXI 8.0U1, etc.
- Execution rate is done. No new issue is found.
* Cryptodev: 
 *Function test: test scenarios including Cryptodev API testing/CompressDev 
ISA-L/QAT/ZLIB PMD Testing/FIPS, etc.
- Execution rate is done. found the 3 issue. 
 *Performance test: test scenarios including Throughput Performance /Cryptodev 
Latency, etc.
- Execution rate is done. No performance drop.


Regards,
Xu, Hailin


RE: [PATCH v3 0/3] net/mlx5: support jump FDB Rx

2025-02-26 Thread Bing Zhao
Thanks a lot, Dariusz.

> -Original Message-
> From: Dariusz Sosnowski 
> Sent: Thursday, February 27, 2025 12:04 AM
> To: Dariusz Sosnowski ; Slava Ovsiienko
> ; Bing Zhao ; Ori Kam
> ; Suanming Mou ; Matan Azrad
> 
> Cc: dev@dpdk.org; Raslan Darawsheh 
> Subject: [PATCH v3 0/3] net/mlx5: support jump FDB Rx
> 
> Before FW introduced the JUMP_FDB_RX action feature, jump from FDB Tx to
> Rx is not allowed. JUMP_FDB_RX feature introduces the internal loopback
> for Tx case and allow the REG C0 C1 B be preserved as well.
> 
> This series adds the jump from FDB Tx to FDB Rx support and allows RSS
> action in FDB domain rule when jump FDB Rx is supported.
> 
> v3:
> - Rebased on v2 of mlx5 Unified FDB patches (series 34723)
> 
> Suanming Mou (3):
>   net/mlx5/hws: support jump FDB Rx
>   net/mlx5: add jump FDB Rx flag
>   net/mlx5: allow FDB RSS
> 
>  doc/guides/nics/mlx5.rst  |  4 
>  drivers/common/mlx5/mlx5_devx_cmds.c  |  8 +++
> drivers/common/mlx5/mlx5_devx_cmds.h  |  1 +
>  drivers/common/mlx5/mlx5_prm.h|  9 +++-
>  drivers/net/mlx5/hws/mlx5dr_action.c  | 26 +-
> drivers/net/mlx5/hws/mlx5dr_action.h  |  4 
>  drivers/net/mlx5/hws/mlx5dr_cmd.c |  9 
>  drivers/net/mlx5/hws/mlx5dr_cmd.h |  2 ++
>  drivers/net/mlx5/hws/mlx5dr_context.c | 17 ++
> drivers/net/mlx5/hws/mlx5dr_context.h |  2 ++
>  drivers/net/mlx5/linux/mlx5_os.c  |  9 ++--
>  drivers/net/mlx5/mlx5.h   |  1 +
>  drivers/net/mlx5/mlx5_flow.c  |  4 ++--
>  drivers/net/mlx5/mlx5_flow_dv.c   | 32 ++-
>  drivers/net/mlx5/mlx5_flow_hw.c   | 28 ++-
>  15 files changed, 125 insertions(+), 31 deletions(-)
> 
> --
> 2.39.5



RE: [PATCH v2 1/3] net/mlx5/hws: support jump FDB Rx

2025-02-26 Thread Bing Zhao



> -Original Message-
> From: Suanming Mou 
> Sent: Wednesday, February 26, 2025 10:01 PM
> To: Dariusz Sosnowski ; Slava Ovsiienko
> ; Bing Zhao ; Ori Kam
> ; Matan Azrad 
> Cc: dev@dpdk.org; Raslan Darawsheh ; Alex Vesker
> 
> Subject: [PATCH v2 1/3] net/mlx5/hws: support jump FDB Rx
> 
> Before FW introduced the JUMP_FDB_RX action feature, jump from FDB Tx to
> Rx is not allowed. JUMP_FDB_RX feature introduces the internal loopback
> for Tx case and allow the REG C0 C1 B be preserved as well.
> 
> This commit adds the JUMP_FDB_RX cap bit check and use JUMP_FDB_RX instead
> of FT for dest table FDB Rx case.
> 
> Signed-off-by: Suanming Mou 
> Signed-off-by: Alex Vesker 
> ---
> 
> v2: fix line lengh.
> 
> ---
>  drivers/common/mlx5/mlx5_prm.h|  9 -
>  drivers/net/mlx5/hws/mlx5dr_action.c  | 26 +-
> drivers/net/mlx5/hws/mlx5dr_action.h  |  4 
>  drivers/net/mlx5/hws/mlx5dr_cmd.c |  9 +
>  drivers/net/mlx5/hws/mlx5dr_cmd.h |  2 ++
>  drivers/net/mlx5/hws/mlx5dr_context.c | 17 +
> drivers/net/mlx5/hws/mlx5dr_context.h |  2 ++
>  7 files changed, 63 insertions(+), 6 deletions(-)
> 
> diff --git a/drivers/common/mlx5/mlx5_prm.h
> b/drivers/common/mlx5/mlx5_prm.h index 3fc3b0cd2a..84e3347794 100644
> --- a/drivers/common/mlx5/mlx5_prm.h
> +++ b/drivers/common/mlx5/mlx5_prm.h
> @@ -2466,7 +2466,8 @@ struct mlx5_ifc_wqe_based_flow_table_cap_bits {
>   u8 reserved_at_60[0x8];
>   u8 max_header_modify_pattern_length[0x8];
>   u8 ste_format[0x10];
> - u8 stc_action_type[0x80];
> + u8 stc_action_type_63_0[0x40];
> + u8 stc_action_type_127_64[0x40];
>   u8 header_insert_type[0x10];
>   u8 header_remove_type[0x10];
>   u8 trivial_match_definer[0x20];
> @@ -3543,6 +3544,11 @@ enum mlx5_ifc_rtc_reparse_mode {
>   MLX5_IFC_RTC_REPARSE_BY_STC = 0x2,
>  };
> 
> +enum mlx5_ifc_stc_action_type_bit_index {
> + MLX5_IFC_STC_ACTION_TYPE_BIT_64_INDEX = 64,
> + MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX = 71, };
> +
>  #define MLX5_IFC_RTC_LINEAR_LOOKUP_TBL_LOG_MAX 16
> 
>  struct mlx5_ifc_rtc_bits {
> @@ -3621,6 +3627,7 @@ enum mlx5_ifc_stc_action_type {
>   MLX5_IFC_STC_ACTION_TYPE_ALLOW = 0x84,
>   MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_VPORT = 0x85,
>   MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_UPLINK = 0x86,
> + MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FLOW_TABLE_FDB_RX = 0x87,
>  };
> 
>  enum mlx5_ifc_stc_reparse_mode {
> diff --git a/drivers/net/mlx5/hws/mlx5dr_action.c
> b/drivers/net/mlx5/hws/mlx5dr_action.c
> index b9452a3ebc..e21db5b327 100644
> --- a/drivers/net/mlx5/hws/mlx5dr_action.c
> +++ b/drivers/net/mlx5/hws/mlx5dr_action.c
> @@ -803,6 +803,9 @@ int mlx5dr_action_root_build_attr(struct
> mlx5dr_rule_action rule_actions[],
> 
>   switch (action->type) {
>   case MLX5DR_ACTION_TYP_TBL:
> + attr[i].type = MLX5DV_FLOW_ACTION_DEST_DEVX;
> + attr[i].obj = action->dest_tbl.devx_obj->obj;
> + break;
>   case MLX5DR_ACTION_TYP_TIR:
>   attr[i].type = MLX5DV_FLOW_ACTION_DEST_DEVX;
>   attr[i].obj = action->devx_obj;
> @@ -1097,6 +1100,17 @@ static void mlx5dr_action_fill_stc_attr(struct
> mlx5dr_action *action,
>   }
>   break;
>   case MLX5DR_ACTION_TYP_TBL:
> + attr->action_offset = MLX5DR_ACTION_OFFSET_HIT;
> + attr->dest_table_id = obj->id;
> + /* Only for unified FDB Rx case */
> + if (mlx5dr_context_cap_stc(action->ctx,
> + MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX)
> &&
> + action->dest_tbl.type == MLX5DR_TABLE_TYPE_FDB_RX)
> + attr->action_type =
> MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FLOW_TABLE_FDB_RX;
> + else
> + attr->action_type = MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FT;
> +
> + break;
>   case MLX5DR_ACTION_TYP_DEST_ARRAY:
>   attr->action_type = MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FT;
>   attr->action_offset = MLX5DR_ACTION_OFFSET_HIT; @@ -1419,17
> +1433,19 @@ mlx5dr_action_create_dest_table(struct mlx5dr_context *ctx,
>   if (!action)
>   return NULL;
> 
> + action->dest_tbl.type = tbl->type;
> +
>   if (mlx5dr_action_is_root_flags(flags)) {
>   if (mlx5dr_context_shared_gvmi_used(ctx))
> - action->devx_obj = tbl->local_ft->obj;
> + action->dest_tbl.devx_obj = tbl->local_ft;
>   else
> - action->devx_obj = tbl->ft->obj;
> + action->dest_tbl.devx_obj = tbl->ft;
>   } else {
> + action->dest_tbl.devx_obj = tbl->ft;
> +
>   ret = mlx5dr_action_create_stcs(action, tbl->ft);
>   if (ret)
>   goto free_action;
> -
> - action->devx_dest.dev

[PATCH v12 2/2] net/af_xdp: Refactor af_xdp_tx_zc

2025-02-26 Thread Ariel Otilibili
Both legs of the loop share the same logic: the common parts are about
reserving and filling both address and length into the description.

This is moved into reserve_and_fill().

Bugzilla ID: 1440
Suggested-by: Maryam Tahhan 
Signed-off-by: Ariel Otilibili 
Acked-by: Maryam Tahhan 
---
 drivers/net/af_xdp/rte_eth_af_xdp.c | 75 -
 1 file changed, 42 insertions(+), 33 deletions(-)

diff --git a/drivers/net/af_xdp/rte_eth_af_xdp.c 
b/drivers/net/af_xdp/rte_eth_af_xdp.c
index 092bcb73aa0a..05115150a7b9 100644
--- a/drivers/net/af_xdp/rte_eth_af_xdp.c
+++ b/drivers/net/af_xdp/rte_eth_af_xdp.c
@@ -536,21 +536,49 @@ kick_tx(struct pkt_tx_queue *txq, struct xsk_ring_cons 
*cq)
}
 }
 
+static inline struct xdp_desc *
+reserve_and_fill(struct pkt_tx_queue *txq, struct rte_mbuf *mbuf,
+struct xsk_umem_info *umem, void **pkt_ptr)
+{
+   struct xdp_desc *desc = NULL;
+   uint64_t addr, offset;
+   uint32_t idx_tx;
+
+   if (!xsk_ring_prod__reserve(&txq->tx, 1, &idx_tx))
+   goto out;
+
+   desc = xsk_ring_prod__tx_desc(&txq->tx, idx_tx);
+   desc->len = mbuf->pkt_len;
+
+   addr = (uint64_t)mbuf - (uint64_t)umem->buffer
+   - umem->mb_pool->header_size;
+   offset = rte_pktmbuf_mtod(mbuf, uint64_t) - (uint64_t)mbuf
+   + umem->mb_pool->header_size;
+
+   if (pkt_ptr)
+   *pkt_ptr = xsk_umem__get_data(umem->buffer, addr + offset);
+
+   offset = offset << XSK_UNALIGNED_BUF_OFFSET_SHIFT;
+   desc->addr = addr | offset;
+
+out:
+   return desc;
+}
+
 #if defined(XDP_UMEM_UNALIGNED_CHUNK_FLAG)
 static uint16_t
 af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
 {
struct pkt_tx_queue *txq = queue;
struct xsk_umem_info *umem = txq->umem;
-   struct rte_mbuf *mbuf;
+   struct rte_mbuf *mbuf, *local_mbuf = NULL;
unsigned long tx_bytes = 0;
int i;
-   uint32_t idx_tx;
uint16_t count = 0;
struct xdp_desc *desc;
-   uint64_t addr, offset;
struct xsk_ring_cons *cq = &txq->pair->cq;
uint32_t free_thresh = cq->size >> 1;
+   void *pkt;
 
if (xsk_cons_nb_avail(cq, free_thresh) >= free_thresh)
pull_umem_cq(umem, XSK_RING_CONS__DEFAULT_NUM_DESCS, cq);
@@ -559,51 +587,32 @@ af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, 
uint16_t nb_pkts)
mbuf = bufs[i];
 
if (mbuf->pool == umem->mb_pool) {
-   if (!xsk_ring_prod__reserve(&txq->tx, 1, &idx_tx)) {
+   desc = reserve_and_fill(txq, mbuf, umem, NULL);
+   if (!desc) {
kick_tx(txq, cq);
-   if (!xsk_ring_prod__reserve(&txq->tx, 1,
-   &idx_tx))
+   desc = reserve_and_fill(txq, mbuf, umem, NULL);
+   if (!desc)
goto out;
}
-   desc = xsk_ring_prod__tx_desc(&txq->tx, idx_tx);
-   desc->len = mbuf->pkt_len;
-   addr = (uint64_t)mbuf - (uint64_t)umem->buffer -
-   umem->mb_pool->header_size;
-   offset = rte_pktmbuf_mtod(mbuf, uint64_t) -
-   (uint64_t)mbuf +
-   umem->mb_pool->header_size;
-   offset = offset << XSK_UNALIGNED_BUF_OFFSET_SHIFT;
-   desc->addr = addr | offset;
+
tx_bytes += desc->len;
count++;
} else {
-   struct rte_mbuf *local_mbuf =
-   rte_pktmbuf_alloc(umem->mb_pool);
-   void *pkt;
-
-   if (local_mbuf == NULL)
+   local_mbuf = rte_pktmbuf_alloc(umem->mb_pool);
+   if (!local_mbuf)
goto out;
 
-   if (!xsk_ring_prod__reserve(&txq->tx, 1, &idx_tx)) {
+   desc = reserve_and_fill(txq, local_mbuf, umem, &pkt);
+   if (!desc) {
rte_pktmbuf_free(local_mbuf);
goto out;
}
 
-   desc = xsk_ring_prod__tx_desc(&txq->tx, idx_tx);
desc->len = mbuf->pkt_len;
-
-   addr = (uint64_t)local_mbuf - (uint64_t)umem->buffer -
-   umem->mb_pool->header_size;
-   offset = rte_pktmbuf_mtod(local_mbuf, uint64_t) -
-   (uint64_t)local_mbuf +
-   umem->mb_pool->header_size;
-   pkt = xsk_umem__g

[PATCH v12 0/2] Fix use after free, and refactor af_xdp_tx_zc

2025-02-26 Thread Ariel Otilibili
Hello,

The series addresses Bugzilla ID 1440 in two steps;

1. Fix use after free.
2. Refactor af_xdp_tx_zc().

Thank you,
---
v12
* fixed style issues 
(http://mails.dpdk.org/archives/test-report/2025-February/857083.html)

v11
* removed assignments in if statements (Stephen Hemminger)

v10 
(https://inbox.dpdk.org/dev/20250223215259.448723-1-ariel.otilib...@6wind.com/)
* no change

v9 
(https://inbox.dpdk.org/dev/20250207104552.1663519-1-ariel.otilib...@6wind.com/)
* desc->len takes its value from mbuf (Maryam Tahhan)

v8 
(https://inbox.dpdk.org/dev/20250206204645.1564535-1-ariel.otilib...@6wind.com/)
* fixed seg fault 
(https://inbox.dpdk.org/dev/CAFdtZitaNGhC5Q10ATNa7xXX1JbuWob=yzrcwmq8ltz+qji...@mail.gmail.com/)
* addressed feedback from Maryam Tahhan

v7 
(https://inbox.dpdk.org/dev/20250201100300.2194018-1-ariel.otilib...@6wind.com/)
* no change
* resent because CI didn't take patch-2 into the series

v6 
(https://inbox.dpdk.org/dev/20250131231018.2163893-1-ariel.otilib...@6wind.com/)
* added missing credits to Maryam Tahhan in v5 

v5 
(https://inbox.dpdk.org/dev/20250131183439.909831-1-ariel.otilib...@6wind.com/)
* fix style issues in the signature of reserve_and_fill() (Stephen Hemminger)

v4 
(https://inbox.dpdk.org/dev/20250130221853.789366-1-ariel.otilib...@6wind.com/)
* redid the refactor (Maryam Tahhan)
* marked the fix as acked (Stephen Hemminger)
* updated .mailmap, my main e-mail is @6wind.com 
(https://inbox.dpdk.org/dev/20250115121152.487360-4-otili...@eurecom.fr/)

v3 
(https://inbox.dpdk.org/dev/20250128231152.249497-1-ariel.otilib...@6wind.com/)
* reworded commit message of patch-1
* addressed feedback of Maryam Tahhan

v2 
(https://inbox.dpdk.org/dev/20250116225151.188214-1-ariel.otilib...@6wind.com/)
* reworded commit messages
* addressed feedback from Stephen Hemminger

v1 (http://inbox.dpdk.org/dev/20250116195640.68885-1-ariel.otilib...@6wind.com/)

Ariel Otilibili (2):
  net/af_xdp: Fix use after free in af_xdp_tx_zc
  net/af_xdp: Refactor af_xdp_tx_zc

 .mailmap|  2 +-
 drivers/net/af_xdp/rte_eth_af_xdp.c | 77 -
 2 files changed, 44 insertions(+), 35 deletions(-)

-- 
2.30.2



[PATCH v11 0/2] Fix use after free, and refactor af_xdp_tx_zc

2025-02-26 Thread Ariel Otilibili
Hello,

The series addresses Bugzilla ID 1440 in two steps;

1. Fix use after free.
2. Refactor af_xdp_tx_zc().

Thank you,
---
v11
* removed assignments in if statements (Stephen Hemminger)

v10 
(https://inbox.dpdk.org/dev/20250223215259.448723-1-ariel.otilib...@6wind.com/)
* no change

v9 
(https://inbox.dpdk.org/dev/20250207104552.1663519-1-ariel.otilib...@6wind.com/)
* desc->len takes its value from mbuf (Maryam Tahhan)

v8 
(https://inbox.dpdk.org/dev/20250206204645.1564535-1-ariel.otilib...@6wind.com/)
* fixed seg fault 
(https://inbox.dpdk.org/dev/CAFdtZitaNGhC5Q10ATNa7xXX1JbuWob=yzrcwmq8ltz+qji...@mail.gmail.com/)
* addressed feedback from Maryam Tahhan

v7 
(https://inbox.dpdk.org/dev/20250201100300.2194018-1-ariel.otilib...@6wind.com/)
* no change
* resent because CI didn't take patch-2 into the series

v6 
(https://inbox.dpdk.org/dev/20250131231018.2163893-1-ariel.otilib...@6wind.com/)
* added missing credits to Maryam Tahhan in v5 

v5 
(https://inbox.dpdk.org/dev/20250131183439.909831-1-ariel.otilib...@6wind.com/)
* fix style issues in the signature of reserve_and_fill() (Stephen Hemminger)

v4 
(https://inbox.dpdk.org/dev/20250130221853.789366-1-ariel.otilib...@6wind.com/)
* redid the refactor (Maryam Tahhan)
* marked the fix as acked (Stephen Hemminger)
* updated .mailmap, my main e-mail is @6wind.com 
(https://inbox.dpdk.org/dev/20250115121152.487360-4-otili...@eurecom.fr/)

v3 
(https://inbox.dpdk.org/dev/20250128231152.249497-1-ariel.otilib...@6wind.com/)
* reworded commit message of patch-1
* addressed feedback of Maryam Tahhan

v2 
(https://inbox.dpdk.org/dev/20250116225151.188214-1-ariel.otilib...@6wind.com/)
* reworded commit messages
* addressed feedback from Stephen Hemminger

v1 (http://inbox.dpdk.org/dev/20250116195640.68885-1-ariel.otilib...@6wind.com/)

Ariel Otilibili (2):
  net/af_xdp: Fix use after free in af_xdp_tx_zc
  net/af_xdp: Refactor af_xdp_tx_zc

 .mailmap|  2 +-
 drivers/net/af_xdp/rte_eth_af_xdp.c | 77 -
 2 files changed, 44 insertions(+), 35 deletions(-)

-- 
2.30.2



Re: [PATCH v6 2/2] net/af_xdp: Refactor af_xdp_tx_zc

2025-02-26 Thread Ariel Otilibili
Hello Stephen,

On Mon, Feb 24, 2025 at 8:25 PM Stephen Hemminger <
step...@networkplumber.org> wrote:

>
> Please avoid doing assignment in a conditional statement, can be error
> prone.
> Surprised checkpatch doesn't complain about it.
>

Thanks for the feedback. checkpatch did catch them.

Here is the new series,
https://inbox.dpdk.org/dev/20250226200841.2342632-3-ariel.otilib...@6wind.com/

Regards,
Ariel

>
> Better as:
> desc = reserve_and_fill(txq, mbuf, umem);
> if (!desc) {
> kick_tx(txq, cq);
> desc = reserve_and_fill(txq, mbuf, umem);
> if (!desc)
> goto out;
>


[PATCH v12 1/2] net/af_xdp: Fix use after free in af_xdp_tx_zc

2025-02-26 Thread Ariel Otilibili
tx_bytes is computed after both legs are tested. This might
produce a use after memory free.

The computation is now moved into each leg.

Bugzilla ID: 1440
Fixes: d8a210774e1d ("net/af_xdp: support unaligned umem chunks")
Signed-off-by: Ariel Otilibili 
Acked-by: Stephen Hemminger 
---
 .mailmap| 2 +-
 drivers/net/af_xdp/rte_eth_af_xdp.c | 4 ++--
 2 files changed, 3 insertions(+), 3 deletions(-)

diff --git a/.mailmap b/.mailmap
index a03d3cfb591b..ea68d6180ccc 100644
--- a/.mailmap
+++ b/.mailmap
@@ -135,7 +135,7 @@ Anupam Kapoor 
 Apeksha Gupta 
 Archana Muniganti  
 Archit Pandey 
-Ariel Otilibili  
+Ariel Otilibili  
 Arkadiusz Kubalewski 
 Arkadiusz Kusztal 
 Arnaud Fiorini 
diff --git a/drivers/net/af_xdp/rte_eth_af_xdp.c 
b/drivers/net/af_xdp/rte_eth_af_xdp.c
index 814398ba4b44..092bcb73aa0a 100644
--- a/drivers/net/af_xdp/rte_eth_af_xdp.c
+++ b/drivers/net/af_xdp/rte_eth_af_xdp.c
@@ -574,6 +574,7 @@ af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, uint16_t 
nb_pkts)
umem->mb_pool->header_size;
offset = offset << XSK_UNALIGNED_BUF_OFFSET_SHIFT;
desc->addr = addr | offset;
+   tx_bytes += desc->len;
count++;
} else {
struct rte_mbuf *local_mbuf =
@@ -601,11 +602,10 @@ af_xdp_tx_zc(void *queue, struct rte_mbuf **bufs, 
uint16_t nb_pkts)
desc->addr = addr | offset;
rte_memcpy(pkt, rte_pktmbuf_mtod(mbuf, void *),
desc->len);
+   tx_bytes += desc->len;
rte_pktmbuf_free(mbuf);
count++;
}
-
-   tx_bytes += mbuf->pkt_len;
}
 
 out:
-- 
2.30.2



[PATCH v3 3/3] config: create top level variable cc_avx2_flags

2025-02-26 Thread Andre Muezerie
Top level variable cc_avx2_flags was created and set according to the
compiler used. It is used in lower level directories like drivers/net
and lib/acl.

Signed-off-by: Andre Muezerie 
Acked-by: Bruce Richardson 
---
 config/x86/meson.build | 6 ++
 drivers/net/bnxt/meson.build   | 2 +-
 drivers/net/enic/meson.build   | 2 +-
 drivers/net/intel/i40e/meson.build | 2 +-
 drivers/net/intel/iavf/meson.build | 2 +-
 drivers/net/intel/ice/meson.build  | 2 +-
 drivers/net/intel/idpf/meson.build | 2 +-
 drivers/net/nfp/meson.build| 2 +-
 drivers/net/octeon_ep/meson.build  | 4 ++--
 lib/acl/meson.build| 2 +-
 10 files changed, 16 insertions(+), 10 deletions(-)

diff --git a/config/x86/meson.build b/config/x86/meson.build
index 47a5b0c04a..941d4d921f 100644
--- a/config/x86/meson.build
+++ b/config/x86/meson.build
@@ -1,6 +1,12 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright(c) 2017-2020 Intel Corporation
 
+if is_ms_compiler
+cc_avx2_flags = ['/arch:AVX2']
+else
+cc_avx2_flags = ['-mavx2']
+endif
+
 # get binutils version for the workaround of Bug 97
 binutils_ok = true
 if is_linux or cc.get_id() == 'gcc'
diff --git a/drivers/net/bnxt/meson.build b/drivers/net/bnxt/meson.build
index e26cf13a65..fd82d0c409 100644
--- a/drivers/net/bnxt/meson.build
+++ b/drivers/net/bnxt/meson.build
@@ -65,7 +65,7 @@ if arch_subdir == 'x86'
 static_rte_bus_pci,
 static_rte_kvargs, static_rte_hash],
 include_directories: includes,
-c_args: [cflags, '-mavx2'])
+c_args: [cflags, cc_avx2_flags])
  objs += bnxt_avx2_lib.extract_objects('bnxt_rxtx_vec_avx2.c')
 elif arch_subdir == 'arm' and dpdk_conf.get('RTE_ARCH_64')
 sources += files('bnxt_rxtx_vec_neon.c')
diff --git a/drivers/net/enic/meson.build b/drivers/net/enic/meson.build
index 1e26338350..cfe5ec170a 100644
--- a/drivers/net/enic/meson.build
+++ b/drivers/net/enic/meson.build
@@ -38,7 +38,7 @@ if dpdk_conf.has('RTE_ARCH_X86_64')
 'enic_rxtx_vec_avx2.c',
 dependencies: [static_rte_ethdev, static_rte_bus_pci],
 include_directories: includes,
-c_args: [cflags, '-mavx2'])
+c_args: [cflags, cc_avx2_flags])
 objs += enic_avx2_lib.extract_objects('enic_rxtx_vec_avx2.c')
 endif
 
diff --git a/drivers/net/intel/i40e/meson.build 
b/drivers/net/intel/i40e/meson.build
index ffa40c5d64..2973ed1a01 100644
--- a/drivers/net/intel/i40e/meson.build
+++ b/drivers/net/intel/i40e/meson.build
@@ -49,7 +49,7 @@ if arch_subdir == 'x86'
 'i40e_rxtx_vec_avx2.c',
 dependencies: [static_rte_ethdev, static_rte_kvargs, 
static_rte_hash],
 include_directories: includes,
-c_args: [cflags, '-mavx2'])
+c_args: [cflags, cc_avx2_flags])
 objs += i40e_avx2_lib.extract_objects('i40e_rxtx_vec_avx2.c')
 
 if cc_has_avx512
diff --git a/drivers/net/intel/iavf/meson.build 
b/drivers/net/intel/iavf/meson.build
index 19cd1cfbc8..f7eac7c57a 100644
--- a/drivers/net/intel/iavf/meson.build
+++ b/drivers/net/intel/iavf/meson.build
@@ -37,7 +37,7 @@ if arch_subdir == 'x86'
 'iavf_rxtx_vec_avx2.c',
 dependencies: [static_rte_ethdev],
 include_directories: includes,
-c_args: [cflags, '-mavx2'])
+c_args: [cflags, cc_avx2_flags])
 objs += iavf_avx2_lib.extract_objects('iavf_rxtx_vec_avx2.c')
 
 if cc_has_avx512
diff --git a/drivers/net/intel/ice/meson.build 
b/drivers/net/intel/ice/meson.build
index a34b7c966a..cbdf38c1c4 100644
--- a/drivers/net/intel/ice/meson.build
+++ b/drivers/net/intel/ice/meson.build
@@ -43,7 +43,7 @@ if arch_subdir == 'x86'
 'ice_rxtx_vec_avx2.c',
 dependencies: [static_rte_ethdev, static_rte_hash],
 include_directories: includes,
-c_args: [cflags, '-mavx2'])
+c_args: [cflags, cc_avx2_flags])
 objs += ice_avx2_lib.extract_objects('ice_rxtx_vec_avx2.c')
 
 if cc_has_avx512
diff --git a/drivers/net/intel/idpf/meson.build 
b/drivers/net/intel/idpf/meson.build
index 802b13035b..4b272d02b1 100644
--- a/drivers/net/intel/idpf/meson.build
+++ b/drivers/net/intel/idpf/meson.build
@@ -23,7 +23,7 @@ if arch_subdir == 'x86' and dpdk_conf.get('RTE_IOVA_IN_MBUF') 
== 1
 'idpf_common_rxtx_avx2.c',
 dependencies: [static_rte_ethdev, static_rte_hash],
 include_directories: includes,
-c_args: [cflags, '-mavx2'])
+c_args: [cflags, cc_avx2_flags])
 objs += idpf_avx2_lib.extract_objects('idpf_common_rxtx_avx2.c')
 
 if cc_has_avx512
diff --git a/drivers/net/nfp/meson.build b/drivers/net/nfp/meson.build
index 39762bd45a..0a12b7dce7 100644
--- a/drivers/net/nfp/meson.build
+++ b/drivers/net/nfp/meson.build
@@ -61,7 +61,7 @@ if arch_subdir == 'x86'
 avx2_sources,
 dependencies: [static_rte_ethdev, static_rte_bus_pci],
 include_director

[PATCH v3 0/3] allow faster instruction sets to be used with MSVC

2025-02-26 Thread Andre Muezerie
Meson.build files were refactored to add MSVC support for the AVX2
instruction set with MSVC: Top level variable cc_avx2_flags was created
and set according to the compiler used. It is used in lower level
directories.

Andre Muezerie (3):
  eal: make compatible with instruction set updates for MSVC
  eal: only use numbers as align parameters for MSVC
  config: create top level variable cc_avx2_flags

 config/x86/meson.build |  6 ++
 drivers/net/bnxt/meson.build   |  2 +-
 drivers/net/enic/meson.build   |  2 +-
 drivers/net/intel/i40e/meson.build |  2 +-
 drivers/net/intel/iavf/meson.build |  2 +-
 drivers/net/intel/ice/meson.build  |  2 +-
 drivers/net/intel/idpf/meson.build |  2 +-
 drivers/net/nfp/meson.build|  2 +-
 drivers/net/octeon_ep/meson.build  |  4 ++--
 lib/acl/meson.build|  2 +-
 lib/eal/common/rte_random.c|  2 ++
 lib/eal/x86/include/rte_vect.h | 11 ++-
 12 files changed, 28 insertions(+), 11 deletions(-)

--
2.48.1.vfs.0.0



[PATCH v3 1/3] eal: make compatible with instruction set updates for MSVC

2025-02-26 Thread Andre Muezerie
After the instruction set updates for MSVC the error below popped up:

../lib/eal/common/rte_random.c(6): fatal error C1083:
Cannot open include file: 'x86intrin.h': No such file or directory

The fix is to not include header x86intrin.h with MSVC.

Signed-off-by: Andre Muezerie 
Acked-by: Bruce Richardson 
---
 lib/eal/common/rte_random.c | 2 ++
 1 file changed, 2 insertions(+)

diff --git a/lib/eal/common/rte_random.c b/lib/eal/common/rte_random.c
index 8e62578176..9354358818 100644
--- a/lib/eal/common/rte_random.c
+++ b/lib/eal/common/rte_random.c
@@ -3,8 +3,10 @@
  */
 
 #ifdef __RDSEED__
+#ifndef RTE_TOOLCHAIN_MSVC
 #include 
 #endif
+#endif
 #include 
 
 #include 
-- 
2.48.1.vfs.0.0



[PATCH v3 2/3] eal: only use numbers as align parameters for MSVC

2025-02-26 Thread Andre Muezerie
After the instruction set updates for MSVC the error below popped up:

..\lib\eal\x86\include\rte_vect.h(82): error C2059: syntax error: '('

The issue is that MSVC does not allow __rte_aligned(RTE_X86_ZMM_SIZE).
It only accepts numbers that are power of 2. So, even though
RTE_X86_ZMM_SIZE represents a number that is a power of two it cannot
be used directly.
https://learn.microsoft.com/en-us/cpp/cpp/align-cpp?view=msvc-170

Signed-off-by: Andre Muezerie 
Acked-by: Bruce Richardson 
---
 lib/eal/x86/include/rte_vect.h | 11 ++-
 1 file changed, 10 insertions(+), 1 deletion(-)

diff --git a/lib/eal/x86/include/rte_vect.h b/lib/eal/x86/include/rte_vect.h
index 70c78e9b77..9ea158b27e 100644
--- a/lib/eal/x86/include/rte_vect.h
+++ b/lib/eal/x86/include/rte_vect.h
@@ -76,9 +76,18 @@ __extension__ ({\
 
 #ifdef __AVX512F__
 
-#define RTE_X86_ZMM_SIZE   (sizeof(__m512i))
+#define RTE_X86_ZMM_SIZE64
 #define RTE_X86_ZMM_MASK   (RTE_X86_ZMM_SIZE - 1)
 
+/*
+ * MSVC does not allow __rte_aligned(sizeof(__m512i)). It only accepts
+ * numbers that are power of 2. So, even though sizeof(__m512i) represents a
+ * number that is a power of two it cannot be used directly.
+ * Ref: https://learn.microsoft.com/en-us/cpp/cpp/align-cpp?view=msvc-170
+ * The static assert below ensures that the hardcoded value defined as
+ * RTE_X86_ZMM_SIZE is equal to sizeof(__m512i).
+ */
+static_assert(RTE_X86_ZMM_SIZE == (sizeof(__m512i)), "Unexpected size of 
__m512i");
 typedef union __rte_aligned(RTE_X86_ZMM_SIZE) __rte_x86_zmm {
__m512i  z;
ymm_ty[RTE_X86_ZMM_SIZE / sizeof(ymm_t)];
-- 
2.48.1.vfs.0.0



Re: [PATCH] net/mlx5: fix queue counter error check

2025-02-26 Thread Raslan Darawsheh
Hi,

From: Dariusz Sosnowski 
Sent: Tuesday, February 25, 2025 11:08 AM
To: Slava Ovsiienko; Bing Zhao; Ori Kam; Suanming Mou; Matan Azrad; Shani Peretz
Cc: dev@dpdk.org; Raslan Darawsheh
Subject: [PATCH] net/mlx5: fix queue counter error check

Whenever queue counter allocation fails, the FW syndrome error
should be checked to determine if maximum number of queue counters
was reached.

Fixes: f0c0731b6d40 ("net/mlx5: add counters for hairpin drop")
Cc: shper...@nvidia.com

Signed-off-by: Dariusz Sosnowski 
Acked-by: Bing Zhao 

Patch applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh




[PATCH v16 14/29] net/rnp: add Rx burst simple support

2025-02-26 Thread Wenbo Cao
add only support simple recv pkts.

Signed-off-by: Wenbo Cao 
---
 drivers/net/rnp/rnp_ethdev.c |   7 ++
 drivers/net/rnp/rnp_rxtx.c   | 130 +++
 drivers/net/rnp/rnp_rxtx.h   |   5 ++
 3 files changed, 142 insertions(+)

diff --git a/drivers/net/rnp/rnp_ethdev.c b/drivers/net/rnp/rnp_ethdev.c
index 089e90543c..f1ec2ad73a 100644
--- a/drivers/net/rnp/rnp_ethdev.c
+++ b/drivers/net/rnp/rnp_ethdev.c
@@ -335,6 +335,8 @@ static int rnp_dev_start(struct rte_eth_dev *eth_dev)
goto rxq_start_failed;
/* enable eth rx flow */
RNP_RX_ETH_ENABLE(hw, lane);
+   rnp_rx_func_select(eth_dev);
+   rnp_tx_func_select(eth_dev);
port->port_stopped = 0;
 
return 0;
@@ -595,6 +597,11 @@ static int rnp_dev_infos_get(struct rte_eth_dev *eth_dev,
/* default port configure */
dev_info->default_rxconf = (struct rte_eth_rxconf) {
.rx_drop_en = 0,
+   .rx_thresh = {
+   .pthresh = RNP_RX_DESC_FETCH_TH,
+   .hthresh = RNP_RX_DESC_FETCH_BURST,
+   },
+   .rx_free_thresh = RNP_DEFAULT_RX_FREE_THRESH,
.offloads = 0,
};
 
diff --git a/drivers/net/rnp/rnp_rxtx.c b/drivers/net/rnp/rnp_rxtx.c
index e65bc06d36..a742646c7e 100644
--- a/drivers/net/rnp/rnp_rxtx.c
+++ b/drivers/net/rnp/rnp_rxtx.c
@@ -636,3 +636,133 @@ int rnp_rx_queue_start(struct rte_eth_dev *eth_dev, 
uint16_t qidx)
 
return 0;
 }
+
+#define RNP_CACHE_FETCH_RX (4)
+static __rte_always_inline int
+rnp_refill_rx_ring(struct rnp_rx_queue *rxq)
+{
+   volatile struct rnp_rx_desc *rxbd;
+   struct rnp_rxsw_entry *rx_swbd;
+   struct rte_eth_dev_data *data;
+   struct rte_mbuf *mb;
+   uint16_t j, i;
+   uint16_t rx_id;
+   int ret;
+
+   rxbd = rxq->rx_bdr + rxq->rxrearm_start;
+   rx_swbd = &rxq->sw_ring[rxq->rxrearm_start];
+   ret = rte_mempool_get_bulk(rxq->mb_pool, (void *)rx_swbd,
+   rxq->rx_free_thresh);
+   data = rte_eth_devices[rxq->attr.port_id].data;
+   if (unlikely(ret != 0)) {
+   if (rxq->rxrearm_nb + rxq->rx_free_thresh >= rxq->attr.nb_desc) 
{
+   for (i = 0; i < RNP_CACHE_FETCH_RX; i++) {
+   rx_swbd[i].mbuf = &rxq->fake_mbuf;
+   rxbd[i].d.pkt_addr = 0;
+   rxbd[i].d.cmd = 0;
+   }
+   }
+   data->rx_mbuf_alloc_failed += rxq->rx_free_thresh;
+   return 0;
+   }
+   for (j = 0; j < rxq->rx_free_thresh; ++j) {
+   mb = rx_swbd[j].mbuf;
+   rte_mbuf_refcnt_set(mb, 1);
+   mb->data_off = RTE_PKTMBUF_HEADROOM;
+   mb->port = rxq->attr.port_id;
+
+   rxbd[j].d.pkt_addr = rnp_get_dma_addr(&rxq->attr, mb);
+   rxbd[j].d.cmd = 0;
+   }
+   rxq->rxrearm_start += rxq->rx_free_thresh;
+   if (rxq->rxrearm_start >= rxq->attr.nb_desc - 1)
+   rxq->rxrearm_start = 0;
+   rxq->rxrearm_nb -= rxq->rx_free_thresh;
+
+   rx_id = (uint16_t)((rxq->rxrearm_start == 0) ?
+   (rxq->attr.nb_desc - 1) : (rxq->rxrearm_start - 1));
+   rte_wmb();
+   RNP_REG_WR(rxq->rx_tailreg, 0, rx_id);
+
+   return j;
+}
+
+static __rte_always_inline uint16_t
+rnp_recv_pkts(void *_rxq, struct rte_mbuf **rx_pkts, uint16_t nb_pkts)
+{
+   struct rnp_rx_queue *rxq = (struct rnp_rx_queue *)_rxq;
+   struct rnp_rxsw_entry *rx_swbd;
+   uint32_t state_cmd[RNP_CACHE_FETCH_RX];
+   uint32_t pkt_len[RNP_CACHE_FETCH_RX] = {0};
+   volatile struct rnp_rx_desc *rxbd;
+   struct rte_mbuf *nmb;
+   int nb_dd, nb_rx = 0;
+   int i, j;
+
+   if (unlikely(!rxq->rxq_started || !rxq->rx_link))
+   return 0;
+   nb_pkts = RTE_ALIGN_FLOOR(nb_pkts, RNP_CACHE_FETCH_RX);
+   rxbd = &rxq->rx_bdr[rxq->rx_tail];
+   rte_prefetch0(rxbd);
+   if (rxq->rxrearm_nb > rxq->rx_free_thresh)
+   rnp_refill_rx_ring(rxq);
+
+   if (!(rxbd->wb.qword1.cmd & RNP_CMD_DD))
+   return 0;
+
+   rx_swbd = &rxq->sw_ring[rxq->rx_tail];
+   for (i = 0; i < nb_pkts;
+   i += RNP_CACHE_FETCH_RX, rxbd += RNP_CACHE_FETCH_RX,
+   rx_swbd += RNP_CACHE_FETCH_RX) {
+   for (j = 0; j < RNP_CACHE_FETCH_RX; j++)
+   state_cmd[j] = rxbd[j].wb.qword1.cmd;
+   rte_atomic_thread_fence(rte_memory_order_acquire);
+
+   for (nb_dd = 0; nb_dd < RNP_CACHE_FETCH_RX &&
+   (state_cmd[nb_dd] & 
rte_cpu_to_le_16(RNP_CMD_DD));
+   nb_dd++)
+   ;
+   for (j = 0; j < nb_dd; j++)
+   pkt_len[j] = rxbd[j].wb.qword1.lens;
+
+   for (j = 0; j < nb_dd; ++j) {
+ 

[PATCH 2/2] net/mlx5: add MPLSoGRE matching in HWS

2025-02-26 Thread Maayan Kashani
Remove limitations in validation and MPLS match translation.

For non relaxed mode, assume MPLSoUDP if no protocol is mentioned.
Set UDP dest port to MPLS type only for UDP protocol.

Signed-off-by: Maayan Kashani 
Acked-by: Dariusz Sosnowski 
---
 doc/guides/nics/mlx5.rst  |  2 --
 drivers/net/mlx5/hws/mlx5dr_definer.c | 28 +--
 drivers/net/mlx5/mlx5_flow.c  |  7 ---
 3 files changed, 13 insertions(+), 24 deletions(-)

diff --git a/doc/guides/nics/mlx5.rst b/doc/guides/nics/mlx5.rst
index 316c3397d2e..204572c885e 100644
--- a/doc/guides/nics/mlx5.rst
+++ b/doc/guides/nics/mlx5.rst
@@ -356,8 +356,6 @@ Limitations
 
 - L3 VXLAN and VXLAN-GPE tunnels cannot be supported together with MPLSoGRE 
and MPLSoUDP.
 
-- MPLSoGRE is not supported in HW steering (``dv_flow_en`` = 2).
-
 - MPLSoUDP with multiple MPLS headers is only supported in HW steering 
(``dv_flow_en`` = 2).
 
 - Match on Geneve header supports the following fields only:
diff --git a/drivers/net/mlx5/hws/mlx5dr_definer.c 
b/drivers/net/mlx5/hws/mlx5dr_definer.c
index 837e0c47bda..ba4053c552f 100644
--- a/drivers/net/mlx5/hws/mlx5dr_definer.c
+++ b/drivers/net/mlx5/hws/mlx5dr_definer.c
@@ -1644,14 +1644,9 @@ mlx5dr_definer_conv_item_mpls(struct 
mlx5dr_definer_conv_data *cd,
 {
const struct rte_flow_item_mpls *m = item->mask;
struct mlx5dr_definer_fc *fc;
-   bool inner = cd->tunnel;
-
-   if (inner) {
-   DR_LOG(ERR, "Inner MPLS item not supported");
-   rte_errno = ENOTSUP;
-   return rte_errno;
-   }
+   bool is_udp;
 
+   /* If no protocol is set - assume MPLSoUDP */
if (!cd->relaxed) {
/* In order to match on MPLS we must match on ip_protocol and 
l4_dport. */
fc = &cd->fc[DR_CALC_FNAME(IP_PROTOCOL, false)];
@@ -1661,14 +1656,17 @@ mlx5dr_definer_conv_item_mpls(struct 
mlx5dr_definer_conv_data *cd,
fc->tag_set = &mlx5dr_definer_udp_protocol_set;
DR_CALC_SET(fc, eth_l2, l4_type_bwc, false);
}
-
-   /* Currently support only MPLSoUDP */
-   fc = &cd->fc[DR_CALC_FNAME(L4_DPORT, false)];
-   if (!fc->tag_set) {
-   fc->item_idx = item_idx;
-   fc->tag_mask_set = &mlx5dr_definer_ones_set;
-   fc->tag_set = &mlx5dr_definer_mpls_udp_port_set;
-   DR_CALC_SET(fc, eth_l4, destination_port, false);
+   is_udp = (fc->tag_set == &mlx5dr_definer_udp_protocol_set);
+
+   if (is_udp) {
+   /* Set UDP dest port to MPLS. */
+   fc = &cd->fc[DR_CALC_FNAME(L4_DPORT, false)];
+   if (!fc->tag_set) {
+   fc->item_idx = item_idx;
+   fc->tag_mask_set = &mlx5dr_definer_ones_set;
+   fc->tag_set = &mlx5dr_definer_mpls_udp_port_set;
+   DR_CALC_SET(fc, eth_l4, destination_port, 
false);
+   }
}
}
 
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 3fbe89a9d42..69d7621d0bf 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -3812,13 +3812,6 @@ mlx5_flow_validate_item_mpls(struct rte_eth_dev *dev 
__rte_unused,
return rte_flow_error_set(error, ENOTSUP,
  RTE_FLOW_ERROR_TYPE_ITEM, 
item,
  "multiple tunnel layers not 
supported");
-   } else {
-   /* Multi-tunnel isn't allowed but MPLS over GRE is an 
exception. */
-   if ((item_flags & MLX5_FLOW_LAYER_TUNNEL) &&
-   !(item_flags & MLX5_FLOW_LAYER_MPLS))
-   return rte_flow_error_set(error, ENOTSUP,
- RTE_FLOW_ERROR_TYPE_ITEM, 
item,
- "multiple tunnel layers not 
supported");
}
if (!mask)
mask = nic_mask;
-- 
2.21.0



[PATCH 1/2] doc: fix MPLS limitations under HWS

2025-02-26 Thread Maayan Kashani
Update missing limitations.

Fixes: 1e2a51f545b6 ("net/mlx5: support MPLS tunnel with HWS")
Cc: sta...@dpdk.org

Signed-off-by: Maayan Kashani 
Acked-by: Dariusz Sosnowski 
---
 doc/guides/nics/mlx5.rst | 8 
 1 file changed, 8 insertions(+)

diff --git a/doc/guides/nics/mlx5.rst b/doc/guides/nics/mlx5.rst
index 690b46a9743..316c3397d2e 100644
--- a/doc/guides/nics/mlx5.rst
+++ b/doc/guides/nics/mlx5.rst
@@ -279,6 +279,14 @@ Limitations
which results in L3 encap.
 d. Only in transfer (switchdev) mode.
 
+  - MPLS:
+
+- RTE_FLOW_ITEM_TYPE_MPLS matching is not supported for mlxcfg 
``FLEX_PARSER_PROFILE_ENABLE`` = 1.
+- RTE_FLOW_ITEM_TYPE_MPLS matching is not supported on group 0.
+- Testpmd commands "set mplsoudp_encap" and "set mplsoudp_decap" support 
only one MPLS.
+- For non relaxed pattern template, MPLSoUDP match is assumed when trying 
to match MPLS w/o former
+  protocol match.
+
 - When using Verbs flow engine (``dv_flow_en`` = 0), flow pattern without any
   specific VLAN will match for VLAN packets as well:
 
-- 
2.21.0



[PATCH] app/test: fix the check of the oop header data

2025-02-26 Thread Arkadiusz Kusztal
The data of the out-of-place header is never checked. Therefore,
the faulty PMD, which overwrites this data, will not be able to
verify that with tests. New checks to support that were added to
the GCM OOP functions.

Fixes: 51e202f0596f ("test/crypto: rename GCM test code")
Cc: sta...@dpdk.org

Signed-off-by: Arkadiusz Kusztal 
---
 app/test/test_cryptodev.c | 29 +++--
 1 file changed, 27 insertions(+), 2 deletions(-)

diff --git a/app/test/test_cryptodev.c b/app/test/test_cryptodev.c
index 441ecc6ad5..6ef43a784d 100644
--- a/app/test/test_cryptodev.c
+++ b/app/test/test_cryptodev.c
@@ -13769,8 +13769,9 @@ test_authenticated_encryption_oop(const struct 
aead_test_data *tdata)
struct crypto_testsuite_params *ts_params = &testsuite_params;
struct crypto_unittest_params *ut_params = &unittest_params;
 
+   uint32_t i;
int retval;
-   uint8_t *ciphertext, *auth_tag;
+   uint8_t *ciphertext, *auth_tag, *buffer_oop;
uint16_t plaintext_pad_len;
struct rte_cryptodev_info dev_info;
 
@@ -13846,6 +13847,18 @@ test_authenticated_encryption_oop(const struct 
aead_test_data *tdata)
ut_params->op->sym->cipher.data.offset);
auth_tag = ciphertext + plaintext_pad_len;
 
+   /* Check if the data within the offset range is not overwritten in the 
OOP */
+   buffer_oop = rte_pktmbuf_mtod(ut_params->obuf, uint8_t *);
+   for (i = 0; i < ut_params->op->sym->cipher.data.offset; i++) {
+   if (buffer_oop[i]) {
+   RTE_LOG(ERR, USER1,
+   "Incorrect value of the output buffer 
header\n");
+   debug_hexdump(stdout, "Incorrect value:", buffer_oop,
+   ut_params->op->sym->cipher.data.offset);
+   return TEST_FAILED;
+   }
+   }
+
debug_hexdump(stdout, "ciphertext:", ciphertext, tdata->ciphertext.len);
debug_hexdump(stdout, "auth tag:", auth_tag, tdata->auth_tag.len);
 
@@ -13878,8 +13891,9 @@ test_authenticated_decryption_oop(const struct 
aead_test_data *tdata)
struct crypto_testsuite_params *ts_params = &testsuite_params;
struct crypto_unittest_params *ut_params = &unittest_params;
 
+   uint32_t i;
int retval;
-   uint8_t *plaintext;
+   uint8_t *plaintext, *buffer_oop;
struct rte_cryptodev_info dev_info;
 
rte_cryptodev_info_get(ts_params->valid_devs[0], &dev_info);
@@ -13957,6 +13971,17 @@ test_authenticated_decryption_oop(const struct 
aead_test_data *tdata)
 
debug_hexdump(stdout, "plaintext:", plaintext, tdata->ciphertext.len);
 
+   /* Check if the data within the offset range is not overwritten in the 
OOP */
+   buffer_oop = rte_pktmbuf_mtod(ut_params->obuf, uint8_t *);
+   for (i = 0; i < ut_params->op->sym->cipher.data.offset; i++) {
+   if (buffer_oop[i]) {
+   RTE_LOG(ERR, USER1,
+   "Incorrect value of the output buffer 
header\n");
+   debug_hexdump(stdout, "Incorrect value:", buffer_oop,
+   ut_params->op->sym->cipher.data.offset);
+   return TEST_FAILED;
+   }
+   }
/* Validate obuf */
TEST_ASSERT_BUFFERS_ARE_EQUAL(
plaintext,
-- 
2.43.0



RE: [PATCH 3/3] net/mlx5: allow FDB RSS

2025-02-26 Thread Bing Zhao



> -Original Message-
> From: Suanming Mou 
> Sent: Tuesday, February 25, 2025 8:45 AM
> To: Dariusz Sosnowski ; Slava Ovsiienko
> ; Bing Zhao ; Ori Kam
> ; Matan Azrad 
> Cc: dev@dpdk.org; Raslan Darawsheh 
> Subject: [PATCH 3/3] net/mlx5: allow FDB RSS
> 
> RSS can be used in FDB Rx rules when JUMP_FDB_RX action to allow jump from
> FDB Tx to FDB Rx table.
> 
> Different with NIC RSS, FDB RSS will not do the internal implicit metadata
> copy.
> 
> This commit enables the FDB RSS if JUMP_FDB_RX is supported.
> 
> Signed-off-by: Suanming Mou 
> ---
>  drivers/net/mlx5/mlx5_flow.c|  4 ++--
>  drivers/net/mlx5/mlx5_flow_dv.c | 32 +---
> drivers/net/mlx5/mlx5_flow_hw.c | 19 ++-
>  3 files changed, 33 insertions(+), 22 deletions(-)
> 
> diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
> index 3fbe89a9d4..9c6a4f39fb 100644
> --- a/drivers/net/mlx5/mlx5_flow.c
> +++ b/drivers/net/mlx5/mlx5_flow.c
> @@ -1977,9 +1977,9 @@ mlx5_flow_validate_action_mark(struct rte_eth_dev
> *dev,
> RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
> "mark action not supported for "
> "egress");
> - if (attr->transfer && mlx5_hws_active(dev))
> + if (attr->transfer && !mlx5_hws_active(dev))
>   return rte_flow_error_set(error, ENOTSUP,
> -   RTE_FLOW_ERROR_TYPE_ATTR_EGRESS, NULL,
> +   RTE_FLOW_ERROR_TYPE_ATTR_TRANSFER, 
> NULL,
> "non-template mark action not 
> supported
> for transfer");
>   return 0;
>  }
> diff --git a/drivers/net/mlx5/mlx5_flow_dv.c
> b/drivers/net/mlx5/mlx5_flow_dv.c index 633c41e358..61d3101ce8 100644
> --- a/drivers/net/mlx5/mlx5_flow_dv.c
> +++ b/drivers/net/mlx5/mlx5_flow_dv.c
> @@ -8939,21 +8939,23 @@ flow_dv_validate(struct rte_eth_dev *dev, const
> struct rte_flow_attr *attr,
> RTE_FLOW_ERROR_TYPE_ACTION,
> NULL,
> "unsupported action MARK");
> - if (action_flags & MLX5_FLOW_ACTION_QUEUE)
> - return rte_flow_error_set(error, ENOTSUP,
> -   RTE_FLOW_ERROR_TYPE_ACTION,
> -   NULL,
> -   "unsupported action QUEUE");
> - if (action_flags & MLX5_FLOW_ACTION_RSS)
> - return rte_flow_error_set(error, ENOTSUP,
> -   RTE_FLOW_ERROR_TYPE_ACTION,
> -   NULL,
> -   "unsupported action RSS");
> - if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
> - return rte_flow_error_set(error, EINVAL,
> -   RTE_FLOW_ERROR_TYPE_ACTION,
> -   actions,
> -   "no fate action is found");
> + if (!priv->jump_fdb_rx_en) {
> + if (action_flags & MLX5_FLOW_ACTION_QUEUE)
> + return rte_flow_error_set(error, ENOTSUP,
> +   
> RTE_FLOW_ERROR_TYPE_ACTION,
> +   NULL,
> +   "unsupported action 
> QUEUE");
> + if (action_flags & MLX5_FLOW_ACTION_RSS)
> + return rte_flow_error_set(error, ENOTSUP,
> +   
> RTE_FLOW_ERROR_TYPE_ACTION,
> +   NULL,
> +   "unsupported action 
> RSS");
> + if (!(action_flags & MLX5_FLOW_FATE_ESWITCH_ACTIONS))
> + return rte_flow_error_set(error, EINVAL,
> +   
> RTE_FLOW_ERROR_TYPE_ACTION,
> +   actions,
> +   "no fate action is 
> found");
> + }
>   } else {
>   if (!(action_flags & MLX5_FLOW_FATE_ACTIONS) && attr->ingress)
>   return rte_flow_error_set(error, EINVAL, diff --git
> a/drivers/net/mlx5/mlx5_flow_hw.c b/drivers/net/mlx5/mlx5_flow_hw.c index
> 83f55ed3e8..a063e5ac9c 100644
> --- a/drivers/net/mlx5/mlx5_flow_hw.c
> +++ b/drivers/net/mlx5/mlx5_flow_hw.c
> @@ -7026,6 +7026,7 @@ mlx5_hw_validate_action_queue(struct rte_eth_dev
> *dev,
> uint64_t action_flags,
>   

RE: [PATCH 2/3] net/mlx5: add jump FDB Rx flag

2025-02-26 Thread Bing Zhao
Hi

> -Original Message-
> From: Suanming Mou 
> Sent: Tuesday, February 25, 2025 8:45 AM
> To: Dariusz Sosnowski ; Slava Ovsiienko
> ; Bing Zhao ; Ori Kam
> ; Matan Azrad 
> Cc: dev@dpdk.org; Raslan Darawsheh 
> Subject: [PATCH 2/3] net/mlx5: add jump FDB Rx flag
> 
> When jump FDB Rx is supported, flow will be able to jump from FDB Tx to
> FDB Rx, in that case the dest action in FDB Rx table should support FDB Tx
> as well.
> 
> Signed-off-by: Suanming Mou 
> ---
>  drivers/common/mlx5/mlx5_devx_cmds.c | 8 
> drivers/common/mlx5/mlx5_devx_cmds.h | 1 +
>  drivers/net/mlx5/linux/mlx5_os.c | 1 +
>  drivers/net/mlx5/mlx5.h  | 1 +
>  drivers/net/mlx5/mlx5_flow_hw.c  | 8 +++-
>  5 files changed, 18 insertions(+), 1 deletion(-)
> 
> diff --git a/drivers/common/mlx5/mlx5_devx_cmds.c
> b/drivers/common/mlx5/mlx5_devx_cmds.c
> index f504b29f31..eb8553e8ad 100644
> --- a/drivers/common/mlx5/mlx5_devx_cmds.c
> +++ b/drivers/common/mlx5/mlx5_devx_cmds.c
> @@ -924,6 +924,7 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
>   uint32_t out[MLX5_ST_SZ_DW(query_hca_cap_out)] = {0};
>   bool hca_cap_2_sup;
>   uint64_t general_obj_types_supported = 0;
> + uint64_t stc_action_type_127_64;
>   void *hcattr;
>   int rc, i;
> 
> @@ -1352,6 +1353,13 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
>   attr->fdb_unified_en = MLX5_GET(wqe_based_flow_table_cap,
>   hcattr,
>   fdb_unified_en);
> + stc_action_type_127_64 = MLX5_GET64(wqe_based_flow_table_cap,
> + hcattr,
> + stc_action_type_127_64);
> + if (stc_action_type_127_64 &
> +(1 <<
> (MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX -
> +   MLX5_IFC_STC_ACTION_TYPE_BIT_64_INDEX)))
> + attr->jump_fdb_rx_en = true;

I assume integer value 1 is enough, since it is a bit field, no need to use 
bool type. 

>   }
>   /* Query HCA attribute for ROCE. */
>   if (attr->roce) {
> diff --git a/drivers/common/mlx5/mlx5_devx_cmds.h
> b/drivers/common/mlx5/mlx5_devx_cmds.h
> index 8de4210fb2..6c726a0d46 100644
> --- a/drivers/common/mlx5/mlx5_devx_cmds.h
> +++ b/drivers/common/mlx5/mlx5_devx_cmds.h
> @@ -326,6 +326,7 @@ struct mlx5_hca_attr {
>   uint32_t lag_rx_port_affinity:1;
>   uint32_t wqe_based_flow_table_sup:1;
>   uint32_t fdb_unified_en:1;
> + uint32_t jump_fdb_rx_en:1;
>   uint8_t max_header_modify_pattern_length;
>   uint64_t system_image_guid;
>   uint32_t log_max_conn_track_offload:5; diff --git
> a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
> index 9410211e3b..4e64026137 100644
> --- a/drivers/net/mlx5/linux/mlx5_os.c
> +++ b/drivers/net/mlx5/linux/mlx5_os.c
> @@ -1718,6 +1718,7 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
>   priv->unified_fdb_en = !!priv->master;
>   DRV_LOG(DEBUG, "port %u: unified FDB %s enabled.",
>   eth_dev->data->port_id, priv->unified_fdb_en ? "is" :
> "isn't");
> + priv->jump_fdb_rx_en = sh->cdev-
> >config.hca_attr.jump_fdb_rx_en;

Can we add some INFO or DEBUG level log to indicate this to the user?

>   if (priv->sh->config.dv_esw_en) {
>   uint32_t usable_bits;
>   uint32_t required_bits;
> diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h index
> f73f6e63ff..545ba48b3c 100644
> --- a/drivers/net/mlx5/mlx5.h
> +++ b/drivers/net/mlx5/mlx5.h
> @@ -1987,6 +1987,7 @@ struct mlx5_priv {
>   uint32_t num_lag_ports:4; /* Number of ports can be bonded. */
>   uint32_t tunnel_enabled:1; /* If tunnel offloading is enabled on
> rxqs. */
>   uint32_t unified_fdb_en:1; /* Unified FDB flag per port. */
> + uint32_t jump_fdb_rx_en:1; /* Jump from FDB Tx to FDB Rx flag per
> +port. */
>   uint16_t domain_id; /* Switch domain identifier. */
>   uint16_t vport_id; /* Associated VF vport index (if any). */
>   uint32_t vport_meta_tag; /* Used for vport index match ove VF LAG.
> */ diff --git a/drivers/net/mlx5/mlx5_flow_hw.c
> b/drivers/net/mlx5/mlx5_flow_hw.c index f0888dbf0e..83f55ed3e8 100644
> --- a/drivers/net/mlx5/mlx5_flow_hw.c
> +++ b/drivers/net/mlx5/mlx5_flow_hw.c
> @@ -9322,6 +9322,7 @@ flow_hw_grp_create_cb(void *tool_ctx, void *cb_ctx)
>   struct mlx5_flow_group *grp_data;
>   struct mlx5dr_table *tbl = NULL;
>   struct mlx5dr_action *jump;
> + uint32_t hws_flags;
>   uint32_t idx = 0;
>   MKSTR(matcher_name, "%s_%s_%u_%u_matcher_list",
> attr->transfer ? "FDB" : "NIC", attr->egress ? "egress" :
> "ingress", @@ -9342,10 +9343,15 @@ flow_hw_grp_create_cb(void *tool_ctx,
> void *cb_ctx)
>   goto error;
>   grp_data->tbl = tbl;
>   if (attr->group) {
>

Re: [PATCH 1/2] doc: fix MPLS limitations under HWS

2025-02-26 Thread Raslan Darawsheh
Hi,

From: Maayan Kashani 
Sent: Wednesday, February 26, 2025 10:02 AM
To: dev@dpdk.org
Cc: Maayan Kashani; Dariusz Sosnowski; Raslan Darawsheh; sta...@dpdk.org; Slava 
Ovsiienko; Bing Zhao; Ori Kam; Suanming Mou; Matan Azrad; Michael Baum
Subject: [PATCH 1/2] doc: fix MPLS limitations under HWS

Update missing limitations.

Fixes: 1e2a51f545b6 ("net/mlx5: support MPLS tunnel with HWS")
Cc: sta...@dpdk.org

Signed-off-by: Maayan Kashani 
Acked-by: Dariusz Sosnowski 

series squashed and applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh



RE: [PATCH V2 4/4] net/mlx5/hws: allow different types in miss validation

2025-02-26 Thread Dariusz Sosnowski



> -Original Message-
> From: Ophir Munk 
> Sent: Wednesday, February 26, 2025 9:39 AM
> To: Dariusz Sosnowski ; Slava Ovsiienko
> ; Bing Zhao ; Ori Kam
> ; Suanming Mou ; Matan Azrad
> 
> Cc: dev@dpdk.org; Raslan Darawsheh 
> Subject: [PATCH V2 4/4] net/mlx5/hws: allow different types in miss validation
> 
> In validating Jump from group A to default miss group B DPDK requests that 
> both
> groups types will be identical. In practice jumping from FDB_UNIFIED to FDB_RX
> or FDB_TX group types (or vice versa) is allowed.
> This commit enables this jumping between the different types.
> 
> Fixes: b81f95ca770d ("net/mlx5/hws: support default miss table")
> 
> Signed-off-by: Ophir Munk 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


[PATCH] crypto/qat: fix incorrect sm3 state size

2025-02-26 Thread Arkadiusz Kusztal
This commit fixes incorrectly set state size within
the QAT PMD.

Fixes: 75fd4bbc94ab ("crypto/qat: support SM3 hash algorithm")
Cc: sta...@dpdk.org

Signed-off-by: Arkadiusz Kusztal 
---
 drivers/crypto/qat/qat_sym_session.c | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/drivers/crypto/qat/qat_sym_session.c 
b/drivers/crypto/qat/qat_sym_session.c
index 9b26f8a89b..7f370f03fb 100644
--- a/drivers/crypto/qat/qat_sym_session.c
+++ b/drivers/crypto/qat/qat_sym_session.c
@@ -2473,7 +2473,8 @@ int qat_sym_cd_auth_set(struct qat_sym_session *cdesc,
qat_hash_get_block_size(cdesc->qat_hash_alg) >> 3;
auth_param->u2.inner_prefix_sz =
qat_hash_get_block_size(cdesc->qat_hash_alg);
-   auth_param->hash_state_sz = digestsize;
+   auth_param->hash_state_sz = (hash_cd_ctrl->outer_prefix_sz +
+   auth_param->u2.inner_prefix_sz) >> 3;
if (qat_dev_gen == QAT_GEN4 || qat_dev_gen == QAT_GEN5 ||
qat_dev_gen == QAT_VQAT) {
ICP_QAT_FW_HASH_FLAG_MODE2_SET(
-- 
2.43.0



RE: [PATCH v2 05/14] net/ipn3ke: remove unnecessary dereference

2025-02-26 Thread Xu, Rosen
Hi,

> -Original Message-
> From: Stephen Hemminger 
> Sent: Thursday, February 20, 2025 10:32 AM
> To: dev@dpdk.org
> Cc: Stephen Hemminger ; Xu, Rosen
> 
> Subject: [PATCH v2 05/14] net/ipn3ke: remove unnecessary dereference
> 
> Since it is a function pointer, the expression *hw->f_mac_read and
> hw->f_mac_read are equivalent in C.
> 
> Remove excession indentation here.
> 
> Signed-off-by: Stephen Hemminger 
> ---
>  drivers/net/intel/ipn3ke/ipn3ke_ethdev.c | 27 +---
>  1 file changed, 5 insertions(+), 22 deletions(-)
> 
> diff --git a/drivers/net/intel/ipn3ke/ipn3ke_ethdev.c
> b/drivers/net/intel/ipn3ke/ipn3ke_ethdev.c
> index baae80d661..c8f06a06cb 100644
> --- a/drivers/net/intel/ipn3ke/ipn3ke_ethdev.c
> +++ b/drivers/net/intel/ipn3ke/ipn3ke_ethdev.c
> @@ -229,34 +229,17 @@ ipn3ke_mtu_set(struct ipn3ke_hw *hw, uint32_t
> mac_num,
>   uint32_t rx;
>   uint32_t tmp;
> 
> - if (!(*hw->f_mac_read) || !(*hw->f_mac_write))
> + if (!hw->f_mac_read || !hw->f_mac_write)
>   return;
> 
> - (*hw->f_mac_read)(hw,
> - &tx,
> - txaddr,
> - mac_num,
> - eth_group_sel);
> + hw->f_mac_read(hw, &tx, txaddr, mac_num, eth_group_sel);
> 
> - (*hw->f_mac_read)(hw,
> - &rx,
> - rxaddr,
> - mac_num,
> - eth_group_sel);
> + hw->f_mac_read(hw, &rx, rxaddr, mac_num, eth_group_sel);
> 
>   tmp = ipn3ke_mtu_cal(tx, rx);
> 
> - (*hw->f_mac_write)(hw,
> - tmp,
> - txaddr,
> - mac_num,
> - eth_group_sel);
> -
> - (*hw->f_mac_write)(hw,
> - tmp,
> - rxaddr,
> - mac_num,
> - eth_group_sel);
> + hw->f_mac_write(hw, tmp, txaddr, mac_num, eth_group_sel);
> + hw->f_mac_write(hw, tmp, rxaddr, mac_num, eth_group_sel);
>  }
> 
>  static void
> --
> 2.47.2

Reviewed-by: Rosen Xu 


Re: [PATCH v2 5/5] acl: make compatible with instruction set updates for MSVC

2025-02-26 Thread Bruce Richardson
On Tue, Feb 25, 2025 at 05:06:56PM -0800, Andre Muezerie wrote:
> Top level 'cc_avx2_flags' was created and holds the correct flags
> depending on the compiler used.
> 
> Signed-off-by: Andre Muezerie 
> ---
Acked-by: Bruce Richardson 

This patch can probably be merged with the patch introducing the new build
variable. It and the previous-but-one patch are both very small and are
probably better merged.



Re: [PATCH v2 4/5] drivers/net: make compatible with instruction set updates for MSVC

2025-02-26 Thread Bruce Richardson
On Tue, Feb 25, 2025 at 05:06:55PM -0800, Andre Muezerie wrote:
> Top level 'cc_avx2_flags' was created and holds the correct flags
> depending on the compiler used.
> 
> Signed-off-by: Andre Muezerie 
> ---
Acked-by: Bruce Richardson 


RE: [PATCH] net/mlx5: support ESP in non-template RSS expansion

2025-02-26 Thread Dariusz Sosnowski



> -Original Message-
> From: Gregory Etelson 
> Sent: Monday, February 17, 2025 8:13 AM
> To: dev@dpdk.org
> Cc: Gregory Etelson ; Maayan Kashani
> ; Raslan Darawsheh 
> Subject: [PATCH] net/mlx5: support ESP in non-template RSS expansion
> 
> External email: Use caution opening links or attachments
> 
> 
> Add support for the ESP protocol in non-template RSS expansion.
> 
> Signed-off-by: Gregory Etelson 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


[PATCH v16 22/29] net/rnp: add supported packet types

2025-02-26 Thread Wenbo Cao
add support parse hw packet types result.

Signed-off-by: Wenbo Cao 
---
 doc/guides/nics/features/rnp.ini  |  1 +
 doc/guides/nics/rnp.rst   |  2 ++
 drivers/net/rnp/base/rnp_bdq_if.h |  4 +++
 drivers/net/rnp/rnp_ethdev.c  | 31 +
 drivers/net/rnp/rnp_rxtx.c| 45 +++
 5 files changed, 83 insertions(+)

diff --git a/doc/guides/nics/features/rnp.ini b/doc/guides/nics/features/rnp.ini
index 00b23b0fa2..eb1c27a3d3 100644
--- a/doc/guides/nics/features/rnp.ini
+++ b/doc/guides/nics/features/rnp.ini
@@ -7,6 +7,7 @@
 Speed capabilities   = Y
 Link status  = Y
 Link status event= Y
+Packet type parsing  = Y
 Basic stats  = Y
 Stats per queue  = Y
 Extended stats   = Y
diff --git a/doc/guides/nics/rnp.rst b/doc/guides/nics/rnp.rst
index ea3270a758..58f4f4a3a8 100644
--- a/doc/guides/nics/rnp.rst
+++ b/doc/guides/nics/rnp.rst
@@ -55,6 +55,7 @@ Features
 - Jumbo frames
 - Scatter-Gather IO support
 - Port hardware statistic
+- Packet type parsing
 
 Prerequisites and Pre-conditions
 
@@ -97,6 +98,7 @@ Listed below are the rte_eth functions supported:
 * ``rte_eth_dev_default_mac_addr_set``
 * ``rte_eth_dev_mac_addr_add``
 * ``rte_eth_dev_mac_addr_remove``
+* ``rte_eth_dev_get_supported_ptypes``
 * ``rte_eth_promiscuous_disable``
 * ``rte_eth_promiscuous_enable``
 * ``rte_eth_allmulticast_enable``
diff --git a/drivers/net/rnp/base/rnp_bdq_if.h 
b/drivers/net/rnp/base/rnp_bdq_if.h
index 61a38323e5..a7d27bdebb 100644
--- a/drivers/net/rnp/base/rnp_bdq_if.h
+++ b/drivers/net/rnp/base/rnp_bdq_if.h
@@ -73,6 +73,7 @@ struct rnp_tx_desc {
 #define RNP_RX_L3TYPE_IPV4 (0x00UL << RNP_RX_L3TYPE_S)
 #define RNP_RX_L3TYPE_IPV6 (0x01UL << RNP_RX_L3TYPE_S)
 #define RNP_RX_L4TYPE_S(6)
+#define RNP_RX_L4TYPE_MASK RTE_GENMASK32(7, 6)
 #define RNP_RX_L4TYPE_TCP  (0x01UL << RNP_RX_L4TYPE_S)
 #define RNP_RX_L4TYPE_SCTP (0x02UL << RNP_RX_L4TYPE_S)
 #define RNP_RX_L4TYPE_UDP  (0x03UL << RNP_RX_L4TYPE_S)
@@ -83,9 +84,12 @@ struct rnp_tx_desc {
 #define RNP_RX_IN_L3_ERR   RTE_BIT32(11)
 #define RNP_RX_IN_L4_ERR   RTE_BIT32(12)
 #define RNP_RX_TUNNEL_TYPE_S   (13)
+#define RNP_RX_TUNNEL_MASK RTE_GENMASK32(14, 13)
 #define RNP_RX_PTYPE_VXLAN (0x01UL << RNP_RX_TUNNEL_TYPE_S)
 #define RNP_RX_PTYPE_NVGRE (0x02UL << RNP_RX_TUNNEL_TYPE_S)
 #define RNP_RX_PTYPE_VLAN  RTE_BIT32(15)
+/* mark_data */
+#define RNP_RX_L3TYPE_VALIDRTE_BIT32(31)
 /* tx data cmd */
 #define RNP_TX_TSO_EN  RTE_BIT32(4)
 #define RNP_TX_L3TYPE_S(5)
diff --git a/drivers/net/rnp/rnp_ethdev.c b/drivers/net/rnp/rnp_ethdev.c
index 0ce9fa52e9..289d9ed867 100644
--- a/drivers/net/rnp/rnp_ethdev.c
+++ b/drivers/net/rnp/rnp_ethdev.c
@@ -1186,6 +1186,36 @@ rnp_dev_mac_addr_remove(struct rte_eth_dev *dev,
rnp_clear_macaddr(port, index);
 }
 
+static uint32_t *rnp_support_ptypes_get(void)
+{
+   static uint32_t ptypes[] = {
+   RTE_PTYPE_L2_ETHER,
+   RTE_PTYPE_L2_ETHER_TIMESYNC,
+   RTE_PTYPE_L3_IPV4_EXT_UNKNOWN,
+   RTE_PTYPE_L3_IPV6_EXT_UNKNOWN,
+   RTE_PTYPE_L4_TCP,
+   RTE_PTYPE_L4_UDP,
+   RTE_PTYPE_L4_SCTP,
+   RTE_PTYPE_TUNNEL_VXLAN,
+   RTE_PTYPE_TUNNEL_GRE,
+   RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN,
+   RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN,
+   RTE_PTYPE_INNER_L4_TCP,
+   RTE_PTYPE_INNER_L4_UDP,
+   RTE_PTYPE_INNER_L4_SCTP,
+   RTE_PTYPE_UNKNOWN,
+   };
+
+   return ptypes;
+}
+
+static const uint32_t *
+rnp_dev_supported_ptypes_get(struct rte_eth_dev *dev __rte_unused,
+size_t *no_of_elements __rte_unused)
+{
+   return rnp_support_ptypes_get();
+}
+
 /* Features supported by this driver */
 static const struct eth_dev_ops rnp_eth_dev_ops = {
.dev_configure= rnp_dev_configure,
@@ -1228,6 +1258,7 @@ static const struct eth_dev_ops rnp_eth_dev_ops = {
.mac_addr_set = rnp_dev_mac_addr_set,
.mac_addr_add = rnp_dev_mac_addr_add,
.mac_addr_remove  = rnp_dev_mac_addr_remove,
+   .dev_supported_ptypes_get = rnp_dev_supported_ptypes_get,
 };
 
 static void
diff --git a/drivers/net/rnp/rnp_rxtx.c b/drivers/net/rnp/rnp_rxtx.c
index 8d67e04990..dd8cde8aff 100644
--- a/drivers/net/rnp/rnp_rxtx.c
+++ b/drivers/net/rnp/rnp_rxtx.c
@@ -639,6 +639,49 @@ int rnp_rx_queue_start(struct rte_eth_dev *eth_dev, 
uint16_t qidx)
return 0;
 }
 
+static __rte_always_inline void
+rnp_dev_rx_parse(struct rnp_rx_queue *rxq __rte_unused,
+struct rte_mbuf *m,
+volatile struct rnp_rx_desc rxbd)
+{
+   uint32_t mark_data = rxbd.wb.qword0.mark_data;
+   uint16_t vlan_tci = rxbd.wb.qword1.vlan_tci;
+   uint32_t cmd = rxbd.wb.q

[PATCH v16 17/29] net/rnp: add Rx scatter segment version

2025-02-26 Thread Wenbo Cao
add support scatter multi segment received.

Signed-off-by: Wenbo Cao 
---
 doc/guides/nics/features/rnp.ini |   1 +
 doc/guides/nics/rnp.rst  |   2 +
 drivers/net/rnp/rnp_ethdev.c |   2 +
 drivers/net/rnp/rnp_rxtx.c   | 130 ++-
 drivers/net/rnp/rnp_rxtx.h   |   2 +
 5 files changed, 135 insertions(+), 2 deletions(-)

diff --git a/doc/guides/nics/features/rnp.ini b/doc/guides/nics/features/rnp.ini
index 6d13370e8f..6453762745 100644
--- a/doc/guides/nics/features/rnp.ini
+++ b/doc/guides/nics/features/rnp.ini
@@ -15,5 +15,6 @@ RSS hash = Y
 RSS key update   = Y
 RSS reta update  = Y
 Inner RSS= Y
+Scattered Rx = Y
 Linux= Y
 x86-64   = Y
diff --git a/doc/guides/nics/rnp.rst b/doc/guides/nics/rnp.rst
index 6aa8b00af6..ebf6971140 100644
--- a/doc/guides/nics/rnp.rst
+++ b/doc/guides/nics/rnp.rst
@@ -51,6 +51,8 @@ Features
 - Promiscuous mode
 - Link state information
 - MTU update
+- Jumbo frames
+- Scatter-Gather IO support
 
 Prerequisites and Pre-conditions
 
diff --git a/drivers/net/rnp/rnp_ethdev.c b/drivers/net/rnp/rnp_ethdev.c
index 4afdccd0ab..3dfc7ef97b 100644
--- a/drivers/net/rnp/rnp_ethdev.c
+++ b/drivers/net/rnp/rnp_ethdev.c
@@ -606,6 +606,8 @@ static int rnp_dev_infos_get(struct rte_eth_dev *eth_dev,
dev_info->reta_size = RNP_RSS_INDIR_SIZE;
/* speed cap info */
dev_info->speed_capa = rnp_get_speed_caps(eth_dev);
+   /* rx support offload cap */
+   dev_info->rx_offload_capa = RTE_ETH_RX_OFFLOAD_SCATTER;
/* default ring configure */
dev_info->default_rxportconf.burst_size = 32;
dev_info->default_txportconf.burst_size = 32;
diff --git a/drivers/net/rnp/rnp_rxtx.c b/drivers/net/rnp/rnp_rxtx.c
index 260c34113a..b2ca9b15fc 100644
--- a/drivers/net/rnp/rnp_rxtx.c
+++ b/drivers/net/rnp/rnp_rxtx.c
@@ -826,7 +826,6 @@ rnp_xmit_simple(void *_txq, struct rte_mbuf **tx_pkts, 
uint16_t nb_pkts)
if (txq->tx_next_rs > txq->attr.nb_desc)
txq->tx_next_rs = txq->tx_rs_thresh - 1;
}
-
txq->tx_tail = i;
 
rte_wmb();
@@ -835,9 +834,136 @@ rnp_xmit_simple(void *_txq, struct rte_mbuf **tx_pkts, 
uint16_t nb_pkts)
return start;
 }
 
+static int
+rnp_rxq_bulk_alloc(struct rnp_rx_queue *rxq,
+  volatile struct rnp_rx_desc *rxbd,
+  struct rnp_rxsw_entry *rxe,
+  bool bulk_alloc)
+{
+   struct rte_eth_dev_data *data;
+   struct rte_mbuf *nmb = NULL;
+   uint16_t update_tail;
+
+   if (!bulk_alloc) {
+   data = rte_eth_devices[rxq->attr.port_id].data;
+   nmb = rte_mbuf_raw_alloc(rxq->mb_pool);
+   if (unlikely(!nmb)) {
+   data->rx_mbuf_alloc_failed++;
+   return -ENOMEM;
+   }
+   rxbd->d.pkt_addr = 0;
+   rxbd->d.cmd = 0;
+   rxe->mbuf = NULL;
+   rxe->mbuf = nmb;
+   rxbd->d.pkt_addr = rnp_get_dma_addr(&rxq->attr, nmb);
+   }
+   if (rxq->rxrearm_nb > rxq->rx_free_thresh) {
+   rxq->rxrearm_nb -= rxq->rx_free_thresh;
+   rxq->rxrearm_start += rxq->rx_free_thresh;
+   if (rxq->rxrearm_start >= rxq->attr.nb_desc)
+   rxq->rxrearm_start = 0;
+   update_tail = (uint16_t)((rxq->rxrearm_start == 0) ?
+   (rxq->attr.nb_desc - 1) : (rxq->rxrearm_start - 
1));
+   rte_io_wmb();
+   RNP_REG_WR(rxq->rx_tailreg, 0, update_tail);
+   }
+
+   return 0;
+}
+
+static __rte_always_inline uint16_t
+rnp_scattered_rx(void *rx_queue, struct rte_mbuf **rx_pkts,
+uint16_t nb_pkts)
+{
+   struct rnp_rx_queue *rxq = (struct rnp_rx_queue *)rx_queue;
+   volatile struct rnp_rx_desc *bd_ring = rxq->rx_bdr;
+   struct rte_mbuf *first_seg = rxq->pkt_first_seg;
+   struct rte_mbuf *last_seg = rxq->pkt_last_seg;
+   struct rnp_rxsw_entry *sw_ring = rxq->sw_ring;
+   volatile struct rnp_rx_desc *rxbd;
+   volatile struct rnp_rx_desc rxd;
+   struct rnp_rxsw_entry *rxe;
+   struct rte_mbuf *rxm;
+   uint16_t rx_pkt_len;
+   uint16_t nb_rx = 0;
+   uint16_t rx_status;
+   uint16_t rx_id;
+
+   if (unlikely(!rxq->rxq_started || !rxq->rx_link))
+   return 0;
+   rx_id = rxq->rx_tail;
+   while (nb_rx < nb_pkts) {
+   rxbd = &bd_ring[rx_id];
+   rx_status = rxbd->wb.qword1.cmd;
+   if (!(rx_status & rte_cpu_to_le_16(RNP_CMD_DD)))
+   break;
+   rte_atomic_thread_fence(rte_memory_order_acquire);
+   rxd = *rxbd;
+   rxe = &sw_ring[rx_id];
+   rxm = rxe->mbuf;
+   if (rnp_rxq_bulk_alloc(rxq, rxbd, rxe, false))
+   break;

RE: [PATCH 2/2] net/mlx5/hws: add DROP action to multi destination actions

2025-02-26 Thread Dariusz Sosnowski



> -Original Message-
> From: Gregory Etelson 
> Sent: Thursday, February 13, 2025 9:33 AM
> To: dev@dpdk.org
> Cc: Gregory Etelson ; Maayan Kashani
> ; Raslan Darawsheh ; Tamar
> Mashiah ; sta...@dpdk.org
> Subject: [PATCH 2/2] net/mlx5/hws: add DROP action to multi destination 
> actions
> 
> External email: Use caution opening links or attachments
> 
> 
> From: Tamar Mashiah 
> 
> Add DROP action to multi destination actions, set NOP destination type when
> DROP is specified.
> 
> cc: sta...@dpdk.org
> 
> Signed-off-by: Tamar Mashiah 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


RE: [PATCH 1/2] net/mlx5: add DROP destination to mirror clone actions

2025-02-26 Thread Dariusz Sosnowski



> -Original Message-
> From: Gregory Etelson 
> Sent: Thursday, February 13, 2025 9:33 AM
> To: dev@dpdk.org
> Cc: Gregory Etelson ; Maayan Kashani
> ; Raslan Darawsheh ; Tamar
> Mashiah ; sta...@dpdk.org
> Subject: [PATCH 1/2] net/mlx5: add DROP destination to mirror clone actions
> 
> External email: Use caution opening links or attachments
> 
> 
> From: Tamar Mashiah 
> 
> Add DROP destination to mirror clone actions.
> 
> cc: sta...@dpdk.org
> 
> Signed-off-by: Tamar Mashiah 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


RE: [PATCH] net/mlx5/hws: change error flow on matcher disconnect

2025-02-26 Thread Dariusz Sosnowski
> -Original Message-
> From: Gregory Etelson 
> Sent: Sunday, February 9, 2025 2:48 PM
> To: dev@dpdk.org
> Cc: Gregory Etelson ; Maayan Kashani
> ; Raslan Darawsheh ; Itamar
> Gozlan 
> Subject: [PATCH] net/mlx5/hws: change error flow on matcher disconnect
> 
> External email: Use caution opening links or attachments
> 
> 
> From: Itamar Gozlan 
> 
> Currently, when a firmware (FW) failure occurs during matcher error flow, the
> system attempts to reconnect the matcher, leading to a segmentation fault. 
> This
> happens because the matcher is freed but remains in the list.
> 
> Example scenario:
> Given matchers M1->M2->M3, if a FW failure occurs:
> 1. System tries to destroy M1 and fails
> 2. M1 remains in the list but is freed
> 3. When destroying M2, it attempts to remove itself and create M1->M3 4. This
> results in a segmentation fault as M1 is already freed
> 
> Signed-off-by: Itamar Gozlan 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


[PATCH V2 0/4] mlx5 unified fdb

2025-02-26 Thread Ophir Munk
V1: initial version
V2: rebase + determine unified fdb based on FW capability

Bing Zhao (1):
  net/mlx5: remove unneeded FDB flag on representor action

Ophir Munk (3):
  common/mlx5: support FDB unified capability query
  net/mlx5: support FDB unified domain
  net/mlx5/hws: allow different types in miss validation

 drivers/common/mlx5/mlx5_devx_cmds.c |   3 +
 drivers/common/mlx5/mlx5_devx_cmds.h |   1 +
 drivers/net/mlx5/hws/mlx5dr_table.c  |  18 +-
 drivers/net/mlx5/linux/mlx5_os.c |  11 +
 drivers/net/mlx5/mlx5.h  |   8 +
 drivers/net/mlx5/mlx5_flow_hw.c  | 521 ---
 drivers/net/mlx5/mlx5_flow_meter.c   |  16 +-
 drivers/net/mlx5/mlx5_flow_quota.c   |   9 +-
 drivers/net/mlx5/mlx5_hws_cnt.c  |   9 +-
 9 files changed, 429 insertions(+), 167 deletions(-)

-- 
2.8.4



[PATCH V2 1/4] common/mlx5: support FDB unified capability query

2025-02-26 Thread Ophir Munk
This commit queries the FW for the new unified FDB mode and saves it in
mlx5 shared device as fdb_unified_en bit.

Signed-off-by: Ophir Munk 
---
 drivers/common/mlx5/mlx5_devx_cmds.c | 3 +++
 drivers/common/mlx5/mlx5_devx_cmds.h | 1 +
 2 files changed, 4 insertions(+)

diff --git a/drivers/common/mlx5/mlx5_devx_cmds.c 
b/drivers/common/mlx5/mlx5_devx_cmds.c
index bba00a9..f504b29 100644
--- a/drivers/common/mlx5/mlx5_devx_cmds.c
+++ b/drivers/common/mlx5/mlx5_devx_cmds.c
@@ -1349,6 +1349,9 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
attr->max_header_modify_pattern_length = 
MLX5_GET(wqe_based_flow_table_cap,
  hcattr,
  
max_header_modify_pattern_length);
+   attr->fdb_unified_en = MLX5_GET(wqe_based_flow_table_cap,
+   hcattr,
+   fdb_unified_en);
}
/* Query HCA attribute for ROCE. */
if (attr->roce) {
diff --git a/drivers/common/mlx5/mlx5_devx_cmds.h 
b/drivers/common/mlx5/mlx5_devx_cmds.h
index 38548b4..8de4210 100644
--- a/drivers/common/mlx5/mlx5_devx_cmds.h
+++ b/drivers/common/mlx5/mlx5_devx_cmds.h
@@ -325,6 +325,7 @@ struct mlx5_hca_attr {
uint32_t cross_vhca:1;
uint32_t lag_rx_port_affinity:1;
uint32_t wqe_based_flow_table_sup:1;
+   uint32_t fdb_unified_en:1;
uint8_t max_header_modify_pattern_length;
uint64_t system_image_guid;
uint32_t log_max_conn_track_offload:5;
-- 
2.8.4



[PATCH V2 2/4] net/mlx5: support FDB unified domain

2025-02-26 Thread Ophir Munk
Legacy FDB domain is enhanced with optional three sub-domains: FDB_RX
FDB_TX and FDB_UNIFIED. Based on FW capability bit mlx5 will work in
either legacy or unified mode. To work in fdb unified mode the flow
attribute transfer bit must be set and the fdb sub domains are
determined by the table specialize flags SPECIALIZE_TRANSFER_WIRE_ORIG
and SPECIALIZE_TRANSFER_VPORT_ORIG.

++-+--+
| WIRE_ORIG  | VPORT_ORIG  | FDB domain   |
++-+--+
| 0  | 0   | FDB_UNIFIED  |
| 0  | 1   | FDB_TX   |
| 1  | 0   | FDB_RX   |
++-+--+

There can be no two groups which only differ in their fdb sub domain.
A table must be created in a group which has the same fdb sub domain.

Signed-off-by: Ophir Munk 
---
 drivers/net/mlx5/linux/mlx5_os.c   |  11 +
 drivers/net/mlx5/mlx5.h|   8 +
 drivers/net/mlx5/mlx5_flow_hw.c| 519 ++---
 drivers/net/mlx5/mlx5_flow_meter.c |  16 +-
 drivers/net/mlx5/mlx5_flow_quota.c |   9 +-
 drivers/net/mlx5/mlx5_hws_cnt.c|   9 +-
 6 files changed, 409 insertions(+), 163 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 091f288..7f13cb2 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -1707,6 +1707,17 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev,
LIST_INIT(&priv->hw_ext_ctrl_flows);
if (priv->sh->config.dv_flow_en == 2) {
 #ifdef HAVE_MLX5_HWS_SUPPORT
+   /*
+* Unified FDB flag is only needed for the actions created on 
the transfer
+* port. proxy port. It is not needed on the following ports:
+*   1. NIC PF / VF / SF
+*   2. in Verbs or DV/DR mode
+*   3. with unsupported FW
+*   4. all representors in HWS
+*/
+   priv->unified_fdb_en = !!priv->master && 
sh->cdev->config.hca_attr.fdb_unified_en;
+   DRV_LOG(DEBUG, "port %u: unified FDB %s enabled.",
+   eth_dev->data->port_id, priv->unified_fdb_en ? "is" : 
"isn't");
if (priv->sh->config.dv_esw_en) {
uint32_t usable_bits;
uint32_t required_bits;
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index 2b590c9..f73f6e6 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -1986,6 +1986,7 @@ struct mlx5_priv {
uint32_t mark_enabled:1; /* If mark action is enabled on rxqs. */
uint32_t num_lag_ports:4; /* Number of ports can be bonded. */
uint32_t tunnel_enabled:1; /* If tunnel offloading is enabled on rxqs. 
*/
+   uint32_t unified_fdb_en:1; /* Unified FDB flag per port. */
uint16_t domain_id; /* Switch domain identifier. */
uint16_t vport_id; /* Associated VF vport index (if any). */
uint32_t vport_meta_tag; /* Used for vport index match ove VF LAG. */
@@ -2226,6 +2227,13 @@ mlx5_is_port_on_mpesw_device(struct mlx5_priv *priv)
return priv->mpesw_port >= 0;
 }
 
+static inline bool
+is_unified_fdb(const struct mlx5_priv *priv)
+{
+   /* Only needed on proxy port in E-Switch mode. */
+   return priv->unified_fdb_en;
+}
+
 /* mlx5.c */
 
 int mlx5_getenv_int(const char *);
diff --git a/drivers/net/mlx5/mlx5_flow_hw.c b/drivers/net/mlx5/mlx5_flow_hw.c
index e72b87d..1cc074d 100644
--- a/drivers/net/mlx5/mlx5_flow_hw.c
+++ b/drivers/net/mlx5/mlx5_flow_hw.c
@@ -282,12 +282,34 @@ mlx5_multi_pattern_reformat_index_to_type(uint32_t ix)
 }
 
 static inline enum mlx5dr_table_type
-get_mlx5dr_table_type(const struct rte_flow_attr *attr)
+get_mlx5dr_fdb_table_type(const struct rte_flow_attr *attr,
+ uint32_t specialize, bool fdb_unified_en)
+{
+   if (fdb_unified_en && !!attr->group) {
+   if ((specialize & (RTE_FLOW_TABLE_SPECIALIZE_TRANSFER_WIRE_ORIG 
|
+  
RTE_FLOW_TABLE_SPECIALIZE_TRANSFER_VPORT_ORIG)) == 0)
+   return MLX5DR_TABLE_TYPE_FDB_UNIFIED;
+   MLX5_ASSERT((specialize & 
(RTE_FLOW_TABLE_SPECIALIZE_TRANSFER_WIRE_ORIG |
+  
RTE_FLOW_TABLE_SPECIALIZE_TRANSFER_VPORT_ORIG)) !=
+ 
(RTE_FLOW_TABLE_SPECIALIZE_TRANSFER_WIRE_ORIG |
+  
RTE_FLOW_TABLE_SPECIALIZE_TRANSFER_VPORT_ORIG));
+   if (specialize & RTE_FLOW_TABLE_SPECIALIZE_TRANSFER_WIRE_ORIG)
+   return MLX5DR_TABLE_TYPE_FDB_RX;
+   if (specialize & RTE_FLOW_TABLE_SPECIALIZE_TRANSFER_VPORT_ORIG)
+   return MLX5DR_TABLE_TYPE_FDB_TX;
+   }
+
+   return MLX5DR_TABLE_TYPE_FDB;
+}
+
+static inline enum mlx5dr_table_type
+get_mlx5dr_table_type(const struct rte_flow_attr *attr, uint32_t

[PATCH V2 3/4] net/mlx5: remove unneeded FDB flag on representor action

2025-02-26 Thread Ophir Munk
From: Bing Zhao 

The "send_to_kernel" action is shared and created during the HWS
configuration stage. By default, the NIC Rx and Tx flags are needed.

There is only one case that the FDB* flags are needed when creating
the action, the same as other shared actions. Only the transfer proxy
port needs the FDB* flags for shared actions.

Removing the unneeded FDB flags will help to save the resources of
memory and STCs.

Fixes: 07836ba156cb ("net/mlx5: fix device checking for send to kernel action")
Cc: jiaw...@nvidia.com

Signed-off-by: Bing Zhao 
---
 drivers/net/mlx5/mlx5_flow_hw.c | 14 ++
 1 file changed, 6 insertions(+), 8 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_flow_hw.c b/drivers/net/mlx5/mlx5_flow_hw.c
index 1cc074d..f0888db 100644
--- a/drivers/net/mlx5/mlx5_flow_hw.c
+++ b/drivers/net/mlx5/mlx5_flow_hw.c
@@ -9649,21 +9649,18 @@ _create_send_to_kernel_actions(struct mlx5_priv *priv, 
int type)
 }
 
 static void
-flow_hw_create_send_to_kernel_actions(struct mlx5_priv *priv)
+flow_hw_create_send_to_kernel_actions(struct mlx5_priv *priv, bool is_proxy)
 {
 #ifdef HAVE_MLX5DV_DR_ACTION_CREATE_DEST_ROOT_TABLE
int i, from, to;
-   bool is_vf_sf_dev = priv->sh->dev_cap.vf || priv->sh->dev_cap.sf;
bool unified_fdb = is_unified_fdb(priv);
 
for (i = MLX5DR_TABLE_TYPE_NIC_RX; i <= MLX5DR_TABLE_TYPE_NIC_TX; i++)
_create_send_to_kernel_actions(priv, i);
 
-   if (priv->sh->config.dv_esw_en && !is_vf_sf_dev) {
-   from = unified_fdb ? MLX5DR_TABLE_TYPE_FDB_RX :
-   MLX5DR_TABLE_TYPE_FDB;
-   to = unified_fdb ? MLX5DR_TABLE_TYPE_FDB_UNIFIED :
- MLX5DR_TABLE_TYPE_FDB;
+   if (is_proxy) {
+   from = unified_fdb ? MLX5DR_TABLE_TYPE_FDB_RX : 
MLX5DR_TABLE_TYPE_FDB;
+   to = unified_fdb ? MLX5DR_TABLE_TYPE_FDB_UNIFIED : 
MLX5DR_TABLE_TYPE_FDB;
for (i = from; i <= to; i++)
_create_send_to_kernel_actions(priv, i);
}
@@ -9674,6 +9671,7 @@ static void
 flow_hw_destroy_send_to_kernel_action(struct mlx5_priv *priv)
 {
int i;
+
for (i = MLX5DR_TABLE_TYPE_NIC_RX; i < MLX5DR_TABLE_TYPE_MAX; i++) {
if (priv->hw_send_to_kernel[i]) {
mlx5dr_action_destroy(priv->hw_send_to_kernel[i]);
@@ -12171,7 +12169,7 @@ __flow_hw_configure(struct rte_eth_dev *dev,
}
}
if (!priv->shared_host)
-   flow_hw_create_send_to_kernel_actions(priv);
+   flow_hw_create_send_to_kernel_actions(priv, is_proxy);
if (port_attr->nb_conn_tracks || (host_priv && host_priv->hws_ctpool)) {
if (mlx5_flow_ct_init(dev, port_attr->nb_conn_tracks, 
nb_q_updated))
goto err;
-- 
2.8.4



[PATCH V2 4/4] net/mlx5/hws: allow different types in miss validation

2025-02-26 Thread Ophir Munk
In validating Jump from group A to default miss group B DPDK requests
that both groups types will be identical. In practice jumping from
FDB_UNIFIED to FDB_RX or FDB_TX group types (or vice versa) is allowed.
This commit enables this jumping between the different types.

Fixes: b81f95ca770d ("net/mlx5/hws: support default miss table")

Signed-off-by: Ophir Munk 
---
 drivers/net/mlx5/hws/mlx5dr_table.c | 18 --
 1 file changed, 16 insertions(+), 2 deletions(-)

diff --git a/drivers/net/mlx5/hws/mlx5dr_table.c 
b/drivers/net/mlx5/hws/mlx5dr_table.c
index b371f42..c1c60b4 100644
--- a/drivers/net/mlx5/hws/mlx5dr_table.c
+++ b/drivers/net/mlx5/hws/mlx5dr_table.c
@@ -626,6 +626,19 @@ int mlx5dr_table_connect_to_miss_table(struct mlx5dr_table 
*src_tbl,
return 0;
 }
 
+static bool mlx5dr_table_set_default_miss_valid_types(enum mlx5dr_table_type 
from,
+ enum mlx5dr_table_type to)
+{
+   if (from == to ||
+   ((from == MLX5DR_TABLE_TYPE_FDB_UNIFIED &&
+(to == MLX5DR_TABLE_TYPE_FDB_RX || to == 
MLX5DR_TABLE_TYPE_FDB_TX)) ||
+(to == MLX5DR_TABLE_TYPE_FDB_UNIFIED &&
+(from == MLX5DR_TABLE_TYPE_FDB_RX || from == 
MLX5DR_TABLE_TYPE_FDB_TX
+   return true;
+
+   return false;
+}
+
 static int mlx5dr_table_set_default_miss_not_valid(struct mlx5dr_table *tbl,
   struct mlx5dr_table 
*miss_tbl)
 {
@@ -637,8 +650,9 @@ static int mlx5dr_table_set_default_miss_not_valid(struct 
mlx5dr_table *tbl,
}
 
if (mlx5dr_table_is_root(tbl) ||
-   (miss_tbl && mlx5dr_table_is_root(miss_tbl)) ||
-   (miss_tbl && miss_tbl->type != tbl->type)) {
+   (miss_tbl &&
+((mlx5dr_table_is_root(miss_tbl)) ||
+!mlx5dr_table_set_default_miss_valid_types(tbl->type, 
miss_tbl->type {
DR_LOG(ERR, "Invalid arguments");
rte_errno = EINVAL;
return -rte_errno;
-- 
2.8.4



RE: [PATCH] net/mlx5: fix packet hardware ptype translation

2025-02-26 Thread Dariusz Sosnowski



> -Original Message-
> From: Gregory Etelson 
> Sent: Sunday, February 9, 2025 3:01 PM
> To: dev@dpdk.org
> Cc: Gregory Etelson ; Maayan Kashani
> ; Raslan Darawsheh ; Slava
> Ovsiienko ; sta...@dpdk.org
> Subject: [PATCH] net/mlx5: fix packet hardware ptype translation
> 
> External email: Use caution opening links or attachments
> 
> 
> From: Viacheslav Ovsiienko 
> 
> Since ConnectX-8 NIC the hardware recognizes L4 protocol in the IP fragments 
> if
> the IP offset is zero and reports the L4 protocol type in CQE fields. This 
> case was
> not supported in the ptype translation table cause not reporting the ptype in 
> the
> mbufs.
> 
> Fixes: ea16068c0064 ("net/mlx5: fix L4 packet type support")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Viacheslav Ovsiienko 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


Re: [PATCH v3] net/mlx5: add eCPRI support

2025-02-26 Thread Raslan Darawsheh
Hi,

From: Bing Zhao 
Sent: Wednesday, February 26, 2025 9:17 AM
To: Slava Ovsiienko; dev@dpdk.org; Raslan Darawsheh
Cc: Ori Kam; Dariusz Sosnowski; Suanming Mou; Matan Azrad; Hamdan Agbariya
Subject: [PATCH v3] net/mlx5: add eCPRI support

Support eCPRI item matching over L2 (ETHER / VLAN) in HWS, both for
template API and backward compatibility API.

Signed-off-by: Bing Zhao 
Acked-by: Dariusz Sosnowski 
---
v2: fix the checkpatch warning of unneeded else
v3: fix single line if condition code style warning
---
 
Patch applied to next-net-mlx,

Kindest regards,
Raslan Darawsheh


RE: [PATCH V2 1/4] common/mlx5: support FDB unified capability query

2025-02-26 Thread Dariusz Sosnowski



> -Original Message-
> From: Ophir Munk 
> Sent: Wednesday, February 26, 2025 9:39 AM
> To: Dariusz Sosnowski ; Slava Ovsiienko
> ; Bing Zhao ; Ori Kam
> ; Suanming Mou ; Matan Azrad
> 
> Cc: dev@dpdk.org; Raslan Darawsheh 
> Subject: [PATCH V2 1/4] common/mlx5: support FDB unified capability query
> 
> This commit queries the FW for the new unified FDB mode and saves it in
> mlx5 shared device as fdb_unified_en bit.
> 
> Signed-off-by: Ophir Munk 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


RE: [PATCH V2 2/4] net/mlx5: support FDB unified domain

2025-02-26 Thread Dariusz Sosnowski



> -Original Message-
> From: Ophir Munk 
> Sent: Wednesday, February 26, 2025 9:39 AM
> To: Dariusz Sosnowski ; Slava Ovsiienko
> ; Bing Zhao ; Ori Kam
> ; Suanming Mou ; Matan Azrad
> 
> Cc: dev@dpdk.org; Raslan Darawsheh 
> Subject: [PATCH V2 2/4] net/mlx5: support FDB unified domain
> 
> Legacy FDB domain is enhanced with optional three sub-domains: FDB_RX
> FDB_TX and FDB_UNIFIED. Based on FW capability bit mlx5 will work in either
> legacy or unified mode. To work in fdb unified mode the flow attribute 
> transfer
> bit must be set and the fdb sub domains are determined by the table specialize
> flags SPECIALIZE_TRANSFER_WIRE_ORIG and
> SPECIALIZE_TRANSFER_VPORT_ORIG.
> 
> ++-+--+
> | WIRE_ORIG  | VPORT_ORIG  | FDB domain   |
> ++-+--+
> | 0  | 0   | FDB_UNIFIED  |
> | 0  | 1   | FDB_TX   |
> | 1  | 0   | FDB_RX   |
> ++-+--+
> 
> There can be no two groups which only differ in their fdb sub domain.
> A table must be created in a group which has the same fdb sub domain.
> 
> Signed-off-by: Ophir Munk 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


RE: [PATCH V2 3/4] net/mlx5: remove unneeded FDB flag on representor action

2025-02-26 Thread Dariusz Sosnowski



> -Original Message-
> From: Ophir Munk 
> Sent: Wednesday, February 26, 2025 9:39 AM
> To: Dariusz Sosnowski ; Slava Ovsiienko
> ; Bing Zhao ; Ori Kam
> ; Suanming Mou ; Matan Azrad
> 
> Cc: dev@dpdk.org; Raslan Darawsheh ; Jiawei(Jonny)
> Wang 
> Subject: [PATCH V2 3/4] net/mlx5: remove unneeded FDB flag on representor
> action
> 
> From: Bing Zhao 
> 
> The "send_to_kernel" action is shared and created during the HWS configuration
> stage. By default, the NIC Rx and Tx flags are needed.
> 
> There is only one case that the FDB* flags are needed when creating the 
> action,
> the same as other shared actions. Only the transfer proxy port needs the FDB*
> flags for shared actions.
> 
> Removing the unneeded FDB flags will help to save the resources of memory and
> STCs.
> 
> Fixes: 07836ba156cb ("net/mlx5: fix device checking for send to kernel 
> action")
> Cc: jiaw...@nvidia.com
> 
> Signed-off-by: Bing Zhao 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


Re: [PATCH 3/6] config: allow faster instruction sets to be used with MSVC

2025-02-26 Thread Bruce Richardson
On Tue, Feb 25, 2025 at 06:01:38PM -0800, Andre Muezerie wrote:
> On Tue, Feb 25, 2025 at 02:28:02PM +, Bruce Richardson wrote:
> > On Mon, Feb 24, 2025 at 01:01:16PM -0800, Andre Muezerie wrote:
> > > Up to now MSVC has being used with the default mode, which uses SSE2
> > > instructions for scalar floating-point and vector calculations.
> > > https://learn.microsoft.com/en-us/cpp/build/reference/arch-x64?view=msvc-170
> > > 
> > > This patch allows users to specify the CPU for which the generated
> > > code should be optimized for in the same way it's done for GCC: by
> > > passing the CPU name.
> > > When no explicit CPU name is passed, 'native' is assumed (like it
> > > happens with GCC) and the code will be optimized for the same CPU
> > > type used to compile the code.
> > > 
> > > MSVC does not provide this functionality natively, so logic was
> > > added to meson.build to handle these differences, detecting which
> > > instruction sets are supported by the CPU(s), passing the best
> > > options to MSVC and setting the correct macros (like __AVX512F__)
> > > so that the DPDK code can rely on them like it is done with GCC.
> > > 
> > > Signed-off-by: Andre Muezerie 
> > > ---
> > 
> > Hi Andre,
> > 
> > couple of initial thoughts inline below.
> > 
> > /Bruce
> > 
> > >  config/x86/meson.build | 364 -
> > >  1 file changed, 325 insertions(+), 39 deletions(-)
> > > 
> > 
> > There is quite a lot of new code to be added here. Might it be worthwhile
> > creating a "config/x86/msvc/" subdirectory with its own meson.build file to
> > handle all the complexities of using it. We can have the common material at
> > the top of the x86/meson.build file, and then do 
> > 
> > if is_ms_compiler
> > subdir(msvc)
> > subdir_done()
> > endif
> > 
> > leaving the rest of the file for the gcc/clang/icx code.
> 
> I think that makes sense, as there's not much common code there that is 
> common to gcc and msvc.
> 



> > 
> > I really don't want to have tables like this to maintain in our code if at
> > all possible. We used to have something a bit similar in DPDK IIRC, but we
> > found it a maintenance nightmare and just switched to using the compiler to
> > do all the work. In our existing builds, we just pass the
> > cpu_instruction_set parameter straight to the -march flag of the compiler.
> > For MSVC support, I believe we should just do the exact same.
> > 
> > Maintaining lists like this will be a problem as new platforms need to be
> > constantly added.
> 
> It's great that when using gcc users can just pass the CPU type to it and
> that it will set all the macros corresponding to that CPU for them. I
> would love to be able to rely on MSVC for that as well, unfortunately MSVC
> does not provide that level of granularity, that's why I came up with the
> idea of having this table.
> 
> Initially I was also very concerned about the amount of data to be stored
> there, and the work required to maintain it. Then I decided to throw away
> all the CPU types that do not have SSE4_2. That reduced the table in half.
> Adding the entries manually was still a lot of work and error prone. So,
> I decided to write some code that uses gcc to build that table for me.
> I'll polish that code and add it to the patch. With that it will be almost
> zero effort to maintain that table. All that will be required is running
> that code on a setup with the latest gcc, providing a file with the CPU
> names. Assuming gcc knows about the latest CPUs a new table will be
> generated and can be pasted in the meson.build file.
> 
> I assume this does not need to be done too often. In the worst case, if it
> happens that DPDK was not updated with the latest CPUs, people can still
> pick an earlier CPU with similar characteristics and have similar (if not
> same) performance.
> 

Thanks. Having it automated certainly helps, but I'm still not fully
convinced that it's a good idea. If everyone else is happy, though, I'm ok
to ignore those concerns :-). Let's see what others think.

> 
> > Do we also look to backport them, because if equivalence
> > with the linux build is necessary then that will have to be done - as on
> > Linux when a new version of GCC comes out, we can then use the new
> > instruction set targets on the old releases of DPDK.
> 
> That would be nice, and I'm willing to help with that. It makes it a
> better user experience if we can minimize the perceived differences
> between the toolsets. I'm not sure if it's a requirement though. If
> you're concerned that this would add too much overhead, it could be
> decided that no such backport should happen.
> 
> > 



> > > +if 'RTE_CPUFLAG_AVX512F' in compile_time_cpuflags
> > > +machine_args += ['/arch:AVX512']
> > > +elif 'RTE_CPUFLAG_AVX2' in compile_time_cpuflags
> > > +machine_args += ['/arch:AVX2']
> > > +elif 'RTE_CPUFLAG_AVX' in compile_time_cpuflags
> > > +machine_args += ['/arch:AVX']

Re: [PATCH v2 1/5] eal: make compatible with instruction set updates for MSVC

2025-02-26 Thread Bruce Richardson
On Tue, Feb 25, 2025 at 05:06:52PM -0800, Andre Muezerie wrote:
> After the instruction set updates for MSVC the error below popped up:
> 
> ../lib/eal/common/rte_random.c(6): fatal error C1083:
> Cannot open include file: 'x86intrin.h': No such file or directory
> 
> The fix is to not include header x86intrin.h with MSVC.
> 
> Signed-off-by: Andre Muezerie 
> ---
Acked-by: Bruce Richardson 



Re: [PATCH v2 2/5] eal: only use numbers as align parameters for MSVC

2025-02-26 Thread Bruce Richardson
On Tue, Feb 25, 2025 at 05:06:53PM -0800, Andre Muezerie wrote:
> After the instruction set updates for MSVC the error below popped up:
> 
> ..\lib\eal\x86\include\rte_vect.h(82): error C2059: syntax error: '('
> 
> The issue is that MSVC does not allow __rte_aligned(RTE_X86_ZMM_SIZE).
> It only accepts numbers that are power of 2. So, even though
> RTE_X86_ZMM_SIZE represents a number that is a power of two it cannot
> be used directly.
> https://learn.microsoft.com/en-us/cpp/cpp/align-cpp?view=msvc-170
> 
> Signed-off-by: Andre Muezerie 
> ---
Acked-by: Bruce Richardson 



Re: [PATCH v2 3/5] config: create top level variable cc_avx2_flags

2025-02-26 Thread Bruce Richardson
On Tue, Feb 25, 2025 at 05:06:54PM -0800, Andre Muezerie wrote:
> Top level variable cc_avx2_flags was created and set according to the
> compiler used. It is used in lower level directories.
> 
> Signed-off-by: Andre Muezerie 
> ---
>  config/x86/meson.build | 6 ++
>  1 file changed, 6 insertions(+)
> 
Acked-by: Bruce Richardson 



RE: [PATCH v2 2/5] eal: only use numbers as align parameters for MSVC

2025-02-26 Thread Konstantin Ananyev



> -Original Message-
> From: Andre Muezerie 
> Sent: Wednesday, February 26, 2025 1:07 AM
> To: andre...@linux.microsoft.com
> Cc: dev@dpdk.org
> Subject: [PATCH v2 2/5] eal: only use numbers as align parameters for MSVC
> 
> After the instruction set updates for MSVC the error below popped up:
> 
> ..\lib\eal\x86\include\rte_vect.h(82): error C2059: syntax error: '('
> 
> The issue is that MSVC does not allow __rte_aligned(RTE_X86_ZMM_SIZE).
> It only accepts numbers that are power of 2. So, even though
> RTE_X86_ZMM_SIZE represents a number that is a power of two it cannot
> be used directly.
> https://learn.microsoft.com/en-us/cpp/cpp/align-cpp?view=msvc-170
> 
> Signed-off-by: Andre Muezerie 
> ---
>  lib/eal/x86/include/rte_vect.h | 11 ++-
>  1 file changed, 10 insertions(+), 1 deletion(-)
> 
> diff --git a/lib/eal/x86/include/rte_vect.h b/lib/eal/x86/include/rte_vect.h
> index 70c78e9b77..0a51c539a4 100644
> --- a/lib/eal/x86/include/rte_vect.h
> +++ b/lib/eal/x86/include/rte_vect.h
> @@ -79,7 +79,16 @@ __extension__ ({\
>  #define RTE_X86_ZMM_SIZE (sizeof(__m512i))
>  #define RTE_X86_ZMM_MASK (RTE_X86_ZMM_SIZE - 1)
> 
> -typedef union __rte_aligned(RTE_X86_ZMM_SIZE) __rte_x86_zmm {
> +/*
> + * MSVC does not allow __rte_aligned(RTE_X86_ZMM_SIZE). It only accepts
> + * numbers that are power of 2. So, even though RTE_X86_ZMM_SIZE represents a
> + * number that is a power of two it cannot be used directly.
> + * Ref: https://learn.microsoft.com/en-us/cpp/cpp/align-cpp?view=msvc-170
> + * The static assert below ensures that RTE_X86_ZMM_SIZE is equal to what is
> + * used in the __rte_aligned() expression.
> + */
> +static_assert(RTE_X86_ZMM_SIZE == 64, "Unexpected size of __m512i");
> +typedef union __rte_aligned(64) __rte_x86_zmm {

Just wonder, would then MSVC understand something like:
#define RTE_X86_ZMM_SIZE64
static_assert(RTE_X86_ZMM_SIZE == sizeof((__m512i), "Unexpected size of 
__m512i");
typedef union __rte_aligned(RTE_X86_ZMM_SIZE) __rte_x86_zmm {
? 

>   __m512i  z;
>   ymm_ty[RTE_X86_ZMM_SIZE / sizeof(ymm_t)];
>   xmm_tx[RTE_X86_ZMM_SIZE / sizeof(xmm_t)];
> --
> 2.48.1.vfs.0.0



RE: [PATCH 1/3] net/mlx5/hws: support jump FDB Rx

2025-02-26 Thread Bing Zhao
Hi Suanming,

PSB

> -Original Message-
> From: Suanming Mou 
> Sent: Tuesday, February 25, 2025 8:45 AM
> To: Dariusz Sosnowski ; Slava Ovsiienko
> ; Bing Zhao ; Ori Kam
> ; Matan Azrad 
> Cc: dev@dpdk.org; Raslan Darawsheh ; Alex Vesker
> 
> Subject: [PATCH 1/3] net/mlx5/hws: support jump FDB Rx
> 
> Before FW introduced the JUMP_FDB_RX action feature, jump from FDB Tx to
> Rx is not allowed. JUMP_FDB_RX feature introduces the internal loopback
> for Tx case and allow the REG C0 C1 B be preserved as well.
> 
> This commit adds the JUMP_FDB_RX cap bit check and use JUMP_FDB_RX instead
> of FT for dest table FDB Rx case.
> 
> Signed-off-by: Suanming Mou 
> Signed-off-by: Alex Vesker 
> ---
>  drivers/common/mlx5/mlx5_prm.h|  9 -
>  drivers/net/mlx5/hws/mlx5dr_action.c  | 26 +-
> drivers/net/mlx5/hws/mlx5dr_action.h  |  4 
>  drivers/net/mlx5/hws/mlx5dr_cmd.c |  9 +
>  drivers/net/mlx5/hws/mlx5dr_cmd.h |  2 ++
>  drivers/net/mlx5/hws/mlx5dr_context.c | 17 +
> drivers/net/mlx5/hws/mlx5dr_context.h |  2 ++
>  7 files changed, 63 insertions(+), 6 deletions(-)
> 
> diff --git a/drivers/common/mlx5/mlx5_prm.h
> b/drivers/common/mlx5/mlx5_prm.h index 3fc3b0cd2a..84e3347794 100644
> --- a/drivers/common/mlx5/mlx5_prm.h
> +++ b/drivers/common/mlx5/mlx5_prm.h
> @@ -2466,7 +2466,8 @@ struct mlx5_ifc_wqe_based_flow_table_cap_bits {
>   u8 reserved_at_60[0x8];
>   u8 max_header_modify_pattern_length[0x8];
>   u8 ste_format[0x10];
> - u8 stc_action_type[0x80];
> + u8 stc_action_type_63_0[0x40];
> + u8 stc_action_type_127_64[0x40];
>   u8 header_insert_type[0x10];
>   u8 header_remove_type[0x10];
>   u8 trivial_match_definer[0x20];
> @@ -3543,6 +3544,11 @@ enum mlx5_ifc_rtc_reparse_mode {
>   MLX5_IFC_RTC_REPARSE_BY_STC = 0x2,
>  };
> 
> +enum mlx5_ifc_stc_action_type_bit_index {
> + MLX5_IFC_STC_ACTION_TYPE_BIT_64_INDEX = 64,
> + MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX = 71, };
> +
>  #define MLX5_IFC_RTC_LINEAR_LOOKUP_TBL_LOG_MAX 16
> 
>  struct mlx5_ifc_rtc_bits {
> @@ -3621,6 +3627,7 @@ enum mlx5_ifc_stc_action_type {
>   MLX5_IFC_STC_ACTION_TYPE_ALLOW = 0x84,
>   MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_VPORT = 0x85,
>   MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_UPLINK = 0x86,
> + MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FLOW_TABLE_FDB_RX = 0x87,
>  };
> 
>  enum mlx5_ifc_stc_reparse_mode {
> diff --git a/drivers/net/mlx5/hws/mlx5dr_action.c
> b/drivers/net/mlx5/hws/mlx5dr_action.c
> index b9452a3ebc..e21db5b327 100644
> --- a/drivers/net/mlx5/hws/mlx5dr_action.c
> +++ b/drivers/net/mlx5/hws/mlx5dr_action.c
> @@ -803,6 +803,9 @@ int mlx5dr_action_root_build_attr(struct
> mlx5dr_rule_action rule_actions[],
> 
>   switch (action->type) {
>   case MLX5DR_ACTION_TYP_TBL:
> + attr[i].type = MLX5DV_FLOW_ACTION_DEST_DEVX;
> + attr[i].obj = action->dest_tbl.devx_obj->obj;
> + break;
>   case MLX5DR_ACTION_TYP_TIR:
>   attr[i].type = MLX5DV_FLOW_ACTION_DEST_DEVX;
>   attr[i].obj = action->devx_obj;
> @@ -1097,6 +1100,17 @@ static void mlx5dr_action_fill_stc_attr(struct
> mlx5dr_action *action,
>   }
>   break;
>   case MLX5DR_ACTION_TYP_TBL:
> + attr->action_offset = MLX5DR_ACTION_OFFSET_HIT;
> + attr->dest_table_id = obj->id;
> + /* Only for unified FDB Rx case */
> + if (mlx5dr_context_cap_stc(action->ctx,
> + MLX5_IFC_STC_ACTION_TYPE_JUMP_FLOW_TABLE_FDB_RX_BIT_INDEX)
> &&
> + action->dest_tbl.type == MLX5DR_TABLE_TYPE_FDB_RX)
> + attr->action_type =
> MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FLOW_TABLE_FDB_RX;
> + else
> + attr->action_type = MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FT;

Fix the long line warnings, same as other places.
Also mlx5dr_context_cap_stc() seems to contain some calculation inside. If the 
context and the CAP will not be changed during the runtime on a specific port. 
Can this CAP BIT checking be done only once and saved for later usage? Since 
the code is used in the flow insertion.WDYT?

> +
> + break;
>   case MLX5DR_ACTION_TYP_DEST_ARRAY:
>   attr->action_type = MLX5_IFC_STC_ACTION_TYPE_JUMP_TO_FT;
>   attr->action_offset = MLX5DR_ACTION_OFFSET_HIT; @@ -1419,17
> +1433,19 @@ mlx5dr_action_create_dest_table(struct mlx5dr_context *ctx,
>   if (!action)
>   return NULL;
> 
> + action->dest_tbl.type = tbl->type;
> +
>   if (mlx5dr_action_is_root_flags(flags)) {
>   if (mlx5dr_context_shared_gvmi_used(ctx))
> - action->devx_obj = tbl->local_ft->obj;
> + action->dest_tbl.devx_obj = tbl->local_ft;
>   else
> - action->devx_obj = tbl->ft->obj

RE: [PATCH v5 1/3] cryptodev: add SM4 GCM

2025-02-26 Thread De Lara Guarch, Pablo



> -Original Message-
> From: Dooley, Brian 
> Sent: Monday, February 24, 2025 6:00 PM
> To: Akhil Goyal ; Fan Zhang 
> Cc: dev@dpdk.org; Ji, Kai ; De Lara Guarch, Pablo
> ; Dooley, Brian 
> Subject: [PATCH v5 1/3] cryptodev: add SM4 GCM
> 
> Added support for symmetric algorithm SM4 GCM in cryptodev.
> 
> Signed-off-by: Brian Dooley 

Acked-by: Pablo de Lara 


RE: [PATCH v5 3/3] app/test: add SM4 GCM tests

2025-02-26 Thread De Lara Guarch, Pablo



> -Original Message-
> From: Dooley, Brian 
> Sent: Monday, February 24, 2025 6:00 PM
> To: Akhil Goyal ; Fan Zhang 
> Cc: dev@dpdk.org; Ji, Kai ; De Lara Guarch, Pablo
> ; Dooley, Brian 
> Subject: [PATCH v5 3/3] app/test: add SM4 GCM tests
> 
> Added SM4-GCM tests for the AESNI MB PMD.
> 
> Signed-off-by: Brian Dooley 

Acked-by: Pablo de Lara 


RE: [PATCH v5 2/3] crypto/ipsec_mb: add SM4 GCM support

2025-02-26 Thread De Lara Guarch, Pablo



> -Original Message-
> From: Dooley, Brian 
> Sent: Monday, February 24, 2025 6:00 PM
> To: Ji, Kai ; De Lara Guarch, Pablo
> 
> Cc: dev@dpdk.org; gak...@marvell.com; Dooley, Brian
> 
> Subject: [PATCH v5 2/3] crypto/ipsec_mb: add SM4 GCM support
> 
> This patch introduces SM4 GCM algorithm support to the AESNI_MB PMD.
> SM4 GCM is available in the v2.0 release of Intel IPsec MB.
> 
> Signed-off-by: Brian Dooley 

Acked-by: Pablo de Lara