RE: [PATCH v4 08/10] test-pmd: declare lcore_count atomic

2025-02-21 Thread Konstantin Ananyev



> Compiling with MSVC results in the error below:
> 
> app/test/test_ring_perf.c(197): error C7712: address argument to atomic
> operation must be a pointer to an atomic integer,
> 'volatile unsigned int *' is not valid
> 
> The fix is to mark lcore_count as atomic.
> 
> Signed-off-by: Andre Muezerie 
> Signed-off-by: Chengwen Feng 
> ---
>  app/test/test_ring_perf.c | 6 +-
>  1 file changed, 1 insertion(+), 5 deletions(-)
> 
> diff --git a/app/test/test_ring_perf.c b/app/test/test_ring_perf.c
> index 57cd04a124..366e256323 100644
> --- a/app/test/test_ring_perf.c
> +++ b/app/test/test_ring_perf.c
> @@ -34,7 +34,7 @@ struct lcore_pair {
>   unsigned c1, c2;
>  };
> 
> -static volatile unsigned lcore_count = 0;
> +static RTE_ATOMIC(unsigned int) lcore_count;

But for not-MSVC builds RTE_ATOMIC() by default means nothing.
If you are removing volatile, then I think, you need to fix the code to ensure
that all reads/writes to this var are done with atomic ops.
Or have both -  volatile and RTE_ATOMIC()


>  static void
>  test_ring_print_test_string(unsigned int api_type, int esize,
> @@ -193,11 +193,7 @@ enqueue_dequeue_bulk_helper(const unsigned int flag, 
> struct thread_params *p)
>   unsigned int n_remaining;
>   const unsigned int bulk_n = bulk_sizes[p->ring_params->bulk_sizes_i];
> 
> -#ifdef RTE_USE_C11_MEM_MODEL
>   if (rte_atomic_fetch_add_explicit(&lcore_count, 1, 
> rte_memory_order_relaxed) + 1 != 2)
> -#else
> - if (__sync_add_and_fetch(&lcore_count, 1) != 2)
> -#endif
>   while(lcore_count != 2)
>   rte_pause();
> 
> --
> 2.48.1.vfs.0.0
> 



RE: [PATCH v2] examples/l3fwd: add option to set Tx burst size

2025-02-21 Thread Konstantin Ananyev


> >> On 2024/12/4 10:06, Jie Hai wrote:
> >>> The application send packets only when the buffer is full, or the
> >>> buffer is empty and the packets to be sent extends TX_PKT_BURST.
> >>> The change of MAX_PKT_BURST make TX buffer size and TX_PKT_BURST
> >>> increase, while the default cache size is 256. The packets in
> >>> the TX direction occupy the cache. As a result, the performance
> >>> deteriorates.
> >>>
> >>> Restore the default Tx burst and add option '--tx-burst' to set
> >>> the Tx burst size. To ensure consistency, rename the option
> >>> '--burst' to '--rx-burst'. The valid range of the user-provided
> >>> value is (0, MAX_PKT_BURST] for both directions.
> >>>
> >>> Fixes: d5c4897ecfb2 ("examples/l3fwd: add option to set RX burst size")
> >>> Cc: sta...@dpdk.org
> >
> > Wouldn't that patch:
> > https://patchwork.dpdk.org/project/dpdk/patch/20250212045416.2393001-1-sivaprasad.tumm...@amd.com/
> > fix that issue too?
> > And probably in a less disruptive manner?
> >
> Thanks for your comment.
> 
> I think different drivers may need different values for better
> performance, it makes sense to add options.

 Well, we used to run l3fwd for more then 10 years by now,
and I don’t' remember that anyone claim that it is needed.
Again, AFAIR, our other sample apps including test-pmd uses
one burst size for both RX and TX.
Do you have any perf numbers to back-up your claim?  
 
> 
> >>> Signed-off-by: Jie Hai 
> >>> ---
> >>>examples/l3fwd/l3fwd.h| 14 +++---
> >>>examples/l3fwd/l3fwd_acl.c|  2 +-
> >>>examples/l3fwd/l3fwd_common.h |  2 +-
> >>>examples/l3fwd/l3fwd_em.c |  2 +-
> >>>examples/l3fwd/l3fwd_fib.c|  2 +-
> >>>examples/l3fwd/l3fwd_lpm.c|  2 +-
> >>>examples/l3fwd/main.c | 89 ++-
> >>>7 files changed, 59 insertions(+), 54 deletions(-)
> >>>
> >>> diff --git a/examples/l3fwd/l3fwd.h b/examples/l3fwd/l3fwd.h
> >>> index 0cce3406ee7d..a4e23b817edf 100644
> >>> --- a/examples/l3fwd/l3fwd.h
> >>> +++ b/examples/l3fwd/l3fwd.h
> >>> @@ -32,10 +32,6 @@
> >>>
> >>>#define VECTOR_SIZE_DEFAULT   MAX_PKT_BURST
> >>>#define VECTOR_TMO_NS_DEFAULT 1E6 /* 1ms */
> >>> -/*
> >>> - * Try to avoid TX buffering if we have at least MAX_TX_BURST packets to 
> >>> send.
> >>> - */
> >>> -#define  MAX_TX_BURST  (MAX_PKT_BURST / 2)
> >>>
> >>>#define NB_SOCKETS8
> >>>
> >>> @@ -116,7 +112,11 @@ extern struct acl_algorithms acl_alg[];
> >>>
> >>>extern uint32_t max_pkt_len;
> >>>
> >>> -extern uint32_t nb_pkt_per_burst;
> >>> +extern uint32_t rx_pkt_burst;
> >>> +/*
> >>> + * Try to avoid TX buffering if we have at least tx_pkt_burst packets to 
> >>> send.
> >>> + */
> >>> +extern uint32_t tx_pkt_burst;
> >>>extern uint32_t mb_mempool_cache_size;
> >>>
> >>>/* Send burst of packets on an output interface */
> >>> @@ -152,8 +152,8 @@ send_single_packet(struct lcore_conf *qconf,
> >>>   len++;
> >>>
> >>>   /* enough pkts to be sent */
> >>> - if (unlikely(len == MAX_PKT_BURST)) {
> >>> - send_burst(qconf, MAX_PKT_BURST, port);
> >>> + if (unlikely(len == tx_pkt_burst)) {
> >>> + send_burst(qconf, tx_pkt_burst, port);
> >>>   len = 0;
> >>>   }
> >>>
> >>> diff --git a/examples/l3fwd/l3fwd_acl.c b/examples/l3fwd/l3fwd_acl.c
> >>> index 4fc4b986cce6..a5af82357a03 100644
> >>> --- a/examples/l3fwd/l3fwd_acl.c
> >>> +++ b/examples/l3fwd/l3fwd_acl.c
> >>> @@ -1136,7 +1136,7 @@ acl_main_loop(__rte_unused void *dummy)
> >>>   portid = qconf->rx_queue_list[i].port_id;
> >>>   queueid = qconf->rx_queue_list[i].queue_id;
> >>>   nb_rx = rte_eth_rx_burst(portid, queueid,
> >>> - pkts_burst, nb_pkt_per_burst);
> >>> + pkts_burst, rx_pkt_burst);
> >>>
> >>>   if (nb_rx > 0) {
> >>>   nb_drop = acl_process_pkts(pkts_burst, 
> >>> hops,
> >>> diff --git a/examples/l3fwd/l3fwd_common.h b/examples/l3fwd/l3fwd_common.h
> >>> index d94e5f135791..34fe70b9415c 100644
> >>> --- a/examples/l3fwd/l3fwd_common.h
> >>> +++ b/examples/l3fwd/l3fwd_common.h
> >>> @@ -71,7 +71,7 @@ send_packetsx4(struct lcore_conf *qconf, uint16_t port, 
> >>> struct rte_mbuf *m[],
> >>>* If TX buffer for that queue is empty, and we have enough 
> >>> packets,
> >>>* then send them straightway.
> >>>*/
> >>> - if (num >= MAX_TX_BURST && len == 0) {
> >>> + if (num >= tx_pkt_burst / 2 && len == 0) {
> >>>   n = rte_eth_tx_burst(port, qconf->tx_queue_id[port], m, 
> >>> num);
> >>>   if (unlikely(n < num)) {
> >>>   do {
> >>> diff --git a/examples/l3fwd/l3fwd_em.c b/examples/l3fwd/l3fwd_em.c
> >>> index da9c45e3a482..ea74506ed971 100644
> >>> --- a/examples/l3fwd/l3fwd_em.c
> >>> +++ b/examples/l3fwd/l3fwd_em.c
> >>

[PATCH] net/ixgbe: fix min Rx/Tx descriptors

2025-02-21 Thread Mingjin Ye
The minimum free packet threshold (tx_free_thresh) and the minimum RS bit
threshold (tx_rs_thresh) both have a default value of 32. Therefore, the
default minimum number of ring descriptors value is 64.

Fixes: dee5f1fd5fc7 ("ixgbe: get queue info and descriptor limits")
Cc: sta...@dpdk.org

Signed-off-by: Mingjin Ye 
---
 doc/guides/nics/ixgbe.rst| 2 +-
 drivers/net/intel/ixgbe/ixgbe_rxtx.h | 2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)

diff --git a/doc/guides/nics/ixgbe.rst b/doc/guides/nics/ixgbe.rst
index c5c6a6c34b..10a0cdd270 100644
--- a/doc/guides/nics/ixgbe.rst
+++ b/doc/guides/nics/ixgbe.rst
@@ -76,7 +76,7 @@ Scattered packets are not supported in this mode.
 If an incoming packet is greater than the maximum acceptable length of one 
"mbuf" data size (by default, the size is 2 KB),
 vPMD for RX would be disabled.
 
-By default, IXGBE_MAX_RING_DESC is set to 8192 and RTE_PMD_IXGBE_RX_MAX_BURST 
is set to 32.
+By default, IXGBE_MAX_RING_DESC is set to 8192 and RTE_PMD_IXGBE_RX_MAX_BURST 
is set to 64.
 
 Windows Prerequisites and Pre-conditions
 
diff --git a/drivers/net/intel/ixgbe/ixgbe_rxtx.h 
b/drivers/net/intel/ixgbe/ixgbe_rxtx.h
index 278f665108..54569c7ade 100644
--- a/drivers/net/intel/ixgbe/ixgbe_rxtx.h
+++ b/drivers/net/intel/ixgbe/ixgbe_rxtx.h
@@ -26,7 +26,7 @@
  * descriptors should meet the following condition:
  *  (num_ring_desc * sizeof(rx/tx descriptor)) % 128 == 0
  */
-#defineIXGBE_MIN_RING_DESC 32
+#defineIXGBE_MIN_RING_DESC 64
 #defineIXGBE_MAX_RING_DESC 8192
 
 #define RTE_PMD_IXGBE_TX_MAX_BURST 32
-- 
2.25.1



Re: [PATCH v4 03/10] test-pmd: fix printf format string mismatch

2025-02-21 Thread Bruce Richardson
On Thu, Feb 20, 2025 at 01:30:57PM -0800, Andre Muezerie wrote:
> Compiling with MSVC results in warnings like the one below:
> 
> app/test-pmd/csumonly.c(1085): warning C4477: 'printf' : format string
> '%d' requires an argument of type 'int',
> but variadic argument 1 has type 'uint64_t'
> 
> Signed-off-by: Andre Muezerie 
> Signed-off-by: Chengwen Feng 
> ---
>  app/test-pmd/csumonly.c | 23 ---
>  1 file changed, 12 insertions(+), 11 deletions(-)
> 
Acked-by: Bruce Richardson 


[PATCH] net/ice: fix ACL filter uninit

2025-02-21 Thread Mingjin Ye
The pf has enabled the ACL filter, so uninit is no longer limited
to the DCF.

Fixes: a9d612291c2d ("net/ice: support IPv4 fragments in ACL filters")

Signed-off-by: Mingjin Ye 
---
 drivers/net/intel/ice/ice_acl_filter.c | 48 +-
 1 file changed, 24 insertions(+), 24 deletions(-)

diff --git a/drivers/net/intel/ice/ice_acl_filter.c 
b/drivers/net/intel/ice/ice_acl_filter.c
index 04c17a98ed..83cb3e36f9 100644
--- a/drivers/net/intel/ice/ice_acl_filter.c
+++ b/drivers/net/intel/ice/ice_acl_filter.c
@@ -278,26 +278,28 @@ ice_acl_prof_init(struct ice_pf *pf)
if (ret)
goto err_add_prof_ipv4_sctp;
 
-   for (i = 0; i < pf->main_vsi->idx; i++) {
-   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, prof_ipv4, i);
-   if (ret)
-   goto err_assoc_prof;
-
-   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, prof_ipv4_frag, i);
-   if (ret)
-   goto err_assoc_prof;
-
-   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, prof_ipv4_udp, i);
-   if (ret)
-   goto err_assoc_prof;
-
-   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, prof_ipv4_tcp, i);
-   if (ret)
-   goto err_assoc_prof;
-
-   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, prof_ipv4_sctp, i);
-   if (ret)
-   goto err_assoc_prof;
+   if (hw->dcf_enabled) {
+   for (i = 0; i < pf->main_vsi->idx; i++) {
+   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, prof_ipv4, 
i);
+   if (ret)
+   goto err_assoc_prof;
+
+   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, 
prof_ipv4_frag, i);
+   if (ret)
+   goto err_assoc_prof;
+
+   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, 
prof_ipv4_udp, i);
+   if (ret)
+   goto err_assoc_prof;
+
+   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, 
prof_ipv4_tcp, i);
+   if (ret)
+   goto err_assoc_prof;
+
+   ret = ice_flow_assoc_prof(hw, ICE_BLK_ACL, 
prof_ipv4_sctp, i);
+   if (ret)
+   goto err_assoc_prof;
+   }
}
return 0;
 
@@ -1082,10 +1084,8 @@ ice_acl_uninit(struct ice_adapter *ad)
struct ice_pf *pf = &ad->pf;
struct ice_hw *hw = ICE_PF_TO_HW(pf);
 
-   if (ad->hw.dcf_enabled) {
-   ice_deinit_acl(pf);
-   ice_acl_prof_free(hw);
-   }
+   ice_deinit_acl(pf);
+   ice_acl_prof_free(hw);
 }
 
 static struct
-- 
2.25.1



RE: [PATCH v3 11/14] rawdev: remove unnecessary deref of function pointers

2025-02-21 Thread Hemant Agrawal
Acked-by: Hemant Agrawal 


Re: [PATCH] net/ixgbe: fix min Rx/Tx descriptors

2025-02-21 Thread Bruce Richardson
On Fri, Feb 21, 2025 at 08:23:18AM +, Mingjin Ye wrote:
> The minimum free packet threshold (tx_free_thresh) and the minimum RS bit
> threshold (tx_rs_thresh) both have a default value of 32. Therefore, the
> default minimum number of ring descriptors value is 64.
> 
> Fixes: dee5f1fd5fc7 ("ixgbe: get queue info and descriptor limits")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Mingjin Ye 
> ---
>  doc/guides/nics/ixgbe.rst| 2 +-
>  drivers/net/intel/ixgbe/ixgbe_rxtx.h | 2 +-
>  2 files changed, 2 insertions(+), 2 deletions(-)
> 
> diff --git a/doc/guides/nics/ixgbe.rst b/doc/guides/nics/ixgbe.rst
> index c5c6a6c34b..10a0cdd270 100644
> --- a/doc/guides/nics/ixgbe.rst
> +++ b/doc/guides/nics/ixgbe.rst
> @@ -76,7 +76,7 @@ Scattered packets are not supported in this mode.
>  If an incoming packet is greater than the maximum acceptable length of one 
> "mbuf" data size (by default, the size is 2 KB),
>  vPMD for RX would be disabled.
>  
> -By default, IXGBE_MAX_RING_DESC is set to 8192 and 
> RTE_PMD_IXGBE_RX_MAX_BURST is set to 32.
> +By default, IXGBE_MAX_RING_DESC is set to 8192 and 
> RTE_PMD_IXGBE_RX_MAX_BURST is set to 64.
>  
>  Windows Prerequisites and Pre-conditions
>  
> diff --git a/drivers/net/intel/ixgbe/ixgbe_rxtx.h 
> b/drivers/net/intel/ixgbe/ixgbe_rxtx.h
> index 278f665108..54569c7ade 100644
> --- a/drivers/net/intel/ixgbe/ixgbe_rxtx.h
> +++ b/drivers/net/intel/ixgbe/ixgbe_rxtx.h
> @@ -26,7 +26,7 @@
>   * descriptors should meet the following condition:
>   *  (num_ring_desc * sizeof(rx/tx descriptor)) % 128 == 0
>   */
> -#define  IXGBE_MIN_RING_DESC 32
> +#define  IXGBE_MIN_RING_DESC 64
>  #define  IXGBE_MAX_RING_DESC 8192

Ack for this part of the change. Just ran a quick test using ixgbe and
things stopped working pretty quickly with an Rx ring size of 32.

With the doc update above dropped:

Acked-by: Bruce Richardson 


Re: [PATCH] net/ixgbe: fix min Rx/Tx descriptors

2025-02-21 Thread Bruce Richardson
On Fri, Feb 21, 2025 at 08:23:18AM +, Mingjin Ye wrote:
> The minimum free packet threshold (tx_free_thresh) and the minimum RS bit
> threshold (tx_rs_thresh) both have a default value of 32. Therefore, the
> default minimum number of ring descriptors value is 64.
> 
> Fixes: dee5f1fd5fc7 ("ixgbe: get queue info and descriptor limits")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Mingjin Ye 
> ---
>  doc/guides/nics/ixgbe.rst| 2 +-
>  drivers/net/intel/ixgbe/ixgbe_rxtx.h | 2 +-
>  2 files changed, 2 insertions(+), 2 deletions(-)
> 
> diff --git a/doc/guides/nics/ixgbe.rst b/doc/guides/nics/ixgbe.rst
> index c5c6a6c34b..10a0cdd270 100644
> --- a/doc/guides/nics/ixgbe.rst
> +++ b/doc/guides/nics/ixgbe.rst
> @@ -76,7 +76,7 @@ Scattered packets are not supported in this mode.
>  If an incoming packet is greater than the maximum acceptable length of one 
> "mbuf" data size (by default, the size is 2 KB),
>  vPMD for RX would be disabled.
>  
> -By default, IXGBE_MAX_RING_DESC is set to 8192 and 
> RTE_PMD_IXGBE_RX_MAX_BURST is set to 32.
> +By default, IXGBE_MAX_RING_DESC is set to 8192 and 
> RTE_PMD_IXGBE_RX_MAX_BURST is set to 64.
>  

Was this part of the patch included by mistake? The RX_MAX_BURST value is
still set to 32 in the code.

However, in terms of doc updates, I believe the text above this point is
incorrect. It states that scattered packets are not supported in vector
drivers, which is no longer correct, as we support them for vector Rx.
Similarly, I think the pre-conditions for use of burst/vector mode are
incorrect compared to what is checked in the code.

>  Windows Prerequisites and Pre-conditions
>  
> diff --git a/drivers/net/intel/ixgbe/ixgbe_rxtx.h 
> b/drivers/net/intel/ixgbe/ixgbe_rxtx.h
> index 278f665108..54569c7ade 100644
> --- a/drivers/net/intel/ixgbe/ixgbe_rxtx.h
> +++ b/drivers/net/intel/ixgbe/ixgbe_rxtx.h
> @@ -26,7 +26,7 @@
>   * descriptors should meet the following condition:
>   *  (num_ring_desc * sizeof(rx/tx descriptor)) % 128 == 0
>   */
> -#define  IXGBE_MIN_RING_DESC 32
> +#define  IXGBE_MIN_RING_DESC 64
>  #define  IXGBE_MAX_RING_DESC 8192
>  
>  #define RTE_PMD_IXGBE_TX_MAX_BURST 32
> -- 
> 2.25.1
> 


[RFC PATCH v18] mempool: fix mempool cache size

2025-02-21 Thread Morten Brørup
NOTE: THIS VERSION DOES NOT BREAK THE API/ABI.

First, a per-lcore mempool cache could hold 50 % more than the cache's
size.
Since application developers do not expect this behavior, it could lead to
application failure.
This patch fixes this bug without breaking the API/ABI, by using the
mempool cache's "size" instead of the "flushthresh" as the threshold for
how many objects can be held in a mempool cache.
Note: The "flushthresh" field can be removed from the cache structure in a
future API/ABI breaking release, which must be announced in advance.

Second, requests to fetch a number of objects from the backend driver
exceeding the cache's size (but less than RTE_MEMPOOL_CACHE_MAX_SIZE) were
copied twice; first to the cache, and from there to the destination.
Such superfluous copying through the mempool cache degrades the
performance in these cases.
This patch also fixes this misbehavior, so when fetching more objects from
the driver than the mempool cache's size, they are fetched directly to the
destination.

The internal macro to calculate the cache flush threshold was updated to
reflect the new flush threshold of 1 * size instead of 1.5 * size.

The function rte_mempool_do_generic_put() for adding objects to a mempool
was modified as follows:
- When determining if the cache has sufficient room for the request
  without flushing, compare to the cache's size (cache->size) instead of
  the obsolete flush threshold (cache->flushthresh).
- The comparison for the request being too big, which is considered
  unlikely, was moved down and out of the code path where the cache has
  sufficient room for the added objects, which is considered the most
  likely code path.

The function rte_mempool_do_generic_get() for getting objects from a
mempool was refactored as follows:
- Handling a request for a constant number of objects was merged with
  handling a request for a nonconstant number of objects, and a note about
  compiler loop unrolling in the constant case was added.
- When determining if the remaining part of a request to be dequeued from
  the backend is too big to be copied via the cache, compare to the
  cache's size (cache->size) instead of the max possible cache size
  (RTE_MEMPOOL_CACHE_MAX_SIZE).
- When refilling the cache, the target fill level was reduced from the
  full cache size to half the cache size. This allows some room for a
  put() request following a get() request where the cache was refilled,
  without "flapping" between draining and refilling the entire cache.
  Note: Before this patch, the distance between the flush threshold and
  the refill level was also half a cache size.
- A copy of cache->len in the local variable "len" is no longer needed,
  so it was removed.

Furthermore, some likely()/unlikely()'s were added to a few inline
functions; most prominently rte_mempool_default_cache(), which is used by
both rte_mempool_put_bulk() and rte_mempool_get_bulk().

And finally, some comments were updated.

Signed-off-by: Morten Brørup 
---
v18:
* Start over from scratch, to avoid API/ABI breakage.
v17:
* Update rearm in idpf driver.
v16:
* Fix bug in rte_mempool_do_generic_put() regarding criteria for flush.
v15:
* Changed back cache bypass limit from n >= RTE_MEMPOOL_CACHE_MAX_SIZE to
  n > RTE_MEMPOOL_CACHE_MAX_SIZE.
* Removed cache size limit from serving via cache.
v14:
* Change rte_mempool_do_generic_put() back from add-then-flush to
  flush-then-add.
  Keep the target cache fill level of ca. 1/2 size of the cache.
v13:
* Target a cache fill level of ca. 1/2 size of the cache when flushing and
  refilling; based on an assumption of equal probability of get and put,
  instead of assuming a higher probability of put being followed by
  another put, and get being followed by another get.
* Reduce the amount of changes to the drivers.
v12:
* Do not init mempool caches with size zero; they don't exist.
  Bug introduced in v10.
v11:
* Removed rte_mempool_do_generic_get_split().
v10:
* Initialize mempool caches, regardless of size zero.
  This to fix compiler warning about out of bounds access.
v9:
* Removed factor 1.5 from description of cache_size parameter to
  rte_mempool_create().
* Refactored rte_mempool_do_generic_put() to eliminate some gotos.
  No functional change.
* Removed check for n >= RTE_MEMPOOL_CACHE_MAX_SIZE in
  rte_mempool_do_generic_get(); it caused the function to fail when the
  request could not be served from the backend alone, but it could be
  served from the cache and the backend.
* Refactored rte_mempool_do_generic_get_split() to make it shorter.
* When getting objects directly from the backend, use burst size aligned
  with either CPU cache line size or mempool cache size.
v8:
* Rewrote rte_mempool_do_generic_put() to get rid of transaction
  splitting. Use a method similar to the existing put method with fill
  followed by flush if overfilled.
  This also made rte_mempool_do_generic_put_split() obsolete.
* When flushing the cache as much as we can, use burst si

RE: [PATCH v4 01/10] eal: add workaround for __builtin_constant_p

2025-02-21 Thread Morten Brørup
Thank you for this updated MSVC variant of the macro, and its detailed 
description.

Acked-by: Morten Brørup 



[PATCH v5 08/10] test-pmd: declare lcore_count atomic

2025-02-21 Thread Andre Muezerie
Compiling with MSVC results in the error below:

app/test/test_ring_perf.c(197): error C7712: address argument to atomic
operation must be a pointer to an atomic integer,
'volatile unsigned int *' is not valid

The fix is to mark lcore_count as atomic.

Signed-off-by: Andre Muezerie 
Signed-off-by: Chengwen Feng 
---
 app/test/test_ring_perf.c | 6 +-
 1 file changed, 1 insertion(+), 5 deletions(-)

diff --git a/app/test/test_ring_perf.c b/app/test/test_ring_perf.c
index 57cd04a124..921aa902c5 100644
--- a/app/test/test_ring_perf.c
+++ b/app/test/test_ring_perf.c
@@ -34,7 +34,7 @@ struct lcore_pair {
unsigned c1, c2;
 };
 
-static volatile unsigned lcore_count = 0;
+static volatile RTE_ATOMIC(unsigned int) lcore_count;
 
 static void
 test_ring_print_test_string(unsigned int api_type, int esize,
@@ -193,11 +193,7 @@ enqueue_dequeue_bulk_helper(const unsigned int flag, 
struct thread_params *p)
unsigned int n_remaining;
const unsigned int bulk_n = bulk_sizes[p->ring_params->bulk_sizes_i];
 
-#ifdef RTE_USE_C11_MEM_MODEL
if (rte_atomic_fetch_add_explicit(&lcore_count, 1, 
rte_memory_order_relaxed) + 1 != 2)
-#else
-   if (__sync_add_and_fetch(&lcore_count, 1) != 2)
-#endif
while(lcore_count != 2)
rte_pause();
 
-- 
2.48.1.vfs.0.0



[PATCH v5 10/10] app: enable app directory to be compiled with MSVC

2025-02-21 Thread Andre Muezerie
Enabled "app" directory to be compiled with MSVC along with all its
contents.

Removed flag Wno-deprecated-declarations which is not needed anymore.

Signed-off-by: Andre Muezerie 
Acked-by: Bruce Richardson 
---
 app/meson.build  | 4 
 app/test-pmd/meson.build | 1 -
 2 files changed, 5 deletions(-)

diff --git a/app/meson.build b/app/meson.build
index e2db888ae1..1798db3ae4 100644
--- a/app/meson.build
+++ b/app/meson.build
@@ -1,10 +1,6 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright(c) 2017-2019 Intel Corporation
 
-if is_ms_compiler
-subdir_done()
-endif
-
 disable_apps = ',' + get_option('disable_apps')
 disable_apps = run_command(list_dir_globs, disable_apps, check: 
true).stdout().split()
 
diff --git a/app/test-pmd/meson.build b/app/test-pmd/meson.build
index f1c36529b4..000548c261 100644
--- a/app/test-pmd/meson.build
+++ b/app/test-pmd/meson.build
@@ -3,7 +3,6 @@
 
 # override default name to drop the hyphen
 name = 'testpmd'
-cflags += '-Wno-deprecated-declarations'
 sources = files(
 '5tswap.c',
 'cmdline.c',
-- 
2.48.1.vfs.0.0



[PATCH v5 09/10] test: add workaround for __builtin_constant_p in test_memcpy_perf

2025-02-21 Thread Andre Muezerie
There's no MSVC equivalent for compiler extension __builtin_constant_p,
so a workaround is needed.

Signed-off-by: Andre Muezerie 
---
 app/test/test_memcpy_perf.c | 54 ++---
 1 file changed, 27 insertions(+), 27 deletions(-)

diff --git a/app/test/test_memcpy_perf.c b/app/test/test_memcpy_perf.c
index 5c05a84619..32c589c4ba 100644
--- a/app/test/test_memcpy_perf.c
+++ b/app/test/test_memcpy_perf.c
@@ -196,37 +196,37 @@ do {  
  \
 } while (0)
 
 /* Run aligned memcpy tests for each cached/uncached permutation */
-#define ALL_PERF_TESTS_FOR_SIZE(n)   \
-do { \
-if (__builtin_constant_p(n)) \
-printf("\nC%6u", (unsigned)n);   \
-else \
-printf("\n%7u", (unsigned)n);\
-SINGLE_PERF_TEST(small_buf_write, 1, 0, small_buf_read, 1, 0, n);\
-SINGLE_PERF_TEST(large_buf_write, 0, 0, small_buf_read, 1, 0, n);\
-SINGLE_PERF_TEST(small_buf_write, 1, 0, large_buf_read, 0, 0, n);\
-SINGLE_PERF_TEST(large_buf_write, 0, 0, large_buf_read, 0, 0, n);\
+#define ALL_PERF_TESTS_FOR_SIZE(n)   \
+do { \
+   if (__rte_constant(n))   \
+   printf("\nC%6u", (unsigned int)n);   \
+   else \
+   printf("\n%7u", (unsigned int)n);\
+   SINGLE_PERF_TEST(small_buf_write, 1, 0, small_buf_read, 1, 0, n);\
+   SINGLE_PERF_TEST(large_buf_write, 0, 0, small_buf_read, 1, 0, n);\
+   SINGLE_PERF_TEST(small_buf_write, 1, 0, large_buf_read, 0, 0, n);\
+   SINGLE_PERF_TEST(large_buf_write, 0, 0, large_buf_read, 0, 0, n);\
 } while (0)
 
 /* Run unaligned memcpy tests for each cached/uncached permutation */
-#define ALL_PERF_TESTS_FOR_SIZE_UNALIGNED(n) \
-do { \
-if (__builtin_constant_p(n)) \
-printf("\nC%6u", (unsigned)n);   \
-else \
-printf("\n%7u", (unsigned)n);\
-SINGLE_PERF_TEST(small_buf_write, 1, 1, small_buf_read, 1, 5, n);\
-SINGLE_PERF_TEST(large_buf_write, 0, 1, small_buf_read, 1, 5, n);\
-SINGLE_PERF_TEST(small_buf_write, 1, 1, large_buf_read, 0, 5, n);\
-SINGLE_PERF_TEST(large_buf_write, 0, 1, large_buf_read, 0, 5, n);\
+#define ALL_PERF_TESTS_FOR_SIZE_UNALIGNED(n) \
+do { \
+   if (__rte_constant(n))   \
+   printf("\nC%6u", (unsigned int)n);   \
+   else \
+   printf("\n%7u", (unsigned int)n);\
+   SINGLE_PERF_TEST(small_buf_write, 1, 1, small_buf_read, 1, 5, n);\
+   SINGLE_PERF_TEST(large_buf_write, 0, 1, small_buf_read, 1, 5, n);\
+   SINGLE_PERF_TEST(small_buf_write, 1, 1, large_buf_read, 0, 5, n);\
+   SINGLE_PERF_TEST(large_buf_write, 0, 1, large_buf_read, 0, 5, n);\
 } while (0)
 
 /* Run memcpy tests for constant length */
-#define ALL_PERF_TEST_FOR_CONSTANT  \
-do {\
-TEST_CONSTANT(6U); TEST_CONSTANT(64U); TEST_CONSTANT(128U); \
-TEST_CONSTANT(192U); TEST_CONSTANT(256U); TEST_CONSTANT(512U);  \
-TEST_CONSTANT(768U); TEST_CONSTANT(1024U); TEST_CONSTANT(1536U);\
+#define ALL_PERF_TEST_FOR_CONSTANT  \
+do {\
+   TEST_CONSTANT(6U); TEST_CONSTANT(64U); TEST_CONSTANT(128U); \
+   TEST_CONSTANT(192U); TEST_CONSTANT(256U); TEST_CONSTANT(512U);  \
+   TEST_CONSTANT(768U); TEST_CONSTANT(1024U); TEST_CONSTANT(1536U);\
 } while (0)
 
 /* Run all memcpy tests for aligned constant cases */
@@ -253,7 +253,7 @@ perf_test_variable_aligned(void)
 {
unsigned i;
for (i = 0; i < RTE_DIM(buf_sizes); i++) {
-   ALL_PERF_TESTS_FOR_SIZE((size_t)buf_sizes[i]);
+   ALL_PERF_TESTS_FOR_SIZE(buf_sizes[i]);
}
 }
 
@@ -263,7 +263,7 @@ perf_test_vari

[PATCH v5 05/10] test-pmd: avoid undefined behavior

2025-02-21 Thread Andre Muezerie
Compiling with MSVC results in warnings like below:

app/test-pmd/cmdline.c(9023): warning C5101: use of preprocessor
directive in function-like macro argument list is undefined behavior

Signed-off-by: Andre Muezerie 
Signed-off-by: Chengwen Feng 
---
 app/test-pmd/cmdline.c | 12 
 app/test-pmd/testpmd.h |  2 ++
 2 files changed, 10 insertions(+), 4 deletions(-)

diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c
index 2afcf916c0..d0b47eac8f 100644
--- a/app/test-pmd/cmdline.c
+++ b/app/test-pmd/cmdline.c
@@ -8980,6 +8980,14 @@ dump_socket_mem(FILE *f)
last_total = total;
 }
 
+#ifdef RTE_EXEC_ENV_WINDOWS
+int
+rte_trace_save(void)
+{
+   return TEST_SKIPPED;
+}
+#endif
+
 static void cmd_dump_parsed(void *parsed_result,
__rte_unused struct cmdline *cl,
__rte_unused void *data)
@@ -9002,10 +9010,8 @@ static void cmd_dump_parsed(void *parsed_result,
rte_devargs_dump(stdout);
else if (!strcmp(res->dump, "dump_lcores"))
rte_lcore_dump(stdout);
-#ifndef RTE_EXEC_ENV_WINDOWS
else if (!strcmp(res->dump, "dump_trace"))
rte_trace_save();
-#endif
else if (!strcmp(res->dump, "dump_log_types"))
rte_log_dump(stdout);
 }
@@ -9020,9 +9026,7 @@ static cmdline_parse_token_string_t cmd_dump_dump =
"dump_mempool#"
"dump_devargs#"
"dump_lcores#"
-#ifndef RTE_EXEC_ENV_WINDOWS
"dump_trace#"
-#endif
"dump_log_types");
 
 static cmdline_parse_inst_t cmd_dump = {
diff --git a/app/test-pmd/testpmd.h b/app/test-pmd/testpmd.h
index a933fb433a..de8e8f8dca 100644
--- a/app/test-pmd/testpmd.h
+++ b/app/test-pmd/testpmd.h
@@ -84,6 +84,8 @@ extern volatile uint8_t f_quit;
 /* Maximum number of pools supported per Rx queue */
 #define MAX_MEMPOOL 8
 
+#define TEST_SKIPPED 77
+
 typedef uint32_t lcoreid_t;
 typedef uint16_t portid_t;
 typedef uint16_t queueid_t;
-- 
2.48.1.vfs.0.0



[PATCH v5 01/10] eal: add workaround for __builtin_constant_p

2025-02-21 Thread Andre Muezerie
There's no MSVC equivalent for compiler extension __builtin_constant_p,
but the same result can be obtained through a clever expression using
_Generic.

This patch redefines the macro __rte_constant when msvc is used and uses
it as a replacement for __builtin_constant_p.

Signed-off-by: Andre Muezerie 
Signed-off-by: Chengwen Feng 
---
 lib/eal/include/generic/rte_pause.h |  2 +-
 lib/eal/include/rte_common.h| 34 -
 2 files changed, 34 insertions(+), 2 deletions(-)

diff --git a/lib/eal/include/generic/rte_pause.h 
b/lib/eal/include/generic/rte_pause.h
index 968c0886d3..9515caadbb 100644
--- a/lib/eal/include/generic/rte_pause.h
+++ b/lib/eal/include/generic/rte_pause.h
@@ -130,7 +130,7 @@ rte_wait_until_equal_64(volatile uint64_t *addr, uint64_t 
expected,
  *  rte_memory_order_acquire and rte_memory_order_relaxed.
  */
 #define RTE_WAIT_UNTIL_MASKED(addr, mask, cond, expected, memorder) do { \
-   RTE_BUILD_BUG_ON(!__builtin_constant_p(memorder));   \
+   RTE_BUILD_BUG_ON(!__rte_constant(memorder)); \
RTE_BUILD_BUG_ON((memorder) != rte_memory_order_acquire &&   \
(memorder) != rte_memory_order_relaxed); \
typeof(*(addr)) expected_value = (expected); \
diff --git a/lib/eal/include/rte_common.h b/lib/eal/include/rte_common.h
index 4be65376a5..386f11ae40 100644
--- a/lib/eal/include/rte_common.h
+++ b/lib/eal/include/rte_common.h
@@ -44,8 +44,40 @@ extern "C" {
 #endif
 #endif
 
+/*
+ * Macro __rte_constant checks if an expression's value can be determined at
+ * compile time. It takes a single argument, the expression to test, and
+ * returns 1 if the expression is a compile-time constant, and 0 otherwise.
+ * For most compilers it uses built-in function __builtin_constant_p, but for
+ * MSVC it uses a different method because MSVC does not have an equivalent
+ * to __builtin_constant_p.
+ *
+ * The trick used with MSVC relies on the way null pointer constants interact
+ * with the type of a ?: expression:
+ * An integer constant expression with the value 0, or such an expression cast
+ * to type void *, is called a null pointer constant.
+ * If both the second and third operands (of the ?: expression) are pointers or
+ * one is a null pointer constant and the other is a pointer, the result type
+ * is a pointer to a type qualified with all the type qualifiers of the types
+ * referenced by both operands. Furthermore, if both operands are pointers to
+ * compatible types or to differently qualified versions of compatible types,
+ * the result type is a pointer to an appropriately qualified version of the
+ * composite type; if one operand is a null pointer constant, the result has
+ * the type of the other operand; otherwise, one operand is a pointer to void
+ * or a qualified version of void, in which case the result type is a pointer
+ * to an appropriately qualified version of void.
+ *
+ * The _Generic keyword then checks the type of the expression
+ * (void *) ((e) * 0ll). It matches this type against the types listed in the
+ * _Generic construct:
+ * - If the type is int *, the result is 1.
+ * - If the type is void *, the result is 0.
+ *
+ * This explanation with some more details can be found at:
+ * 
https://stackoverflow.com/questions/49480442/detecting-integer-constant-expressions-in-macros
+ */
 #ifdef RTE_TOOLCHAIN_MSVC
-#define __rte_constant(e) 0
+#define __rte_constant(e) _Generic((1 ? (void *) ((e) * 0ll) : (int *) 0), int 
* : 1, void * : 0)
 #else
 #define __rte_constant(e) __extension__(__builtin_constant_p(e))
 #endif
-- 
2.48.1.vfs.0.0



[PATCH v5 03/10] test-pmd: fix printf format string mismatch

2025-02-21 Thread Andre Muezerie
Compiling with MSVC results in warnings like the one below:

app/test-pmd/csumonly.c(1085): warning C4477: 'printf' : format string
'%d' requires an argument of type 'int',
but variadic argument 1 has type 'uint64_t'

Signed-off-by: Andre Muezerie 
Signed-off-by: Chengwen Feng 
---
 app/test-pmd/csumonly.c | 23 ---
 1 file changed, 12 insertions(+), 11 deletions(-)

diff --git a/app/test-pmd/csumonly.c b/app/test-pmd/csumonly.c
index d77a140641..5f273e60d7 100644
--- a/app/test-pmd/csumonly.c
+++ b/app/test-pmd/csumonly.c
@@ -1070,7 +1070,7 @@ pkt_burst_checksum_forward(struct fwd_stream *fs)
info.l2_len, rte_be_to_cpu_16(info.ethertype),
info.l3_len, info.l4_proto, info.l4_len, buf);
if (rx_ol_flags & RTE_MBUF_F_RX_LRO)
-   printf("rx: m->lro_segsz=%u\n", m->tso_segsz);
+   printf("rx: m->lro_segsz=%u\n", (unsigned 
int)m->tso_segsz);
if (info.is_tunnel == 1)
printf("rx: outer_l2_len=%d outer_ethertype=%x "
"outer_l3_len=%d\n", info.outer_l2_len,
@@ -1082,28 +1082,29 @@ pkt_burst_checksum_forward(struct fwd_stream *fs)
RTE_ETH_TX_OFFLOAD_TCP_CKSUM |
RTE_ETH_TX_OFFLOAD_SCTP_CKSUM)) ||
info.tso_segsz != 0)
-   printf("tx: m->l2_len=%d m->l3_len=%d "
-   "m->l4_len=%d\n",
-   m->l2_len, m->l3_len, m->l4_len);
+   printf("tx: m->l2_len=%u m->l3_len=%u "
+   "m->l4_len=%u\n",
+   (unsigned int)m->l2_len, (unsigned 
int)m->l3_len,
+   (unsigned int)m->l4_len);
if (info.is_tunnel == 1) {
if ((tx_offloads &
RTE_ETH_TX_OFFLOAD_OUTER_IPV4_CKSUM) ||
(tx_offloads &
RTE_ETH_TX_OFFLOAD_OUTER_UDP_CKSUM) ||
(tx_ol_flags & RTE_MBUF_F_TX_OUTER_IPV6))
-   printf("tx: m->outer_l2_len=%d "
-   "m->outer_l3_len=%d\n",
-   m->outer_l2_len,
-   m->outer_l3_len);
+   printf("tx: m->outer_l2_len=%u "
+   "m->outer_l3_len=%u\n",
+   (unsigned int)m->outer_l2_len,
+   (unsigned int)m->outer_l3_len);
if (info.tunnel_tso_segsz != 0 &&
(m->ol_flags & 
(RTE_MBUF_F_TX_TCP_SEG |
RTE_MBUF_F_TX_UDP_SEG)))
-   printf("tx: m->tso_segsz=%d\n",
-   m->tso_segsz);
+   printf("tx: m->tso_segsz=%u\n",
+   (unsigned int)m->tso_segsz);
} else if (info.tso_segsz != 0 &&
(m->ol_flags & (RTE_MBUF_F_TX_TCP_SEG |
RTE_MBUF_F_TX_UDP_SEG)))
-   printf("tx: m->tso_segsz=%d\n", m->tso_segsz);
+   printf("tx: m->tso_segsz=%u\n", (unsigned 
int)m->tso_segsz);
rte_get_tx_ol_flag_list(m->ol_flags, buf, sizeof(buf));
printf("tx: flags=%s", buf);
printf("\n");
-- 
2.48.1.vfs.0.0



[PATCH v5 04/10] test-pmd: do explicit 64-bit shift to avoid implicit conversion

2025-02-21 Thread Andre Muezerie
Compiling with MSVC results in warnings like the one below:

app/test-pmd/util.c(201): warning C4334: '<<': result of 32-bit shift
implicitly converted to 64 bits (was 64-bit shift intended?)

Signed-off-by: Andre Muezerie 
Signed-off-by: Chengwen Feng 
Acked-by: Bruce Richardson 
---
 app/test-pmd/cmdline.c | 4 ++--
 app/test-pmd/testpmd.c | 2 +-
 app/test-pmd/util.c| 2 +-
 3 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c
index 86d763b66a..2afcf916c0 100644
--- a/app/test-pmd/cmdline.c
+++ b/app/test-pmd/cmdline.c
@@ -12632,11 +12632,11 @@ cmd_config_dynf_specific_parsed(void *parsed_result,
}
old_port_flags = ports[res->port_id].mbuf_dynf;
if (!strcmp(res->value, "set")) {
-   ports[res->port_id].mbuf_dynf |= 1UL << flag;
+   ports[res->port_id].mbuf_dynf |= RTE_BIT64(flag);
if (old_port_flags == 0)
add_tx_dynf_callback(res->port_id);
} else {
-   ports[res->port_id].mbuf_dynf &= ~(1UL << flag);
+   ports[res->port_id].mbuf_dynf &= ~RTE_BIT64(flag);
if (ports[res->port_id].mbuf_dynf == 0)
remove_tx_dynf_callback(res->port_id);
}
diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c
index 0520aba0db..0a5b999c3a 100644
--- a/app/test-pmd/testpmd.c
+++ b/app/test-pmd/testpmd.c
@@ -4152,7 +4152,7 @@ get_eth_dcb_conf(struct rte_eth_conf *eth_conf, enum 
dcb_mode_enable dcb_mode,
for (i = 0; i < vmdq_rx_conf->nb_pool_maps; i++) {
vmdq_rx_conf->pool_map[i].vlan_id = vlan_tags[i];
vmdq_rx_conf->pool_map[i].pools =
-   1 << (i % vmdq_rx_conf->nb_queue_pools);
+   RTE_BIT64(i % vmdq_rx_conf->nb_queue_pools);
}
for (i = 0; i < RTE_ETH_DCB_NUM_USER_PRIORITIES; i++) {
vmdq_rx_conf->dcb_tc[i] = i % num_tcs;
diff --git a/app/test-pmd/util.c b/app/test-pmd/util.c
index 5fa05fad16..3934831b19 100644
--- a/app/test-pmd/util.c
+++ b/app/test-pmd/util.c
@@ -201,7 +201,7 @@ dump_pkt_burst(uint16_t port_id, uint16_t queue, struct 
rte_mbuf *pkts[],
MKDUMPSTR(print_buf, buf_size, cur_len,
  " - dynf %s: %d",
  dynf_names[dynf_index],
- !!(ol_flags & (1UL << dynf_index)));
+ !!(ol_flags & RTE_BIT64(dynf_index)));
}
if (mb->packet_type) {
rte_get_ptype_name(mb->packet_type, buf, sizeof(buf));
-- 
2.48.1.vfs.0.0



[PATCH v5 06/10] test-pmd: avoid non-constant initializer

2025-02-21 Thread Andre Muezerie
Compiling with MSVC results in errors like the one below:

app/test-pmd/cmdline_flow.c(8819): error C2099: initializer
is not a constant

Signed-off-by: Andre Muezerie 
Signed-off-by: Chengwen Feng 
---
 app/test-pmd/cmdline_flow.c | 10 +++---
 1 file changed, 3 insertions(+), 7 deletions(-)

diff --git a/app/test-pmd/cmdline_flow.c b/app/test-pmd/cmdline_flow.c
index e1720e54d7..24323d8891 100644
--- a/app/test-pmd/cmdline_flow.c
+++ b/app/test-pmd/cmdline_flow.c
@@ -8812,8 +8812,6 @@ parse_vc_spec(struct context *ctx, const struct token 
*token,
return -1;
/* Parse parameter types. */
switch (ctx->curr) {
-   static const enum index prefix[] = NEXT_ENTRY(COMMON_PREFIX);
-
case ITEM_PARAM_IS:
index = 0;
objmask = 1;
@@ -8828,7 +8826,7 @@ parse_vc_spec(struct context *ctx, const struct token 
*token,
/* Modify next token to expect a prefix. */
if (ctx->next_num < 2)
return -1;
-   ctx->next[ctx->next_num - 2] = prefix;
+   ctx->next[ctx->next_num - 2] = NEXT_ENTRY(COMMON_PREFIX);
/* Fall through. */
case ITEM_PARAM_MASK:
index = 2;
@@ -9270,7 +9268,6 @@ parse_vc_action_rss_type(struct context *ctx, const 
struct token *token,
  const char *str, unsigned int len,
  void *buf, unsigned int size)
 {
-   static const enum index next[] = NEXT_ENTRY(ACTION_RSS_TYPE);
struct action_rss_data *action_rss_data;
unsigned int i;
 
@@ -9296,7 +9293,7 @@ parse_vc_action_rss_type(struct context *ctx, const 
struct token *token,
/* Repeat token. */
if (ctx->next_num == RTE_DIM(ctx->next))
return -1;
-   ctx->next[ctx->next_num++] = next;
+   ctx->next[ctx->next_num++] = NEXT_ENTRY(ACTION_RSS_TYPE);
if (!ctx->object)
return len;
action_rss_data = ctx->object;
@@ -9314,7 +9311,6 @@ parse_vc_action_rss_queue(struct context *ctx, const 
struct token *token,
  const char *str, unsigned int len,
  void *buf, unsigned int size)
 {
-   static const enum index next[] = NEXT_ENTRY(ACTION_RSS_QUEUE);
struct action_rss_data *action_rss_data;
const struct arg *arg;
int ret;
@@ -9347,7 +9343,7 @@ parse_vc_action_rss_queue(struct context *ctx, const 
struct token *token,
/* Repeat token. */
if (ctx->next_num == RTE_DIM(ctx->next))
return -1;
-   ctx->next[ctx->next_num++] = next;
+   ctx->next[ctx->next_num++] = NEXT_ENTRY(ACTION_RSS_QUEUE);
 end:
if (!ctx->object)
return len;
-- 
2.48.1.vfs.0.0



[PATCH v5 07/10] test-pmd: don't return value from void function

2025-02-21 Thread Andre Muezerie
Compiling with MSVC results in the warning below:

app/test-pmd/cmdline_flow.c(13964): warning C4098: 'cmd_set_raw_parsed':
'void' function returning a value

Signed-off-by: Andre Muezerie 
Signed-off-by: Chengwen Feng 
Acked-by: Bruce Richardson 
---
 app/test-pmd/cmdline_flow.c | 12 
 1 file changed, 8 insertions(+), 4 deletions(-)

diff --git a/app/test-pmd/cmdline_flow.c b/app/test-pmd/cmdline_flow.c
index 24323d8891..15a18db2c7 100644
--- a/app/test-pmd/cmdline_flow.c
+++ b/app/test-pmd/cmdline_flow.c
@@ -13952,11 +13952,15 @@ cmd_set_raw_parsed(const struct buffer *in)
int gtp_psc = -1; /* GTP PSC option index. */
const void *src_spec;
 
-   if (in->command == SET_SAMPLE_ACTIONS)
-   return cmd_set_raw_parsed_sample(in);
+   if (in->command == SET_SAMPLE_ACTIONS) {
+   cmd_set_raw_parsed_sample(in);
+   return;
+   }
else if (in->command == SET_IPV6_EXT_PUSH ||
-in->command == SET_IPV6_EXT_REMOVE)
-   return cmd_set_ipv6_ext_parsed(in);
+in->command == SET_IPV6_EXT_REMOVE) {
+   cmd_set_ipv6_ext_parsed(in);
+   return;
+   }
RTE_ASSERT(in->command == SET_RAW_ENCAP ||
   in->command == SET_RAW_DECAP);
if (in->command == SET_RAW_ENCAP) {
-- 
2.48.1.vfs.0.0



Re: [PATCH v4 08/10] test-pmd: declare lcore_count atomic

2025-02-21 Thread Andre Muezerie
On Fri, Feb 21, 2025 at 08:40:17AM +, Konstantin Ananyev wrote:
> 
> 
> > Compiling with MSVC results in the error below:
> > 
> > app/test/test_ring_perf.c(197): error C7712: address argument to atomic
> > operation must be a pointer to an atomic integer,
> > 'volatile unsigned int *' is not valid
> > 
> > The fix is to mark lcore_count as atomic.
> > 
> > Signed-off-by: Andre Muezerie 
> > Signed-off-by: Chengwen Feng 
> > ---
> >  app/test/test_ring_perf.c | 6 +-
> >  1 file changed, 1 insertion(+), 5 deletions(-)
> > 
> > diff --git a/app/test/test_ring_perf.c b/app/test/test_ring_perf.c
> > index 57cd04a124..366e256323 100644
> > --- a/app/test/test_ring_perf.c
> > +++ b/app/test/test_ring_perf.c
> > @@ -34,7 +34,7 @@ struct lcore_pair {
> > unsigned c1, c2;
> >  };
> > 
> > -static volatile unsigned lcore_count = 0;
> > +static RTE_ATOMIC(unsigned int) lcore_count;
> 
> But for not-MSVC builds RTE_ATOMIC() by default means nothing.
> If you are removing volatile, then I think, you need to fix the code to ensure
> that all reads/writes to this var are done with atomic ops.
> Or have both -  volatile and RTE_ATOMIC()

Good point.
I'll use both volatile and RTE_ATOMIC() in the next series.

> 
> 
> >  static void
> >  test_ring_print_test_string(unsigned int api_type, int esize,
> > @@ -193,11 +193,7 @@ enqueue_dequeue_bulk_helper(const unsigned int flag, 
> > struct thread_params *p)
> > unsigned int n_remaining;
> > const unsigned int bulk_n = bulk_sizes[p->ring_params->bulk_sizes_i];
> > 
> > -#ifdef RTE_USE_C11_MEM_MODEL
> > if (rte_atomic_fetch_add_explicit(&lcore_count, 1, 
> > rte_memory_order_relaxed) + 1 != 2)
> > -#else
> > -   if (__sync_add_and_fetch(&lcore_count, 1) != 2)
> > -#endif
> > while(lcore_count != 2)
> > rte_pause();
> > 
> > --
> > 2.48.1.vfs.0.0
> > 


[PATCH v5 02/10] test_alarm: avoid warning about different qualifiers

2025-02-21 Thread Andre Muezerie
Compiling with MSVC results in the warning below:

app/test/test_alarm.c(54): warning C4090: 'function':
different '_Atomic' qualifiers

The fix is to use a macro to explicitly drop the qualifier.

Signed-off-by: Andre Muezerie 
Signed-off-by: Chengwen Feng 
---
 app/test/test_alarm.c | 12 ++--
 1 file changed, 6 insertions(+), 6 deletions(-)

diff --git a/app/test/test_alarm.c b/app/test/test_alarm.c
index 9ed8c6f72c..6445f713fe 100644
--- a/app/test/test_alarm.c
+++ b/app/test/test_alarm.c
@@ -51,12 +51,12 @@ test_alarm(void)
 "Expected rte_eal_alarm_cancel to fail with null 
callback parameter");
 
/* check if can set a alarm for one second */
-   TEST_ASSERT_SUCCESS(rte_eal_alarm_set(US_PER_SEC, test_alarm_callback, 
&triggered),
-   "Setting one second alarm failed");
+   TEST_ASSERT_SUCCESS(rte_eal_alarm_set(US_PER_SEC, test_alarm_callback,
+   RTE_PTR_UNQUAL(&triggered)), "Setting one second 
alarm failed");
 
/* set a longer alarm that will be canceled. */
-   TEST_ASSERT_SUCCESS(rte_eal_alarm_set(10 * US_PER_SEC, 
test_alarm_callback, &later),
-   "Setting ten second alarm failed");
+   TEST_ASSERT_SUCCESS(rte_eal_alarm_set(10 * US_PER_SEC, 
test_alarm_callback,
+   RTE_PTR_UNQUAL(&later)), "Setting ten second alarm 
failed");
 
/* wait for alarm to happen */
while (rte_atomic_load_explicit(&triggered, rte_memory_order_acquire) 
== false)
@@ -65,11 +65,11 @@ test_alarm(void)
TEST_ASSERT(!rte_atomic_load_explicit(&later, rte_memory_order_acquire),
"Only one alarm should have fired.");
 
-   ret = rte_eal_alarm_cancel(test_alarm_callback, &triggered);
+   ret = rte_eal_alarm_cancel(test_alarm_callback, 
RTE_PTR_UNQUAL(&triggered));
TEST_ASSERT(ret == 0 && rte_errno == ENOENT,
"Canceling alarm after run ret %d: %s", ret, 
rte_strerror(rte_errno));
 
-   ret = rte_eal_alarm_cancel(test_alarm_callback, &later);
+   ret = rte_eal_alarm_cancel(test_alarm_callback, RTE_PTR_UNQUAL(&later));
TEST_ASSERT(ret == 1, "Canceling ten second alarm failed %d: %s",
ret, rte_strerror(rte_errno));
 
-- 
2.48.1.vfs.0.0



[PATCH v5 00/10] enable "app" to be compiled with MSVC

2025-02-21 Thread Andre Muezerie
v5:
- add "volatile" to declaration of lcore_count in test_ring_perf.c.

v4:
- add explanation about the expression used in __rte_constant during v3.

v3:
- use clever expression with _Generic for __rte_constant and use it
  where __builtin_constant_p was being used.
- use %u in csumonly.c and just cast variables to unsigned int.

v2:
- use lcore_count as atomic always, not only when RTE_USE_C11_MEM_MODEL
  is set.
- use PRIu64 in csumonly.c instead of casting variables to int.

This series fixes many issues that prevent the "app" directory
from being compiled with MSVC.

Andre Muezerie (10):
  eal: add workaround for __builtin_constant_p
  test_alarm: avoid warning about different qualifiers
  test-pmd: fix printf format string mismatch
  test-pmd: do explicit 64-bit shift to avoid implicit conversion
  test-pmd: avoid undefined behavior
  test-pmd: avoid non-constant initializer
  test-pmd: don't return value from void function
  test-pmd: declare lcore_count atomic
  test: add workaround for __builtin_constant_p in test_memcpy_perf
  app: enable app directory to be compiled with MSVC

 app/meson.build |  4 ---
 app/test-pmd/cmdline.c  | 16 +
 app/test-pmd/cmdline_flow.c | 22 ++--
 app/test-pmd/csumonly.c | 23 ++--
 app/test-pmd/meson.build|  1 -
 app/test-pmd/testpmd.c  |  2 +-
 app/test-pmd/testpmd.h  |  2 ++
 app/test-pmd/util.c |  2 +-
 app/test/test_alarm.c   | 12 +++
 app/test/test_memcpy_perf.c | 54 ++---
 app/test/test_ring_perf.c   |  6 +---
 lib/eal/include/generic/rte_pause.h |  2 +-
 lib/eal/include/rte_common.h| 34 +-
 13 files changed, 105 insertions(+), 75 deletions(-)

--
2.48.1.vfs.0.0
Series-acked-by: Stephen Hemminger 


[v1] app/crypto-perf: add RSA support

2025-02-21 Thread Gowrishankar Muthukrishnan
From: Akhil Goyal 

Add RSA support in crypto-perf application.

Signed-off-by: Akhil Goyal 
---
 app/test-crypto-perf/cperf_ops.c |  68 
 app/test-crypto-perf/cperf_options.h |   4 +
 app/test-crypto-perf/cperf_options_parsing.c |  38 +++-
 app/test-crypto-perf/cperf_test_common.c |   1 +
 app/test-crypto-perf/cperf_test_vectors.c| 172 +++
 app/test-crypto-perf/cperf_test_vectors.h|  22 +++
 app/test-crypto-perf/main.c  |  19 ++
 7 files changed, 323 insertions(+), 1 deletion(-)

diff --git a/app/test-crypto-perf/cperf_ops.c b/app/test-crypto-perf/cperf_ops.c
index 6d5f510220..67df5c34ff 100644
--- a/app/test-crypto-perf/cperf_ops.c
+++ b/app/test-crypto-perf/cperf_ops.c
@@ -34,6 +34,40 @@ cperf_set_ops_asym_modex(struct rte_crypto_op **ops,
}
 }
 
+static void
+cperf_set_ops_asym_rsa(struct rte_crypto_op **ops,
+  uint32_t src_buf_offset __rte_unused,
+  uint32_t dst_buf_offset __rte_unused, uint16_t nb_ops,
+  void *sess,
+  const struct cperf_options *options,
+  const struct cperf_test_vector *test_vector __rte_unused,
+  uint16_t iv_offset __rte_unused,
+  uint32_t *imix_idx __rte_unused,
+  uint64_t *tsc_start __rte_unused)
+{
+   uint8_t cipher_buf[4096] = {0};
+   uint16_t i;
+
+   for (i = 0; i < nb_ops; i++) {
+   struct rte_crypto_asym_op *asym_op = ops[i]->asym;
+
+   ops[i]->status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED;
+   asym_op->rsa.op_type = options->asym_op_type;
+   asym_op->rsa.message.data = rsa_plaintext.data;
+   asym_op->rsa.message.length = rsa_plaintext.len;
+   if (options->asym_op_type == RTE_CRYPTO_ASYM_OP_SIGN) {
+   asym_op->rsa.sign.data = cipher_buf;
+   asym_op->rsa.sign.length = options->rsa_data->n.length;
+   } else if (options->asym_op_type == RTE_CRYPTO_ASYM_OP_ENCRYPT) 
{
+   asym_op->rsa.cipher.data = cipher_buf;
+   asym_op->rsa.cipher.length = 
options->rsa_data->n.length;
+   } else {
+   printf("RSA DECRYPT/VERIFY not supported");
+   }
+   rte_crypto_op_attach_asym_session(ops[i], sess);
+   }
+}
+
 static void
 cperf_set_ops_asym_ecdsa(struct rte_crypto_op **ops,
   uint32_t src_buf_offset __rte_unused,
@@ -1040,6 +1074,37 @@ cperf_create_session(struct rte_mempool *sess_mp,
return asym_sess;
}
 
+   if (options->op_type == CPERF_ASYM_RSA) {
+   xform.next = NULL;
+   xform.xform_type = RTE_CRYPTO_ASYM_XFORM_RSA;
+   xform.rsa.n.data = options->rsa_data->n.data;
+   xform.rsa.n.length = options->rsa_data->n.length;
+   xform.rsa.e.data = options->rsa_data->e.data;
+   xform.rsa.e.length = options->rsa_data->e.length;
+   xform.rsa.d.data = options->rsa_data->d.data;
+   xform.rsa.d.length = options->rsa_data->d.length;
+   xform.rsa.key_type = options->rsa_data->key_type;
+   if (xform.rsa.key_type == RTE_RSA_KEY_TYPE_QT) {
+   xform.rsa.qt.p.data = options->rsa_data->p.data;
+   xform.rsa.qt.p.length = options->rsa_data->p.length;
+   xform.rsa.qt.q.data = options->rsa_data->q.data;
+   xform.rsa.qt.q.length = options->rsa_data->q.length;
+   xform.rsa.qt.dP.data = options->rsa_data->dp.data;
+   xform.rsa.qt.dP.length = options->rsa_data->dp.length;
+   xform.rsa.qt.dQ.data = options->rsa_data->dq.data;
+   xform.rsa.qt.dQ.length = options->rsa_data->dq.length;
+   xform.rsa.qt.qInv.data = options->rsa_data->qinv.data;
+   xform.rsa.qt.qInv.length = 
options->rsa_data->qinv.length;
+   }
+   ret = rte_cryptodev_asym_session_create(dev_id, &xform,
+   sess_mp, &asym_sess);
+   if (ret < 0) {
+   RTE_LOG(ERR, USER1, "Asym session create failed\n");
+   return NULL;
+   }
+   return asym_sess;
+   }
+
if (options->op_type == CPERF_ASYM_SECP256R1) {
xform.next = NULL;
xform.xform_type = RTE_CRYPTO_ASYM_XFORM_ECDSA;
@@ -1400,6 +1465,9 @@ cperf_get_op_functions(const struct cperf_options 
*options,
case CPERF_ASYM_MODEX:
op_fns->populate_ops = cperf_set_ops_asym_modex;
break;
+   case CPERF_ASYM_RSA:
+   op_fns->populate_ops = cperf_set_ops_asym_rsa;
+   break;
case CPERF_ASYM_SECP256R1:
op_fns->populate_ops = c

Re: [Patch v3] net/mana: use mana_local_data for tracking usage data for primary process

2025-02-21 Thread Stephen Hemminger
On Thu, 20 Feb 2025 15:32:02 -0800
lon...@linuxonhyperv.com wrote:

> From: Long Li 
> 
> The driver uses mana_shared_data for tracking usage count for primary
> process. This is not correct as the mana_shared_data is allocated
> by the primary and is meant to track usage of secondary process by the
> primary process. And it creates a race condition when the device is
> removed because the counter is no longer available if this shared
> memory is freed.
> 
> Move the usage count tracking to mana_local_data and fix the race
> condition in mana_pci_remove().
> 
> Fixes: 517ed6e2d590 ("net/mana: add basic driver with build environment")
> Signed-off-by: Long Li 

Ok, mind if I break the long lines when merging?


[RFC PATCH v20] mempool: fix mempool cache size

2025-02-21 Thread Morten Brørup
NOTE: THIS VERSION DOES NOT BREAK THE API/ABI.

First, a per-lcore mempool cache could hold 50 % more than the cache's
size.
Since application developers do not expect this behavior, it could lead to
application failure.
This patch fixes this bug without breaking the API/ABI, by using the
mempool cache's "size" instead of the "flushthresh" as the threshold for
how many objects can be held in a mempool cache.
Note: The "flushthresh" field can be removed from the cache structure in a
future API/ABI breaking release, which must be announced in advance.

Second, requests to fetch a number of objects from the backend driver
exceeding the cache's size (but less than RTE_MEMPOOL_CACHE_MAX_SIZE) were
copied twice; first to the cache, and from there to the destination.
Such superfluous copying through the mempool cache degrades the
performance in these cases.
This patch also fixes this misbehavior, so when fetching more objects from
the driver than the mempool cache's size, they are fetched directly to the
destination.

The internal macro to calculate the cache flush threshold was updated to
reflect the new flush threshold of 1 * size instead of 1.5 * size.

The function rte_mempool_do_generic_put() for adding objects to a mempool
was modified as follows:
- When determining if the cache has sufficient room for the request
  without flushing, compare to the cache's size (cache->size) instead of
  the obsolete flush threshold (cache->flushthresh).
- The comparison for the request being too big, which is considered
  unlikely, was moved down and out of the code path where the cache has
  sufficient room for the added objects, which is considered the most
  likely code path.
- Added __rte_assume() about the cache size, for compiler optimization
  when "n" is compile time constant.
- Added __rte_assume() about "ret", for compiler optimization of
  rte_mempool_generic_get() considering the return value of
  rte_mempool_do_generic_get().

The function rte_mempool_do_generic_get() for getting objects from a
mempool was refactored as follows:
- Handling a request for a constant number of objects was merged with
  handling a request for a nonconstant number of objects, and a note about
  compiler loop unrolling in the constant case was added.
- When determining if the remaining part of a request to be dequeued from
  the backend is too big to be copied via the cache, compare to the
  cache's size (cache->size) instead of the max possible cache size
  (RTE_MEMPOOL_CACHE_MAX_SIZE).
- When refilling the cache, the target fill level was reduced from the
  full cache size to half the cache size. This allows some room for a
  put() request following a get() request where the cache was refilled,
  without "flapping" between draining and refilling the entire cache.
  Note: Before this patch, the distance between the flush threshold and
  the refill level was also half a cache size.
- A copy of cache->len in the local variable "len" is no longer needed,
  so it was removed.
- Added a group of __rte_assume()'s, for compiler optimization when "n" is
  compile time constant.

Some comments were also updated.

Furthermore, some likely()/unlikely()'s were added to a few inline
functions; most prominently rte_mempool_default_cache(), which is used by
both rte_mempool_put_bulk() and rte_mempool_get_bulk().

And finally, RTE_ASSERT()'s were added to check the return values of the
mempool driver dequeue() and enqueue() operations.

Signed-off-by: Morten Brørup 
---
v20:
* Added more __rte_assume()'s to fix build error with GCC 11.4.1 and
  GCC 11.5.0 in call to mempool_get_bulk() with compile time constant "n"
  larger than RTE_MEMPOOL_CACHE_MAX_SIZE.
v19:
* Added __rte_assume()'s and RTE_ASSERT()'s.
v18:
* Start over from scratch, to avoid API/ABI breakage.
v17:
* Update rearm in idpf driver.
v16:
* Fix bug in rte_mempool_do_generic_put() regarding criteria for flush.
v15:
* Changed back cache bypass limit from n >= RTE_MEMPOOL_CACHE_MAX_SIZE to
  n > RTE_MEMPOOL_CACHE_MAX_SIZE.
* Removed cache size limit from serving via cache.
v14:
* Change rte_mempool_do_generic_put() back from add-then-flush to
  flush-then-add.
  Keep the target cache fill level of ca. 1/2 size of the cache.
v13:
* Target a cache fill level of ca. 1/2 size of the cache when flushing and
  refilling; based on an assumption of equal probability of get and put,
  instead of assuming a higher probability of put being followed by
  another put, and get being followed by another get.
* Reduce the amount of changes to the drivers.
v12:
* Do not init mempool caches with size zero; they don't exist.
  Bug introduced in v10.
v11:
* Removed rte_mempool_do_generic_get_split().
v10:
* Initialize mempool caches, regardless of size zero.
  This to fix compiler warning about out of bounds access.
v9:
* Removed factor 1.5 from description of cache_size parameter to
  rte_mempool_create().
* Refactored rte_mempool_do_generic_put() to eliminate some gotos.
  No functional change.
* Remove

RE: [EXTERNAL] Re: [Patch v3] net/mana: use mana_local_data for tracking usage data for primary process

2025-02-21 Thread Long Li
> Subject: [EXTERNAL] Re: [Patch v3] net/mana: use mana_local_data for tracking
> usage data for primary process
> 
> On Thu, 20 Feb 2025 15:32:02 -0800
> lon...@linuxonhyperv.com wrote:
> 
> > From: Long Li 
> >
> > The driver uses mana_shared_data for tracking usage count for primary
> > process. This is not correct as the mana_shared_data is allocated by
> > the primary and is meant to track usage of secondary process by the
> > primary process. And it creates a race condition when the device is
> > removed because the counter is no longer available if this shared
> > memory is freed.
> >
> > Move the usage count tracking to mana_local_data and fix the race
> > condition in mana_pci_remove().
> >
> > Fixes: 517ed6e2d590 ("net/mana: add basic driver with build
> > environment")
> > Signed-off-by: Long Li 
> 
> Ok, mind if I break the long lines when merging?

Yes, please do.

Sorry I didn't catch this earlier. Will do next time.

Long


Re: [PATCH v1] doc/guides: update monitor PMD mode description

2025-02-21 Thread Bruce Richardson
On Fri, Feb 21, 2025 at 04:54:40PM +, Chris MacNamara wrote:
> A recent CPU change requires an extra enabling step for
> the umonitor instruction on Intel CPUs.
> This is now detailed in the l3 fwd power manager doc.
> 
> Signed-off-by: Chris MacNamara 
> ---
>  .mailmap  | 1 +
>  doc/guides/sample_app_ug/l3_forward_power_man.rst | 5 +
>  2 files changed, 6 insertions(+)
> 
> diff --git a/.mailmap b/.mailmap
> index a03d3cfb59..c4bc38752f 100644
> --- a/.mailmap
> +++ b/.mailmap
> @@ -263,6 +263,7 @@ Christopher Reder 
>  Christoph Gysin 
>  Christos Ricudis 
>  Chris Wright 
> +Chris MacNamara 

Minor nit, this is kept in alphabetical order, so your name should go just
above Chris Wright's one.

>  Chuanshe Zhang 
>  Chuanyu Xue 
>  Chuhong Yao 
> diff --git a/doc/guides/sample_app_ug/l3_forward_power_man.rst 
> b/doc/guides/sample_app_ug/l3_forward_power_man.rst
> index 3271bc2154..d0af28e0ec 100644
> --- a/doc/guides/sample_app_ug/l3_forward_power_man.rst
> +++ b/doc/guides/sample_app_ug/l3_forward_power_man.rst
> @@ -293,6 +293,11 @@ and has three available power management schemes:
>  ``monitor``
>This will use ``rte_power_monitor()`` function to enter
>a power-optimized state (subject to platform support).
> +  On recent Gen 4 Xeon Scalable Processors the umonitor instruction
> +  is disabled by default.
> +  An additional step is required to enable the umonitor instruction.
> +  Writing 0 to bit 6 of register 0x123 will enable umonitor.
> +  `More details are available via Monitor and Umonitor Performance Guidance 
> `_
>  
>  ``pause``
>This will use ``rte_power_pause()`` or ``rte_pause()``
> -- 

I'd suggest this extra info be added as a note, rather than inline in the
text, since it's not applicable to all systems. I'd also suggest that an
equivalent note needs to be added to the programmer's guide doc, in [1].
The info is relevant for those using the library, as well as those just
using the sample app.

/Bruce

[1] 
https://doc.dpdk.org/guides/prog_guide/power_man.html#ethernet-pmd-power-management-api


[PATCH] test/crypto: add negative test for RSA verify op

2025-02-21 Thread Gowrishankar Muthukrishnan
Add negative test for RSA verify operation to check if incorrect
signature is validated.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 app/test/test_cryptodev_asym.c | 27 +--
 1 file changed, 25 insertions(+), 2 deletions(-)

diff --git a/app/test/test_cryptodev_asym.c b/app/test/test_cryptodev_asym.c
index 8977d9d3a5..9b5f3c545e 100644
--- a/app/test/test_cryptodev_asym.c
+++ b/app/test/test_cryptodev_asym.c
@@ -62,7 +62,7 @@ queue_ops_rsa_sign_verify(void *sess)
struct rte_crypto_op *op, *result_op;
struct rte_crypto_asym_op *asym_op;
uint8_t output_buf[TEST_DATA_SIZE];
-   int status = TEST_SUCCESS;
+   int status;
 
/* Set up crypto op data structure */
op = rte_crypto_op_alloc(op_mpool, RTE_CRYPTO_OP_TYPE_ASYMMETRIC);
@@ -129,12 +129,35 @@ queue_ops_rsa_sign_verify(void *sess)
goto error_exit;
}
 
-   status = TEST_SUCCESS;
if (result_op->status != RTE_CRYPTO_OP_STATUS_SUCCESS) {
RTE_LOG(ERR, USER1, "Failed to process sign-verify op\n");
status = TEST_FAILED;
+   goto error_exit;
+   }
+
+   /* Negative test */
+   result_op->asym->rsa.sign.data[0] ^= 0xff;
+   if (rte_cryptodev_enqueue_burst(dev_id, 0, &result_op, 1) != 1) {
+   RTE_LOG(ERR, USER1, "Error sending packet for verify\n");
+   status = TEST_FAILED;
+   goto error_exit;
}
 
+   while (rte_cryptodev_dequeue_burst(dev_id, 0, &result_op, 1) == 0)
+   rte_pause();
+
+   if (result_op == NULL) {
+   RTE_LOG(ERR, USER1, "Failed to process verify op\n");
+   status = TEST_FAILED;
+   goto error_exit;
+   }
+
+   if (result_op->status != RTE_CRYPTO_OP_STATUS_ERROR) {
+   RTE_LOG(ERR, USER1, "Failed to process sign-verify op\n");
+   status = TEST_FAILED;
+   }
+
+   status = TEST_SUCCESS;
 error_exit:
 
rte_crypto_op_free(op);
-- 
2.25.1



Re: Intel FAST_FREE offload question

2025-02-21 Thread Bruce Richardson
On Fri, Feb 21, 2025 at 05:58:21PM +0100, Morten Brørup wrote:
> Intel NIC folks,
> 
> Why do the Intel network drivers, when using
> RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE, fall back to normal freeing when the
> mempool cache is empty (cache->len == 0)? It doesn't make sense to me.
> 
> Example:
> https://git.dpdk.org/dpdk/tree/drivers/net/intel/common/tx.h#n146
> 
Good question. I suspect that it may be a bug and that we meant to check
for size == 0 rather than len == 0.


[v3 1/6] crypto/virtio: add asymmetric RSA support

2025-02-21 Thread Gowrishankar Muthukrishnan
Asymmetric RSA operations (SIGN, VERIFY, ENCRYPT and DECRYPT) are
supported in virtio PMD.

Signed-off-by: Gowrishankar Muthukrishnan 
---
Depends-on: series-34674 ("vhost: add RSA support")

v3:
 - fast path optimizations.
---
 .../virtio/virtio_crypto_capabilities.h   |  19 +
 drivers/crypto/virtio/virtio_cryptodev.c  | 347 +++---
 drivers/crypto/virtio/virtio_cryptodev.h  |   2 +
 drivers/crypto/virtio/virtio_rxtx.c   | 243 ++--
 lib/cryptodev/cryptodev_pmd.h |   6 +
 5 files changed, 539 insertions(+), 78 deletions(-)

diff --git a/drivers/crypto/virtio/virtio_crypto_capabilities.h 
b/drivers/crypto/virtio/virtio_crypto_capabilities.h
index 03c30deefd..1b26ff6720 100644
--- a/drivers/crypto/virtio/virtio_crypto_capabilities.h
+++ b/drivers/crypto/virtio/virtio_crypto_capabilities.h
@@ -48,4 +48,23 @@
}, }\
}
 
+#define VIRTIO_ASYM_CAPABILITIES   \
+   {   /* RSA */   \
+   .op = RTE_CRYPTO_OP_TYPE_ASYMMETRIC,\
+   {.asym = {  \
+   .xform_capa = { \
+   .xform_type = RTE_CRYPTO_ASYM_XFORM_RSA,\
+   .op_types = ((1 << RTE_CRYPTO_ASYM_OP_SIGN) |   \
+   (1 << RTE_CRYPTO_ASYM_OP_VERIFY) |  \
+   (1 << RTE_CRYPTO_ASYM_OP_ENCRYPT) | \
+   (1 << RTE_CRYPTO_ASYM_OP_DECRYPT)), \
+   {.modlen = {\
+   .min = 1,   \
+   .max = 1024,\
+   .increment = 1  \
+   }, }\
+   }   \
+}, }   \
+   }
+
 #endif /* _VIRTIO_CRYPTO_CAPABILITIES_H_ */
diff --git a/drivers/crypto/virtio/virtio_cryptodev.c 
b/drivers/crypto/virtio/virtio_cryptodev.c
index 793f50059f..6a264bc24a 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.c
+++ b/drivers/crypto/virtio/virtio_cryptodev.c
@@ -41,6 +41,11 @@ static void virtio_crypto_sym_clear_session(struct 
rte_cryptodev *dev,
 static int virtio_crypto_sym_configure_session(struct rte_cryptodev *dev,
struct rte_crypto_sym_xform *xform,
struct rte_cryptodev_sym_session *session);
+static void virtio_crypto_asym_clear_session(struct rte_cryptodev *dev,
+   struct rte_cryptodev_asym_session *sess);
+static int virtio_crypto_asym_configure_session(struct rte_cryptodev *dev,
+   struct rte_crypto_asym_xform *xform,
+   struct rte_cryptodev_asym_session *session);
 
 /*
  * The set of PCI devices this driver supports
@@ -53,6 +58,7 @@ static const struct rte_pci_id pci_id_virtio_crypto_map[] = {
 
 static const struct rte_cryptodev_capabilities virtio_capabilities[] = {
VIRTIO_SYM_CAPABILITIES,
+   VIRTIO_ASYM_CAPABILITIES,
RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST()
 };
 
@@ -103,22 +109,24 @@ virtio_crypto_send_command(struct virtqueue *vq,
}
 
/* calculate the length of cipher key */
-   if (cipher_key) {
+   if (session->ctrl.header.algo == VIRTIO_CRYPTO_SERVICE_CIPHER) {
switch (ctrl->u.sym_create_session.op_type) {
case VIRTIO_CRYPTO_SYM_OP_CIPHER:
-   len_cipher_key
-   = ctrl->u.sym_create_session.u.cipher
-   .para.keylen;
+   len_cipher_key = 
ctrl->u.sym_create_session.u.cipher.para.keylen;
break;
case VIRTIO_CRYPTO_SYM_OP_ALGORITHM_CHAINING:
-   len_cipher_key
-   = ctrl->u.sym_create_session.u.chain
-   .para.cipher_param.keylen;
+   len_cipher_key =
+   
ctrl->u.sym_create_session.u.chain.para.cipher_param.keylen;
break;
default:
VIRTIO_CRYPTO_SESSION_LOG_ERR("invalid op type");
return -EINVAL;
}
+   } else if (session->ctrl.header.algo == VIRTIO_CRYPTO_AKCIPHER_RSA) {
+   len_cipher_key = ctrl->u.akcipher_create_session.para.keylen;
+   } else {
+   VIRTIO_CRYPTO_SESSION_LOG_ERR("Invalid crypto service for 
cipher key");
+   return -EINVAL;
}
 
/* calculate the length of auth key *

Re: [PATCH v1] doc/guides: update monitor PMD mode description

2025-02-21 Thread Stephen Hemminger
On Fri, 21 Feb 2025 16:54:40 +
Chris MacNamara  wrote:

> A recent CPU change requires an extra enabling step for
> the umonitor instruction on Intel CPUs.
> This is now detailed in the l3 fwd power manager doc.
> 
> Signed-off-by: Chris MacNamara 
> ---
>  .mailmap  | 1 +
>  doc/guides/sample_app_ug/l3_forward_power_man.rst | 5 +
>  2 files changed, 6 insertions(+)
> 
> diff --git a/.mailmap b/.mailmap
> index a03d3cfb59..c4bc38752f 100644
> --- a/.mailmap
> +++ b/.mailmap
> @@ -263,6 +263,7 @@ Christopher Reder 
>  Christoph Gysin 
>  Christos Ricudis 
>  Chris Wright 
> +Chris MacNamara 
>  Chuanshe Zhang 
>  Chuanyu Xue 
>  Chuhong Yao 
> diff --git a/doc/guides/sample_app_ug/l3_forward_power_man.rst 
> b/doc/guides/sample_app_ug/l3_forward_power_man.rst
> index 3271bc2154..d0af28e0ec 100644
> --- a/doc/guides/sample_app_ug/l3_forward_power_man.rst
> +++ b/doc/guides/sample_app_ug/l3_forward_power_man.rst
> @@ -293,6 +293,11 @@ and has three available power management schemes:
>  ``monitor``
>This will use ``rte_power_monitor()`` function to enter
>a power-optimized state (subject to platform support).
> +  On recent Gen 4 Xeon Scalable Processors the umonitor instruction
> +  is disabled by default.
> +  An additional step is required to enable the umonitor instruction.
> +  Writing 0 to bit 6 of register 0x123 will enable umonitor.
> +  `More details are available via Monitor and Umonitor Performance Guidance 
> `_
>  

This should be done by the power library, what is the point of having a power 
API
if we require applications to test for CPU type and go tweak some CPU register 
bits.
And it may be restricted by OS permissions etc.



RE: [PATCH v5 08/10] test-pmd: declare lcore_count atomic

2025-02-21 Thread Morten Brørup
> From: Andre Muezerie [mailto:andre...@linux.microsoft.com]
> Sent: Friday, 21 February 2025 20.53
> 
> Compiling with MSVC results in the error below:
> 
> app/test/test_ring_perf.c(197): error C7712: address argument to atomic
> operation must be a pointer to an atomic integer,
> 'volatile unsigned int *' is not valid
> 
> The fix is to mark lcore_count as atomic.
> 
> Signed-off-by: Andre Muezerie 
> Signed-off-by: Chengwen Feng 
> ---
>  app/test/test_ring_perf.c | 6 +-
>  1 file changed, 1 insertion(+), 5 deletions(-)
> 
> diff --git a/app/test/test_ring_perf.c b/app/test/test_ring_perf.c
> index 57cd04a124..921aa902c5 100644
> --- a/app/test/test_ring_perf.c
> +++ b/app/test/test_ring_perf.c
> @@ -34,7 +34,7 @@ struct lcore_pair {
>   unsigned c1, c2;
>  };
> 
> -static volatile unsigned lcore_count = 0;
> +static volatile RTE_ATOMIC(unsigned int) lcore_count;

Using rte_atomic_xxx to access the RTE_ATOMIC qualified lcore_count seems more 
clean than making it volatile and sometimes access it non-atomically.
It is only accessed two more times, so not a big change.

Anyway, this could just be my personal preference. No change required if you 
disagree.



RE: Intel FAST_FREE offload question

2025-02-21 Thread Morten Brørup
> From: Bruce Richardson [mailto:bruce.richard...@intel.com]
> Sent: Friday, 21 February 2025 18.10
> 
> On Fri, Feb 21, 2025 at 05:58:21PM +0100, Morten Brørup wrote:
> > Intel NIC folks,
> >
> > Why do the Intel network drivers, when using
> > RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE, fall back to normal freeing when
> the
> > mempool cache is empty (cache->len == 0)? It doesn't make sense to
> me.
> >
> > Example:
> > https://git.dpdk.org/dpdk/tree/drivers/net/intel/common/tx.h#n146
> >
> Good question. I suspect that it may be a bug and that we meant to
> check
> for size == 0 rather than len == 0.

Then checking for cache == NULL suffices, because rte_mempool_default_cache() 
returns NULL if the cache size is 0:
https://elixir.bootlin.com/dpdk/v24.11.1/source/lib/mempool/rte_mempool.h#L1333



[v3 0/6] crypto/virtio: enhancements for RSA and vDPA

2025-02-21 Thread Gowrishankar Muthukrishnan
This patch series enhances virtio crypto PMD to:
 * support RSA
 * support packed virtio ring
 * support vDPA backend

v3:
 - vDPA backend code majorly sourced from virtio net.

Gowrishankar Muthukrishnan (6):
  crypto/virtio: add asymmetric RSA support
  crypto/virtio: refactor queue operations
  crypto/virtio: add packed ring support
  crypto/virtio: add vDPA backend
  test/crypto: add asymmetric tests for virtio PMD
  test/crypto: add tests for virtio user PMD

 app/test/test_cryptodev.c |7 +
 app/test/test_cryptodev.h |1 +
 app/test/test_cryptodev_asym.c|   43 +
 drivers/crypto/virtio/meson.build |8 +
 drivers/crypto/virtio/virtio_crypto_algs.h|2 +-
 .../virtio/virtio_crypto_capabilities.h   |   19 +
 drivers/crypto/virtio/virtio_cryptodev.c  | 1060 +++--
 drivers/crypto/virtio/virtio_cryptodev.h  |   18 +-
 drivers/crypto/virtio/virtio_cvq.c|  228 
 drivers/crypto/virtio/virtio_cvq.h|   33 +
 drivers/crypto/virtio/virtio_logs.h   |6 +-
 drivers/crypto/virtio/virtio_pci.h|   38 +-
 drivers/crypto/virtio/virtio_ring.h   |   65 +-
 drivers/crypto/virtio/virtio_rxtx.c   |  721 ++-
 drivers/crypto/virtio/virtio_rxtx.h   |   13 +
 drivers/crypto/virtio/virtio_user/vhost.h |   90 ++
 .../crypto/virtio/virtio_user/vhost_vdpa.c|  710 +++
 .../virtio/virtio_user/virtio_user_dev.c  |  767 
 .../virtio/virtio_user/virtio_user_dev.h  |   85 ++
 drivers/crypto/virtio/virtio_user_cryptodev.c |  575 +
 drivers/crypto/virtio/virtqueue.c |  229 +++-
 drivers/crypto/virtio/virtqueue.h |  221 +++-
 lib/cryptodev/cryptodev_pmd.h |6 +
 23 files changed, 4453 insertions(+), 492 deletions(-)
 create mode 100644 drivers/crypto/virtio/virtio_cvq.c
 create mode 100644 drivers/crypto/virtio/virtio_cvq.h
 create mode 100644 drivers/crypto/virtio/virtio_rxtx.h
 create mode 100644 drivers/crypto/virtio/virtio_user/vhost.h
 create mode 100644 drivers/crypto/virtio/virtio_user/vhost_vdpa.c
 create mode 100644 drivers/crypto/virtio/virtio_user/virtio_user_dev.c
 create mode 100644 drivers/crypto/virtio/virtio_user/virtio_user_dev.h
 create mode 100644 drivers/crypto/virtio/virtio_user_cryptodev.c

-- 
2.25.1



[v3 4/6] crypto/virtio: add vDPA backend

2025-02-21 Thread Gowrishankar Muthukrishnan
Add vDPA backend to virtio_user crypto.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 drivers/crypto/virtio/meson.build |   7 +
 drivers/crypto/virtio/virtio_cryptodev.c  |  57 +-
 drivers/crypto/virtio/virtio_cryptodev.h  |   3 +
 drivers/crypto/virtio/virtio_logs.h   |   6 +-
 drivers/crypto/virtio/virtio_pci.h|   7 +
 drivers/crypto/virtio/virtio_ring.h   |   6 -
 drivers/crypto/virtio/virtio_user/vhost.h |  90 ++
 .../crypto/virtio/virtio_user/vhost_vdpa.c| 710 
 .../virtio/virtio_user/virtio_user_dev.c  | 767 ++
 .../virtio/virtio_user/virtio_user_dev.h  |  85 ++
 drivers/crypto/virtio/virtio_user_cryptodev.c | 575 +
 11 files changed, 2283 insertions(+), 30 deletions(-)
 create mode 100644 drivers/crypto/virtio/virtio_user/vhost.h
 create mode 100644 drivers/crypto/virtio/virtio_user/vhost_vdpa.c
 create mode 100644 drivers/crypto/virtio/virtio_user/virtio_user_dev.c
 create mode 100644 drivers/crypto/virtio/virtio_user/virtio_user_dev.h
 create mode 100644 drivers/crypto/virtio/virtio_user_cryptodev.c

diff --git a/drivers/crypto/virtio/meson.build 
b/drivers/crypto/virtio/meson.build
index d2c3b3ad07..3763e86746 100644
--- a/drivers/crypto/virtio/meson.build
+++ b/drivers/crypto/virtio/meson.build
@@ -16,3 +16,10 @@ sources = files(
 'virtio_rxtx.c',
 'virtqueue.c',
 )
+
+if is_linux
+sources += files('virtio_user_cryptodev.c',
+'virtio_user/vhost_vdpa.c',
+'virtio_user/virtio_user_dev.c')
+deps += ['bus_vdev']
+endif
diff --git a/drivers/crypto/virtio/virtio_cryptodev.c 
b/drivers/crypto/virtio/virtio_cryptodev.c
index 92fea557ab..bc737f1e68 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.c
+++ b/drivers/crypto/virtio/virtio_cryptodev.c
@@ -544,24 +544,12 @@ virtio_crypto_init_device(struct rte_cryptodev *cryptodev,
return 0;
 }
 
-/*
- * This function is based on probe() function
- * It returns 0 on success.
- */
-static int
-crypto_virtio_create(const char *name, struct rte_pci_device *pci_dev,
-   struct rte_cryptodev_pmd_init_params *init_params)
+int
+crypto_virtio_dev_init(struct rte_cryptodev *cryptodev, uint64_t features,
+   struct rte_pci_device *pci_dev)
 {
-   struct rte_cryptodev *cryptodev;
struct virtio_crypto_hw *hw;
 
-   PMD_INIT_FUNC_TRACE();
-
-   cryptodev = rte_cryptodev_pmd_create(name, &pci_dev->device,
-   init_params);
-   if (cryptodev == NULL)
-   return -ENODEV;
-
cryptodev->driver_id = cryptodev_virtio_driver_id;
cryptodev->dev_ops = &virtio_crypto_dev_ops;
 
@@ -578,16 +566,41 @@ crypto_virtio_create(const char *name, struct 
rte_pci_device *pci_dev,
hw->dev_id = cryptodev->data->dev_id;
hw->virtio_dev_capabilities = virtio_capabilities;
 
-   VIRTIO_CRYPTO_INIT_LOG_DBG("dev %d vendorID=0x%x deviceID=0x%x",
-   cryptodev->data->dev_id, pci_dev->id.vendor_id,
-   pci_dev->id.device_id);
+   if (pci_dev) {
+   /* pci device init */
+   VIRTIO_CRYPTO_INIT_LOG_DBG("dev %d vendorID=0x%x deviceID=0x%x",
+   cryptodev->data->dev_id, pci_dev->id.vendor_id,
+   pci_dev->id.device_id);
 
-   /* pci device init */
-   if (vtpci_cryptodev_init(pci_dev, hw))
+   if (vtpci_cryptodev_init(pci_dev, hw))
+   return -1;
+   }
+
+   if (virtio_crypto_init_device(cryptodev, features) < 0)
return -1;
 
-   if (virtio_crypto_init_device(cryptodev,
-   VIRTIO_CRYPTO_PMD_GUEST_FEATURES) < 0)
+   return 0;
+}
+
+/*
+ * This function is based on probe() function
+ * It returns 0 on success.
+ */
+static int
+crypto_virtio_create(const char *name, struct rte_pci_device *pci_dev,
+   struct rte_cryptodev_pmd_init_params *init_params)
+{
+   struct rte_cryptodev *cryptodev;
+
+   PMD_INIT_FUNC_TRACE();
+
+   cryptodev = rte_cryptodev_pmd_create(name, &pci_dev->device,
+   init_params);
+   if (cryptodev == NULL)
+   return -ENODEV;
+
+   if (crypto_virtio_dev_init(cryptodev, VIRTIO_CRYPTO_PMD_GUEST_FEATURES,
+   pci_dev) < 0)
return -1;
 
rte_cryptodev_pmd_probing_finish(cryptodev);
diff --git a/drivers/crypto/virtio/virtio_cryptodev.h 
b/drivers/crypto/virtio/virtio_cryptodev.h
index f8498246e2..fad73d54a8 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.h
+++ b/drivers/crypto/virtio/virtio_cryptodev.h
@@ -76,4 +76,7 @@ uint16_t virtio_crypto_pkt_rx_burst(void *tx_queue,
struct rte_crypto_op **tx_pkts,
uint16_t nb_pkts);
 
+int crypto_virtio_dev_init(struct rte_cryptodev *cryptodev, uint64_t features,
+   struct rte_pci_device *pci_dev);
+
 #endif /* _VIRTI

[v3 2/6] crypto/virtio: refactor queue operations

2025-02-21 Thread Gowrishankar Muthukrishnan
Move existing control queue operations into a common place
that would be shared with other virtio type of devices.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 drivers/crypto/virtio/meson.build  |   1 +
 drivers/crypto/virtio/virtio_crypto_algs.h |   2 +-
 drivers/crypto/virtio/virtio_cryptodev.c   | 563 -
 drivers/crypto/virtio/virtio_cvq.c | 129 +
 drivers/crypto/virtio/virtio_cvq.h |  33 ++
 drivers/crypto/virtio/virtio_pci.h |   6 +-
 drivers/crypto/virtio/virtio_ring.h|  12 +-
 drivers/crypto/virtio/virtio_rxtx.c|  42 +-
 drivers/crypto/virtio/virtio_rxtx.h|  13 +
 drivers/crypto/virtio/virtqueue.c  | 191 ++-
 drivers/crypto/virtio/virtqueue.h  |  89 +++-
 11 files changed, 691 insertions(+), 390 deletions(-)
 create mode 100644 drivers/crypto/virtio/virtio_cvq.c
 create mode 100644 drivers/crypto/virtio/virtio_cvq.h
 create mode 100644 drivers/crypto/virtio/virtio_rxtx.h

diff --git a/drivers/crypto/virtio/meson.build 
b/drivers/crypto/virtio/meson.build
index 45533c9b89..d2c3b3ad07 100644
--- a/drivers/crypto/virtio/meson.build
+++ b/drivers/crypto/virtio/meson.build
@@ -11,6 +11,7 @@ includes += include_directories('../../../lib/vhost')
 deps += 'bus_pci'
 sources = files(
 'virtio_cryptodev.c',
+'virtio_cvq.c',
 'virtio_pci.c',
 'virtio_rxtx.c',
 'virtqueue.c',
diff --git a/drivers/crypto/virtio/virtio_crypto_algs.h 
b/drivers/crypto/virtio/virtio_crypto_algs.h
index 4c44af3733..3824017ca5 100644
--- a/drivers/crypto/virtio/virtio_crypto_algs.h
+++ b/drivers/crypto/virtio/virtio_crypto_algs.h
@@ -22,7 +22,7 @@ struct virtio_crypto_session {
phys_addr_t phys_addr;
} aad;
 
-   struct virtio_crypto_op_ctrl_req ctrl;
+   struct virtio_pmd_ctrl ctrl;
 };
 
 #endif /* _VIRTIO_CRYPTO_ALGS_H_ */
diff --git a/drivers/crypto/virtio/virtio_cryptodev.c 
b/drivers/crypto/virtio/virtio_cryptodev.c
index 6a264bc24a..6bb76ff15e 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.c
+++ b/drivers/crypto/virtio/virtio_cryptodev.c
@@ -64,211 +64,6 @@ static const struct rte_cryptodev_capabilities 
virtio_capabilities[] = {
 
 uint8_t cryptodev_virtio_driver_id;
 
-#define NUM_ENTRY_SYM_CREATE_SESSION 4
-
-static int
-virtio_crypto_send_command(struct virtqueue *vq,
-   struct virtio_crypto_op_ctrl_req *ctrl, uint8_t *cipher_key,
-   uint8_t *auth_key, struct virtio_crypto_session *session)
-{
-   uint8_t idx = 0;
-   uint8_t needed = 1;
-   uint32_t head = 0;
-   uint32_t len_cipher_key = 0;
-   uint32_t len_auth_key = 0;
-   uint32_t len_ctrl_req = sizeof(struct virtio_crypto_op_ctrl_req);
-   uint32_t len_session_input = sizeof(struct virtio_crypto_session_input);
-   uint32_t len_total = 0;
-   uint32_t input_offset = 0;
-   void *virt_addr_started = NULL;
-   phys_addr_t phys_addr_started;
-   struct vring_desc *desc;
-   uint32_t desc_offset;
-   struct virtio_crypto_session_input *input;
-   int ret;
-
-   PMD_INIT_FUNC_TRACE();
-
-   if (session == NULL) {
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("session is NULL.");
-   return -EINVAL;
-   }
-   /* cipher only is supported, it is available if auth_key is NULL */
-   if (!cipher_key) {
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("cipher key is NULL.");
-   return -EINVAL;
-   }
-
-   head = vq->vq_desc_head_idx;
-   VIRTIO_CRYPTO_INIT_LOG_DBG("vq->vq_desc_head_idx = %d, vq = %p",
-   head, vq);
-
-   if (vq->vq_free_cnt < needed) {
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("Not enough entry");
-   return -ENOSPC;
-   }
-
-   /* calculate the length of cipher key */
-   if (session->ctrl.header.algo == VIRTIO_CRYPTO_SERVICE_CIPHER) {
-   switch (ctrl->u.sym_create_session.op_type) {
-   case VIRTIO_CRYPTO_SYM_OP_CIPHER:
-   len_cipher_key = 
ctrl->u.sym_create_session.u.cipher.para.keylen;
-   break;
-   case VIRTIO_CRYPTO_SYM_OP_ALGORITHM_CHAINING:
-   len_cipher_key =
-   
ctrl->u.sym_create_session.u.chain.para.cipher_param.keylen;
-   break;
-   default:
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("invalid op type");
-   return -EINVAL;
-   }
-   } else if (session->ctrl.header.algo == VIRTIO_CRYPTO_AKCIPHER_RSA) {
-   len_cipher_key = ctrl->u.akcipher_create_session.para.keylen;
-   } else {
-   VIRTIO_CRYPTO_SESSION_LOG_ERR("Invalid crypto service for 
cipher key");
-   return -EINVAL;
-   }
-
-   /* calculate the length of auth key */
-   if (auth_key) {
-   len_auth_key =
-   ctrl->u.sym_

[v3 6/6] test/crypto: add tests for virtio user PMD

2025-02-21 Thread Gowrishankar Muthukrishnan
Reuse virtio_crypto tests for testing virtio_crypto_user PMD.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 app/test/test_cryptodev.c  |  7 +++
 app/test/test_cryptodev.h  |  1 +
 app/test/test_cryptodev_asym.c | 15 +++
 3 files changed, 23 insertions(+)

diff --git a/app/test/test_cryptodev.c b/app/test/test_cryptodev.c
index 441ecc6ad5..60aacdc155 100644
--- a/app/test/test_cryptodev.c
+++ b/app/test/test_cryptodev.c
@@ -19737,6 +19737,12 @@ test_cryptodev_virtio(void)
return run_cryptodev_testsuite(RTE_STR(CRYPTODEV_NAME_VIRTIO_PMD));
 }
 
+static int
+test_cryptodev_virtio_user(void)
+{
+   return run_cryptodev_testsuite(RTE_STR(CRYPTODEV_NAME_VIRTIO_USER_PMD));
+}
+
 static int
 test_cryptodev_aesni_mb(void)
 {
@@ -20074,6 +20080,7 @@ REGISTER_DRIVER_TEST(cryptodev_dpaa_sec_autotest, 
test_cryptodev_dpaa_sec);
 REGISTER_DRIVER_TEST(cryptodev_ccp_autotest, test_cryptodev_ccp);
 REGISTER_DRIVER_TEST(cryptodev_uadk_autotest, test_cryptodev_uadk);
 REGISTER_DRIVER_TEST(cryptodev_virtio_autotest, test_cryptodev_virtio);
+REGISTER_DRIVER_TEST(cryptodev_virtio_user_autotest, 
test_cryptodev_virtio_user);
 REGISTER_DRIVER_TEST(cryptodev_octeontx_autotest, test_cryptodev_octeontx);
 REGISTER_DRIVER_TEST(cryptodev_caam_jr_autotest, test_cryptodev_caam_jr);
 REGISTER_DRIVER_TEST(cryptodev_nitrox_autotest, test_cryptodev_nitrox);
diff --git a/app/test/test_cryptodev.h b/app/test/test_cryptodev.h
index bb54a33d62..f6c7478f19 100644
--- a/app/test/test_cryptodev.h
+++ b/app/test/test_cryptodev.h
@@ -64,6 +64,7 @@
 #define CRYPTODEV_NAME_MVSAM_PMD   crypto_mvsam
 #define CRYPTODEV_NAME_CCP_PMD crypto_ccp
 #define CRYPTODEV_NAME_VIRTIO_PMD  crypto_virtio
+#define CRYPTODEV_NAME_VIRTIO_USER_PMD crypto_virtio_user
 #define CRYPTODEV_NAME_OCTEONTX_SYM_PMDcrypto_octeontx
 #define CRYPTODEV_NAME_CAAM_JR_PMD crypto_caam_jr
 #define CRYPTODEV_NAME_NITROX_PMD  crypto_nitrox_sym
diff --git a/app/test/test_cryptodev_asym.c b/app/test/test_cryptodev_asym.c
index ac47be724f..a98e3dc824 100644
--- a/app/test/test_cryptodev_asym.c
+++ b/app/test/test_cryptodev_asym.c
@@ -4118,9 +4118,24 @@ test_cryptodev_virtio_asym(void)
return unit_test_suite_runner(&cryptodev_virtio_asym_testsuite);
 }
 
+static int
+test_cryptodev_virtio_user_asym(void)
+{
+   gbl_driver_id = rte_cryptodev_driver_id_get(
+   RTE_STR(CRYPTODEV_NAME_VIRTIO_USER_PMD));
+   if (gbl_driver_id == -1) {
+   RTE_LOG(ERR, USER1, "virtio user PMD must be loaded.\n");
+   return TEST_FAILED;
+   }
+
+   /* Use test suite registered for crypto_virtio_user PMD */
+   return unit_test_suite_runner(&cryptodev_virtio_asym_testsuite);
+}
+
 REGISTER_DRIVER_TEST(cryptodev_openssl_asym_autotest, 
test_cryptodev_openssl_asym);
 REGISTER_DRIVER_TEST(cryptodev_qat_asym_autotest, test_cryptodev_qat_asym);
 REGISTER_DRIVER_TEST(cryptodev_octeontx_asym_autotest, 
test_cryptodev_octeontx_asym);
 REGISTER_DRIVER_TEST(cryptodev_cn9k_asym_autotest, test_cryptodev_cn9k_asym);
 REGISTER_DRIVER_TEST(cryptodev_cn10k_asym_autotest, test_cryptodev_cn10k_asym);
 REGISTER_DRIVER_TEST(cryptodev_virtio_asym_autotest, 
test_cryptodev_virtio_asym);
+REGISTER_DRIVER_TEST(cryptodev_virtio_user_asym_autotest, 
test_cryptodev_virtio_user_asym);
-- 
2.25.1



[v3 5/6] test/crypto: add asymmetric tests for virtio PMD

2025-02-21 Thread Gowrishankar Muthukrishnan
Add asymmetric tests for Virtio PMD.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 app/test/test_cryptodev_asym.c | 28 
 1 file changed, 28 insertions(+)

diff --git a/app/test/test_cryptodev_asym.c b/app/test/test_cryptodev_asym.c
index 9b5f3c545e..ac47be724f 100644
--- a/app/test/test_cryptodev_asym.c
+++ b/app/test/test_cryptodev_asym.c
@@ -4023,6 +4023,19 @@ static struct unit_test_suite 
cryptodev_octeontx_asym_testsuite  = {
}
 };
 
+static struct unit_test_suite cryptodev_virtio_asym_testsuite  = {
+   .suite_name = "Crypto Device VIRTIO ASYM Unit Test Suite",
+   .setup = testsuite_setup,
+   .teardown = testsuite_teardown,
+   .unit_test_cases = {
+   TEST_CASE_ST(ut_setup_asym, ut_teardown_asym, test_capability),
+   TEST_CASE_ST(ut_setup_asym, ut_teardown_asym,
+   test_rsa_sign_verify_crt),
+   TEST_CASE_ST(ut_setup_asym, ut_teardown_asym, 
test_rsa_enc_dec_crt),
+   TEST_CASES_END() /**< NULL terminate unit test array */
+   }
+};
+
 static int
 test_cryptodev_openssl_asym(void)
 {
@@ -4091,8 +4104,23 @@ test_cryptodev_cn10k_asym(void)
return unit_test_suite_runner(&cryptodev_octeontx_asym_testsuite);
 }
 
+static int
+test_cryptodev_virtio_asym(void)
+{
+   gbl_driver_id = rte_cryptodev_driver_id_get(
+   RTE_STR(CRYPTODEV_NAME_VIRTIO_PMD));
+   if (gbl_driver_id == -1) {
+   RTE_LOG(ERR, USER1, "virtio PMD must be loaded.\n");
+   return TEST_FAILED;
+   }
+
+   /* Use test suite registered for crypto_virtio PMD */
+   return unit_test_suite_runner(&cryptodev_virtio_asym_testsuite);
+}
+
 REGISTER_DRIVER_TEST(cryptodev_openssl_asym_autotest, 
test_cryptodev_openssl_asym);
 REGISTER_DRIVER_TEST(cryptodev_qat_asym_autotest, test_cryptodev_qat_asym);
 REGISTER_DRIVER_TEST(cryptodev_octeontx_asym_autotest, 
test_cryptodev_octeontx_asym);
 REGISTER_DRIVER_TEST(cryptodev_cn9k_asym_autotest, test_cryptodev_cn9k_asym);
 REGISTER_DRIVER_TEST(cryptodev_cn10k_asym_autotest, test_cryptodev_cn10k_asym);
+REGISTER_DRIVER_TEST(cryptodev_virtio_asym_autotest, 
test_cryptodev_virtio_asym);
-- 
2.25.1



[RFC PATCH v19] mempool: fix mempool cache size

2025-02-21 Thread Morten Brørup
NOTE: THIS VERSION DOES NOT BREAK THE API/ABI.

First, a per-lcore mempool cache could hold 50 % more than the cache's
size.
Since application developers do not expect this behavior, it could lead to
application failure.
This patch fixes this bug without breaking the API/ABI, by using the
mempool cache's "size" instead of the "flushthresh" as the threshold for
how many objects can be held in a mempool cache.
Note: The "flushthresh" field can be removed from the cache structure in a
future API/ABI breaking release, which must be announced in advance.

Second, requests to fetch a number of objects from the backend driver
exceeding the cache's size (but less than RTE_MEMPOOL_CACHE_MAX_SIZE) were
copied twice; first to the cache, and from there to the destination.
Such superfluous copying through the mempool cache degrades the
performance in these cases.
This patch also fixes this misbehavior, so when fetching more objects from
the driver than the mempool cache's size, they are fetched directly to the
destination.

The internal macro to calculate the cache flush threshold was updated to
reflect the new flush threshold of 1 * size instead of 1.5 * size.

The function rte_mempool_do_generic_put() for adding objects to a mempool
was modified as follows:
- When determining if the cache has sufficient room for the request
  without flushing, compare to the cache's size (cache->size) instead of
  the obsolete flush threshold (cache->flushthresh).
- The comparison for the request being too big, which is considered
  unlikely, was moved down and out of the code path where the cache has
  sufficient room for the added objects, which is considered the most
  likely code path.
- Added __rte_assume() about the cache size, for compiler optimization
  when "n" is compile time constant.
- Added __rte_assume() about "ret", for compiler optimization of
  rte_mempool_generic_get() considering the return value of
  rte_mempool_do_generic_get().

The function rte_mempool_do_generic_get() for getting objects from a
mempool was refactored as follows:
- Handling a request for a constant number of objects was merged with
  handling a request for a nonconstant number of objects, and a note about
  compiler loop unrolling in the constant case was added.
- When determining if the remaining part of a request to be dequeued from
  the backend is too big to be copied via the cache, compare to the
  cache's size (cache->size) instead of the max possible cache size
  (RTE_MEMPOOL_CACHE_MAX_SIZE).
- When refilling the cache, the target fill level was reduced from the
  full cache size to half the cache size. This allows some room for a
  put() request following a get() request where the cache was refilled,
  without "flapping" between draining and refilling the entire cache.
  Note: Before this patch, the distance between the flush threshold and
  the refill level was also half a cache size.
- A copy of cache->len in the local variable "len" is no longer needed,
  so it was removed.
- Added a group of __rte_assume()'s, for compiler optimization when "n" is
  compile time constant.

Some comments were also updated.

Furthermore, some likely()/unlikely()'s were added to a few inline
functions; most prominently rte_mempool_default_cache(), which is used by
both rte_mempool_put_bulk() and rte_mempool_get_bulk().

And finally, RTE_ASSERT()'s were added to check the return values of the
mempool driver dequeue() and enqueue() operations.

Signed-off-by: Morten Brørup 
---
v19:
* Added __rte_assume()'s and RTE_ASSERT()'s.
v18:
* Start over from scratch, to avoid API/ABI breakage.
v17:
* Update rearm in idpf driver.
v16:
* Fix bug in rte_mempool_do_generic_put() regarding criteria for flush.
v15:
* Changed back cache bypass limit from n >= RTE_MEMPOOL_CACHE_MAX_SIZE to
  n > RTE_MEMPOOL_CACHE_MAX_SIZE.
* Removed cache size limit from serving via cache.
v14:
* Change rte_mempool_do_generic_put() back from add-then-flush to
  flush-then-add.
  Keep the target cache fill level of ca. 1/2 size of the cache.
v13:
* Target a cache fill level of ca. 1/2 size of the cache when flushing and
  refilling; based on an assumption of equal probability of get and put,
  instead of assuming a higher probability of put being followed by
  another put, and get being followed by another get.
* Reduce the amount of changes to the drivers.
v12:
* Do not init mempool caches with size zero; they don't exist.
  Bug introduced in v10.
v11:
* Removed rte_mempool_do_generic_get_split().
v10:
* Initialize mempool caches, regardless of size zero.
  This to fix compiler warning about out of bounds access.
v9:
* Removed factor 1.5 from description of cache_size parameter to
  rte_mempool_create().
* Refactored rte_mempool_do_generic_put() to eliminate some gotos.
  No functional change.
* Removed check for n >= RTE_MEMPOOL_CACHE_MAX_SIZE in
  rte_mempool_do_generic_get(); it caused the function to fail when the
  request could not be served from the backend alone, but it could be
 

[v3 4/5] vhost: support asymmetric RSA crypto ops

2025-02-21 Thread Gowrishankar Muthukrishnan
Support asymmetric RSA crypto operations in vhost-user.

Signed-off-by: Gowrishankar Muthukrishnan 
---
v3:
 - TLV decode optimization for fast path.
 - virtio_crypto.h changes moved from virtio PMD patch series into this series
   as asymmetric support starts essentially from library.
---

 lib/vhost/vhost_crypto.c  | 492 +++---
 lib/vhost/virtio_crypto.h |  67 ++
 2 files changed, 524 insertions(+), 35 deletions(-)

diff --git a/lib/vhost/vhost_crypto.c b/lib/vhost/vhost_crypto.c
index 05f3c85884..9892603891 100644
--- a/lib/vhost/vhost_crypto.c
+++ b/lib/vhost/vhost_crypto.c
@@ -54,6 +54,14 @@ RTE_LOG_REGISTER_SUFFIX(vhost_crypto_logtype, crypto, INFO);
  */
 #define vhost_crypto_desc vring_desc
 
+struct vhost_crypto_session {
+   union {
+   struct rte_cryptodev_asym_session *asym;
+   struct rte_cryptodev_sym_session *sym;
+   };
+   enum rte_crypto_op_type type;
+};
+
 static int
 cipher_algo_transform(uint32_t virtio_cipher_algo,
enum rte_crypto_cipher_algorithm *algo)
@@ -206,8 +214,10 @@ struct __rte_cache_aligned vhost_crypto {
 
uint64_t last_session_id;
 
-   uint64_t cache_session_id;
-   struct rte_cryptodev_sym_session *cache_session;
+   uint64_t cache_sym_session_id;
+   struct rte_cryptodev_sym_session *cache_sym_session;
+   uint64_t cache_asym_session_id;
+   struct rte_cryptodev_asym_session *cache_asym_session;
/** socket id for the device */
int socket_id;
 
@@ -334,10 +344,11 @@ transform_chain_param(struct rte_crypto_sym_xform *xforms,
 }
 
 static void
-vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
+vhost_crypto_create_sym_sess(struct vhost_crypto *vcrypto,
VhostUserCryptoSessionParam *sess_param)
 {
struct rte_crypto_sym_xform xform1 = {0}, xform2 = {0};
+   struct vhost_crypto_session *vhost_session;
struct rte_cryptodev_sym_session *session;
int ret;
 
@@ -384,42 +395,277 @@ vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
return;
}
 
-   /* insert hash to map */
-   if (rte_hash_add_key_data(vcrypto->session_map,
-   &vcrypto->last_session_id, session) < 0) {
+   vhost_session = rte_zmalloc(NULL, sizeof(*vhost_session), 0);
+   if (vhost_session == NULL) {
+   VC_LOG_ERR("Failed to alloc session memory");
+   goto error_exit;
+   }
+
+   vhost_session->type = RTE_CRYPTO_OP_TYPE_SYMMETRIC;
+   vhost_session->sym = session;
+
+   /* insert session to map */
+   if ((rte_hash_add_key_data(vcrypto->session_map,
+   &vcrypto->last_session_id, vhost_session) < 0)) {
VC_LOG_ERR("Failed to insert session to hash table");
+   goto error_exit;
+   }
+
+   VC_LOG_INFO("Session %"PRIu64" created for vdev %i.",
+   vcrypto->last_session_id, vcrypto->dev->vid);
+
+   sess_param->session_id = vcrypto->last_session_id;
+   vcrypto->last_session_id++;
+   return;
+
+error_exit:
+   if (rte_cryptodev_sym_session_free(vcrypto->cid, session) < 0)
+   VC_LOG_ERR("Failed to free session");
+
+   sess_param->session_id = -VIRTIO_CRYPTO_ERR;
+   rte_free(vhost_session);
+}
+
+static int
+tlv_decode(uint8_t *tlv, uint8_t type, uint8_t **data, size_t *data_len)
+{
+   size_t tlen = -EINVAL, len;
+
+   if (tlv[0] != type)
+   return -EINVAL;
+
+   if (tlv[1] == 0x82) {
+   len = (tlv[2] << 8) | tlv[3];
+   *data = &tlv[4];
+   tlen = len + 4;
+   } else if (tlv[1] == 0x81) {
+   len = tlv[2];
+   *data = &tlv[3];
+   tlen = len + 3;
+   } else {
+   len = tlv[1];
+   *data = &tlv[2];
+   tlen = len + 2;
+   }
+
+   *data_len = len;
+   return tlen;
+}
+
+static int
+virtio_crypto_asym_rsa_der_to_xform(uint8_t *der, size_t der_len,
+   struct rte_crypto_asym_xform *xform)
+{
+   uint8_t *n = NULL, *e = NULL, *d = NULL, *p = NULL, *q = NULL, *dp = 
NULL,
+   *dq = NULL, *qinv = NULL, *v = NULL, *tlv;
+   size_t nlen, elen, dlen, plen, qlen, dplen, dqlen, qinvlen, vlen;
+   int len;
+
+   RTE_SET_USED(der_len);
+
+   if (der[0] != 0x30)
+   return -EINVAL;
+
+   if (der[1] == 0x82)
+   tlv = &der[4];
+   else if (der[1] == 0x81)
+   tlv = &der[3];
+   else
+   return -EINVAL;
+
+   len = tlv_decode(tlv, 0x02, &v, &vlen);
+   if (len < 0 || v[0] != 0x0 || vlen != 1)
+   return -EINVAL;
+
+   tlv = tlv + len;
+   len = tlv_decode(tlv, 0x02, &n, &nlen);
+   if (len < 0)
+   return len;
+
+   tlv = tlv + len;
+   len = tlv_decode(tlv, 0x02, &e, &elen);
+   if (len < 0)
+   return len;
+
+   t

[v3 1/5] vhost: skip crypto op fetch before vring init

2025-02-21 Thread Gowrishankar Muthukrishnan
Until virtio avail ring is initialized (by VHOST_USER_SET_VRING_ADDR),
worker thread should not try to fetch crypto op, which would lead to
memory fault.

Fixes: 939066d9656 ("vhost/crypto: add public function implementation")
Cc: sta...@dpdk.org

Signed-off-by: Gowrishankar Muthukrishnan 
---
 lib/vhost/vhost_crypto.c | 10 ++
 1 file changed, 10 insertions(+)

diff --git a/lib/vhost/vhost_crypto.c b/lib/vhost/vhost_crypto.c
index 3dc41a3bd5..55ea24710e 100644
--- a/lib/vhost/vhost_crypto.c
+++ b/lib/vhost/vhost_crypto.c
@@ -1580,6 +1580,16 @@ rte_vhost_crypto_fetch_requests(int vid, uint32_t qid,
 
vq = dev->virtqueue[qid];
 
+   if (unlikely(vq == NULL)) {
+   VC_LOG_ERR("Invalid virtqueue %u", qid);
+   return 0;
+   }
+
+   if (unlikely(vq->avail == NULL)) {
+   VC_LOG_DBG("Virtqueue ring not yet initialized %u", qid);
+   return 0;
+   }
+
avail_idx = *((volatile uint16_t *)&vq->avail->idx);
start_idx = vq->last_used_idx;
count = avail_idx - start_idx;
-- 
2.25.1



[v3 5/5] examples/vhost_crypto: support asymmetric crypto

2025-02-21 Thread Gowrishankar Muthukrishnan
Support asymmetric crypto operations.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 examples/vhost_crypto/main.c | 50 +++-
 1 file changed, 38 insertions(+), 12 deletions(-)

diff --git a/examples/vhost_crypto/main.c b/examples/vhost_crypto/main.c
index b1fe4120b9..8bdfc40c4b 100644
--- a/examples/vhost_crypto/main.c
+++ b/examples/vhost_crypto/main.c
@@ -59,6 +59,7 @@ struct vhost_crypto_options {
uint32_t nb_los;
uint32_t zero_copy;
uint32_t guest_polling;
+   bool asymmetric_crypto;
 } options;
 
 enum {
@@ -70,6 +71,8 @@ enum {
OPT_ZERO_COPY_NUM,
 #define OPT_POLLING "guest-polling"
OPT_POLLING_NUM,
+#define OPT_ASYM"asymmetric-crypto"
+   OPT_ASYM_NUM,
 };
 
 #define NB_SOCKET_FIELDS   (2)
@@ -202,9 +205,10 @@ vhost_crypto_usage(const char *prgname)
"  --%s ,SOCKET-FILE-PATH\n"
"  --%s (lcore,cdev_id,queue_id)[,(lcore,cdev_id,queue_id)]\n"
"  --%s: zero copy\n"
-   "  --%s: guest polling\n",
+   "  --%s: guest polling\n"
+   "  --%s: asymmetric crypto\n",
prgname, OPT_SOCKET_FILE, OPT_CONFIG,
-   OPT_ZERO_COPY, OPT_POLLING);
+   OPT_ZERO_COPY, OPT_POLLING, OPT_ASYM);
 }
 
 static int
@@ -223,6 +227,8 @@ vhost_crypto_parse_args(int argc, char **argv)
NULL, OPT_ZERO_COPY_NUM},
{OPT_POLLING, no_argument,
NULL, OPT_POLLING_NUM},
+   {OPT_ASYM, no_argument,
+   NULL, OPT_ASYM_NUM},
{NULL, 0, 0, 0}
};
 
@@ -262,6 +268,10 @@ vhost_crypto_parse_args(int argc, char **argv)
options.guest_polling = 1;
break;
 
+   case OPT_ASYM_NUM:
+   options.asymmetric_crypto = true;
+   break;
+
default:
vhost_crypto_usage(prgname);
return -EINVAL;
@@ -376,6 +386,7 @@ vhost_crypto_worker(void *arg)
int callfds[VIRTIO_CRYPTO_MAX_NUM_BURST_VQS];
uint32_t lcore_id = rte_lcore_id();
uint32_t burst_size = MAX_PKT_BURST;
+   enum rte_crypto_op_type cop_type;
uint32_t i, j, k;
uint32_t to_fetch, fetched;
 
@@ -383,9 +394,13 @@ vhost_crypto_worker(void *arg)
 
RTE_LOG(INFO, USER1, "Processing on Core %u started\n", lcore_id);
 
+   cop_type = RTE_CRYPTO_OP_TYPE_SYMMETRIC;
+   if (options.asymmetric_crypto)
+   cop_type = RTE_CRYPTO_OP_TYPE_ASYMMETRIC;
+
for (i = 0; i < NB_VIRTIO_QUEUES; i++) {
if (rte_crypto_op_bulk_alloc(info->cop_pool,
-   RTE_CRYPTO_OP_TYPE_SYMMETRIC, ops[i],
+   cop_type, ops[i],
burst_size) < burst_size) {
RTE_LOG(ERR, USER1, "Failed to alloc cops\n");
ret = -1;
@@ -411,12 +426,11 @@ vhost_crypto_worker(void *arg)
fetched);
if (unlikely(rte_crypto_op_bulk_alloc(
info->cop_pool,
-   RTE_CRYPTO_OP_TYPE_SYMMETRIC,
+   cop_type,
ops[j], fetched) < fetched)) {
RTE_LOG(ERR, USER1, "Failed realloc\n");
return -1;
}
-
fetched = rte_cryptodev_dequeue_burst(
info->cid, info->qid,
ops_deq[j], RTE_MIN(burst_size,
@@ -477,6 +491,7 @@ main(int argc, char *argv[])
struct rte_cryptodev_qp_conf qp_conf;
struct rte_cryptodev_config config;
struct rte_cryptodev_info dev_info;
+   enum rte_crypto_op_type cop_type;
char name[128];
uint32_t i, j, lcore;
int ret;
@@ -539,12 +554,21 @@ main(int argc, char *argv[])
goto error_exit;
}
 
-   snprintf(name, 127, "SESS_POOL_%u", lo->lcore_id);
-   info->sess_pool = rte_cryptodev_sym_session_pool_create(name,
-   SESSION_MAP_ENTRIES,
-   rte_cryptodev_sym_get_private_session_size(
-   info->cid), 0, 0,
-   rte_lcore_to_socket_id(lo->lcore_id));
+   if (!options.asymmetric_crypto) {
+   snprintf(name, 127, "SYM_SESS_POOL_%u", lo->lcore_id);
+   info->sess_pool = 
rte_cryptodev_sym_session_pool_create(name,
+   SESSION_MAP_ENTRIES,

[v3 3/5] examples/vhost_crypto: fix user callbacks

2025-02-21 Thread Gowrishankar Muthukrishnan
In order to handle new vhost user connection, use new_connection
and destroy_connection callbacks.

Fixes: f5188211c721 ("examples/vhost_crypto: add sample application")
Cc: sta...@dpdk.org

Signed-off-by: Gowrishankar Muthukrishnan 
---
v3:
 - decoupled from v2 single patch.
---
 examples/vhost_crypto/main.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/examples/vhost_crypto/main.c b/examples/vhost_crypto/main.c
index 558c09a60f..b1fe4120b9 100644
--- a/examples/vhost_crypto/main.c
+++ b/examples/vhost_crypto/main.c
@@ -362,8 +362,8 @@ destroy_device(int vid)
 }
 
 static const struct rte_vhost_device_ops virtio_crypto_device_ops = {
-   .new_device =  new_device,
-   .destroy_device = destroy_device,
+   .new_connection =  new_device,
+   .destroy_connection = destroy_device,
 };
 
 static int
-- 
2.25.1



[v3 0/5] vhost: add RSA support

2025-02-21 Thread Gowrishankar Muthukrishnan
This patch series supports asymmetric RSA in vhost crypto library.
It also includes changes to improve vhost crypto library:
 * support newer QEMU versions.
 * fix broken vhost_crypto example application.
 * stabilize crypto fastpath operations.

v3:
 - spin off new patches from one single patch in v2.
 - stabilized vhost crypto lib and example app.

Gowrishankar Muthukrishnan (5):
  vhost: skip crypto op fetch before vring init
  vhost: update vhost_user crypto session parameters
  examples/vhost_crypto: fix user callbacks
  vhost: support asymmetric RSA crypto ops
  examples/vhost_crypto: support asymmetric crypto

 examples/vhost_crypto/main.c |  54 +++-
 lib/vhost/vhost_crypto.c | 514 ---
 lib/vhost/vhost_user.h   |  33 ++-
 lib/vhost/virtio_crypto.h|  67 +
 4 files changed, 609 insertions(+), 59 deletions(-)

-- 
2.25.1



[v3 2/5] vhost: update vhost_user crypto session parameters

2025-02-21 Thread Gowrishankar Muthukrishnan
As per requirements on vhost_user spec, session id should be
located at the end of session parameter.

Update VhostUserCryptoSessionParam structure to support newer QEMU.
Due to additional parameters added in QEMU, received payload from
QEMU would be larger than existing payload, hence breaks parsing
vhost_user message.

This patch addresses both of the above problems.

Signed-off-by: Gowrishankar Muthukrishnan 
---
v3:
 - decoupled originally from v2 series single patch.
---

 lib/vhost/vhost_crypto.c | 12 ++--
 lib/vhost/vhost_user.h   | 33 +
 2 files changed, 35 insertions(+), 10 deletions(-)

diff --git a/lib/vhost/vhost_crypto.c b/lib/vhost/vhost_crypto.c
index 55ea24710e..05f3c85884 100644
--- a/lib/vhost/vhost_crypto.c
+++ b/lib/vhost/vhost_crypto.c
@@ -237,7 +237,7 @@ struct vhost_crypto_data_req {
 
 static int
 transform_cipher_param(struct rte_crypto_sym_xform *xform,
-   VhostUserCryptoSessionParam *param)
+   VhostUserCryptoSymSessionParam *param)
 {
int ret;
 
@@ -273,7 +273,7 @@ transform_cipher_param(struct rte_crypto_sym_xform *xform,
 
 static int
 transform_chain_param(struct rte_crypto_sym_xform *xforms,
-   VhostUserCryptoSessionParam *param)
+   VhostUserCryptoSymSessionParam *param)
 {
struct rte_crypto_sym_xform *xform_cipher, *xform_auth;
int ret;
@@ -341,10 +341,10 @@ vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
struct rte_cryptodev_sym_session *session;
int ret;
 
-   switch (sess_param->op_type) {
+   switch (sess_param->u.sym_sess.op_type) {
case VIRTIO_CRYPTO_SYM_OP_NONE:
case VIRTIO_CRYPTO_SYM_OP_CIPHER:
-   ret = transform_cipher_param(&xform1, sess_param);
+   ret = transform_cipher_param(&xform1, &sess_param->u.sym_sess);
if (unlikely(ret)) {
VC_LOG_ERR("Error transform session msg (%i)", ret);
sess_param->session_id = ret;
@@ -352,7 +352,7 @@ vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
}
break;
case VIRTIO_CRYPTO_SYM_OP_ALGORITHM_CHAINING:
-   if (unlikely(sess_param->hash_mode !=
+   if (unlikely(sess_param->u.sym_sess.hash_mode !=
VIRTIO_CRYPTO_SYM_HASH_MODE_AUTH)) {
sess_param->session_id = -VIRTIO_CRYPTO_NOTSUPP;
VC_LOG_ERR("Error transform session message (%i)",
@@ -362,7 +362,7 @@ vhost_crypto_create_sess(struct vhost_crypto *vcrypto,
 
xform1.next = &xform2;
 
-   ret = transform_chain_param(&xform1, sess_param);
+   ret = transform_chain_param(&xform1, &sess_param->u.sym_sess);
if (unlikely(ret)) {
VC_LOG_ERR("Error transform session message (%i)", ret);
sess_param->session_id = ret;
diff --git a/lib/vhost/vhost_user.h b/lib/vhost/vhost_user.h
index 9a905ee5f4..ef486545ba 100644
--- a/lib/vhost/vhost_user.h
+++ b/lib/vhost/vhost_user.h
@@ -99,11 +99,10 @@ typedef struct VhostUserLog {
 /* Comply with Cryptodev-Linux */
 #define VHOST_USER_CRYPTO_MAX_HMAC_KEY_LENGTH  512
 #define VHOST_USER_CRYPTO_MAX_CIPHER_KEY_LENGTH64
+#define VHOST_USER_CRYPTO_MAX_KEY_LENGTH   1024
 
 /* Same structure as vhost-user backend session info */
-typedef struct VhostUserCryptoSessionParam {
-   int64_t session_id;
-   uint32_t op_code;
+typedef struct VhostUserCryptoSymSessionParam {
uint32_t cipher_algo;
uint32_t cipher_key_len;
uint32_t hash_algo;
@@ -114,10 +113,36 @@ typedef struct VhostUserCryptoSessionParam {
uint8_t dir;
uint8_t hash_mode;
uint8_t chaining_dir;
-   uint8_t *ciphe_key;
+   uint8_t *cipher_key;
uint8_t *auth_key;
uint8_t cipher_key_buf[VHOST_USER_CRYPTO_MAX_CIPHER_KEY_LENGTH];
uint8_t auth_key_buf[VHOST_USER_CRYPTO_MAX_HMAC_KEY_LENGTH];
+} VhostUserCryptoSymSessionParam;
+
+
+typedef struct VhostUserCryptoAsymRsaParam {
+   uint32_t padding_algo;
+   uint32_t hash_algo;
+} VhostUserCryptoAsymRsaParam;
+
+typedef struct VhostUserCryptoAsymSessionParam {
+   uint32_t algo;
+   uint32_t key_type;
+   uint32_t key_len;
+   uint8_t *key;
+   union {
+   VhostUserCryptoAsymRsaParam rsa;
+   } u;
+   uint8_t key_buf[VHOST_USER_CRYPTO_MAX_KEY_LENGTH];
+} VhostUserCryptoAsymSessionParam;
+
+typedef struct VhostUserCryptoSessionParam {
+   uint32_t op_code;
+   union {
+   VhostUserCryptoSymSessionParam sym_sess;
+   VhostUserCryptoAsymSessionParam asym_sess;
+   } u;
+   int64_t session_id;
 } VhostUserCryptoSessionParam;
 
 typedef struct VhostUserVringArea {
-- 
2.25.1



Re: [PATCH] sched: fix wrr parameter data type

2025-02-21 Thread Stephen Hemminger
On Fri, 21 Feb 2025 18:17:55 +0530
Megha Ajmera  wrote:

> wrr tokens getting truncated to uint8_t in wrr_store function() due to
> type mismatch. This patch chnages the data type to uint16_t.
> 
> Fixes: e16b06da0908 ("sched: remove WRR from strict priority TC queues")
> 
> Signed-off-by: Megha Ajmera 
> ---
>  lib/sched/rte_sched.c | 2 +-
>  1 file changed, 1 insertion(+), 1 deletion(-)
> 
> diff --git a/lib/sched/rte_sched.c b/lib/sched/rte_sched.c
> index d8ee4e7e91..dcef44b91b 100644
> --- a/lib/sched/rte_sched.c
> +++ b/lib/sched/rte_sched.c
> @@ -66,7 +66,7 @@ struct __rte_cache_aligned rte_sched_pipe {
>   uint64_t tc_credits[RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE];
>  
>   /* Weighted Round Robin (WRR) */
> - uint8_t wrr_tokens[RTE_SCHED_BE_QUEUES_PER_PIPE];
> + uint16_t wrr_tokens[RTE_SCHED_BE_QUEUES_PER_PIPE];
>  
>   /* TC oversubscription */
>   uint64_t tc_ov_credits;

This would be a change in ABI.


[PATCH v2] crypto/openssl: validate incorrect signature in verify op

2025-02-21 Thread Gowrishankar Muthukrishnan
Return correct error status when incorrect signature is
used in RSA verify op.

Fixes: d7bd42f6db19 ("crypto/openssl: update RSA routine with 3.0 EVP API")
Cc: sta...@dpdk.org

Signed-off-by: Gowrishankar Muthukrishnan 
---
v2:
 - added comments.
---
 drivers/crypto/openssl/rte_openssl_pmd.c | 10 --
 1 file changed, 8 insertions(+), 2 deletions(-)

diff --git a/drivers/crypto/openssl/rte_openssl_pmd.c 
b/drivers/crypto/openssl/rte_openssl_pmd.c
index b090611bd0..5bfad92b7c 100644
--- a/drivers/crypto/openssl/rte_openssl_pmd.c
+++ b/drivers/crypto/openssl/rte_openssl_pmd.c
@@ -2803,9 +2803,15 @@ process_openssl_rsa_op_evp(struct rte_crypto_op *cop,
goto err_rsa;
}
 
-   if (EVP_PKEY_verify_recover(rsa_ctx, tmp, &outlen,
+   ret = EVP_PKEY_verify_recover(rsa_ctx, tmp, &outlen,
op->rsa.sign.data,
-   op->rsa.sign.length) <= 0) {
+   op->rsa.sign.length);
+   if (ret <= 0) {
+   /* OpenSSL RSA verification returns one on
+* successful verification, otherwise 0. Hence,
+* this enqueue operation should succeed even if
+* invalid signature has been requested in verify.
+*/
OPENSSL_free(tmp);
goto err_rsa;
}
-- 
2.25.1



[PATCH] sched: fix wrr parameter data type

2025-02-21 Thread Megha Ajmera
wrr tokens getting truncated to uint8_t in wrr_store function() due to
type mismatch. This patch chnages the data type to uint16_t.

Fixes: e16b06da0908 ("sched: remove WRR from strict priority TC queues")

Signed-off-by: Megha Ajmera 
---
 lib/sched/rte_sched.c | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/lib/sched/rte_sched.c b/lib/sched/rte_sched.c
index d8ee4e7e91..dcef44b91b 100644
--- a/lib/sched/rte_sched.c
+++ b/lib/sched/rte_sched.c
@@ -66,7 +66,7 @@ struct __rte_cache_aligned rte_sched_pipe {
uint64_t tc_credits[RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE];
 
/* Weighted Round Robin (WRR) */
-   uint8_t wrr_tokens[RTE_SCHED_BE_QUEUES_PER_PIPE];
+   uint16_t wrr_tokens[RTE_SCHED_BE_QUEUES_PER_PIPE];
 
/* TC oversubscription */
uint64_t tc_ov_credits;
-- 
2.34.1



RE: [PATCH] sched: fix wrr parameter data type

2025-02-21 Thread Singh, Jasvinder



> -Original Message-
> From: Ajmera, Megha 
> Sent: Friday, February 21, 2025 12:48 PM
> To: dev@dpdk.org
> Cc: Singh, Jasvinder ; Dumitrescu, Cristian
> 
> Subject: [PATCH] sched: fix wrr parameter data type
> 
> wrr tokens getting truncated to uint8_t in wrr_store function() due to type
> mismatch. This patch chnages the data type to uint16_t.
> 
> Fixes: e16b06da0908 ("sched: remove WRR from strict priority TC queues")
> 
> Signed-off-by: Megha Ajmera 
> ---
>  lib/sched/rte_sched.c | 2 +-
>  1 file changed, 1 insertion(+), 1 deletion(-)
> 
> diff --git a/lib/sched/rte_sched.c b/lib/sched/rte_sched.c index
> d8ee4e7e91..dcef44b91b 100644
> --- a/lib/sched/rte_sched.c
> +++ b/lib/sched/rte_sched.c
> @@ -66,7 +66,7 @@ struct __rte_cache_aligned rte_sched_pipe {
>   uint64_t tc_credits[RTE_SCHED_TRAFFIC_CLASSES_PER_PIPE];
> 
>   /* Weighted Round Robin (WRR) */
> - uint8_t wrr_tokens[RTE_SCHED_BE_QUEUES_PER_PIPE];
> + uint16_t wrr_tokens[RTE_SCHED_BE_QUEUES_PER_PIPE];
> 
>   /* TC oversubscription */
>   uint64_t tc_ov_credits;
> --
> 2.34.1

Acked-by: Jasvinder Singh 



RE: [PATCH v2] net/mlx5/hws: fix fragmented ptype match

2025-02-21 Thread Dariusz Sosnowski



> -Original Message-
> From: Alexander Kozyrev 
> Sent: Thursday, December 19, 2024 17:06
> To: dev@dpdk.org
> Cc: sta...@dpdk.org; Raslan Darawsheh ; Slava Ovsiienko
> ; Dariusz Sosnowski ; Bing
> Zhao ; Suanming Mou ; Michael
> Baum 
> Subject: [PATCH v2] net/mlx5/hws: fix fragmented ptype match
> 
> Fragmented PTYPE matching requires setting the mask to the exact
> RTE_PTYPE_L4_FRAG value to avoid conflicts with other L4 types.
> Adding L2 or L3 types to the same mask should be allowed, but there is a check
> for the exact value for setting the definer.
> This prevents the fragmented packets from matching in case of L2/L3 mask is
> provided as well. Mask out L2/L3 types when setting L4_FRAG.
> 
> Fixes: 761439a20f net/mlx5/hws: support fragmented packet type matching
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Alexander Kozyrev 

Acked-by: Dariusz Sosnowski 

Best regards,
Dariusz Sosnowski


[PATCH] net/ice: fix incorrect mbuf cleanup function

2025-02-21 Thread Bruce Richardson
The mbuf cleanup function is not being called correctly on close of the
driver, due to the vector_tx flag not being set correctly on each Tx
queue. Normally, this causes no visible problems, because the default is
to have the vector flag unset, which means that, if vector Tx path is
chosen, we only free every second mbuf in the descriptor vector ring - a
subtle memory leak.

The reverse case - where we initially configure a port to use a vector
driver, and then later reconfigure it for scalar - causes a segfault on
close, as the vector cleanup is used on a scalar SW ring, and that
vector path treats the extra metadata in the SW ring as mbufs to be
freed.

This is due to an inconsistency across drivers on port start: in e.g.
the i40e driver, the Rx and Tx functions are chosen before the
individual queues are started, while in ice driver the order is
reversed.  Because the Tx function selection is done after the queues
are initialized, the initialized queues use an old value of the overall
flag indicating if a vector Tx path is being used.

Fix the issue by moving the Rx and Tx path selection before queue start.

Fixes: 552979dfb1c9 ("net/intel: create common Tx queue mbuf cleanup")

Signed-off-by: Bruce Richardson 
---
 drivers/net/intel/ice/ice_ethdev.c | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)

diff --git a/drivers/net/intel/ice/ice_ethdev.c 
b/drivers/net/intel/ice/ice_ethdev.c
index 0e092702cc..3cdfa16f77 100644
--- a/drivers/net/intel/ice/ice_ethdev.c
+++ b/drivers/net/intel/ice/ice_ethdev.c
@@ -3961,6 +3961,9 @@ ice_dev_start(struct rte_eth_dev *dev)
ice_declare_bitmap(pmask, ICE_PROMISC_MAX);
ice_zero_bitmap(pmask, ICE_PROMISC_MAX);
 
+   ice_set_rx_function(dev);
+   ice_set_tx_function(dev);
+
/* program Tx queues' context in hardware */
for (nb_txq = 0; nb_txq < data->nb_tx_queues; nb_txq++) {
ret = ice_tx_queue_start(dev, nb_txq);
@@ -3989,9 +3992,6 @@ ice_dev_start(struct rte_eth_dev *dev)
}
}
 
-   ice_set_rx_function(dev);
-   ice_set_tx_function(dev);
-
mask = RTE_ETH_VLAN_STRIP_MASK | RTE_ETH_VLAN_FILTER_MASK |
RTE_ETH_VLAN_EXTEND_MASK;
if (ice_is_dvm_ena(hw))
-- 
2.43.0



Re: [PATCH v4 01/10] eal: add workaround for __builtin_constant_p

2025-02-21 Thread Andre Muezerie
On Fri, Feb 21, 2025 at 04:31:28PM +0100, Morten Brørup wrote:
> Thank you for this updated MSVC variant of the macro, and its detailed 
> description.
> 
> Acked-by: Morten Brørup 

Sure. Thank you for pushing for this better solution.


[PATCH v1] doc/guides: update monitor PMD mode description

2025-02-21 Thread Chris MacNamara
A recent CPU change requires an extra enabling step for
the umonitor instruction on Intel CPUs.
This is now detailed in the l3 fwd power manager doc.

Signed-off-by: Chris MacNamara 
---
 .mailmap  | 1 +
 doc/guides/sample_app_ug/l3_forward_power_man.rst | 5 +
 2 files changed, 6 insertions(+)

diff --git a/.mailmap b/.mailmap
index a03d3cfb59..c4bc38752f 100644
--- a/.mailmap
+++ b/.mailmap
@@ -263,6 +263,7 @@ Christopher Reder 
 Christoph Gysin 
 Christos Ricudis 
 Chris Wright 
+Chris MacNamara 
 Chuanshe Zhang 
 Chuanyu Xue 
 Chuhong Yao 
diff --git a/doc/guides/sample_app_ug/l3_forward_power_man.rst 
b/doc/guides/sample_app_ug/l3_forward_power_man.rst
index 3271bc2154..d0af28e0ec 100644
--- a/doc/guides/sample_app_ug/l3_forward_power_man.rst
+++ b/doc/guides/sample_app_ug/l3_forward_power_man.rst
@@ -293,6 +293,11 @@ and has three available power management schemes:
 ``monitor``
   This will use ``rte_power_monitor()`` function to enter
   a power-optimized state (subject to platform support).
+  On recent Gen 4 Xeon Scalable Processors the umonitor instruction
+  is disabled by default.
+  An additional step is required to enable the umonitor instruction.
+  Writing 0 to bit 6 of register 0x123 will enable umonitor.
+  `More details are available via Monitor and Umonitor Performance Guidance 
`_
 
 ``pause``
   This will use ``rte_power_pause()`` or ``rte_pause()``
-- 
2.43.0

--
Intel Research and Development Ireland Limited
Registered in Ireland
Registered Office: Collinstown Industrial Park, Leixlip, County Kildare
Registered Number: 308263


This e-mail and any attachments may contain confidential material for the sole
use of the intended recipient(s). Any review or distribution by others is
strictly prohibited. If you are not the intended recipient, please contact the
sender and delete all copies.



Intel FAST_FREE offload question

2025-02-21 Thread Morten Brørup
Intel NIC folks,

Why do the Intel network drivers, when using RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE, 
fall back to normal freeing when the mempool cache is empty (cache->len == 0)? 
It doesn't make sense to me.

Example:
https://git.dpdk.org/dpdk/tree/drivers/net/intel/common/tx.h#n146

Med venlig hilsen / Kind regards,
-Morten Brørup




Re: [PATCH] power: use hugepage memory for queue list entry structure

2025-02-21 Thread Burakov, Anatoly

On 20/02/2025 17:45, Stephen Hemminger wrote:

On Thu, 20 Feb 2025 16:39:52 +
Konstantin Ananyev  wrote:


-Original Message-
From: Stephen Hemminger 
Sent: Thursday, February 20, 2025 4:12 PM
To: lihuisong (C) 
Cc: dev@dpdk.org; tho...@monjalon.net; david.h...@intel.com; 
anatoly.bura...@intel.com; sivaprasad.tumm...@amd.com;
liuyonglong 
Subject: Re: [PATCH] power: use hugepage memory for queue list entry structure

On Thu, 20 Feb 2025 17:01:53 +0800
"lihuisong (C)"  wrote:
   

The queue_list_entry structure data is used in rx_callback of io path
when enable PMD Power Management. However its memory is currently from
normal heap memory. For better performance, use hugepage memory to
replace it.

Signed-off-by: Huisong Li 


How is that in a hot path where this could matter?


AFAIU - it is used in RX/TX callbacks that power library installs,
so I presume will get hit on every eth_rx_burst/tx_burst calls.


The safety rails in rte_malloc() are much less than regular malloc().
I prefer some degree of safety from checkers and malloc library internals.


Didn't get your point - what's suddenly wrong with rte_malloc()?


Coverity and Gcc analyzer treat malloc as special case.
With attributes rte_malloc gets similar treatment but not quite as much.
Also internally, malloc and free have more heap pool sanity checks.
In name of performance, those don't exist in rte_malloc().
Lastly hugepages are limited resource, so they should only be used when needed.


The last thing I would associate with rte_malloc is performance. I'm not 
sure I follow - which "sanity checks" were omitted from rte_malloc "in 
the name of performance" that are present in regular malloc?


--
Thanks,
Anatoly


Re: [PATCH v2 1/7] eal: add queue macro extensions from FreeBSD

2025-02-21 Thread Stephen Hemminger
On Wed, 19 Feb 2025 15:55:17 +0100
Thomas Monjalon  wrote:

> 14/02/2025 18:20, Stephen Hemminger:
> > The Linux version of sys/queue.h is frozen at an older version
> > and is missing the _SAFE macro variants. Several drivers started
> > introducing the own workarounds for this. Should be handled in EAL.
> > 
> > Signed-off-by: Stephen Hemminger   
> 
> We may want to unify with lib/eal/windows/include/sys/queue.h

Do we want to just not use the local Linux version of queue.h entirely?
Maybe just put in current FreeBSD one? It has lots more useful variations.

The point is to avoid broken versions (like one Intel driver) or
having same macro in multiple places.


Re: [PATCH v3 1/2] net: add ptype parse for tunnel packets

2025-02-21 Thread Jie Hai

On 2025/2/15 0:35, Stephen Hemminger wrote:

On Fri, 14 Feb 2025 09:56:37 +0800
Jie Hai  wrote:


Add packet types parse for vxlan/vxlan-gpe/gtp/geneve packets.

Signed-off-by: Jie Hai 
---


Not sure about this.
The original purpose of ptype was to support hardware offload information.

Does any driver do this ptype detection in tunnels?

Would the software detection produce the same result as HW offload?
.
As far as I know, there are two types of packet type: sw packet type and 
hw packet type.


The formmer comes from the call of function rte_net_get_ptype(), which 
parses the

network headers in mbuf data and return its packet type.
All caller(of different drivers) share the same standard.
And it's commonly used in Tx checksum process.

The latter comes from the 'packet_type' field in mbuf.
It is about data really present in the Rx mbuf, and it is defined by drivers
themselves and generally arived from Rx descriptor (or hardware).

Only a few drivers use rte_net_get_ptype() to set packet_type, e.g.

1.With mask RTE_PTYPE_L2_MASK | RTE_PTYPE_L3_MASK | RTE_PTYPE_L4_MASK
drivers\net\netvsc\hn_rxtx.c
drivers\net\pfe\pfe_hif_lib.c  under contidion RTE_LIBRTE_PFE_SW_PARSE

2. With mask RTE_PTYPE_ALL_MASK
drivers\net\tap\rte_eth_tap.c
drivers\net\virtio\virtio_rxtx_packed.h
drivers\net\virtio\virtio_rxtx.c
lib\vhost\virtio_net.c
lib\node\kernel_rx.c


What this patch changes is the sw packet type, only these examples are 
affected.


And they may not have hardware information of packets to report.
Please check whether there is any impact, @all maintainers.



Re: [PATCH] test/ring: fix return uninitialized variable

2025-02-21 Thread fengchengwen
Acked-by: Chengwen Feng 

On 2025/2/19 22:23, Konstantin Ananyev wrote:
> This change addresses the CID 455328: Uninitialized variables  (UNINIT).
> Strictly speaking right now it should never happen, as role_mask is not
> zero. But that might change in future, as more test-cases will be
> introduced. Anyway the fix is trivial.
> 
> Coverity issue: 455328
> Fixes: 70581c355d69 ("test/ring: add unit tests for soring API")
> 
> Signed-off-by: Konstantin Ananyev 



Re: [PATCH v5 09/10] test: add workaround for __builtin_constant_p in test_memcpy_perf

2025-02-21 Thread fengchengwen
Acked-by: Chengwen Feng 

On 2025/2/22 3:52, Andre Muezerie wrote:
> There's no MSVC equivalent for compiler extension __builtin_constant_p,
> so a workaround is needed.
> 
> Signed-off-by: Andre Muezerie 



Re: [PATCH v5 10/10] app: enable app directory to be compiled with MSVC

2025-02-21 Thread fengchengwen
Acked-by: Chengwen Feng 

On 2025/2/22 3:52, Andre Muezerie wrote:
> Enabled "app" directory to be compiled with MSVC along with all its
> contents.
> 
> Removed flag Wno-deprecated-declarations which is not needed anymore.
> 
> Signed-off-by: Andre Muezerie 
> Acked-by: Bruce Richardson 



[PATCH 02/12] net/xsc: remove useless call

2025-02-21 Thread Renyong Wan
Address incorrect expression (USELESS_CALL) reported by Coverity Scan.

Coverity issue: 456590

Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_np.c | 1 -
 1 file changed, 1 deletion(-)

diff --git a/drivers/net/xsc/xsc_np.c b/drivers/net/xsc/xsc_np.c
index d4eb833bf6..f032a0dbc2 100644
--- a/drivers/net/xsc/xsc_np.c
+++ b/drivers/net/xsc/xsc_np.c
@@ -384,7 +384,6 @@ xsc_dev_create_vfos_baselp(struct xsc_dev *xdev)
 void
 xsc_dev_pct_uninit(void)
 {
-   rte_bitmap_free(xsc_pct_mgr.bmp_pct);
rte_free(xsc_pct_mgr.bmp_mem);
 }
 
-- 
2.25.1


[PATCH 01/12] net/xsc: avoid integer overflow

2025-02-21 Thread Renyong Wan
Address the integer overflow issue reported by Coverity Scan.

Coverity issue: 456589

Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_tx.c | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/drivers/net/xsc/xsc_tx.c b/drivers/net/xsc/xsc_tx.c
index 406fa95381..1d31e84d69 100644
--- a/drivers/net/xsc/xsc_tx.c
+++ b/drivers/net/xsc/xsc_tx.c
@@ -50,7 +50,7 @@ xsc_txq_obj_new(struct xsc_dev *xdev, struct xsc_txq_data 
*txq_data,
txq_data->cq_db = cq_info.cq_db;
txq_data->cqn = cq_info.cqn;
txq_data->cqes = cq_info.cqes;
-   txq_data->cqe_m = txq_data->cqe_s - 1;
+   txq_data->cqe_m = (uint16_t)(1 << cq_info.cqe_n) - 1;
 
PMD_DRV_LOG(INFO, "Create tx cq, cqe_s:%d, cqe_n:%d, cq_db=%p, cqn:%d",
txq_data->cqe_s, txq_data->cqe_n,
-- 
2.25.1


[PATCH 05/12] net/xsc: avoid variable is assigned but not used

2025-02-21 Thread Renyong Wan
This issue was reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v1001/

Signed-off-by: Rong Qian 
Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_ethdev.c | 5 +
 1 file changed, 5 insertions(+)

diff --git a/drivers/net/xsc/xsc_ethdev.c b/drivers/net/xsc/xsc_ethdev.c
index 0729d43884..d04fe78d3f 100644
--- a/drivers/net/xsc/xsc_ethdev.c
+++ b/drivers/net/xsc/xsc_ethdev.c
@@ -314,6 +314,11 @@ xsc_ethdev_start(struct rte_eth_dev *dev)
dev->tx_pkt_burst = xsc_tx_burst;
 
ret = xsc_ethdev_enable(dev);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "Failed to enable port: %u",
+   dev->data->port_id);
+   goto error;
+   }
 
return 0;
 
-- 
2.25.1


[PATCH 03/12] net/xsc: address incorrect format warnings

2025-02-21 Thread Renyong Wan
These warnings reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v576/

Signed-off-by: Rong Qian 
Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_dev.c   | 2 +-
 drivers/net/xsc/xsc_rx.c| 4 ++--
 drivers/net/xsc/xsc_tx.c| 4 ++--
 drivers/net/xsc/xsc_vfio.c  | 4 ++--
 drivers/net/xsc/xsc_vfio_mbox.c | 2 +-
 5 files changed, 8 insertions(+), 8 deletions(-)

diff --git a/drivers/net/xsc/xsc_dev.c b/drivers/net/xsc/xsc_dev.c
index c836f2f35a..1e3dd493f9 100644
--- a/drivers/net/xsc/xsc_dev.c
+++ b/drivers/net/xsc/xsc_dev.c
@@ -289,7 +289,7 @@ xsc_dev_repr_ports_probe(struct xsc_dev *xdev, int 
nb_repr_ports, int max_eth_po
 
xdev->num_repr_ports = nb_repr_ports + XSC_PHY_PORT_NUM;
if (xdev->num_repr_ports > max_eth_ports) {
-   PMD_DRV_LOG(ERR, "Repr ports num %u, should be less than max 
%u",
+   PMD_DRV_LOG(ERR, "Repr ports num %d, should be less than max 
%d",
xdev->num_repr_ports, max_eth_ports);
return -EINVAL;
}
diff --git a/drivers/net/xsc/xsc_rx.c b/drivers/net/xsc/xsc_rx.c
index 58a9cc2f26..3230710656 100644
--- a/drivers/net/xsc/xsc_rx.c
+++ b/drivers/net/xsc/xsc_rx.c
@@ -358,7 +358,7 @@ xsc_rss_qp_create(struct xsc_ethdev_priv *priv, int port_id)
xsc_rxq_initialize(xdev, rxq_data);
rxq_data->cq_ci = 0;
priv->dev_data->rx_queue_state[i] = RTE_ETH_QUEUE_STATE_STARTED;
-   PMD_DRV_LOG(INFO, "Port %u create rx qp, wqe_s:%d, wqe_n:%d, 
qp_db=%p, qpn:%d",
+   PMD_DRV_LOG(INFO, "Port %d create rx qp, wqe_s:%d, wqe_n:%d, 
qp_db=%p, qpn:%u",
port_id,
rxq_data->wqe_s, rxq_data->wqe_n,
rxq_data->rq_db, rxq_data->qpn);
@@ -407,7 +407,7 @@ xsc_rxq_rss_obj_new(struct xsc_ethdev_priv *priv, uint16_t 
port_id)
rxq_data->cq_db = cq_info.cq_db;
rxq_data->cqn = cq_info.cqn;
 
-   PMD_DRV_LOG(INFO, "Port %u create rx cq, cqe_s:%d, cqe_n:%d, 
cq_db=%p, cqn:%d",
+   PMD_DRV_LOG(INFO, "Port %u create rx cq, cqe_s:%d, cqe_n:%d, 
cq_db=%p, cqn:%u",
port_id,
rxq_data->cqe_s, rxq_data->cqe_n,
rxq_data->cq_db, rxq_data->cqn);
diff --git a/drivers/net/xsc/xsc_tx.c b/drivers/net/xsc/xsc_tx.c
index 1d31e84d69..0d6a9981d0 100644
--- a/drivers/net/xsc/xsc_tx.c
+++ b/drivers/net/xsc/xsc_tx.c
@@ -52,7 +52,7 @@ xsc_txq_obj_new(struct xsc_dev *xdev, struct xsc_txq_data 
*txq_data,
txq_data->cqes = cq_info.cqes;
txq_data->cqe_m = (uint16_t)(1 << cq_info.cqe_n) - 1;
 
-   PMD_DRV_LOG(INFO, "Create tx cq, cqe_s:%d, cqe_n:%d, cq_db=%p, cqn:%d",
+   PMD_DRV_LOG(INFO, "Create tx cq, cqe_s:%d, cqe_n:%d, cq_db=%p, cqn:%u",
txq_data->cqe_s, txq_data->cqe_n,
txq_data->cq_db, txq_data->cqn);
 
@@ -83,7 +83,7 @@ xsc_txq_obj_new(struct xsc_dev *xdev, struct xsc_txq_data 
*txq_data,
txq_data->wqe_pi = 0;
txq_data->wqe_comp = 0;
 
-   PMD_DRV_LOG(INFO, "Create tx qp, wqe_s:%d, wqe_n:%d, qp_db=%p, qpn:%d",
+   PMD_DRV_LOG(INFO, "Create tx qp, wqe_s:%d, wqe_n:%d, qp_db=%p, qpn:%u",
txq_data->wqe_s, txq_data->wqe_n,
txq_data->qp_db, txq_data->qpn);
return 0;
diff --git a/drivers/net/xsc/xsc_vfio.c b/drivers/net/xsc/xsc_vfio.c
index e8fefb9e94..6c679449af 100644
--- a/drivers/net/xsc/xsc_vfio.c
+++ b/drivers/net/xsc/xsc_vfio.c
@@ -337,7 +337,7 @@ xsc_vfio_modify_qp_status(struct xsc_dev *xdev, uint32_t 
qpn, int num, int opcod
 
ret = xsc_vfio_mbox_exec(xdev, in, in_len, out, out_len);
if (ret != 0 || out->hdr.status != 0) {
-   PMD_DRV_LOG(ERR, "Modify qp status failed, qpn=%d, 
err=%d, out.status=%u",
+   PMD_DRV_LOG(ERR, "Modify qp status failed, qpn=%u, 
err=%d, out.status=%u",
qpn + i, ret, out->hdr.status);
rte_errno = ENOEXEC;
ret = -rte_errno;
@@ -483,7 +483,7 @@ xsc_vfio_rx_cq_create(struct xsc_dev *xdev, struct 
xsc_rx_cq_params *cq_params,
cq_info->cqn = rte_be_to_cpu_32(out->cqn);
cq->cqn = cq_info->cqn;
cq->xdev = xdev;
-   PMD_DRV_LOG(INFO, "Port id=%d, Rx cqe_n:%d, cqn:%d",
+   PMD_DRV_LOG(INFO, "Port id=%d, Rx cqe_n:%d, cqn:%u",
port_id, cq_info->cqe_n, cq_info->cqn);
 
free(in);
diff --git a/drivers/net/xsc/xsc_vfio_mbox.c b/drivers/net/xsc/xsc_vfio_mbox.c
index 454fe64645..c465679527 100644
--- a/drivers/net/xsc/xsc_vfio_mbox.c
+++ b/drivers/net/xsc/xsc_vfio_mbox.c
@@ -515,7 +515,7 @@ xsc_cmdq_req_status_restore(struct xsc_dev *xdev, struct 
xsc_cmd_queue *cmdq)
if (req_pid >= (uint32_t)(1 << cmdq->depth_n) ||
req_cid >= (uint32_t)(1 <

[PATCH 09/12] net/xsc: avoid assign the same value to a variable

2025-02-21 Thread Renyong Wan
This issue was reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v1048/

Signed-off-by: Rong Qian 
Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_np.c | 4 
 drivers/net/xsc/xsc_tx.c | 1 -
 2 files changed, 5 deletions(-)

diff --git a/drivers/net/xsc/xsc_np.c b/drivers/net/xsc/xsc_np.c
index eff7497255..7ff39f83c9 100644
--- a/drivers/net/xsc/xsc_np.c
+++ b/drivers/net/xsc/xsc_np.c
@@ -137,10 +137,6 @@ xsc_rss_hash_template_get(struct rte_eth_rss_conf 
*rss_conf)
rss_hf &= ~XSC_RSS_HASH_BIT_IPV6_SPORT;
}
 
-   if ((rss_conf->rss_hf & RTE_ETH_RSS_LEVEL_PMD_DEFAULT) ||
-   (rss_conf->rss_hf & RTE_ETH_RSS_LEVEL_OUTERMOST))
-   outer = 1;
-
if (rss_conf->rss_hf & RTE_ETH_RSS_LEVEL_INNERMOST)
outer = 0;
 
diff --git a/drivers/net/xsc/xsc_tx.c b/drivers/net/xsc/xsc_tx.c
index 0d6a9981d0..d1a0f3284b 100644
--- a/drivers/net/xsc/xsc_tx.c
+++ b/drivers/net/xsc/xsc_tx.c
@@ -215,7 +215,6 @@ xsc_tx_wqe_ctrl_seg_init(struct xsc_txq_data 
*__rte_restrict txq,
cs->csum_en = 0;
 
if (txq->tso_en == 1 && (mbuf->ol_flags & RTE_MBUF_F_TX_TCP_SEG)) {
-   cs->has_pph = 0;
cs->so_type = 1;
cs->so_hdr_len = mbuf->l2_len + mbuf->l3_len + mbuf->l4_len;
cs->so_data_size = rte_cpu_to_le_16(mbuf->tso_segsz);
-- 
2.25.1


[PATCH 04/12] net/xsc: remove always-true if expressions

2025-02-21 Thread Renyong Wan
This issue was reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v547/

Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_ethdev.c | 13 +
 1 file changed, 5 insertions(+), 8 deletions(-)

diff --git a/drivers/net/xsc/xsc_ethdev.c b/drivers/net/xsc/xsc_ethdev.c
index b9675a5218..0729d43884 100644
--- a/drivers/net/xsc/xsc_ethdev.c
+++ b/drivers/net/xsc/xsc_ethdev.c
@@ -721,16 +721,13 @@ xsc_ethdev_init_one_representor(struct rte_eth_dev 
*eth_dev, void *init_params)
config->tso = 0;
} else {
config->tso = 1;
-   if (config->tso)
-   config->tso_max_payload_sz = 1500;
+   config->tso_max_payload_sz = 1500;
}
 
-   priv->is_representor = (priv->eth_type == RTE_ETH_REPRESENTOR_NONE) ? 0 
: 1;
-   if (priv->is_representor) {
-   eth_dev->data->dev_flags |= RTE_ETH_DEV_REPRESENTOR;
-   eth_dev->data->representor_id = priv->representor_id;
-   eth_dev->data->backer_port_id = eth_dev->data->port_id;
-   }
+   priv->is_representor = 1;
+   eth_dev->data->dev_flags |= RTE_ETH_DEV_REPRESENTOR;
+   eth_dev->data->representor_id = priv->representor_id;
+   eth_dev->data->backer_port_id = eth_dev->data->port_id;
 
eth_dev->dev_ops = &xsc_eth_dev_ops;
eth_dev->rx_pkt_burst = rte_eth_pkt_burst_dummy;
-- 
2.25.1


[PATCH 10/12] net/xsc: avoid initialize by same function

2025-02-21 Thread Renyong Wan
This issue was reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v656

Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_rx.c | 6 ++
 1 file changed, 2 insertions(+), 4 deletions(-)

diff --git a/drivers/net/xsc/xsc_rx.c b/drivers/net/xsc/xsc_rx.c
index c031e56a94..ca832dab23 100644
--- a/drivers/net/xsc/xsc_rx.c
+++ b/drivers/net/xsc/xsc_rx.c
@@ -445,10 +445,8 @@ xsc_rxq_elts_alloc(struct xsc_rxq_data *rxq_data)
 
mbuf->port = rxq_data->port_id;
mbuf->nb_segs = 1;
-   rte_pktmbuf_data_len(mbuf) = 
rte_pktmbuf_data_room_size(rxq_data->mp)
-   - mbuf->data_off;
-   rte_pktmbuf_pkt_len(mbuf) = 
rte_pktmbuf_data_room_size(rxq_data->mp)
-   - mbuf->data_off;
+   rte_pktmbuf_data_len(mbuf) = mbuf->buf_len - mbuf->data_off;
+   rte_pktmbuf_pkt_len(mbuf) = rte_pktmbuf_data_len(mbuf);
(*rxq_data->elts)[i] = mbuf;
}
 
-- 
2.25.1


[PATCH 00/12] net/xsc: Resolve issues from PVS and Coverity Scan

2025-02-21 Thread Renyong Wan
This patch series resolves several issues reported by PVS and Coverity Scan,
which were earlier forwarded to us by Stephen Hemminger.

---
Renyong Wan (12):
  net/xsc: avoid integer overflow
  net/xsc: remove useless call
  net/xsc: address incorrect format warnings
  net/xsc: remove always-true if expressions
  net/xsc: avoid variable is assigned but not used
  net/xsc: check possible null pointer dereference
  net/xsc: avoid potential null pointer before used
  net/xsc: remove always-true part of if expression
  net/xsc: avoid assign the same value to a variable
  net/xsc: avoid initialize by same function
  net/xsc: optimize memcmp returns not 0 check
  net/xsc: avoid pointer cast to unrelated class

 drivers/net/xsc/xsc_dev.c   |  2 +-
 drivers/net/xsc/xsc_ethdev.c| 35 
 drivers/net/xsc/xsc_np.c| 17 +++---
 drivers/net/xsc/xsc_rx.c| 31 ++-
 drivers/net/xsc/xsc_tx.c|  7 +--
 drivers/net/xsc/xsc_vfio.c  | 97 -
 drivers/net/xsc/xsc_vfio_mbox.c |  2 +-
 7 files changed, 111 insertions(+), 80 deletions(-)

-- 
2.25.1


[PATCH 08/12] net/xsc: remove always-true part of if expression

2025-02-21 Thread Renyong Wan
This issue was reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v560/

Signed-off-by: Rong Qian 
Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_rx.c | 2 --
 1 file changed, 2 deletions(-)

diff --git a/drivers/net/xsc/xsc_rx.c b/drivers/net/xsc/xsc_rx.c
index dfb71ed087..c031e56a94 100644
--- a/drivers/net/xsc/xsc_rx.c
+++ b/drivers/net/xsc/xsc_rx.c
@@ -42,8 +42,6 @@ xsc_rx_poll_len(struct xsc_rxq_data *rxq, volatile struct 
xsc_cqe *cqe)
if (unlikely(ret != XSC_CQE_OWNER_SW)) {
if (unlikely(ret == XSC_CQE_OWNER_ERR)) {
++rxq->stats.rx_errors;
-   if (ret == XSC_CQE_OWNER_HW || ret == -1)
-   return 0;
} else {
return 0;
}
-- 
2.25.1


[PATCH 11/12] net/xsc: optimize memcmp returns not 0 check

2025-02-21 Thread Renyong Wan
This issue was reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v526/

Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_ethdev.c | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/drivers/net/xsc/xsc_ethdev.c b/drivers/net/xsc/xsc_ethdev.c
index 44d861e484..28af9c5d15 100644
--- a/drivers/net/xsc/xsc_ethdev.c
+++ b/drivers/net/xsc/xsc_ethdev.c
@@ -663,7 +663,7 @@ xsc_ethdev_mac_addr_add(struct rte_eth_dev *dev, struct 
rte_ether_addr *mac, uin
for (i = 0; i != XSC_MAX_MAC_ADDRESSES; ++i) {
if (i == (int)index)
continue;
-   if (memcmp(&dev->data->mac_addrs[i], mac, sizeof(*mac)))
+   if (memcmp(&dev->data->mac_addrs[i], mac, sizeof(*mac)) != 0)
continue;
/* Address already configured elsewhere, return with error */
rte_errno = EADDRINUSE;
-- 
2.25.1


[PATCH 07/12] net/xsc: avoid potential null pointer before used

2025-02-21 Thread Renyong Wan
This issue reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v595/

Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_np.c   | 2 +-
 drivers/net/xsc/xsc_rx.c   | 2 +-
 drivers/net/xsc/xsc_vfio.c | 2 +-
 3 files changed, 3 insertions(+), 3 deletions(-)

diff --git a/drivers/net/xsc/xsc_np.c b/drivers/net/xsc/xsc_np.c
index f032a0dbc2..eff7497255 100644
--- a/drivers/net/xsc/xsc_np.c
+++ b/drivers/net/xsc/xsc_np.c
@@ -172,13 +172,13 @@ xsc_dev_np_exec(struct xsc_dev *xdev, void *cmd, int len, 
int table, int opmod)
out_len = sizeof(struct xsc_np_mbox_out) + data_len;
cmd_len = RTE_MAX(in_len, out_len);
in = malloc(cmd_len);
-   memset(in, 0, cmd_len);
if (in == NULL) {
rte_errno = ENOMEM;
PMD_DRV_LOG(ERR, "Failed to alloc np cmd memory");
return -rte_errno;
}
 
+   memset(in, 0, cmd_len);
in->hdr.opcode = rte_cpu_to_be_16(XSC_CMD_OP_EXEC_NP);
in->len = rte_cpu_to_be_16(data_len);
 
diff --git a/drivers/net/xsc/xsc_rx.c b/drivers/net/xsc/xsc_rx.c
index 140d7728d6..dfb71ed087 100644
--- a/drivers/net/xsc/xsc_rx.c
+++ b/drivers/net/xsc/xsc_rx.c
@@ -296,13 +296,13 @@ xsc_rss_qp_create(struct xsc_ethdev_priv *priv, int 
port_id)
out_len = sizeof(struct xsc_cmd_create_multiqp_mbox_out) + 
entry_total_len;
cmd_len = RTE_MAX(in_len, out_len);
in = malloc(cmd_len);
-   memset(in, 0, cmd_len);
if (in == NULL) {
rte_errno = ENOMEM;
PMD_DRV_LOG(ERR, "Alloc rss qp create cmd memory failed");
goto error;
}
 
+   memset(in, 0, cmd_len);
in->qp_num = rte_cpu_to_be_16((uint16_t)priv->num_rq);
in->qp_type = XSC_QUEUE_TYPE_RAW;
in->req_len = rte_cpu_to_be_32(cmd_len);
diff --git a/drivers/net/xsc/xsc_vfio.c b/drivers/net/xsc/xsc_vfio.c
index 6c679449af..2cd796ba26 100644
--- a/drivers/net/xsc/xsc_vfio.c
+++ b/drivers/net/xsc/xsc_vfio.c
@@ -431,13 +431,13 @@ xsc_vfio_rx_cq_create(struct xsc_dev *xdev, struct 
xsc_rx_cq_params *cq_params,
}
 
in = malloc(cmd_len);
-   memset(in, 0, cmd_len);
if (in == NULL) {
rte_errno = ENOMEM;
PMD_DRV_LOG(ERR, "Failed to alloc rx cq exec cmd memory");
goto error;
}
 
+   memset(in, 0, cmd_len);
in->hdr.opcode = rte_cpu_to_be_16(XSC_CMD_OP_CREATE_CQ);
in->ctx.eqn = 0;
in->ctx.pa_num = rte_cpu_to_be_16(pa_num);
-- 
2.25.1


[PATCH 12/12] net/xsc: avoid pointer cast to unrelated class

2025-02-21 Thread Renyong Wan
This issue was reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v1027/

Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_np.c   | 10 +++--
 drivers/net/xsc/xsc_rx.c   | 12 ++---
 drivers/net/xsc/xsc_vfio.c | 91 +++---
 3 files changed, 68 insertions(+), 45 deletions(-)

diff --git a/drivers/net/xsc/xsc_np.c b/drivers/net/xsc/xsc_np.c
index 7ff39f83c9..f96797b7a4 100644
--- a/drivers/net/xsc/xsc_np.c
+++ b/drivers/net/xsc/xsc_np.c
@@ -162,18 +162,20 @@ xsc_dev_np_exec(struct xsc_dev *xdev, void *cmd, int len, 
int table, int opmod)
int data_len;
int cmd_len;
int ret;
+   void *cmd_buf;
 
data_len = sizeof(struct xsc_np_data_tl) + len;
in_len = sizeof(struct xsc_np_mbox_in) + data_len;
out_len = sizeof(struct xsc_np_mbox_out) + data_len;
cmd_len = RTE_MAX(in_len, out_len);
-   in = malloc(cmd_len);
-   if (in == NULL) {
+   cmd_buf = malloc(cmd_len);
+   if (cmd_buf == NULL) {
rte_errno = ENOMEM;
PMD_DRV_LOG(ERR, "Failed to alloc np cmd memory");
return -rte_errno;
}
 
+   in = cmd_buf;
memset(in, 0, cmd_len);
in->hdr.opcode = rte_cpu_to_be_16(XSC_CMD_OP_EXEC_NP);
in->len = rte_cpu_to_be_16(data_len);
@@ -185,10 +187,10 @@ xsc_dev_np_exec(struct xsc_dev *xdev, void *cmd, int len, 
int table, int opmod)
if (cmd && len)
memcpy(tl + 1, cmd, len);
 
-   out = (struct xsc_np_mbox_out *)in;
+   out = cmd_buf;
ret = xsc_dev_mailbox_exec(xdev, in, in_len, out, out_len);
 
-   free(in);
+   free(cmd_buf);
return ret;
 }
 
diff --git a/drivers/net/xsc/xsc_rx.c b/drivers/net/xsc/xsc_rx.c
index ca832dab23..a702b9592b 100644
--- a/drivers/net/xsc/xsc_rx.c
+++ b/drivers/net/xsc/xsc_rx.c
@@ -276,6 +276,7 @@ xsc_rss_qp_create(struct xsc_ethdev_priv *priv, int port_id)
struct xsc_dev *xdev = priv->xdev;
struct xsc_hwinfo *hwinfo = &xdev->hwinfo;
char name[RTE_ETH_NAME_MAX_LEN] = { 0 };
+   void *cmd_buf;
 
rxq_data = xsc_rxq_get(priv, 0);
if (rxq_data == NULL)
@@ -293,13 +294,14 @@ xsc_rss_qp_create(struct xsc_ethdev_priv *priv, int 
port_id)
in_len = sizeof(struct xsc_cmd_create_multiqp_mbox_in) + 
entry_total_len;
out_len = sizeof(struct xsc_cmd_create_multiqp_mbox_out) + 
entry_total_len;
cmd_len = RTE_MAX(in_len, out_len);
-   in = malloc(cmd_len);
-   if (in == NULL) {
+   cmd_buf = malloc(cmd_len);
+   if (cmd_buf == NULL) {
rte_errno = ENOMEM;
PMD_DRV_LOG(ERR, "Alloc rss qp create cmd memory failed");
goto error;
}
 
+   in = cmd_buf;
memset(in, 0, cmd_len);
in->qp_num = rte_cpu_to_be_16((uint16_t)priv->num_rq);
in->qp_type = XSC_QUEUE_TYPE_RAW;
@@ -333,7 +335,7 @@ xsc_rss_qp_create(struct xsc_ethdev_priv *priv, int port_id)
}
 
in->hdr.opcode = rte_cpu_to_be_16(XSC_CMD_OP_CREATE_MULTI_QP);
-   out = (struct xsc_cmd_create_multiqp_mbox_out *)in;
+   out = cmd_buf;
ret = xsc_dev_mailbox_exec(xdev, in, in_len, out, out_len);
if (ret != 0 || out->hdr.status != 0) {
PMD_DRV_LOG(ERR,
@@ -365,11 +367,11 @@ xsc_rss_qp_create(struct xsc_ethdev_priv *priv, int 
port_id)
rxq_data->rq_db, rxq_data->qpn);
}
 
-   free(in);
+   free(cmd_buf);
return 0;
 
 error:
-   free(in);
+   free(cmd_buf);
return -rte_errno;
 }
 
diff --git a/drivers/net/xsc/xsc_vfio.c b/drivers/net/xsc/xsc_vfio.c
index 2cd796ba26..8330483523 100644
--- a/drivers/net/xsc/xsc_vfio.c
+++ b/drivers/net/xsc/xsc_vfio.c
@@ -70,22 +70,24 @@ xsc_vfio_hwinfo_init(struct xsc_dev *xdev)
struct xsc_cmd_query_hca_cap_mbox_in *in;
struct xsc_cmd_query_hca_cap_mbox_out *out;
struct xsc_cmd_hca_cap *hca_cap;
+   void *cmd_buf;
 
in_len = sizeof(struct xsc_cmd_query_hca_cap_mbox_in);
out_len = sizeof(struct xsc_cmd_query_hca_cap_mbox_out);
cmd_len = RTE_MAX(in_len, out_len);
 
-   in = malloc(cmd_len);
-   if (in == NULL) {
+   cmd_buf = malloc(cmd_len);
+   if (cmd_buf == NULL) {
PMD_DRV_LOG(ERR, "Failed to alloc dev hwinfo cmd memory");
rte_errno = ENOMEM;
return -rte_errno;
}
 
+   in = cmd_buf;
memset(in, 0, cmd_len);
in->hdr.opcode = rte_cpu_to_be_16(XSC_CMD_OP_QUERY_HCA_CAP);
in->hdr.ver = rte_cpu_to_be_16(XSC_CMD_QUERY_HCA_CAP_V1);
-   out = (struct xsc_cmd_query_hca_cap_mbox_out *)in;
+   out = cmd_buf;
 
ret = xsc_vfio_mbox_exec(xdev, in, in_len, out, out_len);
if (ret != 0 || out->hdr.status != 0) {
@@ -126,7 +128,7 @@ xsc_vfio_hwinfo_init(struct xsc_dev *xdev)
xsc_vfio_pcie_no_init(&xdev->hwinfo);
 
 exit:
-   free(in)

[PATCH 06/12] net/xsc: check possible null pointer dereference

2025-02-21 Thread Renyong Wan
This issue was reported by PVS studio, described as:
https://pvs-studio.com/en/docs/warnings/v522/

Signed-off-by: Rong Qian 
Signed-off-by: Renyong Wan 
---
 drivers/net/xsc/xsc_ethdev.c | 15 +--
 drivers/net/xsc/xsc_rx.c |  5 +
 2 files changed, 18 insertions(+), 2 deletions(-)

diff --git a/drivers/net/xsc/xsc_ethdev.c b/drivers/net/xsc/xsc_ethdev.c
index d04fe78d3f..44d861e484 100644
--- a/drivers/net/xsc/xsc_ethdev.c
+++ b/drivers/net/xsc/xsc_ethdev.c
@@ -137,8 +137,8 @@ xsc_ethdev_enable(struct rte_eth_dev *dev)
int pcie_logic_port = 0;
int qp_set_id;
int repr_id;
-   struct xsc_rxq_data *rxq = xsc_rxq_get(priv, 0);
-   uint16_t rx_qpn = (uint16_t)rxq->qpn;
+   struct xsc_rxq_data *rxq;
+   uint16_t rx_qpn;
int i, vld;
struct xsc_txq_data *txq;
struct xsc_repr_port *repr;
@@ -147,6 +147,11 @@ xsc_ethdev_enable(struct rte_eth_dev *dev)
if (priv->funcid_type != XSC_PHYPORT_MAC_FUNCID)
return -ENODEV;
 
+   rxq = xsc_rxq_get(priv, 0);
+   if (rxq == NULL)
+   return -EINVAL;
+
+   rx_qpn = (uint16_t)rxq->qpn;
hwinfo = &priv->xdev->hwinfo;
repr_id = priv->representor_id;
repr = &priv->xdev->repr_ports[repr_id];
@@ -162,6 +167,8 @@ xsc_ethdev_enable(struct rte_eth_dev *dev)
 
for (i = 0; i < priv->num_sq; i++) {
txq = xsc_txq_get(priv, i);
+   if (txq == NULL)
+   return -EINVAL;
xsc_dev_modify_qp_status(priv->xdev, txq->qpn, 1, 
XSC_CMD_OP_RTR2RTS_QP);
xsc_dev_modify_qp_qostree(priv->xdev, txq->qpn);
xsc_dev_set_qpsetid(priv->xdev, txq->qpn, qp_set_id);
@@ -229,6 +236,8 @@ xsc_txq_start(struct xsc_ethdev_priv *priv)
 
for (i = 0; i != priv->num_sq; ++i) {
txq_data = xsc_txq_get(priv, i);
+   if (txq_data == NULL)
+   goto error;
xsc_txq_elts_alloc(txq_data);
ret = xsc_txq_obj_new(priv->xdev, txq_data, offloads, i);
if (ret < 0)
@@ -270,6 +279,8 @@ xsc_rxq_start(struct xsc_ethdev_priv *priv)
 
for (i = 0; i != priv->num_rq; ++i) {
rxq_data = xsc_rxq_get(priv, i);
+   if (rxq_data == NULL)
+   goto error;
if (dev->data->rx_queue_state[i] != 
RTE_ETH_QUEUE_STATE_STARTED) {
ret = xsc_rxq_elts_alloc(rxq_data);
if (ret != 0)
diff --git a/drivers/net/xsc/xsc_rx.c b/drivers/net/xsc/xsc_rx.c
index 3230710656..140d7728d6 100644
--- a/drivers/net/xsc/xsc_rx.c
+++ b/drivers/net/xsc/xsc_rx.c
@@ -280,6 +280,9 @@ xsc_rss_qp_create(struct xsc_ethdev_priv *priv, int port_id)
char name[RTE_ETH_NAME_MAX_LEN] = { 0 };
 
rxq_data = xsc_rxq_get(priv, 0);
+   if (rxq_data == NULL)
+   return -EINVAL;
+
log_ele = rte_log2_u32(sizeof(struct xsc_wqe_data_seg));
wqe_n = rxq_data->wqe_s;
log_rq_sz = rte_log2_u32(wqe_n * hwinfo->recv_seg_num);
@@ -385,6 +388,8 @@ xsc_rxq_rss_obj_new(struct xsc_ethdev_priv *priv, uint16_t 
port_id)
/* Create CQ */
for (i = 0; i < priv->num_rq; ++i) {
rxq_data = xsc_rxq_get(priv, i);
+   if (rxq_data == NULL)
+   return -EINVAL;
 
memset(&cq_params, 0, sizeof(cq_params));
memset(&cq_info, 0, sizeof(cq_info));
-- 
2.25.1


Re: [PATCH v2 00/16] net/zxdh: updated net zxdh driver

2025-02-21 Thread Stephen Hemminger
On Fri, 21 Feb 2025 10:03:28 +0800
Junlong Wang  wrote:

> V2:
>   - modify CI some error results(checkpatches warnings、Wrong headline format)
>   - fix warnings when enable extra warnings.
>   - modify apply memcpy script for coccinelle and unnecessary init and
> unneccessary cast of void when use malloc.
> 
> V1:
>   - updated net zxdh driver.
> optimize init and some ops.
> provided csum/lro/tso 、extend stats、fw_version、module_info 、meter, etc.
> 
> Junlong Wang (16):
>   net/zxdh: optimize np dtb channel initialization
>   net/zxdh: optimize queue res alloc/free process
>   net/zxdh: optimize link update process
>   net/zxdh: update Rx/Tx to latest
>   net/zxdh: provided PF/VF msg intr callback
>   net/zxdh: optimize MAC ops
>   net/zxdh: optimize promisc ops
>   net/zxdh: optimize VLAN filter/offload ops
>   net/zxdh: optimize RSS/RETA hash config/update/get
>   net/zxdh: optimize MTU set ops
>   net/zxdh: optimize basic stats ops
>   net/zxdh: provided CSUM/TSO/LRO config
>   net/zxdh: provided rxq/txq info get implementations
>   net/zxdh: provide extended stats ops implementations
>   net/zxdh: provide ptypes FW version EEPROM ops
>   net/zxdh: provide meter ops implementations
> 
>  doc/guides/nics/features/zxdh.ini  |   11 +
>  doc/guides/nics/zxdh.rst   |5 +
>  drivers/net/zxdh/meson.build   |1 +
>  drivers/net/zxdh/zxdh_common.c |   46 +-
>  drivers/net/zxdh/zxdh_common.h |3 +
>  drivers/net/zxdh/zxdh_ethdev.c |  724 +---
>  drivers/net/zxdh/zxdh_ethdev.h |   69 +-
>  drivers/net/zxdh/zxdh_ethdev_ops.c |  916 ++---
>  drivers/net/zxdh/zxdh_ethdev_ops.h |   53 +-
>  drivers/net/zxdh/zxdh_msg.c|  985 +-
>  drivers/net/zxdh/zxdh_msg.h|  175 +++-
>  drivers/net/zxdh/zxdh_mtr.c| 1223 
>  drivers/net/zxdh/zxdh_mtr.h|  114 +++
>  drivers/net/zxdh/zxdh_np.c |  811 --
>  drivers/net/zxdh/zxdh_np.h |  262 ++
>  drivers/net/zxdh/zxdh_pci.c|   10 -
>  drivers/net/zxdh/zxdh_queue.c  |  132 +--
>  drivers/net/zxdh/zxdh_queue.h  |  118 +--
>  drivers/net/zxdh/zxdh_rxtx.c   |  695 +---
>  drivers/net/zxdh/zxdh_rxtx.h   |   27 +
>  drivers/net/zxdh/zxdh_tables.c |  378 +++--
>  drivers/net/zxdh/zxdh_tables.h |  220 +++--
>  22 files changed, 6040 insertions(+), 938 deletions(-)
>  create mode 100644 drivers/net/zxdh/zxdh_mtr.c
>  create mode 100644 drivers/net/zxdh/zxdh_mtr.h

Still several checklist items, some of these are likely pre-existing conditions.
The one that matters are the build issue with RTE_ASSERT enabled.

Mark items with:
✔ passed
✘ Failed
o N/A

Basic hygiene
✔ Look at CI results in patchwork
✔ Merge cleanly with git am
✔ Run checkpatches
  - bug in checkpatch around __rte_packed_begin macros
  - warning from stub

✔ Run check-git-log
✔ Run check-symbol-maps.sh
✔ Run check-doc-vs-code
✔ Run check-spdk-tag

Builds
✔ Normal Gcc build; make sure driver is built!
✔ Use latest experimental Gcc 15 to catch new warnings
✔ Clang build using current version (clang-19)
✔ Doc build
o Build for 32 bit x86
o Cross build for Windows (if applicable)
✔ Debug build
x Enable asserts
[1594/3244] Compiling C object 
drivers/libtmp_rte_net_zxdh.a.p/net_zxdh_zxdh_msg.c.o
FAILED: drivers/libtmp_rte_net_zxdh.a.p/net_zxdh_zxdh_msg.c.o
cc -Idrivers/libtmp_rte_net_zxdh.a.p -Idrivers -I../drivers -Idrivers/net/zxdh 
-I../drivers/net/zxdh -Ilib/ethdev -I../lib/ethdev -I. -I.. -Iconfig 
-I../config -Ilib/eal/include -I../lib/eal/include -Ilib/eal/linux/include 
-I../lib/eal/linux/include -Ilib/eal/x86/include -I../lib/eal/x86/include 
-I../kernel/linux -Ilib/eal/common -I../lib/eal/common -Ilib/eal -I../lib/eal 
-Ilib/kvargs -I../lib/kvargs -Ilib/log -I../lib/log -Ilib/metrics 
-I../lib/metrics -Ilib/telemetry -I../lib/telemetry -Ilib/net -I../lib/net 
-Ilib/mbuf -I../lib/mbuf -Ilib/mempool -I../lib/mempool -Ilib/ring 
-I../lib/ring -Ilib/meter -I../lib/meter -Idrivers/bus/pci -I../drivers/bus/pci 
-I../drivers/bus/pci/linux -Ilib/pci -I../lib/pci -Idrivers/bus/vdev 
-I../drivers/bus/vdev -I/usr/include/x86_64-linux-gnu 
-fdiagnostics-color=always -D_FILE_OFFSET_BITS=64 -Wall -Winvalid-pch -Wextra 
-std=c11 -O3 -include rte_config.h -Wvla -Wcast-qual -Wdeprecated -Wformat 
-Wformat-nonliteral -Wformat-security -Wmissing-declarations 
-Wmissing-prototypes -Wnested-externs -Wold-style-definition -Wpointer-arith 
-Wsign-compare -Wstrict-prototypes -Wundef -Wwrite-strings 
-Wno-packed-not-aligned -Wno-missing-field-initializers -D_GNU_SOURCE 
-DRTE_ENABLE_ASSERT -DRTE_LIBRTE_ETHDEV_DEBUG=1 -fPIC -march=native -mrtm 
-DALLOW_EXPERIMENTAL_API -DALLOW_INTERNAL_API -Wno-format-truncation 
-Wno-address-of-packed-member -Wno-vla -DRTE_LOG_DEFAULT_LOGTYPE=pmd.net.zxdh 
-MD -MQ drivers/libtmp_rte_net_zxdh.a.p/net_zxd

[v3 3/6] crypto/virtio: add packed ring support

2025-02-21 Thread Gowrishankar Muthukrishnan
Add packed ring support.

Signed-off-by: Gowrishankar Muthukrishnan 
---
 drivers/crypto/virtio/virtio_cryptodev.c | 125 +++
 drivers/crypto/virtio/virtio_cryptodev.h |  13 +-
 drivers/crypto/virtio/virtio_cvq.c   | 103 +-
 drivers/crypto/virtio/virtio_pci.h   |  25 ++
 drivers/crypto/virtio/virtio_ring.h  |  59 ++-
 drivers/crypto/virtio/virtio_rxtx.c  | 444 ++-
 drivers/crypto/virtio/virtqueue.c|  50 ++-
 drivers/crypto/virtio/virtqueue.h| 134 ++-
 8 files changed, 924 insertions(+), 29 deletions(-)

diff --git a/drivers/crypto/virtio/virtio_cryptodev.c 
b/drivers/crypto/virtio/virtio_cryptodev.c
index 6bb76ff15e..92fea557ab 100644
--- a/drivers/crypto/virtio/virtio_cryptodev.c
+++ b/drivers/crypto/virtio/virtio_cryptodev.c
@@ -869,6 +869,125 @@ virtio_crypto_clear_session(
rte_free(ctrl);
 }
 
+static void
+virtio_crypto_clear_session_packed(
+   struct rte_cryptodev *dev,
+   struct virtio_crypto_op_ctrl_req *ctrl)
+{
+   struct virtio_crypto_hw *hw;
+   struct virtqueue *vq;
+   struct vring_packed_desc *desc;
+   uint8_t *status;
+   uint8_t needed = 1;
+   uint32_t head;
+   uint64_t malloc_phys_addr;
+   uint8_t len_inhdr = sizeof(struct virtio_crypto_inhdr);
+   uint32_t len_op_ctrl_req = sizeof(struct virtio_crypto_op_ctrl_req);
+   uint64_t session_id = ctrl->u.destroy_session.session_id;
+   uint16_t flags;
+   uint8_t nb_descs = 0;
+
+   hw = dev->data->dev_private;
+   vq = virtcrypto_cq_to_vq(hw->cvq);
+   head = vq->vq_avail_idx;
+   flags = vq->vq_packed.cached_flags;
+
+   VIRTIO_CRYPTO_SESSION_LOG_INFO("vq->vq_desc_head_idx = %d, "
+   "vq = %p", vq->vq_desc_head_idx, vq);
+
+   if (vq->vq_free_cnt < needed) {
+   VIRTIO_CRYPTO_SESSION_LOG_ERR(
+   "vq->vq_free_cnt = %d is less than %d, "
+   "not enough", vq->vq_free_cnt, needed);
+   return;
+   }
+
+   malloc_phys_addr = rte_malloc_virt2iova(ctrl);
+
+   /* status part */
+   status = &(((struct virtio_crypto_inhdr *)
+   ((uint8_t *)ctrl + len_op_ctrl_req))->status);
+   *status = VIRTIO_CRYPTO_ERR;
+
+   /* indirect desc vring part */
+   desc = vq->vq_packed.ring.desc;
+
+   /* ctrl request part */
+   desc[head].addr = malloc_phys_addr;
+   desc[head].len = len_op_ctrl_req;
+   desc[head].flags = VRING_DESC_F_NEXT | vq->vq_packed.cached_flags;
+   vq->vq_free_cnt--;
+   nb_descs++;
+   if (++vq->vq_avail_idx >= vq->vq_nentries) {
+   vq->vq_avail_idx -= vq->vq_nentries;
+   vq->vq_packed.cached_flags ^= VRING_PACKED_DESC_F_AVAIL_USED;
+   }
+
+   /* status part */
+   desc[vq->vq_avail_idx].addr = malloc_phys_addr + len_op_ctrl_req;
+   desc[vq->vq_avail_idx].len = len_inhdr;
+   desc[vq->vq_avail_idx].flags = VRING_DESC_F_WRITE;
+   vq->vq_free_cnt--;
+   nb_descs++;
+   if (++vq->vq_avail_idx >= vq->vq_nentries) {
+   vq->vq_avail_idx -= vq->vq_nentries;
+   vq->vq_packed.cached_flags ^= VRING_PACKED_DESC_F_AVAIL_USED;
+   }
+
+   virtqueue_store_flags_packed(&desc[head], VRING_DESC_F_NEXT | flags,
+   vq->hw->weak_barriers);
+
+   virtio_wmb(vq->hw->weak_barriers);
+   virtqueue_notify(vq);
+
+   /* wait for used desc in virtqueue
+* desc_is_used has a load-acquire or rte_io_rmb inside
+*/
+   rte_rmb();
+   while (!desc_is_used(&desc[head], vq)) {
+   rte_rmb();
+   usleep(100);
+   }
+
+   /* now get used descriptors */
+   vq->vq_free_cnt += nb_descs;
+   vq->vq_used_cons_idx += nb_descs;
+   if (vq->vq_used_cons_idx >= vq->vq_nentries) {
+   vq->vq_used_cons_idx -= vq->vq_nentries;
+   vq->vq_packed.used_wrap_counter ^= 1;
+   }
+
+   PMD_INIT_LOG(DEBUG, "vq->vq_free_cnt=%d "
+   "vq->vq_queue_idx=%d "
+   "vq->vq_avail_idx=%d "
+   "vq->vq_used_cons_idx=%d "
+   "vq->vq_packed.cached_flags=0x%x "
+   "vq->vq_packed.used_wrap_counter=%d",
+   vq->vq_free_cnt,
+   vq->vq_queue_index,
+   vq->vq_avail_idx,
+   vq->vq_used_cons_idx,
+   vq->vq_packed.cached_flags,
+   vq->vq_packed.used_wrap_counter);
+
+   if (*status != VIRTIO_CRYPTO_OK) {
+   VIRTIO_CRYPTO_SESSION_LOG_ERR("Close session failed "
+   "status=%"PRIu32", session_id=%"PRIu64"",
+   *status, session_id);
+   rte_free(ctrl);
+   return;
+   }
+
+   VIRTIO_CRYPTO_INIT_LOG_DBG("vq->vq_free_cnt=%d "
+  

RE: [EXTERNAL] [PATCH v8] graph: mcore: optimize graph search

2025-02-21 Thread Kiran Kumar Kokkilagadda



> -Original Message-
> From: Huichao Cai 
> Sent: Friday, February 7, 2025 7:10 AM
> To: Jerin Jacob ; Kiran Kumar Kokkilagadda
> ; Nithin Kumar Dabilpuram
> ; yanzhirun_...@163.com
> Cc: dev@dpdk.org
> Subject: [EXTERNAL] [PATCH v8] graph: mcore: optimize graph search
> 
> In the function __rte_graph_mcore_dispatch_sched_node_enqueue, use a
> slower loop to search for the graph, modify the search logic to record the
> result of the first search, and use this record for subsequent searches to
> improve search speed. 
> In the function __rte_graph_mcore_dispatch_sched_node_enqueue,
> use a slower loop to search for the graph, modify the search logic to record 
> the
> result of the first search, and use this record for subsequent searches to
> improve search speed.
> 
> Signed-off-by: Huichao Cai 
> ---

Acked-by: Kiran Kumar Kokkilagadda 



>  devtools/libabigail.abignore   |  5 +
>  doc/guides/rel_notes/release_25_03.rst |  1 +
>  lib/graph/rte_graph_model_mcore_dispatch.c | 11 +++
>  lib/graph/rte_graph_worker_common.h|  1 +
>  4 files changed, 14 insertions(+), 4 deletions(-)
> 
> diff --git a/devtools/libabigail.abignore b/devtools/libabigail.abignore index
> 21b8cd6113..8876aaee2e 100644
> --- a/devtools/libabigail.abignore
> +++ b/devtools/libabigail.abignore
> @@ -33,3 +33,8 @@
>  
>  ; Temporary exceptions till next major ABI version ;
> 
> +[suppress_type]
> +name = rte_node
> +has_size_change = no
> +has_data_member_inserted_between =
> +{offset_after(original_process), offset_of(xstat_off)}
> \ No newline at end of file
> diff --git a/doc/guides/rel_notes/release_25_03.rst
> b/doc/guides/rel_notes/release_25_03.rst
> index 269ab6f68a..16a888fd19 100644
> --- a/doc/guides/rel_notes/release_25_03.rst
> +++ b/doc/guides/rel_notes/release_25_03.rst
> @@ -150,6 +150,7 @@ ABI Changes
> 
>  * No ABI change that would break compatibility with 24.11.
> 
> +* graph: Added ``graph`` field to the ``dispatch`` structure in the 
> ``rte_node``
> structure.
> 
>  Known Issues
>  
> diff --git a/lib/graph/rte_graph_model_mcore_dispatch.c
> b/lib/graph/rte_graph_model_mcore_dispatch.c
> index a590fc9497..a81d338227 100644
> --- a/lib/graph/rte_graph_model_mcore_dispatch.c
> +++ b/lib/graph/rte_graph_model_mcore_dispatch.c
> @@ -118,11 +118,14 @@
> __rte_graph_mcore_dispatch_sched_node_enqueue(struct rte_node *node,
> struct rte_graph_rq_head *rq)  {
>   const unsigned int lcore_id = node->dispatch.lcore_id;
> - struct rte_graph *graph;
> + struct rte_graph *graph = node->dispatch.graph;
> 
> - SLIST_FOREACH(graph, rq, next)
> - if (graph->dispatch.lcore_id == lcore_id)
> - break;
> + if (unlikely((!graph) || (graph->dispatch.lcore_id != lcore_id))) {
> + SLIST_FOREACH(graph, rq, next)
> + if (graph->dispatch.lcore_id == lcore_id)
> + break;
> + node->dispatch.graph = graph;
> + }
> 
>   return graph != NULL ? __graph_sched_node_enqueue(node, graph) :
> false;  } diff --git a/lib/graph/rte_graph_worker_common.h
> b/lib/graph/rte_graph_worker_common.h
> index d3ec88519d..aef0f65673 100644
> --- a/lib/graph/rte_graph_worker_common.h
> +++ b/lib/graph/rte_graph_worker_common.h
> @@ -110,6 +110,7 @@ struct __rte_cache_aligned rte_node {
>   unsigned int lcore_id;  /**< Node running lcore. */
>   uint64_t total_sched_objs; /**< Number of objects
> scheduled. */
>   uint64_t total_sched_fail; /**< Number of scheduled
> failure. */
> + struct rte_graph *graph;  /**< Graph corresponding to
> lcore_id. */
>   } dispatch;
>   };
> 
> --
> 2.33.0



[PATCH v2 04/14] net/zxdh: modify dtb queue ops

2025-02-21 Thread Bingbin Chen
Modify the implementation of the dtb queue
request and release interfaces,
and add the implementation of queue initialization.

Signed-off-by: Bingbin Chen 
---
 drivers/net/zxdh/zxdh_np.c | 495 +
 drivers/net/zxdh/zxdh_np.h |  97 
 2 files changed, 485 insertions(+), 107 deletions(-)

diff --git a/drivers/net/zxdh/zxdh_np.c b/drivers/net/zxdh/zxdh_np.c
index cf5a192a02..eb57d61898 100644
--- a/drivers/net/zxdh/zxdh_np.c
+++ b/drivers/net/zxdh/zxdh_np.c
@@ -17,16 +17,15 @@
 
 static ZXDH_DEV_MGR_T g_dev_mgr;
 static ZXDH_SDT_MGR_T g_sdt_mgr;
-static uint32_t g_dpp_dtb_int_enable;
 static uint32_t g_table_type[ZXDH_DEV_CHANNEL_MAX][ZXDH_DEV_SDT_ID_MAX];
 static ZXDH_PPU_CLS_BITMAP_T g_ppu_cls_bit_map[ZXDH_DEV_CHANNEL_MAX];
 static ZXDH_DTB_MGR_T *p_dpp_dtb_mgr[ZXDH_DEV_CHANNEL_MAX];
-static ZXDH_RISCV_DTB_MGR *p_riscv_dtb_queue_mgr[ZXDH_DEV_CHANNEL_MAX];
 static ZXDH_SDT_TBL_DATA_T 
g_sdt_info[ZXDH_DEV_CHANNEL_MAX][ZXDH_DEV_SDT_ID_MAX];
 static ZXDH_PPU_STAT_CFG_T g_ppu_stat_cfg[ZXDH_DEV_CHANNEL_MAX];
 static uint64_t g_np_fw_compat_addr[ZXDH_DEV_CHANNEL_MAX];
 static const ZXDH_VERSION_COMPATIBLE_REG_T g_np_sdk_version = {
ZXDH_NPSDK_COMPAT_ITEM_ID, 1, 0, 0, 0, {0} };
+static const uint32_t hardware_ep_id[5] = {5, 6, 7, 8, 9};
 
 static ZXDH_FIELD_T g_smmu0_smmu0_cpu_ind_cmd_reg[] = {
{"cpu_ind_rw", ZXDH_FIELD_FLAG_RW, 31, 1, 0x0, 0x0},
@@ -436,6 +435,25 @@ zxdh_np_dev_opr_spinlock_get(uint32_t dev_id, uint32_t 
type, ZXDH_SPINLOCK_T **p
return ZXDH_OK;
 }
 
+static uint32_t
+zxdh_np_dev_dtb_opr_spinlock_get(uint32_t dev_id, uint32_t type,
+   uint32_t index, ZXDH_SPINLOCK_T **p_spinlock_out)
+{
+   ZXDH_DEV_MGR_T *p_dev_mgr = &g_dev_mgr;
+   ZXDH_DEV_CFG_T *p_dev_info = p_dev_mgr->p_dev_array[dev_id];
+
+   switch (type) {
+   case ZXDH_DEV_SPINLOCK_T_DTB:
+   *p_spinlock_out = &p_dev_info->dtb_queue_spinlock[index];
+   break;
+   default:
+   PMD_DRV_LOG(ERR, "spinlock type is invalid!");
+   return ZXDH_ERR;
+   }
+
+   return ZXDH_OK;
+}
+
 static uint32_t
 zxdh_np_dev_read_channel(uint32_t dev_id, uint32_t addr, uint32_t size, 
uint32_t *p_data)
 {
@@ -767,6 +785,7 @@ zxdh_np_dev_add(uint32_t  dev_id, ZXDH_DEV_TYPE_E dev_type,
 {
ZXDH_DEV_CFG_T *p_dev_info = NULL;
ZXDH_DEV_MGR_T *p_dev_mgr  = NULL;
+   uint32_t i = 0;
 
p_dev_mgr = &g_dev_mgr;
if (!p_dev_mgr->is_init) {
@@ -802,6 +821,10 @@ zxdh_np_dev_add(uint32_t  dev_id, ZXDH_DEV_TYPE_E dev_type,
p_dev_info->p_pcie_read_fun  = zxdh_np_dev_pcie_default_read;
 
zxdh_np_comm_spinlock_create(&p_dev_info->dtb_spinlock);
+
+   for (i = 0; i < ZXDH_DTB_QUEUE_NUM_MAX; i++)
+   
zxdh_np_comm_spinlock_create(&p_dev_info->dtb_queue_spinlock[i]);
+
return 0;
 }
 
@@ -1156,6 +1179,89 @@ zxdh_np_agent_channel_reg_write(uint32_t dev_id,
return ret;
 }
 
+static uint32_t
+zxdh_np_agent_channel_dtb_sync_send(uint32_t dev_id,
+   
ZXDH_AGENT_CHANNEL_DTB_MSG_T *p_msg,
+   uint32_t *p_data,
+   uint32_t rep_len)
+{
+   uint32_t ret = ZXDH_OK;
+
+   ZXDH_AGENT_CHANNEL_MSG_T agent_msg = {0};
+   agent_msg.msg = (void *)p_msg;
+   agent_msg.msg_len = sizeof(ZXDH_AGENT_CHANNEL_DTB_MSG_T);
+
+   ret = zxdh_np_agent_channel_sync_send(dev_id, &agent_msg, p_data, 
rep_len);
+   if (ret != ZXDH_OK) {
+   PMD_DRV_LOG(ERR, "zxdh_np_agent_channel_sync_send failed");
+   return ZXDH_ERR;
+   }
+
+   return ZXDH_OK;
+}
+
+static uint32_t
+zxdh_np_agent_channel_dtb_queue_request(uint32_t dev_id,
+   char 
p_name[32],
+   
uint32_t vport_info,
+   
uint32_t *p_queue_id)
+{
+   uint32_t rc = ZXDH_OK;
+
+   uint32_t rsp_buff[2] = {0};
+   uint32_t msg_result = 0;
+   uint32_t queue_id = 0;
+   ZXDH_AGENT_CHANNEL_DTB_MSG_T msgcfg = {
+   .dev_id  = 0,
+   .type= ZXDH_DTB_MSG,
+   .oper= ZXDH_QUEUE_REQUEST,
+   .vport   = vport_info,
+   };
+   memcpy(msgcfg.name, p_name, strnlen(p_name, ZXDH_PORT_NAME_MAX));
+
+   PMD_DRV_LOG(DEBUG, "msgcfg.name=%s", msgcfg.name);
+
+   rc = zxdh_np_agent_channel_dtb_sync_send(dev_id, &msgcfg, rsp_buff, 
sizeof(rsp_buff));
+   ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, 
"zxdh_np_agent_channel_dtb_sync_send");
+
+   msg_result = rsp_buff[0];
+   queue_id = rsp_buff[1];
+
+   PMD_DRV_LOG(DEBUG, "dev_id: %d, msg_result: %d", dev_id, msg_result);
+   PMD_DRV_LOG(DEBUG, "dev_id: %d, queue_id: %d", dev_id

[PATCH v2 03/14] net/zxdh: add agent channel

2025-02-21 Thread Bingbin Chen
Add agent channel to access (np)network processor registers
that are not mapped by PCIE.

Signed-off-by: Bingbin Chen 
---
 drivers/net/zxdh/zxdh_np.c | 386 -
 drivers/net/zxdh/zxdh_np.h |  78 
 2 files changed, 463 insertions(+), 1 deletion(-)

diff --git a/drivers/net/zxdh/zxdh_np.c b/drivers/net/zxdh/zxdh_np.c
index f6d1a215e3..cf5a192a02 100644
--- a/drivers/net/zxdh/zxdh_np.c
+++ b/drivers/net/zxdh/zxdh_np.c
@@ -298,6 +298,24 @@ zxdh_np_comm_convert32(uint32_t dw_data)
 #define ZXDH_DTB_QUEUE_INIT_FLAG_GET(DEV_ID, QUEUE_ID)   \
(p_dpp_dtb_mgr[(DEV_ID)]->queue_info[(QUEUE_ID)].init_flag)
 
+static inline void
+zxdh_np_comm_spinlock_create(ZXDH_SPINLOCK_T *p_spinlock)
+{
+   rte_spinlock_init(&p_spinlock->spinlock);
+}
+
+static inline void
+zxdh_np_comm_spinlock_lock(ZXDH_SPINLOCK_T *p_spinlock)
+{
+   rte_spinlock_lock(&p_spinlock->spinlock);
+}
+
+static inline void
+zxdh_np_comm_spinlock_unlock(ZXDH_SPINLOCK_T *p_spinlock)
+{
+   rte_spinlock_unlock(&p_spinlock->spinlock);
+}
+
 static uint32_t
 zxdh_np_comm_is_big_endian(void)
 {
@@ -354,6 +372,70 @@ zxdh_np_dev_init(void)
return 0;
 }
 
+static void
+zxdh_np_dev_vport_get(uint32_t dev_id, uint32_t *vport)
+{
+   ZXDH_DEV_MGR_T *p_dev_mgr = &g_dev_mgr;
+   ZXDH_DEV_CFG_T *p_dev_info = p_dev_mgr->p_dev_array[dev_id];
+
+   *vport = p_dev_info->vport;
+}
+
+static void
+zxdh_np_dev_agent_addr_get(uint32_t dev_id, uint64_t *agent_addr)
+{
+   ZXDH_DEV_MGR_T *p_dev_mgr = &g_dev_mgr;
+   ZXDH_DEV_CFG_T *p_dev_info = p_dev_mgr->p_dev_array[dev_id];
+
+   *agent_addr = p_dev_info->agent_addr;
+}
+
+static void
+zxdh_np_dev_fw_bar_msg_num_set(uint32_t dev_id, uint32_t bar_msg_num)
+{
+   ZXDH_DEV_MGR_T *p_dev_mgr = &g_dev_mgr;
+   ZXDH_DEV_CFG_T *p_dev_info = p_dev_mgr->p_dev_array[dev_id];
+
+   p_dev_info->fw_bar_msg_num = bar_msg_num;
+
+   PMD_DRV_LOG(INFO, "fw_bar_msg_num_set:fw support agent msg num = %u!", 
bar_msg_num);
+}
+
+static void
+zxdh_np_dev_fw_bar_msg_num_get(uint32_t dev_id, uint32_t *bar_msg_num)
+{
+   ZXDH_DEV_MGR_T *p_dev_mgr = &g_dev_mgr;
+   ZXDH_DEV_CFG_T *p_dev_info = p_dev_mgr->p_dev_array[dev_id];
+
+   *bar_msg_num = p_dev_info->fw_bar_msg_num;
+}
+
+static uint32_t
+zxdh_np_dev_opr_spinlock_get(uint32_t dev_id, uint32_t type, ZXDH_SPINLOCK_T 
**p_spinlock_out)
+{
+   ZXDH_DEV_MGR_T *p_dev_mgr = &g_dev_mgr;
+   ZXDH_DEV_CFG_T *p_dev_info = p_dev_mgr->p_dev_array[dev_id];
+
+   if (p_dev_info == NULL) {
+   PMD_DRV_LOG(ERR, "Get dev_info[ %d ] fail!", dev_id);
+   return ZXDH_DEV_TYPE_INVALID;
+   }
+
+   switch (type) {
+   case ZXDH_DEV_SPINLOCK_T_DTB:
+   *p_spinlock_out = &p_dev_info->dtb_spinlock;
+   break;
+   case ZXDH_DEV_SPINLOCK_T_SMMU0:
+   *p_spinlock_out = &p_dev_info->smmu0_spinlock;
+   break;
+   default:
+   PMD_DRV_LOG(ERR, "spinlock type is invalid!");
+   return ZXDH_ERR;
+   }
+
+   return ZXDH_OK;
+}
+
 static uint32_t
 zxdh_np_dev_read_channel(uint32_t dev_id, uint32_t addr, uint32_t size, 
uint32_t *p_data)
 {
@@ -719,6 +801,7 @@ zxdh_np_dev_add(uint32_t  dev_id, ZXDH_DEV_TYPE_E dev_type,
p_dev_info->p_pcie_write_fun = zxdh_np_dev_pcie_default_write;
p_dev_info->p_pcie_read_fun  = zxdh_np_dev_pcie_default_read;
 
+   zxdh_np_comm_spinlock_create(&p_dev_info->dtb_spinlock);
return 0;
 }
 
@@ -810,6 +893,269 @@ zxdh_np_ppu_parse_cls_bitmap(uint32_t dev_id,
}
 }
 
+static void
+zxdh_np_agent_msg_prt(uint8_t type, uint32_t rtn)
+{
+   switch (rtn) {
+   case ZXDH_RC_CTRLCH_MSG_LEN_ZERO:
+   PMD_DRV_LOG(ERR, "type[%u]:msg len is zero!", type);
+   break;
+   case ZXDH_RC_CTRLCH_MSG_PRO_ERR:
+   PMD_DRV_LOG(ERR, "type[%u]:msg process error!", type);
+   break;
+   case ZXDH_RC_CTRLCH_MSG_TYPE_NOT_SUPPORT:
+   PMD_DRV_LOG(ERR, "type[%u]:fw not support the msg!", type);
+   break;
+   case ZXDH_RC_CTRLCH_MSG_OPER_NOT_SUPPORT:
+   PMD_DRV_LOG(ERR, "type[%u]:fw not support opr of the msg!", 
type);
+   break;
+   case ZXDH_RC_CTRLCH_MSG_DROP:
+   PMD_DRV_LOG(ERR, "type[%u]:fw not support,drop msg!", type);
+   break;
+   default:
+   break;
+   }
+}
+
+static uint32_t
+zxdh_np_agent_bar_msg_check(uint32_t dev_id, ZXDH_AGENT_CHANNEL_MSG_T *p_msg)
+{
+   uint8_t type = 0;
+   uint32_t bar_msg_num = 0;
+
+   type = *((uint8_t *)(p_msg->msg) + 1);
+   if (type != ZXDH_PCIE_BAR_MSG) {
+   zxdh_np_dev_fw_bar_msg_num_get(dev_id, &bar_msg_num);
+   if (type >= bar_msg_num) {
+   PMD_DRV_LOG(ERR, "type[%u] > fw_bar_msg_num[%u]!", 
type, bar_msg_num);
+   return ZXDH_RC_CTRLCH_MSG_

[PATCH v2 02/14] net/zxdh: support compatibility check

2025-02-21 Thread Bingbin Chen
Add compatibility check between (np)network processor
software and firmware.

Signed-off-by: Bingbin Chen 
---
 drivers/net/zxdh/zxdh_np.c | 94 ++
 drivers/net/zxdh/zxdh_np.h | 12 +
 2 files changed, 106 insertions(+)

diff --git a/drivers/net/zxdh/zxdh_np.c b/drivers/net/zxdh/zxdh_np.c
index edc06beaf9..f6d1a215e3 100644
--- a/drivers/net/zxdh/zxdh_np.c
+++ b/drivers/net/zxdh/zxdh_np.c
@@ -13,6 +13,7 @@
 
 #include "zxdh_np.h"
 #include "zxdh_logs.h"
+#include "zxdh_msg.h"
 
 static ZXDH_DEV_MGR_T g_dev_mgr;
 static ZXDH_SDT_MGR_T g_sdt_mgr;
@@ -23,6 +24,9 @@ static ZXDH_DTB_MGR_T *p_dpp_dtb_mgr[ZXDH_DEV_CHANNEL_MAX];
 static ZXDH_RISCV_DTB_MGR *p_riscv_dtb_queue_mgr[ZXDH_DEV_CHANNEL_MAX];
 static ZXDH_SDT_TBL_DATA_T 
g_sdt_info[ZXDH_DEV_CHANNEL_MAX][ZXDH_DEV_SDT_ID_MAX];
 static ZXDH_PPU_STAT_CFG_T g_ppu_stat_cfg[ZXDH_DEV_CHANNEL_MAX];
+static uint64_t g_np_fw_compat_addr[ZXDH_DEV_CHANNEL_MAX];
+static const ZXDH_VERSION_COMPATIBLE_REG_T g_np_sdk_version = {
+   ZXDH_NPSDK_COMPAT_ITEM_ID, 1, 0, 0, 0, {0} };
 
 static ZXDH_FIELD_T g_smmu0_smmu0_cpu_ind_cmd_reg[] = {
{"cpu_ind_rw", ZXDH_FIELD_FLAG_RW, 31, 1, 0x0, 0x0},
@@ -931,6 +935,91 @@ zxdh_np_addr_calc(uint64_t pcie_vir_baddr, uint32_t 
bar_offset)
return np_addr;
 }
 
+static uint64_t
+zxdh_np_fw_compatible_addr_calc(uint64_t pcie_vir_baddr, uint64_t 
compatible_offset)
+{
+   return (pcie_vir_baddr + compatible_offset);
+}
+
+static void
+zxdh_np_pf_fw_compatible_addr_set(uint32_t dev_id, uint64_t pcie_vir_baddr)
+{
+   uint64_t compatible_offset = ZXDH_DPU_NO_DEBUG_PF_COMPAT_REG_OFFSET;
+   uint64_t compatible_addr = 0;
+
+   compatible_addr = zxdh_np_fw_compatible_addr_calc(pcie_vir_baddr, 
compatible_offset);
+
+   g_np_fw_compat_addr[dev_id] = compatible_addr;
+}
+
+static void
+zxdh_np_fw_compatible_addr_get(uint32_t dev_id, uint64_t *p_compatible_addr)
+{
+   *p_compatible_addr = g_np_fw_compat_addr[dev_id];
+}
+
+static void
+zxdh_np_fw_version_data_read(uint64_t compatible_base_addr,
+   ZXDH_VERSION_COMPATIBLE_REG_T *p_fw_version_data, 
uint32_t module_id)
+{
+   void *fw_addr = NULL;
+   uint64_t module_compatible_addr = 0;
+
+   module_compatible_addr = compatible_base_addr +
+   sizeof(ZXDH_VERSION_COMPATIBLE_REG_T) * (module_id - 1);
+
+   fw_addr = (void *)module_compatible_addr;
+
+   memcpy(p_fw_version_data, fw_addr, 
sizeof(ZXDH_VERSION_COMPATIBLE_REG_T));
+}
+
+static void
+zxdh_np_fw_version_compatible_data_get(uint32_t dev_id,
+   ZXDH_VERSION_COMPATIBLE_REG_T 
*p_version_compatible_value,
+   uint32_t module_id)
+{
+   uint64_t compatible_addr = 0;
+
+   zxdh_np_fw_compatible_addr_get(dev_id, &compatible_addr);
+
+   zxdh_np_fw_version_data_read(compatible_addr, 
p_version_compatible_value, module_id);
+}
+
+static uint32_t
+zxdh_np_np_sdk_version_compatible_check(uint32_t dev_id)
+{
+   ZXDH_VERSION_COMPATIBLE_REG_T fw_version = {0};
+
+   zxdh_np_fw_version_compatible_data_get(dev_id, &fw_version, 
ZXDH_NPSDK_COMPAT_ITEM_ID);
+
+   if (fw_version.version_compatible_item != ZXDH_NPSDK_COMPAT_ITEM_ID) {
+   PMD_DRV_LOG(ERR, "version_compatible_item is not DH_NPSDK.");
+   return ZXDH_ERR;
+   }
+
+   if (g_np_sdk_version.major != fw_version.major) {
+   PMD_DRV_LOG(ERR, "dh_npsdk major:%hhu: is not match fw:%hhu!",
+   g_np_sdk_version.major, fw_version.major);
+   return ZXDH_ERR;
+   }
+
+   if (g_np_sdk_version.fw_minor > fw_version.fw_minor) {
+   PMD_DRV_LOG(ERR, "dh_npsdk fw_minor:%hhu is higher than 
fw:%hhu!",
+   g_np_sdk_version.fw_minor, fw_version.fw_minor);
+   return ZXDH_ERR;
+   }
+
+   if (g_np_sdk_version.drv_minor < fw_version.drv_minor) {
+   PMD_DRV_LOG(ERR, "dh_npsdk drv_minor:%hhu is lower than 
fw:%hhu!",
+   g_np_sdk_version.drv_minor, fw_version.drv_minor);
+   return ZXDH_ERR;
+   }
+
+   PMD_DRV_LOG(INFO, "dh_npsdk compatible check success!");
+
+   return ZXDH_OK;
+}
+
 static ZXDH_RISCV_DTB_MGR *
 zxdh_np_riscv_dtb_queue_mgr_get(uint32_t dev_id)
 {
@@ -2508,5 +2597,10 @@ zxdh_np_host_init(uint32_t dev_id,
agent_addr = ZXDH_PCIE_AGENT_ADDR_OFFSET + 
p_dev_init_ctrl->pcie_vir_addr;
zxdh_np_dev_agent_addr_set(dev_id, agent_addr);
 
+   zxdh_np_pf_fw_compatible_addr_set(dev_id, 
p_dev_init_ctrl->pcie_vir_addr);
+
+   rc = zxdh_np_np_sdk_version_compatible_check(dev_id);
+   ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, 
"zxdh_np_np_sdk_version_compatible_check");
+
return 0;
 }
diff --git a/drivers/net/zxdh/zxdh_np.h b/drivers/net/zxdh/zxdh_np.h
index cc298d0eff..a592436b4f 100644
--- a/drivers/net/zxdh/zxdh_np.h
+++ b/drivers/net/zxdh/zxdh_np.h
@@ -112,6 +112,9 @@
 
 #define ZXDH_SE_OPR_RD   

[PATCH v2 11/14] net/zxdh: get hash table entry result

2025-02-21 Thread Bingbin Chen
Implement hash tables get operation by dtb channel.

Signed-off-by: Bingbin Chen 
---
 drivers/net/zxdh/zxdh_np.c | 479 +
 drivers/net/zxdh/zxdh_np.h |   7 +
 2 files changed, 486 insertions(+)

diff --git a/drivers/net/zxdh/zxdh_np.c b/drivers/net/zxdh/zxdh_np.c
index ea25d431b9..8ac33e88ed 100644
--- a/drivers/net/zxdh/zxdh_np.c
+++ b/drivers/net/zxdh/zxdh_np.c
@@ -1282,6 +1282,40 @@ zxdh_comm_rb_insert(ZXDH_RB_CFG *p_rb_cfg,
return ZXDH_OK;
 }
 
+static uint32_t
+zxdh_comm_rb_search(ZXDH_RB_CFG *p_rb_cfg,
+void  *p_key,
+void  *out_val)
+{
+   int32_t cmprtn   = 0;
+   ZXDH_RB_TN *p_cur_tn = NULL;
+
+   p_cur_tn = p_rb_cfg->p_root;
+
+   while (p_cur_tn) {
+   cmprtn = p_rb_cfg->p_cmpfun(p_key, p_cur_tn->p_key, 
p_rb_cfg->key_size);
+
+   if (cmprtn > 0)
+   p_cur_tn = p_cur_tn->p_right;
+   else if (cmprtn < 0)
+   p_cur_tn = p_cur_tn->p_left;
+   else
+   break;
+   }
+
+   if (!p_cur_tn) {
+   PMD_DRV_LOG(DEBUG, "rb srh fail");
+   return ZXDH_RBT_RC_SRHFAIL;
+   }
+
+   if (p_rb_cfg->is_dynamic)
+   *(ZXDH_RB_TN **)out_val = p_cur_tn;
+   else
+   *(uint32_t *)out_val = zxdh_np_get_tn_lsv(p_cur_tn);
+
+   return ZXDH_OK;
+}
+
 static uint32_t
 zxdh_comm_rb_handle_del(__rte_unused ZXDH_RB_CFG *p_rb_cfg,
ZXDH_RB_TN ***stack_tn,
@@ -3857,6 +3891,38 @@ zxdh_np_dtb_smmu0_dump_info_write(uint32_t dev_id,
return rc;
 }
 
+static uint32_t
+zxdh_np_dtb_zcam_dump_info_write(uint32_t dev_id,
+   uint32_t addr,
+   uint32_t tb_width,
+   uint32_t depth,
+   uint32_t addr_high32,
+   uint32_t addr_low32,
+   uint32_t *p_dump_info)
+{
+   uint32_t rc = ZXDH_OK;
+
+   ZXDH_DTB_ZCAM_DUMP_FORM_T dtb_zcam_dump_form_info = {
+   .valid = 1,
+   .up_type = ZXDH_DTB_DUMP_MODE_ZCAM,
+   .tb_width = tb_width,
+   .sram_addr = addr & 0x1FF,
+   .ram_reg_flag = (addr >> 16) & 0x1,
+   .z_reg_cell_id = (addr >> 9) & 0x3,
+   .zblock_id = (addr >> 11) & 0x7,
+   .zgroup_id = (addr >> 14) & 0x3,
+   .tb_depth = depth,
+   .tb_dst_addr_h = addr_high32,
+   .tb_dst_addr_l = addr_low32,
+   };
+
+   rc = zxdh_np_dtb_write_dump_cmd(dev_id, ZXDH_DTB_DUMP_ZCAM,
+   &dtb_zcam_dump_form_info, p_dump_info);
+   ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, "zxdh_np_dtb_write_dump_cmd");
+
+   return rc;
+}
+
 static uint32_t
 zxdh_np_dtb_se_smmu0_ind_write(uint32_t dev_id,
uint32_t base_addr,
@@ -5899,6 +5965,397 @@ zxdh_np_dtb_eram_data_get(uint32_t dev_id, uint32_t 
queue_id, uint32_t sdt_no,
return rc;
 }
 
+static uint32_t
+zxdh_np_dtb_se_zcam_dma_dump(uint32_t dev_id,
+   uint32_t queue_id,
+   uint32_t addr,
+   uint32_t tb_width,
+   uint32_t depth,
+   uint32_t *p_data,
+   uint32_t *element_id)
+{
+   uint32_t rc = ZXDH_OK;
+   uint32_t dump_dst_phy_haddr = 0;
+   uint32_t dump_dst_phy_laddr = 0;
+   uint32_t queue_item_index = 0;
+   uint32_t data_len = 0;
+   uint32_t desc_len = 0;
+   uint32_t tb_width_len = 0;
+   uint8_t form_buff[ZXDH_DTB_TABLE_CMD_SIZE_BIT / 8] = {0};
+
+   rc = zxdh_np_dtb_tab_up_free_item_get(dev_id, queue_id, 
&queue_item_index);
+   if (rc != ZXDH_OK) {
+   PMD_DRV_LOG(ERR, "zxdh_np_dtb_tab_up_free_item_get failed!");
+   return ZXDH_RC_DTB_QUEUE_ITEM_SW_EMPTY;
+   }
+
+   PMD_DRV_LOG(DEBUG, "table up item queue_element_id is: %d.",
+   queue_item_index);
+
+   *element_id = queue_item_index;
+
+   rc = zxdh_np_dtb_tab_up_item_addr_get(dev_id, queue_id, 
queue_item_index,
+   &dump_dst_phy_haddr, &dump_dst_phy_laddr);
+   ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, "zxdh_np_dtb_tab_up_item_addr_get");
+
+   rc = zxdh_np_dtb_zcam_dump_info_write(dev_id,
+ addr,
+ 
tb_w

[PATCH v2 14/14] net/zxdh: clean stat values

2025-02-21 Thread Bingbin Chen
Implement stat values clean operation by agent channel.

Signed-off-by: Bingbin Chen 
---
 drivers/net/zxdh/zxdh_ethdev_ops.c |   4 +
 drivers/net/zxdh/zxdh_np.c | 302 +
 drivers/net/zxdh/zxdh_np.h |  23 ++-
 3 files changed, 328 insertions(+), 1 deletion(-)

diff --git a/drivers/net/zxdh/zxdh_ethdev_ops.c 
b/drivers/net/zxdh/zxdh_ethdev_ops.c
index 1a21ded58e..284bf27d10 100644
--- a/drivers/net/zxdh/zxdh_ethdev_ops.c
+++ b/drivers/net/zxdh/zxdh_ethdev_ops.c
@@ -1500,11 +1500,15 @@ static int zxdh_hw_stats_reset(struct rte_eth_dev *dev, 
enum zxdh_agent_msg_type
 int zxdh_dev_stats_reset(struct rte_eth_dev *dev)
 {
struct zxdh_hw *hw = dev->data->dev_private;
+   uint64_t stats_data;
 
zxdh_hw_stats_reset(dev, ZXDH_VQM_DEV_STATS_RESET);
if (hw->is_pf)
zxdh_hw_stats_reset(dev, ZXDH_MAC_STATS_RESET);
 
+   zxdh_np_stat_ppu_cnt_get_ex(0, ZXDH_STAT_64_MODE, 0,
+   ZXDH_STAT_RD_CLR_MODE_CLR, (uint32_t *)&stats_data);
+
return 0;
 }
 
diff --git a/drivers/net/zxdh/zxdh_np.c b/drivers/net/zxdh/zxdh_np.c
index 760560d8b7..362814f410 100644
--- a/drivers/net/zxdh/zxdh_np.c
+++ b/drivers/net/zxdh/zxdh_np.c
@@ -2197,6 +2197,8 @@ zxdh_np_dev_add(uint32_t  dev_id, ZXDH_DEV_TYPE_E 
dev_type,
 
zxdh_np_comm_spinlock_create(&p_dev_info->dtb_spinlock);
 
+   zxdh_np_comm_spinlock_create(&p_dev_info->smmu0_spinlock);
+
for (i = 0; i < ZXDH_DTB_QUEUE_NUM_MAX; i++)

zxdh_np_comm_spinlock_create(&p_dev_info->dtb_queue_spinlock[i]);
 
@@ -3238,6 +3240,32 @@ zxdh_np_reg_read(uint32_t dev_id, uint32_t reg_no,
return rc;
 }
 
+static uint32_t
+zxdh_np_reg_read32(uint32_t dev_id, uint32_t reg_no,
+   uint32_t m_offset, uint32_t n_offset, uint32_t *p_data)
+{
+   uint32_t rc = 0;
+   uint32_t addr = 0;
+   ZXDH_REG_T *p_reg_info = &g_dpp_reg_info[reg_no];
+   uint32_t p_buff[ZXDH_REG_DATA_MAX] = {0};
+   uint32_t reg_real_no = p_reg_info->reg_no;
+   uint32_t reg_type = p_reg_info->flags;
+   uint32_t reg_module = p_reg_info->module_no;
+
+   addr = zxdh_np_reg_get_reg_addr(reg_no, m_offset, n_offset);
+
+   if (reg_module == DTB4K) {
+   rc = p_reg_info->p_read_fun(dev_id, addr, p_data);
+   ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, "p_reg_info->p_read_fun");
+   } else {
+   rc = zxdh_np_agent_channel_reg_read(dev_id, reg_type, 
reg_real_no, 4, addr, p_buff);
+   ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, 
"zxdh_np_agent_channel_reg_read");
+   *p_data = p_buff[0];
+   }
+
+   return rc;
+}
+
 static uint32_t
 zxdh_np_dtb_queue_vm_info_get(uint32_t dev_id,
uint32_t queue_id,
@@ -10274,3 +10302,277 @@ zxdh_np_dtb_hash_offline_delete(uint32_t dev_id,
 
return rc;
 }
+
+static uint32_t
+zxdh_np_se_done_status_check(uint32_t dev_id, uint32_t reg_no, uint32_t pos)
+{
+   uint32_t rc = ZXDH_OK;
+
+   uint32_t data = 0;
+   uint32_t rd_cnt = 0;
+   uint32_t done_flag = 0;
+
+   while (!done_flag) {
+   rc = zxdh_np_reg_read32(dev_id, reg_no, 0, 0, &data);
+   if (rc != ZXDH_OK) {
+   PMD_DRV_LOG(ERR, " [ErrorCode:0x%x] !-- 
zxdh_np_reg_read32 Fail!", rc);
+   return rc;
+   }
+
+   done_flag = (data >> pos) & 0x1;
+
+   if (done_flag)
+   break;
+
+   if (rd_cnt > ZXDH_RD_CNT_MAX * ZXDH_RD_CNT_MAX)
+   return ZXDH_ERR;
+
+   rd_cnt++;
+   }
+
+   return rc;
+}
+
+static uint32_t
+zxdh_np_se_smmu0_ind_read(uint32_t dev_id,
+   uint32_t base_addr,
+   uint32_t index,
+   uint32_t rd_mode,
+   uint32_t rd_clr_mode,
+   uint32_t *p_data)
+{
+   uint32_t rc = ZXDH_OK;
+   uint32_t i = 0;
+   uint32_t row_index = 0;
+   uint32_t col_index = 0;
+   uint32_t temp_data[4] = {0};
+   uint32_t *p_temp_data = NULL;
+   ZXDH_SMMU0_SMMU0_CPU_IND_CMD_T cpu_ind_cmd = {0};
+   ZXDH_SPINLOCK_T *p_ind_spinlock = NULL;
+
+   rc = zxdh_np_dev_opr_spinlock_get(dev_id, ZXDH_DEV_SPINLOCK_T_SMMU0, 
&p_ind_spinlock);
+   ZXDH_COMM_CHECK_DEV_RC(dev_id, rc, "zxdh_np_dev_opr_spinlock_get");
+
+   zxdh_np_comm_spinlock_lock(p_ind_spinlock);
+
+   rc = zxdh_np_se_done_status_check(dev_id, 
ZXDH_SMMU0_SMMU0_WR_ARB_CPU_RDYR, 0);
+   if (rc != ZXDH_OK) {
+   PMD_DRV_LOG(ERR, "se done status check failed, rc=0x%x.", rc);
+   zxdh_np_comm_spinlock_unlock(p_ind_spinlock);
+   return ZXDH_ERR;
+   }
+
+   if (rd_clr_mode == ZXDH_RD_MODE_HOLD) {
+   cpu_ind_cmd.cpu_ind_rw = ZXDH_SE_OPR_RD;
+