[dpdk-dev] [PATCH] maintainers: fix missing title for examples

2020-10-06 Thread Olivier Matz
The devtools/get-maintainer.sh script does not work with examples
because there is no title line between them: it returns a longer list
than expected.

Add the missing titles for each example to fix this behavior.

Signed-off-by: Olivier Matz 
---
 MAINTAINERS | 17 +
 1 file changed, 17 insertions(+)

diff --git a/MAINTAINERS b/MAINTAINERS
index 13881aa737..39ddaa5910 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -1564,46 +1564,57 @@ F: doc/guides/tools/proc_info.rst
 Other Example Applications
 --
 
+Ethtool example
 F: examples/ethtool/
 F: doc/guides/sample_app_ug/ethtool.rst
 
+FIPS validation example
 M: Marko Kovacevic 
 F: examples/fips_validation/
 F: doc/guides/sample_app_ug/fips_validation.rst
 
+Flow filtering example
 M: Ori Kam 
 F: examples/flow_filtering/
 F: doc/guides/sample_app_ug/flow_filtering.rst
 
+Helloworld example
 M: Bruce Richardson 
 F: examples/helloworld/
 F: doc/guides/sample_app_ug/hello_world.rst
 
+IPsec security gateway example
 M: Radu Nicolau 
 M: Akhil Goyal 
 F: examples/ipsec-secgw/
 F: doc/guides/sample_app_ug/ipsec_secgw.rst
 
+IPv4 multicast example
 F: examples/ipv4_multicast/
 F: doc/guides/sample_app_ug/ipv4_multicast.rst
 
+L2 forwarding example
 M: Bruce Richardson 
 F: examples/l2fwd/
 F: doc/guides/sample_app_ug/l2_forward_real_virtual.rst
 
+L2 forwarding with cache allocation example
 M: Tomasz Kantecki 
 F: doc/guides/sample_app_ug/l2_forward_cat.rst
 F: examples/l2fwd-cat/
 
+L2 forwarding with eventdev example
 M: Sunil Kumar Kori 
 M: Pavan Nikhilesh 
 T: git://dpdk.org/next/dpdk-next-eventdev
 F: examples/l2fwd-event/
 F: doc/guides/sample_app_ug/l2_forward_event.rst
 
+L3 forwarding example
 F: examples/l3fwd/
 F: doc/guides/sample_app_ug/l3_forward.rst
 
+Link status interrupt example
 F: examples/link_status_interrupt/
 F: doc/guides/sample_app_ug/link_status_intr.rst
 
@@ -1612,26 +1623,32 @@ M: John McNamara 
 F: examples/performance-thread/
 F: doc/guides/sample_app_ug/performance_thread.rst
 
+PTP client example
 M: Kirill Rybalchenko 
 F: examples/ptpclient/
 
+Rx/Tx callbacks example
 M: Bruce Richardson 
 M: John McNamara 
 F: examples/rxtx_callbacks/
 F: doc/guides/sample_app_ug/rxtx_callbacks.rst
 
+Service cores example
 M: Harry van Haaren 
 F: examples/service_cores/
 F: doc/guides/sample_app_ug/service_cores.rst
 
+Skeleton example
 M: Bruce Richardson 
 M: John McNamara 
 F: examples/skeleton/
 F: doc/guides/sample_app_ug/skeleton.rst
 
+TEP termination example
 M: Xiaoyun Li 
 F: examples/tep_termination/
 
+VMDQ examples
 F: examples/vmdq/
 F: doc/guides/sample_app_ug/vmdq_forwarding.rst
 F: examples/vmdq_dcb/
-- 
2.25.1



[dpdk-dev] [PATCH 1/2] mempool: remove v20 ABI

2020-10-06 Thread Olivier Matz
Remove the deprecated v20 ABI of rte_mempool_populate_iova() and
rte_mempool_populate_virt().

Signed-off-by: Olivier Matz 
---
 lib/librte_mempool/meson.build |  2 -
 lib/librte_mempool/rte_mempool.c   | 79 ++
 lib/librte_mempool/rte_mempool_version.map |  7 --
 3 files changed, 5 insertions(+), 83 deletions(-)

diff --git a/lib/librte_mempool/meson.build b/lib/librte_mempool/meson.build
index 7dbe6b9bea..a6e861cbfc 100644
--- a/lib/librte_mempool/meson.build
+++ b/lib/librte_mempool/meson.build
@@ -9,8 +9,6 @@ foreach flag: extra_flags
endif
 endforeach
 
-use_function_versioning = true
-
 sources = files('rte_mempool.c', 'rte_mempool_ops.c',
'rte_mempool_ops_default.c', 'mempool_trace_points.c')
 headers = files('rte_mempool.h', 'rte_mempool_trace.h',
diff --git a/lib/librte_mempool/rte_mempool.c b/lib/librte_mempool/rte_mempool.c
index 7774f0c8da..0e3a2a7635 100644
--- a/lib/librte_mempool/rte_mempool.c
+++ b/lib/librte_mempool/rte_mempool.c
@@ -30,7 +30,6 @@
 #include 
 #include 
 #include 
-#include 
 #include 
 
 
@@ -305,17 +304,12 @@ mempool_ops_alloc_once(struct rte_mempool *mp)
return 0;
 }
 
-__vsym int
-rte_mempool_populate_iova_v21(struct rte_mempool *mp, char *vaddr,
-   rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
-   void *opaque);
-
 /* Add objects in the pool, using a physically contiguous memory
  * zone. Return the number of objects added, or a negative value
  * on error.
  */
-__vsym int
-rte_mempool_populate_iova_v21(struct rte_mempool *mp, char *vaddr,
+int
+rte_mempool_populate_iova(struct rte_mempool *mp, char *vaddr,
rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
void *opaque)
 {
@@ -375,35 +369,6 @@ rte_mempool_populate_iova_v21(struct rte_mempool *mp, char 
*vaddr,
return ret;
 }
 
-BIND_DEFAULT_SYMBOL(rte_mempool_populate_iova, _v21, 21);
-MAP_STATIC_SYMBOL(
-   int rte_mempool_populate_iova(struct rte_mempool *mp, char *vaddr,
-   rte_iova_t iova, size_t len,
-   rte_mempool_memchunk_free_cb_t *free_cb,
-   void *opaque),
-   rte_mempool_populate_iova_v21);
-
-__vsym int
-rte_mempool_populate_iova_v20(struct rte_mempool *mp, char *vaddr,
-   rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
-   void *opaque);
-
-__vsym int
-rte_mempool_populate_iova_v20(struct rte_mempool *mp, char *vaddr,
-   rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
-   void *opaque)
-{
-   int ret;
-
-   ret = rte_mempool_populate_iova_v21(mp, vaddr, iova, len, free_cb,
-   opaque);
-   if (ret == 0)
-   ret = -EINVAL;
-
-   return ret;
-}
-VERSION_SYMBOL(rte_mempool_populate_iova, _v20, 20.0);
-
 static rte_iova_t
 get_iova(void *addr)
 {
@@ -417,16 +382,11 @@ get_iova(void *addr)
return ms->iova + RTE_PTR_DIFF(addr, ms->addr);
 }
 
-__vsym int
-rte_mempool_populate_virt_v21(struct rte_mempool *mp, char *addr,
-   size_t len, size_t pg_sz, rte_mempool_memchunk_free_cb_t *free_cb,
-   void *opaque);
-
 /* Populate the mempool with a virtual area. Return the number of
  * objects added, or a negative value on error.
  */
-__vsym int
-rte_mempool_populate_virt_v21(struct rte_mempool *mp, char *addr,
+int
+rte_mempool_populate_virt(struct rte_mempool *mp, char *addr,
size_t len, size_t pg_sz, rte_mempool_memchunk_free_cb_t *free_cb,
void *opaque)
 {
@@ -459,7 +419,7 @@ rte_mempool_populate_virt_v21(struct rte_mempool *mp, char 
*addr,
break;
}
 
-   ret = rte_mempool_populate_iova_v21(mp, addr + off, iova,
+   ret = rte_mempool_populate_iova(mp, addr + off, iova,
phys_len, free_cb, opaque);
if (ret == 0)
continue;
@@ -477,35 +437,6 @@ rte_mempool_populate_virt_v21(struct rte_mempool *mp, char 
*addr,
rte_mempool_free_memchunks(mp);
return ret;
 }
-BIND_DEFAULT_SYMBOL(rte_mempool_populate_virt, _v21, 21);
-MAP_STATIC_SYMBOL(
-   int rte_mempool_populate_virt(struct rte_mempool *mp,
-   char *addr, size_t len, size_t pg_sz,
-   rte_mempool_memchunk_free_cb_t *free_cb,
-   void *opaque),
-   rte_mempool_populate_virt_v21);
-
-__vsym int
-rte_mempool_populate_virt_v20(struct rte_mempool *mp, char *addr,
-   size_t len, size_t pg_sz, rte_mempool_memchunk_free_cb_t *free_cb,
-   void *opaque);
-
-__vsym int
-rte_mempool_populate_virt_v20(struct rte_mempool *mp, char *addr,
-   size_t len, size_t pg_sz, rte_mempool_memchunk_free_cb_t *free_cb,
-   void *opaque)
-{
-   int ret;
-
-   ret = rte_mempool_populate_virt_v21(mp, addr, len, pg_sz,
-   

[dpdk-dev] [PATCH 2/2] mempool: remove experimental tags

2020-10-06 Thread Olivier Matz
Move symbols introduced in version <= 19.11 in the stable ABI.

Signed-off-by: Olivier Matz 
---
 lib/librte_mempool/rte_mempool.h   | 32 --
 lib/librte_mempool/rte_mempool_version.map | 12 +++-
 2 files changed, 4 insertions(+), 40 deletions(-)

diff --git a/lib/librte_mempool/rte_mempool.h b/lib/librte_mempool/rte_mempool.h
index 9ea7ff934c..c551cf733a 100644
--- a/lib/librte_mempool/rte_mempool.h
+++ b/lib/librte_mempool/rte_mempool.h
@@ -191,9 +191,6 @@ struct rte_mempool_memhdr {
 };
 
 /**
- * @warning
- * @b EXPERIMENTAL: this API may change without prior notice.
- *
  * Additional information about the mempool
  *
  * The structure is cache-line aligned to avoid ABI breakages in
@@ -358,9 +355,6 @@ void rte_mempool_check_cookies(const struct rte_mempool *mp,
 #endif /* RTE_LIBRTE_MEMPOOL_DEBUG */
 
 /**
- * @warning
- * @b EXPERIMENTAL: this API may change without prior notice.
- *
  * @internal Check contiguous object blocks and update cookies or panic.
  *
  * @param mp
@@ -421,9 +415,6 @@ typedef int (*rte_mempool_dequeue_t)(struct rte_mempool *mp,
void **obj_table, unsigned int n);
 
 /**
- * @warning
- * @b EXPERIMENTAL: this API may change without prior notice.
- *
  * Dequeue a number of contiguous object blocks from the external pool.
  */
 typedef int (*rte_mempool_dequeue_contig_blocks_t)(struct rte_mempool *mp,
@@ -462,9 +453,6 @@ typedef ssize_t (*rte_mempool_calc_mem_size_t)(const struct 
rte_mempool *mp,
size_t *min_chunk_size, size_t *align);
 
 /**
- * @warning
- * @b EXPERIMENTAL: this API may change without prior notice.
- *
  * @internal Helper to calculate memory size required to store given
  * number of objects.
  *
@@ -499,7 +487,6 @@ typedef ssize_t (*rte_mempool_calc_mem_size_t)(const struct 
rte_mempool *mp,
  * @return
  *   Required memory size.
  */
-__rte_experimental
 ssize_t rte_mempool_op_calc_mem_size_helper(const struct rte_mempool *mp,
uint32_t obj_num, uint32_t pg_shift, size_t chunk_reserve,
size_t *min_chunk_size, size_t *align);
@@ -569,9 +556,6 @@ typedef int (*rte_mempool_populate_t)(struct rte_mempool 
*mp,
 #define RTE_MEMPOOL_POPULATE_F_ALIGN_OBJ 0x0001
 
 /**
- * @warning
- * @b EXPERIMENTAL: this API may change without prior notice.
- *
  * @internal Helper to populate memory pool object using provided memory
  * chunk: just slice objects one by one, taking care of not
  * crossing page boundaries.
@@ -603,7 +587,6 @@ typedef int (*rte_mempool_populate_t)(struct rte_mempool 
*mp,
  * @return
  *   The number of objects added in mempool.
  */
-__rte_experimental
 int rte_mempool_op_populate_helper(struct rte_mempool *mp,
unsigned int flags, unsigned int max_objs,
void *vaddr, rte_iova_t iova, size_t len,
@@ -621,9 +604,6 @@ int rte_mempool_op_populate_default(struct rte_mempool *mp,
rte_mempool_populate_obj_cb_t *obj_cb, void *obj_cb_arg);
 
 /**
- * @warning
- * @b EXPERIMENTAL: this API may change without prior notice.
- *
  * Get some additional information about a mempool.
  */
 typedef int (*rte_mempool_get_info_t)(const struct rte_mempool *mp,
@@ -846,9 +826,6 @@ int rte_mempool_ops_populate(struct rte_mempool *mp, 
unsigned int max_objs,
 void *obj_cb_arg);
 
 /**
- * @warning
- * @b EXPERIMENTAL: this API may change without prior notice.
- *
  * Wrapper for mempool_ops get_info callback.
  *
  * @param[in] mp
@@ -860,7 +837,6 @@ int rte_mempool_ops_populate(struct rte_mempool *mp, 
unsigned int max_objs,
  *mempool information
  *   - -ENOTSUP - doesn't support get_info ops (valid case).
  */
-__rte_experimental
 int rte_mempool_ops_get_info(const struct rte_mempool *mp,
 struct rte_mempool_info *info);
 
@@ -1577,9 +1553,6 @@ rte_mempool_get(struct rte_mempool *mp, void **obj_p)
 }
 
 /**
- * @warning
- * @b EXPERIMENTAL: this API may change without prior notice.
- *
  * Get a contiguous blocks of objects from the mempool.
  *
  * If cache is enabled, consider to flush it first, to reuse objects
@@ -1601,7 +1574,6 @@ rte_mempool_get(struct rte_mempool *mp, void **obj_p)
  *   - -EOPNOTSUPP: The mempool driver does not support block dequeue
  */
 static __rte_always_inline int
-__rte_experimental
 rte_mempool_get_contig_blocks(struct rte_mempool *mp,
  void **first_obj_table, unsigned int n)
 {
@@ -1786,13 +1758,9 @@ void rte_mempool_walk(void (*func)(struct rte_mempool *, 
void *arg),
  void *arg);
 
 /**
- * @warning
- * @b EXPERIMENTAL: this API may change without prior notice.
- *
  * @internal Get page size used for mempool object allocation.
  * This function is internal to mempool library and mempool drivers.
  */
-__rte_experimental
 int
 rte_mempool_get_page_size(struct rte_mempool *mp, size_t *pg_sz);
 
diff --git a/lib/librte_mempool/rte_mempool_version.map 
b/lib/librte_mempool/rte_memp

[dpdk-dev] [PATCH v2] mbuf: extend meaning of QinQ stripped bit

2020-10-06 Thread Olivier Matz
From: Somnath Kotur 

Clarify the documentation of QinQ flags, and extend the meaning of the
flag: if PKT_RX_QINQ_STRIPPED is set and PKT_RX_VLAN_STRIPPED is unset,
only the outer VLAN is removed from packet data, but both tci are saved
in mbuf->vlan_tci (inner) and mbuf->vlan_tci_outer (outer).

Signed-off-by: Somnath Kotur 
Signed-off-by: Olivier Matz 
---

Hi,

Here is an update of this old patch from Somnath.

Olivier


v2:
- better reword PKT_RX_QINQ_STRIPPED documentation
- also update PKT_RX_QINQ documentation

 lib/librte_mbuf/rte_mbuf_core.h | 21 ++---
 1 file changed, 14 insertions(+), 7 deletions(-)

diff --git a/lib/librte_mbuf/rte_mbuf_core.h b/lib/librte_mbuf/rte_mbuf_core.h
index 8c2c20644d..8f631b84cf 100644
--- a/lib/librte_mbuf/rte_mbuf_core.h
+++ b/lib/librte_mbuf/rte_mbuf_core.h
@@ -126,12 +126,19 @@ extern "C" {
 #define PKT_RX_FDIR_FLX  (1ULL << 14)
 
 /**
- * The 2 vlans have been stripped by the hardware and their tci are
- * saved in mbuf->vlan_tci (inner) and mbuf->vlan_tci_outer (outer).
- * This can only happen if vlan stripping is enabled in the RX
+ * The outer VLAN has been stripped by the hardware and its TCI is
+ * saved in mbuf->vlan_tci_outer.
+ * This can only happen if VLAN stripping is enabled in the Rx
  * configuration of the PMD.
- * When PKT_RX_QINQ_STRIPPED is set, the flags (PKT_RX_VLAN |
- * PKT_RX_VLAN_STRIPPED | PKT_RX_QINQ) must also be set.
+ * When PKT_RX_QINQ_STRIPPED is set, the flags PKT_RX_VLAN and PKT_RX_QINQ
+ * must also be set.
+ *
+ * - If both PKT_RX_QINQ_STRIPPED and PKT_RX_VLAN_STRIPPED are set, the 2 VLANs
+ *   have been stripped by the hardware and their TCIs are saved in
+ *   mbuf->vlan_tci (inner) and mbuf->vlan_tci_outer (outer).
+ * - If PKT_RX_QINQ_STRIPPED is set and PKT_RX_VLAN_STRIPPED is unset, only the
+ *   outer VLAN is removed from packet data, but both tci are saved in
+ *   mbuf->vlan_tci (inner) and mbuf->vlan_tci_outer (outer).
  */
 #define PKT_RX_QINQ_STRIPPED (1ULL << 15)
 
@@ -159,8 +166,8 @@ extern "C" {
 
 /**
  * The RX packet is a double VLAN, and the outer tci has been
- * saved in in mbuf->vlan_tci_outer. If PKT_RX_QINQ set, PKT_RX_VLAN
- * also should be set and inner tci should be saved to mbuf->vlan_tci.
+ * saved in mbuf->vlan_tci_outer. If this flag is set, PKT_RX_VLAN
+ * must also be set and the inner tci is saved in mbuf->vlan_tci.
  * If the flag PKT_RX_QINQ_STRIPPED is also present, both VLANs
  * headers have been stripped from mbuf data, else they are still
  * present.
-- 
2.25.1



Re: [dpdk-dev] [PATCH v2 0/3] simplify unit-testing of rawdevs

2020-10-06 Thread Thomas Monjalon
> At present the "rawdev_autotest" command creates a skeleton rawdev and runs
> a series of tests on the rawdev API and on that rawdev. While the rawdev
> API set includes a "selftest" function, it is not hooked up to this test so
> to test an individual rawdev driver, e.g. ioat, requires that a new test
> command be added.
> 
> This patchset improves the situation by changing the UT to first run the
> existing API tests, but then call selftest on all rawdevs on the system.
> This removes the need for any new test commands for new drivers. If there
> are multiple rawdevs on a system, the sub-set to be tested can be limited
> via existing means such as using the device block/allow EAL parameters or
> similarly via vdev args, etc.
> 
> As part of this change, the ioat rawdev autotest is fixed to allow calling
> on multiple instances inside the one test run, and thereafter the custom
> test command for it is removed as it is no longer necessary.

Applied, thanks




Re: [dpdk-dev] [PATCH] net/vhost: fix xstats wrong after clearing stats

2020-10-06 Thread Maxime Coquelin



On 10/5/20 7:43 PM, David Christensen wrote:
> 
> 
> On 9/23/20 1:07 AM, Maxime Coquelin wrote:
>> Hi David,
>>
>> Could you please post a v2 with Chenbo's comments taken into account?
> 
> Sorry, been out of the office for a bit but now working through my
> backlog. I did make the change but discovered there were additional
> dependencies between stats and xstats that weren't addressed in the v1
> patch, specifically per queue packet/byte counters:
> 
> $ diff before.txt after.txt:
> < tx_good_packets: 24769339
> ---
>> tx_good_packets: 0
> 6c6
> < tx_good_bytes: 1501282012967
> ---
>> tx_good_bytes: 0
> 14,15c14,15
> < tx_q0packets: 24769339
> < tx_q0bytes: 1501282012967
> ---
>> tx_q0packets: 0
>> tx_q0bytes: 0
> 51c51
> < tx_good_packets: 2653125
> ---
>> tx_good_packets: 0
> 53c53
> < tx_good_bytes: 167404266
> ---
>> tx_good_bytes: 0
> 61,62c61,62
> < tx_q0packets: 2653125
> < tx_q0bytes: 167404266
> ---
>> tx_q0packets: 0
>> tx_q0bytes: 0
> 
> I'm looking at how to address these as well for a v2 patch soon.

Thanks Dave!

> Dave
> 



[dpdk-dev] [PATCH 3/3] examples/fips_validation: support self-test only

2020-10-06 Thread Olivier Matz
Make it possible to pass the self-tests when no req path is set.

Signed-off-by: Olivier Matz 
---
 examples/fips_validation/main.c | 22 --
 1 file changed, 20 insertions(+), 2 deletions(-)

diff --git a/examples/fips_validation/main.c b/examples/fips_validation/main.c
index 0a1c8b568c..ee3a890e9e 100644
--- a/examples/fips_validation/main.c
+++ b/examples/fips_validation/main.c
@@ -315,8 +315,21 @@ cryptodev_fips_validate_parse_args(int argc, char **argv)
}
}
 
-   if (env.req_path == NULL || env.rsp_path == NULL ||
-   env.dev_id == UINT32_MAX) {
+   if (env.dev_id == UINT32_MAX) {
+   RTE_LOG(ERR, USER1, "No device specified\n");
+   cryptodev_fips_validate_usage(prgname);
+   return -EINVAL;
+   }
+
+   if ((env.req_path == NULL && env.rsp_path != NULL) ||
+   (env.req_path != NULL && env.rsp_path == NULL)) {
+   RTE_LOG(ERR, USER1, "Missing req path or rsp path\n");
+   cryptodev_fips_validate_usage(prgname);
+   return -EINVAL;
+   }
+
+   if (env.req_path == NULL && env.self_test == 0) {
+   RTE_LOG(ERR, USER1, "--self-test must be set if req path is 
missing\n");
cryptodev_fips_validate_usage(prgname);
return -EINVAL;
}
@@ -348,6 +361,11 @@ main(int argc, char *argv[])
return -1;
}
 
+   if (env.req_path == NULL || env.rsp_path == NULL) {
+   printf("No request, exit.\n");
+   goto exit;
+   }
+
if (!env.is_path_folder) {
printf("Processing file %s... ", env.req_path);
 
-- 
2.25.1



[dpdk-dev] [PATCH 2/3] examples/fips_validation: ignore \r in input files

2020-10-06 Thread Olivier Matz
Some test vectors contain '\r' before '\n' (see link). Ignore them.

Link: https://www.openssl.org/docs/fips/testvectors-linux-2007-10-10.tar.gz
Fixes: 3d0fad56b74a ("examples/fips_validation: add crypto FIPS application")
Cc: sta...@dpdk.org

Signed-off-by: Olivier Matz 
---
 examples/fips_validation/fips_validation.c | 2 ++
 1 file changed, 2 insertions(+)

diff --git a/examples/fips_validation/fips_validation.c 
b/examples/fips_validation/fips_validation.c
index 13f763c9aa..858f581ba3 100644
--- a/examples/fips_validation/fips_validation.c
+++ b/examples/fips_validation/fips_validation.c
@@ -33,6 +33,8 @@ get_file_line(void)
 
if (loc >= MAX_LINE_CHAR - 1)
return -ENOMEM;
+   if (c == '\r')
+   continue;
if (c == '\n')
break;
line[loc++] = c;
-- 
2.25.1



[dpdk-dev] [PATCH 1/3] examples/fips_validation: fix buffer overflow

2020-10-06 Thread Olivier Matz
If the file name is larger than MAX_STRING_SIZE (64), strcpy()
will overwrite the content of memory.

Replace strcpy() by rte_strscpy(), check its return value, and
increase file_name size to 256.

Fixes: 3d0fad56b74a ("examples/fips_validation: add crypto FIPS application")
Cc: sta...@dpdk.org

Signed-off-by: Olivier Matz 
---
 examples/fips_validation/fips_validation.c | 12 ++--
 examples/fips_validation/fips_validation.h |  3 ++-
 2 files changed, 12 insertions(+), 3 deletions(-)

diff --git a/examples/fips_validation/fips_validation.c 
b/examples/fips_validation/fips_validation.c
index 9bdf257b8b..13f763c9aa 100644
--- a/examples/fips_validation/fips_validation.c
+++ b/examples/fips_validation/fips_validation.c
@@ -281,7 +281,11 @@ fips_test_init(const char *req_file_path, const char 
*rsp_file_path,
 
fips_test_clear();
 
-   strcpy(info.file_name, req_file_path);
+   if (rte_strscpy(info.file_name, req_file_path,
+   sizeof(info.file_name)) < 0) {
+   RTE_LOG(ERR, USER1, "Path %s too long\n", req_file_path);
+   return -EINVAL;
+   }
info.algo = FIPS_TEST_ALGO_MAX;
if (parse_file_type(req_file_path) < 0) {
RTE_LOG(ERR, USER1, "File %s type not supported\n",
@@ -307,7 +311,11 @@ fips_test_init(const char *req_file_path, const char 
*rsp_file_path,
return -ENOMEM;
}
 
-   strlcpy(info.device_name, device_name, sizeof(info.device_name));
+   if (rte_strscpy(info.device_name, device_name,
+   sizeof(info.device_name)) < 0) {
+   RTE_LOG(ERR, USER1, "Device name %s too long\n", device_name);
+   return -EINVAL;
+   }
 
if (fips_test_parse_header() < 0) {
RTE_LOG(ERR, USER1, "Failed parsing header\n");
diff --git a/examples/fips_validation/fips_validation.h 
b/examples/fips_validation/fips_validation.h
index 75fa555fa6..deba83eada 100644
--- a/examples/fips_validation/fips_validation.h
+++ b/examples/fips_validation/fips_validation.h
@@ -14,6 +14,7 @@
 #define MAX_NB_TESTS   10240
 #define MAX_BUF_SIZE   2048
 #define MAX_STRING_SIZE64
+#define MAX_FILE_NAME_SIZE 256
 #define MAX_DIGEST_SIZE64
 
 #define POSITIVE_TEST  0
@@ -164,7 +165,7 @@ struct fips_test_interim_info {
uint32_t vec_start_off;
uint32_t nb_vec_lines;
char device_name[MAX_STRING_SIZE];
-   char file_name[MAX_STRING_SIZE];
+   char file_name[MAX_FILE_NAME_SIZE];
 
union {
struct aesavs_interim_data aes_data;
-- 
2.25.1



[dpdk-dev] [PATCH 0/3] examples/fips_validation: misc fixes

2020-10-06 Thread Olivier Matz
Seen while trying to test this example: the first two patches
are fixes (buffer overflow and an incorrect parsing). The last
one makes it possible to only run the self test, without passing
a req file.

Olivier Matz (3):
  examples/fips_validation: fix buffer overflow
  examples/fips_validation: ignore \r in input files
  examples/fips_validation: support self-test only

 examples/fips_validation/fips_validation.c | 14 --
 examples/fips_validation/fips_validation.h |  3 ++-
 examples/fips_validation/main.c| 22 --
 3 files changed, 34 insertions(+), 5 deletions(-)

-- 
2.25.1



Re: [dpdk-dev] [PATCH] examples/vm_power: fix 32-bit build

2020-10-06 Thread Thomas Monjalon
05/10/2020 19:02, Ferruh Yigit:
> Compiler version:
> gcc 10.2.1 "cc (GCC) 10.2.1 20200723 (Red Hat 10.2.1-1)"
> 
> Build error:
> ../examples/vm_power_manager/guest_cli/vm_power_cli_guest.c:346:23:
>  warning: format ‘%ld’ expects argument of type ‘long int’, but
>   argument 4 has type ‘uint64_t’ {aka ‘long long unsigned int’}
>   [-Wformat=]
>   346 |cmdline_printf(cl, "Capabilities of [%d] vcore are:"
>   |   ^
> ..
>   349 |  pkt_caps_list.turbo[i],
>   |  ~~
>   | |
>   | uint64_t {aka long long unsigned int}
> 
> Fixes: 07525d1a047a ("examples/vm_power: send capabilities request from 
> guest")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Ferruh Yigit 
> ---
> Cc: David Marchand 
> Cc: David Hunt 
> 
> Fixed patch is not new, not sure why getting the error now but was not
> getting it before, do you have any clue?
> Or is the CI having this error?

Probably because we don't test this example in 32-bit env.
Note: it requires 32-bit libvirt.

Applied with spaces around PRId64, thanks




Re: [dpdk-dev] [PATCH] examples/vm_power: fix 32-bit build

2020-10-06 Thread David Marchand
On Mon, Oct 5, 2020 at 8:57 PM Ferruh Yigit  wrote:
>
> Compiler version:
> gcc 10.2.1 "cc (GCC) 10.2.1 20200723 (Red Hat 10.2.1-1)"
>
> Build error:
> ../examples/vm_power_manager/guest_cli/vm_power_cli_guest.c:346:23:
>  warning: format ‘%ld’ expects argument of type ‘long int’, but
> argument 4 has type ‘uint64_t’ {aka ‘long long unsigned int’}
> [-Wformat=]
>   346 |cmdline_printf(cl, "Capabilities of [%d] vcore are:"
>   |   ^
> ..
>   349 |  pkt_caps_list.turbo[i],
>   |  ~~
>   | |
>   | uint64_t {aka long long unsigned int}
>
> Fixes: 07525d1a047a ("examples/vm_power: send capabilities request from 
> guest")
> Cc: sta...@dpdk.org
>
> Signed-off-by: Ferruh Yigit 
> ---
> Cc: David Marchand 
> Cc: David Hunt 
>
> Fixed patch is not new, not sure why getting the error now but was not
> getting it before, do you have any clue?
> Or is the CI having this error?

Personally, I do not have a libvirt for 32 bits on my build system.

In Intel CI, I can see a 32 bits job, but I guess there is no libvirt
for i386 available in it.
http://mails.dpdk.org/archives/test-report/2020-October/156516.html

The UNH lab has no 32 bits job.

In Travis, we have a "static" cross compilation job.
But static jobs only build l2fwd and l3fwd examples.

We could switch this job to shared to get all examples built, but we
would still need the libvirt devel headers + libs for i386 to get this
specific example to get built (I tried, but did not find the right
magic to install a i386 package).

We can at least add libvirt-devel for 64 bits in Travis.


> ---
>  examples/vm_power_manager/guest_cli/vm_power_cli_guest.c | 4 ++--
>  1 file changed, 2 insertions(+), 2 deletions(-)
>
> diff --git a/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c 
> b/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c
> index 96c1a1ff69..58029f8fb9 100644
> --- a/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c
> +++ b/examples/vm_power_manager/guest_cli/vm_power_cli_guest.c
> @@ -344,13 +344,13 @@ cmd_query_caps_list_parsed(void *parsed_result,
> unsigned int i;
> for (i = 0; i < pkt_caps_list.num_vcpu; ++i)
> cmdline_printf(cl, "Capabilities of [%d] vcore are:"
> -   " turbo possibility: %ld, is priority 
> core: %ld.\n",
> +   " turbo possibility: %"PRId64", is 
> priority core: %"PRId64".\n",
> i,
> pkt_caps_list.turbo[i],
> pkt_caps_list.priority[i]);
> } else {
> cmdline_printf(cl, "Capabilities of [%d] vcore are:"
> -   " turbo possibility: %ld, is priority core: 
> %ld.\n",
> +   " turbo possibility: %"PRId64", is priority 
> core: %"PRId64".\n",
> lcore_id,
> pkt_caps_list.turbo[lcore_id],
> pkt_caps_list.priority[lcore_id]);
> --
> 2.26.2
>

Installed a libvirt-devel for i686 on a fc32, reproduced and tested the fix.
Tested-by: David Marchand 


-- 
David Marchand



Re: [dpdk-dev] [dpdk-dev v2 1/2] vhost: add backend type in driver start

2020-10-06 Thread Maxime Coquelin
Hi Fan,

On 10/2/20 5:36 PM, Fan Zhang wrote:
> This patch adds an internal driver start function with a newly
> added backend type identifier as parameter. With this way
> different built-in driver types (net, crypto) can be identified.
> 
> Signed-off-by: Fan Zhang 
> ---
>  lib/librte_vhost/socket.c | 39 ---
>  lib/librte_vhost/vhost_user.h |  8 +++
>  2 files changed, 39 insertions(+), 8 deletions(-)
> 
> diff --git a/lib/librte_vhost/socket.c b/lib/librte_vhost/socket.c
> index 0169d3648..a53e536ac 100644
> --- a/lib/librte_vhost/socket.c
> +++ b/lib/librte_vhost/socket.c
> @@ -38,7 +38,7 @@ struct vhost_user_socket {
>   bool is_server;
>   bool reconnect;
>   bool iommu_support;
> - bool use_builtin_virtio_net;
> + enum virtio_backend_type backend_type;
>   bool extbuf;
>   bool linearbuf;
>   bool async_copy;
> @@ -224,7 +224,9 @@ vhost_user_add_connection(int fd, struct 
> vhost_user_socket *vsocket)
>   size = strnlen(vsocket->path, PATH_MAX);
>   vhost_set_ifname(vid, vsocket->path, size);
>  
> - vhost_set_builtin_virtio_net(vid, vsocket->use_builtin_virtio_net);
> + vhost_set_builtin_virtio_net(vid,
> + vsocket->backend_type == VIRTIO_DEV_BUILTIN_NET ?
> + true : false);
>  
>   vhost_attach_vdpa_device(vid, vsocket->vdpa_dev);
>  
> @@ -632,7 +634,7 @@ rte_vhost_driver_disable_features(const char *path, 
> uint64_t features)
>   pthread_mutex_lock(&vhost_user.mutex);
>   vsocket = find_vhost_user_socket(path);
>  
> - /* Note that use_builtin_virtio_net is not affected by this function
> + /* Note that backend type is not affected by this function
>* since callers may want to selectively disable features of the
>* built-in vhost net device backend.
>*/
> @@ -681,7 +683,8 @@ rte_vhost_driver_set_features(const char *path, uint64_t 
> features)
>   /* Anyone setting feature bits is implementing their own vhost
>* device backend.
>*/
> - vsocket->use_builtin_virtio_net = false;
> + if (vsocket->backend_type == VIRTIO_DEV_BUILTIN_NET)
> + vsocket->backend_type = VIRTIO_DEV_UNKNOWN;
>   }
>   pthread_mutex_unlock(&vhost_user.mutex);
>  
> @@ -899,7 +902,7 @@ rte_vhost_driver_register(const char *path, uint64_t 
> flags)
>* rte_vhost_driver_set_features(), which will overwrite following
>* two values.
>*/
> - vsocket->use_builtin_virtio_net = true;
> + vsocket->backend_type = VIRTIO_DEV_BUILTIN_NET;
>   vsocket->supported_features = VIRTIO_NET_SUPPORTED_FEATURES;
>   vsocket->features   = VIRTIO_NET_SUPPORTED_FEATURES;
>   vsocket->protocol_features  = VHOST_USER_PROTOCOL_FEATURES;
> @@ -1117,10 +1120,17 @@ vhost_driver_callback_get(const char *path)
>  }
>  
>  int
> -rte_vhost_driver_start(const char *path)
> +vhost_driver_start(const char *path, enum virtio_backend_type backend_type)
>  {
>   struct vhost_user_socket *vsocket;
>   static pthread_t fdset_tid;
> + int ret;
> +
> + if (backend_type <= VIRTIO_DEV_UNKNOWN ||
> + backend_type > VIRTIO_DEV_BUILTIN_CRYPTO) {
> + VHOST_LOG_CONFIG(ERR, "Wrong backend type\n");
> + return -1;
> + }
>  
>   pthread_mutex_lock(&vhost_user.mutex);
>   vsocket = find_vhost_user_socket(path);
> @@ -1153,7 +1163,20 @@ rte_vhost_driver_start(const char *path)
>   }
>  
>   if (vsocket->is_server)
> - return vhost_user_start_server(vsocket);
> + ret = vhost_user_start_server(vsocket);
>   else
> - return vhost_user_start_client(vsocket);
> + ret = vhost_user_start_client(vsocket);
> +
> + if (ret < 0)
> + return ret;
> +
> + vsocket->backend_type = backend_type;
> +
> + return 0;

Looking at it again, I think patch 1 is not necessary.
Patch 2 should be enough and would be easier to backport.

What do you think?

Regards,
Maxime

> +}
> +
> +int
> +rte_vhost_driver_start(const char *path)
> +{
> + return vhost_driver_start(path, VIRTIO_DEV_BUILTIN_NET);
>  }
> diff --git a/lib/librte_vhost/vhost_user.h b/lib/librte_vhost/vhost_user.h
> index 16fe03f88..9f526e676 100644
> --- a/lib/librte_vhost/vhost_user.h
> +++ b/lib/librte_vhost/vhost_user.h
> @@ -158,6 +158,12 @@ typedef struct VhostUserMsg {
>  /* The version of the protocol we support */
>  #define VHOST_USER_VERSION0x1
>  
> +/* virtio backend types */
> +enum virtio_backend_type {
> + VIRTIO_DEV_UNKNOWN = 0, /* Likely external */
> + VIRTIO_DEV_BUILTIN_NET, /* Virtio-net device */
> + VIRTIO_DEV_BUILTIN_CRYPTO, /* Virtio-crypto device */
> +};
>  
>  /* vhost_user.c */
>  int vhost_user_msg_handler(int vid, int fd);
> @@ -167,5 +173,7 @@ int vhost_user_iotlb_miss(struct virtio_net *dev, 
> uint64_t iova, u

Re: [dpdk-dev] [PATCH v1] event/sw: performance improvements

2020-10-06 Thread Van Haaren, Harry
> -Original Message-
> From: Jerin Jacob 
> Sent: Monday, October 5, 2020 5:35 PM
> To: Nicolau, Radu 
> Cc: Honnappa Nagarahalli ; Richardson, Bruce
> ; Ananyev, Konstantin
> ; Van Haaren, Harry
> ; dev@dpdk.org; jer...@marvell.com; nd
> 
> Subject: Re: [dpdk-dev] [PATCH v1] event/sw: performance improvements
> 
> On Tue, Sep 29, 2020 at 2:32 PM Nicolau, Radu  wrote:
> >
> >
> > On 9/28/2020 5:02 PM, Honnappa Nagarahalli wrote:
> > > 
> > >>> Add minimum burst throughout the scheduler pipeline and a flush counter.
> > >>> Replace ring API calls with local single threaded implementation where
> > >>> possible.
> > >>>
> > >>> Signed-off-by: Radu Nicolau mailto:radu.nico...@intel.com
> > >>>
> > >>> Thanks for the patch, a few comments inline.
> > >>>
> 
> > >>> Why not make these APIs part of the rte_ring library? You could further
> > >> optimize them by keeping the indices on the same cacheline.
> > >>> I'm not sure there is any need for non thread-safe rings outside this
> > >> particular case.
> > >>> [Honnappa] I think if we add the APIs, we will find the use cases.
> > >>> But, more than that, I understand that rte_ring structure is exposed to 
> > >>> the
> > >> application. The reason for doing that is the inline functions that 
> > >> rte_ring
> > >> provides. IMO, we should still maintain modularity and should not use the
> > >> internals of the rte_ring structure outside of the library.
> > >>> +1 to that.
> > >>>
> > >>> BTW, is there any real perf benefit from such micor-optimisation?
> > >> I'd tend to view these as use-case specific, and I'm not sure we should 
> > >> clutter
> > >> up the ring library with yet more functions, especially since they can't 
> > >> be
> > >> mixed with the existing enqueue/dequeue functions, since they don't use
> > >> the head pointers.
> > > IMO, the ring library is pretty organized with the recent addition of 
> > > HTS/RTS
> modes. This can be one of the modes and should allow us to use the existing
> functions (though additional functions are required as well).
> > > The other concern I have is, this implementation can be further optimized 
> > > by
> using a single cache line for the pointers. It uses 2 cache lines just 
> because of the
> layout of the rte_ring structure.
> > > There was a question earlier about the performance improvements of this
> patch? Are there any % performance improvements that can be shared?
> > > It is also possible to change the above functions to use the head/tail 
> > > pointers
> from producer or the consumer cache line alone to check for perf differences.
> >
> > I don't have a % for the final improvement for this change alone, but
> > there was some improvement in the memory overhead measurable during
> > development, which very likely resulted in the whole optimization having
> > more headroom.
> >
> > I agree that this may be further optimized, maybe by having a local
> > implementation of a ring-like container instead.
> 
> Have we decided on the next steps for this patch? Is the plan to
> supersede this patch and have different
> one in rte_ring subsystem,

My preference is to merge this version of the patch;
1) The ring helper functions are stripped to the SW PMD usage, and not valid to 
use in the general.
2) Adding static inline APIs in an LTS without extensive doesn't seem a good 
idea.

If Honnappa is OK with the above solution for 20.11, we can see about moving 
the rings part of the
code to rte_ring library location in 21.02, and give ourselves some time to 
settle the usage/API before
the next LTS.

Regards, -Harry



Re: [dpdk-dev] [PATCH] eal/windows: export all built functions for clang

2020-10-06 Thread Tal Shnaiderman
> Subject: Re: [dpdk-dev] [PATCH] eal/windows: export all built functions for
> clang
> 
> External email: Use caution opening links or attachments
> 
> 
> > From: Ranjit Menon 
> > Subject: Re: [PATCH] eal/windows: export all built functions for clang
> >
> > Hi Tal,
> >
> > On 10/2/2020 12:18 PM, Tal Shnaiderman wrote:
> > > export for clang build all the functions currently built on Windows
> > > and listed in rte_eal_version.map by adding them to
> > > rte_eal_exports.def.
> > >
> > > Signed-off-by: Tal Shnaiderman 
> > > ---
> >
> > While you are at this, can we sort and re-order the functions 
> > alphabetically?
> >
> > For instance, I can see some rte_fbarray*, rte_mem*, per_lcore*
> > functions in multiple locations.
> >
> > Also as an aside, Bruce (in cc:) had suggested in a private thread
> > that we should start looking into auto-generating this .def file now
> > that it has most of the functions defined in the .map file.
> >
> > ranjit m.
> 
> Hi Ranjit,
> 
> I tried to keep the file structure similar to the original rte_eal_version.map
> file, in the map file each DPDK release exported additional functions so the
> whole file isn't sorted but each release section is.
> 
> I not sure however if you can add the comments separating the sections to
> the def file.
> 
> Regarding autogenerating this file, there is actually an open task on it, MSFT
> team are on it and DmitryK even suggested a way to generate it using
> DUMPBIN, however until this is done we should have all the exported
> functions updated in the def file.

Hi Ranjit,

Any other concerns regarding this patch? 

If not can you Ack it?

Thanks,

Tal.


Re: [dpdk-dev] [PATCH] maintainers: fix missing title for examples

2020-10-06 Thread Thomas Monjalon
> The devtools/get-maintainer.sh script does not work with examples
> because there is no title line between them: it returns a longer list
> than expected.
> 
> Add the missing titles for each example to fix this behavior.
> 
> Signed-off-by: Olivier Matz 

Applied, thanks

(reworded VMDQ to VMDq)




Re: [dpdk-dev] [dpdk-dev v2 2/2] vhost/crypto: fix feature negotiation

2020-10-06 Thread Maxime Coquelin



On 10/2/20 5:36 PM, Fan Zhang wrote:
> This patch fixes the feature negotiation for vhost crypto during
> initialization. The patch uses the newly created driver start
> function to inform the driver type with the fixed vhost features.
> In addtion the patch provides a new API specifically used by
> the application to start a vhost-crypto driver.
> 
> Fixes: 939066d96563 ("vhost/crypto: add public function implementation")
> Cc: roy.fan.zh...@intel.com
> 
> Signed-off-by: Fan Zhang 
> ---
>  examples/vhost_crypto/main.c   |  3 +-
>  lib/librte_vhost/rte_vhost_crypto.h| 12 
>  lib/librte_vhost/rte_vhost_version.map |  1 +
>  lib/librte_vhost/vhost_crypto.c| 41 +-
>  4 files changed, 42 insertions(+), 15 deletions(-)
> 
> diff --git a/examples/vhost_crypto/main.c b/examples/vhost_crypto/main.c
> index d78fd9b81..11ad49159 100644
> --- a/examples/vhost_crypto/main.c
> +++ b/examples/vhost_crypto/main.c
> @@ -598,7 +598,8 @@ main(int argc, char *argv[])
>   rte_vhost_driver_callback_register(lo->socket_files[j],
>   &virtio_crypto_device_ops);
>  
> - ret = rte_vhost_driver_start(lo->socket_files[j]);
> + ret = rte_vhost_crypto_driver_start(
> + lo->socket_files[j]);
>   if (ret < 0)  {
>   RTE_LOG(ERR, USER1, "failed to start vhost.\n");
>   goto error_exit;
> diff --git a/lib/librte_vhost/rte_vhost_crypto.h 
> b/lib/librte_vhost/rte_vhost_crypto.h
> index b54d61db6..c809c46a2 100644
> --- a/lib/librte_vhost/rte_vhost_crypto.h
> +++ b/lib/librte_vhost/rte_vhost_crypto.h
> @@ -20,6 +20,18 @@ enum rte_vhost_crypto_zero_copy {
>   RTE_VHOST_CRYPTO_MAX_ZERO_COPY_OPTIONS
>  };
>  
> +/**
> + * Start vhost crypto driver
> + *
> + * @param path
> + *  The vhost-user socket file path
> + * @return
> + *  0 on success, -1 on failure
> + */
> +__rte_experimental
> +int
> +rte_vhost_crypto_driver_start(const char *path);
> +
>  /**
>   *  Create Vhost-crypto instance
>   *
> diff --git a/lib/librte_vhost/rte_vhost_version.map 
> b/lib/librte_vhost/rte_vhost_version.map
> index 55e98e557..9183d6f2f 100644
> --- a/lib/librte_vhost/rte_vhost_version.map
> +++ b/lib/librte_vhost/rte_vhost_version.map
> @@ -55,6 +55,7 @@ EXPERIMENTAL {
>   rte_vhost_driver_get_protocol_features;
>   rte_vhost_driver_get_queue_num;
>   rte_vhost_crypto_create;
> + rte_vhost_crypto_driver_start;
>   rte_vhost_crypto_free;
>   rte_vhost_crypto_fetch_requests;
>   rte_vhost_crypto_finalize_requests;
> diff --git a/lib/librte_vhost/vhost_crypto.c b/lib/librte_vhost/vhost_crypto.c
> index e08f9c6d7..6195958d2 100644
> --- a/lib/librte_vhost/vhost_crypto.c
> +++ b/lib/librte_vhost/vhost_crypto.c
> @@ -35,13 +35,12 @@
>  #define VC_LOG_DBG(fmt, args...)
>  #endif
>  
> -#define VIRTIO_CRYPTO_FEATURES ((1 << VIRTIO_F_NOTIFY_ON_EMPTY) |\
> - (1 << VIRTIO_RING_F_INDIRECT_DESC) |\
> - (1 << VIRTIO_RING_F_EVENT_IDX) |\
> - (1 << VIRTIO_CRYPTO_SERVICE_CIPHER) |   \
> - (1 << VIRTIO_CRYPTO_SERVICE_MAC) |  \
> - (1 << VIRTIO_NET_F_CTRL_VQ) |   \
> - (1 << VHOST_USER_PROTOCOL_F_CONFIG))
> +#define VIRTIO_CRYPTO_FEATURES ((1ULL << VIRTIO_F_NOTIFY_ON_EMPTY) | \
> + (1ULL << VIRTIO_RING_F_INDIRECT_DESC) | \
> + (1ULL << VIRTIO_RING_F_EVENT_IDX) | \
> + (1ULL << VIRTIO_NET_F_CTRL_VQ) |\
> + (1ULL << VIRTIO_F_VERSION_1) |  \
> + (1ULL << VHOST_USER_F_PROTOCOL_FEATURES))
>  
>  #define IOVA_TO_VVA(t, r, a, l, p)   \
>   ((t)(uintptr_t)vhost_iova_to_vva(r->dev, r->vq, a, l, p))
> @@ -1400,6 +1399,27 @@ vhost_crypto_complete_one_vm_requests(struct 
> rte_crypto_op **ops,
>   return processed;
>  }
>  
> +int
> +rte_vhost_crypto_driver_start(const char *path)
> +{
> + uint64_t protocol_features;
> + int ret;
> +
> + ret = rte_vhost_driver_set_features(path, VIRTIO_CRYPTO_FEATURES);
> + if (ret)
> + return -1;

As rte_vhost_driver_set_features is now called on time,
use_builtin_virtio_net is set to false before the connection is
established.

So it should be enough.

> + ret = rte_vhost_driver_get_protocol_features(path, &protocol_features);
> + if (ret)
> + return -1;
> + protocol_features |= (1ULL << VHOST_USER_PROTOCOL_F_CONFIG);
> + ret = rte_vhost_driver_set_protocol_features(path, protocol_features);
> + if (ret)
> + return -1;
> +
> + return vhost_driver_start(path, VIRTIO_DEV_BUILTIN_CRYPTO);

We just have to remove the extra param.

I will to the

Re: [dpdk-dev] [PATCH v2] net: calculate checksums for packets with IPv4 options

2020-10-06 Thread Olivier Matz
On Mon, Oct 05, 2020 at 07:39:45PM -0700, Stephen Hemminger wrote:
> On Tue, 06 Oct 2020 00:55:19 +0200
> Thomas Monjalon  wrote:
> >
> > On 9/1/20 12:47 PM, Michael Pfeiffer wrote:
> > > Currently, rte_ipv4_cksum() and rte_ipv4_udptcp_cksum() assume all IPv4
> > > headers have sizeof(struct rte_ipv4_hdr) bytes. This is not true for
> > > those (rare) packets with IPv4 options. Thus, both IPv4 and TCP/UDP
> > > checksums are calculated wrong.
> > >
> > > This patch fixes the issue by using the actual IPv4 header length from
> > > the packet's IHL field.
> > >
> > > Signed-off-by: Michael Pfeiffer 

Acked-by: Olivier Matz 

> > > -   cksum = rte_raw_cksum(ipv4_hdr, sizeof(struct rte_ipv4_hdr));
> > > +   cksum = rte_raw_cksum(ipv4_hdr, (ipv4_hdr->version_ihl & 0xf) * 
> > > 4);
> >
> > Truly naive questions:
> > - doesn't it deserve a static inline function rte_ipv4_hdr_len()?
>
> Makes sense to have that.

+1

However it could be in another patch: there are ~15 places where it
could be replaced in dpdk.


Re: [dpdk-dev] [PATCH v2 1/2] eventdev: eventdev: express DLB/DLB2 PMD constraints

2020-10-06 Thread Van Haaren, Harry
> -Original Message-
> From: McDaniel, Timothy 
> Sent: Monday, October 5, 2020 9:28 PM
> To: Hemant Agrawal ; Nipun Gupta
> ; Mattias Rönnblom ;
> Jerin Jacob ; Pavan Nikhilesh ;
> Ma, Liang J ; Mccarthy, Peter 
> ;
> Van Haaren, Harry ; Rao, Nikhil
> ; Ray Kinsella ; Neil Horman
> 
> Cc: dev@dpdk.org; Carrillo, Erik G ; Eads, Gage
> 
> Subject: [PATCH v2 1/2] eventdev: eventdev: express DLB/DLB2 PMD constraints
> 
> This commit implements the eventdev ABI changes required by
> the DLB PMD.
> 
> Signed-off-by: Timothy McDaniel 
> ---

I think patches 1/2 and 2/2 will have to be merged into a single patch.
Reason is that compilation fails after 1/2 because apps/examples access
the "disable_implicit_release" field, which is refactored in 1/2.

For review, this split is quite convenient for reviewing - so suggesting we can
merge on apply if no changes are required?


>  drivers/event/sw/sw_evdev.c|  8 ++--
>  drivers/event/sw/sw_evdev_selftest.c   |  6 +--

For SW PMD component and selftests;
Acked-by: Harry van Haaren 


>  static void
> diff --git a/drivers/event/sw/sw_evdev.c b/drivers/event/sw/sw_evdev.c
> index 98dae71..058f568 100644
> --- a/drivers/event/sw/sw_evdev.c
> +++ b/drivers/event/sw/sw_evdev.c
> @@ -175,7 +175,8 @@ sw_port_setup(struct rte_eventdev *dev, uint8_t port_id,
>   }
> 
>   p->inflight_max = conf->new_event_threshold;
> - p->implicit_release = !conf->disable_implicit_release;
> + p->implicit_release = !(conf->event_port_cfg &
> + RTE_EVENT_PORT_CFG_DISABLE_IMPL_REL);
> 
>   /* check if ring exists, same as rx_worker above */
>   snprintf(buf, sizeof(buf), "sw%d_p%u, %s", dev->data->dev_id,
> @@ -508,7 +509,7 @@ sw_port_def_conf(struct rte_eventdev *dev, uint8_t
> port_id,
>   port_conf->new_event_threshold = 1024;
>   port_conf->dequeue_depth = 16;
>   port_conf->enqueue_depth = 16;
> - port_conf->disable_implicit_release = 0;
> + port_conf->event_port_cfg = 0;
>  }
> 
>  static int
> @@ -615,7 +616,8 @@ sw_info_get(struct rte_eventdev *dev, struct
> rte_event_dev_info *info)
> 
>   RTE_EVENT_DEV_CAP_IMPLICIT_RELEASE_DISABLE|
>   RTE_EVENT_DEV_CAP_RUNTIME_PORT_LINK |
>   RTE_EVENT_DEV_CAP_MULTIPLE_QUEUE_PORT |
> - RTE_EVENT_DEV_CAP_NONSEQ_MODE),
> + RTE_EVENT_DEV_CAP_NONSEQ_MODE |
> + RTE_EVENT_DEV_CAP_CARRY_FLOW_ID),
>   };
> 
>   *info = evdev_sw_info;
> diff --git a/drivers/event/sw/sw_evdev_selftest.c
> b/drivers/event/sw/sw_evdev_selftest.c
> index 38c21fa..4a7d823 100644
> --- a/drivers/event/sw/sw_evdev_selftest.c
> +++ b/drivers/event/sw/sw_evdev_selftest.c
> @@ -172,7 +172,6 @@ create_ports(struct test *t, int num_ports)
>   .new_event_threshold = 1024,
>   .dequeue_depth = 32,
>   .enqueue_depth = 64,
> - .disable_implicit_release = 0,
>   };
>   if (num_ports > MAX_PORTS)
>   return -1;
> @@ -1227,7 +1226,6 @@ port_reconfig_credits(struct test *t)
>   .new_event_threshold = 128,
>   .dequeue_depth = 32,
>   .enqueue_depth = 64,
> - .disable_implicit_release = 0,
>   };
>   if (rte_event_port_setup(evdev, 0, &port_conf) < 0) {
>   printf("%d Error setting up port\n", __LINE__);
> @@ -1317,7 +1315,6 @@ port_single_lb_reconfig(struct test *t)
>   .new_event_threshold = 128,
>   .dequeue_depth = 32,
>   .enqueue_depth = 64,
> - .disable_implicit_release = 0,
>   };
>   if (rte_event_port_setup(evdev, 0, &port_conf) < 0) {
>   printf("%d Error setting up port\n", __LINE__);
> @@ -3079,7 +3076,8 @@ worker_loopback(struct test *t, uint8_t
> disable_implicit_release)
>* only be initialized once - and this needs to be set for multiple runs
>*/
>   conf.new_event_threshold = 512;
> - conf.disable_implicit_release = disable_implicit_release;
> + conf.event_port_cfg = disable_implicit_release ?
> + RTE_EVENT_PORT_CFG_DISABLE_IMPL_REL : 0;
> 
>   if (rte_event_port_setup(evdev, 0, &conf) < 0) {
>   printf("Error setting up RX port\n");



Re: [dpdk-dev] [PATCH 1/2] mempool: remove v20 ABI

2020-10-06 Thread Bruce Richardson
On Tue, Oct 06, 2020 at 09:07:49AM +0200, Olivier Matz wrote:
> Remove the deprecated v20 ABI of rte_mempool_populate_iova() and
> rte_mempool_populate_virt().
> 
> Signed-off-by: Olivier Matz 
> ---
>  lib/librte_mempool/meson.build |  2 -
>  lib/librte_mempool/rte_mempool.c   | 79 ++
>  lib/librte_mempool/rte_mempool_version.map |  7 --
>  3 files changed, 5 insertions(+), 83 deletions(-)
> 
Thanks for the cleanup.

Series-acked-by: Bruce Richardson 


Re: [dpdk-dev] [PATCH v2 2/2] eventdev: update app and examples for new eventdev ABI

2020-10-06 Thread Van Haaren, Harry
> -Original Message-
> From: McDaniel, Timothy 
> Sent: Monday, October 5, 2020 9:28 PM
> To: Jerin Jacob ; Van Haaren, Harry
> ; Kovacevic, Marko ;
> Ori Kam ; Richardson, Bruce
> ; Nicolau, Radu ; Akhil
> Goyal ; Kantecki, Tomasz ;
> Sunil Kumar Kori ; Pavan Nikhilesh
> 
> Cc: dev@dpdk.org; Carrillo, Erik G ; Eads, Gage
> ; hemant.agra...@nxp.com
> Subject: [PATCH v2 2/2] eventdev: update app and examples for new eventdev ABI
> 
> Several data structures and constants changed, or were added,
> in the previous patch.  This commit updates the dependent
> apps and examples to use the new ABI.
> 
> Signed-off-by: Timothy McDaniel 
> ---

With this patch applied, the compilation works fine, however runtime fails.
Note that there is a dependency to the following fix which Timothy upstreamed: 
http://patches.dpdk.org/patch/79713/

The above linked patch increases the CTF trace size, and fixes the following 
error:
./dpdk-test
EAL: __rte_trace_point_emit_field():442 CTF field is too long
EAL: __rte_trace_point_register():468 missing rte_trace_emit_header() in 
register fn


>  app/test-eventdev/evt_common.h | 11 
>  app/test-eventdev/test_order_atq.c | 28 +++--
>  app/test-eventdev/test_order_common.c  |  1 +
>  app/test-eventdev/test_order_queue.c   | 29 
> --
>  app/test/test_eventdev.c   |  4 +--
>  .../eventdev_pipeline/pipeline_worker_generic.c|  6 +++--
>  examples/eventdev_pipeline/pipeline_worker_tx.c|  1 +
>  examples/l2fwd-event/l2fwd_event_generic.c |  7 --
>  examples/l2fwd-event/l2fwd_event_internal_port.c   |  6 +++--
>  examples/l3fwd/l3fwd_event_generic.c   |  7 --
>  examples/l3fwd/l3fwd_event_internal_port.c |  6 +++--
>  11 files changed, 80 insertions(+), 26 deletions(-)
> 
> diff --git a/app/test-eventdev/evt_common.h b/app/test-eventdev/evt_common.h
> index f9d7378..a1da1cf 100644
> --- a/app/test-eventdev/evt_common.h
> +++ b/app/test-eventdev/evt_common.h
> @@ -104,6 +104,16 @@ evt_has_all_types_queue(uint8_t dev_id)
>   true : false;
>  }
> 
> +static inline bool
> +evt_has_flow_id(uint8_t dev_id)
> +{
> + struct rte_event_dev_info dev_info;
> +
> + rte_event_dev_info_get(dev_id, &dev_info);
> + return (dev_info.event_dev_cap &
> RTE_EVENT_DEV_CAP_CARRY_FLOW_ID) ?
> + true : false;
> +}
> +
>  static inline int
>  evt_service_setup(uint32_t service_id)
>  {
> @@ -169,6 +179,7 @@ evt_configure_eventdev(struct evt_options *opt, uint8_t
> nb_queues,
>   .dequeue_timeout_ns = opt->deq_tmo_nsec,
>   .nb_event_queues = nb_queues,
>   .nb_event_ports = nb_ports,
> + .nb_single_link_event_port_queues = 0,
>   .nb_events_limit  = info.max_num_events,
>   .nb_event_queue_flows = opt->nb_flows,
>   .nb_event_port_dequeue_depth =
> diff --git a/app/test-eventdev/test_order_atq.c b/app/test-
> eventdev/test_order_atq.c
> index 3366cfc..cfcb1dc 100644
> --- a/app/test-eventdev/test_order_atq.c
> +++ b/app/test-eventdev/test_order_atq.c
> @@ -19,7 +19,7 @@ order_atq_process_stage_0(struct rte_event *const ev)
>  }
> 
>  static int
> -order_atq_worker(void *arg)
> +order_atq_worker(void *arg, const bool flow_id_cap)
>  {
>   ORDER_WORKER_INIT;
>   struct rte_event ev;
> @@ -34,6 +34,9 @@ order_atq_worker(void *arg)
>   continue;
>   }
> 
> + if (!flow_id_cap)
> + ev.flow_id = ev.mbuf->udata64;
> +
>   if (ev.sub_event_type == 0) { /* stage 0 from producer */
>   order_atq_process_stage_0(&ev);
>   while (rte_event_enqueue_burst(dev_id, port, &ev, 1)
> @@ -50,7 +53,7 @@ order_atq_worker(void *arg)
>  }
> 
>  static int
> -order_atq_worker_burst(void *arg)
> +order_atq_worker_burst(void *arg, const bool flow_id_cap)
>  {
>   ORDER_WORKER_INIT;
>   struct rte_event ev[BURST_SIZE];
> @@ -68,6 +71,9 @@ order_atq_worker_burst(void *arg)
>   }
> 
>   for (i = 0; i < nb_rx; i++) {
> + if (!flow_id_cap)
> + ev[i].flow_id = ev[i].mbuf->udata64;
> +
>   if (ev[i].sub_event_type == 0) { /*stage 0 */
>   order_atq_process_stage_0(&ev[i]);
>   } else if (ev[i].sub_event_type == 1) { /* stage 1 */
> @@ -95,11 +101,19 @@ worker_wrapper(void *arg)
>  {
>   struct worker_data *w  = arg;
>   const bool burst = evt_has_burst_mode(w->dev_id);
> -
> - if (burst)
> - return order_atq_worker_burst(arg);
> - else
> - return order_atq_worker(arg);
> + const bool flow_id_cap = evt_has_flow_id(w->dev_id);
> +
> + if (burst) {
> + if (flow_i

Re: [dpdk-dev] [PATCH] build: support i686 target on x86 64-bit hosts

2020-10-06 Thread Thomas Monjalon
25/09/2020 15:27, Lance Richardson:
> Bruce Richardson wrote:
> > On Thu, Sep 24, 2020 at 12:37:42PM -0400, Lance Richardson wrote:
> > > Add meson cross files for building i686 targets using gcc on x86_64
> > > linux hosts.
> > >
> > > Uusage example:
> > >
> > > meson --cross-file config/x86/cross-i686-linux-gcc build-i686
> > > ninja -C build-i686
> > >
> > > Signed-off-by: Lance Richardson 
> > > ---
> > > Red Hat distros use a different name for the 32-bit pkg-config
> > > command from most other distros, maybe there is a better way
> > > to handle this than using separate files.
> > >
> > > Others will probably have better naming suggestions for the files.
> > >
> > Just to note that rather than using cross-files, building i686 on x86_64
> > should be possible by just setting environment variables.
> >
> > For example, on Ubuntu 20.04, this works for me:
> >
> > PKG_CONFIG_LIBDIR=/usr/lib/i386-linux-gnu/pkgconfig CFLAGS='-m32' 
> > LDFLAGS='-m32' meson --werror build-32bit
> > ninja -C build-32bit/
> >
> > For Fedora or Redhat system the PKG_CONFIG_LIBDIR will be different (I
> > think just /usr/lib/pkgconfig), but the rest should work identically.
> 
> Thanks, that does work, although I'll have some trouble remembering it
> for future

Would it help to have this command in devtools/test-meson-builds.sh ?





Re: [dpdk-dev] [dpdk-techboard] [PATCH V5 1/2] dpdk: resolve compiling errors for per-queue stats

2020-10-06 Thread Olivier Matz
Hi,

On Mon, Oct 05, 2020 at 01:23:08PM +0100, Ferruh Yigit wrote:
> On 9/28/2020 4:43 PM, Stephen Hemminger wrote:
> > On Mon, 28 Sep 2020 17:24:26 +0200
> > Thomas Monjalon  wrote:
> > 
> > > 28/09/2020 15:53, Ferruh Yigit:
> > > > On 9/28/2020 10:16 AM, Thomas Monjalon wrote:
> > > > > 28/09/2020 10:59, Ferruh Yigit:
> > > > > > On 9/27/2020 4:16 AM, Min Hu (Connor) wrote:
> > > > > > > From: Huisong Li 
> > > > > > > 
> > > > > > > Currently, only statistics of rx/tx queues with queue_id less than
> > > > > > > RTE_ETHDEV_QUEUE_STAT_CNTRS can be displayed. If there is a 
> > > > > > > certain
> > > > > > > application scenario that it needs to use 256 or more than 256 
> > > > > > > queues
> > > > > > > and display all statistics of rx/tx queue. At this moment, we 
> > > > > > > have to
> > > > > > > change the macro to be equaled to the queue number.
> > > > > > > 
> > > > > > > However, modifying the macro to be greater than 256 will trigger
> > > > > > > many errors and warnings from test-pmd, PMD drivers and 
> > > > > > > librte_ethdev
> > > > > > > during compiling dpdk project. But it is possible and permitted 
> > > > > > > that
> > > > > > > rx/tx queue number is greater than 256 and all statistics of rx/tx
> > > > > > > queue need to be displayed. In addition, the data type of rx/tx 
> > > > > > > queue
> > > > > > > number in rte_eth_dev_configure API is 'uint16_t'. So It is 
> > > > > > > unreasonable
> > > > > > > to use the 'uint8_t' type for variables that control which 
> > > > > > > per-queue
> > > > > > > statistics can be displayed.
> > > > > 
> > > > > The explanation is too much complex and misleading.
> > > > > You mean you cannot increase RTE_ETHDEV_QUEUE_STAT_CNTRS
> > > > > above 256 because it is an 8-bit type?
> > > > > 
> > > > > [...]
> > > > > > > --- a/lib/librte_ethdev/rte_ethdev.h
> > > > > > > +++ b/lib/librte_ethdev/rte_ethdev.h
> > > > > > > int rte_eth_dev_set_tx_queue_stats_mapping(uint16_t port_id,
> > > > > > > - uint16_t tx_queue_id, uint8_t stat_idx);
> > > > > > > + uint16_t tx_queue_id, uint16_t stat_idx);
> > > > > [...]
> > > > > > > int rte_eth_dev_set_rx_queue_stats_mapping(uint16_t port_id,
> > > > > > >  uint16_t rx_queue_id,
> > > > > > > -uint8_t stat_idx);
> > > > > > > +uint16_t stat_idx);
> > > > > [...]
> > > > > > cc'ed tech-board,
> > > > > > 
> > > > > > The patch breaks the ethdev ABI without a deprecation notice from 
> > > > > > previous
> > > > > > release(s).
> > > > > > 
> > > > > > It is mainly a fix to the port_id storage type, which we have 
> > > > > > updated from
> > > > > > uint8_t to uint16_t in past but some seems remained for
> > > > > > 'rte_eth_dev_set_tx_queue_stats_mapping()' &
> > > > > > 'rte_eth_dev_set_rx_queue_stats_mapping()' APIs.
> > > > > 
> > > > > No, it is not related to the port id, but the number of limited stats.
> > > > 
> > > > Right, it is not related to the port id, it is fixing the storage type 
> > > > for index
> > > > used to map the queue stats.
> > > > > > Since the ethdev library already heavily breaks the ABI this 
> > > > > > release, I am for
> > > > > > getting this fix, instead of waiting the fix for one more year.
> > > > > 
> > > > > If stats can be managed for more than 256 queues, I think it means
> > > > > it is not limited. In this case, we probably don't need the API
> > > > > *_queue_stats_mapping which was invented for a limitation of ixgbe.
> > > > > 
> > > > > The problem is probably somewhere else (in testpmd),
> > > > > that's why I am against this patch.
> > > > 
> > > > This patch is not to fix queue stats mapping, I agree there are 
> > > > problems related
> > > > to it, already shared as comment to this set.
> > > > 
> > > > But this patch is to fix the build errors when 
> > > > 'RTE_ETHDEV_QUEUE_STAT_CNTRS'
> > > > needs to set more than 255. Where the build errors seems around the
> > > > stats_mapping APIs.
> > > 
> > > It is not said this API is supposed to manage more than 256 queues 
> > > mapping.
> > > In general we should not need this API.
> > > I think it is solving the wrong problem.
> > 
> > 
> > The original API is a band aid for the limited number of statistics counters
> > in the Intel IXGBE hardware. It crept into to the DPDK as an API. I would 
> > rather
> > have per-queue statistics and make ixgbe say "not supported"
> > 
> 
> The current issue is not directly related to '*_queue_stats_mapping' APIs.
> 
> Problem is not able to set 'RTE_ETHDEV_QUEUE_STAT_CNTRS' > 255.
> User may need to set the 'RTE_ETHDEV_QUEUE_STAT_CNTRS' > 255, since it is
> used to define size of the stats counter.
> "uint64_t q_ipackets[RTE_ETHDEV_QUEUE_STAT_CNTRS];"
> 
> When 'RTE_ETHDEV_QUEUE_STAT_CNTRS' > 255, it gives multiple build errors,
> the one in the ethdev is like [1].
> 
> This can be fixed two ways,
> a) increase the size of 'stat_id

Re: [dpdk-dev] [dpdk-dev v2 2/2] vhost/crypto: fix feature negotiation

2020-10-06 Thread Zhang, Roy Fan
Hi Maxime,

Thanks I will verify it after you applied the patch.

Regards,
Fan

> -Original Message-
> From: Maxime Coquelin 
> Sent: Tuesday, October 6, 2020 9:10 AM
> To: Zhang, Roy Fan ; dev@dpdk.org
> Cc: Xia, Chenbo ; Liu, Changpeng
> ; Yigit, Ferruh ;
> sta...@dpdk.org
> Subject: Re: [dpdk-dev v2 2/2] vhost/crypto: fix feature negotiation
> 
> 
> 
> On 10/2/20 5:36 PM, Fan Zhang wrote:
> > This patch fixes the feature negotiation for vhost crypto during
> > initialization. The patch uses the newly created driver start
> > function to inform the driver type with the fixed vhost features.
> > In addtion the patch provides a new API specifically used by
> > the application to start a vhost-crypto driver.
> >
> > Fixes: 939066d96563 ("vhost/crypto: add public function implementation")
> > Cc: roy.fan.zh...@intel.com
> >
> > Signed-off-by: Fan Zhang 
> > ---
> >  examples/vhost_crypto/main.c   |  3 +-
> >  lib/librte_vhost/rte_vhost_crypto.h| 12 
> >  lib/librte_vhost/rte_vhost_version.map |  1 +
> >  lib/librte_vhost/vhost_crypto.c| 41 +-
> >  4 files changed, 42 insertions(+), 15 deletions(-)
> >
> > diff --git a/examples/vhost_crypto/main.c
> b/examples/vhost_crypto/main.c
> > index d78fd9b81..11ad49159 100644
> > --- a/examples/vhost_crypto/main.c
> > +++ b/examples/vhost_crypto/main.c
> > @@ -598,7 +598,8 @@ main(int argc, char *argv[])
> > rte_vhost_driver_callback_register(lo-
> >socket_files[j],
> > &virtio_crypto_device_ops);
> >
> > -   ret = rte_vhost_driver_start(lo->socket_files[j]);
> > +   ret = rte_vhost_crypto_driver_start(
> > +   lo->socket_files[j]);
> > if (ret < 0)  {
> > RTE_LOG(ERR, USER1, "failed to start
> vhost.\n");
> > goto error_exit;
> > diff --git a/lib/librte_vhost/rte_vhost_crypto.h
> b/lib/librte_vhost/rte_vhost_crypto.h
> > index b54d61db6..c809c46a2 100644
> > --- a/lib/librte_vhost/rte_vhost_crypto.h
> > +++ b/lib/librte_vhost/rte_vhost_crypto.h
> > @@ -20,6 +20,18 @@ enum rte_vhost_crypto_zero_copy {
> > RTE_VHOST_CRYPTO_MAX_ZERO_COPY_OPTIONS
> >  };
> >
> > +/**
> > + * Start vhost crypto driver
> > + *
> > + * @param path
> > + *  The vhost-user socket file path
> > + * @return
> > + *  0 on success, -1 on failure
> > + */
> > +__rte_experimental
> > +int
> > +rte_vhost_crypto_driver_start(const char *path);
> > +
> >  /**
> >   *  Create Vhost-crypto instance
> >   *
> > diff --git a/lib/librte_vhost/rte_vhost_version.map
> b/lib/librte_vhost/rte_vhost_version.map
> > index 55e98e557..9183d6f2f 100644
> > --- a/lib/librte_vhost/rte_vhost_version.map
> > +++ b/lib/librte_vhost/rte_vhost_version.map
> > @@ -55,6 +55,7 @@ EXPERIMENTAL {
> > rte_vhost_driver_get_protocol_features;
> > rte_vhost_driver_get_queue_num;
> > rte_vhost_crypto_create;
> > +   rte_vhost_crypto_driver_start;
> > rte_vhost_crypto_free;
> > rte_vhost_crypto_fetch_requests;
> > rte_vhost_crypto_finalize_requests;
> > diff --git a/lib/librte_vhost/vhost_crypto.c
> b/lib/librte_vhost/vhost_crypto.c
> > index e08f9c6d7..6195958d2 100644
> > --- a/lib/librte_vhost/vhost_crypto.c
> > +++ b/lib/librte_vhost/vhost_crypto.c
> > @@ -35,13 +35,12 @@
> >  #define VC_LOG_DBG(fmt, args...)
> >  #endif
> >
> > -#define VIRTIO_CRYPTO_FEATURES ((1 << VIRTIO_F_NOTIFY_ON_EMPTY)
> | \
> > -   (1 << VIRTIO_RING_F_INDIRECT_DESC) |
>   \
> > -   (1 << VIRTIO_RING_F_EVENT_IDX) |\
> > -   (1 << VIRTIO_CRYPTO_SERVICE_CIPHER) |
>   \
> > -   (1 << VIRTIO_CRYPTO_SERVICE_MAC) |
>   \
> > -   (1 << VIRTIO_NET_F_CTRL_VQ) |
>   \
> > -   (1 << VHOST_USER_PROTOCOL_F_CONFIG))
> > +#define VIRTIO_CRYPTO_FEATURES ((1ULL <<
> VIRTIO_F_NOTIFY_ON_EMPTY) |   \
> > +   (1ULL << VIRTIO_RING_F_INDIRECT_DESC) |
>   \
> > +   (1ULL << VIRTIO_RING_F_EVENT_IDX) |
>   \
> > +   (1ULL << VIRTIO_NET_F_CTRL_VQ) |\
> > +   (1ULL << VIRTIO_F_VERSION_1) |
>   \
> > +   (1ULL << VHOST_USER_F_PROTOCOL_FEATURES))
> >
> >  #define IOVA_TO_VVA(t, r, a, l, p) \
> > ((t)(uintptr_t)vhost_iova_to_vva(r->dev, r->vq, a, l, p))
> > @@ -1400,6 +1399,27 @@
> vhost_crypto_complete_one_vm_requests(struct rte_crypto_op **ops,
> > return processed;
> >  }
> >
> > +int
> > +rte_vhost_crypto_driver_start(const char *path)
> > +{
> > +   uint64_t protocol_features;
> > +   int ret;
> > +
> > +   ret = rte_vhost_driver_set_features(path,
> VIRTIO_CRYPTO_FEATURES);
> > +   if (ret)
> > +   return -1;
> 
> As rte_vhost_driver_set_features is now called on time,
> use_builtin_virtio_net is set to false before the connection is
> established.
> 
> So it should be enough.
> 
> > +   ret = rte

Re: [dpdk-dev] [PATCH 1/1] eal: increase TRACE CTF SIZE to recommended size

2020-10-06 Thread David Marchand
On Mon, Oct 5, 2020 at 10:16 PM Timothy McDaniel
 wrote:
>
> Increase TRACE_CTF_FIELD_SIZE to 448, the recommended size.

Repeating the same sentence in the title and the commitlog does not
give much info.

Plus, what is this "recommendation"?


> Fixes "CTF field is too long" error when running with trace enabled.

Ok, you hit this limit, but it would help to get some context here.
Looking at this patch in the future, we won't know why it was necessary.


>
> Signed-off-by: Timothy McDaniel 
> ---
>  lib/librte_eal/common/eal_trace.h | 2 +-
>  1 file changed, 1 insertion(+), 1 deletion(-)
>
> diff --git a/lib/librte_eal/common/eal_trace.h 
> b/lib/librte_eal/common/eal_trace.h
> index 92c5951..438c2b7 100644
> --- a/lib/librte_eal/common/eal_trace.h
> +++ b/lib/librte_eal/common/eal_trace.h
> @@ -24,7 +24,7 @@
>
>  #define TRACE_PREFIX_LEN 12
>  #define TRACE_DIR_STR_LEN (sizeof("-mm-dd-AM-HH-MM-SS") + 
> TRACE_PREFIX_LEN)
> -#define TRACE_CTF_FIELD_SIZE 384
> +#define TRACE_CTF_FIELD_SIZE 448
>  #define TRACE_POINT_NAME_SIZE 64
>  #define TRACE_CTF_MAGIC 0xC1FC1FC1
>  #define TRACE_MAX_ARGS 32
> --
> 2.6.4
>


-- 
David Marchand



[dpdk-dev] [PATCH] net/mlx5: disable flow dump for Verbs flow

2020-10-06 Thread Viacheslav Ovsiienko
From: Slava Ovsiienko 

-Original Message-
From: Slava Ovsiienko 
Sent: Sunday, October 4, 2020 19:56
To: Xueming Li ; Matan Azrad 
Cc: dev@dpdk.org; Asaf Penso ; Xueming(Steven) Li 
; sta...@dpdk.org
Subject: RE: [PATCH] net/mlx5: disable flow dump for Verbs flow

> -Original Message-
> From: Xueming Li 
> Sent: Thursday, September 10, 2020 7:26
> To: Matan Azrad ; Slava Ovsiienko 
> 
> Cc: dev@dpdk.org; Asaf Penso ; Xueming(Steven) Li 
> ; sta...@dpdk.org
> Subject: [PATCH] net/mlx5: disable flow dump for Verbs flow
> 
> There was a segment fault when dump flows with device argument of 
> dv_flow_en=0. In such case, Verbs flow engine was enabled and fdb 
> resources were not initialized. It's suggested to use mlx_fs_dump for 
> Verbs flow dump.
> 
> This patch adds verbs engine check, prints warning message and return 
> gracefully.
> 
> Cc: sta...@dpdk.org
> Signed-off-by: Xueming Li 
Acked-by: Viacheslav Ovsiienko 


Re: [dpdk-dev] [PATCH 2/3] examples/fips_validation: ignore \r in input files

2020-10-06 Thread Zhang, Roy Fan
Hi Olivier,

The patch looks ok but the test file link you provided in the patch is CAVS 
5.3. 
As mentioned in https://doc.dpdk.org/guides/sample_app_ug/fips_validation.html, 
the supported CAVS supported version is 21.0 (not latest one by newer than 
5.3). In CAVS 21.0 test files there is no '\r' before '\n' (I suppose this is 
for Windows right).

Regards,
Fan

> -Original Message-
> From: Olivier Matz 
> Sent: Tuesday, October 6, 2020 8:42 AM
> To: dev@dpdk.org
> Cc: Kovacevic, Marko ; Akhil Goyal
> ; Zhang, Roy Fan ; Kusztal,
> ArkadiuszX ; sta...@dpdk.org
> Subject: [PATCH 2/3] examples/fips_validation: ignore \r in input files
> 
> Some test vectors contain '\r' before '\n' (see link). Ignore them.
> 
> Link: https://www.openssl.org/docs/fips/testvectors-linux-2007-10-10.tar.gz
> Fixes: 3d0fad56b74a ("examples/fips_validation: add crypto FIPS application")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Olivier Matz 
> ---
>  examples/fips_validation/fips_validation.c | 2 ++
>  1 file changed, 2 insertions(+)
> 
> diff --git a/examples/fips_validation/fips_validation.c
> b/examples/fips_validation/fips_validation.c
> index 13f763c9aa..858f581ba3 100644
> --- a/examples/fips_validation/fips_validation.c
> +++ b/examples/fips_validation/fips_validation.c
> @@ -33,6 +33,8 @@ get_file_line(void)
> 
>   if (loc >= MAX_LINE_CHAR - 1)
>   return -ENOMEM;
> + if (c == '\r')
> + continue;
>   if (c == '\n')
>   break;
>   line[loc++] = c;
> --
> 2.25.1



Re: [dpdk-dev] [PATCH 1/3] examples/fips_validation: fix buffer overflow

2020-10-06 Thread Zhang, Roy Fan



> -Original Message-
> From: Olivier Matz 
> Sent: Tuesday, October 6, 2020 8:42 AM
> To: dev@dpdk.org
> Cc: Kovacevic, Marko ; Akhil Goyal
> ; Zhang, Roy Fan ; Kusztal,
> ArkadiuszX ; sta...@dpdk.org
> Subject: [PATCH 1/3] examples/fips_validation: fix buffer overflow
> 
> If the file name is larger than MAX_STRING_SIZE (64), strcpy()
> will overwrite the content of memory.
> 
> Replace strcpy() by rte_strscpy(), check its return value, and
> increase file_name size to 256.
> 
> Fixes: 3d0fad56b74a ("examples/fips_validation: add crypto FIPS application")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Olivier Matz 
> ---

Acked-by: Fan Zhang 


Re: [dpdk-dev] [PATCH 0/4] remove experimental tags from APIs

2020-10-06 Thread David Marchand
On Fri, Sep 11, 2020 at 7:46 AM Joyce Kong  wrote:
>
> Experimental tags can be removed from APIs in module: ticket lock,
> MCS lock, pause (wait until equal APIs), and RCU (except for defer
> queue APIs) as they have been aroud four releases.
>
> Joyce Kong (4):
>   eal/ticketlock: remove experimental tag
>   eal/mcslock: remove experimental tag
>   rcu: remove experimental tag from QSBR mechanism
>   eal/pause: remove experimental tag from wait until equal
>
>  MAINTAINERS   |  4 +-
>  lib/librte_eal/include/generic/rte_mcslock.h  | 16 --
>  lib/librte_eal/include/generic/rte_pause.h| 12 -
>  .../include/generic/rte_ticketlock.h  |  9 
>  lib/librte_rcu/rte_rcu_qsbr.h | 52 ---
>  lib/librte_rcu/rte_rcu_version.map| 11 +++-
>  6 files changed, 11 insertions(+), 93 deletions(-)
>

Series applied, thanks.


-- 
David Marchand



Re: [dpdk-dev] [PATCH 3/3] examples/fips_validation: support self-test only

2020-10-06 Thread Zhang, Roy Fan
> -Original Message-
> From: Olivier Matz 
> Sent: Tuesday, October 6, 2020 8:42 AM
> To: dev@dpdk.org
> Cc: Kovacevic, Marko ; Akhil Goyal
> ; Zhang, Roy Fan ; Kusztal,
> ArkadiuszX 
> Subject: [PATCH 3/3] examples/fips_validation: support self-test only
> 
> Make it possible to pass the self-tests when no req path is set.
> 
> Signed-off-by: Olivier Matz 
> ---

Acked-by: Fan Zhang 


Re: [dpdk-dev] [dpdk-dev v2 1/2] vhost: add backend type in driver start

2020-10-06 Thread Zhang, Roy Fan
Hi Maxime,

I think you are right :-)

Regards,
Fan

> -Original Message-
> From: Maxime Coquelin 
> Sent: Tuesday, October 6, 2020 8:54 AM
> To: Zhang, Roy Fan ; dev@dpdk.org
> Cc: Xia, Chenbo ; Liu, Changpeng
> ; Yigit, Ferruh ;
> sta...@dpdk.org
> Subject: Re: [dpdk-dev v2 1/2] vhost: add backend type in driver start
> 
> Hi Fan,
> 
> On 10/2/20 5:36 PM, Fan Zhang wrote:
> > This patch adds an internal driver start function with a newly
> > added backend type identifier as parameter. With this way
> > different built-in driver types (net, crypto) can be identified.
> >
> > Signed-off-by: Fan Zhang 
> > ---
> >  lib/librte_vhost/socket.c | 39 ---
> >  lib/librte_vhost/vhost_user.h |  8 +++
> >  2 files changed, 39 insertions(+), 8 deletions(-)
> >
> > diff --git a/lib/librte_vhost/socket.c b/lib/librte_vhost/socket.c
> > index 0169d3648..a53e536ac 100644
> > --- a/lib/librte_vhost/socket.c
> > +++ b/lib/librte_vhost/socket.c
> > @@ -38,7 +38,7 @@ struct vhost_user_socket {
> > bool is_server;
> > bool reconnect;
> > bool iommu_support;
> > -   bool use_builtin_virtio_net;
> > +   enum virtio_backend_type backend_type;
> > bool extbuf;
> > bool linearbuf;
> > bool async_copy;
> > @@ -224,7 +224,9 @@ vhost_user_add_connection(int fd, struct
> vhost_user_socket *vsocket)
> > size = strnlen(vsocket->path, PATH_MAX);
> > vhost_set_ifname(vid, vsocket->path, size);
> >
> > -   vhost_set_builtin_virtio_net(vid, vsocket->use_builtin_virtio_net);
> > +   vhost_set_builtin_virtio_net(vid,
> > +   vsocket->backend_type ==
> VIRTIO_DEV_BUILTIN_NET ?
> > +   true : false);
> >
> > vhost_attach_vdpa_device(vid, vsocket->vdpa_dev);
> >
> > @@ -632,7 +634,7 @@ rte_vhost_driver_disable_features(const char
> *path, uint64_t features)
> > pthread_mutex_lock(&vhost_user.mutex);
> > vsocket = find_vhost_user_socket(path);
> >
> > -   /* Note that use_builtin_virtio_net is not affected by this function
> > +   /* Note that backend type is not affected by this function
> >  * since callers may want to selectively disable features of the
> >  * built-in vhost net device backend.
> >  */
> > @@ -681,7 +683,8 @@ rte_vhost_driver_set_features(const char *path,
> uint64_t features)
> > /* Anyone setting feature bits is implementing their own
> vhost
> >  * device backend.
> >  */
> > -   vsocket->use_builtin_virtio_net = false;
> > +   if (vsocket->backend_type == VIRTIO_DEV_BUILTIN_NET)
> > +   vsocket->backend_type = VIRTIO_DEV_UNKNOWN;
> > }
> > pthread_mutex_unlock(&vhost_user.mutex);
> >
> > @@ -899,7 +902,7 @@ rte_vhost_driver_register(const char *path,
> uint64_t flags)
> >  * rte_vhost_driver_set_features(), which will overwrite following
> >  * two values.
> >  */
> > -   vsocket->use_builtin_virtio_net = true;
> > +   vsocket->backend_type = VIRTIO_DEV_BUILTIN_NET;
> > vsocket->supported_features =
> VIRTIO_NET_SUPPORTED_FEATURES;
> > vsocket->features   = VIRTIO_NET_SUPPORTED_FEATURES;
> > vsocket->protocol_features  = VHOST_USER_PROTOCOL_FEATURES;
> > @@ -1117,10 +1120,17 @@ vhost_driver_callback_get(const char *path)
> >  }
> >
> >  int
> > -rte_vhost_driver_start(const char *path)
> > +vhost_driver_start(const char *path, enum virtio_backend_type
> backend_type)
> >  {
> > struct vhost_user_socket *vsocket;
> > static pthread_t fdset_tid;
> > +   int ret;
> > +
> > +   if (backend_type <= VIRTIO_DEV_UNKNOWN ||
> > +   backend_type > VIRTIO_DEV_BUILTIN_CRYPTO) {
> > +   VHOST_LOG_CONFIG(ERR, "Wrong backend type\n");
> > +   return -1;
> > +   }
> >
> > pthread_mutex_lock(&vhost_user.mutex);
> > vsocket = find_vhost_user_socket(path);
> > @@ -1153,7 +1163,20 @@ rte_vhost_driver_start(const char *path)
> > }
> >
> > if (vsocket->is_server)
> > -   return vhost_user_start_server(vsocket);
> > +   ret = vhost_user_start_server(vsocket);
> > else
> > -   return vhost_user_start_client(vsocket);
> > +   ret = vhost_user_start_client(vsocket);
> > +
> > +   if (ret < 0)
> > +   return ret;
> > +
> > +   vsocket->backend_type = backend_type;
> > +
> > +   return 0;
> 
> Looking at it again, I think patch 1 is not necessary.
> Patch 2 should be enough and would be easier to backport.
> 
> What do you think?
> 
> Regards,
> Maxime
> 
> > +}
> > +
> > +int
> > +rte_vhost_driver_start(const char *path)
> > +{
> > +   return vhost_driver_start(path, VIRTIO_DEV_BUILTIN_NET);
> >  }
> > diff --git a/lib/librte_vhost/vhost_user.h b/lib/librte_vhost/vhost_user.h
> > index 16fe03f88..9f526e676 100644
> > --- a/lib/librte_vhost/vhost_user.h
> > +++ b/lib/librte_vhost/vhost_user.h
> > @@ -158,6 +158,12 @@ typedef struct VhostUserMsg {
> >  /* The version of the protocol we support */
>

Re: [dpdk-dev] [PATCH v1] bus/fslmc: fix dpio close

2020-10-06 Thread Thomas Monjalon
06/08/2020 16:32, Sachin Saxena (OSS):
> On 28-Jul-20 9:54 PM, rohit@nxp.com wrote:
> > From: Rohit Raj 
> >
> > The current state of the DPIO object should be checked
> > before trying to close/disable the object.
> >
> > Fixes: 293c0ca94c36 ("bus/fslmc: support memory backed portals with QBMAN 
> > 5.0")
> > Cc: sta...@dpdk.org
> >
> > Signed-off-by: Rohit Raj 
> 
> Acked-by: Sachin Saxena 

Applied, thanks





Re: [dpdk-dev] [PATCH] net/memif: use abstract socket address

2020-10-06 Thread Ferruh Yigit

On 10/5/2020 7:17 PM, Stephen Hemminger wrote:

On Mon, 5 Oct 2020 14:09:20 +0100
Ferruh Yigit  wrote:


On 10/5/2020 1:39 PM, Jakub Grajciar wrote:

Abstract socket address has no connection with
filesystem pathnames and the socket dissapears
once all open references are closed.

Memif pmd will use abstract socket address by default.
For backwards compatibility use new argument
'socket-abstract=no'
   


Why this backward compatibility is required? How the end user affected from
swithching to abstract sockets?


It would only matter if mixing applications with different versions.


Since when linux supports abstract sockets, does this switch will cause problem
with old kernel versions?


This is not new, it dates back to Linux 2.4 or earlier.



Is there any benefit of the abstract sockets other than socket cleaned
automatically (I assume for unix sockets it is done when file filesystem
reference removed)?


The big one is that applications don't have to blindly unlink the old filesystem
remnant. This means that if application can't bind it means another application
is still running with that name. So abstract sockets are safer.


Abstract sockets are not pathnames so they get handled differently by security
systems (like SELinux and AppArmor). This can be helpful in containers.



Hi Stephen, thank you for clarification.


Re: [dpdk-dev] [DPDK_KMODS v3] linux/igb_uio: add Makefile to build the kernel module

2020-10-06 Thread Ferruh Yigit

On 10/5/2020 6:57 PM, Hariprasad Govindharajan wrote:

With DPDK 20.11 release, the igb_uio module is no more part of DPDK.
There are use cases where this module is required, for example while
testing the virtual ports in OvS, the virtual ports are bound to
igb_uio module inside a VM. So, this patch provides a Makefile
which can be used to build this module and use as needed.

Before building this module, the user is expected to build the
DPDK using meson build system and make sure that the required
libraries are installed in the path /usr/local

Signed-off-by: Hariprasad Govindharajan 
---
This patch will be part of dpdk-kmods repo
https://git.dpdk.org/dpdk-kmods/
---
v3:
Edited the commit message and corrected the mistakes in the variable
definition
v2:
Added more information to the commit message
---
  linux/igb_uio/Makefile | 8 
  1 file changed, 8 insertions(+)
  create mode 100644 linux/igb_uio/Makefile

diff --git a/linux/igb_uio/Makefile b/linux/igb_uio/Makefile
new file mode 100644
index 000..8be32f0
--- /dev/null
+++ b/linux/igb_uio/Makefile
@@ -0,0 +1,8 @@
+DPDK_HEADERS ?= /usr/local/include
+RTE_KERNELDIR ?= /lib/modules/`uname -r`/build
+
+all:
+   make EXTRA_CFLAGS="-I $(DPDK_HEADERS)" -C $(RTE_KERNELDIR)/ 
M=$(PWD)
+
+clean:
+   make -C $(RTE_KERNELDIR)/ M=$(PWD) clean



Reviewed-by: Ferruh Yigit 


Re: [dpdk-dev] [EXT] Re: [PATCH 1/1] eal: increase TRACE CTF SIZE to recommended size

2020-10-06 Thread Sunil Kumar Kori


Regards
Sunil Kumar Kori

>-Original Message-
>From: David Marchand 
>Sent: Tuesday, October 6, 2020 2:15 PM
>To: Timothy McDaniel 
>Cc: Jerin Jacob Kollanukkaran ; Sunil Kumar Kori
>; dev ; Erik Gabriel Carrillo
>; Gage Eads ; Van Haaren
>Harry 
>Subject: [EXT] Re: [dpdk-dev] [PATCH 1/1] eal: increase TRACE CTF SIZE to
>recommended size
>
>External Email
>
>--
>On Mon, Oct 5, 2020 at 10:16 PM Timothy McDaniel
> wrote:
>>
>> Increase TRACE_CTF_FIELD_SIZE to 448, the recommended size.
>
>Repeating the same sentence in the title and the commitlog does not give
>much info.
>
>Plus, what is this "recommendation"?
When analyzed this issue, only one more byte was needed to fix this issue but 
in future similar issue can occur again.
So increasing this value by 64 bytes which actually equals to a cache line. 
That’s why we have suggested this size.

>
>
>> Fixes "CTF field is too long" error when running with trace enabled.
>
>Ok, you hit this limit, but it would help to get some context here.
>Looking at this patch in the future, we won't know why it was necessary.
>
>
>>
>> Signed-off-by: Timothy McDaniel 
>> ---
>>  lib/librte_eal/common/eal_trace.h | 2 +-
>>  1 file changed, 1 insertion(+), 1 deletion(-)
>>
>> diff --git a/lib/librte_eal/common/eal_trace.h
>> b/lib/librte_eal/common/eal_trace.h
>> index 92c5951..438c2b7 100644
>> --- a/lib/librte_eal/common/eal_trace.h
>> +++ b/lib/librte_eal/common/eal_trace.h
>> @@ -24,7 +24,7 @@
>>
>>  #define TRACE_PREFIX_LEN 12
>>  #define TRACE_DIR_STR_LEN (sizeof("-mm-dd-AM-HH-MM-SS") +
>> TRACE_PREFIX_LEN) -#define TRACE_CTF_FIELD_SIZE 384
>> +#define TRACE_CTF_FIELD_SIZE 448
>>  #define TRACE_POINT_NAME_SIZE 64
>>  #define TRACE_CTF_MAGIC 0xC1FC1FC1
>>  #define TRACE_MAX_ARGS 32
>> --
>> 2.6.4
>>
>
>
>--
>David Marchand



Re: [dpdk-dev] [PATCH v2] regex/mlx5: add dynamic memory registration to datapath

2020-10-06 Thread Yuval Avnery
Fixed in version 3

From: Thomas Monjalon 
Sent: Sunday, October 4, 2020 10:38 PM
To: Yuval Avnery 
Cc: Ori Kam ; dev@dpdk.org 
Subject: Re: [dpdk-dev] [PATCH v2] regex/mlx5: add dynamic memory registration 
to datapath

Hi Yuval,

> From: Yuval Avnery 
>
> Currently job data is being copied to pre-registered buffer.
> To avoid memcpy on the datapath, use dynamic memory registration.
>
> This change will reduce latency when sending regex jobs. The first few
> jobs may have high latency due to registration, but assuming all
> following mbufs will arrive from the same mempool/hugepage, there will
> be no further memory registration.
>
> Signed-off-by: Yuval Avnery 
> Acked-by: Ori Kam 

There is a compilation issue:
drivers/regex/mlx5/mlx5_regex.c:200:
undefined reference to `mlx5_os_set_reg_mr_cb'




Re: [dpdk-dev] [PATCH v3 01/18] eal: add max SIMD bitwidth

2020-10-06 Thread Olivier Matz
Hi Ciara,

Please find some comments below.

On Wed, Sep 30, 2020 at 02:03:57PM +0100, Ciara Power wrote:
> This patch adds a max SIMD bitwidth EAL configuration. The API allows
> for an app to set this value. It can also be set using EAL argument
> --force-max-simd-bitwidth, which will lock the value and override any
> modifications made by the app.
> 
> Signed-off-by: Ciara Power 
> 
> ---
> v3:
>   - Added enum value to essentially disable using max SIMD to choose
> paths, intended for use by ARM SVE.
>   - Fixed parsing bitwidth argument to return an error for values
> greater than uint16_t.
> v2: Added to Doxygen comment for API.
> ---
>  lib/librte_eal/common/eal_common_options.c | 64 ++
>  lib/librte_eal/common/eal_internal_cfg.h   |  8 +++
>  lib/librte_eal/common/eal_options.h|  2 +
>  lib/librte_eal/include/rte_eal.h   | 33 +++
>  lib/librte_eal/rte_eal_version.map |  4 ++
>  5 files changed, 111 insertions(+)
> 
> diff --git a/lib/librte_eal/common/eal_common_options.c 
> b/lib/librte_eal/common/eal_common_options.c
> index a5426e1234..e9117a96af 100644
> --- a/lib/librte_eal/common/eal_common_options.c
> +++ b/lib/librte_eal/common/eal_common_options.c
> @@ -102,6 +102,7 @@ eal_long_options[] = {
>   {OPT_MATCH_ALLOCATIONS, 0, NULL, OPT_MATCH_ALLOCATIONS_NUM},
>   {OPT_TELEMETRY, 0, NULL, OPT_TELEMETRY_NUM},
>   {OPT_NO_TELEMETRY,  0, NULL, OPT_NO_TELEMETRY_NUM },
> + {OPT_FORCE_MAX_SIMD_BITWIDTH, 1, NULL, OPT_FORCE_MAX_SIMD_BITWIDTH_NUM},
>   {0, 0, NULL, 0}
>  };
>  
> @@ -1309,6 +1310,34 @@ eal_parse_iova_mode(const char *name)
>   return 0;
>  }
>  
> +static int
> +eal_parse_simd_bitwidth(const char *arg, bool locked)
> +{
> + char *end;
> + unsigned long bitwidth;
> + int ret;
> + struct internal_config *internal_conf =
> + eal_get_internal_configuration();
> +
> + if (arg == NULL || arg[0] == '\0')
> + return -1;
> +
> + errno = 0;
> + bitwidth = strtoul(arg, &end, 0);
> +
> + /* check for errors */
> + if (bitwidth > UINT16_MAX || errno != 0 || end == NULL || *end != '\0')
> + return -1;
> +
> + if (bitwidth == 0)
> + bitwidth = UINT16_MAX;
> + ret = rte_set_max_simd_bitwidth(bitwidth);
> + if (ret < 0)
> + return -1;
> + internal_conf->max_simd_bitwidth.locked = locked;
> + return 0;
> +}
> +
>  static int
>  eal_parse_base_virtaddr(const char *arg)
>  {
> @@ -1707,6 +1736,13 @@ eal_parse_common_option(int opt, const char *optarg,
>   case OPT_NO_TELEMETRY_NUM:
>   conf->no_telemetry = 1;
>   break;
> + case OPT_FORCE_MAX_SIMD_BITWIDTH_NUM:
> + if (eal_parse_simd_bitwidth(optarg, 1) < 0) {
> + RTE_LOG(ERR, EAL, "invalid parameter for --"
> + OPT_FORCE_MAX_SIMD_BITWIDTH "\n");
> + return -1;
> + }
> + break;
>  
>   /* don't know what to do, leave this to caller */
>   default:
> @@ -1903,6 +1939,33 @@ eal_check_common_options(struct internal_config 
> *internal_cfg)
>   return 0;
>  }
>  
> +uint16_t
> +rte_get_max_simd_bitwidth(void)
> +{
> + const struct internal_config *internal_conf =
> + eal_get_internal_configuration();
> + return internal_conf->max_simd_bitwidth.bitwidth;
> +}

Should the return value be enum rte_max_simd_t?
If not, do we really need the enum definition?

> +
> +int
> +rte_set_max_simd_bitwidth(uint16_t bitwidth)
> +{
> + struct internal_config *internal_conf =
> + eal_get_internal_configuration();
> + if (internal_conf->max_simd_bitwidth.locked) {
> + RTE_LOG(NOTICE, EAL, "Cannot set max SIMD bitwidth - user 
> runtime override enabled");
> + return -EPERM;
> + }
> +
> + if (bitwidth != RTE_MAX_SIMD_DISABLE && (bitwidth < RTE_NO_SIMD ||
> + !rte_is_power_of_2(bitwidth))) {
> + RTE_LOG(ERR, EAL, "Invalid bitwidth value!\n");
> + return -EINVAL;
> + }
> + internal_conf->max_simd_bitwidth.bitwidth = bitwidth;
> + return 0;
> +}

Same question, should the parameter be enum rte_max_simd_t?

> +
>  void
>  eal_common_usage(void)
>  {
> @@ -1981,6 +2044,7 @@ eal_common_usage(void)
>  "  --"OPT_BASE_VIRTADDR" Base virtual address\n"
>  "  --"OPT_TELEMETRY"   Enable telemetry support (on by 
> default)\n"
>  "  --"OPT_NO_TELEMETRY"   Disable telemetry support\n"
> +"  --"OPT_FORCE_MAX_SIMD_BITWIDTH" Force the max SIMD bitwidth\n"
>  "\nEAL options for DEBUG use only:\n"
>  "  --"OPT_HUGE_UNLINK"   Unlink hugepage files after init\n"
>  "  --"OPT_NO_HUGE"   Use malloc instead of hugetlbfs\n"
> diff --git a/lib/librte_eal/common/eal

Re: [dpdk-dev] [EXT] Re: [PATCH 1/1] eal: increase TRACE CTF SIZE to recommended size

2020-10-06 Thread David Marchand
On Tue, Oct 6, 2020 at 11:22 AM Sunil Kumar Kori  wrote:
> >On Mon, Oct 5, 2020 at 10:16 PM Timothy McDaniel
> > wrote:
> >>
> >> Increase TRACE_CTF_FIELD_SIZE to 448, the recommended size.
> >
> >Repeating the same sentence in the title and the commitlog does not give
> >much info.
> >
> >Plus, what is this "recommendation"?
> When analyzed this issue, only one more byte was needed to fix this issue but 
> in future similar issue can occur again.
> So increasing this value by 64 bytes which actually equals to a cache line. 
> That’s why we have suggested this size.

384 is aligned to both 64 and 128 bytes cache lines.
448 is only aligned to 64 bytes.

Should we care about 128 bytes cache lines systems?

>
> >
> >
> >> Fixes "CTF field is too long" error when running with trace enabled.
> >
> >Ok, you hit this limit, but it would help to get some context here.
> >Looking at this patch in the future, we won't know why it was necessary.

How about following commitlog:

"""
trace: increase trace point buffer size

The current buffer size is not big enough to accomodate traces for new
additions in the eventdev subsystem.
Increase this buffer size by XXX for reason YYY.
"""


-- 
David Marchand



[dpdk-dev] [PATCH v1] crypto/aesni_mb: fix incorrect clearing of security session

2020-10-06 Thread David Coyle
When destroying a security session, the AESNI-MB PMD attempted to clear
the private aesni_mb session object to remove any key material. However,
the function aesni_mb_pmd_sec_sess_destroy() cleared the security session
object instead of the private session object.

This patch fixes this issue by now clearing the private session object.

Fixes: fda5216fba55 ("crypto/aesni_mb: support DOCSIS protocol")

Signed-off-by: David Coyle 
---
 drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c 
b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
index 2362f0c3c..b11d7f12b 100644
--- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
+++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
@@ -911,7 +911,7 @@ aesni_mb_pmd_sec_sess_destroy(void *dev __rte_unused,
 
if (sess_priv) {
struct rte_mempool *sess_mp = rte_mempool_from_obj(sess_priv);
-   memset(sess, 0, sizeof(struct aesni_mb_session));
+   memset(sess_priv, 0, sizeof(struct aesni_mb_session));
set_sec_session_private_data(sess, NULL);
rte_mempool_put(sess_mp, sess_priv);
}
-- 
2.17.1



Re: [dpdk-dev] [PATCH v4 3/3] ethdev: allow close function to return an error

2020-10-06 Thread Ferruh Yigit

On 10/5/2020 6:08 PM, Thomas Monjalon wrote:

The API function rte_eth_dev_close() was returning void.
The return type is changed to int for notifying of errors.

If an error happens during a close operation,
the status of the port is undefined,
a maximum of resources having been freed.

Signed-off-by: Thomas Monjalon 
Reviewed-by: Liron Himi 
Acked-by: Stephen Hemminger 


<...>


-void
+int
  rte_eth_dev_close(uint16_t port_id)
  {
struct rte_eth_dev *dev;
+   int firsterr, binerr;
+   int *lasterr = &firsterr;
  
-	RTE_ETH_VALID_PORTID_OR_RET(port_id);

+   RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -EINVAL);
dev = &rte_eth_devices[port_id];
  
-	RTE_FUNC_PTR_OR_RET(*dev->dev_ops->dev_close);

-   (*dev->dev_ops->dev_close)(dev);
+   RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_close, -ENOTSUP);
+   *lasterr = (*dev->dev_ops->dev_close)(dev);
+   if (*lasterr != 0)
+   lasterr = &binerr;
  
  	rte_ethdev_trace_close(port_id);

-   rte_eth_dev_release_port(dev);
+   *lasterr = rte_eth_dev_release_port(dev);
+
+   return firsterr;
  }


This may be personal taste but above error handling looks like unnecessary 
complex, what do you think something like below:


close_err = (*dev->dev_ops->dev_close)(dev);
release_err = rte_eth_dev_release_port(dev);
return close_err ? close_err : release_err;


Re: [dpdk-dev] [PATCH v2] net: calculate checksums for packets with IPv4 options

2020-10-06 Thread Michael Pfeiffer
Hi,

On Tue, 2020-10-06 at 10:10 +0200, Olivier Matz wrote:
> On Mon, Oct 05, 2020 at 07:39:45PM -0700, Stephen Hemminger wrote:
> > On Tue, 06 Oct 2020 00:55:19 +0200
> > Thomas Monjalon  wrote:
> > > 
> > > On 9/1/20 12:47 PM, Michael Pfeiffer wrote:
> > > > Currently, rte_ipv4_cksum() and rte_ipv4_udptcp_cksum() assume all IPv4
> > > > headers have sizeof(struct rte_ipv4_hdr) bytes. This is not true for
> > > > those (rare) packets with IPv4 options. Thus, both IPv4 and TCP/UDP
> > > > checksums are calculated wrong.
> > > > 
> > > > This patch fixes the issue by using the actual IPv4 header length from
> > > > the packet's IHL field.
> > > > 
> > > > Signed-off-by: Michael Pfeiffer 
> 
> Acked-by: Olivier Matz 
> 
> > > > -   cksum = rte_raw_cksum(ipv4_hdr, sizeof(struct rte_ipv4_hdr));
> > > > +   cksum = rte_raw_cksum(ipv4_hdr, (ipv4_hdr->version_ihl & 0xf) *
> > > > 4);
> > > 
> > > Truly naive questions:
> > > - doesn't it deserve a static inline function rte_ipv4_hdr_len()?
> > 
> > Makes sense to have that.
> 
> +1
> 
> However it could be in another patch: there are ~15 places where it
> could be replaced in dpdk.

Thank you very much for your feedback. I will prepare a second patch for that
function.




Re: [dpdk-dev] [PATCH 1/2] mempool: remove v20 ABI

2020-10-06 Thread David Marchand
On Tue, Oct 6, 2020 at 9:08 AM Olivier Matz  wrote:
>
> Remove the deprecated v20 ABI of rte_mempool_populate_iova() and
> rte_mempool_populate_virt().
>
> Signed-off-by: Olivier Matz 
> ---
>  lib/librte_mempool/meson.build |  2 -
>  lib/librte_mempool/rte_mempool.c   | 79 ++
>  lib/librte_mempool/rte_mempool_version.map |  7 --
>  3 files changed, 5 insertions(+), 83 deletions(-)
>
> diff --git a/lib/librte_mempool/meson.build b/lib/librte_mempool/meson.build
> index 7dbe6b9bea..a6e861cbfc 100644
> --- a/lib/librte_mempool/meson.build
> +++ b/lib/librte_mempool/meson.build
> @@ -9,8 +9,6 @@ foreach flag: extra_flags
> endif
>  endforeach
>
> -use_function_versioning = true
> -
>  sources = files('rte_mempool.c', 'rte_mempool_ops.c',
> 'rte_mempool_ops_default.c', 'mempool_trace_points.c')
>  headers = files('rte_mempool.h', 'rte_mempool_trace.h',
> diff --git a/lib/librte_mempool/rte_mempool.c 
> b/lib/librte_mempool/rte_mempool.c
> index 7774f0c8da..0e3a2a7635 100644
> --- a/lib/librte_mempool/rte_mempool.c
> +++ b/lib/librte_mempool/rte_mempool.c
> @@ -30,7 +30,6 @@
>  #include 
>  #include 
>  #include 
> -#include 
>  #include 
>
>
> @@ -305,17 +304,12 @@ mempool_ops_alloc_once(struct rte_mempool *mp)
> return 0;
>  }
>
> -__vsym int
> -rte_mempool_populate_iova_v21(struct rte_mempool *mp, char *vaddr,
> -   rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
> -   void *opaque);
> -
>  /* Add objects in the pool, using a physically contiguous memory
>   * zone. Return the number of objects added, or a negative value
>   * on error.
>   */
> -__vsym int
> -rte_mempool_populate_iova_v21(struct rte_mempool *mp, char *vaddr,
> +int
> +rte_mempool_populate_iova(struct rte_mempool *mp, char *vaddr,
> rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
> void *opaque)
>  {
> @@ -375,35 +369,6 @@ rte_mempool_populate_iova_v21(struct rte_mempool *mp, 
> char *vaddr,
> return ret;
>  }
>
> -BIND_DEFAULT_SYMBOL(rte_mempool_populate_iova, _v21, 21);
> -MAP_STATIC_SYMBOL(
> -   int rte_mempool_populate_iova(struct rte_mempool *mp, char *vaddr,
> -   rte_iova_t iova, size_t len,
> -   rte_mempool_memchunk_free_cb_t *free_cb,
> -   void *opaque),
> -   rte_mempool_populate_iova_v21);
> -
> -__vsym int
> -rte_mempool_populate_iova_v20(struct rte_mempool *mp, char *vaddr,
> -   rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
> -   void *opaque);
> -
> -__vsym int
> -rte_mempool_populate_iova_v20(struct rte_mempool *mp, char *vaddr,
> -   rte_iova_t iova, size_t len, rte_mempool_memchunk_free_cb_t *free_cb,
> -   void *opaque)
> -{
> -   int ret;
> -
> -   ret = rte_mempool_populate_iova_v21(mp, vaddr, iova, len, free_cb,
> -   opaque);
> -   if (ret == 0)
> -   ret = -EINVAL;
> -
> -   return ret;
> -}
> -VERSION_SYMBOL(rte_mempool_populate_iova, _v20, 20.0);
> -
>  static rte_iova_t
>  get_iova(void *addr)
>  {
> @@ -417,16 +382,11 @@ get_iova(void *addr)
> return ms->iova + RTE_PTR_DIFF(addr, ms->addr);
>  }
>
> -__vsym int
> -rte_mempool_populate_virt_v21(struct rte_mempool *mp, char *addr,
> -   size_t len, size_t pg_sz, rte_mempool_memchunk_free_cb_t *free_cb,
> -   void *opaque);
> -
>  /* Populate the mempool with a virtual area. Return the number of
>   * objects added, or a negative value on error.
>   */
> -__vsym int
> -rte_mempool_populate_virt_v21(struct rte_mempool *mp, char *addr,
> +int
> +rte_mempool_populate_virt(struct rte_mempool *mp, char *addr,
> size_t len, size_t pg_sz, rte_mempool_memchunk_free_cb_t *free_cb,
> void *opaque)
>  {
> @@ -459,7 +419,7 @@ rte_mempool_populate_virt_v21(struct rte_mempool *mp, 
> char *addr,
> break;
> }
>
> -   ret = rte_mempool_populate_iova_v21(mp, addr + off, iova,
> +   ret = rte_mempool_populate_iova(mp, addr + off, iova,
> phys_len, free_cb, opaque);
> if (ret == 0)
> continue;
> @@ -477,35 +437,6 @@ rte_mempool_populate_virt_v21(struct rte_mempool *mp, 
> char *addr,
> rte_mempool_free_memchunks(mp);
> return ret;
>  }
> -BIND_DEFAULT_SYMBOL(rte_mempool_populate_virt, _v21, 21);
> -MAP_STATIC_SYMBOL(
> -   int rte_mempool_populate_virt(struct rte_mempool *mp,
> -   char *addr, size_t len, size_t pg_sz,
> -   rte_mempool_memchunk_free_cb_t *free_cb,
> -   void *opaque),
> -   rte_mempool_populate_virt_v21);
> -
> -__vsym int
> -rte_mempool_populate_virt_v20(struct rte_mempool *mp, char *addr,
> -   size_t len, size_t pg_sz, rte_mempool_memchunk_free_cb_t 

Re: [dpdk-dev] [PATCH v3 17/18] net: add checks for max SIMD bitwidth

2020-10-06 Thread Olivier Matz
Hi,

On Wed, Sep 30, 2020 at 02:04:13PM +0100, Ciara Power wrote:
> When choosing a vector path to take, an extra condition must be
> satisfied to ensure the max SIMD bitwidth allows for the CPU enabled
> path.
> 
> The vector path was initially chosen in RTE_INIT, however this is no
> longer suitable as we cannot check the max SIMD bitwidth at that time.
> The default chosen in RTE_INIT is now scalar. For best performance
> and to use vector paths, apps must explicitly call the set algorithm
> function before using other functions from this library, as this is
> where vector handlers are now chosen.
> 
> Suggested-by: Jasvinder Singh 
> 
> Signed-off-by: Ciara Power 
> 
> ---
> v3:
>   - Moved choosing vector paths out of RTE_INIT.
>   - Moved checking max_simd_bitwidth into the set_alg function.
> ---
>  lib/librte_net/rte_net_crc.c | 26 +-
>  lib/librte_net/rte_net_crc.h |  3 ++-
>  2 files changed, 19 insertions(+), 10 deletions(-)
> 
> diff --git a/lib/librte_net/rte_net_crc.c b/lib/librte_net/rte_net_crc.c
> index 9fd4794a9d..241eb16399 100644
> --- a/lib/librte_net/rte_net_crc.c
> +++ b/lib/librte_net/rte_net_crc.c
> @@ -9,6 +9,7 @@
>  #include 
>  #include 
>  #include 
> +#include 
>  
>  #if defined(RTE_ARCH_X86_64) && defined(RTE_MACHINE_CPUFLAG_PCLMULQDQ)
>  #define X86_64_SSE42_PCLMULQDQ 1
> @@ -60,6 +61,9 @@ static rte_net_crc_handler handlers_neon[] = {
>  };
>  #endif
>  
> +static uint16_t max_simd_bitwidth;
> +#define RTE_LOGTYPE_NET RTE_LOGTYPE_USER1

RTE_LOG_REGISTER() should be used instead.

> +
>  /**
>   * Reflect the bits about the middle
>   *
> @@ -145,18 +149,26 @@ rte_crc32_eth_handler(const uint8_t *data, uint32_t 
> data_len)
>  void
>  rte_net_crc_set_alg(enum rte_net_crc_alg alg)
>  {
> + if (max_simd_bitwidth == 0)
> + max_simd_bitwidth = rte_get_max_simd_bitwidth();
> +
>   switch (alg) {
>  #ifdef X86_64_SSE42_PCLMULQDQ
>   case RTE_NET_CRC_SSE42:
> - handlers = handlers_sse42;
> - break;
> + if (max_simd_bitwidth >= RTE_MAX_128_SIMD) {
> + handlers = handlers_sse42;
> + return;
> + }
> + RTE_LOG(INFO, NET, "Max SIMD Bitwidth too low, using scalar\n");

If max_simd_bitwidth is too low, it will keep the previous value.
I think we should avoid to say "using scalar" in the log, even if it is
correct today. For instance, when the avx implementation will be added,
the log will become wrong.


>  #elif defined ARM64_NEON_PMULL
>   /* fall-through */
>   case RTE_NET_CRC_NEON:
> - if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_PMULL)) {
> + if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_PMULL) &&
> + max_simd_bitwidth >= RTE_MAX_128_SIMD) {
>   handlers = handlers_neon;
> - break;
> + return;
>   }
> + RTE_LOG(INFO, NET, "Max SIMD Bitwidth too low or CPU flag not 
> enabled, using scalar\n");
>  #endif
>   /* fall-through */
>   case RTE_NET_CRC_SCALAR:
> @@ -184,19 +196,15 @@ rte_net_crc_calc(const void *data,
>  /* Select highest available crc algorithm as default one */
>  RTE_INIT(rte_net_crc_init)
>  {
> - enum rte_net_crc_alg alg = RTE_NET_CRC_SCALAR;
> -
>   rte_net_crc_scalar_init();
>  
>  #ifdef X86_64_SSE42_PCLMULQDQ
> - alg = RTE_NET_CRC_SSE42;
>   rte_net_crc_sse42_init();
>  #elif defined ARM64_NEON_PMULL
>   if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_PMULL)) {
> - alg = RTE_NET_CRC_NEON;
>   rte_net_crc_neon_init();
>   }
>  #endif
>  
> - rte_net_crc_set_alg(alg);
> + rte_net_crc_set_alg(RTE_NET_CRC_SCALAR);
>  }
> diff --git a/lib/librte_net/rte_net_crc.h b/lib/librte_net/rte_net_crc.h
> index 16e85ca970..7a45ebe193 100644
> --- a/lib/librte_net/rte_net_crc.h
> +++ b/lib/librte_net/rte_net_crc.h
> @@ -28,7 +28,8 @@ enum rte_net_crc_alg {
>  /**
>   * This API set the CRC computation algorithm (i.e. scalar version,
>   * x86 64-bit sse4.2 intrinsic version, etc.) and internal data
> - * structure.
> + * structure. This should be called before any other functions, to
> + * choose the algorithm for best performance.
>   *
>   * @param alg
>   *   This parameter is used to select the CRC implementation version.
> -- 
> 2.17.1
> 


Re: [dpdk-dev] [EXT] Re: [PATCH 1/1] eal: increase TRACE CTF SIZE to recommended size

2020-10-06 Thread Jerin Jacob
On Tue, Oct 6, 2020 at 3:10 PM David Marchand  wrote:
>
> On Tue, Oct 6, 2020 at 11:22 AM Sunil Kumar Kori  wrote:
> > >On Mon, Oct 5, 2020 at 10:16 PM Timothy McDaniel
> > > wrote:
> > >>
> > >> Increase TRACE_CTF_FIELD_SIZE to 448, the recommended size.
> > >
> > >Repeating the same sentence in the title and the commitlog does not give
> > >much info.
> > >
> > >Plus, what is this "recommendation"?
> > When analyzed this issue, only one more byte was needed to fix this issue 
> > but in future similar issue can occur again.
> > So increasing this value by 64 bytes which actually equals to a cache line. 
> > That’s why we have suggested this size.
>
> 384 is aligned to both 64 and 128 bytes cache lines.
> 448 is only aligned to 64 bytes.
>
> Should we care about 128 bytes cache lines systems?

it is on a slow path. 448 is OK.

>
> >
> > >
> > >
> > >> Fixes "CTF field is too long" error when running with trace enabled.
> > >
> > >Ok, you hit this limit, but it would help to get some context here.
> > >Looking at this patch in the future, we won't know why it was necessary.
>
> How about following commitlog:
>
> """
> trace: increase trace point buffer size
>
> The current buffer size is not big enough to accomodate traces for new
> additions in the eventdev subsystem.
> Increase this buffer size by XXX for reason YYY.
> """

Looks good to me.

>
>
> --
> David Marchand
>


Re: [dpdk-dev] [PATCH v3 17/18] net: add checks for max SIMD bitwidth

2020-10-06 Thread Olivier Matz
Hi,

On Thu, Oct 01, 2020 at 02:19:37PM +, Power, Ciara wrote:
> Hi David,
> 
> Thanks for reviewing, 
> 
> >-Original Message-
> >From: Coyle, David 
> >Sent: Thursday 1 October 2020 15:17
> >To: Singh, Jasvinder ; Power, Ciara
> >; dev@dpdk.org
> >Cc: Power, Ciara ; Olivier Matz
> >; O'loingsigh, Mairtin
> >; Ryan, Brendan ;
> >Richardson, Bruce 
> >Subject: RE: [dpdk-dev] [PATCH v3 17/18] net: add checks for max SIMD
> >bitwidth
> >
> >Hi Jasvinder/Ciara
> >
> >> -Original Message-
> >> From: Singh, Jasvinder 
> >> >
> >> > > From: dev  On Behalf Of Ciara Power When
> >> > > choosing a vector path to take, an extra condition must be
> >> > > satisfied to ensure the max SIMD bitwidth allows for the CPU enabled
> >path.
> >> > >
> >> > > The vector path was initially chosen in RTE_INIT, however this is
> >> > > no longer suitable as we cannot check the max SIMD bitwidth at that
> >time.
> >> > > The default chosen in RTE_INIT is now scalar. For best performance
> >> > > and to use vector paths, apps must explicitly call the set
> >> > > algorithm function before using other functions from this library,
> >> > > as this is where vector handlers are now chosen.
> >> >
> >> > [DC] Has it been decided that it is ok to now require applications
> >> > to pick the CRC algorithm they want to use?
> >> >
> >> > An application which previously automatically got SSE4.2 CRC, for
> >> > example, will now automatically only get scalar.
> >> >
> >> > If this is ok, this should probably be called out explicitly in
> >> > release notes as it may not be Immediately noticeable to users that
> >> > they now need to select the CRC algo.
> >> >
> >> > Actually, in general, the release notes need to be updated for this
> >> patchset.
> >>
> >> The decision to move rte_set_alg() out of RTE_INIT was taken to avoid
> >> check on max_simd_bitwidth in data path for every single time when
> >> crc_calc() api is invoked. Based on my understanding,
> >> max_simd_bitwidth is set after eal init, and when used in crc_calc(),
> >> it might override the default crc algo set during RTE_INIT. Therefore,
> >> to avoid extra check on max_simd_bitwidth in data path,  better option
> >> will be to use this static configuration one time after eal init in the 
> >> set_algo
> >API.
> >
> >[DC] Yes that is a good change to have made to avoid extra datapath checks.
> >
> >Based on off-list discussion, I now also know the reason behind now
> >defaulting to scalar CRC in RTE_INIT. If a higher bitwidth CRC was chosen by
> >RTE_INIT (e.g.
> >SSE4.2 CRC) but the max_simd_bitwidth was then set to RTE_NO_SIMD (64)
> >through the EAL parameter or call to rte_set_max_simd_bitwidth(), then
> >there is a mismatch if rte_net_crc_set_alg() is not then called to 
> >reconfigure
> >the CRC. Defaulting to scalar avoids this mismatch and works on all archs
> >
> >As I mentioned before, I think this needs to be called out in release notes, 
> >as
> >it's an under-the-hood change which could cause app performance to drop if
> >app developers aren't aware of it - the API itself hasn't changed, so they 
> >may
> >not read the doxygen :)
> >
> 
> Yes that is a good point, I can add to the release notes for this to call it 
> out. 

I don't think it is a good idea to have the scalar crc by default.
To me, the fastest available CRC has to be enabled by default.

I understand the technical reason why you did it like this however: the
SIMD bitwidth may not be known at the time the
RTE_INIT(rte_net_crc_init) function is called.

A simple approach to solve this issue would be to initialize the
rte_net_crc_handler pointer to a handlers_default. The first time a crc
is called, the rte_crc32_*_default_handler() function would check the
configured SIMD bitwidth, and set the handler to the correct one, to
avoid to do the test for next time.

This approach still does not solve the case where the SIMD bitwidth is
modified during the life of the application. In this case, a callback
would have to be registered to notify SIMD bitwidth changes... but I
don't think it is worth to do it. Instead, it can be documented that
rte_set_max_simd_bitwidth() has to be called early, before
rte_eal_init().



> >>
> >>
> >> > >
> >> > > Suggested-by: Jasvinder Singh 
> >> > >
> >> > > Signed-off-by: Ciara Power 
> >> > >
> >> > > ---
> >> > > v3:
> >> > >   - Moved choosing vector paths out of RTE_INIT.
> >> > >   - Moved checking max_simd_bitwidth into the set_alg function.
> >> > > ---
> >> > >  lib/librte_net/rte_net_crc.c | 26 +-
> >> > > lib/librte_net/rte_net_crc.h |  3 ++-
> >> > >  2 files changed, 19 insertions(+), 10 deletions(-)
> >> > >
> >> > > diff --git a/lib/librte_net/rte_net_crc.c
> >> > > b/lib/librte_net/rte_net_crc.c index
> >> > > 9fd4794a9d..241eb16399 100644
> >> > > --- a/lib/librte_net/rte_net_crc.c
> >> > > +++ b/lib/librte_net/rte_net_crc.c
> >> >
> >> > 
> >> >
> >> > > @@ -145,18 +149,26 @@ rte_crc32_eth_handler(const uint8_t *

[dpdk-dev] [PATCH 0/2] baseband: fix segfault in Intel drivers

2020-10-06 Thread Maxime Coquelin
This series fixes segfaults in fpga_5gnr_fec and fpga_lte_fec
drivers when bbdev debug is enabled.

Maxime Coquelin (2):
  baseband/fpga_5gnr_fec: fix segfaults with debug
  baseband/fpga_lte_fec: fix segfaults with debug

 drivers/baseband/fpga_5gnr_fec/rte_fpga_5gnr_fec.c | 4 ++--
 drivers/baseband/fpga_lte_fec/fpga_lte_fec.c   | 4 ++--
 2 files changed, 4 insertions(+), 4 deletions(-)

-- 
2.26.2



[dpdk-dev] [PATCH 1/2] baseband/fpga_5gnr_fec: fix segfaults with debug

2020-10-06 Thread Maxime Coquelin
When RTE_LIBRTE_BBDEV_DEBUG is enabled, rte_device's driver
pointer is dereferenced twice in fpga_5gnr_fec's probe callback.
It causes a segmentation fault because this pointer is only
assigned after probe callback call.

This patch makes use of rte_pci_driver pointer instead.

Fixes: 0b5927cbcba7 ("baseband/fpga_5gnr_fec: add PMD for FPGA 5GNR FEC")

Signed-off-by: Maxime Coquelin 
---
 drivers/baseband/fpga_5gnr_fec/rte_fpga_5gnr_fec.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/drivers/baseband/fpga_5gnr_fec/rte_fpga_5gnr_fec.c 
b/drivers/baseband/fpga_5gnr_fec/rte_fpga_5gnr_fec.c
index 61f9c04ba2..11ee4d3e10 100644
--- a/drivers/baseband/fpga_5gnr_fec/rte_fpga_5gnr_fec.c
+++ b/drivers/baseband/fpga_5gnr_fec/rte_fpga_5gnr_fec.c
@@ -1839,7 +1839,7 @@ fpga_5gnr_fec_init(struct rte_bbdev *dev, struct 
rte_pci_driver *drv)
 
rte_bbdev_log_debug(
"Init device %s [%s] @ virtaddr %p phyaddr %#"PRIx64,
-   dev->device->driver->name, dev->data->name,
+   drv->driver.name, dev->data->name,
(void *)pci_dev->mem_resource[0].addr,
pci_dev->mem_resource[0].phys_addr);
 }
@@ -1895,7 +1895,7 @@ fpga_5gnr_fec_probe(struct rte_pci_driver *pci_drv,
((uint16_t)(version_id >> 16)), ((uint16_t)version_id));
 
 #ifdef RTE_LIBRTE_BBDEV_DEBUG
-   if (!strcmp(bbdev->device->driver->name,
+   if (!strcmp(pci_drv->driver.name,
RTE_STR(FPGA_5GNR_FEC_PF_DRIVER_NAME)))
print_static_reg_debug_info(d->mmio_base);
 #endif
-- 
2.26.2



[dpdk-dev] [PATCH 2/2] baseband/fpga_lte_fec: fix segfaults with debug

2020-10-06 Thread Maxime Coquelin
When RTE_LIBRTE_BBDEV_DEBUG is enabled, rte_device's driver
pointer is dereferenced twice in fpga_lte_fec's probe callback.
It causes a segmentation fault because this pointer is only
assigned after probe callback call.

This patch makes use of rte_pci_driver pointer instead

Fixes: efd453698c49 ("baseband/fpga_lte_fec: add driver for FEC on FPGA")
Cc: sta...@dpdk.org

Signed-off-by: Maxime Coquelin 
---
 drivers/baseband/fpga_lte_fec/fpga_lte_fec.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/drivers/baseband/fpga_lte_fec/fpga_lte_fec.c 
b/drivers/baseband/fpga_lte_fec/fpga_lte_fec.c
index 37018b9c7f..c80721be96 100644
--- a/drivers/baseband/fpga_lte_fec/fpga_lte_fec.c
+++ b/drivers/baseband/fpga_lte_fec/fpga_lte_fec.c
@@ -2328,7 +2328,7 @@ fpga_lte_fec_init(struct rte_bbdev *dev, struct 
rte_pci_driver *drv)
 
rte_bbdev_log_debug(
"Init device %s [%s] @ virtaddr %p phyaddr %#"PRIx64,
-   dev->device->driver->name, dev->data->name,
+   drv->driver.name, dev->data->name,
(void *)pci_dev->mem_resource[0].addr,
pci_dev->mem_resource[0].phys_addr);
 }
@@ -2383,7 +2383,7 @@ fpga_lte_fec_probe(struct rte_pci_driver *pci_drv,
((uint16_t)(version_id >> 16)), ((uint16_t)version_id));
 
 #ifdef RTE_LIBRTE_BBDEV_DEBUG
-   if (!strcmp(bbdev->device->driver->name,
+   if (!strcmp(pci_drv->driver.name,
RTE_STR(FPGA_LTE_FEC_PF_DRIVER_NAME)))
print_static_reg_debug_info(d->mmio_base);
 #endif
-- 
2.26.2



Re: [dpdk-dev] [PATCH 2/3] examples/fips_validation: ignore \r in input files

2020-10-06 Thread Olivier Matz
Hi Fan,

On Tue, Oct 06, 2020 at 08:47:10AM +, Zhang, Roy Fan wrote:
> Hi Olivier,
>
> > -Original Message-
> > From: Olivier Matz 
> > Sent: Tuesday, October 6, 2020 8:42 AM
> > To: dev@dpdk.org
> > Cc: Kovacevic, Marko ; Akhil Goyal
> > ; Zhang, Roy Fan ; Kusztal,
> > ArkadiuszX ; sta...@dpdk.org
> > Subject: [PATCH 2/3] examples/fips_validation: ignore \r in input files
> > 
> > Some test vectors contain '\r' before '\n' (see link). Ignore them.
> > 
> > Link: https://www.openssl.org/docs/fips/testvectors-linux-2007-10-10.tar.gz
> > Fixes: 3d0fad56b74a ("examples/fips_validation: add crypto FIPS 
> > application")
> > Cc: sta...@dpdk.org
> > 
> > Signed-off-by: Olivier Matz 
> > ---
> >  examples/fips_validation/fips_validation.c | 2 ++
> >  1 file changed, 2 insertions(+)
> > 
> > diff --git a/examples/fips_validation/fips_validation.c
> > b/examples/fips_validation/fips_validation.c
> > index 13f763c9aa..858f581ba3 100644
> > --- a/examples/fips_validation/fips_validation.c
> > +++ b/examples/fips_validation/fips_validation.c
> > @@ -33,6 +33,8 @@ get_file_line(void)
> > 
> > if (loc >= MAX_LINE_CHAR - 1)
> > return -ENOMEM;
> > +   if (c == '\r')
> > +   continue;
> > if (c == '\n')
> > break;
> > line[loc++] = c;
> > --
> 
>
> The patch looks ok but the test file link you provided in the patch is CAVS
> 5.3.
>
> As mentioned in
> https://doc.dpdk.org/guides/sample_app_ug/fips_validation.html, the supported
> CAVS supported version is 21.0 (not latest one by newer than 5.3). In CAVS
> 21.0 test files there is no '\r' before '\n' (I suppose this is for Windows
> right).

Thank you for your feedback.

I'm ok to drop this patch from the patchset if you feel it's useless, or
I can update the commit log with the information you provide, to clarify
that it should not happen with the supported version of CAVS.

Please let me know what you prefer.


Thanks,
Olivier


Re: [dpdk-dev] [PATCH v1] event/sw: performance improvements

2020-10-06 Thread Ananyev, Konstantin

> 
> > -Original Message-
> > From: Jerin Jacob 
> > Sent: Monday, October 5, 2020 5:35 PM
> > To: Nicolau, Radu 
> > Cc: Honnappa Nagarahalli ; Richardson, Bruce
> > ; Ananyev, Konstantin
> > ; Van Haaren, Harry
> > ; dev@dpdk.org; jer...@marvell.com; nd
> > 
> > Subject: Re: [dpdk-dev] [PATCH v1] event/sw: performance improvements
> >
> > On Tue, Sep 29, 2020 at 2:32 PM Nicolau, Radu  
> > wrote:
> > >
> > >
> > > On 9/28/2020 5:02 PM, Honnappa Nagarahalli wrote:
> > > > 
> > > >>> Add minimum burst throughout the scheduler pipeline and a flush 
> > > >>> counter.
> > > >>> Replace ring API calls with local single threaded implementation where
> > > >>> possible.
> > > >>>
> > > >>> Signed-off-by: Radu Nicolau mailto:radu.nico...@intel.com
> > > >>>
> > > >>> Thanks for the patch, a few comments inline.
> > > >>>
> >
> > > >>> Why not make these APIs part of the rte_ring library? You could 
> > > >>> further
> > > >> optimize them by keeping the indices on the same cacheline.
> > > >>> I'm not sure there is any need for non thread-safe rings outside this
> > > >> particular case.
> > > >>> [Honnappa] I think if we add the APIs, we will find the use cases.
> > > >>> But, more than that, I understand that rte_ring structure is exposed 
> > > >>> to the
> > > >> application. The reason for doing that is the inline functions that 
> > > >> rte_ring
> > > >> provides. IMO, we should still maintain modularity and should not use 
> > > >> the
> > > >> internals of the rte_ring structure outside of the library.
> > > >>> +1 to that.
> > > >>>
> > > >>> BTW, is there any real perf benefit from such micor-optimisation?
> > > >> I'd tend to view these as use-case specific, and I'm not sure we 
> > > >> should clutter
> > > >> up the ring library with yet more functions, especially since they 
> > > >> can't be
> > > >> mixed with the existing enqueue/dequeue functions, since they don't use
> > > >> the head pointers.
> > > > IMO, the ring library is pretty organized with the recent addition of 
> > > > HTS/RTS
> > modes. This can be one of the modes and should allow us to use the existing
> > functions (though additional functions are required as well).
> > > > The other concern I have is, this implementation can be further 
> > > > optimized by
> > using a single cache line for the pointers. It uses 2 cache lines just 
> > because of the
> > layout of the rte_ring structure.
> > > > There was a question earlier about the performance improvements of this
> > patch? Are there any % performance improvements that can be shared?
> > > > It is also possible to change the above functions to use the head/tail 
> > > > pointers
> > from producer or the consumer cache line alone to check for perf 
> > differences.
> > >
> > > I don't have a % for the final improvement for this change alone, but
> > > there was some improvement in the memory overhead measurable during
> > > development, which very likely resulted in the whole optimization having
> > > more headroom.
> > >
> > > I agree that this may be further optimized, maybe by having a local
> > > implementation of a ring-like container instead.
> >
> > Have we decided on the next steps for this patch? Is the plan to
> > supersede this patch and have different
> > one in rte_ring subsystem,
> 
> My preference is to merge this version of the patch;
> 1) The ring helper functions are stripped to the SW PMD usage, and not valid 
> to use in the general.
> 2) Adding static inline APIs in an LTS without extensive doesn't seem a good 
> idea.
> 
> If Honnappa is OK with the above solution for 20.11, we can see about moving 
> the rings part of the
> code to rte_ring library location in 21.02, and give ourselves some time to 
> settle the usage/API before
> the next LTS.
> 

As ring library maintainer I share Honnappa concern that another library not 
uses public ring API,
but instead accesses ring internals directly. Obviously such coding practice is 
not welcomed
as it makes harder to maintain/extend ring library in future.
About 2) - these new API can(/shoud) be marked an experimental anyway. 
As another thing - it is still unclear what a performance gain we are talking 
about here.
Is it really worth it comparing to just using SP/SC?





Re: [dpdk-dev] [PATCH v4] kernel/linux: remove igb_uio

2020-10-06 Thread Ferruh Yigit

On 10/5/2020 10:42 AM, Thomas Monjalon wrote:

As decided in the Technical Board in November 2019,
the kernel module igb_uio is moved to the dpdk-kmods repository
in the /linux/igb_uio/ directory.

Minutes of Technical Board meeting:
https://mails.dpdk.org/archives/dev/2019-November/151763.html

Signed-off-by: Thomas Monjalon 
Acked-by: Stephen Hemminger 
---
v4: Cc more maintainers
v3: update more docs and provide a link to the new repo
v2: update few docs (including release notes)
---
  MAINTAINERS|   1 -
  doc/guides/bbdevs/fpga_5gnr_fec.rst|   3 +-
  doc/guides/bbdevs/fpga_lte_fec.rst |   3 +-
  doc/guides/cryptodevs/ccp.rst  |   3 +-
  doc/guides/cryptodevs/qat.rst  |   3 +-
  doc/guides/howto/lm_bond_virtio_sriov.rst  |   2 +-
  doc/guides/howto/lm_virtio_vhost_user.rst  |   2 +-
  doc/guides/howto/openwrt.rst   |   5 -
  doc/guides/linux_gsg/enable_func.rst   |   3 +-
  doc/guides/linux_gsg/linux_drivers.rst |  23 +-
  doc/guides/nics/build_and_test.rst |   2 +-
  doc/guides/nics/ena.rst|   4 +-
  doc/guides/rel_notes/deprecation.rst   |   7 -
  doc/guides/rel_notes/release_20_11.rst |   4 +-
  doc/guides/sample_app_ug/multi_process.rst |   2 -
  drivers/bus/pci/bsd/pci.c  |   2 +-
  kernel/linux/igb_uio/Kbuild|   2 -
  kernel/linux/igb_uio/compat.h  | 154 -
  kernel/linux/igb_uio/igb_uio.c | 660 -
  kernel/linux/igb_uio/meson.build   |  20 -
  kernel/linux/meson.build   |   2 +-
  21 files changed, 21 insertions(+), 886 deletions(-)
  delete mode 100644 kernel/linux/igb_uio/Kbuild
  delete mode 100644 kernel/linux/igb_uio/compat.h
  delete mode 100644 kernel/linux/igb_uio/igb_uio.c
  delete mode 100644 kernel/linux/igb_uio/meson.build



Patch doesn't apply cleanly, not sure why, CI seems having same problem:
http://mails.dpdk.org/archives/test-report/2020-October/156378.html

<...>


index 7789c572bb..080b44955a 100644
--- a/doc/guides/linux_gsg/linux_drivers.rst
+++ b/doc/guides/linux_gsg/linux_drivers.rst
@@ -27,33 +27,20 @@ can provide the uio capability. This module can be loaded 
using the command:
  
  ``uio_pci_generic`` module doesn't support the creation of virtual functions.
  
-As an alternative to the ``uio_pci_generic``, the DPDK also includes the igb_uio

-module which can be found in the kernel/linux subdirectory referred to above. 
It can
-be loaded as shown below:
+As an alternative to the ``uio_pci_generic``, there is the ``igb_uio`` module
+which can be found in the repository `dpdk-kmods 
`_.
+It can be loaded as shown below:
  
  .. code-block:: console
  
  sudo modprobe uio

-sudo insmod /kernel/linux/igb_uio/igb_uio.ko
-
-.. note::
-
-   Building DPDK Linux kernel modules is disabled by default starting from 
DPDK 20.02.
-   To enable them again, the config option "enable_kmods" needs to be set
-   in the meson build configuration.
-   See :ref:`adjusting_build_options` for details on how to set/clear build 
options.
-   It is planned to move ``igb_uio`` module to a different git repository.
-
-.. note::
-
-For some devices which lack support for legacy interrupts, e.g. virtual 
function
-(VF) devices, the ``igb_uio`` module may be needed in place of 
``uio_pci_generic``.
+sudo insmod igb_uio.ko
  
  .. note::
  
 If UEFI secure boot is enabled, the Linux kernel may disallow the use of

 UIO on the system. Therefore, devices for use by DPDK should be bound to 
the
-   ``vfio-pci`` kernel module rather than ``igb_uio`` or ``uio_pci_generic``.
+   ``vfio-pci`` kernel module rather than any UIO-based module.
 For more details see :ref:`linux_gsg_binding_kernel` below.


Should we change the order of the kernel drivers, currently first section is 
"UIO" and second section is "VFIO". Since we encoruage vfio and move igb_uio out 
of the repo, should we list the "VfIO" first?


<...>


index 4e61431c6c..243bd940a4 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -122,9 +122,11 @@ Removed Items
  
  * vhost: Dequeue zero-copy support has been removed.
  
+* kernel: The module ``igb_uio`` has been moved to the git repository

+``dpdk-kmods`` in a new directory ``linux/igb_uio``.
+



Doc build giving warning on this:
WARNING: Bullet list ends without a blank line; unexpected unindent.


Except from above minor comments looks good to me,
Reviewed-by: Ferruh Yigit 


Re: [dpdk-dev] [PATCH v2 2/3] test/event_crypto_adapter: fix function arguments

2020-10-06 Thread Gujjar, Abhinandan S
Hi Ankur,

Overall, it looks good. 
You have missed out fixing caps args in one place line #299 of 
test_sessionless_with_op_forward_mode().

> -Original Message-
> From: Ankur Dwivedi 
> Sent: Friday, September 25, 2020 12:59 PM
> To: dev@dpdk.org
> Cc: Gujjar, Abhinandan S ;
> jer...@marvell.com; ano...@marvell.com; Ankur Dwivedi
> 
> Subject: [PATCH v2 2/3] test/event_crypto_adapter: fix function arguments
> 
> The arguments passed to rte_event_crypto_adapter_caps_get() and
> rte_event_crypto_adapter_create() are incorrect.
> 
> In the rte_event_crypto_adapter_caps_get(), event device id should be the
> first argument and cryptodev id should be the second argument.
> In the rte_event_crypto_adapter_create(), the event device id should be
> the second argument.
> 
> Signed-off-by: Ankur Dwivedi 
> ---
>  app/test/test_event_crypto_adapter.c | 18 +-
>  1 file changed, 9 insertions(+), 9 deletions(-)
> 
> diff --git a/app/test/test_event_crypto_adapter.c
> b/app/test/test_event_crypto_adapter.c
> index 3c64742b3..572ef9c27 100644
> --- a/app/test/test_event_crypto_adapter.c
> +++ b/app/test/test_event_crypto_adapter.c
> @@ -209,8 +209,8 @@ test_op_forward_mode(uint8_t session_less)
>   &cipher_xform,
> params.session_priv_mpool);
>   TEST_ASSERT_SUCCESS(ret, "Failed to init session\n");
> 
> - ret =
> rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID,
> - evdev, &cap);
> + ret = rte_event_crypto_adapter_caps_get(evdev,
> TEST_CDEV_ID,
> + &cap);
>   TEST_ASSERT_SUCCESS(ret, "Failed to get adapter
> capabilities\n");
> 
>   if (cap &
> RTE_EVENT_CRYPTO_ADAPTER_CAP_SESSION_PRIVATE_DATA) { @@ -321,7
> +321,7 @@ test_session_with_op_forward_mode(void)
>   uint32_t cap;
>   int ret;
> 
> - ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID,
> evdev, &cap);
> + ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID,
> &cap);
>   TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
> 
>   if (!(cap &
> RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD) && @@ -
> 410,8 +410,8 @@ test_op_new_mode(uint8_t session_less)
>   params.session_mpool);
>   TEST_ASSERT_NOT_NULL(sess, "Session creation failed\n");
> 
> - ret =
> rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID,
> - evdev, &cap);
> + ret = rte_event_crypto_adapter_caps_get(evdev,
> TEST_CDEV_ID,
> + &cap);
>   TEST_ASSERT_SUCCESS(ret, "Failed to get adapter
> capabilities\n");
> 
>   if (cap &
> RTE_EVENT_CRYPTO_ADAPTER_CAP_SESSION_PRIVATE_DATA) { @@ -460,7
> +460,7 @@ test_sessionless_with_op_new_mode(void)
>   uint32_t cap;
>   int ret;
> 
> - ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID,
> evdev, &cap);
> + ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID,
> &cap);
>   TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
> 
>   if (!(cap &
> RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD) && @@ -
> 486,7 +486,7 @@ test_session_with_op_new_mode(void)
>   uint32_t cap;
>   int ret;
> 
> - ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID,
> evdev, &cap);
> + ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID,
> &cap);
>   TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
> 
>   if (!(cap &
> RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD) && @@ -
> 706,7 +706,7 @@ test_crypto_adapter_create(void)
> 
>   /* Create adapter with default port creation callback */
>   ret = rte_event_crypto_adapter_create(TEST_ADAPTER_ID,
> -   TEST_CDEV_ID,
> +   evdev,
> &conf, 0);
>   TEST_ASSERT_SUCCESS(ret, "Failed to create event crypto
> adapter\n");
> 
> @@ -719,7 +719,7 @@ test_crypto_adapter_qp_add_del(void)
>   uint32_t cap;
>   int ret;
> 
> - ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID,
> evdev, &cap);
> + ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID,
> &cap);
>   TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
> 
>   if (cap &
> RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_QP_EV_BIND) {
> --
> 2.28.0



Re: [dpdk-dev] [PATCH v2 3/3] test/event_crypto_adapter: free resources during exit

2020-10-06 Thread Gujjar, Abhinandan S
Hi Ankur,

Please see comment below.

> -Original Message-
> From: Ankur Dwivedi 
> Sent: Friday, September 25, 2020 12:59 PM
> To: dev@dpdk.org
> Cc: Gujjar, Abhinandan S ;
> jer...@marvell.com; ano...@marvell.com; Ankur Dwivedi
> 
> Subject: [PATCH v2 3/3] test/event_crypto_adapter: free resources during
> exit
> 
> The resources held by crypto adapter should be freed when the test suite
> exits.
> 
> Signed-off-by: Ankur Dwivedi 
> ---
>  app/test/test_event_crypto_adapter.c | 15 +++
>  1 file changed, 15 insertions(+)
> 
> diff --git a/app/test/test_event_crypto_adapter.c
> b/app/test/test_event_crypto_adapter.c
> index 572ef9c27..4b09ab3b7 100644
> --- a/app/test/test_event_crypto_adapter.c
> +++ b/app/test/test_event_crypto_adapter.c
> @@ -899,6 +899,20 @@ testsuite_setup(void)
>   return TEST_SUCCESS;
>  }
> 
> +static void
> +crypto_adapter_teardown(void)
> +{
> + int ret;
> +
> + crypto_adapter_setup_done = 0;
Call rte_event_crypto_adapter_stop() here?
> + ret =
> rte_event_crypto_adapter_queue_pair_del(TEST_ADAPTER_ID,
> + TEST_CDEV_ID, TEST_CDEV_QP_ID);
> + if (ret < 0)
> + RTE_LOG(ERR, USER1, "Failed to delete queue pair!");
> +
> + rte_event_crypto_adapter_free(TEST_ADAPTER_ID);
> +}
> +
>  static void
>  crypto_teardown(void)
>  {
> @@ -941,6 +955,7 @@ eventdev_teardown(void)  static void
>  testsuite_teardown(void)
>  {
> + crypto_adapter_teardown();
>   crypto_teardown();
>   eventdev_teardown();
>  }
> --
> 2.28.0



Re: [dpdk-dev] [PATCH v3 01/10] ethdev: add flow shared action API

2020-10-06 Thread Andrey Vesnovaty
Hi Ori,

PSB

Thanks,
Andrey

> -Original Message-
> From: Ori Kam 
> Sent: Sunday, October 4, 2020 2:10 PM
> To: Andrey Vesnovaty ; dev@dpdk.org
> Cc: j...@marvell.com; jerinjac...@gmail.com; NBU-Contact-Thomas Monjalon
> ; ferruh.yi...@intel.com;
> step...@networkplumber.org; bruce.richard...@intel.com; Slava Ovsiienko
> ; andrey.vesnov...@gmail.com; m...@ashroe.eu;
> nhor...@tuxdriver.com; ajit.khapa...@broadcom.com;
> samik.gu...@broadcom.com; Andrey Vesnovaty ;
> Andrew Rybchenko ; Ori Kam
> 
> Subject: RE: [PATCH v3 01/10] ethdev: add flow shared action API
> 
> Hi Andrey,
> 
> You are missing the doc update (rte_flow)

Got it. Will update in release_notes_20.11.rst & testpmd.

> PSB for more comments.
> 
> Best,
> Ori
> 
> > -Original Message-
> > From: Andrey Vesnovaty 
> > Sent: Sunday, October 4, 2020 1:06 AM
> > Subject: [PATCH v3 01/10] ethdev: add flow shared action API
> >
> > This commit introduces extension of DPDK flow action API enabling
> > sharing of single rte_flow_action in multiple flows. The API intended for
> > PMDs where multiple HW offloaded flows can reuse the same HW
> > essence/object representing flow action and modification of such an
> > essence/object effects all the rules using it.
> >
> > Motivation and example
> > ===
> > Adding or removing one or more queues to RSS used by multiple flow rules
> > imposes per rule toll for current DPDK flow API; the scenario requires
> > for each flow sharing cloned RSS action:
> > - call `rte_flow_destroy()`
> > - call `rte_flow_create()` with modified RSS action
> >
> > API for sharing action and its in-place update benefits:
> > - reduce the overhead of multiple RSS flow rules reconfiguration
> > - optimize resource utilization by sharing action across of multiple
> >   flows
> >
> > Change description
> > ===
> >
> > Shared action
> > ===
> > In order to represent flow action shared by multiple flows new action
> > type RTE_FLOW_ACTION_TYPE_SHARED is introduced (see `enum
> > rte_flow_action_type`).
> > Actually the introduced API decouples action from any specific flow and
> > enables sharing of single action by its handle across multiple flows.
> >
> > Shared action create/use/destroy
> > ===
> > Shared action may be reused by some or none flow rules at any given
> > moment, i.e. shared action reside outside of the context of any flow.
> > Shared action represent HW resources/objects used for action offloading
> > implementation.
> > API for shared action create (see `rte_flow_shared_action_create()`):
> > - should allocate HW resources and make related initializations required
> >   for shared action implementation.
> > - make necessary preparations to maintain shared access to
> >   the action resources, configuration and state.
> > API for shared action destroy (see `rte_flow_shared_action_destroy()`)
> > should release HW resources and make related cleanups required for shared
> > action implementation.
> >
> > In order to share some flow action reuse the handle of type
> > `struct rte_flow_shared_action` returned by
> > rte_flow_shared_action_create() as a `conf` field of
> > `struct rte_flow_action` (see "example" section).
> >
> > If some shared action not used by any flow rule all resources allocated
> > by the shared action can be released by rte_flow_shared_action_destroy()
> > (see "example" section). The shared action handle passed as argument to
> > destroy API should not be used any further i.e. result of the usage is
> > undefined.
> >
> > Shared action re-configuration
> > ===
> > Shared action behavior defined by its configuration can be updated via
> > rte_flow_shared_action_update() (see "example" section). The shared
> > action update operation modifies HW related resources/objects allocated
> > on the action creation. The number of operations performed by the update
> > operation should not be dependent on number of flows sharing the related
> > action. On return of shared action update API action behavior should be
> > according to updated configuration for all flows sharing the action.
> >
> > Shared action query
> > ===
> > Provide separate API to query shared action sate (see
> 
> I think sate --> state, also it is not clear if we are sharing both 
> configuration and
> state
> or only configuration. For example what will happen if we create counter and
> attached it
> to number of flow? Does it count for all flow so counter = flow1 + flow2?
> Or it is only for configuration which means that one each flow gets different
> counter?
> For counter there is a shared bit in the action so this can be used to 
> determent if
> the state
> should be shared,  but counter doesn't have configuration to share.
> What happens with aging where we have state and configuration?
> 
> I'm O.K with starting just with configuration, but it should be clear.
>

Shared action state considered shared till it's defined differently.
So when shared action queried it's considered query for state.
Configuration of shre

Re: [dpdk-dev] [PATCH] test/event_crypto_adapter: fix conf issue

2020-10-06 Thread Gujjar, Abhinandan S
Hi Akhil,

If the patch is ok, could you please ack it?

Regards
Abhinandan

> -Original Message-
> From: Gujjar, Abhinandan S
> Sent: Wednesday, September 23, 2020 12:41 PM
> To: Akhil Goyal ; dev@dpdk.org; jer...@marvell.com
> Cc: Vangati, Narender 
> Subject: RE: [dpdk-dev] [PATCH] test/event_crypto_adapter: fix conf issue
> 
> Hi Akhil,
> 
> Recently, I observed that test was failing with sw adapter.
> I noticed that, it was because of some changes went in related to crypto in
> the past releases.
> One of the tests, configures a session in turn calls
> null_crypto_set_session_parameters()
> which checks for params was failing. As part of fixing & updating the xforms
> to address this issue, thought of getting rid of algo as well, to make the 
> test
> simpler. The test continues to use null cryptodev and does not deal with any
> encryption or auth.
> 
> Regards
> Abhinandan
> 
> > -Original Message-
> > From: Akhil Goyal 
> > Sent: Wednesday, September 23, 2020 12:59 AM
> > To: Gujjar, Abhinandan S ; dev@dpdk.org;
> > jer...@marvell.com
> > Cc: Vangati, Narender 
> > Subject: RE: [dpdk-dev] [PATCH] test/event_crypto_adapter: fix conf
> > issue
> >
> > Hi Abhinandan,
> >
> > > This patch updates the xform with right configuration.
> > > For session based ops, sym session pool is created with valid
> > > userdata size.
> > >
> > > Signed-off-by: Abhinandan Gujjar 
> > > ---
> > The patch description do not match with the patch.
> > I can see that you are changing the cipher algo from AES-CBC to NULL
> > Without explaining the reason.
> > Could you please clarify the need of this patch.
> >
> > Regards,
> > Akhil



Re: [dpdk-dev] [PATCH v3 02/10] ethdev: add conf arg to shared action icreate API

2020-10-06 Thread Andrey Vesnovaty
PSB
> -Original Message-
> From: Ori Kam 
> Sent: Sunday, October 4, 2020 2:11 PM
> To: Andrey Vesnovaty ; dev@dpdk.org
> Cc: j...@marvell.com; jerinjac...@gmail.com; NBU-Contact-Thomas Monjalon
> ; ferruh.yi...@intel.com;
> step...@networkplumber.org; bruce.richard...@intel.com; Slava Ovsiienko
> ; andrey.vesnov...@gmail.com; m...@ashroe.eu;
> nhor...@tuxdriver.com; ajit.khapa...@broadcom.com;
> samik.gu...@broadcom.com; Ori Kam ; Andrew
> Rybchenko 
> Subject: RE: [PATCH v3 02/10] ethdev: add conf arg to shared action icreate 
> API
> 
> Hi Andrey
> 
> Why is this patch not part of the previous one?

Got it. Will be squashed.

> 
> Best,
> Ori
> > -Original Message-
> > From: Andrey Vesnovaty 
> > Sent: Sunday, October 4, 2020 1:06 AM
> > To: dev@dpdk.org
> > Cc: j...@marvell.com; jerinjac...@gmail.com; NBU-Contact-Thomas
> Monjalon
> > ; ferruh.yi...@intel.com;
> > step...@networkplumber.org; bruce.richard...@intel.com; Ori Kam
> > ; Slava Ovsiienko ;
> > andrey.vesnov...@gmail.com; m...@ashroe.eu; nhor...@tuxdriver.com;
> > ajit.khapa...@broadcom.com; samik.gu...@broadcom.com; Ori Kam
> > ; Andrew Rybchenko 
> > Subject: [PATCH v3 02/10] ethdev: add conf arg to shared action icreate API
> >
> > Add configuration argument to shared action create interface.
> > Currently there is only ingress & egress fields but more fields can be
> > added later. Shared action configuration & implementation are PMD
> > specific.
> >
> > Signed-off-by: Andrey Vesnovaty 
> > ---
> >  lib/librte_ethdev/rte_flow.c|  4 +++-
> >  lib/librte_ethdev/rte_flow.h| 17 +++--
> >  lib/librte_ethdev/rte_flow_driver.h |  5 +++--
> >  3 files changed, 21 insertions(+), 5 deletions(-)
> >
> > diff --git a/lib/librte_ethdev/rte_flow.c b/lib/librte_ethdev/rte_flow.c
> > index ba3f01f7c7..9afa8905df 100644
> > --- a/lib/librte_ethdev/rte_flow.c
> > +++ b/lib/librte_ethdev/rte_flow.c
> > @@ -1255,6 +1255,7 @@ rte_flow_get_aged_flows(uint16_t port_id, void
> > **contexts,
> >
> >  struct rte_flow_shared_action *
> >  rte_flow_shared_action_create(uint16_t port_id,
> > + const struct rte_flow_shared_action_conf *conf,
> >   const struct rte_flow_action *action,
> >   struct rte_flow_error *error)
> >  {
> > @@ -1265,7 +1266,8 @@ rte_flow_shared_action_create(uint16_t port_id,
> > if (unlikely(!ops))
> > return NULL;
> > if (likely(!!ops->shared_action_create)) {
> > -   shared_action = ops->shared_action_create(dev, action, error);
> > +   shared_action = ops->shared_action_create(dev, conf, action,
> > + error);
> > if (shared_action == NULL)
> > flow_err(port_id, -rte_errno, error);
> > return shared_action;
> > diff --git a/lib/librte_ethdev/rte_flow.h b/lib/librte_ethdev/rte_flow.h
> > index 02391316cb..8a2db4f6da 100644
> > --- a/lib/librte_ethdev/rte_flow.h
> > +++ b/lib/librte_ethdev/rte_flow.h
> > @@ -3380,6 +3380,16 @@ int
> >  rte_flow_get_aged_flows(uint16_t port_id, void **contexts,
> > uint32_t nb_contexts, struct rte_flow_error *error);
> >
> > +/**
> > + * Specify shared action configuration
> > + */
> > +struct rte_flow_shared_action_conf {
> > +   uint32_t ingress:1;
> > +   /**< Action valid for rules applied to ingress traffic. */
> > +   uint32_t egress:1;
> > +   /**< Action valid for rules applied to egress traffic. */
> > +};
> > +
> >  /**
> >   * @warning
> >   * @b EXPERIMENTAL: this API may change without prior notice.
> > @@ -3388,6 +3398,8 @@ rte_flow_get_aged_flows(uint16_t port_id, void
> > **contexts,
> >   *
> >   * @param[in] port_id
> >   *The port identifier of the Ethernet device.
> > + * @param[in] conf
> > + *   Shared action configuration.
> >   * @param[in] action
> >   *   Action configuration for shared action creation.
> >   * @param[out] error
> > @@ -3404,6 +3416,7 @@ rte_flow_get_aged_flows(uint16_t port_id, void
> > **contexts,
> >  __rte_experimental
> >  struct rte_flow_shared_action *
> >  rte_flow_shared_action_create(uint16_t port_id,
> > + const struct rte_flow_shared_action_conf *conf,
> >   const struct rte_flow_action *action,
> >   struct rte_flow_error *error);
> >
> > @@ -3432,8 +3445,8 @@ rte_flow_shared_action_create(uint16_t port_id,
> >  __rte_experimental
> >  int
> >  rte_flow_shared_action_destroy(uint16_t port_id,
> > - struct rte_flow_shared_action *action,
> > - struct rte_flow_error *error);
> > +  struct rte_flow_shared_action *action,
> > +  struct rte_flow_error *error);
> >
> >  /**
> >   * @warning
> > diff --git a/lib/librte_ethdev/rte_flow_driver.h
> > b/lib/librte_ethdev/rte_flow_driver.h
> > index 72bfc3b7a7..adaace47ea 100644
> > --

Re: [dpdk-dev] [PATCH v3 00/10] RTE flow shared action

2020-10-06 Thread Andrey Vesnovaty
PSB
> -Original Message-
> From: Ori Kam 
> Sent: Sunday, October 4, 2020 2:14 PM
> To: Andrey Vesnovaty ; dev@dpdk.org
> Cc: j...@marvell.com; jerinjac...@gmail.com; NBU-Contact-Thomas Monjalon
> ; ferruh.yi...@intel.com;
> step...@networkplumber.org; bruce.richard...@intel.com; Slava Ovsiienko
> ; andrey.vesnov...@gmail.com; m...@ashroe.eu;
> nhor...@tuxdriver.com; ajit.khapa...@broadcom.com;
> samik.gu...@broadcom.com
> Subject: RE: [dpdk-dev] [PATCH v3 00/10] RTE flow shared action
> 
> Hi Andrey,
> 
> I suggest that you split this patchset to two different ones.
> One for the Mellanox PMD
> and one RTE  just in each patch set reference the second patch.
> This why it is possible to merge only the RTE level, if other PMD will use it.

Got it.
About to publish RTE level v4 part in next couple of hours.
Mellanox PMD part will be published tomorrow. 

> 
> Best,
> Ori
> 
> > -Original Message-
> > From: dev  On Behalf Of Andrey Vesnovaty
> > Sent: Sunday, October 4, 2020 1:06 AM
> > To: dev@dpdk.org
> > Cc: j...@marvell.com; jerinjac...@gmail.com; NBU-Contact-Thomas
> Monjalon
> > ; ferruh.yi...@intel.com;
> > step...@networkplumber.org; bruce.richard...@intel.com; Ori Kam
> > ; Slava Ovsiienko ;
> > andrey.vesnov...@gmail.com; m...@ashroe.eu; nhor...@tuxdriver.com;
> > ajit.khapa...@broadcom.com; samik.gu...@broadcom.com
> > Subject: [dpdk-dev] [PATCH v3 00/10] RTE flow shared action
> >
> > This patchset introduces shared action for RTE flow.
> >
> > V3 changes:
> > - implement testpmd for shared action
> > - fix flow_filtering example application
> > - add conf arg to shared action create API
> >
> > Notes: PMD implementation should be considered as draft
> >
> > Andrey Vesnovaty (10):
> >   ethdev: add flow shared action API
> >   ethdev: add conf arg to shared action icreate API
> >   common/mlx5: modify advanced Rx object via DevX
> >   net/mlx5: modify hash Rx queue objects
> >   net/mlx5: shared action PMD
> >   net/mlx5: shared action PMD create conf arg
> >   net/mlx5: driver support for shared action
> >   net/mlx5: shared action create conf drv support
> >   examples/flow_filtering: utilize shared RSS action
> >   app/testpmd: support shared action
> >
> >  app/test-pmd/cmdline_flow.c   | 262 ++-
> >  app/test-pmd/config.c | 217 ++
> >  app/test-pmd/testpmd.h|  19 +
> >  doc/guides/sample_app_ug/flow_filtering.rst   |  62 +-
> >  drivers/common/mlx5/mlx5_devx_cmds.c  |  84 +++
> >  drivers/common/mlx5/mlx5_devx_cmds.h  |  10 +
> >  drivers/common/mlx5/mlx5_prm.h|  29 +
> >  .../common/mlx5/rte_common_mlx5_version.map   |   1 +
> >  drivers/net/mlx5/mlx5.c   |   1 +
> >  drivers/net/mlx5/mlx5.h   |   6 +
> >  drivers/net/mlx5/mlx5_defs.h  |   3 +
> >  drivers/net/mlx5/mlx5_devx.c  | 178 -
> >  drivers/net/mlx5/mlx5_flow.c  | 497 -
> >  drivers/net/mlx5/mlx5_flow.h  |  86 +++
> >  drivers/net/mlx5/mlx5_flow_dv.c   | 684 +-
> >  drivers/net/mlx5/mlx5_rxq.c   | 103 +++
> >  drivers/net/mlx5/mlx5_rxtx.h  |   5 +-
> >  examples/flow_filtering/flow_blocks.c |  81 ++-
> >  examples/flow_filtering/main.c|  13 +-
> >  lib/librte_ethdev/rte_ethdev_version.map  |   4 +
> >  lib/librte_ethdev/rte_flow.c  |  84 +++
> >  lib/librte_ethdev/rte_flow.h  | 161 -
> >  lib/librte_ethdev/rte_flow_driver.h   |  23 +
> >  23 files changed, 2489 insertions(+), 124 deletions(-)
> >
> > --
> > 2.26.2



Re: [dpdk-dev] [PATCH v2 3/3] test/event_crypto_adapter: free resources during exit

2020-10-06 Thread Gujjar, Abhinandan S



> -Original Message-
> From: Gujjar, Abhinandan S
> Sent: Tuesday, October 6, 2020 3:52 PM
> To: 'Ankur Dwivedi' ; dev@dpdk.org
> Cc: jer...@marvell.com; ano...@marvell.com
> Subject: RE: [PATCH v2 3/3] test/event_crypto_adapter: free resources
> during exit
> 
> Hi Ankur,
> 
> Please see comment below.
> 
> > -Original Message-
> > From: Ankur Dwivedi 
> > Sent: Friday, September 25, 2020 12:59 PM
> > To: dev@dpdk.org
> > Cc: Gujjar, Abhinandan S ;
> > jer...@marvell.com; ano...@marvell.com; Ankur Dwivedi
> > 
> > Subject: [PATCH v2 3/3] test/event_crypto_adapter: free resources
> > during exit
> >
> > The resources held by crypto adapter should be freed when the test
> > suite exits.
> >
> > Signed-off-by: Ankur Dwivedi 
> > ---
> >  app/test/test_event_crypto_adapter.c | 15 +++
> >  1 file changed, 15 insertions(+)
> >
> > diff --git a/app/test/test_event_crypto_adapter.c
> > b/app/test/test_event_crypto_adapter.c
> > index 572ef9c27..4b09ab3b7 100644
> > --- a/app/test/test_event_crypto_adapter.c
> > +++ b/app/test/test_event_crypto_adapter.c
> > @@ -899,6 +899,20 @@ testsuite_setup(void)
> > return TEST_SUCCESS;
> >  }
> >
> > +static void
> > +crypto_adapter_teardown(void)
> > +{
> > +   int ret;
> > +

Below APIs could return error. So, set this flag at the end?
> > +   crypto_adapter_setup_done = 0; 

> Call rte_event_crypto_adapter_stop() here?
> > +   ret =
> > rte_event_crypto_adapter_queue_pair_del(TEST_ADAPTER_ID,
> > +   TEST_CDEV_ID, TEST_CDEV_QP_ID);
> > +   if (ret < 0)
> > +   RTE_LOG(ERR, USER1, "Failed to delete queue pair!");
> > +
> > +   rte_event_crypto_adapter_free(TEST_ADAPTER_ID);
> > +}
> > +
> >  static void
> >  crypto_teardown(void)
> >  {
> > @@ -941,6 +955,7 @@ eventdev_teardown(void)  static void
> >  testsuite_teardown(void)
> >  {
> > +   crypto_adapter_teardown();
> > crypto_teardown();
> > eventdev_teardown();
> >  }
> > --
> > 2.28.0



Re: [dpdk-dev] [PATCH v3 09/10] examples/flow_filtering: utilize shared RSS action

2020-10-06 Thread Andrey Vesnovaty
Hi Ori,

PSB

Thanks,
Andrey

> -Original Message-
> From: Ori Kam 
> Sent: Sunday, October 4, 2020 2:22 PM
> To: Andrey Vesnovaty ; dev@dpdk.org
> Cc: j...@marvell.com; jerinjac...@gmail.com; NBU-Contact-Thomas Monjalon
> ; ferruh.yi...@intel.com;
> step...@networkplumber.org; bruce.richard...@intel.com; Slava Ovsiienko
> ; andrey.vesnov...@gmail.com; m...@ashroe.eu;
> nhor...@tuxdriver.com; ajit.khapa...@broadcom.com;
> samik.gu...@broadcom.com; Andrey Vesnovaty ;
> Marko Kovacevic ; Ori Kam
> ; Radu Nicolau ; Akhil Goyal
> ; Tomasz Kantecki ; Sunil
> Kumar Kori ; Pavan Nikhilesh
> ; John McNamara 
> Subject: RE: [PATCH v3 09/10] examples/flow_filtering: utilize shared RSS 
> action
> 
> Hi Andrey,
> 
> I don't think this change should be part of the series.

Now there is PMD support for shared action & having some example isn't critical
For this patch series. Removing it. 

> After you update this example will only work if the PMD support shared action.
> Not all PMD supports it, and in any case this example is to show basic rules 
> and
> RSS
> your code changes the app that the RSS action can't be checked.

There is a fallback for PMDs without shared action support but I that
agree, it would be better to provide separate example for shared RSS.

> 
> Best,
> Ori
> 
> 
> > -Original Message-
> > From: Andrey Vesnovaty 
> > Sent: Sunday, October 4, 2020 1:06 AM
> > Subject: [PATCH v3 09/10] examples/flow_filtering: utilize shared RSS action
> >
> > From: Andrey Vesnovaty 
> >
> > This commit give very first shared RSS action usage example.
> > Queue action used by the flow replaced by shared RSS action
> > having single queue. On each RX burst queue switched 0 <-> 1
> > utilizing rte_flow_shared_action_update() API. User supposed
> > to observe consistent queue switches on each packet burst.
> >
> > Signed-off-by: Andrey Vesnovaty 
> > Signed-off-by: Andrey Vesnovaty 
> > ---
> >  doc/guides/sample_app_ug/flow_filtering.rst | 62 +---
> >  examples/flow_filtering/flow_blocks.c   | 81 +
> >  examples/flow_filtering/main.c  | 13 +++-
> >  3 files changed, 128 insertions(+), 28 deletions(-)
> >
> > diff --git a/doc/guides/sample_app_ug/flow_filtering.rst
> > b/doc/guides/sample_app_ug/flow_filtering.rst
> > index 5e5a6cd8a0..cfe9334717 100644
> > --- a/doc/guides/sample_app_ug/flow_filtering.rst
> > +++ b/doc/guides/sample_app_ug/flow_filtering.rst
> > @@ -106,7 +106,7 @@ following code:
> >  .. code-block:: c
> >
> > /* create flow for send packet with */
> > -   flow = generate_ipv4_flow(port_id, selected_queue,
> > +   flow = generate_ipv4_flow(port_id, shared_action,
> >  SRC_IP, EMPTY_MASK,
> >  DEST_IP, FULL_MASK, &error);
> > if (!flow) {
> > @@ -242,7 +242,7 @@ The Ethernet port is configured with default settings
> > using the
> > rxq_conf = dev_info.default_rxconf;
> > rxq_conf.offloads = port_conf.rxmode.offloads;
> >
> > -For this example we are configuring number of rx and tx queues that are
> > connected
> > +For this example we are configuring 2 rx and 2 tx queues that are connected
> >  to a single port.
> >
> >  .. code-block:: c
> > @@ -270,13 +270,22 @@ to a single port.
> >}
> > }
> >
> > +Before we create the flow we create shared action in order to send it as
> > +actions argument when creating a flow. The action is single queue RSS 
> > action
> > +similar to action queue with the only difference that shared RSS action
> > +provides update capability after action creation.
> > +
> > +.. code-block:: c
> > +
> > +   shared_action = rte_flow_shared_action_create(port_id, &action, &error);
> > +
> >  In the next step we create and apply the flow rule. which is to send 
> > packets
> >  with destination ip equals to 192.168.1.1 to queue number 1. The detail
> >  explanation of the ``generate_ipv4_flow()`` appears later in this document:
> >
> >  .. code-block:: c
> >
> > -   flow = generate_ipv4_flow(port_id, selected_queue,
> > +   flow = generate_ipv4_flow(port_id, shared_action,
> >   SRC_IP, EMPTY_MASK,
> >   DEST_IP, FULL_MASK, &error);
> >
> > @@ -339,6 +348,21 @@ looks like the following:
> > printf("\n");
> > rte_pktmbuf_free(m);
> > }
> > +   if (rss_queue[0] == 0) {
> > +   printf(">>> switching queue 0 
> > -> 1\n");
> > +   rss_queue[0] = 1;
> > +   } else {
> > +   printf(">>> switching queue 1 
> > -> 0\n");
> > +   rss_queue[0] = 0;
> > +   }
> > +   ret = r

Re: [dpdk-dev] [PATCH v3 10/10] app/testpmd: support shared action

2020-10-06 Thread Andrey Vesnovaty
Hi Ori.

testpmd documentation will be added in v4.

Thanks,
Andrey

> -Original Message-
> From: Ori Kam 
> Sent: Sunday, October 4, 2020 3:05 PM
> To: Ori Kam ; Andrey Vesnovaty ;
> dev@dpdk.org
> Cc: j...@marvell.com; jerinjac...@gmail.com; NBU-Contact-Thomas Monjalon
> ; ferruh.yi...@intel.com;
> step...@networkplumber.org; bruce.richard...@intel.com; Slava Ovsiienko
> ; andrey.vesnov...@gmail.com; m...@ashroe.eu;
> nhor...@tuxdriver.com; ajit.khapa...@broadcom.com;
> samik.gu...@broadcom.com; Ori Kam ; Wenzhuo Lu
> ; Beilei Xing ; Bernard
> Iremonger 
> Subject: RE: [PATCH v3 10/10] app/testpmd: support shared action
> 
> Hi Andrey,
> 
> I think you are missing documentation:
> Do/guides/testpmd_app_ug/testpmd_funcs.rst
> Best,
> Ori
> 
> > -Original Message-
> > From: Ori Kam 
> > Sent: Sunday, October 4, 2020 2:28 PM
> .x...@intel.com>; Bernard
> > Iremonger 
> > Subject: RE: [PATCH v3 10/10] app/testpmd: support shared action
> >
> > Hi Andrey
> >
> > > -Original Message-
> > > From: Andrey Vesnovaty 
> > > Sent: Sunday, October 4, 2020 1:06 AM
> > > Subject: [PATCH v3 10/10] app/testpmd: support shared action
> > >
> > > This patch adds shared action support to testpmd CLI.
> > >
> > > All shared actions created via testpmd CLI assigned ID for further
> > > reference in other CLI commands. Shared action ID supplied as CLI
> > > argument or assigned by testpmd is similar to flow ID & limited to
> > > scope of testpdm CLI.
> > >
> > > Create shared action syntax:
> > > flow shared_action (port) create {action_id (shared_action_id)} \
> > > (action) / end
> > >
> > > Create shared action examples:
> > > flow shared_action 0 create action_id 100 \
> > > rss queues 1 2 end / end
> > > This creates shared rss action with id 100 on port 0.
> > >
> > > flow shared_action 0 create action_id \
> > > rss queues 0 1 end / end
> > > This creates shared rss action with id assigned by tetspmd
> > > on port 0.
> > >
> > > Update shared action syntax:
> > > flow shared_action (port) update (shared_action_id) (action) / end
> > >
> > > Update shared action example:
> > > flow shared_action 0 update 100 rss queues 0 3 end / end
> > > This updates shared rss action having id 100 on port 0
> > > with rss to queues 0 3 (in create example rss queues were
> > > 1 & 2).
> > >
> > > Destroy shared action syntax:
> > > flow shared_action (port) destroy action_id (shared_action_id) \
> > > { action_id (shared_action_id) [...]}
> > >
> > > Update shared action example:
> > > flow shared_action 0 destroy action_id 100 action_id 101
> > > This destroys shared actions having id 100 & 101
> > >
> > > Query shared action syntax:
> > > flow shared_action (port) query (shared_action_id)
> > >
> > > Query shared action example:
> > > flow shared_action 0 query 100
> > > This queries shared actions having id 100
> > >
> > > Use shared action as flow action syntax:
> > > flow create (port) ... / end actions {action / [...]} \
> > > shared (action_id) / {action / [...]} end
> > >
> > > Use shared action as flow action example:
> > > flow create 0 ingress pattern ... / end \
> > > actions shared 100 / end
> > > This creates flow rule where rss action is shared rss action
> > > having id 100.
> > >
> > > All shared action CLIs report status of the command.
> > > Shared action query CLI output depends on action type.
> > >
> > > Signed-off-by: Andrey Vesnovaty 
> > > ---
> >
> > Acked-by: Ori Kam 
> > Thanks,
> > Ori


Re: [dpdk-dev] [PATCH 1/3 v2] ethdev: add rx offload to drop error packets

2020-10-06 Thread Nipun Gupta


> -Original Message-
> From: Jerin Jacob 
> Sent: Monday, October 5, 2020 9:40 PM
> To: Stephen Hemminger 
> Cc: Nipun Gupta ; dpdk-dev ; Thomas
> Monjalon ; Ferruh Yigit ;
> Andrew Rybchenko ; Hemant Agrawal
> ; Sachin Saxena ; Rohit
> Raj 
> Subject: Re: [dpdk-dev] [PATCH 1/3 v2] ethdev: add rx offload to drop error
> packets
> 
> On Mon, Oct 5, 2020 at 9:05 PM Stephen Hemminger
>  wrote:
> >
> > On Mon,  5 Oct 2020 12:45:04 +0530
> > nipun.gu...@nxp.com wrote:
> >
> > > From: Nipun Gupta 
> > >
> > > This change adds a RX offload capability, which once enabled,
> > > hardware will drop the packets in case there of any error in
> > > the packet such as L3 checksum error or L4 checksum.
> 
> IMO, Providing additional support up to the level to choose the errors
> to drops give more control to the application. Meaning,
> L1 errors such as FCS error
> L2 errors ..
> L3 errors such checksum
> i.e ethdev spec need to have  error level supported by PMD and the
> application can set the layers interested to drop.

Agree, but 'DEV_RX_OFFLOAD_ERR_PKT_DROP' shall also be there to drop all the 
error packets? Maybe we can rename it to DEV_RX_OFFLOAD_ALL_ERR_PKT_DROP.

Currently we have not planned to add separate knobs for separate error in
the driver, maybe we can define them separately, or we need have them in
this series itself?

> 
> > >
> > > Signed-off-by: Nipun Gupta 
> > > Signed-off-by: Rohit Raj 
> > > ---
> > > These patches are based over series:
> > >
> https://eur01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fpatchwo
> rk.dpdk.org%2Fpatch%2F78630%2F&data=02%7C01%7Cnipun.gupta%40nx
> p.com%7C90b516fd465c48945e7008d869492b3e%7C686ea1d3bc2b4c6fa92cd9
> 9c5c301635%7C0%7C0%7C637375110263097933&sdata=RBQswMBsfpM6
> nyKur%2FaHvOMvNK7RU%2BRyhHt%2FXBsP1OM%3D&reserved=0
> > >
> > > Changes in v2:
> > >  - Add support in DPAA1 driver (patch 2/3)
> > >  - Add support and config parameter in testpmd (patch 3/3)
> > >
> > >  lib/librte_ethdev/rte_ethdev.h | 1 +
> > >  1 file changed, 1 insertion(+)
> >
> > Maybe this should be an rte_flow match/action which would then make it
> > more flexible?
> 
> I think, it is not based on any Patten matching. So IMO, it should be best if 
> it
> is part of RX offload.
> 
> >
> > There is not much of a performance gain for this in real life and
> > if only one driver supports it then I am not convinced this is needed.
> 
> Marvell HW has this feature.


[dpdk-dev] [PATCH] crypto/aesni_mb: support AES-CCM-256

2020-10-06 Thread Pablo de Lara
This patch adds support for AES-CCM-256 when using AESNI-MB

Signed-off-by: Pablo de Lara 
---
 doc/guides/cryptodevs/features/aesni_mb.ini| 1 +
 doc/guides/rel_notes/release_20_11.rst | 4 
 drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c | 5 +
 3 files changed, 10 insertions(+)

diff --git a/doc/guides/cryptodevs/features/aesni_mb.ini 
b/doc/guides/cryptodevs/features/aesni_mb.ini
index 38d255aff..58afb203e 100644
--- a/doc/guides/cryptodevs/features/aesni_mb.ini
+++ b/doc/guides/cryptodevs/features/aesni_mb.ini
@@ -55,6 +55,7 @@ AES GMAC = Y
 ;
 [AEAD]
 AES CCM (128) = Y
+AES CCM (256) = Y
 AES GCM (128) = Y
 AES GCM (192) = Y
 AES GCM (256) = Y
diff --git a/doc/guides/rel_notes/release_20_11.rst 
b/doc/guides/rel_notes/release_20_11.rst
index 6d8c24413..6a2d000d3 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -210,6 +210,10 @@ API Changes
 
 * bpf: ``RTE_BPF_XTYPE_NUM`` has been dropped from ``rte_bpf_xtype``.
 
+* **Updated the AESNI MB crypto PMD.**
+
+  * Updated the AESNI MB PMD with AES-256 CCM algorithm.
+
 
 ABI Changes
 ---
diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c 
b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
index 2362f0c3c..7759a9873 100644
--- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
+++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
@@ -400,8 +400,13 @@ static const struct rte_cryptodev_capabilities 
aesni_mb_pmd_capabilities[] = {
.block_size = 16,
.key_size = {
.min = 16,
+#if IMB_VERSION(0, 54, 2) <= IMB_VERSION_NUM
+   .max = 32,
+   .increment = 16
+#else
.max = 16,
.increment = 0
+#endif
},
.digest_size = {
.min = 4,
-- 
2.25.1



Re: [dpdk-dev] [PATCH v4] kernel/linux: remove igb_uio

2020-10-06 Thread Thomas Monjalon
06/10/2020 12:16, Ferruh Yigit:
> On 10/5/2020 10:42 AM, Thomas Monjalon wrote:
> > As decided in the Technical Board in November 2019,
> > the kernel module igb_uio is moved to the dpdk-kmods repository
> > in the /linux/igb_uio/ directory.

[...]
> Patch doesn't apply cleanly, not sure why, CI seems having same problem:
> http://mails.dpdk.org/archives/test-report/2020-October/156378.html

That's because I used --irreversible-delete to minimize the size of the patch.

[...]
> >  If UEFI secure boot is enabled, the Linux kernel may disallow the use 
> > of
> >  UIO on the system. Therefore, devices for use by DPDK should be bound 
> > to the
> > -   ``vfio-pci`` kernel module rather than ``igb_uio`` or 
> > ``uio_pci_generic``.
> > +   ``vfio-pci`` kernel module rather than any UIO-based module.
> >  For more details see :ref:`linux_gsg_binding_kernel` below.
> 
> Should we change the order of the kernel drivers, currently first section is 
> "UIO" and second section is "VFIO". Since we encoruage vfio and move igb_uio 
> out 
> of the repo, should we list the "VfIO" first?

Absolutely.
In general we need to encourage more VFIO and remove some references
to igb_uio. But I don't want to do such "marketing" change in this patch.
The goal of this patch is just to move the code.

[...]
> > +* kernel: The module ``igb_uio`` has been moved to the git repository
> > +``dpdk-kmods`` in a new directory ``linux/igb_uio``.
> > +
> 
> Doc build giving warning on this:
> WARNING: Bullet list ends without a blank line; unexpected unindent.

Probably a merge issue. That's fine on my side.

> Except from above minor comments looks good to me,
> Reviewed-by: Ferruh Yigit 

Do we agree I can merge without any change?




Re: [dpdk-dev] [PATCH v4 3/3] ethdev: allow close function to return an error

2020-10-06 Thread Thomas Monjalon
06/10/2020 11:43, Ferruh Yigit:
> On 10/5/2020 6:08 PM, Thomas Monjalon wrote:
> > The API function rte_eth_dev_close() was returning void.
> > The return type is changed to int for notifying of errors.
> > 
> > If an error happens during a close operation,
> > the status of the port is undefined,
> > a maximum of resources having been freed.
> > 
> > Signed-off-by: Thomas Monjalon 
> > Reviewed-by: Liron Himi 
> > Acked-by: Stephen Hemminger 
> 
> <...>
> 
> > -void
> > +int
> >   rte_eth_dev_close(uint16_t port_id)
> >   {
> > struct rte_eth_dev *dev;
> > +   int firsterr, binerr;
> > +   int *lasterr = &firsterr;
> >   
> > -   RTE_ETH_VALID_PORTID_OR_RET(port_id);
> > +   RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -EINVAL);
> > dev = &rte_eth_devices[port_id];
> >   
> > -   RTE_FUNC_PTR_OR_RET(*dev->dev_ops->dev_close);
> > -   (*dev->dev_ops->dev_close)(dev);
> > +   RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->dev_close, -ENOTSUP);
> > +   *lasterr = (*dev->dev_ops->dev_close)(dev);
> > +   if (*lasterr != 0)
> > +   lasterr = &binerr;
> >   
> > rte_ethdev_trace_close(port_id);
> > -   rte_eth_dev_release_port(dev);
> > +   *lasterr = rte_eth_dev_release_port(dev);
> > +
> > +   return firsterr;
> >   }
> 
> This may be personal taste but above error handling looks like unnecessary 
> complex, what do you think something like below:
> 
> close_err = (*dev->dev_ops->dev_close)(dev);
> release_err = rte_eth_dev_release_port(dev);
> return close_err ? close_err : release_err;

This is what I did first. Then thought about this "elegant" handling.
The pointer solution is just one more line and is more future proof.

I'm fine with any choice. Andrew?




[dpdk-dev] [PATCH 1/3] crypto/aesni_mb: fix CCM digest size check

2020-10-06 Thread Pablo de Lara
Digest size for CCM was being checked for other algorithms
apart from CCM.

Fixes: c4c0c312a823 ("crypto/aesni_mb: check for invalid digest size")
Cc: sta...@dpdk.org

Signed-off-by: Pablo de Lara 
---
 drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c | 29 +++---
 1 file changed, 15 insertions(+), 14 deletions(-)

diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c 
b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
index 1bddbcf74..784278719 100644
--- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
+++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
@@ -564,6 +564,14 @@ aesni_mb_set_session_aead_parameters(const MB_MGR *mb_mgr,
return -EINVAL;
}
 
+   /* Set IV parameters */
+   sess->iv.offset = xform->aead.iv.offset;
+   sess->iv.length = xform->aead.iv.length;
+
+   /* Set digest sizes */
+   sess->auth.req_digest_len = xform->aead.digest_length;
+   sess->auth.gen_digest_len = sess->auth.req_digest_len;
+
switch (xform->aead.algo) {
case RTE_CRYPTO_AEAD_AES_CCM:
sess->cipher.mode = CCM;
@@ -582,6 +590,13 @@ aesni_mb_set_session_aead_parameters(const MB_MGR *mb_mgr,
return -EINVAL;
}
 
+   /* CCM digests must be between 4 and 16 and an even number */
+   if (sess->auth.req_digest_len < AES_CCM_DIGEST_MIN_LEN ||
+   sess->auth.req_digest_len > 
AES_CCM_DIGEST_MAX_LEN ||
+   (sess->auth.req_digest_len & 1) == 1) {
+   AESNI_MB_LOG(ERR, "Invalid digest size\n");
+   return -EINVAL;
+   }
break;
 
case RTE_CRYPTO_AEAD_AES_GCM:
@@ -616,20 +631,6 @@ aesni_mb_set_session_aead_parameters(const MB_MGR *mb_mgr,
return -ENOTSUP;
}
 
-   /* Set IV parameters */
-   sess->iv.offset = xform->aead.iv.offset;
-   sess->iv.length = xform->aead.iv.length;
-
-   sess->auth.req_digest_len = xform->aead.digest_length;
-   /* CCM digests must be between 4 and 16 and an even number */
-   if (sess->auth.req_digest_len < AES_CCM_DIGEST_MIN_LEN ||
-   sess->auth.req_digest_len > AES_CCM_DIGEST_MAX_LEN ||
-   (sess->auth.req_digest_len & 1) == 1) {
-   AESNI_MB_LOG(ERR, "Invalid digest size\n");
-   return -EINVAL;
-   }
-   sess->auth.gen_digest_len = sess->auth.req_digest_len;
-
return 0;
 }
 
-- 
2.25.1



[dpdk-dev] [PATCH 2/3] crypto/aesni_mb: fix GCM digest size check

2020-10-06 Thread Pablo de Lara
GCM digest sizes should be between 1 and 16 bytes.

Fixes: 7b2d4706c90e ("crypto/aesni_mb: support newer library version only")
Cc: sta...@dpdk.org

Signed-off-by: Pablo de Lara 
---
 .../crypto/aesni_mb/aesni_mb_pmd_private.h|  4 ++--
 drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c| 22 +--
 .../crypto/aesni_mb/rte_aesni_mb_pmd_ops.c|  8 +++
 3 files changed, 16 insertions(+), 18 deletions(-)

diff --git a/drivers/crypto/aesni_mb/aesni_mb_pmd_private.h 
b/drivers/crypto/aesni_mb/aesni_mb_pmd_private.h
index e0c7b4f7c..8c5acfc51 100644
--- a/drivers/crypto/aesni_mb/aesni_mb_pmd_private.h
+++ b/drivers/crypto/aesni_mb/aesni_mb_pmd_private.h
@@ -80,7 +80,7 @@ static const unsigned auth_truncated_digest_byte_lengths[] = {
[AES_CMAC]  = 12,
[AES_CCM]   = 8,
[NULL_HASH] = 0,
-   [AES_GMAC]  = 16,
+   [AES_GMAC]  = 12,
[PLAIN_SHA1]= 20,
[PLAIN_SHA_224] = 28,
[PLAIN_SHA_256] = 32,
@@ -111,7 +111,7 @@ static const unsigned auth_digest_byte_lengths[] = {
[AES_XCBC]  = 16,
[AES_CMAC]  = 16,
[AES_CCM]   = 16,
-   [AES_GMAC]  = 12,
+   [AES_GMAC]  = 16,
[NULL_HASH] = 0,
[PLAIN_SHA1]= 20,
[PLAIN_SHA_224] = 28,
diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c 
b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
index 784278719..fa364530e 100644
--- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
+++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
@@ -209,19 +209,11 @@ aesni_mb_set_session_auth_parameters(const MB_MGR *mb_mgr,
sess->cipher.direction = DECRYPT;
 
sess->auth.algo = AES_GMAC;
-   /*
-* Multi-buffer lib supports 8, 12 and 16 bytes of digest.
-* If size requested is different, generate the full digest
-* (16 bytes) in a temporary location and then memcpy
-* the requested number of bytes.
-*/
-   if (sess->auth.req_digest_len != 16 &&
-   sess->auth.req_digest_len != 12 &&
-   sess->auth.req_digest_len != 8) {
-   sess->auth.gen_digest_len = 16;
-   } else {
-   sess->auth.gen_digest_len = sess->auth.req_digest_len;
+   if (sess->auth.req_digest_len > 
get_digest_byte_length(AES_GMAC)) {
+   AESNI_MB_LOG(ERR, "Invalid digest size\n");
+   return -EINVAL;
}
+   sess->auth.gen_digest_len = sess->auth.req_digest_len;
sess->iv.length = xform->auth.iv.length;
sess->iv.offset = xform->auth.iv.offset;
 
@@ -624,6 +616,12 @@ aesni_mb_set_session_aead_parameters(const MB_MGR *mb_mgr,
return -EINVAL;
}
 
+   /* GCM digest size must be between 1 and 16 */
+   if (sess->auth.req_digest_len == 0 ||
+   sess->auth.req_digest_len > 16) {
+   AESNI_MB_LOG(ERR, "Invalid digest size\n");
+   return -EINVAL;
+   }
break;
 
default:
diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c 
b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
index 2362f0c3c..3e4282954 100644
--- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
+++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
@@ -455,9 +455,9 @@ static const struct rte_cryptodev_capabilities 
aesni_mb_pmd_capabilities[] = {
.increment = 8
},
.digest_size = {
-   .min = 8,
+   .min = 1,
.max = 16,
-   .increment = 4
+   .increment = 1
},
.aad_size = {
.min = 0,
@@ -485,9 +485,9 @@ static const struct rte_cryptodev_capabilities 
aesni_mb_pmd_capabilities[] = {
.increment = 8
},
.digest_size = {
-   .min = 8,
+   .min = 1,
.max = 16,
-   .increment = 4
+   .increment = 1
},
.iv_size = {
.min = 12,
-- 
2.25.1



[dpdk-dev] [PATCH 3/3] crypto/aesni_mb: support Chacha20-Poly1305

2020-10-06 Thread Pablo de Lara
Add support for Chacha20-Poly1305 AEAD algorithm.

Signed-off-by: Pablo de Lara 
---
 doc/guides/cryptodevs/aesni_mb.rst|  1 +
 doc/guides/cryptodevs/features/aesni_mb.ini   | 10 +--
 doc/guides/rel_notes/release_20_11.rst|  3 +
 drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c| 63 ---
 .../crypto/aesni_mb/rte_aesni_mb_pmd_ops.c| 32 ++
 5 files changed, 97 insertions(+), 12 deletions(-)

diff --git a/doc/guides/cryptodevs/aesni_mb.rst 
b/doc/guides/cryptodevs/aesni_mb.rst
index 15388d20a..cf7ad5d57 100644
--- a/doc/guides/cryptodevs/aesni_mb.rst
+++ b/doc/guides/cryptodevs/aesni_mb.rst
@@ -54,6 +54,7 @@ AEAD algorithms:
 
 * RTE_CRYPTO_AEAD_AES_CCM
 * RTE_CRYPTO_AEAD_AES_GCM
+* RTE_CRYPTO_AEAD_CHACHA20_POLY1305
 
 Protocol offloads:
 
diff --git a/doc/guides/cryptodevs/features/aesni_mb.ini 
b/doc/guides/cryptodevs/features/aesni_mb.ini
index 38d255aff..2e8305709 100644
--- a/doc/guides/cryptodevs/features/aesni_mb.ini
+++ b/doc/guides/cryptodevs/features/aesni_mb.ini
@@ -54,11 +54,11 @@ AES GMAC = Y
 ; Supported AEAD algorithms of the 'aesni_mb' crypto driver.
 ;
 [AEAD]
-AES CCM (128) = Y
-AES GCM (128) = Y
-AES GCM (192) = Y
-AES GCM (256) = Y
-
+AES CCM (128) = Y
+AES GCM (128) = Y
+AES GCM (192) = Y
+AES GCM (256) = Y
+CHACHA20-POLY1305 = Y
 ;
 ; Supported Asymmetric algorithms of the 'aesni_mb' crypto driver.
 ;
diff --git a/doc/guides/rel_notes/release_20_11.rst 
b/doc/guides/rel_notes/release_20_11.rst
index 6d8c24413..f606c9a74 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -210,6 +210,9 @@ API Changes
 
 * bpf: ``RTE_BPF_XTYPE_NUM`` has been dropped from ``rte_bpf_xtype``.
 
+* **Updated the AESNI MB crypto PMD.**
+
+  * Added support for Chacha20-Poly1305.
 
 ABI Changes
 ---
diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c 
b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
index fa364530e..7b4d5f148 100644
--- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
+++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd.c
@@ -125,6 +125,18 @@ aesni_mb_get_chain_order(const struct rte_crypto_sym_xform 
*xform)
return AESNI_MB_OP_NOT_SUPPORTED;
 }
 
+static inline int
+is_aead_algo(JOB_HASH_ALG hash_alg, JOB_CIPHER_MODE cipher_mode)
+{
+#if IMB_VERSION(0, 54, 3) <= IMB_VERSION_NUM
+   return (hash_alg == IMB_AUTH_CHACHA20_POLY1305 || hash_alg == AES_CCM ||
+   (hash_alg == AES_GMAC && cipher_mode == GCM));
+#else
+   return ((hash_alg == AES_GMAC && cipher_mode == GCM) ||
+   hash_alg == AES_CCM);
+#endif
+}
+
 /** Set session authentication parameters */
 static int
 aesni_mb_set_session_auth_parameters(const MB_MGR *mb_mgr,
@@ -624,6 +636,24 @@ aesni_mb_set_session_aead_parameters(const MB_MGR *mb_mgr,
}
break;
 
+#if IMB_VERSION(0, 54, 3) <= IMB_VERSION_NUM
+   case RTE_CRYPTO_AEAD_CHACHA20_POLY1305:
+   sess->cipher.mode = IMB_CIPHER_CHACHA20_POLY1305;
+   sess->auth.algo = IMB_AUTH_CHACHA20_POLY1305;
+
+   if (xform->aead.key.length != 32) {
+   AESNI_MB_LOG(ERR, "Invalid key length");
+   return -EINVAL;
+   }
+   sess->cipher.key_length_in_bytes = 32;
+   memcpy(sess->cipher.expanded_aes_keys.encode,
+   xform->aead.key.data, 32);
+   if (sess->auth.req_digest_len != 16) {
+   AESNI_MB_LOG(ERR, "Invalid digest size\n");
+   return -EINVAL;
+   }
+   break;
+#endif
default:
AESNI_MB_LOG(ERR, "Unsupported aead mode parameter");
return -ENOTSUP;
@@ -1122,6 +1152,8 @@ set_mb_job_params(JOB_AES_HMAC *job, struct aesni_mb_qp 
*qp,
/* Set authentication parameters */
job->hash_alg = session->auth.algo;
 
+   const int aead = is_aead_algo(job->hash_alg, job->cipher_mode);
+
switch (job->hash_alg) {
case AES_XCBC:
job->u.XCBC._k1_expanded = session->auth.xcbc.k1_expanded;
@@ -1168,6 +1200,14 @@ set_mb_job_params(JOB_AES_HMAC *job, struct aesni_mb_qp 
*qp,
job->aes_dec_key_expanded = &session->cipher.gcm_key;
break;
 
+#if IMB_VERSION(0, 54, 3) <= IMB_VERSION_NUM
+   case IMB_AUTH_CHACHA20_POLY1305:
+   job->u.CHACHA20_POLY1305.aad = op->sym->aead.aad.data;
+   job->u.CHACHA20_POLY1305.aad_len_in_bytes = 
session->aead.aad_len;
+   job->aes_enc_key_expanded = 
session->cipher.expanded_aes_keys.encode;
+   job->aes_dec_key_expanded = 
session->cipher.expanded_aes_keys.encode;
+   break;
+#endif
default:
job->u.HMAC._hashed_auth_key_xor_ipad = 
session->auth.pads.inner;
job->u.HMAC._hashed_auth_key_xor_opad = 
session->auth.pads.outer;
@@ -1199,8 +1239,7 @@ set_mb_job_params(JOB_AES_HMAC *job

Re: [dpdk-dev] [PATCH v2 00/56] net: txgbe PMD

2020-10-06 Thread Ferruh Yigit

On 10/5/2020 1:08 PM, Jiawen Wu wrote:

v2: re-order patches and fix some known problems
v1: introduce txgbe PMD

jiawenwu (56):
   net/txgbe: add build and doc infrastructure
   net/txgbe: add ethdev probe and remove
   net/txgbe: add device init and uninit
   net/txgbe: add error types and registers
   net/txgbe: add mac type and bus lan id
   net/txgbe: add HW infrastructure and dummy function
   net/txgbe: add EEPROM functions
   net/txgbe: add HW init and reset operation
   net/txgbe: add PHY init
   net/txgbe: add module identify
   net/txgbe: add PHY reset
   net/txgbe: add info get operation
   net/txgbe: add interrupt operation
   net/txgbe: add device configure operation
   net/txgbe: add link status change
   net/txgbe: add multi-speed link setup
   net/txgbe: add autoc read and write
   net/txgbe: add MAC address operations
   net/txgbe: add unicast hash bitmap
   net/txgbe: add RX and TX init
   net/txgbe: add RX and TX queues setup and release
   net/txgbe: add RX and TX start and stop
   net/txgbe: add packet type
   net/txgbe: fill simple transmit function
   net/txgbe: fill transmit function with hardware offload
   net/txgbe: fill TX prepare funtion
   net/txgbe: fill receive functions
   net/txgbe: add device start operation
   net/txgbe: add RX and TX data path start and stop
   net/txgbe: add device stop and close operations
   net/txgbe: support RX interrupt
   net/txgbe: add RX and TX queue info get
   net/txgbe: add device stats get
   net/txgbe: add device xstats get
   net/txgbe: add queue stats mapping
   net/txgbe: add VLAN handle support
   net/txgbe: add SWFW semaphore and lock
   net/txgbe: add PF module init and uninit for SRIOV
   net/txgbe: add process mailbox operation
   net/txgbe: add PF module configure for SRIOV
   net/txgbe: add VMDq configure
   net/txgbe: add RSS support
   net/txgbe: add DCB support
   net/txgbe: add flow control support
   net/txgbe: add FC auto negotiation support
   net/txgbe: add priority flow control support
   net/txgbe: add device promiscuous and allmulticast mode
   net/txgbe: add MTU set operation
   net/txgbe: add FW version get operation
   net/txgbe: add EEPROM info get operation
   net/txgbe: add register dump support
   net/txgbe: support device LED on and off
   net/txgbe: add mirror rule operations
   net/txgbe: add PTP support
   net/txgbe: add DCB info get operation
   net/txgbe: add Rx and Tx descriptor status



Hi Jiawen,

Before going into more detailed reviews, the patchset conflicts with some recent 
changes in the main repo [1], can you please rebase on top of the latest head of 
the repo?


Also DPDK syntax/style check scripts are giving errors, can you please fix them 
too? You should run following to check:

./devtools/checkpatches.sh -n56
./devtools/check-git-log.sh -n56
(This one needs codespell package to show spelling errors)



[1] mainly the list is:
1) PMD close behavior change,
   - .dev_close changes
   - RTE_ETH_DEV_CLOSE_REMOVE flag removed

2) Some dev_ops moved to ethdev struct
   - .rx_queue_count
   - .rx_descriptor_done
   - .rx_descriptor_status
   - .tx_descriptor_status



Re: [dpdk-dev] [PATCH v2 39/56] net/txgbe: add process mailbox operation

2020-10-06 Thread Ferruh Yigit

On 10/5/2020 1:08 PM, Jiawen Wu wrote:

From: jiawenwu 

Add check operation for vf function level reset, mailbox messages and ack from 
vf. Waiting to process the messages.

Signed-off-by: jiawenwu 


<...>


--- /dev/null
+++ b/drivers/net/txgbe/rte_pmd_txgbe.c
@@ -0,0 +1,11 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2015-2020
+ */
+
+#include 
+
+#include "base/txgbe.h"
+#include "txgbe_ethdev.h"
+#include "rte_pmd_txgbe.h"
+


Why 'rte_pmd_txgbe.c' added at all? As far as I can see there is no PMD specific 
API, so no need to have this file.


Re: [dpdk-dev] [PATCH v1] crypto/aesni_mb: fix incorrect clearing of security session

2020-10-06 Thread De Lara Guarch, Pablo
Hi David,


> -Original Message-
> From: Coyle, David 
> Sent: Tuesday, October 6, 2020 10:42 AM
> To: De Lara Guarch, Pablo ; Doherty, Declan
> 
> Cc: dev@dpdk.org; akhil.go...@nxp.com; Coyle, David
> 
> Subject: [PATCH v1] crypto/aesni_mb: fix incorrect clearing of security 
> session
> 
> When destroying a security session, the AESNI-MB PMD attempted to clear the
> private aesni_mb session object to remove any key material. However, the
> function aesni_mb_pmd_sec_sess_destroy() cleared the security session object
> instead of the private session object.
> 
> This patch fixes this issue by now clearing the private session object.
> 
> Fixes: fda5216fba55 ("crypto/aesni_mb: support DOCSIS protocol")
> 
> Signed-off-by: David Coyle 

Patch looks good, but you need to CC stable, as this should be backported since 
the issue was introduced in the previous release.
So, add Cc: sta...@dpdk.org after Fixes: fda... and send a v2.
Apart from that, you can keep my ack:

Acked-by: Pablo de Lara 


Re: [dpdk-dev] [PATCH v4] kernel/linux: remove igb_uio

2020-10-06 Thread Ferruh Yigit

On 10/6/2020 11:54 AM, Thomas Monjalon wrote:

06/10/2020 12:16, Ferruh Yigit:

On 10/5/2020 10:42 AM, Thomas Monjalon wrote:

As decided in the Technical Board in November 2019,
the kernel module igb_uio is moved to the dpdk-kmods repository
in the /linux/igb_uio/ directory.


[...]

Patch doesn't apply cleanly, not sure why, CI seems having same problem:
http://mails.dpdk.org/archives/test-report/2020-October/156378.html


That's because I used --irreversible-delete to minimize the size of the patch.

[...]

  If UEFI secure boot is enabled, the Linux kernel may disallow the use of
  UIO on the system. Therefore, devices for use by DPDK should be bound to 
the
-   ``vfio-pci`` kernel module rather than ``igb_uio`` or ``uio_pci_generic``.
+   ``vfio-pci`` kernel module rather than any UIO-based module.
  For more details see :ref:`linux_gsg_binding_kernel` below.


Should we change the order of the kernel drivers, currently first section is
"UIO" and second section is "VFIO". Since we encoruage vfio and move igb_uio out
of the repo, should we list the "VfIO" first?


Absolutely.
In general we need to encourage more VFIO and remove some references
to igb_uio. But I don't want to do such "marketing" change in this patch.
The goal of this patch is just to move the code.



OK.


[...]

+* kernel: The module ``igb_uio`` has been moved to the git repository
+``dpdk-kmods`` in a new directory ``linux/igb_uio``.
+


Doc build giving warning on this:
WARNING: Bullet list ends without a blank line; unexpected unindent.


Probably a merge issue. That's fine on my side.



I can see it is wrong in the patch itself, please double check in your end to be 
sure warning is not hidden in doc build.

I am good as long as there is no warning.


Except from above minor comments looks good to me,
Reviewed-by: Ferruh Yigit 


Do we agree I can merge without any change?



Agree if the above doc warning is gone.


[dpdk-dev] [PATCH 2/6] net/mlx5: optimize shared counter memory

2020-10-06 Thread Suanming Mou
Currently, when a counter is allocated, the counter list entry memory
will not be used until the counter is released and added back to the
counter container free list. In this case, if a counter is allocated
as shared counter, the shared information can be save to the counter
list entry memory. This adjustment engages to save more memory for
the shared counter.

One more thing is that now the shared counter is only available for
single counter as shared counter may be applied to both root and
none root table, but batch counter with offset is not supported in
root table in some old OFED version. The batch counter with offset
is now fully supported in the current OFED. This commit is also
the initialize change for the batch counter with offeset can be
shared in the later change.

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/mlx5.h| 33 +---
 drivers/net/mlx5/mlx5_flow_dv.c| 78 +++---
 drivers/net/mlx5/mlx5_flow_verbs.c | 60 +
 3 files changed, 86 insertions(+), 85 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index 27c8f45..fe6bd88 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -270,6 +270,10 @@ struct mlx5_drop {
 #define MLX5_COUNTERS_PER_POOL 512
 #define MLX5_MAX_PENDING_QUERIES 4
 #define MLX5_CNT_CONTAINER_RESIZE 64
+#define MLX5_CNT_SHARED_OFFSET 0x8000
+#define IS_SHARED_CNT(cnt) (!!((cnt) & MLX5_CNT_SHARED_OFFSET))
+#define IS_BATCH_CNT(cnt) (((cnt) & (MLX5_CNT_SHARED_OFFSET - 1)) >= \
+  MLX5_CNT_BATCH_OFFSET)
 #define CNT_SIZE (sizeof(struct mlx5_flow_counter))
 #define CNTEXT_SIZE (sizeof(struct mlx5_flow_counter_ext))
 #define AGE_SIZE (sizeof(struct mlx5_age_param))
@@ -348,11 +352,21 @@ struct flow_counter_stats {
uint64_t bytes;
 };
 
+/* Shared counters information for counters. */
+struct mlx5_flow_counter_shared {
+   uint32_t ref_cnt; /**< Reference counter. */
+   uint32_t id; /**< User counter ID. */
+};
+
 struct mlx5_flow_counter_pool;
 /* Generic counters information. */
 struct mlx5_flow_counter {
-   TAILQ_ENTRY(mlx5_flow_counter) next;
-   /**< Pointer to the next flow counter structure. */
+   union {
+   TAILQ_ENTRY(mlx5_flow_counter) next;
+   /**< Pointer to the next flow counter structure. */
+   struct mlx5_flow_counter_shared shared_info;
+   /**< Shared counter information. */
+   };
union {
uint64_t hits; /**< Reset value of hits packets. */
struct mlx5_flow_counter_pool *pool; /**< Counter pool. */
@@ -361,22 +375,15 @@ struct mlx5_flow_counter {
void *action; /**< Pointer to the dv action. */
 };
 
-/* Extend counters information for none batch counters. */
+/* Extend counters information for none batch fallback counters. */
 struct mlx5_flow_counter_ext {
-   uint32_t shared:1; /**< Share counter ID with other flow rules. */
-   uint32_t batch: 1;
uint32_t skipped:1; /* This counter is skipped or not. */
-   /**< Whether the counter was allocated by batch command. */
-   uint32_t ref_cnt:29; /**< Reference counter. */
-   uint32_t id; /**< User counter ID. */
-   union {  /**< Holds the counters for the rule. */
 #if defined(HAVE_IBV_DEVICE_COUNTERS_SET_V42)
-   struct ibv_counter_set *cs;
+   struct ibv_counter_set *cs;
 #elif defined(HAVE_IBV_DEVICE_COUNTERS_SET_V45)
-   struct ibv_counters *cs;
+   struct ibv_counters *cs;
 #endif
-   struct mlx5_devx_obj *dcs; /**< Counter Devx object. */
-   };
+   struct mlx5_devx_obj *dcs; /**< Counter Devx object. */
 };
 
 TAILQ_HEAD(mlx5_counters, mlx5_flow_counter);
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 1bd3899..10be990 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -4172,8 +4172,9 @@ struct field_modify_info modify_tcp[] = {
struct mlx5_flow_counter_pool *pool;
uint32_t batch = 0;
 
-   idx--;
-   if (idx >= MLX5_CNT_BATCH_OFFSET) {
+   /* Decrease to original index and clear shared bit. */
+   idx = (idx - 1) & (MLX5_CNT_SHARED_OFFSET - 1);
+   if (IS_BATCH_CNT(idx)) {
idx -= MLX5_CNT_BATCH_OFFSET;
batch = 1;
}
@@ -4408,7 +4409,7 @@ struct field_modify_info modify_tcp[] = {
 
cnt = flow_dv_counter_get_by_idx(dev, counter, &pool);
MLX5_ASSERT(pool);
-   if (counter < MLX5_CNT_BATCH_OFFSET) {
+   if (!IS_BATCH_CNT(counter)) {
cnt_ext = MLX5_CNT_TO_CNT_EXT(pool, cnt);
if (priv->counter_fallback)
return mlx5_devx_cmd_flow_counter_query(cnt_ext->dcs, 0,
@@ -4696,29 +4697,19 @@ struct field_modify_info modify_tcp[] = {
  *   Pointer to the Ethernet device structure.
  * @param[in] id
  *   The shared counter ID 

[dpdk-dev] [PATCH 0/6] net/mlx5: make counter thread safe

2020-10-06 Thread Suanming Mou
The mlx5 PMD is going to support multiple-thread flow operations.
This patch set makes the counter action to be thread safe.

Suanming Mou (6):
  net/mlx5: locate aging pools in the general container
  net/mlx5: optimize shared counter memory
  net/mlx5: remove single counter container
  net/mlx5: synchronize flow counter pool creation
  net/mlx5: make three level table thread safe
  net/mlx5: make shared counters thread safe

 drivers/net/mlx5/linux/mlx5_os.c   |  36 ++-
 drivers/net/mlx5/mlx5.c|  38 ++-
 drivers/net/mlx5/mlx5.h|  81 +++---
 drivers/net/mlx5/mlx5_flow.c   | 266 +-
 drivers/net/mlx5/mlx5_flow.h   |   2 +
 drivers/net/mlx5/mlx5_flow_dv.c| 555 ++---
 drivers/net/mlx5/mlx5_flow_verbs.c |  78 +++---
 drivers/net/mlx5/mlx5_utils.c  | 144 +++---
 drivers/net/mlx5/mlx5_utils.h  |  52 ++--
 9 files changed, 618 insertions(+), 634 deletions(-)

-- 
1.8.3.1



[dpdk-dev] [PATCH 3/6] net/mlx5: remove single counter container

2020-10-06 Thread Suanming Mou
A flow counter which was allocated by a batch API couldn't be assigned
to a flow in the root table (group 0) in old rdma-core version.
Hence, a root table flow counter required PMD mechanism to manage
counters which were allocated singly.

Currently, the batch counters have already been supported in root table
includes a new rdma-core version with MLX5_FLOW_ACTION_COUNTER_OFFSET
enum and with a kernel driver includes
MLX5_IB_ATTR_CREATE_FLOW_ARR_COUNTERS_DEVX_OFFSET enum.

When the PMD uses rdma-core API to assign a batch counter to a root
table flow using invalid counter offset, it should get an error only
if the batch counter assignment for root table is supported.
Using this trial in the initialization time can help to detect the
support.

Using the above trial, if the support is valid, remove the management of
single counter container in the fast counter mechanism. Otherwise, move
the counter mechanism to fallback mode.

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/linux/mlx5_os.c   |  36 +++-
 drivers/net/mlx5/mlx5.c|  39 ++---
 drivers/net/mlx5/mlx5.h|  27 +--
 drivers/net/mlx5/mlx5_flow.c   | 152 ++---
 drivers/net/mlx5/mlx5_flow.h   |   2 +
 drivers/net/mlx5/mlx5_flow_dv.c| 337 +
 drivers/net/mlx5/mlx5_flow_verbs.c |  26 +--
 7 files changed, 273 insertions(+), 346 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 188a6d4..43d173b 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -512,6 +512,32 @@
 }
 
 /**
+ * DV flow counter mode detect and config.
+ *
+ * @param dev
+ *   Pointer to rte_eth_dev structure.
+ *
+ */
+static void
+mlx5_flow_counter_mode_config(struct rte_eth_dev *dev)
+{
+   struct mlx5_priv *priv = dev->data->dev_private;
+
+   /* If devx is not supported, counters are not working. */
+   if (!priv->config.devx)
+   return;
+   priv->counter_fallback = 0;
+   if (!priv->config.hca_attr.flow_counters_dump ||
+   (mlx5_flow_discover_counter_offset_support(dev) == -ENOTSUP))
+   priv->counter_fallback = 1;
+#ifndef HAVE_IBV_DEVX_ASYNC
+   priv->counter_fallback = 1;
+#endif
+   if (priv->counter_fallback)
+   DRV_LOG(INFO, "Use fall-back DV counter management");
+}
+
+/**
  * Spawn an Ethernet device from Verbs information.
  *
  * @param dpdk_dev
@@ -979,19 +1005,11 @@
DRV_LOG(INFO, "Rx CQE padding is enabled");
}
if (config->devx) {
-   priv->counter_fallback = 0;
err = mlx5_devx_cmd_query_hca_attr(sh->ctx, &config->hca_attr);
if (err) {
err = -err;
goto error;
}
-   if (!config->hca_attr.flow_counters_dump)
-   priv->counter_fallback = 1;
-#ifndef HAVE_IBV_DEVX_ASYNC
-   priv->counter_fallback = 1;
-#endif
-   if (priv->counter_fallback)
-   DRV_LOG(INFO, "Use fall-back DV counter management");
/* Check for LRO support. */
if (config->dest_tir && config->hca_attr.lro_cap &&
config->dv_flow_en) {
@@ -1364,6 +1382,8 @@
goto error;
}
}
+   if (priv->config.dv_flow_en)
+   mlx5_flow_counter_mode_config(eth_dev);
return eth_dev;
 error:
if (priv) {
diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index 5e3569d..96cebba 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -458,20 +458,18 @@ struct mlx5_flow_id_pool *
 static void
 mlx5_flow_counters_mng_init(struct mlx5_dev_ctx_shared *sh)
 {
-   int i, j;
+   int i;
 
memset(&sh->cmng, 0, sizeof(sh->cmng));
TAILQ_INIT(&sh->cmng.flow_counters);
-   for (i = 0; i < MLX5_CCONT_TYPE_MAX; ++i) {
-   sh->cmng.ccont[i].min_id = MLX5_CNT_BATCH_OFFSET;
-   sh->cmng.ccont[i].max_id = -1;
-   sh->cmng.ccont[i].last_pool_idx = POOL_IDX_INVALID;
-   TAILQ_INIT(&sh->cmng.ccont[i].pool_list);
-   rte_spinlock_init(&sh->cmng.ccont[i].resize_sl);
-   for (j = 0; j < MLX5_COUNTER_TYPE_MAX; j++)
-   TAILQ_INIT(&sh->cmng.ccont[i].counters[j]);
-   rte_spinlock_init(&sh->cmng.ccont[i].csl);
-   }
+   sh->cmng.min_id = MLX5_CNT_BATCH_OFFSET;
+   sh->cmng.max_id = -1;
+   sh->cmng.last_pool_idx = POOL_IDX_INVALID;
+   TAILQ_INIT(&sh->cmng.pool_list);
+   rte_spinlock_init(&sh->cmng.resize_sl);
+   for (i = 0; i < MLX5_COUNTER_TYPE_MAX; i++)
+   TAILQ_INIT(&sh->cmng.counters[i]);
+   rte_spinlock_init(&sh->cmng.csl);
 }
 
 /**
@@ -501,7 +499,6 @@ struct mlx5_flow_id_pool *
 mlx5_flow_counters_mng_close(struct mlx5_dev_ctx_shared *sh)
 {
struct mlx5_counter_stat

[dpdk-dev] [PATCH 5/6] net/mlx5: make three level table thread safe

2020-10-06 Thread Suanming Mou
This commit adds thread safety support in three level table using
spinlock and reference counter for each the table entry.

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/mlx5_utils.c | 144 +++---
 drivers/net/mlx5/mlx5_utils.h |  52 ++-
 2 files changed, 142 insertions(+), 54 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_utils.c b/drivers/net/mlx5/mlx5_utils.c
index fefe833..f3c259d 100644
--- a/drivers/net/mlx5/mlx5_utils.c
+++ b/drivers/net/mlx5/mlx5_utils.c
@@ -551,26 +551,23 @@ struct mlx5_l3t_tbl *
tbl->type = type;
switch (type) {
case MLX5_L3T_TYPE_WORD:
-   l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_word) +
- sizeof(uint16_t) * MLX5_L3T_ET_SIZE;
+   l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_word);
l3t_ip_cfg.type = "mlx5_l3t_e_tbl_w";
break;
case MLX5_L3T_TYPE_DWORD:
-   l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_dword) +
- sizeof(uint32_t) * MLX5_L3T_ET_SIZE;
+   l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_dword);
l3t_ip_cfg.type = "mlx5_l3t_e_tbl_dw";
break;
case MLX5_L3T_TYPE_QWORD:
-   l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_qword) +
- sizeof(uint64_t) * MLX5_L3T_ET_SIZE;
+   l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_qword);
l3t_ip_cfg.type = "mlx5_l3t_e_tbl_qw";
break;
default:
-   l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_ptr) +
- sizeof(void *) * MLX5_L3T_ET_SIZE;
+   l3t_ip_cfg.size = sizeof(struct mlx5_l3t_entry_ptr);
l3t_ip_cfg.type = "mlx5_l3t_e_tbl_tpr";
break;
}
+   rte_spinlock_init(&tbl->sl);
tbl->eip = mlx5_ipool_create(&l3t_ip_cfg);
if (!tbl->eip) {
rte_errno = ENOMEM;
@@ -620,46 +617,63 @@ struct mlx5_l3t_tbl *
mlx5_free(tbl);
 }
 
-uint32_t
+int32_t
 mlx5_l3t_get_entry(struct mlx5_l3t_tbl *tbl, uint32_t idx,
   union mlx5_l3t_data *data)
 {
struct mlx5_l3t_level_tbl *g_tbl, *m_tbl;
+   struct mlx5_l3t_entry_word *w_e_tbl;
+   struct mlx5_l3t_entry_dword *dw_e_tbl;
+   struct mlx5_l3t_entry_qword *qw_e_tbl;
+   struct mlx5_l3t_entry_ptr *ptr_e_tbl;
void *e_tbl;
uint32_t entry_idx;
+   int32_t ret = -1;
 
+   rte_spinlock_lock(&tbl->sl);
g_tbl = tbl->tbl;
if (!g_tbl)
-   return -1;
+   goto out;
m_tbl = g_tbl->tbl[(idx >> MLX5_L3T_GT_OFFSET) & MLX5_L3T_GT_MASK];
if (!m_tbl)
-   return -1;
+   goto out;
e_tbl = m_tbl->tbl[(idx >> MLX5_L3T_MT_OFFSET) & MLX5_L3T_MT_MASK];
if (!e_tbl)
-   return -1;
+   goto out;
+   ret = 0;
entry_idx = idx & MLX5_L3T_ET_MASK;
switch (tbl->type) {
case MLX5_L3T_TYPE_WORD:
-   data->word = ((struct mlx5_l3t_entry_word *)e_tbl)->entry
-[entry_idx];
+   w_e_tbl = (struct mlx5_l3t_entry_word *)e_tbl;
+   data->word = w_e_tbl->entry[entry_idx].data;
+   if (w_e_tbl->entry[entry_idx].data)
+   w_e_tbl->entry[entry_idx].ref_cnt++;
break;
case MLX5_L3T_TYPE_DWORD:
-   data->dword = ((struct mlx5_l3t_entry_dword *)e_tbl)->entry
-[entry_idx];
+   dw_e_tbl = (struct mlx5_l3t_entry_dword *)e_tbl;
+   data->dword = dw_e_tbl->entry[entry_idx].data;
+   if (dw_e_tbl->entry[entry_idx].data)
+   dw_e_tbl->entry[entry_idx].ref_cnt++;
break;
case MLX5_L3T_TYPE_QWORD:
-   data->qword = ((struct mlx5_l3t_entry_qword *)e_tbl)->entry
- [entry_idx];
+   qw_e_tbl = (struct mlx5_l3t_entry_qword *)e_tbl;
+   data->qword = qw_e_tbl->entry[entry_idx].data;
+   if (qw_e_tbl->entry[entry_idx].data)
+   qw_e_tbl->entry[entry_idx].ref_cnt++;
break;
default:
-   data->ptr = ((struct mlx5_l3t_entry_ptr *)e_tbl)->entry
-   [entry_idx];
+   ptr_e_tbl = (struct mlx5_l3t_entry_ptr *)e_tbl;
+   data->ptr = ptr_e_tbl->entry[entry_idx].data;
+   if (ptr_e_tbl->entry[entry_idx].data)
+   ptr_e_tbl->entry[entry_idx].ref_cnt++;
break;
}
-   return 0;
+out:
+   rte_spinlock_unlock(&tbl->sl);
+   return ret;
 }
 
-void
+int32_t
 mlx5_l3t_clear_entry(struct mlx5_l3t_tbl *tbl, uint32_t idx)
 {
struct mlx5_l3t_level_tbl *g_tbl, *m_tbl;
@@ -670,36 +684,54 @@ struct mlx5_l3t_tb

[dpdk-dev] [PATCH 6/6] net/mlx5: make shared counters thread safe

2020-10-06 Thread Suanming Mou
The shared counters save the counter index to three level table. As
three level table supports multiple-thread opertations now, the shared
counters can take advantage of the table to support multiple-thread.

Once multiple threads saves the same ID counter to the same table entry
at the same time, only one will be success, others will get the EEXIST
errno with entry reference counter increase. In this case, the other
duplicate created counter will be released.

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/mlx5.h|  1 -
 drivers/net/mlx5/mlx5_flow_dv.c| 62 +++---
 drivers/net/mlx5/mlx5_flow_verbs.c | 19 +++-
 3 files changed, 21 insertions(+), 61 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index 8c951e2..6e0b2e2 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -348,7 +348,6 @@ struct flow_counter_stats {
 
 /* Shared counters information for counters. */
 struct mlx5_flow_counter_shared {
-   uint32_t ref_cnt; /**< Reference counter. */
uint32_t id; /**< User counter ID. */
 };
 
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 31d7fe4..3adb905 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -4461,28 +4461,6 @@ struct field_modify_info modify_tcp[] = {
 }
 
 /**
- * Search for existed shared counter.
- *
- * @param[in] dev
- *   Pointer to the Ethernet device structure.
- * @param[in] id
- *   The shared counter ID to search.
- *
- * @return
- *   0 if not existed, otherwise shared counter index.
- */
-static uint32_t
-flow_dv_counter_shared_search(struct rte_eth_dev *dev, uint32_t id)
-{
-   struct mlx5_priv *priv = dev->data->dev_private;
-   union mlx5_l3t_data data;
-
-   if (mlx5_l3t_get_entry(priv->sh->cnt_id_tbl, id, &data))
-   return 0;
-   return data.dword;
-}
-
-/**
  * Allocate a flow counter.
  *
  * @param[in] dev
@@ -4510,24 +4488,15 @@ struct field_modify_info modify_tcp[] = {
enum mlx5_counter_type cnt_type =
age ? MLX5_COUNTER_TYPE_AGE : MLX5_COUNTER_TYPE_ORIGIN;
uint32_t cnt_idx;
+   union mlx5_l3t_data data;
 
if (!priv->config.devx) {
rte_errno = ENOTSUP;
return 0;
}
-   if (shared) {
-   cnt_idx = flow_dv_counter_shared_search(dev, id);
-   if (cnt_idx) {
-   cnt_free = flow_dv_counter_get_by_idx(dev, cnt_idx,
- NULL);
-   if (cnt_free->shared_info.ref_cnt + 1 == 0) {
-   rte_errno = E2BIG;
-   return 0;
-   }
-   cnt_free->shared_info.ref_cnt++;
-   return cnt_idx;
-   }
-   }
+   if (shared && !mlx5_l3t_get_entry(priv->sh->cnt_id_tbl, id, &data) &&
+   data.dword)
+   return data.dword;
/* Get free counters from container. */
rte_spinlock_lock(&cmng->csl);
cnt_free = TAILQ_FIRST(&cmng->counters[cnt_type]);
@@ -4566,12 +4535,18 @@ struct field_modify_info modify_tcp[] = {
 &cnt_free->bytes))
goto err;
if (shared) {
-   union mlx5_l3t_data data;
-
data.dword = cnt_idx;
-   if (mlx5_l3t_set_entry(priv->sh->cnt_id_tbl, id, &data))
+   if (mlx5_l3t_set_entry(priv->sh->cnt_id_tbl, id, &data)) {
+   if (rte_errno == EEXIST) {
+   cnt_free->pool = pool;
+   rte_spinlock_lock(&cmng->csl);
+   TAILQ_INSERT_TAIL(&cmng->counters[cnt_type],
+ cnt_free, next);
+   rte_spinlock_unlock(&cmng->csl);
+   return data.dword;
+   }
goto err;
-   cnt_free->shared_info.ref_cnt = 1;
+   }
cnt_free->shared_info.id = id;
cnt_idx |= MLX5_CNT_SHARED_OFFSET;
}
@@ -4667,12 +4642,9 @@ struct field_modify_info modify_tcp[] = {
return;
cnt = flow_dv_counter_get_by_idx(dev, counter, &pool);
MLX5_ASSERT(pool);
-   if (IS_SHARED_CNT(counter)) {
-   if (--cnt->shared_info.ref_cnt)
-   return;
-   mlx5_l3t_clear_entry(priv->sh->cnt_id_tbl,
-cnt->shared_info.id);
-   }
+   if (IS_SHARED_CNT(counter) &&
+   mlx5_l3t_clear_entry(priv->sh->cnt_id_tbl, cnt->shared_info.id))
+   return;
if (IS_AGE_POOL(pool))
flow_dv_counter_remove_from_age(dev, counter, cnt);
cnt->pool = pool;
diff --git a/drivers/net/mlx5/mlx5_flow_ver

[dpdk-dev] [PATCH 1/6] net/mlx5: locate aging pools in the general container

2020-10-06 Thread Suanming Mou
Commit [1] introduced different container for the aging counter
pools. In order to save container memory the aging counter pools
can be located in the general pool container.

This patch locates the aging counter pools in the general pool
container. Remove the aging container management.

[1] commit fd143711a6ea ("net/mlx5: separate aging counter pool range")

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/mlx5.c|  7 ++--
 drivers/net/mlx5/mlx5.h| 17 +
 drivers/net/mlx5/mlx5_flow.c   | 19 +++---
 drivers/net/mlx5/mlx5_flow_dv.c| 78 ++
 drivers/net/mlx5/mlx5_flow_verbs.c |  4 +-
 5 files changed, 57 insertions(+), 68 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index 01ead6e..5e3569d 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -458,7 +458,7 @@ struct mlx5_flow_id_pool *
 static void
 mlx5_flow_counters_mng_init(struct mlx5_dev_ctx_shared *sh)
 {
-   int i;
+   int i, j;
 
memset(&sh->cmng, 0, sizeof(sh->cmng));
TAILQ_INIT(&sh->cmng.flow_counters);
@@ -468,7 +468,8 @@ struct mlx5_flow_id_pool *
sh->cmng.ccont[i].last_pool_idx = POOL_IDX_INVALID;
TAILQ_INIT(&sh->cmng.ccont[i].pool_list);
rte_spinlock_init(&sh->cmng.ccont[i].resize_sl);
-   TAILQ_INIT(&sh->cmng.ccont[i].counters);
+   for (j = 0; j < MLX5_COUNTER_TYPE_MAX; j++)
+   TAILQ_INIT(&sh->cmng.ccont[i].counters[j]);
rte_spinlock_init(&sh->cmng.ccont[i].csl);
}
 }
@@ -513,7 +514,7 @@ struct mlx5_flow_id_pool *
}
for (i = 0; i < MLX5_CCONT_TYPE_MAX; ++i) {
struct mlx5_flow_counter_pool *pool;
-   uint32_t batch = !!(i > 1);
+   uint32_t batch = (i == MLX5_CCONT_TYPE_BATCH);
 
if (!sh->cmng.ccont[i].pools)
continue;
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index bd91e16..27c8f45 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -270,7 +270,6 @@ struct mlx5_drop {
 #define MLX5_COUNTERS_PER_POOL 512
 #define MLX5_MAX_PENDING_QUERIES 4
 #define MLX5_CNT_CONTAINER_RESIZE 64
-#define MLX5_CNT_AGE_OFFSET 0x8000
 #define CNT_SIZE (sizeof(struct mlx5_flow_counter))
 #define CNTEXT_SIZE (sizeof(struct mlx5_flow_counter_ext))
 #define AGE_SIZE (sizeof(struct mlx5_age_param))
@@ -279,7 +278,6 @@ struct mlx5_drop {
 #define CNT_POOL_TYPE_AGE  (1 << 1)
 #define IS_EXT_POOL(pool) (((pool)->type) & CNT_POOL_TYPE_EXT)
 #define IS_AGE_POOL(pool) (((pool)->type) & CNT_POOL_TYPE_AGE)
-#define MLX_CNT_IS_AGE(counter) ((counter) & MLX5_CNT_AGE_OFFSET ? 1 : 0)
 #define MLX5_CNT_LEN(pool) \
(CNT_SIZE + \
(IS_AGE_POOL(pool) ? AGE_SIZE : 0) + \
@@ -322,17 +320,20 @@ enum {
AGE_TMOUT, /* Timeout, wait for rte_flow_get_aged_flows and destroy. */
 };
 
-#define MLX5_CNT_CONTAINER(sh, batch, age) (&(sh)->cmng.ccont \
-   [(batch) * 2 + (age)])
+#define MLX5_CNT_CONTAINER(sh, batch) (&(sh)->cmng.ccont[batch])
 
 enum {
MLX5_CCONT_TYPE_SINGLE,
-   MLX5_CCONT_TYPE_SINGLE_FOR_AGE,
MLX5_CCONT_TYPE_BATCH,
-   MLX5_CCONT_TYPE_BATCH_FOR_AGE,
MLX5_CCONT_TYPE_MAX,
 };
 
+enum mlx5_counter_type {
+   MLX5_COUNTER_TYPE_ORIGIN,
+   MLX5_COUNTER_TYPE_AGE,
+   MLX5_COUNTER_TYPE_MAX,
+};
+
 /* Counter age parameter. */
 struct mlx5_age_param {
rte_atomic16_t state; /**< Age state. */
@@ -427,7 +428,8 @@ struct mlx5_pools_container {
int max_id; /* The maximum counter ID in the pools. */
rte_spinlock_t resize_sl; /* The resize lock. */
rte_spinlock_t csl; /* The counter free list lock. */
-   struct mlx5_counters counters; /* Free counter list. */
+   struct mlx5_counters counters[MLX5_COUNTER_TYPE_MAX];
+   /* Free counter list. */
struct mlx5_counter_pools pool_list; /* Counter pool list. */
struct mlx5_flow_counter_pool **pools; /* Counter pool array. */
struct mlx5_counter_stats_mem_mng *mem_mng;
@@ -441,7 +443,6 @@ struct mlx5_flow_counter_mng {
uint8_t pending_queries;
uint8_t batch;
uint16_t pool_index;
-   uint8_t age;
uint8_t query_thread_on;
LIST_HEAD(mem_mngs, mlx5_counter_stats_mem_mng) mem_mngs;
LIST_HEAD(stat_raws, mlx5_counter_stats_raw) free_stat_raws;
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index ffa7646..db7fc8f 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -5940,7 +5940,6 @@ struct mlx5_meter_domains_infos *
uint16_t offset;
int ret;
uint8_t batch = sh->cmng.batch;
-   uint8_t age = sh->cmng.age;
uint16_t pool_index = sh->cmng.pool_index;
struct mlx5_pools_container *cont;
struct mlx5_flow_counter_pool *pool;
@@ -5949,7

[dpdk-dev] [PATCH v2] crypto/aesni_mb: fix incorrect clearing of security session

2020-10-06 Thread David Coyle
When destroying a security session, the AESNI-MB PMD attempted to clear
the private aesni_mb session object to remove any key material. However,
the function aesni_mb_pmd_sec_sess_destroy() cleared the security session
object instead of the private session object.

This patch fixes this issue by now clearing the private session object.

Fixes: fda5216fba55 ("crypto/aesni_mb: support DOCSIS protocol")
Cc: sta...@dpdk.org

Signed-off-by: David Coyle 
Acked-by: Pablo de Lara 
---
v2:
* CC'ing sta...@dpdk.org

 drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c 
b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
index 2362f0c3c..b11d7f12b 100644
--- a/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
+++ b/drivers/crypto/aesni_mb/rte_aesni_mb_pmd_ops.c
@@ -911,7 +911,7 @@ aesni_mb_pmd_sec_sess_destroy(void *dev __rte_unused,
 
if (sess_priv) {
struct rte_mempool *sess_mp = rte_mempool_from_obj(sess_priv);
-   memset(sess, 0, sizeof(struct aesni_mb_session));
+   memset(sess_priv, 0, sizeof(struct aesni_mb_session));
set_sec_session_private_data(sess, NULL);
rte_mempool_put(sess_mp, sess_priv);
}
-- 
2.17.1



[dpdk-dev] [PATCH 4/6] net/mlx5: synchronize flow counter pool creation

2020-10-06 Thread Suanming Mou
Currently, counter operations are not thread safe as the counter
pools' array resize is not protected.

This commit protects the container pools' array resize using a spinlock.
The original counter pool statistic memory allocate is moved to the
host thread in order to minimize the critical section. Since that pool
statistic memory is required only in query time. The container pools'
array should be resized by the user threads, the new pool may be used
by other rte_flow APIs before the host thread resize is done, if the
pool is not saved to the pools' array, the specified counter memory will
not be found as the pool is not saved to the counter management pool
array. The pool raw statistic memory will be filled in host thread.

The shared counters will be protected in other commit.

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/mlx5.c|  12 ++-
 drivers/net/mlx5/mlx5.h|  11 ++-
 drivers/net/mlx5/mlx5_flow.c   | 127 ++--
 drivers/net/mlx5/mlx5_flow_dv.c| 146 ++---
 drivers/net/mlx5/mlx5_flow_verbs.c |   5 +-
 5 files changed, 160 insertions(+), 141 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index 96cebba..79c5563 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -465,8 +465,7 @@ struct mlx5_flow_id_pool *
sh->cmng.min_id = MLX5_CNT_BATCH_OFFSET;
sh->cmng.max_id = -1;
sh->cmng.last_pool_idx = POOL_IDX_INVALID;
-   TAILQ_INIT(&sh->cmng.pool_list);
-   rte_spinlock_init(&sh->cmng.resize_sl);
+   rte_spinlock_init(&sh->cmng.pool_update_sl);
for (i = 0; i < MLX5_COUNTER_TYPE_MAX; i++)
TAILQ_INIT(&sh->cmng.counters[i]);
rte_spinlock_init(&sh->cmng.csl);
@@ -499,7 +498,7 @@ struct mlx5_flow_id_pool *
 mlx5_flow_counters_mng_close(struct mlx5_dev_ctx_shared *sh)
 {
struct mlx5_counter_stats_mem_mng *mng;
-   int j;
+   int i, j;
int retries = 1024;
 
rte_errno = 0;
@@ -512,9 +511,10 @@ struct mlx5_flow_id_pool *
 
if (sh->cmng.pools) {
struct mlx5_flow_counter_pool *pool;
+   int n_valid = sh->cmng.n_valid;
 
-   pool = TAILQ_FIRST(&sh->cmng.pool_list);
-   while (pool) {
+   for (i = 0; i < n_valid; ++i) {
+   pool = sh->cmng.pools[i];
if (!IS_EXT_POOL(pool) && pool->min_dcs)
claim_zero(mlx5_devx_cmd_destroy
   (pool->min_dcs));
@@ -530,9 +530,7 @@ struct mlx5_flow_id_pool *
   (MLX5_GET_POOL_CNT_EXT
(pool, j)->dcs));
}
-   TAILQ_REMOVE(&sh->cmng.pool_list, pool, next);
mlx5_free(pool);
-   pool = TAILQ_FIRST(&sh->cmng.pool_list);
}
mlx5_free(sh->cmng.pools);
}
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index a3d4ad9..8c951e2 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -395,7 +395,11 @@ struct mlx5_flow_counter_pool {
volatile uint32_t query_gen:1; /* Query round. */
rte_spinlock_t sl; /* The pool lock. */
struct mlx5_counter_stats_raw *raw;
-   struct mlx5_counter_stats_raw *raw_hw; /* The raw on HW working. */
+   union {
+   struct rte_eth_dev *dev; /* The counter pool create device. */
+   struct mlx5_counter_stats_raw *raw_hw;
+   /* The raw on HW working. */
+   };
 };
 
 struct mlx5_counter_stats_raw;
@@ -419,16 +423,15 @@ struct mlx5_counter_stats_raw {
 
 /* Counter global management structure. */
 struct mlx5_flow_counter_mng {
-   rte_atomic16_t n_valid; /* Number of valid pools. */
+   volatile uint16_t n_valid; /* Number of valid pools. */
uint16_t n; /* Number of pools. */
uint16_t last_pool_idx; /* Last used pool index */
int min_id; /* The minimum counter ID in the pools. */
int max_id; /* The maximum counter ID in the pools. */
-   rte_spinlock_t resize_sl; /* The resize lock. */
+   rte_spinlock_t pool_update_sl; /* The pool update lock. */
rte_spinlock_t csl; /* The counter free list lock. */
struct mlx5_counters counters[MLX5_COUNTER_TYPE_MAX];
/* Free counter list. */
-   struct mlx5_counter_pools pool_list; /* Counter pool list. */
struct mlx5_flow_counter_pool **pools; /* Counter pool array. */
struct mlx5_counter_stats_mem_mng *mem_mng;
/* Hold the memory management for the next allocated pools raws. */
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index c280f56..a9664b8 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -13,6 +13,7 @@
 #include 
 #in

Re: [dpdk-dev] [PATCH v1] crypto/aesni_mb: fix incorrect clearing of security session

2020-10-06 Thread Coyle, David
Hi Pablo

> >
> > When destroying a security session, the AESNI-MB PMD attempted to
> > clear the private aesni_mb session object to remove any key material.
> > However, the function aesni_mb_pmd_sec_sess_destroy() cleared the
> > security session object instead of the private session object.
> >
> > This patch fixes this issue by now clearing the private session object.
> >
> > Fixes: fda5216fba55 ("crypto/aesni_mb: support DOCSIS protocol")
> >
> > Signed-off-by: David Coyle 
> 
> Patch looks good, but you need to CC stable, as this should be backported
> since the issue was introduced in the previous release.
> So, add Cc: sta...@dpdk.org after Fixes: fda... and send a v2.
> Apart from that, you can keep my ack:

[DC] Done, thanks for pointing that out

> 
> Acked-by: Pablo de Lara 


[dpdk-dev] [PATCH 00/25] net/mlx5: support multiple-thread flow operations

2020-10-06 Thread Suanming Mou
This patch set contains multiple-thread flow operations support
for the flow objects.

Suanming Mou (10):
  net/mlx5: use thread safe index pool for flow objects
  net/mlx5: fix redundant Direct Verbs resources allocate
  net/mlx5: make header reformat action thread safe
  net/mlx5: remove unused hash list operations
  net/mlx5: make Rx queue thread safe
  net/mlx5: create global jump action
  net/mlx5: create global default miss action
  net/mlx5: create global drop action
  net/mlx5: make meter action thread safe
  net/mlx5: make VLAN network interface thread safe

Xueming Li (15):
  net/mlx5: use thread specific flow context
  net/mlx5: reuse flow Id as hairpin Id
  net/mlx5: indexed pool supports zero size entry
  net/mlx5: use indexed pool for RSS flow ID
  net/mlx5: make rte flow list thread safe
  net/mlx5: support concurrent access for hash list
  net/mlx5: make flow table cache thread safe
  net/mlx5: make flow tag list thread safe
  net/mlx5: make flow modify action list thread safe
  net/mlx5: make metadata copy flow list thread safe
  net/mlx5: introduce thread safe linked list cache
  net/mlx5: make matcher list thread safe
  net/mlx5: make port ID action cache thread safe
  net/mlx5: make push VLAN action cache thread safe
  net/mlx5: remove shared context lock

 drivers/net/mlx5/linux/mlx5_os.c  |   95 ++-
 drivers/net/mlx5/linux/mlx5_vlan_os.c |   24 +-
 drivers/net/mlx5/mlx5.c   |  268 +-
 drivers/net/mlx5/mlx5.h   |   54 +-
 drivers/net/mlx5/mlx5_flow.c  |  500 ++--
 drivers/net/mlx5/mlx5_flow.h  |  117 ++-
 drivers/net/mlx5/mlx5_flow_dv.c   | 1450 ++---
 drivers/net/mlx5/mlx5_flow_meter.c|   72 +-
 drivers/net/mlx5/mlx5_flow_verbs.c|   78 +-
 drivers/net/mlx5/mlx5_rxq.c   |  234 +++---
 drivers/net/mlx5/mlx5_rxtx.h  |   20 +-
 drivers/net/mlx5/mlx5_utils.c |  350 ++--
 drivers/net/mlx5/mlx5_utils.h |  302 +--
 13 files changed, 1837 insertions(+), 1727 deletions(-)

-- 
1.8.3.1



[dpdk-dev] [PATCH 02/25] net/mlx5: use thread specific flow context

2020-10-06 Thread Suanming Mou
From: Xueming Li 

As part of multi-thread flow support, this patch moves flow intermediate
data to thread specific, makes them a flow context. The context is
allocated per thread, destroyed along with thread life-cycle.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/linux/mlx5_os.c   |   5 --
 drivers/net/mlx5/mlx5.c|   2 -
 drivers/net/mlx5/mlx5.h|   6 --
 drivers/net/mlx5/mlx5_flow.c   | 134 +
 drivers/net/mlx5/mlx5_flow.h   |  15 -
 drivers/net/mlx5/mlx5_flow_dv.c|  26 ---
 drivers/net/mlx5/mlx5_flow_verbs.c |  24 ---
 7 files changed, 133 insertions(+), 79 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 188a6d4..4276964 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -1330,11 +1330,6 @@
err = ENOTSUP;
goto error;
}
-   /*
-* Allocate the buffer for flow creating, just once.
-* The allocation must be done before any flow creating.
-*/
-   mlx5_flow_alloc_intermediate(eth_dev);
/* Query availability of metadata reg_c's. */
err = mlx5_flow_discover_mreg_c(eth_dev);
if (err < 0) {
diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index c9fc085..16719e6 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -1375,8 +1375,6 @@ struct mlx5_dev_ctx_shared *
 */
mlx5_flow_list_flush(dev, &priv->flows, true);
mlx5_flow_meter_flush(dev, NULL);
-   /* Free the intermediate buffers for flow creation. */
-   mlx5_flow_free_intermediate(dev);
/* Prevent crashes when queues are still in use. */
dev->rx_pkt_burst = removed_rx_burst;
dev->tx_pkt_burst = removed_tx_burst;
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index bd91e16..0080ac8 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -793,10 +793,6 @@ struct mlx5_priv {
struct mlx5_drop drop_queue; /* Flow drop queues. */
uint32_t flows; /* RTE Flow rules. */
uint32_t ctrl_flows; /* Control flow rules. */
-   void *inter_flows; /* Intermediate resources for flow creation. */
-   void *rss_desc; /* Intermediate rss description resources. */
-   int flow_idx; /* Intermediate device flow index. */
-   int flow_nested_idx; /* Intermediate device flow index, nested. */
struct mlx5_obj_ops obj_ops; /* HW objects operations. */
LIST_HEAD(rxq, mlx5_rxq_ctrl) rxqsctrl; /* DPDK Rx queues. */
LIST_HEAD(rxqobj, mlx5_rxq_obj) rxqsobj; /* Verbs/DevX Rx queues. */
@@ -1020,8 +1016,6 @@ int mlx5_dev_filter_ctrl(struct rte_eth_dev *dev,
 void mlx5_flow_stop(struct rte_eth_dev *dev, uint32_t *list);
 int mlx5_flow_start_default(struct rte_eth_dev *dev);
 void mlx5_flow_stop_default(struct rte_eth_dev *dev);
-void mlx5_flow_alloc_intermediate(struct rte_eth_dev *dev);
-void mlx5_flow_free_intermediate(struct rte_eth_dev *dev);
 int mlx5_flow_verify(struct rte_eth_dev *dev);
 int mlx5_ctrl_flow_source_queue(struct rte_eth_dev *dev, uint32_t queue);
 int mlx5_ctrl_flow_vlan(struct rte_eth_dev *dev,
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index ffa7646..546 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -306,6 +306,13 @@ struct mlx5_flow_tunnel_info {
},
 };
 
+/* Key of thread specific flow workspace data. */
+static pthread_key_t key_workspace;
+
+/* Thread specific flow workspace data once initialization data. */
+static pthread_once_t key_workspace_init;
+
+
 /**
  * Translate tag ID to register.
  *
@@ -4348,16 +4355,18 @@ struct mlx5_flow_tunnel_info {
uint8_t buffer[2048];
} items_tx;
struct rte_flow_expand_rss *buf = &expand_buffer.buf;
-   struct mlx5_flow_rss_desc *rss_desc = &((struct mlx5_flow_rss_desc *)
- priv->rss_desc)[!!priv->flow_idx];
+   struct mlx5_flow_rss_desc *rss_desc;
const struct rte_flow_action *p_actions_rx = actions;
uint32_t i;
uint32_t idx = 0;
int hairpin_flow;
uint32_t hairpin_id = 0;
struct rte_flow_attr attr_tx = { .priority = 0 };
+   struct mlx5_flow_workspace *wks = mlx5_flow_get_thread_workspace();
int ret;
 
+   MLX5_ASSERT(wks);
+   rss_desc = &wks->rss_desc[!!wks->flow_idx];
hairpin_flow = flow_check_hairpin_split(dev, attr, actions);
ret = flow_drv_validate(dev, attr, items, p_actions_rx,
external, hairpin_flow, error);
@@ -4383,9 +4392,25 @@ struct mlx5_flow_tunnel_info {
flow->hairpin_flow_id = hairpin_id;
MLX5_ASSERT(flow->drv_type > MLX5_FLOW_TYPE_MIN &&
flow->drv_type < MLX5_FLOW_TYPE_MAX);
-   memset(rss_desc, 0, sizeof(*rss_desc));
+   memset(rs

[dpdk-dev] [PATCH 01/25] net/mlx5: use thread safe index pool for flow objects

2020-10-06 Thread Suanming Mou
As mlx5 PMD is changed to be thread safe, all the flow-related
sub-objects inside the PMD should be thread safe. This commit
changes the index memory pools' lock configuration to be enabled.
That makes the index pool be thread safe.

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/mlx5.c | 18 +-
 1 file changed, 9 insertions(+), 9 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index 01ead6e..c9fc085 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -191,7 +191,7 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.trunk_size = 64,
.grow_trunk = 3,
.grow_shift = 2,
-   .need_lock = 0,
+   .need_lock = 1,
.release_mem_en = 1,
.malloc = mlx5_malloc,
.free = mlx5_free,
@@ -202,7 +202,7 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.trunk_size = 64,
.grow_trunk = 3,
.grow_shift = 2,
-   .need_lock = 0,
+   .need_lock = 1,
.release_mem_en = 1,
.malloc = mlx5_malloc,
.free = mlx5_free,
@@ -213,7 +213,7 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.trunk_size = 64,
.grow_trunk = 3,
.grow_shift = 2,
-   .need_lock = 0,
+   .need_lock = 1,
.release_mem_en = 1,
.malloc = mlx5_malloc,
.free = mlx5_free,
@@ -224,7 +224,7 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.trunk_size = 64,
.grow_trunk = 3,
.grow_shift = 2,
-   .need_lock = 0,
+   .need_lock = 1,
.release_mem_en = 1,
.malloc = mlx5_malloc,
.free = mlx5_free,
@@ -235,7 +235,7 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.trunk_size = 64,
.grow_trunk = 3,
.grow_shift = 2,
-   .need_lock = 0,
+   .need_lock = 1,
.release_mem_en = 1,
.malloc = mlx5_malloc,
.free = mlx5_free,
@@ -247,7 +247,7 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.trunk_size = 64,
.grow_trunk = 3,
.grow_shift = 2,
-   .need_lock = 0,
+   .need_lock = 1,
.release_mem_en = 1,
.malloc = mlx5_malloc,
.free = mlx5_free,
@@ -258,7 +258,7 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.trunk_size = 64,
.grow_trunk = 3,
.grow_shift = 2,
-   .need_lock = 0,
+   .need_lock = 1,
.release_mem_en = 1,
.malloc = mlx5_malloc,
.free = mlx5_free,
@@ -269,7 +269,7 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.trunk_size = 64,
.grow_trunk = 3,
.grow_shift = 2,
-   .need_lock = 0,
+   .need_lock = 1,
.release_mem_en = 1,
.malloc = mlx5_malloc,
.free = mlx5_free,
@@ -284,7 +284,7 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.trunk_size = 64,
.grow_trunk = 3,
.grow_shift = 2,
-   .need_lock = 0,
+   .need_lock = 1,
.release_mem_en = 1,
.malloc = mlx5_malloc,
.free = mlx5_free,
-- 
1.8.3.1



[dpdk-dev] [PATCH 03/25] net/mlx5: reuse flow Id as hairpin Id

2020-10-06 Thread Suanming Mou
From: Xueming Li 

Hairpin flow matching required a unique flow ID for matching.
This patch reuses flow ID as hairpin flow ID, this will save some code
to generate a separate hairpin ID, also saves flow memory by removing
hairpin ID.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/mlx5.c  | 11 ---
 drivers/net/mlx5/mlx5.h  |  1 -
 drivers/net/mlx5/mlx5_flow.c | 32 ++--
 drivers/net/mlx5/mlx5_flow.h |  5 +
 4 files changed, 11 insertions(+), 38 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index 16719e6..6c5c04d 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -952,13 +952,6 @@ struct mlx5_dev_ctx_shared *
MLX5_ASSERT(sh->devx_rx_uar);
MLX5_ASSERT(mlx5_os_get_devx_uar_base_addr(sh->devx_rx_uar));
}
-   sh->flow_id_pool = mlx5_flow_id_pool_alloc
-   ((1 << HAIRPIN_FLOW_ID_BITS) - 1);
-   if (!sh->flow_id_pool) {
-   DRV_LOG(ERR, "can't create flow id pool");
-   err = ENOMEM;
-   goto error;
-   }
 #ifndef RTE_ARCH_64
/* Initialize UAR access locks for 32bit implementations. */
rte_spinlock_init(&sh->uar_lock_cq);
@@ -1020,8 +1013,6 @@ struct mlx5_dev_ctx_shared *
claim_zero(mlx5_glue->dealloc_pd(sh->pd));
if (sh->ctx)
claim_zero(mlx5_glue->close_device(sh->ctx));
-   if (sh->flow_id_pool)
-   mlx5_flow_id_pool_release(sh->flow_id_pool);
mlx5_free(sh);
MLX5_ASSERT(err > 0);
rte_errno = err;
@@ -1092,8 +1083,6 @@ struct mlx5_dev_ctx_shared *
mlx5_glue->devx_free_uar(sh->devx_rx_uar);
if (sh->ctx)
claim_zero(mlx5_glue->close_device(sh->ctx));
-   if (sh->flow_id_pool)
-   mlx5_flow_id_pool_release(sh->flow_id_pool);
pthread_mutex_destroy(&sh->txpp.mutex);
mlx5_free(sh);
return;
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index 0080ac8..a3ec994 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -652,7 +652,6 @@ struct mlx5_dev_ctx_shared {
void *devx_comp; /* DEVX async comp obj. */
struct mlx5_devx_obj *tis; /* TIS object. */
struct mlx5_devx_obj *td; /* Transport domain. */
-   struct mlx5_flow_id_pool *flow_id_pool; /* Flow ID pool. */
void *tx_uar; /* Tx/packet pacing shared UAR. */
struct mlx5_flex_parser_profiles fp[MLX5_FLEX_PARSER_MAX];
/* Flex parser profiles information. */
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 546..f0a6a57 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -3426,9 +3426,8 @@ struct mlx5_flow_tunnel_info {
   struct rte_flow_action actions_rx[],
   struct rte_flow_action actions_tx[],
   struct rte_flow_item pattern_tx[],
-  uint32_t *flow_id)
+  uint32_t flow_id)
 {
-   struct mlx5_priv *priv = dev->data->dev_private;
const struct rte_flow_action_raw_encap *raw_encap;
const struct rte_flow_action_raw_decap *raw_decap;
struct mlx5_rte_flow_action_set_tag *set_tag;
@@ -3438,7 +3437,6 @@ struct mlx5_flow_tunnel_info {
char *addr;
int encap = 0;
 
-   mlx5_flow_id_get(priv->sh->flow_id_pool, flow_id);
for (; actions->type != RTE_FLOW_ACTION_TYPE_END; actions++) {
switch (actions->type) {
case RTE_FLOW_ACTION_TYPE_VXLAN_ENCAP:
@@ -3507,7 +3505,7 @@ struct mlx5_flow_tunnel_info {
set_tag = (void *)actions_rx;
set_tag->id = mlx5_flow_get_reg_id(dev, MLX5_HAIRPIN_RX, 0, NULL);
MLX5_ASSERT(set_tag->id > REG_NON);
-   set_tag->data = *flow_id;
+   set_tag->data = flow_id;
tag_action->conf = set_tag;
/* Create Tx item list. */
rte_memcpy(actions_tx, actions, sizeof(struct rte_flow_action));
@@ -3516,7 +3514,7 @@ struct mlx5_flow_tunnel_info {
item->type = (enum rte_flow_item_type)
 MLX5_RTE_FLOW_ITEM_TYPE_TAG;
tag_item = (void *)addr;
-   tag_item->data = *flow_id;
+   tag_item->data = flow_id;
tag_item->id = mlx5_flow_get_reg_id(dev, MLX5_HAIRPIN_TX, 0, NULL);
MLX5_ASSERT(set_tag->id > REG_NON);
item->spec = tag_item;
@@ -4360,7 +4358,6 @@ struct mlx5_flow_tunnel_info {
uint32_t i;
uint32_t idx = 0;
int hairpin_flow;
-   uint32_t hairpin_id = 0;
struct rte_flow_attr attr_tx = { .priority = 0 };
struct mlx5_flow_workspace *wks = mlx5_flow_get_thread_workspace();
int ret;
@@ -4372,24 +4369,22 @@ struct mlx5_flow_tunnel_info {
external, hairpin_flow, error);
if (ret < 0)
return 0;
+   flow = mlx5_ipool_zmalloc(priv->sh->ipool[MLX5_IPOOL_RTE_FLOW

[dpdk-dev] [PATCH 04/25] net/mlx5: indexed pool supports zero size entry

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To make indexed pool to be used as ID generator, this patch allows entry
size to be zero.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/mlx5_utils.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_utils.c b/drivers/net/mlx5/mlx5_utils.c
index fefe833..0a75fa6 100644
--- a/drivers/net/mlx5/mlx5_utils.c
+++ b/drivers/net/mlx5/mlx5_utils.c
@@ -230,7 +230,7 @@ struct mlx5_indexed_pool *
struct mlx5_indexed_pool *pool;
uint32_t i;
 
-   if (!cfg || !cfg->size || (!cfg->malloc ^ !cfg->free) ||
+   if (!cfg || (!cfg->malloc ^ !cfg->free) ||
(cfg->trunk_size && ((cfg->trunk_size & (cfg->trunk_size - 1)) ||
((__builtin_ffs(cfg->trunk_size) + TRUNK_IDX_BITS) > 32
return NULL;
@@ -391,7 +391,7 @@ struct mlx5_indexed_pool *
 {
void *entry = mlx5_ipool_malloc(pool, idx);
 
-   if (entry)
+   if (entry && pool->cfg.size)
memset(entry, 0, pool->cfg.size);
return entry;
 }
-- 
1.8.3.1



[dpdk-dev] [PATCH 07/25] net/mlx5: support concurrent access for hash list

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To support multi-thread flow operations, this patch adds following to
support concurrent access to hash list:
1. list level read/write lock
2. entry reference count
3. entry create/match/remove callback
4. remove insert/lookup/remove function which are not thread safe
5. add register/unregister function to support entry reuse

For better performance, lookup function uses read lock to
allow concurrent lookup from different thread, all other hash list
modification functions uses write lock which blocks concurrent
modification from other thread.

The exact objects change will be applied in the next patches.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/linux/mlx5_os.c |  27 ---
 drivers/net/mlx5/mlx5.c  |  13 ++--
 drivers/net/mlx5/mlx5_flow.c |   7 +-
 drivers/net/mlx5/mlx5_flow_dv.c  |   6 +-
 drivers/net/mlx5/mlx5_utils.c| 154 ---
 drivers/net/mlx5/mlx5_utils.h| 142 +---
 6 files changed, 272 insertions(+), 77 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 94c1e38..13b5a3f 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -237,14 +237,16 @@
return err;
/* Create tags hash list table. */
snprintf(s, sizeof(s), "%s_tags", sh->ibdev_name);
-   sh->tag_table = mlx5_hlist_create(s, MLX5_TAGS_HLIST_ARRAY_SIZE);
+   sh->tag_table = mlx5_hlist_create(s, MLX5_TAGS_HLIST_ARRAY_SIZE, 0,
+ false, NULL, NULL, NULL);
if (!sh->tag_table) {
DRV_LOG(ERR, "tags with hash creation failed.");
err = ENOMEM;
goto error;
}
snprintf(s, sizeof(s), "%s_hdr_modify", sh->ibdev_name);
-   sh->modify_cmds = mlx5_hlist_create(s, MLX5_FLOW_HDR_MODIFY_HTABLE_SZ);
+   sh->modify_cmds = mlx5_hlist_create(s, MLX5_FLOW_HDR_MODIFY_HTABLE_SZ,
+   0, false, NULL, NULL, NULL);
if (!sh->modify_cmds) {
DRV_LOG(ERR, "hdr modify hash creation failed");
err = ENOMEM;
@@ -252,7 +254,8 @@
}
snprintf(s, sizeof(s), "%s_encaps_decaps", sh->ibdev_name);
sh->encaps_decaps = mlx5_hlist_create(s,
- MLX5_FLOW_ENCAP_DECAP_HTABLE_SZ);
+ MLX5_FLOW_ENCAP_DECAP_HTABLE_SZ,
+ 0, false, NULL, NULL, NULL);
if (!sh->encaps_decaps) {
DRV_LOG(ERR, "encap decap hash creation failed");
err = ENOMEM;
@@ -332,16 +335,16 @@
sh->pop_vlan_action = NULL;
}
if (sh->encaps_decaps) {
-   mlx5_hlist_destroy(sh->encaps_decaps, NULL, NULL);
+   mlx5_hlist_destroy(sh->encaps_decaps);
sh->encaps_decaps = NULL;
}
if (sh->modify_cmds) {
-   mlx5_hlist_destroy(sh->modify_cmds, NULL, NULL);
+   mlx5_hlist_destroy(sh->modify_cmds);
sh->modify_cmds = NULL;
}
if (sh->tag_table) {
/* tags should be destroyed with flow before. */
-   mlx5_hlist_destroy(sh->tag_table, NULL, NULL);
+   mlx5_hlist_destroy(sh->tag_table);
sh->tag_table = NULL;
}
mlx5_free_table_hash_list(priv);
@@ -393,16 +396,16 @@
pthread_mutex_destroy(&sh->dv_mutex);
 #endif /* HAVE_MLX5DV_DR */
if (sh->encaps_decaps) {
-   mlx5_hlist_destroy(sh->encaps_decaps, NULL, NULL);
+   mlx5_hlist_destroy(sh->encaps_decaps);
sh->encaps_decaps = NULL;
}
if (sh->modify_cmds) {
-   mlx5_hlist_destroy(sh->modify_cmds, NULL, NULL);
+   mlx5_hlist_destroy(sh->modify_cmds);
sh->modify_cmds = NULL;
}
if (sh->tag_table) {
/* tags should be destroyed with flow before. */
-   mlx5_hlist_destroy(sh->tag_table, NULL, NULL);
+   mlx5_hlist_destroy(sh->tag_table);
sh->tag_table = NULL;
}
mlx5_free_table_hash_list(priv);
@@ -1343,7 +1346,9 @@
mlx5_flow_ext_mreg_supported(eth_dev) &&
priv->sh->dv_regc0_mask) {
priv->mreg_cp_tbl = mlx5_hlist_create(MLX5_FLOW_MREG_HNAME,
- MLX5_FLOW_MREG_HTABLE_SZ);
+ MLX5_FLOW_MREG_HTABLE_SZ,
+ 0, false,
+ NULL, NULL, NULL);
if (!priv->mreg_cp_tbl) {
err = ENOMEM;
goto error;
@@ -1353,7 +1358,7 @@
 error:
if (priv) {
if (priv->mreg_cp_tbl)
-

[dpdk-dev] [PATCH 06/25] net/mlx5: make rte flow list thread safe

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To support multi-thread flow operations, this patch introduces list lock
for the rte_flow list manages all the rte_flow handlers.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/linux/mlx5_os.c |  1 +
 drivers/net/mlx5/mlx5.h  |  1 +
 drivers/net/mlx5/mlx5_flow.c | 10 --
 3 files changed, 10 insertions(+), 2 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index bfd5276..94c1e38 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -1250,6 +1250,7 @@
  MLX5_MAX_MAC_ADDRESSES);
priv->flows = 0;
priv->ctrl_flows = 0;
+   rte_spinlock_init(&priv->flow_list_lock);
TAILQ_INIT(&priv->flow_meters);
TAILQ_INIT(&priv->flow_meter_profiles);
/* Hint libmlx5 to use PMD allocator for data plane resources */
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index 0532577..464d2cf 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -784,6 +784,7 @@ struct mlx5_priv {
struct mlx5_drop drop_queue; /* Flow drop queues. */
uint32_t flows; /* RTE Flow rules. */
uint32_t ctrl_flows; /* Control flow rules. */
+   rte_spinlock_t flow_list_lock;
struct mlx5_obj_ops obj_ops; /* HW objects operations. */
LIST_HEAD(rxq, mlx5_rxq_ctrl) rxqsctrl; /* DPDK Rx queues. */
LIST_HEAD(rxqobj, mlx5_rxq_obj) rxqsobj; /* Verbs/DevX Rx queues. */
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index c1fbc80..2790c32 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -4479,9 +4479,12 @@ struct mlx5_flow_tunnel_info {
if (ret < 0)
goto error;
}
-   if (list)
+   if (list) {
+   rte_spinlock_lock(&priv->flow_list_lock);
ILIST_INSERT(priv->sh->ipool[MLX5_IPOOL_RTE_FLOW], list, idx,
 flow, next);
+   rte_spinlock_unlock(&priv->flow_list_lock);
+   }
flow_rxq_flags_set(dev, flow);
/* Nested flow creation index recovery. */
wks->flow_idx = wks->flow_nested_idx;
@@ -4637,9 +4640,12 @@ struct rte_flow *
if (dev->data->dev_started)
flow_rxq_flags_trim(dev, flow);
flow_drv_destroy(dev, flow);
-   if (list)
+   if (list) {
+   rte_spinlock_lock(&priv->flow_list_lock);
ILIST_REMOVE(priv->sh->ipool[MLX5_IPOOL_RTE_FLOW], list,
 flow_idx, flow, next);
+   rte_spinlock_unlock(&priv->flow_list_lock);
+   }
flow_mreg_del_copy_action(dev, flow);
if (flow->fdir) {
LIST_FOREACH(priv_fdir_flow, &priv->fdir_flows, next) {
-- 
1.8.3.1



[dpdk-dev] [PATCH 05/25] net/mlx5: use indexed pool for RSS flow ID

2020-10-06 Thread Suanming Mou
From: Xueming Li 

The flow ID generation API used an integer pool to save released ID, The
only usage is to generate RSS flow ID. To support multiple flow, it has
to be enhanced to be thread safe.

Indexed pool could be used to generate unique ID by setting size of pool
entry to zero. Since bitmap is used, an extra benefits is saving memory
to about one bit per entry. Further more indexed pool could be thread
safe by enabling lock.

This patch leverages indexed pool to generate RSS flow ID, removes
unused flow ID generating API.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/linux/mlx5_os.c |  13 
 drivers/net/mlx5/mlx5.c  | 125 ++-
 drivers/net/mlx5/mlx5.h  |  11 +---
 drivers/net/mlx5/mlx5_flow.c |  47 +--
 drivers/net/mlx5/mlx5_flow.h |   5 --
 5 files changed, 21 insertions(+), 180 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 4276964..bfd5276 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -1287,17 +1287,6 @@
err = mlx5_alloc_shared_dr(priv);
if (err)
goto error;
-   /*
-* RSS id is shared with meter flow id. Meter flow id can only
-* use the 24 MSB of the register.
-*/
-   priv->qrss_id_pool = mlx5_flow_id_pool_alloc(UINT32_MAX >>
-MLX5_MTR_COLOR_BITS);
-   if (!priv->qrss_id_pool) {
-   DRV_LOG(ERR, "can't create flow id pool");
-   err = ENOMEM;
-   goto error;
-   }
}
if (config->devx && config->dv_flow_en && config->dest_tir) {
priv->obj_ops = devx_obj_ops;
@@ -1372,8 +1361,6 @@
close(priv->nl_socket_rdma);
if (priv->vmwa_context)
mlx5_vlan_vmwa_exit(priv->vmwa_context);
-   if (priv->qrss_id_pool)
-   mlx5_flow_id_pool_release(priv->qrss_id_pool);
if (own_domain_id)
claim_zero(rte_eth_switch_domain_free(priv->domain_id));
mlx5_free(priv);
diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index 6c5c04d..b3d1638 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -299,6 +299,11 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list =
.free = mlx5_free,
.type = "rte_flow_ipool",
},
+   {
+   .size = 0,
+   .need_lock = 1,
+   .type = "mlx5_flow_rss_id_ipool",
+   },
 };
 
 
@@ -307,126 +312,6 @@ static LIST_HEAD(, mlx5_dev_ctx_shared) mlx5_dev_ctx_list 
=
 
 #define MLX5_FLOW_TABLE_HLIST_ARRAY_SIZE 4096
 
-/**
- * Allocate ID pool structure.
- *
- * @param[in] max_id
- *   The maximum id can be allocated from the pool.
- *
- * @return
- *   Pointer to pool object, NULL value otherwise.
- */
-struct mlx5_flow_id_pool *
-mlx5_flow_id_pool_alloc(uint32_t max_id)
-{
-   struct mlx5_flow_id_pool *pool;
-   void *mem;
-
-   pool = mlx5_malloc(MLX5_MEM_ZERO, sizeof(*pool),
-  RTE_CACHE_LINE_SIZE, SOCKET_ID_ANY);
-   if (!pool) {
-   DRV_LOG(ERR, "can't allocate id pool");
-   rte_errno  = ENOMEM;
-   return NULL;
-   }
-   mem = mlx5_malloc(MLX5_MEM_ZERO,
- MLX5_FLOW_MIN_ID_POOL_SIZE * sizeof(uint32_t),
- RTE_CACHE_LINE_SIZE, SOCKET_ID_ANY);
-   if (!mem) {
-   DRV_LOG(ERR, "can't allocate mem for id pool");
-   rte_errno  = ENOMEM;
-   goto error;
-   }
-   pool->free_arr = mem;
-   pool->curr = pool->free_arr;
-   pool->last = pool->free_arr + MLX5_FLOW_MIN_ID_POOL_SIZE;
-   pool->base_index = 0;
-   pool->max_id = max_id;
-   return pool;
-error:
-   mlx5_free(pool);
-   return NULL;
-}
-
-/**
- * Release ID pool structure.
- *
- * @param[in] pool
- *   Pointer to flow id pool object to free.
- */
-void
-mlx5_flow_id_pool_release(struct mlx5_flow_id_pool *pool)
-{
-   mlx5_free(pool->free_arr);
-   mlx5_free(pool);
-}
-
-/**
- * Generate ID.
- *
- * @param[in] pool
- *   Pointer to flow id pool.
- * @param[out] id
- *   The generated ID.
- *
- * @return
- *   0 on success, error value otherwise.
- */
-uint32_t
-mlx5_flow_id_get(struct mlx5_flow_id_pool *pool, uint32_t *id)
-{
-   if (pool->curr == pool->free_arr) {
-   if (pool->base_index == pool->max_id) {
-   rte_errno  = ENOMEM;
-   DRV_LOG(ERR, "no free id");
-   return -rte_errno;
-   }
-   *id = ++pool->base_index;
-   return 0;
-   }
-   *id = *(--pool->curr);
-   return 0;
-}
-
-/**
- * Release I

[dpdk-dev] [PATCH 08/25] net/mlx5: make flow table cache thread safe

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To support multi-thread flow insertion/removal, this patch uses thread
safe hash list API for flow table cache hash list.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/mlx5.c | 102 
 drivers/net/mlx5/mlx5.h |   2 +-
 drivers/net/mlx5/mlx5_flow.h|  17 +
 drivers/net/mlx5/mlx5_flow_dv.c | 147 
 4 files changed, 105 insertions(+), 163 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index ddf236a..61e5e69 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -976,7 +976,7 @@ struct mlx5_dev_ctx_shared *
 }
 
 /**
- * Destroy table hash list and all the root entries per domain.
+ * Destroy table hash list.
  *
  * @param[in] priv
  *   Pointer to the private device data structure.
@@ -985,46 +985,9 @@ struct mlx5_dev_ctx_shared *
 mlx5_free_table_hash_list(struct mlx5_priv *priv)
 {
struct mlx5_dev_ctx_shared *sh = priv->sh;
-   struct mlx5_flow_tbl_data_entry *tbl_data;
-   union mlx5_flow_tbl_key table_key = {
-   {
-   .table_id = 0,
-   .reserved = 0,
-   .domain = 0,
-   .direction = 0,
-   }
-   };
-   struct mlx5_hlist_entry *pos;
 
if (!sh->flow_tbls)
return;
-   pos = mlx5_hlist_lookup(sh->flow_tbls, table_key.v64, NULL);
-   if (pos) {
-   tbl_data = container_of(pos, struct mlx5_flow_tbl_data_entry,
-   entry);
-   MLX5_ASSERT(tbl_data);
-   mlx5_hlist_remove(sh->flow_tbls, pos);
-   mlx5_free(tbl_data);
-   }
-   table_key.direction = 1;
-   pos = mlx5_hlist_lookup(sh->flow_tbls, table_key.v64, NULL);
-   if (pos) {
-   tbl_data = container_of(pos, struct mlx5_flow_tbl_data_entry,
-   entry);
-   MLX5_ASSERT(tbl_data);
-   mlx5_hlist_remove(sh->flow_tbls, pos);
-   mlx5_free(tbl_data);
-   }
-   table_key.direction = 0;
-   table_key.domain = 1;
-   pos = mlx5_hlist_lookup(sh->flow_tbls, table_key.v64, NULL);
-   if (pos) {
-   tbl_data = container_of(pos, struct mlx5_flow_tbl_data_entry,
-   entry);
-   MLX5_ASSERT(tbl_data);
-   mlx5_hlist_remove(sh->flow_tbls, pos);
-   mlx5_free(tbl_data);
-   }
mlx5_hlist_destroy(sh->flow_tbls);
 }
 
@@ -1039,80 +1002,45 @@ struct mlx5_dev_ctx_shared *
  *   Zero on success, positive error code otherwise.
  */
 int
-mlx5_alloc_table_hash_list(struct mlx5_priv *priv)
+mlx5_alloc_table_hash_list(struct mlx5_priv *priv __rte_unused)
 {
+   int err = 0;
+   /* Tables are only used in DV and DR modes. */
+#ifdef HAVE_IBV_FLOW_DV_SUPPORT
struct mlx5_dev_ctx_shared *sh = priv->sh;
char s[MLX5_HLIST_NAMESIZE];
-   int err = 0;
 
MLX5_ASSERT(sh);
snprintf(s, sizeof(s), "%s_flow_table", priv->sh->ibdev_name);
sh->flow_tbls = mlx5_hlist_create(s, MLX5_FLOW_TABLE_HLIST_ARRAY_SIZE,
- 0, false, NULL, NULL, NULL);
+   0, false, flow_dv_tbl_create_cb, NULL,
+   flow_dv_tbl_remove_cb);
if (!sh->flow_tbls) {
DRV_LOG(ERR, "flow tables with hash creation failed.");
err = ENOMEM;
return err;
}
+   sh->flow_tbls->ctx = sh;
 #ifndef HAVE_MLX5DV_DR
+   struct rte_flow_error error;
+   struct rte_eth_dev *dev = &rte_eth_devices[priv->dev_data->port_id];
+
/*
 * In case we have not DR support, the zero tables should be created
 * because DV expect to see them even if they cannot be created by
 * RDMA-CORE.
 */
-   union mlx5_flow_tbl_key table_key = {
-   {
-   .table_id = 0,
-   .reserved = 0,
-   .domain = 0,
-   .direction = 0,
-   }
-   };
-   struct mlx5_flow_tbl_data_entry *tbl_data = mlx5_malloc(MLX5_MEM_ZERO,
- sizeof(*tbl_data), 0,
- SOCKET_ID_ANY);
-
-   if (!tbl_data) {
-   err = ENOMEM;
-   goto error;
-   }
-   tbl_data->entry.key = table_key.v64;
-   err = mlx5_hlist_insert(sh->flow_tbls, &tbl_data->entry);
-   if (err)
-   goto error;
-   rte_atomic32_init(&tbl_data->tbl.refcnt);
-   rte_atomic32_inc(&tbl_data->tbl.refcnt);
-   table_key.direction = 1;
-   tbl_data = mlx5_malloc(MLX5_MEM_ZERO, sizeof(*tbl_data), 0,
-  SOCKET_ID_ANY);
-   if (!tbl_data) {
+   if (!flow_dv_tbl_resource_get(dev

[dpdk-dev] [PATCH 09/25] net/mlx5: fix redundant Direct Verbs resources allocate

2020-10-06 Thread Suanming Mou
All table, tag, header modify, header reformat are supported only on DV
mode. For the OFED version doesn't support these, create the related
redundant DV resources waste the memory.

Add the code section in the HAVE_IBV_FLOW_DV_SUPPORT macro to avoid the
redundant resources allocation.

Fixes: 2eb4d0107acc ("net/mlx5: refactor PCI probing on Linux")
Cc: sta...@dpdk.org
Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/linux/mlx5_os.c | 5 -
 1 file changed, 4 insertions(+), 1 deletion(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 13b5a3f..d828035 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -234,7 +234,9 @@
DRV_LOG(DEBUG, "sh->flow_tbls[%p] already created, reuse\n",
(void *)sh->flow_tbls);
if (err)
-   return err;
+   goto error;
+   /* The resources below are only valid with DV support. */
+#ifdef HAVE_IBV_FLOW_DV_SUPPORT
/* Create tags hash list table. */
snprintf(s, sizeof(s), "%s_tags", sh->ibdev_name);
sh->tag_table = mlx5_hlist_create(s, MLX5_TAGS_HLIST_ARRAY_SIZE, 0,
@@ -261,6 +263,7 @@
err = ENOMEM;
goto error;
}
+#endif
 #ifdef HAVE_MLX5DV_DR
void *domain;
 
-- 
1.8.3.1



[dpdk-dev] [PATCH 10/25] net/mlx5: make flow tag list thread safe

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To support multi-thread flow insertion, this pathc updates flow tag list
to use thread safe hash list.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/linux/mlx5_os.c |  6 ++-
 drivers/net/mlx5/mlx5_flow.h |  5 +++
 drivers/net/mlx5/mlx5_flow_dv.c  | 97 +++-
 3 files changed, 56 insertions(+), 52 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index d828035..39bd16b 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -225,7 +225,7 @@
 mlx5_alloc_shared_dr(struct mlx5_priv *priv)
 {
struct mlx5_dev_ctx_shared *sh = priv->sh;
-   char s[MLX5_HLIST_NAMESIZE];
+   char s[MLX5_HLIST_NAMESIZE] __rte_unused;
int err = 0;
 
if (!sh->flow_tbls)
@@ -240,12 +240,14 @@
/* Create tags hash list table. */
snprintf(s, sizeof(s), "%s_tags", sh->ibdev_name);
sh->tag_table = mlx5_hlist_create(s, MLX5_TAGS_HLIST_ARRAY_SIZE, 0,
- false, NULL, NULL, NULL);
+ true, flow_dv_tag_create_cb, NULL,
+ flow_dv_tag_remove_cb);
if (!sh->tag_table) {
DRV_LOG(ERR, "tags with hash creation failed.");
err = ENOMEM;
goto error;
}
+   sh->tag_table->ctx = sh;
snprintf(s, sizeof(s), "%s_hdr_modify", sh->ibdev_name);
sh->modify_cmds = mlx5_hlist_create(s, MLX5_FLOW_HDR_MODIFY_HTABLE_SZ,
0, false, NULL, NULL, NULL);
diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h
index f661d1e..c92a40b 100644
--- a/drivers/net/mlx5/mlx5_flow.h
+++ b/drivers/net/mlx5/mlx5_flow.h
@@ -1091,4 +1091,9 @@ struct mlx5_flow_tbl_resource 
*flow_dv_tbl_resource_get(struct rte_eth_dev *dev,
uint32_t table_id, uint8_t egress, uint8_t transfer,
uint8_t dummy, struct rte_flow_error *error);
 
+struct mlx5_hlist_entry *flow_dv_tag_create_cb(struct mlx5_hlist *list,
+  uint64_t key, void *cb_ctx);
+void flow_dv_tag_remove_cb(struct mlx5_hlist *list,
+  struct mlx5_hlist_entry *entry);
+
 #endif /* RTE_PMD_MLX5_FLOW_H_ */
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index fa19873..23ec983 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -7825,6 +7825,35 @@ struct mlx5_flow_tbl_resource *
return 0;
 }
 
+struct mlx5_hlist_entry *
+flow_dv_tag_create_cb(struct mlx5_hlist *list, uint64_t key, void *ctx)
+{
+   struct mlx5_dev_ctx_shared *sh = list->ctx;
+   struct rte_flow_error *error = ctx;
+   struct mlx5_flow_dv_tag_resource *entry;
+   uint32_t idx = 0;
+   int ret;
+
+   entry = mlx5_ipool_zmalloc(sh->ipool[MLX5_IPOOL_TAG], &idx);
+   if (!entry) {
+   rte_flow_error_set(error, ENOMEM,
+  RTE_FLOW_ERROR_TYPE_UNSPECIFIED, NULL,
+  "cannot allocate resource memory");
+   return NULL;
+   }
+   entry->idx = idx;
+   ret = mlx5_flow_os_create_flow_action_tag(key,
+ &entry->action);
+   if (ret) {
+   mlx5_ipool_free(sh->ipool[MLX5_IPOOL_TAG], idx);
+   rte_flow_error_set(error, ENOMEM,
+  RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
+  NULL, "cannot create action");
+   return NULL;
+   }
+   return &entry->entry;
+}
+
 /**
  * Find existing tag resource or create and register a new one.
  *
@@ -7848,54 +7877,32 @@ struct mlx5_flow_tbl_resource *
 struct rte_flow_error *error)
 {
struct mlx5_priv *priv = dev->data->dev_private;
-   struct mlx5_dev_ctx_shared *sh = priv->sh;
struct mlx5_flow_dv_tag_resource *cache_resource;
struct mlx5_hlist_entry *entry;
-   int ret;
 
-   /* Lookup a matching resource from cache. */
-   entry = mlx5_hlist_lookup(sh->tag_table, (uint64_t)tag_be24, NULL);
+   entry = mlx5_hlist_register(priv->sh->tag_table, tag_be24, error);
if (entry) {
cache_resource = container_of
(entry, struct mlx5_flow_dv_tag_resource, entry);
-   rte_atomic32_inc(&cache_resource->refcnt);
dev_flow->handle->dvh.rix_tag = cache_resource->idx;
dev_flow->dv.tag_resource = cache_resource;
-   DRV_LOG(DEBUG, "cached tag resource %p: refcnt now %d++",
-   (void *)cache_resource,
-   rte_atomic32_read(&cache_resource->refcnt));
return 0;
}
-   /* Register new resource. */
-   cache_resource = mlx5_ipool_zmalloc(sh->ipool[MLX5_IPOOL_

[dpdk-dev] [PATCH 13/25] net/mlx5: make header reformat action thread safe

2020-10-06 Thread Suanming Mou
This commit applies the thread safe hash list to the header reformat
action. That makes the theader reformat action to be thread safe.

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/linux/mlx5_os.c |   6 +-
 drivers/net/mlx5/mlx5_flow.h |   6 ++
 drivers/net/mlx5/mlx5_flow_dv.c  | 181 +--
 3 files changed, 111 insertions(+), 82 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 50d3d99..24cf348 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -262,12 +262,16 @@
snprintf(s, sizeof(s), "%s_encaps_decaps", sh->ibdev_name);
sh->encaps_decaps = mlx5_hlist_create(s,
  MLX5_FLOW_ENCAP_DECAP_HTABLE_SZ,
- 0, false, NULL, NULL, NULL);
+ 0, true,
+ flow_dv_encap_decap_create_cb,
+ flow_dv_encap_decap_match_cb,
+ flow_dv_encap_decap_remove_cb);
if (!sh->encaps_decaps) {
DRV_LOG(ERR, "encap decap hash creation failed");
err = ENOMEM;
goto error;
}
+   sh->encaps_decaps->ctx = sh;
 #endif
 #ifdef HAVE_MLX5DV_DR
void *domain;
diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h
index 41969c2..1fe0b30 100644
--- a/drivers/net/mlx5/mlx5_flow.h
+++ b/drivers/net/mlx5/mlx5_flow.h
@@ -1105,4 +1105,10 @@ struct mlx5_hlist_entry *flow_dv_mreg_create_cb(struct 
mlx5_hlist *list,
 void flow_dv_mreg_remove_cb(struct mlx5_hlist *list,
struct mlx5_hlist_entry *entry);
 
+int flow_dv_encap_decap_match_cb(struct mlx5_hlist *list,
+struct mlx5_hlist_entry *entry, void *cb_ctx);
+struct mlx5_hlist_entry *flow_dv_encap_decap_create_cb(struct mlx5_hlist *list,
+   uint64_t key, void *cb_ctx);
+void flow_dv_encap_decap_remove_cb(struct mlx5_hlist *list,
+  struct mlx5_hlist_entry *entry);
 #endif /* RTE_PMD_MLX5_FLOW_H_ */
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index d19f697..b884d8c 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -2548,21 +2548,24 @@ struct field_modify_info modify_tcp[] = {
 /**
  * Match encap_decap resource.
  *
+ * @param list
+ *   Pointer to the hash list.
  * @param entry
  *   Pointer to exist resource entry object.
- * @param ctx
+ * @param ctx_cb
  *   Pointer to new encap_decap resource.
  *
  * @return
- *   0 on matching, -1 otherwise.
+ *   0 on matching, none-zero otherwise.
  */
-static int
-flow_dv_encap_decap_resource_match(struct mlx5_hlist_entry *entry, void *ctx)
+int
+flow_dv_encap_decap_match_cb(struct mlx5_hlist *list __rte_unused,
+struct mlx5_hlist_entry *entry, void *cb_ctx)
 {
-   struct mlx5_flow_dv_encap_decap_resource *resource;
+   struct mlx5_flow_cb_ctx *ctx = cb_ctx;
+   struct mlx5_flow_dv_encap_decap_resource *resource = ctx->data;
struct mlx5_flow_dv_encap_decap_resource *cache_resource;
 
-   resource = (struct mlx5_flow_dv_encap_decap_resource *)ctx;
cache_resource = container_of(entry,
  struct mlx5_flow_dv_encap_decap_resource,
  entry);
@@ -2579,6 +2582,63 @@ struct field_modify_info modify_tcp[] = {
 }
 
 /**
+ * Allocate encap_decap resource.
+ *
+ * @param list
+ *   Pointer to the hash list.
+ * @param entry
+ *   Pointer to exist resource entry object.
+ * @param ctx_cb
+ *   Pointer to new encap_decap resource.
+ *
+ * @return
+ *   0 on matching, none-zero otherwise.
+ */
+struct mlx5_hlist_entry *
+flow_dv_encap_decap_create_cb(struct mlx5_hlist *list,
+ uint64_t key __rte_unused,
+ void *cb_ctx)
+{
+   struct mlx5_dev_ctx_shared *sh = list->ctx;
+   struct mlx5_flow_cb_ctx *ctx = cb_ctx;
+   struct mlx5dv_dr_domain *domain;
+   struct mlx5_flow_dv_encap_decap_resource *resource = ctx->data;
+   struct mlx5_flow_dv_encap_decap_resource *cache_resource;
+   uint32_t idx;
+   int ret;
+
+   if (resource->ft_type == MLX5DV_FLOW_TABLE_TYPE_FDB)
+   domain = sh->fdb_domain;
+   else if (resource->ft_type == MLX5DV_FLOW_TABLE_TYPE_NIC_RX)
+   domain = sh->rx_domain;
+   else
+   domain = sh->tx_domain;
+   /* Register new encap/decap resource. */
+   cache_resource = mlx5_ipool_zmalloc(sh->ipool[MLX5_IPOOL_DECAP_ENCAP],
+  &idx);
+   if (!cache_resource) {
+   rte_flow_error_set(ctx->error, ENOMEM,
+  RTE_FLOW_ERROR_TYPE_UNSPECIFIED, 

[dpdk-dev] [PATCH 11/25] net/mlx5: make flow modify action list thread safe

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To support multi-thread flow insertion, this patch updates flow modify
action list to use thread safe hash list.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/linux/mlx5_os.c |   5 +-
 drivers/net/mlx5/mlx5_flow.h |  13 ++-
 drivers/net/mlx5/mlx5_flow_dv.c  | 191 +--
 3 files changed, 96 insertions(+), 113 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 39bd16b..744b1dd 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -250,12 +250,15 @@
sh->tag_table->ctx = sh;
snprintf(s, sizeof(s), "%s_hdr_modify", sh->ibdev_name);
sh->modify_cmds = mlx5_hlist_create(s, MLX5_FLOW_HDR_MODIFY_HTABLE_SZ,
-   0, false, NULL, NULL, NULL);
+   0, true, flow_dv_modify_create_cb,
+   flow_dv_modify_match_cb,
+   flow_dv_modify_remove_cb);
if (!sh->modify_cmds) {
DRV_LOG(ERR, "hdr modify hash creation failed");
err = ENOMEM;
goto error;
}
+   sh->modify_cmds->ctx = sh;
snprintf(s, sizeof(s), "%s_encaps_decaps", sh->ibdev_name);
sh->encaps_decaps = mlx5_hlist_create(s,
  MLX5_FLOW_ENCAP_DECAP_HTABLE_SZ,
diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h
index c92a40b..7effacc 100644
--- a/drivers/net/mlx5/mlx5_flow.h
+++ b/drivers/net/mlx5/mlx5_flow.h
@@ -451,10 +451,8 @@ struct mlx5_flow_dv_tag_resource {
 /* Modify resource structure */
 struct mlx5_flow_dv_modify_hdr_resource {
struct mlx5_hlist_entry entry;
-   /* Pointer to next element. */
-   rte_atomic32_t refcnt; /**< Reference counter. */
-   void *action;
-   /**< Modify header action object. */
+   void *action; /**< Modify header action object. */
+   /* Key area for hash list matching: */
uint8_t ft_type; /**< Flow table type, Rx or Tx. */
uint32_t actions_num; /**< Number of modification actions. */
uint64_t flags; /**< Flags for RDMA API. */
@@ -1096,4 +1094,11 @@ struct mlx5_hlist_entry *flow_dv_tag_create_cb(struct 
mlx5_hlist *list,
 void flow_dv_tag_remove_cb(struct mlx5_hlist *list,
   struct mlx5_hlist_entry *entry);
 
+int flow_dv_modify_match_cb(struct mlx5_hlist *list,
+   struct mlx5_hlist_entry *entry, void *cb_ctx);
+struct mlx5_hlist_entry *flow_dv_modify_create_cb(struct mlx5_hlist *list,
+ uint64_t key, void *ctx);
+void flow_dv_modify_remove_cb(struct mlx5_hlist *list,
+ struct mlx5_hlist_entry *entry);
+
 #endif /* RTE_PMD_MLX5_FLOW_H_ */
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 23ec983..d19f697 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -4010,35 +4010,72 @@ struct field_modify_info modify_tcp[] = {
 /**
  * Match modify-header resource.
  *
+ * @param list
+ *   Pointer to the hash list.
  * @param entry
  *   Pointer to exist resource entry object.
  * @param ctx
  *   Pointer to new modify-header resource.
  *
  * @return
- *   0 on matching, -1 otherwise.
+ *   0 on matching, non-zero otherwise.
  */
-static int
-flow_dv_modify_hdr_resource_match(struct mlx5_hlist_entry *entry, void *ctx)
+int
+flow_dv_modify_match_cb(struct mlx5_hlist *list __rte_unused,
+   struct mlx5_hlist_entry *entry, void *cb_ctx)
 {
-   struct mlx5_flow_dv_modify_hdr_resource *resource;
-   struct mlx5_flow_dv_modify_hdr_resource *cache_resource;
-   uint32_t actions_len;
+   struct mlx5_flow_cb_ctx *ctx = cb_ctx;
+   struct mlx5_flow_dv_modify_hdr_resource *ref = ctx->data;
+   struct mlx5_flow_dv_modify_hdr_resource *resource =
+   container_of(entry, typeof(*resource), entry);
+   uint32_t key_len = sizeof(*ref) - offsetof(typeof(*ref), ft_type);
 
-   resource = (struct mlx5_flow_dv_modify_hdr_resource *)ctx;
-   cache_resource = container_of(entry,
- struct mlx5_flow_dv_modify_hdr_resource,
- entry);
-   actions_len = resource->actions_num * sizeof(resource->actions[0]);
-   if (resource->entry.key == cache_resource->entry.key &&
-   resource->ft_type == cache_resource->ft_type &&
-   resource->actions_num == cache_resource->actions_num &&
-   resource->flags == cache_resource->flags &&
-   !memcmp((const void *)resource->actions,
-   (const void *)cache_resource->actions,
-   actions_len))
-   return 0;
-   return -1;
+   key_len += ref->actions_num * sizeof(ref->actions[0]);
+   ret

[dpdk-dev] [PATCH 15/25] net/mlx5: introduce thread safe linked list cache

2020-10-06 Thread Suanming Mou
From: Xueming Li 

New API of linked list for cache:
- optimized for small amount cache list
- optimized for read-most list
- thread safe
- since number of entries are limited, entries allocated by API
- for dynamic entry size, pass 0 as entry size, then the creation
callback allocate the entry.
- since number of entries are limited, no need to use indexed pool to
allocate memory. API will remove entry and free with mlx5_free.
- search API is not supposed to be used in multi-thread

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/mlx5_utils.c | 170 +
 drivers/net/mlx5/mlx5_utils.h | 172 ++
 2 files changed, 342 insertions(+)

diff --git a/drivers/net/mlx5/mlx5_utils.c b/drivers/net/mlx5/mlx5_utils.c
index 387a988..c0b4ae5 100644
--- a/drivers/net/mlx5/mlx5_utils.c
+++ b/drivers/net/mlx5/mlx5_utils.c
@@ -223,6 +223,176 @@ struct mlx5_hlist_entry*
mlx5_free(h);
 }
 
+/* Cache list /
+
+static struct mlx5_cache_entry *
+mlx5_clist_default_create_cb(struct mlx5_cache_list *list,
+struct mlx5_cache_entry *entry __rte_unused,
+void *ctx __rte_unused)
+{
+   return mlx5_malloc(MLX5_MEM_ZERO, list->entry_sz, 0, SOCKET_ID_ANY);
+}
+
+static void
+mlx5_clist_default_remove_cb(struct mlx5_cache_list *list __rte_unused,
+struct mlx5_cache_entry *entry)
+{
+   mlx5_free(entry);
+}
+
+int
+mlx5_cache_list_init(struct mlx5_cache_list *list, const char *name,
+uint32_t entry_size, void *ctx,
+mlx5_cache_create_cb cb_create,
+mlx5_cache_match_cb cb_match,
+mlx5_cache_remove_cb cb_remove)
+{
+   MLX5_ASSERT(list);
+   if (!cb_match || (!cb_create ^ !cb_remove))
+   return -1;
+   if (name)
+   snprintf(list->name, sizeof(list->name), "%s", name);
+   list->entry_sz = entry_size;
+   list->ctx = ctx;
+   list->cb_create = cb_create ? cb_create : mlx5_clist_default_create_cb;
+   list->cb_match = cb_match;
+   list->cb_remove = cb_remove ? cb_remove : mlx5_clist_default_remove_cb;
+   rte_rwlock_init(&list->lock);
+   DRV_LOG(DEBUG, "Cache list %s initialized.", list->name);
+   LIST_INIT(&list->head);
+   return 0;
+}
+
+static struct mlx5_cache_entry *
+__cache_lookup(struct mlx5_cache_list *list, void *ctx, bool reuse)
+{
+   struct mlx5_cache_entry *entry;
+
+   LIST_FOREACH(entry, &list->head, next) {
+   if (!__atomic_load_n(&entry->ref_cnt, __ATOMIC_RELAXED))
+   /* Ignore entry in middle of removal */
+   continue;
+   if (list->cb_match(list, entry, ctx))
+   continue;
+   if (reuse) {
+   __atomic_add_fetch(&entry->ref_cnt, 1,
+  __ATOMIC_RELAXED);
+   DRV_LOG(DEBUG, "cache list %s entry %p ref++: %u",
+   list->name, (void *)entry, entry->ref_cnt);
+   }
+   break;
+   }
+   return entry;
+}
+
+static struct mlx5_cache_entry *
+cache_lookup(struct mlx5_cache_list *list, void *ctx, bool reuse)
+{
+   struct mlx5_cache_entry *entry;
+
+   rte_rwlock_read_lock(&list->lock);
+   entry = __cache_lookup(list, ctx, reuse);
+   rte_rwlock_read_unlock(&list->lock);
+   return entry;
+}
+
+struct mlx5_cache_entry *
+mlx5_cache_lookup(struct mlx5_cache_list *list, void *ctx)
+{
+   return __cache_lookup(list, ctx, false);
+}
+
+struct mlx5_cache_entry *
+mlx5_cache_register(struct mlx5_cache_list *list, void *ctx)
+{
+   struct mlx5_cache_entry *entry;
+   uint32_t prev_gen_cnt = 0;
+
+   MLX5_ASSERT(list);
+   prev_gen_cnt = __atomic_load_n(&list->gen_cnt, __ATOMIC_ACQUIRE);
+   /* Lookup with read lock, reuse if found. */
+   entry = cache_lookup(list, ctx, true);
+   if (entry)
+   return entry;
+   /* Not found, append with write lock - block read from other threads. */
+   rte_rwlock_write_lock(&list->lock);
+   /* If list changed by other threads before lock, search again. */
+   if (prev_gen_cnt != __atomic_load_n(&list->gen_cnt, __ATOMIC_ACQUIRE)) {
+   /* Lookup and reuse w/o read lock */
+   entry = __cache_lookup(list, ctx, true);
+   if (entry)
+   goto done;
+   }
+   entry = list->cb_create(list, entry, ctx);
+   if (!entry) {
+   if (list->entry_sz)
+   mlx5_free(entry);
+   else if (list->cb_remove)
+   list->cb_remove(list, entry);
+   DRV_LOG(ERR, "Failed to init cache list %s entry %p",
+   list->name, (void *)entry);
+   entry = NULL;
+  

[dpdk-dev] [PATCH 14/25] net/mlx5: remove unused hash list operations

2020-10-06 Thread Suanming Mou
In previous commits the hash list objects have been converted
to new thread safe hash list. The legacy hash list code can be
removed now.

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/mlx5_utils.c | 38 -
 drivers/net/mlx5/mlx5_utils.h | 66 ---
 2 files changed, 104 deletions(-)

diff --git a/drivers/net/mlx5/mlx5_utils.c b/drivers/net/mlx5/mlx5_utils.c
index 4eb3db0..387a988 100644
--- a/drivers/net/mlx5/mlx5_utils.c
+++ b/drivers/net/mlx5/mlx5_utils.c
@@ -164,44 +164,6 @@ struct mlx5_hlist_entry*
return entry;
 }
 
-struct mlx5_hlist_entry *
-mlx5_hlist_lookup_ex(struct mlx5_hlist *h, uint64_t key,
-mlx5_hlist_match_callback_fn cb, void *ctx)
-{
-   uint32_t idx;
-   struct mlx5_hlist_head *first;
-   struct mlx5_hlist_entry *node;
-
-   MLX5_ASSERT(h && cb && ctx);
-   idx = rte_hash_crc_8byte(key, 0) & h->mask;
-   first = &h->heads[idx];
-   LIST_FOREACH(node, first, next) {
-   if (!cb(node, ctx))
-   return node;
-   }
-   return NULL;
-}
-
-int
-mlx5_hlist_insert_ex(struct mlx5_hlist *h, struct mlx5_hlist_entry *entry,
-mlx5_hlist_match_callback_fn cb, void *ctx)
-{
-   uint32_t idx;
-   struct mlx5_hlist_head *first;
-   struct mlx5_hlist_entry *node;
-
-   MLX5_ASSERT(h && entry && cb && ctx);
-   idx = rte_hash_crc_8byte(entry->key, 0) & h->mask;
-   first = &h->heads[idx];
-   /* No need to reuse the lookup function. */
-   LIST_FOREACH(node, first, next) {
-   if (!cb(node, ctx))
-   return -EEXIST;
-   }
-   LIST_INSERT_HEAD(first, entry, next);
-   return 0;
-}
-
 int
 mlx5_hlist_unregister(struct mlx5_hlist *h, struct mlx5_hlist_entry *entry)
 {
diff --git a/drivers/net/mlx5/mlx5_utils.h b/drivers/net/mlx5/mlx5_utils.h
index 8719dee..479dd10 100644
--- a/drivers/net/mlx5/mlx5_utils.h
+++ b/drivers/net/mlx5/mlx5_utils.h
@@ -21,12 +21,6 @@
 
 #include "mlx5_defs.h"
 
-#define mlx5_hlist_remove(h, e) \
-   mlx5_hlist_unregister(h, e)
-
-#define mlx5_hlist_insert(h, e) \
-   mlx5_hlist_register(h, 0, e)
-
 /* Convert a bit number to the corresponding 64-bit mask */
 #define MLX5_BITSHIFT(v) (UINT64_C(1) << (v))
 
@@ -267,23 +261,6 @@ struct mlx5_hlist_entry {
 /** Structure for hash head. */
 LIST_HEAD(mlx5_hlist_head, mlx5_hlist_entry);
 
-/** Type of function that is used to handle the data before freeing. */
-typedef void (*mlx5_hlist_destroy_callback_fn)(void *p, void *ctx);
-
-/**
- * Type of function for user defined matching.
- *
- * @param entry
- *   The entry in the list.
- * @param ctx
- *   The pointer to new entry context.
- *
- * @return
- *   0 if matching, -1 otherwise.
- */
-typedef int (*mlx5_hlist_match_callback_fn)(struct mlx5_hlist_entry *entry,
-void *ctx);
-
 /**
  * Type of callback function for entry removal.
  *
@@ -405,49 +382,6 @@ struct mlx5_hlist_entry *mlx5_hlist_lookup(struct 
mlx5_hlist *h, uint64_t key,
   void *ctx);
 
 /**
- * Extended routine to search an entry matching the context with
- * user defined match function.
- *
- * @param h
- *   Pointer to the hast list table.
- * @param key
- *   Key for the searching entry.
- * @param cb
- *   Callback function to match the node with context.
- * @param ctx
- *   Common context parameter used by callback function.
- *
- * @return
- *   Pointer of the hlist entry if found, NULL otherwise.
- */
-struct mlx5_hlist_entry *mlx5_hlist_lookup_ex(struct mlx5_hlist *h,
- uint64_t key,
- mlx5_hlist_match_callback_fn cb,
- void *ctx);
-
-/**
- * Extended routine to insert an entry to the list with key collisions.
- *
- * For the list have key collision, the extra user defined match function
- * allows node with same key will be inserted.
- *
- * @param h
- *   Pointer to the hast list table.
- * @param entry
- *   Entry to be inserted into the hash list table.
- * @param cb
- *   Callback function to match the node with context.
- * @param ctx
- *   Common context parameter used by callback function.
- *
- * @return
- *   - zero for success.
- *   - -EEXIST if the entry is already inserted.
- */
-int mlx5_hlist_insert_ex(struct mlx5_hlist *h, struct mlx5_hlist_entry *entry,
-mlx5_hlist_match_callback_fn cb, void *ctx);
-
-/**
  * Insert an entry to the hash list table, the entry is only part of whole data
  * element and a 64B key is used for matching. User should construct the key or
  * give a calculated hash signature and guarantee there is no collision.
-- 
1.8.3.1



[dpdk-dev] [PATCH 12/25] net/mlx5: make metadata copy flow list thread safe

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To support multi-thread flow insertion, this patch updates metadata copy
flow list to use thread safe hash list.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/linux/mlx5_os.c |   6 +-
 drivers/net/mlx5/mlx5_flow.c | 276 +++
 drivers/net/mlx5/mlx5_flow.h |   6 +-
 3 files changed, 139 insertions(+), 149 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 744b1dd..50d3d99 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -1354,13 +1354,13 @@
mlx5_flow_ext_mreg_supported(eth_dev) &&
priv->sh->dv_regc0_mask) {
priv->mreg_cp_tbl = mlx5_hlist_create(MLX5_FLOW_MREG_HNAME,
- MLX5_FLOW_MREG_HTABLE_SZ,
- 0, false,
- NULL, NULL, NULL);
+   MLX5_FLOW_MREG_HTABLE_SZ, 0, false,
+   flow_dv_mreg_create_cb, NULL, flow_dv_mreg_remove_cb);
if (!priv->mreg_cp_tbl) {
err = ENOMEM;
goto error;
}
+   priv->mreg_cp_tbl->ctx = eth_dev;
}
return eth_dev;
 error:
diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c
index 3a3b783..3808e2b 100644
--- a/drivers/net/mlx5/mlx5_flow.c
+++ b/drivers/net/mlx5/mlx5_flow.c
@@ -2950,36 +2950,17 @@ struct mlx5_flow_tunnel_info {
 flow_list_destroy(struct rte_eth_dev *dev, uint32_t *list,
  uint32_t flow_idx);
 
-/**
- * Add a flow of copying flow metadata registers in RX_CP_TBL.
- *
- * As mark_id is unique, if there's already a registered flow for the mark_id,
- * return by increasing the reference counter of the resource. Otherwise, 
create
- * the resource (mcp_res) and flow.
- *
- * Flow looks like,
- *   - If ingress port is ANY and reg_c[1] is mark_id,
- * flow_tag := mark_id, reg_b := reg_c[0] and jump to RX_ACT_TBL.
- *
- * For default flow (zero mark_id), flow is like,
- *   - If ingress port is ANY,
- * reg_b := reg_c[0] and jump to RX_ACT_TBL.
- *
- * @param dev
- *   Pointer to Ethernet device.
- * @param mark_id
- *   ID of MARK action, zero means default flow for META.
- * @param[out] error
- *   Perform verbose error reporting if not NULL.
- *
- * @return
- *   Associated resource on success, NULL otherwise and rte_errno is set.
- */
-static struct mlx5_flow_mreg_copy_resource *
-flow_mreg_add_copy_action(struct rte_eth_dev *dev, uint32_t mark_id,
- struct rte_flow_error *error)
+struct mlx5_hlist_entry *
+flow_dv_mreg_create_cb(struct mlx5_hlist *list, uint64_t key,
+  void *cb_ctx)
 {
+   struct rte_eth_dev *dev = list->ctx;
struct mlx5_priv *priv = dev->data->dev_private;
+   struct mlx5_flow_cb_ctx *ctx = cb_ctx;
+   struct mlx5_flow_mreg_copy_resource *mcp_res;
+   uint32_t idx = 0;
+   int ret;
+   uint32_t mark_id = key;
struct rte_flow_attr attr = {
.group = MLX5_FLOW_MREG_CP_TABLE_GROUP,
.ingress = 1,
@@ -3003,30 +2984,22 @@ struct mlx5_flow_tunnel_info {
struct rte_flow_action actions[] = {
[3] = { .type = RTE_FLOW_ACTION_TYPE_END, },
};
-   struct mlx5_flow_mreg_copy_resource *mcp_res;
-   uint32_t idx = 0;
-   int ret;
 
/* Fill the register fileds in the flow. */
-   ret = mlx5_flow_get_reg_id(dev, MLX5_FLOW_MARK, 0, error);
+   ret = mlx5_flow_get_reg_id(ctx->dev, MLX5_FLOW_MARK, 0, ctx->error);
if (ret < 0)
return NULL;
tag_spec.id = ret;
-   ret = mlx5_flow_get_reg_id(dev, MLX5_METADATA_RX, 0, error);
+   ret = mlx5_flow_get_reg_id(ctx->dev, MLX5_METADATA_RX, 0, ctx->error);
if (ret < 0)
return NULL;
cp_mreg.src = ret;
-   /* Check if already registered. */
-   MLX5_ASSERT(priv->mreg_cp_tbl);
-   mcp_res = (void *)mlx5_hlist_lookup(priv->mreg_cp_tbl, mark_id, NULL);
-   if (mcp_res) {
-   /* For non-default rule. */
-   if (mark_id != MLX5_DEFAULT_COPY_ID)
-   mcp_res->refcnt++;
-   MLX5_ASSERT(mark_id != MLX5_DEFAULT_COPY_ID ||
-   mcp_res->refcnt == 1);
-   return mcp_res;
+   mcp_res = mlx5_ipool_zmalloc(priv->sh->ipool[MLX5_IPOOL_MCP], &idx);
+   if (!mcp_res) {
+   rte_errno = ENOMEM;
+   return NULL;
}
+   mcp_res->idx = idx;
/* Provide the full width of FLAG specific value. */
if (mark_id == (priv->sh->dv_regc0_mask & MLX5_FLOW_MARK_DEFAULT))
tag_spec.data = MLX5_FLOW_MARK_DEFAULT;
@@ -3076,39 +3049,68 @@ struct mlx5_flow_tunnel_info {
.type = RTE_FLOW_ACTION_TYPE_END,

[dpdk-dev] [PATCH 16/25] net/mlx5: make Rx queue thread safe

2020-10-06 Thread Suanming Mou
This commit applies the cache linked list to Rx queue to make it thread
safe.

Signed-off-by: Suanming Mou 
---
 drivers/net/mlx5/linux/mlx5_os.c   |   5 +
 drivers/net/mlx5/mlx5.c|   1 +
 drivers/net/mlx5/mlx5.h|  24 +++-
 drivers/net/mlx5/mlx5_flow.h   |  16 ---
 drivers/net/mlx5/mlx5_flow_dv.c|  20 +---
 drivers/net/mlx5/mlx5_flow_verbs.c |  19 +--
 drivers/net/mlx5/mlx5_rxq.c| 234 ++---
 drivers/net/mlx5/mlx5_rxtx.h   |  20 ++--
 8 files changed, 164 insertions(+), 175 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 24cf348..db7b0de 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -1335,6 +1335,10 @@
err = ENOTSUP;
goto error;
}
+   mlx5_cache_list_init(&priv->hrxqs, "hrxq", 0, eth_dev,
+mlx5_hrxq_create_cb,
+mlx5_hrxq_match_cb,
+mlx5_hrxq_remove_cb);
/* Query availability of metadata reg_c's. */
err = mlx5_flow_discover_mreg_c(eth_dev);
if (err < 0) {
@@ -1381,6 +1385,7 @@
mlx5_vlan_vmwa_exit(priv->vmwa_context);
if (own_domain_id)
claim_zero(rte_eth_switch_domain_free(priv->domain_id));
+   mlx5_cache_list_destroy(&priv->hrxqs);
mlx5_free(priv);
if (eth_dev != NULL)
eth_dev->data->dev_private = NULL;
diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c
index 61e5e69..fc9c5a9 100644
--- a/drivers/net/mlx5/mlx5.c
+++ b/drivers/net/mlx5/mlx5.c
@@ -1219,6 +1219,7 @@ struct mlx5_dev_ctx_shared *
close(priv->nl_socket_rdma);
if (priv->vmwa_context)
mlx5_vlan_vmwa_exit(priv->vmwa_context);
+   mlx5_cache_list_destroy(&priv->hrxqs);
ret = mlx5_hrxq_verify(dev);
if (ret)
DRV_LOG(WARNING, "port %u some hash Rx queue still remain",
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index f11d783..97729a8 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -61,6 +61,13 @@ enum mlx5_reclaim_mem_mode {
MLX5_RCM_AGGR, /* Reclaim PMD and rdma-core level. */
 };
 
+/* Hash list callback context */
+struct mlx5_flow_cb_ctx {
+   struct rte_eth_dev *dev;
+   struct rte_flow_error *error;
+   void *data;
+};
+
 /* Device attributes used in mlx5 PMD */
 struct mlx5_dev_attr {
uint64_tdevice_cap_flags_ex;
@@ -664,6 +671,18 @@ struct mlx5_proc_priv {
 /* MTR list. */
 TAILQ_HEAD(mlx5_flow_meters, mlx5_flow_meter);
 
+/* RSS description. */
+struct mlx5_flow_rss_desc {
+   uint32_t level;
+   uint32_t queue_num; /**< Number of entries in @p queue. */
+   uint64_t types; /**< Specific RSS hash types (see ETH_RSS_*). */
+   uint64_t hash_fields; /* Verbs Hash fields. */
+   uint8_t key[MLX5_RSS_HASH_KEY_LEN]; /**< RSS hash key. */
+   uint32_t key_len; /**< RSS hash key len. */
+   uint32_t tunnel; /**< Queue in tunnel. */
+   uint16_t *queue; /**< Destination queues. */
+};
+
 #define MLX5_PROC_PRIV(port_id) \
((struct mlx5_proc_priv *)rte_eth_devices[port_id].process_private)
 
@@ -710,7 +729,7 @@ struct mlx5_ind_table_obj {
 
 /* Hash Rx queue. */
 struct mlx5_hrxq {
-   ILIST_ENTRY(uint32_t)next; /* Index to the next element. */
+   struct mlx5_cache_entry entry; /* Cache entry. */
rte_atomic32_t refcnt; /* Reference counter. */
struct mlx5_ind_table_obj *ind_table; /* Indirection table. */
RTE_STD_C11
@@ -723,6 +742,7 @@ struct mlx5_hrxq {
 #endif
uint64_t hash_fields; /* Verbs Hash fields. */
uint32_t rss_key_len; /* Hash key length in bytes. */
+   uint32_t idx; /* Hash Rx queue index*/
uint8_t rss_key[]; /* Hash key. */
 };
 
@@ -788,7 +808,7 @@ struct mlx5_priv {
struct mlx5_obj_ops obj_ops; /* HW objects operations. */
LIST_HEAD(rxq, mlx5_rxq_ctrl) rxqsctrl; /* DPDK Rx queues. */
LIST_HEAD(rxqobj, mlx5_rxq_obj) rxqsobj; /* Verbs/DevX Rx queues. */
-   uint32_t hrxqs; /* Verbs Hash Rx queues. */
+   struct mlx5_cache_list hrxqs; /* Verbs Hash Rx queues. */
LIST_HEAD(txq, mlx5_txq_ctrl) txqsctrl; /* DPDK Tx queues. */
LIST_HEAD(txqobj, mlx5_txq_obj) txqsobj; /* Verbs/DevX Tx queues. */
/* Indirection tables. */
diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h
index 1fe0b30..6ec0e72 100644
--- a/drivers/net/mlx5/mlx5_flow.h
+++ b/drivers/net/mlx5/mlx5_flow.h
@@ -368,13 +368,6 @@ enum mlx5_flow_fate_type {
MLX5_FLOW_FATE_MAX,
 };
 
-/* Hash list callback context */
-struct mlx5_flow_cb_ctx {
-   struct rte_eth_dev *dev;
-   struct rte_flow_error *error;
-   void *data;
-};
-
 /* Matcher PRM representation */
 struct

[dpdk-dev] [PATCH 19/25] net/mlx5: make push VLAN action cache thread safe

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To support multi-thread flow insertion, this patch converts push VLAN
action cache list to thread safe cache list.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/linux/mlx5_os.c |   7 ++
 drivers/net/mlx5/mlx5.h  |   2 +-
 drivers/net/mlx5/mlx5_flow.h |  13 +++-
 drivers/net/mlx5/mlx5_flow_dv.c  | 157 +--
 4 files changed, 102 insertions(+), 77 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index 3618e54..63ea55b 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -243,6 +243,12 @@
 flow_dv_port_id_create_cb,
 flow_dv_port_id_match_cb,
 flow_dv_port_id_remove_cb);
+   /* Init push vlan action cache list. */
+   snprintf(s, sizeof(s), "%s_push_vlan_action_cache", sh->ibdev_name);
+   mlx5_cache_list_init(&sh->push_vlan_action_list, s, 0, sh,
+flow_dv_push_vlan_create_cb,
+flow_dv_push_vlan_match_cb,
+flow_dv_push_vlan_remove_cb);
/* Create tags hash list table. */
snprintf(s, sizeof(s), "%s_tags", sh->ibdev_name);
sh->tag_table = mlx5_hlist_create(s, MLX5_TAGS_HLIST_ARRAY_SIZE, 0,
@@ -427,6 +433,7 @@
sh->tag_table = NULL;
}
mlx5_cache_list_destroy(&sh->port_id_action_list);
+   mlx5_cache_list_destroy(&sh->push_vlan_action_list);
mlx5_free_table_hash_list(priv);
 }
 
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index ab44fa0..b2312cf 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -641,7 +641,7 @@ struct mlx5_dev_ctx_shared {
struct mlx5_hlist *modify_cmds;
struct mlx5_hlist *tag_table;
struct mlx5_cache_list port_id_action_list; /* Port ID action cache. */
-   uint32_t push_vlan_action_list; /* List of push VLAN actions. */
+   struct mlx5_cache_list push_vlan_action_list; /* Push VLAN actions. */
struct mlx5_flow_counter_mng cmng; /* Counters management structure. */
struct mlx5_flow_default_miss_resource default_miss;
/* Default miss action resource structure. */
diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h
index 50aa7ea..2e060e6 100644
--- a/drivers/net/mlx5/mlx5_flow.h
+++ b/drivers/net/mlx5/mlx5_flow.h
@@ -479,12 +479,11 @@ struct mlx5_flow_dv_port_id_action_resource {
 
 /* Push VLAN action resource structure */
 struct mlx5_flow_dv_push_vlan_action_resource {
-   ILIST_ENTRY(uint32_t)next;
-   /* Pointer to next element. */
-   rte_atomic32_t refcnt; /**< Reference counter. */
+   struct mlx5_cache_entry entry; /* Cache entry. */
void *action; /**< Action object. */
uint8_t ft_type; /**< Flow table type, Rx, Tx or FDB. */
rte_be32_t vlan_tag; /**< VLAN tag value. */
+   uint32_t idx; /**< Indexed pool memory index. */
 };
 
 /* Metadata register copy table entry. */
@@ -1107,4 +1106,12 @@ struct mlx5_cache_entry 
*flow_dv_port_id_create_cb(struct mlx5_cache_list *list,
 void flow_dv_port_id_remove_cb(struct mlx5_cache_list *list,
   struct mlx5_cache_entry *entry);
 
+int flow_dv_push_vlan_match_cb(struct mlx5_cache_list *list,
+  struct mlx5_cache_entry *entry, void *cb_ctx);
+struct mlx5_cache_entry *flow_dv_push_vlan_create_cb
+   (struct mlx5_cache_list *list,
+struct mlx5_cache_entry *entry, void *cb_ctx);
+void flow_dv_push_vlan_remove_cb(struct mlx5_cache_list *list,
+struct mlx5_cache_entry *entry);
+
 #endif /* RTE_PMD_MLX5_FLOW_H_ */
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 7882bb4..7c9b9190 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -2862,6 +2862,58 @@ struct mlx5_cache_entry *
return 0;
 }
 
+int
+flow_dv_push_vlan_match_cb(struct mlx5_cache_list *list __rte_unused,
+struct mlx5_cache_entry *entry, void *cb_ctx)
+{
+   struct mlx5_flow_cb_ctx *ctx = cb_ctx;
+   struct mlx5_flow_dv_push_vlan_action_resource *ref = ctx->data;
+   struct mlx5_flow_dv_push_vlan_action_resource *res =
+   container_of(entry, typeof(*res), entry);
+
+   return ref->vlan_tag != res->vlan_tag || ref->ft_type != res->ft_type;
+}
+
+struct mlx5_cache_entry *
+flow_dv_push_vlan_create_cb(struct mlx5_cache_list *list,
+ struct mlx5_cache_entry *entry __rte_unused,
+ void *cb_ctx)
+{
+   struct mlx5_dev_ctx_shared *sh = list->ctx;
+   struct mlx5_flow_cb_ctx *ctx = cb_ctx;
+   struct mlx5_flow_dv_push_vlan_action_resource *ref = ctx->data;
+   struct mlx5_flow_dv_push_vlan_action_resource *cache;
+

[dpdk-dev] [PATCH 18/25] net/mlx5: make port ID action cache thread safe

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To support multi-thread flow insertion, this patch convert port id
action cache list to thread safe cache list.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/linux/mlx5_os.c |   7 ++
 drivers/net/mlx5/mlx5.h  |   2 +-
 drivers/net/mlx5/mlx5_flow.h |  15 +++--
 drivers/net/mlx5/mlx5_flow_dv.c  | 140 +--
 4 files changed, 94 insertions(+), 70 deletions(-)

diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c
index db7b0de..3618e54 100644
--- a/drivers/net/mlx5/linux/mlx5_os.c
+++ b/drivers/net/mlx5/linux/mlx5_os.c
@@ -237,6 +237,12 @@
goto error;
/* The resources below are only valid with DV support. */
 #ifdef HAVE_IBV_FLOW_DV_SUPPORT
+   /* Init port id action cache list. */
+   snprintf(s, sizeof(s), "%s_port_id_action_cache", sh->ibdev_name);
+   mlx5_cache_list_init(&sh->port_id_action_list, s, 0, sh,
+flow_dv_port_id_create_cb,
+flow_dv_port_id_match_cb,
+flow_dv_port_id_remove_cb);
/* Create tags hash list table. */
snprintf(s, sizeof(s), "%s_tags", sh->ibdev_name);
sh->tag_table = mlx5_hlist_create(s, MLX5_TAGS_HLIST_ARRAY_SIZE, 0,
@@ -420,6 +426,7 @@
mlx5_hlist_destroy(sh->tag_table);
sh->tag_table = NULL;
}
+   mlx5_cache_list_destroy(&sh->port_id_action_list);
mlx5_free_table_hash_list(priv);
 }
 
diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index ee2211b..ab44fa0 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -640,7 +640,7 @@ struct mlx5_dev_ctx_shared {
struct mlx5_hlist *encaps_decaps; /* Encap/decap action hash list. */
struct mlx5_hlist *modify_cmds;
struct mlx5_hlist *tag_table;
-   uint32_t port_id_action_list; /* List of port ID actions. */
+   struct mlx5_cache_list port_id_action_list; /* Port ID action cache. */
uint32_t push_vlan_action_list; /* List of push VLAN actions. */
struct mlx5_flow_counter_mng cmng; /* Counters management structure. */
struct mlx5_flow_default_miss_resource default_miss;
diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h
index fe9b6a6..50aa7ea 100644
--- a/drivers/net/mlx5/mlx5_flow.h
+++ b/drivers/net/mlx5/mlx5_flow.h
@@ -471,12 +471,10 @@ struct mlx5_flow_dv_jump_tbl_resource {
 
 /* Port ID resource structure. */
 struct mlx5_flow_dv_port_id_action_resource {
-   ILIST_ENTRY(uint32_t)next;
-   /* Pointer to next element. */
-   rte_atomic32_t refcnt; /**< Reference counter. */
-   void *action;
-   /**< Action object. */
+   struct mlx5_cache_entry entry;
+   void *action; /**< Action object. */
uint32_t port_id; /**< Port ID value. */
+   uint32_t idx; /**< Indexed pool memory index. */
 };
 
 /* Push VLAN action resource structure */
@@ -1102,4 +1100,11 @@ struct mlx5_cache_entry 
*flow_dv_matcher_create_cb(struct mlx5_cache_list *list,
 void flow_dv_matcher_remove_cb(struct mlx5_cache_list *list,
   struct mlx5_cache_entry *entry);
 
+int flow_dv_port_id_match_cb(struct mlx5_cache_list *list,
+struct mlx5_cache_entry *entry, void *cb_ctx);
+struct mlx5_cache_entry *flow_dv_port_id_create_cb(struct mlx5_cache_list 
*list,
+   struct mlx5_cache_entry *entry, void *cb_ctx);
+void flow_dv_port_id_remove_cb(struct mlx5_cache_list *list,
+  struct mlx5_cache_entry *entry);
+
 #endif /* RTE_PMD_MLX5_FLOW_H_ */
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 3774e46..7882bb4 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -2741,6 +2741,52 @@ struct mlx5_hlist_entry *
return 0;
 }
 
+int
+flow_dv_port_id_match_cb(struct mlx5_cache_list *list __rte_unused,
+struct mlx5_cache_entry *entry, void *cb_ctx)
+{
+   struct mlx5_flow_cb_ctx *ctx = cb_ctx;
+   struct mlx5_flow_dv_port_id_action_resource *ref = ctx->data;
+   struct mlx5_flow_dv_port_id_action_resource *res =
+   container_of(entry, typeof(*res), entry);
+
+   return ref->port_id != res->port_id;
+}
+
+struct mlx5_cache_entry *
+flow_dv_port_id_create_cb(struct mlx5_cache_list *list,
+ struct mlx5_cache_entry *entry __rte_unused,
+ void *cb_ctx)
+{
+   struct mlx5_dev_ctx_shared *sh = list->ctx;
+   struct mlx5_flow_cb_ctx *ctx = cb_ctx;
+   struct mlx5_flow_dv_port_id_action_resource *ref = ctx->data;
+   struct mlx5_flow_dv_port_id_action_resource *cache;
+   uint32_t idx;
+   int ret;
+
+   /* Register new port id action resource. */
+   cache = mlx5_ipool_zmalloc(sh->ipool[MLX5_IPOOL_PORT_ID], &idx);
+   if (!cache) {
+   rte_fl

[dpdk-dev] [PATCH 17/25] net/mlx5: make matcher list thread safe

2020-10-06 Thread Suanming Mou
From: Xueming Li 

To support multi-thread flow insertion, this path converts matcher list
to use thread safe cache list API.

Signed-off-by: Xueming Li 
---
 drivers/net/mlx5/mlx5.h |   3 +
 drivers/net/mlx5/mlx5_flow.h|  15 ++-
 drivers/net/mlx5/mlx5_flow_dv.c | 209 +---
 3 files changed, 126 insertions(+), 101 deletions(-)

diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h
index 97729a8..ee2211b 100644
--- a/drivers/net/mlx5/mlx5.h
+++ b/drivers/net/mlx5/mlx5.h
@@ -32,6 +32,9 @@
 #include "mlx5_os.h"
 #include "mlx5_autoconf.h"
 
+
+#define MLX5_SH(dev) (((struct mlx5_priv *)(dev)->data->dev_private)->sh)
+
 enum mlx5_ipool_index {
 #ifdef HAVE_IBV_FLOW_DV_SUPPORT
MLX5_IPOOL_DECAP_ENCAP = 0, /* Pool for encap/decap resource. */
diff --git a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h
index 6ec0e72..fe9b6a6 100644
--- a/drivers/net/mlx5/mlx5_flow.h
+++ b/drivers/net/mlx5/mlx5_flow.h
@@ -378,11 +378,9 @@ struct mlx5_flow_dv_match_params {
 
 /* Matcher structure. */
 struct mlx5_flow_dv_matcher {
-   LIST_ENTRY(mlx5_flow_dv_matcher) next;
-   /**< Pointer to the next element. */
+   struct mlx5_cache_entry entry; /**< Pointer to the next element. */
struct mlx5_flow_tbl_resource *tbl;
/**< Pointer to the table(group) the matcher associated with. */
-   rte_atomic32_t refcnt; /**< Reference counter. */
void *matcher_object; /**< Pointer to DV matcher */
uint16_t crc; /**< CRC of key. */
uint16_t priority; /**< Priority of matcher. */
@@ -512,11 +510,12 @@ struct mlx5_flow_tbl_data_entry {
/**< hash list entry, 64-bits key inside. */
struct mlx5_flow_tbl_resource tbl;
/**< flow table resource. */
-   LIST_HEAD(matchers, mlx5_flow_dv_matcher) matchers;
+   struct mlx5_cache_list matchers;
/**< matchers' header associated with the flow table. */
struct mlx5_flow_dv_jump_tbl_resource jump;
/**< jump resource, at most one for each table created. */
uint32_t idx; /**< index for the indexed mempool. */
+   uint8_t direction; /**< table direction, 0: ingress, 1: egress. */
 };
 
 /* Verbs specification header. */
@@ -1095,4 +1094,12 @@ struct mlx5_hlist_entry 
*flow_dv_encap_decap_create_cb(struct mlx5_hlist *list,
uint64_t key, void *cb_ctx);
 void flow_dv_encap_decap_remove_cb(struct mlx5_hlist *list,
   struct mlx5_hlist_entry *entry);
+
+int flow_dv_matcher_match_cb(struct mlx5_cache_list *list,
+struct mlx5_cache_entry *entry, void *ctx);
+struct mlx5_cache_entry *flow_dv_matcher_create_cb(struct mlx5_cache_list 
*list,
+   struct mlx5_cache_entry *entry, void *ctx);
+void flow_dv_matcher_remove_cb(struct mlx5_cache_list *list,
+  struct mlx5_cache_entry *entry);
+
 #endif /* RTE_PMD_MLX5_FLOW_H_ */
diff --git a/drivers/net/mlx5/mlx5_flow_dv.c b/drivers/net/mlx5/mlx5_flow_dv.c
index 5092130..3774e46 100644
--- a/drivers/net/mlx5/mlx5_flow_dv.c
+++ b/drivers/net/mlx5/mlx5_flow_dv.c
@@ -70,7 +70,7 @@
 };
 
 static int
-flow_dv_tbl_resource_release(struct rte_eth_dev *dev,
+flow_dv_tbl_resource_release(struct mlx5_dev_ctx_shared *sh,
 struct mlx5_flow_tbl_resource *tbl);
 
 static int
@@ -2730,7 +2730,7 @@ struct mlx5_hlist_entry *
(void *)&tbl_data->jump, cnt);
} else {
/* old jump should not make the table ref++. */
-   flow_dv_tbl_resource_release(dev, &tbl_data->tbl);
+   flow_dv_tbl_resource_release(MLX5_SH(dev), &tbl_data->tbl);
MLX5_ASSERT(tbl_data->jump.action);
DRV_LOG(DEBUG, "existed jump table resource %p: refcnt %d++",
(void *)&tbl_data->jump, cnt);
@@ -7608,6 +7608,7 @@ struct mlx5_hlist_entry *
return NULL;
}
tbl_data->idx = idx;
+   tbl_data->direction = key.direction;
tbl = &tbl_data->tbl;
if (key.dummy)
return &tbl_data->entry;
@@ -7625,6 +7626,13 @@ struct mlx5_hlist_entry *
mlx5_ipool_free(sh->ipool[MLX5_IPOOL_JUMP], idx);
return NULL;
}
+   MKSTR(matcher_name, "%s_%s_%u_matcher_cache",
+ key.domain ? "FDB" : "NIC", key.direction ? "egress" : "ingress",
+ key.table_id);
+   mlx5_cache_list_init(&tbl_data->matchers, matcher_name, 0, sh,
+flow_dv_matcher_create_cb,
+flow_dv_matcher_match_cb,
+flow_dv_matcher_remove_cb);
rte_atomic32_init(&tbl_data->jump.refcnt);
return &tbl_data->entry;
 }
@@ -7684,14 +7692,15 @@ struct mlx5_flow_tbl_resource *
MLX5_ASSERT(entry && sh);
if (tbl_data->tbl.obj)
mlx5_flow_os_destroy_flow_tbl(tbl_data->tbl.obj);

  1   2   3   >