Re: [PATCH 0/2] fix build disabling common/mlx5

2022-10-30 Thread Andrew Rybchenko

On 10/30/22 02:17, Thomas Monjalon wrote:

Andrew reported a build failure when disabling mlx5 common driver.
It is a blocker for -rc2 release.

While fixing the use of a variable across mlx5 drivers in first patch,
the consistency of its use is improved in a second patch.

Thomas Monjalon (2):
   net/mlx5: fix disabling common/mlx5 dependency
   common/mlx5: move Meson config initialization and check

  drivers/common/mlx5/linux/meson.build   | 2 --
  drivers/common/mlx5/meson.build | 2 ++
  drivers/common/mlx5/windows/meson.build | 4 
  drivers/net/mlx5/hws/meson.build| 4 
  drivers/net/mlx5/meson.build| 8 +---
  5 files changed, 11 insertions(+), 9 deletions(-)



Many thanks, it solves my build problem:

Series-tested-by: Andrew Rybchenko 



[PATCH v2 1/2] common/mlx5: fix build disabling

2022-10-30 Thread Thomas Monjalon
If the dependency common/mlx5 is explicitly disabled,
but net/mlx5 is not explicitly disabled,
Meson will read the full recipe of net/mlx5
and will fail when accessing a variable from common/mlx5:
drivers/net/mlx5/meson.build:76:4: ERROR: Unknown variable "mlx5_config".

The solution is to stop parsing net/mlx5 if common/mlx5 is disabled.
The deps array must be defined before stopping, in order to automatically
disable the build of net/mlx5 and print the reason.

The same protection is applied to other mlx5 drivers,
so it will allow using the variable mlx5_config in future.

Fixes: 22681deead3e ("net/mlx5/hws: enable hardware steering")

Reported-by: Andrew Rybchenko 
Signed-off-by: Thomas Monjalon 
Tested-by: Andrew Rybchenko 
---
 drivers/compress/mlx5/meson.build | 5 +
 drivers/crypto/mlx5/meson.build   | 5 +
 drivers/net/mlx5/meson.build  | 5 +
 drivers/regex/mlx5/meson.build| 5 +
 drivers/vdpa/mlx5/meson.build | 5 +
 5 files changed, 25 insertions(+)

diff --git a/drivers/compress/mlx5/meson.build 
b/drivers/compress/mlx5/meson.build
index 7aac329986..49ce3aff46 100644
--- a/drivers/compress/mlx5/meson.build
+++ b/drivers/compress/mlx5/meson.build
@@ -9,6 +9,11 @@ endif
 
 fmt_name = 'mlx5_compress'
 deps += ['common_mlx5', 'eal', 'compressdev']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 sources = files(
 'mlx5_compress.c',
 )
diff --git a/drivers/crypto/mlx5/meson.build b/drivers/crypto/mlx5/meson.build
index 9d9c9c00bc..7521c4c671 100644
--- a/drivers/crypto/mlx5/meson.build
+++ b/drivers/crypto/mlx5/meson.build
@@ -9,6 +9,11 @@ endif
 
 fmt_name = 'mlx5_crypto'
 deps += ['common_mlx5', 'eal', 'cryptodev']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 sources = files(
 'mlx5_crypto.c',
 'mlx5_crypto_dek.c',
diff --git a/drivers/net/mlx5/meson.build b/drivers/net/mlx5/meson.build
index ff84448186..fa15158039 100644
--- a/drivers/net/mlx5/meson.build
+++ b/drivers/net/mlx5/meson.build
@@ -9,6 +9,11 @@ if not (is_linux or is_windows)
 endif
 
 deps += ['hash', 'common_mlx5']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 headers = files('rte_pmd_mlx5.h')
 sources = files(
 'mlx5.c',
diff --git a/drivers/regex/mlx5/meson.build b/drivers/regex/mlx5/meson.build
index e553dcb83d..70edc5b6da 100644
--- a/drivers/regex/mlx5/meson.build
+++ b/drivers/regex/mlx5/meson.build
@@ -8,6 +8,11 @@ if not is_linux
 endif
 
 deps += ['common_mlx5', 'eal', 'regexdev']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 sources = files(
 'mlx5_regex.c',
 'mlx5_rxp.c',
diff --git a/drivers/vdpa/mlx5/meson.build b/drivers/vdpa/mlx5/meson.build
index 9d8dbb1a82..54a4eac6f4 100644
--- a/drivers/vdpa/mlx5/meson.build
+++ b/drivers/vdpa/mlx5/meson.build
@@ -8,6 +8,11 @@ if not is_linux
 endif
 
 deps += ['hash', 'common_mlx5', 'vhost', 'pci', 'eal', 'sched']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 sources = files(
 'mlx5_vdpa.c',
 'mlx5_vdpa_mem.c',
-- 
2.36.1



[PATCH v2 0/2] fix build disabling common/mlx5

2022-10-30 Thread Thomas Monjalon
Andrew reported a build failure when disabling mlx5 common driver.
It is a blocker for -rc2 release.

While fixing the use of a variable across mlx5 drivers in first patch,
the consistency of its use is improved in a second patch.

v2: apply the same protection to other mlx5 drivers

Thomas Monjalon (2):
  common/mlx5: fix build disabling
  common/mlx5: move Meson config initialization and check

 drivers/common/mlx5/linux/meson.build   | 2 --
 drivers/common/mlx5/meson.build | 2 ++
 drivers/common/mlx5/windows/meson.build | 4 
 drivers/compress/mlx5/meson.build   | 5 +
 drivers/crypto/mlx5/meson.build | 5 +
 drivers/net/mlx5/hws/meson.build| 4 
 drivers/net/mlx5/meson.build| 9 ++---
 drivers/regex/mlx5/meson.build  | 5 +
 drivers/vdpa/mlx5/meson.build   | 5 +
 9 files changed, 32 insertions(+), 9 deletions(-)

-- 
2.36.1



[PATCH v2 2/2] common/mlx5: move Meson config initialization and check

2022-10-30 Thread Thomas Monjalon
The variable mlx5_config may be used by other mlx5 drivers
and should be always initialized.
By moving its initialization (with configuration file generation),
it is made consistent for Linux and Windows builds.

And the check of mlx5_config in net/mlx5 is moved at the top of
net/mlx5/hws/meson.build so HWS requirements are in the right context.

Signed-off-by: Thomas Monjalon 
Tested-by: Andrew Rybchenko 
---
 drivers/common/mlx5/linux/meson.build   | 2 --
 drivers/common/mlx5/meson.build | 2 ++
 drivers/common/mlx5/windows/meson.build | 4 
 drivers/net/mlx5/hws/meson.build| 4 
 drivers/net/mlx5/meson.build| 4 +---
 5 files changed, 7 insertions(+), 9 deletions(-)

diff --git a/drivers/common/mlx5/linux/meson.build 
b/drivers/common/mlx5/linux/meson.build
index 84e2a1ad8c..2b57b299bd 100644
--- a/drivers/common/mlx5/linux/meson.build
+++ b/drivers/common/mlx5/linux/meson.build
@@ -8,7 +8,6 @@ dlopen_ibverbs = (get_option('ibverbs_link') == 'dlopen')
 LIB_GLUE_BASE = 'librte_common_mlx5_glue.so'
 LIB_GLUE_VERSION = abi_version
 LIB_GLUE = LIB_GLUE_BASE + '.' + LIB_GLUE_VERSION
-mlx5_config = configuration_data()
 if dlopen_ibverbs
 dpdk_conf.set('RTE_IBVERBS_LINK_DLOPEN', 1)
 cflags += [
@@ -232,7 +231,6 @@ foreach arg:has_member_args
 file_prefix = '#include <' + arg[1] + '>'
 mlx5_config.set(arg[0], cc.has_member(arg[2], arg[3], prefix : 
file_prefix, dependencies: libs))
 endforeach
-configure_file(output : 'mlx5_autoconf.h', configuration : mlx5_config)
 
 # Build Glue Library
 if dlopen_ibverbs
diff --git a/drivers/common/mlx5/meson.build b/drivers/common/mlx5/meson.build
index 6ddbde7e8f..d7ca21d2cf 100644
--- a/drivers/common/mlx5/meson.build
+++ b/drivers/common/mlx5/meson.build
@@ -37,4 +37,6 @@ else
 cflags += [ '-UPEDANTIC' ]
 endif
 
+mlx5_config = configuration_data()
 subdir(exec_env)
+configure_file(output : 'mlx5_autoconf.h', configuration : mlx5_config)
diff --git a/drivers/common/mlx5/windows/meson.build 
b/drivers/common/mlx5/windows/meson.build
index edbbaa9ae1..cc486014a8 100644
--- a/drivers/common/mlx5/windows/meson.build
+++ b/drivers/common/mlx5/windows/meson.build
@@ -39,7 +39,3 @@ if get_option('buildtype').contains('debug')
 else
 cflags += [ '-UPEDANTIC' ]
 endif
-
-# Generate an empty mlx5_autoconf.h file for compatibility with Linux
-config = configuration_data()
-configure_file(output : 'mlx5_autoconf.h', configuration : config)
diff --git a/drivers/net/mlx5/hws/meson.build b/drivers/net/mlx5/hws/meson.build
index d2bb864fd2..38776d5163 100644
--- a/drivers/net/mlx5/hws/meson.build
+++ b/drivers/net/mlx5/hws/meson.build
@@ -1,6 +1,10 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright (c) 2022 NVIDIA Corporation & Affiliates
 
+if not (is_linux and mlx5_config.get('HAVE_IBV_FLOW_DV_SUPPORT', false))
+subdir_done()
+endif
+
 includes += include_directories('.')
 sources += files(
 'mlx5dr_context.c',
diff --git a/drivers/net/mlx5/meson.build b/drivers/net/mlx5/meson.build
index fa15158039..f1aab18f82 100644
--- a/drivers/net/mlx5/meson.build
+++ b/drivers/net/mlx5/meson.build
@@ -78,6 +78,4 @@ testpmd_sources += files('mlx5_testpmd.c')
 
 subdir(exec_env)
 
-if (is_linux and mlx5_config.get('HAVE_IBV_FLOW_DV_SUPPORT', false))
-subdir('hws')
-endif
+subdir('hws')
-- 
2.36.1



RE: Copy-pasted code should be updated

2022-10-30 Thread Morten Brørup
> From: Honnappa Nagarahalli [mailto:honnappa.nagaraha...@arm.com]
> Sent: Tuesday, 11 October 2022 23.48
> 
> 
> 
> >
> > Dear Intel PMD maintainers (CC: techboard),
> >
> > I strongly recommend that you update the code you copy-pasted from
> the
> > mempool library to your PMDs, so they reflect the new and improved
> > mempool cache behavior [1]. When choosing to copy-paste code from a
> core
> > library, you should feel obliged to keep your copied code matching
> the source
> > code you copied it from!
> >
> > Also, as reported in bug #1052, you forgot to copy-paste the
> instrumentation,
> > thereby 1. making the mempool debug statistics invalid and 2.
> omitting the
> > mempool accesses from the trace when using your PMDs. :-(
> We are working on mempool APIs to expose the per core cache memory to
> PMD so that the buffers can be copied directly. We are planning to fix
> this duplication as part of that.

Is the copy-paste bug fix going to make it for 22.11?

Otherwise, these PMDs are managing the mempool cache differently than the 
mempool library does. (And the mempool library instrumentation will remain 
partially bypassed for these PMDs.) This should be mentioned as a know bug in 
the release notes.

> 
> >
> > Alternatively, just remove the copy-pasted code and use the mempool
> > library's API instead. ;-)
> >
> > The direct re-arm code also contains copy-pasted mempool cache
> handling
> > code - which was accepted with the argument that the same code was
> > already copy-pasted elsewhere. I don't know if the direct re-arm code
> also
> > needs updating... Authors of that patch (CC to this email), please
> coordinate
> > with the PMD maintainers.
> Direct-rearm patch is not accepted yet.
> 
> >
> > PS:  As noted in the 22.11-rc1 release notes, more changes to the
> mempool
> > library [2] may be coming.
> >
> > [1]:
> > https://patches.dpdk.org/project/dpdk/patch/20221007104450.2567961-1-
> > andrew.rybche...@oktetlabs.ru/
> >
> > [2]: https://patches.dpdk.org/project/dpdk/list/?series=25063
> >
> > -Morten
> 



Re: [PATCH] MAINTAINERS: remove sthemmin@microsoft

2022-10-30 Thread Thomas Monjalon
28/10/2022 17:40, Stephen Hemminger:
> These are my last couple of days at Microsoft.
> Remove the old email from MAINTAINERS.
> Will no longer have free access to Azure to work on Netvsc.
> 
> Signed-off-by: Stephen Hemminger 

Applied, looks like the networkplumber is free :)




Re: [PATCH] maintainers: change maintainer for event ethdev Rx/Tx adapters

2022-10-30 Thread Thomas Monjalon
21/10/2022 13:35, Jay Jayatheerthan:
> Harish is the new maintainer of Rx/Tx adapters due to role change of Jay
> 
> Signed-off-by: Jay Jayatheerthan 

Please could we have an approval from the new maintainer?
An ack would make things clear and accepted.





RE: [PATCH v4 2/2] mempool: optimized debug statistics

2022-10-30 Thread Morten Brørup
> From: Morten Brørup [mailto:m...@smartsharesystems.com]
> Sent: Friday, 28 October 2022 08.42
> 
> When built with debug enabled (RTE_LIBRTE_MEMPOOL_DEBUG defined), the
> performance of mempools with caches is improved as follows.
> 
> Accessing objects in the mempool is likely to increment either the
> put_bulk and put_objs or the get_success_bulk and get_success_objs
> debug statistics counters.
> 
> By adding an alternative set of these counters to the mempool cache
> structure, accessing the dedicated debug statistics structure is
> avoided in
> the likely cases where these counters are incremented.
> 
> The trick here is that the cache line holding the mempool cache
> structure
> is accessed anyway, in order to update the "len" field. Updating some
> debug statistics counters in the same cache line has lower performance
> cost than accessing the debug statistics counters in the dedicated
> debug
> statistics structure, i.e. in another cache line.
> 
> Running mempool_perf_autotest on a VMware virtual server shows an avg.
> increase of 6.4 % in rate_persec for the tests with cache. (Only when
> built with debug enabled, obviously!)
> 
> For the tests without cache, the avg. increase in rate_persec is 0.8 %.
> I
> assume this is noise from the test environment.
> 
> v4:
> * Fix spelling and repeated word in commit message, caught by
> checkpatch.
> v3:
> * Try to fix git reference by making part of a series.
> * Add --in-reply-to v1 when sending email.
> v2:
> * Fix spelling and repeated word in commit message, caught by
> checkpatch.
> 
> Signed-off-by: Morten Brørup 
> ---
>  lib/mempool/rte_mempool.c |  7 +
>  lib/mempool/rte_mempool.h | 55 +++
>  2 files changed, 51 insertions(+), 11 deletions(-)
> 
> diff --git a/lib/mempool/rte_mempool.c b/lib/mempool/rte_mempool.c
> index 21c94a2b9f..7b8c00a022 100644
> --- a/lib/mempool/rte_mempool.c
> +++ b/lib/mempool/rte_mempool.c
> @@ -1285,6 +1285,13 @@ rte_mempool_dump(FILE *f, struct rte_mempool
> *mp)
>   sum.get_fail_objs += mp->stats[lcore_id].get_fail_objs;
>   sum.get_success_blks += mp-
> >stats[lcore_id].get_success_blks;
>   sum.get_fail_blks += mp->stats[lcore_id].get_fail_blks;
> + /* Add the fast access statistics, if local caches exist */
> + if (mp->cache_size != 0) {
> + sum.put_bulk += mp->local_cache[lcore_id].put_bulk;
> + sum.put_objs += mp->local_cache[lcore_id].put_objs;
> + sum.get_success_bulk += mp-
> >local_cache[lcore_id].get_success_bulk;
> + sum.get_success_objs += mp-
> >local_cache[lcore_id].get_success_objs;
> + }
>   }
>   fprintf(f, "  stats:\n");
>   fprintf(f, "put_bulk=%"PRIu64"\n", sum.put_bulk);
> diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
> index 3725a72951..d84087bc92 100644
> --- a/lib/mempool/rte_mempool.h
> +++ b/lib/mempool/rte_mempool.h
> @@ -86,6 +86,14 @@ struct rte_mempool_cache {
>   uint32_t size;/**< Size of the cache */
>   uint32_t flushthresh; /**< Threshold before we flush excess
> elements */
>   uint32_t len; /**< Current cache count */
> +#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> + uint32_t unused;
> + /* Fast access statistics, only for likely events */
> + uint64_t put_bulk; /**< Number of puts. */
> + uint64_t put_objs; /**< Number of objects
> successfully put. */
> + uint64_t get_success_bulk; /**< Successful allocation number.
> */
> + uint64_t get_success_objs; /**< Objects successfully
> allocated. */
> +#endif
>   /**
>* Cache objects
>*
> @@ -1327,13 +1335,19 @@ rte_mempool_do_generic_put(struct rte_mempool
> *mp, void * const *obj_table,
>  {
>   void **cache_objs;
> 
> + /* No cache provided */
> + if (unlikely(cache == NULL))
> + goto driver_enqueue;
> +
> +#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
>   /* increment stat now, adding in mempool always success */
> - RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
> - RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
> + cache->put_bulk += 1;
> + cache->put_objs += n;
> +#endif
> 
> - /* No cache provided or the request itself is too big for the
> cache */
> - if (unlikely(cache == NULL || n > cache->flushthresh))
> - goto driver_enqueue;
> + /* The request is too big for the cache */
> + if (unlikely(n > cache->flushthresh))
> + goto driver_enqueue_stats_incremented;
> 
>   /*
>* The cache follows the following algorithm:
> @@ -1358,6 +1372,12 @@ rte_mempool_do_generic_put(struct rte_mempool
> *mp, void * const *obj_table,
> 
>  driver_enqueue:
> 
> + /* increment stat now, adding in mempool always success */
> + RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
> + RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
> +
> +driver_enqueue_stats_incremented:
> +

Re: [PATCH v2 0/2] fix build disabling common/mlx5

2022-10-30 Thread David Marchand
On Sun, Oct 30, 2022 at 9:27 AM Thomas Monjalon  wrote:
>
> Andrew reported a build failure when disabling mlx5 common driver.
> It is a blocker for -rc2 release.
>
> While fixing the use of a variable across mlx5 drivers in first patch,
> the consistency of its use is improved in a second patch.
>
> v2: apply the same protection to other mlx5 drivers
>
> Thomas Monjalon (2):
>   common/mlx5: fix build disabling
>   common/mlx5: move Meson config initialization and check
>
>  drivers/common/mlx5/linux/meson.build   | 2 --
>  drivers/common/mlx5/meson.build | 2 ++
>  drivers/common/mlx5/windows/meson.build | 4 
>  drivers/compress/mlx5/meson.build   | 5 +
>  drivers/crypto/mlx5/meson.build | 5 +
>  drivers/net/mlx5/hws/meson.build| 4 
>  drivers/net/mlx5/meson.build| 9 ++---
>  drivers/regex/mlx5/meson.build  | 5 +
>  drivers/vdpa/mlx5/meson.build   | 5 +
>  9 files changed, 32 insertions(+), 9 deletions(-)

For the series,
Reviewed-by: David Marchand 


-- 
David Marchand



Re: [PATCH v4 2/2] mempool: optimized debug statistics

2022-10-30 Thread Thomas Monjalon
30/10/2022 10:09, Morten Brørup:
> > From: Morten Brørup [mailto:m...@smartsharesystems.com]
> > Sent: Friday, 28 October 2022 08.42
> > 
> > When built with debug enabled (RTE_LIBRTE_MEMPOOL_DEBUG defined), the
> > performance of mempools with caches is improved as follows.
> > 
> > Accessing objects in the mempool is likely to increment either the
> > put_bulk and put_objs or the get_success_bulk and get_success_objs
> > debug statistics counters.
> > 
> > By adding an alternative set of these counters to the mempool cache
> > structure, accessing the dedicated debug statistics structure is
> > avoided in
> > the likely cases where these counters are incremented.
> > 
> > The trick here is that the cache line holding the mempool cache
> > structure
> > is accessed anyway, in order to update the "len" field. Updating some
> > debug statistics counters in the same cache line has lower performance
> > cost than accessing the debug statistics counters in the dedicated
> > debug
> > statistics structure, i.e. in another cache line.
> > 
> > Running mempool_perf_autotest on a VMware virtual server shows an avg.
> > increase of 6.4 % in rate_persec for the tests with cache. (Only when
> > built with debug enabled, obviously!)
> > 
> > For the tests without cache, the avg. increase in rate_persec is 0.8 %.
> > I
> > assume this is noise from the test environment.
> > 
> > v4:
> > * Fix spelling and repeated word in commit message, caught by
> > checkpatch.
> > v3:
> > * Try to fix git reference by making part of a series.
> > * Add --in-reply-to v1 when sending email.
> > v2:
> > * Fix spelling and repeated word in commit message, caught by
> > checkpatch.
> > 
> > Signed-off-by: Morten Brørup 
> 
> I am retracting this second part of the patch series, and reopening the 
> original patch instead. This second part is probably not going to make it to 
> 22.11 anyway.

Indeed, I have decided to take patch 1 only, which is reviewed.

> Instead, I am going to provide another patch series (after 22.11) to split 
> the current RTE_LIBRTE_MEMPOOL_DEBUG define in two: RTE_LIBRTE_MEMPOOL_STATS 
> for statistics, and RTE_LIBRTE_MEMPOOL_DEBUG for debugging. And then this 
> patch can be added to the RTE_LIBRTE_MEMPOOL_STATS.





RE: [PATCH v2 1/2] common/mlx5: fix build disabling

2022-10-30 Thread Matan Azrad



> If the dependency common/mlx5 is explicitly disabled, but net/mlx5 is not
> explicitly disabled, Meson will read the full recipe of net/mlx5 and will fail
> when accessing a variable from common/mlx5:
> drivers/net/mlx5/meson.build:76:4: ERROR: Unknown variable
> "mlx5_config".
> 
> The solution is to stop parsing net/mlx5 if common/mlx5 is disabled.
> The deps array must be defined before stopping, in order to automatically
> disable the build of net/mlx5 and print the reason.
> 
> The same protection is applied to other mlx5 drivers, so it will allow using 
> the
> variable mlx5_config in future.
> 
> Fixes: 22681deead3e ("net/mlx5/hws: enable hardware steering")
> 
> Reported-by: Andrew Rybchenko 
> Signed-off-by: Thomas Monjalon 
> Tested-by: Andrew Rybchenko 
Acked-by: Matan Azrad 


RE: [PATCH v2 2/2] common/mlx5: move Meson config initialization and check

2022-10-30 Thread Matan Azrad


> The variable mlx5_config may be used by other mlx5 drivers and should be
> always initialized.
> By moving its initialization (with configuration file generation), it is made
> consistent for Linux and Windows builds.
> 
> And the check of mlx5_config in net/mlx5 is moved at the top of
> net/mlx5/hws/meson.build so HWS requirements are in the right context.
> 
> Signed-off-by: Thomas Monjalon 
> Tested-by: Andrew Rybchenko 
Acked-by: Matan Azrad 


Re: [PATCH v4 1/2] mempool: cache align mempool cache objects

2022-10-30 Thread Thomas Monjalon
28/10/2022 08:41, Morten Brørup:
> Add __rte_cache_aligned to the objs array.
> 
> It makes no difference in the general case, but if get/put operations are
> always 32 objects, it will reduce the number of memory (or last level
> cache) accesses from five to four 64 B cache lines for every get/put
> operation.
> 
> For readability reasons, an example using 16 objects follows:
> 
> Currently, with 16 objects (128B), we access to 3
> cache lines:
> 
>   ┌┐
>   │len │
> cache ││---
> line0 ││ ^
>   ││ |
>   ├┤ | 16 objects
>   ││ | 128B
> cache ││ |
> line1 ││ |
>   ││ |
>   ├┤ |
>   ││_v_
> cache ││
> line2 ││
>   ││
>   └┘
> 
> With the alignment, it is also 3 cache lines:
> 
>   ┌┐
>   │len │
> cache ││
> line0 ││
>   ││
>   ├┤---
>   ││ ^
> cache ││ |
> line1 ││ |
>   ││ |
>   ├┤ | 16 objects
>   ││ | 128B
> cache ││ |
> line2 ││ |
>   ││ v
>   └┘---
> 
> However, accessing the objects at the bottom of the mempool cache is a
> special case, where cache line0 is also used for objects.
> 
> Consider the next burst (and any following bursts):
> 
> Current:
>   ┌┐
>   │len │
> cache ││
> line0 ││
>   ││
>   ├┤
>   ││
> cache ││
> line1 ││
>   ││
>   ├┤
>   ││
> cache ││---
> line2 ││ ^
>   ││ |
>   ├┤ | 16 objects
>   ││ | 128B
> cache ││ |
> line3 ││ |
>   ││ |
>   ├┤ |
>   ││_v_
> cache ││
> line4 ││
>   ││
>   └┘
> 4 cache lines touched, incl. line0 for len.
> 
> With the proposed alignment:
>   ┌┐
>   │len │
> cache ││
> line0 ││
>   ││
>   ├┤
>   ││
> cache ││
> line1 ││
>   ││
>   ├┤
>   ││
> cache ││
> line2 ││
>   ││
>   ├┤
>   ││---
> cache ││ ^
> line3 ││ |
>   ││ | 16 objects
>   ├┤ | 128B
>   ││ |
> cache ││ |
> line4 ││ |
>   ││_v_
>   └┘
> Only 3 cache lines touched, incl. line0 for len.
> 
> Credits go to Olivier Matz for the nice ASCII graphics.
> 
> v4:
> * No changes. Added reviewed- and acked-by tags.
> v3:
> * No changes. Made part of a series.
> v2:
> * No such version.
> 
> Signed-off-by: Morten Brørup 
> Reviewed-by: Andrew Rybchenko 
> Acked-by: Olivier Matz 

Applied only this first patch, thanks.
The second patch needs more time.

[Bug 1089] mlx5 Windows crash creating flows in multiple threads

2022-10-30 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1089

Tal Shnaiderman (tal...@nvidia.com) changed:

   What|Removed |Added

 Status|UNCONFIRMED |RESOLVED
 Resolution|--- |FIXED

--- Comment #2 from Tal Shnaiderman (tal...@nvidia.com) ---
Resolved in
http://git.dpdk.org/dpdk/commit/?id=5976328d91c3616b1ad841a9181e1da23a2980bf

-- 
You are receiving this mail because:
You are the assignee for the bug.

[PATCH v3 0/2] fix build disabling common/mlx5

2022-10-30 Thread Thomas Monjalon
Andrew reported a build failure when disabling mlx5 common driver.
It is a blocker for -rc2 release.

While fixing the use of a variable across mlx5 drivers in first patch,
the consistency of its use is improved in a second patch.

v2: apply the same protection to other mlx5 drivers
v3: fix ibverbs_link=dlopen (include directory was missing)

Thomas Monjalon (2):
  common/mlx5: fix build disabling
  common/mlx5: move build config initialization and check

 drivers/common/mlx5/linux/meson.build   | 3 +--
 drivers/common/mlx5/meson.build | 2 ++
 drivers/common/mlx5/windows/meson.build | 4 
 drivers/compress/mlx5/meson.build   | 5 +
 drivers/crypto/mlx5/meson.build | 5 +
 drivers/net/mlx5/hws/meson.build| 4 
 drivers/net/mlx5/meson.build| 9 ++---
 drivers/regex/mlx5/meson.build  | 5 +
 drivers/vdpa/mlx5/meson.build   | 5 +
 9 files changed, 33 insertions(+), 9 deletions(-)

-- 
2.36.1



[PATCH v3 1/2] common/mlx5: fix build disabling

2022-10-30 Thread Thomas Monjalon
If the dependency common/mlx5 is explicitly disabled,
but net/mlx5 is not explicitly disabled,
Meson will read the full recipe of net/mlx5
and will fail when accessing a variable from common/mlx5:
drivers/net/mlx5/meson.build:76:4: ERROR: Unknown variable "mlx5_config".

The solution is to stop parsing net/mlx5 if common/mlx5 is disabled.
The deps array must be defined before stopping, in order to automatically
disable the build of net/mlx5 and print the reason.

The same protection is applied to other mlx5 drivers,
so it will allow using the variable mlx5_config in future.

Fixes: 22681deead3e ("net/mlx5/hws: enable hardware steering")

Reported-by: Andrew Rybchenko 
Signed-off-by: Thomas Monjalon 
Tested-by: Andrew Rybchenko 
Reviewed-by: David Marchand 
Acked-by: Matan Azrad 
---
 drivers/compress/mlx5/meson.build | 5 +
 drivers/crypto/mlx5/meson.build   | 5 +
 drivers/net/mlx5/meson.build  | 5 +
 drivers/regex/mlx5/meson.build| 5 +
 drivers/vdpa/mlx5/meson.build | 5 +
 5 files changed, 25 insertions(+)

diff --git a/drivers/compress/mlx5/meson.build 
b/drivers/compress/mlx5/meson.build
index 7aac329986..49ce3aff46 100644
--- a/drivers/compress/mlx5/meson.build
+++ b/drivers/compress/mlx5/meson.build
@@ -9,6 +9,11 @@ endif
 
 fmt_name = 'mlx5_compress'
 deps += ['common_mlx5', 'eal', 'compressdev']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 sources = files(
 'mlx5_compress.c',
 )
diff --git a/drivers/crypto/mlx5/meson.build b/drivers/crypto/mlx5/meson.build
index 9d9c9c00bc..7521c4c671 100644
--- a/drivers/crypto/mlx5/meson.build
+++ b/drivers/crypto/mlx5/meson.build
@@ -9,6 +9,11 @@ endif
 
 fmt_name = 'mlx5_crypto'
 deps += ['common_mlx5', 'eal', 'cryptodev']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 sources = files(
 'mlx5_crypto.c',
 'mlx5_crypto_dek.c',
diff --git a/drivers/net/mlx5/meson.build b/drivers/net/mlx5/meson.build
index ff84448186..fa15158039 100644
--- a/drivers/net/mlx5/meson.build
+++ b/drivers/net/mlx5/meson.build
@@ -9,6 +9,11 @@ if not (is_linux or is_windows)
 endif
 
 deps += ['hash', 'common_mlx5']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 headers = files('rte_pmd_mlx5.h')
 sources = files(
 'mlx5.c',
diff --git a/drivers/regex/mlx5/meson.build b/drivers/regex/mlx5/meson.build
index e553dcb83d..70edc5b6da 100644
--- a/drivers/regex/mlx5/meson.build
+++ b/drivers/regex/mlx5/meson.build
@@ -8,6 +8,11 @@ if not is_linux
 endif
 
 deps += ['common_mlx5', 'eal', 'regexdev']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 sources = files(
 'mlx5_regex.c',
 'mlx5_rxp.c',
diff --git a/drivers/vdpa/mlx5/meson.build b/drivers/vdpa/mlx5/meson.build
index 9d8dbb1a82..54a4eac6f4 100644
--- a/drivers/vdpa/mlx5/meson.build
+++ b/drivers/vdpa/mlx5/meson.build
@@ -8,6 +8,11 @@ if not is_linux
 endif
 
 deps += ['hash', 'common_mlx5', 'vhost', 'pci', 'eal', 'sched']
+if not ('mlx5' in common_drivers)
+# avoid referencing undefined variables from common/mlx5
+subdir_done()
+endif
+
 sources = files(
 'mlx5_vdpa.c',
 'mlx5_vdpa_mem.c',
-- 
2.36.1



[PATCH v3 2/2] common/mlx5: move build config initialization and check

2022-10-30 Thread Thomas Monjalon
The variable mlx5_config may be used by other mlx5 drivers
and should be always initialized.
By moving its initialization (with configuration file generation),
it is made consistent for Linux and Windows builds.

And the check of mlx5_config in net/mlx5 is moved at the top of
net/mlx5/hws/meson.build so HWS requirements are in the right context.

Signed-off-by: Thomas Monjalon 
Tested-by: Andrew Rybchenko 
Reviewed-by: David Marchand 
Acked-by: Matan Azrad 
---
 drivers/common/mlx5/linux/meson.build   | 3 +--
 drivers/common/mlx5/meson.build | 2 ++
 drivers/common/mlx5/windows/meson.build | 4 
 drivers/net/mlx5/hws/meson.build| 4 
 drivers/net/mlx5/meson.build| 4 +---
 5 files changed, 8 insertions(+), 9 deletions(-)

diff --git a/drivers/common/mlx5/linux/meson.build 
b/drivers/common/mlx5/linux/meson.build
index 84e2a1ad8c..7e1575efc8 100644
--- a/drivers/common/mlx5/linux/meson.build
+++ b/drivers/common/mlx5/linux/meson.build
@@ -8,7 +8,6 @@ dlopen_ibverbs = (get_option('ibverbs_link') == 'dlopen')
 LIB_GLUE_BASE = 'librte_common_mlx5_glue.so'
 LIB_GLUE_VERSION = abi_version
 LIB_GLUE = LIB_GLUE_BASE + '.' + LIB_GLUE_VERSION
-mlx5_config = configuration_data()
 if dlopen_ibverbs
 dpdk_conf.set('RTE_IBVERBS_LINK_DLOPEN', 1)
 cflags += [
@@ -232,7 +231,6 @@ foreach arg:has_member_args
 file_prefix = '#include <' + arg[1] + '>'
 mlx5_config.set(arg[0], cc.has_member(arg[2], arg[3], prefix : 
file_prefix, dependencies: libs))
 endforeach
-configure_file(output : 'mlx5_autoconf.h', configuration : mlx5_config)
 
 # Build Glue Library
 if dlopen_ibverbs
@@ -243,6 +241,7 @@ if dlopen_ibverbs
 dlopen_install_dir = [ eal_pmd_path + '-glue' ]
 dlopen_includes = [global_inc]
 dlopen_includes += 
include_directories('../../../../lib/eal/include/generic')
+dlopen_includes += include_directories('..')
 shared_lib = shared_library(
 dlopen_lib_name,
 dlopen_sources,
diff --git a/drivers/common/mlx5/meson.build b/drivers/common/mlx5/meson.build
index 6ddbde7e8f..60ccd95cbc 100644
--- a/drivers/common/mlx5/meson.build
+++ b/drivers/common/mlx5/meson.build
@@ -37,4 +37,6 @@ else
 cflags += [ '-UPEDANTIC' ]
 endif
 
+mlx5_config = configuration_data()
 subdir(exec_env)
+configure_file(output: 'mlx5_autoconf.h', configuration: mlx5_config)
diff --git a/drivers/common/mlx5/windows/meson.build 
b/drivers/common/mlx5/windows/meson.build
index edbbaa9ae1..cc486014a8 100644
--- a/drivers/common/mlx5/windows/meson.build
+++ b/drivers/common/mlx5/windows/meson.build
@@ -39,7 +39,3 @@ if get_option('buildtype').contains('debug')
 else
 cflags += [ '-UPEDANTIC' ]
 endif
-
-# Generate an empty mlx5_autoconf.h file for compatibility with Linux
-config = configuration_data()
-configure_file(output : 'mlx5_autoconf.h', configuration : config)
diff --git a/drivers/net/mlx5/hws/meson.build b/drivers/net/mlx5/hws/meson.build
index d2bb864fd2..38776d5163 100644
--- a/drivers/net/mlx5/hws/meson.build
+++ b/drivers/net/mlx5/hws/meson.build
@@ -1,6 +1,10 @@
 # SPDX-License-Identifier: BSD-3-Clause
 # Copyright (c) 2022 NVIDIA Corporation & Affiliates
 
+if not (is_linux and mlx5_config.get('HAVE_IBV_FLOW_DV_SUPPORT', false))
+subdir_done()
+endif
+
 includes += include_directories('.')
 sources += files(
 'mlx5dr_context.c',
diff --git a/drivers/net/mlx5/meson.build b/drivers/net/mlx5/meson.build
index fa15158039..f1aab18f82 100644
--- a/drivers/net/mlx5/meson.build
+++ b/drivers/net/mlx5/meson.build
@@ -78,6 +78,4 @@ testpmd_sources += files('mlx5_testpmd.c')
 
 subdir(exec_env)
 
-if (is_linux and mlx5_config.get('HAVE_IBV_FLOW_DV_SUPPORT', false))
-subdir('hws')
-endif
+subdir('hws')
-- 
2.36.1



[PATCH] mempool: split statistics from debug

2022-10-30 Thread Morten Brørup
Split statistics from debug, to make mempool statistics available without
the performance cost of continuously validating the cookies in the mempool
elements.

Signed-off-by: Morten Brørup 
---
 config/rte_config.h   |  2 ++
 lib/mempool/rte_mempool.c | 10 +-
 lib/mempool/rte_mempool.h | 14 +++---
 3 files changed, 14 insertions(+), 12 deletions(-)

diff --git a/config/rte_config.h b/config/rte_config.h
index ae56a86394..1b12d30557 100644
--- a/config/rte_config.h
+++ b/config/rte_config.h
@@ -47,6 +47,8 @@
 
 /* mempool defines */
 #define RTE_MEMPOOL_CACHE_MAX_SIZE 512
+// RTE_LIBRTE_MEMPOOL_STATS is not set
+// RTE_LIBRTE_MEMPOOL_DEBUG is not set
 
 /* mbuf defines */
 #define RTE_MBUF_DEFAULT_MEMPOOL_OPS "ring_mp_mc"
diff --git a/lib/mempool/rte_mempool.c b/lib/mempool/rte_mempool.c
index 21c94a2b9f..f180257a54 100644
--- a/lib/mempool/rte_mempool.c
+++ b/lib/mempool/rte_mempool.c
@@ -818,8 +818,8 @@ rte_mempool_create_empty(const char *name, unsigned n, 
unsigned elt_size,
  RTE_CACHE_LINE_MASK) != 0);
RTE_BUILD_BUG_ON((sizeof(struct rte_mempool_cache) &
  RTE_CACHE_LINE_MASK) != 0);
-#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
-   RTE_BUILD_BUG_ON((sizeof(struct rte_mempool_debug_stats) &
+#ifdef RTE_LIBRTE_MEMPOOL_STATS
+   RTE_BUILD_BUG_ON((sizeof(struct rte_mempool_stats) &
  RTE_CACHE_LINE_MASK) != 0);
RTE_BUILD_BUG_ON((offsetof(struct rte_mempool, stats) &
  RTE_CACHE_LINE_MASK) != 0);
@@ -1221,9 +1221,9 @@ rte_mempool_audit(struct rte_mempool *mp)
 void
 rte_mempool_dump(FILE *f, struct rte_mempool *mp)
 {
-#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+#ifdef RTE_LIBRTE_MEMPOOL_STATS
struct rte_mempool_info info;
-   struct rte_mempool_debug_stats sum;
+   struct rte_mempool_stats sum;
unsigned lcore_id;
 #endif
struct rte_mempool_memhdr *memhdr;
@@ -1269,7 +1269,7 @@ rte_mempool_dump(FILE *f, struct rte_mempool *mp)
fprintf(f, "  common_pool_count=%u\n", common_count);
 
/* sum and dump statistics */
-#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+#ifdef RTE_LIBRTE_MEMPOOL_STATS
rte_mempool_ops_get_info(mp, &info);
memset(&sum, 0, sizeof(sum));
for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 3725a72951..89640e48e5 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -56,14 +56,14 @@ extern "C" {
 #define RTE_MEMPOOL_HEADER_COOKIE2  0xf2eef2eedadd2e55ULL /**< Header cookie. 
*/
 #define RTE_MEMPOOL_TRAILER_COOKIE  0xadd2e55badbadbadULL /**< Trailer 
cookie.*/
 
-#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+#ifdef RTE_LIBRTE_MEMPOOL_STATS
 /**
  * A structure that stores the mempool statistics (per-lcore).
  * Note: Cache stats (put_cache_bulk/objs, get_cache_bulk/objs) are not
  * captured since they can be calculated from other stats.
  * For example: put_cache_objs = put_objs - put_common_pool_objs.
  */
-struct rte_mempool_debug_stats {
+struct rte_mempool_stats {
uint64_t put_bulk; /**< Number of puts. */
uint64_t put_objs; /**< Number of objects successfully put. 
*/
uint64_t put_common_pool_bulk; /**< Number of bulks enqueued in common 
pool. */
@@ -237,9 +237,9 @@ struct rte_mempool {
uint32_t nb_mem_chunks;  /**< Number of memory chunks */
struct rte_mempool_memhdr_list mem_list; /**< List of memory chunks */
 
-#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+#ifdef RTE_LIBRTE_MEMPOOL_STATS
/** Per-lcore statistics. */
-   struct rte_mempool_debug_stats stats[RTE_MAX_LCORE];
+   struct rte_mempool_stats stats[RTE_MAX_LCORE];
 #endif
 }  __rte_cache_aligned;
 
@@ -293,16 +293,16 @@ struct rte_mempool {
| RTE_MEMPOOL_F_NO_IOVA_CONTIG \
)
 /**
- * @internal When debug is enabled, store some statistics.
+ * @internal When stats are enabled, store some statistics.
  *
  * @param mp
  *   Pointer to the memory pool.
  * @param name
  *   Name of the statistics field to increment in the memory pool.
  * @param n
- *   Number to add to the object-oriented statistics.
+ *   Number to add to the statistics.
  */
-#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+#ifdef RTE_LIBRTE_MEMPOOL_STATS
 #define RTE_MEMPOOL_STAT_ADD(mp, name, n) do {  \
unsigned __lcore_id = rte_lcore_id();   \
if (__lcore_id < RTE_MAX_LCORE) {   \
-- 
2.17.1



RE: [PATCH v3 1/2] common/mlx5: fix build disabling

2022-10-30 Thread Alex Vesker



> -Original Message-
> From: Thomas Monjalon 
> Sent: Sunday, October 30, 2022 1:08 PM
> To: dev@dpdk.org
> Cc: Alex Vesker ; ferruh.yi...@amd.com;
> andrew.rybche...@oktetlabs.ru; Raslan Darawsheh ;
> david.march...@redhat.com; Matan Azrad ; Fan
> Zhang ; Ashish Gupta
> ; Slava Ovsiienko ;
> Ori Kam 
> Subject: [PATCH v3 1/2] common/mlx5: fix build disabling
> 
> If the dependency common/mlx5 is explicitly disabled, but net/mlx5 is not
> explicitly disabled, Meson will read the full recipe of net/mlx5 and will fail
> when accessing a variable from common/mlx5:
> drivers/net/mlx5/meson.build:76:4: ERROR: Unknown variable
> "mlx5_config".
> 
> The solution is to stop parsing net/mlx5 if common/mlx5 is disabled.
> The deps array must be defined before stopping, in order to automatically
> disable the build of net/mlx5 and print the reason.
> 
> The same protection is applied to other mlx5 drivers, so it will allow using 
> the
> variable mlx5_config in future.
> 
> Fixes: 22681deead3e ("net/mlx5/hws: enable hardware steering")
> 
> Reported-by: Andrew Rybchenko 
> Signed-off-by: Thomas Monjalon 
> Tested-by: Andrew Rybchenko 
> Reviewed-by: David Marchand 
> Acked-by: Matan Azrad 
> ---
>  drivers/compress/mlx5/meson.build | 5 +
>  drivers/crypto/mlx5/meson.build   | 5 +
>  drivers/net/mlx5/meson.build  | 5 +
>  drivers/regex/mlx5/meson.build| 5 +
>  drivers/vdpa/mlx5/meson.build | 5 +
>  5 files changed, 25 insertions(+)
> 
> diff --git a/drivers/compress/mlx5/meson.build
> b/drivers/compress/mlx5/meson.build
> index 7aac329986..49ce3aff46 100644
> --- a/drivers/compress/mlx5/meson.build
> +++ b/drivers/compress/mlx5/meson.build
> @@ -9,6 +9,11 @@ endif
> 
>  fmt_name = 'mlx5_compress'
>  deps += ['common_mlx5', 'eal', 'compressdev']
> +if not ('mlx5' in common_drivers)
> +# avoid referencing undefined variables from common/mlx5
> +subdir_done()
> +endif
> +
>  sources = files(
>  'mlx5_compress.c',
>  )
> diff --git a/drivers/crypto/mlx5/meson.build
> b/drivers/crypto/mlx5/meson.build index 9d9c9c00bc..7521c4c671 100644
> --- a/drivers/crypto/mlx5/meson.build
> +++ b/drivers/crypto/mlx5/meson.build
> @@ -9,6 +9,11 @@ endif
> 
>  fmt_name = 'mlx5_crypto'
>  deps += ['common_mlx5', 'eal', 'cryptodev']
> +if not ('mlx5' in common_drivers)
> +# avoid referencing undefined variables from common/mlx5
> +subdir_done()
> +endif
> +
>  sources = files(
>  'mlx5_crypto.c',
>  'mlx5_crypto_dek.c',
> diff --git a/drivers/net/mlx5/meson.build b/drivers/net/mlx5/meson.build
> index ff84448186..fa15158039 100644
> --- a/drivers/net/mlx5/meson.build
> +++ b/drivers/net/mlx5/meson.build
> @@ -9,6 +9,11 @@ if not (is_linux or is_windows)  endif
> 
>  deps += ['hash', 'common_mlx5']
> +if not ('mlx5' in common_drivers)
> +# avoid referencing undefined variables from common/mlx5
> +subdir_done()
> +endif
> +
>  headers = files('rte_pmd_mlx5.h')
>  sources = files(
>  'mlx5.c',
> diff --git a/drivers/regex/mlx5/meson.build
> b/drivers/regex/mlx5/meson.build index e553dcb83d..70edc5b6da 100644
> --- a/drivers/regex/mlx5/meson.build
> +++ b/drivers/regex/mlx5/meson.build
> @@ -8,6 +8,11 @@ if not is_linux
>  endif
> 
>  deps += ['common_mlx5', 'eal', 'regexdev']
> +if not ('mlx5' in common_drivers)
> +# avoid referencing undefined variables from common/mlx5
> +subdir_done()
> +endif
> +
>  sources = files(
>  'mlx5_regex.c',
>  'mlx5_rxp.c',
> diff --git a/drivers/vdpa/mlx5/meson.build
> b/drivers/vdpa/mlx5/meson.build index 9d8dbb1a82..54a4eac6f4 100644
> --- a/drivers/vdpa/mlx5/meson.build
> +++ b/drivers/vdpa/mlx5/meson.build
> @@ -8,6 +8,11 @@ if not is_linux
>  endif
> 
>  deps += ['hash', 'common_mlx5', 'vhost', 'pci', 'eal', 'sched']
> +if not ('mlx5' in common_drivers)
> +# avoid referencing undefined variables from common/mlx5
> +subdir_done()
> +endif
> +
>  sources = files(
>  'mlx5_vdpa.c',
>  'mlx5_vdpa_mem.c',
> --
> 2.36.1

Acked-by: Alex Vesker 


RE: [PATCH v3 2/2] common/mlx5: move build config initialization and check

2022-10-30 Thread Alex Vesker



> -Original Message-
> From: Thomas Monjalon 
> Sent: Sunday, October 30, 2022 1:08 PM
> To: dev@dpdk.org
> Cc: Alex Vesker ; ferruh.yi...@amd.com;
> andrew.rybche...@oktetlabs.ru; Raslan Darawsheh ;
> david.march...@redhat.com; Matan Azrad ; Slava
> Ovsiienko 
> Subject: [PATCH v3 2/2] common/mlx5: move build config initialization and
> check
> 
> The variable mlx5_config may be used by other mlx5 drivers and should be
> always initialized.
> By moving its initialization (with configuration file generation), it is made
> consistent for Linux and Windows builds.
> 
> And the check of mlx5_config in net/mlx5 is moved at the top of
> net/mlx5/hws/meson.build so HWS requirements are in the right context.
> 
> Signed-off-by: Thomas Monjalon 
> Tested-by: Andrew Rybchenko 
> Reviewed-by: David Marchand 
> Acked-by: Matan Azrad 
> ---
>  drivers/common/mlx5/linux/meson.build   | 3 +--
>  drivers/common/mlx5/meson.build | 2 ++
>  drivers/common/mlx5/windows/meson.build | 4 
>  drivers/net/mlx5/hws/meson.build| 4 
>  drivers/net/mlx5/meson.build| 4 +---
>  5 files changed, 8 insertions(+), 9 deletions(-)
> 
> diff --git a/drivers/common/mlx5/linux/meson.build
> b/drivers/common/mlx5/linux/meson.build
> index 84e2a1ad8c..7e1575efc8 100644
> --- a/drivers/common/mlx5/linux/meson.build
> +++ b/drivers/common/mlx5/linux/meson.build
> @@ -8,7 +8,6 @@ dlopen_ibverbs = (get_option('ibverbs_link') == 'dlopen')
> LIB_GLUE_BASE = 'librte_common_mlx5_glue.so'
>  LIB_GLUE_VERSION = abi_version
>  LIB_GLUE = LIB_GLUE_BASE + '.' + LIB_GLUE_VERSION -mlx5_config =
> configuration_data()  if dlopen_ibverbs
>  dpdk_conf.set('RTE_IBVERBS_LINK_DLOPEN', 1)
>  cflags += [
> @@ -232,7 +231,6 @@ foreach arg:has_member_args
>  file_prefix = '#include <' + arg[1] + '>'
>  mlx5_config.set(arg[0], cc.has_member(arg[2], arg[3], prefix : 
> file_prefix,
> dependencies: libs))  endforeach -configure_file(output : 'mlx5_autoconf.h',
> configuration : mlx5_config)
> 
>  # Build Glue Library
>  if dlopen_ibverbs
> @@ -243,6 +241,7 @@ if dlopen_ibverbs
>  dlopen_install_dir = [ eal_pmd_path + '-glue' ]
>  dlopen_includes = [global_inc]
>  dlopen_includes += 
> include_directories('../../../../lib/eal/include/generic')
> +dlopen_includes += include_directories('..')
>  shared_lib = shared_library(
>  dlopen_lib_name,
>  dlopen_sources,
> diff --git a/drivers/common/mlx5/meson.build
> b/drivers/common/mlx5/meson.build index 6ddbde7e8f..60ccd95cbc
> 100644
> --- a/drivers/common/mlx5/meson.build
> +++ b/drivers/common/mlx5/meson.build
> @@ -37,4 +37,6 @@ else
>  cflags += [ '-UPEDANTIC' ]
>  endif
> 
> +mlx5_config = configuration_data()
>  subdir(exec_env)
> +configure_file(output: 'mlx5_autoconf.h', configuration: mlx5_config)
> diff --git a/drivers/common/mlx5/windows/meson.build
> b/drivers/common/mlx5/windows/meson.build
> index edbbaa9ae1..cc486014a8 100644
> --- a/drivers/common/mlx5/windows/meson.build
> +++ b/drivers/common/mlx5/windows/meson.build
> @@ -39,7 +39,3 @@ if get_option('buildtype').contains('debug')
>  else
>  cflags += [ '-UPEDANTIC' ]
>  endif
> -
> -# Generate an empty mlx5_autoconf.h file for compatibility with Linux -
> config = configuration_data() -configure_file(output : 'mlx5_autoconf.h',
> configuration : config) diff --git a/drivers/net/mlx5/hws/meson.build
> b/drivers/net/mlx5/hws/meson.build
> index d2bb864fd2..38776d5163 100644
> --- a/drivers/net/mlx5/hws/meson.build
> +++ b/drivers/net/mlx5/hws/meson.build
> @@ -1,6 +1,10 @@
>  # SPDX-License-Identifier: BSD-3-Clause  # Copyright (c) 2022 NVIDIA
> Corporation & Affiliates
> 
> +if not (is_linux and mlx5_config.get('HAVE_IBV_FLOW_DV_SUPPORT', false))
> +subdir_done()
> +endif
> +
>  includes += include_directories('.')
>  sources += files(
>  'mlx5dr_context.c',
> diff --git a/drivers/net/mlx5/meson.build b/drivers/net/mlx5/meson.build
> index fa15158039..f1aab18f82 100644
> --- a/drivers/net/mlx5/meson.build
> +++ b/drivers/net/mlx5/meson.build
> @@ -78,6 +78,4 @@ testpmd_sources += files('mlx5_testpmd.c')
> 
>  subdir(exec_env)
> 
> -if (is_linux and mlx5_config.get('HAVE_IBV_FLOW_DV_SUPPORT', false))
> -subdir('hws')
> -endif
> +subdir('hws')
> --
> 2.36.1

Acked-by: Alex Vesker 


RE: [PATCH] mempool: split statistics from debug

2022-10-30 Thread Morten Brørup
> From: Morten Brørup [mailto:m...@smartsharesystems.com]
> Sent: Sunday, 30 October 2022 12.55
> 
> Split statistics from debug, to make mempool statistics available
> without
> the performance cost of continuously validating the cookies in the
> mempool
> elements.

mempool_perf_autotest shows that the rate_persec drops to a third (-66 %) when 
enabling full mempool debug - quite prohibitive!

With this patch, the performance cost is much lower. I don't have detailed test 
results, but the initial tests without mempool cache show a performance drop of 
-11 %. Significant, but not prohibitive.

> 
> Signed-off-by: Morten Brørup 
> ---
>  config/rte_config.h   |  2 ++
>  lib/mempool/rte_mempool.c | 10 +-
>  lib/mempool/rte_mempool.h | 14 +++---
>  3 files changed, 14 insertions(+), 12 deletions(-)
> 
> diff --git a/config/rte_config.h b/config/rte_config.h
> index ae56a86394..1b12d30557 100644
> --- a/config/rte_config.h
> +++ b/config/rte_config.h
> @@ -47,6 +47,8 @@
> 
>  /* mempool defines */
>  #define RTE_MEMPOOL_CACHE_MAX_SIZE 512
> +// RTE_LIBRTE_MEMPOOL_STATS is not set
> +// RTE_LIBRTE_MEMPOOL_DEBUG is not set
> 
>  /* mbuf defines */
>  #define RTE_MBUF_DEFAULT_MEMPOOL_OPS "ring_mp_mc"
> diff --git a/lib/mempool/rte_mempool.c b/lib/mempool/rte_mempool.c
> index 21c94a2b9f..f180257a54 100644
> --- a/lib/mempool/rte_mempool.c
> +++ b/lib/mempool/rte_mempool.c
> @@ -818,8 +818,8 @@ rte_mempool_create_empty(const char *name, unsigned
> n, unsigned elt_size,
> RTE_CACHE_LINE_MASK) != 0);
>   RTE_BUILD_BUG_ON((sizeof(struct rte_mempool_cache) &
> RTE_CACHE_LINE_MASK) != 0);
> -#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> - RTE_BUILD_BUG_ON((sizeof(struct rte_mempool_debug_stats) &
> +#ifdef RTE_LIBRTE_MEMPOOL_STATS
> + RTE_BUILD_BUG_ON((sizeof(struct rte_mempool_stats) &
> RTE_CACHE_LINE_MASK) != 0);
>   RTE_BUILD_BUG_ON((offsetof(struct rte_mempool, stats) &
> RTE_CACHE_LINE_MASK) != 0);
> @@ -1221,9 +1221,9 @@ rte_mempool_audit(struct rte_mempool *mp)
>  void
>  rte_mempool_dump(FILE *f, struct rte_mempool *mp)
>  {
> -#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> +#ifdef RTE_LIBRTE_MEMPOOL_STATS
>   struct rte_mempool_info info;
> - struct rte_mempool_debug_stats sum;
> + struct rte_mempool_stats sum;
>   unsigned lcore_id;
>  #endif
>   struct rte_mempool_memhdr *memhdr;
> @@ -1269,7 +1269,7 @@ rte_mempool_dump(FILE *f, struct rte_mempool *mp)
>   fprintf(f, "  common_pool_count=%u\n", common_count);
> 
>   /* sum and dump statistics */
> -#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> +#ifdef RTE_LIBRTE_MEMPOOL_STATS
>   rte_mempool_ops_get_info(mp, &info);
>   memset(&sum, 0, sizeof(sum));
>   for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) {
> diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
> index 3725a72951..89640e48e5 100644
> --- a/lib/mempool/rte_mempool.h
> +++ b/lib/mempool/rte_mempool.h
> @@ -56,14 +56,14 @@ extern "C" {
>  #define RTE_MEMPOOL_HEADER_COOKIE2  0xf2eef2eedadd2e55ULL /**< Header
> cookie. */
>  #define RTE_MEMPOOL_TRAILER_COOKIE  0xadd2e55badbadbadULL /**< Trailer
> cookie.*/
> 
> -#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> +#ifdef RTE_LIBRTE_MEMPOOL_STATS
>  /**
>   * A structure that stores the mempool statistics (per-lcore).
>   * Note: Cache stats (put_cache_bulk/objs, get_cache_bulk/objs) are
> not
>   * captured since they can be calculated from other stats.
>   * For example: put_cache_objs = put_objs - put_common_pool_objs.
>   */
> -struct rte_mempool_debug_stats {
> +struct rte_mempool_stats {
>   uint64_t put_bulk; /**< Number of puts. */
>   uint64_t put_objs; /**< Number of objects
> successfully put. */
>   uint64_t put_common_pool_bulk; /**< Number of bulks enqueued in
> common pool. */
> @@ -237,9 +237,9 @@ struct rte_mempool {
>   uint32_t nb_mem_chunks;  /**< Number of memory chunks */
>   struct rte_mempool_memhdr_list mem_list; /**< List of memory
> chunks */
> 
> -#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> +#ifdef RTE_LIBRTE_MEMPOOL_STATS
>   /** Per-lcore statistics. */
> - struct rte_mempool_debug_stats stats[RTE_MAX_LCORE];
> + struct rte_mempool_stats stats[RTE_MAX_LCORE];
>  #endif
>  }  __rte_cache_aligned;
> 
> @@ -293,16 +293,16 @@ struct rte_mempool {
>   | RTE_MEMPOOL_F_NO_IOVA_CONTIG \
>   )
>  /**
> - * @internal When debug is enabled, store some statistics.
> + * @internal When stats are enabled, store some statistics.
>   *
>   * @param mp
>   *   Pointer to the memory pool.
>   * @param name
>   *   Name of the statistics field to increment in the memory pool.
>   * @param n
> - *   Number to add to the object-oriented statistics.
> + *   Number to add to the statistics.
>   */
> -#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> +#ifdef RTE_LIBRTE_MEMPOOL_STATS
>  #define RTE_MEMPOOL_STAT_ADD(mp, name, n) do {  \
> 

Re: [PATCH v3 0/2] fix build disabling common/mlx5

2022-10-30 Thread Thomas Monjalon
30/10/2022 12:08, Thomas Monjalon:
> Andrew reported a build failure when disabling mlx5 common driver.
> It is a blocker for -rc2 release.
> 
> While fixing the use of a variable across mlx5 drivers in first patch,
> the consistency of its use is improved in a second patch.
> 
> v2: apply the same protection to other mlx5 drivers
> v3: fix ibverbs_link=dlopen (include directory was missing)
> 
> Thomas Monjalon (2):
>   common/mlx5: fix build disabling
>   common/mlx5: move build config initialization and check

Applied

There is a CI error with Fedora 35 but it seems unrelated.




Re: [PATCH v12 04/16] baseband/acc: introduce PMD for ACC200

2022-10-30 Thread Thomas Monjalon
12/10/2022 19:59, Nicolas Chautru:
> +Bind PF UIO driver(s)
> +~
> +
> +Install the DPDK igb_uio driver, bind it with the PF PCI device ID and use
> +``lspci`` to confirm the PF device is under use by ``igb_uio`` DPDK UIO 
> driver.

igb_uio is not recommended.
Please focus on VFIO first.

> +The igb_uio driver may be bound to the PF PCI device using one of two methods
> +for ACC200:
> +
> +
> +1. PCI functions (physical or virtual, depending on the use case) can be 
> bound
> +to the UIO driver by repeating this command for every function.
> +
> +.. code-block:: console
> +
> +  cd 
> +  insmod ./build/kmod/igb_uio.ko
> +  echo "8086 57c0" > /sys/bus/pci/drivers/igb_uio/new_id
> +  lspci -vd8086:57c0
> +
> +
> +2. Another way to bind PF with DPDK UIO driver is by using the 
> ``dpdk-devbind.py`` tool
> +
> +.. code-block:: console
> +
> +  cd 
> +  ./usertools/dpdk-devbind.py -b igb_uio :f7:00.0
> +
> +where the PCI device ID (example: :f7:00.0) is obtained using lspci 
> -vd8086:57c0

This binding is not specific to the driver.
It would be better to refer to the Linux guide
instead of duplicating it again and again.

> +In a similar way the PF may be bound with vfio-pci as any PCIe device.

You could mention igb_uio here.
Is there any advantage in using igb_uio?




[PATCH] net/mlx5/hws: remove deprecated rte_atomic

2022-10-30 Thread Alex Vesker
The use of rte_atomic functions is deprecated and is not
required in HWS code. HWS refcounts are used only during
control and always under lock.

Fixes: f8c8a6d8440d ("net/mlx5/hws: add action object")
Signed-off-by: Alex Vesker 
---
 drivers/net/mlx5/hws/mlx5dr_action.c  | 8 +++-
 drivers/net/mlx5/hws/mlx5dr_action.h  | 2 +-
 drivers/net/mlx5/hws/mlx5dr_pat_arg.c | 8 +++-
 drivers/net/mlx5/hws/mlx5dr_pat_arg.h | 2 +-
 4 files changed, 8 insertions(+), 12 deletions(-)

diff --git a/drivers/net/mlx5/hws/mlx5dr_action.c 
b/drivers/net/mlx5/hws/mlx5dr_action.c
index 755d5d09cf..a9e12aa1f5 100644
--- a/drivers/net/mlx5/hws/mlx5dr_action.c
+++ b/drivers/net/mlx5/hws/mlx5dr_action.c
@@ -83,7 +83,7 @@ static int mlx5dr_action_get_shared_stc_nic(struct 
mlx5dr_context *ctx,
 
pthread_spin_lock(&ctx->ctrl_lock);
if (ctx->common_res[tbl_type].shared_stc[stc_type]) {
-   
rte_atomic32_add(&ctx->common_res[tbl_type].shared_stc[stc_type]->refcount, 1);
+   ctx->common_res[tbl_type].shared_stc[stc_type]->refcount++;
pthread_spin_unlock(&ctx->ctrl_lock);
return 0;
}
@@ -123,9 +123,7 @@ static int mlx5dr_action_get_shared_stc_nic(struct 
mlx5dr_context *ctx,
}
 
ctx->common_res[tbl_type].shared_stc[stc_type] = shared_stc;
-
-   
rte_atomic32_init(&ctx->common_res[tbl_type].shared_stc[stc_type]->refcount);
-   
rte_atomic32_set(&ctx->common_res[tbl_type].shared_stc[stc_type]->refcount, 1);
+   ctx->common_res[tbl_type].shared_stc[stc_type]->refcount = 1;
 
pthread_spin_unlock(&ctx->ctrl_lock);
 
@@ -145,7 +143,7 @@ static void mlx5dr_action_put_shared_stc_nic(struct 
mlx5dr_context *ctx,
struct mlx5dr_action_shared_stc *shared_stc;
 
pthread_spin_lock(&ctx->ctrl_lock);
-   if 
(!rte_atomic32_dec_and_test(&ctx->common_res[tbl_type].shared_stc[stc_type]->refcount))
 {
+   if (--ctx->common_res[tbl_type].shared_stc[stc_type]->refcount) {
pthread_spin_unlock(&ctx->ctrl_lock);
return;
}
diff --git a/drivers/net/mlx5/hws/mlx5dr_action.h 
b/drivers/net/mlx5/hws/mlx5dr_action.h
index f14d91f994..3b31ffc90e 100644
--- a/drivers/net/mlx5/hws/mlx5dr_action.h
+++ b/drivers/net/mlx5/hws/mlx5dr_action.h
@@ -70,7 +70,7 @@ struct mlx5dr_action_default_stc {
 
 struct mlx5dr_action_shared_stc {
struct mlx5dr_pool_chunk remove_header;
-   rte_atomic32_t refcount;
+   uint32_t refcount;
 };
 
 struct mlx5dr_actions_apply_data {
diff --git a/drivers/net/mlx5/hws/mlx5dr_pat_arg.c 
b/drivers/net/mlx5/hws/mlx5dr_pat_arg.c
index 46fdc8ce68..df451f1ae0 100644
--- a/drivers/net/mlx5/hws/mlx5dr_pat_arg.c
+++ b/drivers/net/mlx5/hws/mlx5dr_pat_arg.c
@@ -128,7 +128,7 @@ mlx5dr_pat_get_existing_cached_pattern(struct 
mlx5dr_pattern_cache *cache,
/* LRU: move it to be first in the list */
LIST_REMOVE(cached_pattern, next);
LIST_INSERT_HEAD(&cache->head, cached_pattern, next);
-   rte_atomic32_add(&cached_pattern->refcount, 1);
+   cached_pattern->refcount++;
}
 
return cached_pattern;
@@ -179,9 +179,7 @@ mlx5dr_pat_add_pattern_to_cache(struct mlx5dr_pattern_cache 
*cache,
   num_of_actions * MLX5DR_MODIFY_ACTION_SIZE);
 
LIST_INSERT_HEAD(&cache->head, cached_pattern, next);
-
-   rte_atomic32_init(&cached_pattern->refcount);
-   rte_atomic32_set(&cached_pattern->refcount, 1);
+   cached_pattern->refcount = 1;
 
return cached_pattern;
 
@@ -212,7 +210,7 @@ mlx5dr_pat_put_pattern(struct mlx5dr_pattern_cache *cache,
goto out;
}
 
-   if (!rte_atomic32_dec_and_test(&cached_pattern->refcount))
+   if (--cached_pattern->refcount)
goto out;
 
mlx5dr_pat_remove_pattern(cached_pattern);
diff --git a/drivers/net/mlx5/hws/mlx5dr_pat_arg.h 
b/drivers/net/mlx5/hws/mlx5dr_pat_arg.h
index 8a4670427f..d9353e9a3e 100644
--- a/drivers/net/mlx5/hws/mlx5dr_pat_arg.h
+++ b/drivers/net/mlx5/hws/mlx5dr_pat_arg.h
@@ -35,7 +35,7 @@ struct mlx5dr_pat_cached_pattern {
uint8_t *data;
uint16_t num_of_actions;
} mh_data;
-   rte_atomic32_t refcount;
+   uint32_t refcount;
LIST_ENTRY(mlx5dr_pat_cached_pattern) next;
 };
 
-- 
2.18.1



Re: [PATCH] mempool: split statistics from debug

2022-10-30 Thread Stephen Hemminger
On Sun, 30 Oct 2022 15:04:18 +0100
Morten Brørup  wrote:

> > From: Morten Brørup [mailto:m...@smartsharesystems.com]
> > Sent: Sunday, 30 October 2022 12.55
> > 
> > Split statistics from debug, to make mempool statistics available
> > without
> > the performance cost of continuously validating the cookies in the
> > mempool
> > elements.  
> 
> mempool_perf_autotest shows that the rate_persec drops to a third (-66 %) 
> when enabling full mempool debug - quite prohibitive!
> 
> With this patch, the performance cost is much lower. I don't have detailed 
> test results, but the initial tests without mempool cache show a performance 
> drop of -11 %. Significant, but not prohibitive.


One trick to avoid conditional in fast path would be to add a dummy stats[] per 
core.

Another would be to move the fast path get/put stats into the mempool_cache.
The current model has stats[] per cpu always on another cache line.

diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 1f5707f46a21..87905b7286a6 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -236,8 +236,10 @@ struct rte_mempool {
struct rte_mempool_memhdr_list mem_list; /**< List of memory chunks */
 
 #ifdef RTE_LIBRTE_MEMPOOL_DEBUG
-   /** Per-lcore statistics. */
-   struct rte_mempool_debug_stats stats[RTE_MAX_LCORE];
+   /** Per-lcore statistics.
+* Allocate one additional per-cpu slot for non-DPDK threads
+*/
+   struct rte_mempool_debug_stats stats[RTE_MAX_LCORE + 1];
 #endif
 }  __rte_cache_aligned;
 
@@ -302,10 +304,7 @@ struct rte_mempool {
  */
 #ifdef RTE_LIBRTE_MEMPOOL_DEBUG
 #define RTE_MEMPOOL_STAT_ADD(mp, name, n) do {  \
-   unsigned __lcore_id = rte_lcore_id();   \
-   if (__lcore_id < RTE_MAX_LCORE) {   \
-   mp->stats[__lcore_id].name += n;\
-   }   \
+   (mp)->stats[rte_lcore_id()].name += n;
} while (0)
 #else
 #define RTE_MEMPOOL_STAT_ADD(mp, name, n) do {} while (0)


RE: [PATCH] mempool: split statistics from debug

2022-10-30 Thread Morten Brørup
> From: Stephen Hemminger [mailto:step...@networkplumber.org]
> Sent: Sunday, 30 October 2022 17.13
> 
> On Sun, 30 Oct 2022 15:04:18 +0100
> Morten Brørup  wrote:
> 
> > > From: Morten Brørup [mailto:m...@smartsharesystems.com]
> > > Sent: Sunday, 30 October 2022 12.55
> > >
> > > Split statistics from debug, to make mempool statistics available
> > > without
> > > the performance cost of continuously validating the cookies in the
> > > mempool
> > > elements.
> >
> > mempool_perf_autotest shows that the rate_persec drops to a third (-
> 66 %) when enabling full mempool debug - quite prohibitive!
> >
> > With this patch, the performance cost is much lower. I don't have
> detailed test results, but the initial tests without mempool cache show
> a performance drop of -11 %. Significant, but not prohibitive.
> 
> 
> One trick to avoid conditional in fast path would be to add a dummy
> stats[] per core.
> 
> Another would be to move the fast path get/put stats into the
> mempool_cache.
> The current model has stats[] per cpu always on another cache line.

I submitted a patch to move the likely get/put stats to the mempool cache [1], 
but retracted it shortly after. I realized that splitting the stats from the 
debug cookies had much higher performance effect, so we should do this first. 
We can move statistics counters into the mempool_cache as part two.

[1]: 
https://patchwork.dpdk.org/project/dpdk/patch/20221028064152.98341-2...@smartsharesystems.com/

Also, it might be too late for 22.11 to move stats to the mempool cache. 
However, splitting the stats from the debug cookies is extremely simple, so 
that should be accepted for 22.11 (if reviewed properly).

> 
> diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
> index 1f5707f46a21..87905b7286a6 100644
> --- a/lib/mempool/rte_mempool.h
> +++ b/lib/mempool/rte_mempool.h
> @@ -236,8 +236,10 @@ struct rte_mempool {
> struct rte_mempool_memhdr_list mem_list; /**< List of memory
> chunks */
> 
>  #ifdef RTE_LIBRTE_MEMPOOL_DEBUG
> -   /** Per-lcore statistics. */
> -   struct rte_mempool_debug_stats stats[RTE_MAX_LCORE];
> +   /** Per-lcore statistics.
> +* Allocate one additional per-cpu slot for non-DPDK threads
> +*/
> +   struct rte_mempool_debug_stats stats[RTE_MAX_LCORE + 1];

Excellent! As a bonus, this also fixes a bug in the statistics: Non-DPDK thread 
operations were not counted.

>  #endif
>  }  __rte_cache_aligned;
> 
> @@ -302,10 +304,7 @@ struct rte_mempool {
>   */
>  #ifdef RTE_LIBRTE_MEMPOOL_DEBUG
>  #define RTE_MEMPOOL_STAT_ADD(mp, name, n) do {  \
> -   unsigned __lcore_id = rte_lcore_id();   \
> -   if (__lcore_id < RTE_MAX_LCORE) {   \
> -   mp->stats[__lcore_id].name += n;\
> -   }   \
> +   (mp)->stats[rte_lcore_id()].name += n;

I suppose the index must be offset by one, for rte_lcore_id() returning 
LCORE_ID_ANY (=UINT32_MAX) to be stored at offset zero:
+   (mp)->stats[rte_lcore_id() + 1].name += n;

> } while (0)
>  #else
>  #define RTE_MEMPOOL_STAT_ADD(mp, name, n) do {} while (0)

I have marked this patch as Changes Requested, and will submit a v2 patch 
series with this improvement - but not today, considering the local time.

NB: I am aware that the loop in rte_mempool_dump() in rte_mempool.c must also 
be updated to account for the additional slot in the stats array. I'll check if 
there are similar effects elsewhere in the library.



RE: Copy-pasted code should be updated

2022-10-30 Thread Honnappa Nagarahalli

> >
> > >
> > > Dear Intel PMD maintainers (CC: techboard),
> > >
> > > I strongly recommend that you update the code you copy-pasted from
> > the
> > > mempool library to your PMDs, so they reflect the new and improved
> > > mempool cache behavior [1]. When choosing to copy-paste code from a
> > core
> > > library, you should feel obliged to keep your copied code matching
> > the source
> > > code you copied it from!
> > >
> > > Also, as reported in bug #1052, you forgot to copy-paste the
> > instrumentation,
> > > thereby 1. making the mempool debug statistics invalid and 2.
> > omitting the
> > > mempool accesses from the trace when using your PMDs. :-(
> > We are working on mempool APIs to expose the per core cache memory to
> > PMD so that the buffers can be copied directly. We are planning to fix
> > this duplication as part of that.
> 
> Is the copy-paste bug fix going to make it for 22.11?
It will not make it to 22.11. It is targeted for 23.02.

> 
> Otherwise, these PMDs are managing the mempool cache differently than
> the mempool library does. (And the mempool library instrumentation will
> remain partially bypassed for these PMDs.) This should be mentioned as a
> know bug in the release notes.
Agree

> 
> >
> > >
> > > Alternatively, just remove the copy-pasted code and use the mempool
> > > library's API instead. ;-)
> > >
> > > The direct re-arm code also contains copy-pasted mempool cache
> > handling
> > > code - which was accepted with the argument that the same code was
> > > already copy-pasted elsewhere. I don't know if the direct re-arm
> > > code
> > also
> > > needs updating... Authors of that patch (CC to this email), please
> > coordinate
> > > with the PMD maintainers.
> > Direct-rearm patch is not accepted yet.
> >
> > >
> > > PS:  As noted in the 22.11-rc1 release notes, more changes to the
> > mempool
> > > library [2] may be coming.
> > >
> > > [1]:
> > >
> https://patches.dpdk.org/project/dpdk/patch/20221007104450.2567961-1
> > > -
> > > andrew.rybche...@oktetlabs.ru/
> > >
> > > [2]: https://patches.dpdk.org/project/dpdk/list/?series=25063
> > >
> > > -Morten
> >



RE: [PATCH v15 00/18] add support for idpf PMD in DPDK

2022-10-30 Thread Xing, Beilei


> -Original Message-
> From: Andrew Rybchenko 
> Sent: Saturday, October 29, 2022 10:48 PM
> To: Xing, Beilei ; Wu, Jingjing 
> Cc: dev@dpdk.org; Thomas Monjalon 
> Subject: Re: [PATCH v15 00/18] add support for idpf PMD in DPDK
> 
> On 10/29/22 06:27, beilei.x...@intel.com wrote:
> > From: Beilei Xing 
> >
> > This patchset introduced the idpf (Infrastructure Data Path Function) PMD
> in DPDK for Intel® IPU E2000 (Device ID: 0x1452).
> > The Intel® IPU E2000 targets to deliver high performance under real
> workloads with security and isolation.
> > Please refer to
> > https://www.intel.com/content/www/us/en/products/network-
> io/infrastruc
> > ture-processing-units/asic/e2000-asic.html
> > for more information.
> >
> > Linux upstream is still ongoing, previous work refers to
> https://patchwork.ozlabs.org/project/intel-wired-
> lan/patch/20220128001009.721392-20-alan.br...@intel.com/.
> >
> > v2-v4:
> > fixed some coding style issues and did some refactors.
> >
> > v5:
> > fixed typo.
> >
> > v6-v9:
> > fixed build errors and coding style issues.
> >
> > v11:
> >   - move shared code to common/idpf/base
> >   - Create one vport if there's no vport devargs
> >   - Refactor if conditions according to coding style
> >   - Refactor virtual channel return values
> >   - Refine dev_stop function
> >   - Refine RSS lut/key
> >   - Fix build error
> >
> > v12:
> >   - Refine dev_configure
> >   - Fix coding style according to the comments
> >   - Re-order patch
> >   - Romove dev_supported_ptypes_get
> >
> > v13:
> >   - refine dev_start/stop and queue_start/stop
> >   - fix timestamp offload
> >
> > v14:
> >   - fix wrong position for rte_validate_tx_offload
> >
> > v15:
> >   - refine the return value for ethdev ops.
> >   - removce forward static declarations.
> >   - refine get caps.
> >   - fix lock/unlock handling.
> 
> Applied to dpdk-next-net/main, thanks.
> 
> I've a number of concerns:
>   * conditional compilation IDPF_RX_PTYPE_OFFLOAD in [PATCH v15 17/18]

Will remove the conditional compilation

> net/idpf: add AVX512 data path for single queue model
>   * the same prefix used for functions in common/idpf/base and net/idpf
> drivers

I think the name of PMD and common library can be the same, right?

>   * common/idpf/base uses own defines for negative errno (defined as a
> number with corresponding errno in a comment). Strictly speaking it is not
> the same, but work fine in a majority of cases

Make sense, will remove the own defines.
Thanks for your review, I saw the status in patchwork has been accepted, but 
didn't see idpf in dpdk-next-net, will send v16 to address the comments first.

> 
> So, final decision will be done by Thomas on pulling to main tree.


[PATCH v16 00/18] add support for idpf PMD in DPDK

2022-10-30 Thread beilei . xing
From: Beilei Xing 

This patchset introduced the idpf (Infrastructure Data Path Function) PMD in 
DPDK for Intel® IPU E2000 (Device ID: 0x1452).
The Intel® IPU E2000 targets to deliver high performance under real workloads 
with security and isolation.
Please refer to
https://www.intel.com/content/www/us/en/products/network-io/infrastructure-processing-units/asic/e2000-asic.html
for more information.

Linux upstream is still ongoing, previous work refers to 
https://patchwork.ozlabs.org/project/intel-wired-lan/patch/20220128001009.721392-20-alan.br...@intel.com/.

v2-v4:
fixed some coding style issues and did some refactors.

v5:
fixed typo.

v6-v9:
fixed build errors and coding style issues.

v11:
 - move shared code to common/idpf/base
 - Create one vport if there's no vport devargs
 - Refactor if conditions according to coding style
 - Refactor virtual channel return values
 - Refine dev_stop function
 - Refine RSS lut/key
 - Fix build error

v12:
 - Refine dev_configure
 - Fix coding style according to the comments
 - Re-order patch
 - Romove dev_supported_ptypes_get

v13:
 - refine dev_start/stop and queue_start/stop
 - fix timestamp offload

v14:
 - fix wrong position for rte_validate_tx_offload

v15:
 - refine the return value for ethdev ops.
 - removce forward static declarations.
 - refine get caps.
 - fix lock/unlock handling.

v16:
 - refine errno in shared code
 - remove the conditional compilation IDPF_RX_PTYPE_OFFLOAD

Junfeng Guo (18):
  common/idpf: introduce common library
  net/idpf: add support for device initialization
  net/idpf: add Tx queue setup
  net/idpf: add Rx queue setup
  net/idpf: add support for device start and stop
  net/idpf: add support for queue start
  net/idpf: add support for queue stop
  net/idpf: add queue release
  net/idpf: add support for MTU configuration
  net/idpf: add support for basic Rx datapath
  net/idpf: add support for basic Tx datapath
  net/idpf: support parsing packet type
  net/idpf: add support for write back based on ITR expire
  net/idpf: add support for RSS
  net/idpf: add support for Rx offloading
  net/idpf: add support for Tx offloading
  net/idpf: add AVX512 data path for single queue model
  net/idpf: add support for timestamp offload

 MAINTAINERS   |9 +
 doc/guides/nics/features/idpf.ini |   17 +
 doc/guides/nics/idpf.rst  |   85 +
 doc/guides/nics/index.rst |1 +
 doc/guides/rel_notes/release_22_11.rst|6 +
 drivers/common/idpf/base/idpf_alloc.h |   22 +
 drivers/common/idpf/base/idpf_common.c|  364 +++
 drivers/common/idpf/base/idpf_controlq.c  |  691 
 drivers/common/idpf/base/idpf_controlq.h  |  224 ++
 drivers/common/idpf/base/idpf_controlq_api.h  |  207 ++
 .../common/idpf/base/idpf_controlq_setup.c|  179 +
 drivers/common/idpf/base/idpf_devids.h|   18 +
 drivers/common/idpf/base/idpf_lan_pf_regs.h   |  134 +
 drivers/common/idpf/base/idpf_lan_txrx.h  |  428 +++
 drivers/common/idpf/base/idpf_lan_vf_regs.h   |  114 +
 drivers/common/idpf/base/idpf_osdep.h |  364 +++
 drivers/common/idpf/base/idpf_prototype.h |   45 +
 drivers/common/idpf/base/idpf_type.h  |  106 +
 drivers/common/idpf/base/meson.build  |   14 +
 drivers/common/idpf/base/siov_regs.h  |   47 +
 drivers/common/idpf/base/virtchnl.h   | 2866 +
 drivers/common/idpf/base/virtchnl2.h  | 1462 +
 drivers/common/idpf/base/virtchnl2_lan_desc.h |  606 
 .../common/idpf/base/virtchnl_inline_ipsec.h  |  567 
 drivers/common/idpf/meson.build   |4 +
 drivers/common/idpf/version.map   |   12 +
 drivers/common/meson.build|1 +
 drivers/net/idpf/idpf_ethdev.c| 1293 
 drivers/net/idpf/idpf_ethdev.h|  252 ++
 drivers/net/idpf/idpf_logs.h  |   56 +
 drivers/net/idpf/idpf_rxtx.c  | 2308 +
 drivers/net/idpf/idpf_rxtx.h  |  291 ++
 drivers/net/idpf/idpf_rxtx_vec_avx512.c   |  871 +
 drivers/net/idpf/idpf_rxtx_vec_common.h   |  100 +
 drivers/net/idpf/idpf_vchnl.c | 1416 
 drivers/net/idpf/meson.build  |   44 +
 drivers/net/idpf/version.map  |3 +
 drivers/net/meson.build   |1 +
 38 files changed, 15228 insertions(+)
 create mode 100644 doc/guides/nics/features/idpf.ini
 create mode 100644 doc/guides/nics/idpf.rst
 create mode 100644 drivers/common/idpf/base/idpf_alloc.h
 create mode 100644 drivers/common/idpf/base/idpf_common.c
 create mode 100644 drivers/common/idpf/base/idpf_controlq.c
 create mode 100644 drivers/common/idpf/base/idpf_controlq.h
 create mode 100644 drivers/common/idpf/base/idpf_controlq_api.h
 create mode 100644 drivers/common/idpf/base/idpf_controlq_setup.c
 create mode 100644 drivers/common/idpf/base/id

[PATCH v16 03/18] net/idpf: add Tx queue setup

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for tx_queue_setup ops.

In the single queue model, the same descriptor queue is used by SW to
post buffer descriptors to HW and by HW to post completed descriptors
to SW.

In the split queue model, "RX buffer queues" are used to pass
descriptor buffers from SW to HW while Rx queues are used only to
pass the descriptor completions, that is, descriptors that point
to completed buffers, from HW to SW. This is contrary to the single
queue model in which Rx queues are used for both purposes.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  13 ++
 drivers/net/idpf/idpf_rxtx.c   | 364 +
 drivers/net/idpf/idpf_rxtx.h   |  70 +++
 drivers/net/idpf/meson.build   |   1 +
 4 files changed, 448 insertions(+)
 create mode 100644 drivers/net/idpf/idpf_rxtx.c
 create mode 100644 drivers/net/idpf/idpf_rxtx.h

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 035f563275..54f20d30ca 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -11,6 +11,7 @@
 #include 
 
 #include "idpf_ethdev.h"
+#include "idpf_rxtx.h"
 
 #define IDPF_TX_SINGLE_Q   "tx_single"
 #define IDPF_RX_SINGLE_Q   "rx_single"
@@ -42,6 +43,17 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
dev_info->max_mtu = dev_info->max_rx_pktlen - IDPF_ETH_OVERHEAD;
dev_info->min_mtu = RTE_ETHER_MIN_MTU;
 
+   dev_info->default_txconf = (struct rte_eth_txconf) {
+   .tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
+   .tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH,
+   };
+
+   dev_info->tx_desc_lim = (struct rte_eth_desc_lim) {
+   .nb_max = IDPF_MAX_RING_DESC,
+   .nb_min = IDPF_MIN_RING_DESC,
+   .nb_align = IDPF_ALIGN_RING_DESC,
+   };
+
return 0;
 }
 
@@ -631,6 +643,7 @@ idpf_adapter_init(struct rte_pci_device *pci_dev, struct 
idpf_adapter *adapter)
 static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_configure  = idpf_dev_configure,
.dev_close  = idpf_dev_close,
+   .tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
 };
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
new file mode 100644
index 00..4afa0a2560
--- /dev/null
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -0,0 +1,364 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2022 Intel Corporation
+ */
+
+#include 
+#include 
+
+#include "idpf_ethdev.h"
+#include "idpf_rxtx.h"
+
+static int
+check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
+   uint16_t tx_free_thresh)
+{
+   /* TX descriptors will have their RS bit set after tx_rs_thresh
+* descriptors have been used. The TX descriptor ring will be cleaned
+* after tx_free_thresh descriptors are used or if the number of
+* descriptors required to transmit a packet is greater than the
+* number of free TX descriptors.
+*
+* The following constraints must be satisfied:
+*  - tx_rs_thresh must be less than the size of the ring minus 2.
+*  - tx_free_thresh must be less than the size of the ring minus 3.
+*  - tx_rs_thresh must be less than or equal to tx_free_thresh.
+*  - tx_rs_thresh must be a divisor of the ring size.
+*
+* One descriptor in the TX ring is used as a sentinel to avoid a H/W
+* race condition, hence the maximum threshold constraints. When set
+* to zero use default values.
+*/
+   if (tx_rs_thresh >= (nb_desc - 2)) {
+   PMD_INIT_LOG(ERR, "tx_rs_thresh (%u) must be less than the "
+"number of TX descriptors (%u) minus 2",
+tx_rs_thresh, nb_desc);
+   return -EINVAL;
+   }
+   if (tx_free_thresh >= (nb_desc - 3)) {
+   PMD_INIT_LOG(ERR, "tx_free_thresh (%u) must be less than the "
+"number of TX descriptors (%u) minus 3.",
+tx_free_thresh, nb_desc);
+   return -EINVAL;
+   }
+   if (tx_rs_thresh > tx_free_thresh) {
+   PMD_INIT_LOG(ERR, "tx_rs_thresh (%u) must be less than or "
+"equal to tx_free_thresh (%u).",
+tx_rs_thresh, tx_free_thresh);
+   return -EINVAL;
+   }
+   if ((nb_desc % tx_rs_thresh) != 0) {
+   PMD_INIT_LOG(ERR, "tx_rs_thresh (%u) must be a divisor of the "
+"number of TX descriptors (%u).",
+tx_rs_thresh, nb_desc);
+   return -EINVAL;
+   }
+
+   return 0;
+}
+
+static void
+reset_split_tx_descq(struct idpf_tx_queue *txq)
+{
+   struct id

[PATCH v16 02/18] net/idpf: add support for device initialization

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Support device init and add the following dev ops:
 - dev_configure
 - dev_close
 - dev_infos_get

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Xiao Wang 
Signed-off-by: Wenjun Wu 
Signed-off-by: Junfeng Guo 
---
 MAINTAINERS|   9 +
 doc/guides/nics/features/idpf.ini  |   9 +
 doc/guides/nics/idpf.rst   |  66 ++
 doc/guides/nics/index.rst  |   1 +
 doc/guides/rel_notes/release_22_11.rst |   6 +
 drivers/net/idpf/idpf_ethdev.c | 891 +
 drivers/net/idpf/idpf_ethdev.h | 189 ++
 drivers/net/idpf/idpf_logs.h   |  56 ++
 drivers/net/idpf/idpf_vchnl.c  | 416 
 drivers/net/idpf/meson.build   |  15 +
 drivers/net/idpf/version.map   |   3 +
 drivers/net/meson.build|   1 +
 12 files changed, 1662 insertions(+)
 create mode 100644 doc/guides/nics/features/idpf.ini
 create mode 100644 doc/guides/nics/idpf.rst
 create mode 100644 drivers/net/idpf/idpf_ethdev.c
 create mode 100644 drivers/net/idpf/idpf_ethdev.h
 create mode 100644 drivers/net/idpf/idpf_logs.h
 create mode 100644 drivers/net/idpf/idpf_vchnl.c
 create mode 100644 drivers/net/idpf/meson.build
 create mode 100644 drivers/net/idpf/version.map

diff --git a/MAINTAINERS b/MAINTAINERS
index bdf233c9f8..cc66db25e8 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -770,6 +770,15 @@ F: drivers/net/ice/
 F: doc/guides/nics/ice.rst
 F: doc/guides/nics/features/ice.ini
 
+Intel idpf
+M: Jingjing Wu 
+M: Beilei Xing 
+T: git://dpdk.org/next/dpdk-next-net-intel
+F: drivers/net/idpf/
+F: drivers/common/idpf/
+F: doc/guides/nics/idpf.rst
+F: doc/guides/nics/features/idpf.ini
+
 Intel igc
 M: Junfeng Guo 
 M: Simei Su 
diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
new file mode 100644
index 00..46aab2eb61
--- /dev/null
+++ b/doc/guides/nics/features/idpf.ini
@@ -0,0 +1,9 @@
+;
+; Supported features of the 'idpf' network poll mode driver.
+;
+; Refer to default.ini for the full list of available PMD features.
+;
+[Features]
+Linux= Y
+x86-32   = Y
+x86-64   = Y
diff --git a/doc/guides/nics/idpf.rst b/doc/guides/nics/idpf.rst
new file mode 100644
index 00..c1001d5d0c
--- /dev/null
+++ b/doc/guides/nics/idpf.rst
@@ -0,0 +1,66 @@
+..  SPDX-License-Identifier: BSD-3-Clause
+Copyright(c) 2022 Intel Corporation.
+
+IDPF Poll Mode Driver
+==
+
+The [*EXPERIMENTAL*] idpf PMD (**librte_net_idpf**) provides poll mode driver 
support for
+Intel® Infrastructure Processing Unit (Intel® IPU) E2000.
+
+
+Linux Prerequisites
+---
+
+- Follow the DPDK :ref:`Getting Started Guide for Linux ` to setup 
the basic DPDK environment.
+
+- To get better performance on Intel platforms, please follow the "How to get 
best performance with NICs on Intel platforms"
+  section of the :ref:`Getting Started Guide for Linux `.
+
+
+Pre-Installation Configuration
+--
+
+Runtime Config Options
+~~
+
+- ``vport`` (default ``0``)
+
+  The IDPF PMD supports creation of multiple vports for one PCI device, each 
vport
+  corresponds to a single ethdev. Using the ``devargs`` parameter ``vport`` 
the user
+  can specify the vports with specific ID to be created, for example::
+
+-a ca:00.0,vport=[0,2,3]
+
+  Then idpf PMD will create 3 vports (ethdevs) for device ca:00.0.
+  NOTE: If the parameter is not provided, the vport 0 will be created by 
default.
+
+- ``rx_single`` (default ``0``)
+
+  There're two queue modes supported by Intel® IPU Ethernet ES2000 Series, 
single queue
+  mode and split queue mode for Rx queue. User can choose Rx queue mode by the 
``devargs``
+  parameter ``rx_single``.
+
+-a ca:00.0,rx_single=1
+
+  Then idpf PMD will configure Rx queue with single queue mode. Otherwise, 
split queue
+  mode is chosen by default.
+
+- ``tx_single`` (default ``0``)
+
+  There're two queue modes supported by Intel® IPU Ethernet ES2000 Series, 
single queue
+  mode and split queue mode for Tx queue. User can choose Tx queue mode by the 
``devargs``
+  parameter ``tx_single``.
+
+-a ca:00.0,tx_single=1
+
+  Then idpf PMD will configure Tx queue with single queue mode. Otherwise, 
split queue
+  mode is chosen by default.
+
+
+Driver compilation and testing
+--
+
+Refer to the document :ref:`compiling and testing a PMD for a NIC 
`
+for details.
+
+
diff --git a/doc/guides/nics/index.rst b/doc/guides/nics/index.rst
index 4d40ea29a3..12841ce407 100644
--- a/doc/guides/nics/index.rst
+++ b/doc/guides/nics/index.rst
@@ -34,6 +34,7 @@ Network Interface Controller Drivers
 hns3
 i40e
 ice
+idpf
 igb
 igc
 ionic
diff --git a/doc/guides/rel_notes/release_22_11.rst 
b/doc/guides/rel_notes/release_22_11.rst
index 28812e092f..77674f2b06 100644
--- a/doc/guides/rel_notes/rele

[PATCH v16 04/18] net/idpf: add Rx queue setup

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for rx_queue_setup ops.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  11 +
 drivers/net/idpf/idpf_rxtx.c   | 400 +
 drivers/net/idpf/idpf_rxtx.h   |  46 
 3 files changed, 457 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 54f20d30ca..fb5cd1b111 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -48,12 +48,22 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
.tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH,
};
 
+   dev_info->default_rxconf = (struct rte_eth_rxconf) {
+   .rx_free_thresh = IDPF_DEFAULT_RX_FREE_THRESH,
+   };
+
dev_info->tx_desc_lim = (struct rte_eth_desc_lim) {
.nb_max = IDPF_MAX_RING_DESC,
.nb_min = IDPF_MIN_RING_DESC,
.nb_align = IDPF_ALIGN_RING_DESC,
};
 
+   dev_info->rx_desc_lim = (struct rte_eth_desc_lim) {
+   .nb_max = IDPF_MAX_RING_DESC,
+   .nb_min = IDPF_MIN_RING_DESC,
+   .nb_align = IDPF_ALIGN_RING_DESC,
+   };
+
return 0;
 }
 
@@ -643,6 +653,7 @@ idpf_adapter_init(struct rte_pci_device *pci_dev, struct 
idpf_adapter *adapter)
 static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_configure  = idpf_dev_configure,
.dev_close  = idpf_dev_close,
+   .rx_queue_setup = idpf_rx_queue_setup,
.tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
 };
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 4afa0a2560..25dd5d85d5 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -8,6 +8,21 @@
 #include "idpf_ethdev.h"
 #include "idpf_rxtx.h"
 
+static int
+check_rx_thresh(uint16_t nb_desc, uint16_t thresh)
+{
+   /* The following constraints must be satisfied:
+*   thresh < rxq->nb_rx_desc
+*/
+   if (thresh >= nb_desc) {
+   PMD_INIT_LOG(ERR, "rx_free_thresh (%u) must be less than %u",
+thresh, nb_desc);
+   return -EINVAL;
+   }
+
+   return 0;
+}
+
 static int
 check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
uint16_t tx_free_thresh)
@@ -56,6 +71,87 @@ check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
return 0;
 }
 
+static void
+reset_split_rx_descq(struct idpf_rx_queue *rxq)
+{
+   uint16_t len;
+   uint32_t i;
+
+   if (rxq == NULL)
+   return;
+
+   len = rxq->nb_rx_desc + IDPF_RX_MAX_BURST;
+
+   for (i = 0; i < len * sizeof(struct virtchnl2_rx_flex_desc_adv_nic_3);
+i++)
+   ((volatile char *)rxq->rx_ring)[i] = 0;
+
+   rxq->rx_tail = 0;
+   rxq->expected_gen_id = 1;
+}
+
+static void
+reset_split_rx_bufq(struct idpf_rx_queue *rxq)
+{
+   uint16_t len;
+   uint32_t i;
+
+   if (rxq == NULL)
+   return;
+
+   len = rxq->nb_rx_desc + IDPF_RX_MAX_BURST;
+
+   for (i = 0; i < len * sizeof(struct virtchnl2_splitq_rx_buf_desc);
+i++)
+   ((volatile char *)rxq->rx_ring)[i] = 0;
+
+   memset(&rxq->fake_mbuf, 0x0, sizeof(rxq->fake_mbuf));
+
+   for (i = 0; i < IDPF_RX_MAX_BURST; i++)
+   rxq->sw_ring[rxq->nb_rx_desc + i] = &rxq->fake_mbuf;
+
+   /* The next descriptor id which can be received. */
+   rxq->rx_next_avail = 0;
+
+   /* The next descriptor id which can be refilled. */
+   rxq->rx_tail = 0;
+   /* The number of descriptors which can be refilled. */
+   rxq->nb_rx_hold = rxq->nb_rx_desc - 1;
+
+   rxq->bufq1 = NULL;
+   rxq->bufq2 = NULL;
+}
+
+static void
+reset_single_rx_queue(struct idpf_rx_queue *rxq)
+{
+   uint16_t len;
+   uint32_t i;
+
+   if (rxq == NULL)
+   return;
+
+   len = rxq->nb_rx_desc + IDPF_RX_MAX_BURST;
+
+   for (i = 0; i < len * sizeof(struct virtchnl2_singleq_rx_buf_desc);
+i++)
+   ((volatile char *)rxq->rx_ring)[i] = 0;
+
+   memset(&rxq->fake_mbuf, 0x0, sizeof(rxq->fake_mbuf));
+
+   for (i = 0; i < IDPF_RX_MAX_BURST; i++)
+   rxq->sw_ring[rxq->nb_rx_desc + i] = &rxq->fake_mbuf;
+
+   rxq->rx_tail = 0;
+   rxq->nb_rx_hold = 0;
+
+   if (rxq->pkt_first_seg != NULL)
+   rte_pktmbuf_free(rxq->pkt_first_seg);
+
+   rxq->pkt_first_seg = NULL;
+   rxq->pkt_last_seg = NULL;
+}
+
 static void
 reset_split_tx_descq(struct idpf_tx_queue *txq)
 {
@@ -145,6 +241,310 @@ reset_single_tx_queue(struct idpf_tx_queue *txq)
txq->next_rs = txq->rs_thresh - 1;
 }
 
+static int
+idpf_rx_split_bufq_setup(struct rte_eth_dev *dev, struct idpf_rx_queue *bufq,
+   

[PATCH v16 05/18] net/idpf: add support for device start and stop

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add dev ops dev_start, dev_stop and link_update.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c | 55 ++
 drivers/net/idpf/idpf_rxtx.c   | 20 +
 2 files changed, 75 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index fb5cd1b111..621bf9aad5 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -29,6 +29,22 @@ static const char * const idpf_valid_args[] = {
NULL
 };
 
+static int
+idpf_dev_link_update(struct rte_eth_dev *dev,
+__rte_unused int wait_to_complete)
+{
+   struct rte_eth_link new_link;
+
+   memset(&new_link, 0, sizeof(new_link));
+
+   new_link.link_speed = RTE_ETH_SPEED_NUM_NONE;
+   new_link.link_duplex = RTE_ETH_LINK_FULL_DUPLEX;
+   new_link.link_autoneg = !(dev->data->dev_conf.link_speeds &
+ RTE_ETH_LINK_SPEED_FIXED);
+
+   return rte_eth_linkstatus_set(dev, &new_link);
+}
+
 static int
 idpf_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
 {
@@ -267,6 +283,42 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_dev_start(struct rte_eth_dev *dev)
+{
+   struct idpf_vport *vport = dev->data->dev_private;
+   int ret;
+
+   if (dev->data->mtu > vport->max_mtu) {
+   PMD_DRV_LOG(ERR, "MTU should be less than %d", vport->max_mtu);
+   return -EINVAL;
+   }
+
+   vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
+
+   /* TODO: start queues */
+
+   ret = idpf_vc_ena_dis_vport(vport, true);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR, "Failed to enable vport");
+   return ret;
+   }
+
+   return 0;
+}
+
+static int
+idpf_dev_stop(struct rte_eth_dev *dev)
+{
+   struct idpf_vport *vport = dev->data->dev_private;
+
+   idpf_vc_ena_dis_vport(vport, false);
+
+   /* TODO: stop queues */
+
+   return 0;
+}
+
 static int
 idpf_dev_close(struct rte_eth_dev *dev)
 {
@@ -656,6 +708,9 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.rx_queue_setup = idpf_rx_queue_setup,
.tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
+   .dev_start  = idpf_dev_start,
+   .dev_stop   = idpf_dev_stop,
+   .link_update= idpf_dev_link_update,
 };
 
 static uint16_t
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 25dd5d85d5..3528d2f2c7 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -334,6 +334,11 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   if (rx_conf->rx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -465,6 +470,11 @@ idpf_rx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   if (rx_conf->rx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -569,6 +579,11 @@ idpf_tx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   if (tx_conf->tx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf split txq",
 sizeof(struct idpf_tx_queue),
@@ -691,6 +706,11 @@ idpf_tx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   if (tx_conf->tx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf txq",
 sizeof(struct idpf_tx_queue),
-- 
2.26.2



[PATCH v16 06/18] net/idpf: add support for queue start

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for these device ops:
 - rx_queue_start
 - tx_queue_start

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  42 +++-
 drivers/net/idpf/idpf_ethdev.h |   9 +
 drivers/net/idpf/idpf_rxtx.c   | 237 +++--
 drivers/net/idpf/idpf_rxtx.h   |   6 +
 drivers/net/idpf/idpf_vchnl.c  | 447 +
 5 files changed, 720 insertions(+), 21 deletions(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 621bf9aad5..0400ed611f 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -283,6 +283,39 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_start_queues(struct rte_eth_dev *dev)
+{
+   struct idpf_rx_queue *rxq;
+   struct idpf_tx_queue *txq;
+   int err = 0;
+   int i;
+
+   for (i = 0; i < dev->data->nb_tx_queues; i++) {
+   txq = dev->data->tx_queues[i];
+   if (txq == NULL || txq->tx_deferred_start)
+   continue;
+   err = idpf_tx_queue_start(dev, i);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR, "Fail to start Tx queue %u", i);
+   return err;
+   }
+   }
+
+   for (i = 0; i < dev->data->nb_rx_queues; i++) {
+   rxq = dev->data->rx_queues[i];
+   if (rxq == NULL || rxq->rx_deferred_start)
+   continue;
+   err = idpf_rx_queue_start(dev, i);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR, "Fail to start Rx queue %u", i);
+   return err;
+   }
+   }
+
+   return err;
+}
+
 static int
 idpf_dev_start(struct rte_eth_dev *dev)
 {
@@ -296,11 +329,16 @@ idpf_dev_start(struct rte_eth_dev *dev)
 
vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
 
-   /* TODO: start queues */
+   ret = idpf_start_queues(dev);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR, "Failed to start queues");
+   return ret;
+   }
 
ret = idpf_vc_ena_dis_vport(vport, true);
if (ret != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
+   /* TODO: stop queues */
return ret;
}
 
@@ -711,6 +749,8 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_start  = idpf_dev_start,
.dev_stop   = idpf_dev_stop,
.link_update= idpf_dev_link_update,
+   .rx_queue_start = idpf_rx_queue_start,
+   .tx_queue_start = idpf_tx_queue_start,
 };
 
 static uint16_t
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 84ae6641e2..96c22009e9 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -24,7 +24,9 @@
 #define IDPF_DEFAULT_TXQ_NUM   16
 
 #define IDPF_INVALID_VPORT_IDX 0x
+#define IDPF_TXQ_PER_GRP   1
 #define IDPF_TX_COMPLQ_PER_GRP 1
+#define IDPF_RXQ_PER_GRP   1
 #define IDPF_RX_BUFQ_PER_GRP   2
 
 #define IDPF_CTLQ_ID   -1
@@ -182,6 +184,13 @@ int idpf_vc_check_api_version(struct idpf_adapter 
*adapter);
 int idpf_vc_get_caps(struct idpf_adapter *adapter);
 int idpf_vc_create_vport(struct idpf_adapter *adapter);
 int idpf_vc_destroy_vport(struct idpf_vport *vport);
+int idpf_vc_config_rxqs(struct idpf_vport *vport);
+int idpf_vc_config_rxq(struct idpf_vport *vport, uint16_t rxq_id);
+int idpf_vc_config_txqs(struct idpf_vport *vport);
+int idpf_vc_config_txq(struct idpf_vport *vport, uint16_t txq_id);
+int idpf_switch_queue(struct idpf_vport *vport, uint16_t qid,
+ bool rx, bool on);
+int idpf_vc_ena_dis_queues(struct idpf_vport *vport, bool enable);
 int idpf_vc_ena_dis_vport(struct idpf_vport *vport, bool enable);
 int idpf_read_one_msg(struct idpf_adapter *adapter, uint32_t ops,
  uint16_t buf_len, uint8_t *buf);
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 3528d2f2c7..6d954afd9d 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -334,11 +334,6 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
-   if (rx_conf->rx_deferred_start) {
-   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
-   return -EINVAL;
-   }
-
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -354,6 +349,7 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, uint16_t 
queue_idx,
rxq->rx_free_thresh = rx_free_thresh;
rxq->queue_id = vport->chunks_info.rx_start_qid + queue_idx;
rxq->port_id = dev-

[PATCH v16 07/18] net/idpf: add support for queue stop

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for these device ops:
 - rx_queue_stop
 - tx_queue_stop

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  17 ++--
 drivers/net/idpf/idpf_rxtx.c   | 148 +
 drivers/net/idpf/idpf_rxtx.h   |  13 +++
 drivers/net/idpf/idpf_vchnl.c  |  69 +++
 4 files changed, 242 insertions(+), 5 deletions(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 0400ed611f..9f1e1e6a18 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -324,7 +324,8 @@ idpf_dev_start(struct rte_eth_dev *dev)
 
if (dev->data->mtu > vport->max_mtu) {
PMD_DRV_LOG(ERR, "MTU should be less than %d", vport->max_mtu);
-   return -EINVAL;
+   ret = -EINVAL;
+   goto err_mtu;
}
 
vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
@@ -332,17 +333,21 @@ idpf_dev_start(struct rte_eth_dev *dev)
ret = idpf_start_queues(dev);
if (ret != 0) {
PMD_DRV_LOG(ERR, "Failed to start queues");
-   return ret;
+   goto err_mtu;
}
 
ret = idpf_vc_ena_dis_vport(vport, true);
if (ret != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
-   /* TODO: stop queues */
-   return ret;
+   goto err_vport;
}
 
return 0;
+
+err_vport:
+   idpf_stop_queues(dev);
+err_mtu:
+   return ret;
 }
 
 static int
@@ -352,7 +357,7 @@ idpf_dev_stop(struct rte_eth_dev *dev)
 
idpf_vc_ena_dis_vport(vport, false);
 
-   /* TODO: stop queues */
+   idpf_stop_queues(dev);
 
return 0;
 }
@@ -751,6 +756,8 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.link_update= idpf_dev_link_update,
.rx_queue_start = idpf_rx_queue_start,
.tx_queue_start = idpf_tx_queue_start,
+   .rx_queue_stop  = idpf_rx_queue_stop,
+   .tx_queue_stop  = idpf_tx_queue_stop,
 };
 
 static uint16_t
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 6d954afd9d..8d5ec41a1f 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -71,6 +71,55 @@ check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
return 0;
 }
 
+static void
+release_rxq_mbufs(struct idpf_rx_queue *rxq)
+{
+   uint16_t i;
+
+   if (rxq->sw_ring == NULL)
+   return;
+
+   for (i = 0; i < rxq->nb_rx_desc; i++) {
+   if (rxq->sw_ring[i] != NULL) {
+   rte_pktmbuf_free_seg(rxq->sw_ring[i]);
+   rxq->sw_ring[i] = NULL;
+   }
+   }
+}
+
+static void
+release_txq_mbufs(struct idpf_tx_queue *txq)
+{
+   uint16_t nb_desc, i;
+
+   if (txq == NULL || txq->sw_ring == NULL) {
+   PMD_DRV_LOG(DEBUG, "Pointer to rxq or sw_ring is NULL");
+   return;
+   }
+
+   if (txq->sw_nb_desc != 0) {
+   /* For split queue model, descriptor ring */
+   nb_desc = txq->sw_nb_desc;
+   } else {
+   /* For single queue model */
+   nb_desc = txq->nb_tx_desc;
+   }
+   for (i = 0; i < nb_desc; i++) {
+   if (txq->sw_ring[i].mbuf != NULL) {
+   rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf);
+   txq->sw_ring[i].mbuf = NULL;
+   }
+   }
+}
+
+static const struct idpf_rxq_ops def_rxq_ops = {
+   .release_mbufs = release_rxq_mbufs,
+};
+
+static const struct idpf_txq_ops def_txq_ops = {
+   .release_mbufs = release_txq_mbufs,
+};
+
 static void
 reset_split_rx_descq(struct idpf_rx_queue *rxq)
 {
@@ -122,6 +171,14 @@ reset_split_rx_bufq(struct idpf_rx_queue *rxq)
rxq->bufq2 = NULL;
 }
 
+static inline void
+reset_split_rx_queue(struct idpf_rx_queue *rxq)
+{
+   reset_split_rx_descq(rxq);
+   reset_split_rx_bufq(rxq->bufq1);
+   reset_split_rx_bufq(rxq->bufq2);
+}
+
 static void
 reset_single_rx_queue(struct idpf_rx_queue *rxq)
 {
@@ -301,6 +358,7 @@ idpf_rx_split_bufq_setup(struct rte_eth_dev *dev, struct 
idpf_rx_queue *bufq,
bufq->q_set = true;
bufq->qrx_tail = hw->hw_addr + (vport->chunks_info.rx_buf_qtail_start +
 queue_idx * vport->chunks_info.rx_buf_qtail_spacing);
+   bufq->ops = &def_rxq_ops;
 
/* TODO: allow bulk or vec */
 
@@ -527,6 +585,7 @@ idpf_rx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
dev->data->rx_queues[queue_idx] = rxq;
rxq->qrx_tail = hw->hw_addr + (vport->chunks_info.rx_qtail_start +
queue_idx * vport->chunks_info.rx_qtail_spacing);
+   rxq->ops = &def_rxq_ops;
 
return 0;
 }
@@ -621,6 +680,7 @@ idpf_tx_split_queue_setup(struct

[PATCH v16 08/18] net/idpf: add queue release

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for queue operations:
 - rx_queue_release
 - tx_queue_release

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  2 +
 drivers/net/idpf/idpf_rxtx.c   | 81 ++
 drivers/net/idpf/idpf_rxtx.h   |  3 ++
 3 files changed, 86 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 9f1e1e6a18..1485f40e71 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -758,6 +758,8 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.tx_queue_start = idpf_tx_queue_start,
.rx_queue_stop  = idpf_rx_queue_stop,
.tx_queue_stop  = idpf_tx_queue_stop,
+   .rx_queue_release   = idpf_dev_rx_queue_release,
+   .tx_queue_release   = idpf_dev_tx_queue_release,
 };
 
 static uint16_t
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 8d5ec41a1f..053409b99a 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -171,6 +171,51 @@ reset_split_rx_bufq(struct idpf_rx_queue *rxq)
rxq->bufq2 = NULL;
 }
 
+static void
+idpf_rx_queue_release(void *rxq)
+{
+   struct idpf_rx_queue *q = rxq;
+
+   if (q == NULL)
+   return;
+
+   /* Split queue */
+   if (q->bufq1 != NULL && q->bufq2 != NULL) {
+   q->bufq1->ops->release_mbufs(q->bufq1);
+   rte_free(q->bufq1->sw_ring);
+   rte_memzone_free(q->bufq1->mz);
+   rte_free(q->bufq1);
+   q->bufq2->ops->release_mbufs(q->bufq2);
+   rte_free(q->bufq2->sw_ring);
+   rte_memzone_free(q->bufq2->mz);
+   rte_free(q->bufq2);
+   rte_memzone_free(q->mz);
+   rte_free(q);
+   return;
+   }
+
+   /* Single queue */
+   q->ops->release_mbufs(q);
+   rte_free(q->sw_ring);
+   rte_memzone_free(q->mz);
+   rte_free(q);
+}
+
+static void
+idpf_tx_queue_release(void *txq)
+{
+   struct idpf_tx_queue *q = txq;
+
+   if (q == NULL)
+   return;
+
+   rte_free(q->complq);
+   q->ops->release_mbufs(q);
+   rte_free(q->sw_ring);
+   rte_memzone_free(q->mz);
+   rte_free(q);
+}
+
 static inline void
 reset_split_rx_queue(struct idpf_rx_queue *rxq)
 {
@@ -392,6 +437,12 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed */
+   if (dev->data->rx_queues[queue_idx] != NULL) {
+   idpf_rx_queue_release(dev->data->rx_queues[queue_idx]);
+   dev->data->rx_queues[queue_idx] = NULL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -524,6 +575,12 @@ idpf_rx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed */
+   if (dev->data->rx_queues[queue_idx] != NULL) {
+   idpf_rx_queue_release(dev->data->rx_queues[queue_idx]);
+   dev->data->rx_queues[queue_idx] = NULL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -630,6 +687,12 @@ idpf_tx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed. */
+   if (dev->data->tx_queues[queue_idx] != NULL) {
+   idpf_tx_queue_release(dev->data->tx_queues[queue_idx]);
+   dev->data->tx_queues[queue_idx] = NULL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf split txq",
 sizeof(struct idpf_tx_queue),
@@ -754,6 +817,12 @@ idpf_tx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed. */
+   if (dev->data->tx_queues[queue_idx] != NULL) {
+   idpf_tx_queue_release(dev->data->tx_queues[queue_idx]);
+   dev->data->tx_queues[queue_idx] = NULL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf txq",
 sizeof(struct idpf_tx_queue),
@@ -1102,6 +1171,18 @@ idpf_tx_queue_stop(struct rte_eth_dev *dev, uint16_t 
tx_queue_id)
return 0;
 }
 
+void
+idpf_dev_rx_queue_release(struct rte_eth_dev *dev, uint16_t qid)
+{
+   idpf_rx_queue_r

[PATCH v16 09/18] net/idpf: add support for MTU configuration

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add dev ops mtu_set.

Signed-off-by: Beilei Xing 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |  1 +
 drivers/net/idpf/idpf_ethdev.c| 13 +
 2 files changed, 14 insertions(+)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index 46aab2eb61..d722c49fde 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -4,6 +4,7 @@
 ; Refer to default.ini for the full list of available PMD features.
 ;
 [Features]
+MTU update   = Y
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 1485f40e71..856f3d7266 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -83,6 +83,18 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
return 0;
 }
 
+static int
+idpf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu __rte_unused)
+{
+   /* mtu setting is forbidden if port is start */
+   if (dev->data->dev_started) {
+   PMD_DRV_LOG(ERR, "port must be stopped before configuration");
+   return -EBUSY;
+   }
+
+   return 0;
+}
+
 static int
 idpf_init_vport_req_info(struct rte_eth_dev *dev)
 {
@@ -760,6 +772,7 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.tx_queue_stop  = idpf_tx_queue_stop,
.rx_queue_release   = idpf_dev_rx_queue_release,
.tx_queue_release   = idpf_dev_tx_queue_release,
+   .mtu_set= idpf_dev_mtu_set,
 };
 
 static uint16_t
-- 
2.26.2



[PATCH v16 10/18] net/idpf: add support for basic Rx datapath

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add basic Rx support in split queue mode and single queue mode.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |   2 +
 drivers/net/idpf/idpf_rxtx.c   | 273 +
 drivers/net/idpf/idpf_rxtx.h   |   7 +-
 3 files changed, 281 insertions(+), 1 deletion(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 856f3d7266..2f1f95 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -348,6 +348,8 @@ idpf_dev_start(struct rte_eth_dev *dev)
goto err_mtu;
}
 
+   idpf_set_rx_function(dev);
+
ret = idpf_vc_ena_dis_vport(vport, true);
if (ret != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 053409b99a..ea499c4d37 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1208,3 +1208,276 @@ idpf_stop_queues(struct rte_eth_dev *dev)
PMD_DRV_LOG(WARNING, "Fail to stop Tx queue %d", i);
}
 }
+
+static void
+idpf_split_rx_bufq_refill(struct idpf_rx_queue *rx_bufq)
+{
+   volatile struct virtchnl2_splitq_rx_buf_desc *rx_buf_ring;
+   volatile struct virtchnl2_splitq_rx_buf_desc *rx_buf_desc;
+   uint16_t nb_refill = rx_bufq->rx_free_thresh;
+   uint16_t nb_desc = rx_bufq->nb_rx_desc;
+   uint16_t next_avail = rx_bufq->rx_tail;
+   struct rte_mbuf *nmb[rx_bufq->rx_free_thresh];
+   struct rte_eth_dev *dev;
+   uint64_t dma_addr;
+   uint16_t delta;
+   int i;
+
+   if (rx_bufq->nb_rx_hold < rx_bufq->rx_free_thresh)
+   return;
+
+   rx_buf_ring = rx_bufq->rx_ring;
+   delta = nb_desc - next_avail;
+   if (unlikely(delta < nb_refill)) {
+   if (likely(rte_pktmbuf_alloc_bulk(rx_bufq->mp, nmb, delta) == 
0)) {
+   for (i = 0; i < delta; i++) {
+   rx_buf_desc = &rx_buf_ring[next_avail + i];
+   rx_bufq->sw_ring[next_avail + i] = nmb[i];
+   dma_addr = 
rte_cpu_to_le_64(rte_mbuf_data_iova_default(nmb[i]));
+   rx_buf_desc->hdr_addr = 0;
+   rx_buf_desc->pkt_addr = dma_addr;
+   }
+   nb_refill -= delta;
+   next_avail = 0;
+   rx_bufq->nb_rx_hold -= delta;
+   } else {
+   dev = &rte_eth_devices[rx_bufq->port_id];
+   dev->data->rx_mbuf_alloc_failed += nb_desc - next_avail;
+   PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u 
queue_id=%u",
+  rx_bufq->port_id, rx_bufq->queue_id);
+   return;
+   }
+   }
+
+   if (nb_desc - next_avail >= nb_refill) {
+   if (likely(rte_pktmbuf_alloc_bulk(rx_bufq->mp, nmb, nb_refill) 
== 0)) {
+   for (i = 0; i < nb_refill; i++) {
+   rx_buf_desc = &rx_buf_ring[next_avail + i];
+   rx_bufq->sw_ring[next_avail + i] = nmb[i];
+   dma_addr = 
rte_cpu_to_le_64(rte_mbuf_data_iova_default(nmb[i]));
+   rx_buf_desc->hdr_addr = 0;
+   rx_buf_desc->pkt_addr = dma_addr;
+   }
+   next_avail += nb_refill;
+   rx_bufq->nb_rx_hold -= nb_refill;
+   } else {
+   dev = &rte_eth_devices[rx_bufq->port_id];
+   dev->data->rx_mbuf_alloc_failed += nb_desc - next_avail;
+   PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u 
queue_id=%u",
+  rx_bufq->port_id, rx_bufq->queue_id);
+   }
+   }
+
+   IDPF_PCI_REG_WRITE(rx_bufq->qrx_tail, next_avail);
+
+   rx_bufq->rx_tail = next_avail;
+}
+
+uint16_t
+idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
+ uint16_t nb_pkts)
+{
+   volatile struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc_ring;
+   volatile struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc;
+   uint16_t pktlen_gen_bufq_id;
+   struct idpf_rx_queue *rxq;
+   struct rte_mbuf *rxm;
+   uint16_t rx_id_bufq1;
+   uint16_t rx_id_bufq2;
+   uint16_t pkt_len;
+   uint16_t bufq_id;
+   uint16_t gen_id;
+   uint16_t rx_id;
+   uint16_t nb_rx;
+
+   nb_rx = 0;
+   rxq = rx_queue;
+
+   if (unlikely(rxq == NULL) || unlikely(!rxq->q_started))
+   return nb_rx;
+
+   rx_id = rxq->rx_tail;
+   rx_id_bufq1 = rxq->bufq1->rx_next_avail;
+   rx_id_bufq2 = rxq->bufq2->rx_next_avail;
+   rx_desc_ring = rxq->rx_ring;
+
+  

[PATCH v16 11/18] net/idpf: add support for basic Tx datapath

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add basic Tx support in split queue mode and single queue mode.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |   3 +
 drivers/net/idpf/idpf_ethdev.h |   1 +
 drivers/net/idpf/idpf_rxtx.c   | 357 +
 drivers/net/idpf/idpf_rxtx.h   |  10 +
 4 files changed, 371 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 2f1f95..f9f6fe1162 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -59,6 +59,8 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
dev_info->max_mtu = dev_info->max_rx_pktlen - IDPF_ETH_OVERHEAD;
dev_info->min_mtu = RTE_ETHER_MIN_MTU;
 
+   dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
+
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
.tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH,
@@ -349,6 +351,7 @@ idpf_dev_start(struct rte_eth_dev *dev)
}
 
idpf_set_rx_function(dev);
+   idpf_set_tx_function(dev);
 
ret = idpf_vc_ena_dis_vport(vport, true);
if (ret != 0) {
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 96c22009e9..af0a8e2970 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -35,6 +35,7 @@
 
 #define IDPF_MIN_BUF_SIZE  1024
 #define IDPF_MAX_FRAME_SIZE9728
+#define IDPF_MIN_FRAME_SIZE14
 
 #define IDPF_NUM_MACADDR_MAX   64
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index ea499c4d37..f55d2143b9 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1365,6 +1365,148 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
return nb_rx;
 }
 
+static inline void
+idpf_split_tx_free(struct idpf_tx_queue *cq)
+{
+   volatile struct idpf_splitq_tx_compl_desc *compl_ring = cq->compl_ring;
+   volatile struct idpf_splitq_tx_compl_desc *txd;
+   uint16_t next = cq->tx_tail;
+   struct idpf_tx_entry *txe;
+   struct idpf_tx_queue *txq;
+   uint16_t gen, qid, q_head;
+   uint8_t ctype;
+
+   txd = &compl_ring[next];
+   gen = (rte_le_to_cpu_16(txd->qid_comptype_gen) &
+   IDPF_TXD_COMPLQ_GEN_M) >> IDPF_TXD_COMPLQ_GEN_S;
+   if (gen != cq->expected_gen_id)
+   return;
+
+   ctype = (rte_le_to_cpu_16(txd->qid_comptype_gen) &
+   IDPF_TXD_COMPLQ_COMPL_TYPE_M) >> IDPF_TXD_COMPLQ_COMPL_TYPE_S;
+   qid = (rte_le_to_cpu_16(txd->qid_comptype_gen) &
+   IDPF_TXD_COMPLQ_QID_M) >> IDPF_TXD_COMPLQ_QID_S;
+   q_head = rte_le_to_cpu_16(txd->q_head_compl_tag.compl_tag);
+   txq = cq->txqs[qid - cq->tx_start_qid];
+
+   switch (ctype) {
+   case IDPF_TXD_COMPLT_RE:
+   if (q_head == 0)
+   txq->last_desc_cleaned = txq->nb_tx_desc - 1;
+   else
+   txq->last_desc_cleaned = q_head - 1;
+   if (unlikely((txq->last_desc_cleaned % 32) == 0)) {
+   PMD_DRV_LOG(ERR, "unexpected desc (head = %u) 
completion.",
+   q_head);
+   return;
+   }
+
+   break;
+   case IDPF_TXD_COMPLT_RS:
+   txq->nb_free++;
+   txq->nb_used--;
+   txe = &txq->sw_ring[q_head];
+   if (txe->mbuf != NULL) {
+   rte_pktmbuf_free_seg(txe->mbuf);
+   txe->mbuf = NULL;
+   }
+   break;
+   default:
+   PMD_DRV_LOG(ERR, "unknown completion type.");
+   return;
+   }
+
+   if (++next == cq->nb_tx_desc) {
+   next = 0;
+   cq->expected_gen_id ^= 1;
+   }
+
+   cq->tx_tail = next;
+}
+
+uint16_t
+idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
+ uint16_t nb_pkts)
+{
+   struct idpf_tx_queue *txq = (struct idpf_tx_queue *)tx_queue;
+   volatile struct idpf_flex_tx_sched_desc *txr;
+   volatile struct idpf_flex_tx_sched_desc *txd;
+   struct idpf_tx_entry *sw_ring;
+   struct idpf_tx_entry *txe, *txn;
+   uint16_t nb_used, tx_id, sw_id;
+   struct rte_mbuf *tx_pkt;
+   uint16_t nb_to_clean;
+   uint16_t nb_tx = 0;
+
+   if (unlikely(txq == NULL) || unlikely(!txq->q_started))
+   return nb_tx;
+
+   txr = txq->desc_ring;
+   sw_ring = txq->sw_ring;
+   tx_id = txq->tx_tail;
+   sw_id = txq->sw_tail;
+   txe = &sw_ring[sw_id];
+
+   for (nb_tx = 0; nb_tx < nb_pkts; nb_tx++) {
+   tx_pkt = tx_pkts[nb_tx];
+
+   if (txq->nb_free <= txq->free_thresh) {
+   /* TODO: Need to refine
+

[PATCH v16 12/18] net/idpf: support parsing packet type

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Parse packet type during receiving packets.

Signed-off-by: Wenjun Wu 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |   6 +
 drivers/net/idpf/idpf_ethdev.h |   6 +
 drivers/net/idpf/idpf_rxtx.c   |  11 ++
 drivers/net/idpf/idpf_rxtx.h   |   5 +
 drivers/net/idpf/idpf_vchnl.c  | 240 +
 5 files changed, 268 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index f9f6fe1162..d0821ec3f3 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -686,6 +686,12 @@ idpf_adapter_init(struct rte_pci_device *pci_dev, struct 
idpf_adapter *adapter)
goto err_api;
}
 
+   ret = idpf_get_pkt_type(adapter);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to set ptype table");
+   goto err_api;
+   }
+
adapter->caps = rte_zmalloc("idpf_caps",
sizeof(struct virtchnl2_get_capabilities), 0);
if (adapter->caps == NULL) {
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index af0a8e2970..db9af58f72 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -39,6 +39,8 @@
 
 #define IDPF_NUM_MACADDR_MAX   64
 
+#define IDPF_MAX_PKT_TYPE  1024
+
 #define IDPF_VLAN_TAG_SIZE 4
 #define IDPF_ETH_OVERHEAD \
(RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + IDPF_VLAN_TAG_SIZE * 2)
@@ -125,6 +127,8 @@ struct idpf_adapter {
/* Max config queue number per VC message */
uint32_t max_rxq_per_msg;
uint32_t max_txq_per_msg;
+
+   uint32_t ptype_tbl[IDPF_MAX_PKT_TYPE] __rte_cache_min_aligned;
 };
 
 TAILQ_HEAD(idpf_adapter_list, idpf_adapter);
@@ -182,6 +186,7 @@ atomic_set_cmd(struct idpf_adapter *adapter, enum 
virtchnl_ops ops)
 struct idpf_adapter *idpf_find_adapter(struct rte_pci_device *pci_dev);
 void idpf_handle_virtchnl_msg(struct rte_eth_dev *dev);
 int idpf_vc_check_api_version(struct idpf_adapter *adapter);
+int idpf_get_pkt_type(struct idpf_adapter *adapter);
 int idpf_vc_get_caps(struct idpf_adapter *adapter);
 int idpf_vc_create_vport(struct idpf_adapter *adapter);
 int idpf_vc_destroy_vport(struct idpf_vport *vport);
@@ -193,6 +198,7 @@ int idpf_switch_queue(struct idpf_vport *vport, uint16_t 
qid,
  bool rx, bool on);
 int idpf_vc_ena_dis_queues(struct idpf_vport *vport, bool enable);
 int idpf_vc_ena_dis_vport(struct idpf_vport *vport, bool enable);
+int idpf_vc_query_ptype_info(struct idpf_adapter *adapter);
 int idpf_read_one_msg(struct idpf_adapter *adapter, uint32_t ops,
  uint16_t buf_len, uint8_t *buf);
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index f55d2143b9..a980714060 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1281,6 +1281,7 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
volatile struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc;
uint16_t pktlen_gen_bufq_id;
struct idpf_rx_queue *rxq;
+   const uint32_t *ptype_tbl;
struct rte_mbuf *rxm;
uint16_t rx_id_bufq1;
uint16_t rx_id_bufq2;
@@ -1300,6 +1301,7 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rx_id_bufq1 = rxq->bufq1->rx_next_avail;
rx_id_bufq2 = rxq->bufq2->rx_next_avail;
rx_desc_ring = rxq->rx_ring;
+   ptype_tbl = rxq->adapter->ptype_tbl;
 
while (nb_rx < nb_pkts) {
rx_desc = &rx_desc_ring[rx_id];
@@ -1347,6 +1349,10 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rxm->next = NULL;
rxm->nb_segs = 1;
rxm->port = rxq->port_id;
+   rxm->packet_type =
+   ptype_tbl[(rte_le_to_cpu_16(rx_desc->ptype_err_fflags0) 
&
+  VIRTCHNL2_RX_FLEX_DESC_ADV_PTYPE_M) >>
+ VIRTCHNL2_RX_FLEX_DESC_ADV_PTYPE_S];
 
rx_pkts[nb_rx++] = rxm;
}
@@ -1533,6 +1539,7 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
volatile union virtchnl2_rx_desc *rxdp;
union virtchnl2_rx_desc rxd;
struct idpf_rx_queue *rxq;
+   const uint32_t *ptype_tbl;
uint16_t rx_id, nb_hold;
struct rte_eth_dev *dev;
uint16_t rx_packet_len;
@@ -1551,6 +1558,7 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
 
rx_id = rxq->rx_tail;
rx_ring = rxq->rx_ring;
+   ptype_tbl = rxq->adapter->ptype_tbl;
 
while (nb_rx < nb_pkts) {
rxdp = &rx_ring[rx_id];
@@ -1603,6 +1611,9 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rxm->pkt_len = rx_packet_len;
rxm->data_len = rx_packet_len;
rxm->port = rxq->port_id;
+   rxm->packet_type =
+   

[PATCH v16 13/18] net/idpf: add support for write back based on ITR expire

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Enable write back on ITR expire, then packets can be received one by
one.

Signed-off-by: Beilei Xing 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c | 120 +
 drivers/net/idpf/idpf_ethdev.h |  13 
 drivers/net/idpf/idpf_vchnl.c  | 113 +++
 3 files changed, 246 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index d0821ec3f3..957cc10616 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -297,6 +297,90 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_config_rx_queues_irqs(struct rte_eth_dev *dev)
+{
+   struct idpf_vport *vport = dev->data->dev_private;
+   struct idpf_adapter *adapter = vport->adapter;
+   struct virtchnl2_queue_vector *qv_map;
+   struct idpf_hw *hw = &adapter->hw;
+   uint32_t dynctl_reg_start;
+   uint32_t itrn_reg_start;
+   uint32_t dynctl_val, itrn_val;
+   uint16_t i;
+
+   qv_map = rte_zmalloc("qv_map",
+   dev->data->nb_rx_queues *
+   sizeof(struct virtchnl2_queue_vector), 0);
+   if (qv_map == NULL) {
+   PMD_DRV_LOG(ERR, "Failed to allocate %d queue-vector map",
+   dev->data->nb_rx_queues);
+   goto qv_map_alloc_err;
+   }
+
+   /* Rx interrupt disabled, Map interrupt only for writeback */
+
+   /* The capability flags adapter->caps->other_caps should be
+* compared with bit VIRTCHNL2_CAP_WB_ON_ITR here. The if
+* condition should be updated when the FW can return the
+* correct flag bits.
+*/
+   dynctl_reg_start =
+   vport->recv_vectors->vchunks.vchunks->dynctl_reg_start;
+   itrn_reg_start =
+   vport->recv_vectors->vchunks.vchunks->itrn_reg_start;
+   dynctl_val = IDPF_READ_REG(hw, dynctl_reg_start);
+   PMD_DRV_LOG(DEBUG, "Value of dynctl_reg_start is 0x%x",
+   dynctl_val);
+   itrn_val = IDPF_READ_REG(hw, itrn_reg_start);
+   PMD_DRV_LOG(DEBUG, "Value of itrn_reg_start is 0x%x", itrn_val);
+   /* Force write-backs by setting WB_ON_ITR bit in DYN_CTL
+* register. WB_ON_ITR and INTENA are mutually exclusive
+* bits. Setting WB_ON_ITR bits means TX and RX Descs
+* are written back based on ITR expiration irrespective
+* of INTENA setting.
+*/
+   /* TBD: need to tune INTERVAL value for better performance. */
+   if (itrn_val != 0)
+   IDPF_WRITE_REG(hw,
+  dynctl_reg_start,
+  VIRTCHNL2_ITR_IDX_0  <<
+  PF_GLINT_DYN_CTL_ITR_INDX_S |
+  PF_GLINT_DYN_CTL_WB_ON_ITR_M |
+  itrn_val <<
+  PF_GLINT_DYN_CTL_INTERVAL_S);
+   else
+   IDPF_WRITE_REG(hw,
+  dynctl_reg_start,
+  VIRTCHNL2_ITR_IDX_0  <<
+  PF_GLINT_DYN_CTL_ITR_INDX_S |
+  PF_GLINT_DYN_CTL_WB_ON_ITR_M |
+  IDPF_DFLT_INTERVAL <<
+  PF_GLINT_DYN_CTL_INTERVAL_S);
+
+   for (i = 0; i < dev->data->nb_rx_queues; i++) {
+   /* map all queues to the same vector */
+   qv_map[i].queue_id = vport->chunks_info.rx_start_qid + i;
+   qv_map[i].vector_id =
+   vport->recv_vectors->vchunks.vchunks->start_vector_id;
+   }
+   vport->qv_map = qv_map;
+
+   if (idpf_vc_config_irq_map_unmap(vport, true) != 0) {
+   PMD_DRV_LOG(ERR, "config interrupt mapping failed");
+   goto config_irq_map_err;
+   }
+
+   return 0;
+
+config_irq_map_err:
+   rte_free(vport->qv_map);
+   vport->qv_map = NULL;
+
+qv_map_alloc_err:
+   return -1;
+}
+
 static int
 idpf_start_queues(struct rte_eth_dev *dev)
 {
@@ -334,6 +418,10 @@ static int
 idpf_dev_start(struct rte_eth_dev *dev)
 {
struct idpf_vport *vport = dev->data->dev_private;
+   struct idpf_adapter *adapter = vport->adapter;
+   uint16_t num_allocated_vectors =
+   adapter->caps->num_allocated_vectors;
+   uint16_t req_vecs_num;
int ret;
 
if (dev->data->mtu > vport->max_mtu) {
@@ -344,6 +432,27 @@ idpf_dev_start(struct rte_eth_dev *dev)
 
vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
 
+   req_vecs_num = IDPF_DFLT_Q_VEC_NUM;
+   if (req_vecs_num + adapter->used_vecs_num > num_allocated_vectors) {
+   PMD_DRV_LOG(ERR, "The accumulated request vectors' number 
should be less than %d",
+   num_allocated_vectors);
+   ret = -EINVAL;
+   goto err_mtu;
+   }
+
+   ret = idpf_vc_alloc_vectors(vpo

[PATCH v16 14/18] net/idpf: add support for RSS

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add RSS support.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c | 120 -
 drivers/net/idpf/idpf_ethdev.h |  26 +++
 drivers/net/idpf/idpf_vchnl.c  | 113 +++
 3 files changed, 258 insertions(+), 1 deletion(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 957cc10616..58560ea404 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -59,6 +59,8 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
dev_info->max_mtu = dev_info->max_rx_pktlen - IDPF_ETH_OVERHEAD;
dev_info->min_mtu = RTE_ETHER_MIN_MTU;
 
+   dev_info->flow_type_rss_offloads = IDPF_RSS_OFFLOAD_ALL;
+
dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
@@ -169,6 +171,8 @@ idpf_parse_devarg_id(char *name)
return val;
 }
 
+#define IDPF_RSS_KEY_LEN 52
+
 static int
 idpf_init_vport(struct rte_eth_dev *dev)
 {
@@ -189,6 +193,10 @@ idpf_init_vport(struct rte_eth_dev *dev)
vport->max_mtu = vport_info->max_mtu;
rte_memcpy(vport->default_mac_addr,
   vport_info->default_mac_addr, ETH_ALEN);
+   vport->rss_algorithm = vport_info->rss_algorithm;
+   vport->rss_key_size = RTE_MIN(IDPF_RSS_KEY_LEN,
+vport_info->rss_key_size);
+   vport->rss_lut_size = vport_info->rss_lut_size;
vport->sw_idx = idx;
 
for (i = 0; i < vport_info->chunks.num_chunks; i++) {
@@ -246,17 +254,110 @@ idpf_init_vport(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_config_rss(struct idpf_vport *vport)
+{
+   int ret;
+
+   ret = idpf_vc_set_rss_key(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS key");
+   return ret;
+   }
+
+   ret = idpf_vc_set_rss_lut(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS lut");
+   return ret;
+   }
+
+   ret = idpf_vc_set_rss_hash(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS hash");
+   return ret;
+   }
+
+   return ret;
+}
+
+static int
+idpf_init_rss(struct idpf_vport *vport)
+{
+   struct rte_eth_rss_conf *rss_conf;
+   uint16_t i, nb_q, lut_size;
+   int ret = 0;
+
+   rss_conf = &vport->dev_data->dev_conf.rx_adv_conf.rss_conf;
+   nb_q = vport->dev_data->nb_rx_queues;
+
+   vport->rss_key = rte_zmalloc("rss_key",
+vport->rss_key_size, 0);
+   if (vport->rss_key == NULL) {
+   PMD_INIT_LOG(ERR, "Failed to allocate RSS key");
+   ret = -ENOMEM;
+   goto err_alloc_key;
+   }
+
+   lut_size = vport->rss_lut_size;
+   vport->rss_lut = rte_zmalloc("rss_lut",
+sizeof(uint32_t) * lut_size, 0);
+   if (vport->rss_lut == NULL) {
+   PMD_INIT_LOG(ERR, "Failed to allocate RSS lut");
+   ret = -ENOMEM;
+   goto err_alloc_lut;
+   }
+
+   if (rss_conf->rss_key == NULL) {
+   for (i = 0; i < vport->rss_key_size; i++)
+   vport->rss_key[i] = (uint8_t)rte_rand();
+   } else if (rss_conf->rss_key_len != vport->rss_key_size) {
+   PMD_INIT_LOG(ERR, "Invalid RSS key length in RSS configuration, 
should be %d",
+vport->rss_key_size);
+   ret = -EINVAL;
+   goto err_cfg_key;
+   } else {
+   rte_memcpy(vport->rss_key, rss_conf->rss_key,
+  vport->rss_key_size);
+   }
+
+   for (i = 0; i < lut_size; i++)
+   vport->rss_lut[i] = i % nb_q;
+
+   vport->rss_hf = IDPF_DEFAULT_RSS_HASH_EXPANDED;
+
+   ret = idpf_config_rss(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS");
+   goto err_cfg_key;
+   }
+
+   return ret;
+
+err_cfg_key:
+   rte_free(vport->rss_lut);
+   vport->rss_lut = NULL;
+err_alloc_lut:
+   rte_free(vport->rss_key);
+   vport->rss_key = NULL;
+err_alloc_key:
+   return ret;
+}
+
 static int
 idpf_dev_configure(struct rte_eth_dev *dev)
 {
+   struct idpf_vport *vport = dev->data->dev_private;
struct rte_eth_conf *conf = &dev->data->dev_conf;
+   struct idpf_adapter *adapter = vport->adapter;
+   int ret;
 
if (conf->link_speeds & RTE_ETH_LINK_SPEED_FIXED) {
PMD_INIT_LOG(ERR, "Setting link speed is not supported");
return -ENOTSUP;
}
 
-   if (dev->data->nb_rx_queues == 1 && conf->rxmode.mq_mode != 
RTE_ETH_MQ_RX_NONE) {
+   if ((dev->data->nb_rx_queues == 1 && conf->rxmode.mq_mode != 
R

[PATCH v16 15/18] net/idpf: add support for Rx offloading

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add Rx offloading support:
 - support CHKSUM and RSS offload for split queue model
 - support CHKSUM offload for single queue model

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |   5 ++
 drivers/net/idpf/idpf_ethdev.c|   6 ++
 drivers/net/idpf/idpf_rxtx.c  | 123 ++
 drivers/net/idpf/idpf_vchnl.c |  18 +
 4 files changed, 152 insertions(+)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index d722c49fde..868571654f 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -3,8 +3,13 @@
 ;
 ; Refer to default.ini for the full list of available PMD features.
 ;
+; A feature with "P" indicates only be supported when non-vector path
+; is selected.
+;
 [Features]
 MTU update   = Y
+L3 checksum offload  = P
+L4 checksum offload  = P
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 58560ea404..a09f104425 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -61,6 +61,12 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
 
dev_info->flow_type_rss_offloads = IDPF_RSS_OFFLOAD_ALL;
 
+   dev_info->rx_offload_capa =
+   RTE_ETH_RX_OFFLOAD_IPV4_CKSUM   |
+   RTE_ETH_RX_OFFLOAD_UDP_CKSUM|
+   RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
+   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM;
+
dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index a980714060..f15e61a785 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1209,6 +1209,73 @@ idpf_stop_queues(struct rte_eth_dev *dev)
}
 }
 
+#define IDPF_RX_FLEX_DESC_ADV_STATUS0_XSUM_S   \
+   (RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_IPE_S) | \
+RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_L4E_S) | \
+RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EIPE_S) |\
+RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EUDPE_S))
+
+static inline uint64_t
+idpf_splitq_rx_csum_offload(uint8_t err)
+{
+   uint64_t flags = 0;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_L3L4P_S)) == 0))
+   return flags;
+
+   if (likely((err & IDPF_RX_FLEX_DESC_ADV_STATUS0_XSUM_S) == 0)) {
+   flags |= (RTE_MBUF_F_RX_IP_CKSUM_GOOD |
+ RTE_MBUF_F_RX_L4_CKSUM_GOOD);
+   return flags;
+   }
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_IPE_S)) != 0))
+   flags |= RTE_MBUF_F_RX_IP_CKSUM_BAD;
+   else
+   flags |= RTE_MBUF_F_RX_IP_CKSUM_GOOD;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_L4E_S)) != 0))
+   flags |= RTE_MBUF_F_RX_L4_CKSUM_BAD;
+   else
+   flags |= RTE_MBUF_F_RX_L4_CKSUM_GOOD;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EIPE_S)) != 0))
+   flags |= RTE_MBUF_F_RX_OUTER_IP_CKSUM_BAD;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EUDPE_S)) != 0))
+   flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_BAD;
+   else
+   flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_GOOD;
+
+   return flags;
+}
+
+#define IDPF_RX_FLEX_DESC_ADV_HASH1_S  0
+#define IDPF_RX_FLEX_DESC_ADV_HASH2_S  16
+#define IDPF_RX_FLEX_DESC_ADV_HASH3_S  24
+
+static inline uint64_t
+idpf_splitq_rx_rss_offload(struct rte_mbuf *mb,
+  volatile struct virtchnl2_rx_flex_desc_adv_nic_3 
*rx_desc)
+{
+   uint8_t status_err0_qw0;
+   uint64_t flags = 0;
+
+   status_err0_qw0 = rx_desc->status_err0_qw0;
+
+   if ((status_err0_qw0 & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_RSS_VALID_S)) != 0) {
+   flags |= RTE_MBUF_F_RX_RSS_HASH;
+   mb->hash.rss = (rte_le_to_cpu_16(rx_desc->hash1) <<
+   IDPF_RX_FLEX_DESC_ADV_HASH1_S) |
+   ((uint32_t)(rx_desc->ff2_mirrid_hash2.hash2) <<
+IDPF_RX_FLEX_DESC_ADV_HASH2_S) |
+   ((uint32_t)(rx_desc->hash3) <<
+IDPF_RX_FLEX_DESC_ADV_HASH3_S);
+   }
+
+   return flags;
+}
+
 static void
 idpf_split_rx_bufq_refill(struct idpf_rx_queue *rx_bufq)
 {
@@ -1282,9 +1349,11 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
uint16_t pktlen_gen_bufq_id;
struct idpf_rx_queue *rxq;
const uint32_t *ptype_tbl;
+   uint8_t status_err0_qw1;
struct rte_mb

[PATCH v16 16/18] net/idpf: add support for Tx offloading

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add Tx offloading support:
 - support TSO for single queue model and split queue model.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |   1 +
 drivers/net/idpf/idpf_ethdev.c|   4 +-
 drivers/net/idpf/idpf_rxtx.c  | 128 +-
 drivers/net/idpf/idpf_rxtx.h  |  22 +
 4 files changed, 152 insertions(+), 3 deletions(-)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index 868571654f..d82b4aa0ff 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -8,6 +8,7 @@
 ;
 [Features]
 MTU update   = Y
+TSO  = P
 L3 checksum offload  = P
 L4 checksum offload  = P
 Linux= Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index a09f104425..084426260c 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -67,7 +67,9 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM;
 
-   dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
+   dev_info->tx_offload_capa =
+   RTE_ETH_TX_OFFLOAD_TCP_TSO  |
+   RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index f15e61a785..cc296d7ab1 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1506,6 +1506,49 @@ idpf_split_tx_free(struct idpf_tx_queue *cq)
cq->tx_tail = next;
 }
 
+/* Check if the context descriptor is needed for TX offloading */
+static inline uint16_t
+idpf_calc_context_desc(uint64_t flags)
+{
+   if ((flags & RTE_MBUF_F_TX_TCP_SEG) != 0)
+   return 1;
+
+   return 0;
+}
+
+/* set TSO context descriptor
+ */
+static inline void
+idpf_set_splitq_tso_ctx(struct rte_mbuf *mbuf,
+   union idpf_tx_offload tx_offload,
+   volatile union idpf_flex_tx_ctx_desc *ctx_desc)
+{
+   uint16_t cmd_dtype;
+   uint32_t tso_len;
+   uint8_t hdr_len;
+
+   if (tx_offload.l4_len == 0) {
+   PMD_TX_LOG(DEBUG, "L4 length set to 0");
+   return;
+   }
+
+   hdr_len = tx_offload.l2_len +
+   tx_offload.l3_len +
+   tx_offload.l4_len;
+   cmd_dtype = IDPF_TX_DESC_DTYPE_FLEX_TSO_CTX |
+   IDPF_TX_FLEX_CTX_DESC_CMD_TSO;
+   tso_len = mbuf->pkt_len - hdr_len;
+
+   ctx_desc->tso.qw1.cmd_dtype = rte_cpu_to_le_16(cmd_dtype);
+   ctx_desc->tso.qw0.hdr_len = hdr_len;
+   ctx_desc->tso.qw0.mss_rt =
+   rte_cpu_to_le_16((uint16_t)mbuf->tso_segsz &
+IDPF_TXD_FLEX_CTX_MSS_RT_M);
+   ctx_desc->tso.qw0.flex_tlen =
+   rte_cpu_to_le_32(tso_len &
+IDPF_TXD_FLEX_CTX_MSS_RT_M);
+}
+
 uint16_t
 idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
  uint16_t nb_pkts)
@@ -1514,11 +1557,14 @@ idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
volatile struct idpf_flex_tx_sched_desc *txr;
volatile struct idpf_flex_tx_sched_desc *txd;
struct idpf_tx_entry *sw_ring;
+   union idpf_tx_offload tx_offload = {0};
struct idpf_tx_entry *txe, *txn;
uint16_t nb_used, tx_id, sw_id;
struct rte_mbuf *tx_pkt;
uint16_t nb_to_clean;
uint16_t nb_tx = 0;
+   uint64_t ol_flags;
+   uint16_t nb_ctx;
 
if (unlikely(txq == NULL) || unlikely(!txq->q_started))
return nb_tx;
@@ -1548,7 +1594,29 @@ idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
 
if (txq->nb_free < tx_pkt->nb_segs)
break;
-   nb_used = tx_pkt->nb_segs;
+
+   ol_flags = tx_pkt->ol_flags;
+   tx_offload.l2_len = tx_pkt->l2_len;
+   tx_offload.l3_len = tx_pkt->l3_len;
+   tx_offload.l4_len = tx_pkt->l4_len;
+   tx_offload.tso_segsz = tx_pkt->tso_segsz;
+   /* Calculate the number of context descriptors needed. */
+   nb_ctx = idpf_calc_context_desc(ol_flags);
+   nb_used = tx_pkt->nb_segs + nb_ctx;
+
+   /* context descriptor */
+   if (nb_ctx != 0) {
+   volatile union idpf_flex_tx_ctx_desc *ctx_desc =
+   (volatile union idpf_flex_tx_ctx_desc *)&txr[tx_id];
+
+   if ((ol_flags & RTE_MBUF_F_TX_TCP_SEG) != 0)
+   idpf_set_splitq_tso_ctx(tx_pkt, tx_offload,
+   

[PATCH v16 17/18] net/idpf: add AVX512 data path for single queue model

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support of AVX512 vector data path for single queue model.

Signed-off-by: Wenjun Wu 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/idpf.rst|  19 +
 drivers/net/idpf/idpf_ethdev.c  |   3 +-
 drivers/net/idpf/idpf_ethdev.h  |   5 +
 drivers/net/idpf/idpf_rxtx.c| 145 
 drivers/net/idpf/idpf_rxtx.h|  21 +
 drivers/net/idpf/idpf_rxtx_vec_avx512.c | 871 
 drivers/net/idpf/idpf_rxtx_vec_common.h | 100 +++
 drivers/net/idpf/meson.build|  28 +
 8 files changed, 1191 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/idpf/idpf_rxtx_vec_avx512.c
 create mode 100644 drivers/net/idpf/idpf_rxtx_vec_common.h

diff --git a/doc/guides/nics/idpf.rst b/doc/guides/nics/idpf.rst
index c1001d5d0c..3039c61748 100644
--- a/doc/guides/nics/idpf.rst
+++ b/doc/guides/nics/idpf.rst
@@ -64,3 +64,22 @@ Refer to the document :ref:`compiling and testing a PMD for 
a NIC tx_offload_capa =
RTE_ETH_TX_OFFLOAD_TCP_TSO  |
-   RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
+   RTE_ETH_TX_OFFLOAD_MULTI_SEGS   |
+   RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 8d0804f603..7d54e5db60 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -162,6 +162,11 @@ struct idpf_adapter {
uint32_t max_txq_per_msg;
 
uint32_t ptype_tbl[IDPF_MAX_PKT_TYPE] __rte_cache_min_aligned;
+
+   bool rx_vec_allowed;
+   bool tx_vec_allowed;
+   bool rx_use_avx512;
+   bool tx_use_avx512;
 };
 
 TAILQ_HEAD(idpf_adapter_list, idpf_adapter);
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index cc296d7ab1..9e20f2b9d3 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -4,9 +4,11 @@
 
 #include 
 #include 
+#include 
 
 #include "idpf_ethdev.h"
 #include "idpf_rxtx.h"
+#include "idpf_rxtx_vec_common.h"
 
 static int
 check_rx_thresh(uint16_t nb_desc, uint16_t thresh)
@@ -252,6 +254,8 @@ reset_single_rx_queue(struct idpf_rx_queue *rxq)
 
rxq->pkt_first_seg = NULL;
rxq->pkt_last_seg = NULL;
+   rxq->rxrearm_start = 0;
+   rxq->rxrearm_nb = 0;
 }
 
 static void
@@ -2073,25 +2077,166 @@ idpf_prep_pkts(__rte_unused void *tx_queue, struct 
rte_mbuf **tx_pkts,
return i;
 }
 
+static void __rte_cold
+release_rxq_mbufs_vec(struct idpf_rx_queue *rxq)
+{
+   const uint16_t mask = rxq->nb_rx_desc - 1;
+   uint16_t i;
+
+   if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc)
+   return;
+
+   /* free all mbufs that are valid in the ring */
+   if (rxq->rxrearm_nb == 0) {
+   for (i = 0; i < rxq->nb_rx_desc; i++) {
+   if (rxq->sw_ring[i] != NULL)
+   rte_pktmbuf_free_seg(rxq->sw_ring[i]);
+   }
+   } else {
+   for (i = rxq->rx_tail; i != rxq->rxrearm_start; i = (i + 1) & 
mask) {
+   if (rxq->sw_ring[i] != NULL)
+   rte_pktmbuf_free_seg(rxq->sw_ring[i]);
+   }
+   }
+
+   rxq->rxrearm_nb = rxq->nb_rx_desc;
+
+   /* set all entries to NULL */
+   memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc);
+}
+
+static const struct idpf_rxq_ops def_singleq_rx_ops_vec = {
+   .release_mbufs = release_rxq_mbufs_vec,
+};
+
+static inline int
+idpf_singleq_rx_vec_setup_default(struct idpf_rx_queue *rxq)
+{
+   uintptr_t p;
+   struct rte_mbuf mb_def = { .buf_addr = 0 }; /* zeroed mbuf */
+
+   mb_def.nb_segs = 1;
+   mb_def.data_off = RTE_PKTMBUF_HEADROOM;
+   mb_def.port = rxq->port_id;
+   rte_mbuf_refcnt_set(&mb_def, 1);
+
+   /* prevent compiler reordering: rearm_data covers previous fields */
+   rte_compiler_barrier();
+   p = (uintptr_t)&mb_def.rearm_data;
+   rxq->mbuf_initializer = *(uint64_t *)p;
+   return 0;
+}
+
+int __rte_cold
+idpf_singleq_rx_vec_setup(struct idpf_rx_queue *rxq)
+{
+   rxq->ops = &def_singleq_rx_ops_vec;
+   return idpf_singleq_rx_vec_setup_default(rxq);
+}
+
 void
 idpf_set_rx_function(struct rte_eth_dev *dev)
 {
struct idpf_vport *vport = dev->data->dev_private;
+#ifdef RTE_ARCH_X86
+   struct idpf_adapter *ad = vport->adapter;
+   struct idpf_rx_queue *rxq;
+   int i;
+
+   if (idpf_rx_vec_dev_check_default(dev) == IDPF_VECTOR_PATH &&
+   rte_vect_get_max_simd_bitwidth() >= RTE_VECT_SIMD_128) {
+   ad->rx_vec_allowed = true;
+
+   if (rte_vect_get_max_simd_bitwidth() >= RTE_VECT_SIMD_512)
+#ifdef CC_AVX512_SUPPORT
+   if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_AVX512F) == 1 
&&
+  

[PATCH v16 18/18] net/idpf: add support for timestamp offload

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for timestamp offload.

Signed-off-by: Wenjing Qiao 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |  1 +
 drivers/net/idpf/idpf_ethdev.c|  5 +-
 drivers/net/idpf/idpf_ethdev.h|  3 ++
 drivers/net/idpf/idpf_rxtx.c  | 65 ++
 drivers/net/idpf/idpf_rxtx.h  | 90 +++
 5 files changed, 163 insertions(+), 1 deletion(-)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index d82b4aa0ff..099fd7f216 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -11,6 +11,7 @@ MTU update   = Y
 TSO  = P
 L3 checksum offload  = P
 L4 checksum offload  = P
+Timestamp offload= P
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index cd4ebcc2c6..50aac65daf 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -22,6 +22,8 @@ rte_spinlock_t idpf_adapter_lock;
 struct idpf_adapter_list idpf_adapter_list;
 bool idpf_adapter_list_init;
 
+uint64_t idpf_timestamp_dynflag;
+
 static const char * const idpf_valid_args[] = {
IDPF_TX_SINGLE_Q,
IDPF_RX_SINGLE_Q,
@@ -65,7 +67,8 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
RTE_ETH_RX_OFFLOAD_IPV4_CKSUM   |
RTE_ETH_RX_OFFLOAD_UDP_CKSUM|
RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
-   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM;
+   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM |
+   RTE_ETH_RX_OFFLOAD_TIMESTAMP;
 
dev_info->tx_offload_capa =
RTE_ETH_TX_OFFLOAD_TCP_TSO  |
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 7d54e5db60..ccdf4abe40 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -167,6 +167,9 @@ struct idpf_adapter {
bool tx_vec_allowed;
bool rx_use_avx512;
bool tx_use_avx512;
+
+   /* For PTP */
+   uint64_t time_hw;
 };
 
 TAILQ_HEAD(idpf_adapter_list, idpf_adapter);
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 9e20f2b9d3..bafa007faf 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -10,6 +10,8 @@
 #include "idpf_rxtx.h"
 #include "idpf_rxtx_vec_common.h"
 
+static int idpf_timestamp_dynfield_offset = -1;
+
 static int
 check_rx_thresh(uint16_t nb_desc, uint16_t thresh)
 {
@@ -900,6 +902,24 @@ idpf_tx_queue_setup(struct rte_eth_dev *dev, uint16_t 
queue_idx,
return idpf_tx_split_queue_setup(dev, queue_idx, nb_desc,
 socket_id, tx_conf);
 }
+
+static int
+idpf_register_ts_mbuf(struct idpf_rx_queue *rxq)
+{
+   int err;
+   if ((rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) != 0) {
+   /* Register mbuf field and flag for Rx timestamp */
+   err = 
rte_mbuf_dyn_rx_timestamp_register(&idpf_timestamp_dynfield_offset,
+
&idpf_timestamp_dynflag);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR,
+   "Cannot register mbuf field/flag for 
timestamp");
+   return -EINVAL;
+   }
+   }
+   return 0;
+}
+
 static int
 idpf_alloc_single_rxq_mbufs(struct idpf_rx_queue *rxq)
 {
@@ -993,6 +1013,13 @@ idpf_rx_queue_init(struct rte_eth_dev *dev, uint16_t 
rx_queue_id)
return -EINVAL;
}
 
+   err = idpf_register_ts_mbuf(rxq);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR, "fail to regidter timestamp mbuf %u",
+   rx_queue_id);
+   return -EIO;
+   }
+
if (rxq->bufq1 == NULL) {
/* Single queue */
err = idpf_alloc_single_rxq_mbufs(rxq);
@@ -1354,6 +1381,7 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
struct idpf_rx_queue *rxq;
const uint32_t *ptype_tbl;
uint8_t status_err0_qw1;
+   struct idpf_adapter *ad;
struct rte_mbuf *rxm;
uint16_t rx_id_bufq1;
uint16_t rx_id_bufq2;
@@ -1363,9 +1391,11 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
uint16_t gen_id;
uint16_t rx_id;
uint16_t nb_rx;
+   uint64_t ts_ns;
 
nb_rx = 0;
rxq = rx_queue;
+   ad = rxq->adapter;
 
if (unlikely(rxq == NULL) || unlikely(!rxq->q_started))
return nb_rx;
@@ -1376,6 +1406,9 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rx_desc_ring = rxq->rx_ring;
ptype_tbl = rxq->adapter->ptype_tbl;
 
+   if ((rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) != 0)
+   rxq->hw_register

[PATCH] net/gve: fix pointers dereference before null check

2022-10-30 Thread Junfeng Guo
The pointers 'rxq' and 'txq' are dereferenced before the null check.
Fixed the logic in this patch.

Fixes: 4bec2d0b5572 ("net/gve: support queue operations")

Signed-off-by: Junfeng Guo 
---
 drivers/net/gve/gve_rx.c | 3 ++-
 drivers/net/gve/gve_tx.c | 3 ++-
 2 files changed, 4 insertions(+), 2 deletions(-)

diff --git a/drivers/net/gve/gve_rx.c b/drivers/net/gve/gve_rx.c
index ea397d68fa..4c5b8c517d 100644
--- a/drivers/net/gve/gve_rx.c
+++ b/drivers/net/gve/gve_rx.c
@@ -150,7 +150,7 @@ gve_rx_burst(void *rx_queue, struct rte_mbuf **rx_pkts, 
uint16_t nb_pkts)
 static inline void
 gve_reset_rxq(struct gve_rx_queue *rxq)
 {
-   struct rte_mbuf **sw_ring = rxq->sw_ring;
+   struct rte_mbuf **sw_ring;
uint32_t size, i;
 
if (rxq == NULL) {
@@ -166,6 +166,7 @@ gve_reset_rxq(struct gve_rx_queue *rxq)
for (i = 0; i < size; i++)
((volatile char *)rxq->rx_data_ring)[i] = 0;
 
+   sw_ring = rxq->sw_ring;
for (i = 0; i < rxq->nb_rx_desc; i++)
sw_ring[i] = NULL;
 
diff --git a/drivers/net/gve/gve_tx.c b/drivers/net/gve/gve_tx.c
index cd0bdaa2ad..4420a17192 100644
--- a/drivers/net/gve/gve_tx.c
+++ b/drivers/net/gve/gve_tx.c
@@ -463,7 +463,7 @@ gve_tx_burst(void *tx_queue, struct rte_mbuf **tx_pkts, 
uint16_t nb_pkts)
 static inline void
 gve_reset_txq(struct gve_tx_queue *txq)
 {
-   struct rte_mbuf **sw_ring = txq->sw_ring;
+   struct rte_mbuf **sw_ring;
uint32_t size, i;
 
if (txq == NULL) {
@@ -475,6 +475,7 @@ gve_reset_txq(struct gve_tx_queue *txq)
for (i = 0; i < size; i++)
((volatile char *)txq->tx_desc_ring)[i] = 0;
 
+   sw_ring = txq->sw_ring;
for (i = 0; i < txq->nb_tx_desc; i++) {
sw_ring[i] = NULL;
if (txq->is_gqi_qpl) {
-- 
2.34.1



[PATCH v17 00/18] add support for idpf PMD in DPDK

2022-10-30 Thread beilei . xing
From: Beilei Xing 

This patchset introduced the idpf (Infrastructure Data Path Function) PMD in 
DPDK for Intel® IPU E2000 (Device ID: 0x1452).
The Intel® IPU E2000 targets to deliver high performance under real workloads 
with security and isolation.
Please refer to
https://www.intel.com/content/www/us/en/products/network-io/infrastructure-processing-units/asic/e2000-asic.html
for more information.

Linux upstream is still ongoing, previous work refers to 
https://patchwork.ozlabs.org/project/intel-wired-lan/patch/20220128001009.721392-20-alan.br...@intel.com/.

v2-v4:
fixed some coding style issues and did some refactors.

v5:
fixed typo.

v6-v9:
fixed build errors and coding style issues.

v11:
 - move shared code to common/idpf/base
 - Create one vport if there's no vport devargs
 - Refactor if conditions according to coding style
 - Refactor virtual channel return values
 - Refine dev_stop function
 - Refine RSS lut/key
 - Fix build error

v12:
 - Refine dev_configure
 - Fix coding style according to the comments
 - Re-order patch
 - Romove dev_supported_ptypes_get

v13:
 - refine dev_start/stop and queue_start/stop
 - fix timestamp offload

v14:
 - fix wrong position for rte_validate_tx_offload

v15:
 - refine the return value for ethdev ops.
 - removce forward static declarations.
 - refine get caps.
 - fix lock/unlock handling.

v16:
 - refine errno in shared code
 - remove the conditional compilation IDPF_RX_PTYPE_OFFLOAD

v17:
 - fix build error on FreeBSD

Junfeng Guo (18):
  common/idpf: introduce common library
  net/idpf: add support for device initialization
  net/idpf: add Tx queue setup
  net/idpf: add Rx queue setup
  net/idpf: add support for device start and stop
  net/idpf: add support for queue start
  net/idpf: add support for queue stop
  net/idpf: add queue release
  net/idpf: add support for MTU configuration
  net/idpf: add support for basic Rx datapath
  net/idpf: add support for basic Tx datapath
  net/idpf: support parsing packet type
  net/idpf: add support for write back based on ITR expire
  net/idpf: add support for RSS
  net/idpf: add support for Rx offloading
  net/idpf: add support for Tx offloading
  net/idpf: add AVX512 data path for single queue model
  net/idpf: add support for timestamp offload

 MAINTAINERS   |9 +
 doc/guides/nics/features/idpf.ini |   17 +
 doc/guides/nics/idpf.rst  |   85 +
 doc/guides/nics/index.rst |1 +
 doc/guides/rel_notes/release_22_11.rst|6 +
 drivers/common/idpf/base/idpf_alloc.h |   22 +
 drivers/common/idpf/base/idpf_common.c|  364 +++
 drivers/common/idpf/base/idpf_controlq.c  |  691 
 drivers/common/idpf/base/idpf_controlq.h  |  224 ++
 drivers/common/idpf/base/idpf_controlq_api.h  |  207 ++
 .../common/idpf/base/idpf_controlq_setup.c|  179 +
 drivers/common/idpf/base/idpf_devids.h|   18 +
 drivers/common/idpf/base/idpf_lan_pf_regs.h   |  134 +
 drivers/common/idpf/base/idpf_lan_txrx.h  |  428 +++
 drivers/common/idpf/base/idpf_lan_vf_regs.h   |  114 +
 drivers/common/idpf/base/idpf_osdep.h |  364 +++
 drivers/common/idpf/base/idpf_prototype.h |   45 +
 drivers/common/idpf/base/idpf_type.h  |  106 +
 drivers/common/idpf/base/meson.build  |   14 +
 drivers/common/idpf/base/siov_regs.h  |   47 +
 drivers/common/idpf/base/virtchnl.h   | 2866 +
 drivers/common/idpf/base/virtchnl2.h  | 1462 +
 drivers/common/idpf/base/virtchnl2_lan_desc.h |  606 
 .../common/idpf/base/virtchnl_inline_ipsec.h  |  567 
 drivers/common/idpf/meson.build   |4 +
 drivers/common/idpf/version.map   |   12 +
 drivers/common/meson.build|1 +
 drivers/net/idpf/idpf_ethdev.c| 1293 
 drivers/net/idpf/idpf_ethdev.h|  252 ++
 drivers/net/idpf/idpf_logs.h  |   56 +
 drivers/net/idpf/idpf_rxtx.c  | 2308 +
 drivers/net/idpf/idpf_rxtx.h  |  291 ++
 drivers/net/idpf/idpf_rxtx_vec_avx512.c   |  871 +
 drivers/net/idpf/idpf_rxtx_vec_common.h   |  100 +
 drivers/net/idpf/idpf_vchnl.c | 1416 
 drivers/net/idpf/meson.build  |   44 +
 drivers/net/idpf/version.map  |3 +
 drivers/net/meson.build   |1 +
 38 files changed, 15228 insertions(+)
 create mode 100644 doc/guides/nics/features/idpf.ini
 create mode 100644 doc/guides/nics/idpf.rst
 create mode 100644 drivers/common/idpf/base/idpf_alloc.h
 create mode 100644 drivers/common/idpf/base/idpf_common.c
 create mode 100644 drivers/common/idpf/base/idpf_controlq.c
 create mode 100644 drivers/common/idpf/base/idpf_controlq.h
 create mode 100644 drivers/common/idpf/base/idpf_controlq_api.h
 create mode 100644 drivers/common/idpf/base/idpf_controlq_setup.c
 create mod

[PATCH v17 02/18] net/idpf: add support for device initialization

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Support device init and add the following dev ops:
 - dev_configure
 - dev_close
 - dev_infos_get

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Xiao Wang 
Signed-off-by: Wenjun Wu 
Signed-off-by: Junfeng Guo 
---
 MAINTAINERS|   9 +
 doc/guides/nics/features/idpf.ini  |   9 +
 doc/guides/nics/idpf.rst   |  66 ++
 doc/guides/nics/index.rst  |   1 +
 doc/guides/rel_notes/release_22_11.rst |   6 +
 drivers/net/idpf/idpf_ethdev.c | 891 +
 drivers/net/idpf/idpf_ethdev.h | 189 ++
 drivers/net/idpf/idpf_logs.h   |  56 ++
 drivers/net/idpf/idpf_vchnl.c  | 416 
 drivers/net/idpf/meson.build   |  15 +
 drivers/net/idpf/version.map   |   3 +
 drivers/net/meson.build|   1 +
 12 files changed, 1662 insertions(+)
 create mode 100644 doc/guides/nics/features/idpf.ini
 create mode 100644 doc/guides/nics/idpf.rst
 create mode 100644 drivers/net/idpf/idpf_ethdev.c
 create mode 100644 drivers/net/idpf/idpf_ethdev.h
 create mode 100644 drivers/net/idpf/idpf_logs.h
 create mode 100644 drivers/net/idpf/idpf_vchnl.c
 create mode 100644 drivers/net/idpf/meson.build
 create mode 100644 drivers/net/idpf/version.map

diff --git a/MAINTAINERS b/MAINTAINERS
index bdf233c9f8..cc66db25e8 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -770,6 +770,15 @@ F: drivers/net/ice/
 F: doc/guides/nics/ice.rst
 F: doc/guides/nics/features/ice.ini
 
+Intel idpf
+M: Jingjing Wu 
+M: Beilei Xing 
+T: git://dpdk.org/next/dpdk-next-net-intel
+F: drivers/net/idpf/
+F: drivers/common/idpf/
+F: doc/guides/nics/idpf.rst
+F: doc/guides/nics/features/idpf.ini
+
 Intel igc
 M: Junfeng Guo 
 M: Simei Su 
diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
new file mode 100644
index 00..46aab2eb61
--- /dev/null
+++ b/doc/guides/nics/features/idpf.ini
@@ -0,0 +1,9 @@
+;
+; Supported features of the 'idpf' network poll mode driver.
+;
+; Refer to default.ini for the full list of available PMD features.
+;
+[Features]
+Linux= Y
+x86-32   = Y
+x86-64   = Y
diff --git a/doc/guides/nics/idpf.rst b/doc/guides/nics/idpf.rst
new file mode 100644
index 00..c1001d5d0c
--- /dev/null
+++ b/doc/guides/nics/idpf.rst
@@ -0,0 +1,66 @@
+..  SPDX-License-Identifier: BSD-3-Clause
+Copyright(c) 2022 Intel Corporation.
+
+IDPF Poll Mode Driver
+==
+
+The [*EXPERIMENTAL*] idpf PMD (**librte_net_idpf**) provides poll mode driver 
support for
+Intel® Infrastructure Processing Unit (Intel® IPU) E2000.
+
+
+Linux Prerequisites
+---
+
+- Follow the DPDK :ref:`Getting Started Guide for Linux ` to setup 
the basic DPDK environment.
+
+- To get better performance on Intel platforms, please follow the "How to get 
best performance with NICs on Intel platforms"
+  section of the :ref:`Getting Started Guide for Linux `.
+
+
+Pre-Installation Configuration
+--
+
+Runtime Config Options
+~~
+
+- ``vport`` (default ``0``)
+
+  The IDPF PMD supports creation of multiple vports for one PCI device, each 
vport
+  corresponds to a single ethdev. Using the ``devargs`` parameter ``vport`` 
the user
+  can specify the vports with specific ID to be created, for example::
+
+-a ca:00.0,vport=[0,2,3]
+
+  Then idpf PMD will create 3 vports (ethdevs) for device ca:00.0.
+  NOTE: If the parameter is not provided, the vport 0 will be created by 
default.
+
+- ``rx_single`` (default ``0``)
+
+  There're two queue modes supported by Intel® IPU Ethernet ES2000 Series, 
single queue
+  mode and split queue mode for Rx queue. User can choose Rx queue mode by the 
``devargs``
+  parameter ``rx_single``.
+
+-a ca:00.0,rx_single=1
+
+  Then idpf PMD will configure Rx queue with single queue mode. Otherwise, 
split queue
+  mode is chosen by default.
+
+- ``tx_single`` (default ``0``)
+
+  There're two queue modes supported by Intel® IPU Ethernet ES2000 Series, 
single queue
+  mode and split queue mode for Tx queue. User can choose Tx queue mode by the 
``devargs``
+  parameter ``tx_single``.
+
+-a ca:00.0,tx_single=1
+
+  Then idpf PMD will configure Tx queue with single queue mode. Otherwise, 
split queue
+  mode is chosen by default.
+
+
+Driver compilation and testing
+--
+
+Refer to the document :ref:`compiling and testing a PMD for a NIC 
`
+for details.
+
+
diff --git a/doc/guides/nics/index.rst b/doc/guides/nics/index.rst
index 4d40ea29a3..12841ce407 100644
--- a/doc/guides/nics/index.rst
+++ b/doc/guides/nics/index.rst
@@ -34,6 +34,7 @@ Network Interface Controller Drivers
 hns3
 i40e
 ice
+idpf
 igb
 igc
 ionic
diff --git a/doc/guides/rel_notes/release_22_11.rst 
b/doc/guides/rel_notes/release_22_11.rst
index 28812e092f..77674f2b06 100644
--- a/doc/guides/rel_notes/rele

[PATCH v17 04/18] net/idpf: add Rx queue setup

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for rx_queue_setup ops.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  11 +
 drivers/net/idpf/idpf_rxtx.c   | 400 +
 drivers/net/idpf/idpf_rxtx.h   |  46 
 3 files changed, 457 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 54f20d30ca..fb5cd1b111 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -48,12 +48,22 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
.tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH,
};
 
+   dev_info->default_rxconf = (struct rte_eth_rxconf) {
+   .rx_free_thresh = IDPF_DEFAULT_RX_FREE_THRESH,
+   };
+
dev_info->tx_desc_lim = (struct rte_eth_desc_lim) {
.nb_max = IDPF_MAX_RING_DESC,
.nb_min = IDPF_MIN_RING_DESC,
.nb_align = IDPF_ALIGN_RING_DESC,
};
 
+   dev_info->rx_desc_lim = (struct rte_eth_desc_lim) {
+   .nb_max = IDPF_MAX_RING_DESC,
+   .nb_min = IDPF_MIN_RING_DESC,
+   .nb_align = IDPF_ALIGN_RING_DESC,
+   };
+
return 0;
 }
 
@@ -643,6 +653,7 @@ idpf_adapter_init(struct rte_pci_device *pci_dev, struct 
idpf_adapter *adapter)
 static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_configure  = idpf_dev_configure,
.dev_close  = idpf_dev_close,
+   .rx_queue_setup = idpf_rx_queue_setup,
.tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
 };
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 4afa0a2560..25dd5d85d5 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -8,6 +8,21 @@
 #include "idpf_ethdev.h"
 #include "idpf_rxtx.h"
 
+static int
+check_rx_thresh(uint16_t nb_desc, uint16_t thresh)
+{
+   /* The following constraints must be satisfied:
+*   thresh < rxq->nb_rx_desc
+*/
+   if (thresh >= nb_desc) {
+   PMD_INIT_LOG(ERR, "rx_free_thresh (%u) must be less than %u",
+thresh, nb_desc);
+   return -EINVAL;
+   }
+
+   return 0;
+}
+
 static int
 check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
uint16_t tx_free_thresh)
@@ -56,6 +71,87 @@ check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
return 0;
 }
 
+static void
+reset_split_rx_descq(struct idpf_rx_queue *rxq)
+{
+   uint16_t len;
+   uint32_t i;
+
+   if (rxq == NULL)
+   return;
+
+   len = rxq->nb_rx_desc + IDPF_RX_MAX_BURST;
+
+   for (i = 0; i < len * sizeof(struct virtchnl2_rx_flex_desc_adv_nic_3);
+i++)
+   ((volatile char *)rxq->rx_ring)[i] = 0;
+
+   rxq->rx_tail = 0;
+   rxq->expected_gen_id = 1;
+}
+
+static void
+reset_split_rx_bufq(struct idpf_rx_queue *rxq)
+{
+   uint16_t len;
+   uint32_t i;
+
+   if (rxq == NULL)
+   return;
+
+   len = rxq->nb_rx_desc + IDPF_RX_MAX_BURST;
+
+   for (i = 0; i < len * sizeof(struct virtchnl2_splitq_rx_buf_desc);
+i++)
+   ((volatile char *)rxq->rx_ring)[i] = 0;
+
+   memset(&rxq->fake_mbuf, 0x0, sizeof(rxq->fake_mbuf));
+
+   for (i = 0; i < IDPF_RX_MAX_BURST; i++)
+   rxq->sw_ring[rxq->nb_rx_desc + i] = &rxq->fake_mbuf;
+
+   /* The next descriptor id which can be received. */
+   rxq->rx_next_avail = 0;
+
+   /* The next descriptor id which can be refilled. */
+   rxq->rx_tail = 0;
+   /* The number of descriptors which can be refilled. */
+   rxq->nb_rx_hold = rxq->nb_rx_desc - 1;
+
+   rxq->bufq1 = NULL;
+   rxq->bufq2 = NULL;
+}
+
+static void
+reset_single_rx_queue(struct idpf_rx_queue *rxq)
+{
+   uint16_t len;
+   uint32_t i;
+
+   if (rxq == NULL)
+   return;
+
+   len = rxq->nb_rx_desc + IDPF_RX_MAX_BURST;
+
+   for (i = 0; i < len * sizeof(struct virtchnl2_singleq_rx_buf_desc);
+i++)
+   ((volatile char *)rxq->rx_ring)[i] = 0;
+
+   memset(&rxq->fake_mbuf, 0x0, sizeof(rxq->fake_mbuf));
+
+   for (i = 0; i < IDPF_RX_MAX_BURST; i++)
+   rxq->sw_ring[rxq->nb_rx_desc + i] = &rxq->fake_mbuf;
+
+   rxq->rx_tail = 0;
+   rxq->nb_rx_hold = 0;
+
+   if (rxq->pkt_first_seg != NULL)
+   rte_pktmbuf_free(rxq->pkt_first_seg);
+
+   rxq->pkt_first_seg = NULL;
+   rxq->pkt_last_seg = NULL;
+}
+
 static void
 reset_split_tx_descq(struct idpf_tx_queue *txq)
 {
@@ -145,6 +241,310 @@ reset_single_tx_queue(struct idpf_tx_queue *txq)
txq->next_rs = txq->rs_thresh - 1;
 }
 
+static int
+idpf_rx_split_bufq_setup(struct rte_eth_dev *dev, struct idpf_rx_queue *bufq,
+   

[PATCH v17 05/18] net/idpf: add support for device start and stop

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add dev ops dev_start, dev_stop and link_update.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c | 55 ++
 drivers/net/idpf/idpf_rxtx.c   | 20 +
 2 files changed, 75 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index fb5cd1b111..621bf9aad5 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -29,6 +29,22 @@ static const char * const idpf_valid_args[] = {
NULL
 };
 
+static int
+idpf_dev_link_update(struct rte_eth_dev *dev,
+__rte_unused int wait_to_complete)
+{
+   struct rte_eth_link new_link;
+
+   memset(&new_link, 0, sizeof(new_link));
+
+   new_link.link_speed = RTE_ETH_SPEED_NUM_NONE;
+   new_link.link_duplex = RTE_ETH_LINK_FULL_DUPLEX;
+   new_link.link_autoneg = !(dev->data->dev_conf.link_speeds &
+ RTE_ETH_LINK_SPEED_FIXED);
+
+   return rte_eth_linkstatus_set(dev, &new_link);
+}
+
 static int
 idpf_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info)
 {
@@ -267,6 +283,42 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_dev_start(struct rte_eth_dev *dev)
+{
+   struct idpf_vport *vport = dev->data->dev_private;
+   int ret;
+
+   if (dev->data->mtu > vport->max_mtu) {
+   PMD_DRV_LOG(ERR, "MTU should be less than %d", vport->max_mtu);
+   return -EINVAL;
+   }
+
+   vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
+
+   /* TODO: start queues */
+
+   ret = idpf_vc_ena_dis_vport(vport, true);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR, "Failed to enable vport");
+   return ret;
+   }
+
+   return 0;
+}
+
+static int
+idpf_dev_stop(struct rte_eth_dev *dev)
+{
+   struct idpf_vport *vport = dev->data->dev_private;
+
+   idpf_vc_ena_dis_vport(vport, false);
+
+   /* TODO: stop queues */
+
+   return 0;
+}
+
 static int
 idpf_dev_close(struct rte_eth_dev *dev)
 {
@@ -656,6 +708,9 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.rx_queue_setup = idpf_rx_queue_setup,
.tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
+   .dev_start  = idpf_dev_start,
+   .dev_stop   = idpf_dev_stop,
+   .link_update= idpf_dev_link_update,
 };
 
 static uint16_t
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 25dd5d85d5..3528d2f2c7 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -334,6 +334,11 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   if (rx_conf->rx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -465,6 +470,11 @@ idpf_rx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   if (rx_conf->rx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -569,6 +579,11 @@ idpf_tx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   if (tx_conf->tx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf split txq",
 sizeof(struct idpf_tx_queue),
@@ -691,6 +706,11 @@ idpf_tx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   if (tx_conf->tx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf txq",
 sizeof(struct idpf_tx_queue),
-- 
2.26.2



[PATCH v17 03/18] net/idpf: add Tx queue setup

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for tx_queue_setup ops.

In the single queue model, the same descriptor queue is used by SW to
post buffer descriptors to HW and by HW to post completed descriptors
to SW.

In the split queue model, "RX buffer queues" are used to pass
descriptor buffers from SW to HW while Rx queues are used only to
pass the descriptor completions, that is, descriptors that point
to completed buffers, from HW to SW. This is contrary to the single
queue model in which Rx queues are used for both purposes.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  13 ++
 drivers/net/idpf/idpf_rxtx.c   | 364 +
 drivers/net/idpf/idpf_rxtx.h   |  70 +++
 drivers/net/idpf/meson.build   |   1 +
 4 files changed, 448 insertions(+)
 create mode 100644 drivers/net/idpf/idpf_rxtx.c
 create mode 100644 drivers/net/idpf/idpf_rxtx.h

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 035f563275..54f20d30ca 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -11,6 +11,7 @@
 #include 
 
 #include "idpf_ethdev.h"
+#include "idpf_rxtx.h"
 
 #define IDPF_TX_SINGLE_Q   "tx_single"
 #define IDPF_RX_SINGLE_Q   "rx_single"
@@ -42,6 +43,17 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
dev_info->max_mtu = dev_info->max_rx_pktlen - IDPF_ETH_OVERHEAD;
dev_info->min_mtu = RTE_ETHER_MIN_MTU;
 
+   dev_info->default_txconf = (struct rte_eth_txconf) {
+   .tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
+   .tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH,
+   };
+
+   dev_info->tx_desc_lim = (struct rte_eth_desc_lim) {
+   .nb_max = IDPF_MAX_RING_DESC,
+   .nb_min = IDPF_MIN_RING_DESC,
+   .nb_align = IDPF_ALIGN_RING_DESC,
+   };
+
return 0;
 }
 
@@ -631,6 +643,7 @@ idpf_adapter_init(struct rte_pci_device *pci_dev, struct 
idpf_adapter *adapter)
 static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_configure  = idpf_dev_configure,
.dev_close  = idpf_dev_close,
+   .tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
 };
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
new file mode 100644
index 00..4afa0a2560
--- /dev/null
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -0,0 +1,364 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2022 Intel Corporation
+ */
+
+#include 
+#include 
+
+#include "idpf_ethdev.h"
+#include "idpf_rxtx.h"
+
+static int
+check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
+   uint16_t tx_free_thresh)
+{
+   /* TX descriptors will have their RS bit set after tx_rs_thresh
+* descriptors have been used. The TX descriptor ring will be cleaned
+* after tx_free_thresh descriptors are used or if the number of
+* descriptors required to transmit a packet is greater than the
+* number of free TX descriptors.
+*
+* The following constraints must be satisfied:
+*  - tx_rs_thresh must be less than the size of the ring minus 2.
+*  - tx_free_thresh must be less than the size of the ring minus 3.
+*  - tx_rs_thresh must be less than or equal to tx_free_thresh.
+*  - tx_rs_thresh must be a divisor of the ring size.
+*
+* One descriptor in the TX ring is used as a sentinel to avoid a H/W
+* race condition, hence the maximum threshold constraints. When set
+* to zero use default values.
+*/
+   if (tx_rs_thresh >= (nb_desc - 2)) {
+   PMD_INIT_LOG(ERR, "tx_rs_thresh (%u) must be less than the "
+"number of TX descriptors (%u) minus 2",
+tx_rs_thresh, nb_desc);
+   return -EINVAL;
+   }
+   if (tx_free_thresh >= (nb_desc - 3)) {
+   PMD_INIT_LOG(ERR, "tx_free_thresh (%u) must be less than the "
+"number of TX descriptors (%u) minus 3.",
+tx_free_thresh, nb_desc);
+   return -EINVAL;
+   }
+   if (tx_rs_thresh > tx_free_thresh) {
+   PMD_INIT_LOG(ERR, "tx_rs_thresh (%u) must be less than or "
+"equal to tx_free_thresh (%u).",
+tx_rs_thresh, tx_free_thresh);
+   return -EINVAL;
+   }
+   if ((nb_desc % tx_rs_thresh) != 0) {
+   PMD_INIT_LOG(ERR, "tx_rs_thresh (%u) must be a divisor of the "
+"number of TX descriptors (%u).",
+tx_rs_thresh, nb_desc);
+   return -EINVAL;
+   }
+
+   return 0;
+}
+
+static void
+reset_split_tx_descq(struct idpf_tx_queue *txq)
+{
+   struct id

[PATCH v17 08/18] net/idpf: add queue release

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for queue operations:
 - rx_queue_release
 - tx_queue_release

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  2 +
 drivers/net/idpf/idpf_rxtx.c   | 81 ++
 drivers/net/idpf/idpf_rxtx.h   |  3 ++
 3 files changed, 86 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 9f1e1e6a18..1485f40e71 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -758,6 +758,8 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.tx_queue_start = idpf_tx_queue_start,
.rx_queue_stop  = idpf_rx_queue_stop,
.tx_queue_stop  = idpf_tx_queue_stop,
+   .rx_queue_release   = idpf_dev_rx_queue_release,
+   .tx_queue_release   = idpf_dev_tx_queue_release,
 };
 
 static uint16_t
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 8d5ec41a1f..053409b99a 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -171,6 +171,51 @@ reset_split_rx_bufq(struct idpf_rx_queue *rxq)
rxq->bufq2 = NULL;
 }
 
+static void
+idpf_rx_queue_release(void *rxq)
+{
+   struct idpf_rx_queue *q = rxq;
+
+   if (q == NULL)
+   return;
+
+   /* Split queue */
+   if (q->bufq1 != NULL && q->bufq2 != NULL) {
+   q->bufq1->ops->release_mbufs(q->bufq1);
+   rte_free(q->bufq1->sw_ring);
+   rte_memzone_free(q->bufq1->mz);
+   rte_free(q->bufq1);
+   q->bufq2->ops->release_mbufs(q->bufq2);
+   rte_free(q->bufq2->sw_ring);
+   rte_memzone_free(q->bufq2->mz);
+   rte_free(q->bufq2);
+   rte_memzone_free(q->mz);
+   rte_free(q);
+   return;
+   }
+
+   /* Single queue */
+   q->ops->release_mbufs(q);
+   rte_free(q->sw_ring);
+   rte_memzone_free(q->mz);
+   rte_free(q);
+}
+
+static void
+idpf_tx_queue_release(void *txq)
+{
+   struct idpf_tx_queue *q = txq;
+
+   if (q == NULL)
+   return;
+
+   rte_free(q->complq);
+   q->ops->release_mbufs(q);
+   rte_free(q->sw_ring);
+   rte_memzone_free(q->mz);
+   rte_free(q);
+}
+
 static inline void
 reset_split_rx_queue(struct idpf_rx_queue *rxq)
 {
@@ -392,6 +437,12 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed */
+   if (dev->data->rx_queues[queue_idx] != NULL) {
+   idpf_rx_queue_release(dev->data->rx_queues[queue_idx]);
+   dev->data->rx_queues[queue_idx] = NULL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -524,6 +575,12 @@ idpf_rx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed */
+   if (dev->data->rx_queues[queue_idx] != NULL) {
+   idpf_rx_queue_release(dev->data->rx_queues[queue_idx]);
+   dev->data->rx_queues[queue_idx] = NULL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -630,6 +687,12 @@ idpf_tx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed. */
+   if (dev->data->tx_queues[queue_idx] != NULL) {
+   idpf_tx_queue_release(dev->data->tx_queues[queue_idx]);
+   dev->data->tx_queues[queue_idx] = NULL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf split txq",
 sizeof(struct idpf_tx_queue),
@@ -754,6 +817,12 @@ idpf_tx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed. */
+   if (dev->data->tx_queues[queue_idx] != NULL) {
+   idpf_tx_queue_release(dev->data->tx_queues[queue_idx]);
+   dev->data->tx_queues[queue_idx] = NULL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf txq",
 sizeof(struct idpf_tx_queue),
@@ -1102,6 +1171,18 @@ idpf_tx_queue_stop(struct rte_eth_dev *dev, uint16_t 
tx_queue_id)
return 0;
 }
 
+void
+idpf_dev_rx_queue_release(struct rte_eth_dev *dev, uint16_t qid)
+{
+   idpf_rx_queue_r

[PATCH v17 07/18] net/idpf: add support for queue stop

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for these device ops:
 - rx_queue_stop
 - tx_queue_stop

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  17 ++--
 drivers/net/idpf/idpf_rxtx.c   | 148 +
 drivers/net/idpf/idpf_rxtx.h   |  13 +++
 drivers/net/idpf/idpf_vchnl.c  |  69 +++
 4 files changed, 242 insertions(+), 5 deletions(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 0400ed611f..9f1e1e6a18 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -324,7 +324,8 @@ idpf_dev_start(struct rte_eth_dev *dev)
 
if (dev->data->mtu > vport->max_mtu) {
PMD_DRV_LOG(ERR, "MTU should be less than %d", vport->max_mtu);
-   return -EINVAL;
+   ret = -EINVAL;
+   goto err_mtu;
}
 
vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
@@ -332,17 +333,21 @@ idpf_dev_start(struct rte_eth_dev *dev)
ret = idpf_start_queues(dev);
if (ret != 0) {
PMD_DRV_LOG(ERR, "Failed to start queues");
-   return ret;
+   goto err_mtu;
}
 
ret = idpf_vc_ena_dis_vport(vport, true);
if (ret != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
-   /* TODO: stop queues */
-   return ret;
+   goto err_vport;
}
 
return 0;
+
+err_vport:
+   idpf_stop_queues(dev);
+err_mtu:
+   return ret;
 }
 
 static int
@@ -352,7 +357,7 @@ idpf_dev_stop(struct rte_eth_dev *dev)
 
idpf_vc_ena_dis_vport(vport, false);
 
-   /* TODO: stop queues */
+   idpf_stop_queues(dev);
 
return 0;
 }
@@ -751,6 +756,8 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.link_update= idpf_dev_link_update,
.rx_queue_start = idpf_rx_queue_start,
.tx_queue_start = idpf_tx_queue_start,
+   .rx_queue_stop  = idpf_rx_queue_stop,
+   .tx_queue_stop  = idpf_tx_queue_stop,
 };
 
 static uint16_t
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 6d954afd9d..8d5ec41a1f 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -71,6 +71,55 @@ check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
return 0;
 }
 
+static void
+release_rxq_mbufs(struct idpf_rx_queue *rxq)
+{
+   uint16_t i;
+
+   if (rxq->sw_ring == NULL)
+   return;
+
+   for (i = 0; i < rxq->nb_rx_desc; i++) {
+   if (rxq->sw_ring[i] != NULL) {
+   rte_pktmbuf_free_seg(rxq->sw_ring[i]);
+   rxq->sw_ring[i] = NULL;
+   }
+   }
+}
+
+static void
+release_txq_mbufs(struct idpf_tx_queue *txq)
+{
+   uint16_t nb_desc, i;
+
+   if (txq == NULL || txq->sw_ring == NULL) {
+   PMD_DRV_LOG(DEBUG, "Pointer to rxq or sw_ring is NULL");
+   return;
+   }
+
+   if (txq->sw_nb_desc != 0) {
+   /* For split queue model, descriptor ring */
+   nb_desc = txq->sw_nb_desc;
+   } else {
+   /* For single queue model */
+   nb_desc = txq->nb_tx_desc;
+   }
+   for (i = 0; i < nb_desc; i++) {
+   if (txq->sw_ring[i].mbuf != NULL) {
+   rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf);
+   txq->sw_ring[i].mbuf = NULL;
+   }
+   }
+}
+
+static const struct idpf_rxq_ops def_rxq_ops = {
+   .release_mbufs = release_rxq_mbufs,
+};
+
+static const struct idpf_txq_ops def_txq_ops = {
+   .release_mbufs = release_txq_mbufs,
+};
+
 static void
 reset_split_rx_descq(struct idpf_rx_queue *rxq)
 {
@@ -122,6 +171,14 @@ reset_split_rx_bufq(struct idpf_rx_queue *rxq)
rxq->bufq2 = NULL;
 }
 
+static inline void
+reset_split_rx_queue(struct idpf_rx_queue *rxq)
+{
+   reset_split_rx_descq(rxq);
+   reset_split_rx_bufq(rxq->bufq1);
+   reset_split_rx_bufq(rxq->bufq2);
+}
+
 static void
 reset_single_rx_queue(struct idpf_rx_queue *rxq)
 {
@@ -301,6 +358,7 @@ idpf_rx_split_bufq_setup(struct rte_eth_dev *dev, struct 
idpf_rx_queue *bufq,
bufq->q_set = true;
bufq->qrx_tail = hw->hw_addr + (vport->chunks_info.rx_buf_qtail_start +
 queue_idx * vport->chunks_info.rx_buf_qtail_spacing);
+   bufq->ops = &def_rxq_ops;
 
/* TODO: allow bulk or vec */
 
@@ -527,6 +585,7 @@ idpf_rx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
dev->data->rx_queues[queue_idx] = rxq;
rxq->qrx_tail = hw->hw_addr + (vport->chunks_info.rx_qtail_start +
queue_idx * vport->chunks_info.rx_qtail_spacing);
+   rxq->ops = &def_rxq_ops;
 
return 0;
 }
@@ -621,6 +680,7 @@ idpf_tx_split_queue_setup(struct

[PATCH v17 06/18] net/idpf: add support for queue start

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for these device ops:
 - rx_queue_start
 - tx_queue_start

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  42 +++-
 drivers/net/idpf/idpf_ethdev.h |   9 +
 drivers/net/idpf/idpf_rxtx.c   | 237 +++--
 drivers/net/idpf/idpf_rxtx.h   |   6 +
 drivers/net/idpf/idpf_vchnl.c  | 447 +
 5 files changed, 720 insertions(+), 21 deletions(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 621bf9aad5..0400ed611f 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -283,6 +283,39 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_start_queues(struct rte_eth_dev *dev)
+{
+   struct idpf_rx_queue *rxq;
+   struct idpf_tx_queue *txq;
+   int err = 0;
+   int i;
+
+   for (i = 0; i < dev->data->nb_tx_queues; i++) {
+   txq = dev->data->tx_queues[i];
+   if (txq == NULL || txq->tx_deferred_start)
+   continue;
+   err = idpf_tx_queue_start(dev, i);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR, "Fail to start Tx queue %u", i);
+   return err;
+   }
+   }
+
+   for (i = 0; i < dev->data->nb_rx_queues; i++) {
+   rxq = dev->data->rx_queues[i];
+   if (rxq == NULL || rxq->rx_deferred_start)
+   continue;
+   err = idpf_rx_queue_start(dev, i);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR, "Fail to start Rx queue %u", i);
+   return err;
+   }
+   }
+
+   return err;
+}
+
 static int
 idpf_dev_start(struct rte_eth_dev *dev)
 {
@@ -296,11 +329,16 @@ idpf_dev_start(struct rte_eth_dev *dev)
 
vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
 
-   /* TODO: start queues */
+   ret = idpf_start_queues(dev);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR, "Failed to start queues");
+   return ret;
+   }
 
ret = idpf_vc_ena_dis_vport(vport, true);
if (ret != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
+   /* TODO: stop queues */
return ret;
}
 
@@ -711,6 +749,8 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_start  = idpf_dev_start,
.dev_stop   = idpf_dev_stop,
.link_update= idpf_dev_link_update,
+   .rx_queue_start = idpf_rx_queue_start,
+   .tx_queue_start = idpf_tx_queue_start,
 };
 
 static uint16_t
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 84ae6641e2..96c22009e9 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -24,7 +24,9 @@
 #define IDPF_DEFAULT_TXQ_NUM   16
 
 #define IDPF_INVALID_VPORT_IDX 0x
+#define IDPF_TXQ_PER_GRP   1
 #define IDPF_TX_COMPLQ_PER_GRP 1
+#define IDPF_RXQ_PER_GRP   1
 #define IDPF_RX_BUFQ_PER_GRP   2
 
 #define IDPF_CTLQ_ID   -1
@@ -182,6 +184,13 @@ int idpf_vc_check_api_version(struct idpf_adapter 
*adapter);
 int idpf_vc_get_caps(struct idpf_adapter *adapter);
 int idpf_vc_create_vport(struct idpf_adapter *adapter);
 int idpf_vc_destroy_vport(struct idpf_vport *vport);
+int idpf_vc_config_rxqs(struct idpf_vport *vport);
+int idpf_vc_config_rxq(struct idpf_vport *vport, uint16_t rxq_id);
+int idpf_vc_config_txqs(struct idpf_vport *vport);
+int idpf_vc_config_txq(struct idpf_vport *vport, uint16_t txq_id);
+int idpf_switch_queue(struct idpf_vport *vport, uint16_t qid,
+ bool rx, bool on);
+int idpf_vc_ena_dis_queues(struct idpf_vport *vport, bool enable);
 int idpf_vc_ena_dis_vport(struct idpf_vport *vport, bool enable);
 int idpf_read_one_msg(struct idpf_adapter *adapter, uint32_t ops,
  uint16_t buf_len, uint8_t *buf);
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 3528d2f2c7..6d954afd9d 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -334,11 +334,6 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
-   if (rx_conf->rx_deferred_start) {
-   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
-   return -EINVAL;
-   }
-
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -354,6 +349,7 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, uint16_t 
queue_idx,
rxq->rx_free_thresh = rx_free_thresh;
rxq->queue_id = vport->chunks_info.rx_start_qid + queue_idx;
rxq->port_id = dev-

[PATCH v17 09/18] net/idpf: add support for MTU configuration

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add dev ops mtu_set.

Signed-off-by: Beilei Xing 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |  1 +
 drivers/net/idpf/idpf_ethdev.c| 13 +
 2 files changed, 14 insertions(+)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index 46aab2eb61..d722c49fde 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -4,6 +4,7 @@
 ; Refer to default.ini for the full list of available PMD features.
 ;
 [Features]
+MTU update   = Y
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 1485f40e71..856f3d7266 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -83,6 +83,18 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
return 0;
 }
 
+static int
+idpf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu __rte_unused)
+{
+   /* mtu setting is forbidden if port is start */
+   if (dev->data->dev_started) {
+   PMD_DRV_LOG(ERR, "port must be stopped before configuration");
+   return -EBUSY;
+   }
+
+   return 0;
+}
+
 static int
 idpf_init_vport_req_info(struct rte_eth_dev *dev)
 {
@@ -760,6 +772,7 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.tx_queue_stop  = idpf_tx_queue_stop,
.rx_queue_release   = idpf_dev_rx_queue_release,
.tx_queue_release   = idpf_dev_tx_queue_release,
+   .mtu_set= idpf_dev_mtu_set,
 };
 
 static uint16_t
-- 
2.26.2



[PATCH v17 10/18] net/idpf: add support for basic Rx datapath

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add basic Rx support in split queue mode and single queue mode.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |   2 +
 drivers/net/idpf/idpf_rxtx.c   | 273 +
 drivers/net/idpf/idpf_rxtx.h   |   7 +-
 3 files changed, 281 insertions(+), 1 deletion(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 856f3d7266..2f1f95 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -348,6 +348,8 @@ idpf_dev_start(struct rte_eth_dev *dev)
goto err_mtu;
}
 
+   idpf_set_rx_function(dev);
+
ret = idpf_vc_ena_dis_vport(vport, true);
if (ret != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 053409b99a..ea499c4d37 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1208,3 +1208,276 @@ idpf_stop_queues(struct rte_eth_dev *dev)
PMD_DRV_LOG(WARNING, "Fail to stop Tx queue %d", i);
}
 }
+
+static void
+idpf_split_rx_bufq_refill(struct idpf_rx_queue *rx_bufq)
+{
+   volatile struct virtchnl2_splitq_rx_buf_desc *rx_buf_ring;
+   volatile struct virtchnl2_splitq_rx_buf_desc *rx_buf_desc;
+   uint16_t nb_refill = rx_bufq->rx_free_thresh;
+   uint16_t nb_desc = rx_bufq->nb_rx_desc;
+   uint16_t next_avail = rx_bufq->rx_tail;
+   struct rte_mbuf *nmb[rx_bufq->rx_free_thresh];
+   struct rte_eth_dev *dev;
+   uint64_t dma_addr;
+   uint16_t delta;
+   int i;
+
+   if (rx_bufq->nb_rx_hold < rx_bufq->rx_free_thresh)
+   return;
+
+   rx_buf_ring = rx_bufq->rx_ring;
+   delta = nb_desc - next_avail;
+   if (unlikely(delta < nb_refill)) {
+   if (likely(rte_pktmbuf_alloc_bulk(rx_bufq->mp, nmb, delta) == 
0)) {
+   for (i = 0; i < delta; i++) {
+   rx_buf_desc = &rx_buf_ring[next_avail + i];
+   rx_bufq->sw_ring[next_avail + i] = nmb[i];
+   dma_addr = 
rte_cpu_to_le_64(rte_mbuf_data_iova_default(nmb[i]));
+   rx_buf_desc->hdr_addr = 0;
+   rx_buf_desc->pkt_addr = dma_addr;
+   }
+   nb_refill -= delta;
+   next_avail = 0;
+   rx_bufq->nb_rx_hold -= delta;
+   } else {
+   dev = &rte_eth_devices[rx_bufq->port_id];
+   dev->data->rx_mbuf_alloc_failed += nb_desc - next_avail;
+   PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u 
queue_id=%u",
+  rx_bufq->port_id, rx_bufq->queue_id);
+   return;
+   }
+   }
+
+   if (nb_desc - next_avail >= nb_refill) {
+   if (likely(rte_pktmbuf_alloc_bulk(rx_bufq->mp, nmb, nb_refill) 
== 0)) {
+   for (i = 0; i < nb_refill; i++) {
+   rx_buf_desc = &rx_buf_ring[next_avail + i];
+   rx_bufq->sw_ring[next_avail + i] = nmb[i];
+   dma_addr = 
rte_cpu_to_le_64(rte_mbuf_data_iova_default(nmb[i]));
+   rx_buf_desc->hdr_addr = 0;
+   rx_buf_desc->pkt_addr = dma_addr;
+   }
+   next_avail += nb_refill;
+   rx_bufq->nb_rx_hold -= nb_refill;
+   } else {
+   dev = &rte_eth_devices[rx_bufq->port_id];
+   dev->data->rx_mbuf_alloc_failed += nb_desc - next_avail;
+   PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u 
queue_id=%u",
+  rx_bufq->port_id, rx_bufq->queue_id);
+   }
+   }
+
+   IDPF_PCI_REG_WRITE(rx_bufq->qrx_tail, next_avail);
+
+   rx_bufq->rx_tail = next_avail;
+}
+
+uint16_t
+idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
+ uint16_t nb_pkts)
+{
+   volatile struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc_ring;
+   volatile struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc;
+   uint16_t pktlen_gen_bufq_id;
+   struct idpf_rx_queue *rxq;
+   struct rte_mbuf *rxm;
+   uint16_t rx_id_bufq1;
+   uint16_t rx_id_bufq2;
+   uint16_t pkt_len;
+   uint16_t bufq_id;
+   uint16_t gen_id;
+   uint16_t rx_id;
+   uint16_t nb_rx;
+
+   nb_rx = 0;
+   rxq = rx_queue;
+
+   if (unlikely(rxq == NULL) || unlikely(!rxq->q_started))
+   return nb_rx;
+
+   rx_id = rxq->rx_tail;
+   rx_id_bufq1 = rxq->bufq1->rx_next_avail;
+   rx_id_bufq2 = rxq->bufq2->rx_next_avail;
+   rx_desc_ring = rxq->rx_ring;
+
+  

[PATCH v17 12/18] net/idpf: support parsing packet type

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Parse packet type during receiving packets.

Signed-off-by: Wenjun Wu 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |   6 +
 drivers/net/idpf/idpf_ethdev.h |   6 +
 drivers/net/idpf/idpf_rxtx.c   |  11 ++
 drivers/net/idpf/idpf_rxtx.h   |   5 +
 drivers/net/idpf/idpf_vchnl.c  | 240 +
 5 files changed, 268 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index f9f6fe1162..d0821ec3f3 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -686,6 +686,12 @@ idpf_adapter_init(struct rte_pci_device *pci_dev, struct 
idpf_adapter *adapter)
goto err_api;
}
 
+   ret = idpf_get_pkt_type(adapter);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to set ptype table");
+   goto err_api;
+   }
+
adapter->caps = rte_zmalloc("idpf_caps",
sizeof(struct virtchnl2_get_capabilities), 0);
if (adapter->caps == NULL) {
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index af0a8e2970..db9af58f72 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -39,6 +39,8 @@
 
 #define IDPF_NUM_MACADDR_MAX   64
 
+#define IDPF_MAX_PKT_TYPE  1024
+
 #define IDPF_VLAN_TAG_SIZE 4
 #define IDPF_ETH_OVERHEAD \
(RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + IDPF_VLAN_TAG_SIZE * 2)
@@ -125,6 +127,8 @@ struct idpf_adapter {
/* Max config queue number per VC message */
uint32_t max_rxq_per_msg;
uint32_t max_txq_per_msg;
+
+   uint32_t ptype_tbl[IDPF_MAX_PKT_TYPE] __rte_cache_min_aligned;
 };
 
 TAILQ_HEAD(idpf_adapter_list, idpf_adapter);
@@ -182,6 +186,7 @@ atomic_set_cmd(struct idpf_adapter *adapter, enum 
virtchnl_ops ops)
 struct idpf_adapter *idpf_find_adapter(struct rte_pci_device *pci_dev);
 void idpf_handle_virtchnl_msg(struct rte_eth_dev *dev);
 int idpf_vc_check_api_version(struct idpf_adapter *adapter);
+int idpf_get_pkt_type(struct idpf_adapter *adapter);
 int idpf_vc_get_caps(struct idpf_adapter *adapter);
 int idpf_vc_create_vport(struct idpf_adapter *adapter);
 int idpf_vc_destroy_vport(struct idpf_vport *vport);
@@ -193,6 +198,7 @@ int idpf_switch_queue(struct idpf_vport *vport, uint16_t 
qid,
  bool rx, bool on);
 int idpf_vc_ena_dis_queues(struct idpf_vport *vport, bool enable);
 int idpf_vc_ena_dis_vport(struct idpf_vport *vport, bool enable);
+int idpf_vc_query_ptype_info(struct idpf_adapter *adapter);
 int idpf_read_one_msg(struct idpf_adapter *adapter, uint32_t ops,
  uint16_t buf_len, uint8_t *buf);
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index f55d2143b9..a980714060 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1281,6 +1281,7 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
volatile struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc;
uint16_t pktlen_gen_bufq_id;
struct idpf_rx_queue *rxq;
+   const uint32_t *ptype_tbl;
struct rte_mbuf *rxm;
uint16_t rx_id_bufq1;
uint16_t rx_id_bufq2;
@@ -1300,6 +1301,7 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rx_id_bufq1 = rxq->bufq1->rx_next_avail;
rx_id_bufq2 = rxq->bufq2->rx_next_avail;
rx_desc_ring = rxq->rx_ring;
+   ptype_tbl = rxq->adapter->ptype_tbl;
 
while (nb_rx < nb_pkts) {
rx_desc = &rx_desc_ring[rx_id];
@@ -1347,6 +1349,10 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rxm->next = NULL;
rxm->nb_segs = 1;
rxm->port = rxq->port_id;
+   rxm->packet_type =
+   ptype_tbl[(rte_le_to_cpu_16(rx_desc->ptype_err_fflags0) 
&
+  VIRTCHNL2_RX_FLEX_DESC_ADV_PTYPE_M) >>
+ VIRTCHNL2_RX_FLEX_DESC_ADV_PTYPE_S];
 
rx_pkts[nb_rx++] = rxm;
}
@@ -1533,6 +1539,7 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
volatile union virtchnl2_rx_desc *rxdp;
union virtchnl2_rx_desc rxd;
struct idpf_rx_queue *rxq;
+   const uint32_t *ptype_tbl;
uint16_t rx_id, nb_hold;
struct rte_eth_dev *dev;
uint16_t rx_packet_len;
@@ -1551,6 +1558,7 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
 
rx_id = rxq->rx_tail;
rx_ring = rxq->rx_ring;
+   ptype_tbl = rxq->adapter->ptype_tbl;
 
while (nb_rx < nb_pkts) {
rxdp = &rx_ring[rx_id];
@@ -1603,6 +1611,9 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rxm->pkt_len = rx_packet_len;
rxm->data_len = rx_packet_len;
rxm->port = rxq->port_id;
+   rxm->packet_type =
+   

[PATCH v17 11/18] net/idpf: add support for basic Tx datapath

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add basic Tx support in split queue mode and single queue mode.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |   3 +
 drivers/net/idpf/idpf_ethdev.h |   1 +
 drivers/net/idpf/idpf_rxtx.c   | 357 +
 drivers/net/idpf/idpf_rxtx.h   |  10 +
 4 files changed, 371 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 2f1f95..f9f6fe1162 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -59,6 +59,8 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
dev_info->max_mtu = dev_info->max_rx_pktlen - IDPF_ETH_OVERHEAD;
dev_info->min_mtu = RTE_ETHER_MIN_MTU;
 
+   dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
+
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
.tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH,
@@ -349,6 +351,7 @@ idpf_dev_start(struct rte_eth_dev *dev)
}
 
idpf_set_rx_function(dev);
+   idpf_set_tx_function(dev);
 
ret = idpf_vc_ena_dis_vport(vport, true);
if (ret != 0) {
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 96c22009e9..af0a8e2970 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -35,6 +35,7 @@
 
 #define IDPF_MIN_BUF_SIZE  1024
 #define IDPF_MAX_FRAME_SIZE9728
+#define IDPF_MIN_FRAME_SIZE14
 
 #define IDPF_NUM_MACADDR_MAX   64
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index ea499c4d37..f55d2143b9 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1365,6 +1365,148 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
return nb_rx;
 }
 
+static inline void
+idpf_split_tx_free(struct idpf_tx_queue *cq)
+{
+   volatile struct idpf_splitq_tx_compl_desc *compl_ring = cq->compl_ring;
+   volatile struct idpf_splitq_tx_compl_desc *txd;
+   uint16_t next = cq->tx_tail;
+   struct idpf_tx_entry *txe;
+   struct idpf_tx_queue *txq;
+   uint16_t gen, qid, q_head;
+   uint8_t ctype;
+
+   txd = &compl_ring[next];
+   gen = (rte_le_to_cpu_16(txd->qid_comptype_gen) &
+   IDPF_TXD_COMPLQ_GEN_M) >> IDPF_TXD_COMPLQ_GEN_S;
+   if (gen != cq->expected_gen_id)
+   return;
+
+   ctype = (rte_le_to_cpu_16(txd->qid_comptype_gen) &
+   IDPF_TXD_COMPLQ_COMPL_TYPE_M) >> IDPF_TXD_COMPLQ_COMPL_TYPE_S;
+   qid = (rte_le_to_cpu_16(txd->qid_comptype_gen) &
+   IDPF_TXD_COMPLQ_QID_M) >> IDPF_TXD_COMPLQ_QID_S;
+   q_head = rte_le_to_cpu_16(txd->q_head_compl_tag.compl_tag);
+   txq = cq->txqs[qid - cq->tx_start_qid];
+
+   switch (ctype) {
+   case IDPF_TXD_COMPLT_RE:
+   if (q_head == 0)
+   txq->last_desc_cleaned = txq->nb_tx_desc - 1;
+   else
+   txq->last_desc_cleaned = q_head - 1;
+   if (unlikely((txq->last_desc_cleaned % 32) == 0)) {
+   PMD_DRV_LOG(ERR, "unexpected desc (head = %u) 
completion.",
+   q_head);
+   return;
+   }
+
+   break;
+   case IDPF_TXD_COMPLT_RS:
+   txq->nb_free++;
+   txq->nb_used--;
+   txe = &txq->sw_ring[q_head];
+   if (txe->mbuf != NULL) {
+   rte_pktmbuf_free_seg(txe->mbuf);
+   txe->mbuf = NULL;
+   }
+   break;
+   default:
+   PMD_DRV_LOG(ERR, "unknown completion type.");
+   return;
+   }
+
+   if (++next == cq->nb_tx_desc) {
+   next = 0;
+   cq->expected_gen_id ^= 1;
+   }
+
+   cq->tx_tail = next;
+}
+
+uint16_t
+idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
+ uint16_t nb_pkts)
+{
+   struct idpf_tx_queue *txq = (struct idpf_tx_queue *)tx_queue;
+   volatile struct idpf_flex_tx_sched_desc *txr;
+   volatile struct idpf_flex_tx_sched_desc *txd;
+   struct idpf_tx_entry *sw_ring;
+   struct idpf_tx_entry *txe, *txn;
+   uint16_t nb_used, tx_id, sw_id;
+   struct rte_mbuf *tx_pkt;
+   uint16_t nb_to_clean;
+   uint16_t nb_tx = 0;
+
+   if (unlikely(txq == NULL) || unlikely(!txq->q_started))
+   return nb_tx;
+
+   txr = txq->desc_ring;
+   sw_ring = txq->sw_ring;
+   tx_id = txq->tx_tail;
+   sw_id = txq->sw_tail;
+   txe = &sw_ring[sw_id];
+
+   for (nb_tx = 0; nb_tx < nb_pkts; nb_tx++) {
+   tx_pkt = tx_pkts[nb_tx];
+
+   if (txq->nb_free <= txq->free_thresh) {
+   /* TODO: Need to refine
+

[PATCH v17 13/18] net/idpf: add support for write back based on ITR expire

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Enable write back on ITR expire, then packets can be received one by
one.

Signed-off-by: Beilei Xing 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c | 120 +
 drivers/net/idpf/idpf_ethdev.h |  13 
 drivers/net/idpf/idpf_vchnl.c  | 113 +++
 3 files changed, 246 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index d0821ec3f3..957cc10616 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -297,6 +297,90 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_config_rx_queues_irqs(struct rte_eth_dev *dev)
+{
+   struct idpf_vport *vport = dev->data->dev_private;
+   struct idpf_adapter *adapter = vport->adapter;
+   struct virtchnl2_queue_vector *qv_map;
+   struct idpf_hw *hw = &adapter->hw;
+   uint32_t dynctl_reg_start;
+   uint32_t itrn_reg_start;
+   uint32_t dynctl_val, itrn_val;
+   uint16_t i;
+
+   qv_map = rte_zmalloc("qv_map",
+   dev->data->nb_rx_queues *
+   sizeof(struct virtchnl2_queue_vector), 0);
+   if (qv_map == NULL) {
+   PMD_DRV_LOG(ERR, "Failed to allocate %d queue-vector map",
+   dev->data->nb_rx_queues);
+   goto qv_map_alloc_err;
+   }
+
+   /* Rx interrupt disabled, Map interrupt only for writeback */
+
+   /* The capability flags adapter->caps->other_caps should be
+* compared with bit VIRTCHNL2_CAP_WB_ON_ITR here. The if
+* condition should be updated when the FW can return the
+* correct flag bits.
+*/
+   dynctl_reg_start =
+   vport->recv_vectors->vchunks.vchunks->dynctl_reg_start;
+   itrn_reg_start =
+   vport->recv_vectors->vchunks.vchunks->itrn_reg_start;
+   dynctl_val = IDPF_READ_REG(hw, dynctl_reg_start);
+   PMD_DRV_LOG(DEBUG, "Value of dynctl_reg_start is 0x%x",
+   dynctl_val);
+   itrn_val = IDPF_READ_REG(hw, itrn_reg_start);
+   PMD_DRV_LOG(DEBUG, "Value of itrn_reg_start is 0x%x", itrn_val);
+   /* Force write-backs by setting WB_ON_ITR bit in DYN_CTL
+* register. WB_ON_ITR and INTENA are mutually exclusive
+* bits. Setting WB_ON_ITR bits means TX and RX Descs
+* are written back based on ITR expiration irrespective
+* of INTENA setting.
+*/
+   /* TBD: need to tune INTERVAL value for better performance. */
+   if (itrn_val != 0)
+   IDPF_WRITE_REG(hw,
+  dynctl_reg_start,
+  VIRTCHNL2_ITR_IDX_0  <<
+  PF_GLINT_DYN_CTL_ITR_INDX_S |
+  PF_GLINT_DYN_CTL_WB_ON_ITR_M |
+  itrn_val <<
+  PF_GLINT_DYN_CTL_INTERVAL_S);
+   else
+   IDPF_WRITE_REG(hw,
+  dynctl_reg_start,
+  VIRTCHNL2_ITR_IDX_0  <<
+  PF_GLINT_DYN_CTL_ITR_INDX_S |
+  PF_GLINT_DYN_CTL_WB_ON_ITR_M |
+  IDPF_DFLT_INTERVAL <<
+  PF_GLINT_DYN_CTL_INTERVAL_S);
+
+   for (i = 0; i < dev->data->nb_rx_queues; i++) {
+   /* map all queues to the same vector */
+   qv_map[i].queue_id = vport->chunks_info.rx_start_qid + i;
+   qv_map[i].vector_id =
+   vport->recv_vectors->vchunks.vchunks->start_vector_id;
+   }
+   vport->qv_map = qv_map;
+
+   if (idpf_vc_config_irq_map_unmap(vport, true) != 0) {
+   PMD_DRV_LOG(ERR, "config interrupt mapping failed");
+   goto config_irq_map_err;
+   }
+
+   return 0;
+
+config_irq_map_err:
+   rte_free(vport->qv_map);
+   vport->qv_map = NULL;
+
+qv_map_alloc_err:
+   return -1;
+}
+
 static int
 idpf_start_queues(struct rte_eth_dev *dev)
 {
@@ -334,6 +418,10 @@ static int
 idpf_dev_start(struct rte_eth_dev *dev)
 {
struct idpf_vport *vport = dev->data->dev_private;
+   struct idpf_adapter *adapter = vport->adapter;
+   uint16_t num_allocated_vectors =
+   adapter->caps->num_allocated_vectors;
+   uint16_t req_vecs_num;
int ret;
 
if (dev->data->mtu > vport->max_mtu) {
@@ -344,6 +432,27 @@ idpf_dev_start(struct rte_eth_dev *dev)
 
vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
 
+   req_vecs_num = IDPF_DFLT_Q_VEC_NUM;
+   if (req_vecs_num + adapter->used_vecs_num > num_allocated_vectors) {
+   PMD_DRV_LOG(ERR, "The accumulated request vectors' number 
should be less than %d",
+   num_allocated_vectors);
+   ret = -EINVAL;
+   goto err_mtu;
+   }
+
+   ret = idpf_vc_alloc_vectors(vpo

[PATCH v17 14/18] net/idpf: add support for RSS

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add RSS support.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c | 120 -
 drivers/net/idpf/idpf_ethdev.h |  26 +++
 drivers/net/idpf/idpf_vchnl.c  | 113 +++
 3 files changed, 258 insertions(+), 1 deletion(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 957cc10616..58560ea404 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -59,6 +59,8 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
dev_info->max_mtu = dev_info->max_rx_pktlen - IDPF_ETH_OVERHEAD;
dev_info->min_mtu = RTE_ETHER_MIN_MTU;
 
+   dev_info->flow_type_rss_offloads = IDPF_RSS_OFFLOAD_ALL;
+
dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
@@ -169,6 +171,8 @@ idpf_parse_devarg_id(char *name)
return val;
 }
 
+#define IDPF_RSS_KEY_LEN 52
+
 static int
 idpf_init_vport(struct rte_eth_dev *dev)
 {
@@ -189,6 +193,10 @@ idpf_init_vport(struct rte_eth_dev *dev)
vport->max_mtu = vport_info->max_mtu;
rte_memcpy(vport->default_mac_addr,
   vport_info->default_mac_addr, ETH_ALEN);
+   vport->rss_algorithm = vport_info->rss_algorithm;
+   vport->rss_key_size = RTE_MIN(IDPF_RSS_KEY_LEN,
+vport_info->rss_key_size);
+   vport->rss_lut_size = vport_info->rss_lut_size;
vport->sw_idx = idx;
 
for (i = 0; i < vport_info->chunks.num_chunks; i++) {
@@ -246,17 +254,110 @@ idpf_init_vport(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_config_rss(struct idpf_vport *vport)
+{
+   int ret;
+
+   ret = idpf_vc_set_rss_key(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS key");
+   return ret;
+   }
+
+   ret = idpf_vc_set_rss_lut(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS lut");
+   return ret;
+   }
+
+   ret = idpf_vc_set_rss_hash(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS hash");
+   return ret;
+   }
+
+   return ret;
+}
+
+static int
+idpf_init_rss(struct idpf_vport *vport)
+{
+   struct rte_eth_rss_conf *rss_conf;
+   uint16_t i, nb_q, lut_size;
+   int ret = 0;
+
+   rss_conf = &vport->dev_data->dev_conf.rx_adv_conf.rss_conf;
+   nb_q = vport->dev_data->nb_rx_queues;
+
+   vport->rss_key = rte_zmalloc("rss_key",
+vport->rss_key_size, 0);
+   if (vport->rss_key == NULL) {
+   PMD_INIT_LOG(ERR, "Failed to allocate RSS key");
+   ret = -ENOMEM;
+   goto err_alloc_key;
+   }
+
+   lut_size = vport->rss_lut_size;
+   vport->rss_lut = rte_zmalloc("rss_lut",
+sizeof(uint32_t) * lut_size, 0);
+   if (vport->rss_lut == NULL) {
+   PMD_INIT_LOG(ERR, "Failed to allocate RSS lut");
+   ret = -ENOMEM;
+   goto err_alloc_lut;
+   }
+
+   if (rss_conf->rss_key == NULL) {
+   for (i = 0; i < vport->rss_key_size; i++)
+   vport->rss_key[i] = (uint8_t)rte_rand();
+   } else if (rss_conf->rss_key_len != vport->rss_key_size) {
+   PMD_INIT_LOG(ERR, "Invalid RSS key length in RSS configuration, 
should be %d",
+vport->rss_key_size);
+   ret = -EINVAL;
+   goto err_cfg_key;
+   } else {
+   rte_memcpy(vport->rss_key, rss_conf->rss_key,
+  vport->rss_key_size);
+   }
+
+   for (i = 0; i < lut_size; i++)
+   vport->rss_lut[i] = i % nb_q;
+
+   vport->rss_hf = IDPF_DEFAULT_RSS_HASH_EXPANDED;
+
+   ret = idpf_config_rss(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS");
+   goto err_cfg_key;
+   }
+
+   return ret;
+
+err_cfg_key:
+   rte_free(vport->rss_lut);
+   vport->rss_lut = NULL;
+err_alloc_lut:
+   rte_free(vport->rss_key);
+   vport->rss_key = NULL;
+err_alloc_key:
+   return ret;
+}
+
 static int
 idpf_dev_configure(struct rte_eth_dev *dev)
 {
+   struct idpf_vport *vport = dev->data->dev_private;
struct rte_eth_conf *conf = &dev->data->dev_conf;
+   struct idpf_adapter *adapter = vport->adapter;
+   int ret;
 
if (conf->link_speeds & RTE_ETH_LINK_SPEED_FIXED) {
PMD_INIT_LOG(ERR, "Setting link speed is not supported");
return -ENOTSUP;
}
 
-   if (dev->data->nb_rx_queues == 1 && conf->rxmode.mq_mode != 
RTE_ETH_MQ_RX_NONE) {
+   if ((dev->data->nb_rx_queues == 1 && conf->rxmode.mq_mode != 
R

[PATCH v17 15/18] net/idpf: add support for Rx offloading

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add Rx offloading support:
 - support CHKSUM and RSS offload for split queue model
 - support CHKSUM offload for single queue model

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |   5 ++
 drivers/net/idpf/idpf_ethdev.c|   6 ++
 drivers/net/idpf/idpf_rxtx.c  | 123 ++
 drivers/net/idpf/idpf_vchnl.c |  18 +
 4 files changed, 152 insertions(+)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index d722c49fde..868571654f 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -3,8 +3,13 @@
 ;
 ; Refer to default.ini for the full list of available PMD features.
 ;
+; A feature with "P" indicates only be supported when non-vector path
+; is selected.
+;
 [Features]
 MTU update   = Y
+L3 checksum offload  = P
+L4 checksum offload  = P
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 58560ea404..a09f104425 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -61,6 +61,12 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
 
dev_info->flow_type_rss_offloads = IDPF_RSS_OFFLOAD_ALL;
 
+   dev_info->rx_offload_capa =
+   RTE_ETH_RX_OFFLOAD_IPV4_CKSUM   |
+   RTE_ETH_RX_OFFLOAD_UDP_CKSUM|
+   RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
+   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM;
+
dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index a980714060..f15e61a785 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1209,6 +1209,73 @@ idpf_stop_queues(struct rte_eth_dev *dev)
}
 }
 
+#define IDPF_RX_FLEX_DESC_ADV_STATUS0_XSUM_S   \
+   (RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_IPE_S) | \
+RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_L4E_S) | \
+RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EIPE_S) |\
+RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EUDPE_S))
+
+static inline uint64_t
+idpf_splitq_rx_csum_offload(uint8_t err)
+{
+   uint64_t flags = 0;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_L3L4P_S)) == 0))
+   return flags;
+
+   if (likely((err & IDPF_RX_FLEX_DESC_ADV_STATUS0_XSUM_S) == 0)) {
+   flags |= (RTE_MBUF_F_RX_IP_CKSUM_GOOD |
+ RTE_MBUF_F_RX_L4_CKSUM_GOOD);
+   return flags;
+   }
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_IPE_S)) != 0))
+   flags |= RTE_MBUF_F_RX_IP_CKSUM_BAD;
+   else
+   flags |= RTE_MBUF_F_RX_IP_CKSUM_GOOD;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_L4E_S)) != 0))
+   flags |= RTE_MBUF_F_RX_L4_CKSUM_BAD;
+   else
+   flags |= RTE_MBUF_F_RX_L4_CKSUM_GOOD;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EIPE_S)) != 0))
+   flags |= RTE_MBUF_F_RX_OUTER_IP_CKSUM_BAD;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EUDPE_S)) != 0))
+   flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_BAD;
+   else
+   flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_GOOD;
+
+   return flags;
+}
+
+#define IDPF_RX_FLEX_DESC_ADV_HASH1_S  0
+#define IDPF_RX_FLEX_DESC_ADV_HASH2_S  16
+#define IDPF_RX_FLEX_DESC_ADV_HASH3_S  24
+
+static inline uint64_t
+idpf_splitq_rx_rss_offload(struct rte_mbuf *mb,
+  volatile struct virtchnl2_rx_flex_desc_adv_nic_3 
*rx_desc)
+{
+   uint8_t status_err0_qw0;
+   uint64_t flags = 0;
+
+   status_err0_qw0 = rx_desc->status_err0_qw0;
+
+   if ((status_err0_qw0 & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_RSS_VALID_S)) != 0) {
+   flags |= RTE_MBUF_F_RX_RSS_HASH;
+   mb->hash.rss = (rte_le_to_cpu_16(rx_desc->hash1) <<
+   IDPF_RX_FLEX_DESC_ADV_HASH1_S) |
+   ((uint32_t)(rx_desc->ff2_mirrid_hash2.hash2) <<
+IDPF_RX_FLEX_DESC_ADV_HASH2_S) |
+   ((uint32_t)(rx_desc->hash3) <<
+IDPF_RX_FLEX_DESC_ADV_HASH3_S);
+   }
+
+   return flags;
+}
+
 static void
 idpf_split_rx_bufq_refill(struct idpf_rx_queue *rx_bufq)
 {
@@ -1282,9 +1349,11 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
uint16_t pktlen_gen_bufq_id;
struct idpf_rx_queue *rxq;
const uint32_t *ptype_tbl;
+   uint8_t status_err0_qw1;
struct rte_mb

[PATCH v17 16/18] net/idpf: add support for Tx offloading

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add Tx offloading support:
 - support TSO for single queue model and split queue model.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |   1 +
 drivers/net/idpf/idpf_ethdev.c|   4 +-
 drivers/net/idpf/idpf_rxtx.c  | 128 +-
 drivers/net/idpf/idpf_rxtx.h  |  22 +
 4 files changed, 152 insertions(+), 3 deletions(-)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index 868571654f..d82b4aa0ff 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -8,6 +8,7 @@
 ;
 [Features]
 MTU update   = Y
+TSO  = P
 L3 checksum offload  = P
 L4 checksum offload  = P
 Linux= Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index a09f104425..084426260c 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -67,7 +67,9 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM;
 
-   dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
+   dev_info->tx_offload_capa =
+   RTE_ETH_TX_OFFLOAD_TCP_TSO  |
+   RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index f15e61a785..cc296d7ab1 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1506,6 +1506,49 @@ idpf_split_tx_free(struct idpf_tx_queue *cq)
cq->tx_tail = next;
 }
 
+/* Check if the context descriptor is needed for TX offloading */
+static inline uint16_t
+idpf_calc_context_desc(uint64_t flags)
+{
+   if ((flags & RTE_MBUF_F_TX_TCP_SEG) != 0)
+   return 1;
+
+   return 0;
+}
+
+/* set TSO context descriptor
+ */
+static inline void
+idpf_set_splitq_tso_ctx(struct rte_mbuf *mbuf,
+   union idpf_tx_offload tx_offload,
+   volatile union idpf_flex_tx_ctx_desc *ctx_desc)
+{
+   uint16_t cmd_dtype;
+   uint32_t tso_len;
+   uint8_t hdr_len;
+
+   if (tx_offload.l4_len == 0) {
+   PMD_TX_LOG(DEBUG, "L4 length set to 0");
+   return;
+   }
+
+   hdr_len = tx_offload.l2_len +
+   tx_offload.l3_len +
+   tx_offload.l4_len;
+   cmd_dtype = IDPF_TX_DESC_DTYPE_FLEX_TSO_CTX |
+   IDPF_TX_FLEX_CTX_DESC_CMD_TSO;
+   tso_len = mbuf->pkt_len - hdr_len;
+
+   ctx_desc->tso.qw1.cmd_dtype = rte_cpu_to_le_16(cmd_dtype);
+   ctx_desc->tso.qw0.hdr_len = hdr_len;
+   ctx_desc->tso.qw0.mss_rt =
+   rte_cpu_to_le_16((uint16_t)mbuf->tso_segsz &
+IDPF_TXD_FLEX_CTX_MSS_RT_M);
+   ctx_desc->tso.qw0.flex_tlen =
+   rte_cpu_to_le_32(tso_len &
+IDPF_TXD_FLEX_CTX_MSS_RT_M);
+}
+
 uint16_t
 idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
  uint16_t nb_pkts)
@@ -1514,11 +1557,14 @@ idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
volatile struct idpf_flex_tx_sched_desc *txr;
volatile struct idpf_flex_tx_sched_desc *txd;
struct idpf_tx_entry *sw_ring;
+   union idpf_tx_offload tx_offload = {0};
struct idpf_tx_entry *txe, *txn;
uint16_t nb_used, tx_id, sw_id;
struct rte_mbuf *tx_pkt;
uint16_t nb_to_clean;
uint16_t nb_tx = 0;
+   uint64_t ol_flags;
+   uint16_t nb_ctx;
 
if (unlikely(txq == NULL) || unlikely(!txq->q_started))
return nb_tx;
@@ -1548,7 +1594,29 @@ idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
 
if (txq->nb_free < tx_pkt->nb_segs)
break;
-   nb_used = tx_pkt->nb_segs;
+
+   ol_flags = tx_pkt->ol_flags;
+   tx_offload.l2_len = tx_pkt->l2_len;
+   tx_offload.l3_len = tx_pkt->l3_len;
+   tx_offload.l4_len = tx_pkt->l4_len;
+   tx_offload.tso_segsz = tx_pkt->tso_segsz;
+   /* Calculate the number of context descriptors needed. */
+   nb_ctx = idpf_calc_context_desc(ol_flags);
+   nb_used = tx_pkt->nb_segs + nb_ctx;
+
+   /* context descriptor */
+   if (nb_ctx != 0) {
+   volatile union idpf_flex_tx_ctx_desc *ctx_desc =
+   (volatile union idpf_flex_tx_ctx_desc *)&txr[tx_id];
+
+   if ((ol_flags & RTE_MBUF_F_TX_TCP_SEG) != 0)
+   idpf_set_splitq_tso_ctx(tx_pkt, tx_offload,
+   

[PATCH v17 17/18] net/idpf: add AVX512 data path for single queue model

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support of AVX512 vector data path for single queue model.

Signed-off-by: Wenjun Wu 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/idpf.rst|  19 +
 drivers/net/idpf/idpf_ethdev.c  |   3 +-
 drivers/net/idpf/idpf_ethdev.h  |   5 +
 drivers/net/idpf/idpf_rxtx.c| 145 
 drivers/net/idpf/idpf_rxtx.h|  21 +
 drivers/net/idpf/idpf_rxtx_vec_avx512.c | 871 
 drivers/net/idpf/idpf_rxtx_vec_common.h | 100 +++
 drivers/net/idpf/meson.build|  28 +
 8 files changed, 1191 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/idpf/idpf_rxtx_vec_avx512.c
 create mode 100644 drivers/net/idpf/idpf_rxtx_vec_common.h

diff --git a/doc/guides/nics/idpf.rst b/doc/guides/nics/idpf.rst
index c1001d5d0c..3039c61748 100644
--- a/doc/guides/nics/idpf.rst
+++ b/doc/guides/nics/idpf.rst
@@ -64,3 +64,22 @@ Refer to the document :ref:`compiling and testing a PMD for 
a NIC tx_offload_capa =
RTE_ETH_TX_OFFLOAD_TCP_TSO  |
-   RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
+   RTE_ETH_TX_OFFLOAD_MULTI_SEGS   |
+   RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 8d0804f603..7d54e5db60 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -162,6 +162,11 @@ struct idpf_adapter {
uint32_t max_txq_per_msg;
 
uint32_t ptype_tbl[IDPF_MAX_PKT_TYPE] __rte_cache_min_aligned;
+
+   bool rx_vec_allowed;
+   bool tx_vec_allowed;
+   bool rx_use_avx512;
+   bool tx_use_avx512;
 };
 
 TAILQ_HEAD(idpf_adapter_list, idpf_adapter);
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index cc296d7ab1..9e20f2b9d3 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -4,9 +4,11 @@
 
 #include 
 #include 
+#include 
 
 #include "idpf_ethdev.h"
 #include "idpf_rxtx.h"
+#include "idpf_rxtx_vec_common.h"
 
 static int
 check_rx_thresh(uint16_t nb_desc, uint16_t thresh)
@@ -252,6 +254,8 @@ reset_single_rx_queue(struct idpf_rx_queue *rxq)
 
rxq->pkt_first_seg = NULL;
rxq->pkt_last_seg = NULL;
+   rxq->rxrearm_start = 0;
+   rxq->rxrearm_nb = 0;
 }
 
 static void
@@ -2073,25 +2077,166 @@ idpf_prep_pkts(__rte_unused void *tx_queue, struct 
rte_mbuf **tx_pkts,
return i;
 }
 
+static void __rte_cold
+release_rxq_mbufs_vec(struct idpf_rx_queue *rxq)
+{
+   const uint16_t mask = rxq->nb_rx_desc - 1;
+   uint16_t i;
+
+   if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc)
+   return;
+
+   /* free all mbufs that are valid in the ring */
+   if (rxq->rxrearm_nb == 0) {
+   for (i = 0; i < rxq->nb_rx_desc; i++) {
+   if (rxq->sw_ring[i] != NULL)
+   rte_pktmbuf_free_seg(rxq->sw_ring[i]);
+   }
+   } else {
+   for (i = rxq->rx_tail; i != rxq->rxrearm_start; i = (i + 1) & 
mask) {
+   if (rxq->sw_ring[i] != NULL)
+   rte_pktmbuf_free_seg(rxq->sw_ring[i]);
+   }
+   }
+
+   rxq->rxrearm_nb = rxq->nb_rx_desc;
+
+   /* set all entries to NULL */
+   memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc);
+}
+
+static const struct idpf_rxq_ops def_singleq_rx_ops_vec = {
+   .release_mbufs = release_rxq_mbufs_vec,
+};
+
+static inline int
+idpf_singleq_rx_vec_setup_default(struct idpf_rx_queue *rxq)
+{
+   uintptr_t p;
+   struct rte_mbuf mb_def = { .buf_addr = 0 }; /* zeroed mbuf */
+
+   mb_def.nb_segs = 1;
+   mb_def.data_off = RTE_PKTMBUF_HEADROOM;
+   mb_def.port = rxq->port_id;
+   rte_mbuf_refcnt_set(&mb_def, 1);
+
+   /* prevent compiler reordering: rearm_data covers previous fields */
+   rte_compiler_barrier();
+   p = (uintptr_t)&mb_def.rearm_data;
+   rxq->mbuf_initializer = *(uint64_t *)p;
+   return 0;
+}
+
+int __rte_cold
+idpf_singleq_rx_vec_setup(struct idpf_rx_queue *rxq)
+{
+   rxq->ops = &def_singleq_rx_ops_vec;
+   return idpf_singleq_rx_vec_setup_default(rxq);
+}
+
 void
 idpf_set_rx_function(struct rte_eth_dev *dev)
 {
struct idpf_vport *vport = dev->data->dev_private;
+#ifdef RTE_ARCH_X86
+   struct idpf_adapter *ad = vport->adapter;
+   struct idpf_rx_queue *rxq;
+   int i;
+
+   if (idpf_rx_vec_dev_check_default(dev) == IDPF_VECTOR_PATH &&
+   rte_vect_get_max_simd_bitwidth() >= RTE_VECT_SIMD_128) {
+   ad->rx_vec_allowed = true;
+
+   if (rte_vect_get_max_simd_bitwidth() >= RTE_VECT_SIMD_512)
+#ifdef CC_AVX512_SUPPORT
+   if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_AVX512F) == 1 
&&
+  

[PATCH v17 18/18] net/idpf: add support for timestamp offload

2022-10-30 Thread beilei . xing
From: Junfeng Guo 

Add support for timestamp offload.

Signed-off-by: Wenjing Qiao 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |  1 +
 drivers/net/idpf/idpf_ethdev.c|  5 +-
 drivers/net/idpf/idpf_ethdev.h|  3 ++
 drivers/net/idpf/idpf_rxtx.c  | 65 ++
 drivers/net/idpf/idpf_rxtx.h  | 90 +++
 5 files changed, 163 insertions(+), 1 deletion(-)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index d82b4aa0ff..099fd7f216 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -11,6 +11,7 @@ MTU update   = Y
 TSO  = P
 L3 checksum offload  = P
 L4 checksum offload  = P
+Timestamp offload= P
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index cd4ebcc2c6..50aac65daf 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -22,6 +22,8 @@ rte_spinlock_t idpf_adapter_lock;
 struct idpf_adapter_list idpf_adapter_list;
 bool idpf_adapter_list_init;
 
+uint64_t idpf_timestamp_dynflag;
+
 static const char * const idpf_valid_args[] = {
IDPF_TX_SINGLE_Q,
IDPF_RX_SINGLE_Q,
@@ -65,7 +67,8 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
RTE_ETH_RX_OFFLOAD_IPV4_CKSUM   |
RTE_ETH_RX_OFFLOAD_UDP_CKSUM|
RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
-   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM;
+   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM |
+   RTE_ETH_RX_OFFLOAD_TIMESTAMP;
 
dev_info->tx_offload_capa =
RTE_ETH_TX_OFFLOAD_TCP_TSO  |
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 7d54e5db60..ccdf4abe40 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -167,6 +167,9 @@ struct idpf_adapter {
bool tx_vec_allowed;
bool rx_use_avx512;
bool tx_use_avx512;
+
+   /* For PTP */
+   uint64_t time_hw;
 };
 
 TAILQ_HEAD(idpf_adapter_list, idpf_adapter);
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 9e20f2b9d3..bafa007faf 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -10,6 +10,8 @@
 #include "idpf_rxtx.h"
 #include "idpf_rxtx_vec_common.h"
 
+static int idpf_timestamp_dynfield_offset = -1;
+
 static int
 check_rx_thresh(uint16_t nb_desc, uint16_t thresh)
 {
@@ -900,6 +902,24 @@ idpf_tx_queue_setup(struct rte_eth_dev *dev, uint16_t 
queue_idx,
return idpf_tx_split_queue_setup(dev, queue_idx, nb_desc,
 socket_id, tx_conf);
 }
+
+static int
+idpf_register_ts_mbuf(struct idpf_rx_queue *rxq)
+{
+   int err;
+   if ((rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) != 0) {
+   /* Register mbuf field and flag for Rx timestamp */
+   err = 
rte_mbuf_dyn_rx_timestamp_register(&idpf_timestamp_dynfield_offset,
+
&idpf_timestamp_dynflag);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR,
+   "Cannot register mbuf field/flag for 
timestamp");
+   return -EINVAL;
+   }
+   }
+   return 0;
+}
+
 static int
 idpf_alloc_single_rxq_mbufs(struct idpf_rx_queue *rxq)
 {
@@ -993,6 +1013,13 @@ idpf_rx_queue_init(struct rte_eth_dev *dev, uint16_t 
rx_queue_id)
return -EINVAL;
}
 
+   err = idpf_register_ts_mbuf(rxq);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR, "fail to regidter timestamp mbuf %u",
+   rx_queue_id);
+   return -EIO;
+   }
+
if (rxq->bufq1 == NULL) {
/* Single queue */
err = idpf_alloc_single_rxq_mbufs(rxq);
@@ -1354,6 +1381,7 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
struct idpf_rx_queue *rxq;
const uint32_t *ptype_tbl;
uint8_t status_err0_qw1;
+   struct idpf_adapter *ad;
struct rte_mbuf *rxm;
uint16_t rx_id_bufq1;
uint16_t rx_id_bufq2;
@@ -1363,9 +1391,11 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
uint16_t gen_id;
uint16_t rx_id;
uint16_t nb_rx;
+   uint64_t ts_ns;
 
nb_rx = 0;
rxq = rx_queue;
+   ad = rxq->adapter;
 
if (unlikely(rxq == NULL) || unlikely(!rxq->q_started))
return nb_rx;
@@ -1376,6 +1406,9 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rx_desc_ring = rxq->rx_ring;
ptype_tbl = rxq->adapter->ptype_tbl;
 
+   if ((rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) != 0)
+   rxq->hw_register

[PATCH v1] net/iavf: fix refine protocol header error

2022-10-30 Thread Steve Yang
Protocol header count should be changed when tunnel level is larger than 1.

Fixes: 13a7dcddd8ee ("net/iavf: fix taninted scalar")

Signed-off-by: Steve Yang 
---
 drivers/net/iavf/iavf_hash.c | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/drivers/net/iavf/iavf_hash.c b/drivers/net/iavf/iavf_hash.c
index 71c80d2c75..67b05313eb 100644
--- a/drivers/net/iavf/iavf_hash.c
+++ b/drivers/net/iavf/iavf_hash.c
@@ -1210,7 +1210,7 @@ iavf_refine_proto_hdrs_by_pattern(struct 
virtchnl_proto_hdrs *proto_hdrs,
struct virtchnl_proto_hdr *hdr2;
int i, shift_count = 1;
int tun_lvl = proto_hdrs->tunnel_level;
-   int phdrs_count = proto_hdrs->count;
+   int phdrs_count = 0;
 
if (!(phint & IAVF_PHINT_GTPU_MSK) && !(phint & IAVF_PHINT_GRE))
return;
@@ -1219,6 +1219,7 @@ iavf_refine_proto_hdrs_by_pattern(struct 
virtchnl_proto_hdrs *proto_hdrs,
if (phint & IAVF_PHINT_LAYERS_MSK)
shift_count = 2;
 
+   phdrs_count = proto_hdrs->count;
/* shift headers layer */
for (i = phdrs_count - 1 + shift_count;
 i > shift_count - 1; i--) {
-- 
2.25.1