Re: [dpdk-dev] [PATCH] net/bnxt: add support for 200G PAM4 link

2020-09-25 Thread Ajit Khaparde
On Thu, Sep 24, 2020 at 4:53 PM Ajit Khaparde
 wrote:
>
> Thor based NICs can support PAM4 as wells as NRZ link negotiation.
> With this patch we are adding support for 200G link speeds based on
> PAM4 signaling. While PAM4 can negotiate speeds for 50G and 100G as
> well, the PMD will use NRZ signalling for these speeds.
>
> Signed-off-by: Ajit Khaparde 
> Reviewed-by: Somnath Kotur 
> --
> v1->v2: Updated release notes
> ---
Patch applied to dpdk-next-net-brcm.

>  doc/guides/rel_notes/release_20_11.rst |   6 +
>  drivers/net/bnxt/bnxt.h|   7 ++
>  drivers/net/bnxt/bnxt_ethdev.c |   6 +-
>  drivers/net/bnxt/bnxt_hwrm.c   |  77 +---
>  drivers/net/bnxt/bnxt_hwrm.h   |   7 ++
>  drivers/net/bnxt/hsi_struct_def_dpdk.h | 168 +++--
>  6 files changed, 220 insertions(+), 51 deletions(-)
>
> diff --git a/doc/guides/rel_notes/release_20_11.rst 
> b/doc/guides/rel_notes/release_20_11.rst
> index c6642f5f9..a02c40dcd 100644
> --- a/doc/guides/rel_notes/release_20_11.rst
> +++ b/doc/guides/rel_notes/release_20_11.rst
> @@ -78,6 +78,12 @@ New Features
>  ``--portmask=N``
>  where N represents the hexadecimal bitmask of ports used.
>
> +* **Updated Broadcom bnxt driver.**
> +
> +  Updated the Broadcom bnxt driver with new features and improvements, 
> including:
> +
> +  * Added support for 200G PAM4 link speed.
> +
>
>  Removed Items
>  -
> diff --git a/drivers/net/bnxt/bnxt.h b/drivers/net/bnxt/bnxt.h
> index 3dded37bb..19d4774c5 100644
> --- a/drivers/net/bnxt/bnxt.h
> +++ b/drivers/net/bnxt/bnxt.h
> @@ -292,6 +292,13 @@ struct bnxt_link_info {
> uint32_tpreemphasis;
> uint8_t phy_type;
> uint8_t media_type;
> +   uint16_tsupport_auto_speeds;
> +   uint8_t link_signal_mode;
> +   uint16_tforce_pam4_link_speed;
> +   uint16_tsupport_pam4_speeds;
> +   uint16_tauto_pam4_link_speeds;
> +   uint16_tsupport_pam4_auto_speeds;
> +   uint8_t req_signal_mode;
>  };
>
>  #define BNXT_COS_QUEUE_COUNT   8
> diff --git a/drivers/net/bnxt/bnxt_ethdev.c b/drivers/net/bnxt/bnxt_ethdev.c
> index bb6b25dff..af392bffc 100644
> --- a/drivers/net/bnxt/bnxt_ethdev.c
> +++ b/drivers/net/bnxt/bnxt_ethdev.c
> @@ -864,7 +864,11 @@ uint32_t bnxt_get_speed_capabilities(struct bnxt *bp)
> speed_capa |= ETH_LINK_SPEED_50G;
> if (link_speed & HWRM_PORT_PHY_QCFG_OUTPUT_SUPPORT_SPEEDS_100GB)
> speed_capa |= ETH_LINK_SPEED_100G;
> -   if (link_speed & HWRM_PORT_PHY_QCFG_OUTPUT_SUPPORT_SPEEDS_200GB)
> +   if (link_speed & HWRM_PORT_PHY_QCFG_OUTPUT_SUPPORT_PAM4_SPEEDS_50G)
> +   speed_capa |= ETH_LINK_SPEED_50G;
> +   if (link_speed & HWRM_PORT_PHY_QCFG_OUTPUT_SUPPORT_PAM4_SPEEDS_100G)
> +   speed_capa |= ETH_LINK_SPEED_100G;
> +   if (link_speed & HWRM_PORT_PHY_QCFG_OUTPUT_SUPPORT_PAM4_SPEEDS_200G)
> speed_capa |= ETH_LINK_SPEED_200G;
>
> if (bp->link_info->auto_mode ==
> diff --git a/drivers/net/bnxt/bnxt_hwrm.c b/drivers/net/bnxt/bnxt_hwrm.c
> index f25cdf05c..a9d5fd77e 100644
> --- a/drivers/net/bnxt/bnxt_hwrm.c
> +++ b/drivers/net/bnxt/bnxt_hwrm.c
> @@ -1260,16 +1260,24 @@ static int bnxt_hwrm_port_phy_cfg(struct bnxt *bp, 
> struct bnxt_link_info *conf)
> }
>
> req.flags = rte_cpu_to_le_32(conf->phy_flags);
> -   req.force_link_speed = rte_cpu_to_le_16(conf->link_speed);
> -   enables |= HWRM_PORT_PHY_CFG_INPUT_ENABLES_AUTO_MODE;
> /*
>  * Note, ChiMP FW 20.2.1 and 20.2.2 return an error when we 
> set
>  * any auto mode, even "none".
>  */
> if (!conf->link_speed) {
> /* No speeds specified. Enable AutoNeg - all speeds */
> +   enables |= HWRM_PORT_PHY_CFG_INPUT_ENABLES_AUTO_MODE;
> req.auto_mode =
> HWRM_PORT_PHY_CFG_INPUT_AUTO_MODE_ALL_SPEEDS;
> +   } else {
> +   if (bp->link_info->link_signal_mode) {
> +   enables |=
> +   HWRM_PORT_PHY_CFG_IN_EN_FORCE_PAM4_LINK_SPEED;
> +   req.force_pam4_link_speed =
> +   rte_cpu_to_le_16(conf->link_speed);
> +   }
> +   req.force_link_speed =
> +   rte_cpu_to_le_16(conf->link_speed);
> }
> /* AutoNeg - Advertise speeds specified. */
> if (conf->auto_link_speed_mask &&
> @@ -1278,9 +1286,20 @@ static int bnxt_hwrm_port_phy_cfg(struct bnxt *bp, 
> struct bnxt_link_info *conf)
> 

Re: [dpdk-dev] [PATCH] net/bnx2x: enhance old ethdev close behavior

2020-09-25 Thread Thomas Monjalon
25/09/2020 06:17, Rasesh Mody:
> Set RTE_ETH_DEV_CLOSE_REMOVE upon probe so all the private resources
> for the port can be freed by rte_eth_dev_close(). With this change the
> private port resources are released in the .dev_close callback.
> 
> Signed-off-by: Rasesh Mody 
> ---
>  static int eth_bnx2x_dev_uninit(struct rte_eth_dev *eth_dev)
>  {
> - /* mac_addrs must not be freed alone because part of dev_private */
> - eth_dev->data->mac_addrs = NULL;
> + struct bnx2x_softc *sc = eth_dev->data->dev_private;
> +
> + PMD_INIT_FUNC_TRACE(sc);
> +
> + /* only uninitialize in the primary process */
> + if (rte_eal_process_type() != RTE_PROC_PRIMARY)
> + return 0;
> +
> + bnx2x_dev_close(eth_dev);

The check for primary process should be in the close function.

About the title, I would suggest net/bnx2x: release port upon close




[dpdk-dev] [PATCH v2 3/3] test/event_crypto_adapter: free resources during exit

2020-09-25 Thread Ankur Dwivedi
The resources held by crypto adapter should be freed when the
test suite exits.

Signed-off-by: Ankur Dwivedi 
---
 app/test/test_event_crypto_adapter.c | 15 +++
 1 file changed, 15 insertions(+)

diff --git a/app/test/test_event_crypto_adapter.c 
b/app/test/test_event_crypto_adapter.c
index 572ef9c27..4b09ab3b7 100644
--- a/app/test/test_event_crypto_adapter.c
+++ b/app/test/test_event_crypto_adapter.c
@@ -899,6 +899,20 @@ testsuite_setup(void)
return TEST_SUCCESS;
 }
 
+static void
+crypto_adapter_teardown(void)
+{
+   int ret;
+
+   crypto_adapter_setup_done = 0;
+   ret = rte_event_crypto_adapter_queue_pair_del(TEST_ADAPTER_ID,
+   TEST_CDEV_ID, TEST_CDEV_QP_ID);
+   if (ret < 0)
+   RTE_LOG(ERR, USER1, "Failed to delete queue pair!");
+
+   rte_event_crypto_adapter_free(TEST_ADAPTER_ID);
+}
+
 static void
 crypto_teardown(void)
 {
@@ -941,6 +955,7 @@ eventdev_teardown(void)
 static void
 testsuite_teardown(void)
 {
+   crypto_adapter_teardown();
crypto_teardown();
eventdev_teardown();
 }
-- 
2.28.0



[dpdk-dev] [PATCH v2 2/3] test/event_crypto_adapter: fix function arguments

2020-09-25 Thread Ankur Dwivedi
The arguments passed to rte_event_crypto_adapter_caps_get() and
rte_event_crypto_adapter_create() are incorrect.

In the rte_event_crypto_adapter_caps_get(), event device id should
be the first argument and cryptodev id should be the second argument.
In the rte_event_crypto_adapter_create(), the event device id should
be the second argument.

Signed-off-by: Ankur Dwivedi 
---
 app/test/test_event_crypto_adapter.c | 18 +-
 1 file changed, 9 insertions(+), 9 deletions(-)

diff --git a/app/test/test_event_crypto_adapter.c 
b/app/test/test_event_crypto_adapter.c
index 3c64742b3..572ef9c27 100644
--- a/app/test/test_event_crypto_adapter.c
+++ b/app/test/test_event_crypto_adapter.c
@@ -209,8 +209,8 @@ test_op_forward_mode(uint8_t session_less)
&cipher_xform, params.session_priv_mpool);
TEST_ASSERT_SUCCESS(ret, "Failed to init session\n");
 
-   ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID,
-   evdev, &cap);
+   ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID,
+   &cap);
TEST_ASSERT_SUCCESS(ret, "Failed to get adapter 
capabilities\n");
 
if (cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_SESSION_PRIVATE_DATA) {
@@ -321,7 +321,7 @@ test_session_with_op_forward_mode(void)
uint32_t cap;
int ret;
 
-   ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID, evdev, &cap);
+   ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID, &cap);
TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
 
if (!(cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD) &&
@@ -410,8 +410,8 @@ test_op_new_mode(uint8_t session_less)
params.session_mpool);
TEST_ASSERT_NOT_NULL(sess, "Session creation failed\n");
 
-   ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID,
-   evdev, &cap);
+   ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID,
+   &cap);
TEST_ASSERT_SUCCESS(ret, "Failed to get adapter 
capabilities\n");
 
if (cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_SESSION_PRIVATE_DATA) {
@@ -460,7 +460,7 @@ test_sessionless_with_op_new_mode(void)
uint32_t cap;
int ret;
 
-   ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID, evdev, &cap);
+   ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID, &cap);
TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
 
if (!(cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD) &&
@@ -486,7 +486,7 @@ test_session_with_op_new_mode(void)
uint32_t cap;
int ret;
 
-   ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID, evdev, &cap);
+   ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID, &cap);
TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
 
if (!(cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD) &&
@@ -706,7 +706,7 @@ test_crypto_adapter_create(void)
 
/* Create adapter with default port creation callback */
ret = rte_event_crypto_adapter_create(TEST_ADAPTER_ID,
- TEST_CDEV_ID,
+ evdev,
  &conf, 0);
TEST_ASSERT_SUCCESS(ret, "Failed to create event crypto adapter\n");
 
@@ -719,7 +719,7 @@ test_crypto_adapter_qp_add_del(void)
uint32_t cap;
int ret;
 
-   ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID, evdev, &cap);
+   ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID, &cap);
TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
 
if (cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_QP_EV_BIND) {
-- 
2.28.0



[dpdk-dev] [PATCH v2 0/3] Bug fixes in event crypto adapter test application

2020-09-25 Thread Ankur Dwivedi
This patch series resolves bugs in the event crypto adapter test
application.

v2:
* In the "test/event_crypto_adapter: return error with unsupported mode"
  patch the following changes were done:
  - Skip mismatch of mode and capability for software eventdev.
  - In test_crypto_adapter_conf_op_new_mode() and
test_crypto_adapter_conf_op_forward_mode(),the local variable
for holding return of test_crypto_adapter_conf() is removed.
  - Minor change in commit message.

Ankur Dwivedi (3):
  test/event_crypto_adapter: return error with unsupported mode
  test/event_crypto_adapter: fix function arguments
  test/event_crypto_adapter: free resources during exit

 app/test/test_event_crypto_adapter.c | 66 
 1 file changed, 47 insertions(+), 19 deletions(-)

-- 
2.28.0



[dpdk-dev] [PATCH v2 1/3] test/event_crypto_adapter: return error with unsupported mode

2020-09-25 Thread Ankur Dwivedi
The capability of a hardware event device should be checked before
creating a event crypto adapter in a particular mode. The test case
returns error if the mode is not supported.

Signed-off-by: Ankur Dwivedi 
---
 app/test/test_event_crypto_adapter.c | 33 +++-
 1 file changed, 23 insertions(+), 10 deletions(-)

diff --git a/app/test/test_event_crypto_adapter.c 
b/app/test/test_event_crypto_adapter.c
index 8d42462d8..3c64742b3 100644
--- a/app/test/test_event_crypto_adapter.c
+++ b/app/test/test_event_crypto_adapter.c
@@ -750,15 +750,30 @@ configure_event_crypto_adapter(enum 
rte_event_crypto_adapter_mode mode)
uint32_t cap;
int ret;
 
+   ret = rte_event_crypto_adapter_caps_get(evdev, TEST_CDEV_ID, &cap);
+   TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
+
+   /* Skip mode and capability mismatch check for SW eventdev */
+   if (!(cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_NEW) &&
+   !(cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD) &&
+   !(cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_QP_EV_BIND))
+   goto adapter_create;
+
+   if ((mode == RTE_EVENT_CRYPTO_ADAPTER_OP_FORWARD) &&
+   !(cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_FWD))
+   return -ENOTSUP;
+
+   if ((mode == RTE_EVENT_CRYPTO_ADAPTER_OP_NEW) &&
+   !(cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_OP_NEW))
+   return -ENOTSUP;
+
+adapter_create:
/* Create adapter with default port creation callback */
ret = rte_event_crypto_adapter_create(TEST_ADAPTER_ID,
- TEST_CDEV_ID,
+ evdev,
  &conf, mode);
TEST_ASSERT_SUCCESS(ret, "Failed to create event crypto adapter\n");
 
-   ret = rte_event_crypto_adapter_caps_get(TEST_ADAPTER_ID, evdev, &cap);
-   TEST_ASSERT_SUCCESS(ret, "Failed to get adapter capabilities\n");
-
if (cap & RTE_EVENT_CRYPTO_ADAPTER_CAP_INTERNAL_PORT_QP_EV_BIND) {
ret = rte_event_crypto_adapter_queue_pair_add(TEST_ADAPTER_ID,
TEST_CDEV_ID, TEST_CDEV_QP_ID, &response_info);
@@ -813,6 +828,8 @@ test_crypto_adapter_conf(enum rte_event_crypto_adapter_mode 
mode)
TEST_ASSERT(ret >= 0, "Failed to link queue %d "
"port=%u\n", qid,
params.crypto_event_port_id);
+   } else {
+   return ret;
}
crypto_adapter_setup_done = 1;
}
@@ -847,10 +864,8 @@ test_crypto_adapter_conf_op_forward_mode(void)
enum rte_event_crypto_adapter_mode mode;
 
mode = RTE_EVENT_CRYPTO_ADAPTER_OP_FORWARD;
-   TEST_ASSERT_SUCCESS(test_crypto_adapter_conf(mode),
-   "Failed to config crypto adapter");
 
-   return TEST_SUCCESS;
+   return test_crypto_adapter_conf(mode);
 }
 
 static int
@@ -859,10 +874,8 @@ test_crypto_adapter_conf_op_new_mode(void)
enum rte_event_crypto_adapter_mode mode;
 
mode = RTE_EVENT_CRYPTO_ADAPTER_OP_NEW;
-   TEST_ASSERT_SUCCESS(test_crypto_adapter_conf(mode),
-   "Failed to config crypto adapter");
 
-   return TEST_SUCCESS;
+   return test_crypto_adapter_conf(mode);
 }
 
 
-- 
2.28.0



Re: [dpdk-dev] Meson Minimum Version

2020-09-25 Thread Morten Brørup
> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Bruce Richardson
> Sent: Thursday, September 24, 2020 5:43 PM
> 
> On Thu, Sep 24, 2020 at 03:32:41PM +, John Alexander wrote:
> >Hi,
> >Regarding the subproject local patch support, yes, it's only
> supported
> >since Meson 0.55:
> >https://mesonbuild.com/Wrap-dependency-system-manual.html We pip
> >installed 0.55 Meson.
> >I have a number of subsequent patches that depend on this
> particular
> >pthreads library to advance the Windows DPDK support. Locally, we
> have
> >testpmd (minus mmap'd external memory currently) running against
> the
> >Intel i40e PMD (XL710 4x10Gbps SPF+ NIC) on Windows on our local
> DPDK
> >fork (based off 20.08-rc2 using Microsoft's latest NetUIO driver).
> We
> >have 47 of the 51 RTE libraries building and have had l2fwd,
> l3fwd,
> >ipv4_multicast and almost all of the regression tests
> compiling+linking
> >too. I'd like to push as much of the Windows EAL work we've done
> >upstream if I can (after a bit of tidying up :).
> >I've also coded up a meson build patch for the Jansson JSON parser
> used
> >by the RTE metrics library (the config.h generation was quite
> fiddly!)
> >That's ready to go. We get nice meson syntax as follows to specify
> a
> >fallback if the library isn't installed locally:
> >jansson = dependency('jansson', required: false, fallback :
> ['jansson',
> >'jansson_static_dep'])
> >I believe the meson command line enables disabling fallbacks if
> people
> >would prefer not to use them (--wrap-mode=nofallback).
> >Kind regards,
> >John.
> 
> Hi again, John,
> 
> thanks for the full reply - the work you have sounds really good, and
> the
> possibilities using the wrap support are definitely of interest.
> [Though
> since jansson is only used for the legacy parts of the telemetry
> library
> I'm not sure we want to wrap it just for that - the later telemetry
> work
> from this year doesn't depend on jansson. Then again, the
> vm_power_manager
> example also uses it too...]. It's probably something that would be
> especially useful for software dependencies that aren't normally
> packaged.
> 
> I've added techboard on CC to previous reply, so hopefully we'll get
> some
> thoughts from others.
> 

A buggy C compiler might generate buggy code, and the compiler induced bugs may 
not be caught during development (but hopefully during testing). I have seen 
this happen, and I still consider it a realistic scenario. This is the 
strongest argument for using a trusted version of the C compiler, which is 
supported by the popular GNU/Linux distributors (Red Hat etc.), or sufficiently 
vetted to be included in the Crosstool-NG project, or similar.

Meson is a relatively simple tool, so I personally wouldn't mind overwriting 
the distro-supported version on our build systems with a new version. (I 
haven't worked much with Meson, but assume that the new version of Meson is 
backwards compatible with its earlier versions.)

In short, my primary concern is: What could realistically go wrong if the 
required version of Meson is buggy?

Bruce, you have worked for quite a while with Meson/Ninja by now, so perhaps 
you can assess this risk based on your experience.


> Regards,
> /Bruce
> 
> PS: please don't top-post in replying and keep plain text email if
> possible. Thanks.
> 
> >--
> 
> >Date: Thu, 24 Sep 2020 15:38:30 +0100
> >From: Bruce Richardson 
> >To: John Alexander 
> >Cc: "dev@dpdk.org" , techbo...@dpdk.org
> >Subject: Re: [dpdk-dev] Meson Minimum Version
> >Message-ID: <20200924143830.gd...@bricha3-mobl.ger.corp.intel.com>
> >Content-Type: text/plain; charset=us-ascii
> >On Thu, Sep 24, 2020 at 02:22:03PM +, John Alexander wrote:
> >> Hi,
> >>
> >> I've submitted a patch that uses new features of Meson,
> specifically
> >the directory patch aspect of the subproject feature. This
> requires a
> >minimum Meson version of 0.55.0. How do we go about getting the
> >community to accept a more recent version of Meson and getting the
> >Travis server upgraded too so the CI builds succeed?
> >>
> >> Patch link for reference: http://patches.dpdk.org/patch/78675/
> >>
> >Hi John,
> >from what I understand the specific dependency on 0.55 is the
> support
> >for local patchfiles for the wrapped software, and that previous
> >versions only supported using patches pulled remotely - is that
> >correct?
> >While I'm in favour of incrementing the minimum meson version in
> >general, since 0.55 is the very latest version I am worried about
> any
> >impacts that might have, since it will basically mean that
> everyone
> >building DPDK has to pull meson from pip rather than being able to
> use
> >a distro-supplied version. Updating to something a 

Re: [dpdk-dev] [dpdk-dev v10 2/4] cryptodev: add raw crypto data-path APIs

2020-09-25 Thread Dybkowski, AdamX
> -Original Message-
> From: Zhang, Roy Fan 
> Sent: Thursday, 24 September, 2020 18:34
> To: dev@dpdk.org
> Cc: akhil.go...@nxp.com; Trahe, Fiona ; Kusztal,
> ArkadiuszX ; Dybkowski, AdamX
> ; ano...@marvell.com; Ananyev, Konstantin
> ; Zhang, Roy Fan
> ; Bronowski, PiotrX
> 
> Subject: [dpdk-dev v10 2/4] cryptodev: add raw crypto data-path APIs
> 
> This patch adds raw data-path APIs for enqueue and dequeue operations to
> cryptodev. The APIs support flexible user-define enqueue and dequeue
> behaviors.
> 
> Signed-off-by: Fan Zhang 
> Signed-off-by: Piotr Bronowski 

Acked-by: Adam Dybkowski 



Re: [dpdk-dev] [dpdk-dev v10 1/4] cryptodev: change crypto symmetric vector structure

2020-09-25 Thread Dybkowski, AdamX
> -Original Message-
> From: Zhang, Roy Fan 
> Sent: Thursday, 24 September, 2020 18:34
> To: dev@dpdk.org
> Cc: akhil.go...@nxp.com; Trahe, Fiona ; Kusztal,
> ArkadiuszX ; Dybkowski, AdamX
> ; ano...@marvell.com; Ananyev, Konstantin
> ; Zhang, Roy Fan
> 
> Subject: [dpdk-dev v10 1/4] cryptodev: change crypto symmetric vector
> structure
> 
> This patch updates ``rte_crypto_sym_vec`` structure to add support for both
> cpu_crypto synchrounous operation and asynchronous raw data-path APIs.
> The patch also includes AESNI-MB and AESNI-GCM PMD changes, unit test
> changes and documentation updates.
> 
> Signed-off-by: Fan Zhang 

Acked-by: Adam Dybkowski 



Re: [dpdk-dev] [dpdk-dev v10 4/4] test/crypto: add unit-test for cryptodev raw API test

2020-09-25 Thread Dybkowski, AdamX
> -Original Message-
> From: Zhang, Roy Fan 
> Sent: Thursday, 24 September, 2020 18:34
> To: dev@dpdk.org
> Cc: akhil.go...@nxp.com; Trahe, Fiona ; Kusztal,
> ArkadiuszX ; Dybkowski, AdamX
> ; ano...@marvell.com; Ananyev, Konstantin
> ; Zhang, Roy Fan
> 
> Subject: [dpdk-dev v10 4/4] test/crypto: add unit-test for cryptodev raw API
> test
> 
> This patch adds the cryptodev raw API test support to unit test.
> In addtion a new test-case for QAT PMD for the test type is
> enabled.
> 
> Signed-off-by: Fan Zhang 

Acked-by: Adam Dybkowski 



Re: [dpdk-dev] [dpdk-dev v10 3/4] crypto/qat: add raw crypto data-path API support

2020-09-25 Thread Dybkowski, AdamX
> -Original Message-
> From: Zhang, Roy Fan 
> Sent: Thursday, 24 September, 2020 18:34
> To: dev@dpdk.org
> Cc: akhil.go...@nxp.com; Trahe, Fiona ; Kusztal,
> ArkadiuszX ; Dybkowski, AdamX
> ; ano...@marvell.com; Ananyev, Konstantin
> ; Zhang, Roy Fan
> 
> Subject: [dpdk-dev v10 3/4] crypto/qat: add raw crypto data-path API
> support
> 
> This patch updates QAT PMD to add raw data-path API support.
> 
> Signed-off-by: Fan Zhang 

Acked-by: Adam Dybkowski 



Re: [dpdk-dev] Meson Minimum Version

2020-09-25 Thread Bruce Richardson
On Fri, Sep 25, 2020 at 09:31:53AM +0200, Morten Brørup wrote:
> > From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Bruce Richardson
> > Sent: Thursday, September 24, 2020 5:43 PM
> > 
> > On Thu, Sep 24, 2020 at 03:32:41PM +, John Alexander wrote:
> > >Hi,
> > >Regarding the subproject local patch support, yes, it's only
> > supported
> > >since Meson 0.55:
> > >https://mesonbuild.com/Wrap-dependency-system-manual.html We pip
> > >installed 0.55 Meson.
> > >I have a number of subsequent patches that depend on this
> > particular
> > >pthreads library to advance the Windows DPDK support. Locally, we
> > have
> > >testpmd (minus mmap'd external memory currently) running against
> > the
> > >Intel i40e PMD (XL710 4x10Gbps SPF+ NIC) on Windows on our local
> > DPDK
> > >fork (based off 20.08-rc2 using Microsoft's latest NetUIO driver).
> > We
> > >have 47 of the 51 RTE libraries building and have had l2fwd,
> > l3fwd,
> > >ipv4_multicast and almost all of the regression tests
> > compiling+linking
> > >too. I'd like to push as much of the Windows EAL work we've done
> > >upstream if I can (after a bit of tidying up :).
> > >I've also coded up a meson build patch for the Jansson JSON parser
> > used
> > >by the RTE metrics library (the config.h generation was quite
> > fiddly!)
> > >That's ready to go. We get nice meson syntax as follows to specify
> > a
> > >fallback if the library isn't installed locally:
> > >jansson = dependency('jansson', required: false, fallback :
> > ['jansson',
> > >'jansson_static_dep'])
> > >I believe the meson command line enables disabling fallbacks if
> > people
> > >would prefer not to use them (--wrap-mode=nofallback).
> > >Kind regards,
> > >John.
> > 
> > Hi again, John,
> > 
> > thanks for the full reply - the work you have sounds really good, and
> > the
> > possibilities using the wrap support are definitely of interest.
> > [Though
> > since jansson is only used for the legacy parts of the telemetry
> > library
> > I'm not sure we want to wrap it just for that - the later telemetry
> > work
> > from this year doesn't depend on jansson. Then again, the
> > vm_power_manager
> > example also uses it too...]. It's probably something that would be
> > especially useful for software dependencies that aren't normally
> > packaged.
> > 
> > I've added techboard on CC to previous reply, so hopefully we'll get
> > some
> > thoughts from others.
> > 
> 
> A buggy C compiler might generate buggy code, and the compiler induced bugs 
> may not be caught during development (but hopefully during testing). I have 
> seen this happen, and I still consider it a realistic scenario. This is the 
> strongest argument for using a trusted version of the C compiler, which is 
> supported by the popular GNU/Linux distributors (Red Hat etc.), or 
> sufficiently vetted to be included in the Crosstool-NG project, or similar.
> 
> Meson is a relatively simple tool, so I personally wouldn't mind overwriting 
> the distro-supported version on our build systems with a new version. (I 
> haven't worked much with Meson, but assume that the new version of Meson is 
> backwards compatible with its earlier versions.)
> 
> In short, my primary concern is: What could realistically go wrong if the 
> required version of Meson is buggy?
> 
> Bruce, you have worked for quite a while with Meson/Ninja by now, so perhaps 
> you can assess this risk based on your experience.
> 
I'd say the risk in this case is small, especially since I see that 0.56 of
meson is well under way for development and may well be released before
DPDK 20.11. Generally backwards compatibilty of meson is excellent as they
have comprehensive test suite for all features.

Rather than any bugginess, my concern was purely requiring people to update
meson using "pip3", but I suppose that's not really a big deal, and when
using pip update it defaults to just updating the copy for the local user,
not system-wide.

Updating to a later meson will give us access to a few other nice features
we can use also (from earlier releases than 0.55), such as support for
"break" and "continue" keywords in loops, which can simplify out lib or
driver building code, perhaps, or the "console" parameter for custom
targets which should improve the output visibility when builing our docs.

/Bruce


[dpdk-dev] [PATCH V13 0/3] add FEC support

2020-09-25 Thread Min Hu (Connor)
This series add FEC support for ethdev.

Min Hu (Connor) (3):
  ethdev: introduce FEC API
  net/hns3: support FEC
  app/testpmd: add FEC command

 app/test-pmd/cmdline.c   | 223 +++
 app/test-pmd/config.c|  91 
 app/test-pmd/testpmd.h   |   2 +
 doc/guides/rel_notes/release_20_11.rst   |  10 +
 drivers/net/hns3/hns3_cmd.h  |  19 +-
 drivers/net/hns3/hns3_ethdev.c   | 356 +++
 drivers/net/hns3/hns3_ethdev.h   |   1 +
 lib/librte_ethdev/rte_ethdev.c   |  43 
 lib/librte_ethdev/rte_ethdev.h   |  94 
 lib/librte_ethdev/rte_ethdev_driver.h|  80 +++
 lib/librte_ethdev/rte_ethdev_version.map |   3 +
 11 files changed, 921 insertions(+), 1 deletion(-)

-- 
2.7.4



[dpdk-dev] [PATCH V13 1/3] ethdev: introduce FEC API

2020-09-25 Thread Min Hu (Connor)
This patch adds Forward error correction(FEC) support for ethdev.
Introduce APIs which support query and config FEC information in
hardware.

Signed-off-by: Min Hu (Connor) 
Reviewed-by: Wei Hu (Xavier) 
Reviewed-by: Chengwen Feng 
Reviewed-by: Chengchang Tang 
---
v12->v13:
change fec get capa API.
fix comment styles.

---
v10->v11:
allow to report capabilities per link speed.
specify what should be reported if link is down
when get FEC.
change mode to capa bitmask.

---
v9->v10:
add macro RTE_ETH_FEC_MODE_CAPA_MASK(x) to indicate
different FEC mode capa.

---
v8->v9:
added reviewed-by and acked-by.

---
v7->v8:
put AUTO just after NOFEC in rte_fec_mode definition.

---
v6->v7:
deleted RTE_ETH_FEC_NUM to prevent ABI breakage.
add new macro to indicate translation from fec mode
to capa.

---
v5->v6:
modified release notes.
deleted check duplicated for FEC API
fixed code styles according to DPDK coding style.
added _eth prefix.

---
v4->v5:
Modifies FEC capa definitions using macros.
Add RTE_ prefix for public FEC mode enum.
add release notes about FEC for dpdk20_11.

---
v2->v3:
add function return value "-ENOTSUP" for API.

---
 doc/guides/rel_notes/release_20_11.rst   |  5 ++
 lib/librte_ethdev/rte_ethdev.c   | 43 +++
 lib/librte_ethdev/rte_ethdev.h   | 94 
 lib/librte_ethdev/rte_ethdev_driver.h| 80 +++
 lib/librte_ethdev/rte_ethdev_version.map |  3 +
 5 files changed, 225 insertions(+)

diff --git a/doc/guides/rel_notes/release_20_11.rst 
b/doc/guides/rel_notes/release_20_11.rst
index c6642f5..1f04bd5 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -78,6 +78,11 @@ New Features
 ``--portmask=N``
 where N represents the hexadecimal bitmask of ports used.
 
+* **Added the FEC API, for a generic FEC query and config.**
+
+  Added the FEC API which provides functions for query FEC capabilities and
+  current FEC mode from device. Also, API for configuring FEC mode is also 
provided.
+
 
 Removed Items
 -
diff --git a/lib/librte_ethdev/rte_ethdev.c b/lib/librte_ethdev/rte_ethdev.c
index dfe5c1b..86ead87 100644
--- a/lib/librte_ethdev/rte_ethdev.c
+++ b/lib/librte_ethdev/rte_ethdev.c
@@ -3679,6 +3679,49 @@ rte_eth_led_off(uint16_t port_id)
return eth_err(port_id, (*dev->dev_ops->dev_led_off)(dev));
 }
 
+int
+rte_eth_fec_get_capability(uint16_t port_id,
+  struct rte_eth_fec_capa *speed_fec_capa,
+  unsigned int num)
+{
+   struct rte_eth_dev *dev;
+   int ret;
+
+   RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+   dev = &rte_eth_devices[port_id];
+   RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->fec_get_capability, -ENOTSUP);
+   ret = (*dev->dev_ops->fec_get_capability)(dev, speed_fec_capa, num);
+   if (ret > (int)num)
+   RTE_ETHDEV_LOG(ERR, "Insufficient num, num should be no less 
than %d\n",
+  ret);
+
+   return ret;
+}
+
+int
+rte_eth_fec_get(uint16_t port_id, uint32_t *mode)
+{
+   struct rte_eth_dev *dev;
+
+   if (mode == NULL)
+   return -EINVAL;
+
+   RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+   dev = &rte_eth_devices[port_id];
+   RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->fec_get, -ENOTSUP);
+   return eth_err(port_id, (*dev->dev_ops->fec_get)(dev, mode));
+}
+
+int
+rte_eth_fec_set(uint16_t port_id, uint32_t mode)
+{
+   struct rte_eth_dev *dev;
+
+   dev = &rte_eth_devices[port_id];
+   RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->fec_set, -ENOTSUP);
+   return eth_err(port_id, (*dev->dev_ops->fec_set)(dev, mode));
+}
+
 /*
  * Returns index into MAC address array of addr. Use 00:00:00:00:00:00 to find
  * an empty spot.
diff --git a/lib/librte_ethdev/rte_ethdev.h b/lib/librte_ethdev/rte_ethdev.h
index 645a186..04525a8 100644
--- a/lib/librte_ethdev/rte_ethdev.h
+++ b/lib/librte_ethdev/rte_ethdev.h
@@ -1544,6 +1544,29 @@ struct rte_eth_dcb_info {
struct rte_eth_dcb_tc_queue_mapping tc_queue;
 };
 
+/**
+ * This enum indicates the possible Forward Error Correction (FEC) modes
+ * of an ethdev port.
+ */
+enum rte_eth_fec_mode {
+   RTE_ETH_FEC_NOFEC = 0,  /**< FEC is off */
+   RTE_ETH_FEC_AUTO,   /**< FEC autonegotiation modes */
+   RTE_ETH_FEC_BASER,  /**< FEC using common algorithm */
+   RTE_ETH_FEC_RS, /**< FEC using RS algorithm */
+};
+
+/* Translate from FEC mode to FEC capa */
+#define RTE_ETH_FEC_MODE_TO_CAPA(x)(1U << (x))
+
+/* This macro indicates FEC capa mask */
+#define RTE_ETH_FEC_MODE_CAPA_MASK(x)  (1U << (RTE_ETH_FEC_ ## x))
+
+/* A structure used to get capabilities per link speed */
+struct rte_eth_fec_capa {
+   uint32_t speed; /**< Link speed (see ETH_SPEED_NUM_*) */
+   uint32_t capa;  /**< FEC capabilities bitmask (see RTE_FEC_CAPA_*) */
+};
+
 #define RTE_ETH_ALL RTE_M

[dpdk-dev] [PATCH V13 3/3] app/testpmd: add FEC command

2020-09-25 Thread Min Hu (Connor)
This commit adds testpmd capability to query and config FEC
function of device. This includes:
- show FEC capabilities, example:
testpmd> show port 0 fec capabilities
- show FEC mode, example:
testpmd> show port 0 fec_mode
- config FEC mode, example:
testpmd> set port  0 

where:

auto|off|rs|baser are four kinds of FEC mode which dev
support according to MAC link speed.

Signed-off-by: Min Hu (Connor) 
Reviewed-by: Wei Hu (Xavier) 
Reviewed-by: Chengwen Feng 
Reviewed-by: Chengchang Tang 
---
v12->v13:
change fec get capa interface.

---
v10->v11:
change mode to capa bitmask.

---
v8->v9:
added acked-by.

---
v6->v7:
used RTE_DIM(fec_mode_name) instead of RTE_ETH_FEC_NUM

---
v5->v6:
fixed code styles according to DPDK coding style.
added _eth prefix.

---
v4->v5:
Add RTE_ prefix for public FEC mode enum.

---
v3->v4:
adjust the display format of FEC mode

---
v2->v3:
adjust the display format of FEC capability.

---
 app/test-pmd/cmdline.c | 223 +
 app/test-pmd/config.c  |  91 
 app/test-pmd/testpmd.h |   2 +
 3 files changed, 316 insertions(+)

diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c
index 5f93409..407513c 100644
--- a/app/test-pmd/cmdline.c
+++ b/app/test-pmd/cmdline.c
@@ -19161,6 +19161,226 @@ cmdline_parse_inst_t cmd_show_tx_metadata = {
},
 };
 
+/* *** show fec capability per port configuration *** */
+struct cmd_show_fec_capability_result {
+   cmdline_fixed_string_t cmd_show;
+   cmdline_fixed_string_t cmd_port;
+   cmdline_fixed_string_t cmd_fec;
+   cmdline_fixed_string_t cmd_keyword;
+   portid_t cmd_pid;
+};
+
+static void
+cmd_show_fec_capability_parsed(void *parsed_result,
+   __rte_unused struct cmdline *cl,
+   __rte_unused void *data)
+{
+#define FEC_CAP_NUM 2
+   struct cmd_show_fec_capability_result *res = parsed_result;
+   struct rte_eth_fec_capa speed_fec_capa[FEC_CAP_NUM];
+   unsigned int num = FEC_CAP_NUM;
+   unsigned int ret_num;
+   int ret;
+
+   if (!rte_eth_dev_is_valid_port(res->cmd_pid)) {
+   printf("Invalid port id %u\n", res->cmd_pid);
+   return;
+   }
+
+   ret = rte_eth_fec_get_capability(res->cmd_pid, speed_fec_capa, num);
+   if (ret == -ENOTSUP) {
+   printf("Function not implemented\n");
+   return;
+   } else if (ret < 0) {
+   printf("Get FEC capability failed\n");
+   return;
+   }
+
+   ret_num = (unsigned int)ret;
+   show_fec_capability(ret_num, speed_fec_capa);
+}
+
+cmdline_parse_token_string_t cmd_show_fec_capability_show =
+   TOKEN_STRING_INITIALIZER(struct cmd_show_fec_capability_result,
+   cmd_show, "show");
+cmdline_parse_token_string_t cmd_show_fec_capability_port =
+   TOKEN_STRING_INITIALIZER(struct cmd_show_fec_capability_result,
+   cmd_port, "port");
+cmdline_parse_token_num_t cmd_show_fec_capability_pid =
+   TOKEN_NUM_INITIALIZER(struct cmd_show_fec_capability_result,
+   cmd_pid, UINT16);
+cmdline_parse_token_string_t cmd_show_fec_capability_fec =
+   TOKEN_STRING_INITIALIZER(struct cmd_show_fec_capability_result,
+   cmd_fec, "fec");
+cmdline_parse_token_string_t cmd_show_fec_capability_keyword =
+   TOKEN_STRING_INITIALIZER(struct cmd_show_fec_capability_result,
+   cmd_keyword, "capabilities");
+
+cmdline_parse_inst_t cmd_show_capability = {
+   .f = cmd_show_fec_capability_parsed,
+   .data = NULL,
+   .help_str = "show port  fec capabilities",
+   .tokens = {
+   (void *)&cmd_show_fec_capability_show,
+   (void *)&cmd_show_fec_capability_port,
+   (void *)&cmd_show_fec_capability_pid,
+   (void *)&cmd_show_fec_capability_fec,
+   (void *)&cmd_show_fec_capability_keyword,
+   NULL,
+   },
+};
+
+/* *** show fec mode per port configuration *** */
+struct cmd_show_fec_metadata_result {
+   cmdline_fixed_string_t cmd_show;
+   cmdline_fixed_string_t cmd_port;
+   cmdline_fixed_string_t cmd_keyword;
+   portid_t cmd_pid;
+};
+
+static void
+cmd_show_fec_mode_parsed(void *parsed_result,
+   __rte_unused struct cmdline *cl,
+   __rte_unused void *data)
+{
+#define FEC_NAME_SIZE 16
+   struct cmd_show_fec_metadata_result *res = parsed_result;
+   uint32_t mode;
+   char buf[FEC_NAME_SIZE];
+   int ret;
+
+   if (!rte_eth_dev_is_valid_port(res->cmd_pid)) {
+   printf("Invalid port id %u\n", res->cmd_pid);
+   return;
+   }
+   ret = rte_eth_fec_get(res->cmd_pid, &mode);
+   if (ret == -ENOTSUP) {
+   printf("Function not implemented\n");
+   return;
+   } else if (ret < 0) {
+   printf("Get FEC

[dpdk-dev] [PATCH V13 2/3] net/hns3: support FEC

2020-09-25 Thread Min Hu (Connor)
Forward error correction (FEC) is a bit error correction mode.
It adds error correction information to data packets at the
transmit end, and uses the error correction information to correct
the bit errors generated during data packet transmission at the
receive end. This improves signal quality but also brings a delay
to signals. This function can be enabled or disabled as required.

This patch adds FEC support for ethdev.Introduce ethdev
operations which support query and config FEC information in
hardware.

Signed-off-by: Min Hu (Connor) 
Reviewed-by: Wei Hu (Xavier) 
Reviewed-by: Chengwen Feng 
Reviewed-by: Chengchang Tang 
---
v12->v13:
change fec get capa interface.

---
v11->v12:
fix coding warning.

v10->v11:
allow to report capabilities per link speed.
specify what should be reported if link is down
when get FEC.
change mode to capa bitmask.

---
v9->v10:
use RTE_ETH_FEC_MODE_CAPA_MASK(x) which is just defined.

---
v7->v8:
FEC mode order defined in hns3 hardware is inconsistend with
that defined in the ethdev library. So the sequence needs
to be converted.

---
v5->v6:
fixed code styles according to DPDK coding style.
added _eth prefix.

---
v4->v5:
Data type for fec_cap is changed from uint8_t
to uint32_t for possible future expansion.

---
v2->v3:
adjust the return value of function.

---
 doc/guides/rel_notes/release_20_11.rst |   5 +
 drivers/net/hns3/hns3_cmd.h|  19 +-
 drivers/net/hns3/hns3_ethdev.c | 356 +
 drivers/net/hns3/hns3_ethdev.h |   1 +
 4 files changed, 380 insertions(+), 1 deletion(-)

diff --git a/doc/guides/rel_notes/release_20_11.rst 
b/doc/guides/rel_notes/release_20_11.rst
index 1f04bd5..eec6930 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -83,6 +83,11 @@ New Features
   Added the FEC API which provides functions for query FEC capabilities and
   current FEC mode from device. Also, API for configuring FEC mode is also 
provided.
 
+* **Added hns3 FEC PMD, for supporting query and config FEC mode.**
+
+  Added the FEC PMD which provides functions for query FEC capabilities and
+  current FEC mode from device. Also, PMD for configuring FEC mode is also 
provided.
+
 
 Removed Items
 -
diff --git a/drivers/net/hns3/hns3_cmd.h b/drivers/net/hns3/hns3_cmd.h
index 87d6053..a1e9604 100644
--- a/drivers/net/hns3/hns3_cmd.h
+++ b/drivers/net/hns3/hns3_cmd.h
@@ -100,6 +100,7 @@ enum hns3_opcode_type {
HNS3_OPC_QUERY_LINK_STATUS  = 0x0307,
HNS3_OPC_CONFIG_MAX_FRM_SIZE= 0x0308,
HNS3_OPC_CONFIG_SPEED_DUP   = 0x0309,
+   HNS3_OPC_CONFIG_FEC_MODE= 0x031A,
 
/* PFC/Pause commands */
HNS3_OPC_CFG_MAC_PAUSE_EN   = 0x0701,
@@ -684,9 +685,25 @@ struct hns3_config_auto_neg_cmd {
uint8_t   rsv[20];
 };
 
+#define HNS3_MAC_CFG_FEC_AUTO_EN_B 0
+#define HNS3_MAC_CFG_FEC_MODE_S1
+#define HNS3_MAC_CFG_FEC_MODE_MGENMASK(3, 1)
+#define HNS3_MAC_FEC_OFF   0
+#define HNS3_MAC_FEC_BASER 1
+#define HNS3_MAC_FEC_RS2
+
 struct hns3_sfp_speed_cmd {
uint32_t  sfp_speed;
-   uint32_t  rsv[5];
+   uint8_t   query_type; /* 0: sfp speed, 1: active fec */
+   uint8_t   active_fec; /* current FEC mode */
+   uint16_t  rsv1;
+   uint32_t  rsv2[4];
+};
+
+/* Configure FEC mode, opcode:0x031A */
+struct hns3_config_fec_cmd {
+   uint8_t fec_mode;
+   uint8_t rsv[23];
 };
 
 #define HNS3_MAC_MGR_MASK_VLAN_B   BIT(0)
diff --git a/drivers/net/hns3/hns3_ethdev.c b/drivers/net/hns3/hns3_ethdev.c
index 73d5042..d3650bc 100644
--- a/drivers/net/hns3/hns3_ethdev.c
+++ b/drivers/net/hns3/hns3_ethdev.c
@@ -63,6 +63,11 @@
 #define HNS3_RESET_WAIT_MS 100
 #define HNS3_RESET_WAIT_CNT200
 
+/* FEC mode order defined in HNS3 hardware */
+#define HNS3_HW_FEC_MODE_NOFEC  0
+#define HNS3_HW_FEC_MODE_BASER  1
+#define HNS3_HW_FEC_MODE_RS 2
+
 enum hns3_evt_cause {
HNS3_VECTOR0_EVENT_RST,
HNS3_VECTOR0_EVENT_MBX,
@@ -70,6 +75,34 @@ enum hns3_evt_cause {
HNS3_VECTOR0_EVENT_OTHER,
 };
 
+static const struct rte_eth_fec_capa speed_fec_capa_tbl[] = {
+   { ETH_SPEED_NUM_10G, RTE_ETH_FEC_MODE_CAPA_MASK(NOFEC) |
+RTE_ETH_FEC_MODE_CAPA_MASK(AUTO) |
+RTE_ETH_FEC_MODE_CAPA_MASK(BASER) },
+
+   { ETH_SPEED_NUM_25G, RTE_ETH_FEC_MODE_CAPA_MASK(NOFEC) |
+RTE_ETH_FEC_MODE_CAPA_MASK(AUTO) |
+RTE_ETH_FEC_MODE_CAPA_MASK(BASER) |
+RTE_ETH_FEC_MODE_CAPA_MASK(RS) },
+
+   { ETH_SPEED_NUM_40G, RTE_ETH_FEC_MODE_CAPA_MASK(NOFEC) |
+RTE_ETH_FEC_MODE_CAPA_MASK(AUTO) |
+RTE_ETH_FEC_MODE_CAPA_MASK(BASER) },
+
+   { ETH_SPEED_NUM_50G, RTE_ETH_FEC_MODE_CAPA_MASK(NOFEC) |
+RTE_E

Re: [dpdk-dev] [PATCH 5/8] net/bnxt: add a null ptr check in bnxt PCI probe

2020-09-25 Thread Ferruh Yigit

On 9/25/2020 3:04 AM, Somnath Kotur wrote:

On Thu, Sep 24, 2020 at 8:17 PM Ferruh Yigit  wrote:


On 9/22/2020 8:06 AM, Somnath Kotur wrote:

Check for devargs before invoking rep port probe.

Fixes: 6dc83230b43b ("net/bnxt: support port representor data path")

Signed-off-by: Somnath Kotur 
Reviewed-by: Venkat Duvvuru 
---
   drivers/net/bnxt/bnxt_ethdev.c | 4 
   1 file changed, 4 insertions(+)

diff --git a/drivers/net/bnxt/bnxt_ethdev.c b/drivers/net/bnxt/bnxt_ethdev.c
index db2f0dd..84eba0b 100644
--- a/drivers/net/bnxt/bnxt_ethdev.c
+++ b/drivers/net/bnxt/bnxt_ethdev.c
@@ -6147,6 +6147,10 @@ static int bnxt_pci_probe(struct rte_pci_driver *pci_drv 
__rte_unused,
   }
   PMD_DRV_LOG(DEBUG, "BNXT Port:%d pci probe\n",
   backing_eth_dev->data->port_id);
+
+ if (!pci_dev->device.devargs)
+ return ret;
+


There is already a null check at the beginning of the function because
of the same thing (port representors), should they be combined?


No, this is to catch the corner case if/when 'backing_eth_dev' is
already allocated , so code would unconditionally call
bnxt_rep_port_probe()
irrespective of devargs being there or not, the check at this point
helps prevent that

And devargs being not NULL does not really mean it has arguments related
to the port representors, it may have other device devargs. Perhaps
'eth_da' can  be used to check?

eth_da is a local var in this function, so perhaps 'num_rep' i.e
invoke bnxt_rep_port_probe only if num_rep > 0 ?


+1


Please let me know if you want me to do a respin of this patch alone
or will you be doing this minor change while merging it in?


Please send a new version of this patch alone. Thanks.


Re: [dpdk-dev] [EXT] Re: [RFC PATCH] app/testpmd: tx pkt clones parameter in flowgen

2020-09-25 Thread Igor Russkikh


>> +} else {
>> +nb_clones--;
>> +}
>> +pkts_burst[nb_pkt] = pkt;
>>  
>>  next_flow = (next_flow + 1) % cfg_n_flows;
>>  }
> 
> 
> This doesn't look safe. You can't just send same mbuf N times without
> incrementing
> the reference count.

Hi Stephen,

You are right, of course we need refcnt increment here.

It looks like it worked for me, because occasionally its possible to execute
rte_pktmbuf_free multiple times over mbuf with refcnt==1 without any immediate
side effects.

Will repost.

Thanks,
  Igor


Re: [dpdk-dev] [PATCH V12 1/4] ethdev: introduce FEC API

2020-09-25 Thread Min Hu (Connor)

HI,Andrew,
I fix it in V13 according to your advice.
Thanks for  your patient review. Please check it out again.

By the way, there is always a coding check warning, like this:

WARNING:PREFER_FALLTHROUGH: Prefer 'fallthrough;' over fallthrough comment
#264: FILE: drivers/net/hns3/hns3_ethdev.c:5601:
+   /* fallthrough */

total: 0 errors, 1 warnings, 0 checks, 439 lines checked.

I have tried some ways, but it does not help.
Could you giver some advice?
thanks.


在 2020/9/24 22:46, Andrew Rybchenko 写道:

On 9/24/20 4:05 PM, Min Hu (Connor) wrote:

This patch adds Forward error correction(FEC) support for ethdev.
Introduce APIs which support query and config FEC information in
hardware.


Almost good now. See my notes below.
Many thanks for hard work and patience.



Signed-off-by: Min Hu (Connor) 
Reviewed-by: Wei Hu (Xavier) 
Reviewed-by: Chengwen Feng 
Reviewed-by: Chengchang Tang 
Reviewed-by: Ajit Khaparde 
Acked-by: Konstantin Ananyev 


I think that tags for Ajit and Konstantin should be
dropped, since API changes significantly after their review.


---
v10->v11:
allow to report capabilities per link speed.
specify what should be reported if link is down
when get FEC.
change mode to capa bitmask.

---
v9->v10:
add macro RTE_ETH_FEC_MODE_CAPA_MASK(x) to indicate
different FEC mode capa.

---
v8->v9:
added reviewed-by and acked-by.

---
v7->v8:
put AUTO just after NOFEC in rte_fec_mode definition.

---
v6->v7:
deleted RTE_ETH_FEC_NUM to prevent ABI breakage.
add new macro to indicate translation from fec mode
to capa.

---
v5->v6:
modified release notes.
deleted check duplicated for FEC API
fixed code styles according to DPDK coding style.
added _eth prefix.

---
v4->v5:
Modifies FEC capa definitions using macros.
Add RTE_ prefix for public FEC mode enum.
add release notes about FEC for dpdk20_11.

---
v2->v3:
add function return value "-ENOTSUP" for API.

---
  lib/librte_ethdev/rte_ethdev.c   | 37 +
  lib/librte_ethdev/rte_ethdev.h   | 91 
  lib/librte_ethdev/rte_ethdev_driver.h| 82 
  lib/librte_ethdev/rte_ethdev_version.map |  3 ++
  4 files changed, 213 insertions(+)

diff --git a/lib/librte_ethdev/rte_ethdev.c b/lib/librte_ethdev/rte_ethdev.c
index dfe5c1b..b614bfc 100644
--- a/lib/librte_ethdev/rte_ethdev.c
+++ b/lib/librte_ethdev/rte_ethdev.c
@@ -3679,6 +3679,43 @@ rte_eth_led_off(uint16_t port_id)
return eth_err(port_id, (*dev->dev_ops->dev_led_off)(dev));
  }
  
+int

+rte_eth_fec_get_capability(uint16_t port_id, uint32_t *num,
+  struct rte_eth_fec_capa *speed_fec_capa)
+{
+   struct rte_eth_dev *dev;
+
+   if (num == NULL || speed_fec_capa == NULL)
+   return -EINVAL;


I think it is OK to have speed_fec_cap==NULL if *num is 0.
I.e. a request to get number of required array entries.


+
+   RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+   dev = &rte_eth_devices[port_id];
+   RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->fec_get_capability, -ENOTSUP);
+   return eth_err(port_id, (*dev->dev_ops->fec_get_capability)(dev, num,
+   speed_fec_capa));
+}
+
+int
+rte_eth_fec_get(uint16_t port_id, uint32_t *mode)
+{
+   struct rte_eth_dev *dev;


I think it would be good to check that mode is not NULL here.


+
+   RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
+   dev = &rte_eth_devices[port_id];
+   RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->fec_get, -ENOTSUP);
+   return eth_err(port_id, (*dev->dev_ops->fec_get)(dev, mode));
+}
+
+int
+rte_eth_fec_set(uint16_t port_id, uint32_t mode)
+{
+   struct rte_eth_dev *dev;
+
+   dev = &rte_eth_devices[port_id];
+   RTE_FUNC_PTR_OR_ERR_RET(*dev->dev_ops->fec_set, -ENOTSUP);
+   return eth_err(port_id, (*dev->dev_ops->fec_set)(dev, mode));
+}
+
  /*
   * Returns index into MAC address array of addr. Use 00:00:00:00:00:00 to find
   * an empty spot.
diff --git a/lib/librte_ethdev/rte_ethdev.h b/lib/librte_ethdev/rte_ethdev.h
index 645a186..104181d 100644
--- a/lib/librte_ethdev/rte_ethdev.h
+++ b/lib/librte_ethdev/rte_ethdev.h
@@ -1544,6 +1544,29 @@ struct rte_eth_dcb_info {
struct rte_eth_dcb_tc_queue_mapping tc_queue;
  };
  
+/**

+ * This enum indicates the possible (forward error correction)FEC modes


(forward error correction)FEC -> Forward Error Correction (FEC)


+ * of an ethdev port.
+ */
+enum rte_eth_fec_mode {
+   RTE_ETH_FEC_NOFEC = 0,  /**< FEC is off */
+   RTE_ETH_FEC_AUTO,   /**< FEC autonegotiation modes */
+   RTE_ETH_FEC_BASER,  /**< FEC using common algorithm */
+   RTE_ETH_FEC_RS, /**< FEC using RS algorithm */
+};
+
+/* Translate from FEC mode to FEC capa */
+#define RTE_ETH_FEC_MODE_TO_CAPA(x)(1U << (x))
+
+/* This macro indicates FEC capa mask*/


Add missing space before */


+#define RTE_ETH_FEC_MOD

Re: [dpdk-dev] [PATCH 10/20] net/null: release port upon close

2020-09-25 Thread Ferruh Yigit

On 9/24/2020 10:58 PM, Thomas Monjalon wrote:

23/09/2020 22:47, Thomas Monjalon:

23/09/2020 18:44, Ferruh Yigit:

On 9/13/2020 11:07 PM, Thomas Monjalon wrote:

The flag RTE_ETH_DEV_CLOSE_REMOVE is set so all port resources
can be freed by rte_eth_dev_close().

Signed-off-by: Thomas Monjalon 
---
--- a/drivers/net/null/rte_eth_null.c
+++ b/drivers/net/null/rte_eth_null.c
+static int
+eth_dev_close(struct rte_eth_dev *dev)
+{
+   PMD_LOG(INFO, "Closing null ethdev on NUMA socket %u",
+   rte_socket_id());
+
+   /* mac_addrs must not be freed alone because part of dev_private */
+   dev->data->mac_addrs = NULL;
+
+   return 0;
+}


should check 'RTE_PROC_PRIMARY' in 'eth_dev_close()'?


Yes, looks to be a miss for this new function.


Sorry no, this function is not freeing any shared data,
so no restriction on secondary process.



"dev->data->mac_addrs = NULL", won't this change the shared data?



Re: [dpdk-dev] Meson Minimum Version

2020-09-25 Thread Morten Brørup
> From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Bruce Richardson
> Sent: Friday, September 25, 2020 10:41 AM
> To: Morten Brørup
> 
> On Fri, Sep 25, 2020 at 09:31:53AM +0200, Morten Brørup wrote:
> > > From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Bruce
> Richardson
> > > Sent: Thursday, September 24, 2020 5:43 PM
> > >
> > > On Thu, Sep 24, 2020 at 03:32:41PM +, John Alexander wrote:
> > > >Hi,
> > > >Regarding the subproject local patch support, yes, it's only
> > > supported
> > > >since Meson 0.55:
> > > >https://mesonbuild.com/Wrap-dependency-system-manual.html We
> pip
> > > >installed 0.55 Meson.
> > > >I have a number of subsequent patches that depend on this
> > > particular
> > > >pthreads library to advance the Windows DPDK support. Locally,
> we
> > > have
> > > >testpmd (minus mmap'd external memory currently) running
> against
> > > the
> > > >Intel i40e PMD (XL710 4x10Gbps SPF+ NIC) on Windows on our
> local
> > > DPDK
> > > >fork (based off 20.08-rc2 using Microsoft's latest NetUIO
> driver).
> > > We
> > > >have 47 of the 51 RTE libraries building and have had l2fwd,
> > > l3fwd,
> > > >ipv4_multicast and almost all of the regression tests
> > > compiling+linking
> > > >too. I'd like to push as much of the Windows EAL work we've
> done
> > > >upstream if I can (after a bit of tidying up :).
> > > >I've also coded up a meson build patch for the Jansson JSON
> parser
> > > used
> > > >by the RTE metrics library (the config.h generation was quite
> > > fiddly!)
> > > >That's ready to go. We get nice meson syntax as follows to
> specify
> > > a
> > > >fallback if the library isn't installed locally:
> > > >jansson = dependency('jansson', required: false, fallback :
> > > ['jansson',
> > > >'jansson_static_dep'])
> > > >I believe the meson command line enables disabling fallbacks
> if
> > > people
> > > >would prefer not to use them (--wrap-mode=nofallback).
> > > >Kind regards,
> > > >John.
> > >
> > > Hi again, John,
> > >
> > > thanks for the full reply - the work you have sounds really good,
> and
> > > the
> > > possibilities using the wrap support are definitely of interest.
> > > [Though
> > > since jansson is only used for the legacy parts of the telemetry
> > > library
> > > I'm not sure we want to wrap it just for that - the later telemetry
> > > work
> > > from this year doesn't depend on jansson. Then again, the
> > > vm_power_manager
> > > example also uses it too...]. It's probably something that would be
> > > especially useful for software dependencies that aren't normally
> > > packaged.
> > >
> > > I've added techboard on CC to previous reply, so hopefully we'll
> get
> > > some
> > > thoughts from others.
> > >
> >
> > A buggy C compiler might generate buggy code, and the compiler
> induced bugs may not be caught during development (but hopefully during
> testing). I have seen this happen, and I still consider it a realistic
> scenario. This is the strongest argument for using a trusted version of
> the C compiler, which is supported by the popular GNU/Linux
> distributors (Red Hat etc.), or sufficiently vetted to be included in
> the Crosstool-NG project, or similar.
> >
> > Meson is a relatively simple tool, so I personally wouldn't mind
> overwriting the distro-supported version on our build systems with a
> new version. (I haven't worked much with Meson, but assume that the new
> version of Meson is backwards compatible with its earlier versions.)
> >
> > In short, my primary concern is: What could realistically go wrong if
> the required version of Meson is buggy?
> >
> > Bruce, you have worked for quite a while with Meson/Ninja by now, so
> perhaps you can assess this risk based on your experience.
> >
> I'd say the risk in this case is small, especially since I see that
> 0.56 of
> meson is well under way for development and may well be released before
> DPDK 20.11. Generally backwards compatibilty of meson is excellent as
> they
> have comprehensive test suite for all features.
> 

Thank you for the qualified risk assessment, Bruce.

> Rather than any bugginess, my concern was purely requiring people to
> update
> meson using "pip3", but I suppose that's not really a big deal, and
> when
> using pip update it defaults to just updating the copy for the local
> user,
> not system-wide.
> 
> Updating to a later meson will give us access to a few other nice
> features
> we can use also (from earlier releases than 0.55), such as support for
> "break" and "continue" keywords in loops, which can simplify out lib or
> driver building code, perhaps, or the "console" parameter for custom
> targets which should improve the output visibility when builing our
> docs.
> 
> /Bruce

I vote for requiring the later Meson version.

-Morten



Re: [dpdk-dev] [PATCH V2 1/4] ethdev: fix compiling errors for per-queue statistics

2020-09-25 Thread Min Hu (Connor)

Hi, Ferruh Yigit,

Your method is very practical. Thanks!
But we found that testpmd fails to start with "--rx-queue-stats-mapping" 
and "--tx-queue-stats-mapping". We've found the cause and modified it. 
These modifications are being tested.


Currently, the following modifications are related to framework APIs.
1)struct 'rte_eth_dcb_tc_queue_mapping'
2)fix 'rte_eth_dev_set_rx/tx_queue_stats_mapping' function to resolve
the problem that DPDK project fails to build when 
'RTE_ETHDEV_QUEUE_STAT_CNTRS' > 256.


According to my current analysis, these modifications for queue stats 
mapping in testpmd do not involve the modification of framework APIs. 
Therefore, I think we should give priority to focusing on the current 
changes to the DPDK framework APIs.


After all, DPDK-20.11 is approaching. What do you think?


在 2020/9/23 17:18, Ferruh Yigit 写道:

On 9/23/2020 3:31 AM, Min Hu (Connor) wrote:



在 2020/9/5 2:31, Ferruh Yigit 写道:

On 9/4/2020 12:32 PM, Min Hu (Connor) wrote:

From: Huisong Li 

Currently, only statistics of rx/tx queues with queue_id less than
RTE_ETHDEV_QUEUE_STAT_CNTRS can be displayed. If there is a certain
application scenario that it needs to use 256 or more than 256 queues
and display all statistics of rx/tx queue. At this moment, we have to
change the macro to be equaled to the queue number.

However, modifying the macro to be greater than 256 will trigger
many errors and warnings from test-pmd, PMD driver and librte_ethdev
during compiling dpdk project. But it is possible and permited that
rx/tx queue number is greater than 256 and all statistics of rx/tx
queue need to be displayed. In addition, the data type of rx/tx queue
number in rte_eth_dev_configure API is 'uint16_t'. So It is 
unreasonable

to use the 'uint8_t' type for variables that control which per-queue
statistics can be displayed.

Fixes: ed30d9b691b2 ("app/testpmd: add stats per queue")
Fixes: 09c7e63a71f9 ("net/memif: introduce memory interface PMD")
Fixes: abf7275bbaa2 ("ixgbe: move to drivers/net/")
Fixes: e6defdfddc3b ("net/igc: enable statistics")
Fixes: 2265e4b4e84b ("net/octeontx2: add basic stats operation")
Fixes: 6c3169a3dc04 ("virtio: move to drivers/net/")
Cc: sta...@dpdk.org

Signed-off-by: Huisong Li 
Signed-off-by: Min Hu (Connor) 
Reviewed-by: Wei Hu (Xavier) 
Reviewed-by: Dongdong Liu 


The patch mostly looks good and it enables build with
'RTE_ETHDEV_QUEUE_STAT_CNTRS' > 256.
Only I put a comment for a testpmd change to enable the "set 
stat_qmap" command

for map value > 256.


BUT there are many things to fix in the queue stats mapping, since 
you are
already on it can you help on a few things on testpmd related to it, 
if you have

time for it?

1) Getting queue stats shouldn't require stats mapping, it should be 
controlled
separately. Many PMDs doesn't require/do the stats mapping but they 
still can
collect the per queue stats, which can be displayed independent from 
mapping.


2) Even you map only one queue, while displaying stats it will display
'RTE_ETHDEV_QUEUE_STAT_CNTRS' queues, and when that number is high it 
makes hard

to see the actual interested values.
If there is no mapping, it should display min(number_of_queues,
RTE_ETHDEV_QUEUE_STAT_CNTRS).
If there is mapping it should display queues that mapping done, this 
may require

adding a new 'active' field to 'struct queue_stats_mappings'.

3) Why 'struct queue_stats_mappings' is cache aligned, is it really 
needed?


4) The mapping arrays, 'tx_queue_stats_mappings_array' &
'rx_queue_stats_mappings_array' are global and their size is based on 
fixed max
port and queue size assumptions, can those mapping array be done per 
port and

'RTE_ETHDEV_QUEUE_STAT_CNTRS' size per port?


It does seem to be unreasonable. We also try do it, and found that
it is hard to control per queue stats and queue stats mapping, 
separately.

For details, see next V3 patch.


I will check v3, but as far as know stats mapping exist because of 
limitations of some NICs, that there are N queues but M stats registers 
where N > M. So need to map some queues to stat registers, and there is 
N:1 relation there, so multiple queue stats can be represented in single 
stat register.


But for rest of the NICs it should be possible to display the queue 
stats independent from the mapping.



We're working on debugging this stats_mapping setting. During this 
process,

we have a problem that starting testpmd with --rx/tx-queue-stats-mapping
parameter fails. log as follows:
[root]$ ./testpmd -l 1,3,4,8,12 -n 4 -w :04:00.0 --file-prefix=lee 
--log-level=7

  -- -i --rxq=6 --txq=6 --burst=64 --rxd=2048 --txd=2048 --nb-cores=4
  --rx-queue-stats-mapping=(0,2,2) --tx-queue-stats-mapping=(0,3,3)

-bash: syntax error near unexpected token `('
[root]$


Above seems bash error, using '' can help for it, like 
--rx-queue-stats-mapping='(0,2,2)'

.


[dpdk-dev] [RFC PATCH v2] app/testpmd: tx pkt clones parameter in flowgen

2020-09-25 Thread Igor Russkikh
When testing high performance numbers, it is often that CPU performance
limits the max values device can reach (both in pps and in gbps)

Here instead of recreating each packet separately, we use clones counter
to resend the same mbuf to the line multiple times.

PMDs handle that transparently due to reference counting inside of mbuf.

Verified on Marvell qede and atlantic PMDs.

v2: increment ref counter for each mbuf pointer copy

Signed-off-by: Igor Russkikh 
---
 app/test-pmd/flowgen.c| 101 ++
 app/test-pmd/parameters.c |  12 +++
 app/test-pmd/testpmd.c|   1 +
 app/test-pmd/testpmd.h|   1 +
 doc/guides/testpmd_app_ug/run_app.rst |   7 ++
 5 files changed, 75 insertions(+), 47 deletions(-)

diff --git a/app/test-pmd/flowgen.c b/app/test-pmd/flowgen.c
index acf3e2460..f639155c7 100644
--- a/app/test-pmd/flowgen.c
+++ b/app/test-pmd/flowgen.c
@@ -94,6 +94,7 @@ pkt_burst_flow_gen(struct fwd_stream *fs)
uint16_t nb_rx;
uint16_t nb_tx;
uint16_t nb_pkt;
+   uint16_t nb_clones = nb_pkt_clones;
uint16_t i;
uint32_t retry;
uint64_t tx_offloads;
@@ -123,53 +124,59 @@ pkt_burst_flow_gen(struct fwd_stream *fs)
ol_flags |= PKT_TX_MACSEC;
 
for (nb_pkt = 0; nb_pkt < nb_pkt_per_burst; nb_pkt++) {
-   pkt = rte_mbuf_raw_alloc(mbp);
-   if (!pkt)
-   break;
-
-   pkt->data_len = pkt_size;
-   pkt->next = NULL;
-
-   /* Initialize Ethernet header. */
-   eth_hdr = rte_pktmbuf_mtod(pkt, struct rte_ether_hdr *);
-   rte_ether_addr_copy(&cfg_ether_dst, ð_hdr->d_addr);
-   rte_ether_addr_copy(&cfg_ether_src, ð_hdr->s_addr);
-   eth_hdr->ether_type = rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4);
-
-   /* Initialize IP header. */
-   ip_hdr = (struct rte_ipv4_hdr *)(eth_hdr + 1);
-   memset(ip_hdr, 0, sizeof(*ip_hdr));
-   ip_hdr->version_ihl = RTE_IPV4_VHL_DEF;
-   ip_hdr->type_of_service = 0;
-   ip_hdr->fragment_offset = 0;
-   ip_hdr->time_to_live= IP_DEFTTL;
-   ip_hdr->next_proto_id   = IPPROTO_UDP;
-   ip_hdr->packet_id   = 0;
-   ip_hdr->src_addr= rte_cpu_to_be_32(cfg_ip_src);
-   ip_hdr->dst_addr= rte_cpu_to_be_32(cfg_ip_dst +
-  next_flow);
-   ip_hdr->total_length= RTE_CPU_TO_BE_16(pkt_size -
-  sizeof(*eth_hdr));
-   ip_hdr->hdr_checksum= ip_sum((unaligned_uint16_t *)ip_hdr,
-sizeof(*ip_hdr));
-
-   /* Initialize UDP header. */
-   udp_hdr = (struct rte_udp_hdr *)(ip_hdr + 1);
-   udp_hdr->src_port   = rte_cpu_to_be_16(cfg_udp_src);
-   udp_hdr->dst_port   = rte_cpu_to_be_16(cfg_udp_dst);
-   udp_hdr->dgram_cksum= 0; /* No UDP checksum. */
-   udp_hdr->dgram_len  = RTE_CPU_TO_BE_16(pkt_size -
-  sizeof(*eth_hdr) -
-  sizeof(*ip_hdr));
-   pkt->nb_segs= 1;
-   pkt->pkt_len= pkt_size;
-   pkt->ol_flags   &= EXT_ATTACHED_MBUF;
-   pkt->ol_flags   |= ol_flags;
-   pkt->vlan_tci   = vlan_tci;
-   pkt->vlan_tci_outer = vlan_tci_outer;
-   pkt->l2_len = sizeof(struct rte_ether_hdr);
-   pkt->l3_len = sizeof(struct rte_ipv4_hdr);
-   pkts_burst[nb_pkt]  = pkt;
+   if (!nb_pkt || !nb_clones) {
+   nb_clones = nb_pkt_clones;
+   pkt = rte_mbuf_raw_alloc(mbp);
+   if (!pkt)
+   break;
+
+   pkt->data_len = pkt_size;
+   pkt->next = NULL;
+
+   /* Initialize Ethernet header. */
+   eth_hdr = rte_pktmbuf_mtod(pkt, struct rte_ether_hdr *);
+   rte_ether_addr_copy(&cfg_ether_dst, ð_hdr->d_addr);
+   rte_ether_addr_copy(&cfg_ether_src, ð_hdr->s_addr);
+   eth_hdr->ether_type = 
rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4);
+
+   /* Initialize IP header. */
+   ip_hdr = (struct rte_ipv4_hdr *)(eth_hdr + 1);
+   memset(ip_hdr, 0, sizeof(*ip_hdr));
+   ip_hdr->version_ihl = RTE_IPV4_VHL_DEF;
+   ip_hdr->type_of_service = 0;
+   ip_hdr->fragment_offset = 0;
+   ip

Re: [dpdk-dev] [PATCH] build: support i686 target on x86 64-bit hosts

2020-09-25 Thread Bruce Richardson
On Thu, Sep 24, 2020 at 12:37:42PM -0400, Lance Richardson wrote:
> Add meson cross files for building i686 targets using gcc on x86_64
> linux hosts.
> 
> Uusage example:
> 
> meson --cross-file config/x86/cross-i686-linux-gcc build-i686
> ninja -C build-i686
> 
> Signed-off-by: Lance Richardson 
> ---
> Red Hat distros use a different name for the 32-bit pkg-config
> command from most other distros, maybe there is a better way
> to handle this than using separate files.
> 
> Others will probably have better naming suggestions for the files.
> 
Just to note that rather than using cross-files, building i686 on x86_64
should be possible by just setting environment variables.

For example, on Ubuntu 20.04, this works for me:

PKG_CONFIG_LIBDIR=/usr/lib/i386-linux-gnu/pkgconfig CFLAGS='-m32' 
LDFLAGS='-m32' meson --werror build-32bit
ninja -C build-32bit/

For Fedora or Redhat system the PKG_CONFIG_LIBDIR will be different (I
think just /usr/lib/pkgconfig), but the rest should work identically.

/Bruce


Re: [dpdk-dev] [PATCH v1 1/3] lib/mirror: introduce traffic mirror API

2020-09-25 Thread MEHAN, MUNISH
Acked-by: MEHAN, MUNISH 


On 9/8/20, 8:28 PM, "Patrick Fu"  wrote:

Network Test Access Point (TAP) is the network monitoring service
commonly adotpted in SDN-based network infrastructures. When VMs are
inter-connected over virtual switches, TAP requires vSwitch to mirror
out network traffics from specific workload VM ports to the TAP
device/VM ports.

This patch introduce 2 a new APIs to support high-throughput packet
mirroring:
 - rte_mirror_register()
 - rte_mirror_unregister()
Applications use the new API to setup the source traffic port and
the mirror traffic port. Packets flowing over the registered
source ports will be transmited to the mirror port when registration
succeeds.

Signed-off-by: Liang-min Wang 
Signed-off-by: Patrick Fu 
Signed-off-by: Timothy Miskell 
---
 config/common_base   |   5 +
 lib/Makefile |   2 +
 lib/librte_mirror/Makefile   |  20 ++
 lib/librte_mirror/meson.build|   6 +
 lib/librte_mirror/rte_mirror.c   | 297 +++
 lib/librte_mirror/rte_mirror.h   | 111 +
 lib/librte_mirror/rte_mirror_version.map |   7 +
 lib/meson.build  |   2 +-
 8 files changed, 449 insertions(+), 1 deletion(-)
 create mode 100644 lib/librte_mirror/Makefile
 create mode 100644 lib/librte_mirror/meson.build
 create mode 100644 lib/librte_mirror/rte_mirror.c
 create mode 100644 lib/librte_mirror/rte_mirror.h
 create mode 100644 lib/librte_mirror/rte_mirror_version.map

diff --git a/config/common_base b/config/common_base
index fbf0ee70c..60217cc93 100644
--- a/config/common_base
+++ b/config/common_base
@@ -1110,6 +1110,11 @@ CONFIG_RTE_LIBRTE_GRAPH_STATS=y
 #
 CONFIG_RTE_LIBRTE_NODE=y
 
+#
+# Compile librte_mirror
+#
+CONFIG_RTE_LIBRTE_MIRROR=y
+
 #
 # Compile the test application
 #
diff --git a/lib/Makefile b/lib/Makefile
index 8f5b68a2d..24175dd98 100644
--- a/lib/Makefile
+++ b/lib/Makefile
@@ -112,6 +112,8 @@ DIRS-$(CONFIG_RTE_LIBRTE_REORDER) += librte_reorder
 DEPDIRS-librte_reorder := librte_eal librte_mempool librte_mbuf
 DIRS-$(CONFIG_RTE_LIBRTE_PDUMP) += librte_pdump
 DEPDIRS-librte_pdump := librte_eal librte_mempool librte_mbuf librte_ethdev
+DIRS-$(CONFIG_RTE_LIBRTE_MIRROR) += librte_mirror
+DEPDIRS-librte_mirror := librte_eal librte_mempool librte_mbuf 
librte_ethdev
 DIRS-$(CONFIG_RTE_LIBRTE_GSO) += librte_gso
 DEPDIRS-librte_gso := librte_eal librte_mbuf librte_ethdev librte_net
 DEPDIRS-librte_gso += librte_mempool
diff --git a/lib/librte_mirror/Makefile b/lib/librte_mirror/Makefile
new file mode 100644
index 0..3ca1db734
--- /dev/null
+++ b/lib/librte_mirror/Makefile
@@ -0,0 +1,20 @@
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(c) 2020- Intel Corporation
+
+include $(RTE_SDK)/mk/rte.vars.mk
+
+# library name
+LIB = librte_mirror.a
+
+CFLAGS += $(WERROR_FLAGS) -I$(SRCDIR) -O3
+LDLIBS += -lrte_eal -lrte_mbuf -lrte_mempool -lrte_ethdev
+
+EXPORT_MAP := rte_mirror_version.map
+
+# all source are stored in SRCS-y
+SRCS-$(CONFIG_RTE_LIBRTE_MIRROR) := rte_mirror.c
+
+# install this header file
+SYMLINK-$(CONFIG_RTE_LIBRTE_MIRROR)-include := rte_mirror.h
+
+include $(RTE_SDK)/mk/rte.lib.mk
diff --git a/lib/librte_mirror/meson.build b/lib/librte_mirror/meson.build
new file mode 100644
index 0..2d0ba0ed2
--- /dev/null
+++ b/lib/librte_mirror/meson.build
@@ -0,0 +1,6 @@
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(c) 2020 Intel Corporation
+
+sources = files('rte_mirror.c')
+headers = files('rte_mirror.h')
+deps += ['ethdev']
diff --git a/lib/librte_mirror/rte_mirror.c b/lib/librte_mirror/rte_mirror.c
new file mode 100644
index 0..f2a85976a
--- /dev/null
+++ b/lib/librte_mirror/rte_mirror.c
@@ -0,0 +1,297 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2020- Intel Corporation
+ */
+
+#include 
+#include 
+#include 
+#include 
+#include 
+
+#include "rte_mirror.h"
+
+/* Macro for printing using RTE_LOG */
+extern int mirror_logtype;
+#define MIRROR_LOG(level, fmt, args...)\
+   rte_log(RTE_LOG_ ## level, mirror_logtype, "%s(): " fmt,\
+   __func__, ## args)
+
+#define MAC_ADDR_MAP   0xULL
+#define VLAN_INSERT_FLAG   (PKT_TX_VLAN_PKT)
+#define is_mac_addr_match(a, b) (((a^b)&MAC_ADDR_MAP) == 0)
+#define INIT_MIRROR_DB_SIZE8
+#define INVALID_PORT_ID0x
+#define SRC_MAC_OFFSET 6
+#define DST_MAC_OFF

Re: [dpdk-dev] [PATCH] app/test-flow-perf: configurable rule batches

2020-09-25 Thread Wisam Monther
Hi Georgios,

>-Original Message-
>From: george@gmail.com  On Behalf Of Georgios
>Katsikas
>Sent: Tuesday, September 8, 2020 9:13 AM
>To: wis...@mellanox.com
>Cc: dev@dpdk.org; Georgios Katsikas 
>Subject: [PATCH] app/test-flow-perf: configurable rule batches

Please use 'app/flow-perf' as a prefix.

BRs,
Wisam Jaddo


Re: [dpdk-dev] [PATCH] app/flow-perf: configurable rule batches

2020-09-25 Thread Wisam Monther
Hi,

>-Original Message-
>From: george@gmail.com  On Behalf Of Georgios
>Katsikas
>Sent: Thursday, September 24, 2020 12:11 PM
>To: wis...@mellanox.com
>Cc: dev@dpdk.org; Georgios Katsikas 
>Subject: [PATCH] app/flow-perf: configurable rule batches
>
>* One can now configure the number of rules per batch
>* Refactored flow_count variable to rules_count as it is related to the newly
>added rules_batch variable
>* Added default values to usage function
>
>Signed-off-by: Georgios Katsikas 

Acked-by: Wisam Jaddo 

BRs,
Wisam Jaddo


Re: [dpdk-dev] [PATCH v3 0/4] Remove RTE_MACHINE_CPUFLAG_ macros

2020-09-25 Thread David Marchand
On Thu, Sep 24, 2020 at 10:19 AM Radu Nicolau  wrote:
>
> Remove RTE_MACHINE_CPUFLAG_ macros from the build.
> Deprecation notice sent, pasted here for reference:
>
> * build macros: The macros defining RTE_MACHINE_CPUFLAG_* will be removed
>   from the build. The information provided by these macros is available
>   through standard compiler macros. For example, RTE_MACHINE_CPUFLAG_SSE3
>   duplicates the compiler-provided macro __SSE3__.
>
>
> Radu Nicolau (4):
>   x86: change cpuflag macros to compiler macros
>   arm: change cpuflag macros to compiler macros
>   ppc: change cpuflag macros to compiler macros
>   doc: remove reference to RTE_MACHINE_CPUFLAG
>
>  app/test-pmd/macswap.c   |  2 +-
>  app/test/test_memcpy_perf.c  |  8 
>  config/arm/meson.build   |  6 --
>  config/ppc/meson.build   |  2 --
>  config/x86/meson.build   |  2 --
>  doc/guides/prog_guide/writing_efficient_code.rst |  2 --
>  doc/guides/rel_notes/release_20_11.rst   |  1 +
>  drivers/net/enic/meson.build |  2 +-
>  drivers/net/i40e/meson.build |  2 +-
>  drivers/net/iavf/meson.build |  2 +-
>  drivers/net/ice/meson.build  |  2 +-
>  drivers/net/ixgbe/ixgbe_ethdev.c |  2 +-
>  examples/l3fwd/l3fwd.h   |  2 +-
>  examples/l3fwd/l3fwd_em.c| 16 
>  examples/l3fwd/l3fwd_em_hlm.h|  2 +-
>  examples/l3fwd/l3fwd_em_sequential.h |  2 +-
>  examples/l3fwd/l3fwd_lpm.c   |  6 +++---
>  lib/librte_acl/meson.build   |  2 +-
>  lib/librte_eal/arm/include/rte_memcpy_32.h   |  2 +-
>  lib/librte_eal/common/rte_random.c   |  4 ++--
>  lib/librte_eal/x86/include/rte_memcpy.h  |  8 
>  lib/librte_efd/rte_efd_x86.h |  2 +-
>  lib/librte_hash/rte_cuckoo_hash.c|  4 ++--
>  lib/librte_hash/rte_hash_crc.h   |  2 +-
>  lib/librte_hash/rte_thash.h  |  4 ++--
>  lib/librte_member/rte_member.h   |  2 +-
>  lib/librte_member/rte_member_ht.c| 10 +-
>  lib/librte_member/rte_member_x86.h   |  2 +-
>  lib/librte_net/rte_net_crc.c |  4 ++--
>  lib/librte_node/ip4_lookup.c |  2 +-
>  lib/librte_sched/rte_sched.c |  2 +-
>  lib/librte_table/rte_lru_arm64.h |  2 +-
>  lib/librte_table/rte_table_hash_func.h   |  2 +-
>  33 files changed, 52 insertions(+), 63 deletions(-)

Acked-by: David Marchand 

Rewrote the release note update based on the deprecation notice.
Dropped the notice.

Applied, thanks.

-- 
David Marchand



Re: [dpdk-dev] Meson Minimum Version

2020-09-25 Thread Dmitry Kozlyuk
On Fri, 25 Sep 2020 09:41:22 +0100, Bruce Richardson wrote:
> On Fri, Sep 25, 2020 at 09:31:53AM +0200, Morten Brørup wrote:
> > > From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Bruce Richardson
[snip]
> > In short, my primary concern is: What could realistically go wrong if the 
> > required version of Meson is buggy?
> > 
> > Bruce, you have worked for quite a while with Meson/Ninja by now, so 
> > perhaps you can assess this risk based on your experience.
> >   
> I'd say the risk in this case is small, especially since I see that 0.56 of
> meson is well under way for development and may well be released before
> DPDK 20.11. Generally backwards compatibilty of meson is excellent as they
> have comprehensive test suite for all features.
>
> Rather than any bugginess, my concern was purely requiring people to update
> meson using "pip3", but I suppose that's not really a big deal, and when
> using pip update it defaults to just updating the copy for the local user,
> not system-wide.

Speaking for Windows, at least twice this year there were incompatibilities
between _minor_ versions of Meson, due to admitted bugs in Meson. However,
IMO this is an argument for using just _exact_ version, not necessarily an old
one. Pip facilitates this better than OS package manager, because developer
controls the version and can easily switch, regardless of distro updates.
Thus, John's upgrade suggestion and transition to pip both look reasonable.



Re: [dpdk-dev] Meson Minimum Version

2020-09-25 Thread Bruce Richardson
On Fri, Sep 25, 2020 at 12:19:33PM +0300, Dmitry Kozlyuk wrote:
> On Fri, 25 Sep 2020 09:41:22 +0100, Bruce Richardson wrote:
> > On Fri, Sep 25, 2020 at 09:31:53AM +0200, Morten Brørup wrote:
> > > > From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Bruce Richardson
> [snip]
> > > In short, my primary concern is: What could realistically go wrong if the 
> > > required version of Meson is buggy?
> > > 
> > > Bruce, you have worked for quite a while with Meson/Ninja by now, so 
> > > perhaps you can assess this risk based on your experience.
> > >   
> > I'd say the risk in this case is small, especially since I see that 0.56 of
> > meson is well under way for development and may well be released before
> > DPDK 20.11. Generally backwards compatibilty of meson is excellent as they
> > have comprehensive test suite for all features.
> >
> > Rather than any bugginess, my concern was purely requiring people to update
> > meson using "pip3", but I suppose that's not really a big deal, and when
> > using pip update it defaults to just updating the copy for the local user,
> > not system-wide.
> 
> Speaking for Windows, at least twice this year there were incompatibilities
> between _minor_ versions of Meson, due to admitted bugs in Meson. However,
> IMO this is an argument for using just _exact_ version, not necessarily an old
> one. Pip facilitates this better than OS package manager, because developer
> controls the version and can easily switch, regardless of distro updates.
> Thus, John's upgrade suggestion and transition to pip both look reasonable.
> 

Seems that meson must be a bit more fragile on windows then, which is a
pity (and perhaps their regression tests aren't as good as I thought).
On Linux, since version 0.40 I think only one version, 0.47.0, caused an
issue for us, which was fixed in 0.47.1.
However, having a recommended version to use can work too.


Re: [dpdk-dev] [PATCH V2 1/4] ethdev: fix compiling errors for per-queue statistics

2020-09-25 Thread Ferruh Yigit

On 9/25/2020 9:58 AM, Min Hu (Connor) wrote:

Hi, Ferruh Yigit,

Your method is very practical. Thanks!
But we found that testpmd fails to start with "--rx-queue-stats-mapping" 
and "--tx-queue-stats-mapping". We've found the cause and modified it. 
These modifications are being tested.


Currently, the following modifications are related to framework APIs.
1)struct 'rte_eth_dcb_tc_queue_mapping'
2)fix 'rte_eth_dev_set_rx/tx_queue_stats_mapping' function to resolve
the problem that DPDK project fails to build when 
'RTE_ETHDEV_QUEUE_STAT_CNTRS' > 256.


According to my current analysis, these modifications for queue stats 
mapping in testpmd do not involve the modification of framework APIs. 
Therefore, I think we should give priority to focusing on the current 
changes to the DPDK framework APIs.


After all, DPDK-20.11 is approaching. What do you think?



Makes sense, agree to continue with this patchset and queue stats 
mapping related changes can be done later.




在 2020/9/23 17:18, Ferruh Yigit 写道:

On 9/23/2020 3:31 AM, Min Hu (Connor) wrote:



在 2020/9/5 2:31, Ferruh Yigit 写道:

On 9/4/2020 12:32 PM, Min Hu (Connor) wrote:

From: Huisong Li 

Currently, only statistics of rx/tx queues with queue_id less than
RTE_ETHDEV_QUEUE_STAT_CNTRS can be displayed. If there is a certain
application scenario that it needs to use 256 or more than 256 queues
and display all statistics of rx/tx queue. At this moment, we have to
change the macro to be equaled to the queue number.

However, modifying the macro to be greater than 256 will trigger
many errors and warnings from test-pmd, PMD driver and librte_ethdev
during compiling dpdk project. But it is possible and permited that
rx/tx queue number is greater than 256 and all statistics of rx/tx
queue need to be displayed. In addition, the data type of rx/tx queue
number in rte_eth_dev_configure API is 'uint16_t'. So It is 
unreasonable

to use the 'uint8_t' type for variables that control which per-queue
statistics can be displayed.

Fixes: ed30d9b691b2 ("app/testpmd: add stats per queue")
Fixes: 09c7e63a71f9 ("net/memif: introduce memory interface PMD")
Fixes: abf7275bbaa2 ("ixgbe: move to drivers/net/")
Fixes: e6defdfddc3b ("net/igc: enable statistics")
Fixes: 2265e4b4e84b ("net/octeontx2: add basic stats operation")
Fixes: 6c3169a3dc04 ("virtio: move to drivers/net/")
Cc: sta...@dpdk.org

Signed-off-by: Huisong Li 
Signed-off-by: Min Hu (Connor) 
Reviewed-by: Wei Hu (Xavier) 
Reviewed-by: Dongdong Liu 


The patch mostly looks good and it enables build with
'RTE_ETHDEV_QUEUE_STAT_CNTRS' > 256.
Only I put a comment for a testpmd change to enable the "set 
stat_qmap" command

for map value > 256.


BUT there are many things to fix in the queue stats mapping, since 
you are
already on it can you help on a few things on testpmd related to it, 
if you have

time for it?

1) Getting queue stats shouldn't require stats mapping, it should be 
controlled
separately. Many PMDs doesn't require/do the stats mapping but they 
still can
collect the per queue stats, which can be displayed independent from 
mapping.


2) Even you map only one queue, while displaying stats it will display
'RTE_ETHDEV_QUEUE_STAT_CNTRS' queues, and when that number is high 
it makes hard

to see the actual interested values.
If there is no mapping, it should display min(number_of_queues,
RTE_ETHDEV_QUEUE_STAT_CNTRS).
If there is mapping it should display queues that mapping done, this 
may require

adding a new 'active' field to 'struct queue_stats_mappings'.

3) Why 'struct queue_stats_mappings' is cache aligned, is it really 
needed?


4) The mapping arrays, 'tx_queue_stats_mappings_array' &
'rx_queue_stats_mappings_array' are global and their size is based 
on fixed max
port and queue size assumptions, can those mapping array be done per 
port and

'RTE_ETHDEV_QUEUE_STAT_CNTRS' size per port?


It does seem to be unreasonable. We also try do it, and found that
it is hard to control per queue stats and queue stats mapping, 
separately.

For details, see next V3 patch.


I will check v3, but as far as know stats mapping exist because of 
limitations of some NICs, that there are N queues but M stats 
registers where N > M. So need to map some queues to stat registers, 
and there is N:1 relation there, so multiple queue stats can be 
represented in single stat register.


But for rest of the NICs it should be possible to display the queue 
stats independent from the mapping.



We're working on debugging this stats_mapping setting. During this 
process,

we have a problem that starting testpmd with --rx/tx-queue-stats-mapping
parameter fails. log as follows:
[root]$ ./testpmd -l 1,3,4,8,12 -n 4 -w :04:00.0 
--file-prefix=lee --log-level=7

  -- -i --rxq=6 --txq=6 --burst=64 --rxd=2048 --txd=2048 --nb-cores=4
  --rx-queue-stats-mapping=(0,2,2) --tx-queue-stats-mapping=(0,3,3)

-bash: syntax error near unexpected token `('
[root]$


Above seems bash error, using '' can help for it, like 
--rx-q

Re: [dpdk-dev] [PATCH V3 0/4] change data type in TC queue

2020-09-25 Thread Ferruh Yigit

On 9/23/2020 1:59 PM, Min Hu (Connor) wrote:

From: Huisong Li 

This series change data type in TC rxq and TC txq,
fix compiling errors for per-queue statistics.

Huisong Li (4):
   dpdk: fix compiling errors for per-queue statistics
   ethdev: change   data type in TC rxq and TC txq
   doc: announce modified field in DCB TC queue mapping
   doc: announce modified in queue_stats_mapping api



The change request from previous version seems not addressed, can you 
please check:

http://inbox.dpdk.org/dev/93479bbe-7ebe-a400-2805-5debd96a0...@intel.com/
Basically it requires following change:
 diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c
 index 4df42dbc28..23e624f442 100644
 --- a/app/test-pmd/cmdline.c
 +++ b/app/test-pmd/cmdline.c
 @@ -8346,7 +8346,7 @@ cmdline_parse_token_num_t cmd_setqmap_queueid =
   queue_id, UINT16);
  cmdline_parse_token_num_t cmd_setqmap_mapvalue =
 TOKEN_NUM_INITIALIZER(struct cmd_set_qmap_result,
 - map_value, UINT8);
 + map_value, UINT16);

  cmdline_parse_inst_t cmd_set_qmap = {
 .f = cmd_set_qmap_parsed,



Also you can drop deprecation notes as explained here:
http://inbox.dpdk.org/dev/38b50f43-0b31-02a8-87ec-d937d83d0...@intel.com/


Thanks,
ferruh


[dpdk-dev] [PATCH v3] net/af_xdp: custom XDP program loading

2020-09-25 Thread Ciara Loftus
The new 'xdp_prog=' vdev arg allows the user to specify the path to
a custom XDP program to be set on the device, instead of the default libbpf
one. The program must have an XSK_MAP of name 'xsks_map' which will allow
for the redirection of some packets to userspace and thus the PMD, using
some criteria defined in the program. This can be useful for filtering
purposes, for example if we only want a subset of packets to reach
userspace or to drop or process a subset of packets in the kernel.

Note: a netdev may only load one program.

Signed-off-by: Ciara Loftus 
Tested-by: Xuekun Hu 
---
v3:
* Rebased to next-net

v2:
* Modified error checking for strnlen return.
* Fixed copyright header edits

 doc/guides/nics/af_xdp.rst  |  1 +
 drivers/net/af_xdp/rte_eth_af_xdp.c | 99 +++--
 2 files changed, 95 insertions(+), 5 deletions(-)

diff --git a/doc/guides/nics/af_xdp.rst b/doc/guides/nics/af_xdp.rst
index 78a088468f..be268fe7ff 100644
--- a/doc/guides/nics/af_xdp.rst
+++ b/doc/guides/nics/af_xdp.rst
@@ -34,6 +34,7 @@ The following options can be provided to set up an af_xdp 
port in DPDK.
 *   ``queue_count`` - total netdev queue number (optional, default 1);
 *   ``shared_umem`` - PMD will attempt to share UMEM with others (optional,
 default 0);
+*   ``xdp_prog`` - path to custom xdp program (optional, default none);
 
 Prerequisites
 -
diff --git a/drivers/net/af_xdp/rte_eth_af_xdp.c 
b/drivers/net/af_xdp/rte_eth_af_xdp.c
index 01f462b465..8dfdadb3c2 100644
--- a/drivers/net/af_xdp/rte_eth_af_xdp.c
+++ b/drivers/net/af_xdp/rte_eth_af_xdp.c
@@ -126,6 +126,8 @@ struct pmd_internals {
int max_queue_cnt;
int combined_queue_cnt;
bool shared_umem;
+   char prog_path[PATH_MAX];
+   bool custom_prog_configured;
 
struct rte_ether_addr eth_addr;
 
@@ -137,12 +139,14 @@ struct pmd_internals {
 #define ETH_AF_XDP_START_QUEUE_ARG "start_queue"
 #define ETH_AF_XDP_QUEUE_COUNT_ARG "queue_count"
 #define ETH_AF_XDP_SHARED_UMEM_ARG "shared_umem"
+#define ETH_AF_XDP_PROG_ARG"xdp_prog"
 
 static const char * const valid_arguments[] = {
ETH_AF_XDP_IFACE_ARG,
ETH_AF_XDP_START_QUEUE_ARG,
ETH_AF_XDP_QUEUE_COUNT_ARG,
ETH_AF_XDP_SHARED_UMEM_ARG,
+   ETH_AF_XDP_PROG_ARG,
NULL
 };
 
@@ -1021,6 +1025,45 @@ xsk_umem_info *xdp_umem_configure(struct pmd_internals 
*internals,
return NULL;
 }
 
+static int
+load_custom_xdp_prog(const char *prog_path, int if_index)
+{
+   int ret, prog_fd = -1;
+   struct bpf_object *obj;
+   struct bpf_map *map;
+
+   ret = bpf_prog_load(prog_path, BPF_PROG_TYPE_XDP, &obj, &prog_fd);
+   if (ret) {
+   AF_XDP_LOG(ERR, "Failed to load program %s\n", prog_path);
+   return ret;
+   }
+
+   /*
+* The loaded program must provision for a map of xsks, such that some
+* traffic can be redirected to userspace. When the xsk is created,
+* libbpf inserts it into the map.
+*/
+   map = bpf_object__find_map_by_name(obj, "xsks_map");
+   if (!map) {
+   AF_XDP_LOG(ERR, "Failed to find xsks_map in %s\n", prog_path);
+   return -1;
+   }
+
+   /* Link the program with the given network device */
+   ret = bpf_set_link_xdp_fd(if_index, prog_fd,
+   XDP_FLAGS_UPDATE_IF_NOEXIST);
+   if (ret) {
+   AF_XDP_LOG(ERR, "Failed to set prog fd %d on interface\n",
+   prog_fd);
+   return -1;
+   }
+
+   AF_XDP_LOG(INFO, "Successfully loaded XDP program %s with fd %d\n",
+   prog_path, prog_fd);
+
+   return 0;
+}
+
 static int
 xsk_configure(struct pmd_internals *internals, struct pkt_rx_queue *rxq,
  int ring_size)
@@ -1046,6 +1089,18 @@ xsk_configure(struct pmd_internals *internals, struct 
pkt_rx_queue *rxq,
cfg.bind_flags |= XDP_USE_NEED_WAKEUP;
 #endif
 
+   if (strnlen(internals->prog_path, PATH_MAX) &&
+   !internals->custom_prog_configured) {
+   ret = load_custom_xdp_prog(internals->prog_path,
+  internals->if_index);
+   if (ret) {
+   AF_XDP_LOG(ERR, "Failed to load custom XDP program 
%s\n",
+   internals->prog_path);
+   goto err;
+   }
+   internals->custom_prog_configured = 1;
+   }
+
if (internals->shared_umem)
ret = create_shared_socket(&rxq->xsk, internals->if_name,
rxq->xsk_queue_idx, rxq->umem->umem, &rxq->rx,
@@ -1264,6 +1319,30 @@ parse_name_arg(const char *key __rte_unused,
return 0;
 }
 
+/** parse xdp prog argument */
+static int
+parse_prog_arg(const char *key __rte_unused,
+  

Re: [dpdk-dev] [PATCH v1] event/sw: performance improvements

2020-09-25 Thread Bruce Richardson
On Thu, Sep 24, 2020 at 06:02:48PM +, Ananyev, Konstantin wrote:
> 
> 
> 
> 
> 
> Add minimum burst throughout the scheduler pipeline and a flush counter.
> Replace ring API calls with local single threaded implementation where
> possible.
> 
> Signed-off-by: Radu Nicolau mailto:radu.nico...@intel.com
> 
> Thanks for the patch, a few comments inline.
> 
> ---
>  drivers/event/sw/sw_evdev.h   | 11 +++-
>  drivers/event/sw/sw_evdev_scheduler.c | 83
> +++
>  2 files changed, 81 insertions(+), 13 deletions(-)
> 
> diff --git a/drivers/event/sw/sw_evdev.h b/drivers/event/sw/sw_evdev.h
> index 7c77b2495..95e51065f 100644
> --- a/drivers/event/sw/sw_evdev.h
> +++ b/drivers/event/sw/sw_evdev.h
> @@ -29,7 +29,13 @@
>  /* report dequeue burst sizes in buckets */  #define
> SW_DEQ_STAT_BUCKET_SHIFT 2
>  /* how many packets pulled from port by sched */ -#define
> SCHED_DEQUEUE_BURST_SIZE 32
> +#define SCHED_DEQUEUE_BURST_SIZE 64
> +
> +#define SCHED_MIN_BURST_SIZE 8
> +#define SCHED_NO_ENQ_CYCLE_FLUSH 256
> +/* set SCHED_DEQUEUE_BURST_SIZE to 64 or 128 when setting this to 1*/
> +#define SCHED_REFILL_ONCE_PER_CALL 1
> 
> Is it possible to make the above #define a runtime option?
> Eg, --vdev event_sw,refill_iter=1
> 
> That would allow packaged versions of DPDK to be usable in both modes.
> 
> +
> 
>  #define SW_PORT_HIST_LIST (MAX_SW_PROD_Q_DEPTH) /* size of our
> history list */  #define NUM_SAMPLES 64 /* how many data points use
> for average stats */ @@ -214,6 +220,9 @@ struct sw_evdev {
>      uint32_t xstats_count_mode_port;
>      uint32_t xstats_count_mode_queue;
> 
> +    uint16_t sched_flush_count;
> +    uint16_t sched_min_burst;
> +
>      /* Contains all ports - load balanced and directed */
>      struct sw_port ports[SW_PORTS_MAX] __rte_cache_aligned;
> 
> diff --git a/drivers/event/sw/sw_evdev_scheduler.c
> b/drivers/event/sw/sw_evdev_scheduler.c
> index cff747da8..ca6d1caff 100644
> --- a/drivers/event/sw/sw_evdev_scheduler.c
> +++ b/drivers/event/sw/sw_evdev_scheduler.c
> @@ -26,6 +26,29 @@
>  /* use cheap bit mixing, we only need to lose a few bits */  #define
> SW_HASH_FLOWID(f) (((f) ^ (f >> 10)) & FLOWID_MASK)
> 
> +
> +/* single object enq and deq for non MT ring */ static
> +__rte_always_inline void sw_nonmt_ring_dequeue(struct rte_ring *r,
> +void **obj) {
> +    if ((r->prod.tail - r->cons.tail) < 1)
> +    return;
> +    void **ring = (void **)&r[1];
> +    *obj = ring[r->cons.tail & r->mask];
> +    r->cons.tail++;
> +}
> +static __rte_always_inline int
> +sw_nonmt_ring_enqueue(struct rte_ring *r, void *obj) {
> +    if ((r->capacity + r->cons.tail - r->prod.tail) < 1)
> +    return 0;
> +    void **ring = (void **)&r[1];
> +    ring[r->prod.tail & r->mask] = obj;
> +    r->prod.tail++;
> +    return 1;
> +
> Why not make these APIs part of the rte_ring library? You could further 
> optimize them by keeping the indices on the same cacheline.
> I'm not sure there is any need for non thread-safe rings outside this 
> particular case.
> [Honnappa] I think if we add the APIs, we will find the use cases.
> But, more than that, I understand that rte_ring structure is exposed to the 
> application. The reason for doing that is the inline functions that rte_ring 
> provides. IMO, we should still maintain modularity and should not use the 
> internals of the rte_ring structure outside of the library.
> 
> +1 to that.
> 
> BTW, is there any real perf benefit from such micor-optimisation?

I'd tend to view these as use-case specific, and I'm not sure we should
clutter up the ring library with yet more functions, especially since they
can't be mixed with the existing enqueue/dequeue functions, since they
don't use the head pointers.



Re: [dpdk-dev] [PATCH 1/7] raw/dpaa2_qdma: change DPAA2 QDMA APIs to rawdev ops

2020-09-25 Thread Hemant Agrawal

Acked-by: Hemant Agrawal 




Re: [dpdk-dev] [PATCH 0/7] raw/dpaa2_qdma: driver enhancement

2020-09-25 Thread Hemant Agrawal

Series-

Acked-by: Hemant Agrawal 

On 9/7/2020 2:55 PM, Gagandeep Singh wrote:

In this patchset, we have done some changes in dpaa2_qdma driver
related to rawdev APIs, optimizations, scatter-gather support on TX,
enqueue without wait.

Gagandeep Singh (2):
   raw/dpaa2_qdma: change DPAA2 QDMA APIs to rawdev ops
   raw/dpaa2_qdma: memset to only required memory

Jun Yang (5):
   raw/dpaa2_qdma: refactor the code
   raw/dpaa2_qdma: optimize IOVA conversion
   raw/dpaa2_qdma: support scatter gather in enqueue
   raw/dpaa2_qdma: support FLE pool per queue
   raw/dpaa2_qdma: support enqueue without response wait

  drivers/bus/fslmc/portal/dpaa2_hw_pvt.h |   18 +-
  drivers/raw/dpaa2_qdma/dpaa2_qdma.c | 1824 ---
  drivers/raw/dpaa2_qdma/dpaa2_qdma.h |  128 +-
  drivers/raw/dpaa2_qdma/rte_pmd_dpaa2_qdma.h |  231 +---
  4 files changed, 1257 insertions(+), 944 deletions(-)



[dpdk-dev] [PATCH v3 02/25] raw/ioat: fix missing close function

2020-09-25 Thread Bruce Richardson
From: Kevin Laatz 

When rte_rawdev_pmd_release() is called, rte_rawdev_close() looks for a
dev_close function for the device causing a segmentation fault when no
close() function is implemented for a driver.

This patch resolves the issue by adding a stub function ioat_dev_close().

Fixes: f687e842e328 ("raw/ioat: introduce IOAT driver")
Cc: sta...@dpdk.org

Reported-by: Sunil Pai G 
Signed-off-by: Kevin Laatz 
---
 drivers/raw/ioat/ioat_rawdev.c | 7 +++
 1 file changed, 7 insertions(+)

diff --git a/drivers/raw/ioat/ioat_rawdev.c b/drivers/raw/ioat/ioat_rawdev.c
index 7f1a15436..0732b059f 100644
--- a/drivers/raw/ioat/ioat_rawdev.c
+++ b/drivers/raw/ioat/ioat_rawdev.c
@@ -203,6 +203,12 @@ ioat_xstats_reset(struct rte_rawdev *dev, const uint32_t 
*ids, uint32_t nb_ids)
return 0;
 }
 
+static int
+ioat_dev_close(struct rte_rawdev *dev __rte_unused)
+{
+   return 0;
+}
+
 extern int ioat_rawdev_test(uint16_t dev_id);
 
 static int
@@ -212,6 +218,7 @@ ioat_rawdev_create(const char *name, struct rte_pci_device 
*dev)
.dev_configure = ioat_dev_configure,
.dev_start = ioat_dev_start,
.dev_stop = ioat_dev_stop,
+   .dev_close = ioat_dev_close,
.dev_info_get = ioat_dev_info_get,
.xstats_get = ioat_xstats_get,
.xstats_get_names = ioat_xstats_get_names,
-- 
2.25.1



[dpdk-dev] [PATCH v3 00/25] raw/ioat: enhancements and new hardware support

2020-09-25 Thread Bruce Richardson
This patchset adds some small enhancements, some rework and also support
for new hardware to the ioat rawdev driver. Most rework and enhancements
are largely self-explanatory from the individual patches.

The new hardware support is for the Intel(R) DSA accelerator which will be
present in future Intel processors. A description of this new hardware is
covered in [1]. Functions specific to the new hardware use the "idxd"
prefix, for consistency with the kernel driver.

[1] https://01.org/blogs/2019/introducing-intel-data-streaming-accelerator

---
V3:
 * More doc updates including release note updates throughout the set
 * Added in fill operation
 * Added in fix for missing close operation
 * Added in fix for doc building to ensure ioat is in in the index

V2:
 * Included documentation additions in the set
 * Split off the rawdev unit test changes to a separate patchset for easier
   review
 * General code improvements and cleanups

Bruce Richardson (19):
  doc/api: add ioat driver to index
  raw/ioat: enable use from C++ code
  raw/ioat: include extra info in error messages
  raw/ioat: split header for readability
  raw/ioat: rename functions to be operation-agnostic
  raw/ioat: add separate API for fence call
  raw/ioat: make the HW register spec private
  raw/ioat: add skeleton for VFIO/UIO based DSA device
  raw/ioat: include example configuration script
  raw/ioat: create rawdev instances on idxd PCI probe
  raw/ioat: add datapath data structures for idxd devices
  raw/ioat: add configure function for idxd devices
  raw/ioat: add start and stop functions for idxd devices
  raw/ioat: add data path for idxd devices
  raw/ioat: add info function for idxd devices
  raw/ioat: create separate statistics structure
  raw/ioat: move xstats functions to common file
  raw/ioat: add xstats tracking for idxd devices
  raw/ioat: clean up use of common test function

Cheng Jiang (1):
  raw/ioat: add a flag to control copying handle parameters

Kevin Laatz (5):
  raw/ioat: fix missing close function
  usertools/dpdk-devbind.py: add support for DSA HW
  raw/ioat: add vdev probe for DSA/idxd devices
  raw/ioat: create rawdev instances for idxd vdevs
  raw/ioat: add fill operation

 doc/api/doxy-api-index.md |   1 +
 doc/api/doxy-api.conf.in  |   1 +
 doc/guides/rawdevs/ioat.rst   | 163 +++--
 doc/guides/rel_notes/release_20_11.rst|  23 +
 drivers/raw/ioat/dpdk_idxd_cfg.py |  79 +++
 drivers/raw/ioat/idxd_pci.c   | 345 ++
 drivers/raw/ioat/idxd_vdev.c  | 233 +++
 drivers/raw/ioat/ioat_common.c| 244 +++
 drivers/raw/ioat/ioat_private.h   |  82 +++
 drivers/raw/ioat/ioat_rawdev.c|  92 +--
 drivers/raw/ioat/ioat_rawdev_test.c   | 112 +++-
 .../raw/ioat/{rte_ioat_spec.h => ioat_spec.h} |  90 ++-
 drivers/raw/ioat/meson.build  |  15 +-
 drivers/raw/ioat/rte_ioat_rawdev.h| 221 +++
 drivers/raw/ioat/rte_ioat_rawdev_fns.h| 599 ++
 examples/ioat/ioatfwd.c   |  16 +-
 lib/librte_eal/include/rte_common.h   |   1 +
 usertools/dpdk-devbind.py |   4 +-
 18 files changed, 1971 insertions(+), 350 deletions(-)
 create mode 100755 drivers/raw/ioat/dpdk_idxd_cfg.py
 create mode 100644 drivers/raw/ioat/idxd_pci.c
 create mode 100644 drivers/raw/ioat/idxd_vdev.c
 create mode 100644 drivers/raw/ioat/ioat_common.c
 create mode 100644 drivers/raw/ioat/ioat_private.h
 rename drivers/raw/ioat/{rte_ioat_spec.h => ioat_spec.h} (74%)
 create mode 100644 drivers/raw/ioat/rte_ioat_rawdev_fns.h

-- 
2.25.1



[dpdk-dev] [PATCH v3 01/25] doc/api: add ioat driver to index

2020-09-25 Thread Bruce Richardson
Add the ioat driver to the doxygen documentation.

Cc: sta...@dpdk.org

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 doc/api/doxy-api-index.md | 1 +
 doc/api/doxy-api.conf.in  | 1 +
 2 files changed, 2 insertions(+)

diff --git a/doc/api/doxy-api-index.md b/doc/api/doxy-api-index.md
index b855a8f3b..06e49f438 100644
--- a/doc/api/doxy-api-index.md
+++ b/doc/api/doxy-api-index.md
@@ -44,6 +44,7 @@ The public API headers are grouped by topics:
   [ixgbe]  (@ref rte_pmd_ixgbe.h),
   [i40e]   (@ref rte_pmd_i40e.h),
   [ice](@ref rte_pmd_ice.h),
+  [ioat]   (@ref rte_ioat_rawdev.h),
   [bnxt]   (@ref rte_pmd_bnxt.h),
   [dpaa]   (@ref rte_pmd_dpaa.h),
   [dpaa2]  (@ref rte_pmd_dpaa2.h),
diff --git a/doc/api/doxy-api.conf.in b/doc/api/doxy-api.conf.in
index 42d38919d..f87336365 100644
--- a/doc/api/doxy-api.conf.in
+++ b/doc/api/doxy-api.conf.in
@@ -18,6 +18,7 @@ INPUT   = @TOPDIR@/doc/api/doxy-api-index.md \
   @TOPDIR@/drivers/net/softnic \
   @TOPDIR@/drivers/raw/dpaa2_cmdif \
   @TOPDIR@/drivers/raw/dpaa2_qdma \
+  @TOPDIR@/drivers/raw/ioat \
   @TOPDIR@/lib/librte_eal/include \
   @TOPDIR@/lib/librte_eal/include/generic \
   @TOPDIR@/lib/librte_acl \
-- 
2.25.1



[dpdk-dev] [PATCH v3 03/25] raw/ioat: enable use from C++ code

2020-09-25 Thread Bruce Richardson
To allow the header file to be used from C++ code we need to ensure all
typecasts are explicit, and include an 'extern "C"' guard.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/rte_ioat_rawdev.h | 23 +--
 1 file changed, 17 insertions(+), 6 deletions(-)

diff --git a/drivers/raw/ioat/rte_ioat_rawdev.h 
b/drivers/raw/ioat/rte_ioat_rawdev.h
index f765a6557..3d8419271 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev.h
@@ -5,6 +5,10 @@
 #ifndef _RTE_IOAT_RAWDEV_H_
 #define _RTE_IOAT_RAWDEV_H_
 
+#ifdef __cplusplus
+extern "C" {
+#endif
+
 /**
  * @file rte_ioat_rawdev.h
  *
@@ -100,7 +104,8 @@ rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, 
phys_addr_t dst,
unsigned int length, uintptr_t src_hdl, uintptr_t dst_hdl,
int fence)
 {
-   struct rte_ioat_rawdev *ioat = rte_rawdevs[dev_id].dev_private;
+   struct rte_ioat_rawdev *ioat =
+   (struct rte_ioat_rawdev 
*)rte_rawdevs[dev_id].dev_private;
unsigned short read = ioat->next_read;
unsigned short write = ioat->next_write;
unsigned short mask = ioat->ring_size - 1;
@@ -141,7 +146,8 @@ rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, 
phys_addr_t dst,
 static inline void
 rte_ioat_do_copies(int dev_id)
 {
-   struct rte_ioat_rawdev *ioat = rte_rawdevs[dev_id].dev_private;
+   struct rte_ioat_rawdev *ioat =
+   (struct rte_ioat_rawdev 
*)rte_rawdevs[dev_id].dev_private;
ioat->desc_ring[(ioat->next_write - 1) & (ioat->ring_size - 1)].u
.control.completion_update = 1;
rte_compiler_barrier();
@@ -190,7 +196,8 @@ static inline int
 rte_ioat_completed_copies(int dev_id, uint8_t max_copies,
uintptr_t *src_hdls, uintptr_t *dst_hdls)
 {
-   struct rte_ioat_rawdev *ioat = rte_rawdevs[dev_id].dev_private;
+   struct rte_ioat_rawdev *ioat =
+   (struct rte_ioat_rawdev 
*)rte_rawdevs[dev_id].dev_private;
unsigned short mask = (ioat->ring_size - 1);
unsigned short read = ioat->next_read;
unsigned short end_read, count;
@@ -212,13 +219,13 @@ rte_ioat_completed_copies(int dev_id, uint8_t max_copies,
__m128i hdls0 = _mm_load_si128(&ioat->hdls[read & mask]);
__m128i hdls1 = _mm_load_si128(&ioat->hdls[(read + 1) & mask]);
 
-   _mm_storeu_si128((void *)&src_hdls[i],
+   _mm_storeu_si128((__m128i *)&src_hdls[i],
_mm_unpacklo_epi64(hdls0, hdls1));
-   _mm_storeu_si128((void *)&dst_hdls[i],
+   _mm_storeu_si128((__m128i *)&dst_hdls[i],
_mm_unpackhi_epi64(hdls0, hdls1));
}
for (; i < count; i++, read++) {
-   uintptr_t *hdls = (void *)&ioat->hdls[read & mask];
+   uintptr_t *hdls = (uintptr_t *)&ioat->hdls[read & mask];
src_hdls[i] = hdls[0];
dst_hdls[i] = hdls[1];
}
@@ -228,4 +235,8 @@ rte_ioat_completed_copies(int dev_id, uint8_t max_copies,
return count;
 }
 
+#ifdef __cplusplus
+}
+#endif
+
 #endif /* _RTE_IOAT_RAWDEV_H_ */
-- 
2.25.1



[dpdk-dev] [PATCH v3 04/25] raw/ioat: include extra info in error messages

2020-09-25 Thread Bruce Richardson
In case of any failures, include the function name and the line number of
the error message in the message, to make tracking down the failure easier.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/ioat_rawdev_test.c | 53 +++--
 1 file changed, 35 insertions(+), 18 deletions(-)

diff --git a/drivers/raw/ioat/ioat_rawdev_test.c 
b/drivers/raw/ioat/ioat_rawdev_test.c
index c463a82ad..8e7fd96af 100644
--- a/drivers/raw/ioat/ioat_rawdev_test.c
+++ b/drivers/raw/ioat/ioat_rawdev_test.c
@@ -13,6 +13,23 @@ int ioat_rawdev_test(uint16_t dev_id); /* pre-define to keep 
compiler happy */
 static struct rte_mempool *pool;
 static unsigned short expected_ring_size;
 
+#define PRINT_ERR(...) print_err(__func__, __LINE__, __VA_ARGS__)
+
+static inline int
+__rte_format_printf(3, 4)
+print_err(const char *func, int lineno, const char *format, ...)
+{
+   va_list ap;
+   int ret;
+
+   ret = fprintf(stderr, "In %s:%d - ", func, lineno);
+   va_start(ap, format);
+   ret += vfprintf(stderr, format, ap);
+   va_end(ap);
+
+   return ret;
+}
+
 static int
 test_enqueue_copies(int dev_id)
 {
@@ -42,7 +59,7 @@ test_enqueue_copies(int dev_id)
(uintptr_t)src,
(uintptr_t)dst,
0 /* no fence */) != 1) {
-   printf("Error with rte_ioat_enqueue_copy\n");
+   PRINT_ERR("Error with rte_ioat_enqueue_copy\n");
return -1;
}
rte_ioat_do_copies(dev_id);
@@ -50,18 +67,18 @@ test_enqueue_copies(int dev_id)
 
if (rte_ioat_completed_copies(dev_id, 1, (void *)&completed[0],
(void *)&completed[1]) != 1) {
-   printf("Error with rte_ioat_completed_copies\n");
+   PRINT_ERR("Error with rte_ioat_completed_copies\n");
return -1;
}
if (completed[0] != src || completed[1] != dst) {
-   printf("Error with completions: got (%p, %p), not 
(%p,%p)\n",
+   PRINT_ERR("Error with completions: got (%p, %p), not 
(%p,%p)\n",
completed[0], completed[1], src, dst);
return -1;
}
 
for (i = 0; i < length; i++)
if (dst_data[i] != src_data[i]) {
-   printf("Data mismatch at char %u\n", i);
+   PRINT_ERR("Data mismatch at char %u\n", i);
return -1;
}
rte_pktmbuf_free(src);
@@ -94,7 +111,7 @@ test_enqueue_copies(int dev_id)
(uintptr_t)srcs[i],
(uintptr_t)dsts[i],
0 /* nofence */) != 1) {
-   printf("Error with rte_ioat_enqueue_copy for 
buffer %u\n",
+   PRINT_ERR("Error with rte_ioat_enqueue_copy for 
buffer %u\n",
i);
return -1;
}
@@ -104,18 +121,18 @@ test_enqueue_copies(int dev_id)
 
if (rte_ioat_completed_copies(dev_id, 64, (void *)completed_src,
(void *)completed_dst) != RTE_DIM(srcs)) {
-   printf("Error with rte_ioat_completed_copies\n");
+   PRINT_ERR("Error with rte_ioat_completed_copies\n");
return -1;
}
for (i = 0; i < RTE_DIM(srcs); i++) {
char *src_data, *dst_data;
 
if (completed_src[i] != srcs[i]) {
-   printf("Error with source pointer %u\n", i);
+   PRINT_ERR("Error with source pointer %u\n", i);
return -1;
}
if (completed_dst[i] != dsts[i]) {
-   printf("Error with dest pointer %u\n", i);
+   PRINT_ERR("Error with dest pointer %u\n", i);
return -1;
}
 
@@ -123,7 +140,7 @@ test_enqueue_copies(int dev_id)
dst_data = rte_pktmbuf_mtod(dsts[i], char *);
for (j = 0; j < length; j++)
if (src_data[j] != dst_data[j]) {
-   printf("Error with copy of packet %u, 
byte %u\n",
+   PRINT_ERR("Error with copy of packet 
%u, byte %u\n",
i, j);
return -1;
}
@@ -150,26 +167,26 @@ ioat_rawdev

[dpdk-dev] [PATCH v3 07/25] raw/ioat: rename functions to be operation-agnostic

2020-09-25 Thread Bruce Richardson
Since the hardware supported by the ioat driver is capable of operations
other than just copies, we can rename the doorbell and completion-return
functions to not have "copies" in their names. These functions are not
copy-specific, and so would apply for other operations which may be added
later to the driver.

Also add a suitable warning using deprecation attribute for any code using
the old functions names.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 

---
Note: The checkpatches warning on this patch is a false positive due to
the addition of the new __rte_deprecated_msg macro in rte_common.h
---
 doc/guides/rawdevs/ioat.rst| 16 
 doc/guides/rel_notes/release_20_11.rst |  9 +
 doc/guides/sample_app_ug/ioat.rst  |  8 
 drivers/raw/ioat/ioat_rawdev_test.c| 12 ++--
 drivers/raw/ioat/rte_ioat_rawdev.h | 14 +++---
 drivers/raw/ioat/rte_ioat_rawdev_fns.h | 20 
 examples/ioat/ioatfwd.c|  4 ++--
 lib/librte_eal/include/rte_common.h|  1 +
 8 files changed, 53 insertions(+), 31 deletions(-)

diff --git a/doc/guides/rawdevs/ioat.rst b/doc/guides/rawdevs/ioat.rst
index af00d77fb..3db5f5d09 100644
--- a/doc/guides/rawdevs/ioat.rst
+++ b/doc/guides/rawdevs/ioat.rst
@@ -157,9 +157,9 @@ Performing Data Copies
 ~~~
 
 To perform data copies using IOAT rawdev devices, the functions
-``rte_ioat_enqueue_copy()`` and ``rte_ioat_do_copies()`` should be used.
+``rte_ioat_enqueue_copy()`` and ``rte_ioat_perform_ops()`` should be used.
 Once copies have been completed, the completion will be reported back when
-the application calls ``rte_ioat_completed_copies()``.
+the application calls ``rte_ioat_completed_ops()``.
 
 The ``rte_ioat_enqueue_copy()`` function enqueues a single copy to the
 device ring for copying at a later point. The parameters to that function
@@ -172,11 +172,11 @@ pointers if packet data is being copied.
 
 While the ``rte_ioat_enqueue_copy()`` function enqueues a copy operation on
 the device ring, the copy will not actually be performed until after the
-application calls the ``rte_ioat_do_copies()`` function. This function
+application calls the ``rte_ioat_perform_ops()`` function. This function
 informs the device hardware of the elements enqueued on the ring, and the
 device will begin to process them. It is expected that, for efficiency
 reasons, a burst of operations will be enqueued to the device via multiple
-enqueue calls between calls to the ``rte_ioat_do_copies()`` function.
+enqueue calls between calls to the ``rte_ioat_perform_ops()`` function.
 
 The following code from ``test_ioat_rawdev.c`` demonstrates how to enqueue
 a burst of copies to the device and start the hardware processing of them:
@@ -210,10 +210,10 @@ a burst of copies to the device and start the hardware 
processing of them:
 return -1;
 }
 }
-rte_ioat_do_copies(dev_id);
+rte_ioat_perform_ops(dev_id);
 
 To retrieve information about completed copies, the API
-``rte_ioat_completed_copies()`` should be used. This API will return to the
+``rte_ioat_completed_ops()`` should be used. This API will return to the
 application a set of completion handles passed in when the relevant copies
 were enqueued.
 
@@ -223,9 +223,9 @@ is correct before freeing the data buffers using the 
returned handles:
 
 .. code-block:: C
 
-if (rte_ioat_completed_copies(dev_id, 64, (void *)completed_src,
+if (rte_ioat_completed_ops(dev_id, 64, (void *)completed_src,
 (void *)completed_dst) != RTE_DIM(srcs)) {
-printf("Error with rte_ioat_completed_copies\n");
+printf("Error with rte_ioat_completed_ops\n");
 return -1;
 }
 for (i = 0; i < RTE_DIM(srcs); i++) {
diff --git a/doc/guides/rel_notes/release_20_11.rst 
b/doc/guides/rel_notes/release_20_11.rst
index 196209f63..c99c0b33f 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -83,6 +83,11 @@ New Features
   The ioat rawdev driver has been updated and enhanced. Changes include:
 
   * Added a per-device configuration flag to disable management of 
user-provided completion handles
+  * Renamed the ``rte_ioat_do_copies()`` API to ``rte_ioat_perform_ops()``,
+and renamed the ``rte_ioat_completed_copies()`` API to 
``rte_ioat_completed_ops()``
+to better reflect the APIs' purposes, and remove the implication that
+they are limited to copy operations only.
+[Note: The old API is still provided but marked as deprecated in the code]
 
 
 Removed Items
@@ -178,6 +183,10 @@ API Changes
 
 * bpf: ``RTE_BPF_XTYPE_NUM`` has been dropped from ``rte_bpf_xtype``.
 
+* raw/ioat: As noted above, the ``rte_ioat_do_copies()`` and
+  ``rte_ioat_completed_copies()`` functions have been renamed to
+  ``rte_ioat_perform_ops()`` and ``rte_ioat_completed_op

[dpdk-dev] [PATCH v3 05/25] raw/ioat: add a flag to control copying handle parameters

2020-09-25 Thread Bruce Richardson
From: Cheng Jiang 

Add a flag which controls whether rte_ioat_enqueue_copy and
rte_ioat_completed_copies function should process handle parameters. Not
doing so can improve the performance when handle parameters are not
necessary.

Signed-off-by: Cheng Jiang 
Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 doc/guides/rawdevs/ioat.rst|  3 ++
 doc/guides/rel_notes/release_20_11.rst |  6 
 drivers/raw/ioat/ioat_rawdev.c |  2 ++
 drivers/raw/ioat/rte_ioat_rawdev.h | 45 +++---
 4 files changed, 45 insertions(+), 11 deletions(-)

diff --git a/doc/guides/rawdevs/ioat.rst b/doc/guides/rawdevs/ioat.rst
index c46460ff4..af00d77fb 100644
--- a/doc/guides/rawdevs/ioat.rst
+++ b/doc/guides/rawdevs/ioat.rst
@@ -129,6 +129,9 @@ output, the ``dev_private`` structure element cannot be 
NULL, and must
 point to a valid ``rte_ioat_rawdev_config`` structure, containing the ring
 size to be used by the device. The ring size must be a power of two,
 between 64 and 4096.
+If it is not needed, the tracking by the driver of user-provided completion
+handles may be disabled by setting the ``hdls_disable`` flag in
+the configuration structure also.
 
 The following code shows how the device is configured in
 ``test_ioat_rawdev.c``:
diff --git a/doc/guides/rel_notes/release_20_11.rst 
b/doc/guides/rel_notes/release_20_11.rst
index 4eb3224a7..196209f63 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -78,6 +78,12 @@ New Features
 ``--portmask=N``
 where N represents the hexadecimal bitmask of ports used.
 
+* **Updated ioat rawdev driver**
+
+  The ioat rawdev driver has been updated and enhanced. Changes include:
+
+  * Added a per-device configuration flag to disable management of 
user-provided completion handles
+
 
 Removed Items
 -
diff --git a/drivers/raw/ioat/ioat_rawdev.c b/drivers/raw/ioat/ioat_rawdev.c
index 0732b059f..ea9f51ffc 100644
--- a/drivers/raw/ioat/ioat_rawdev.c
+++ b/drivers/raw/ioat/ioat_rawdev.c
@@ -58,6 +58,7 @@ ioat_dev_configure(const struct rte_rawdev *dev, 
rte_rawdev_obj_t config,
return -EINVAL;
 
ioat->ring_size = params->ring_size;
+   ioat->hdls_disable = params->hdls_disable;
if (ioat->desc_ring != NULL) {
rte_memzone_free(ioat->desc_mz);
ioat->desc_ring = NULL;
@@ -122,6 +123,7 @@ ioat_dev_info_get(struct rte_rawdev *dev, rte_rawdev_obj_t 
dev_info,
return -EINVAL;
 
cfg->ring_size = ioat->ring_size;
+   cfg->hdls_disable = ioat->hdls_disable;
return 0;
 }
 
diff --git a/drivers/raw/ioat/rte_ioat_rawdev.h 
b/drivers/raw/ioat/rte_ioat_rawdev.h
index 3d8419271..28ce95cc9 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev.h
@@ -38,7 +38,8 @@ extern "C" {
  * an ioat rawdev instance.
  */
 struct rte_ioat_rawdev_config {
-   unsigned short ring_size;
+   unsigned short ring_size; /**< size of job submission descriptor ring */
+   bool hdls_disable;/**< if set, ignore user-supplied handle params */
 };
 
 /**
@@ -56,6 +57,7 @@ struct rte_ioat_rawdev {
 
unsigned short ring_size;
struct rte_ioat_generic_hw_desc *desc_ring;
+   bool hdls_disable;
__m128i *hdls; /* completion handles for returning to user */
 
 
@@ -88,10 +90,14 @@ struct rte_ioat_rawdev {
  *   The length of the data to be copied
  * @param src_hdl
  *   An opaque handle for the source data, to be returned when this operation
- *   has been completed and the user polls for the completion details
+ *   has been completed and the user polls for the completion details.
+ *   NOTE: If hdls_disable configuration option for the device is set, this
+ *   parameter is ignored.
  * @param dst_hdl
  *   An opaque handle for the destination data, to be returned when this
- *   operation has been completed and the user polls for the completion details
+ *   operation has been completed and the user polls for the completion 
details.
+ *   NOTE: If hdls_disable configuration option for the device is set, this
+ *   parameter is ignored.
  * @param fence
  *   A flag parameter indicating that hardware should not begin to perform any
  *   subsequently enqueued copy operations until after this operation has
@@ -126,8 +132,10 @@ rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, 
phys_addr_t dst,
desc->u.control_raw = (uint32_t)((!!fence << 4) | (!(write & 0xF)) << 
3);
desc->src_addr = src;
desc->dest_addr = dst;
+   if (!ioat->hdls_disable)
+   ioat->hdls[write] = _mm_set_epi64x((int64_t)dst_hdl,
+   (int64_t)src_hdl);
 
-   ioat->hdls[write] = _mm_set_epi64x((int64_t)dst_hdl, (int64_t)src_hdl);
rte_prefetch0(&ioat->desc_ring[ioat->next_write & mask]);
 
ioat->enqueued++;
@@ -174,19 +182,29 @@ rte_ioat_get_last_completed(struct rte_ioat_rawdev *

[dpdk-dev] [PATCH v3 06/25] raw/ioat: split header for readability

2020-09-25 Thread Bruce Richardson
Rather than having a single long complicated header file for general use we
can split things so that there is one header with all the publicly needed
information - data structs and function prototypes - while the rest of the
internal details are put separately. This makes it easier to read,
understand and use the APIs.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---

Some versions of checkpatch show errors about spacing in this patch,
however, it appears that these are false positives.
---
 drivers/raw/ioat/meson.build   |   1 +
 drivers/raw/ioat/rte_ioat_rawdev.h | 147 +-
 drivers/raw/ioat/rte_ioat_rawdev_fns.h | 168 +
 3 files changed, 175 insertions(+), 141 deletions(-)
 create mode 100644 drivers/raw/ioat/rte_ioat_rawdev_fns.h

diff --git a/drivers/raw/ioat/meson.build b/drivers/raw/ioat/meson.build
index 0878418ae..f66e9b605 100644
--- a/drivers/raw/ioat/meson.build
+++ b/drivers/raw/ioat/meson.build
@@ -8,4 +8,5 @@ sources = files('ioat_rawdev.c',
 deps += ['rawdev', 'bus_pci', 'mbuf']
 
 install_headers('rte_ioat_rawdev.h',
+   'rte_ioat_rawdev_fns.h',
'rte_ioat_spec.h')
diff --git a/drivers/raw/ioat/rte_ioat_rawdev.h 
b/drivers/raw/ioat/rte_ioat_rawdev.h
index 28ce95cc9..7067b352f 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev.h
@@ -18,12 +18,7 @@ extern "C" {
  * @b EXPERIMENTAL: these structures and APIs may change without prior notice
  */
 
-#include 
-#include 
-#include 
-#include 
-#include 
-#include "rte_ioat_spec.h"
+#include 
 
 /** Name of the device driver */
 #define IOAT_PMD_RAWDEV_NAME rawdev_ioat
@@ -42,38 +37,6 @@ struct rte_ioat_rawdev_config {
bool hdls_disable;/**< if set, ignore user-supplied handle params */
 };
 
-/**
- * @internal
- * Structure representing a device instance
- */
-struct rte_ioat_rawdev {
-   struct rte_rawdev *rawdev;
-   const struct rte_memzone *mz;
-   const struct rte_memzone *desc_mz;
-
-   volatile struct rte_ioat_registers *regs;
-   phys_addr_t status_addr;
-   phys_addr_t ring_addr;
-
-   unsigned short ring_size;
-   struct rte_ioat_generic_hw_desc *desc_ring;
-   bool hdls_disable;
-   __m128i *hdls; /* completion handles for returning to user */
-
-
-   unsigned short next_read;
-   unsigned short next_write;
-
-   /* some statistics for tracking, if added/changed update xstats fns*/
-   uint64_t enqueue_failed __rte_cache_aligned;
-   uint64_t enqueued;
-   uint64_t started;
-   uint64_t completed;
-
-   /* to report completions, the device will write status back here */
-   volatile uint64_t status __rte_cache_aligned;
-};
-
 /**
  * Enqueue a copy operation onto the ioat device
  *
@@ -108,39 +71,7 @@ struct rte_ioat_rawdev {
 static inline int
 rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, phys_addr_t dst,
unsigned int length, uintptr_t src_hdl, uintptr_t dst_hdl,
-   int fence)
-{
-   struct rte_ioat_rawdev *ioat =
-   (struct rte_ioat_rawdev 
*)rte_rawdevs[dev_id].dev_private;
-   unsigned short read = ioat->next_read;
-   unsigned short write = ioat->next_write;
-   unsigned short mask = ioat->ring_size - 1;
-   unsigned short space = mask + read - write;
-   struct rte_ioat_generic_hw_desc *desc;
-
-   if (space == 0) {
-   ioat->enqueue_failed++;
-   return 0;
-   }
-
-   ioat->next_write = write + 1;
-   write &= mask;
-
-   desc = &ioat->desc_ring[write];
-   desc->size = length;
-   /* set descriptor write-back every 16th descriptor */
-   desc->u.control_raw = (uint32_t)((!!fence << 4) | (!(write & 0xF)) << 
3);
-   desc->src_addr = src;
-   desc->dest_addr = dst;
-   if (!ioat->hdls_disable)
-   ioat->hdls[write] = _mm_set_epi64x((int64_t)dst_hdl,
-   (int64_t)src_hdl);
-
-   rte_prefetch0(&ioat->desc_ring[ioat->next_write & mask]);
-
-   ioat->enqueued++;
-   return 1;
-}
+   int fence);
 
 /**
  * Trigger hardware to begin performing enqueued copy operations
@@ -152,32 +83,7 @@ rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, 
phys_addr_t dst,
  *   The rawdev device id of the ioat instance
  */
 static inline void
-rte_ioat_do_copies(int dev_id)
-{
-   struct rte_ioat_rawdev *ioat =
-   (struct rte_ioat_rawdev 
*)rte_rawdevs[dev_id].dev_private;
-   ioat->desc_ring[(ioat->next_write - 1) & (ioat->ring_size - 1)].u
-   .control.completion_update = 1;
-   rte_compiler_barrier();
-   ioat->regs->dmacount = ioat->next_write;
-   ioat->started = ioat->enqueued;
-}
-
-/**
- * @internal
- * Returns the index of the last completed operation.
- */
-static inline int
-rte_ioat_get_last_completed(struct rte_ioat_rawdev *ioat, int *error)
-{

[dpdk-dev] [PATCH v3 10/25] usertools/dpdk-devbind.py: add support for DSA HW

2020-09-25 Thread Bruce Richardson
From: Kevin Laatz 

Intel Data Streaming Accelerator (Intel DSA) is a high-performance data
copy and transformation accelerator which will be integrated in future
Intel processors [1].

Add DSA device support to dpdk-devbind.py script.

[1] https://01.org/blogs/2019/introducing-intel-data-streaming-accelerator

Signed-off-by: Bruce Richardson 
Signed-off-by: Kevin Laatz 
---
 doc/guides/rel_notes/release_20_11.rst | 2 ++
 usertools/dpdk-devbind.py  | 4 +++-
 2 files changed, 5 insertions(+), 1 deletion(-)

diff --git a/doc/guides/rel_notes/release_20_11.rst 
b/doc/guides/rel_notes/release_20_11.rst
index 3868529ac..4d8b78154 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -82,6 +82,8 @@ New Features
 
   The ioat rawdev driver has been updated and enhanced. Changes include:
 
+  * Added support for Intel\ |reg| Data Streaming Accelerator hardware.
+For more information, see 
https://01.org/blogs/2019/introducing-intel-data-streaming-accelerator
   * Added a per-device configuration flag to disable management of 
user-provided completion handles
   * Renamed the ``rte_ioat_do_copies()`` API to ``rte_ioat_perform_ops()``,
 and renamed the ``rte_ioat_completed_copies()`` API to 
``rte_ioat_completed_ops()``
diff --git a/usertools/dpdk-devbind.py b/usertools/dpdk-devbind.py
index 094c2ffc8..f2916bef5 100755
--- a/usertools/dpdk-devbind.py
+++ b/usertools/dpdk-devbind.py
@@ -53,6 +53,8 @@
   'SVendor': None, 'SDevice': None}
 intel_ioat_icx = {'Class': '08', 'Vendor': '8086', 'Device': '0b00',
   'SVendor': None, 'SDevice': None}
+intel_idxd_spr = {'Class': '08', 'Vendor': '8086', 'Device': '0b25',
+  'SVendor': None, 'SDevice': None}
 intel_ntb_skx = {'Class': '06', 'Vendor': '8086', 'Device': '201c',
   'SVendor': None, 'SDevice': None}
 
@@ -62,7 +64,7 @@
 eventdev_devices = [cavium_sso, cavium_tim, octeontx2_sso]
 mempool_devices = [cavium_fpa, octeontx2_npa]
 compress_devices = [cavium_zip]
-misc_devices = [intel_ioat_bdw, intel_ioat_skx, intel_ioat_icx, intel_ntb_skx, 
octeontx2_dma]
+misc_devices = [intel_ioat_bdw, intel_ioat_skx, intel_ioat_icx, 
intel_idxd_spr, intel_ntb_skx, octeontx2_dma]
 
 # global dict ethernet devices present. Dictionary indexed by PCI address.
 # Each device within this is itself a dictionary of device properties
-- 
2.25.1



[dpdk-dev] [PATCH v3 12/25] raw/ioat: add vdev probe for DSA/idxd devices

2020-09-25 Thread Bruce Richardson
From: Kevin Laatz 

The Intel DSA devices can be exposed to userspace via kernel driver, so can
be used without having to bind them to vfio/uio. Therefore we add support
for using those kernel-configured devices as vdevs, taking as parameter the
individual HW work queue to be used by the vdev.

Signed-off-by: Kevin Laatz 
Signed-off-by: Bruce Richardson 
---
 doc/guides/rawdevs/ioat.rst  |  68 +--
 drivers/raw/ioat/idxd_vdev.c | 123 +++
 drivers/raw/ioat/meson.build |   6 +-
 3 files changed, 192 insertions(+), 5 deletions(-)
 create mode 100644 drivers/raw/ioat/idxd_vdev.c

diff --git a/doc/guides/rawdevs/ioat.rst b/doc/guides/rawdevs/ioat.rst
index b898f98d5..5b8d27980 100644
--- a/doc/guides/rawdevs/ioat.rst
+++ b/doc/guides/rawdevs/ioat.rst
@@ -37,9 +37,62 @@ No additional compilation steps are necessary.
 Device Setup
 -
 
+Depending on support provided by the PMD, HW devices can either use the kernel 
configured driver
+or be bound to a user-space IO driver for use.
+For example, Intel\ |reg| DSA devices can use the IDXD kernel driver or 
DPDK-supported drivers,
+such as ``vfio-pci``.
+
+Intel\ |reg| DSA devices using idxd kernel driver
+~~
+
+To use a Intel\ |reg| DSA device bound to the IDXD kernel driver, the device 
must first be configured.
+The `accel-config `_ utility library can 
be used for configuration.
+
+.. note::
+The device configuration can also be done by directly interacting with 
the sysfs nodes.
+
+There are some mandatory configuration steps before being able to use a device 
with an application.
+The internal engines, which do the copies or other operations,
+and the work-queues, which are used by applications to assign work to the 
device,
+need to be assigned to groups, and the various other configuration options,
+such as priority or queue depth, need to be set for each queue.
+
+To assign an engine to a group::
+
+$ accel-config config-engine dsa0/engine0.0 --group-id=0
+$ accel-config config-engine dsa0/engine0.1 --group-id=1
+
+To assign work queues to groups for passing descriptors to the engines a 
similar accel-config command can be used.
+However, the work queues also need to be configured depending on the use-case.
+Some configuration options include:
+
+* mode (Dedicated/Shared): Indicates whether a WQ may accept jobs from 
multiple queues simultaneously.
+* priority: WQ priority between 1 and 15. Larger value means higher priority.
+* wq-size: the size of the WQ. Sum of all WQ sizes must be less that the 
total-size defined by the device.
+* type: WQ type (kernel/mdev/user). Determines how the device is presented.
+* name: identifier given to the WQ.
+
+Example configuration for a work queue::
+
+$ accel-config config-wq dsa0/wq0.0 --group-id=0 \
+   --mode=dedicated --priority=10 --wq-size=8 \
+   --type=user --name=app1
+
+Once the devices have been configured, they need to be enabled::
+
+$ accel-config enable-device dsa0
+$ accel-config enable-wq dsa0/wq0.0
+
+Check the device configuration::
+
+$ accel-config list
+
+Devices using VFIO/UIO drivers
+~~~
+
 The HW devices to be used will need to be bound to a user-space IO driver for 
use.
 The ``dpdk-devbind.py`` script can be used to view the state of the devices
-and to bind them to a suitable DPDK-supported kernel driver, such as 
``vfio-pci``.
+and to bind them to a suitable DPDK-supported driver, such as ``vfio-pci``.
 For example::
 
$ dpdk-devbind.py -b vfio-pci 00:04.0 00:04.1
@@ -47,9 +100,16 @@ For example::
 Device Probing and Initialization
 ~~
 
-Once bound to a suitable kernel device driver, the HW devices will be found
-as part of the PCI scan done at application initialization time. No vdev
-parameters need to be passed to create or initialize the device.
+For devices bound to a suitable DPDK-supported VFIO/UIO driver, the HW devices 
will
+be found as part of the device scan done at application initialization time 
without
+the need to pass parameters to the application.
+
+If the device is bound to the IDXD kernel driver (and previously configured 
with sysfs),
+then a specific work queue needs to be passed to the application via a vdev 
parameter.
+This vdev parameter take the driver name and work queue name as parameters.
+For example, to use work queue 0 on Intel\ |reg| DSA instance 0::
+
+$ dpdk-test --no-pci --vdev=rawdev_idxd,wq=0.0
 
 Once probed successfully, the device will appear as a ``rawdev``, that is a
 "raw device type" inside DPDK, and can be accessed using APIs from the
diff --git a/drivers/raw/ioat/idxd_vdev.c b/drivers/raw/ioat/idxd_vdev.c
new file mode 100644
index 0..0509fc084
--- /dev/null
+++ b/drivers/raw/ioat/idxd_vdev.c
@@ -0,0 +1,123 @@
+/* SPDX-License-Identifier: BSD-3-

[dpdk-dev] [PATCH v3 09/25] raw/ioat: make the HW register spec private

2020-09-25 Thread Bruce Richardson
Only a few definitions from the hardware spec are actually used in the
driver runtime, so we can copy over those few and make the rest of the spec
a private header in the driver.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/ioat_rawdev.c|  3 ++
 .../raw/ioat/{rte_ioat_spec.h => ioat_spec.h} | 26 ---
 drivers/raw/ioat/meson.build  |  3 +-
 drivers/raw/ioat/rte_ioat_rawdev_fns.h| 43 +--
 4 files changed, 44 insertions(+), 31 deletions(-)
 rename drivers/raw/ioat/{rte_ioat_spec.h => ioat_spec.h} (91%)

diff --git a/drivers/raw/ioat/ioat_rawdev.c b/drivers/raw/ioat/ioat_rawdev.c
index ea9f51ffc..aa59b731f 100644
--- a/drivers/raw/ioat/ioat_rawdev.c
+++ b/drivers/raw/ioat/ioat_rawdev.c
@@ -4,10 +4,12 @@
 
 #include 
 #include 
+#include 
 #include 
 #include 
 
 #include "rte_ioat_rawdev.h"
+#include "ioat_spec.h"
 
 static struct rte_pci_driver ioat_pmd_drv;
 
@@ -268,6 +270,7 @@ ioat_rawdev_create(const char *name, struct rte_pci_device 
*dev)
ioat->rawdev = rawdev;
ioat->mz = mz;
ioat->regs = dev->mem_resource[0].addr;
+   ioat->doorbell = &ioat->regs->dmacount;
ioat->ring_size = 0;
ioat->desc_ring = NULL;
ioat->status_addr = ioat->mz->iova +
diff --git a/drivers/raw/ioat/rte_ioat_spec.h b/drivers/raw/ioat/ioat_spec.h
similarity index 91%
rename from drivers/raw/ioat/rte_ioat_spec.h
rename to drivers/raw/ioat/ioat_spec.h
index c6e7929b2..9645e16d4 100644
--- a/drivers/raw/ioat/rte_ioat_spec.h
+++ b/drivers/raw/ioat/ioat_spec.h
@@ -86,32 +86,6 @@ struct rte_ioat_registers {
 
 #define RTE_IOAT_CHANCMP_ALIGN 8   /* CHANCMP address must 
be 64-bit aligned */
 
-struct rte_ioat_generic_hw_desc {
-   uint32_t size;
-   union {
-   uint32_t control_raw;
-   struct {
-   uint32_t int_enable: 1;
-   uint32_t src_snoop_disable: 1;
-   uint32_t dest_snoop_disable: 1;
-   uint32_t completion_update: 1;
-   uint32_t fence: 1;
-   uint32_t reserved2: 1;
-   uint32_t src_page_break: 1;
-   uint32_t dest_page_break: 1;
-   uint32_t bundle: 1;
-   uint32_t dest_dca: 1;
-   uint32_t hint: 1;
-   uint32_t reserved: 13;
-   uint32_t op: 8;
-   } control;
-   } u;
-   uint64_t src_addr;
-   uint64_t dest_addr;
-   uint64_t next;
-   uint64_t op_specific[4];
-};
-
 struct rte_ioat_dma_hw_desc {
uint32_t size;
union {
diff --git a/drivers/raw/ioat/meson.build b/drivers/raw/ioat/meson.build
index f66e9b605..06636f8a9 100644
--- a/drivers/raw/ioat/meson.build
+++ b/drivers/raw/ioat/meson.build
@@ -8,5 +8,4 @@ sources = files('ioat_rawdev.c',
 deps += ['rawdev', 'bus_pci', 'mbuf']
 
 install_headers('rte_ioat_rawdev.h',
-   'rte_ioat_rawdev_fns.h',
-   'rte_ioat_spec.h')
+   'rte_ioat_rawdev_fns.h')
diff --git a/drivers/raw/ioat/rte_ioat_rawdev_fns.h 
b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
index 466721a23..c6e0b9a58 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev_fns.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
@@ -8,7 +8,36 @@
 #include 
 #include 
 #include 
-#include "rte_ioat_spec.h"
+
+/**
+ * @internal
+ * Structure representing a device descriptor
+ */
+struct rte_ioat_generic_hw_desc {
+   uint32_t size;
+   union {
+   uint32_t control_raw;
+   struct {
+   uint32_t int_enable: 1;
+   uint32_t src_snoop_disable: 1;
+   uint32_t dest_snoop_disable: 1;
+   uint32_t completion_update: 1;
+   uint32_t fence: 1;
+   uint32_t reserved2: 1;
+   uint32_t src_page_break: 1;
+   uint32_t dest_page_break: 1;
+   uint32_t bundle: 1;
+   uint32_t dest_dca: 1;
+   uint32_t hint: 1;
+   uint32_t reserved: 13;
+   uint32_t op: 8;
+   } control;
+   } u;
+   uint64_t src_addr;
+   uint64_t dest_addr;
+   uint64_t next;
+   uint64_t op_specific[4];
+};
 
 /**
  * @internal
@@ -19,7 +48,7 @@ struct rte_ioat_rawdev {
const struct rte_memzone *mz;
const struct rte_memzone *desc_mz;
 
-   volatile struct rte_ioat_registers *regs;
+   volatile uint16_t *doorbell;
phys_addr_t status_addr;
phys_addr_t ring_addr;
 
@@ -40,8 +69,16 @@ struct rte_ioat_rawdev {
 
/* to report completions, the device will write status back here */
volatile uint64_t status __rte_cache_aligned;
+
+   /* pointer to the register bar */
+   volatile stru

[dpdk-dev] [PATCH v3 08/25] raw/ioat: add separate API for fence call

2020-09-25 Thread Bruce Richardson
Rather than having the fence signalled via a flag on a descriptor - which
requires reading the docs to find out whether the flag needs to go on the
last descriptor before, or the first descriptor after the fence - we can
instead add a separate fence API call. This becomes unambiguous to use,
since the fence call explicitly comes between two other enqueue calls. It
also allows more freedom of implementation in the driver code.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 doc/guides/rawdevs/ioat.rst|  3 +--
 doc/guides/rel_notes/release_20_11.rst |  4 
 drivers/raw/ioat/ioat_rawdev_test.c|  6 ++
 drivers/raw/ioat/rte_ioat_rawdev.h | 26 --
 drivers/raw/ioat/rte_ioat_rawdev_fns.h | 22 +++---
 examples/ioat/ioatfwd.c| 12 
 6 files changed, 50 insertions(+), 23 deletions(-)

diff --git a/doc/guides/rawdevs/ioat.rst b/doc/guides/rawdevs/ioat.rst
index 3db5f5d09..71bca0b28 100644
--- a/doc/guides/rawdevs/ioat.rst
+++ b/doc/guides/rawdevs/ioat.rst
@@ -203,8 +203,7 @@ a burst of copies to the device and start the hardware 
processing of them:
 dsts[i]->buf_iova + dsts[i]->data_off,
 length,
 (uintptr_t)srcs[i],
-(uintptr_t)dsts[i],
-0 /* nofence */) != 1) {
+(uintptr_t)dsts[i]) != 1) {
 printf("Error with rte_ioat_enqueue_copy for buffer 
%u\n",
 i);
 return -1;
diff --git a/doc/guides/rel_notes/release_20_11.rst 
b/doc/guides/rel_notes/release_20_11.rst
index c99c0b33f..3868529ac 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -88,6 +88,10 @@ New Features
 to better reflect the APIs' purposes, and remove the implication that
 they are limited to copy operations only.
 [Note: The old API is still provided but marked as deprecated in the code]
+  * Added a new API ``rte_ioat_fence()`` to add a fence between operations.
+This API replaces the ``fence`` flag parameter in the 
``rte_ioat_enqueue_copies()`` function,
+and is clearer as there is no ambiguity as to whether the flag should be
+set on the last operation before the fence or the first operation after it.
 
 
 Removed Items
diff --git a/drivers/raw/ioat/ioat_rawdev_test.c 
b/drivers/raw/ioat/ioat_rawdev_test.c
index bb40eab6b..8ff546803 100644
--- a/drivers/raw/ioat/ioat_rawdev_test.c
+++ b/drivers/raw/ioat/ioat_rawdev_test.c
@@ -57,8 +57,7 @@ test_enqueue_copies(int dev_id)
dst->buf_iova + dst->data_off,
length,
(uintptr_t)src,
-   (uintptr_t)dst,
-   0 /* no fence */) != 1) {
+   (uintptr_t)dst) != 1) {
PRINT_ERR("Error with rte_ioat_enqueue_copy\n");
return -1;
}
@@ -109,8 +108,7 @@ test_enqueue_copies(int dev_id)
dsts[i]->buf_iova + dsts[i]->data_off,
length,
(uintptr_t)srcs[i],
-   (uintptr_t)dsts[i],
-   0 /* nofence */) != 1) {
+   (uintptr_t)dsts[i]) != 1) {
PRINT_ERR("Error with rte_ioat_enqueue_copy for 
buffer %u\n",
i);
return -1;
diff --git a/drivers/raw/ioat/rte_ioat_rawdev.h 
b/drivers/raw/ioat/rte_ioat_rawdev.h
index 5b2c47e8c..6b891cd44 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev.h
@@ -61,17 +61,31 @@ struct rte_ioat_rawdev_config {
  *   operation has been completed and the user polls for the completion 
details.
  *   NOTE: If hdls_disable configuration option for the device is set, this
  *   parameter is ignored.
- * @param fence
- *   A flag parameter indicating that hardware should not begin to perform any
- *   subsequently enqueued copy operations until after this operation has
- *   completed
  * @return
  *   Number of operations enqueued, either 0 or 1
  */
 static inline int
 rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, phys_addr_t dst,
-   unsigned int length, uintptr_t src_hdl, uintptr_t dst_hdl,
-   int fence);
+   unsigned int length, uintptr_t src_hdl, uintptr_t dst_hdl);
+
+/**
+ * Add a fence to force ordering between operations
+ *
+ * This adds a fence to a sequence of operations to enforce ordering, such that
+ * all operations enqueued before the fence must be completed before operations
+ * after the f

[dpdk-dev] [PATCH v3 11/25] raw/ioat: add skeleton for VFIO/UIO based DSA device

2020-09-25 Thread Bruce Richardson
Add in the basic probe/remove skeleton code for DSA devices which are bound
directly to vfio or uio driver. The kernel module for supporting these uses
the "idxd" name, so that name is used as function and file prefix to avoid
conflict with existing "ioat" prefixed functions.

Since we are adding new files to the driver and there will be common
definitions shared between the various files, we create a new internal
header file ioat_private.h to hold common macros and function prototypes.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 doc/guides/rawdevs/ioat.rst | 69 ++---
 drivers/raw/ioat/idxd_pci.c | 56 ++
 drivers/raw/ioat/ioat_private.h | 27 +
 drivers/raw/ioat/ioat_rawdev.c  |  9 +
 drivers/raw/ioat/meson.build|  6 ++-
 5 files changed, 108 insertions(+), 59 deletions(-)
 create mode 100644 drivers/raw/ioat/idxd_pci.c
 create mode 100644 drivers/raw/ioat/ioat_private.h

diff --git a/doc/guides/rawdevs/ioat.rst b/doc/guides/rawdevs/ioat.rst
index 71bca0b28..b898f98d5 100644
--- a/doc/guides/rawdevs/ioat.rst
+++ b/doc/guides/rawdevs/ioat.rst
@@ -3,10 +3,12 @@
 
 .. include:: 
 
-IOAT Rawdev Driver for Intel\ |reg| QuickData Technology
-==
+IOAT Rawdev Driver
+===
 
 The ``ioat`` rawdev driver provides a poll-mode driver (PMD) for Intel\ |reg|
+Data Streaming Accelerator `(Intel DSA)
+`_ and 
for Intel\ |reg|
 QuickData Technology, part of Intel\ |reg| I/O Acceleration Technology
 `(Intel I/OAT)
 
`_.
@@ -17,61 +19,30 @@ be done by software, freeing up CPU cycles for other tasks.
 Hardware Requirements
 --
 
-On Linux, the presence of an Intel\ |reg| QuickData Technology hardware can
-be detected by checking the output of the ``lspci`` command, where the
-hardware will be often listed as "Crystal Beach DMA" or "CBDMA". For
-example, on a system with Intel\ |reg| Xeon\ |reg| CPU E5-2699 v4 @2.20GHz,
-lspci shows:
-
-.. code-block:: console
-
-  # lspci | grep DMA
-  00:04.0 System peripheral: Intel Corporation Xeon E7 v4/Xeon E5 v4/Xeon E3 
v4/Xeon D Crystal Beach DMA Channel 0 (rev 01)
-  00:04.1 System peripheral: Intel Corporation Xeon E7 v4/Xeon E5 v4/Xeon E3 
v4/Xeon D Crystal Beach DMA Channel 1 (rev 01)
-  00:04.2 System peripheral: Intel Corporation Xeon E7 v4/Xeon E5 v4/Xeon E3 
v4/Xeon D Crystal Beach DMA Channel 2 (rev 01)
-  00:04.3 System peripheral: Intel Corporation Xeon E7 v4/Xeon E5 v4/Xeon E3 
v4/Xeon D Crystal Beach DMA Channel 3 (rev 01)
-  00:04.4 System peripheral: Intel Corporation Xeon E7 v4/Xeon E5 v4/Xeon E3 
v4/Xeon D Crystal Beach DMA Channel 4 (rev 01)
-  00:04.5 System peripheral: Intel Corporation Xeon E7 v4/Xeon E5 v4/Xeon E3 
v4/Xeon D Crystal Beach DMA Channel 5 (rev 01)
-  00:04.6 System peripheral: Intel Corporation Xeon E7 v4/Xeon E5 v4/Xeon E3 
v4/Xeon D Crystal Beach DMA Channel 6 (rev 01)
-  00:04.7 System peripheral: Intel Corporation Xeon E7 v4/Xeon E5 v4/Xeon E3 
v4/Xeon D Crystal Beach DMA Channel 7 (rev 01)
-
-On a system with Intel\ |reg| Xeon\ |reg| Gold 6154 CPU @ 3.00GHz, lspci
-shows:
-
-.. code-block:: console
-
-  # lspci | grep DMA
-  00:04.0 System peripheral: Intel Corporation Sky Lake-E CBDMA Registers (rev 
04)
-  00:04.1 System peripheral: Intel Corporation Sky Lake-E CBDMA Registers (rev 
04)
-  00:04.2 System peripheral: Intel Corporation Sky Lake-E CBDMA Registers (rev 
04)
-  00:04.3 System peripheral: Intel Corporation Sky Lake-E CBDMA Registers (rev 
04)
-  00:04.4 System peripheral: Intel Corporation Sky Lake-E CBDMA Registers (rev 
04)
-  00:04.5 System peripheral: Intel Corporation Sky Lake-E CBDMA Registers (rev 
04)
-  00:04.6 System peripheral: Intel Corporation Sky Lake-E CBDMA Registers (rev 
04)
-  00:04.7 System peripheral: Intel Corporation Sky Lake-E CBDMA Registers (rev 
04)
-
+The ``dpdk-devbind.py`` script, included with DPDK,
+can be used to show the presence of supported hardware.
+Running ``dpdk-devbind.py --status-dev misc`` will show all the miscellaneous,
+or rawdev-based devices on the system.
+For Intel\ |reg| QuickData Technology devices, the hardware will be often 
listed as "Crystal Beach DMA",
+or "CBDMA".
+For Intel\ |reg| DSA devices, they are currently (at time of writing) 
appearing as devices with type "0b25",
+due to the absence of pci-id database entries for them at this point.
 
 Compilation
 
 
-For builds done with ``make``, the driver compilation is enabled by the
-``CONFIG_RTE_LIBRTE_PMD_IOAT_RAWDEV`` build configuration option. This is
-enabled by default in builds for x86 platforms, and disabled in other
-configurations.
-
-For builds using ``meson`` and ``ninja``, the driver will be built when the
-target platform is x86-based.
+For builds usi

[dpdk-dev] [PATCH v3 13/25] raw/ioat: include example configuration script

2020-09-25 Thread Bruce Richardson
Devices managed by the idxd kernel driver must be configured for DPDK use
before it can be used by the ioat driver. This example script serves both
as a quick way to get the driver set up with a simple configuration, and as
the basis for users to modify it and create their own configuration
scripts.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 doc/guides/rawdevs/ioat.rst   |  2 +
 drivers/raw/ioat/dpdk_idxd_cfg.py | 79 +++
 2 files changed, 81 insertions(+)
 create mode 100755 drivers/raw/ioat/dpdk_idxd_cfg.py

diff --git a/doc/guides/rawdevs/ioat.rst b/doc/guides/rawdevs/ioat.rst
index 5b8d27980..7c2a2d457 100644
--- a/doc/guides/rawdevs/ioat.rst
+++ b/doc/guides/rawdevs/ioat.rst
@@ -50,6 +50,8 @@ The `accel-config `_ 
utility library can b
 
 .. note::
 The device configuration can also be done by directly interacting with 
the sysfs nodes.
+An example of how this may be done can be seen in the script 
``dpdk_idxd_cfg.py``
+included in the driver source directory.
 
 There are some mandatory configuration steps before being able to use a device 
with an application.
 The internal engines, which do the copies or other operations,
diff --git a/drivers/raw/ioat/dpdk_idxd_cfg.py 
b/drivers/raw/ioat/dpdk_idxd_cfg.py
new file mode 100755
index 0..bce4bb5bd
--- /dev/null
+++ b/drivers/raw/ioat/dpdk_idxd_cfg.py
@@ -0,0 +1,79 @@
+#!/usr/bin/env python3
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(c) 2020 Intel Corporation
+
+"""
+Configure an entire Intel DSA instance, using idxd kernel driver, for DPDK use
+"""
+
+import sys
+import argparse
+import os
+import os.path
+
+
+class SysfsDir:
+"Used to read/write paths in a sysfs directory"
+def __init__(self, path):
+self.path = path
+
+def read_int(self, filename):
+"Return a value from sysfs file"
+with open(os.path.join(self.path, filename)) as f:
+return int(f.readline())
+
+def write_values(self, values):
+"write dictionary, where key is filename and value is value to write"
+for filename, contents in values.items():
+with open(os.path.join(self.path, filename), "w") as f:
+f.write(str(contents))
+
+
+def configure_dsa(dsa_id, queues):
+"Configure the DSA instance with appropriate number of queues"
+dsa_dir = SysfsDir(f"/sys/bus/dsa/devices/dsa{dsa_id}")
+drv_dir = SysfsDir("/sys/bus/dsa/drivers/dsa")
+
+max_groups = dsa_dir.read_int("max_groups")
+max_engines = dsa_dir.read_int("max_engines")
+max_queues = dsa_dir.read_int("max_work_queues")
+max_tokens = dsa_dir.read_int("max_tokens")
+
+# we want one engine per group
+nb_groups = min(max_engines, max_groups)
+for grp in range(nb_groups):
+dsa_dir.write_values({f"engine{dsa_id}.{grp}/group_id": grp})
+
+nb_queues = min(queues, max_queues)
+if queues > nb_queues:
+print(f"Setting number of queues to max supported value: {max_queues}")
+
+# configure each queue
+for q in range(nb_queues):
+wq_dir = SysfsDir(os.path.join(dsa_dir.path, f"wq{dsa_id}.{q}"))
+wq_dir.write_values({"group_id": q % nb_groups,
+ "type": "user",
+ "mode": "dedicated",
+ "name": f"dpdk_wq{dsa_id}.{q}",
+ "priority": 1,
+ "size": int(max_tokens / nb_queues)})
+
+# enable device and then queues
+drv_dir.write_values({"bind": f"dsa{dsa_id}"})
+for q in range(nb_queues):
+drv_dir.write_values({"bind": f"wq{dsa_id}.{q}"})
+
+
+def main(args):
+"Main function, does arg parsing and calls config function"
+arg_p = argparse.ArgumentParser(
+description="Configure whole DSA device instance for DPDK use")
+arg_p.add_argument('dsa_id', type=int, help="DSA instance number")
+arg_p.add_argument('-q', metavar='queues', type=int, default=255,
+   help="Number of queues to set up")
+parsed_args = arg_p.parse_args(args[1:])
+configure_dsa(parsed_args.dsa_id, parsed_args.q)
+
+
+if __name__ == "__main__":
+main(sys.argv)
-- 
2.25.1



[dpdk-dev] [PATCH v3 17/25] raw/ioat: add configure function for idxd devices

2020-09-25 Thread Bruce Richardson
Add configure function for idxd devices, taking the same parameters as the
existing configure function for ioat. The ring_size parameter is used to
compute the maximum number of bursts to be supported by the driver, given
that the hardware works on individual bursts of descriptors at a time.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/idxd_pci.c|  1 +
 drivers/raw/ioat/idxd_vdev.c   |  1 +
 drivers/raw/ioat/ioat_common.c | 64 ++
 drivers/raw/ioat/ioat_private.h|  3 ++
 drivers/raw/ioat/rte_ioat_rawdev_fns.h |  1 +
 5 files changed, 70 insertions(+)

diff --git a/drivers/raw/ioat/idxd_pci.c b/drivers/raw/ioat/idxd_pci.c
index 113ee98e8..2d8dbc2f2 100644
--- a/drivers/raw/ioat/idxd_pci.c
+++ b/drivers/raw/ioat/idxd_pci.c
@@ -55,6 +55,7 @@ static const struct rte_rawdev_ops idxd_pci_ops = {
.dev_close = idxd_rawdev_close,
.dev_selftest = idxd_rawdev_test,
.dump = idxd_dev_dump,
+   .dev_configure = idxd_dev_configure,
 };
 
 /* each portal uses 4 x 4k pages */
diff --git a/drivers/raw/ioat/idxd_vdev.c b/drivers/raw/ioat/idxd_vdev.c
index 31d8916d0..aa6a5a9aa 100644
--- a/drivers/raw/ioat/idxd_vdev.c
+++ b/drivers/raw/ioat/idxd_vdev.c
@@ -34,6 +34,7 @@ static const struct rte_rawdev_ops idxd_vdev_ops = {
.dev_close = idxd_rawdev_close,
.dev_selftest = idxd_rawdev_test,
.dump = idxd_dev_dump,
+   .dev_configure = idxd_dev_configure,
 };
 
 static void *
diff --git a/drivers/raw/ioat/ioat_common.c b/drivers/raw/ioat/ioat_common.c
index 672241351..5173c331c 100644
--- a/drivers/raw/ioat/ioat_common.c
+++ b/drivers/raw/ioat/ioat_common.c
@@ -44,6 +44,70 @@ idxd_dev_dump(struct rte_rawdev *dev, FILE *f)
return 0;
 }
 
+int
+idxd_dev_configure(const struct rte_rawdev *dev,
+   rte_rawdev_obj_t config, size_t config_size)
+{
+   struct idxd_rawdev *idxd = dev->dev_private;
+   struct rte_idxd_rawdev *rte_idxd = &idxd->public;
+   struct rte_ioat_rawdev_config *cfg = config;
+   uint16_t max_desc = cfg->ring_size;
+   uint16_t max_batches = max_desc / BATCH_SIZE;
+   uint16_t i;
+
+   if (config_size != sizeof(*cfg))
+   return -EINVAL;
+
+   if (dev->started) {
+   IOAT_PMD_ERR("%s: Error, device is started.", __func__);
+   return -EAGAIN;
+   }
+
+   rte_idxd->hdls_disable = cfg->hdls_disable;
+
+   /* limit the batches to what can be stored in hardware */
+   if (max_batches > idxd->max_batches) {
+   IOAT_PMD_DEBUG("Ring size of %u is too large for this device, 
need to limit to %u batches of %u",
+   max_desc, idxd->max_batches, BATCH_SIZE);
+   max_batches = idxd->max_batches;
+   max_desc = max_batches * BATCH_SIZE;
+   }
+   if (!rte_is_power_of_2(max_desc))
+   max_desc = rte_align32pow2(max_desc);
+   IOAT_PMD_DEBUG("Rawdev %u using %u descriptors in %u batches",
+   dev->dev_id, max_desc, max_batches);
+
+   /* in case we are reconfiguring a device, free any existing memory */
+   rte_free(rte_idxd->batch_ring);
+   rte_free(rte_idxd->hdl_ring);
+
+   rte_idxd->batch_ring = rte_zmalloc(NULL,
+   sizeof(*rte_idxd->batch_ring) * max_batches, 0);
+   if (rte_idxd->batch_ring == NULL)
+   return -ENOMEM;
+
+   rte_idxd->hdl_ring = rte_zmalloc(NULL,
+   sizeof(*rte_idxd->hdl_ring) * max_desc, 0);
+   if (rte_idxd->hdl_ring == NULL) {
+   rte_free(rte_idxd->batch_ring);
+   rte_idxd->batch_ring = NULL;
+   return -ENOMEM;
+   }
+   rte_idxd->batch_ring_sz = max_batches;
+   rte_idxd->hdl_ring_sz = max_desc;
+
+   for (i = 0; i < rte_idxd->batch_ring_sz; i++) {
+   struct rte_idxd_desc_batch *b = &rte_idxd->batch_ring[i];
+   b->batch_desc.completion = rte_mem_virt2iova(&b->comp);
+   b->batch_desc.desc_addr = rte_mem_virt2iova(&b->null_desc);
+   b->batch_desc.op_flags = (idxd_op_batch << IDXD_CMD_OP_SHIFT) |
+   IDXD_FLAG_COMPLETION_ADDR_VALID |
+   IDXD_FLAG_REQUEST_COMPLETION;
+   }
+
+   return 0;
+}
+
 int
 idxd_rawdev_create(const char *name, struct rte_device *dev,
   const struct idxd_rawdev *base_idxd,
diff --git a/drivers/raw/ioat/ioat_private.h b/drivers/raw/ioat/ioat_private.h
index f521c85a1..aba70d8d7 100644
--- a/drivers/raw/ioat/ioat_private.h
+++ b/drivers/raw/ioat/ioat_private.h
@@ -59,6 +59,9 @@ extern int idxd_rawdev_create(const char *name, struct 
rte_device *dev,
 
 extern int idxd_rawdev_close(struct rte_rawdev *dev);
 
+extern int idxd_dev_configure(const struct rte_rawdev *dev,
+   rte_rawdev_obj_t

[dpdk-dev] [PATCH v3 14/25] raw/ioat: create rawdev instances on idxd PCI probe

2020-09-25 Thread Bruce Richardson
When a matching device is found via PCI probe create a rawdev instance for
each queue on the hardware. Use empty self-test function for these devices
so that the overall rawdev_autotest does not report failures.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/idxd_pci.c| 237 -
 drivers/raw/ioat/ioat_common.c |  68 +++
 drivers/raw/ioat/ioat_private.h|  33 
 drivers/raw/ioat/ioat_rawdev_test.c|   7 +
 drivers/raw/ioat/ioat_spec.h   |  64 +++
 drivers/raw/ioat/meson.build   |   1 +
 drivers/raw/ioat/rte_ioat_rawdev_fns.h |  35 +++-
 7 files changed, 442 insertions(+), 3 deletions(-)
 create mode 100644 drivers/raw/ioat/ioat_common.c

diff --git a/drivers/raw/ioat/idxd_pci.c b/drivers/raw/ioat/idxd_pci.c
index 1a30e9c31..6752959ed 100644
--- a/drivers/raw/ioat/idxd_pci.c
+++ b/drivers/raw/ioat/idxd_pci.c
@@ -3,8 +3,10 @@
  */
 
 #include 
+#include 
 
 #include "ioat_private.h"
+#include "ioat_spec.h"
 
 #define IDXD_VENDOR_ID 0x8086
 #define IDXD_DEVICE_ID_SPR 0x0B25
@@ -16,17 +18,246 @@ const struct rte_pci_id pci_id_idxd_map[] = {
{ .vendor_id = 0, /* sentinel */ },
 };
 
+static inline int
+idxd_pci_dev_command(struct idxd_rawdev *idxd, enum rte_idxd_cmds command)
+{
+   uint8_t err_code;
+   uint16_t qid = idxd->qid;
+   int i = 0;
+
+   if (command >= idxd_disable_wq && command <= idxd_reset_wq)
+   qid = (1 << qid);
+   rte_spinlock_lock(&idxd->u.pci->lk);
+   idxd->u.pci->regs->cmd = (command << IDXD_CMD_SHIFT) | qid;
+
+   do {
+   rte_pause();
+   err_code = idxd->u.pci->regs->cmdstatus;
+   if (++i >= 1000) {
+   IOAT_PMD_ERR("Timeout waiting for command response from 
HW");
+   rte_spinlock_unlock(&idxd->u.pci->lk);
+   return err_code;
+   }
+   } while (idxd->u.pci->regs->cmdstatus & CMDSTATUS_ACTIVE_MASK);
+   rte_spinlock_unlock(&idxd->u.pci->lk);
+
+   return err_code & CMDSTATUS_ERR_MASK;
+}
+
+static int
+idxd_is_wq_enabled(struct idxd_rawdev *idxd)
+{
+   uint32_t state = idxd->u.pci->wq_regs[idxd->qid].wqcfg[WQ_STATE_IDX];
+   return ((state >> WQ_STATE_SHIFT) & WQ_STATE_MASK) == 0x1;
+}
+
+static const struct rte_rawdev_ops idxd_pci_ops = {
+   .dev_close = idxd_rawdev_close,
+   .dev_selftest = idxd_rawdev_test,
+};
+
+/* each portal uses 4 x 4k pages */
+#define IDXD_PORTAL_SIZE (4096 * 4)
+
+static int
+init_pci_device(struct rte_pci_device *dev, struct idxd_rawdev *idxd)
+{
+   struct idxd_pci_common *pci;
+   uint8_t nb_groups, nb_engines, nb_wqs;
+   uint16_t grp_offset, wq_offset; /* how far into bar0 the regs are */
+   uint16_t wq_size, total_wq_size;
+   uint8_t lg2_max_batch, lg2_max_copy_size;
+   unsigned int i, err_code;
+
+   pci = malloc(sizeof(*pci));
+   if (pci == NULL) {
+   IOAT_PMD_ERR("%s: Can't allocate memory", __func__);
+   goto err;
+   }
+   rte_spinlock_init(&pci->lk);
+
+   /* assign the bar registers, and then configure device */
+   pci->regs = dev->mem_resource[0].addr;
+   grp_offset = (uint16_t)pci->regs->offsets[0];
+   pci->grp_regs = RTE_PTR_ADD(pci->regs, grp_offset * 0x100);
+   wq_offset = (uint16_t)(pci->regs->offsets[0] >> 16);
+   pci->wq_regs = RTE_PTR_ADD(pci->regs, wq_offset * 0x100);
+   pci->portals = dev->mem_resource[2].addr;
+
+   /* sanity check device status */
+   if (pci->regs->gensts & GENSTS_DEV_STATE_MASK) {
+   /* need function-level-reset (FLR) or is enabled */
+   IOAT_PMD_ERR("Device status is not disabled, cannot init");
+   goto err;
+   }
+   if (pci->regs->cmdstatus & CMDSTATUS_ACTIVE_MASK) {
+   /* command in progress */
+   IOAT_PMD_ERR("Device has a command in progress, cannot init");
+   goto err;
+   }
+
+   /* read basic info about the hardware for use when configuring */
+   nb_groups = (uint8_t)pci->regs->grpcap;
+   nb_engines = (uint8_t)pci->regs->engcap;
+   nb_wqs = (uint8_t)(pci->regs->wqcap >> 16);
+   total_wq_size = (uint16_t)pci->regs->wqcap;
+   lg2_max_copy_size = (uint8_t)(pci->regs->gencap >> 16) & 0x1F;
+   lg2_max_batch = (uint8_t)(pci->regs->gencap >> 21) & 0x0F;
+
+   IOAT_PMD_DEBUG("nb_groups = %u, nb_engines = %u, nb_wqs = %u",
+   nb_groups, nb_engines, nb_wqs);
+
+   /* zero out any old config */
+   for (i = 0; i < nb_groups; i++) {
+   pci->grp_regs[i].grpengcfg = 0;
+   pci->grp_regs[i].grpwqcfg[0] = 0;
+   }
+   for (i = 0; i < nb_wqs; i++)
+   pci->wq_regs[i].wqcfg[0] = 0;
+
+   /* put each engine into a separate group to avoid reordering */
+   if (nb_groups > nb_engines)

[dpdk-dev] [PATCH v3 15/25] raw/ioat: create rawdev instances for idxd vdevs

2020-09-25 Thread Bruce Richardson
From: Kevin Laatz 

For each vdev (DSA work queue) instance, create a rawdev instance.

Signed-off-by: Kevin Laatz 
Signed-off-by: Bruce Richardson 
---
 drivers/raw/ioat/idxd_vdev.c| 106 +++-
 drivers/raw/ioat/ioat_private.h |   4 ++
 2 files changed, 109 insertions(+), 1 deletion(-)

diff --git a/drivers/raw/ioat/idxd_vdev.c b/drivers/raw/ioat/idxd_vdev.c
index 0509fc084..d2d588916 100644
--- a/drivers/raw/ioat/idxd_vdev.c
+++ b/drivers/raw/ioat/idxd_vdev.c
@@ -2,6 +2,12 @@
  * Copyright(c) 2020 Intel Corporation
  */
 
+#include 
+#include 
+#include 
+#include 
+
+#include 
 #include 
 #include 
 #include 
@@ -24,6 +30,36 @@ struct idxd_vdev_args {
uint8_t wq_id;
 };
 
+static const struct rte_rawdev_ops idxd_vdev_ops = {
+   .dev_close = idxd_rawdev_close,
+   .dev_selftest = idxd_rawdev_test,
+};
+
+static void *
+idxd_vdev_mmap_wq(struct idxd_vdev_args *args)
+{
+   void *addr;
+   char path[PATH_MAX];
+   int fd;
+
+   snprintf(path, sizeof(path), "/dev/dsa/wq%u.%u",
+   args->device_id, args->wq_id);
+   fd = open(path, O_RDWR);
+   if (fd < 0) {
+   IOAT_PMD_ERR("Failed to open device path");
+   return NULL;
+   }
+
+   addr = mmap(NULL, 0x1000, PROT_WRITE, MAP_SHARED | MAP_POPULATE, fd, 0);
+   close(fd);
+   if (addr == MAP_FAILED) {
+   IOAT_PMD_ERR("Failed to mmap device");
+   return NULL;
+   }
+
+   return addr;
+}
+
 static int
 idxd_rawdev_parse_wq(const char *key __rte_unused, const char *value,
  void *extra_args)
@@ -70,10 +106,32 @@ idxd_vdev_parse_params(struct rte_kvargs *kvlist, struct 
idxd_vdev_args *args)
return -EINVAL;
 }
 
+static int
+idxd_vdev_get_max_batches(struct idxd_vdev_args *args)
+{
+   char sysfs_path[PATH_MAX];
+   FILE *f;
+   int ret;
+
+   snprintf(sysfs_path, sizeof(sysfs_path),
+   "/sys/bus/dsa/devices/wq%u.%u/size",
+   args->device_id, args->wq_id);
+   f = fopen(sysfs_path, "r");
+   if (f == NULL)
+   return -1;
+
+   if (fscanf(f, "%d", &ret) != 1)
+   ret = -1;
+
+   fclose(f);
+   return ret;
+}
+
 static int
 idxd_rawdev_probe_vdev(struct rte_vdev_device *vdev)
 {
struct rte_kvargs *kvlist;
+   struct idxd_rawdev idxd = {0};
struct idxd_vdev_args vdev_args;
const char *name;
int ret = 0;
@@ -96,13 +154,32 @@ idxd_rawdev_probe_vdev(struct rte_vdev_device *vdev)
return -EINVAL;
}
 
+   idxd.qid = vdev_args.wq_id;
+   idxd.u.vdev.dsa_id = vdev_args.device_id;
+   idxd.max_batches = idxd_vdev_get_max_batches(&vdev_args);
+
+   idxd.public.portal = idxd_vdev_mmap_wq(&vdev_args);
+   if (idxd.public.portal == NULL) {
+   IOAT_PMD_ERR("WQ mmap failed");
+   return -ENOENT;
+   }
+
+   ret = idxd_rawdev_create(name, &vdev->device, &idxd, &idxd_vdev_ops);
+   if (ret) {
+   IOAT_PMD_ERR("Failed to create rawdev %s", name);
+   return ret;
+   }
+
return 0;
 }
 
 static int
 idxd_rawdev_remove_vdev(struct rte_vdev_device *vdev)
 {
+   struct idxd_rawdev *idxd;
const char *name;
+   struct rte_rawdev *rdev;
+   int ret = 0;
 
name = rte_vdev_device_name(vdev);
if (name == NULL)
@@ -110,7 +187,34 @@ idxd_rawdev_remove_vdev(struct rte_vdev_device *vdev)
 
IOAT_PMD_INFO("Remove DSA vdev %p", name);
 
-   return 0;
+   rdev = rte_rawdev_pmd_get_named_dev(name);
+   if (!rdev) {
+   IOAT_PMD_ERR("Invalid device name (%s)", name);
+   return -EINVAL;
+   }
+
+   idxd = rdev->dev_private;
+
+   /* free context and memory */
+   if (rdev->dev_private != NULL) {
+   IOAT_PMD_DEBUG("Freeing device driver memory");
+   rdev->dev_private = NULL;
+
+   if (munmap(idxd->public.portal, 0x1000) < 0) {
+   IOAT_PMD_ERR("Error unmapping portal");
+   ret = -errno;
+   }
+
+   rte_free(idxd->public.batch_ring);
+   rte_free(idxd->public.hdl_ring);
+
+   rte_memzone_free(idxd->mz);
+   }
+
+   if (rte_rawdev_pmd_release(rdev))
+   IOAT_PMD_ERR("Device cleanup failed");
+
+   return ret;
 }
 
 struct rte_vdev_driver idxd_rawdev_drv_vdev = {
diff --git a/drivers/raw/ioat/ioat_private.h b/drivers/raw/ioat/ioat_private.h
index 53f00a9f3..6f7bdb499 100644
--- a/drivers/raw/ioat/ioat_private.h
+++ b/drivers/raw/ioat/ioat_private.h
@@ -45,6 +45,10 @@ struct idxd_rawdev {
uint16_t max_batches;
 
union {
+   struct {
+   unsigned int dsa_id;
+   } vdev;
+
struct idxd_pci_common *pci;
} u;
 };
-- 
2.25.1



[dpdk-dev] [PATCH v3 16/25] raw/ioat: add datapath data structures for idxd devices

2020-09-25 Thread Bruce Richardson
Add in the relevant data structures for the data path for DSA devices. Also
include a device dump function to output the status of each device.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/idxd_pci.c|  1 +
 drivers/raw/ioat/idxd_vdev.c   |  1 +
 drivers/raw/ioat/ioat_common.c | 34 +++
 drivers/raw/ioat/ioat_private.h|  2 +
 drivers/raw/ioat/ioat_rawdev_test.c|  3 +-
 drivers/raw/ioat/rte_ioat_rawdev_fns.h | 80 ++
 6 files changed, 120 insertions(+), 1 deletion(-)

diff --git a/drivers/raw/ioat/idxd_pci.c b/drivers/raw/ioat/idxd_pci.c
index 6752959ed..113ee98e8 100644
--- a/drivers/raw/ioat/idxd_pci.c
+++ b/drivers/raw/ioat/idxd_pci.c
@@ -54,6 +54,7 @@ idxd_is_wq_enabled(struct idxd_rawdev *idxd)
 static const struct rte_rawdev_ops idxd_pci_ops = {
.dev_close = idxd_rawdev_close,
.dev_selftest = idxd_rawdev_test,
+   .dump = idxd_dev_dump,
 };
 
 /* each portal uses 4 x 4k pages */
diff --git a/drivers/raw/ioat/idxd_vdev.c b/drivers/raw/ioat/idxd_vdev.c
index d2d588916..31d8916d0 100644
--- a/drivers/raw/ioat/idxd_vdev.c
+++ b/drivers/raw/ioat/idxd_vdev.c
@@ -33,6 +33,7 @@ struct idxd_vdev_args {
 static const struct rte_rawdev_ops idxd_vdev_ops = {
.dev_close = idxd_rawdev_close,
.dev_selftest = idxd_rawdev_test,
+   .dump = idxd_dev_dump,
 };
 
 static void *
diff --git a/drivers/raw/ioat/ioat_common.c b/drivers/raw/ioat/ioat_common.c
index c3aa015ed..672241351 100644
--- a/drivers/raw/ioat/ioat_common.c
+++ b/drivers/raw/ioat/ioat_common.c
@@ -14,6 +14,36 @@ idxd_rawdev_close(struct rte_rawdev *dev __rte_unused)
return 0;
 }
 
+int
+idxd_dev_dump(struct rte_rawdev *dev, FILE *f)
+{
+   struct idxd_rawdev *idxd = dev->dev_private;
+   struct rte_idxd_rawdev *rte_idxd = &idxd->public;
+   int i;
+
+   fprintf(f, "Raw Device #%d\n", dev->dev_id);
+   fprintf(f, "Driver: %s\n\n", dev->driver_name);
+
+   fprintf(f, "Portal: %p\n", rte_idxd->portal);
+   fprintf(f, "Batch Ring size: %u\n", rte_idxd->batch_ring_sz);
+   fprintf(f, "Comp Handle Ring size: %u\n\n", rte_idxd->hdl_ring_sz);
+
+   fprintf(f, "Next batch: %u\n", rte_idxd->next_batch);
+   fprintf(f, "Next batch to be completed: %u\n", 
rte_idxd->next_completed);
+   for (i = 0; i < rte_idxd->batch_ring_sz; i++) {
+   struct rte_idxd_desc_batch *b = &rte_idxd->batch_ring[i];
+   fprintf(f, "Batch %u @%p: submitted=%u, op_count=%u, 
hdl_end=%u\n",
+   i, b, b->submitted, b->op_count, b->hdl_end);
+   }
+
+   fprintf(f, "\n");
+   fprintf(f, "Next free hdl: %u\n", rte_idxd->next_free_hdl);
+   fprintf(f, "Last completed hdl: %u\n", rte_idxd->last_completed_hdl);
+   fprintf(f, "Next returned hdl: %u\n", rte_idxd->next_ret_hdl);
+
+   return 0;
+}
+
 int
 idxd_rawdev_create(const char *name, struct rte_device *dev,
   const struct idxd_rawdev *base_idxd,
@@ -25,6 +55,10 @@ idxd_rawdev_create(const char *name, struct rte_device *dev,
char mz_name[RTE_MEMZONE_NAMESIZE];
int ret = 0;
 
+   RTE_BUILD_BUG_ON(sizeof(struct rte_idxd_hw_desc) != 64);
+   RTE_BUILD_BUG_ON(offsetof(struct rte_idxd_hw_desc, size) != 32);
+   RTE_BUILD_BUG_ON(sizeof(struct rte_idxd_completion) != 32);
+
if (!name) {
IOAT_PMD_ERR("Invalid name of the device!");
ret = -EINVAL;
diff --git a/drivers/raw/ioat/ioat_private.h b/drivers/raw/ioat/ioat_private.h
index 6f7bdb499..f521c85a1 100644
--- a/drivers/raw/ioat/ioat_private.h
+++ b/drivers/raw/ioat/ioat_private.h
@@ -61,4 +61,6 @@ extern int idxd_rawdev_close(struct rte_rawdev *dev);
 
 extern int idxd_rawdev_test(uint16_t dev_id);
 
+extern int idxd_dev_dump(struct rte_rawdev *dev, FILE *f);
+
 #endif /* _IOAT_PRIVATE_H_ */
diff --git a/drivers/raw/ioat/ioat_rawdev_test.c 
b/drivers/raw/ioat/ioat_rawdev_test.c
index 7cd0f4abf..a9132a8f1 100644
--- a/drivers/raw/ioat/ioat_rawdev_test.c
+++ b/drivers/raw/ioat/ioat_rawdev_test.c
@@ -261,7 +261,8 @@ ioat_rawdev_test(uint16_t dev_id)
 }
 
 int
-idxd_rawdev_test(uint16_t dev_id __rte_unused)
+idxd_rawdev_test(uint16_t dev_id)
 {
+   rte_rawdev_dump(dev_id, stdout);
return 0;
 }
diff --git a/drivers/raw/ioat/rte_ioat_rawdev_fns.h 
b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
index fa2eb5334..178c432dd 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev_fns.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
@@ -90,6 +90,86 @@ struct rte_ioat_rawdev {
 #define RTE_IOAT_CHANSTS_HALTED0x3
 #define RTE_IOAT_CHANSTS_ARMED 0x4
 
+/*
+ * Defines used in the data path for interacting with hardware.
+ */
+#define IDXD_CMD_OP_SHIFT 24
+enum rte_idxd_ops {
+   idxd_op_nop = 0,
+   idxd_op_batch,
+   idxd_op_drain,
+   idxd_op_memmove,
+   idxd_

[dpdk-dev] [PATCH v3 18/25] raw/ioat: add start and stop functions for idxd devices

2020-09-25 Thread Bruce Richardson
Add the start and stop functions for DSA hardware devices using the
vfio/uio kernel drivers. For vdevs using the idxd kernel driver, the device
must be started using sysfs before the device node appears for vdev use -
making start/stop functions in the driver unnecessary.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/idxd_pci.c | 50 +
 1 file changed, 50 insertions(+)

diff --git a/drivers/raw/ioat/idxd_pci.c b/drivers/raw/ioat/idxd_pci.c
index 2d8dbc2f2..e4eecdc3a 100644
--- a/drivers/raw/ioat/idxd_pci.c
+++ b/drivers/raw/ioat/idxd_pci.c
@@ -51,11 +51,61 @@ idxd_is_wq_enabled(struct idxd_rawdev *idxd)
return ((state >> WQ_STATE_SHIFT) & WQ_STATE_MASK) == 0x1;
 }
 
+static void
+idxd_pci_dev_stop(struct rte_rawdev *dev)
+{
+   struct idxd_rawdev *idxd = dev->dev_private;
+   uint8_t err_code;
+
+   if (!idxd_is_wq_enabled(idxd)) {
+   IOAT_PMD_ERR("Work queue %d already disabled", idxd->qid);
+   return;
+   }
+
+   err_code = idxd_pci_dev_command(idxd, idxd_disable_wq);
+   if (err_code || idxd_is_wq_enabled(idxd)) {
+   IOAT_PMD_ERR("Failed disabling work queue %d, error code: %#x",
+   idxd->qid, err_code);
+   return;
+   }
+   IOAT_PMD_DEBUG("Work queue %d disabled OK", idxd->qid);
+}
+
+static int
+idxd_pci_dev_start(struct rte_rawdev *dev)
+{
+   struct idxd_rawdev *idxd = dev->dev_private;
+   uint8_t err_code;
+
+   if (idxd_is_wq_enabled(idxd)) {
+   IOAT_PMD_WARN("WQ %d already enabled", idxd->qid);
+   return 0;
+   }
+
+   if (idxd->public.batch_ring == NULL) {
+   IOAT_PMD_ERR("WQ %d has not been fully configured", idxd->qid);
+   return -EINVAL;
+   }
+
+   err_code = idxd_pci_dev_command(idxd, idxd_enable_wq);
+   if (err_code || !idxd_is_wq_enabled(idxd)) {
+   IOAT_PMD_ERR("Failed enabling work queue %d, error code: %#x",
+   idxd->qid, err_code);
+   return err_code == 0 ? -1 : err_code;
+   }
+
+   IOAT_PMD_DEBUG("Work queue %d enabled OK", idxd->qid);
+
+   return 0;
+}
+
 static const struct rte_rawdev_ops idxd_pci_ops = {
.dev_close = idxd_rawdev_close,
.dev_selftest = idxd_rawdev_test,
.dump = idxd_dev_dump,
.dev_configure = idxd_dev_configure,
+   .dev_start = idxd_pci_dev_start,
+   .dev_stop = idxd_pci_dev_stop,
 };
 
 /* each portal uses 4 x 4k pages */
-- 
2.25.1



[dpdk-dev] [PATCH v3 20/25] raw/ioat: add info function for idxd devices

2020-09-25 Thread Bruce Richardson
Add the info get function for DSA devices, returning just the ring size
info about the device, same as is returned for existing IOAT/CBDMA devices.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/idxd_pci.c |  1 +
 drivers/raw/ioat/idxd_vdev.c|  1 +
 drivers/raw/ioat/ioat_common.c  | 18 ++
 drivers/raw/ioat/ioat_private.h |  3 +++
 4 files changed, 23 insertions(+)

diff --git a/drivers/raw/ioat/idxd_pci.c b/drivers/raw/ioat/idxd_pci.c
index e4eecdc3a..475763e3c 100644
--- a/drivers/raw/ioat/idxd_pci.c
+++ b/drivers/raw/ioat/idxd_pci.c
@@ -106,6 +106,7 @@ static const struct rte_rawdev_ops idxd_pci_ops = {
.dev_configure = idxd_dev_configure,
.dev_start = idxd_pci_dev_start,
.dev_stop = idxd_pci_dev_stop,
+   .dev_info_get = idxd_dev_info_get,
 };
 
 /* each portal uses 4 x 4k pages */
diff --git a/drivers/raw/ioat/idxd_vdev.c b/drivers/raw/ioat/idxd_vdev.c
index aa6a5a9aa..5fbbd8e25 100644
--- a/drivers/raw/ioat/idxd_vdev.c
+++ b/drivers/raw/ioat/idxd_vdev.c
@@ -35,6 +35,7 @@ static const struct rte_rawdev_ops idxd_vdev_ops = {
.dev_selftest = idxd_rawdev_test,
.dump = idxd_dev_dump,
.dev_configure = idxd_dev_configure,
+   .dev_info_get = idxd_dev_info_get,
 };
 
 static void *
diff --git a/drivers/raw/ioat/ioat_common.c b/drivers/raw/ioat/ioat_common.c
index 6a4e2979f..b5cea2fda 100644
--- a/drivers/raw/ioat/ioat_common.c
+++ b/drivers/raw/ioat/ioat_common.c
@@ -44,6 +44,24 @@ idxd_dev_dump(struct rte_rawdev *dev, FILE *f)
return 0;
 }
 
+int
+idxd_dev_info_get(struct rte_rawdev *dev, rte_rawdev_obj_t dev_info,
+   size_t info_size)
+{
+   struct rte_ioat_rawdev_config *cfg = dev_info;
+   struct idxd_rawdev *idxd = dev->dev_private;
+   struct rte_idxd_rawdev *rte_idxd = &idxd->public;
+
+   if (info_size != sizeof(*cfg))
+   return -EINVAL;
+
+   if (cfg != NULL) {
+   cfg->ring_size = rte_idxd->hdl_ring_sz;
+   cfg->hdls_disable = rte_idxd->hdls_disable;
+   }
+   return 0;
+}
+
 int
 idxd_dev_configure(const struct rte_rawdev *dev,
rte_rawdev_obj_t config, size_t config_size)
diff --git a/drivers/raw/ioat/ioat_private.h b/drivers/raw/ioat/ioat_private.h
index aba70d8d7..0f80d60bf 100644
--- a/drivers/raw/ioat/ioat_private.h
+++ b/drivers/raw/ioat/ioat_private.h
@@ -62,6 +62,9 @@ extern int idxd_rawdev_close(struct rte_rawdev *dev);
 extern int idxd_dev_configure(const struct rte_rawdev *dev,
rte_rawdev_obj_t config, size_t config_size);
 
+extern int idxd_dev_info_get(struct rte_rawdev *dev, rte_rawdev_obj_t dev_info,
+   size_t info_size);
+
 extern int idxd_rawdev_test(uint16_t dev_id);
 
 extern int idxd_dev_dump(struct rte_rawdev *dev, FILE *f);
-- 
2.25.1



[dpdk-dev] [PATCH v3 21/25] raw/ioat: create separate statistics structure

2020-09-25 Thread Bruce Richardson
Rather than having the xstats as fields inside the main driver structure,
create a separate structure type for them.

As part of the change, when updating the stats functions referring to the
stats by the old path, we can simplify them to use the id to directly index
into the stats structure, making the code shorter and simpler.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/ioat_rawdev.c | 40 +++---
 drivers/raw/ioat/rte_ioat_rawdev_fns.h | 30 ---
 2 files changed, 29 insertions(+), 41 deletions(-)

diff --git a/drivers/raw/ioat/ioat_rawdev.c b/drivers/raw/ioat/ioat_rawdev.c
index 0097be87e..4ea913fff 100644
--- a/drivers/raw/ioat/ioat_rawdev.c
+++ b/drivers/raw/ioat/ioat_rawdev.c
@@ -132,16 +132,14 @@ ioat_xstats_get(const struct rte_rawdev *dev, const 
unsigned int ids[],
uint64_t values[], unsigned int n)
 {
const struct rte_ioat_rawdev *ioat = dev->dev_private;
+   const uint64_t *stats = (const void *)&ioat->xstats;
unsigned int i;
 
for (i = 0; i < n; i++) {
-   switch (ids[i]) {
-   case 0: values[i] = ioat->enqueue_failed; break;
-   case 1: values[i] = ioat->enqueued; break;
-   case 2: values[i] = ioat->started; break;
-   case 3: values[i] = ioat->completed; break;
-   default: values[i] = 0; break;
-   }
+   if (ids[i] < sizeof(ioat->xstats)/sizeof(*stats))
+   values[i] = stats[ids[i]];
+   else
+   values[i] = 0;
}
return n;
 }
@@ -167,35 +165,17 @@ static int
 ioat_xstats_reset(struct rte_rawdev *dev, const uint32_t *ids, uint32_t nb_ids)
 {
struct rte_ioat_rawdev *ioat = dev->dev_private;
+   uint64_t *stats = (void *)&ioat->xstats;
unsigned int i;
 
if (!ids) {
-   ioat->enqueue_failed = 0;
-   ioat->enqueued = 0;
-   ioat->started = 0;
-   ioat->completed = 0;
+   memset(&ioat->xstats, 0, sizeof(ioat->xstats));
return 0;
}
 
-   for (i = 0; i < nb_ids; i++) {
-   switch (ids[i]) {
-   case 0:
-   ioat->enqueue_failed = 0;
-   break;
-   case 1:
-   ioat->enqueued = 0;
-   break;
-   case 2:
-   ioat->started = 0;
-   break;
-   case 3:
-   ioat->completed = 0;
-   break;
-   default:
-   IOAT_PMD_WARN("Invalid xstat id - cannot reset value");
-   break;
-   }
-   }
+   for (i = 0; i < nb_ids; i++)
+   if (ids[i] < sizeof(ioat->xstats)/sizeof(*stats))
+   stats[ids[i]] = 0;
 
return 0;
 }
diff --git a/drivers/raw/ioat/rte_ioat_rawdev_fns.h 
b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
index 36ba876ea..89bfc8d21 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev_fns.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
@@ -49,17 +49,31 @@ enum rte_ioat_dev_type {
RTE_IDXD_DEV,
 };
 
+/**
+ * @internal
+ * some statistics for tracking, if added/changed update xstats fns
+ */
+struct rte_ioat_xstats {
+   uint64_t enqueue_failed;
+   uint64_t enqueued;
+   uint64_t started;
+   uint64_t completed;
+};
+
 /**
  * @internal
  * Structure representing an IOAT device instance
  */
 struct rte_ioat_rawdev {
+   /* common fields at the top - match those in rte_idxd_rawdev */
enum rte_ioat_dev_type type;
+   struct rte_ioat_xstats xstats;
+
struct rte_rawdev *rawdev;
const struct rte_memzone *mz;
const struct rte_memzone *desc_mz;
 
-   volatile uint16_t *doorbell;
+   volatile uint16_t *doorbell __rte_cache_aligned;
phys_addr_t status_addr;
phys_addr_t ring_addr;
 
@@ -72,12 +86,6 @@ struct rte_ioat_rawdev {
unsigned short next_read;
unsigned short next_write;
 
-   /* some statistics for tracking, if added/changed update xstats fns*/
-   uint64_t enqueue_failed __rte_cache_aligned;
-   uint64_t enqueued;
-   uint64_t started;
-   uint64_t completed;
-
/* to report completions, the device will write status back here */
volatile uint64_t status __rte_cache_aligned;
 
@@ -209,7 +217,7 @@ __ioat_enqueue_copy(int dev_id, phys_addr_t src, 
phys_addr_t dst,
struct rte_ioat_generic_hw_desc *desc;
 
if (space == 0) {
-   ioat->enqueue_failed++;
+   ioat->xstats.enqueue_failed++;
return 0;
}
 
@@ -228,7 +236,7 @@ __ioat_enqueue_copy(int dev_id, phys_addr_t src, 
phys_addr_t dst,
(int64_t)src_hdl);
rte_prefetch0(&ioat->desc_ring[ioat->next_write & mask]);

[dpdk-dev] [PATCH v3 22/25] raw/ioat: move xstats functions to common file

2020-09-25 Thread Bruce Richardson
The xstats functions can be used by all ioat devices so move them from the
ioat_rawdev.c file to ioat_common.c, and add the function prototypes to the
internal header file.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/ioat_common.c  | 59 +
 drivers/raw/ioat/ioat_private.h | 10 ++
 drivers/raw/ioat/ioat_rawdev.c  | 58 
 3 files changed, 69 insertions(+), 58 deletions(-)

diff --git a/drivers/raw/ioat/ioat_common.c b/drivers/raw/ioat/ioat_common.c
index b5cea2fda..142e171bc 100644
--- a/drivers/raw/ioat/ioat_common.c
+++ b/drivers/raw/ioat/ioat_common.c
@@ -5,9 +5,68 @@
 #include 
 #include 
 #include 
+#include 
 
 #include "ioat_private.h"
 
+static const char * const xstat_names[] = {
+   "failed_enqueues", "successful_enqueues",
+   "copies_started", "copies_completed"
+};
+
+int
+ioat_xstats_get(const struct rte_rawdev *dev, const unsigned int ids[],
+   uint64_t values[], unsigned int n)
+{
+   const struct rte_ioat_rawdev *ioat = dev->dev_private;
+   const uint64_t *stats = (const void *)&ioat->xstats;
+   unsigned int i;
+
+   for (i = 0; i < n; i++) {
+   if (ids[i] > sizeof(ioat->xstats)/sizeof(*stats))
+   values[i] = 0;
+   else
+   values[i] = stats[ids[i]];
+   }
+   return n;
+}
+
+int
+ioat_xstats_get_names(const struct rte_rawdev *dev,
+   struct rte_rawdev_xstats_name *names,
+   unsigned int size)
+{
+   unsigned int i;
+
+   RTE_SET_USED(dev);
+   if (size < RTE_DIM(xstat_names))
+   return RTE_DIM(xstat_names);
+
+   for (i = 0; i < RTE_DIM(xstat_names); i++)
+   strlcpy(names[i].name, xstat_names[i], sizeof(names[i]));
+
+   return RTE_DIM(xstat_names);
+}
+
+int
+ioat_xstats_reset(struct rte_rawdev *dev, const uint32_t *ids, uint32_t nb_ids)
+{
+   struct rte_ioat_rawdev *ioat = dev->dev_private;
+   uint64_t *stats = (void *)&ioat->xstats;
+   unsigned int i;
+
+   if (!ids) {
+   memset(&ioat->xstats, 0, sizeof(ioat->xstats));
+   return 0;
+   }
+
+   for (i = 0; i < nb_ids; i++)
+   if (ids[i] < sizeof(ioat->xstats)/sizeof(*stats))
+   stats[ids[i]] = 0;
+
+   return 0;
+}
+
 int
 idxd_rawdev_close(struct rte_rawdev *dev __rte_unused)
 {
diff --git a/drivers/raw/ioat/ioat_private.h b/drivers/raw/ioat/ioat_private.h
index 0f80d60bf..ab9a3e6cc 100644
--- a/drivers/raw/ioat/ioat_private.h
+++ b/drivers/raw/ioat/ioat_private.h
@@ -53,6 +53,16 @@ struct idxd_rawdev {
} u;
 };
 
+int ioat_xstats_get(const struct rte_rawdev *dev, const unsigned int ids[],
+   uint64_t values[], unsigned int n);
+
+int ioat_xstats_get_names(const struct rte_rawdev *dev,
+   struct rte_rawdev_xstats_name *names,
+   unsigned int size);
+
+int ioat_xstats_reset(struct rte_rawdev *dev, const uint32_t *ids,
+   uint32_t nb_ids);
+
 extern int idxd_rawdev_create(const char *name, struct rte_device *dev,
   const struct idxd_rawdev *idxd,
   const struct rte_rawdev_ops *ops);
diff --git a/drivers/raw/ioat/ioat_rawdev.c b/drivers/raw/ioat/ioat_rawdev.c
index 4ea913fff..dd2543c80 100644
--- a/drivers/raw/ioat/ioat_rawdev.c
+++ b/drivers/raw/ioat/ioat_rawdev.c
@@ -122,64 +122,6 @@ ioat_dev_info_get(struct rte_rawdev *dev, rte_rawdev_obj_t 
dev_info,
return 0;
 }
 
-static const char * const xstat_names[] = {
-   "failed_enqueues", "successful_enqueues",
-   "copies_started", "copies_completed"
-};
-
-static int
-ioat_xstats_get(const struct rte_rawdev *dev, const unsigned int ids[],
-   uint64_t values[], unsigned int n)
-{
-   const struct rte_ioat_rawdev *ioat = dev->dev_private;
-   const uint64_t *stats = (const void *)&ioat->xstats;
-   unsigned int i;
-
-   for (i = 0; i < n; i++) {
-   if (ids[i] < sizeof(ioat->xstats)/sizeof(*stats))
-   values[i] = stats[ids[i]];
-   else
-   values[i] = 0;
-   }
-   return n;
-}
-
-static int
-ioat_xstats_get_names(const struct rte_rawdev *dev,
-   struct rte_rawdev_xstats_name *names,
-   unsigned int size)
-{
-   unsigned int i;
-
-   RTE_SET_USED(dev);
-   if (size < RTE_DIM(xstat_names))
-   return RTE_DIM(xstat_names);
-
-   for (i = 0; i < RTE_DIM(xstat_names); i++)
-   strlcpy(names[i].name, xstat_names[i], sizeof(names[i]));
-
-   return RTE_DIM(xstat_names);
-}
-
-static int
-ioat_xstats_reset(struct rte_rawdev *dev, const uint32_t *ids, uint32_t nb_ids)
-{
-   struct rte_ioat_rawdev *ioat = dev->dev_private;
-   uint64_t *stats = (void *)&ioat->xstats;
-   unsigned int i;
-
-   if 

[dpdk-dev] [PATCH v3 19/25] raw/ioat: add data path for idxd devices

2020-09-25 Thread Bruce Richardson
Add support for doing copies using DSA hardware. This is implemented by
just switching on the device type field at the start of the inline
functions. Since there is no hardware which will have both device types
present this branch will always be predictable after the first call,
meaning it has little to no perf penalty.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/ioat_common.c |   1 +
 drivers/raw/ioat/ioat_rawdev.c |   1 +
 drivers/raw/ioat/rte_ioat_rawdev_fns.h | 201 +++--
 3 files changed, 192 insertions(+), 11 deletions(-)

diff --git a/drivers/raw/ioat/ioat_common.c b/drivers/raw/ioat/ioat_common.c
index 5173c331c..6a4e2979f 100644
--- a/drivers/raw/ioat/ioat_common.c
+++ b/drivers/raw/ioat/ioat_common.c
@@ -153,6 +153,7 @@ idxd_rawdev_create(const char *name, struct rte_device *dev,
 
idxd = rawdev->dev_private;
*idxd = *base_idxd; /* copy over the main fields already passed in */
+   idxd->public.type = RTE_IDXD_DEV;
idxd->rawdev = rawdev;
idxd->mz = mz;
 
diff --git a/drivers/raw/ioat/ioat_rawdev.c b/drivers/raw/ioat/ioat_rawdev.c
index 1fe32278d..0097be87e 100644
--- a/drivers/raw/ioat/ioat_rawdev.c
+++ b/drivers/raw/ioat/ioat_rawdev.c
@@ -260,6 +260,7 @@ ioat_rawdev_create(const char *name, struct rte_pci_device 
*dev)
rawdev->driver_name = dev->device.driver->name;
 
ioat = rawdev->dev_private;
+   ioat->type = RTE_IOAT_DEV;
ioat->rawdev = rawdev;
ioat->mz = mz;
ioat->regs = dev->mem_resource[0].addr;
diff --git a/drivers/raw/ioat/rte_ioat_rawdev_fns.h 
b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
index e9cdce016..36ba876ea 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev_fns.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
@@ -196,8 +196,8 @@ struct rte_idxd_rawdev {
 /*
  * Enqueue a copy operation onto the ioat device
  */
-static inline int
-rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, phys_addr_t dst,
+static __rte_always_inline int
+__ioat_enqueue_copy(int dev_id, phys_addr_t src, phys_addr_t dst,
unsigned int length, uintptr_t src_hdl, uintptr_t dst_hdl)
 {
struct rte_ioat_rawdev *ioat =
@@ -233,8 +233,8 @@ rte_ioat_enqueue_copy(int dev_id, phys_addr_t src, 
phys_addr_t dst,
 }
 
 /* add fence to last written descriptor */
-static inline int
-rte_ioat_fence(int dev_id)
+static __rte_always_inline int
+__ioat_fence(int dev_id)
 {
struct rte_ioat_rawdev *ioat =
(struct rte_ioat_rawdev 
*)rte_rawdevs[dev_id].dev_private;
@@ -252,8 +252,8 @@ rte_ioat_fence(int dev_id)
 /*
  * Trigger hardware to begin performing enqueued operations
  */
-static inline void
-rte_ioat_perform_ops(int dev_id)
+static __rte_always_inline void
+__ioat_perform_ops(int dev_id)
 {
struct rte_ioat_rawdev *ioat =
(struct rte_ioat_rawdev 
*)rte_rawdevs[dev_id].dev_private;
@@ -268,8 +268,8 @@ rte_ioat_perform_ops(int dev_id)
  * @internal
  * Returns the index of the last completed operation.
  */
-static inline int
-rte_ioat_get_last_completed(struct rte_ioat_rawdev *ioat, int *error)
+static __rte_always_inline int
+__ioat_get_last_completed(struct rte_ioat_rawdev *ioat, int *error)
 {
uint64_t status = ioat->status;
 
@@ -283,8 +283,8 @@ rte_ioat_get_last_completed(struct rte_ioat_rawdev *ioat, 
int *error)
 /*
  * Returns details of operations that have been completed
  */
-static inline int
-rte_ioat_completed_ops(int dev_id, uint8_t max_copies,
+static __rte_always_inline int
+__ioat_completed_ops(int dev_id, uint8_t max_copies,
uintptr_t *src_hdls, uintptr_t *dst_hdls)
 {
struct rte_ioat_rawdev *ioat =
@@ -295,7 +295,7 @@ rte_ioat_completed_ops(int dev_id, uint8_t max_copies,
int error;
int i = 0;
 
-   end_read = (rte_ioat_get_last_completed(ioat, &error) + 1) & mask;
+   end_read = (__ioat_get_last_completed(ioat, &error) + 1) & mask;
count = (end_read - (read & mask)) & mask;
 
if (error) {
@@ -332,6 +332,185 @@ rte_ioat_completed_ops(int dev_id, uint8_t max_copies,
return count;
 }
 
+static __rte_always_inline int
+__idxd_write_desc(int dev_id, const struct rte_idxd_hw_desc *desc,
+   const struct rte_idxd_user_hdl *hdl)
+{
+   struct rte_idxd_rawdev *idxd =
+   (struct rte_idxd_rawdev 
*)rte_rawdevs[dev_id].dev_private;
+   struct rte_idxd_desc_batch *b = &idxd->batch_ring[idxd->next_batch];
+
+   /* check for room in the handle ring */
+   if (((idxd->next_free_hdl + 1) & (idxd->hdl_ring_sz - 1)) == 
idxd->next_ret_hdl)
+   goto failed;
+
+   /* check for space in current batch */
+   if (b->op_count >= BATCH_SIZE)
+   goto failed;
+
+   /* check that we can actually use the current batch */
+   if (b->submitted)
+   goto failed;
+
+   /* write the descriptor */
+   b->ops[b->op_count++]

[dpdk-dev] [PATCH v3 23/25] raw/ioat: add xstats tracking for idxd devices

2020-09-25 Thread Bruce Richardson
Add update of the relevant stats for the data path functions and point the
overall device struct xstats function pointers to the existing ioat
functions.

At this point, all necessary hooks for supporting the existing unit tests
are in place so call them for each device.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/idxd_pci.c| 3 +++
 drivers/raw/ioat/idxd_vdev.c   | 3 +++
 drivers/raw/ioat/ioat_rawdev_test.c| 2 +-
 drivers/raw/ioat/rte_ioat_rawdev_fns.h | 6 ++
 4 files changed, 13 insertions(+), 1 deletion(-)

diff --git a/drivers/raw/ioat/idxd_pci.c b/drivers/raw/ioat/idxd_pci.c
index 475763e3c..5dfa4ffab 100644
--- a/drivers/raw/ioat/idxd_pci.c
+++ b/drivers/raw/ioat/idxd_pci.c
@@ -107,6 +107,9 @@ static const struct rte_rawdev_ops idxd_pci_ops = {
.dev_start = idxd_pci_dev_start,
.dev_stop = idxd_pci_dev_stop,
.dev_info_get = idxd_dev_info_get,
+   .xstats_get = ioat_xstats_get,
+   .xstats_get_names = ioat_xstats_get_names,
+   .xstats_reset = ioat_xstats_reset,
 };
 
 /* each portal uses 4 x 4k pages */
diff --git a/drivers/raw/ioat/idxd_vdev.c b/drivers/raw/ioat/idxd_vdev.c
index 5fbbd8e25..3a5cc94b4 100644
--- a/drivers/raw/ioat/idxd_vdev.c
+++ b/drivers/raw/ioat/idxd_vdev.c
@@ -36,6 +36,9 @@ static const struct rte_rawdev_ops idxd_vdev_ops = {
.dump = idxd_dev_dump,
.dev_configure = idxd_dev_configure,
.dev_info_get = idxd_dev_info_get,
+   .xstats_get = ioat_xstats_get,
+   .xstats_get_names = ioat_xstats_get_names,
+   .xstats_reset = ioat_xstats_reset,
 };
 
 static void *
diff --git a/drivers/raw/ioat/ioat_rawdev_test.c 
b/drivers/raw/ioat/ioat_rawdev_test.c
index a9132a8f1..0b172f318 100644
--- a/drivers/raw/ioat/ioat_rawdev_test.c
+++ b/drivers/raw/ioat/ioat_rawdev_test.c
@@ -264,5 +264,5 @@ int
 idxd_rawdev_test(uint16_t dev_id)
 {
rte_rawdev_dump(dev_id, stdout);
-   return 0;
+   return ioat_rawdev_test(dev_id);
 }
diff --git a/drivers/raw/ioat/rte_ioat_rawdev_fns.h 
b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
index 89bfc8d21..d0045d8a4 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev_fns.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev_fns.h
@@ -184,6 +184,8 @@ struct rte_idxd_user_hdl {
  */
 struct rte_idxd_rawdev {
enum rte_ioat_dev_type type;
+   struct rte_ioat_xstats xstats;
+
void *portal; /* address to write the batch descriptor */
 
/* counters to track the batches and the individual op handles */
@@ -369,9 +371,11 @@ __idxd_write_desc(int dev_id, const struct 
rte_idxd_hw_desc *desc,
if (++idxd->next_free_hdl == idxd->hdl_ring_sz)
idxd->next_free_hdl = 0;
 
+   idxd->xstats.enqueued++;
return 1;
 
 failed:
+   idxd->xstats.enqueue_failed++;
rte_errno = ENOSPC;
return 0;
 }
@@ -429,6 +433,7 @@ __idxd_perform_ops(int dev_id)
 
if (++idxd->next_batch == idxd->batch_ring_sz)
idxd->next_batch = 0;
+   idxd->xstats.started = idxd->xstats.enqueued;
 }
 
 static __rte_always_inline int
@@ -466,6 +471,7 @@ __idxd_completed_ops(int dev_id, uint8_t max_ops,
 
idxd->next_ret_hdl = h_idx;
 
+   idxd->xstats.completed += n;
return n;
 }
 
-- 
2.25.1



[dpdk-dev] [PATCH v3 25/25] raw/ioat: add fill operation

2020-09-25 Thread Bruce Richardson
From: Kevin Laatz 

Add fill operation enqueue support for IOAT and IDXD. The fill enqueue is
similar to the copy enqueue, but takes a 'pattern' rather than a source
address to transfer to the destination address. This patch also includes an
additional test case for the new operation type.

Signed-off-by: Kevin Laatz 
Reviewed-by: Bruce Richardson 
---
 doc/guides/rawdevs/ioat.rst| 10 
 doc/guides/rel_notes/release_20_11.rst |  2 +
 drivers/raw/ioat/ioat_rawdev_test.c| 44 +
 drivers/raw/ioat/rte_ioat_rawdev.h | 26 +++
 drivers/raw/ioat/rte_ioat_rawdev_fns.h | 65 --
 5 files changed, 142 insertions(+), 5 deletions(-)

diff --git a/doc/guides/rawdevs/ioat.rst b/doc/guides/rawdevs/ioat.rst
index 7c2a2d457..250cfc48a 100644
--- a/doc/guides/rawdevs/ioat.rst
+++ b/doc/guides/rawdevs/ioat.rst
@@ -285,6 +285,16 @@ is correct before freeing the data buffers using the 
returned handles:
 }
 
 
+Filling an Area of Memory
+~~
+
+The IOAT driver also has support for the ``fill`` operation, where an area
+of memory is overwritten, or filled, with a short pattern of data.
+Fill operations can be performed in much the same was as copy operations
+described above, just using the ``rte_ioat_enqueue_fill()`` function rather
+than the ``rte_ioat_enqueue_copy()`` function.
+
+
 Querying Device Statistics
 ~~~
 
diff --git a/doc/guides/rel_notes/release_20_11.rst 
b/doc/guides/rel_notes/release_20_11.rst
index 4d8b78154..dd65b779d 100644
--- a/doc/guides/rel_notes/release_20_11.rst
+++ b/doc/guides/rel_notes/release_20_11.rst
@@ -84,6 +84,8 @@ New Features
 
   * Added support for Intel\ |reg| Data Streaming Accelerator hardware.
 For more information, see 
https://01.org/blogs/2019/introducing-intel-data-streaming-accelerator
+  * Added support for the fill operation via the API 
``rte_ioat_enqueue_fill()``,
+where the hardware fills an area of memory with a repeating pattern.
   * Added a per-device configuration flag to disable management of 
user-provided completion handles
   * Renamed the ``rte_ioat_do_copies()`` API to ``rte_ioat_perform_ops()``,
 and renamed the ``rte_ioat_completed_copies()`` API to 
``rte_ioat_completed_ops()``
diff --git a/drivers/raw/ioat/ioat_rawdev_test.c 
b/drivers/raw/ioat/ioat_rawdev_test.c
index 7be6f2a2d..64269af55 100644
--- a/drivers/raw/ioat/ioat_rawdev_test.c
+++ b/drivers/raw/ioat/ioat_rawdev_test.c
@@ -152,6 +152,46 @@ test_enqueue_copies(int dev_id)
return 0;
 }
 
+static int
+test_enqueue_fill(int dev_id)
+{
+   const unsigned int length[] = {8, 64, 1024, 50, 100, 89};
+   struct rte_mbuf *dst = rte_pktmbuf_alloc(pool);
+   char *dst_data = rte_pktmbuf_mtod(dst, char *);
+   uint64_t pattern = 0xfedcba9876543210;
+   unsigned int i, j;
+
+   for (i = 0; i < RTE_DIM(length); i++) {
+   /* reset dst_data */
+   memset(dst_data, 0, length[i]);
+
+   /* perform the fill operation */
+   if (rte_ioat_enqueue_fill(dev_id, pattern,
+   dst->buf_iova + dst->data_off, length[i],
+   (uintptr_t)dst) != 1) {
+   PRINT_ERR("Error with rte_ioat_enqueue_fill\n");
+   return -1;
+   }
+
+   rte_ioat_perform_ops(dev_id);
+   usleep(100);
+
+   /* check the result */
+   for (j = 0; j < length[i]; j++) {
+   char pat_byte = ((char *)&pattern)[j % 8];
+   if (dst_data[j] != pat_byte) {
+   PRINT_ERR("Error with fill operation (length = 
%u): got (%x), not (%x)\n",
+   length[i], dst_data[j],
+   pat_byte);
+   return -1;
+   }
+   }
+   }
+
+   rte_pktmbuf_free(dst);
+   return 0;
+}
+
 int
 ioat_rawdev_test(uint16_t dev_id)
 {
@@ -238,6 +278,10 @@ ioat_rawdev_test(uint16_t dev_id)
}
printf("\n");
 
+   /* test enqueue fill operation */
+   if (test_enqueue_fill(dev_id) != 0)
+   goto err;
+
rte_rawdev_stop(dev_id);
if (rte_rawdev_xstats_reset(dev_id, NULL, 0) != 0) {
PRINT_ERR("Error resetting xstat values\n");
diff --git a/drivers/raw/ioat/rte_ioat_rawdev.h 
b/drivers/raw/ioat/rte_ioat_rawdev.h
index 6b891cd44..b7632ebf3 100644
--- a/drivers/raw/ioat/rte_ioat_rawdev.h
+++ b/drivers/raw/ioat/rte_ioat_rawdev.h
@@ -37,6 +37,32 @@ struct rte_ioat_rawdev_config {
bool hdls_disable;/**< if set, ignore user-supplied handle params */
 };
 
+/**
+ * Enqueue a fill operation onto the ioat device
+ *
+ * This queues up a fill operation to be performed by hardware, but does not
+ * trigger hardware to begin that operation.
+ *
+ * @param de

[dpdk-dev] [PATCH v3 24/25] raw/ioat: clean up use of common test function

2020-09-25 Thread Bruce Richardson
Now that all devices can pass the same set of unit tests, eliminate the
temporary idxd_rawdev_test function and move the prototype for
ioat_rawdev_test to the proper internal header file, to be used by all
device instances.

Signed-off-by: Bruce Richardson 
Reviewed-by: Kevin Laatz 
---
 drivers/raw/ioat/idxd_pci.c | 2 +-
 drivers/raw/ioat/idxd_vdev.c| 2 +-
 drivers/raw/ioat/ioat_private.h | 4 ++--
 drivers/raw/ioat/ioat_rawdev.c  | 2 --
 drivers/raw/ioat/ioat_rawdev_test.c | 7 ---
 5 files changed, 4 insertions(+), 13 deletions(-)

diff --git a/drivers/raw/ioat/idxd_pci.c b/drivers/raw/ioat/idxd_pci.c
index 5dfa4ffab..c0da1f977 100644
--- a/drivers/raw/ioat/idxd_pci.c
+++ b/drivers/raw/ioat/idxd_pci.c
@@ -101,7 +101,7 @@ idxd_pci_dev_start(struct rte_rawdev *dev)
 
 static const struct rte_rawdev_ops idxd_pci_ops = {
.dev_close = idxd_rawdev_close,
-   .dev_selftest = idxd_rawdev_test,
+   .dev_selftest = ioat_rawdev_test,
.dump = idxd_dev_dump,
.dev_configure = idxd_dev_configure,
.dev_start = idxd_pci_dev_start,
diff --git a/drivers/raw/ioat/idxd_vdev.c b/drivers/raw/ioat/idxd_vdev.c
index 3a5cc94b4..989d9e45f 100644
--- a/drivers/raw/ioat/idxd_vdev.c
+++ b/drivers/raw/ioat/idxd_vdev.c
@@ -32,7 +32,7 @@ struct idxd_vdev_args {
 
 static const struct rte_rawdev_ops idxd_vdev_ops = {
.dev_close = idxd_rawdev_close,
-   .dev_selftest = idxd_rawdev_test,
+   .dev_selftest = ioat_rawdev_test,
.dump = idxd_dev_dump,
.dev_configure = idxd_dev_configure,
.dev_info_get = idxd_dev_info_get,
diff --git a/drivers/raw/ioat/ioat_private.h b/drivers/raw/ioat/ioat_private.h
index ab9a3e6cc..a74bc0422 100644
--- a/drivers/raw/ioat/ioat_private.h
+++ b/drivers/raw/ioat/ioat_private.h
@@ -63,6 +63,8 @@ int ioat_xstats_get_names(const struct rte_rawdev *dev,
 int ioat_xstats_reset(struct rte_rawdev *dev, const uint32_t *ids,
uint32_t nb_ids);
 
+extern int ioat_rawdev_test(uint16_t dev_id);
+
 extern int idxd_rawdev_create(const char *name, struct rte_device *dev,
   const struct idxd_rawdev *idxd,
   const struct rte_rawdev_ops *ops);
@@ -75,8 +77,6 @@ extern int idxd_dev_configure(const struct rte_rawdev *dev,
 extern int idxd_dev_info_get(struct rte_rawdev *dev, rte_rawdev_obj_t dev_info,
size_t info_size);
 
-extern int idxd_rawdev_test(uint16_t dev_id);
-
 extern int idxd_dev_dump(struct rte_rawdev *dev, FILE *f);
 
 #endif /* _IOAT_PRIVATE_H_ */
diff --git a/drivers/raw/ioat/ioat_rawdev.c b/drivers/raw/ioat/ioat_rawdev.c
index dd2543c80..2c88b4369 100644
--- a/drivers/raw/ioat/ioat_rawdev.c
+++ b/drivers/raw/ioat/ioat_rawdev.c
@@ -128,8 +128,6 @@ ioat_dev_close(struct rte_rawdev *dev __rte_unused)
return 0;
 }
 
-extern int ioat_rawdev_test(uint16_t dev_id);
-
 static int
 ioat_rawdev_create(const char *name, struct rte_pci_device *dev)
 {
diff --git a/drivers/raw/ioat/ioat_rawdev_test.c 
b/drivers/raw/ioat/ioat_rawdev_test.c
index 0b172f318..7be6f2a2d 100644
--- a/drivers/raw/ioat/ioat_rawdev_test.c
+++ b/drivers/raw/ioat/ioat_rawdev_test.c
@@ -259,10 +259,3 @@ ioat_rawdev_test(uint16_t dev_id)
free(ids);
return -1;
 }
-
-int
-idxd_rawdev_test(uint16_t dev_id)
-{
-   rte_rawdev_dump(dev_id, stdout);
-   return ioat_rawdev_test(dev_id);
-}
-- 
2.25.1



Re: [dpdk-dev] [PATCH v3 02/25] raw/ioat: fix missing close function

2020-09-25 Thread Bruce Richardson
On Fri, Sep 25, 2020 at 12:08:47PM +0100, Bruce Richardson wrote:
> From: Kevin Laatz 
> 
> When rte_rawdev_pmd_release() is called, rte_rawdev_close() looks for a
> dev_close function for the device causing a segmentation fault when no
> close() function is implemented for a driver.
> 
> This patch resolves the issue by adding a stub function ioat_dev_close().
> 
> Fixes: f687e842e328 ("raw/ioat: introduce IOAT driver")
> Cc: sta...@dpdk.org
> 
> Reported-by: Sunil Pai G 
> Signed-off-by: Kevin Laatz 
> ---
Forgot to add my own reviewed-by to this:

Reviewed-by: Bruce Richardson 


Re: [dpdk-dev] [PATCH v3 02/25] raw/ioat: fix missing close function

2020-09-25 Thread Pai G, Sunil
> -Original Message-
> From: Richardson, Bruce 
> Sent: Friday, September 25, 2020 4:39 PM
> To: dev@dpdk.org
> Cc: Fu, Patrick ; Laatz, Kevin ;
> sta...@dpdk.org; Pai G, Sunil 
> Subject: [PATCH v3 02/25] raw/ioat: fix missing close function
> 
> From: Kevin Laatz 
> 
> When rte_rawdev_pmd_release() is called, rte_rawdev_close() looks for a
> dev_close function for the device causing a segmentation fault when no
> close() function is implemented for a driver.
> 
> This patch resolves the issue by adding a stub function ioat_dev_close().
> 
> Fixes: f687e842e328 ("raw/ioat: introduce IOAT driver")
> Cc: sta...@dpdk.org
> 
> Reported-by: Sunil Pai G 
> Signed-off-by: Kevin Laatz 
> ---
>  drivers/raw/ioat/ioat_rawdev.c | 7 +++
>  1 file changed, 7 insertions(+)
> 
> diff --git a/drivers/raw/ioat/ioat_rawdev.c b/drivers/raw/ioat/ioat_rawdev.c
> index 7f1a15436..0732b059f 100644
> --- a/drivers/raw/ioat/ioat_rawdev.c
> +++ b/drivers/raw/ioat/ioat_rawdev.c
> @@ -203,6 +203,12 @@ ioat_xstats_reset(struct rte_rawdev *dev, const
> uint32_t *ids, uint32_t nb_ids)
>   return 0;
>  }
> 
> +static int
> +ioat_dev_close(struct rte_rawdev *dev __rte_unused) {
> + return 0;
> +}
> +
>  extern int ioat_rawdev_test(uint16_t dev_id);
> 
>  static int
> @@ -212,6 +218,7 @@ ioat_rawdev_create(const char *name, struct
> rte_pci_device *dev)
>   .dev_configure = ioat_dev_configure,
>   .dev_start = ioat_dev_start,
>   .dev_stop = ioat_dev_stop,
> + .dev_close = ioat_dev_close,
>   .dev_info_get = ioat_dev_info_get,
>   .xstats_get = ioat_xstats_get,
>   .xstats_get_names = ioat_xstats_get_names,
> --
> 2.25.1
Acked-by: Sunil Pai G 


[dpdk-dev] [PATCH] examples/vhost_blk: fix unchecked return value

2020-09-25 Thread wangyunjian
From: Yunjian Wang 

This checks the return value from the function
rte_vhost_driver_start.

Coverity issue: 362027
Fixes: c19beb3f38cd ("examples/vhost_blk: introduce vhost storage sample")
Cc: sta...@dpdk.org

Signed-off-by: Yunjian Wang 
---
 examples/vhost_blk/vhost_blk.c | 6 +-
 1 file changed, 5 insertions(+), 1 deletion(-)

diff --git a/examples/vhost_blk/vhost_blk.c b/examples/vhost_blk/vhost_blk.c
index f4c59437a..8f5d61a58 100644
--- a/examples/vhost_blk/vhost_blk.c
+++ b/examples/vhost_blk/vhost_blk.c
@@ -877,7 +877,11 @@ int main(int argc, char *argv[])
 
signal(SIGINT, signal_handler);
 
-   rte_vhost_driver_start(dev_pathname);
+   ret = rte_vhost_driver_start(dev_pathname);
+   if (ret < 0) {
+   fprintf(stderr, "Failed to start vhost driver.\n");
+   return -1;
+   }
 
/* loop for exit the application */
while (1)
-- 
2.23.0



Re: [dpdk-dev] [PATCH] bus/pci: fix mapping BAR containing MSI-X table

2020-09-25 Thread Burakov, Anatoly

On 25-Sep-20 3:14 AM, Hyong Youb Kim wrote:

When the BAR contains MSI-X table, pci_vfio_mmap_bar() tries to skip
the table and map the rest. "map around it" is the phrase used in the
source. The function splits the BAR into two regions: the region
before the table (first part or memreg[0]) and the region after the
table (second part or memreg[1]).

For hardware that has MSI-X vector table offset 0, the first part does
not exist (memreg[0].size == 0).

   Capabilities: [60] MSI-X: Enable- Count=48 Masked-
  Vector table: BAR=2 offset=
  PBA: BAR=2 offset=1000

The mapping part of the function maps the first part, if it
exists. Then, it maps the second part, if it exists and "if mapping the
first part succeeded".

The recent change that replaces MAP_FAILED with NULL breaks the "if
mapping the first part succeeded" condition (1) in the snippet below.

 void *map_addr = NULL;
 if (memreg[0].size) {
/* actual map of first part */
map_addr = pci_map_resource(...);
 }

 /* if there's a second part, try to map it */
 if (map_addr != NULL  // -- (1)
&& memreg[1].offset && memreg[1].size) {
[...]
 }

 if (map_addr == NULL) {
 RTE_LOG(ERR, EAL, "Failed to map pci BAR%d\n",
 bar_index);
 return -1;
 }

When the first part does not exist, (1) sees map_addr is still NULL,
and the function fails. This behavior is a regression and fails
probing hardware with vector table offset 0.

Previously, (1) was "map_addr != MAP_FAILED", which meant
pci_map_resource() was actually attempted and failed. So, expand (1)
to check if the first part exists as well, to match the semantics of
MAP_FAILED.

Bugzilla ID: 539
Fixes: e200535c1ca3 ("mem: drop mapping API workaround")

Signed-off-by: Hyong Youb Kim 
---


LGTM

Acked-by: Anatoly Burakov 

--
Thanks,
Anatoly


[dpdk-dev] [PATCH v6 0/6] qede: SR-IOV PF driver support

2020-09-25 Thread Manish Chopra
Hi,

This series adds SR-IOV PF pmd driver support to have VF pmd
driver work over PF pmd driver instances in order to run the
adapter completely under DPDK environment for one of the use
cases like ovs-dpdk.

This is very initial bring-up with following testing covered -

* Enable/Disable SR-IOV VFs through igb_uio sysfs hook.
* Load VFs, run fastpath, teardown VFs in hypervisor and guest VM.
* VF FLR flow (in case of VF PCI passthrough to the guest VM)
* Bulletin mechanism tested to communicate link changes to the VFs.

Note that this series is intended for upcoming DPDK release (20.11)
Please consider applying this series to dpdk-next-net-mrvl.git

V5->V6:
===

* Re-based it on latest dpdk-next-net-mrvl.git
  (driver/net/qede/Makefile change removed from patch #3, file no longer exist)
* Added Reviewed-by tag and renamed rte_pci_find_next_ext_capability to
  rte_pci_find_ext_capability as per comment from Nithin Kumar Dabilpuram

V4->V5:
===

* drivers/raw/ifpga was escaped to compile, test-robot reported
  below build failure which is fixed within patch #1 in this series.

'drivers/drivers@@tmp_rte_rawdev_ifpga@sta/raw_ifpga_ifpga_rawdev.c.o' -c 
../drivers/raw/ifpga/ifpga_rawdev.c
../drivers/raw/ifpga/ifpga_rawdev.c: In function 
‘ifpga_pci_find_next_ext_capability’:
../drivers/raw/ifpga/ifpga_rawdev.c:166:34: error: comparison between signed 
and unsigned integer expressions [-Werror=sign-compare]
   if (RTE_PCI_EXT_CAP_ID(header) == cap && pos != start)
  ^~
cc1: all warnings being treated as errors

V3->V4:
===

* Fixed meson/ninja build failure for AMD arch.
* Incorporated cosmetic changes pointed by Gaetan.

V2->V3:
===

* Instead of adding complete pci_regs.h file from linux, define handful
  PCI symbols (prefixed with RTE_) in dpdk which are required to implement
  generic rte_pci_find_next_ext_capability() API (patch #1) and to enable
  various other qede driver's OSAL_* defines (patch #2)
  (essentially, removed the patch #1 in v2)

* Fixed meson/ninja compilation issues reported

* Incorporated comments by Gaetan for rte_pci_find_next_ext_capability()

V1->V2: (Incorporated comments from Jerin Jacob)


* Added rte_pci_regs.h file (copy of linux/pci_regs.h) under
  lib/librte_pci to remove the dependency of dpdk on user headers

* Added generic API to find PCI extended capability and use
  that in the drivers, removed individual functions implemented
  by the drivers

Thanks,
Manish

Manish Chopra (6):
  drivers: add generic API to find PCI extended cap
  net/qede: define PCI config space specific osals
  net/qede: configure VFs on hardware
  net/qede: add infrastructure support for VF load
  net/qede: initialize VF MAC and link
  net/qede: add VF FLR support

 doc/guides/nics/features/qede.ini  |   1 +
 doc/guides/nics/qede.rst   |   7 +-
 drivers/bus/pci/pci_common.c   |  43 
 drivers/bus/pci/rte_bus_pci.h  |  19 ++
 drivers/bus/pci/rte_bus_pci_version.map|   6 +
 drivers/net/ice/ice_ethdev.c   |  51 +
 drivers/net/nfp/nfpcore/nfp_cpp_pcie_ops.c |  48 +
 drivers/net/qede/base/bcm_osal.c   |  31 +++
 drivers/net/qede/base/bcm_osal.h   |  29 ++-
 drivers/net/qede/base/ecore.h  |   7 +
 drivers/net/qede/base/ecore_dev.c  |   6 +-
 drivers/net/qede/base/ecore_iov_api.h  |   3 +
 drivers/net/qede/base/ecore_sriov.c|  28 +--
 drivers/net/qede/base/meson.build  |   2 +-
 drivers/net/qede/meson.build   |   1 +
 drivers/net/qede/qede_ethdev.c |  37 +++-
 drivers/net/qede/qede_ethdev.h |   1 +
 drivers/net/qede/qede_if.h |   1 +
 drivers/net/qede/qede_main.c   |  12 +-
 drivers/net/qede/qede_sriov.c  | 219 +
 drivers/net/qede/qede_sriov.h  |  22 +++
 drivers/raw/ifpga/ifpga_rawdev.c   |  17 +-
 lib/librte_pci/rte_pci.h   |  35 
 23 files changed, 485 insertions(+), 141 deletions(-)
 create mode 100644 drivers/net/qede/qede_sriov.c
 create mode 100644 drivers/net/qede/qede_sriov.h

-- 
2.17.1



[dpdk-dev] [PATCH v6 3/6] net/qede: configure VFs on hardware

2020-09-25 Thread Manish Chopra
Based on number of VFs enabled at PCI, PF-PMD driver instance
enables/configures those VFs from hardware perspective, such
that in later patches they could get required HW access to
communicate with PFs for slowpath configuration and run the
fastpath themsleves.

This patch also add two new qede IOV files [qede_sriov(.c|.h)]
under qede directory to add non-base driver IOV APIs/contents there.

Signed-off-by: Manish Chopra 
Signed-off-by: Igor Russkikh 
Signed-off-by: Rasesh Mody 
---
 drivers/net/qede/meson.build   |  1 +
 drivers/net/qede/qede_ethdev.c |  1 +
 drivers/net/qede/qede_ethdev.h |  1 +
 drivers/net/qede/qede_if.h |  1 +
 drivers/net/qede/qede_main.c   |  1 +
 drivers/net/qede/qede_sriov.c  | 85 ++
 drivers/net/qede/qede_sriov.h  |  9 
 7 files changed, 99 insertions(+)
 create mode 100644 drivers/net/qede/qede_sriov.c
 create mode 100644 drivers/net/qede/qede_sriov.h

diff --git a/drivers/net/qede/meson.build b/drivers/net/qede/meson.build
index 05c9bff73..ff0ac0b03 100644
--- a/drivers/net/qede/meson.build
+++ b/drivers/net/qede/meson.build
@@ -11,6 +11,7 @@ sources = files(
'qede_rxtx.c',
'qede_debug.c',
'qede_regs.c',
+   'qede_sriov.c',
 )
 
 if cc.has_argument('-Wno-format-nonliteral')
diff --git a/drivers/net/qede/qede_ethdev.c b/drivers/net/qede/qede_ethdev.c
index 59f1746ee..27ec42e2c 100644
--- a/drivers/net/qede/qede_ethdev.c
+++ b/drivers/net/qede/qede_ethdev.c
@@ -2699,6 +2699,7 @@ static int qede_common_dev_init(struct rte_eth_dev 
*eth_dev, bool is_vf)
adapter->vxlan.enable = false;
adapter->geneve.enable = false;
adapter->ipgre.enable = false;
+   qed_ops->sriov_configure(edev, pci_dev->max_vfs);
}
 
DP_INFO(edev, "MAC address : %02x:%02x:%02x:%02x:%02x:%02x\n",
diff --git a/drivers/net/qede/qede_ethdev.h b/drivers/net/qede/qede_ethdev.h
index 76c5dae3b..4fb77b05c 100644
--- a/drivers/net/qede/qede_ethdev.h
+++ b/drivers/net/qede/qede_ethdev.h
@@ -34,6 +34,7 @@
 #include "base/ecore_l2.h"
 #include "base/ecore_vf.h"
 
+#include "qede_sriov.h"
 #include "qede_logs.h"
 #include "qede_if.h"
 #include "qede_rxtx.h"
diff --git a/drivers/net/qede/qede_if.h b/drivers/net/qede/qede_if.h
index c5ae3fb2e..1693a243f 100644
--- a/drivers/net/qede/qede_if.h
+++ b/drivers/net/qede/qede_if.h
@@ -82,6 +82,7 @@ struct qed_eth_ops {
const struct qed_common_ops *common;
int (*fill_dev_info)(struct ecore_dev *edev,
 struct qed_dev_eth_info *info);
+   void (*sriov_configure)(struct ecore_dev *edev, int num_vfs);
 };
 
 struct qed_link_params {
diff --git a/drivers/net/qede/qede_main.c b/drivers/net/qede/qede_main.c
index 94061ec11..c1135822b 100644
--- a/drivers/net/qede/qede_main.c
+++ b/drivers/net/qede/qede_main.c
@@ -821,6 +821,7 @@ const struct qed_common_ops qed_common_ops_pass = {
 const struct qed_eth_ops qed_eth_ops_pass = {
INIT_STRUCT_FIELD(common, &qed_common_ops_pass),
INIT_STRUCT_FIELD(fill_dev_info, &qed_fill_eth_dev_info),
+   INIT_STRUCT_FIELD(sriov_configure, &qed_sriov_configure),
 };
 
 const struct qed_eth_ops *qed_get_eth_ops(void)
diff --git a/drivers/net/qede/qede_sriov.c b/drivers/net/qede/qede_sriov.c
new file mode 100644
index 0..ba4384e90
--- /dev/null
+++ b/drivers/net/qede/qede_sriov.c
@@ -0,0 +1,85 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (c) 2020 Marvell.
+ * All rights reserved.
+ * www.marvell.com
+ */
+
+#include "qede_sriov.h"
+
+static void qed_sriov_enable_qid_config(struct ecore_hwfn *hwfn,
+   u16 vfid,
+   struct ecore_iov_vf_init_params *params)
+{
+   u16 num_pf_l2_queues, base, i;
+
+   /* Since we have an equal resource distribution per-VF, and we assume
+* PF has acquired its first queues, we start setting sequentially from
+* there.
+*/
+   num_pf_l2_queues = (u16)FEAT_NUM(hwfn, ECORE_PF_L2_QUE);
+
+   base = num_pf_l2_queues + vfid * params->num_queues;
+   params->rel_vf_id = vfid;
+
+   for (i = 0; i < params->num_queues; i++) {
+   params->req_rx_queue[i] = base + i;
+   params->req_tx_queue[i] = base + i;
+   }
+
+   /* PF uses indices 0 for itself; Set vport/RSS afterwards */
+   params->vport_id = vfid + 1;
+   params->rss_eng_id = vfid + 1;
+}
+
+static void qed_sriov_enable(struct ecore_dev *edev, int num)
+{
+   struct ecore_iov_vf_init_params params;
+   struct ecore_hwfn *p_hwfn;
+   struct ecore_ptt *p_ptt;
+   int i, j, rc;
+
+   if ((u32)num >= RESC_NUM(&edev->hwfns[0], ECORE_VPORT)) {
+   DP_NOTICE(edev, false, "Can start at most %d VFs\n",
+ RESC_NUM(&edev->hwfns[0], ECORE_VPORT) - 1);
+   return;
+   }
+
+   OSAL_MEMSET(¶ms, 0, sizeof(struct ecore_iov_vf_init_params));

[dpdk-dev] [PATCH v6 2/6] net/qede: define PCI config space specific osals

2020-09-25 Thread Manish Chopra
This patch defines various PCI config space access APIs
in order to read and find IOV specific PCI capabilities.

With these definitions implemented, it enables the base
driver to do SR-IOV specific initialization and HW specific
configuration required from PF-PMD driver instance.

Signed-off-by: Manish Chopra 
Signed-off-by: Igor Russkikh 
Signed-off-by: Rasesh Mody 
---
 drivers/net/qede/base/bcm_osal.h| 14 +-
 drivers/net/qede/base/ecore.h   |  3 +++
 drivers/net/qede/base/ecore_dev.c   |  6 +++---
 drivers/net/qede/base/ecore_sriov.c | 28 +++-
 drivers/net/qede/base/meson.build   |  2 +-
 drivers/net/qede/qede_main.c|  1 +
 lib/librte_pci/rte_pci.h| 23 +--
 7 files changed, 53 insertions(+), 24 deletions(-)

diff --git a/drivers/net/qede/base/bcm_osal.h b/drivers/net/qede/base/bcm_osal.h
index ded4fb0f5..681d2d5b6 100644
--- a/drivers/net/qede/base/bcm_osal.h
+++ b/drivers/net/qede/base/bcm_osal.h
@@ -21,6 +21,7 @@
 #include 
 #include 
 #include 
+#include 
 
 /* Forward declaration */
 struct ecore_dev;
@@ -285,11 +286,14 @@ typedef struct osal_list_t {
OSAL_LIST_PUSH_HEAD(new_entry, list)
 
 /* PCI config space */
-
-#define OSAL_PCI_READ_CONFIG_BYTE(dev, address, dst) nothing
-#define OSAL_PCI_READ_CONFIG_WORD(dev, address, dst) nothing
-#define OSAL_PCI_READ_CONFIG_DWORD(dev, address, dst) nothing
-#define OSAL_PCI_FIND_EXT_CAPABILITY(dev, pcie_id) 0
+#define OSAL_PCI_READ_CONFIG_BYTE(dev, address, dst) \
+   rte_pci_read_config((dev)->pci_dev, dst, 1, address)
+#define OSAL_PCI_READ_CONFIG_WORD(dev, address, dst) \
+   rte_pci_read_config((dev)->pci_dev, dst, 2, address)
+#define OSAL_PCI_READ_CONFIG_DWORD(dev, address, dst) \
+   rte_pci_read_config((dev)->pci_dev, dst, 4, address)
+#define OSAL_PCI_FIND_EXT_CAPABILITY(dev, cap) \
+   rte_pci_find_ext_capability((dev)->pci_dev, cap)
 #define OSAL_PCI_FIND_CAPABILITY(dev, pcie_id) 0
 #define OSAL_PCI_WRITE_CONFIG_WORD(dev, address, val) nothing
 #define OSAL_BAR_SIZE(dev, bar_id) 0
diff --git a/drivers/net/qede/base/ecore.h b/drivers/net/qede/base/ecore.h
index 63bd7466a..750e99a8f 100644
--- a/drivers/net/qede/base/ecore.h
+++ b/drivers/net/qede/base/ecore.h
@@ -937,6 +937,9 @@ struct ecore_dev {
struct ecore_dbg_featuredbg_features[DBG_FEATURE_NUM];
struct ecore_dbg_params dbg_params;
osal_mutex_tdbg_lock;
+
+   /* DPDK specific ecore field */
+   struct rte_pci_device   *pci_dev;
 };
 
 enum ecore_hsi_def_type {
diff --git a/drivers/net/qede/base/ecore_dev.c 
b/drivers/net/qede/base/ecore_dev.c
index 35a8394de..e895dee40 100644
--- a/drivers/net/qede/base/ecore_dev.c
+++ b/drivers/net/qede/base/ecore_dev.c
@@ -2787,7 +2787,7 @@ static enum _ecore_status_t ecore_hw_init_chip(struct 
ecore_dev *p_dev,
return ECORE_IO;
}
 
-   OSAL_PCI_READ_CONFIG_WORD(p_dev, pos + PCI_EXP_DEVCTL, &ctrl);
+   OSAL_PCI_READ_CONFIG_WORD(p_dev, pos + RTE_PCI_EXP_DEVCTL, &ctrl);
wr_mbs = (ctrl & PCI_EXP_DEVCTL_PAYLOAD) >> 5;
ecore_wr(p_hwfn, p_ptt, PSWRQ2_REG_WR_MBS0, wr_mbs);
 
@@ -5499,9 +5499,9 @@ static enum _ecore_status_t ecore_get_dev_info(struct 
ecore_hwfn *p_hwfn,
u32 tmp;
 
/* Read Vendor Id / Device Id */
-   OSAL_PCI_READ_CONFIG_WORD(p_dev, PCICFG_VENDOR_ID_OFFSET,
+   OSAL_PCI_READ_CONFIG_WORD(p_dev, RTE_PCI_VENDOR_ID,
  &p_dev->vendor_id);
-   OSAL_PCI_READ_CONFIG_WORD(p_dev, PCICFG_DEVICE_ID_OFFSET,
+   OSAL_PCI_READ_CONFIG_WORD(p_dev, RTE_PCI_DEVICE_ID,
  &p_dev->device_id);
 
/* Determine type */
diff --git a/drivers/net/qede/base/ecore_sriov.c 
b/drivers/net/qede/base/ecore_sriov.c
index e60257e19..dac4cbee8 100644
--- a/drivers/net/qede/base/ecore_sriov.c
+++ b/drivers/net/qede/base/ecore_sriov.c
@@ -417,15 +417,16 @@ static enum _ecore_status_t ecore_iov_pci_cfg_info(struct 
ecore_dev *p_dev)
int pos = iov->pos;
 
DP_VERBOSE(p_dev, ECORE_MSG_IOV, "sriov ext pos %d\n", pos);
-   OSAL_PCI_READ_CONFIG_WORD(p_dev, pos + PCI_SRIOV_CTRL, &iov->ctrl);
+   OSAL_PCI_READ_CONFIG_WORD(p_dev, pos + RTE_PCI_SRIOV_CTRL, &iov->ctrl);
 
+   OSAL_PCI_READ_CONFIG_WORD(p_dev, pos + RTE_PCI_SRIOV_TOTAL_VF,
+ &iov->total_vfs);
OSAL_PCI_READ_CONFIG_WORD(p_dev,
- pos + PCI_SRIOV_TOTAL_VF, &iov->total_vfs);
-   OSAL_PCI_READ_CONFIG_WORD(p_dev,
- pos + PCI_SRIOV_INITIAL_VF,
+ pos + RTE_PCI_SRIOV_INITIAL_VF,
  &iov->initial_vfs);
 
-   OSAL_PCI_READ_CONFIG_WORD(p_dev, pos + PCI_SRIOV_NUM_VF, &iov->num_vfs);
+   OSAL_PCI_READ_CONFIG_WORD(p_dev, pos + RTE_PCI_SRIOV_NUM_VF,
+ &iov->num_vfs);
if (iov->n

Re: [dpdk-dev] [PATCH v2] net/bnxt: add a check in bnxt PCI probe

2020-09-25 Thread Ferruh Yigit

On 9/25/2020 11:40 AM, Somnath Kotur wrote:

Check for num_rep before invoking rep port probe.

Fixes: 6dc83230b43b ("net/bnxt: support port representor data path")

Signed-off-by: Somnath Kotur 
Reviewed-by: Venkat Duvvuru 


Reviewed-by: Ferruh Yigit 

Applied to dpdk-next-net/main, thanks.


[dpdk-dev] [PATCH v6 1/6] drivers: add generic API to find PCI extended cap

2020-09-25 Thread Manish Chopra
By adding generic API, this patch removes individual
functions/defines implemented by drivers to find extended
PCI capabilities.

Signed-off-by: Manish Chopra 
Signed-off-by: Igor Russkikh 
Reviewed-by: Gaetan Rivet 
---
 drivers/bus/pci/pci_common.c   | 43 ++
 drivers/bus/pci/rte_bus_pci.h  | 19 
 drivers/bus/pci/rte_bus_pci_version.map|  6 +++
 drivers/net/ice/ice_ethdev.c   | 51 +-
 drivers/net/nfp/nfpcore/nfp_cpp_pcie_ops.c | 48 +---
 drivers/raw/ifpga/ifpga_rawdev.c   | 17 +++-
 lib/librte_pci/rte_pci.h   | 16 +++
 7 files changed, 94 insertions(+), 106 deletions(-)

diff --git a/drivers/bus/pci/pci_common.c b/drivers/bus/pci/pci_common.c
index a8e5fd52c..b686aad6a 100644
--- a/drivers/bus/pci/pci_common.c
+++ b/drivers/bus/pci/pci_common.c
@@ -665,6 +665,49 @@ rte_pci_get_iommu_class(void)
return iova_mode;
 }
 
+off_t
+rte_pci_find_ext_capability(struct rte_pci_device *dev, uint32_t cap)
+{
+   off_t offset = RTE_PCI_CFG_SPACE_SIZE;
+   uint32_t header;
+   int ttl;
+
+   /* minimum 8 bytes per capability */
+   ttl = (RTE_PCI_CFG_SPACE_EXP_SIZE - RTE_PCI_CFG_SPACE_SIZE) / 8;
+
+   if (rte_pci_read_config(dev, &header, 4, offset) < 0) {
+   RTE_LOG(ERR, EAL, "error in reading extended capabilities\n");
+   return -1;
+   }
+
+   /*
+* If we have no capabilities, this is indicated by cap ID,
+* cap version and next pointer all being 0.
+*/
+   if (header == 0)
+   return 0;
+
+   while (ttl != 0) {
+   if (RTE_PCI_EXT_CAP_ID(header) == cap)
+   return offset;
+
+   offset = RTE_PCI_EXT_CAP_NEXT(header);
+
+   if (offset < RTE_PCI_CFG_SPACE_SIZE)
+   break;
+
+   if (rte_pci_read_config(dev, &header, 4, offset) < 0) {
+   RTE_LOG(ERR, EAL,
+   "error in reading extended capabilities\n");
+   return -1;
+   }
+
+   ttl--;
+   }
+
+   return 0;
+}
+
 struct rte_pci_bus rte_pci_bus = {
.bus = {
.scan = rte_pci_scan,
diff --git a/drivers/bus/pci/rte_bus_pci.h b/drivers/bus/pci/rte_bus_pci.h
index 29bea6d70..5bb1854d6 100644
--- a/drivers/bus/pci/rte_bus_pci.h
+++ b/drivers/bus/pci/rte_bus_pci.h
@@ -224,6 +224,25 @@ void rte_pci_unmap_device(struct rte_pci_device *dev);
  */
 void rte_pci_dump(FILE *f);
 
+/**
+ * Find device's extended PCI capability.
+ *
+ *  @param dev
+ *A pointer to rte_pci_device structure.
+ *
+ *  @param cap
+ *Extended capability to be found, which can be any from
+ *RTE_PCI_EXT_CAP_ID_*, defined in librte_pci.
+ *
+ *  @return
+ *  > 0: The offset of the next matching extended capability structure
+ *   within the device's PCI configuration space.
+ *  < 0: An error in PCI config space read.
+ *  = 0: Device does not support it.
+ */
+__rte_experimental
+off_t rte_pci_find_ext_capability(struct rte_pci_device *dev, uint32_t cap);
+
 /**
  * Register a PCI driver.
  *
diff --git a/drivers/bus/pci/rte_bus_pci_version.map 
b/drivers/bus/pci/rte_bus_pci_version.map
index 5b75d2367..f33ed0abd 100644
--- a/drivers/bus/pci/rte_bus_pci_version.map
+++ b/drivers/bus/pci/rte_bus_pci_version.map
@@ -16,3 +16,9 @@ DPDK_21 {
 
local: *;
 };
+
+EXPERIMENTAL {
+   global:
+
+   rte_pci_find_ext_capability;
+};
diff --git a/drivers/net/ice/ice_ethdev.c b/drivers/net/ice/ice_ethdev.c
index cfd357b05..6513cc36b 100644
--- a/drivers/net/ice/ice_ethdev.c
+++ b/drivers/net/ice/ice_ethdev.c
@@ -1753,53 +1753,6 @@ ice_pf_setup(struct ice_pf *pf)
return 0;
 }
 
-/* PCIe configuration space setting */
-#define PCI_CFG_SPACE_SIZE  256
-#define PCI_CFG_SPACE_EXP_SIZE  4096
-#define PCI_EXT_CAP_ID(header)  (int)((header) & 0x)
-#define PCI_EXT_CAP_NEXT(header)(((header) >> 20) & 0xffc)
-#define PCI_EXT_CAP_ID_DSN  0x03
-
-static int
-ice_pci_find_next_ext_capability(struct rte_pci_device *dev, int cap)
-{
-   uint32_t header;
-   int ttl;
-   int pos = PCI_CFG_SPACE_SIZE;
-
-   /* minimum 8 bytes per capability */
-   ttl = (PCI_CFG_SPACE_EXP_SIZE - PCI_CFG_SPACE_SIZE) / 8;
-
-   if (rte_pci_read_config(dev, &header, 4, pos) < 0) {
-   PMD_INIT_LOG(ERR, "ice error reading extended capabilities\n");
-   return -1;
-   }
-
-   /*
-* If we have no capabilities, this is indicated by cap ID,
-* cap version and next pointer all being 0.
-*/
-   if (header == 0)
-   return 0;
-
-   while (ttl-- > 0) {
-   if (PCI_EXT_CAP_ID(header) == cap)
-   return pos;
-
-   pos = PCI_EXT_CAP_NEXT(header);
-
-   if (pos < PCI_CFG_SPACE_SIZE)
-  

[dpdk-dev] [PATCH v6 4/6] net/qede: add infrastructure support for VF load

2020-09-25 Thread Manish Chopra
This patch adds necessary infrastructure support (required to handle
messages from VF and sending ramrod on behalf of VF's configuration
request from alarm handler context) to start/load the VF-PMD driver
instance on top of PF-PMD driver instance.

Signed-off-by: Manish Chopra 
Signed-off-by: Igor Russkikh 
Signed-off-by: Rasesh Mody 
---
 drivers/net/qede/base/bcm_osal.c  | 26 
 drivers/net/qede/base/bcm_osal.h  | 11 +++--
 drivers/net/qede/base/ecore.h |  4 ++
 drivers/net/qede/base/ecore_iov_api.h |  3 ++
 drivers/net/qede/qede_ethdev.c|  2 +
 drivers/net/qede/qede_main.c  |  4 +-
 drivers/net/qede/qede_sriov.c | 61 +++
 drivers/net/qede/qede_sriov.h | 16 ++-
 8 files changed, 121 insertions(+), 6 deletions(-)

diff --git a/drivers/net/qede/base/bcm_osal.c b/drivers/net/qede/base/bcm_osal.c
index 65837b53d..ef47339df 100644
--- a/drivers/net/qede/base/bcm_osal.c
+++ b/drivers/net/qede/base/bcm_osal.c
@@ -14,6 +14,32 @@
 #include "ecore_iov_api.h"
 #include "ecore_mcp_api.h"
 #include "ecore_l2_api.h"
+#include "../qede_sriov.h"
+
+int osal_pf_vf_msg(struct ecore_hwfn *p_hwfn)
+{
+   int rc;
+
+   rc = qed_schedule_iov(p_hwfn, QED_IOV_WQ_MSG_FLAG);
+   if (rc) {
+   DP_VERBOSE(p_hwfn, ECORE_MSG_IOV,
+  "Failed to schedule alarm handler rc=%d\n", rc);
+   }
+
+   return rc;
+}
+
+void osal_poll_mode_dpc(osal_int_ptr_t hwfn_cookie)
+{
+   struct ecore_hwfn *p_hwfn = (struct ecore_hwfn *)hwfn_cookie;
+
+   if (!p_hwfn)
+   return;
+
+   OSAL_SPIN_LOCK(&p_hwfn->spq_lock);
+   ecore_int_sp_dpc((osal_int_ptr_t)(p_hwfn));
+   OSAL_SPIN_UNLOCK(&p_hwfn->spq_lock);
+}
 
 /* Array of memzone pointers */
 static const struct rte_memzone *ecore_mz_mapping[RTE_MAX_MEMZONE];
diff --git a/drivers/net/qede/base/bcm_osal.h b/drivers/net/qede/base/bcm_osal.h
index 681d2d5b6..e7212f40c 100644
--- a/drivers/net/qede/base/bcm_osal.h
+++ b/drivers/net/qede/base/bcm_osal.h
@@ -177,9 +177,12 @@ typedef pthread_mutex_t osal_mutex_t;
 
 /* DPC */
 
+void osal_poll_mode_dpc(osal_int_ptr_t hwfn_cookie);
 #define OSAL_DPC_ALLOC(hwfn) OSAL_ALLOC(hwfn, GFP, sizeof(osal_dpc_t))
-#define OSAL_DPC_INIT(dpc, hwfn) nothing
-#define OSAL_POLL_MODE_DPC(hwfn) nothing
+#define OSAL_DPC_INIT(dpc, hwfn) \
+   OSAL_SPIN_LOCK_INIT(&(hwfn)->spq_lock)
+#define OSAL_POLL_MODE_DPC(hwfn) \
+   osal_poll_mode_dpc((osal_int_ptr_t)(p_hwfn))
 #define OSAL_DPC_SYNC(hwfn) nothing
 
 /* Lists */
@@ -344,10 +347,12 @@ u32 qede_find_first_zero_bit(u32 *bitmap, u32 length);
 
 /* SR-IOV channel */
 
+int osal_pf_vf_msg(struct ecore_hwfn *p_hwfn);
 #define OSAL_VF_FLR_UPDATE(hwfn) nothing
 #define OSAL_VF_SEND_MSG2PF(dev, done, msg, reply_addr, msg_size, reply_size) 0
 #define OSAL_VF_CQE_COMPLETION(_dev_p, _cqe, _protocol)(0)
-#define OSAL_PF_VF_MSG(hwfn, vfid) 0
+#define OSAL_PF_VF_MSG(hwfn, vfid) \
+   osal_pf_vf_msg(hwfn)
 #define OSAL_PF_VF_MALICIOUS(hwfn, vfid) nothing
 #define OSAL_IOV_CHK_UCAST(hwfn, vfid, params) 0
 #define OSAL_IOV_POST_START_VPORT(hwfn, vf, vport_id, opaque_fid) nothing
diff --git a/drivers/net/qede/base/ecore.h b/drivers/net/qede/base/ecore.h
index 750e99a8f..6c8e6d407 100644
--- a/drivers/net/qede/base/ecore.h
+++ b/drivers/net/qede/base/ecore.h
@@ -714,6 +714,10 @@ struct ecore_hwfn {
 
/* @DPDK */
struct ecore_ptt*p_arfs_ptt;
+
+   /* DPDK specific, not the part of vanilla ecore */
+   osal_spinlock_t spq_lock;
+   u32 iov_task_flags;
 };
 
 enum ecore_mf_mode {
diff --git a/drivers/net/qede/base/ecore_iov_api.h 
b/drivers/net/qede/base/ecore_iov_api.h
index 545001812..bd7c5703f 100644
--- a/drivers/net/qede/base/ecore_iov_api.h
+++ b/drivers/net/qede/base/ecore_iov_api.h
@@ -14,6 +14,9 @@
 #define ECORE_ETH_VF_NUM_VLAN_FILTERS 2
 #define ECORE_VF_ARRAY_LENGTH (3)
 
+#define ECORE_VF_ARRAY_GET_VFID(arr, vfid) \
+   (((arr)[(vfid) / 64]) & (1ULL << ((vfid) % 64)))
+
 #define IS_VF(p_dev)   ((p_dev)->b_is_vf)
 #define IS_PF(p_dev)   (!((p_dev)->b_is_vf))
 #ifdef CONFIG_ECORE_SRIOV
diff --git a/drivers/net/qede/qede_ethdev.c b/drivers/net/qede/qede_ethdev.c
index 27ec42e2c..3cc141954 100644
--- a/drivers/net/qede/qede_ethdev.c
+++ b/drivers/net/qede/qede_ethdev.c
@@ -281,7 +281,9 @@ qede_fw_version_get(struct rte_eth_dev *dev, char *fw_ver, 
size_t fw_size)
 
 static void qede_interrupt_action(struct ecore_hwfn *p_hwfn)
 {
+   OSAL_SPIN_LOCK(&p_hwfn->spq_lock);
ecore_int_sp_dpc((osal_int_ptr_t)(p_hwfn));
+   OSAL_SPIN_UNLOCK(&p_hwfn->spq_lock);
 }
 
 static void
diff --git a/drivers/net/qede/qede_main.c b/drivers/net/qede/qede_main.c
index c1135822b..6a7bfc188 100644
--- a/drivers/net/qede/qede_main.c
+++ b/drivers/net/qede/qede_main.c
@@ -221,7 +221,9 @@ static void qed_stop_iov_task(struct ecore_dev *edev)
 
for_each_hwfn(edev, i) {
p_

[dpdk-dev] [PATCH v6 5/6] net/qede: initialize VF MAC and link

2020-09-25 Thread Manish Chopra
This patch configures VFs with random mac if no MAC is
provided by the PF/bulletin. This also adds required bulletin
APIs by PF-PMD driver to communicate LINK properties/changes to
the VFs through bulletin update mechanism.

With these changes, VF-PMD instance is able to run
fastpath over PF-PMD driver instance.

Signed-off-by: Manish Chopra 
Signed-off-by: Igor Russkikh 
Signed-off-by: Rasesh Mody 
---
 drivers/net/qede/qede_ethdev.c | 34 -
 drivers/net/qede/qede_main.c   |  6 +++-
 drivers/net/qede/qede_sriov.c  | 55 ++
 drivers/net/qede/qede_sriov.h  |  1 +
 4 files changed, 94 insertions(+), 2 deletions(-)

diff --git a/drivers/net/qede/qede_ethdev.c b/drivers/net/qede/qede_ethdev.c
index 3cc141954..150c70d94 100644
--- a/drivers/net/qede/qede_ethdev.c
+++ b/drivers/net/qede/qede_ethdev.c
@@ -2477,6 +2477,24 @@ static void qede_update_pf_params(struct ecore_dev *edev)
qed_ops->common->update_pf_params(edev, &pf_params);
 }
 
+static void qede_generate_random_mac_addr(struct rte_ether_addr *mac_addr)
+{
+   uint64_t random;
+
+   /* Set Organizationally Unique Identifier (OUI) prefix. */
+   mac_addr->addr_bytes[0] = 0x00;
+   mac_addr->addr_bytes[1] = 0x09;
+   mac_addr->addr_bytes[2] = 0xC0;
+
+   /* Force indication of locally assigned MAC address. */
+   mac_addr->addr_bytes[0] |= RTE_ETHER_LOCAL_ADMIN_ADDR;
+
+   /* Generate the last 3 bytes of the MAC address with a random number. */
+   random = rte_rand();
+
+   memcpy(&mac_addr->addr_bytes[3], &random, 3);
+}
+
 static int qede_common_dev_init(struct rte_eth_dev *eth_dev, bool is_vf)
 {
struct rte_pci_device *pci_dev;
@@ -2489,7 +2507,7 @@ static int qede_common_dev_init(struct rte_eth_dev 
*eth_dev, bool is_vf)
uint8_t bulletin_change;
uint8_t vf_mac[RTE_ETHER_ADDR_LEN];
uint8_t is_mac_forced;
-   bool is_mac_exist;
+   bool is_mac_exist = false;
/* Fix up ecore debug level */
uint32_t dp_module = ~0 & ~ECORE_MSG_HW;
uint8_t dp_level = ECORE_LEVEL_VERBOSE;
@@ -2667,6 +2685,20 @@ static int qede_common_dev_init(struct rte_eth_dev 
*eth_dev, bool is_vf)
DP_ERR(edev, "No VF macaddr assigned\n");
}
}
+
+   /* If MAC doesn't exist from PF, generate random one */
+   if (!is_mac_exist) {
+   struct rte_ether_addr *mac_addr;
+
+   mac_addr = (struct rte_ether_addr *)&vf_mac;
+   qede_generate_random_mac_addr(mac_addr);
+
+   rte_ether_addr_copy(mac_addr,
+   ð_dev->data->mac_addrs[0]);
+
+   rte_ether_addr_copy(ð_dev->data->mac_addrs[0],
+   &adapter->primary_mac);
+   }
}
 
eth_dev->dev_ops = (is_vf) ? &qede_eth_vf_dev_ops : &qede_eth_dev_ops;
diff --git a/drivers/net/qede/qede_main.c b/drivers/net/qede/qede_main.c
index 6a7bfc188..0d3cea6e4 100644
--- a/drivers/net/qede/qede_main.c
+++ b/drivers/net/qede/qede_main.c
@@ -651,8 +651,12 @@ void qed_link_update(struct ecore_hwfn *hwfn)
struct ecore_dev *edev = hwfn->p_dev;
struct qede_dev *qdev = (struct qede_dev *)edev;
struct rte_eth_dev *dev = (struct rte_eth_dev *)qdev->ethdev;
+   int rc;
+
+   rc = qede_link_update(dev, 0);
+   qed_inform_vf_link_state(hwfn);
 
-   if (!qede_link_update(dev, 0))
+   if (!rc)
rte_eth_dev_callback_process(dev, RTE_ETH_EVENT_INTR_LSC, NULL);
 }
 
diff --git a/drivers/net/qede/qede_sriov.c b/drivers/net/qede/qede_sriov.c
index 6d620dde8..93f7a2a55 100644
--- a/drivers/net/qede/qede_sriov.c
+++ b/drivers/net/qede/qede_sriov.c
@@ -126,6 +126,28 @@ static void qed_handle_vf_msg(struct ecore_hwfn *hwfn)
ecore_ptt_release(hwfn, ptt);
 }
 
+static void qed_handle_bulletin_post(struct ecore_hwfn *hwfn)
+{
+   struct ecore_ptt *ptt;
+   int i;
+
+   ptt = ecore_ptt_acquire(hwfn);
+   if (!ptt) {
+   DP_NOTICE(hwfn, true, "PTT acquire failed\n");
+   qed_schedule_iov(hwfn, QED_IOV_WQ_BULLETIN_UPDATE_FLAG);
+   return;
+   }
+
+   /* TODO - at the moment update bulletin board of all VFs.
+* if this proves to costly, we can mark VFs that need their
+* bulletins updated.
+*/
+   ecore_for_each_vf(hwfn, i)
+   ecore_iov_post_vf_bulletin(hwfn, i, ptt);
+
+   ecore_ptt_release(hwfn, ptt);
+}
+
 void qed_iov_pf_task(void *arg)
 {
struct ecore_hwfn *p_hwfn = arg;
@@ -134,6 +156,13 @@ void qed_iov_pf_task(void *arg)
OSAL_CLEAR_BIT(QED_IOV_WQ_MSG_FLAG, &p_hwfn->iov_task_flags);
qed_handle_vf_msg(p_hwfn);
}
+
+   if (OSAL_GET_BIT(QED_IOV_WQ_BULLETIN_UPDATE_FLAG,
+&p_hwfn->iov_task_fla

[dpdk-dev] [PATCH v6 6/6] net/qede: add VF FLR support

2020-09-25 Thread Manish Chopra
This patch adds required bit to handle VF FLR
indication from Management FW (MFW) of the device

With that VFs were able to load in VM (VF attached as PCI
passthrough to the guest VM) followed by FLR successfully

Updated the docs/guides with the feature support

Signed-off-by: Manish Chopra 
Signed-off-by: Igor Russkikh 
Signed-off-by: Rasesh Mody 
---
 doc/guides/nics/features/qede.ini |  1 +
 doc/guides/nics/qede.rst  |  7 +--
 drivers/net/qede/base/bcm_osal.c  |  5 +
 drivers/net/qede/base/bcm_osal.h  |  4 +++-
 drivers/net/qede/qede_sriov.c | 18 ++
 5 files changed, 28 insertions(+), 7 deletions(-)

diff --git a/doc/guides/nics/features/qede.ini 
b/doc/guides/nics/features/qede.ini
index f8716523e..46fba8e6c 100644
--- a/doc/guides/nics/features/qede.ini
+++ b/doc/guides/nics/features/qede.ini
@@ -32,6 +32,7 @@ Basic stats  = Y
 Extended stats   = Y
 Stats per queue  = Y
 Registers dump   = Y
+SR-IOV   = Y
 Multiprocess aware   = Y
 Linux UIO= Y
 Linux VFIO   = Y
diff --git a/doc/guides/nics/qede.rst b/doc/guides/nics/qede.rst
index 5b2f86895..e874915c2 100644
--- a/doc/guides/nics/qede.rst
+++ b/doc/guides/nics/qede.rst
@@ -34,18 +34,13 @@ Supported Features
 - VLAN offload - Filtering and stripping
 - N-tuple filter and flow director (limited support)
 - NPAR (NIC Partitioning)
-- SR-IOV VF
+- SR-IOV PF and VF
 - GRE Tunneling offload
 - GENEVE Tunneling offload
 - VXLAN Tunneling offload
 - MPLSoUDP Tx Tunneling offload
 - Generic flow API
 
-Non-supported Features
---
-
-- SR-IOV PF
-
 Co-existence considerations
 ---
 
diff --git a/drivers/net/qede/base/bcm_osal.c b/drivers/net/qede/base/bcm_osal.c
index ef47339df..44a8692f5 100644
--- a/drivers/net/qede/base/bcm_osal.c
+++ b/drivers/net/qede/base/bcm_osal.c
@@ -29,6 +29,11 @@ int osal_pf_vf_msg(struct ecore_hwfn *p_hwfn)
return rc;
 }
 
+void osal_vf_flr_update(struct ecore_hwfn *p_hwfn)
+{
+   qed_schedule_iov(p_hwfn, QED_IOV_WQ_FLR_FLAG);
+}
+
 void osal_poll_mode_dpc(osal_int_ptr_t hwfn_cookie)
 {
struct ecore_hwfn *p_hwfn = (struct ecore_hwfn *)hwfn_cookie;
diff --git a/drivers/net/qede/base/bcm_osal.h b/drivers/net/qede/base/bcm_osal.h
index e7212f40c..575d4d099 100644
--- a/drivers/net/qede/base/bcm_osal.h
+++ b/drivers/net/qede/base/bcm_osal.h
@@ -348,7 +348,9 @@ u32 qede_find_first_zero_bit(u32 *bitmap, u32 length);
 /* SR-IOV channel */
 
 int osal_pf_vf_msg(struct ecore_hwfn *p_hwfn);
-#define OSAL_VF_FLR_UPDATE(hwfn) nothing
+void osal_vf_flr_update(struct ecore_hwfn *p_hwfn);
+#define OSAL_VF_FLR_UPDATE(hwfn) \
+   osal_vf_flr_update(hwfn)
 #define OSAL_VF_SEND_MSG2PF(dev, done, msg, reply_addr, msg_size, reply_size) 0
 #define OSAL_VF_CQE_COMPLETION(_dev_p, _cqe, _protocol)(0)
 #define OSAL_PF_VF_MSG(hwfn, vfid) \
diff --git a/drivers/net/qede/qede_sriov.c b/drivers/net/qede/qede_sriov.c
index 93f7a2a55..0b99a8d6f 100644
--- a/drivers/net/qede/qede_sriov.c
+++ b/drivers/net/qede/qede_sriov.c
@@ -151,6 +151,7 @@ static void qed_handle_bulletin_post(struct ecore_hwfn 
*hwfn)
 void qed_iov_pf_task(void *arg)
 {
struct ecore_hwfn *p_hwfn = arg;
+   int rc;
 
if (OSAL_GET_BIT(QED_IOV_WQ_MSG_FLAG, &p_hwfn->iov_task_flags)) {
OSAL_CLEAR_BIT(QED_IOV_WQ_MSG_FLAG, &p_hwfn->iov_task_flags);
@@ -163,6 +164,23 @@ void qed_iov_pf_task(void *arg)
   &p_hwfn->iov_task_flags);
qed_handle_bulletin_post(p_hwfn);
}
+
+   if (OSAL_GET_BIT(QED_IOV_WQ_FLR_FLAG, &p_hwfn->iov_task_flags)) {
+   struct ecore_ptt *p_ptt = ecore_ptt_acquire(p_hwfn);
+
+   OSAL_CLEAR_BIT(QED_IOV_WQ_FLR_FLAG, &p_hwfn->iov_task_flags);
+
+   if (!p_ptt) {
+   qed_schedule_iov(p_hwfn, QED_IOV_WQ_FLR_FLAG);
+   return;
+   }
+
+   rc = ecore_iov_vf_flr_cleanup(p_hwfn, p_ptt);
+   if (rc)
+   qed_schedule_iov(p_hwfn, QED_IOV_WQ_FLR_FLAG);
+
+   ecore_ptt_release(p_hwfn, p_ptt);
+   }
 }
 
 int qed_schedule_iov(struct ecore_hwfn *p_hwfn, enum qed_iov_wq_flag flag)
-- 
2.17.1



[dpdk-dev] IPSEC-SECGW application

2020-09-25 Thread Rohit kamble
HI Guys ,

I am trying to use a DPDK based *ipsec-secgw application*. Can anyone
please provide me with all the configurations, setup/topology and about the
NIC drivers.
I already went through the documentation of dpdk, but I am  seeing several
issues regarding ipsec-secgw applications.

Thanks,
Rohit


[dpdk-dev] [PATCH v2 3/8] net/virtio: move backend type selection to ethdev

2020-09-25 Thread Maxime Coquelin
From: Adrian Moreno 

This is a preparation patch with no functional change.

Use an enum instead of a boolean for the backend type.
Move the detection logic to the ethdev layer (where it is needed for the
first time).
The virtio_user_dev stores the backend type in the virtio_user_dev
struct so the type is only determined once

Signed-off-by: Maxime Coquelin 
Signed-off-by: Adrian Moreno 
---
 .../net/virtio/virtio_user/virtio_user_dev.c  | 37 ---
 .../net/virtio/virtio_user/virtio_user_dev.h  | 11 +-
 drivers/net/virtio/virtio_user_ethdev.c   | 28 +-
 3 files changed, 50 insertions(+), 26 deletions(-)

diff --git a/drivers/net/virtio/virtio_user/virtio_user_dev.c 
b/drivers/net/virtio/virtio_user/virtio_user_dev.c
index 2a0c861085..7cb93b1b33 100644
--- a/drivers/net/virtio/virtio_user/virtio_user_dev.c
+++ b/drivers/net/virtio/virtio_user/virtio_user_dev.c
@@ -111,17 +111,6 @@ virtio_user_queue_setup(struct virtio_user_dev *dev,
return 0;
 }
 
-int
-is_vhost_user_by_type(const char *path)
-{
-   struct stat sb;
-
-   if (stat(path, &sb) == -1)
-   return 0;
-
-   return S_ISSOCK(sb.st_mode);
-}
-
 int
 virtio_user_start_device(struct virtio_user_dev *dev)
 {
@@ -144,7 +133,8 @@ virtio_user_start_device(struct virtio_user_dev *dev)
rte_mcfg_mem_read_lock();
pthread_mutex_lock(&dev->mutex);
 
-   if (is_vhost_user_by_type(dev->path) && dev->vhostfd < 0)
+   if (dev->backend_type == VIRTIO_USER_BACKEND_VHOST_USER
+   && dev->vhostfd < 0)
goto error;
 
/* Step 0: tell vhost to create queues */
@@ -360,16 +350,16 @@ virtio_user_dev_setup(struct virtio_user_dev *dev)
dev->tapfds = NULL;
 
if (dev->is_server) {
-   if (access(dev->path, F_OK) == 0 &&
-   !is_vhost_user_by_type(dev->path)) {
-   PMD_DRV_LOG(ERR, "Server mode doesn't support 
vhost-kernel!");
+   if (dev->backend_type != VIRTIO_USER_BACKEND_VHOST_USER) {
+   PMD_DRV_LOG(ERR, "Server mode only supports 
vhost-user!");
return -1;
}
dev->ops = &virtio_ops_user;
} else {
-   if (is_vhost_user_by_type(dev->path)) {
+   if (dev->backend_type == VIRTIO_USER_BACKEND_VHOST_USER) {
dev->ops = &virtio_ops_user;
-   } else {
+   } else if (dev->backend_type ==
+   VIRTIO_USER_BACKEND_VHOST_KERNEL) {
dev->ops = &virtio_ops_kernel;
 
dev->vhostfds = malloc(dev->max_queue_pairs *
@@ -430,7 +420,8 @@ virtio_user_dev_setup(struct virtio_user_dev *dev)
 int
 virtio_user_dev_init(struct virtio_user_dev *dev, char *path, int queues,
 int cq, int queue_size, const char *mac, char **ifname,
-int server, int mrg_rxbuf, int in_order, int packed_vq)
+int server, int mrg_rxbuf, int in_order, int packed_vq,
+enum virtio_user_backend_type backend_type)
 {
uint64_t protocol_features = 0;
 
@@ -445,6 +436,8 @@ virtio_user_dev_init(struct virtio_user_dev *dev, char 
*path, int queues,
dev->frontend_features = 0;
dev->unsupported_features = ~VIRTIO_USER_SUPPORTED_FEATURES;
dev->protocol_features = VIRTIO_USER_SUPPORTED_PROTOCOL_FEATURES;
+   dev->backend_type = backend_type;
+
parse_mac(dev, mac);
 
if (*ifname) {
@@ -457,7 +450,7 @@ virtio_user_dev_init(struct virtio_user_dev *dev, char 
*path, int queues,
return -1;
}
 
-   if (!is_vhost_user_by_type(dev->path))
+   if (dev->backend_type != VIRTIO_USER_BACKEND_VHOST_USER)
dev->unsupported_features |=
(1ULL << VHOST_USER_F_PROTOCOL_FEATURES);
 
@@ -539,7 +532,7 @@ virtio_user_dev_init(struct virtio_user_dev *dev, char 
*path, int queues,
}
 
/* The backend will not report this feature, we add it explicitly */
-   if (is_vhost_user_by_type(dev->path))
+   if (dev->backend_type == VIRTIO_USER_BACKEND_VHOST_USER)
dev->frontend_features |= (1ull << VIRTIO_NET_F_STATUS);
 
/*
@@ -792,7 +785,7 @@ virtio_user_send_status_update(struct virtio_user_dev *dev, 
uint8_t status)
uint64_t arg = status;
 
/* Vhost-user only for now */
-   if (!is_vhost_user_by_type(dev->path))
+   if (dev->backend_type != VIRTIO_USER_BACKEND_VHOST_USER)
return 0;
 
if (!(dev->protocol_features & (1ULL << VHOST_USER_PROTOCOL_F_STATUS)))
@@ -815,7 +808,7 @@ virtio_user_update_status(struct virtio_user_dev *dev)
int err;
 
/* Vhost-user only for now */
-   if (!is_vhost_user_by_type(dev->path))
+   if (dev->backend_type != VIRTIO_USER_BACKEND_VHOST_USER)
return 0;
 
if (!(dev->protocol_fe

[dpdk-dev] [PATCH v2 2/8] net/virtio: introduce DMA ops

2020-09-25 Thread Maxime Coquelin
Add DMA map/unmap callbacks to the virtio_user pmd, which could
be leveraged by vdev bus driver to map memory for backend
devices with DMA capability.

Signed-off-by: Maxime Coquelin 
---
 drivers/net/virtio/virtio_user/vhost.h  |  4 ++
 drivers/net/virtio/virtio_user_ethdev.c | 54 +
 2 files changed, 58 insertions(+)

diff --git a/drivers/net/virtio/virtio_user/vhost.h 
b/drivers/net/virtio/virtio_user/vhost.h
index 8f49ef4574..2e71995a79 100644
--- a/drivers/net/virtio/virtio_user/vhost.h
+++ b/drivers/net/virtio/virtio_user/vhost.h
@@ -105,6 +105,10 @@ struct virtio_user_backend_ops {
int (*enable_qp)(struct virtio_user_dev *dev,
 uint16_t pair_idx,
 int enable);
+   int (*dma_map)(struct virtio_user_dev *dev, void *addr,
+ uint64_t iova, size_t len);
+   int (*dma_unmap)(struct virtio_user_dev *dev, void *addr,
+ uint64_t iova, size_t len);
 };
 
 extern struct virtio_user_backend_ops virtio_ops_user;
diff --git a/drivers/net/virtio/virtio_user_ethdev.c 
b/drivers/net/virtio/virtio_user_ethdev.c
index 87f6cb6950..60d17af888 100644
--- a/drivers/net/virtio/virtio_user_ethdev.c
+++ b/drivers/net/virtio/virtio_user_ethdev.c
@@ -818,9 +818,63 @@ virtio_user_pmd_remove(struct rte_vdev_device *vdev)
return 0;
 }
 
+static int virtio_user_pmd_dma_map(struct rte_vdev_device *vdev, void *addr,
+   uint64_t iova, size_t len)
+{
+   const char *name;
+   struct rte_eth_dev *eth_dev;
+   struct virtio_user_dev *dev;
+   struct virtio_hw *hw;
+
+   if (!vdev)
+   return -EINVAL;
+
+   name = rte_vdev_device_name(vdev);
+   eth_dev = rte_eth_dev_allocated(name);
+   /* Port has already been released by close. */
+   if (!eth_dev)
+   return 0;
+
+   hw = (struct virtio_hw *)eth_dev->data->dev_private;
+   dev = hw->virtio_user_dev;
+
+   if (dev->ops->dma_map)
+   return dev->ops->dma_map(dev, addr, iova, len);
+
+   return 0;
+}
+
+static int virtio_user_pmd_dma_unmap(struct rte_vdev_device *vdev, void *addr,
+   uint64_t iova, size_t len)
+{
+   const char *name;
+   struct rte_eth_dev *eth_dev;
+   struct virtio_user_dev *dev;
+   struct virtio_hw *hw;
+
+   if (!vdev)
+   return -EINVAL;
+
+   name = rte_vdev_device_name(vdev);
+   eth_dev = rte_eth_dev_allocated(name);
+   /* Port has already been released by close. */
+   if (!eth_dev)
+   return 0;
+
+   hw = (struct virtio_hw *)eth_dev->data->dev_private;
+   dev = hw->virtio_user_dev;
+
+   if (dev->ops->dma_unmap)
+   return dev->ops->dma_unmap(dev, addr, iova, len);
+
+   return 0;
+}
+
 static struct rte_vdev_driver virtio_user_driver = {
.probe = virtio_user_pmd_probe,
.remove = virtio_user_pmd_remove,
+   .dma_map = virtio_user_pmd_dma_map,
+   .dma_unmap = virtio_user_pmd_dma_unmap,
 };
 
 RTE_PMD_REGISTER_VDEV(net_virtio_user, virtio_user_driver);
-- 
2.26.2



[dpdk-dev] [PATCH v2 1/8] bus/vdev: add DMA mapping ops

2020-09-25 Thread Maxime Coquelin
Add DMA map/unmap operation callbacks to the vdev bus, which
could be used by DMA capable vdev drivers.

Signed-off-by: Maxime Coquelin 
---
 drivers/bus/vdev/rte_bus_vdev.h | 46 +++--
 drivers/bus/vdev/vdev.c | 52 +
 2 files changed, 95 insertions(+), 3 deletions(-)

diff --git a/drivers/bus/vdev/rte_bus_vdev.h b/drivers/bus/vdev/rte_bus_vdev.h
index 78a032cea8..d14eeb41b0 100644
--- a/drivers/bus/vdev/rte_bus_vdev.h
+++ b/drivers/bus/vdev/rte_bus_vdev.h
@@ -63,14 +63,54 @@ typedef int (rte_vdev_probe_t)(struct rte_vdev_device *dev);
  */
 typedef int (rte_vdev_remove_t)(struct rte_vdev_device *dev);
 
+/**
+ * Driver-specific DMA mapping. After a successful call the device
+ * will be able to read/write from/to this segment.
+ *
+ * @param dev
+ *   Pointer to the Virtual device.
+ * @param addr
+ *   Starting virtual address of memory to be mapped.
+ * @param iova
+ *   Starting IOVA address of memory to be mapped.
+ * @param len
+ *   Length of memory segment being mapped.
+ * @return
+ *   - 0 On success.
+ *   - Negative value and rte_errno is set otherwise.
+ */
+typedef int (rte_vdev_dma_map_t)(struct rte_vdev_device *dev, void *addr,
+   uint64_t iova, size_t len);
+
+/**
+ * Driver-specific DMA un-mapping. After a successful call the device
+ * will not be able to read/write from/to this segment.
+ *
+ * @param dev
+ *   Pointer to the Virtual device.
+ * @param addr
+ *   Starting virtual address of memory to be unmapped.
+ * @param iova
+ *   Starting IOVA address of memory to be unmapped.
+ * @param len
+ *   Length of memory segment being unmapped.
+ * @return
+ *   - 0 On success.
+ *   - Negative value and rte_errno is set otherwise.
+ */
+typedef int (rte_vdev_dma_unmap_t)(struct rte_vdev_device *dev, void *addr,
+ uint64_t iova, size_t len);
+
 /**
  * A virtual device driver abstraction.
  */
 struct rte_vdev_driver {
TAILQ_ENTRY(rte_vdev_driver) next; /**< Next in list. */
-   struct rte_driver driver;  /**< Inherited general driver. */
-   rte_vdev_probe_t *probe;   /**< Virtual device probe function. */
-   rte_vdev_remove_t *remove; /**< Virtual device remove function. */
+   struct rte_driver driver;/**< Inherited general driver. */
+   rte_vdev_probe_t *probe; /**< Virtual device probe function. */
+   rte_vdev_remove_t *remove;   /**< Virtual device remove function. */
+   rte_vdev_dma_map_t *dma_map; /**< Virtual device DMA map function. 
*/
+   rte_vdev_dma_unmap_t *dma_unmap; /**< Virtual device DMA unmap 
function. */
 };
 
 /**
diff --git a/drivers/bus/vdev/vdev.c b/drivers/bus/vdev/vdev.c
index d746149a2e..0e94ea9d26 100644
--- a/drivers/bus/vdev/vdev.c
+++ b/drivers/bus/vdev/vdev.c
@@ -134,6 +134,56 @@ vdev_parse(const char *name, void *addr)
return driver == NULL;
 }
 
+static int
+vdev_dma_map(struct rte_device *dev, void *addr, uint64_t iova, size_t len)
+{
+   struct rte_vdev_device *vdev = RTE_DEV_TO_VDEV(dev);
+   const struct rte_vdev_driver *driver;
+
+   if (!vdev) {
+   rte_errno = EINVAL;
+   return -1;
+   }
+
+   if (!vdev->device.driver) {
+   VDEV_LOG(DEBUG, "no driver attach to device %s", dev->name);
+   return 1;
+   }
+
+   driver = container_of(vdev->device.driver, const struct rte_vdev_driver,
+   driver);
+
+   if (driver->dma_map)
+   return driver->dma_map(vdev, addr, iova, len);
+
+   return 0;
+}
+
+static int
+vdev_dma_unmap(struct rte_device *dev, void *addr, uint64_t iova, size_t len)
+{
+   struct rte_vdev_device *vdev = RTE_DEV_TO_VDEV(dev);
+   const struct rte_vdev_driver *driver;
+
+   if (!vdev) {
+   rte_errno = EINVAL;
+   return -1;
+   }
+
+   if (!vdev->device.driver) {
+   VDEV_LOG(DEBUG, "no driver attach to device %s", dev->name);
+   return 1;
+   }
+
+   driver = container_of(vdev->device.driver, const struct rte_vdev_driver,
+   driver);
+
+   if (driver->dma_unmap)
+   return driver->dma_unmap(vdev, addr, iova, len);
+
+   return 0;
+}
+
 static int
 vdev_probe_all_drivers(struct rte_vdev_device *dev)
 {
@@ -551,6 +601,8 @@ static struct rte_bus rte_vdev_bus = {
.plug = vdev_plug,
.unplug = vdev_unplug,
.parse = vdev_parse,
+   .dma_map = vdev_dma_map,
+   .dma_unmap = vdev_dma_unmap,
.dev_iterate = rte_vdev_dev_iterate,
 };
 
-- 
2.26.2



[dpdk-dev] [PATCH v2 0/8] virtio-user: introduce vhost-vdpa backend

2020-09-25 Thread Maxime Coquelin
vhost-vDPA is a new vhost backend type introduced by vDPA kernel
framework, which provides abstruction to the vDPA devices and
exposes to userspace a unified control interface through char devs.

This patch set adds vhost-vdpa backend type to the virtio_user.
A set of vhost-vdpa specific ops callbacks are attached to the
virtio_user according to the runtime dynamic check result of the
backend type. DMA memory map/unmap callbacks are added to both
vdev bus driver and virtio_user pmd to support address mapping.
In addition, minor fixes to existing virtio control path are also
implemented to make the new backend work.

This is a collaborative work done with Patrick Fu from Intel and
Adrian Moreno from Red Hat. Thanks to them for their contributions.

The series has been tested with vdpasim and Intel IFC Kernel vDPA
drivers, and more lightly with Mellanox mlx5_vdpa on ConnectX-6 Dx.

Changes in v2:
-
 * Split backend-type patch (Adrian)
 * Fix get_status size (Chenbo)
 * Various minro fixes (Chenbo)

Adrian Moreno (1):
  net/virtio: move backend type selection to ethdev

Maxime Coquelin (7):
  bus/vdev: add DMA mapping ops
  net/virtio: introduce DMA ops
  net/virtio: introduce Vhost-vDPA backend type
  net/virtio: check protocol feature in user backend
  net/virtio: adapt Virtio-user status size
  net/virtio: split virtio-user start
  net/virtio: introduce Vhost-vDPA backend

 drivers/bus/vdev/rte_bus_vdev.h   |  46 ++-
 drivers/bus/vdev/vdev.c   |  52 +++
 drivers/net/virtio/meson.build|   1 +
 drivers/net/virtio/virtio_user/vhost.h|   5 +
 drivers/net/virtio/virtio_user/vhost_user.c   |   6 +-
 drivers/net/virtio/virtio_user/vhost_vdpa.c   | 298 ++
 .../net/virtio/virtio_user/virtio_user_dev.c  | 117 ---
 .../net/virtio/virtio_user/virtio_user_dev.h  |  13 +-
 drivers/net/virtio/virtio_user_ethdev.c   | 126 +++-
 9 files changed, 607 insertions(+), 57 deletions(-)
 create mode 100644 drivers/net/virtio/virtio_user/vhost_vdpa.c

-- 
2.26.2



[dpdk-dev] [PATCH v2 4/8] net/virtio: introduce Vhost-vDPA backend type

2020-09-25 Thread Maxime Coquelin
Backend type is determined by checking char-device major numbers

Signed-off-by: Maxime Coquelin 
Signed-off-by: Adrian Moreno 
---
 .../net/virtio/virtio_user/virtio_user_dev.h  |  1 +
 drivers/net/virtio/virtio_user_ethdev.c   | 48 +--
 2 files changed, 45 insertions(+), 4 deletions(-)

diff --git a/drivers/net/virtio/virtio_user/virtio_user_dev.h 
b/drivers/net/virtio/virtio_user/virtio_user_dev.h
index 575bf430c0..1c8c98b1cd 100644
--- a/drivers/net/virtio/virtio_user/virtio_user_dev.h
+++ b/drivers/net/virtio/virtio_user/virtio_user_dev.h
@@ -14,6 +14,7 @@ enum virtio_user_backend_type {
VIRTIO_USER_BACKEND_UNKNOWN,
VIRTIO_USER_BACKEND_VHOST_USER,
VIRTIO_USER_BACKEND_VHOST_KERNEL,
+   VIRTIO_USER_BACKEND_VHOST_VDPA,
 };
 
 struct virtio_user_queue {
diff --git a/drivers/net/virtio/virtio_user_ethdev.c 
b/drivers/net/virtio/virtio_user_ethdev.c
index 7bf065f34b..3a51afd711 100644
--- a/drivers/net/virtio/virtio_user_ethdev.c
+++ b/drivers/net/virtio/virtio_user_ethdev.c
@@ -6,7 +6,9 @@
 #include 
 #include 
 #include 
+#include 
 #include 
+#include 
 #include 
 
 #include 
@@ -519,17 +521,55 @@ get_integer_arg(const char *key __rte_unused,
return -errno;
 }
 
+static uint32_t
+vdpa_dynamic_major_num(void)
+{
+   FILE *fp;
+   char *line = NULL;
+   size_t size;
+   char name[11];
+   bool found = false;
+   uint32_t num;
+
+   fp = fopen("/proc/devices", "r");
+   if (fp == NULL) {
+   PMD_INIT_LOG(ERR, "Cannot open /proc/devices: %s",
+strerror(errno));
+   return UNNAMED_MAJOR;
+   }
+
+   while (getline(&line, &size, fp) > 0) {
+   char *stripped = line + strspn(line, " ");
+   if ((sscanf(stripped, "%u %10s", &num, name) == 2) &&
+   (strncmp(name, "vhost-vdpa", 10) == 0)) {
+   found = true;
+   break;
+   }
+   }
+   fclose(fp);
+   return found ? num : UNNAMED_MAJOR;
+}
+
 static enum virtio_user_backend_type
 virtio_user_backend_type(const char *path)
 {
struct stat sb;
 
-   if (stat(path, &sb) == -1)
+   if (stat(path, &sb) == -1) {
+   PMD_INIT_LOG(ERR, "Stat fails: %s (%s)\n", path,
+strerror(errno));
return VIRTIO_USER_BACKEND_UNKNOWN;
+   }
 
-   return S_ISSOCK(sb.st_mode)?
-   VIRTIO_USER_BACKEND_VHOST_USER :
-   VIRTIO_USER_BACKEND_VHOST_KERNEL;
+   if (S_ISSOCK(sb.st_mode)) {
+   return VIRTIO_USER_BACKEND_VHOST_USER;
+   } else if (S_ISCHR(sb.st_mode)) {
+   if (major(sb.st_rdev) == MISC_MAJOR)
+   return VIRTIO_USER_BACKEND_VHOST_KERNEL;
+   if (major(sb.st_rdev) == vdpa_dynamic_major_num())
+   return VIRTIO_USER_BACKEND_VHOST_VDPA;
+   }
+   return VIRTIO_USER_BACKEND_UNKNOWN;
 }
 
 static struct rte_eth_dev *
-- 
2.26.2



Re: [dpdk-dev] [PATCH v3 0/8] fix distributor synchronization issues

2020-09-25 Thread David Marchand
Hello Lukasz,

On Wed, Sep 23, 2020 at 3:25 PM Lukasz Wojciechowski
 wrote:
>
> During review and verification of the patch created by Sarosh Arif:
> "test_distributor: prevent memory leakages from the pool" I found out
> that running distributor unit tests multiple times in a row causes fails.
> So I investigated all the issues I found.
>
> There are few synchronization issues that might cause deadlocks
> or corrupted data. They are fixed with this set of patches for both tests
> and librte_distributor library.
>
> ---
> v3:
> * add missing acked and tested by statements from v1
>
> v2:
> * assign NULL to freed mbufs in distributor test
> * fix handshake check on legacy single distributor
>  rte_distributor_return_pkt_single()
> * add patch 7 passing NULL to legacy API calls if no bufs are returned
> * add patch 8 fixing API documentation
>
> Lukasz Wojciechowski (8):
>   app/test: fix deadlock in distributor test
>   app/test: synchronize statistics between lcores
>   app/test: fix freeing mbufs in distributor tests
>   app/test: collect return mbufs in distributor test

For these patches, we can use the "test/distributor: " prefix, and we
then avoid repeating "in distributor test"


>   distributor: fix missing handshake synchronization
>   distributor: fix handshake deadlock
>   distributor: do not use oldpkt when not needed
>   distributor: align API documentation with code

Thanks for working on those fixes !

Here is a suggestion:

- we can move this new patch 7 before patch 3 in the series, and
update the unit test:
 * by passing NULL to the first call to rte_distributor_get_pkt(),
there is no need for buf[] array init in handle_work(),
handle_work_with_free_mbufs() and handle_work_for_shutdown_test(),
 * at all points of those functions the buf[] array then contains only
[0, num[ valid entries,
 * bonus point, this makes the UT check passing NULL oldpkt,

- the former patch 3 is then easier to do since there is no need for
buf[] array clearing,

This gives the following diff, wdyt?

diff --git a/app/test/test_distributor.c b/app/test/test_distributor.c
index f31b54edf3..b7ab93ecbe 100644
--- a/app/test/test_distributor.c
+++ b/app/test/test_distributor.c
@@ -65,13 +65,10 @@ handle_work(void *arg)
struct rte_mbuf *buf[8] __rte_cache_aligned;
struct worker_params *wp = arg;
struct rte_distributor *db = wp->dist;
-   unsigned int count = 0, num = 0;
+   unsigned int count = 0, num;
unsigned int id = __atomic_fetch_add(&worker_idx, 1, __ATOMIC_RELAXED);
-   int i;

-   for (i = 0; i < 8; i++)
-   buf[i] = NULL;
-   num = rte_distributor_get_pkt(db, id, buf, buf, num);
+   num = rte_distributor_get_pkt(db, id, buf, NULL, 0);
while (!quit) {
__atomic_fetch_add(&worker_stats[id].handled_packets, num,
__ATOMIC_ACQ_REL);
@@ -277,22 +274,17 @@ handle_work_with_free_mbufs(void *arg)
struct rte_mbuf *buf[8] __rte_cache_aligned;
struct worker_params *wp = arg;
struct rte_distributor *d = wp->dist;
+   unsigned int num;
unsigned int i;
-   unsigned int num = 0;
unsigned int id = __atomic_fetch_add(&worker_idx, 1, __ATOMIC_RELAXED);

-   for (i = 0; i < 8; i++)
-   buf[i] = NULL;
-   num = rte_distributor_get_pkt(d, id, buf, buf, 0);
+   num = rte_distributor_get_pkt(d, id, buf, NULL, 0);
while (!quit) {
__atomic_fetch_add(&worker_stats[id].handled_packets, num,
__ATOMIC_ACQ_REL);
-   for (i = 0; i < num; i++) {
+   for (i = 0; i < num; i++)
rte_pktmbuf_free(buf[i]);
-   buf[i] = NULL;
-   }
-   num = rte_distributor_get_pkt(d,
-   id, buf, buf, 0);
+   num = rte_distributor_get_pkt(d, id, buf, NULL, 0);
}
__atomic_fetch_add(&worker_stats[id].handled_packets, num,
__ATOMIC_ACQ_REL);
@@ -347,15 +339,13 @@ handle_work_for_shutdown_test(void *arg)
struct rte_mbuf *buf[8] __rte_cache_aligned;
struct worker_params *wp = arg;
struct rte_distributor *d = wp->dist;
-   unsigned int num = 0;
+   unsigned int num;
unsigned int i;
unsigned int zero_id = 0;
const unsigned int id = __atomic_fetch_add(&worker_idx, 1,
__ATOMIC_RELAXED);
-   for (i = 0; i < 8; i++)
-   buf[i] = NULL;

-   num = rte_distributor_get_pkt(d, id, buf, buf, 0);
+   num = rte_distributor_get_pkt(d, id, buf, NULL, 0);

zero_id = __atomic_load_n(&zero_idx, __ATOMIC_ACQUIRE);
if (id == zero_id && num > 0) {
@@ -369,12 +359,9 @@ handle_work_for_shutdown_test(void *arg)
while (!quit && !(id == zero_id && zero_quit)) {
__atomic_fetch_add(&worker_stats[id].handled_packets, num,
   

[dpdk-dev] [PATCH v2 5/8] net/virtio: check protocol feature in user backend

2020-09-25 Thread Maxime Coquelin
When sending set status message, move protocol feature check
to vhost_user to be compatible with different backend types.

Signed-off-by: Maxime Coquelin 
---
 drivers/net/virtio/virtio_user/vhost_user.c  | 6 +-
 drivers/net/virtio/virtio_user/virtio_user_dev.c | 6 --
 2 files changed, 5 insertions(+), 7 deletions(-)

diff --git a/drivers/net/virtio/virtio_user/vhost_user.c 
b/drivers/net/virtio/virtio_user/vhost_user.c
index 12b6c7dbcf..ef290c357b 100644
--- a/drivers/net/virtio/virtio_user/vhost_user.c
+++ b/drivers/net/virtio/virtio_user/vhost_user.c
@@ -277,9 +277,13 @@ vhost_user_sock(struct virtio_user_dev *dev,
msg.size = 0;
 
switch (req) {
+   case VHOST_USER_GET_STATUS:
+   if (!(dev->protocol_features &
+   (1ULL << VHOST_USER_PROTOCOL_F_STATUS)))
+   return 0;
+   /* Fallthrough */
case VHOST_USER_GET_FEATURES:
case VHOST_USER_GET_PROTOCOL_FEATURES:
-   case VHOST_USER_GET_STATUS:
need_reply = 1;
break;
 
diff --git a/drivers/net/virtio/virtio_user/virtio_user_dev.c 
b/drivers/net/virtio/virtio_user/virtio_user_dev.c
index 7cb93b1b33..0128a5fefb 100644
--- a/drivers/net/virtio/virtio_user/virtio_user_dev.c
+++ b/drivers/net/virtio/virtio_user/virtio_user_dev.c
@@ -788,9 +788,6 @@ virtio_user_send_status_update(struct virtio_user_dev *dev, 
uint8_t status)
if (dev->backend_type != VIRTIO_USER_BACKEND_VHOST_USER)
return 0;
 
-   if (!(dev->protocol_features & (1ULL << VHOST_USER_PROTOCOL_F_STATUS)))
-   return 0;
-
ret = dev->ops->send_request(dev, VHOST_USER_SET_STATUS, &arg);
if (ret) {
PMD_INIT_LOG(ERR, "VHOST_USER_SET_STATUS failed (%d): %s", ret,
@@ -811,9 +808,6 @@ virtio_user_update_status(struct virtio_user_dev *dev)
if (dev->backend_type != VIRTIO_USER_BACKEND_VHOST_USER)
return 0;
 
-   if (!(dev->protocol_features & (1UL << VHOST_USER_PROTOCOL_F_STATUS)))
-   return 0;
-
err = dev->ops->send_request(dev, VHOST_USER_GET_STATUS, &ret);
if (err) {
PMD_INIT_LOG(ERR, "VHOST_USER_GET_STATUS failed (%d): %s", err,
-- 
2.26.2



[dpdk-dev] [PATCH v2 8/8] net/virtio: introduce Vhost-vDPA backend

2020-09-25 Thread Maxime Coquelin
vhost-vDPA is a new virtio backend type introduced by vDPA kernel
framework, which provides abstruction to the vDPA devices and
exposes an unified control interface through a char dev.

This patch adds support to the vhost-vDPA backend. As similar to
the existing vhost kernel backend, a set of virtio_user ops were
introduced for vhost-vDPA backend to handle device specific operations
such as:
 - device setup
 - ioctl message handling
 - queue pair enabling
 - dma map/unmap
vDPA relevant ioctl codes and data structures are also defined in
this patch.

Signed-off-by: Maxime Coquelin 
---
 drivers/net/virtio/meson.build|   1 +
 drivers/net/virtio/virtio_user/vhost.h|   1 +
 drivers/net/virtio/virtio_user/vhost_vdpa.c   | 298 ++
 .../net/virtio/virtio_user/virtio_user_dev.c  |   6 +
 4 files changed, 306 insertions(+)
 create mode 100644 drivers/net/virtio/virtio_user/vhost_vdpa.c

diff --git a/drivers/net/virtio/meson.build b/drivers/net/virtio/meson.build
index 3fd6051f4b..eaed46373d 100644
--- a/drivers/net/virtio/meson.build
+++ b/drivers/net/virtio/meson.build
@@ -42,6 +42,7 @@ if is_linux
'virtio_user/vhost_kernel.c',
'virtio_user/vhost_kernel_tap.c',
'virtio_user/vhost_user.c',
+   'virtio_user/vhost_vdpa.c',
'virtio_user/virtio_user_dev.c')
deps += ['bus_vdev']
 endif
diff --git a/drivers/net/virtio/virtio_user/vhost.h 
b/drivers/net/virtio/virtio_user/vhost.h
index 2e71995a79..210a3704e7 100644
--- a/drivers/net/virtio/virtio_user/vhost.h
+++ b/drivers/net/virtio/virtio_user/vhost.h
@@ -113,5 +113,6 @@ struct virtio_user_backend_ops {
 
 extern struct virtio_user_backend_ops virtio_ops_user;
 extern struct virtio_user_backend_ops virtio_ops_kernel;
+extern struct virtio_user_backend_ops virtio_ops_vdpa;
 
 #endif
diff --git a/drivers/net/virtio/virtio_user/vhost_vdpa.c 
b/drivers/net/virtio/virtio_user/vhost_vdpa.c
new file mode 100644
index 00..d592ca8817
--- /dev/null
+++ b/drivers/net/virtio/virtio_user/vhost_vdpa.c
@@ -0,0 +1,298 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2020 Red Hat Inc.
+ */
+
+#include 
+#include 
+#include 
+#include 
+#include 
+
+#include 
+
+#include "vhost.h"
+#include "virtio_user_dev.h"
+
+/* vhost kernel & vdpa ioctls */
+#define VHOST_VIRTIO 0xAF
+#define VHOST_GET_FEATURES _IOR(VHOST_VIRTIO, 0x00, __u64)
+#define VHOST_SET_FEATURES _IOW(VHOST_VIRTIO, 0x00, __u64)
+#define VHOST_SET_OWNER _IO(VHOST_VIRTIO, 0x01)
+#define VHOST_RESET_OWNER _IO(VHOST_VIRTIO, 0x02)
+#define VHOST_SET_MEM_TABLE _IOW(VHOST_VIRTIO, 0x03, void *)
+#define VHOST_SET_LOG_BASE _IOW(VHOST_VIRTIO, 0x04, __u64)
+#define VHOST_SET_LOG_FD _IOW(VHOST_VIRTIO, 0x07, int)
+#define VHOST_SET_VRING_NUM _IOW(VHOST_VIRTIO, 0x10, struct vhost_vring_state)
+#define VHOST_SET_VRING_ADDR _IOW(VHOST_VIRTIO, 0x11, struct vhost_vring_addr)
+#define VHOST_SET_VRING_BASE _IOW(VHOST_VIRTIO, 0x12, struct vhost_vring_state)
+#define VHOST_GET_VRING_BASE _IOWR(VHOST_VIRTIO, 0x12, struct 
vhost_vring_state)
+#define VHOST_SET_VRING_KICK _IOW(VHOST_VIRTIO, 0x20, struct vhost_vring_file)
+#define VHOST_SET_VRING_CALL _IOW(VHOST_VIRTIO, 0x21, struct vhost_vring_file)
+#define VHOST_SET_VRING_ERR _IOW(VHOST_VIRTIO, 0x22, struct vhost_vring_file)
+#define VHOST_NET_SET_BACKEND _IOW(VHOST_VIRTIO, 0x30, struct vhost_vring_file)
+#define VHOST_VDPA_GET_DEVICE_ID _IOR(VHOST_VIRTIO, 0x70, __u32)
+#define VHOST_VDPA_GET_STATUS _IOR(VHOST_VIRTIO, 0x71, __u8)
+#define VHOST_VDPA_SET_STATUS _IOW(VHOST_VIRTIO, 0x72, __u8)
+#define VHOST_VDPA_SET_VRING_ENABLE_IOW(VHOST_VIRTIO, 0x75, \
+struct vhost_vring_state)
+
+static uint64_t vhost_req_user_to_vdpa[] = {
+   [VHOST_USER_SET_OWNER] = VHOST_SET_OWNER,
+   [VHOST_USER_RESET_OWNER] = VHOST_RESET_OWNER,
+   [VHOST_USER_SET_FEATURES] = VHOST_SET_FEATURES,
+   [VHOST_USER_GET_FEATURES] = VHOST_GET_FEATURES,
+   [VHOST_USER_SET_VRING_CALL] = VHOST_SET_VRING_CALL,
+   [VHOST_USER_SET_VRING_NUM] = VHOST_SET_VRING_NUM,
+   [VHOST_USER_SET_VRING_BASE] = VHOST_SET_VRING_BASE,
+   [VHOST_USER_GET_VRING_BASE] = VHOST_GET_VRING_BASE,
+   [VHOST_USER_SET_VRING_ADDR] = VHOST_SET_VRING_ADDR,
+   [VHOST_USER_SET_VRING_KICK] = VHOST_SET_VRING_KICK,
+   [VHOST_USER_SET_MEM_TABLE] = VHOST_SET_MEM_TABLE,
+   [VHOST_USER_SET_STATUS] = VHOST_VDPA_SET_STATUS,
+   [VHOST_USER_GET_STATUS] = VHOST_VDPA_GET_STATUS,
+   [VHOST_USER_SET_VRING_ENABLE] = VHOST_VDPA_SET_VRING_ENABLE,
+};
+
+/* no alignment requirement */
+struct vhost_iotlb_msg {
+   uint64_t iova;
+   uint64_t size;
+   uint64_t uaddr;
+#define VHOST_ACCESS_RO  0x1
+#define VHOST_ACCESS_WO  0x2
+#define VHOST_ACCESS_RW  0x3
+   uint8_t perm;
+#define VHOST_IOTLB_MISS   1
+#define VHOST_IOTLB_UPDATE 2
+#define VHOST_IOTLB_INVALIDATE 3
+#define VHOST_IOTL

[dpdk-dev] [PATCH v2 7/8] net/virtio: split virtio-user start

2020-09-25 Thread Maxime Coquelin
Move feature bit settings in device start out as an standalone
function, so that feature bit could be negotiated at device
feature_ok status.

Signed-off-by: Maxime Coquelin 
---
 .../net/virtio/virtio_user/virtio_user_dev.c  | 50 ---
 .../net/virtio/virtio_user/virtio_user_dev.h  |  1 +
 drivers/net/virtio/virtio_user_ethdev.c   |  4 ++
 3 files changed, 37 insertions(+), 18 deletions(-)

diff --git a/drivers/net/virtio/virtio_user/virtio_user_dev.c 
b/drivers/net/virtio/virtio_user/virtio_user_dev.c
index 5988bc92fe..81257a3996 100644
--- a/drivers/net/virtio/virtio_user/virtio_user_dev.c
+++ b/drivers/net/virtio/virtio_user/virtio_user_dev.c
@@ -112,25 +112,11 @@ virtio_user_queue_setup(struct virtio_user_dev *dev,
 }
 
 int
-virtio_user_start_device(struct virtio_user_dev *dev)
+virtio_user_dev_set_features(struct virtio_user_dev *dev)
 {
uint64_t features;
-   int ret;
+   int ret = -1;
 
-   /*
-* XXX workaround!
-*
-* We need to make sure that the locks will be
-* taken in the correct order to avoid deadlocks.
-*
-* Before releasing this lock, this thread should
-* not trigger any memory hotplug events.
-*
-* This is a temporary workaround, and should be
-* replaced when we get proper supports from the
-* memory subsystem in the future.
-*/
-   rte_mcfg_mem_read_lock();
pthread_mutex_lock(&dev->mutex);
 
if (dev->backend_type == VIRTIO_USER_BACKEND_VHOST_USER
@@ -141,10 +127,8 @@ virtio_user_start_device(struct virtio_user_dev *dev)
if (virtio_user_queue_setup(dev, virtio_user_create_queue) < 0)
goto error;
 
-   /* Step 1: negotiate protocol features & set features */
features = dev->features;
 
-
/* Strip VIRTIO_NET_F_MAC, as MAC address is handled in vdev init */
features &= ~(1ull << VIRTIO_NET_F_MAC);
/* Strip VIRTIO_NET_F_CTRL_VQ, as devices do not really need to know */
@@ -154,6 +138,36 @@ virtio_user_start_device(struct virtio_user_dev *dev)
if (ret < 0)
goto error;
PMD_DRV_LOG(INFO, "set features: %" PRIx64, features);
+error:
+   pthread_mutex_unlock(&dev->mutex);
+
+   return ret;
+}
+
+int
+virtio_user_start_device(struct virtio_user_dev *dev)
+{
+   int ret;
+
+   /*
+* XXX workaround!
+*
+* We need to make sure that the locks will be
+* taken in the correct order to avoid deadlocks.
+*
+* Before releasing this lock, this thread should
+* not trigger any memory hotplug events.
+*
+* This is a temporary workaround, and should be
+* replaced when we get proper supports from the
+* memory subsystem in the future.
+*/
+   rte_mcfg_mem_read_lock();
+   pthread_mutex_lock(&dev->mutex);
+
+   if (dev->backend_type == VIRTIO_USER_BACKEND_VHOST_USER &&
+   dev->vhostfd < 0)
+   goto error;
 
/* Step 2: share memory regions */
ret = dev->ops->send_request(dev, VHOST_USER_SET_MEM_TABLE, NULL);
diff --git a/drivers/net/virtio/virtio_user/virtio_user_dev.h 
b/drivers/net/virtio/virtio_user/virtio_user_dev.h
index 1c8c98b1cd..3e9d1a1eb3 100644
--- a/drivers/net/virtio/virtio_user/virtio_user_dev.h
+++ b/drivers/net/virtio/virtio_user/virtio_user_dev.h
@@ -68,6 +68,7 @@ struct virtio_user_dev {
boolstarted;
 };
 
+int virtio_user_dev_set_features(struct virtio_user_dev *dev);
 int virtio_user_start_device(struct virtio_user_dev *dev);
 int virtio_user_stop_device(struct virtio_user_dev *dev);
 int virtio_user_dev_init(struct virtio_user_dev *dev, char *path, int queues,
diff --git a/drivers/net/virtio/virtio_user_ethdev.c 
b/drivers/net/virtio/virtio_user_ethdev.c
index 3a51afd711..f56fc238c4 100644
--- a/drivers/net/virtio/virtio_user_ethdev.c
+++ b/drivers/net/virtio/virtio_user_ethdev.c
@@ -269,7 +269,11 @@ static void
 virtio_user_set_status(struct virtio_hw *hw, uint8_t status)
 {
struct virtio_user_dev *dev = virtio_user_get_dev(hw);
+   uint8_t old_status = dev->status;
 
+   if (status & VIRTIO_CONFIG_STATUS_FEATURES_OK &&
+   ~old_status & VIRTIO_CONFIG_STATUS_FEATURES_OK)
+   virtio_user_dev_set_features(dev);
if (status & VIRTIO_CONFIG_STATUS_DRIVER_OK)
virtio_user_start_device(dev);
else if (status == VIRTIO_CONFIG_STATUS_RESET)
-- 
2.26.2



[dpdk-dev] [PATCH v2 6/8] net/virtio: adapt Virtio-user status size

2020-09-25 Thread Maxime Coquelin
Set proper payload size for set/get status message. The payload
size varies according to backend types.

Signed-off-by: Maxime Coquelin 
Signed-off-by: Patrick Fu 
---
 .../net/virtio/virtio_user/virtio_user_dev.c  | 34 +--
 1 file changed, 23 insertions(+), 11 deletions(-)

diff --git a/drivers/net/virtio/virtio_user/virtio_user_dev.c 
b/drivers/net/virtio/virtio_user/virtio_user_dev.c
index 0128a5fefb..5988bc92fe 100644
--- a/drivers/net/virtio/virtio_user/virtio_user_dev.c
+++ b/drivers/net/virtio/virtio_user/virtio_user_dev.c
@@ -784,11 +784,15 @@ virtio_user_send_status_update(struct virtio_user_dev 
*dev, uint8_t status)
int ret;
uint64_t arg = status;
 
-   /* Vhost-user only for now */
-   if (dev->backend_type != VIRTIO_USER_BACKEND_VHOST_USER)
+   if (dev->backend_type == VIRTIO_USER_BACKEND_VHOST_USER)
+   ret = dev->ops->send_request(dev,
+   VHOST_USER_SET_STATUS, &arg);
+   else if (dev->backend_type == VIRTIO_USER_BACKEND_VHOST_VDPA)
+   ret = dev->ops->send_request(dev,
+   VHOST_USER_SET_STATUS, &status);
+   else
return 0;
 
-   ret = dev->ops->send_request(dev, VHOST_USER_SET_STATUS, &arg);
if (ret) {
PMD_INIT_LOG(ERR, "VHOST_USER_SET_STATUS failed (%d): %s", ret,
 strerror(errno));
@@ -802,24 +806,32 @@ int
 virtio_user_update_status(struct virtio_user_dev *dev)
 {
uint64_t ret;
+   uint8_t status;
int err;
 
-   /* Vhost-user only for now */
-   if (dev->backend_type != VIRTIO_USER_BACKEND_VHOST_USER)
+   if (dev->backend_type == VIRTIO_USER_BACKEND_VHOST_USER) {
+   err = dev->ops->send_request(dev, VHOST_USER_GET_STATUS, &ret);
+   if (!err && ret > UINT8_MAX) {
+   PMD_INIT_LOG(ERR, "Invalid VHOST_USER_GET_STATUS "
+   "response 0x%" PRIx64 "\n", ret);
+   return -1;
+   }
+
+   status = ret;
+   } else if (dev->backend_type == VIRTIO_USER_BACKEND_VHOST_VDPA) {
+   err = dev->ops->send_request(dev, VHOST_USER_GET_STATUS,
+   &status);
+   } else {
return 0;
+   }
 
-   err = dev->ops->send_request(dev, VHOST_USER_GET_STATUS, &ret);
if (err) {
PMD_INIT_LOG(ERR, "VHOST_USER_GET_STATUS failed (%d): %s", err,
 strerror(errno));
return -1;
}
-   if (ret > UINT8_MAX) {
-   PMD_INIT_LOG(ERR, "Invalid VHOST_USER_GET_STATUS response 0x%" 
PRIx64 "\n", ret);
-   return -1;
-   }
 
-   dev->status = ret;
+   dev->status = status;
PMD_INIT_LOG(DEBUG, "Updated Device Status(0x%08x):\n"
"\t-RESET: %u\n"
"\t-ACKNOWLEDGE: %u\n"
-- 
2.26.2



Re: [dpdk-dev] [PATCH] net/ring: fix unchecked return value

2020-09-25 Thread Ferruh Yigit

On 9/22/2020 6:20 PM, Kevin Laatz wrote:

Add a check for the return value of the sscanf call in
parse_internal_args(), returning an error if we don't get the expected
result.

Coverity issue: 362049
Fixes: 96cb19521147 ("net/ring: use EAL APIs in PMD specific API")
Cc: sta...@dpdk.org

Signed-off-by: Kevin Laatz 
---
  drivers/net/ring/rte_eth_ring.c | 3 ++-
  1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/drivers/net/ring/rte_eth_ring.c b/drivers/net/ring/rte_eth_ring.c
index 40fe1ca4ba..62060e46ce 100644
--- a/drivers/net/ring/rte_eth_ring.c
+++ b/drivers/net/ring/rte_eth_ring.c
@@ -539,7 +539,8 @@ parse_internal_args(const char *key __rte_unused, const 
char *value,
struct ring_internal_args **internal_args = data;
void *args;
  
-	sscanf(value, "%p", &args);

+   if (sscanf(value, "%p", &args) != 1)
+   return -1;


I am aware that this is just to fix the coverity error to check the 
return value, BUT :)


The internal args is mainly to pass the information get by API 
('rte_eth_from_ring()') to ring probe. And the main information to pass 
is the ring.
It may be possible to pass the ring_name only and eliminate internal 
args completely, the driver already has a way to pass ring name:

"nodeaction=name:node:ATTACH" devargs.

If you have time, can you please check if it can be possible to fill and 
pass the nodeaction devargs in 'rte_eth_from_rings()' and eliminate the 
'internal' devargs completely?
If it works, as a bonus it will resolve above coverity issue by removing 
it :)




[dpdk-dev] [PATCH v4 2/6] app/testpmd: fix VLAN offload configuration when config fail

2020-09-25 Thread Wei Hu (Xavier)
From: Chengchang Tang 

When failing to configure VLAN offloads after the port was started, there
is no need to update the port configuration. Currently, when user
configure an unsupported VLAN offloads and fails, and then restart the
port, it will fails since the configuration has been refreshed.

This patch makes the function return directly insead of refreshing the
configuration when execution fails.

Fixes: 384161e00627 ("app/testpmd: adjust on the fly VLAN configuration")
Cc: sta...@dpdk.org

Signed-off-by: Chengchang Tang 
Signed-off-by: Wei Hu (Xavier) 
Reviewed-by: Ferruh Yigit 
---
 app/test-pmd/config.c | 16 
 1 file changed, 12 insertions(+), 4 deletions(-)

diff --git a/app/test-pmd/config.c b/app/test-pmd/config.c
index 2ee40df..6496d2f 100644
--- a/app/test-pmd/config.c
+++ b/app/test-pmd/config.c
@@ -3392,9 +3392,11 @@ vlan_extend_set(portid_t port_id, int on)
}
 
diag = rte_eth_dev_set_vlan_offload(port_id, vlan_offload);
-   if (diag < 0)
+   if (diag < 0) {
printf("rx_vlan_extend_set(port_pi=%d, on=%d) failed "
   "diag=%d\n", port_id, on, diag);
+   return;
+   }
ports[port_id].dev_conf.rxmode.offloads = port_rx_offloads;
 }
 
@@ -3419,9 +3421,11 @@ rx_vlan_strip_set(portid_t port_id, int on)
}
 
diag = rte_eth_dev_set_vlan_offload(port_id, vlan_offload);
-   if (diag < 0)
+   if (diag < 0) {
printf("rx_vlan_strip_set(port_pi=%d, on=%d) failed "
   "diag=%d\n", port_id, on, diag);
+   return;
+   }
ports[port_id].dev_conf.rxmode.offloads = port_rx_offloads;
 }
 
@@ -3460,9 +3464,11 @@ rx_vlan_filter_set(portid_t port_id, int on)
}
 
diag = rte_eth_dev_set_vlan_offload(port_id, vlan_offload);
-   if (diag < 0)
+   if (diag < 0) {
printf("rx_vlan_filter_set(port_pi=%d, on=%d) failed "
   "diag=%d\n", port_id, on, diag);
+   return;
+   }
ports[port_id].dev_conf.rxmode.offloads = port_rx_offloads;
 }
 
@@ -3487,9 +3493,11 @@ rx_vlan_qinq_strip_set(portid_t port_id, int on)
}
 
diag = rte_eth_dev_set_vlan_offload(port_id, vlan_offload);
-   if (diag < 0)
+   if (diag < 0) {
printf("%s(port_pi=%d, on=%d) failed "
   "diag=%d\n", __func__, port_id, on, diag);
+   return;
+   }
ports[port_id].dev_conf.rxmode.offloads = port_rx_offloads;
 }
 
-- 
2.9.5



[dpdk-dev] [PATCH v4 0/6] minor fixes for testpmd

2020-09-25 Thread Wei Hu (Xavier)
This series are minor fixes for testpmd application.

Chengchang Tang (5):
  app/testpmd: fix missing verification of port id
  app/testpmd: fix VLAN offload configuration when config fail
  app/testpmd: remove restriction on txpkts set
  app/testpmd: fix packet header in txonly mode
  app/testpmd: fix valid desc id check

Huisong Li (1):
  app/testpmd: fix displaying Rx Tx queues information

 app/test-pmd/cmdline.c |   9 ++
 app/test-pmd/config.c  | 226 ++---
 app/test-pmd/txonly.c  |  32 +++
 3 files changed, 218 insertions(+), 49 deletions(-)

-- 
2.9.5



[dpdk-dev] [PATCH v4 1/6] app/testpmd: fix missing verification of port id

2020-09-25 Thread Wei Hu (Xavier)
From: Chengchang Tang 

To set Tx vlan offloads, it is required to stop port firstly. But before
checking whether the port is stopped, the port id entered by the user
is not checked for validity. When the port id is illegal, it would lead
to a segmentation fault since it attempts to access a member of
non-existent port.

This patch adds verification of port id in tx vlan offloads and remove
duplicated check.

Fixes: 597f9fafe13b ("app/testpmd: convert to new Tx offloads API")
Cc: sta...@dpdk.org

Signed-off-by: Chengchang Tang 
Signed-off-by: Wei Hu (Xavier) 
Reviewed-by: Ferruh Yigit 
---
 app/test-pmd/cmdline.c | 9 +
 app/test-pmd/config.c  | 6 --
 2 files changed, 9 insertions(+), 6 deletions(-)

diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c
index 5f93409..43b7c2f 100644
--- a/app/test-pmd/cmdline.c
+++ b/app/test-pmd/cmdline.c
@@ -4291,6 +4291,9 @@ cmd_tx_vlan_set_parsed(void *parsed_result,
 {
struct cmd_tx_vlan_set_result *res = parsed_result;
 
+   if (port_id_is_invalid(res->port_id, ENABLED_WARN))
+   return;
+
if (!port_is_stopped(res->port_id)) {
printf("Please stop port %d first\n", res->port_id);
return;
@@ -4345,6 +4348,9 @@ cmd_tx_vlan_set_qinq_parsed(void *parsed_result,
 {
struct cmd_tx_vlan_set_qinq_result *res = parsed_result;
 
+   if (port_id_is_invalid(res->port_id, ENABLED_WARN))
+   return;
+
if (!port_is_stopped(res->port_id)) {
printf("Please stop port %d first\n", res->port_id);
return;
@@ -4458,6 +4464,9 @@ cmd_tx_vlan_reset_parsed(void *parsed_result,
 {
struct cmd_tx_vlan_reset_result *res = parsed_result;
 
+   if (port_id_is_invalid(res->port_id, ENABLED_WARN))
+   return;
+
if (!port_is_stopped(res->port_id)) {
printf("Please stop port %d first\n", res->port_id);
return;
diff --git a/app/test-pmd/config.c b/app/test-pmd/config.c
index 2d9a456..2ee40df 100644
--- a/app/test-pmd/config.c
+++ b/app/test-pmd/config.c
@@ -3547,8 +3547,6 @@ tx_vlan_set(portid_t port_id, uint16_t vlan_id)
struct rte_eth_dev_info dev_info;
int ret;
 
-   if (port_id_is_invalid(port_id, ENABLED_WARN))
-   return;
if (vlan_id_is_invalid(vlan_id))
return;
 
@@ -3579,8 +3577,6 @@ tx_qinq_set(portid_t port_id, uint16_t vlan_id, uint16_t 
vlan_id_outer)
struct rte_eth_dev_info dev_info;
int ret;
 
-   if (port_id_is_invalid(port_id, ENABLED_WARN))
-   return;
if (vlan_id_is_invalid(vlan_id))
return;
if (vlan_id_is_invalid(vlan_id_outer))
@@ -3606,8 +3602,6 @@ tx_qinq_set(portid_t port_id, uint16_t vlan_id, uint16_t 
vlan_id_outer)
 void
 tx_vlan_reset(portid_t port_id)
 {
-   if (port_id_is_invalid(port_id, ENABLED_WARN))
-   return;
ports[port_id].dev_conf.txmode.offloads &=
~(DEV_TX_OFFLOAD_VLAN_INSERT |
  DEV_TX_OFFLOAD_QINQ_INSERT);
-- 
2.9.5



[dpdk-dev] [PATCH v4 4/6] app/testpmd: fix packet header in txonly mode

2020-09-25 Thread Wei Hu (Xavier)
From: Chengchang Tang 

In txonly forward mode, the packet header is fixed by the initial
setting, including the packet length and checksum. So when the packets
varies, this may cause a packet header error. Currently, there are two
methods in txonly mode to randomly change the packets.
1. Set txsplit random and txpkts (x[,y]*), the number of segments
   each packets will be a random value between 1 and total number of
   segments determined by txpkts settings.
   The step as follows:
 a) ./testpmd -w xxx -l xx -n 4 -- -i --disable-device-start
 b) port config 0 tx_offload multi_segs on
 c) set fwd txonly
 d) set txsplit rand
 e) set txpkts 2048,2048,2048,2048
 f) start
The nb_segs of the packets sent by testpmd will be 1~4. The real packet
length will be 2048, 4096, 6144 and 8192. But in fact the packet length
in ip header and udp header will be fixed by 8178 and 8158.

2. Set txonly-multi-flow. the ip address will be varied to generate
   multiple flow.
   The step as follows:
 a) ./testpmd -w xxx -l xx -n 4 -- -i --txonly-multi-flow
 b) set fwd txonly
 c) start
The ip address of each pkts will change randomly, but since the header
is fixed, the checksum may be a error value.

Therefore, this patch adds a function to update the packet length and
check sum in the pkts header when the txsplit mode is set to rand or
multi-flow is set.

Fixes: 82010ef55e7c ("app/testpmd: make txonly mode generate multiple flows")
Fixes: 79bec05b32b7 ("app/testpmd: add ability to split outgoing packets")
Cc: sta...@dpdk.org

Signed-off-by: Chengchang Tang 
Signed-off-by: Wei Hu (Xavier) 
---
 app/test-pmd/txonly.c | 32 
 1 file changed, 32 insertions(+)

diff --git a/app/test-pmd/txonly.c b/app/test-pmd/txonly.c
index 45def72..d55ee7c 100644
--- a/app/test-pmd/txonly.c
+++ b/app/test-pmd/txonly.c
@@ -156,6 +156,34 @@ setup_pkt_udp_ip_headers(struct rte_ipv4_hdr *ip_hdr,
ip_hdr->hdr_checksum = (uint16_t) ip_cksum;
 }
 
+static inline void
+update_pkt_header(struct rte_mbuf *pkt, uint32_t total_pkt_len)
+{
+   struct rte_ipv4_hdr *ip_hdr;
+   struct rte_udp_hdr *udp_hdr;
+   uint16_t pkt_data_len;
+   uint16_t pkt_len;
+
+   pkt_data_len = (uint16_t) (total_pkt_len - (
+   sizeof(struct rte_ether_hdr) +
+   sizeof(struct rte_ipv4_hdr) +
+   sizeof(struct rte_udp_hdr)));
+   /* updata udp pkt length */
+   udp_hdr = rte_pktmbuf_mtod_offset(pkt, struct rte_udp_hdr *,
+   sizeof(struct rte_ether_hdr) +
+   sizeof(struct rte_ipv4_hdr));
+   pkt_len = (uint16_t) (pkt_data_len + sizeof(struct rte_udp_hdr));
+   udp_hdr->dgram_len = RTE_CPU_TO_BE_16(pkt_len);
+
+   /* updata ip pkt length and csum */
+   ip_hdr = rte_pktmbuf_mtod_offset(pkt, struct rte_ipv4_hdr *,
+   sizeof(struct rte_ether_hdr));
+   ip_hdr->hdr_checksum = 0;
+   pkt_len = (uint16_t) (pkt_len + sizeof(struct rte_ipv4_hdr));
+   ip_hdr->total_length = RTE_CPU_TO_BE_16(pkt_len);
+   ip_hdr->hdr_checksum = rte_ipv4_cksum(ip_hdr);
+}
+
 static inline bool
 pkt_burst_prepare(struct rte_mbuf *pkt, struct rte_mempool *mbp,
struct rte_ether_hdr *eth_hdr, const uint16_t vlan_tci,
@@ -223,6 +251,10 @@ pkt_burst_prepare(struct rte_mbuf *pkt, struct rte_mempool 
*mbp,
copy_buf_to_pkt(&pkt_udp_hdr, sizeof(pkt_udp_hdr), pkt,
sizeof(struct rte_ether_hdr) +
sizeof(struct rte_ipv4_hdr));
+
+   if (unlikely(tx_pkt_split == TX_PKT_SPLIT_RND) || txonly_multi_flow)
+   update_pkt_header(pkt, pkt_len);
+
if (unlikely(timestamp_enable)) {
uint64_t skew = RTE_PER_LCORE(timestamp_qskew);
struct {
-- 
2.9.5



[dpdk-dev] [PATCH v4 3/6] app/testpmd: remove restriction on txpkts set

2020-09-25 Thread Wei Hu (Xavier)
From: Chengchang Tang 

Currently, if nb_txd is not set, the txpkts is not allowed to be set
because the nb_txd is used to avoid the numer of segments exceed the Tx
ring size and the default value of nb_txd is 0. And there is a bug that
nb_txd is the global configuration for Tx ring size and the ring size
could be changed by some command per queue. So these valid check is
unreliable and introduced unnecessary constraints.

This patch adds a valid check function to use the real Tx ring size to
check the validity of txpkts.

Fixes: af75078fece3 ("first public release")
Cc: sta...@dpdk.org

Signed-off-by: Chengchang Tang 
Signed-off-by: Wei Hu (Xavier) 
---
v3 -> v4:
 add check 'rte_eth_rx_queue_info_get()' return value and
 if it is '-ENOSTUP' calculate the 'ring_size'.
v3:  initial version.
---
 app/test-pmd/config.c | 64 +++
 1 file changed, 60 insertions(+), 4 deletions(-)

diff --git a/app/test-pmd/config.c b/app/test-pmd/config.c
index 6496d2f..8ebb927 100644
--- a/app/test-pmd/config.c
+++ b/app/test-pmd/config.c
@@ -1893,6 +1893,38 @@ tx_queue_id_is_invalid(queueid_t txq_id)
 }
 
 static int
+get_tx_ring_size(portid_t port_id, queueid_t txq_id, uint16_t *ring_size)
+{
+   struct rte_port *port = &ports[port_id];
+   struct rte_eth_txq_info tx_qinfo;
+   int ret;
+
+   ret = rte_eth_tx_queue_info_get(port_id, txq_id, &tx_qinfo);
+   if (ret == 0) {
+   *ring_size = tx_qinfo.nb_desc;
+   return ret;
+   }
+
+   if (ret != -ENOTSUP)
+   return ret;
+   /*
+* If the rte_eth_tx_queue_info_get is not support for this PMD,
+* ring_size stored in testpmd will be used for validity verification.
+* When configure the txq by rte_eth_tx_queue_setup with nb_tx_desc
+* being 0, it will use a default value provided by PMDs to setup this
+* txq. If the default value is 0, it will use the
+* RTE_ETH_DEV_FALLBACK_TX_RINGSIZE to setup this txq.
+*/
+   if (port->nb_tx_desc[txq_id])
+   *ring_size = port->nb_tx_desc[txq_id];
+   else if (port->dev_info.default_txportconf.ring_size)
+   *ring_size = port->dev_info.default_txportconf.ring_size;
+   else
+   *ring_size = RTE_ETH_DEV_FALLBACK_TX_RINGSIZE;
+   return 0;
+}
+
+static int
 rx_desc_id_is_invalid(uint16_t rxdesc_id)
 {
if (rxdesc_id < nb_rxd)
@@ -2986,17 +3018,41 @@ show_tx_pkt_segments(void)
printf("Split packet: %s\n", split);
 }
 
+static bool
+nb_segs_is_invalid(unsigned int nb_segs)
+{
+   uint16_t ring_size;
+   uint16_t queue_id;
+   uint16_t port_id;
+   int ret;
+
+   RTE_ETH_FOREACH_DEV(port_id) {
+   for (queue_id = 0; queue_id < nb_txq; queue_id++) {
+   ret = get_tx_ring_size(port_id, queue_id, &ring_size);
+
+   if (ret)
+   return true;
+
+   if (ring_size < nb_segs) {
+   printf("nb segments per TX packets=%u >= "
+  "TX queue(%u) ring_size=%u - ignored\n",
+  nb_segs, queue_id, ring_size);
+   return true;
+   }
+   }
+   }
+
+   return false;
+}
+
 void
 set_tx_pkt_segments(unsigned *seg_lengths, unsigned nb_segs)
 {
uint16_t tx_pkt_len;
unsigned i;
 
-   if (nb_segs >= (unsigned) nb_txd) {
-   printf("nb segments per TX packets=%u >= nb_txd=%u - ignored\n",
-  nb_segs, (unsigned int) nb_txd);
+   if (nb_segs_is_invalid(nb_segs))
return;
-   }
 
/*
 * Check that each segment length is greater or equal than
-- 
2.9.5



[dpdk-dev] [PATCH v4 5/6] app/testpmd: fix valid desc id check

2020-09-25 Thread Wei Hu (Xavier)
From: Chengchang Tang 

The number of desc is a per queue configuration. But in the check function,
nb_txd & nb_rxd are used to check whether the desc_id is valid. nb_txd &
nb_rxd are the global configuration of number of desc. If the queue
configuration is changed by cmdline liks: "port config xx txq xx ring_size
xxx", the real value will be changed.

This patch use the real value to check whether the desc_id is valid. And if
these are not configured by user. It will use the default value to check
it, since the rte_eth_rx_queue_setup & rte_eth_tx_queue_setup will use a
default value to confiure the queue if nb_rx_desc or nb_tx_desc is zero.

Fixes: af75078fece3 ("first public release")
Cc: sta...@dpdk.org

Signed-off-by: Chengchang Tang 
Signed-off-by: Wei Hu (Xavier) 
---
v3 -> v4:
add check 'rte_eth_rx_queue_info_get()' return value and
 if it is '-ENOSTUP' calculate the 'ring_size'.
v3: initial version.
---
 app/test-pmd/config.c | 76 +++
 1 file changed, 58 insertions(+), 18 deletions(-)

diff --git a/app/test-pmd/config.c b/app/test-pmd/config.c
index 8ebb927..791f8d8 100644
--- a/app/test-pmd/config.c
+++ b/app/test-pmd/config.c
@@ -1893,6 +1893,38 @@ tx_queue_id_is_invalid(queueid_t txq_id)
 }
 
 static int
+get_rx_ring_size(portid_t port_id, queueid_t rxq_id, uint16_t *ring_size)
+{
+   struct rte_port *port = &ports[port_id];
+   struct rte_eth_rxq_info rx_qinfo;
+   int ret;
+
+   ret = rte_eth_rx_queue_info_get(port_id, rxq_id, &rx_qinfo);
+   if (ret == 0) {
+   *ring_size = rx_qinfo.nb_desc;
+   return ret;
+   }
+
+   if (ret != -ENOTSUP)
+   return ret;
+   /*
+* If the rte_eth_rx_queue_info_get is not support for this PMD,
+* ring_size stored in testpmd will be used for validity verification.
+* When configure the rxq by rte_eth_rx_queue_setup with nb_rx_desc
+* being 0, it will use a default value provided by PMDs to setup this
+* rxq. If the default value is 0, it will use the
+* RTE_ETH_DEV_FALLBACK_RX_RINGSIZE to setup this rxq.
+*/
+   if (port->nb_rx_desc[rxq_id])
+   *ring_size = port->nb_rx_desc[rxq_id];
+   else if (port->dev_info.default_rxportconf.ring_size)
+   *ring_size = port->dev_info.default_rxportconf.ring_size;
+   else
+   *ring_size = RTE_ETH_DEV_FALLBACK_RX_RINGSIZE;
+   return 0;
+}
+
+static int
 get_tx_ring_size(portid_t port_id, queueid_t txq_id, uint16_t *ring_size)
 {
struct rte_port *port = &ports[port_id];
@@ -1925,22 +1957,38 @@ get_tx_ring_size(portid_t port_id, queueid_t txq_id, 
uint16_t *ring_size)
 }
 
 static int
-rx_desc_id_is_invalid(uint16_t rxdesc_id)
+rx_desc_id_is_invalid(portid_t port_id, queueid_t rxq_id, uint16_t rxdesc_id)
 {
-   if (rxdesc_id < nb_rxd)
+   uint16_t ring_size;
+   int ret;
+
+   ret = get_rx_ring_size(port_id, rxq_id, &ring_size);
+   if (ret)
+   return 1;
+
+   if (rxdesc_id < ring_size)
return 0;
-   printf("Invalid RX descriptor %d (must be < nb_rxd=%d)\n",
-  rxdesc_id, nb_rxd);
+
+   printf("Invalid RX descriptor %u (must be < ring_size=%u)\n",
+  rxdesc_id, ring_size);
return 1;
 }
 
 static int
-tx_desc_id_is_invalid(uint16_t txdesc_id)
+tx_desc_id_is_invalid(portid_t port_id, queueid_t txq_id, uint16_t txdesc_id)
 {
-   if (txdesc_id < nb_txd)
+   uint16_t ring_size;
+   int ret;
+
+   ret = get_tx_ring_size(port_id, txq_id, &ring_size);
+   if (ret)
+   return 1;
+
+   if (txdesc_id < ring_size)
return 0;
-   printf("Invalid TX descriptor %d (must be < nb_txd=%d)\n",
-  txdesc_id, nb_txd);
+
+   printf("Invalid TX descriptor %u (must be < ring_size=%u)\n",
+  txdesc_id, ring_size);
return 1;
 }
 
@@ -2061,11 +2109,7 @@ rx_ring_desc_display(portid_t port_id, queueid_t rxq_id, 
uint16_t rxd_id)
 {
const struct rte_memzone *rx_mz;
 
-   if (port_id_is_invalid(port_id, ENABLED_WARN))
-   return;
-   if (rx_queue_id_is_invalid(rxq_id))
-   return;
-   if (rx_desc_id_is_invalid(rxd_id))
+   if (rx_desc_id_is_invalid(port_id, rxq_id, rxd_id))
return;
rx_mz = ring_dma_zone_lookup("rx_ring", port_id, rxq_id);
if (rx_mz == NULL)
@@ -2078,11 +2122,7 @@ tx_ring_desc_display(portid_t port_id, queueid_t txq_id, 
uint16_t txd_id)
 {
const struct rte_memzone *tx_mz;
 
-   if (port_id_is_invalid(port_id, ENABLED_WARN))
-   return;
-   if (tx_queue_id_is_invalid(txq_id))
-   return;
-   if (tx_desc_id_is_invalid(txd_id))
+   if (tx_desc_id_is_invalid(port_id, txq_id, txd_id))
return;
tx_mz = ring_dma_zone_lookup("tx_ring", port_id, txq_id);
if (tx_mz == 

[dpdk-dev] [PATCH v4 6/6] app/testpmd: fix displaying Rx Tx queues information

2020-09-25 Thread Wei Hu (Xavier)
From: Huisong Li 

Currently, the information of Rx/Tx queues from PMD driver is not displayed
exactly in the rxtx_config_display function. Because "ports[pid].rx_conf"
and "ports[pid].tx_conf" maintained in testpmd application may be not the
value actually used by PMD driver. For instance, user does not set a field,
but PMD driver has to use the default value.

This patch fixes rxtx_config_display so that the information of Rx/Tx
queues can be really displayed for the PMD driver that implement
.rxq_info_get and .txq_info_get ops callback function.

Fixes: 75c530c1bd5351 ("app/testpmd: fix port configuration print")
Fixes: d44f8a485f5d1f ("app/testpmd: enable per queue configure")
Cc: sta...@dpdk.org

Signed-off-by: Huisong Li 
Signed-off-by: Wei Hu (Xavier) 
Reviewed-by: Ferruh Yigit 
---
 app/test-pmd/config.c | 64 +--
 1 file changed, 47 insertions(+), 17 deletions(-)

diff --git a/app/test-pmd/config.c b/app/test-pmd/config.c
index 791f8d8..6a0058c 100644
--- a/app/test-pmd/config.c
+++ b/app/test-pmd/config.c
@@ -2163,10 +2163,17 @@ rxtx_config_display(void)
struct rte_eth_txconf *tx_conf = &ports[pid].tx_conf[0];
uint16_t *nb_rx_desc = &ports[pid].nb_rx_desc[0];
uint16_t *nb_tx_desc = &ports[pid].nb_tx_desc[0];
-   uint16_t nb_rx_desc_tmp;
-   uint16_t nb_tx_desc_tmp;
struct rte_eth_rxq_info rx_qinfo;
struct rte_eth_txq_info tx_qinfo;
+   uint16_t rx_free_thresh_tmp;
+   uint16_t tx_free_thresh_tmp;
+   uint16_t tx_rs_thresh_tmp;
+   uint16_t nb_rx_desc_tmp;
+   uint16_t nb_tx_desc_tmp;
+   uint64_t offloads_tmp;
+   uint8_t pthresh_tmp;
+   uint8_t hthresh_tmp;
+   uint8_t wthresh_tmp;
int32_t rc;
 
/* per port config */
@@ -2180,41 +2187,64 @@ rxtx_config_display(void)
/* per rx queue config only for first queue to be less verbose 
*/
for (qid = 0; qid < 1; qid++) {
rc = rte_eth_rx_queue_info_get(pid, qid, &rx_qinfo);
-   if (rc)
+   if (rc) {
nb_rx_desc_tmp = nb_rx_desc[qid];
-   else
+   rx_free_thresh_tmp =
+   rx_conf[qid].rx_free_thresh;
+   pthresh_tmp = rx_conf[qid].rx_thresh.pthresh;
+   hthresh_tmp = rx_conf[qid].rx_thresh.hthresh;
+   wthresh_tmp = rx_conf[qid].rx_thresh.wthresh;
+   offloads_tmp = rx_conf[qid].offloads;
+   } else {
nb_rx_desc_tmp = rx_qinfo.nb_desc;
+   rx_free_thresh_tmp =
+   rx_qinfo.conf.rx_free_thresh;
+   pthresh_tmp = rx_qinfo.conf.rx_thresh.pthresh;
+   hthresh_tmp = rx_qinfo.conf.rx_thresh.hthresh;
+   wthresh_tmp = rx_qinfo.conf.rx_thresh.wthresh;
+   offloads_tmp = rx_qinfo.conf.offloads;
+   }
 
printf("RX queue: %d\n", qid);
printf("  RX desc=%d - RX free threshold=%d\n",
-   nb_rx_desc_tmp, rx_conf[qid].rx_free_thresh);
+   nb_rx_desc_tmp, rx_free_thresh_tmp);
printf("  RX threshold registers: pthresh=%d 
hthresh=%d "
" wthresh=%d\n",
-   rx_conf[qid].rx_thresh.pthresh,
-   rx_conf[qid].rx_thresh.hthresh,
-   rx_conf[qid].rx_thresh.wthresh);
-   printf("  RX Offloads=0x%"PRIx64"\n",
-   rx_conf[qid].offloads);
+   pthresh_tmp, hthresh_tmp, wthresh_tmp);
+   printf("  RX Offloads=0x%"PRIx64"\n", offloads_tmp);
}
 
/* per tx queue config only for first queue to be less verbose 
*/
for (qid = 0; qid < 1; qid++) {
rc = rte_eth_tx_queue_info_get(pid, qid, &tx_qinfo);
-   if (rc)
+   if (rc) {
nb_tx_desc_tmp = nb_tx_desc[qid];
-   else
+   tx_free_thresh_tmp =
+   tx_conf[qid].tx_free_thresh;
+   pthresh_tmp = tx_conf[qid].tx_thresh.pthresh;
+   hthresh_tmp = tx_conf[qid].tx_thresh.hthresh;
+   wthresh_tmp = tx_conf[qid].tx_thresh.wthresh;
+   

[dpdk-dev] [PATCH V4 1/2] dpdk: resolve compiling errors for per-queue stats

2020-09-25 Thread Min Hu (Connor)
From: Huisong Li 

Currently, only statistics of rx/tx queues with queue_id less than
RTE_ETHDEV_QUEUE_STAT_CNTRS can be displayed. If there is a certain
application scenario that it needs to use 256 or more than 256 queues
and display all statistics of rx/tx queue. At this moment, we have to
change the macro to be equaled to the queue number.

However, modifying the macro to be greater than 256 will trigger
many errors and warnings from test-pmd, PMD drivers and librte_ethdev
during compiling dpdk project. But it is possible and permitted that
rx/tx queue number is greater than 256 and all statistics of rx/tx
queue need to be displayed. In addition, the data type of rx/tx queue
number in rte_eth_dev_configure API is 'uint16_t'. So It is unreasonable
to use the 'uint8_t' type for variables that control which per-queue
statistics can be displayed.

Fixes: ed30d9b691b2 ("app/testpmd: add stats per queue")
Fixes: 09c7e63a71f9 ("net/memif: introduce memory interface PMD")
Fixes: abf7275bbaa2 ("ixgbe: move to drivers/net/")
Fixes: e6defdfddc3b ("net/igc: enable statistics")
Fixes: 2265e4b4e84b ("net/octeontx2: add basic stats operation")
Fixes: 6c3169a3dc04 ("virtio: move to drivers/net/")

Signed-off-by: Huisong Li 
Signed-off-by: Min Hu (Connor) 
Reviewed-by: Wei Hu (Xavier) 
Reviewed-by: Dongdong Liu 
---
v3->v4:
add a change in cmd_setqmap_mapvalue.

---
v2->v3:
change 'uint8_t i' to 'uint16_t i' in nic_stats_display function

---
 app/proc-info/main.c  | 2 +-
 app/test-pmd/cmdline.c| 4 ++--
 app/test-pmd/config.c | 4 ++--
 app/test-pmd/testpmd.c| 2 +-
 app/test-pmd/testpmd.h| 5 +++--
 drivers/net/igc/igc_ethdev.c  | 4 ++--
 drivers/net/ixgbe/ixgbe_ethdev.c  | 4 ++--
 drivers/net/memif/rte_eth_memif.c | 2 +-
 drivers/net/octeontx2/otx2_ethdev.h   | 2 +-
 drivers/net/octeontx2/otx2_stats.c| 2 +-
 drivers/net/virtio/virtio_ethdev.c| 4 ++--
 lib/librte_ethdev/rte_ethdev.c| 6 +++---
 lib/librte_ethdev/rte_ethdev.h| 4 ++--
 lib/librte_ethdev/rte_ethdev_driver.h | 2 +-
 14 files changed, 24 insertions(+), 23 deletions(-)

diff --git a/app/proc-info/main.c b/app/proc-info/main.c
index 64fb83b..26d9355 100644
--- a/app/proc-info/main.c
+++ b/app/proc-info/main.c
@@ -348,7 +348,7 @@ static void
 nic_stats_display(uint16_t port_id)
 {
struct rte_eth_stats stats;
-   uint8_t i;
+   uint16_t i;
 
static const char *nic_stats_border = "";
 
diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c
index 08e123f..23e624f 100644
--- a/app/test-pmd/cmdline.c
+++ b/app/test-pmd/cmdline.c
@@ -8315,7 +8315,7 @@ struct cmd_set_qmap_result {
cmdline_fixed_string_t what;
portid_t port_id;
uint16_t queue_id;
-   uint8_t map_value;
+   uint16_t map_value;
 };
 
 static void
@@ -8346,7 +8346,7 @@ cmdline_parse_token_num_t cmd_setqmap_queueid =
  queue_id, UINT16);
 cmdline_parse_token_num_t cmd_setqmap_mapvalue =
TOKEN_NUM_INITIALIZER(struct cmd_set_qmap_result,
- map_value, UINT8);
+ map_value, UINT16);
 
 cmdline_parse_inst_t cmd_set_qmap = {
.f = cmd_set_qmap_parsed,
diff --git a/app/test-pmd/config.c b/app/test-pmd/config.c
index 17a6efe..dfe5627 100644
--- a/app/test-pmd/config.c
+++ b/app/test-pmd/config.c
@@ -161,7 +161,7 @@ nic_stats_display(portid_t port_id)
uint64_t mpps_rx, mpps_tx, mbps_rx, mbps_tx;
struct rte_eth_stats stats;
struct rte_port *port = &ports[port_id];
-   uint8_t i;
+   uint16_t i;
 
static const char *nic_stats_border = "";
 
@@ -3742,7 +3742,7 @@ tx_vlan_pvid_set(portid_t port_id, uint16_t vlan_id, int 
on)
 }
 
 void
-set_qmap(portid_t port_id, uint8_t is_rx, uint16_t queue_id, uint8_t map_value)
+set_qmap(portid_t port_id, uint8_t is_rx, uint16_t queue_id, uint16_t 
map_value)
 {
uint16_t i;
uint8_t existing_mapping_found = 0;
diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c
index fe6450c..4b26c5c 100644
--- a/app/test-pmd/testpmd.c
+++ b/app/test-pmd/testpmd.c
@@ -1840,7 +1840,7 @@ fwd_stats_display(void)
fwd_cycles += fs->core_cycles;
}
for (i = 0; i < cur_fwd_config.nb_fwd_ports; i++) {
-   uint8_t j;
+   uint16_t j;
 
pt_id = fwd_ports_ids[i];
port = &ports[pt_id];
diff --git a/app/test-pmd/testpmd.h b/app/test-pmd/testpmd.h
index c7e7e41..9ad002c 100644
--- a/app/test-pmd/testpmd.h
+++ b/app/test-pmd/testpmd.h
@@ -279,7 +279,7 @@ enum dcb_mode_enable
 struct queue_stats_mappings {
portid_t port_id;
uint16_t queue_id;
-   uint8_t stats_counter_id;
+   uint16_t stats_counter_id;
 } __rte_cache_aligned;
 
 extern struct queue_stats_mappings tx_queue_stats_mappings_array[];
@@ -794,7 +794,8 @@ 

[dpdk-dev] [PATCH V4 2/2] ethdev: change data type in TC rxq and TC txq

2020-09-25 Thread Min Hu (Connor)
From: Huisong Li 

Currently, base and nb_queue in the tc_rxq and tc_txq information
of queue and TC mapping on both TX and RX paths are uint8_t.
However, these data will be truncated when queue number under a TC
is greater than 256. So it is necessary for base and nb_queue to
change from uint8_t to uint16_t.

Fixes: 01eb53eefeb40e8 ("ethdev: rename folder to library name")

Signed-off-by: Huisong Li 
Signed-off-by: Min Hu (Connor) 
Reviewed-by: Wei Hu (Xavier) 
Reviewed-by: Dongdong Liu 
---
 lib/librte_ethdev/rte_ethdev.h | 8 
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git a/lib/librte_ethdev/rte_ethdev.h b/lib/librte_ethdev/rte_ethdev.h
index ff3a616..2270460 100644
--- a/lib/librte_ethdev/rte_ethdev.h
+++ b/lib/librte_ethdev/rte_ethdev.h
@@ -1522,13 +1522,13 @@ struct rte_eth_xstat_name {
 struct rte_eth_dcb_tc_queue_mapping {
/** rx queues assigned to tc per Pool */
struct {
-   uint8_t base;
-   uint8_t nb_queue;
+   uint16_t base;
+   uint16_t nb_queue;
} tc_rxq[ETH_MAX_VMDQ_POOL][ETH_DCB_NUM_TCS];
/** rx queues assigned to tc per Pool */
struct {
-   uint8_t base;
-   uint8_t nb_queue;
+   uint16_t base;
+   uint16_t nb_queue;
} tc_txq[ETH_MAX_VMDQ_POOL][ETH_DCB_NUM_TCS];
 };
 
-- 
2.7.4



[dpdk-dev] [PATCH V4 0/2] change data type in TC queue

2020-09-25 Thread Min Hu (Connor)
From: Huisong Li 

This series change data type in TC rxq and TC txq,
fix compiling errors for per-queue statistics.

Huisong Li (2):
  dpdk: resolve compiling errors for per-queue stats
  ethdev: change data type in TC rxq and TC txq

 app/proc-info/main.c  |  2 +-
 app/test-pmd/cmdline.c|  4 ++--
 app/test-pmd/config.c |  4 ++--
 app/test-pmd/testpmd.c|  2 +-
 app/test-pmd/testpmd.h|  5 +++--
 drivers/net/igc/igc_ethdev.c  |  4 ++--
 drivers/net/ixgbe/ixgbe_ethdev.c  |  4 ++--
 drivers/net/memif/rte_eth_memif.c |  2 +-
 drivers/net/octeontx2/otx2_ethdev.h   |  2 +-
 drivers/net/octeontx2/otx2_stats.c|  2 +-
 drivers/net/virtio/virtio_ethdev.c|  4 ++--
 lib/librte_ethdev/rte_ethdev.c|  6 +++---
 lib/librte_ethdev/rte_ethdev.h| 12 ++--
 lib/librte_ethdev/rte_ethdev_driver.h |  2 +-
 14 files changed, 28 insertions(+), 27 deletions(-)

-- 
2.7.4



Re: [dpdk-dev] [PATCH 10/20] net/null: release port upon close

2020-09-25 Thread Thomas Monjalon
25/09/2020 10:52, Ferruh Yigit:
> On 9/24/2020 10:58 PM, Thomas Monjalon wrote:
> > 23/09/2020 22:47, Thomas Monjalon:
> >> 23/09/2020 18:44, Ferruh Yigit:
> >>> On 9/13/2020 11:07 PM, Thomas Monjalon wrote:
>  The flag RTE_ETH_DEV_CLOSE_REMOVE is set so all port resources
>  can be freed by rte_eth_dev_close().
> 
>  Signed-off-by: Thomas Monjalon 
>  ---
>  --- a/drivers/net/null/rte_eth_null.c
>  +++ b/drivers/net/null/rte_eth_null.c
>  +static int
>  +eth_dev_close(struct rte_eth_dev *dev)
>  +{
>  +PMD_LOG(INFO, "Closing null ethdev on NUMA socket %u",
>  +rte_socket_id());
>  +
>  +/* mac_addrs must not be freed alone because part of 
>  dev_private */
>  +dev->data->mac_addrs = NULL;
>  +
>  +return 0;
>  +}
> >>>
> >>> should check 'RTE_PROC_PRIMARY' in 'eth_dev_close()'?
> >>
> >> Yes, looks to be a miss for this new function.
> > 
> > Sorry no, this function is not freeing any shared data,
> > so no restriction on secondary process.
> 
> "dev->data->mac_addrs = NULL", won't this change the shared data?

Absolutely, you're right.
Will fix in v2.




Re: [dpdk-dev] [PATCH V4 0/2] change data type in TC queue

2020-09-25 Thread Ferruh Yigit

On 9/25/2020 1:51 PM, Min Hu (Connor) wrote:

From: Huisong Li 

This series change data type in TC rxq and TC txq,
fix compiling errors for per-queue statistics.

Huisong Li (2):
   dpdk: resolve compiling errors for per-queue stats
   ethdev: change data type in TC rxq and TC txq

  app/proc-info/main.c  |  2 +-
  app/test-pmd/cmdline.c|  4 ++--
  app/test-pmd/config.c |  4 ++--
  app/test-pmd/testpmd.c|  2 +-
  app/test-pmd/testpmd.h|  5 +++--
  drivers/net/igc/igc_ethdev.c  |  4 ++--
  drivers/net/ixgbe/ixgbe_ethdev.c  |  4 ++--
  drivers/net/memif/rte_eth_memif.c |  2 +-
  drivers/net/octeontx2/otx2_ethdev.h   |  2 +-
  drivers/net/octeontx2/otx2_stats.c|  2 +-
  drivers/net/virtio/virtio_ethdev.c|  4 ++--
  lib/librte_ethdev/rte_ethdev.c|  6 +++---
  lib/librte_ethdev/rte_ethdev.h| 12 ++--
  lib/librte_ethdev/rte_ethdev_driver.h |  2 +-
  14 files changed, 28 insertions(+), 27 deletions(-)



While you are dropping the deprecation notice, it seems dropped the 
release notes updates too, they were good to have.


If you don't mind can you add the release notes updated back, and 
instead of making them separate patch they can be be squashed into 
existing patches.


Thanks again.


Re: [dpdk-dev] [PATCH] build: support i686 target on x86 64-bit hosts

2020-09-25 Thread Bruce Richardson
On Fri, Sep 25, 2020 at 09:27:25AM -0400, Lance Richardson wrote:
> On Fri, Sep 25, 2020 at 5:13 AM Bruce Richardson
>  wrote:
> >
> > On Thu, Sep 24, 2020 at 12:37:42PM -0400, Lance Richardson wrote:
> > > Add meson cross files for building i686 targets using gcc on x86_64
> > > linux hosts.
> > >
> > > Uusage example:
> > >
> > > meson --cross-file config/x86/cross-i686-linux-gcc build-i686
> > > ninja -C build-i686
> > >
> > > Signed-off-by: Lance Richardson 
> > > ---
> > > Red Hat distros use a different name for the 32-bit pkg-config
> > > command from most other distros, maybe there is a better way
> > > to handle this than using separate files.
> > >
> > > Others will probably have better naming suggestions for the files.
> > >
> > Just to note that rather than using cross-files, building i686 on x86_64
> > should be possible by just setting environment variables.
> >
> > For example, on Ubuntu 20.04, this works for me:
> >
> > PKG_CONFIG_LIBDIR=/usr/lib/i386-linux-gnu/pkgconfig CFLAGS='-m32' 
> > LDFLAGS='-m32' meson --werror build-32bit
> > ninja -C build-32bit/
> >
> > For Fedora or Redhat system the PKG_CONFIG_LIBDIR will be different (I
> > think just /usr/lib/pkgconfig), but the rest should work identically.
> >
> > /Bruce
> 
> Thanks, that does work, although I'll have some trouble remembering it
> for future
> use (unlike using cross-files). Googling for "PKG_CONFIG_LIBDIR meson dpdk"
> did lead me to this discussion, which makes it sound like cross-files
> might still
> worthy of consideration:
> 
>https://patches.dpdk.org/patch/69067/
> 

Yes, absolutely the cross-file is worthy of consideration. The most awkward
part in both cases is getting the 32-bit pkg-config right.


Re: [dpdk-dev] [PATCH v3 0/4] use C11 atomic builtins for libs

2020-09-25 Thread David Marchand
On Thu, Sep 24, 2020 at 7:40 AM Phil Yang  wrote:
>
> Since rte_atomicXX APIs are not allowed to be used[1][2], use C11 atomic
> builtins instead in eal, bbdev, power, and ethdev libs.
>
> [1] 
> http://code.dpdk.org/dpdk/latest/source/doc/guides/rel_notes/deprecation.rst#L87
> [2] http://code.dpdk.org/dpdk/latest/source/devtools/checkpatches.sh#L80
>
> v3:
> remove unnecessary rte_atomic.h headers. (David)
>
> v2:
> Fix Clang int-conversion warning.
>
> v1:
> Initial version.
>
> Phil Yang (4):
>   eal: use C11 atomic builtins for already initialized check
>   bbdev: use C11 atomic builtins for device processing counter
>   power: use C11 atomic builtins for power in use state update
>   ethdev: use C11 atomic builtins for link status update
>
>  lib/librte_bbdev/rte_bbdev.c|  5 ++--
>  lib/librte_bbdev/rte_bbdev.h|  4 +--
>  lib/librte_eal/freebsd/eal.c| 19 +++---
>  lib/librte_eal/linux/eal.c  | 21 +++
>  lib/librte_ethdev/rte_ethdev_driver.h   | 19 --
>  lib/librte_power/power_acpi_cpufreq.c   | 45 
> +
>  lib/librte_power/power_pstate_cpufreq.c | 45 
> +
>  7 files changed, 100 insertions(+), 58 deletions(-)
>

Series applied, thanks Phil.


-- 
David Marchand



  1   2   >