Re: [PATCH v3 1/1] app/testpmd: control passing Rx metadata to PMD

2022-10-27 Thread Thomas Monjalon
17/10/2022 10:32, Andrew Rybchenko:
> On 10/6/22 21:35, Hanumanth Pothula wrote:
> > Presently, Rx metadata is sent to PMD by default, leading
> > to a performance drop as processing for the same in rx path
> > takes extra cycles.
> > 
> > Hence, introducing command line argument, 'nic-to-pmd-rx-metadata'
> > to control passing rx metadata to PMD. By default it’s disabled.
> > 
> > Signed-off-by: Hanumanth Pothula 
> 
> Acked-by: Andrew Rybchenko 
> 
> Applied to dpdk-next-net/main, thanks.

I'm not sure this patch is really acceptable.
It is disabling Rx metadata by default just for benchmarking reason
because your driver is doing some processing even if metadata is not required.

From a user perspective, if a command requesting metadata is entered,
it won't work until we enable this new option on startup.
It looks terrible.

Please tell me I misunderstood something.




[PATCH v14 00/18] add support for idpf PMD in DPDK

2022-10-27 Thread Junfeng Guo
This patchset introduced the idpf (Infrastructure Data Path Function)
PMD in DPDK for Intel® IPU E2000 (Device ID: 0x1452).
The Intel® IPU E2000 targets to deliver high performance under real
workloads with security and isolation.
Please refer to
https://www.intel.com/content/www/us/en/products/network-io/infrastructure-processing-units/asic/e2000-asic.html
for more information.

Linux upstream is still ongoing, previous work refers to
https://patchwork.ozlabs.org/project/intel-wired-lan/patch/20220128001009.721392-20-alan.br...@intel.com/.

v2-v4:
fixed some coding style issues and did some refactors.

v5:
fixed typo.

v6-v9:
fixed build errors and coding style issues.

v11:
 - move shared code to common/idpf/base
 - Create one vport if there's no vport devargs
 - Refactor if conditions according to coding style
 - Refactor virtual channel return values
 - Refine dev_stop function
 - Refine RSS lut/key
 - Fix build error

v12:
 - Refine dev_configure
 - Fix coding style accoding to the comments
 - Re-order patch
 - Romove dev_supported_ptypes_get

v13:
 - refine dev_start/stop and queue_start/stop
 - fix timestamp offload

v14:
 - fix wrong position for rte_validate_tx_offload

Junfeng Guo (18):
  common/idpf: introduce common library
  net/idpf: add support for device initialization
  net/idpf: add Tx queue setup
  net/idpf: add Rx queue setup
  net/idpf: add support for device start and stop
  net/idpf: add support for queue start
  net/idpf: add support for queue stop
  net/idpf: add queue release
  net/idpf: add support for MTU configuration
  net/idpf: add support for basic Rx datapath
  net/idpf: add support for basic Tx datapath
  net/idpf: support parsing packet type
  net/idpf: add support for write back based on ITR expire
  net/idpf: add support for RSS
  net/idpf: add support for Rx offloading
  net/idpf: add support for Tx offloading
  net/idpf: add AVX512 data path for single queue model
  net/idpf: add support for timestamp offload

 MAINTAINERS   |9 +
 doc/guides/nics/features/idpf.ini |   18 +
 doc/guides/nics/idpf.rst  |   85 +
 doc/guides/nics/index.rst |1 +
 doc/guides/rel_notes/release_22_11.rst|6 +
 drivers/common/idpf/base/idpf_alloc.h |   22 +
 drivers/common/idpf/base/idpf_common.c|  364 +++
 drivers/common/idpf/base/idpf_controlq.c  |  691 
 drivers/common/idpf/base/idpf_controlq.h  |  224 ++
 drivers/common/idpf/base/idpf_controlq_api.h  |  234 ++
 .../common/idpf/base/idpf_controlq_setup.c|  179 +
 drivers/common/idpf/base/idpf_devids.h|   18 +
 drivers/common/idpf/base/idpf_lan_pf_regs.h   |  134 +
 drivers/common/idpf/base/idpf_lan_txrx.h  |  428 +++
 drivers/common/idpf/base/idpf_lan_vf_regs.h   |  114 +
 drivers/common/idpf/base/idpf_osdep.h |  364 +++
 drivers/common/idpf/base/idpf_prototype.h |   45 +
 drivers/common/idpf/base/idpf_type.h  |  106 +
 drivers/common/idpf/base/meson.build  |   14 +
 drivers/common/idpf/base/siov_regs.h  |   47 +
 drivers/common/idpf/base/virtchnl.h   | 2866 +
 drivers/common/idpf/base/virtchnl2.h  | 1462 +
 drivers/common/idpf/base/virtchnl2_lan_desc.h |  606 
 .../common/idpf/base/virtchnl_inline_ipsec.h  |  567 
 drivers/common/idpf/meson.build   |4 +
 drivers/common/idpf/version.map   |   12 +
 drivers/common/meson.build|1 +
 drivers/net/idpf/idpf_ethdev.c| 1280 
 drivers/net/idpf/idpf_ethdev.h|  252 ++
 drivers/net/idpf/idpf_logs.h  |   56 +
 drivers/net/idpf/idpf_rxtx.c  | 2308 +
 drivers/net/idpf/idpf_rxtx.h  |  291 ++
 drivers/net/idpf/idpf_rxtx_vec_avx512.c   |  871 +
 drivers/net/idpf/idpf_rxtx_vec_common.h   |  100 +
 drivers/net/idpf/idpf_vchnl.c | 1430 
 drivers/net/idpf/meson.build  |   44 +
 drivers/net/idpf/version.map  |3 +
 drivers/net/meson.build   |1 +
 38 files changed, 15257 insertions(+)
 create mode 100644 doc/guides/nics/features/idpf.ini
 create mode 100644 doc/guides/nics/idpf.rst
 create mode 100644 drivers/common/idpf/base/idpf_alloc.h
 create mode 100644 drivers/common/idpf/base/idpf_common.c
 create mode 100644 drivers/common/idpf/base/idpf_controlq.c
 create mode 100644 drivers/common/idpf/base/idpf_controlq.h
 create mode 100644 drivers/common/idpf/base/idpf_controlq_api.h
 create mode 100644 drivers/common/idpf/base/idpf_controlq_setup.c
 create mode 100644 drivers/common/idpf/base/idpf_devids.h
 create mode 100644 drivers/common/idpf/base/idpf_lan_pf_regs.h
 create mode 100644 drivers/common/idpf/base/idpf_lan_txrx.h
 create mode 100644 drivers/common/idpf/base/idpf_lan_vf_regs.h
 create mode 100644 drivers/common/idpf/base/idpf_osdep.h

[PATCH v14 02/18] net/idpf: add support for device initialization

2022-10-27 Thread Junfeng Guo
Support device init and add the following dev ops:
 - dev_configure
 - dev_close
 - dev_infos_get

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Xiao Wang 
Signed-off-by: Wenjun Wu 
Signed-off-by: Junfeng Guo 
---
 MAINTAINERS|   9 +
 doc/guides/nics/features/idpf.ini  |   9 +
 doc/guides/nics/idpf.rst   |  66 ++
 doc/guides/nics/index.rst  |   1 +
 doc/guides/rel_notes/release_22_11.rst |   6 +
 drivers/net/idpf/idpf_ethdev.c | 886 +
 drivers/net/idpf/idpf_ethdev.h | 189 ++
 drivers/net/idpf/idpf_logs.h   |  56 ++
 drivers/net/idpf/idpf_vchnl.c  | 468 +
 drivers/net/idpf/meson.build   |  15 +
 drivers/net/idpf/version.map   |   3 +
 drivers/net/meson.build|   1 +
 12 files changed, 1709 insertions(+)
 create mode 100644 doc/guides/nics/features/idpf.ini
 create mode 100644 doc/guides/nics/idpf.rst
 create mode 100644 drivers/net/idpf/idpf_ethdev.c
 create mode 100644 drivers/net/idpf/idpf_ethdev.h
 create mode 100644 drivers/net/idpf/idpf_logs.h
 create mode 100644 drivers/net/idpf/idpf_vchnl.c
 create mode 100644 drivers/net/idpf/meson.build
 create mode 100644 drivers/net/idpf/version.map

diff --git a/MAINTAINERS b/MAINTAINERS
index 92b381bc30..34f8b9cc61 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -764,6 +764,15 @@ F: drivers/net/ice/
 F: doc/guides/nics/ice.rst
 F: doc/guides/nics/features/ice.ini
 
+Intel idpf
+M: Jingjing Wu 
+M: Beilei Xing 
+T: git://dpdk.org/next/dpdk-next-net-intel
+F: drivers/net/idpf/
+F: drivers/common/idpf/
+F: doc/guides/nics/idpf.rst
+F: doc/guides/nics/features/idpf.ini
+
 Intel igc
 M: Junfeng Guo 
 M: Simei Su 
diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
new file mode 100644
index 00..46aab2eb61
--- /dev/null
+++ b/doc/guides/nics/features/idpf.ini
@@ -0,0 +1,9 @@
+;
+; Supported features of the 'idpf' network poll mode driver.
+;
+; Refer to default.ini for the full list of available PMD features.
+;
+[Features]
+Linux= Y
+x86-32   = Y
+x86-64   = Y
diff --git a/doc/guides/nics/idpf.rst b/doc/guides/nics/idpf.rst
new file mode 100644
index 00..c1001d5d0c
--- /dev/null
+++ b/doc/guides/nics/idpf.rst
@@ -0,0 +1,66 @@
+..  SPDX-License-Identifier: BSD-3-Clause
+Copyright(c) 2022 Intel Corporation.
+
+IDPF Poll Mode Driver
+==
+
+The [*EXPERIMENTAL*] idpf PMD (**librte_net_idpf**) provides poll mode driver 
support for
+Intel® Infrastructure Processing Unit (Intel® IPU) E2000.
+
+
+Linux Prerequisites
+---
+
+- Follow the DPDK :ref:`Getting Started Guide for Linux ` to setup 
the basic DPDK environment.
+
+- To get better performance on Intel platforms, please follow the "How to get 
best performance with NICs on Intel platforms"
+  section of the :ref:`Getting Started Guide for Linux `.
+
+
+Pre-Installation Configuration
+--
+
+Runtime Config Options
+~~
+
+- ``vport`` (default ``0``)
+
+  The IDPF PMD supports creation of multiple vports for one PCI device, each 
vport
+  corresponds to a single ethdev. Using the ``devargs`` parameter ``vport`` 
the user
+  can specify the vports with specific ID to be created, for example::
+
+-a ca:00.0,vport=[0,2,3]
+
+  Then idpf PMD will create 3 vports (ethdevs) for device ca:00.0.
+  NOTE: If the parameter is not provided, the vport 0 will be created by 
default.
+
+- ``rx_single`` (default ``0``)
+
+  There're two queue modes supported by Intel® IPU Ethernet ES2000 Series, 
single queue
+  mode and split queue mode for Rx queue. User can choose Rx queue mode by the 
``devargs``
+  parameter ``rx_single``.
+
+-a ca:00.0,rx_single=1
+
+  Then idpf PMD will configure Rx queue with single queue mode. Otherwise, 
split queue
+  mode is chosen by default.
+
+- ``tx_single`` (default ``0``)
+
+  There're two queue modes supported by Intel® IPU Ethernet ES2000 Series, 
single queue
+  mode and split queue mode for Tx queue. User can choose Tx queue mode by the 
``devargs``
+  parameter ``tx_single``.
+
+-a ca:00.0,tx_single=1
+
+  Then idpf PMD will configure Tx queue with single queue mode. Otherwise, 
split queue
+  mode is chosen by default.
+
+
+Driver compilation and testing
+--
+
+Refer to the document :ref:`compiling and testing a PMD for a NIC 
`
+for details.
+
+
diff --git a/doc/guides/nics/index.rst b/doc/guides/nics/index.rst
index 32c7544968..b4dd5ea522 100644
--- a/doc/guides/nics/index.rst
+++ b/doc/guides/nics/index.rst
@@ -33,6 +33,7 @@ Network Interface Controller Drivers
 hns3
 i40e
 ice
+idpf
 igb
 igc
 ionic
diff --git a/doc/guides/rel_notes/release_22_11.rst 
b/doc/guides/rel_notes/release_22_11.rst
index 1c3daf141d..4af790de37 100644
--- a/doc/guides/rel_notes/release_22_11.rst
+++ b

[PATCH v14 03/18] net/idpf: add Tx queue setup

2022-10-27 Thread Junfeng Guo
Add support for tx_queue_setup ops.

In the single queue model, the same descriptor queue is used by SW to
post buffer descriptors to HW and by HW to post completed descriptors
to SW.

In the split queue model, "RX buffer queues" are used to pass
descriptor buffers from SW to HW while Rx queues are used only to
pass the descriptor completions, that is, descriptors that point
to completed buffers, from HW to SW. This is contrary to the single
queue model in which Rx queues are used for both purposes.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  13 ++
 drivers/net/idpf/idpf_rxtx.c   | 364 +
 drivers/net/idpf/idpf_rxtx.h   |  70 +++
 drivers/net/idpf/idpf_vchnl.c  |   2 -
 drivers/net/idpf/meson.build   |   1 +
 5 files changed, 448 insertions(+), 2 deletions(-)
 create mode 100644 drivers/net/idpf/idpf_rxtx.c
 create mode 100644 drivers/net/idpf/idpf_rxtx.h

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 112dccf5c9..11f8b4ba1c 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -11,6 +11,7 @@
 #include 
 
 #include "idpf_ethdev.h"
+#include "idpf_rxtx.h"
 
 #define IDPF_TX_SINGLE_Q   "tx_single"
 #define IDPF_RX_SINGLE_Q   "rx_single"
@@ -37,6 +38,7 @@ static void idpf_adapter_rel(struct idpf_adapter *adapter);
 static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_configure  = idpf_dev_configure,
.dev_close  = idpf_dev_close,
+   .tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
 };
 
@@ -56,6 +58,17 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
 
dev_info->max_mac_addrs = IDPF_NUM_MACADDR_MAX;
 
+   dev_info->default_txconf = (struct rte_eth_txconf) {
+   .tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
+   .tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH,
+   };
+
+   dev_info->tx_desc_lim = (struct rte_eth_desc_lim) {
+   .nb_max = IDPF_MAX_RING_DESC,
+   .nb_min = IDPF_MIN_RING_DESC,
+   .nb_align = IDPF_ALIGN_RING_DESC,
+   };
+
return 0;
 }
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
new file mode 100644
index 00..4afa0a2560
--- /dev/null
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -0,0 +1,364 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2022 Intel Corporation
+ */
+
+#include 
+#include 
+
+#include "idpf_ethdev.h"
+#include "idpf_rxtx.h"
+
+static int
+check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
+   uint16_t tx_free_thresh)
+{
+   /* TX descriptors will have their RS bit set after tx_rs_thresh
+* descriptors have been used. The TX descriptor ring will be cleaned
+* after tx_free_thresh descriptors are used or if the number of
+* descriptors required to transmit a packet is greater than the
+* number of free TX descriptors.
+*
+* The following constraints must be satisfied:
+*  - tx_rs_thresh must be less than the size of the ring minus 2.
+*  - tx_free_thresh must be less than the size of the ring minus 3.
+*  - tx_rs_thresh must be less than or equal to tx_free_thresh.
+*  - tx_rs_thresh must be a divisor of the ring size.
+*
+* One descriptor in the TX ring is used as a sentinel to avoid a H/W
+* race condition, hence the maximum threshold constraints. When set
+* to zero use default values.
+*/
+   if (tx_rs_thresh >= (nb_desc - 2)) {
+   PMD_INIT_LOG(ERR, "tx_rs_thresh (%u) must be less than the "
+"number of TX descriptors (%u) minus 2",
+tx_rs_thresh, nb_desc);
+   return -EINVAL;
+   }
+   if (tx_free_thresh >= (nb_desc - 3)) {
+   PMD_INIT_LOG(ERR, "tx_free_thresh (%u) must be less than the "
+"number of TX descriptors (%u) minus 3.",
+tx_free_thresh, nb_desc);
+   return -EINVAL;
+   }
+   if (tx_rs_thresh > tx_free_thresh) {
+   PMD_INIT_LOG(ERR, "tx_rs_thresh (%u) must be less than or "
+"equal to tx_free_thresh (%u).",
+tx_rs_thresh, tx_free_thresh);
+   return -EINVAL;
+   }
+   if ((nb_desc % tx_rs_thresh) != 0) {
+   PMD_INIT_LOG(ERR, "tx_rs_thresh (%u) must be a divisor of the "
+"number of TX descriptors (%u).",
+tx_rs_thresh, nb_desc);
+   return -EINVAL;
+   }
+
+   return 0;
+}
+
+static void
+reset_split_tx_descq(struct idpf_tx_queue *txq)
+{
+   struct idpf_tx_entry *txe;
+   uint32_t i, size;
+

[PATCH v14 04/18] net/idpf: add Rx queue setup

2022-10-27 Thread Junfeng Guo
Add support for rx_queue_setup ops.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  11 +
 drivers/net/idpf/idpf_rxtx.c   | 400 +
 drivers/net/idpf/idpf_rxtx.h   |  46 
 3 files changed, 457 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 11f8b4ba1c..0585153f69 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -38,6 +38,7 @@ static void idpf_adapter_rel(struct idpf_adapter *adapter);
 static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_configure  = idpf_dev_configure,
.dev_close  = idpf_dev_close,
+   .rx_queue_setup = idpf_rx_queue_setup,
.tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
 };
@@ -63,12 +64,22 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
.tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH,
};
 
+   dev_info->default_rxconf = (struct rte_eth_rxconf) {
+   .rx_free_thresh = IDPF_DEFAULT_RX_FREE_THRESH,
+   };
+
dev_info->tx_desc_lim = (struct rte_eth_desc_lim) {
.nb_max = IDPF_MAX_RING_DESC,
.nb_min = IDPF_MIN_RING_DESC,
.nb_align = IDPF_ALIGN_RING_DESC,
};
 
+   dev_info->rx_desc_lim = (struct rte_eth_desc_lim) {
+   .nb_max = IDPF_MAX_RING_DESC,
+   .nb_min = IDPF_MIN_RING_DESC,
+   .nb_align = IDPF_ALIGN_RING_DESC,
+   };
+
return 0;
 }
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 4afa0a2560..25dd5d85d5 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -8,6 +8,21 @@
 #include "idpf_ethdev.h"
 #include "idpf_rxtx.h"
 
+static int
+check_rx_thresh(uint16_t nb_desc, uint16_t thresh)
+{
+   /* The following constraints must be satisfied:
+*   thresh < rxq->nb_rx_desc
+*/
+   if (thresh >= nb_desc) {
+   PMD_INIT_LOG(ERR, "rx_free_thresh (%u) must be less than %u",
+thresh, nb_desc);
+   return -EINVAL;
+   }
+
+   return 0;
+}
+
 static int
 check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
uint16_t tx_free_thresh)
@@ -56,6 +71,87 @@ check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
return 0;
 }
 
+static void
+reset_split_rx_descq(struct idpf_rx_queue *rxq)
+{
+   uint16_t len;
+   uint32_t i;
+
+   if (rxq == NULL)
+   return;
+
+   len = rxq->nb_rx_desc + IDPF_RX_MAX_BURST;
+
+   for (i = 0; i < len * sizeof(struct virtchnl2_rx_flex_desc_adv_nic_3);
+i++)
+   ((volatile char *)rxq->rx_ring)[i] = 0;
+
+   rxq->rx_tail = 0;
+   rxq->expected_gen_id = 1;
+}
+
+static void
+reset_split_rx_bufq(struct idpf_rx_queue *rxq)
+{
+   uint16_t len;
+   uint32_t i;
+
+   if (rxq == NULL)
+   return;
+
+   len = rxq->nb_rx_desc + IDPF_RX_MAX_BURST;
+
+   for (i = 0; i < len * sizeof(struct virtchnl2_splitq_rx_buf_desc);
+i++)
+   ((volatile char *)rxq->rx_ring)[i] = 0;
+
+   memset(&rxq->fake_mbuf, 0x0, sizeof(rxq->fake_mbuf));
+
+   for (i = 0; i < IDPF_RX_MAX_BURST; i++)
+   rxq->sw_ring[rxq->nb_rx_desc + i] = &rxq->fake_mbuf;
+
+   /* The next descriptor id which can be received. */
+   rxq->rx_next_avail = 0;
+
+   /* The next descriptor id which can be refilled. */
+   rxq->rx_tail = 0;
+   /* The number of descriptors which can be refilled. */
+   rxq->nb_rx_hold = rxq->nb_rx_desc - 1;
+
+   rxq->bufq1 = NULL;
+   rxq->bufq2 = NULL;
+}
+
+static void
+reset_single_rx_queue(struct idpf_rx_queue *rxq)
+{
+   uint16_t len;
+   uint32_t i;
+
+   if (rxq == NULL)
+   return;
+
+   len = rxq->nb_rx_desc + IDPF_RX_MAX_BURST;
+
+   for (i = 0; i < len * sizeof(struct virtchnl2_singleq_rx_buf_desc);
+i++)
+   ((volatile char *)rxq->rx_ring)[i] = 0;
+
+   memset(&rxq->fake_mbuf, 0x0, sizeof(rxq->fake_mbuf));
+
+   for (i = 0; i < IDPF_RX_MAX_BURST; i++)
+   rxq->sw_ring[rxq->nb_rx_desc + i] = &rxq->fake_mbuf;
+
+   rxq->rx_tail = 0;
+   rxq->nb_rx_hold = 0;
+
+   if (rxq->pkt_first_seg != NULL)
+   rte_pktmbuf_free(rxq->pkt_first_seg);
+
+   rxq->pkt_first_seg = NULL;
+   rxq->pkt_last_seg = NULL;
+}
+
 static void
 reset_split_tx_descq(struct idpf_tx_queue *txq)
 {
@@ -145,6 +241,310 @@ reset_single_tx_queue(struct idpf_tx_queue *txq)
txq->next_rs = txq->rs_thresh - 1;
 }
 
+static int
+idpf_rx_split_bufq_setup(struct rte_eth_dev *dev, struct idpf_rx_queue *bufq,
+uint16_t queue_idx, uint16_t r

[PATCH v14 05/18] net/idpf: add support for device start and stop

2022-10-27 Thread Junfeng Guo
Add dev ops dev_start, dev_stop and link_update.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c | 55 ++
 drivers/net/idpf/idpf_rxtx.c   | 20 +
 2 files changed, 75 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 0585153f69..3430d00e92 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -30,17 +30,38 @@ static const char * const idpf_valid_args[] = {
 };
 
 static int idpf_dev_configure(struct rte_eth_dev *dev);
+static int idpf_dev_start(struct rte_eth_dev *dev);
+static int idpf_dev_stop(struct rte_eth_dev *dev);
 static int idpf_dev_close(struct rte_eth_dev *dev);
 static int idpf_dev_info_get(struct rte_eth_dev *dev,
 struct rte_eth_dev_info *dev_info);
 static void idpf_adapter_rel(struct idpf_adapter *adapter);
 
+static int
+idpf_dev_link_update(struct rte_eth_dev *dev,
+__rte_unused int wait_to_complete)
+{
+   struct rte_eth_link new_link;
+
+   memset(&new_link, 0, sizeof(new_link));
+
+   new_link.link_speed = RTE_ETH_SPEED_NUM_NONE;
+   new_link.link_duplex = RTE_ETH_LINK_FULL_DUPLEX;
+   new_link.link_autoneg = !(dev->data->dev_conf.link_speeds &
+ RTE_ETH_LINK_SPEED_FIXED);
+
+   return rte_eth_linkstatus_set(dev, &new_link);
+}
+
 static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_configure  = idpf_dev_configure,
+   .dev_start  = idpf_dev_start,
+   .dev_stop   = idpf_dev_stop,
.dev_close  = idpf_dev_close,
.rx_queue_setup = idpf_rx_queue_setup,
.tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
+   .link_update= idpf_dev_link_update,
 };
 
 static int
@@ -284,6 +305,40 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_dev_start(struct rte_eth_dev *dev)
+{
+   struct idpf_vport *vport = dev->data->dev_private;
+
+   if (dev->data->mtu > vport->max_mtu) {
+   PMD_DRV_LOG(ERR, "MTU should be less than %d", vport->max_mtu);
+   return -1;
+   }
+
+   vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
+
+   /* TODO: start queues */
+
+   if (idpf_vc_ena_dis_vport(vport, true) != 0) {
+   PMD_DRV_LOG(ERR, "Failed to enable vport");
+   return -1;
+   }
+
+   return 0;
+}
+
+static int
+idpf_dev_stop(struct rte_eth_dev *dev)
+{
+   struct idpf_vport *vport = dev->data->dev_private;
+
+   idpf_vc_ena_dis_vport(vport, false);
+
+   /* TODO: stop queues */
+
+   return 0;
+}
+
 static int
 idpf_dev_close(struct rte_eth_dev *dev)
 {
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 25dd5d85d5..3528d2f2c7 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -334,6 +334,11 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   if (rx_conf->rx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -465,6 +470,11 @@ idpf_rx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   if (rx_conf->rx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -569,6 +579,11 @@ idpf_tx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   if (tx_conf->tx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   return -EINVAL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf split txq",
 sizeof(struct idpf_tx_queue),
@@ -691,6 +706,11 @@ idpf_tx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   if (tx_conf->tx_deferred_start) {
+   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
+   retur

[PATCH v14 06/18] net/idpf: add support for queue start

2022-10-27 Thread Junfeng Guo
Add support for these device ops:
 - rx_queue_start
 - tx_queue_start

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  40 ++-
 drivers/net/idpf/idpf_ethdev.h |   9 +
 drivers/net/idpf/idpf_rxtx.c   | 237 +++--
 drivers/net/idpf/idpf_rxtx.h   |   6 +
 drivers/net/idpf/idpf_vchnl.c  | 447 +
 5 files changed, 718 insertions(+), 21 deletions(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 3430d00e92..abbf519977 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -58,6 +58,8 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_start  = idpf_dev_start,
.dev_stop   = idpf_dev_stop,
.dev_close  = idpf_dev_close,
+   .rx_queue_start = idpf_rx_queue_start,
+   .tx_queue_start = idpf_tx_queue_start,
.rx_queue_setup = idpf_rx_queue_setup,
.tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
@@ -305,6 +307,39 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_start_queues(struct rte_eth_dev *dev)
+{
+   struct idpf_rx_queue *rxq;
+   struct idpf_tx_queue *txq;
+   int err = 0;
+   int i;
+
+   for (i = 0; i < dev->data->nb_tx_queues; i++) {
+   txq = dev->data->tx_queues[i];
+   if (txq == NULL || txq->tx_deferred_start)
+   continue;
+   err = idpf_tx_queue_start(dev, i);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR, "Fail to start Tx queue %u", i);
+   return err;
+   }
+   }
+
+   for (i = 0; i < dev->data->nb_rx_queues; i++) {
+   rxq = dev->data->rx_queues[i];
+   if (rxq == NULL || rxq->rx_deferred_start)
+   continue;
+   err = idpf_rx_queue_start(dev, i);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR, "Fail to start Rx queue %u", i);
+   return err;
+   }
+   }
+
+   return err;
+}
+
 static int
 idpf_dev_start(struct rte_eth_dev *dev)
 {
@@ -317,7 +352,10 @@ idpf_dev_start(struct rte_eth_dev *dev)
 
vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
 
-   /* TODO: start queues */
+   if (idpf_start_queues(dev) != 0) {
+   PMD_DRV_LOG(ERR, "Failed to start queues");
+   return -1;
+   }
 
if (idpf_vc_ena_dis_vport(vport, true) != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 84ae6641e2..96c22009e9 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -24,7 +24,9 @@
 #define IDPF_DEFAULT_TXQ_NUM   16
 
 #define IDPF_INVALID_VPORT_IDX 0x
+#define IDPF_TXQ_PER_GRP   1
 #define IDPF_TX_COMPLQ_PER_GRP 1
+#define IDPF_RXQ_PER_GRP   1
 #define IDPF_RX_BUFQ_PER_GRP   2
 
 #define IDPF_CTLQ_ID   -1
@@ -182,6 +184,13 @@ int idpf_vc_check_api_version(struct idpf_adapter 
*adapter);
 int idpf_vc_get_caps(struct idpf_adapter *adapter);
 int idpf_vc_create_vport(struct idpf_adapter *adapter);
 int idpf_vc_destroy_vport(struct idpf_vport *vport);
+int idpf_vc_config_rxqs(struct idpf_vport *vport);
+int idpf_vc_config_rxq(struct idpf_vport *vport, uint16_t rxq_id);
+int idpf_vc_config_txqs(struct idpf_vport *vport);
+int idpf_vc_config_txq(struct idpf_vport *vport, uint16_t txq_id);
+int idpf_switch_queue(struct idpf_vport *vport, uint16_t qid,
+ bool rx, bool on);
+int idpf_vc_ena_dis_queues(struct idpf_vport *vport, bool enable);
 int idpf_vc_ena_dis_vport(struct idpf_vport *vport, bool enable);
 int idpf_read_one_msg(struct idpf_adapter *adapter, uint32_t ops,
  uint16_t buf_len, uint8_t *buf);
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 3528d2f2c7..6d954afd9d 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -334,11 +334,6 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
-   if (rx_conf->rx_deferred_start) {
-   PMD_INIT_LOG(ERR, "Queue start is not supported currently.");
-   return -EINVAL;
-   }
-
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -354,6 +349,7 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, uint16_t 
queue_idx,
rxq->rx_free_thresh = rx_free_thresh;
rxq->queue_id = vport->chunks_info.rx_start_qid + queue_idx;
rxq-

[PATCH v14 07/18] net/idpf: add support for queue stop

2022-10-27 Thread Junfeng Guo
Add support for these device ops:
 - rx_queue_stop
 - tx_queue_stop

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |   1 +
 drivers/net/idpf/idpf_ethdev.c|  14 ++-
 drivers/net/idpf/idpf_rxtx.c  | 148 ++
 drivers/net/idpf/idpf_rxtx.h  |  13 +++
 drivers/net/idpf/idpf_vchnl.c |  69 ++
 5 files changed, 241 insertions(+), 4 deletions(-)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index 46aab2eb61..c25fa5de3f 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -4,6 +4,7 @@
 ; Refer to default.ini for the full list of available PMD features.
 ;
 [Features]
+Queue start/stop = Y
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index abbf519977..05f087a03c 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -59,7 +59,9 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.dev_stop   = idpf_dev_stop,
.dev_close  = idpf_dev_close,
.rx_queue_start = idpf_rx_queue_start,
+   .rx_queue_stop  = idpf_rx_queue_stop,
.tx_queue_start = idpf_tx_queue_start,
+   .tx_queue_stop  = idpf_tx_queue_stop,
.rx_queue_setup = idpf_rx_queue_setup,
.tx_queue_setup = idpf_tx_queue_setup,
.dev_infos_get  = idpf_dev_info_get,
@@ -347,22 +349,26 @@ idpf_dev_start(struct rte_eth_dev *dev)
 
if (dev->data->mtu > vport->max_mtu) {
PMD_DRV_LOG(ERR, "MTU should be less than %d", vport->max_mtu);
-   return -1;
+   goto err_mtu;
}
 
vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
 
if (idpf_start_queues(dev) != 0) {
PMD_DRV_LOG(ERR, "Failed to start queues");
-   return -1;
+   goto err_mtu;
}
 
if (idpf_vc_ena_dis_vport(vport, true) != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
-   return -1;
+   goto err_vport;
}
 
return 0;
+err_vport:
+   idpf_stop_queues(dev);
+err_mtu:
+   return -1;
 }
 
 static int
@@ -372,7 +378,7 @@ idpf_dev_stop(struct rte_eth_dev *dev)
 
idpf_vc_ena_dis_vport(vport, false);
 
-   /* TODO: stop queues */
+   idpf_stop_queues(dev);
 
return 0;
 }
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 6d954afd9d..8d5ec41a1f 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -71,6 +71,55 @@ check_tx_thresh(uint16_t nb_desc, uint16_t tx_rs_thresh,
return 0;
 }
 
+static void
+release_rxq_mbufs(struct idpf_rx_queue *rxq)
+{
+   uint16_t i;
+
+   if (rxq->sw_ring == NULL)
+   return;
+
+   for (i = 0; i < rxq->nb_rx_desc; i++) {
+   if (rxq->sw_ring[i] != NULL) {
+   rte_pktmbuf_free_seg(rxq->sw_ring[i]);
+   rxq->sw_ring[i] = NULL;
+   }
+   }
+}
+
+static void
+release_txq_mbufs(struct idpf_tx_queue *txq)
+{
+   uint16_t nb_desc, i;
+
+   if (txq == NULL || txq->sw_ring == NULL) {
+   PMD_DRV_LOG(DEBUG, "Pointer to rxq or sw_ring is NULL");
+   return;
+   }
+
+   if (txq->sw_nb_desc != 0) {
+   /* For split queue model, descriptor ring */
+   nb_desc = txq->sw_nb_desc;
+   } else {
+   /* For single queue model */
+   nb_desc = txq->nb_tx_desc;
+   }
+   for (i = 0; i < nb_desc; i++) {
+   if (txq->sw_ring[i].mbuf != NULL) {
+   rte_pktmbuf_free_seg(txq->sw_ring[i].mbuf);
+   txq->sw_ring[i].mbuf = NULL;
+   }
+   }
+}
+
+static const struct idpf_rxq_ops def_rxq_ops = {
+   .release_mbufs = release_rxq_mbufs,
+};
+
+static const struct idpf_txq_ops def_txq_ops = {
+   .release_mbufs = release_txq_mbufs,
+};
+
 static void
 reset_split_rx_descq(struct idpf_rx_queue *rxq)
 {
@@ -122,6 +171,14 @@ reset_split_rx_bufq(struct idpf_rx_queue *rxq)
rxq->bufq2 = NULL;
 }
 
+static inline void
+reset_split_rx_queue(struct idpf_rx_queue *rxq)
+{
+   reset_split_rx_descq(rxq);
+   reset_split_rx_bufq(rxq->bufq1);
+   reset_split_rx_bufq(rxq->bufq2);
+}
+
 static void
 reset_single_rx_queue(struct idpf_rx_queue *rxq)
 {
@@ -301,6 +358,7 @@ idpf_rx_split_bufq_setup(struct rte_eth_dev *dev, struct 
idpf_rx_queue *bufq,
bufq->q_set = true;
bufq->qrx_tail = hw->hw_addr + (vport->chunks_info.rx_buf_qtail_start +
 queue_idx * vport->chunks_info.rx_buf_qtail_spacing);
+  

[PATCH v14 08/18] net/idpf: add queue release

2022-10-27 Thread Junfeng Guo
Add support for queue operations:
 - rx_queue_release
 - tx_queue_release

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |  2 +
 drivers/net/idpf/idpf_rxtx.c   | 81 ++
 drivers/net/idpf/idpf_rxtx.h   |  3 ++
 3 files changed, 86 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 05f087a03c..dee1e03d43 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -63,7 +63,9 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.tx_queue_start = idpf_tx_queue_start,
.tx_queue_stop  = idpf_tx_queue_stop,
.rx_queue_setup = idpf_rx_queue_setup,
+   .rx_queue_release   = idpf_dev_rx_queue_release,
.tx_queue_setup = idpf_tx_queue_setup,
+   .tx_queue_release   = idpf_dev_tx_queue_release,
.dev_infos_get  = idpf_dev_info_get,
.link_update= idpf_dev_link_update,
 };
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 8d5ec41a1f..053409b99a 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -171,6 +171,51 @@ reset_split_rx_bufq(struct idpf_rx_queue *rxq)
rxq->bufq2 = NULL;
 }
 
+static void
+idpf_rx_queue_release(void *rxq)
+{
+   struct idpf_rx_queue *q = rxq;
+
+   if (q == NULL)
+   return;
+
+   /* Split queue */
+   if (q->bufq1 != NULL && q->bufq2 != NULL) {
+   q->bufq1->ops->release_mbufs(q->bufq1);
+   rte_free(q->bufq1->sw_ring);
+   rte_memzone_free(q->bufq1->mz);
+   rte_free(q->bufq1);
+   q->bufq2->ops->release_mbufs(q->bufq2);
+   rte_free(q->bufq2->sw_ring);
+   rte_memzone_free(q->bufq2->mz);
+   rte_free(q->bufq2);
+   rte_memzone_free(q->mz);
+   rte_free(q);
+   return;
+   }
+
+   /* Single queue */
+   q->ops->release_mbufs(q);
+   rte_free(q->sw_ring);
+   rte_memzone_free(q->mz);
+   rte_free(q);
+}
+
+static void
+idpf_tx_queue_release(void *txq)
+{
+   struct idpf_tx_queue *q = txq;
+
+   if (q == NULL)
+   return;
+
+   rte_free(q->complq);
+   q->ops->release_mbufs(q);
+   rte_free(q->sw_ring);
+   rte_memzone_free(q->mz);
+   rte_free(q);
+}
+
 static inline void
 reset_split_rx_queue(struct idpf_rx_queue *rxq)
 {
@@ -392,6 +437,12 @@ idpf_rx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed */
+   if (dev->data->rx_queues[queue_idx] != NULL) {
+   idpf_rx_queue_release(dev->data->rx_queues[queue_idx]);
+   dev->data->rx_queues[queue_idx] = NULL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -524,6 +575,12 @@ idpf_rx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_rx_thresh(nb_desc, rx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed */
+   if (dev->data->rx_queues[queue_idx] != NULL) {
+   idpf_rx_queue_release(dev->data->rx_queues[queue_idx]);
+   dev->data->rx_queues[queue_idx] = NULL;
+   }
+
/* Setup Rx description queue */
rxq = rte_zmalloc_socket("idpf rxq",
 sizeof(struct idpf_rx_queue),
@@ -630,6 +687,12 @@ idpf_tx_split_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed. */
+   if (dev->data->tx_queues[queue_idx] != NULL) {
+   idpf_tx_queue_release(dev->data->tx_queues[queue_idx]);
+   dev->data->tx_queues[queue_idx] = NULL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf split txq",
 sizeof(struct idpf_tx_queue),
@@ -754,6 +817,12 @@ idpf_tx_single_queue_setup(struct rte_eth_dev *dev, 
uint16_t queue_idx,
if (check_tx_thresh(nb_desc, tx_rs_thresh, tx_free_thresh) != 0)
return -EINVAL;
 
+   /* Free memory if needed. */
+   if (dev->data->tx_queues[queue_idx] != NULL) {
+   idpf_tx_queue_release(dev->data->tx_queues[queue_idx]);
+   dev->data->tx_queues[queue_idx] = NULL;
+   }
+
/* Allocate the TX queue data structure. */
txq = rte_zmalloc_socket("idpf txq",
 sizeof(struct idpf_tx_queue),
@@ -1102,6 +1171,18 @@ idpf_tx_queue_stop(struct rte_eth_dev *d

[PATCH v14 09/18] net/idpf: add support for MTU configuration

2022-10-27 Thread Junfeng Guo
Add dev ops mtu_set.

Signed-off-by: Beilei Xing 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |  1 +
 drivers/net/idpf/idpf_ethdev.c| 14 ++
 2 files changed, 15 insertions(+)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index c25fa5de3f..c8f4a0d6ed 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -5,6 +5,7 @@
 ;
 [Features]
 Queue start/stop = Y
+MTU update   = Y
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index dee1e03d43..94f4671057 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -35,6 +35,7 @@ static int idpf_dev_stop(struct rte_eth_dev *dev);
 static int idpf_dev_close(struct rte_eth_dev *dev);
 static int idpf_dev_info_get(struct rte_eth_dev *dev,
 struct rte_eth_dev_info *dev_info);
+static int idpf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu);
 static void idpf_adapter_rel(struct idpf_adapter *adapter);
 
 static int
@@ -68,6 +69,7 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.tx_queue_release   = idpf_dev_tx_queue_release,
.dev_infos_get  = idpf_dev_info_get,
.link_update= idpf_dev_link_update,
+   .mtu_set= idpf_dev_mtu_set,
 };
 
 static int
@@ -110,6 +112,18 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
return 0;
 }
 
+static int
+idpf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu __rte_unused)
+{
+   /* mtu setting is forbidden if port is start */
+   if (dev->data->dev_started) {
+   PMD_DRV_LOG(ERR, "port must be stopped before configuration");
+   return -EBUSY;
+   }
+
+   return 0;
+}
+
 static int
 idpf_init_vport_req_info(struct rte_eth_dev *dev)
 {
-- 
2.34.1



[PATCH v14 10/18] net/idpf: add support for basic Rx datapath

2022-10-27 Thread Junfeng Guo
Add basic Rx support in split queue mode and single queue mode.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |   2 +
 drivers/net/idpf/idpf_rxtx.c   | 273 +
 drivers/net/idpf/idpf_rxtx.h   |   7 +-
 3 files changed, 281 insertions(+), 1 deletion(-)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 94f4671057..df2a760673 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -375,6 +375,8 @@ idpf_dev_start(struct rte_eth_dev *dev)
goto err_mtu;
}
 
+   idpf_set_rx_function(dev);
+
if (idpf_vc_ena_dis_vport(vport, true) != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
goto err_vport;
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 053409b99a..ea499c4d37 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1208,3 +1208,276 @@ idpf_stop_queues(struct rte_eth_dev *dev)
PMD_DRV_LOG(WARNING, "Fail to stop Tx queue %d", i);
}
 }
+
+static void
+idpf_split_rx_bufq_refill(struct idpf_rx_queue *rx_bufq)
+{
+   volatile struct virtchnl2_splitq_rx_buf_desc *rx_buf_ring;
+   volatile struct virtchnl2_splitq_rx_buf_desc *rx_buf_desc;
+   uint16_t nb_refill = rx_bufq->rx_free_thresh;
+   uint16_t nb_desc = rx_bufq->nb_rx_desc;
+   uint16_t next_avail = rx_bufq->rx_tail;
+   struct rte_mbuf *nmb[rx_bufq->rx_free_thresh];
+   struct rte_eth_dev *dev;
+   uint64_t dma_addr;
+   uint16_t delta;
+   int i;
+
+   if (rx_bufq->nb_rx_hold < rx_bufq->rx_free_thresh)
+   return;
+
+   rx_buf_ring = rx_bufq->rx_ring;
+   delta = nb_desc - next_avail;
+   if (unlikely(delta < nb_refill)) {
+   if (likely(rte_pktmbuf_alloc_bulk(rx_bufq->mp, nmb, delta) == 
0)) {
+   for (i = 0; i < delta; i++) {
+   rx_buf_desc = &rx_buf_ring[next_avail + i];
+   rx_bufq->sw_ring[next_avail + i] = nmb[i];
+   dma_addr = 
rte_cpu_to_le_64(rte_mbuf_data_iova_default(nmb[i]));
+   rx_buf_desc->hdr_addr = 0;
+   rx_buf_desc->pkt_addr = dma_addr;
+   }
+   nb_refill -= delta;
+   next_avail = 0;
+   rx_bufq->nb_rx_hold -= delta;
+   } else {
+   dev = &rte_eth_devices[rx_bufq->port_id];
+   dev->data->rx_mbuf_alloc_failed += nb_desc - next_avail;
+   PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u 
queue_id=%u",
+  rx_bufq->port_id, rx_bufq->queue_id);
+   return;
+   }
+   }
+
+   if (nb_desc - next_avail >= nb_refill) {
+   if (likely(rte_pktmbuf_alloc_bulk(rx_bufq->mp, nmb, nb_refill) 
== 0)) {
+   for (i = 0; i < nb_refill; i++) {
+   rx_buf_desc = &rx_buf_ring[next_avail + i];
+   rx_bufq->sw_ring[next_avail + i] = nmb[i];
+   dma_addr = 
rte_cpu_to_le_64(rte_mbuf_data_iova_default(nmb[i]));
+   rx_buf_desc->hdr_addr = 0;
+   rx_buf_desc->pkt_addr = dma_addr;
+   }
+   next_avail += nb_refill;
+   rx_bufq->nb_rx_hold -= nb_refill;
+   } else {
+   dev = &rte_eth_devices[rx_bufq->port_id];
+   dev->data->rx_mbuf_alloc_failed += nb_desc - next_avail;
+   PMD_RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u 
queue_id=%u",
+  rx_bufq->port_id, rx_bufq->queue_id);
+   }
+   }
+
+   IDPF_PCI_REG_WRITE(rx_bufq->qrx_tail, next_avail);
+
+   rx_bufq->rx_tail = next_avail;
+}
+
+uint16_t
+idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
+ uint16_t nb_pkts)
+{
+   volatile struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc_ring;
+   volatile struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc;
+   uint16_t pktlen_gen_bufq_id;
+   struct idpf_rx_queue *rxq;
+   struct rte_mbuf *rxm;
+   uint16_t rx_id_bufq1;
+   uint16_t rx_id_bufq2;
+   uint16_t pkt_len;
+   uint16_t bufq_id;
+   uint16_t gen_id;
+   uint16_t rx_id;
+   uint16_t nb_rx;
+
+   nb_rx = 0;
+   rxq = rx_queue;
+
+   if (unlikely(rxq == NULL) || unlikely(!rxq->q_started))
+   return nb_rx;
+
+   rx_id = rxq->rx_tail;
+   rx_id_bufq1 = rxq->bufq1->rx_next_avail;
+   rx_id_bufq2 = rxq->bufq2->rx_next_avail;
+   rx_desc_ring = rxq->rx_ring;
+
+   wh

[PATCH v14 11/18] net/idpf: add support for basic Tx datapath

2022-10-27 Thread Junfeng Guo
Add basic Tx support in split queue mode and single queue mode.

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |   3 +
 drivers/net/idpf/idpf_ethdev.h |   1 +
 drivers/net/idpf/idpf_rxtx.c   | 357 +
 drivers/net/idpf/idpf_rxtx.h   |  10 +
 4 files changed, 371 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index df2a760673..6fb56e584d 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -88,6 +88,8 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
 
dev_info->max_mac_addrs = IDPF_NUM_MACADDR_MAX;
 
+   dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
+
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
.tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH,
@@ -376,6 +378,7 @@ idpf_dev_start(struct rte_eth_dev *dev)
}
 
idpf_set_rx_function(dev);
+   idpf_set_tx_function(dev);
 
if (idpf_vc_ena_dis_vport(vport, true) != 0) {
PMD_DRV_LOG(ERR, "Failed to enable vport");
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 96c22009e9..af0a8e2970 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -35,6 +35,7 @@
 
 #define IDPF_MIN_BUF_SIZE  1024
 #define IDPF_MAX_FRAME_SIZE9728
+#define IDPF_MIN_FRAME_SIZE14
 
 #define IDPF_NUM_MACADDR_MAX   64
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index ea499c4d37..f55d2143b9 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1365,6 +1365,148 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
return nb_rx;
 }
 
+static inline void
+idpf_split_tx_free(struct idpf_tx_queue *cq)
+{
+   volatile struct idpf_splitq_tx_compl_desc *compl_ring = cq->compl_ring;
+   volatile struct idpf_splitq_tx_compl_desc *txd;
+   uint16_t next = cq->tx_tail;
+   struct idpf_tx_entry *txe;
+   struct idpf_tx_queue *txq;
+   uint16_t gen, qid, q_head;
+   uint8_t ctype;
+
+   txd = &compl_ring[next];
+   gen = (rte_le_to_cpu_16(txd->qid_comptype_gen) &
+   IDPF_TXD_COMPLQ_GEN_M) >> IDPF_TXD_COMPLQ_GEN_S;
+   if (gen != cq->expected_gen_id)
+   return;
+
+   ctype = (rte_le_to_cpu_16(txd->qid_comptype_gen) &
+   IDPF_TXD_COMPLQ_COMPL_TYPE_M) >> IDPF_TXD_COMPLQ_COMPL_TYPE_S;
+   qid = (rte_le_to_cpu_16(txd->qid_comptype_gen) &
+   IDPF_TXD_COMPLQ_QID_M) >> IDPF_TXD_COMPLQ_QID_S;
+   q_head = rte_le_to_cpu_16(txd->q_head_compl_tag.compl_tag);
+   txq = cq->txqs[qid - cq->tx_start_qid];
+
+   switch (ctype) {
+   case IDPF_TXD_COMPLT_RE:
+   if (q_head == 0)
+   txq->last_desc_cleaned = txq->nb_tx_desc - 1;
+   else
+   txq->last_desc_cleaned = q_head - 1;
+   if (unlikely((txq->last_desc_cleaned % 32) == 0)) {
+   PMD_DRV_LOG(ERR, "unexpected desc (head = %u) 
completion.",
+   q_head);
+   return;
+   }
+
+   break;
+   case IDPF_TXD_COMPLT_RS:
+   txq->nb_free++;
+   txq->nb_used--;
+   txe = &txq->sw_ring[q_head];
+   if (txe->mbuf != NULL) {
+   rte_pktmbuf_free_seg(txe->mbuf);
+   txe->mbuf = NULL;
+   }
+   break;
+   default:
+   PMD_DRV_LOG(ERR, "unknown completion type.");
+   return;
+   }
+
+   if (++next == cq->nb_tx_desc) {
+   next = 0;
+   cq->expected_gen_id ^= 1;
+   }
+
+   cq->tx_tail = next;
+}
+
+uint16_t
+idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
+ uint16_t nb_pkts)
+{
+   struct idpf_tx_queue *txq = (struct idpf_tx_queue *)tx_queue;
+   volatile struct idpf_flex_tx_sched_desc *txr;
+   volatile struct idpf_flex_tx_sched_desc *txd;
+   struct idpf_tx_entry *sw_ring;
+   struct idpf_tx_entry *txe, *txn;
+   uint16_t nb_used, tx_id, sw_id;
+   struct rte_mbuf *tx_pkt;
+   uint16_t nb_to_clean;
+   uint16_t nb_tx = 0;
+
+   if (unlikely(txq == NULL) || unlikely(!txq->q_started))
+   return nb_tx;
+
+   txr = txq->desc_ring;
+   sw_ring = txq->sw_ring;
+   tx_id = txq->tx_tail;
+   sw_id = txq->sw_tail;
+   txe = &sw_ring[sw_id];
+
+   for (nb_tx = 0; nb_tx < nb_pkts; nb_tx++) {
+   tx_pkt = tx_pkts[nb_tx];
+
+   if (txq->nb_free <= txq->free_thresh) {
+   /* TODO: Need to refine
+* 1. free and clean: Better to decide

[PATCH v14 12/18] net/idpf: support parsing packet type

2022-10-27 Thread Junfeng Guo
Parse packet type during receiving packets.

Signed-off-by: Wenjun Wu 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c |   6 +
 drivers/net/idpf/idpf_ethdev.h |   6 +
 drivers/net/idpf/idpf_rxtx.c   |  11 ++
 drivers/net/idpf/idpf_rxtx.h   |   5 +
 drivers/net/idpf/idpf_vchnl.c  | 240 +
 5 files changed, 268 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 6fb56e584d..630bdabcd4 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -709,6 +709,12 @@ idpf_adapter_init(struct rte_pci_device *pci_dev, struct 
idpf_adapter *adapter)
goto err_api;
}
 
+   ret = idpf_get_pkt_type(adapter);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to set ptype table");
+   goto err_api;
+   }
+
adapter->caps = rte_zmalloc("idpf_caps",
sizeof(struct virtchnl2_get_capabilities), 0);
if (adapter->caps == NULL) {
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index af0a8e2970..db9af58f72 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -39,6 +39,8 @@
 
 #define IDPF_NUM_MACADDR_MAX   64
 
+#define IDPF_MAX_PKT_TYPE  1024
+
 #define IDPF_VLAN_TAG_SIZE 4
 #define IDPF_ETH_OVERHEAD \
(RTE_ETHER_HDR_LEN + RTE_ETHER_CRC_LEN + IDPF_VLAN_TAG_SIZE * 2)
@@ -125,6 +127,8 @@ struct idpf_adapter {
/* Max config queue number per VC message */
uint32_t max_rxq_per_msg;
uint32_t max_txq_per_msg;
+
+   uint32_t ptype_tbl[IDPF_MAX_PKT_TYPE] __rte_cache_min_aligned;
 };
 
 TAILQ_HEAD(idpf_adapter_list, idpf_adapter);
@@ -182,6 +186,7 @@ atomic_set_cmd(struct idpf_adapter *adapter, enum 
virtchnl_ops ops)
 struct idpf_adapter *idpf_find_adapter(struct rte_pci_device *pci_dev);
 void idpf_handle_virtchnl_msg(struct rte_eth_dev *dev);
 int idpf_vc_check_api_version(struct idpf_adapter *adapter);
+int idpf_get_pkt_type(struct idpf_adapter *adapter);
 int idpf_vc_get_caps(struct idpf_adapter *adapter);
 int idpf_vc_create_vport(struct idpf_adapter *adapter);
 int idpf_vc_destroy_vport(struct idpf_vport *vport);
@@ -193,6 +198,7 @@ int idpf_switch_queue(struct idpf_vport *vport, uint16_t 
qid,
  bool rx, bool on);
 int idpf_vc_ena_dis_queues(struct idpf_vport *vport, bool enable);
 int idpf_vc_ena_dis_vport(struct idpf_vport *vport, bool enable);
+int idpf_vc_query_ptype_info(struct idpf_adapter *adapter);
 int idpf_read_one_msg(struct idpf_adapter *adapter, uint32_t ops,
  uint16_t buf_len, uint8_t *buf);
 
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index f55d2143b9..a980714060 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1281,6 +1281,7 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
volatile struct virtchnl2_rx_flex_desc_adv_nic_3 *rx_desc;
uint16_t pktlen_gen_bufq_id;
struct idpf_rx_queue *rxq;
+   const uint32_t *ptype_tbl;
struct rte_mbuf *rxm;
uint16_t rx_id_bufq1;
uint16_t rx_id_bufq2;
@@ -1300,6 +1301,7 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rx_id_bufq1 = rxq->bufq1->rx_next_avail;
rx_id_bufq2 = rxq->bufq2->rx_next_avail;
rx_desc_ring = rxq->rx_ring;
+   ptype_tbl = rxq->adapter->ptype_tbl;
 
while (nb_rx < nb_pkts) {
rx_desc = &rx_desc_ring[rx_id];
@@ -1347,6 +1349,10 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rxm->next = NULL;
rxm->nb_segs = 1;
rxm->port = rxq->port_id;
+   rxm->packet_type =
+   ptype_tbl[(rte_le_to_cpu_16(rx_desc->ptype_err_fflags0) 
&
+  VIRTCHNL2_RX_FLEX_DESC_ADV_PTYPE_M) >>
+ VIRTCHNL2_RX_FLEX_DESC_ADV_PTYPE_S];
 
rx_pkts[nb_rx++] = rxm;
}
@@ -1533,6 +1539,7 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
volatile union virtchnl2_rx_desc *rxdp;
union virtchnl2_rx_desc rxd;
struct idpf_rx_queue *rxq;
+   const uint32_t *ptype_tbl;
uint16_t rx_id, nb_hold;
struct rte_eth_dev *dev;
uint16_t rx_packet_len;
@@ -1551,6 +1558,7 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
 
rx_id = rxq->rx_tail;
rx_ring = rxq->rx_ring;
+   ptype_tbl = rxq->adapter->ptype_tbl;
 
while (nb_rx < nb_pkts) {
rxdp = &rx_ring[rx_id];
@@ -1603,6 +1611,9 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rxm->pkt_len = rx_packet_len;
rxm->data_len = rx_packet_len;
rxm->port = rxq->port_id;
+   rxm->packet_type =
+   
ptype_tbl[(

[PATCH v14 13/18] net/idpf: add support for write back based on ITR expire

2022-10-27 Thread Junfeng Guo
Enable write back on ITR expire, then packets can be received one by
one.

Signed-off-by: Beilei Xing 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c | 116 +
 drivers/net/idpf/idpf_ethdev.h |  13 
 drivers/net/idpf/idpf_vchnl.c  | 111 +++
 3 files changed, 240 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 630bdabcd4..8045073780 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -327,6 +327,90 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_config_rx_queues_irqs(struct rte_eth_dev *dev)
+{
+   struct idpf_vport *vport = dev->data->dev_private;
+   struct idpf_adapter *adapter = vport->adapter;
+   struct virtchnl2_queue_vector *qv_map;
+   struct idpf_hw *hw = &adapter->hw;
+   uint32_t dynctl_reg_start;
+   uint32_t itrn_reg_start;
+   uint32_t dynctl_val, itrn_val;
+   uint16_t i;
+
+   qv_map = rte_zmalloc("qv_map",
+   dev->data->nb_rx_queues *
+   sizeof(struct virtchnl2_queue_vector), 0);
+   if (qv_map == NULL) {
+   PMD_DRV_LOG(ERR, "Failed to allocate %d queue-vector map",
+   dev->data->nb_rx_queues);
+   goto qv_map_alloc_err;
+   }
+
+   /* Rx interrupt disabled, Map interrupt only for writeback */
+
+   /* The capability flags adapter->caps->other_caps here should be
+* compared with bit VIRTCHNL2_CAP_WB_ON_ITR. The if condition should
+* be updated when the FW can return correct flag bits.
+*/
+   if (adapter->caps->other_caps != 0) {
+   dynctl_reg_start =
+   vport->recv_vectors->vchunks.vchunks->dynctl_reg_start;
+   itrn_reg_start =
+   vport->recv_vectors->vchunks.vchunks->itrn_reg_start;
+   dynctl_val = IDPF_READ_REG(hw, dynctl_reg_start);
+   PMD_DRV_LOG(DEBUG, "Value of dynctl_reg_start is 0x%x",
+   dynctl_val);
+   itrn_val = IDPF_READ_REG(hw, itrn_reg_start);
+   PMD_DRV_LOG(DEBUG, "Value of itrn_reg_start is 0x%x", itrn_val);
+   /* Force write-backs by setting WB_ON_ITR bit in DYN_CTL
+* register. WB_ON_ITR and INTENA are mutually exclusive
+* bits. Setting WB_ON_ITR bits means TX and RX Descs
+* are written back based on ITR expiration irrespective
+* of INTENA setting.
+*/
+   /* TBD: need to tune INTERVAL value for better performance. */
+   if (itrn_val != 0)
+   IDPF_WRITE_REG(hw,
+  dynctl_reg_start,
+  VIRTCHNL2_ITR_IDX_0  <<
+  PF_GLINT_DYN_CTL_ITR_INDX_S |
+  PF_GLINT_DYN_CTL_WB_ON_ITR_M |
+  itrn_val <<
+  PF_GLINT_DYN_CTL_INTERVAL_S);
+   else
+   IDPF_WRITE_REG(hw,
+  dynctl_reg_start,
+  VIRTCHNL2_ITR_IDX_0  <<
+  PF_GLINT_DYN_CTL_ITR_INDX_S |
+  PF_GLINT_DYN_CTL_WB_ON_ITR_M |
+  IDPF_DFLT_INTERVAL <<
+  PF_GLINT_DYN_CTL_INTERVAL_S);
+   }
+   for (i = 0; i < dev->data->nb_rx_queues; i++) {
+   /* map all queues to the same vector */
+   qv_map[i].queue_id = vport->chunks_info.rx_start_qid + i;
+   qv_map[i].vector_id =
+   vport->recv_vectors->vchunks.vchunks->start_vector_id;
+   }
+   vport->qv_map = qv_map;
+
+   if (idpf_vc_config_irq_map_unmap(vport, true) != 0) {
+   PMD_DRV_LOG(ERR, "config interrupt mapping failed");
+   goto config_irq_map_err;
+   }
+
+   return 0;
+
+config_irq_map_err:
+   rte_free(vport->qv_map);
+   vport->qv_map = NULL;
+
+qv_map_alloc_err:
+   return -1;
+}
+
 static int
 idpf_start_queues(struct rte_eth_dev *dev)
 {
@@ -364,6 +448,10 @@ static int
 idpf_dev_start(struct rte_eth_dev *dev)
 {
struct idpf_vport *vport = dev->data->dev_private;
+   struct idpf_adapter *adapter = vport->adapter;
+   uint16_t num_allocated_vectors =
+   adapter->caps->num_allocated_vectors;
+   uint16_t req_vecs_num;
 
if (dev->data->mtu > vport->max_mtu) {
PMD_DRV_LOG(ERR, "MTU should be less than %d", vport->max_mtu);
@@ -372,6 +460,23 @@ idpf_dev_start(struct rte_eth_dev *dev)
 
vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD;
 
+   req_vecs_num = IDPF_DFLT_Q_VEC_NUM;
+  

[PATCH v14 14/18] net/idpf: add support for RSS

2022-10-27 Thread Junfeng Guo
Add RSS support.

Signed-off-by: Beilei Xing 
Signed-off-by: Junfeng Guo 
---
 drivers/net/idpf/idpf_ethdev.c | 116 +
 drivers/net/idpf/idpf_ethdev.h |  26 
 drivers/net/idpf/idpf_vchnl.c  |  97 +++
 3 files changed, 239 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 8045073780..ea7d3bfd7e 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -86,6 +86,7 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
dev_info->max_mtu = dev_info->max_rx_pktlen - IDPF_ETH_OVERHEAD;
dev_info->min_mtu = RTE_ETHER_MIN_MTU;
 
+   dev_info->flow_type_rss_offloads = IDPF_RSS_OFFLOAD_ALL;
dev_info->max_mac_addrs = IDPF_NUM_MACADDR_MAX;
 
dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
@@ -198,6 +199,8 @@ idpf_parse_devarg_id(char *name)
return val;
 }
 
+#define IDPF_RSS_KEY_LEN 52
+
 static int
 idpf_init_vport(struct rte_eth_dev *dev)
 {
@@ -218,6 +221,10 @@ idpf_init_vport(struct rte_eth_dev *dev)
vport->max_mtu = vport_info->max_mtu;
rte_memcpy(vport->default_mac_addr,
   vport_info->default_mac_addr, ETH_ALEN);
+   vport->rss_algorithm = vport_info->rss_algorithm;
+   vport->rss_key_size = RTE_MIN(IDPF_RSS_KEY_LEN,
+vport_info->rss_key_size);
+   vport->rss_lut_size = vport_info->rss_lut_size;
vport->sw_idx = idx;
 
for (i = 0; i < vport_info->chunks.num_chunks; i++) {
@@ -275,10 +282,102 @@ idpf_init_vport(struct rte_eth_dev *dev)
return 0;
 }
 
+static int
+idpf_config_rss(struct idpf_vport *vport)
+{
+   int ret;
+
+   ret = idpf_vc_set_rss_key(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS key");
+   return ret;
+   }
+
+   ret = idpf_vc_set_rss_lut(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS lut");
+   return ret;
+   }
+
+   ret = idpf_vc_set_rss_hash(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS hash");
+   return ret;
+   }
+
+   return ret;
+}
+
+static int
+idpf_init_rss(struct idpf_vport *vport)
+{
+   struct rte_eth_rss_conf *rss_conf;
+   uint16_t i, nb_q, lut_size;
+   int ret = 0;
+
+   rss_conf = &vport->dev_data->dev_conf.rx_adv_conf.rss_conf;
+   nb_q = vport->dev_data->nb_rx_queues;
+
+   vport->rss_key = rte_zmalloc("rss_key",
+vport->rss_key_size, 0);
+   if (vport->rss_key == NULL) {
+   PMD_INIT_LOG(ERR, "Failed to allocate RSS key");
+   ret = -ENOMEM;
+   goto err_alloc_key;
+   }
+
+   lut_size = vport->rss_lut_size;
+   vport->rss_lut = rte_zmalloc("rss_lut",
+sizeof(uint32_t) * lut_size, 0);
+   if (vport->rss_lut == NULL) {
+   PMD_INIT_LOG(ERR, "Failed to allocate RSS lut");
+   ret = -ENOMEM;
+   goto err_alloc_lut;
+   }
+
+   if (rss_conf->rss_key == NULL) {
+   for (i = 0; i < vport->rss_key_size; i++)
+   vport->rss_key[i] = (uint8_t)rte_rand();
+   } else if (rss_conf->rss_key_len != vport->rss_key_size) {
+   PMD_INIT_LOG(ERR, "Invalid RSS key length in RSS configuration, 
should be %d",
+vport->rss_key_size);
+   ret = -EINVAL;
+   goto err_cfg_key;
+   } else {
+   rte_memcpy(vport->rss_key, rss_conf->rss_key,
+  vport->rss_key_size);
+   }
+
+   for (i = 0; i < lut_size; i++)
+   vport->rss_lut[i] = i % nb_q;
+
+   vport->rss_hf = IDPF_DEFAULT_RSS_HASH_EXPANDED;
+
+   ret = idpf_config_rss(vport);
+   if (ret != 0) {
+   PMD_INIT_LOG(ERR, "Failed to configure RSS");
+   goto err_cfg_key;
+   }
+
+   return ret;
+
+err_cfg_key:
+   rte_free(vport->rss_lut);
+   vport->rss_lut = NULL;
+err_alloc_lut:
+   rte_free(vport->rss_key);
+   vport->rss_key = NULL;
+err_alloc_key:
+   return ret;
+}
+
 static int
 idpf_dev_configure(struct rte_eth_dev *dev)
 {
+   struct idpf_vport *vport = dev->data->dev_private;
struct rte_eth_conf *conf = &dev->data->dev_conf;
+   struct idpf_adapter *adapter = vport->adapter;
+   int ret;
 
if (conf->link_speeds & RTE_ETH_LINK_SPEED_FIXED) {
PMD_INIT_LOG(ERR, "Setting link speed is not supported");
@@ -324,6 +423,17 @@ idpf_dev_configure(struct rte_eth_dev *dev)
return -1;
}
 
+   if (adapter->caps->rss_caps != 0 && dev->data->nb_rx_queues != 0) {
+   ret = idpf_init_rss(vport);
+   if (ret != 0) {
+  

[PATCH v14 15/18] net/idpf: add support for Rx offloading

2022-10-27 Thread Junfeng Guo
Add Rx offloading support:
 - support CHKSUM and RSS offload for split queue model
 - support CHKSUM offload for single queue model

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |   5 ++
 drivers/net/idpf/idpf_ethdev.c|   6 ++
 drivers/net/idpf/idpf_rxtx.c  | 123 ++
 3 files changed, 134 insertions(+)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index c8f4a0d6ed..6eefac9529 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -3,9 +3,14 @@
 ;
 ; Refer to default.ini for the full list of available PMD features.
 ;
+; A feature with "P" indicates only be supported when non-vector path
+; is selected.
+;
 [Features]
 Queue start/stop = Y
 MTU update   = Y
+L3 checksum offload  = P
+L4 checksum offload  = P
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index ea7d3bfd7e..88f3db3267 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -89,6 +89,12 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
dev_info->flow_type_rss_offloads = IDPF_RSS_OFFLOAD_ALL;
dev_info->max_mac_addrs = IDPF_NUM_MACADDR_MAX;
 
+   dev_info->rx_offload_capa =
+   RTE_ETH_RX_OFFLOAD_IPV4_CKSUM   |
+   RTE_ETH_RX_OFFLOAD_UDP_CKSUM|
+   RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
+   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM;
+
dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index a980714060..f15e61a785 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1209,6 +1209,73 @@ idpf_stop_queues(struct rte_eth_dev *dev)
}
 }
 
+#define IDPF_RX_FLEX_DESC_ADV_STATUS0_XSUM_S   \
+   (RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_IPE_S) | \
+RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_L4E_S) | \
+RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EIPE_S) |\
+RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EUDPE_S))
+
+static inline uint64_t
+idpf_splitq_rx_csum_offload(uint8_t err)
+{
+   uint64_t flags = 0;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_L3L4P_S)) == 0))
+   return flags;
+
+   if (likely((err & IDPF_RX_FLEX_DESC_ADV_STATUS0_XSUM_S) == 0)) {
+   flags |= (RTE_MBUF_F_RX_IP_CKSUM_GOOD |
+ RTE_MBUF_F_RX_L4_CKSUM_GOOD);
+   return flags;
+   }
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_IPE_S)) != 0))
+   flags |= RTE_MBUF_F_RX_IP_CKSUM_BAD;
+   else
+   flags |= RTE_MBUF_F_RX_IP_CKSUM_GOOD;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_L4E_S)) != 0))
+   flags |= RTE_MBUF_F_RX_L4_CKSUM_BAD;
+   else
+   flags |= RTE_MBUF_F_RX_L4_CKSUM_GOOD;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EIPE_S)) != 0))
+   flags |= RTE_MBUF_F_RX_OUTER_IP_CKSUM_BAD;
+
+   if (unlikely((err & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_EUDPE_S)) != 0))
+   flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_BAD;
+   else
+   flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_GOOD;
+
+   return flags;
+}
+
+#define IDPF_RX_FLEX_DESC_ADV_HASH1_S  0
+#define IDPF_RX_FLEX_DESC_ADV_HASH2_S  16
+#define IDPF_RX_FLEX_DESC_ADV_HASH3_S  24
+
+static inline uint64_t
+idpf_splitq_rx_rss_offload(struct rte_mbuf *mb,
+  volatile struct virtchnl2_rx_flex_desc_adv_nic_3 
*rx_desc)
+{
+   uint8_t status_err0_qw0;
+   uint64_t flags = 0;
+
+   status_err0_qw0 = rx_desc->status_err0_qw0;
+
+   if ((status_err0_qw0 & 
RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_RSS_VALID_S)) != 0) {
+   flags |= RTE_MBUF_F_RX_RSS_HASH;
+   mb->hash.rss = (rte_le_to_cpu_16(rx_desc->hash1) <<
+   IDPF_RX_FLEX_DESC_ADV_HASH1_S) |
+   ((uint32_t)(rx_desc->ff2_mirrid_hash2.hash2) <<
+IDPF_RX_FLEX_DESC_ADV_HASH2_S) |
+   ((uint32_t)(rx_desc->hash3) <<
+IDPF_RX_FLEX_DESC_ADV_HASH3_S);
+   }
+
+   return flags;
+}
+
 static void
 idpf_split_rx_bufq_refill(struct idpf_rx_queue *rx_bufq)
 {
@@ -1282,9 +1349,11 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
uint16_t pktlen_gen_bufq_id;
struct idpf_rx_queue *rxq;
const uint32_t *ptype_tbl;
+   uint8_t status_err0_qw1;

[PATCH v14 16/18] net/idpf: add support for Tx offloading

2022-10-27 Thread Junfeng Guo
Add Tx offloading support:
 - support TSO

Signed-off-by: Beilei Xing 
Signed-off-by: Xiaoyun Li 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |   1 +
 drivers/net/idpf/idpf_ethdev.c|   4 +-
 drivers/net/idpf/idpf_rxtx.c  | 128 +-
 drivers/net/idpf/idpf_rxtx.h  |  22 +
 4 files changed, 152 insertions(+), 3 deletions(-)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index 6eefac9529..86c182021f 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -9,6 +9,7 @@
 [Features]
 Queue start/stop = Y
 MTU update   = Y
+TSO  = P
 L3 checksum offload  = P
 L4 checksum offload  = P
 Linux= Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 88f3db3267..868f28d003 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -95,7 +95,9 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM;
 
-   dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
+   dev_info->tx_offload_capa =
+   RTE_ETH_TX_OFFLOAD_TCP_TSO  |
+   RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index f15e61a785..cc296d7ab1 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -1506,6 +1506,49 @@ idpf_split_tx_free(struct idpf_tx_queue *cq)
cq->tx_tail = next;
 }
 
+/* Check if the context descriptor is needed for TX offloading */
+static inline uint16_t
+idpf_calc_context_desc(uint64_t flags)
+{
+   if ((flags & RTE_MBUF_F_TX_TCP_SEG) != 0)
+   return 1;
+
+   return 0;
+}
+
+/* set TSO context descriptor
+ */
+static inline void
+idpf_set_splitq_tso_ctx(struct rte_mbuf *mbuf,
+   union idpf_tx_offload tx_offload,
+   volatile union idpf_flex_tx_ctx_desc *ctx_desc)
+{
+   uint16_t cmd_dtype;
+   uint32_t tso_len;
+   uint8_t hdr_len;
+
+   if (tx_offload.l4_len == 0) {
+   PMD_TX_LOG(DEBUG, "L4 length set to 0");
+   return;
+   }
+
+   hdr_len = tx_offload.l2_len +
+   tx_offload.l3_len +
+   tx_offload.l4_len;
+   cmd_dtype = IDPF_TX_DESC_DTYPE_FLEX_TSO_CTX |
+   IDPF_TX_FLEX_CTX_DESC_CMD_TSO;
+   tso_len = mbuf->pkt_len - hdr_len;
+
+   ctx_desc->tso.qw1.cmd_dtype = rte_cpu_to_le_16(cmd_dtype);
+   ctx_desc->tso.qw0.hdr_len = hdr_len;
+   ctx_desc->tso.qw0.mss_rt =
+   rte_cpu_to_le_16((uint16_t)mbuf->tso_segsz &
+IDPF_TXD_FLEX_CTX_MSS_RT_M);
+   ctx_desc->tso.qw0.flex_tlen =
+   rte_cpu_to_le_32(tso_len &
+IDPF_TXD_FLEX_CTX_MSS_RT_M);
+}
+
 uint16_t
 idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
  uint16_t nb_pkts)
@@ -1514,11 +1557,14 @@ idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
volatile struct idpf_flex_tx_sched_desc *txr;
volatile struct idpf_flex_tx_sched_desc *txd;
struct idpf_tx_entry *sw_ring;
+   union idpf_tx_offload tx_offload = {0};
struct idpf_tx_entry *txe, *txn;
uint16_t nb_used, tx_id, sw_id;
struct rte_mbuf *tx_pkt;
uint16_t nb_to_clean;
uint16_t nb_tx = 0;
+   uint64_t ol_flags;
+   uint16_t nb_ctx;
 
if (unlikely(txq == NULL) || unlikely(!txq->q_started))
return nb_tx;
@@ -1548,7 +1594,29 @@ idpf_splitq_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
 
if (txq->nb_free < tx_pkt->nb_segs)
break;
-   nb_used = tx_pkt->nb_segs;
+
+   ol_flags = tx_pkt->ol_flags;
+   tx_offload.l2_len = tx_pkt->l2_len;
+   tx_offload.l3_len = tx_pkt->l3_len;
+   tx_offload.l4_len = tx_pkt->l4_len;
+   tx_offload.tso_segsz = tx_pkt->tso_segsz;
+   /* Calculate the number of context descriptors needed. */
+   nb_ctx = idpf_calc_context_desc(ol_flags);
+   nb_used = tx_pkt->nb_segs + nb_ctx;
+
+   /* context descriptor */
+   if (nb_ctx != 0) {
+   volatile union idpf_flex_tx_ctx_desc *ctx_desc =
+   (volatile union idpf_flex_tx_ctx_desc *)&txr[tx_id];
+
+   if ((ol_flags & RTE_MBUF_F_TX_TCP_SEG) != 0)
+   idpf_set_splitq_tso_ctx(tx_pkt, tx_offload,
+   ctx_desc);
+
+   tx_id+

[PATCH v14 17/18] net/idpf: add AVX512 data path for single queue model

2022-10-27 Thread Junfeng Guo
Add support of AVX512 vector data path for single queue model.

Signed-off-by: Wenjun Wu 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/idpf.rst|  19 +
 drivers/net/idpf/idpf_ethdev.c  |   3 +-
 drivers/net/idpf/idpf_ethdev.h  |   5 +
 drivers/net/idpf/idpf_rxtx.c| 145 
 drivers/net/idpf/idpf_rxtx.h|  21 +
 drivers/net/idpf/idpf_rxtx_vec_avx512.c | 871 
 drivers/net/idpf/idpf_rxtx_vec_common.h | 100 +++
 drivers/net/idpf/meson.build|  28 +
 8 files changed, 1191 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/idpf/idpf_rxtx_vec_avx512.c
 create mode 100644 drivers/net/idpf/idpf_rxtx_vec_common.h

diff --git a/doc/guides/nics/idpf.rst b/doc/guides/nics/idpf.rst
index c1001d5d0c..3039c61748 100644
--- a/doc/guides/nics/idpf.rst
+++ b/doc/guides/nics/idpf.rst
@@ -64,3 +64,22 @@ Refer to the document :ref:`compiling and testing a PMD for 
a NIC tx_offload_capa =
RTE_ETH_TX_OFFLOAD_TCP_TSO  |
-   RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
+   RTE_ETH_TX_OFFLOAD_MULTI_SEGS   |
+   RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE;
 
dev_info->default_txconf = (struct rte_eth_txconf) {
.tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH,
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 8d0804f603..7d54e5db60 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -162,6 +162,11 @@ struct idpf_adapter {
uint32_t max_txq_per_msg;
 
uint32_t ptype_tbl[IDPF_MAX_PKT_TYPE] __rte_cache_min_aligned;
+
+   bool rx_vec_allowed;
+   bool tx_vec_allowed;
+   bool rx_use_avx512;
+   bool tx_use_avx512;
 };
 
 TAILQ_HEAD(idpf_adapter_list, idpf_adapter);
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index cc296d7ab1..9e20f2b9d3 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -4,9 +4,11 @@
 
 #include 
 #include 
+#include 
 
 #include "idpf_ethdev.h"
 #include "idpf_rxtx.h"
+#include "idpf_rxtx_vec_common.h"
 
 static int
 check_rx_thresh(uint16_t nb_desc, uint16_t thresh)
@@ -252,6 +254,8 @@ reset_single_rx_queue(struct idpf_rx_queue *rxq)
 
rxq->pkt_first_seg = NULL;
rxq->pkt_last_seg = NULL;
+   rxq->rxrearm_start = 0;
+   rxq->rxrearm_nb = 0;
 }
 
 static void
@@ -2073,25 +2077,166 @@ idpf_prep_pkts(__rte_unused void *tx_queue, struct 
rte_mbuf **tx_pkts,
return i;
 }
 
+static void __rte_cold
+release_rxq_mbufs_vec(struct idpf_rx_queue *rxq)
+{
+   const uint16_t mask = rxq->nb_rx_desc - 1;
+   uint16_t i;
+
+   if (rxq->sw_ring == NULL || rxq->rxrearm_nb >= rxq->nb_rx_desc)
+   return;
+
+   /* free all mbufs that are valid in the ring */
+   if (rxq->rxrearm_nb == 0) {
+   for (i = 0; i < rxq->nb_rx_desc; i++) {
+   if (rxq->sw_ring[i] != NULL)
+   rte_pktmbuf_free_seg(rxq->sw_ring[i]);
+   }
+   } else {
+   for (i = rxq->rx_tail; i != rxq->rxrearm_start; i = (i + 1) & 
mask) {
+   if (rxq->sw_ring[i] != NULL)
+   rte_pktmbuf_free_seg(rxq->sw_ring[i]);
+   }
+   }
+
+   rxq->rxrearm_nb = rxq->nb_rx_desc;
+
+   /* set all entries to NULL */
+   memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc);
+}
+
+static const struct idpf_rxq_ops def_singleq_rx_ops_vec = {
+   .release_mbufs = release_rxq_mbufs_vec,
+};
+
+static inline int
+idpf_singleq_rx_vec_setup_default(struct idpf_rx_queue *rxq)
+{
+   uintptr_t p;
+   struct rte_mbuf mb_def = { .buf_addr = 0 }; /* zeroed mbuf */
+
+   mb_def.nb_segs = 1;
+   mb_def.data_off = RTE_PKTMBUF_HEADROOM;
+   mb_def.port = rxq->port_id;
+   rte_mbuf_refcnt_set(&mb_def, 1);
+
+   /* prevent compiler reordering: rearm_data covers previous fields */
+   rte_compiler_barrier();
+   p = (uintptr_t)&mb_def.rearm_data;
+   rxq->mbuf_initializer = *(uint64_t *)p;
+   return 0;
+}
+
+int __rte_cold
+idpf_singleq_rx_vec_setup(struct idpf_rx_queue *rxq)
+{
+   rxq->ops = &def_singleq_rx_ops_vec;
+   return idpf_singleq_rx_vec_setup_default(rxq);
+}
+
 void
 idpf_set_rx_function(struct rte_eth_dev *dev)
 {
struct idpf_vport *vport = dev->data->dev_private;
+#ifdef RTE_ARCH_X86
+   struct idpf_adapter *ad = vport->adapter;
+   struct idpf_rx_queue *rxq;
+   int i;
+
+   if (idpf_rx_vec_dev_check_default(dev) == IDPF_VECTOR_PATH &&
+   rte_vect_get_max_simd_bitwidth() >= RTE_VECT_SIMD_128) {
+   ad->rx_vec_allowed = true;
+
+   if (rte_vect_get_max_simd_bitwidth() >= RTE_VECT_SIMD_512)
+#ifdef CC_AVX512_SUPPORT
+   if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_AVX512F) == 1 
&&
+   rte_cpu

[PATCH v14 18/18] net/idpf: add support for timestamp offload

2022-10-27 Thread Junfeng Guo
Add support for timestamp offload.

Signed-off-by: Wenjing Qiao 
Signed-off-by: Junfeng Guo 
---
 doc/guides/nics/features/idpf.ini |  1 +
 drivers/net/idpf/idpf_ethdev.c|  5 +-
 drivers/net/idpf/idpf_ethdev.h|  3 ++
 drivers/net/idpf/idpf_rxtx.c  | 65 ++
 drivers/net/idpf/idpf_rxtx.h  | 90 +++
 5 files changed, 163 insertions(+), 1 deletion(-)

diff --git a/doc/guides/nics/features/idpf.ini 
b/doc/guides/nics/features/idpf.ini
index 86c182021f..5fbad9600c 100644
--- a/doc/guides/nics/features/idpf.ini
+++ b/doc/guides/nics/features/idpf.ini
@@ -12,6 +12,7 @@ MTU update   = Y
 TSO  = P
 L3 checksum offload  = P
 L4 checksum offload  = P
+Timestamp offload= P
 Linux= Y
 x86-32   = Y
 x86-64   = Y
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 62b13b602e..84e7421d76 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -22,6 +22,8 @@ rte_spinlock_t idpf_adapter_lock;
 struct idpf_adapter_list idpf_adapter_list;
 bool idpf_adapter_list_init;
 
+uint64_t idpf_timestamp_dynflag;
+
 static const char * const idpf_valid_args[] = {
IDPF_TX_SINGLE_Q,
IDPF_RX_SINGLE_Q,
@@ -93,7 +95,8 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct 
rte_eth_dev_info *dev_info)
RTE_ETH_RX_OFFLOAD_IPV4_CKSUM   |
RTE_ETH_RX_OFFLOAD_UDP_CKSUM|
RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
-   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM;
+   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM |
+   RTE_ETH_RX_OFFLOAD_TIMESTAMP;
 
dev_info->tx_offload_capa =
RTE_ETH_TX_OFFLOAD_TCP_TSO  |
diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h
index 7d54e5db60..ccdf4abe40 100644
--- a/drivers/net/idpf/idpf_ethdev.h
+++ b/drivers/net/idpf/idpf_ethdev.h
@@ -167,6 +167,9 @@ struct idpf_adapter {
bool tx_vec_allowed;
bool rx_use_avx512;
bool tx_use_avx512;
+
+   /* For PTP */
+   uint64_t time_hw;
 };
 
 TAILQ_HEAD(idpf_adapter_list, idpf_adapter);
diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c
index 9e20f2b9d3..bafa007faf 100644
--- a/drivers/net/idpf/idpf_rxtx.c
+++ b/drivers/net/idpf/idpf_rxtx.c
@@ -10,6 +10,8 @@
 #include "idpf_rxtx.h"
 #include "idpf_rxtx_vec_common.h"
 
+static int idpf_timestamp_dynfield_offset = -1;
+
 static int
 check_rx_thresh(uint16_t nb_desc, uint16_t thresh)
 {
@@ -900,6 +902,24 @@ idpf_tx_queue_setup(struct rte_eth_dev *dev, uint16_t 
queue_idx,
return idpf_tx_split_queue_setup(dev, queue_idx, nb_desc,
 socket_id, tx_conf);
 }
+
+static int
+idpf_register_ts_mbuf(struct idpf_rx_queue *rxq)
+{
+   int err;
+   if ((rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) != 0) {
+   /* Register mbuf field and flag for Rx timestamp */
+   err = 
rte_mbuf_dyn_rx_timestamp_register(&idpf_timestamp_dynfield_offset,
+
&idpf_timestamp_dynflag);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR,
+   "Cannot register mbuf field/flag for 
timestamp");
+   return -EINVAL;
+   }
+   }
+   return 0;
+}
+
 static int
 idpf_alloc_single_rxq_mbufs(struct idpf_rx_queue *rxq)
 {
@@ -993,6 +1013,13 @@ idpf_rx_queue_init(struct rte_eth_dev *dev, uint16_t 
rx_queue_id)
return -EINVAL;
}
 
+   err = idpf_register_ts_mbuf(rxq);
+   if (err != 0) {
+   PMD_DRV_LOG(ERR, "fail to regidter timestamp mbuf %u",
+   rx_queue_id);
+   return -EIO;
+   }
+
if (rxq->bufq1 == NULL) {
/* Single queue */
err = idpf_alloc_single_rxq_mbufs(rxq);
@@ -1354,6 +1381,7 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
struct idpf_rx_queue *rxq;
const uint32_t *ptype_tbl;
uint8_t status_err0_qw1;
+   struct idpf_adapter *ad;
struct rte_mbuf *rxm;
uint16_t rx_id_bufq1;
uint16_t rx_id_bufq2;
@@ -1363,9 +1391,11 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
uint16_t gen_id;
uint16_t rx_id;
uint16_t nb_rx;
+   uint64_t ts_ns;
 
nb_rx = 0;
rxq = rx_queue;
+   ad = rxq->adapter;
 
if (unlikely(rxq == NULL) || unlikely(!rxq->q_started))
return nb_rx;
@@ -1376,6 +1406,9 @@ idpf_splitq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rx_desc_ring = rxq->rx_ring;
ptype_tbl = rxq->adapter->ptype_tbl;
 
+   if ((rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) != 0)
+   rxq->hw_register_set = 1;
+

Re: [PATCH] bus/vdev: automatically add eth alias for net drivers

2022-10-27 Thread David Marchand
On Thu, Oct 20, 2022 at 1:52 PM Ferruh Yigit  wrote:
> >> Honestly I think the status quo is OK:
> >> We have some aliases in some PMD for some historical reason
> >> and everybody looks OK with that. Isn't it?
> >>
> >
> > Well, the inconsistency bugs me a little, but if others feel the status quo
> > is ok, I'm ok with that.
>
> In my perspective this is for cleanup, and new PMDs keep adding alias
> because they are copying from existing drivers.
> Except from above there is no harm to have alias.

Do we have a "valid" case of adding new aliases?
I don't think it is the case, so we can warn of new aliases
introduction in checkpatches.sh.

At worse, if a valid case is identified later, checkpatches.sh is only
a warning in patchwork and maintainers will manually review this
warning.


-- 
David Marchand



Re: [RFC] pcapng: record received RSS hash in pcap file

2022-10-27 Thread David Marchand
On Tue, Jul 26, 2022 at 11:37 PM Stephen Hemminger
 wrote:
>
> There is an option for recording RSS hash with packets in the
> pcapng standard. This implements this for all received packets.
>
> There is a corner case that can not be addressed with current
> DPDK API's. If using rte_flow() and some hardware it is possible
> to write a flow rule that uses another hash function like XOR.
> But there is no API that records this, or provides the algorithm
> info on a per-packet basis.
>
> The current version of wireshark does not display the recorded
> hash option. But if we build it they will come.

I updated the commitlog adding some info about "yet to be released
4.1" wireshark.

>
> Signed-off-by: Stephen Hemminger 
Tested-by: Ben Magistro 

Applied, thanks.


-- 
David Marchand



Re: [PATCH] mempool: cache align mempool cache objects

2022-10-27 Thread Olivier Matz
Hi Morten,

On Wed, Oct 26, 2022 at 04:44:36PM +0200, Morten Brørup wrote:
> Add __rte_cache_aligned to the objs array.
> 
> It makes no difference in the general case, but if get/put operations are
> always 32 objects, it will reduce the number of memory (or last level
> cache) accesses from five to four 64 B cache lines for every get/put
> operation.
> 
> For readability reasons, an example using 16 objects follows:
> 
> Currently, with 16 objects (128B), we access to 3
> cache lines:
> 
>   ┌┐
>   │len │
> cache ││---
> line0 ││ ^
>   ││ |
>   ├┤ | 16 objects
>   ││ | 128B
> cache ││ |
> line1 ││ |
>   ││ |
>   ├┤ |
>   ││_v_
> cache ││
> line2 ││
>   ││
>   └┘
> 
> With the alignment, it is also 3 cache lines:
> 
>   ┌┐
>   │len │
> cache ││
> line0 ││
>   ││
>   ├┤---
>   ││ ^
> cache ││ |
> line1 ││ |
>   ││ |
>   ├┤ | 16 objects
>   ││ | 128B
> cache ││ |
> line2 ││ |
>   ││ v
>   └┘---
> 
> However, accessing the objects at the bottom of the mempool cache is a
> special case, where cache line0 is also used for objects.
> 
> Consider the next burst (and any following bursts):
> 
> Current:
>   ┌┐
>   │len │
> cache ││
> line0 ││
>   ││
>   ├┤
>   ││
> cache ││
> line1 ││
>   ││
>   ├┤
>   ││
> cache ││---
> line2 ││ ^
>   ││ |
>   ├┤ | 16 objects
>   ││ | 128B
> cache ││ |
> line3 ││ |
>   ││ |
>   ├┤ |
>   ││_v_
> cache ││
> line4 ││
>   ││
>   └┘
> 4 cache lines touched, incl. line0 for len.
> 
> With the proposed alignment:
>   ┌┐
>   │len │
> cache ││
> line0 ││
>   ││
>   ├┤
>   ││
> cache ││
> line1 ││
>   ││
>   ├┤
>   ││
> cache ││
> line2 ││
>   ││
>   ├┤
>   ││---
> cache ││ ^
> line3 ││ |
>   ││ | 16 objects
>   ├┤ | 128B
>   ││ |
> cache ││ |
> line4 ││ |
>   ││_v_
>   └┘
> Only 3 cache lines touched, incl. line0 for len.

I understand your logic, but are we sure that having an application that
works with bulks of 32 means that the cache will stay aligned to 32
elements for the whole life of the application?

In an application, the alignment of the cache can change if you have
any of:
- software queues (reassembly for instance)
- packet duplication (bridge, multicast)
- locally generated packets (keepalive, control protocol)
- pipeline to other cores

Even with testpmd, which work by bulk of 32, I can see that the size
of the cache filling is not aligned to 32. Right after starting the
application, we already have this:

  internal cache infos:
cache_size=250
cache_count[0]=231

This is probably related to the hw rx rings size, number of queues,
number of ports.

The "250" default value for cache size in testpmd is questionable, but
with --mbcache=256, the behavior is similar.

Also, when we transmit to a NIC, the mbufs are not returned immediatly
to the pool, they may stay in the hw tx ring during some time, which is
a driver decision.

After processing traffic on cores 8 and 24 with this testpmd, I get:
cache_count[0]=231
cache_count[8]=123
cache_count[24]=122

In my opinion, it is not realistic to think that the mempool cache will
remain aligned to cachelines. In these conditions, it looks better to
keep the structure packed to avoid wasting memory.

Olivier


> 
> Credits go to Olivier Matz for the nice ASCII graphics.
> 
> Signed-off-by: Morten Brørup 
> ---
>  lib/mempool/rte_mempool.h | 6 --
>  1 file changed, 4 insertions(+), 2 deletions(-)
> 
> diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
> index 1f5707f46a..3725a72951 100644
> --- a/lib/mempool/rte_mempool.h
> +++ b/lib/mempool/rte_mempool.h
> @@ -86,11 +86,13 @@ struct rte_mempool_cache {
>   uint32_t size;/**< Size of the cache */
>   uint32_t flushthresh; /**< Threshold before we flush excess elements */
>   uint32_t len; /**< Current cache count */
> - /*
> + /**
> +  * Cache objects
> +  *
>* Cache is allocated to this size to allow it to overflow in certain
>* cases to avoid needless emptying of cache.
>*/
> - void *objs[RTE_MEMPOOL_CACHE_MAX_SIZE * 2]; /**< Cache objects */
> + void *objs[RTE_MEMPOOL_CACHE_MAX_SIZE * 2] __rte_cache_aligned;
>  } __rte_ca

Re: [PATCH] bus/vdev: automatically add eth alias for net drivers

2022-10-27 Thread Ferruh Yigit

On 10/27/2022 8:58 AM, David Marchand wrote:

On Thu, Oct 20, 2022 at 1:52 PM Ferruh Yigit  wrote:

Honestly I think the status quo is OK:
We have some aliases in some PMD for some historical reason
and everybody looks OK with that. Isn't it?



Well, the inconsistency bugs me a little, but if others feel the status quo
is ok, I'm ok with that.


In my perspective this is for cleanup, and new PMDs keep adding alias
because they are copying from existing drivers.
Except from above there is no harm to have alias.


Do we have a "valid" case of adding new aliases?
I don't think it is the case, so we can warn of new aliases
introduction in checkpatches.sh.



I commented a few of them to drop alias.
checkpatch can be an option, but my intention was to drop old code to 
reduce noise, not to add more :)


OK to keep the alias if removing it will cause more trouble.


At worse, if a valid case is identified later, checkpatches.sh is only
a warning in patchwork and maintainers will manually review this
warning.






RE: [PATCH v4 2/2] net/ice: fix vlan offload

2022-10-27 Thread Huang, ZhiminX
> -Original Message-
> From: Mingjin Ye 
> Sent: Thursday, October 27, 2022 1:10 AM
> To: dev@dpdk.org
> Cc: sta...@dpdk.org; Zhou, YidingX ; Ye, MingjinX
> ; Richardson, Bruce ;
> Konstantin Ananyev ; Yang, Qiming
> ; Zhang, Qi Z ; Lu, Wenzhuo
> ; Junyu Jiang ; Rong, Leyi
> ; Wisam Jaddo ; Xia, Chenbo
> ; Hemant Agrawal ;
> Jerin Jacob ; Ajit Khaparde
> 
> Subject: [PATCH v4 2/2] net/ice: fix vlan offload
> 
> The vlan tag and flag in Rx descriptor are not processed on vector path, then
> the upper application can't fetch the tci from mbuf.
> 
> This commit add handling of vlan RX offloading.
> 
> Fixes: c68a52b8b38c ("net/ice: support vector SSE in Rx")
> Fixes: ece1f8a8f1c8 ("net/ice: switch to flexible descriptor in SSE path")
> Fixes: 12443386a0b0 ("net/ice: support flex Rx descriptor RxDID22")
> Fixes: 214f452f7d5f ("net/ice: add AVX2 offload Rx")
> Fixes: 7f85d5ebcfe1 ("net/ice: add AVX512 vector path")
> Fixes: 295968d17407 ("ethdev: add namespace")
> Fixes: 808a17b3c1e6 ("net/ice: add Rx AVX512 offload path")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Mingjin Ye 
> 
> v3:
>   * Fix macros in ice_rxtx_vec_sse.c source file.
> v4:
>   * Fix ice_rx_desc_to_olflags_v define in ice_rxtx_vec_sse.c source file.
> ---
Tested-by: Zhimin Huang 




RE: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq

2022-10-27 Thread Huang, ZhiminX
> -Original Message-
> From: Mingjin Ye 
> Sent: Thursday, October 27, 2022 1:10 AM
> To: dev@dpdk.org
> Cc: sta...@dpdk.org; Zhou, YidingX ; Ye, MingjinX
> ; Singh, Aman Deep ;
> Zhang, Yuying 
> Subject: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq
> 
> After setting vlan offload in testpmd, the result is not updated to rxq. 
> Therefore,
> the queue needs to be reconfigured after executing the "vlan offload" related
> commands.
> 
> Fixes: a47aa8b97afe ("app/testpmd: add vlan offload support")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Mingjin Ye 
> ---
Tested-by: Zhimin Huang 




RE: [PATCH v3 1/4] telemetry: support boolean type

2022-10-27 Thread Power, Ciara



> -Original Message-
> From: Richardson, Bruce 
> Sent: Tuesday 25 October 2022 11:39
> To: David Marchand 
> Cc: Mattias Rönnblom ; dev@dpdk.org; Morten
> Brørup ; Power, Ciara
> 
> Subject: Re: [PATCH v3 1/4] telemetry: support boolean type
> 
> On Tue, Oct 25, 2022 at 11:43:27AM +0200, David Marchand wrote:
> > On Tue, Oct 25, 2022 at 11:34 AM Mattias Rönnblom
>  wrote:
> > >
> > > On 2022-10-25 11:00, David Marchand wrote:
> > > > Add the boolean type RTE_TEL_BOOL_VAL for values in arrays and
> dicts.
> > > >
> > > > Signed-off-by: David Marchand 
> > > > Acked-by: Morten Brørup 
> > > > Acked-by: Bruce Richardson 
> > > > Acked-by: Ciara Power 
> > > > ---
> > > > Changes since v1:
> > > > - fixed doxygen description,

> > I'll let the telemetry maintainers decide on the fate of those patches.
> > Thanks.
> >
> Well, given the lateness in the release process, if it's a choice between
> getting the patches in as-is, vs not getting them in at all, I think having 
> them
> in 22.11 is the better choice. While there is cleanup that can be done, I 
> don't
> see this patch as making the overall code any worse.
> 
> My 2c.
> 
> /Bruce

Agreed, it's a good addition to the library, +1 to having it in 22.11

Thanks,
Ciara


Re: [PATCH] net/mana: disable driver by default

2022-10-27 Thread Ferruh Yigit

On 10/27/2022 7:13 AM, Thomas Monjalon wrote:

27/10/2022 00:57, Ferruh Yigit:

On 10/26/2022 10:53 PM, Thomas Monjalon wrote:

13/10/2022 18:30, Long Li:

Subject: [PATCH] net/mana: disable driver by default

Driver is disabled by default because its dependencies are not upstreamed
yet, code is available for development and investigation.

When all dependencies are upstreamed, driver can be enabled back.

Fixes: 517ed6e2d590 ("net/mana: add basic driver with build environment")

Signed-off-by: Ferruh Yigit 


Acked-by: Long Li 


Acked-by: Thomas Monjalon 




Applied to dpdk-next-net/main, thanks.


Looks like we cannot dereference the doc:
doc/guides/nics/mana.rst:document isn't included in any toctree




ack

Will fix in next-net, and add mana.rst back to index to remove the warning.



Re: [PATCH] app/testpmd: fix testpmd crash when quit with mlx5 avail_thresh enabled

2022-10-27 Thread Singh, Aman Deep



On 10/24/2022 8:44 AM, Spike Du wrote:

When testpmd quit with mlx5 avail_thresh enabled, a rte timer handler
delays to reconfigure rx queue to re-arm this event. However at the same
time, testpmd is destroying rx queues.
It's never a valid use case for mlx5 avail_thresh. Before testpmd quit,
user should disable avail_thresh configuration to not handle the events.
This is documented in mlx5 driver guide.

To avoid the crash in such use case, check port status, if it is not
RTE_PORT_STARTED, don't process the avail_thresh event.

Fixes: 0edfc9b08316 ("app/testpmd: add Host Shaper command")

Signed-off-by: Spike Du 


Acked-by: Aman Singh


---
  drivers/net/mlx5/mlx5_testpmd.c | 9 +
  1 file changed, 9 insertions(+)

diff --git a/drivers/net/mlx5/mlx5_testpmd.c b/drivers/net/mlx5/mlx5_testpmd.c
index ed84583..1a9ec78 100644
--- a/drivers/net/mlx5/mlx5_testpmd.c
+++ b/drivers/net/mlx5/mlx5_testpmd.c
@@ -25,6 +25,7 @@
  
  static uint8_t host_shaper_avail_thresh_triggered[RTE_MAX_ETHPORTS];

  #define SHAPER_DISABLE_DELAY_US 10 /* 100ms */
+extern struct rte_port *ports;


This might not be needed, as we have testpmd.h included.

  
  /**

   * Disable the host shaper and re-arm available descriptor threshold event.
@@ -39,7 +40,15 @@
uint16_t port_id = port_rxq_id & 0x;
uint16_t qid = (port_rxq_id >> 16) & 0x;
struct rte_eth_rxq_info qinfo;
+   struct rte_port *port;
  
+	port = &ports[port_id];

+   if (port->port_status != RTE_PORT_STARTED) {
+   printf("%s port_status(%d) is incorrect, stop avail_thresh "
+  "event processing.\n",
+  __func__, port->port_status);
+   return;
+   }
printf("%s disable shaper\n", __func__);
if (rte_eth_rx_queue_info_get(port_id, qid, &qinfo)) {
printf("rx_queue_info_get returns error\n");




RE: [PATCH] mempool: cache align mempool cache objects

2022-10-27 Thread Morten Brørup
> From: Olivier Matz [mailto:olivier.m...@6wind.com]
> Sent: Thursday, 27 October 2022 10.35
> 
> Hi Morten,
> 
> On Wed, Oct 26, 2022 at 04:44:36PM +0200, Morten Brørup wrote:
> > Add __rte_cache_aligned to the objs array.
> >
> > It makes no difference in the general case, but if get/put operations
> are
> > always 32 objects, it will reduce the number of memory (or last level
> > cache) accesses from five to four 64 B cache lines for every get/put
> > operation.
> >
> > For readability reasons, an example using 16 objects follows:
> >
> > Currently, with 16 objects (128B), we access to 3
> > cache lines:
> >
> >   ┌┐
> >   │len │
> > cache ││---
> > line0 ││ ^
> >   ││ |
> >   ├┤ | 16 objects
> >   ││ | 128B
> > cache ││ |
> > line1 ││ |
> >   ││ |
> >   ├┤ |
> >   ││_v_
> > cache ││
> > line2 ││
> >   ││
> >   └┘
> >
> > With the alignment, it is also 3 cache lines:
> >
> >   ┌┐
> >   │len │
> > cache ││
> > line0 ││
> >   ││
> >   ├┤---
> >   ││ ^
> > cache ││ |
> > line1 ││ |
> >   ││ |
> >   ├┤ | 16 objects
> >   ││ | 128B
> > cache ││ |
> > line2 ││ |
> >   ││ v
> >   └┘---
> >
> > However, accessing the objects at the bottom of the mempool cache is
> a
> > special case, where cache line0 is also used for objects.
> >
> > Consider the next burst (and any following bursts):
> >
> > Current:
> >   ┌┐
> >   │len │
> > cache ││
> > line0 ││
> >   ││
> >   ├┤
> >   ││
> > cache ││
> > line1 ││
> >   ││
> >   ├┤
> >   ││
> > cache ││---
> > line2 ││ ^
> >   ││ |
> >   ├┤ | 16 objects
> >   ││ | 128B
> > cache ││ |
> > line3 ││ |
> >   ││ |
> >   ├┤ |
> >   ││_v_
> > cache ││
> > line4 ││
> >   ││
> >   └┘
> > 4 cache lines touched, incl. line0 for len.
> >
> > With the proposed alignment:
> >   ┌┐
> >   │len │
> > cache ││
> > line0 ││
> >   ││
> >   ├┤
> >   ││
> > cache ││
> > line1 ││
> >   ││
> >   ├┤
> >   ││
> > cache ││
> > line2 ││
> >   ││
> >   ├┤
> >   ││---
> > cache ││ ^
> > line3 ││ |
> >   ││ | 16 objects
> >   ├┤ | 128B
> >   ││ |
> > cache ││ |
> > line4 ││ |
> >   ││_v_
> >   └┘
> > Only 3 cache lines touched, incl. line0 for len.
> 
> I understand your logic, but are we sure that having an application
> that
> works with bulks of 32 means that the cache will stay aligned to 32
> elements for the whole life of the application?
> 
> In an application, the alignment of the cache can change if you have
> any of:
> - software queues (reassembly for instance)
> - packet duplication (bridge, multicast)
> - locally generated packets (keepalive, control protocol)
> - pipeline to other cores
> 
> Even with testpmd, which work by bulk of 32, I can see that the size
> of the cache filling is not aligned to 32. Right after starting the
> application, we already have this:
> 
>   internal cache infos:
> cache_size=250
> cache_count[0]=231
> 
> This is probably related to the hw rx rings size, number of queues,
> number of ports.
> 
> The "250" default value for cache size in testpmd is questionable, but
> with --mbcache=256, the behavior is similar.
> 
> Also, when we transmit to a NIC, the mbufs are not returned immediatly
> to the pool, they may stay in the hw tx ring during some time, which is
> a driver decision.
> 
> After processing traffic on cores 8 and 24 with this testpmd, I get:
> cache_count[0]=231
> cache_count[8]=123
> cache_count[24]=122
> 
> In my opinion, it is not realistic to think that the mempool cache will
> remain aligned to cachelines. In these conditions, it looks better to
> keep the structure packed to avoid wasting memory.

I agree that is a special use case to only access the mempool cache in bursts 
of 32 objects, so the accesses are always cache line aligned. (Generalized, the 
burst size must not be 32; a burst size that is a multiple of 
RTE_CACHE_LINE_SIZE/sizeof(void*), i.e. a burst size of 8 on a 64-bit 
architecture, will do.)

Adding a hole of 52 byte per mempool cache is nothing, considering that the 
mempool cache already uses 8 KB (RTE_MEMPOOL_CACHE_MAX_SIZE * 2 * sizeof(void*) 
= 1024 * 8 byte) for the objects.

Also - assuming that memory allocations are cache line aligned - the 52 byte of 

回复: 回复: [PATCH v1 3/3] examples/l3fwd-power: enable PMD power mgmt on Arm

2022-10-27 Thread Feifei Wang


> -邮件原件-
> 发件人: Thomas Monjalon 
> 发送时间: Friday, October 21, 2022 4:42 AM
> 收件人: David Marchand 
> 抄送: Hunt, David ; Ruifeng Wang
> ; dev@dpdk.org; nd ; Feifei
> Wang 
> 主题: Re: 回复: [PATCH v1 3/3] examples/l3fwd-power: enable PMD power
> mgmt on Arm
> 
> 11/10/2022 09:56, Feifei Wang:
> > David Marchand 
> > > > On 25/08/2022 07:42, Feifei Wang wrote:
> > > > > --- a/examples/l3fwd-power/main.c
> > > > > +++ b/examples/l3fwd-power/main.c
> > > > > @@ -432,8 +432,16 @@ static void
> > > > >   signal_exit_now(int sigtype)
> > > > >   {
> > > > >
> > > > > - if (sigtype == SIGINT)
> > > > > + if (sigtype == SIGINT) {
> > > > > +#if defined(RTE_ARCH_ARM64)
> > >
> > > Having a arch specific behavior in the application shows that there
> > > is something wrong either in the API, or in the Arm implementation of
> the API.
> > > I don't think this is a good solution.
> > >
> > > Can't we find a better alternative? By changing the API probably?
> > Sorry I do not understand ' shows that there is something wrong either in
> the API'
> 
> David means the application developer should not have to be aware of the
> arch differences.
> When you use an API, you don't check how it is implemented, and you are
> not supposed to use #ifdef.
> The API must be arch-agnostic.

Ok, Understand. Thanks for the explanation.
> 
> > Here we call ' rte_power_monitor_wakeup' API is due to that we need to
> > wake up all cores from WFE instructions in arm, and then l3fwd can exit
> correctly.
> >
> > This is due to that arm arch is different from x86, if there is no
> > packets received, x86's 'UMONITOR' can automatically exit from energy-
> saving state after waiting for a period of time.
> > But arm's 'WFE' can not exit automatically. It will wait 'SEV'
> > instructions in wake_up API to wake up it.
> >
> > Finally, if user want to exit l3fwd by  'SIGINT' in arm, main core
> > should firstly call 'wake_up' API to force worker cores to exit from energy-
> saving state.
> > Otherwise, the worker will stay in the energy-saving state forever if no
> packet is received.
> 
> Please find a way to have a common API,
> even if the API implementation is empty in x86 case.

Yes, I think what we need to do is not a create a new API, it is to look
for a correct location to call 'rte_power_monitor_wakeup'. 

> 
> > >
> > >
> > > > > + /**
> > > > > +  * wake_up api does not need input parameter on Arm,
> > > > > +  * so 0 is meaningless here.
> > > > > +  */
> > > > > + rte_power_monitor_wakeup(0); #endif
> > > > >   quit_signal = true;
> > > > > + }
> 
> 



回复: [PATCH v1 3/3] examples/l3fwd-power: enable PMD power mgmt on Arm

2022-10-27 Thread Feifei Wang


> -邮件原件-
> 发件人: Stephen Hemminger 
> 发送时间: Friday, October 21, 2022 6:10 AM
> 收件人: Feifei Wang 
> 抄送: David Hunt ; dev@dpdk.org; nd
> ; Ruifeng Wang 
> 主题: Re: [PATCH v1 3/3] examples/l3fwd-power: enable PMD power mgmt
> on Arm
> 
> On Thu, 25 Aug 2022 14:42:51 +0800
> Feifei Wang  wrote:
> 
> > diff --git a/examples/l3fwd-power/main.c b/examples/l3fwd-
> power/main.c
> > index 887c6eae3f..2bd0d700f0 100644
> > --- a/examples/l3fwd-power/main.c
> > +++ b/examples/l3fwd-power/main.c
> > @@ -432,8 +432,16 @@ static void
> >  signal_exit_now(int sigtype)
> >  {
> >
> > -   if (sigtype == SIGINT)
> > +   if (sigtype == SIGINT) {
> > +#if defined(RTE_ARCH_ARM64)
> > +   /**
> > +* wake_up api does not need input parameter on Arm,
> > +* so 0 is meaningless here.
> > +*/
> > +   rte_power_monitor_wakeup(0);
> > +#endif
> > quit_signal = true;
> > +   }
> >
> 
> This method is problematic. There is no guarantee that power monitor is
> async signal safe.

Agree with this. We will put 'rte_power_monitor_wakeup' out of signal_exit.
And put it after that main_lcore exit 'main_empty_poll_loop':
-
rte_eal_mp_remote_launch(main_telemetry_loop, NULL, CALL_MAIN);


%wake up all worker calls from low power state.
rte_power_monitor_wakeup(0);

if (app_mode == APP_MODE_EMPTY_POLL || app_mode == APP_MODE_TELEMETRY)
launch_timer(rte_lcore_id());

RTE_LCORE_FOREACH_WORKER(lcore_id) {
if (rte_eal_wait_lcore(lcore_id) < 0)
return -1;
}
-




[Bug 1116] Failsafe PMD: Crashes with multiple secondary processes

2022-10-27 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1116

Bug ID: 1116
   Summary: Failsafe PMD: Crashes with multiple secondary
processes
   Product: DPDK
   Version: 21.11
  Hardware: All
OS: Linux
Status: UNCONFIRMED
  Severity: major
  Priority: Normal
 Component: ethdev
  Assignee: dev@dpdk.org
  Reporter: madhuker.myt...@oracle.com
  Target Milestone: ---

On Hyper-V Linux VM's with failsafe/tap PMD crashes when we launch mutiple(two)
secondary processes.
Primary process able to launch well and then after when we launch secondary
processes, getting a crash on eal-intr-thread.


here are failure logs before crash:

 EAL: Fail to recv reply for request
/var/run/dpdk/oracusbc/mp_socket:eal_dev_mp_request

 EAL: Cannot send request to primary

 EAL: Failed to send hotplug request to primary

 net_failsafe: Failed to probe devargs net_tap_vsc0{*}
===

With DPDK-21.11 version.

Regards,
Madhuker.

-- 
You are receiving this mail because:
You are the assignee for the bug.

[PATCH] flow_classify: mark library as deprecated

2022-10-27 Thread David Marchand
This library has no maintainer and, for now, nobody expressed interest
in taking over.
Mark this experimental library as deprecated and announce plan for
removal in v23.11.

Signed-off-by: David Marchand 
---
 MAINTAINERS | 3 +--
 doc/guides/prog_guide/flow_classify_lib.rst | 9 +
 doc/guides/rel_notes/deprecation.rst| 6 ++
 lib/flow_classify/rte_flow_classify.c   | 3 +++
 lib/meson.build | 5 -
 meson_options.txt   | 2 +-
 6 files changed, 24 insertions(+), 4 deletions(-)

diff --git a/MAINTAINERS b/MAINTAINERS
index 2bd4a55f1b..89768d15ee 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -1448,8 +1448,7 @@ F: doc/guides/prog_guide/ipsec_lib.rst
 M: Vladimir Medvedkin 
 F: app/test-sad/
 
-Flow Classify - EXPERIMENTAL
-M: Bernard Iremonger 
+Flow Classify - EXPERIMENTAL - UNMAINTAINED
 F: lib/flow_classify/
 F: app/test/test_flow_classify*
 F: doc/guides/prog_guide/flow_classify_lib.rst
diff --git a/doc/guides/prog_guide/flow_classify_lib.rst 
b/doc/guides/prog_guide/flow_classify_lib.rst
index 7dae0bc8c6..ad2e10ec26 100644
--- a/doc/guides/prog_guide/flow_classify_lib.rst
+++ b/doc/guides/prog_guide/flow_classify_lib.rst
@@ -4,6 +4,15 @@
 Flow Classification Library
 ===
 
+.. note::
+
+   The Flow Classification library is deprecated and will be removed in future.
+   See :doc:`../rel_notes/deprecation`.
+
+   It is disabled by default in the DPDK build.
+   To re-enable the library, remove 'flow_classify' from the "disable_libs"
+   meson option when configuring a build.
+
 DPDK provides a Flow Classification library that provides the ability
 to classify an input packet by matching it against a set of Flow rules.
 
diff --git a/doc/guides/rel_notes/deprecation.rst 
b/doc/guides/rel_notes/deprecation.rst
index 05cacb3ea8..e2efa2f8b0 100644
--- a/doc/guides/rel_notes/deprecation.rst
+++ b/doc/guides/rel_notes/deprecation.rst
@@ -96,3 +96,9 @@ Deprecation Notices
   to have another parameter ``qp_id`` to return the queue pair ID
   which got error interrupt to the application,
   so that application can reset that particular queue pair.
+
+* flow_classify: The flow_classify library and example have no maintainer.
+  The library is experimental and, as such, it could be removed from DPDK.
+  Its removal has been postponed to let potential users report interest
+  in maintaining it.
+  In the absence of such interest, this library will be removed in DPDK 23.11.
diff --git a/lib/flow_classify/rte_flow_classify.c 
b/lib/flow_classify/rte_flow_classify.c
index e3667306e5..60ca319f24 100644
--- a/lib/flow_classify/rte_flow_classify.c
+++ b/lib/flow_classify/rte_flow_classify.c
@@ -259,6 +259,9 @@ rte_flow_classifier_create(struct 
rte_flow_classifier_params *params)
struct rte_flow_classifier *cls;
int ret;
 
+   RTE_FLOW_CLASSIFY_LOG(WARNING,
+   "WARNING: flow_classify is deprecated and will be removed in 
DPDK 23.11\n");
+
/* Check input parameters */
ret = rte_flow_classifier_check_params(params);
if (ret != 0) {
diff --git a/lib/meson.build b/lib/meson.build
index c51cdc24fa..1f802bc456 100644
--- a/lib/meson.build
+++ b/lib/meson.build
@@ -85,7 +85,10 @@ optional_libs = [
 'vhost',
 ]
 
-dpdk_libs_deprecated += ['kni']
+dpdk_libs_deprecated += [
+'flow_classify',
+'kni',
+]
 
 disabled_libs = []
 opt_disabled_libs = run_command(list_dir_globs, get_option('disable_libs'),
diff --git a/meson_options.txt b/meson_options.txt
index 5de1b3ed23..ce26a8a2ec 100644
--- a/meson_options.txt
+++ b/meson_options.txt
@@ -8,7 +8,7 @@ option('developer_mode', type: 'feature', description:
'turn on additional build checks relevant for DPDK developers')
 option('disable_drivers', type: 'string', value: '', description:
'Comma-separated list of drivers to explicitly disable.')
-option('disable_libs', type: 'string', value: 'kni', description:
+option('disable_libs', type: 'string', value: 'flow_classify,kni', description:
'Comma-separated list of libraries to explicitly disable. [NOTE: not 
all libs can be disabled]')
 option('drivers_install_subdir', type: 'string', value: 'dpdk/pmds-', 
description:
'Subdirectory of libdir where to install PMDs. Defaults to using a 
versioned subdirectory.')
-- 
2.37.3



RE: [EXT] [PATCH v2 0/4] crypto/qat: extend asymmetric crypto pmd

2022-10-27 Thread Akhil Goyal
> This patchset extends Intel QuickAssist Technology asymmetric crypto PMD.
> Following features were added:
> - ECDH algorithm handling
> - EC point verification
> 
> Depends-on: series-25284 ("crypto/qat: fix uncleared cookies in asym")
> 
Series applied to dpdk-next-crypto

Thanks.


RE: [dpdk-dev v1] crypto/qat: fix of qat build request session in mp

2022-10-27 Thread Akhil Goyal
> > Subject: [dpdk-dev v1] crypto/qat: fix of qat build request session in mp
> >
> > This patch fix the session pointer passed in set_session() when ctx has NULL
> > build request pointer in multi-processes scenario.
> >
> > Fixes: fb3b9f492205 ("crypto/qat: rework burst data path")
> > Cc: sta...@dpdk.org
> >
> > Signed-off-by: Kai Ji 
> > ---
> 
> Acked-by: Ciara Power 

Applied to dpdk-next-crypto
Thanks.


RE: [EXT] [dpdk-dev v5] crypto/ipsec_mb: multi-process IPC request handler

2022-10-27 Thread Akhil Goyal
> As the queue pair used in secondary process needs to be set up by
> the primary process, this patch adds an IPC register function to help
> secondary process to send out queue-pair setup request to primary
> process via IPC request messages. A new "qp_in_used_pid" param stores
> the PID to provide the ownership of the queue-pair so that only the PID
> matched queue-pair can be free'd in the request.
> 
> Signed-off-by: Kai Ji 
> Acked-by: Arek Kusztal 
> Acked-by: Pablo de Lara 
> ---
> v5:
> - minor updates and typo fix
Applied to dpdk-next-crypto

Thanks.


RE: [EXT] [PATCH 1/2] app/test-crypto-perf: fix number of scheduler sessions

2022-10-27 Thread Akhil Goyal
> Subject: RE: [EXT] [PATCH 1/2] app/test-crypto-perf: fix number of scheduler
> sessions
> 
> > The scheduler PMD needs 1 session header,
> > along with a session per worker.
> >
> > After the session rework, this change was made to other apps,
> > for example l2fwd-crypto, but was missed in test-crypto-perf.
> >
> > Fixes: bdce2564dbf7 ("cryptodev: rework session framework")
> > Cc: gak...@marvell.com
> >
> > Signed-off-by: Ciara Power 
> Acked-by: Akhil Goyal 

Series Applied to dpdk-next-crypto

Thanks.


RE: [PATCH v2] app/test: add support to run multi-seg inline IPsec test

2022-10-27 Thread Akhil Goyal
> Subject: [PATCH v2] app/test: add support to run multi-seg inline IPsec test
> 
> Add support to run multi-seg inline ipsec test using
> new test command `inline_ipsec_sg_autotest`
> 
> Signed-off-by: Nithin Dabilpuram 
> ---
Acked-by: Akhil Goyal 

Applied to dpdk-next-crypto

Thanks.


RE: [EXT] [PATCH v1] examples/fips_validation: parse block error fix

2022-10-27 Thread Akhil Goyal
> Acked-by: Gowrishankar Muthukrishnan 
> 
> > When parsing request files check for file type. This fix will remove
> > dependence on command line parameter for using libjansson
> >
> > Fixes: 0f42f3d6034c ("examples/fips_validation: share callback with multiple
> > keys")
> > Cc: gmuthukri...@marvell.com
> > Signed-off-by: Brian Dooley 
> > ---
Applied to dpdk-next-crypto

Thanks.


Re: [PATCH 2/2] license: add MIT license exception for gve driver

2022-10-27 Thread Thomas Monjalon
19/10/2022 20:42, Stephen Hemminger:
> Tech Board and Governing Board approved license a exception
> for new Google driver. More general approval for MIT
> usage will be handled as needed.
> 
> Signed-off-by: Stephen Hemminger 
> ---
> +4. MIT   10/19/202210/18/2022
> drivers/net/gve/base/*

Series applied, thanks.





Re: [PATCH] flow_classify: mark library as deprecated

2022-10-27 Thread Ferruh Yigit

On 10/27/2022 10:46 AM, David Marchand wrote:

This library has no maintainer and, for now, nobody expressed interest
in taking over.
Mark this experimental library as deprecated and announce plan for
removal in v23.11.

Signed-off-by: David Marchand


Acked-by: Ferruh Yigit 


[PATCH v1] crypto/qat: fix null hash algorithm digest size

2022-10-27 Thread Brian Dooley
Add check for null hash algorithm digest size. Digest size should be 4B
or request will be rejected.

Fixes: 1703e94ac5ce ("qat: add driver for QuickAssist devices")
Cc: declan.dohe...@intel.com
Cc: sta...@dpdk.org
Signed-off-by: Brian Dooley 
---
 drivers/crypto/qat/qat_sym_session.c | 19 +++
 1 file changed, 15 insertions(+), 4 deletions(-)

diff --git a/drivers/crypto/qat/qat_sym_session.c 
b/drivers/crypto/qat/qat_sym_session.c
index d96122b208..ff4b537706 100644
--- a/drivers/crypto/qat/qat_sym_session.c
+++ b/drivers/crypto/qat/qat_sym_session.c
@@ -2040,7 +2040,12 @@ int qat_sym_cd_auth_set(struct qat_sym_session *cdesc,
hash_offset = cdesc->cd_cur_ptr-((uint8_t *)&cdesc->cd);
hash = (struct icp_qat_hw_auth_setup *)cdesc->cd_cur_ptr;
hash->auth_config.reserved = 0;
-   hash->auth_config.config =
+   if (cdesc->qat_hash_alg == ICP_QAT_HW_AUTH_ALGO_NULL)
+   hash->auth_config.config =
+   ICP_QAT_HW_AUTH_CONFIG_BUILD(cdesc->auth_mode,
+   cdesc->qat_hash_alg, 4);
+   else
+   hash->auth_config.config =
ICP_QAT_HW_AUTH_CONFIG_BUILD(cdesc->auth_mode,
cdesc->qat_hash_alg, digestsize);
 
@@ -2408,10 +2413,16 @@ int qat_sym_cd_auth_set(struct qat_sym_session *cdesc,
/* Auth CD config setup */
hash_cd_ctrl->hash_cfg_offset = hash_offset >> 3;
hash_cd_ctrl->hash_flags = ICP_QAT_FW_AUTH_HDR_FLAG_NO_NESTED;
-   hash_cd_ctrl->inner_res_sz = digestsize;
-   hash_cd_ctrl->final_sz = digestsize;
hash_cd_ctrl->inner_state1_sz = state1_size;
-   auth_param->auth_res_sz = digestsize;
+   if (cdesc->qat_hash_alg == ICP_QAT_HW_AUTH_ALGO_NULL) {
+   hash_cd_ctrl->inner_res_sz = 4;
+   hash_cd_ctrl->final_sz = 4;
+   auth_param->auth_res_sz = 4;
+   } else {
+   hash_cd_ctrl->inner_res_sz = digestsize;
+   hash_cd_ctrl->final_sz = digestsize;
+   auth_param->auth_res_sz = digestsize;
+   }
 
hash_cd_ctrl->inner_state2_sz  = state2_size;
hash_cd_ctrl->inner_state2_offset = hash_cd_ctrl->hash_cfg_offset +
-- 
2.25.1



RE: [EXT] [PATCH v7 0/6] crypto/uadk: introduce uadk crypto driver

2022-10-27 Thread Akhil Goyal
> Introduce a new crypto PMD for hardware accelerators based on UADK [1].
> 
> UADK is a framework for user applications to access hardware accelerators.
> UADK relies on IOMMU SVA (Shared Virtual Address) feature, which share
> the same page table between IOMMU and MMU.
> Thereby user application can directly use virtual address for device dma,
> which enhances the performance as well as easy usability.
> 
> [1] https://urldefense.proofpoint.com/v2/url?u=https-
> 3A__github.com_Linaro_uadk&d=DwIDAg&c=nKjWec2b6R0mOyPaz7xtfQ&r=Dn
> L7Si2wl_PRwpZ9TWey3eu68gBzn7DkPwuqhd6WNyo&m=AwapMJzvfUDL49W9f
> mtmvIGZMarGUdjRSliBhG4tep6Uh5wN2zYrZRJ4JpbrVolg&s=MxSabEjMMkeRr0L
> Z7c1gSmaJmJ1_dIdJIKrst98pKzk&e=
> 
> Test:
> sudo dpdk-test --vdev=crypto_uadk (--log-level=6)
> RTE>>cryptodev_uadk_autotest
> RTE>>quit
> 
> Update in v7:
> 05: fix key_size of SHA384 HMAC
> 
> Update in v6:
> Akhil helped review the v5, and address all the comments from Akhil
> 01: add rel_notes info, update uadk.rst, put struct in header, etc.
> 02: move struct to header
> 04: move struct to header, remove RTE_CRYPTODEV_FF_SYM_SESSIONLESS
> 05: fixed key_size in HMAC mode
> 06: updated app/test/meson.build and uadk.rst in the mean time.
> 
> Update in v5
> Patch 1 fix the build issue when uadk is installed to specific folder
> And update doc accordingly
> 
> Update in v4:
> Akril suggest dpdk use pkg-config, So
> Enable uadk support x86 local build, and support pkg-config.
> Use pkg-config feature for the uadk crypto pmd.
> Add build uadk library steps in doc
> Test on both x86 and arm.
> x86 can build and install, but can not test since no device.
> 
> Resend v3:
> Rebase on next/for-main, which just merged the series
> "cryptodev: rework session framework".
> 
> Update in v3:
> Split patches according to Akhil's suggestions
> Please split the patches as below.
> 1. introduce driver - create files with meson.build and with probe/remove
>and device ops defined but not implemented.
>You do not need to write empty functions.
>Add basic documentation also which defines what the driver is.
>You can explain the build dependency here.
> 2. define queue structs and setup/remove APIs
> 3. Add data path
> 4. implement cipher op. Add capabilities and documentation of what is
>supported in each of the patches. Add feature flags etc.
> 5. implement auth,  add capabilities and documentation
> 6. test app changes.
> 
> Update in v2:
> Change uadk_supported_platform to uadk_crypto_version, which matches
> better
> than platform.
> enum uadk_crypto_version {
>   UADK_CRYPTO_V2,
>   UADK_CRYPTO_V3,
> };
> 
> Update in v1, compared with rfc
> 
> Suggested from Akhil Goyal 
> Only consider crypto PMD first
> Split patch into small (individually compiled) patches.
> Update MAINTAINERS and doc/guides/cryptodevs/features/uadk.ini
> 
> Zhangfei Gao (6):
>   crypto/uadk: introduce uadk crypto driver
>   crypto/uadk: support basic operations
>   crypto/uadk: support enqueue/dequeue operations
>   crypto/uadk: support cipher algorithms
>   crypto/uadk: support auth algorithms
>   test/crypto: support uadk PMD
> 
>  MAINTAINERS   |6 +
>  app/test/meson.build  |1 +
>  app/test/test_cryptodev.c |7 +
>  app/test/test_cryptodev.h |1 +
>  doc/guides/cryptodevs/features/uadk.ini   |   55 +
>  doc/guides/cryptodevs/index.rst   |1 +
>  doc/guides/cryptodevs/uadk.rst|   96 ++
>  doc/guides/rel_notes/release_22_11.rst|6 +
>  drivers/crypto/meson.build|1 +
>  drivers/crypto/uadk/meson.build   |   30 +
>  drivers/crypto/uadk/uadk_crypto_pmd.c | 1081 +
>  drivers/crypto/uadk/uadk_crypto_pmd_private.h |   79 ++
>  drivers/crypto/uadk/version.map   |3 +
>  13 files changed, 1367 insertions(+)
>  create mode 100644 doc/guides/cryptodevs/features/uadk.ini
>  create mode 100644 doc/guides/cryptodevs/uadk.rst
>  create mode 100644 drivers/crypto/uadk/meson.build
>  create mode 100644 drivers/crypto/uadk/uadk_crypto_pmd.c
>  create mode 100644 drivers/crypto/uadk/uadk_crypto_pmd_private.h
>  create mode 100644 drivers/crypto/uadk/version.map

Series Applied to dpdk-next-crypto
Welcome to dpdk.
Thanks.


RE: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq

2022-10-27 Thread Ye, MingjinX
Hi lihuisong,

This means that queue offloads need to update by recalling dev_configure
and setup target queues.
Can you tell me, where is the limitation?

Thanks,
Mingjin

> -Original Message-
> From: lihuisong (C) 
> Sent: 2022年10月26日 17:53
> To: Ye, MingjinX ; dev@dpdk.org
> Cc: sta...@dpdk.org; Zhou, YidingX ; Singh, Aman
> Deep ; Zhang, Yuying
> 
> Subject: Re: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq
> 
> 
> 在 2022/10/27 1:10, Mingjin Ye 写道:
> > After setting vlan offload in testpmd, the result is not updated to
> > rxq. Therefore, the queue needs to be reconfigured after executing the
> > "vlan offload" related commands.
> >
> > Fixes: a47aa8b97afe ("app/testpmd: add vlan offload support")
> > Cc: sta...@dpdk.org
> >
> > Signed-off-by: Mingjin Ye 
> > ---
> >   app/test-pmd/cmdline.c | 1 +
> >   1 file changed, 1 insertion(+)
> >
> > diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c index
> > 17be2de402..ce125f549f 100644
> > --- a/app/test-pmd/cmdline.c
> > +++ b/app/test-pmd/cmdline.c
> > @@ -4133,6 +4133,7 @@ cmd_vlan_offload_parsed(void *parsed_result,
> > else
> > vlan_extend_set(port_id, on);
> >
> > +   cmd_reconfig_device_queue(port_id, 1, 1);
> This means that queue offloads need to upadte by re-calling dev_configure
> and setup all queues, right?
> If it is, this adds a usage limitation.
> > return;
> >   }
> >


[PATCH v2] examples/ipsec-secgw: support per SA HW reassembly

2022-10-27 Thread Rahul Bhansali
This add the support of hardware reassembly per SA basis.
In SA rule, new parameter reassembly_en is added to enable
HW reassembly per SA.
For example:
sa in  aead_algo  aead_key  mode ipv4-tunnel src 
dst  type inline-protocol-offload port_id  reassembly_en

Stats counter frag_dropped will represent the number of fragment
drop in case of reassembly failures.

Signed-off-by: Rahul Bhansali 
---
Changes in v2:
- updated patch title, and moved extern of ip_reassembly_dynflag next to
  ip_reassembly_dynfield_offset.

 doc/guides/sample_app_ug/ipsec_secgw.rst | 12 ++-
 examples/ipsec-secgw/ipsec-secgw.c   | 30 +
 examples/ipsec-secgw/ipsec-secgw.h   | 43 +++-
 examples/ipsec-secgw/ipsec.c |  2 ++
 examples/ipsec-secgw/ipsec.h |  4 ++-
 examples/ipsec-secgw/ipsec_worker.c  | 33 +-
 examples/ipsec-secgw/ipsec_worker.h  |  5 +++
 examples/ipsec-secgw/sa.c| 12 ++-
 8 files changed, 129 insertions(+), 12 deletions(-)

diff --git a/doc/guides/sample_app_ug/ipsec_secgw.rst 
b/doc/guides/sample_app_ug/ipsec_secgw.rst
index 2a1aeae7c5..4d889e8496 100644
--- a/doc/guides/sample_app_ug/ipsec_secgw.rst
+++ b/doc/guides/sample_app_ug/ipsec_secgw.rst
@@ -537,7 +537,7 @@ The SA rule syntax is shown as follows:

 sa  
  
-   
+

 where each options means:

@@ -793,6 +793,16 @@ where each options means:

* *esn N* N is the initial ESN value

+ 
+
+ * Option to enable HW reassembly per SA.
+
+ * Optional: Yes, it is disabled by default
+
+ * Syntax:
+
+   * *reassembly_en*
+
 Example SA rules:

 .. code-block:: console
diff --git a/examples/ipsec-secgw/ipsec-secgw.c 
b/examples/ipsec-secgw/ipsec-secgw.c
index 1d9982a786..24d895451a 100644
--- a/examples/ipsec-secgw/ipsec-secgw.c
+++ b/examples/ipsec-secgw/ipsec-secgw.c
@@ -274,6 +274,7 @@ static void
 print_stats_cb(__rte_unused void *param)
 {
uint64_t total_packets_dropped, total_packets_tx, total_packets_rx;
+   uint64_t total_frag_packets_dropped = 0;
float burst_percent, rx_per_call, tx_per_call;
unsigned int coreid;

@@ -303,6 +304,7 @@ print_stats_cb(__rte_unused void *param)
   "\nPackets received: %20"PRIu64
   "\nPackets sent: %24"PRIu64
   "\nPackets dropped: %21"PRIu64
+  "\nFrag Packets dropped: %16"PRIu64
   "\nBurst percent: %23.2f"
   "\nPackets per Rx call: %17.2f"
   "\nPackets per Tx call: %17.2f",
@@ -310,21 +312,25 @@ print_stats_cb(__rte_unused void *param)
   core_statistics[coreid].rx,
   core_statistics[coreid].tx,
   core_statistics[coreid].dropped,
+  core_statistics[coreid].frag_dropped,
   burst_percent,
   rx_per_call,
   tx_per_call);

total_packets_dropped += core_statistics[coreid].dropped;
+   total_frag_packets_dropped += 
core_statistics[coreid].frag_dropped;
total_packets_tx += core_statistics[coreid].tx;
total_packets_rx += core_statistics[coreid].rx;
}
printf("\nAggregate statistics ==="
   "\nTotal packets received: %14"PRIu64
   "\nTotal packets sent: %18"PRIu64
-  "\nTotal packets dropped: %15"PRIu64,
+  "\nTotal packets dropped: %15"PRIu64
+  "\nTotal frag packets dropped: %10"PRIu64,
   total_packets_rx,
   total_packets_tx,
-  total_packets_dropped);
+  total_packets_dropped,
+  total_frag_packets_dropped);
printf("\n\n");

rte_eal_alarm_set(stats_interval * US_PER_S, print_stats_cb, NULL);
@@ -1867,7 +1873,8 @@ parse_ptype_cb(uint16_t port __rte_unused, uint16_t queue 
__rte_unused,
 }

 static void
-port_init(uint16_t portid, uint64_t req_rx_offloads, uint64_t req_tx_offloads)
+port_init(uint16_t portid, uint64_t req_rx_offloads, uint64_t req_tx_offloads,
+ uint8_t hw_reassembly)
 {
struct rte_eth_dev_info dev_info;
struct rte_eth_txconf *txconf;
@@ -1877,6 +1884,7 @@ port_init(uint16_t portid, uint64_t req_rx_offloads, 
uint64_t req_tx_offloads)
struct lcore_conf *qconf;
struct rte_ether_addr ethaddr;
struct rte_eth_conf local_port_conf = port_conf;
+   struct rte_eth_ip_reassembly_params reass_capa = {0};
int ptype_supported;

ret = rte_eth_dev_info_get(portid, &dev_info);
@@ -2053,6 +2061,12 @@ port_init(uint16_t portid, uint64_t req_rx_offloads, 
uint64_t req_tx_offloads)

}
}
+
+   if (hw_reassembly) 

RE: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq

2022-10-27 Thread Ye, MingjinX
Hi All,

Could you please review and provide suggestions if any.

Thanks,
Mingjin

> -Original Message-
> From: Ye, MingjinX 
> Sent: 2022年10月25日 1:27
> To: dev@dpdk.org
> Cc: sta...@dpdk.org; Zhou, YidingX ; Ye, MingjinX
> ; Singh, Aman Deep
> ; Zhang, Yuying 
> Subject: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq
> 
> After setting vlan offload in testpmd, the result is not updated to rxq.
> Therefore, the queue needs to be reconfigured after executing the "vlan
> offload" related commands.
> 
> Fixes: a47aa8b97afe ("app/testpmd: add vlan offload support")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Mingjin Ye 
> ---
>  app/test-pmd/cmdline.c | 1 +
>  1 file changed, 1 insertion(+)
> 
> diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c index
> 17be2de402..ce125f549f 100644
> --- a/app/test-pmd/cmdline.c
> +++ b/app/test-pmd/cmdline.c
> @@ -4133,6 +4133,7 @@ cmd_vlan_offload_parsed(void *parsed_result,
>   else
>   vlan_extend_set(port_id, on);
> 
> + cmd_reconfig_device_queue(port_id, 1, 1);
>   return;
>  }
> 
> --
> 2.34.1



Re: [PATCH 0/3] remove more experimental tags

2022-10-27 Thread David Marchand
On Tue, Oct 18, 2022 at 3:58 AM Stephen Hemminger
 wrote:
>
> Remove some of the older experimental tags for 22.11
>
> Stephen Hemminger (3):
>   eal: remove experimental from  rte_epoll_wait_interruptible
>   eal: make rte_log_list_types not experimental
>   rwlock: make trylock operations no longer experimental
>
>  lib/eal/include/generic/rte_rwlock.h | 12 ++--
>  lib/eal/include/rte_epoll.h  |  1 -
>  lib/eal/include/rte_log.h|  4 
>  lib/eal/version.map  |  4 ++--
>  4 files changed, 4 insertions(+), 17 deletions(-)

Series applied.

I added to my todolist that we need a cleanup for unneeded inclusion
of rte_compat.h (the headers touched in this series are not the only
one concerned).


-- 
David Marchand



Re: [PATCH] ci: combine static and shared linking build tests

2022-10-27 Thread David Marchand
On Thu, Oct 20, 2022 at 5:34 PM Aaron Conole  wrote:
> > Save some cpu time and disk by testing linking against static and shared
> > library in single environments.
> >
> > The .ci/linux-build.sh is modified so it reconfigures an existing build
> > directory: an empty DEF_LIB= means that static and shared builds are
> > to be tested.
> >
> > ABI checks, documentation generation and unit tests are disabled for
> > static builds as they would be redundant with the check against
> > dynamically linked binaries, if any.
> >
> > Note:
> > - --cross-file is an option that can be passed to meson only when
> >   creating a build environment,
> > - for some other reason, --buildtype and other non -D options are only
> >   accepted when setting up a build directory with meson. When
> >   reconfiguring, only their -D$option forms are accepted,
> >
> > Signed-off-by: David Marchand 
> Acked-by: Bruce Richardson 
> Acked-by: Aaron Conole 

Applied, thanks.


-- 
David Marchand



Re: [PATCH] mempool: cache align mempool cache objects

2022-10-27 Thread Olivier Matz
On Thu, Oct 27, 2022 at 11:22:07AM +0200, Morten Brørup wrote:
> > From: Olivier Matz [mailto:olivier.m...@6wind.com]
> > Sent: Thursday, 27 October 2022 10.35
> > 
> > Hi Morten,
> > 
> > On Wed, Oct 26, 2022 at 04:44:36PM +0200, Morten Brørup wrote:
> > > Add __rte_cache_aligned to the objs array.
> > >
> > > It makes no difference in the general case, but if get/put operations
> > are
> > > always 32 objects, it will reduce the number of memory (or last level
> > > cache) accesses from five to four 64 B cache lines for every get/put
> > > operation.
> > >
> > > For readability reasons, an example using 16 objects follows:
> > >
> > > Currently, with 16 objects (128B), we access to 3
> > > cache lines:
> > >
> > >   ┌┐
> > >   │len │
> > > cache ││---
> > > line0 ││ ^
> > >   ││ |
> > >   ├┤ | 16 objects
> > >   ││ | 128B
> > > cache ││ |
> > > line1 ││ |
> > >   ││ |
> > >   ├┤ |
> > >   ││_v_
> > > cache ││
> > > line2 ││
> > >   ││
> > >   └┘
> > >
> > > With the alignment, it is also 3 cache lines:
> > >
> > >   ┌┐
> > >   │len │
> > > cache ││
> > > line0 ││
> > >   ││
> > >   ├┤---
> > >   ││ ^
> > > cache ││ |
> > > line1 ││ |
> > >   ││ |
> > >   ├┤ | 16 objects
> > >   ││ | 128B
> > > cache ││ |
> > > line2 ││ |
> > >   ││ v
> > >   └┘---
> > >
> > > However, accessing the objects at the bottom of the mempool cache is
> > a
> > > special case, where cache line0 is also used for objects.
> > >
> > > Consider the next burst (and any following bursts):
> > >
> > > Current:
> > >   ┌┐
> > >   │len │
> > > cache ││
> > > line0 ││
> > >   ││
> > >   ├┤
> > >   ││
> > > cache ││
> > > line1 ││
> > >   ││
> > >   ├┤
> > >   ││
> > > cache ││---
> > > line2 ││ ^
> > >   ││ |
> > >   ├┤ | 16 objects
> > >   ││ | 128B
> > > cache ││ |
> > > line3 ││ |
> > >   ││ |
> > >   ├┤ |
> > >   ││_v_
> > > cache ││
> > > line4 ││
> > >   ││
> > >   └┘
> > > 4 cache lines touched, incl. line0 for len.
> > >
> > > With the proposed alignment:
> > >   ┌┐
> > >   │len │
> > > cache ││
> > > line0 ││
> > >   ││
> > >   ├┤
> > >   ││
> > > cache ││
> > > line1 ││
> > >   ││
> > >   ├┤
> > >   ││
> > > cache ││
> > > line2 ││
> > >   ││
> > >   ├┤
> > >   ││---
> > > cache ││ ^
> > > line3 ││ |
> > >   ││ | 16 objects
> > >   ├┤ | 128B
> > >   ││ |
> > > cache ││ |
> > > line4 ││ |
> > >   ││_v_
> > >   └┘
> > > Only 3 cache lines touched, incl. line0 for len.
> > 
> > I understand your logic, but are we sure that having an application
> > that
> > works with bulks of 32 means that the cache will stay aligned to 32
> > elements for the whole life of the application?
> > 
> > In an application, the alignment of the cache can change if you have
> > any of:
> > - software queues (reassembly for instance)
> > - packet duplication (bridge, multicast)
> > - locally generated packets (keepalive, control protocol)
> > - pipeline to other cores
> > 
> > Even with testpmd, which work by bulk of 32, I can see that the size
> > of the cache filling is not aligned to 32. Right after starting the
> > application, we already have this:
> > 
> >   internal cache infos:
> > cache_size=250
> > cache_count[0]=231
> > 
> > This is probably related to the hw rx rings size, number of queues,
> > number of ports.
> > 
> > The "250" default value for cache size in testpmd is questionable, but
> > with --mbcache=256, the behavior is similar.
> > 
> > Also, when we transmit to a NIC, the mbufs are not returned immediatly
> > to the pool, they may stay in the hw tx ring during some time, which is
> > a driver decision.
> > 
> > After processing traffic on cores 8 and 24 with this testpmd, I get:
> > cache_count[0]=231
> > cache_count[8]=123
> > cache_count[24]=122
> > 
> > In my opinion, it is not realistic to think that the mempool cache will
> > remain aligned to cachelines. In these conditions, it looks better to
> > keep the structure packed to avoid wasting memory.
> 
> I agree that is a special use case to only access the mempool cache in
> bursts of 32 objects, so the accesses are always cache line
> aligned. (Generalized, the burst size must not be 32; a burst size
> that is

Re: [PATCH] telemetry: make usable from C++

2022-10-27 Thread David Marchand
On Wed, Oct 12, 2022 at 2:05 PM Markus Theil  wrote:
>
> From: Markus Theil 
>
> Add missing include in order to make C++ compilers
> happy.
>

We have build checks for headers (-Dcheck_includes meson option), like for C++:

$ cat $BUILDDIR/buildtools/chkincs/chkincs-cpp.p/rte_telemetry.cpp
#include "/home/dmarchan/dpdk/lib/telemetry/rte_telemetry.h"

[917/1289] ccache c++ -Ibuildtools/chkincs/chkincs-cpp.p
-Ibuildtools/chkincs -I../../../dpdk/buildtools/chkincs
-Iexamples/vmdq_dcb -I../../../dpdk/examples/vmdq_dcb
-I../../../dpdk/examples/common -Idrivers/bus/vdev
-I../../../dpdk/drivers/bus/vdev -I. -I../../../dpdk -Iconfig
-I../../../dpdk/config -Ilib/eal/include
-I../../../dpdk/lib/eal/include -Ilib/eal/linux/include
-I../../../dpdk/lib/eal/linux/include -Ilib/eal/x86/include
-I../../../dpdk/lib/eal/x86/include -Ilib/eal/common
-I../../../dpdk/lib/eal/common -Ilib/eal -I../../../dpdk/lib/eal
-Ilib/kvargs -I../../../dpdk/lib/kvargs -Ilib/metrics
-I../../../dpdk/lib/metrics -Ilib/telemetry
-I../../../dpdk/lib/telemetry -Idrivers/bus/pci
-I../../../dpdk/drivers/bus/pci -I../../../dpdk/drivers/bus/pci/linux
-Ilib/pci -I../../../dpdk/lib/pci -Ilib/ring -I../../../dpdk/lib/ring
-Ilib/rcu -I../../../dpdk/lib/rcu -Ilib/mempool
-I../../../dpdk/lib/mempool -Ilib/mbuf -I../../../dpdk/lib/mbuf
-Ilib/net -I../../../dpdk/lib/net -Ilib/meter
-I../../../dpdk/lib/meter -Ilib/ethdev -I../../../dpdk/lib/ethdev
-Ilib/cmdline -I../../../dpdk/lib/cmdline -Ilib/hash
-I../../../dpdk/lib/hash -Ilib/timer -I../../../dpdk/lib/timer
-Ilib/acl -I../../../dpdk/lib/acl -Ilib/bbdev
-I../../../dpdk/lib/bbdev -Ilib/bitratestats
-I../../../dpdk/lib/bitratestats -Ilib/bpf -I../../../dpdk/lib/bpf
-Ilib/cfgfile -I../../../dpdk/lib/cfgfile -Ilib/compressdev
-I../../../dpdk/lib/compressdev -Ilib/cryptodev
-I../../../dpdk/lib/cryptodev -Ilib/distributor
-I../../../dpdk/lib/distributor -Ilib/efd -I../../../dpdk/lib/efd
-Ilib/eventdev -I../../../dpdk/lib/eventdev -Ilib/gpudev
-I../../../dpdk/lib/gpudev -Ilib/gro -I../../../dpdk/lib/gro -Ilib/gso
-I../../../dpdk/lib/gso -Ilib/ip_frag -I../../../dpdk/lib/ip_frag
-Ilib/jobstats -I../../../dpdk/lib/jobstats -Ilib/kni
-I../../../dpdk/lib/kni -Ilib/latencystats
-I../../../dpdk/lib/latencystats -Ilib/lpm -I../../../dpdk/lib/lpm
-Ilib/member -I../../../dpdk/lib/member -Ilib/pcapng
-I../../../dpdk/lib/pcapng -Ilib/power -I../../../dpdk/lib/power
-Ilib/rawdev -I../../../dpdk/lib/rawdev -Ilib/regexdev
-I../../../dpdk/lib/regexdev -Ilib/dmadev -I../../../dpdk/lib/dmadev
-Ilib/rib -I../../../dpdk/lib/rib -Ilib/reorder
-I../../../dpdk/lib/reorder -Ilib/sched -I../../../dpdk/lib/sched
-Ilib/security -I../../../dpdk/lib/security -Ilib/stack
-I../../../dpdk/lib/stack -Ilib/vhost -I../../../dpdk/lib/vhost
-Ilib/ipsec -I../../../dpdk/lib/ipsec -Ilib/fib
-I../../../dpdk/lib/fib -Ilib/port -I../../../dpdk/lib/port
-Ilib/pdump -I../../../dpdk/lib/pdump -Ilib/table
-I../../../dpdk/lib/table -Ilib/pipeline -I../../../dpdk/lib/pipeline
-Ilib/flow_classify -I../../../dpdk/lib/flow_classify -Ilib/graph
-I../../../dpdk/lib/graph -Ilib/node -I../../../dpdk/lib/node
-fdiagnostics-color=always -D_FILE_OFFSET_BITS=64 -Wall -Winvalid-pch
-Wnon-virtual-dtor -Wextra -Werror -O2 -g -include rte_config.h
-march=nehalem -DALLOW_EXPERIMENTAL_API -DALLOW_INTERNAL_API -MD -MQ
buildtools/chkincs/chkincs-cpp.p/meson-generated_rte_telemetry.cpp.o
-MF buildtools/chkincs/chkincs-cpp.p/meson-generated_rte_telemetry.cpp.o.d
-o buildtools/chkincs/chkincs-cpp.p/meson-generated_rte_telemetry.cpp.o
-c buildtools/chkincs/chkincs-cpp.p/rte_telemetry.cpp

Besides, I fail to see the need for stddef.h.

In which setup / case, did you get a compilation issue?


-- 
David Marchand



RE: [PATCH] mempool: cache align mempool cache objects

2022-10-27 Thread Morten Brørup
> From: Olivier Matz [mailto:olivier.m...@6wind.com]
> Sent: Thursday, 27 October 2022 13.43
> 
> On Thu, Oct 27, 2022 at 11:22:07AM +0200, Morten Brørup wrote:
> > > From: Olivier Matz [mailto:olivier.m...@6wind.com]
> > > Sent: Thursday, 27 October 2022 10.35
> > >
> > > Hi Morten,
> > >
> > > On Wed, Oct 26, 2022 at 04:44:36PM +0200, Morten Brørup wrote:
> > > > Add __rte_cache_aligned to the objs array.
> > > >
> > > > It makes no difference in the general case, but if get/put
> operations
> > > are
> > > > always 32 objects, it will reduce the number of memory (or last
> level
> > > > cache) accesses from five to four 64 B cache lines for every
> get/put
> > > > operation.
> > > >
> > > > For readability reasons, an example using 16 objects follows:
> > > >
> > > > Currently, with 16 objects (128B), we access to 3
> > > > cache lines:
> > > >
> > > >   ┌┐
> > > >   │len │
> > > > cache ││---
> > > > line0 ││ ^
> > > >   ││ |
> > > >   ├┤ | 16 objects
> > > >   ││ | 128B
> > > > cache ││ |
> > > > line1 ││ |
> > > >   ││ |
> > > >   ├┤ |
> > > >   ││_v_
> > > > cache ││
> > > > line2 ││
> > > >   ││
> > > >   └┘
> > > >
> > > > With the alignment, it is also 3 cache lines:
> > > >
> > > >   ┌┐
> > > >   │len │
> > > > cache ││
> > > > line0 ││
> > > >   ││
> > > >   ├┤---
> > > >   ││ ^
> > > > cache ││ |
> > > > line1 ││ |
> > > >   ││ |
> > > >   ├┤ | 16 objects
> > > >   ││ | 128B
> > > > cache ││ |
> > > > line2 ││ |
> > > >   ││ v
> > > >   └┘---
> > > >
> > > > However, accessing the objects at the bottom of the mempool cache
> is
> > > a
> > > > special case, where cache line0 is also used for objects.
> > > >
> > > > Consider the next burst (and any following bursts):
> > > >
> > > > Current:
> > > >   ┌┐
> > > >   │len │
> > > > cache ││
> > > > line0 ││
> > > >   ││
> > > >   ├┤
> > > >   ││
> > > > cache ││
> > > > line1 ││
> > > >   ││
> > > >   ├┤
> > > >   ││
> > > > cache ││---
> > > > line2 ││ ^
> > > >   ││ |
> > > >   ├┤ | 16 objects
> > > >   ││ | 128B
> > > > cache ││ |
> > > > line3 ││ |
> > > >   ││ |
> > > >   ├┤ |
> > > >   ││_v_
> > > > cache ││
> > > > line4 ││
> > > >   ││
> > > >   └┘
> > > > 4 cache lines touched, incl. line0 for len.
> > > >
> > > > With the proposed alignment:
> > > >   ┌┐
> > > >   │len │
> > > > cache ││
> > > > line0 ││
> > > >   ││
> > > >   ├┤
> > > >   ││
> > > > cache ││
> > > > line1 ││
> > > >   ││
> > > >   ├┤
> > > >   ││
> > > > cache ││
> > > > line2 ││
> > > >   ││
> > > >   ├┤
> > > >   ││---
> > > > cache ││ ^
> > > > line3 ││ |
> > > >   ││ | 16 objects
> > > >   ├┤ | 128B
> > > >   ││ |
> > > > cache ││ |
> > > > line4 ││ |
> > > >   ││_v_
> > > >   └┘
> > > > Only 3 cache lines touched, incl. line0 for len.
> > >
> > > I understand your logic, but are we sure that having an application
> > > that
> > > works with bulks of 32 means that the cache will stay aligned to 32
> > > elements for the whole life of the application?
> > >
> > > In an application, the alignment of the cache can change if you
> have
> > > any of:
> > > - software queues (reassembly for instance)
> > > - packet duplication (bridge, multicast)
> > > - locally generated packets (keepalive, control protocol)
> > > - pipeline to other cores
> > >
> > > Even with testpmd, which work by bulk of 32, I can see that the
> size
> > > of the cache filling is not aligned to 32. Right after starting the
> > > application, we already have this:
> > >
> > >   internal cache infos:
> > > cache_size=250
> > > cache_count[0]=231
> > >
> > > This is probably related to the hw rx rings size, number of queues,
> > > number of ports.
> > >
> > > The "250" default value for cache size in testpmd is questionable,
> but
> > > with --mbcache=256, the behavior is similar.
> > >
> > > Also, when we transmit to a NIC, the mbufs are not returned
> immediatly
> > > to the pool, they may stay in the hw tx ring during some time,
> which is
> > > a driver decision.
> > >
> > > After processing traffic on cores 8 and 24 with this testpmd, I
> get:
> > > cache_count[0]=231
> > > cache_count[8]=123
> > > cache_count[24]=122
> > >
> > > 

RE: [PATCH] flow_classify: mark library as deprecated

2022-10-27 Thread Konstantin Ananyev



> 
> This library has no maintainer and, for now, nobody expressed interest
> in taking over.
> Mark this experimental library as deprecated and announce plan for
> removal in v23.11.
> 
> Signed-off-by: David Marchand 
> ---

 Acked-by: Konstantin Ananyev 

> --
> 2.37.3



Re: [PATCH v3 1/1] app/testpmd: control passing Rx metadata to PMD

2022-10-27 Thread Thomas Monjalon
27/10/2022 09:34, Thomas Monjalon:
> 17/10/2022 10:32, Andrew Rybchenko:
> > On 10/6/22 21:35, Hanumanth Pothula wrote:
> > > Presently, Rx metadata is sent to PMD by default, leading
> > > to a performance drop as processing for the same in rx path
> > > takes extra cycles.
> > > 
> > > Hence, introducing command line argument, 'nic-to-pmd-rx-metadata'
> > > to control passing rx metadata to PMD. By default it’s disabled.
> > > 
> > > Signed-off-by: Hanumanth Pothula 
> > 
> > Acked-by: Andrew Rybchenko 
> > 
> > Applied to dpdk-next-net/main, thanks.
> 
> I'm not sure this patch is really acceptable.
> It is disabling Rx metadata by default just for benchmarking reason
> because your driver is doing some processing even if metadata is not required.
> 
> From a user perspective, if a command requesting metadata is entered,
> it won't work until we enable this new option on startup.
> It looks terrible.
> 
> Please tell me I misunderstood something.

While pulling, I see that the name is not compliant with others.
I think it should start with "enable-", use hyphens and be sorted.

I'll drop it from the pull for now, we can have it in -rc3.




RE: [PATCH v1] crypto/qat: fix null hash algorithm digest size

2022-10-27 Thread Power, Ciara



> -Original Message-
> From: Brian Dooley 
> Sent: Thursday 27 October 2022 11:50
> To: Ji, Kai 
> Cc: dev@dpdk.org; sta...@dpdk.org; gak...@marvell.com; Kusztal,
> ArkadiuszX ; Dooley, Brian
> ; Doherty, Declan 
> Subject: [PATCH v1] crypto/qat: fix null hash algorithm digest size
> 
> Add check for null hash algorithm digest size. Digest size should be 4B or
> request will be rejected.
> 
> Fixes: 1703e94ac5ce ("qat: add driver for QuickAssist devices")
> Cc: declan.dohe...@intel.com
> Cc: sta...@dpdk.org
> Signed-off-by: Brian Dooley 
> ---

Acked-by: Ciara Power 


Re: [PATCH v4 1/2] build: allow to conditionally build apps

2022-10-27 Thread David Marchand
On Fri, Oct 14, 2022 at 9:51 AM Markus Theil  wrote:
>
> Makes apps configurable from meson, like already
> possible for drivers.
>
> Signed-off-by: Markus Theil 

With these new options, it's hard to tell which and why some
application is built (or not).
Can we have an output similar to libraries and drivers (see toplevel
meson.build) ?


-- 
David Marchand



Re: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq

2022-10-27 Thread Singh, Aman Deep




On 10/27/2022 2:06 PM, Huang, ZhiminX wrote:

-Original Message-
From: Mingjin Ye 
Sent: Thursday, October 27, 2022 1:10 AM
To: dev@dpdk.org
Cc: sta...@dpdk.org; Zhou, YidingX ; Ye, MingjinX
; Singh, Aman Deep ;
Zhang, Yuying 
Subject: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq

After setting vlan offload in testpmd, the result is not updated to rxq. 
Therefore,
the queue needs to be reconfigured after executing the "vlan offload" related
commands.

Fixes: a47aa8b97afe ("app/testpmd: add vlan offload support")
Cc: sta...@dpdk.org

Signed-off-by: Mingjin Ye 


Acked-by: Aman Singh


---

Tested-by: Zhimin Huang 






[PATCH v1] dts: fix ssh timeout and output

2022-10-27 Thread Juraj Linkeš
We were always waiting the full timeout because we always waited for
the MAGIC_PROMPT, which is only usable for cleaning the session of
previous data. That also meant that the actual prompt was appended to
the output, resulting in output being '[PEXPECT]#' when the executed
command produced empty output. Both are fixed by using the MAGIC_PROMPT
only in self._clean_session.

Fixes: e81a070a1c7d ("dts: add ssh session module")

Signed-off-by: Juraj Linkeš 
---
 dts/framework/remote_session/ssh_session.py | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/dts/framework/remote_session/ssh_session.py 
b/dts/framework/remote_session/ssh_session.py
index f71acfb1ca..dc37e9772b 100644
--- a/dts/framework/remote_session/ssh_session.py
+++ b/dts/framework/remote_session/ssh_session.py
@@ -116,7 +116,9 @@ def send_expect_base(self, command: str, prompt: str, 
timeout: float) -> str:
 return before
 
 def _clean_session(self) -> None:
+self.session.PROMPT = self.magic_prompt
 self.get_output(timeout=0.01)
+self.session.PROMPT = self.session.UNIQUE_PROMPT
 
 def _send_line(self, command: str) -> None:
 if not self.is_alive():
@@ -134,7 +136,6 @@ def get_output(self, timeout: float = 15) -> str:
 """
 Get all output before timeout
 """
-self.session.PROMPT = self.magic_prompt
 try:
 self.session.prompt(timeout)
 except Exception:
@@ -143,7 +144,6 @@ def get_output(self, timeout: float = 15) -> str:
 before = self._get_output()
 self._flush()
 
-self.logger.debug(before)
 return before
 
 def _get_output(self) -> str:
-- 
2.30.2



RE: [PATCH v1] crypto/qat: fix null hash algorithm digest size

2022-10-27 Thread Akhil Goyal
> > Add check for null hash algorithm digest size. Digest size should be 4B or
> > request will be rejected.
> >
> > Fixes: 1703e94ac5ce ("qat: add driver for QuickAssist devices")
> > Cc: declan.dohe...@intel.com
> > Cc: sta...@dpdk.org
> > Signed-off-by: Brian Dooley 
> > ---
> 
> Acked-by: Ciara Power 
Applied to dpdk-next-crypto

Thanks.


RE: [PATCH v2] examples/ipsec-secgw: support per SA HW reassembly

2022-10-27 Thread Akhil Goyal
> Subject: [PATCH v2] examples/ipsec-secgw: support per SA HW reassembly
> 
> This add the support of hardware reassembly per SA basis.
> In SA rule, new parameter reassembly_en is added to enable
> HW reassembly per SA.
> For example:
> sa in  aead_algo  aead_key  mode ipv4-tunnel src 
> dst  type inline-protocol-offload port_id  reassembly_en
> 
> Stats counter frag_dropped will represent the number of fragment
> drop in case of reassembly failures.
> 
> Signed-off-by: Rahul Bhansali 
> ---
Acked-by: Akhil Goyal 

Applied to dpdk-next-crypto

Thanks.


Re: [PATCH] flow_classify: mark library as deprecated

2022-10-27 Thread Thomas Monjalon
27/10/2022 11:46, David Marchand:
> This library has no maintainer and, for now, nobody expressed interest
> in taking over.
> Mark this experimental library as deprecated and announce plan for
> removal in v23.11.
> 
> Signed-off-by: David Marchand 

It adds enough warnings to reach potential users.

Acked-by: Thomas Monjalon 




Re: [PATCH v4 1/2] build: allow to conditionally build apps

2022-10-27 Thread David Marchand
On Thu, Oct 27, 2022 at 3:13 PM David Marchand
 wrote:
>
> On Fri, Oct 14, 2022 at 9:51 AM Markus Theil  
> wrote:
> >
> > Makes apps configurable from meson, like already
> > possible for drivers.
> >
> > Signed-off-by: Markus Theil 
>
> With these new options, it's hard to tell which and why some
> application is built (or not).
> Can we have an output similar to libraries and drivers (see toplevel
> meson.build) ?

WDYT of followup patch: https://github.com/david-marchand/dpdk/commit/apps


-- 
David Marchand



RE: [EXT] [PATCH v2 1/1] doc: update to include bbdev snippet using literalinclude

2022-10-27 Thread Akhil Goyal
> Adding code snippet using literalinclude so that to keep
> automatically these structures in doc in sync with the
> bbdev source code.
> 
> Signed-off-by: Nicolas Chautru 
Applied to dpdk-next-crypto

Thanks.


Re: [PATCH] mempool: cache align mempool cache objects

2022-10-27 Thread Olivier Matz
On Thu, Oct 27, 2022 at 02:11:29PM +0200, Morten Brørup wrote:
> > From: Olivier Matz [mailto:olivier.m...@6wind.com]
> > Sent: Thursday, 27 October 2022 13.43
> > 
> > On Thu, Oct 27, 2022 at 11:22:07AM +0200, Morten Brørup wrote:
> > > > From: Olivier Matz [mailto:olivier.m...@6wind.com]
> > > > Sent: Thursday, 27 October 2022 10.35
> > > >
> > > > Hi Morten,
> > > >
> > > > On Wed, Oct 26, 2022 at 04:44:36PM +0200, Morten Brørup wrote:
> > > > > Add __rte_cache_aligned to the objs array.
> > > > >
> > > > > It makes no difference in the general case, but if get/put
> > operations
> > > > are
> > > > > always 32 objects, it will reduce the number of memory (or last
> > level
> > > > > cache) accesses from five to four 64 B cache lines for every
> > get/put
> > > > > operation.
> > > > >
> > > > > For readability reasons, an example using 16 objects follows:
> > > > >
> > > > > Currently, with 16 objects (128B), we access to 3
> > > > > cache lines:
> > > > >
> > > > >   ┌┐
> > > > >   │len │
> > > > > cache ││---
> > > > > line0 ││ ^
> > > > >   ││ |
> > > > >   ├┤ | 16 objects
> > > > >   ││ | 128B
> > > > > cache ││ |
> > > > > line1 ││ |
> > > > >   ││ |
> > > > >   ├┤ |
> > > > >   ││_v_
> > > > > cache ││
> > > > > line2 ││
> > > > >   ││
> > > > >   └┘
> > > > >
> > > > > With the alignment, it is also 3 cache lines:
> > > > >
> > > > >   ┌┐
> > > > >   │len │
> > > > > cache ││
> > > > > line0 ││
> > > > >   ││
> > > > >   ├┤---
> > > > >   ││ ^
> > > > > cache ││ |
> > > > > line1 ││ |
> > > > >   ││ |
> > > > >   ├┤ | 16 objects
> > > > >   ││ | 128B
> > > > > cache ││ |
> > > > > line2 ││ |
> > > > >   ││ v
> > > > >   └┘---
> > > > >
> > > > > However, accessing the objects at the bottom of the mempool cache
> > is
> > > > a
> > > > > special case, where cache line0 is also used for objects.
> > > > >
> > > > > Consider the next burst (and any following bursts):
> > > > >
> > > > > Current:
> > > > >   ┌┐
> > > > >   │len │
> > > > > cache ││
> > > > > line0 ││
> > > > >   ││
> > > > >   ├┤
> > > > >   ││
> > > > > cache ││
> > > > > line1 ││
> > > > >   ││
> > > > >   ├┤
> > > > >   ││
> > > > > cache ││---
> > > > > line2 ││ ^
> > > > >   ││ |
> > > > >   ├┤ | 16 objects
> > > > >   ││ | 128B
> > > > > cache ││ |
> > > > > line3 ││ |
> > > > >   ││ |
> > > > >   ├┤ |
> > > > >   ││_v_
> > > > > cache ││
> > > > > line4 ││
> > > > >   ││
> > > > >   └┘
> > > > > 4 cache lines touched, incl. line0 for len.
> > > > >
> > > > > With the proposed alignment:
> > > > >   ┌┐
> > > > >   │len │
> > > > > cache ││
> > > > > line0 ││
> > > > >   ││
> > > > >   ├┤
> > > > >   ││
> > > > > cache ││
> > > > > line1 ││
> > > > >   ││
> > > > >   ├┤
> > > > >   ││
> > > > > cache ││
> > > > > line2 ││
> > > > >   ││
> > > > >   ├┤
> > > > >   ││---
> > > > > cache ││ ^
> > > > > line3 ││ |
> > > > >   ││ | 16 objects
> > > > >   ├┤ | 128B
> > > > >   ││ |
> > > > > cache ││ |
> > > > > line4 ││ |
> > > > >   ││_v_
> > > > >   └┘
> > > > > Only 3 cache lines touched, incl. line0 for len.
> > > >
> > > > I understand your logic, but are we sure that having an application
> > > > that
> > > > works with bulks of 32 means that the cache will stay aligned to 32
> > > > elements for the whole life of the application?
> > > >
> > > > In an application, the alignment of the cache can change if you
> > have
> > > > any of:
> > > > - software queues (reassembly for instance)
> > > > - packet duplication (bridge, multicast)
> > > > - locally generated packets (keepalive, control protocol)
> > > > - pipeline to other cores
> > > >
> > > > Even with testpmd, which work by bulk of 32, I can see that the
> > size
> > > > of the cache filling is not aligned to 32. Right after starting the
> > > > application, we already have this:
> > > >
> > > >   internal cache infos:
> > > > cache_size=250
> > > > cache_count[0]=231
> > > >
> > > > This is probably related to the hw rx rings size, number of queues,
> > > > number of ports.
> > > >
> > > > The "250" default value for cache size in testpmd is questionable,
> > but
> > > > with --mbcache=256, the behavior is similar.

Re: [PATCH] app/testpmd: fix testpmd receive jumbo frame packets

2022-10-27 Thread Singh, Aman Deep

Hi Jie,

Thanks for the patch.


On 10/25/2022 7:35 AM, Jie Wang wrote:

For NIC I40E_10G-10G_BASE_T_X722, when testpmd is configured with
link speed, it cannot receive jumbo frame packets.


Why only the jumbo frame are effected and not other pkts.



Because it has changed the link status of the ports if it was
configured with link speed.


If we configure to same link speed, will it still have issue ?



When exiting testpmd that it automatically stops packet forwarding
and stops all the ports. But it doesn't update the link status of the
ports. If stop the ports first that it will update the link status.

This patch fix the error that testpmd will update the link status of
the ports when it exits.

Fixes: d3a274ce9dee ("app/testpmd: handle SIGINT and SIGTERM")
Fixes: 284c908cc588 ("app/testpmd: request device removal interrupt")
Cc: sta...@dpdk.org

Signed-off-by: Jie Wang 
---
  app/test-pmd/testpmd.c | 4 ++--
  1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c
index 97adafacd0..c348a3f328 100644
--- a/app/test-pmd/testpmd.c
+++ b/app/test-pmd/testpmd.c
@@ -3548,7 +3548,7 @@ pmd_test_exit(void)
}
  #endif
if (ports != NULL) {
-   no_link_check = 1;
+   no_link_check = 0;
RTE_ETH_FOREACH_DEV(pt_id) {
printf("\nStopping port %d...\n", pt_id);
fflush(stdout);
@@ -3675,7 +3675,7 @@ rmv_port_callback(void *arg)
need_to_start = 1;
stop_packet_forwarding();
}
-   no_link_check = 1;
+   no_link_check = 0;


Well, here we are undoing a previous change done for 284c908cc588
Won't that issue come back.


stop_port(port_id);
no_link_check = org_no_link_check;
  




[pull-request] next-crypto 22.11 rc2

2022-10-27 Thread Akhil Goyal
The following changes since commit 9082336048e8c5779c1b4f8930041a42e9197348:

  net/nfp: support new solution for tunnel decap action (2022-10-25 10:53:33 
+0200)

are available in the Git repository at:

  http://dpdk.org/git/next/dpdk-next-crypto 

for you to fetch changes up to ee18061c78af2997585c7a59c03bae397f16:

  doc: include bbdev code snippet using literalinclude (2022-10-27 19:54:50 
+0530)


Ali Alnubani (1):
  examples/l2fwd-crypto: fix typo in error message

Anoob Joseph (5):
  crypto/cnxk: align HW accessible fields to ROC
  crypto/cnxk: fix failure from session rework
  crypto/cnxk: verify IV length during session create
  crypto/cnxk: remove dead code
  test/crypto: fix PDCP vectors

Arek Kusztal (9):
  crypto/qat: fix uncleared cookies in asym
  crypto/qat: fix unnecessary session check
  crypto/qat: fix not set RSA lengths
  common/qat: read HW slice configuration
  crypto/qat: read HW slice configuration
  crypto/qat: make immutable parameters constant
  crypto/qat: improve asym cookies free
  crypto/qat: add ECDH key exchange algorithm
  crypto/qat: add ECDH public key verification

Brian Dooley (5):
  crypto/qat: reallocate on OpenSSL version check
  examples/fips_validation: add parsing for AES GMAC
  examples/fips_validation: add parsing for AES CTR
  examples/fips_validation: fix parse block error
  crypto/qat: fix null hash algorithm digest size

Ciara Power (6):
  test/crypto: fix wireless auth digest segment
  crypto/ipsec_mb: support remaining SGL
  test/crypto: add OOP SNOW3G SGL cases
  test/crypto: add remaining blockcipher SGL cases
  test/crypto-perf: fix number of scheduler sessions
  crypto/scheduler: fix op session in pkt size distr mode

Gowrishankar Muthukrishnan (6):
  examples/fips_validation: add asymmetric validation
  examples/fips_validation: encode digest with hash OID
  examples/fips_validation: randomize message for conformance test
  crypto/cnxk: support exponent type private key
  examples/fips_validation: fix GMAC decryption output
  examples/fips_validation: validate ECDSA

Hernan Vargas (28):
  baseband/acc100: fix ring availability calculation
  baseband/acc100: check AQ availability
  baseband/acc100: fix memory leak
  baseband/acc100: add LDPC encoder padding function
  baseband/acc100: check turbo dec/enc input
  baseband/acc100: add null checks
  baseband/acc100: enforce additional check on FCW
  baseband/acc100: fix ring/queue allocation
  baseband/acc100: fix input length for CRC24B
  baseband/acc100: fix clearing PF IR outside handler
  baseband/acc100: fix device min alignment
  baseband/acc100: fix null HARQ input case
  baseband/acc100: fix pointer after free
  baseband/acc100: update debug print for LDPC FCW
  baseband/acc100: add enqueue status
  baseband/acc100: support scatter-gather
  baseband/acc100: add HARQ index helper function
  baseband/acc100: enable input validation by default
  baseband/acc100: support LDPC transport block
  baseband/acc100: update LDPC enc/dec validation
  baseband/acc100: implement configurable queue depth
  baseband/acc100: add queue stop operation
  baseband/acc100: update uplink CB input length
  baseband/acc100: update log messages
  baseband/acc100: store FCW from first CB descriptor
  baseband/acc100: update device info
  baseband/acc100: add ring companion address
  baseband/acc100: configure PMON control registers

Kai Ji (2):
  crypto/qat: fix multi-process build request session
  crypto/ipsec_mb: add multi-process IPC request handler

Nic Chautru (16):
  baseband/acc100: remove unused registers
  baseband/acc100: refactor to segregate common code
  baseband/acc: rename directory from acc100 to acc
  baseband/acc: introduce ACC200
  baseband/acc: add ACC200 HW register definitions
  baseband/acc: add ACC200 info get function
  baseband/acc: add ACC200 queue configuration
  baseband/acc: add LDPC processing functions
  baseband/acc: add LTE processing functions
  baseband/acc: support FFT operations
  baseband/acc: support interrupt
  baseband/acc: add device status and vf2pf comms
  baseband/acc: add PF configure companion function
  baseband/acc: simplify meson dependency
  baseband/acc: add descriptor index helpers
  test/bbdev: add FFT operations cases

Nicolas Chautru (2):
  baseband/turbo_sw: remove Flexran SDK meson option
  doc: include bbdev code snippet using literalinclude

Nithin Dabilpuram (1):
  app/test: add multi-seg inline IPsec cases

Radu Nicolau (1):
  examples/ipsec-secgw: fix Tx checksum offload flag

Rahul Bhansali (1):
  examples/ipsec-secgw: support per SA HW reassembly

Srujana Challa 

RE: [dpdk-dev v5] crypto/ipsec_mb: multi-process IPC request handler

2022-10-27 Thread Konstantin Ananyev



> 
> As the queue pair used in secondary process needs to be set up by
> the primary process, this patch adds an IPC register function to help
> secondary process to send out queue-pair setup request to primary
> process via IPC request messages. A new "qp_in_used_pid" param stores
> the PID to provide the ownership of the queue-pair so that only the PID
> matched queue-pair can be free'd in the request.

As I can see that patch was already merged, but still: 
How we suppose  to guarantee synchronization with such approach?
Let say secondary process sends an IPC message to configure 
crypto-queue, and at the same moment primary process (for whatever reason)
decides to reconfigure same crypto-dev.
Is there any way to prevent such race-conditions to happen?

> 
> Signed-off-by: Kai Ji 
> Acked-by: Arek Kusztal 
> Acked-by: Pablo de Lara 
> ---
> v5:
> - minor updates and typo fix
> 
> v4:
> - review comments resolved
> 
> v3:
> - remove shared memzone as qp_conf params can be passed directly from
> ipc message.
> 
> v2:
> - add in shared memzone for data exchange between multi-process
> ---
>  drivers/crypto/ipsec_mb/ipsec_mb_ops.c | 129 -
>  drivers/crypto/ipsec_mb/ipsec_mb_private.c |  24 +++-
>  drivers/crypto/ipsec_mb/ipsec_mb_private.h |  38 +-
>  3 files changed, 185 insertions(+), 6 deletions(-)
> 
> diff --git a/drivers/crypto/ipsec_mb/ipsec_mb_ops.c 
> b/drivers/crypto/ipsec_mb/ipsec_mb_ops.c
> index cedcaa2742..bf18d692bd 100644
> --- a/drivers/crypto/ipsec_mb/ipsec_mb_ops.c
> +++ b/drivers/crypto/ipsec_mb/ipsec_mb_ops.c
> @@ -3,6 +3,7 @@
>   */
> 
>  #include 
> +#include 
> 
>  #include 
>  #include 
> @@ -93,6 +94,46 @@ ipsec_mb_info_get(struct rte_cryptodev *dev,
>   }
>  }
> 
> +static int
> +ipsec_mb_secondary_qp_op(int dev_id, int qp_id,
> + const struct rte_cryptodev_qp_conf *qp_conf,
> + int socket_id, enum ipsec_mb_mp_req_type op_type)
> +{
> + int ret;
> + struct rte_mp_msg qp_req_msg;
> + struct rte_mp_msg *qp_resp_msg;
> + struct rte_mp_reply qp_resp;
> + struct ipsec_mb_mp_param *req_param;
> + struct ipsec_mb_mp_param *resp_param;
> + struct timespec ts = {.tv_sec = 1, .tv_nsec = 0};
> +
> + memset(&qp_req_msg, 0, sizeof(IPSEC_MB_MP_MSG));
> + memcpy(qp_req_msg.name, IPSEC_MB_MP_MSG, sizeof(IPSEC_MB_MP_MSG));
> + req_param = (struct ipsec_mb_mp_param *)&qp_req_msg.param;
> +
> + qp_req_msg.len_param = sizeof(struct ipsec_mb_mp_param);
> + req_param->type = op_type;
> + req_param->dev_id = dev_id;
> + req_param->qp_id = qp_id;
> + req_param->socket_id = socket_id;
> + req_param->process_id = getpid();
> + if (qp_conf) {
> + req_param->nb_descriptors = qp_conf->nb_descriptors;
> + req_param->mp_session = (void *)qp_conf->mp_session;
> + }
> +
> + qp_req_msg.num_fds = 0;
> + ret = rte_mp_request_sync(&qp_req_msg, &qp_resp, &ts);
> + if (ret) {
> + RTE_LOG(ERR, USER1, "Create MR request to primary process 
> failed.");
> + return -1;
> + }
> + qp_resp_msg = &qp_resp.msgs[0];
> + resp_param = (struct ipsec_mb_mp_param *)qp_resp_msg->param;
> +
> + return resp_param->result;
> +}
> +
>  /** Release queue pair */
>  int
>  ipsec_mb_qp_release(struct rte_cryptodev *dev, uint16_t qp_id)
> @@ -100,7 +141,10 @@ ipsec_mb_qp_release(struct rte_cryptodev *dev, uint16_t 
> qp_id)
>   struct ipsec_mb_qp *qp = dev->data->queue_pairs[qp_id];
>   struct rte_ring *r = NULL;
> 
> - if (qp != NULL && rte_eal_process_type() == RTE_PROC_PRIMARY) {
> + if (qp != NULL)
> + return 0;
> +
> + if (rte_eal_process_type() == RTE_PROC_PRIMARY) {
>   r = rte_ring_lookup(qp->name);
>   rte_ring_free(r);
> 
> @@ -115,6 +159,9 @@ ipsec_mb_qp_release(struct rte_cryptodev *dev, uint16_t 
> qp_id)
>  #endif
>   rte_free(qp);
>   dev->data->queue_pairs[qp_id] = NULL;
> + } else { /* secondary process */
> + return ipsec_mb_secondary_qp_op(dev->data->dev_id, qp_id,
> + NULL, 0, RTE_IPSEC_MB_MP_REQ_QP_FREE);
>   }
>   return 0;
>  }
> @@ -222,9 +269,13 @@ ipsec_mb_qp_setup(struct rte_cryptodev *dev, uint16_t 
> qp_id,
>  #endif
>   qp = dev->data->queue_pairs[qp_id];
>   if (qp == NULL) {
> - IPSEC_MB_LOG(ERR, "Primary process hasn't configured 
> device qp.");
> - return -EINVAL;
> + IPSEC_MB_LOG(DEBUG, "Secondary process setting up 
> device qp.");
> + return ipsec_mb_secondary_qp_op(dev->data->dev_id, 
> qp_id,
> + qp_conf, socket_id, 
> RTE_IPSEC_MB_MP_REQ_QP_SET);
>   }
> +
> + IPSEC_MB_LOG(ERR, "Queue pair already setup'ed.");
> + return -EINVAL;
>   } else {
>   /* Free memory 

[PATCH] mempool: optimized debug statistics

2022-10-27 Thread Morten Brørup
When built with debug enabled (RTE_LIBRTE_MEMPOOL_DEBUG defined), the
performance of mempools with caches is improved as follows.

Accessing objects in the mempool is likely to increment either the
put_bulk and put_objs or the get_success_bulk and get_success_objs
debug statistics counters.

By adding an alternative set of these counters to the mempool cache
structure, accesing the dedicated debug statistics structure is avoided in
the likely cases where these counters are incremented.

The trick here is that the cache line holding the mempool cache structure
is accessed anyway, in order to update the "len" field. Updating some
debug statistics counters in the same cache line has lower performance
cost than accessing the debug statistics counters in the dedicated debug
statistics structure, i.e. in another cache line.

Running mempool_perf_autotest on a VMware virtual server shows an avg.
increase of 6.4 % in rate_persec for the tests with cache. (Only when
built with with debug enabled, obviously!)

For the tests without cache, the avg. increase in rate_persec is 0.8 %. I
assume this is noise from the test environment.

Signed-off-by: Morten Brørup 
---
 lib/mempool/rte_mempool.c |  7 +
 lib/mempool/rte_mempool.h | 55 +++
 2 files changed, 51 insertions(+), 11 deletions(-)

diff --git a/lib/mempool/rte_mempool.c b/lib/mempool/rte_mempool.c
index 21c94a2b9f..7b8c00a022 100644
--- a/lib/mempool/rte_mempool.c
+++ b/lib/mempool/rte_mempool.c
@@ -1285,6 +1285,13 @@ rte_mempool_dump(FILE *f, struct rte_mempool *mp)
sum.get_fail_objs += mp->stats[lcore_id].get_fail_objs;
sum.get_success_blks += mp->stats[lcore_id].get_success_blks;
sum.get_fail_blks += mp->stats[lcore_id].get_fail_blks;
+   /* Add the fast access statistics, if local caches exist */
+   if (mp->cache_size != 0) {
+   sum.put_bulk += mp->local_cache[lcore_id].put_bulk;
+   sum.put_objs += mp->local_cache[lcore_id].put_objs;
+   sum.get_success_bulk += 
mp->local_cache[lcore_id].get_success_bulk;
+   sum.get_success_objs += 
mp->local_cache[lcore_id].get_success_objs;
+   }
}
fprintf(f, "  stats:\n");
fprintf(f, "put_bulk=%"PRIu64"\n", sum.put_bulk);
diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 3725a72951..d84087bc92 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -86,6 +86,14 @@ struct rte_mempool_cache {
uint32_t size;/**< Size of the cache */
uint32_t flushthresh; /**< Threshold before we flush excess elements */
uint32_t len; /**< Current cache count */
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+   uint32_t unused;
+   /* Fast access statistics, only for likely events */
+   uint64_t put_bulk; /**< Number of puts. */
+   uint64_t put_objs; /**< Number of objects successfully put. 
*/
+   uint64_t get_success_bulk; /**< Successful allocation number. */
+   uint64_t get_success_objs; /**< Objects successfully allocated. */
+#endif
/**
 * Cache objects
 *
@@ -1327,13 +1335,19 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void 
* const *obj_table,
 {
void **cache_objs;
 
+   /* No cache provided */
+   if (unlikely(cache == NULL))
+   goto driver_enqueue;
+
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
/* increment stat now, adding in mempool always success */
-   RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
-   RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
+   cache->put_bulk += 1;
+   cache->put_objs += n;
+#endif
 
-   /* No cache provided or the request itself is too big for the cache */
-   if (unlikely(cache == NULL || n > cache->flushthresh))
-   goto driver_enqueue;
+   /* The request is too big for the cache */
+   if (unlikely(n > cache->flushthresh))
+   goto driver_enqueue_stats_incremented;
 
/*
 * The cache follows the following algorithm:
@@ -1358,6 +1372,12 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void 
* const *obj_table,
 
 driver_enqueue:
 
+   /* increment stat now, adding in mempool always success */
+   RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
+   RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
+
+driver_enqueue_stats_incremented:
+
/* push objects to the backend */
rte_mempool_ops_enqueue_bulk(mp, obj_table, n);
 }
@@ -1464,8 +1484,10 @@ rte_mempool_do_generic_get(struct rte_mempool *mp, void 
**obj_table,
if (remaining == 0) {
/* The entire request is satisfied from the cache. */
 
-   RTE_MEMPOOL_STAT_ADD(mp, get_success_bulk, 1);
-   RTE_MEMPOOL_STAT_ADD(mp, get_success_objs, n);
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+   cache->get_success_bulk += 1;
+ 

[PATCH v2] mempool: optimized debug statistics

2022-10-27 Thread Morten Brørup
When built with debug enabled (RTE_LIBRTE_MEMPOOL_DEBUG defined), the
performance of mempools with caches is improved as follows.

Accessing objects in the mempool is likely to increment either the
put_bulk and put_objs or the get_success_bulk and get_success_objs
debug statistics counters.

By adding an alternative set of these counters to the mempool cache
structure, accessing the dedicated debug statistics structure is avoided
in the likely cases where these counters are incremented.

The trick here is that the cache line holding the mempool cache structure
is accessed anyway, in order to update the "len" field. Updating some
debug statistics counters in the same cache line has lower performance
cost than accessing the debug statistics counters in the dedicated debug
statistics structure, i.e. in another cache line.

Running mempool_perf_autotest on a VMware virtual server shows an avg.
increase of 6.4 % in rate_persec for the tests with cache. (Only when
built with debug enabled, obviously!)

For the tests without cache, the avg. increase in rate_persec is 0.8 %. I
assume this is noise from the test environment.

v2:
* Fix spelling and repeated word in commit message, caught by checkpatch.

Signed-off-by: Morten Brørup 
---
 lib/mempool/rte_mempool.c |  7 +
 lib/mempool/rte_mempool.h | 55 +++
 2 files changed, 51 insertions(+), 11 deletions(-)

diff --git a/lib/mempool/rte_mempool.c b/lib/mempool/rte_mempool.c
index 21c94a2b9f..7b8c00a022 100644
--- a/lib/mempool/rte_mempool.c
+++ b/lib/mempool/rte_mempool.c
@@ -1285,6 +1285,13 @@ rte_mempool_dump(FILE *f, struct rte_mempool *mp)
sum.get_fail_objs += mp->stats[lcore_id].get_fail_objs;
sum.get_success_blks += mp->stats[lcore_id].get_success_blks;
sum.get_fail_blks += mp->stats[lcore_id].get_fail_blks;
+   /* Add the fast access statistics, if local caches exist */
+   if (mp->cache_size != 0) {
+   sum.put_bulk += mp->local_cache[lcore_id].put_bulk;
+   sum.put_objs += mp->local_cache[lcore_id].put_objs;
+   sum.get_success_bulk += 
mp->local_cache[lcore_id].get_success_bulk;
+   sum.get_success_objs += 
mp->local_cache[lcore_id].get_success_objs;
+   }
}
fprintf(f, "  stats:\n");
fprintf(f, "put_bulk=%"PRIu64"\n", sum.put_bulk);
diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 3725a72951..d84087bc92 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -86,6 +86,14 @@ struct rte_mempool_cache {
uint32_t size;/**< Size of the cache */
uint32_t flushthresh; /**< Threshold before we flush excess elements */
uint32_t len; /**< Current cache count */
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+   uint32_t unused;
+   /* Fast access statistics, only for likely events */
+   uint64_t put_bulk; /**< Number of puts. */
+   uint64_t put_objs; /**< Number of objects successfully put. 
*/
+   uint64_t get_success_bulk; /**< Successful allocation number. */
+   uint64_t get_success_objs; /**< Objects successfully allocated. */
+#endif
/**
 * Cache objects
 *
@@ -1327,13 +1335,19 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void 
* const *obj_table,
 {
void **cache_objs;
 
+   /* No cache provided */
+   if (unlikely(cache == NULL))
+   goto driver_enqueue;
+
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
/* increment stat now, adding in mempool always success */
-   RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
-   RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
+   cache->put_bulk += 1;
+   cache->put_objs += n;
+#endif
 
-   /* No cache provided or the request itself is too big for the cache */
-   if (unlikely(cache == NULL || n > cache->flushthresh))
-   goto driver_enqueue;
+   /* The request is too big for the cache */
+   if (unlikely(n > cache->flushthresh))
+   goto driver_enqueue_stats_incremented;
 
/*
 * The cache follows the following algorithm:
@@ -1358,6 +1372,12 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void 
* const *obj_table,
 
 driver_enqueue:
 
+   /* increment stat now, adding in mempool always success */
+   RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
+   RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
+
+driver_enqueue_stats_incremented:
+
/* push objects to the backend */
rte_mempool_ops_enqueue_bulk(mp, obj_table, n);
 }
@@ -1464,8 +1484,10 @@ rte_mempool_do_generic_get(struct rte_mempool *mp, void 
**obj_table,
if (remaining == 0) {
/* The entire request is satisfied from the cache. */
 
-   RTE_MEMPOOL_STAT_ADD(mp, get_success_bulk, 1);
-   RTE_MEMPOOL_STAT_ADD(mp, get_success_objs, n);
+#ifdef RTE_LI

DPDK Release Status Meeting 2022-10-27

2022-10-27 Thread Mcnamara, John
Release status meeting minutes 2022-10-27
=

Agenda:
* Release Dates
* Subtrees
* Roadmaps
* LTS
* Defects
* Opens

Participants:
* ARM
* Canonical [No]
* Debian/Microsoft
* Intel
* Marvell
* Nvidia
* Red Hat
* Xilinx/AMD


Release Dates
-

The following are the proposed current dates for 22.11:

* V1 deadline: 24 August   2022
* RC1:  7 October  2022
* RC2: 28 October  2022 - moved from 24 October
* RC3: 31 October  2022
* Release: 16 November 2022

Subtrees



* next-net
  * NFP several patches merged
  * Indirect Aging patches under review
* New version sent
  * New drivers in this release:
* Microsoft Azure Network Adapter (MANA) PMD
  * Compilation will be disabled until dependencies become available
* Google Virtual Ethernet (GVE)
  * Merged
* idpf (Infrastructure Data Path Function) PMD
  * Under review
  * New updates sent
* net/qdma PMD
  * Awaiting reply from submitter
  * Most likely next release
  * Memif driver needs reviews from maintainer

* next-net-intel
  * ~20 patches, mainly bug fixes, waiting for merge

* next-net-mlx
  * PR sent for RC2

* next-eventdev
  * PR merged for RC2

* next-virtio
  * PR ready for RC2

* next-crypto
  * 60-70 patches merged and ready for pull
  * New Huawei UADK driver ready for merge
  * Will be ready for pull on Friday 28 October

* main
  * RC2 on Friday 28 October
  * Mempool changes merged
  * Meson version upgraded to 0.53.2.
  * KNI deprecation warnings added for compile and runtime
  * We should also upgrade the minimum required kernel version for 22.11
* Suggestions welcome
* Probably the earliest LTS kernel: 4.14
  * Looking for RC1 test reports from different vendors


Proposed Schedule for 2023

23.03
  * Proposal deadline (RFC/v1 patches): 30 December 2022
  * API freeze (-rc1): 10 February 2023
  * PMD features freeze (-rc2): 24 February 2023
  * Built-in applications features freeze (-rc3): 3 March 2023
  * Release: 10 March 2023
23.07
  * Proposal deadline (RFC/v1 patches): 7 April 2023
  * API freeze (-rc1): 26 May 2023
  * PMD features freeze (-rc2): 16 June 2023
  * Built-in applications features freeze (-rc3): 23 June 2023
  * Release: 14 July 2023
23.11
  * Proposal deadline (RFC/v1 patches): 11 August 2023
  * API freeze (-rc1): 29 September 2023
  * PMD features freeze (-rc2): 20 October 2023
  * Built-in applications features freeze (-rc3): 27 October 2023
  * Release: 17 November 2023


Other
-

* TBA

LTS
---

LTSes released on August 29th.

Next releases will be:

* 21.11.3

* 20.11.7

* 19.11.14

Discussion needed on whether we will continue the 3 year LTS cycle that
was trialled for 19.11 to the 20.11 LTS cycle.


* Distros
  * v20.11 in Debian 11
  * Ubuntu 22.04 contains 21.11

Defects
---

* Bugzilla links, 'Bugs',  added for hosted projects
  * https://www.dpdk.org/hosted-projects/


Opens
-

* None


DPDK Release Status Meetings


The DPDK Release Status Meeting is intended for DPDK Committers to discuss the
status of the master tree and sub-trees, and for project managers to track
progress or milestone dates.

The meeting occurs on every Thursday at 9:30 UTC over Jitsi on 
https://meet.jit.si/DPDK

You don't need an invite to join the meeting but if you want a calendar 
reminder just
send an email to "John McNamara john.mcnam...@intel.com" for the invite.




[Bug 1117] af-packet driver fix breaks rte_pktmbuf_prepend()

2022-10-27 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1117

Bug ID: 1117
   Summary: af-packet driver fix breaks rte_pktmbuf_prepend()
   Product: DPDK
   Version: 21.11
  Hardware: All
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: ethdev
  Assignee: dev@dpdk.org
  Reporter: bly...@gmail.com
  Target Milestone: ---

The following commit causes rte_pktmbuf_prepend() to fail when prepending the
full headroom reserved when creating the mbuf pool.

https://git.dpdk.org/dpdk-stable/commit/?h=21.11&id=d41d39bcf76900deb6aa000e929fd65734254201

Of issue is if the received frame was originally tagged (and vlan-stripped by
af-socket connection) the above commit does a vlan reinsert action as part of
eth_af_packet_rx(). This moves the SOF forward into the "reserved" space by
four bytes. For example, if the reserved space was 128 bytes and the next stage
of processing is a generic pipeline handling a variety of interface types, the
generic prepend of the reserved 128-byte meta-data space will fail. This means
you cannot use a common pipeline and have to have an exception now for
af-packet rx handling as compared to physical NIC, tap, vhost, etc. Ideally,
the initial SOF offset should be consistent relative to your design's reserved
headroom, regardless of the vlan tagged status of the frame or the interface
type it was received on.

-- 
You are receiving this mail because:
You are the assignee for the bug.

Re: release candidate 22.11-rc1

2022-10-27 Thread Thinh Tran

Hello,

IBM - Power Systems
DPDK 22.11.0-rc1


* Basic PF on Mellanox: No new issues or regressions were seen.
* Performance: not tested.
* OS: RHEL 8.5  kernel: 4.18.0-348.el8.ppc64le
with gcc version 8.5.0 20210514 (Red Hat 8.5.0-10)
  RHEL 9.0  kernel: 5.14.0-70.13.1.el9_0.ppc64le
with gcc version 11.2.1 20220127 (Red Hat 11.2.1-9)

Systems tested:
 - IBM Power9 PowerNV 9006-22P
NICs:
 - Mellanox Technologies MT28800 Family [ConnectX-5 Ex]
 - firmware version: 16.34.1002
 - MLNX_OFED_LINUX-5.7-1.0.2.1 (OFED-5.7-1.0.2)

 - LPARs on IBM Power10 CHRP IBM,9105-22A
NICs:
- Mellanox Technologies MT28800 Family [ConnectX-5 Ex]
- firmware version: 16.34.1002
- MLNX_OFED_LINUX-5.7-1.0.2.1 (OFED-5.7-1.0.2)

Regards,
Thinh Tran

On 10/10/2022 8:50 PM, Thomas Monjalon wrote:

A new DPDK release candidate is ready for testing:
https://git.dpdk.org/dpdk/tag/?id=v22.11-rc1

There are 737 new patches in this snapshot,
including many API/ABI compatibility breakages.
This release won't be ABI-compatible with previous ones.

Release notes:
https://doc.dpdk.org/guides/rel_notes/release_22_11.html

Highlights of 22.11-rc1:
- LoongArch build
- Intel uncore frequency control
- multiple mbuf pools per Rx queue
- Rx buffer split based on protocol
- hardware congestion management
- hairpin memory configuration
- Rx/Tx descriptor dump
- flow API extensions
- MACsec processing offload
- ShangMi crypto algorithms
- baseband FFT operations
- eventdev Tx queue start/stop
- eventdev crypto vectorization
- NitroSketch membership

Some work is in progress to optimize the mempool cache.
Some patches are part of -rc1, and more could be merged in -rc2.
Please measure the performance of this release candidate,
and check these mempool patches:
https://patches.dpdk.org/project/dpdk/list/?series=25063

Please test and report issues on bugs.dpdk.org.

DPDK 22.11-rc2 is expected in two weeks.

Thank you everyone




Re: [EXT] [PATCH v7 0/6] crypto/uadk: introduce uadk crypto driver

2022-10-27 Thread Zhangfei Gao




On 2022/10/27 下午6:53, Akhil Goyal wrote:

Introduce a new crypto PMD for hardware accelerators based on UADK [1].

UADK is a framework for user applications to access hardware accelerators.
UADK relies on IOMMU SVA (Shared Virtual Address) feature, which share
the same page table between IOMMU and MMU.
Thereby user application can directly use virtual address for device dma,
which enhances the performance as well as easy usability.

[1] https://urldefense.proofpoint.com/v2/url?u=https-
3A__github.com_Linaro_uadk&d=DwIDAg&c=nKjWec2b6R0mOyPaz7xtfQ&r=Dn
L7Si2wl_PRwpZ9TWey3eu68gBzn7DkPwuqhd6WNyo&m=AwapMJzvfUDL49W9f
mtmvIGZMarGUdjRSliBhG4tep6Uh5wN2zYrZRJ4JpbrVolg&s=MxSabEjMMkeRr0L
Z7c1gSmaJmJ1_dIdJIKrst98pKzk&e=

Test:
sudo dpdk-test --vdev=crypto_uadk (--log-level=6)
RTE>>cryptodev_uadk_autotest
RTE>>quit

Update in v7:
05: fix key_size of SHA384 HMAC

Update in v6:
Akhil helped review the v5, and address all the comments from Akhil
01: add rel_notes info, update uadk.rst, put struct in header, etc.
02: move struct to header
04: move struct to header, remove RTE_CRYPTODEV_FF_SYM_SESSIONLESS
05: fixed key_size in HMAC mode
06: updated app/test/meson.build and uadk.rst in the mean time.

Update in v5
Patch 1 fix the build issue when uadk is installed to specific folder
And update doc accordingly

Update in v4:
Akril suggest dpdk use pkg-config, So
Enable uadk support x86 local build, and support pkg-config.
Use pkg-config feature for the uadk crypto pmd.
Add build uadk library steps in doc
Test on both x86 and arm.
x86 can build and install, but can not test since no device.

Resend v3:
Rebase on next/for-main, which just merged the series
"cryptodev: rework session framework".

Update in v3:
Split patches according to Akhil's suggestions
Please split the patches as below.
1. introduce driver - create files with meson.build and with probe/remove
and device ops defined but not implemented.
You do not need to write empty functions.
Add basic documentation also which defines what the driver is.
You can explain the build dependency here.
2. define queue structs and setup/remove APIs
3. Add data path
4. implement cipher op. Add capabilities and documentation of what is
supported in each of the patches. Add feature flags etc.
5. implement auth,  add capabilities and documentation
6. test app changes.

Update in v2:
Change uadk_supported_platform to uadk_crypto_version, which matches
better
than platform.
enum uadk_crypto_version {
UADK_CRYPTO_V2,
UADK_CRYPTO_V3,
};

Update in v1, compared with rfc

Suggested from Akhil Goyal 
Only consider crypto PMD first
Split patch into small (individually compiled) patches.
Update MAINTAINERS and doc/guides/cryptodevs/features/uadk.ini

Zhangfei Gao (6):
   crypto/uadk: introduce uadk crypto driver
   crypto/uadk: support basic operations
   crypto/uadk: support enqueue/dequeue operations
   crypto/uadk: support cipher algorithms
   crypto/uadk: support auth algorithms
   test/crypto: support uadk PMD

  MAINTAINERS   |6 +
  app/test/meson.build  |1 +
  app/test/test_cryptodev.c |7 +
  app/test/test_cryptodev.h |1 +
  doc/guides/cryptodevs/features/uadk.ini   |   55 +
  doc/guides/cryptodevs/index.rst   |1 +
  doc/guides/cryptodevs/uadk.rst|   96 ++
  doc/guides/rel_notes/release_22_11.rst|6 +
  drivers/crypto/meson.build|1 +
  drivers/crypto/uadk/meson.build   |   30 +
  drivers/crypto/uadk/uadk_crypto_pmd.c | 1081 +
  drivers/crypto/uadk/uadk_crypto_pmd_private.h |   79 ++
  drivers/crypto/uadk/version.map   |3 +
  13 files changed, 1367 insertions(+)
  create mode 100644 doc/guides/cryptodevs/features/uadk.ini
  create mode 100644 doc/guides/cryptodevs/uadk.rst
  create mode 100644 drivers/crypto/uadk/meson.build
  create mode 100644 drivers/crypto/uadk/uadk_crypto_pmd.c
  create mode 100644 drivers/crypto/uadk/uadk_crypto_pmd_private.h
  create mode 100644 drivers/crypto/uadk/version.map

Series Applied to dpdk-next-crypto
Welcome to dpdk.


That's Great
Thanks for your guidance and patience.

Originally, the UADK only consider ARM usage and need cross-compile on x86.
With you guidance, now the UADK support pkg-config, and can build and 
install on both ARM and X86.


Thanks Akhil.



RE: [PATCH v11 15/18] net/idpf: add support for Rx offloading

2022-10-27 Thread Xing, Beilei


> -Original Message-
> From: Andrew Rybchenko 
> Sent: Tuesday, October 25, 2022 6:04 PM
> To: Guo, Junfeng ; Zhang, Qi Z
> ; Wu, Jingjing ; Xing, Beilei
> 
> Cc: dev@dpdk.org; Li, Xiaoyun 
> Subject: Re: [PATCH v11 15/18] net/idpf: add support for Rx offloading
> 
> On 10/24/22 16:12, Junfeng Guo wrote:
> > Add Rx offloading support:
> >   - support CHKSUM and RSS offload for split queue model
> >   - support CHKSUM offload for single queue model
> >
> > Signed-off-by: Beilei Xing 
> > Signed-off-by: Xiaoyun Li 
> > Signed-off-by: Junfeng Guo 
> > ---
> >   doc/guides/nics/features/idpf.ini |   2 +
> >   drivers/net/idpf/idpf_ethdev.c|   9 ++-
> >   drivers/net/idpf/idpf_rxtx.c  | 122 ++
> >   3 files changed, 132 insertions(+), 1 deletion(-)
> >
> > diff --git a/doc/guides/nics/features/idpf.ini
> > b/doc/guides/nics/features/idpf.ini
> > index d4eb9b374c..c86d9378ea 100644
> > --- a/doc/guides/nics/features/idpf.ini
> > +++ b/doc/guides/nics/features/idpf.ini
> > @@ -9,5 +9,7 @@
> >   [Features]
> >   Queue start/stop = Y
> >   MTU update   = Y
> > +L3 checksum offload  = P
> > +L4 checksum offload  = P
> 
> RSS hash missing
> 
> >   Packet type parsing  = Y
> >   Linux= Y
> > diff --git a/drivers/net/idpf/idpf_ethdev.c
> > b/drivers/net/idpf/idpf_ethdev.c index 739cf31d65..d8cc423a23 100644
> > --- a/drivers/net/idpf/idpf_ethdev.c
> > +++ b/drivers/net/idpf/idpf_ethdev.c
> > @@ -94,7 +94,14 @@ idpf_dev_info_get(struct rte_eth_dev *dev, struct
> rte_eth_dev_info *dev_info)
> > dev_info->max_mac_addrs = IDPF_NUM_MACADDR_MAX;
> > dev_info->dev_capa =
> RTE_ETH_DEV_CAPA_RUNTIME_RX_QUEUE_SETUP |
> > RTE_ETH_DEV_CAPA_RUNTIME_TX_QUEUE_SETUP;
> > -   dev_info->rx_offload_capa = 0;
> > +
> > +   dev_info->rx_offload_capa =
> > +   RTE_ETH_RX_OFFLOAD_IPV4_CKSUM   |
> > +   RTE_ETH_RX_OFFLOAD_UDP_CKSUM|
> > +   RTE_ETH_RX_OFFLOAD_TCP_CKSUM|
> > +   RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM |
> > +   RTE_ETH_RX_OFFLOAD_RSS_HASH;
> 
> As I understand you know mode here and you should not report offload
> which is not supported in current mode (RSS vs single queue).

Yes, will remove RTE_ETH_RX_OFFLOAD_RSS_HASH here since RSS hash is not 
supported in single queue mode currently,
So we needn't to add 'RSS hash = Y' in idpf.ini, right?

> 
> > +
> > dev_info->tx_offload_capa = RTE_ETH_TX_OFFLOAD_MULTI_SEGS;
> >
> > dev_info->default_rxconf = (struct rte_eth_rxconf) {
> 
> [snip]



RE: [PATCH v2] app/testpmd: fix protocol headers display for Rx buffer split

2022-10-27 Thread Wang, YuanX
Hi All,

Could you please review and provide suggestions if any.

Thanks,
Yuan

> -Original Message-
> From: Wang, YuanX 
> Sent: Tuesday, October 18, 2022 10:50 PM
> To: andrew.rybche...@oktetlabs.ru; Singh, Aman Deep
> ; Zhang, Yuying 
> Cc: dev@dpdk.org; Ding, Xuan ; Tang, Yaqi
> ; Wang, YuanX 
> Subject: [PATCH v2] app/testpmd: fix protocol headers display for Rx buffer
> split
> 
> The "show config rxhdrs" cmd displays the configured protocol headers that
> are used for protocol-based buffer split.
> However, it shows "inner-ipv6" as "inner-ipv4".
> 
> This patch fixes that by adjusting the order of condition judgments.
> 
> Fixes: 52e2e7edcf48 ("app/testpmd: add protocol-based buffer split")
> 
> Signed-off-by: Yuan Wang 
> Tested-by: Yaqi Tang 
> 
> ---
> v2: add fixline.
> 
> ---
>  app/test-pmd/config.c | 24 
>  1 file changed, 12 insertions(+), 12 deletions(-)
> 
> diff --git a/app/test-pmd/config.c b/app/test-pmd/config.c index
> 0f7dbd698f..82fbbc9944 100644
> --- a/app/test-pmd/config.c
> +++ b/app/test-pmd/config.c
> @@ -4937,15 +4937,6 @@ static const char *get_ptype_str(uint32_t ptype)
>   else if ((ptype & RTE_PTYPE_L2_ETHER) == RTE_PTYPE_L2_ETHER)
>   return "eth";
> 
> - else if ((ptype & (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_TCP)) ==
> - (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_TCP))
> - return "inner-ipv4-tcp";
> - else if ((ptype & (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_UDP)) ==
> - (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_UDP))
> - return "inner-ipv4-udp";
> - else if ((ptype & (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_SCTP)) ==
> - (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_SCTP))
> - return "inner-ipv4-sctp";
>   else if ((ptype & (RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_TCP)) ==
>   (RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_TCP))
>   return "inner-ipv6-tcp";
> @@ -4955,18 +4946,27 @@ static const char *get_ptype_str(uint32_t ptype)
>   else if ((ptype & (RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_SCTP)) ==
>   (RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_SCTP))
>   return "inner-ipv6-sctp";
> + else if ((ptype & (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_TCP)) ==
> + (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_TCP))
> + return "inner-ipv4-tcp";
> + else if ((ptype & (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_UDP)) ==
> + (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_UDP))
> + return "inner-ipv4-udp";
> + else if ((ptype & (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_SCTP)) ==
> + (RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
> RTE_PTYPE_INNER_L4_SCTP))
> + return "inner-ipv4-sctp";
>   else if ((ptype & RTE_PTYPE_INNER_L4_TCP) ==
> RTE_PTYPE_INNER_L4_TCP)
>   return "inner-tcp";
>   else if ((ptype & RTE_PTYPE_INNER_L4_UDP) ==
> RTE_PTYPE_INNER_L4_UDP)
>   return "inner-udp";
>   else if ((ptype & RTE_PTYPE_INNER_L4_SCTP) ==
> RTE_PTYPE_INNER_L4_SCTP)
>   return "inner-sctp";
> + else if ((ptype & RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN) ==
> + RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN)
> + return "inner-ipv6";
>   else if ((ptype & RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN) ==
>   RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN)
>   return "inner-ipv4";
> - else if ((ptype & RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN) ==
> - RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN)
> - return "inner-ipv6";
>   else if ((ptype & RTE_PTYPE_INNER_L2_ETHER) ==
> RTE_PTYPE_INNER_L2_ETHER)
>   return "inner-eth";
>   else if ((ptype & RTE_PTYPE_TUNNEL_GRENAT) ==
> RTE_PTYPE_TUNNEL_GRENAT)
> --
> 2.25.1



Re: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq

2022-10-27 Thread lihuisong (C)



在 2022/10/27 19:02, Ye, MingjinX 写道:

Hi lihuisong,

This means that queue offloads need to update by recalling dev_configure
and setup target queues.

Why not update queue offloads in PMD?

Can you tell me, where is the limitation?
According to other Rx/Tx offload configurations, this may not be a 
limitation.

But it seems to create a dependency on user usage.

Port VLAN releated offloads are set by ethdev ops. There is no requirement
in ehedev layer that this port needs to stopped when set these offloads.
Now it depends on user does recall dev_configure and setup queues to update
queue offloads because of setting these offloads.


Thanks,
Mingjin


-Original Message-
From: lihuisong (C) 
Sent: 2022年10月26日 17:53
To: Ye, MingjinX ; dev@dpdk.org
Cc: sta...@dpdk.org; Zhou, YidingX ; Singh, Aman
Deep ; Zhang, Yuying

Subject: Re: [PATCH v4 1/2] app/testpmd: fix vlan offload of rxq


在 2022/10/27 1:10, Mingjin Ye 写道:

After setting vlan offload in testpmd, the result is not updated to
rxq. Therefore, the queue needs to be reconfigured after executing the
"vlan offload" related commands.

Fixes: a47aa8b97afe ("app/testpmd: add vlan offload support")
Cc: sta...@dpdk.org

Signed-off-by: Mingjin Ye 
---
   app/test-pmd/cmdline.c | 1 +
   1 file changed, 1 insertion(+)

diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c index
17be2de402..ce125f549f 100644
--- a/app/test-pmd/cmdline.c
+++ b/app/test-pmd/cmdline.c
@@ -4133,6 +4133,7 @@ cmd_vlan_offload_parsed(void *parsed_result,
else
vlan_extend_set(port_id, on);

+   cmd_reconfig_device_queue(port_id, 1, 1);

In addition, I have some comments:
1) Normally, the parsed function of testpmd command needed to re-config 
port and

queue needs to check if port status is STOPED. Why don't you add this check?
If the check is not exist, queue offloads are not updated until the next 
port

stop/start command is executed. Right?

2) Why is the queue-based VLAN offload API not used?
   Like, rte_eth_dev_set_vlan_strip_on_queue. It seems that this kind 
of API is

   dedicated to do this.

This means that queue offloads need to upadte by re-calling dev_configure
and setup all queues, right?
If it is, this adds a usage limitation.

return;
   }



RE: [PATCH] app/testpmd: fix testpmd receive jumbo frame packets

2022-10-27 Thread Wang, Jie1X
Hi Singh,

> -Original Message-
> From: Singh, Aman Deep 
> Sent: Thursday, October 27, 2022 11:38 PM
> To: Wang, Jie1X ; dev@dpdk.org
> Cc: Yang, SteveX ; Zhang, Qi Z ;
> Yang, Qiming ; Zhang, Yuying
> ; sta...@dpdk.org
> Subject: Re: [PATCH] app/testpmd: fix testpmd receive jumbo frame packets
> 
> Hi Jie,
> 
> Thanks for the patch.
> 
> 
> On 10/25/2022 7:35 AM, Jie Wang wrote:
> > For NIC I40E_10G-10G_BASE_T_X722, when testpmd is configured with link
> > speed, it cannot receive jumbo frame packets.
> 
> Why only the jumbo frame are effected and not other pkts.
> 

I don't know why only the jumbo frame are effected, when set the link speed, it 
will update the link status of the NIC ports.
And this phenomenon can only be reproduced on this NIC 
I40E_10G-10G_BASE_T_X722, and other NICs do not have this phenomenon.
 
> >
> > Because it has changed the link status of the ports if it was
> > configured with link speed.
> 
> If we configure to same link speed, will it still have issue ?
> 

Yes, it is.

> >
> > When exiting testpmd that it automatically stops packet forwarding and
> > stops all the ports. But it doesn't update the link status of the
> > ports. If stop the ports first that it will update the link status.
> >
> > This patch fix the error that testpmd will update the link status of
> > the ports when it exits.
> >
> > Fixes: d3a274ce9dee ("app/testpmd: handle SIGINT and SIGTERM")
> > Fixes: 284c908cc588 ("app/testpmd: request device removal interrupt")
> > Cc: sta...@dpdk.org
> >
> > Signed-off-by: Jie Wang 
> > ---
> >   app/test-pmd/testpmd.c | 4 ++--
> >   1 file changed, 2 insertions(+), 2 deletions(-)
> >
> > diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c index
> > 97adafacd0..c348a3f328 100644
> > --- a/app/test-pmd/testpmd.c
> > +++ b/app/test-pmd/testpmd.c
> > @@ -3548,7 +3548,7 @@ pmd_test_exit(void)
> > }
> >   #endif
> > if (ports != NULL) {
> > -   no_link_check = 1;
> > +   no_link_check = 0;
> > RTE_ETH_FOREACH_DEV(pt_id) {
> > printf("\nStopping port %d...\n", pt_id);
> > fflush(stdout);
> > @@ -3675,7 +3675,7 @@ rmv_port_callback(void *arg)
> > need_to_start = 1;
> > stop_packet_forwarding();
> > }
> > -   no_link_check = 1;
> > +   no_link_check = 0;
> 
> Well, here we are undoing a previous change done for 284c908cc588 Won't that
> issue come back.
> 
> > stop_port(port_id);
> > no_link_check = org_no_link_check;
> >



[PATCH] net/bonding: fix slave device Rx/Tx offload configuration

2022-10-27 Thread Huisong Li
Normally, the Rx/Tx offload capability of bonding interface is
the intersection of the capability of all slave devices. And
Rx/Tx offloads configuration of slave device comes from bonding
interface. But now there is a risk that slave device retains its
previous offload configurations which is not within the offload
configurations of bond interface.

Fixes: 57b156540f51 ("net/bonding: fix offloading configuration")
Cc: sta...@dpdk.org

Signed-off-by: Huisong Li 
---
 drivers/net/bonding/rte_eth_bond_pmd.c | 17 -
 1 file changed, 4 insertions(+), 13 deletions(-)

diff --git a/drivers/net/bonding/rte_eth_bond_pmd.c 
b/drivers/net/bonding/rte_eth_bond_pmd.c
index dc74852137..ca87490065 100644
--- a/drivers/net/bonding/rte_eth_bond_pmd.c
+++ b/drivers/net/bonding/rte_eth_bond_pmd.c
@@ -1741,20 +1741,11 @@ slave_configure(struct rte_eth_dev *bonded_eth_dev,
slave_eth_dev->data->dev_conf.link_speeds =
bonded_eth_dev->data->dev_conf.link_speeds;
 
-   slave_eth_dev->data->dev_conf.txmode.offloads |=
-   bonded_eth_dev->data->dev_conf.txmode.offloads;
-
-   slave_eth_dev->data->dev_conf.txmode.offloads &=
-   (bonded_eth_dev->data->dev_conf.txmode.offloads |
-   ~internals->tx_offload_capa);
-
-   slave_eth_dev->data->dev_conf.rxmode.offloads |=
-   bonded_eth_dev->data->dev_conf.rxmode.offloads;
-
-   slave_eth_dev->data->dev_conf.rxmode.offloads &=
-   (bonded_eth_dev->data->dev_conf.rxmode.offloads |
-   ~internals->rx_offload_capa);
+   slave_eth_dev->data->dev_conf.txmode.offloads =
+   bonded_eth_dev->data->dev_conf.txmode.offloads;
 
+   slave_eth_dev->data->dev_conf.rxmode.offloads =
+   bonded_eth_dev->data->dev_conf.rxmode.offloads;
 
nb_rx_queues = bonded_eth_dev->data->nb_rx_queues;
nb_tx_queues = bonded_eth_dev->data->nb_tx_queues;
-- 
2.33.0



[PATCH v2] net/ice: fix rx scalar path offload parse

2022-10-27 Thread Mingjin Ye
The status_0 field of the rx descriptor is incorrectly parsed as the error
field, resulting in a parsing error of offload features.

This patch fixes parsing of wrong fields.

Fixes: daa02b5cddbb ("mbuf: add namespace to offload flags")
Fixes: dbf3c0e77a22 ("net/ice: handle Rx flex descriptor")
Cc: sta...@dpdk.org

Signed-off-by: Mingjin Ye 
---
 drivers/net/ice/ice_rxtx.c | 20 ++--
 1 file changed, 10 insertions(+), 10 deletions(-)

diff --git a/drivers/net/ice/ice_rxtx.c b/drivers/net/ice/ice_rxtx.c
index 71e5c6f5d6..3c558b32bd 100644
--- a/drivers/net/ice/ice_rxtx.c
+++ b/drivers/net/ice/ice_rxtx.c
@@ -1485,36 +1485,36 @@ ice_rx_queue_count(void *rx_queue)
 (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EIPE_S) |  \
 (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EUDPE_S) | \
 (1 << ICE_RX_FLEX_DESC_STATUS0_RXE_S))
-
+#define ICE_RXD_QW1_ERROR_SHIFT19
 /* Rx L3/L4 checksum */
 static inline uint64_t
 ice_rxd_error_to_pkt_flags(uint16_t stat_err0)
 {
uint64_t flags = 0;
 
-   /* check if HW has decoded the packet and checksum */
-   if (unlikely(!(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_L3L4P_S
-   return 0;
+   uint64_t error_bits = (stat_err0 >> ICE_RXD_QW1_ERROR_SHIFT) & 0x7D;
 
-   if (likely(!(stat_err0 & ICE_RX_FLEX_ERR0_BITS))) {
-   flags |= (RTE_MBUF_F_RX_IP_CKSUM_GOOD | 
RTE_MBUF_F_RX_L4_CKSUM_GOOD);
+   if (likely(!(error_bits & ICE_RX_FLEX_ERR0_BITS))) {
+   flags |= (RTE_MBUF_F_RX_IP_CKSUM_GOOD |
+   RTE_MBUF_F_RX_L4_CKSUM_GOOD |
+   RTE_MBUF_F_RX_OUTER_L4_CKSUM_GOOD);
return flags;
}
 
-   if (unlikely(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_IPE_S)))
+   if (unlikely(error_bits & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_IPE_S)))
flags |= RTE_MBUF_F_RX_IP_CKSUM_BAD;
else
flags |= RTE_MBUF_F_RX_IP_CKSUM_GOOD;
 
-   if (unlikely(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_L4E_S)))
+   if (unlikely(error_bits & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_L4E_S)))
flags |= RTE_MBUF_F_RX_L4_CKSUM_BAD;
else
flags |= RTE_MBUF_F_RX_L4_CKSUM_GOOD;
 
-   if (unlikely(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EIPE_S)))
+   if (unlikely(error_bits & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EIPE_S)))
flags |= RTE_MBUF_F_RX_OUTER_IP_CKSUM_BAD;
 
-   if (unlikely(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EUDPE_S)))
+   if (unlikely(error_bits & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EUDPE_S)))
flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_BAD;
else
flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_GOOD;
-- 
2.34.1



[PATCH v2] net/ice: fix rx scalar path offload parse

2022-10-27 Thread Mingjin Ye
The status_0 field of the rx descriptor is incorrectly parsed as the error
field, which causes the parsing error of offload features.

This patch is to fixe the parsing of wrong fields.

Fixes: daa02b5cddbb ("mbuf: add namespace to offload flags")
Fixes: dbf3c0e77a22 ("net/ice: handle Rx flex descriptor")
Cc: sta...@dpdk.org

Signed-off-by: Mingjin Ye 
---
 drivers/net/ice/ice_rxtx.c | 20 ++--
 1 file changed, 10 insertions(+), 10 deletions(-)

diff --git a/drivers/net/ice/ice_rxtx.c b/drivers/net/ice/ice_rxtx.c
index 71e5c6f5d6..3c558b32bd 100644
--- a/drivers/net/ice/ice_rxtx.c
+++ b/drivers/net/ice/ice_rxtx.c
@@ -1485,36 +1485,36 @@ ice_rx_queue_count(void *rx_queue)
 (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EIPE_S) |  \
 (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EUDPE_S) | \
 (1 << ICE_RX_FLEX_DESC_STATUS0_RXE_S))
-
+#define ICE_RXD_QW1_ERROR_SHIFT19
 /* Rx L3/L4 checksum */
 static inline uint64_t
 ice_rxd_error_to_pkt_flags(uint16_t stat_err0)
 {
uint64_t flags = 0;
 
-   /* check if HW has decoded the packet and checksum */
-   if (unlikely(!(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_L3L4P_S
-   return 0;
+   uint64_t error_bits = (stat_err0 >> ICE_RXD_QW1_ERROR_SHIFT) & 0x7D;
 
-   if (likely(!(stat_err0 & ICE_RX_FLEX_ERR0_BITS))) {
-   flags |= (RTE_MBUF_F_RX_IP_CKSUM_GOOD | 
RTE_MBUF_F_RX_L4_CKSUM_GOOD);
+   if (likely(!(error_bits & ICE_RX_FLEX_ERR0_BITS))) {
+   flags |= (RTE_MBUF_F_RX_IP_CKSUM_GOOD |
+   RTE_MBUF_F_RX_L4_CKSUM_GOOD |
+   RTE_MBUF_F_RX_OUTER_L4_CKSUM_GOOD);
return flags;
}
 
-   if (unlikely(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_IPE_S)))
+   if (unlikely(error_bits & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_IPE_S)))
flags |= RTE_MBUF_F_RX_IP_CKSUM_BAD;
else
flags |= RTE_MBUF_F_RX_IP_CKSUM_GOOD;
 
-   if (unlikely(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_L4E_S)))
+   if (unlikely(error_bits & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_L4E_S)))
flags |= RTE_MBUF_F_RX_L4_CKSUM_BAD;
else
flags |= RTE_MBUF_F_RX_L4_CKSUM_GOOD;
 
-   if (unlikely(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EIPE_S)))
+   if (unlikely(error_bits & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EIPE_S)))
flags |= RTE_MBUF_F_RX_OUTER_IP_CKSUM_BAD;
 
-   if (unlikely(stat_err0 & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EUDPE_S)))
+   if (unlikely(error_bits & (1 << ICE_RX_FLEX_DESC_STATUS0_XSUM_EUDPE_S)))
flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_BAD;
else
flags |= RTE_MBUF_F_RX_OUTER_L4_CKSUM_GOOD;
-- 
2.34.1



[PATCH] app/testpmd: update bond port configurations when add slave

2022-10-27 Thread Huisong Li
Some capabilities (like, rx_offload_capa and tx_offload_capa) of
bonding device in dev_info is zero when no slave is added. And its
capability will be updated when add a new slave device.

The capability to update dynamically can introduce some probelms if
not handled properly. For example, the reconfig() is called to initialize
bonding port configurations when create a bonding device. The global
tx_mode is assigned to dev_conf.txmode. The DEV_TX_OFFLOAD_MBUF_FAST_FREE
which is the default value of global tx_mode.offloads in testpmd is removed
from bonding device configuration because of zero rx_offload_capa.
As a result, this offload isn't set to bonding device.

Generally, port configurations of bonding device must be within the
intersection of the capability of all slave devices. If use original port
configurations, the removed capabilities because of adding a new slave may
cause failure when re-initialize bonding device.

So port configurations of bonding device also need to be updated because of
the added and removed capabilities. In addition, this also helps to ensure
consistency between testpmd and bonding device.

Signed-off-by: Huisong Li 
---
 app/test-pmd/testpmd.c| 40 +++
 app/test-pmd/testpmd.h|  3 +-
 drivers/net/bonding/bonding_testpmd.c |  2 ++
 3 files changed, 44 insertions(+), 1 deletion(-)

diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c
index 97adafacd0..7324b8865c 100644
--- a/app/test-pmd/testpmd.c
+++ b/app/test-pmd/testpmd.c
@@ -2805,6 +2805,41 @@ fill_xstats_display_info(void)
fill_xstats_display_info_for_port(pi);
 }
 
+/*
+ * Some capabilities (like, rx_offload_capa and tx_offload_capa) of bonding
+ * device in dev_info is zero when no slave is added. And its capability of
+ * will be updated when add a new slave device. So adding a device slave need
+ * to update the port configurations of bonding device.
+ */
+static void
+update_bonding_port_dev_conf(portid_t bond_pid)
+{
+#ifdef RTE_NET_BOND
+   struct rte_port *port = &ports[bond_pid];
+   uint16_t i;
+   int ret;
+
+   ret = eth_dev_info_get_print_err(bond_pid, &port->dev_info);
+   if (ret != 0) {
+   fprintf(stderr, "Failed to get dev info for port = %u\n",
+   bond_pid);
+   return;
+   }
+
+   if (port->dev_info.tx_offload_capa & RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE)
+   port->dev_conf.txmode.offloads |=
+   RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE;
+   /* Apply Tx offloads configuration */
+   for (i = 0; i < port->dev_info.max_tx_queues; i++)
+   port->txq[i].conf.offloads = port->dev_conf.txmode.offloads;
+
+   port->dev_conf.rx_adv_conf.rss_conf.rss_hf &=
+   port->dev_info.flow_type_rss_offloads;
+#else
+   RTE_SET_USED(bond_pid);
+#endif
+}
+
 int
 start_port(portid_t pid)
 {
@@ -2869,6 +2904,11 @@ start_port(portid_t pid)
return -1;
}
 
+   if (port->bond_flag == 1 && port->update_conf == 1) {
+   update_bonding_port_dev_conf(pi);
+   port->update_conf = 0;
+   }
+
/* configure port */
diag = eth_dev_configure_mp(pi, nb_rxq + nb_hairpinq,
 nb_txq + nb_hairpinq,
diff --git a/app/test-pmd/testpmd.h b/app/test-pmd/testpmd.h
index 7fef96f9b1..82714119e8 100644
--- a/app/test-pmd/testpmd.h
+++ b/app/test-pmd/testpmd.h
@@ -316,7 +316,8 @@ struct rte_port {
queueid_t   queue_nb; /**< nb. of queues for flow rules */
uint32_tqueue_sz; /**< size of a queue for flow rules */
uint8_t slave_flag : 1, /**< bonding slave port */
-   bond_flag : 1; /**< port is bond device */
+   bond_flag : 1, /**< port is bond device */
+   update_conf : 1; /**< need to update bonding 
device configuration */
struct port_template*pattern_templ_list; /**< Pattern templates. */
struct port_template*actions_templ_list; /**< Actions templates. */
struct port_table   *table_list; /**< Flow tables. */
diff --git a/drivers/net/bonding/bonding_testpmd.c 
b/drivers/net/bonding/bonding_testpmd.c
index 3941f4cf23..9529e16fb6 100644
--- a/drivers/net/bonding/bonding_testpmd.c
+++ b/drivers/net/bonding/bonding_testpmd.c
@@ -625,6 +625,7 @@ static void cmd_add_bonding_slave_parsed(void 
*parsed_result,
slave_port_id, master_port_id);
return;
}
+   ports[master_port_id].update_conf = 1;
init_port_config();
set_port_slave_flag(slave_port_id);
 }
@@ -762,6 +763,7 @@ static void cmd_create_bonded_device_parsed(void 
*parsed

[Bug 1101] [dpdk-22.11.0rc1][meson test] driver-tests/eventdev_selftest_sw test failed

2022-10-27 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1101

jiang,yu (yux.ji...@intel.com) changed:

   What|Removed |Added

 CC||yux.ji...@intel.com
 Status|UNCONFIRMED |RESOLVED
 Resolution|--- |FIXED

--- Comment #7 from jiang,yu (yux.ji...@intel.com) ---
Fixed in
http://git.dpdk.org/dpdk/commit/?id=ab059e82e12f2a9dc7561960004819de68b37110

-- 
You are receiving this mail because:
You are the assignee for the bug.

[PATCH V2] app/testpmd: update bond port configurations when add slave

2022-10-27 Thread Huisong Li
Some capabilities (like, rx_offload_capa and tx_offload_capa) of bonding
device in dev_info is zero when no slave is added. And its capability will
be updated when add a new slave device.

The capability to update dynamically may introduce some problems if not
handled properly. For example, the reconfig() is called to initialize
bonding port configurations when create a bonding device. The global
tx_mode is assigned to dev_conf.txmode. The DEV_TX_OFFLOAD_MBUF_FAST_FREE
which is the default value of global tx_mode.offloads in testpmd is removed
from bonding device configuration because of zero rx_offload_capa.
As a result, this offload isn't set to bonding device.

Generally, port configurations of bonding device must be within the
intersection of the capability of all slave devices. If use original port
configurations, the removed capabilities because of adding a new slave may
cause failure when re-initialize bonding device.

So port configurations of bonding device also need to be updated because of
the added and removed capabilities. In addition, this also helps to ensure
consistency between testpmd and bonding device.

Signed-off-by: Huisong Li 
---
 - v2: fix a spelling error in commit log

---
 app/test-pmd/testpmd.c| 40 +++
 app/test-pmd/testpmd.h|  3 +-
 drivers/net/bonding/bonding_testpmd.c |  2 ++
 3 files changed, 44 insertions(+), 1 deletion(-)

diff --git a/app/test-pmd/testpmd.c b/app/test-pmd/testpmd.c
index 97adafacd0..7324b8865c 100644
--- a/app/test-pmd/testpmd.c
+++ b/app/test-pmd/testpmd.c
@@ -2805,6 +2805,41 @@ fill_xstats_display_info(void)
fill_xstats_display_info_for_port(pi);
 }
 
+/*
+ * Some capabilities (like, rx_offload_capa and tx_offload_capa) of bonding
+ * device in dev_info is zero when no slave is added. And its capability of
+ * will be updated when add a new slave device. So adding a device slave need
+ * to update the port configurations of bonding device.
+ */
+static void
+update_bonding_port_dev_conf(portid_t bond_pid)
+{
+#ifdef RTE_NET_BOND
+   struct rte_port *port = &ports[bond_pid];
+   uint16_t i;
+   int ret;
+
+   ret = eth_dev_info_get_print_err(bond_pid, &port->dev_info);
+   if (ret != 0) {
+   fprintf(stderr, "Failed to get dev info for port = %u\n",
+   bond_pid);
+   return;
+   }
+
+   if (port->dev_info.tx_offload_capa & RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE)
+   port->dev_conf.txmode.offloads |=
+   RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE;
+   /* Apply Tx offloads configuration */
+   for (i = 0; i < port->dev_info.max_tx_queues; i++)
+   port->txq[i].conf.offloads = port->dev_conf.txmode.offloads;
+
+   port->dev_conf.rx_adv_conf.rss_conf.rss_hf &=
+   port->dev_info.flow_type_rss_offloads;
+#else
+   RTE_SET_USED(bond_pid);
+#endif
+}
+
 int
 start_port(portid_t pid)
 {
@@ -2869,6 +2904,11 @@ start_port(portid_t pid)
return -1;
}
 
+   if (port->bond_flag == 1 && port->update_conf == 1) {
+   update_bonding_port_dev_conf(pi);
+   port->update_conf = 0;
+   }
+
/* configure port */
diag = eth_dev_configure_mp(pi, nb_rxq + nb_hairpinq,
 nb_txq + nb_hairpinq,
diff --git a/app/test-pmd/testpmd.h b/app/test-pmd/testpmd.h
index 7fef96f9b1..82714119e8 100644
--- a/app/test-pmd/testpmd.h
+++ b/app/test-pmd/testpmd.h
@@ -316,7 +316,8 @@ struct rte_port {
queueid_t   queue_nb; /**< nb. of queues for flow rules */
uint32_tqueue_sz; /**< size of a queue for flow rules */
uint8_t slave_flag : 1, /**< bonding slave port */
-   bond_flag : 1; /**< port is bond device */
+   bond_flag : 1, /**< port is bond device */
+   update_conf : 1; /**< need to update bonding 
device configuration */
struct port_template*pattern_templ_list; /**< Pattern templates. */
struct port_template*actions_templ_list; /**< Actions templates. */
struct port_table   *table_list; /**< Flow tables. */
diff --git a/drivers/net/bonding/bonding_testpmd.c 
b/drivers/net/bonding/bonding_testpmd.c
index 3941f4cf23..9529e16fb6 100644
--- a/drivers/net/bonding/bonding_testpmd.c
+++ b/drivers/net/bonding/bonding_testpmd.c
@@ -625,6 +625,7 @@ static void cmd_add_bonding_slave_parsed(void 
*parsed_result,
slave_port_id, master_port_id);
return;
}
+   ports[master_port_id].update_conf = 1;
init_port_config();
set_port_slave_flag(slave_port_id);
 }
@@ -762,6 +763,7 @@ static voi

[PATCH v3 1/2] mempool: cache align mempool cache objects

2022-10-27 Thread Morten Brørup
Add __rte_cache_aligned to the objs array.

It makes no difference in the general case, but if get/put operations are
always 32 objects, it will reduce the number of memory (or last level
cache) accesses from five to four 64 B cache lines for every get/put
operation.

For readability reasons, an example using 16 objects follows:

Currently, with 16 objects (128B), we access to 3
cache lines:

  ┌┐
  │len │
cache ││---
line0 ││ ^
  ││ |
  ├┤ | 16 objects
  ││ | 128B
cache ││ |
line1 ││ |
  ││ |
  ├┤ |
  ││_v_
cache ││
line2 ││
  ││
  └┘

With the alignment, it is also 3 cache lines:

  ┌┐
  │len │
cache ││
line0 ││
  ││
  ├┤---
  ││ ^
cache ││ |
line1 ││ |
  ││ |
  ├┤ | 16 objects
  ││ | 128B
cache ││ |
line2 ││ |
  ││ v
  └┘---

However, accessing the objects at the bottom of the mempool cache is a
special case, where cache line0 is also used for objects.

Consider the next burst (and any following bursts):

Current:
  ┌┐
  │len │
cache ││
line0 ││
  ││
  ├┤
  ││
cache ││
line1 ││
  ││
  ├┤
  ││
cache ││---
line2 ││ ^
  ││ |
  ├┤ | 16 objects
  ││ | 128B
cache ││ |
line3 ││ |
  ││ |
  ├┤ |
  ││_v_
cache ││
line4 ││
  ││
  └┘
4 cache lines touched, incl. line0 for len.

With the proposed alignment:
  ┌┐
  │len │
cache ││
line0 ││
  ││
  ├┤
  ││
cache ││
line1 ││
  ││
  ├┤
  ││
cache ││
line2 ││
  ││
  ├┤
  ││---
cache ││ ^
line3 ││ |
  ││ | 16 objects
  ├┤ | 128B
  ││ |
cache ││ |
line4 ││ |
  ││_v_
  └┘
Only 3 cache lines touched, incl. line0 for len.

Credits go to Olivier Matz for the nice ASCII graphics.

v3:
* No changes. Made part of a series.
v2:
* No such version.

Signed-off-by: Morten Brørup 
---
 lib/mempool/rte_mempool.h | 6 --
 1 file changed, 4 insertions(+), 2 deletions(-)

diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 1f5707f46a..3725a72951 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -86,11 +86,13 @@ struct rte_mempool_cache {
uint32_t size;/**< Size of the cache */
uint32_t flushthresh; /**< Threshold before we flush excess elements */
uint32_t len; /**< Current cache count */
-   /*
+   /**
+* Cache objects
+*
 * Cache is allocated to this size to allow it to overflow in certain
 * cases to avoid needless emptying of cache.
 */
-   void *objs[RTE_MEMPOOL_CACHE_MAX_SIZE * 2]; /**< Cache objects */
+   void *objs[RTE_MEMPOOL_CACHE_MAX_SIZE * 2] __rte_cache_aligned;
 } __rte_cache_aligned;
 
 /**
-- 
2.17.1



[PATCH v3 2/2] mempool: optimized debug statistics

2022-10-27 Thread Morten Brørup
When built with debug enabled (RTE_LIBRTE_MEMPOOL_DEBUG defined), the
performance of mempools with caches is improved as follows.

Accessing objects in the mempool is likely to increment either the
put_bulk and put_objs or the get_success_bulk and get_success_objs
debug statistics counters.

By adding an alternative set of these counters to the mempool cache
structure, accesing the dedicated debug statistics structure is avoided in
the likely cases where these counters are incremented.

The trick here is that the cache line holding the mempool cache structure
is accessed anyway, in order to update the "len" field. Updating some
debug statistics counters in the same cache line has lower performance
cost than accessing the debug statistics counters in the dedicated debug
statistics structure, i.e. in another cache line.

Running mempool_perf_autotest on a VMware virtual server shows an avg.
increase of 6.4 % in rate_persec for the tests with cache. (Only when
built with with debug enabled, obviously!)

For the tests without cache, the avg. increase in rate_persec is 0.8 %. I
assume this is noise from the test environment.

v3:
* Try to fix git reference by making part of a series.
* Add --in-reply-to v1 when sending email.
v2:
* Fix spelling and repeated word in commit message, caught by checkpatch.

Signed-off-by: Morten Brørup 
---
 lib/mempool/rte_mempool.c |  7 +
 lib/mempool/rte_mempool.h | 55 +++
 2 files changed, 51 insertions(+), 11 deletions(-)

diff --git a/lib/mempool/rte_mempool.c b/lib/mempool/rte_mempool.c
index 21c94a2b9f..7b8c00a022 100644
--- a/lib/mempool/rte_mempool.c
+++ b/lib/mempool/rte_mempool.c
@@ -1285,6 +1285,13 @@ rte_mempool_dump(FILE *f, struct rte_mempool *mp)
sum.get_fail_objs += mp->stats[lcore_id].get_fail_objs;
sum.get_success_blks += mp->stats[lcore_id].get_success_blks;
sum.get_fail_blks += mp->stats[lcore_id].get_fail_blks;
+   /* Add the fast access statistics, if local caches exist */
+   if (mp->cache_size != 0) {
+   sum.put_bulk += mp->local_cache[lcore_id].put_bulk;
+   sum.put_objs += mp->local_cache[lcore_id].put_objs;
+   sum.get_success_bulk += 
mp->local_cache[lcore_id].get_success_bulk;
+   sum.get_success_objs += 
mp->local_cache[lcore_id].get_success_objs;
+   }
}
fprintf(f, "  stats:\n");
fprintf(f, "put_bulk=%"PRIu64"\n", sum.put_bulk);
diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 3725a72951..d84087bc92 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -86,6 +86,14 @@ struct rte_mempool_cache {
uint32_t size;/**< Size of the cache */
uint32_t flushthresh; /**< Threshold before we flush excess elements */
uint32_t len; /**< Current cache count */
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+   uint32_t unused;
+   /* Fast access statistics, only for likely events */
+   uint64_t put_bulk; /**< Number of puts. */
+   uint64_t put_objs; /**< Number of objects successfully put. 
*/
+   uint64_t get_success_bulk; /**< Successful allocation number. */
+   uint64_t get_success_objs; /**< Objects successfully allocated. */
+#endif
/**
 * Cache objects
 *
@@ -1327,13 +1335,19 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void 
* const *obj_table,
 {
void **cache_objs;
 
+   /* No cache provided */
+   if (unlikely(cache == NULL))
+   goto driver_enqueue;
+
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
/* increment stat now, adding in mempool always success */
-   RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
-   RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
+   cache->put_bulk += 1;
+   cache->put_objs += n;
+#endif
 
-   /* No cache provided or the request itself is too big for the cache */
-   if (unlikely(cache == NULL || n > cache->flushthresh))
-   goto driver_enqueue;
+   /* The request is too big for the cache */
+   if (unlikely(n > cache->flushthresh))
+   goto driver_enqueue_stats_incremented;
 
/*
 * The cache follows the following algorithm:
@@ -1358,6 +1372,12 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void 
* const *obj_table,
 
 driver_enqueue:
 
+   /* increment stat now, adding in mempool always success */
+   RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
+   RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
+
+driver_enqueue_stats_incremented:
+
/* push objects to the backend */
rte_mempool_ops_enqueue_bulk(mp, obj_table, n);
 }
@@ -1464,8 +1484,10 @@ rte_mempool_do_generic_get(struct rte_mempool *mp, void 
**obj_table,
if (remaining == 0) {
/* The entire request is satisfied from the cache. */
 
-   RTE_MEMPOOL_STAT_A

[PATCH v4 1/2] mempool: cache align mempool cache objects

2022-10-27 Thread Morten Brørup
Add __rte_cache_aligned to the objs array.

It makes no difference in the general case, but if get/put operations are
always 32 objects, it will reduce the number of memory (or last level
cache) accesses from five to four 64 B cache lines for every get/put
operation.

For readability reasons, an example using 16 objects follows:

Currently, with 16 objects (128B), we access to 3
cache lines:

  ┌┐
  │len │
cache ││---
line0 ││ ^
  ││ |
  ├┤ | 16 objects
  ││ | 128B
cache ││ |
line1 ││ |
  ││ |
  ├┤ |
  ││_v_
cache ││
line2 ││
  ││
  └┘

With the alignment, it is also 3 cache lines:

  ┌┐
  │len │
cache ││
line0 ││
  ││
  ├┤---
  ││ ^
cache ││ |
line1 ││ |
  ││ |
  ├┤ | 16 objects
  ││ | 128B
cache ││ |
line2 ││ |
  ││ v
  └┘---

However, accessing the objects at the bottom of the mempool cache is a
special case, where cache line0 is also used for objects.

Consider the next burst (and any following bursts):

Current:
  ┌┐
  │len │
cache ││
line0 ││
  ││
  ├┤
  ││
cache ││
line1 ││
  ││
  ├┤
  ││
cache ││---
line2 ││ ^
  ││ |
  ├┤ | 16 objects
  ││ | 128B
cache ││ |
line3 ││ |
  ││ |
  ├┤ |
  ││_v_
cache ││
line4 ││
  ││
  └┘
4 cache lines touched, incl. line0 for len.

With the proposed alignment:
  ┌┐
  │len │
cache ││
line0 ││
  ││
  ├┤
  ││
cache ││
line1 ││
  ││
  ├┤
  ││
cache ││
line2 ││
  ││
  ├┤
  ││---
cache ││ ^
line3 ││ |
  ││ | 16 objects
  ├┤ | 128B
  ││ |
cache ││ |
line4 ││ |
  ││_v_
  └┘
Only 3 cache lines touched, incl. line0 for len.

Credits go to Olivier Matz for the nice ASCII graphics.

v4:
* No changes. Added reviewed- and acked-by tags.
v3:
* No changes. Made part of a series.
v2:
* No such version.

Signed-off-by: Morten Brørup 
Reviewed-by: Andrew Rybchenko 
Acked-by: Olivier Matz 
---
 lib/mempool/rte_mempool.h | 6 --
 1 file changed, 4 insertions(+), 2 deletions(-)

diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 1f5707f46a..3725a72951 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -86,11 +86,13 @@ struct rte_mempool_cache {
uint32_t size;/**< Size of the cache */
uint32_t flushthresh; /**< Threshold before we flush excess elements */
uint32_t len; /**< Current cache count */
-   /*
+   /**
+* Cache objects
+*
 * Cache is allocated to this size to allow it to overflow in certain
 * cases to avoid needless emptying of cache.
 */
-   void *objs[RTE_MEMPOOL_CACHE_MAX_SIZE * 2]; /**< Cache objects */
+   void *objs[RTE_MEMPOOL_CACHE_MAX_SIZE * 2] __rte_cache_aligned;
 } __rte_cache_aligned;
 
 /**
-- 
2.17.1



[PATCH v4 2/2] mempool: optimized debug statistics

2022-10-27 Thread Morten Brørup
When built with debug enabled (RTE_LIBRTE_MEMPOOL_DEBUG defined), the
performance of mempools with caches is improved as follows.

Accessing objects in the mempool is likely to increment either the
put_bulk and put_objs or the get_success_bulk and get_success_objs
debug statistics counters.

By adding an alternative set of these counters to the mempool cache
structure, accessing the dedicated debug statistics structure is avoided in
the likely cases where these counters are incremented.

The trick here is that the cache line holding the mempool cache structure
is accessed anyway, in order to update the "len" field. Updating some
debug statistics counters in the same cache line has lower performance
cost than accessing the debug statistics counters in the dedicated debug
statistics structure, i.e. in another cache line.

Running mempool_perf_autotest on a VMware virtual server shows an avg.
increase of 6.4 % in rate_persec for the tests with cache. (Only when
built with debug enabled, obviously!)

For the tests without cache, the avg. increase in rate_persec is 0.8 %. I
assume this is noise from the test environment.

v4:
* Fix spelling and repeated word in commit message, caught by checkpatch.
v3:
* Try to fix git reference by making part of a series.
* Add --in-reply-to v1 when sending email.
v2:
* Fix spelling and repeated word in commit message, caught by checkpatch.

Signed-off-by: Morten Brørup 
---
 lib/mempool/rte_mempool.c |  7 +
 lib/mempool/rte_mempool.h | 55 +++
 2 files changed, 51 insertions(+), 11 deletions(-)

diff --git a/lib/mempool/rte_mempool.c b/lib/mempool/rte_mempool.c
index 21c94a2b9f..7b8c00a022 100644
--- a/lib/mempool/rte_mempool.c
+++ b/lib/mempool/rte_mempool.c
@@ -1285,6 +1285,13 @@ rte_mempool_dump(FILE *f, struct rte_mempool *mp)
sum.get_fail_objs += mp->stats[lcore_id].get_fail_objs;
sum.get_success_blks += mp->stats[lcore_id].get_success_blks;
sum.get_fail_blks += mp->stats[lcore_id].get_fail_blks;
+   /* Add the fast access statistics, if local caches exist */
+   if (mp->cache_size != 0) {
+   sum.put_bulk += mp->local_cache[lcore_id].put_bulk;
+   sum.put_objs += mp->local_cache[lcore_id].put_objs;
+   sum.get_success_bulk += 
mp->local_cache[lcore_id].get_success_bulk;
+   sum.get_success_objs += 
mp->local_cache[lcore_id].get_success_objs;
+   }
}
fprintf(f, "  stats:\n");
fprintf(f, "put_bulk=%"PRIu64"\n", sum.put_bulk);
diff --git a/lib/mempool/rte_mempool.h b/lib/mempool/rte_mempool.h
index 3725a72951..d84087bc92 100644
--- a/lib/mempool/rte_mempool.h
+++ b/lib/mempool/rte_mempool.h
@@ -86,6 +86,14 @@ struct rte_mempool_cache {
uint32_t size;/**< Size of the cache */
uint32_t flushthresh; /**< Threshold before we flush excess elements */
uint32_t len; /**< Current cache count */
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
+   uint32_t unused;
+   /* Fast access statistics, only for likely events */
+   uint64_t put_bulk; /**< Number of puts. */
+   uint64_t put_objs; /**< Number of objects successfully put. 
*/
+   uint64_t get_success_bulk; /**< Successful allocation number. */
+   uint64_t get_success_objs; /**< Objects successfully allocated. */
+#endif
/**
 * Cache objects
 *
@@ -1327,13 +1335,19 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void 
* const *obj_table,
 {
void **cache_objs;
 
+   /* No cache provided */
+   if (unlikely(cache == NULL))
+   goto driver_enqueue;
+
+#ifdef RTE_LIBRTE_MEMPOOL_DEBUG
/* increment stat now, adding in mempool always success */
-   RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
-   RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
+   cache->put_bulk += 1;
+   cache->put_objs += n;
+#endif
 
-   /* No cache provided or the request itself is too big for the cache */
-   if (unlikely(cache == NULL || n > cache->flushthresh))
-   goto driver_enqueue;
+   /* The request is too big for the cache */
+   if (unlikely(n > cache->flushthresh))
+   goto driver_enqueue_stats_incremented;
 
/*
 * The cache follows the following algorithm:
@@ -1358,6 +1372,12 @@ rte_mempool_do_generic_put(struct rte_mempool *mp, void 
* const *obj_table,
 
 driver_enqueue:
 
+   /* increment stat now, adding in mempool always success */
+   RTE_MEMPOOL_STAT_ADD(mp, put_bulk, 1);
+   RTE_MEMPOOL_STAT_ADD(mp, put_objs, n);
+
+driver_enqueue_stats_incremented:
+
/* push objects to the backend */
rte_mempool_ops_enqueue_bulk(mp, obj_table, n);
 }
@@ -1464,8 +1484,10 @@ rte_mempool_do_generic_get(struct rte_mempool *mp, void 
**obj_table,
if (remaining == 0) {
/* The entire re

Re: [PATCH v4 1/2] build: allow to conditionally build apps

2022-10-27 Thread Markus Theil

On 10/27/22 16:22, David Marchand wrote:

On Thu, Oct 27, 2022 at 3:13 PM David Marchand
 wrote:

On Fri, Oct 14, 2022 at 9:51 AM Markus Theil  wrote:

Makes apps configurable from meson, like already
possible for drivers.

Signed-off-by: Markus Theil 

With these new options, it's hard to tell which and why some
application is built (or not).
Can we have an output similar to libraries and drivers (see toplevel
meson.build) ?

WDYT of followup patch: https://github.com/david-marchand/dpdk/commit/apps



LGTM


Re: [PATCH] telemetry: make usable from C++

2022-10-27 Thread Markus Theil

On 10/27/22 14:04, David Marchand wrote:

On Wed, Oct 12, 2022 at 2:05 PM Markus Theil  wrote:

From: Markus Theil 

Add missing include in order to make C++ compilers
happy.


We have build checks for headers (-Dcheck_includes meson option), like for C++:

$ cat $BUILDDIR/buildtools/chkincs/chkincs-cpp.p/rte_telemetry.cpp
#include "/home/dmarchan/dpdk/lib/telemetry/rte_telemetry.h"

[917/1289] ccache c++ -Ibuildtools/chkincs/chkincs-cpp.p
-Ibuildtools/chkincs -I../../../dpdk/buildtools/chkincs
-Iexamples/vmdq_dcb -I../../../dpdk/examples/vmdq_dcb
-I../../../dpdk/examples/common -Idrivers/bus/vdev
-I../../../dpdk/drivers/bus/vdev -I. -I../../../dpdk -Iconfig
-I../../../dpdk/config -Ilib/eal/include
-I../../../dpdk/lib/eal/include -Ilib/eal/linux/include
-I../../../dpdk/lib/eal/linux/include -Ilib/eal/x86/include
-I../../../dpdk/lib/eal/x86/include -Ilib/eal/common
-I../../../dpdk/lib/eal/common -Ilib/eal -I../../../dpdk/lib/eal
-Ilib/kvargs -I../../../dpdk/lib/kvargs -Ilib/metrics
-I../../../dpdk/lib/metrics -Ilib/telemetry
-I../../../dpdk/lib/telemetry -Idrivers/bus/pci
-I../../../dpdk/drivers/bus/pci -I../../../dpdk/drivers/bus/pci/linux
-Ilib/pci -I../../../dpdk/lib/pci -Ilib/ring -I../../../dpdk/lib/ring
-Ilib/rcu -I../../../dpdk/lib/rcu -Ilib/mempool
-I../../../dpdk/lib/mempool -Ilib/mbuf -I../../../dpdk/lib/mbuf
-Ilib/net -I../../../dpdk/lib/net -Ilib/meter
-I../../../dpdk/lib/meter -Ilib/ethdev -I../../../dpdk/lib/ethdev
-Ilib/cmdline -I../../../dpdk/lib/cmdline -Ilib/hash
-I../../../dpdk/lib/hash -Ilib/timer -I../../../dpdk/lib/timer
-Ilib/acl -I../../../dpdk/lib/acl -Ilib/bbdev
-I../../../dpdk/lib/bbdev -Ilib/bitratestats
-I../../../dpdk/lib/bitratestats -Ilib/bpf -I../../../dpdk/lib/bpf
-Ilib/cfgfile -I../../../dpdk/lib/cfgfile -Ilib/compressdev
-I../../../dpdk/lib/compressdev -Ilib/cryptodev
-I../../../dpdk/lib/cryptodev -Ilib/distributor
-I../../../dpdk/lib/distributor -Ilib/efd -I../../../dpdk/lib/efd
-Ilib/eventdev -I../../../dpdk/lib/eventdev -Ilib/gpudev
-I../../../dpdk/lib/gpudev -Ilib/gro -I../../../dpdk/lib/gro -Ilib/gso
-I../../../dpdk/lib/gso -Ilib/ip_frag -I../../../dpdk/lib/ip_frag
-Ilib/jobstats -I../../../dpdk/lib/jobstats -Ilib/kni
-I../../../dpdk/lib/kni -Ilib/latencystats
-I../../../dpdk/lib/latencystats -Ilib/lpm -I../../../dpdk/lib/lpm
-Ilib/member -I../../../dpdk/lib/member -Ilib/pcapng
-I../../../dpdk/lib/pcapng -Ilib/power -I../../../dpdk/lib/power
-Ilib/rawdev -I../../../dpdk/lib/rawdev -Ilib/regexdev
-I../../../dpdk/lib/regexdev -Ilib/dmadev -I../../../dpdk/lib/dmadev
-Ilib/rib -I../../../dpdk/lib/rib -Ilib/reorder
-I../../../dpdk/lib/reorder -Ilib/sched -I../../../dpdk/lib/sched
-Ilib/security -I../../../dpdk/lib/security -Ilib/stack
-I../../../dpdk/lib/stack -Ilib/vhost -I../../../dpdk/lib/vhost
-Ilib/ipsec -I../../../dpdk/lib/ipsec -Ilib/fib
-I../../../dpdk/lib/fib -Ilib/port -I../../../dpdk/lib/port
-Ilib/pdump -I../../../dpdk/lib/pdump -Ilib/table
-I../../../dpdk/lib/table -Ilib/pipeline -I../../../dpdk/lib/pipeline
-Ilib/flow_classify -I../../../dpdk/lib/flow_classify -Ilib/graph
-I../../../dpdk/lib/graph -Ilib/node -I../../../dpdk/lib/node
-fdiagnostics-color=always -D_FILE_OFFSET_BITS=64 -Wall -Winvalid-pch
-Wnon-virtual-dtor -Wextra -Werror -O2 -g -include rte_config.h
-march=nehalem -DALLOW_EXPERIMENTAL_API -DALLOW_INTERNAL_API -MD -MQ
buildtools/chkincs/chkincs-cpp.p/meson-generated_rte_telemetry.cpp.o
-MF buildtools/chkincs/chkincs-cpp.p/meson-generated_rte_telemetry.cpp.o.d
-o buildtools/chkincs/chkincs-cpp.p/meson-generated_rte_telemetry.cpp.o
-c buildtools/chkincs/chkincs-cpp.p/rte_telemetry.cpp

Besides, I fail to see the need for stddef.h.

In which setup / case, did you get a compilation issue?


I traced back, that we need this only for some internal patches which 
expose the JSON output of telemetry inside our DPDK application in order 
to get SFF digital optics module data without piping the telemetry 
socket back into our application. It can be therefore ignored and was 
posted by mistake. I'm archiving it.