[PATCH 1/1] net/{octeon_ep,enetfec}: remove unused value

2024-12-18 Thread Ariel Otilibili
Coverity issue: 385414, 374016
Fixes: c836a7ba33e ("net/octeon_ep: support mailbox between VF and PF")
Fixes: bb5b5bf1e5c ("net/enetfec: support queue configuration")
Signed-off-by: Ariel Otilibili 
--
Cc: sta...@dpdk.org
Cc: Vamsi Attunuru 
Cc: Apeksha Gupta 
Cc: Sachin Saxena 
---
 drivers/net/enetfec/enet_ethdev.c   | 1 -
 drivers/net/octeon_ep/otx_ep_mbox.c | 3 ---
 2 files changed, 4 deletions(-)

diff --git a/drivers/net/enetfec/enet_ethdev.c 
b/drivers/net/enetfec/enet_ethdev.c
index 91c0f60490..0d27b63953 100644
--- a/drivers/net/enetfec/enet_ethdev.c
+++ b/drivers/net/enetfec/enet_ethdev.c
@@ -415,7 +415,6 @@ enetfec_tx_queue_setup(struct rte_eth_dev *dev,
offset_des_active_txq[queue_idx];
bd_base = (struct bufdesc *)(((uintptr_t)bd_base) + size);
txq->bd.last = (struct bufdesc *)(((uintptr_t)bd_base) - dsize);
-   bdp = txq->bd.base;
bdp = txq->bd.cur;
 
for (i = 0; i < txq->bd.ring_size; i++) {
diff --git a/drivers/net/octeon_ep/otx_ep_mbox.c 
b/drivers/net/octeon_ep/otx_ep_mbox.c
index 64a51c1fd0..1d7e08d2cc 100644
--- a/drivers/net/octeon_ep/otx_ep_mbox.c
+++ b/drivers/net/octeon_ep/otx_ep_mbox.c
@@ -256,9 +256,6 @@ int otx_ep_mbox_get_link_info(struct rte_eth_dev *eth_dev,
}
link->link_status = RTE_ETH_LINK_UP;
link->link_duplex = RTE_ETH_LINK_FULL_DUPLEX;
-   link->link_autoneg = (link_info.autoneg ==
- OTX_EP_LINK_AUTONEG) ? RTE_ETH_LINK_AUTONEG : 
RTE_ETH_LINK_FIXED;
-
link->link_autoneg = link_info.autoneg;
link->link_speed = link_info.speed;
return 0;
-- 
2.47.1



[PATCH 0/1] net/{octeon_ep,enetfec}: remove unused value

2024-12-18 Thread Ariel Otilibili
Hello,

This patch clears out the Coverity issues 385414 & 374016.

Thank you,

Ariel Otilibili (1):
  net/{octeon_ep,enetfec}: remove unused value

 drivers/net/enetfec/enet_ethdev.c   | 1 -
 drivers/net/octeon_ep/otx_ep_mbox.c | 3 ---
 2 files changed, 4 deletions(-)

-- 
2.47.1



RE: [PATCH] net/nfp: implement the burst mode get operation

2024-12-18 Thread Chaoyong He
> On Wed, 18 Dec 2024 14:30:36 +0800
> Chaoyong He  wrote:
> 
> > +int
> > +nfp_net_rx_burst_mode_get(struct rte_eth_dev *eth_dev,
> > +   uint16_t queue_id __rte_unused,
> > +   struct rte_eth_burst_mode *mode)
> > +{
> > +   eth_rx_burst_t pkt_burst;
> > +
> > +   pkt_burst = eth_dev->rx_pkt_burst;
> > +   if (pkt_burst == nfp_net_recv_pkts) {
> > +   snprintf(mode->info, RTE_ETH_BURST_MODE_INFO_SIZE,
> "%s",
> > +   "Scalar");
> > +   } else if (pkt_burst == nfp_net_vec_avx2_recv_pkts) {
> > +   snprintf(mode->info, RTE_ETH_BURST_MODE_INFO_SIZE,
> "%s",
> > +   "Vector AVX2");
> > +   } else {
> > +   return -EINVAL;
> > +   }
> > +
> > +   return 0;
> > +}
> > +
> 
> The coccinelle script want to replace that snprintf with strlcpy
> 
> Also don't need {} for one line statement but its ok as is.

Okay, I will send a new version patch.


Re: [PATCH] net/virtio: fix Rx checksum calculation

2024-12-18 Thread Mattias Rönnblom

On 2024-12-18 09:59, Maxime Coquelin wrote:

Hi,

On 12/18/24 08:34, Wangyunjian(wangyunjian,TongTu) wrote:

-Original Message-
From: Maxime Coquelin [mailto:maxime.coque...@redhat.com]
Sent: Tuesday, December 17, 2024 11:33 PM
To: dev@dpdk.org
Cc: Olivier Matz ; Maxime Gouin
; Maxime Coquelin

Subject: [PATCH] net/virtio: fix Rx checksum calculation

From: Olivier Matz 

If hdr->csum_start is larger than packet length, the len argument passed
to rte_raw_cksum_mbuf() overflows and causes a segmentation fault.

Ignore checksum computation in this case.

CVE-2024-11614

Fixes: ca7036b4af3a ("vhost: fix offload flags in Rx path")

Signed-off-by: Maxime Gouin 
Signed-off-by: Olivier Matz 
Reviewed-by: Maxime Coquelin 
---
  lib/vhost/virtio_net.c | 3 +++
  1 file changed, 3 insertions(+)

diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index d764d4bc6a..69901ab3b5 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -2823,6 +2823,9 @@ vhost_dequeue_offload(struct virtio_net *dev,
struct virtio_net_hdr *hdr,
   */
  uint16_t csum = 0, off;

+    if (hdr->csum_start >= rte_pktmbuf_pkt_len(m))
+    return;
+


The hdr->csum_start does two successive reads from user space to read
a variable length data structure. The result overflow if the data 
structure

changes between the two reads.



You don't know if the resulting object code will perform one, two or 
more loads from memory.


If you want to be sure it's exactly one load, you need to go through a 
volatile pointer. This seems like a use case for RTE_READ_ONCE() (which 
doesn't exist).


An alternative is to do a relaxed atomic load, which is really what you 
ask for.


We can prevent double fetch issue by using the temporary variable 
csum_start.




I don't think that makes a difference. The compiler is still free to 
generate object code which load from the same location multiple times.



Right, that's a good catch! The exploitation od this issue seem
difficult though.

We may systematically copy the full header, as we only do it for ones
not contiguous in host VA space.



I think you would need to go through a volatile pointer here as well, to 
make sure the copy actually occur. At least if the target is 
stack-allocated object.



What do you think? Are you willing to contribute a fix?

Thanks,
Maxime


Thanks,
Yunjian


  if (rte_raw_cksum_mbuf(m, hdr->csum_start,
  rte_pktmbuf_pkt_len(m) - hdr->csum_start, &csum) <
0)
  return;
--
2.47.0








[PATCH] bus/fslmc: fix use after rte_free

2024-12-18 Thread Stephen Hemminger
The cleanup loop would derefence the dpio_dev after freeing.
Use TAILQ_FOREACH_SAFE to fix that.
Found by building with sanitizer undefined flag.

Fixes: e55d0494ab98 ("bus/fslmc: support secondary process")
Cc: shreyansh.j...@nxp.com
Cc: sta...@dpdk.org
Signed-off-by: Stephen Hemminger 
---
 drivers/bus/fslmc/portal/dpaa2_hw_dpio.c | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/drivers/bus/fslmc/portal/dpaa2_hw_dpio.c 
b/drivers/bus/fslmc/portal/dpaa2_hw_dpio.c
index 2dfcf7a498..fb7d1968d1 100644
--- a/drivers/bus/fslmc/portal/dpaa2_hw_dpio.c
+++ b/drivers/bus/fslmc/portal/dpaa2_hw_dpio.c
@@ -403,6 +403,7 @@ dpaa2_create_dpio_device(int vdev_fd,
struct rte_dpaa2_device *obj)
 {
struct dpaa2_dpio_dev *dpio_dev = NULL;
+   struct dpaa2_dpio_dev *dpio_tmp;
struct vfio_region_info reg_info = { .argsz = sizeof(reg_info)};
struct qbman_swp_desc p_des;
struct dpio_attr attr;
@@ -588,7 +589,7 @@ dpaa2_create_dpio_device(int vdev_fd,
rte_free(dpio_dev);
 
/* For each element in the list, cleanup */
-   TAILQ_FOREACH(dpio_dev, &dpio_dev_list, next) {
+   RTE_TAILQ_FOREACH_SAFE(dpio_dev, &dpio_dev_list, next, dpio_tmp) {
if (dpio_dev->dpio) {
dpio_disable(dpio_dev->dpio, CMD_PRI_LOW,
dpio_dev->token);
-- 
2.45.2



Re: [PATCH v2] MAINTAINERS: change maintainer for next-net

2024-12-18 Thread Thomas Monjalon
10/12/2024 21:41, Stephen Hemminger:
> Change of maintainers for next-net tree.
> 
> Signed-off-by: Stephen Hemminger 
> Acked-by: Ferruh Yigit 
> ---
>  Next-net Tree
> -M: Ferruh Yigit 
> -M: Andrew Rybchenko 
> +M: Stephen Hemminger 
>  T: git://dpdk.org/next/dpdk-next-net

Applied, thanks all for the help.




Re: [PATCH 0/3] Defer lcore variables allocation

2024-12-18 Thread Mattias Rönnblom

On 2024-12-16 10:49, David Marchand wrote:

On Mon, Dec 16, 2024 at 10:42 AM Burakov, Anatoly
 wrote:


On 12/5/2024 6:57 PM, David Marchand wrote:

As I had reported in rc2, the lcore variables allocation have a
noticeable impact on applications consuming DPDK, even when such
applications does not use DPDK, or use features associated to
some lcore variables.

While the amount has been reduced in a rush before rc2,
there are still cases when the increased memory footprint is noticed
like in scaling tests.
See https://bugs.launchpad.net/ubuntu/+source/dpdk/+bug/2090931


lcore variable allocations in constructor is a bad idea, as the
application consuming DPDK has no control over such allocation:
linking some code does not mean that all of it will be used at runtime.

The general question on whether lcore variables in constructor should
be forbidden, is left to a later discussion.

For now, this series only focus on fixing subsystems using lcore
variables so that those allocations are deferred either in rte_eal_init()
or in the path that does require such lcore variables.




An idle question: would this have any consequences in use case of eal
init -> eal cleanup -> eal init with different arguments?


Hum, interesting question.

I would say that initialising lcore variables in constructors means
that this usecase is broken, since lcore variables are freed in
eal_lcore_var_cleanup().




After rte_eal_cleanup() is called, no DPDK calls may be made.

So, with the current API, there is no such use case to break.




[PATCH] drivers/common: avoid truncation of constant value

2024-12-18 Thread Andre Muezerie
This issue was flagged by MSVC warning below:

drivers\common\idpf\base/virtchnl2.h(269): warning C4309:
'initializing': truncation of constant value

The problem is that 64-bit numbers are initialized in an enum.

The C11 standard states: The expression that defines the value of an
enumeration constant shall be an integer constant expression that
has a value representable as an int. [Section 6.7.2.2]

As a result compilers end up truncating these numbers. MSVC and Clang
do that. Gcc has an extension that makes it use a 64-bit size for the
enum to avoid the truncation.

At the moment this truncation is only a real problem for
VIRTCHNL2_CAP_OEM, but clearly all these capability flags are
intended to be 64 bits long and therefore should better not be defined
in an enum.

The fix is to use "defines" for them, as the code was using about
7 months ago before enums were introduced here. In this patch the
flag names and values are being preserved, only the enum is being
removed.

Signed-off-by: Andre Muezerie 
---
 drivers/common/idpf/base/virtchnl2.h | 59 ++--
 1 file changed, 30 insertions(+), 29 deletions(-)

diff --git a/drivers/common/idpf/base/virtchnl2.h 
b/drivers/common/idpf/base/virtchnl2.h
index 3285a2b674..b6e28ade97 100644
--- a/drivers/common/idpf/base/virtchnl2.h
+++ b/drivers/common/idpf/base/virtchnl2.h
@@ -239,35 +239,36 @@ enum virtchnl2_cap_rsc {
VIRTCHNL2_CAP_RSC_IPV6_SCTP = BIT(3),
 };
 
-/* Other capability flags */
-enum virtchnl2_cap_other {
-   VIRTCHNL2_CAP_RDMA  = BIT_ULL(0),
-   VIRTCHNL2_CAP_SRIOV = BIT_ULL(1),
-   VIRTCHNL2_CAP_MACFILTER = BIT_ULL(2),
-   VIRTCHNL2_CAP_FLOW_DIRECTOR = BIT_ULL(3),
-   VIRTCHNL2_CAP_SPLITQ_QSCHED = BIT_ULL(4),
-   VIRTCHNL2_CAP_CRC   = BIT_ULL(5),
-   VIRTCHNL2_CAP_FLOW_STEER= BIT_ULL(6),
-   VIRTCHNL2_CAP_WB_ON_ITR = BIT_ULL(7),
-   VIRTCHNL2_CAP_PROMISC   = BIT_ULL(8),
-   VIRTCHNL2_CAP_LINK_SPEED= BIT_ULL(9),
-   VIRTCHNL2_CAP_INLINE_IPSEC  = BIT_ULL(10),
-   VIRTCHNL2_CAP_LARGE_NUM_QUEUES  = BIT_ULL(11),
-   /* Require additional info */
-   VIRTCHNL2_CAP_VLAN  = BIT_ULL(12),
-   VIRTCHNL2_CAP_PTP   = BIT_ULL(13),
-   VIRTCHNL2_CAP_ADV_RSS   = BIT_ULL(15),
-   VIRTCHNL2_CAP_FDIR  = BIT_ULL(16),
-   VIRTCHNL2_CAP_RX_FLEX_DESC  = BIT_ULL(17),
-   VIRTCHNL2_CAP_PTYPE = BIT_ULL(18),
-   VIRTCHNL2_CAP_LOOPBACK  = BIT_ULL(19),
-   /* Enable miss completion types plus ability to detect a miss completion
-* if a reserved bit is set in a standard completion's tag.
-*/
-   VIRTCHNL2_CAP_MISS_COMPL_TAG= BIT_ULL(20),
-   /* This must be the last capability */
-   VIRTCHNL2_CAP_OEM   = BIT_ULL(63),
-};
+/* Other capability flags.
+ * Note: Not using an enum for these flags because these are 64-bit numbers,
+ * which might not fit in an int the enum maps to.
+ */
+#define VIRTCHNL2_CAP_RDMA RTE_BIT64(0)
+#define VIRTCHNL2_CAP_SRIOVRTE_BIT64(1)
+#define VIRTCHNL2_CAP_MACFILTERRTE_BIT64(2)
+#define VIRTCHNL2_CAP_FLOW_DIRECTORRTE_BIT64(3)
+#define VIRTCHNL2_CAP_SPLITQ_QSCHEDRTE_BIT64(4)
+#define VIRTCHNL2_CAP_CRC  RTE_BIT64(5)
+#define VIRTCHNL2_CAP_FLOW_STEER   RTE_BIT64(6)
+#define VIRTCHNL2_CAP_WB_ON_ITRRTE_BIT64(7)
+#define VIRTCHNL2_CAP_PROMISC  RTE_BIT64(8)
+#define VIRTCHNL2_CAP_LINK_SPEED   RTE_BIT64(9)
+#define VIRTCHNL2_CAP_INLINE_IPSEC RTE_BIT64(10)
+#define VIRTCHNL2_CAP_LARGE_NUM_QUEUES RTE_BIT64(11)
+/* Require additional info */
+#define VIRTCHNL2_CAP_VLAN RTE_BIT64(12)
+#define VIRTCHNL2_CAP_PTP  RTE_BIT64(13)
+#define VIRTCHNL2_CAP_ADV_RSS  RTE_BIT64(15)
+#define VIRTCHNL2_CAP_FDIR RTE_BIT64(16)
+#define VIRTCHNL2_CAP_RX_FLEX_DESC RTE_BIT64(17)
+#define VIRTCHNL2_CAP_PTYPERTE_BIT64(18)
+#define VIRTCHNL2_CAP_LOOPBACK RTE_BIT64(19)
+/* Enable miss completion types plus ability to detect a miss completion
+ * if a reserved bit is set in a standard completion's tag.
+ */
+#define VIRTCHNL2_CAP_MISS_COMPL_TAG   RTE_BIT64(20)
+/* This must be the last capability */
+#define VIRTCHNL2_CAP_OEM  RTE_BIT64(63)
 
 /**
  * enum virtchnl2_action_types - Available actions for sideband flow steering
-- 
2.47.0.vfs.0.3



Re: [PATCH v2 2/5] random: defer seeding to EAL init

2024-12-18 Thread Stephen Hemminger
On Tue, 17 Dec 2024 09:59:49 +0100
David Marchand  wrote:

> The RNG is documented as being seeded as part of EAL init.
> 
> Move the initialisation (seeding) helper out of a constructor and
> call it explicitly from rte_eal_init() as it was done before commit
> 3f002f069612 ("eal: replace libc-based random generation with LFSR").
> 
> This also moves the unconditional lcore variable allocation out of a
> constructor.
> 
> While at it, mark local symbol rand_state as static.
> 
> Fixes: 29c39cd3d54d ("random: keep PRNG state in lcore variable")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: David Marchand 
> Reviewed-by: Mattias Rönnblom 
> Acked-by: Anatoly Burakov 

Probably need to add a check to rte_random() so it crashes
if called before initialization, rather than returning an un-random
number which could be a hidden long term bug.


Re: [PATCH] net/virtio: fix Rx checksum calculation

2024-12-18 Thread Stephen Hemminger
On Wed, 18 Dec 2024 10:20:47 +0100
Olivier Matz  wrote:

> Hi,
> 
> On Wed, Dec 18, 2024 at 09:59:05AM +0100, Maxime Coquelin wrote:
> > Hi,
> > 
> > On 12/18/24 08:34, Wangyunjian(wangyunjian,TongTu) wrote:  
> > > > -Original Message-
> > > > From: Maxime Coquelin [mailto:maxime.coque...@redhat.com]
> > > > Sent: Tuesday, December 17, 2024 11:33 PM
> > > > To: dev@dpdk.org
> > > > Cc: Olivier Matz ; Maxime Gouin
> > > > ; Maxime Coquelin
> > > > 
> > > > Subject: [PATCH] net/virtio: fix Rx checksum calculation
> > > > 
> > > > From: Olivier Matz 
> > > > 
> > > > If hdr->csum_start is larger than packet length, the len argument passed
> > > > to rte_raw_cksum_mbuf() overflows and causes a segmentation fault.
> > > > 
> > > > Ignore checksum computation in this case.
> > > > 
> > > > CVE-2024-11614
> > > > 
> > > > Fixes: ca7036b4af3a ("vhost: fix offload flags in Rx path")
> > > > 
> > > > Signed-off-by: Maxime Gouin 
> > > > Signed-off-by: Olivier Matz 
> > > > Reviewed-by: Maxime Coquelin 
> > > > ---
> > > >   lib/vhost/virtio_net.c | 3 +++
> > > >   1 file changed, 3 insertions(+)
> > > > 
> > > > diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
> > > > index d764d4bc6a..69901ab3b5 100644
> > > > --- a/lib/vhost/virtio_net.c
> > > > +++ b/lib/vhost/virtio_net.c
> > > > @@ -2823,6 +2823,9 @@ vhost_dequeue_offload(struct virtio_net *dev,
> > > > struct virtio_net_hdr *hdr,
> > > >  */
> > > > uint16_t csum = 0, off;
> > > > 
> > > > +   if (hdr->csum_start >= rte_pktmbuf_pkt_len(m))
> > > > +   return;
> > > > +  
> > > 
> > > The hdr->csum_start does two successive reads from user space to read
> > > a variable length data structure. The result overflow if the data 
> > > structure
> > > changes between the two reads.
> > > 
> > > We can prevent double fetch issue by using the temporary variable 
> > > csum_start.  
> 
> This is an interesting remark, thanks!
> 
> However, in practical, I'd say that the hdr->csum_start is fetched in a 
> register
> only once if using optimized compilation, because the compiler has no reason 
> to
> think that hdr->csum_start can be modified.
> 
> Olivier

True, but security never depend on optimization.
Needs a fetch and compiler barrier to be truly safe against compilers.


Re: [PATCH] net/virtio: fix Rx checksum calculation

2024-12-18 Thread Olivier Matz
Hi,

On Wed, Dec 18, 2024 at 09:59:05AM +0100, Maxime Coquelin wrote:
> Hi,
> 
> On 12/18/24 08:34, Wangyunjian(wangyunjian,TongTu) wrote:
> > > -Original Message-
> > > From: Maxime Coquelin [mailto:maxime.coque...@redhat.com]
> > > Sent: Tuesday, December 17, 2024 11:33 PM
> > > To: dev@dpdk.org
> > > Cc: Olivier Matz ; Maxime Gouin
> > > ; Maxime Coquelin
> > > 
> > > Subject: [PATCH] net/virtio: fix Rx checksum calculation
> > > 
> > > From: Olivier Matz 
> > > 
> > > If hdr->csum_start is larger than packet length, the len argument passed
> > > to rte_raw_cksum_mbuf() overflows and causes a segmentation fault.
> > > 
> > > Ignore checksum computation in this case.
> > > 
> > > CVE-2024-11614
> > > 
> > > Fixes: ca7036b4af3a ("vhost: fix offload flags in Rx path")
> > > 
> > > Signed-off-by: Maxime Gouin 
> > > Signed-off-by: Olivier Matz 
> > > Reviewed-by: Maxime Coquelin 
> > > ---
> > >   lib/vhost/virtio_net.c | 3 +++
> > >   1 file changed, 3 insertions(+)
> > > 
> > > diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
> > > index d764d4bc6a..69901ab3b5 100644
> > > --- a/lib/vhost/virtio_net.c
> > > +++ b/lib/vhost/virtio_net.c
> > > @@ -2823,6 +2823,9 @@ vhost_dequeue_offload(struct virtio_net *dev,
> > > struct virtio_net_hdr *hdr,
> > >*/
> > >   uint16_t csum = 0, off;
> > > 
> > > + if (hdr->csum_start >= rte_pktmbuf_pkt_len(m))
> > > + return;
> > > +
> > 
> > The hdr->csum_start does two successive reads from user space to read
> > a variable length data structure. The result overflow if the data structure
> > changes between the two reads.
> > 
> > We can prevent double fetch issue by using the temporary variable 
> > csum_start.

This is an interesting remark, thanks!

However, in practical, I'd say that the hdr->csum_start is fetched in a register
only once if using optimized compilation, because the compiler has no reason to
think that hdr->csum_start can be modified.

Olivier

> 
> Right, that's a good catch! The exploitation od this issue seem
> difficult though.
> 
> We may systematically copy the full header, as we only do it for ones
> not contiguous in host VA space.
> 
> What do you think? Are you willing to contribute a fix?
> 
> Thanks,
> Maxime
> > 
> > Thanks,
> > Yunjian
> > 
> > >   if (rte_raw_cksum_mbuf(m, hdr->csum_start,
> > >   rte_pktmbuf_pkt_len(m) - 
> > > hdr->csum_start, &csum) <
> > > 0)
> > >   return;
> > > --
> > > 2.47.0
> > 
> 


Re: [PATCH v2 4/5] power: reduce memory footprint of per-lcore state

2024-12-18 Thread Burakov, Anatoly

On 12/17/2024 9:59 AM, David Marchand wrote:

Now that the per-lcore state was moved into a lcore variable,
there is no reason to align a per-lcore state on a cache line to avoid
false sharing.
Remove this alignment and save a few bytes.

Fixes: 130643319579 ("power: keep per-lcore state in lcore variable")

Signed-off-by: David Marchand 
---
  lib/power/rte_power_pmd_mgmt.c | 2 +-
  1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/lib/power/rte_power_pmd_mgmt.c b/lib/power/rte_power_pmd_mgmt.c
index 369ce3c354..8ec925bd65 100644
--- a/lib/power/rte_power_pmd_mgmt.c
+++ b/lib/power/rte_power_pmd_mgmt.c
@@ -56,7 +56,7 @@ struct queue_list_entry {
const struct rte_eth_rxtx_callback *cb;
  };
  
-struct __rte_cache_aligned pmd_core_cfg {

+struct pmd_core_cfg {
TAILQ_HEAD(queue_list_head, queue_list_entry) head;
/**< List of queues associated with this lcore */
size_t n_queues;

Acked-by: Anatoly Burakov 

--
Thanks,
Anatoly


Re: [PATCH v2 3/5] power: defer lcore variable allocation

2024-12-18 Thread Burakov, Anatoly

On 12/17/2024 9:59 AM, David Marchand wrote:

The lcore variable in this code unit is only used through
rte_power_ethdev_pmgmt_queue_*() public symbols.

Defer the unconditional lcore variable allocation in those symbols.

Fixes: 130643319579 ("power: keep per-lcore state in lcore variable")
Cc: sta...@dpdk.org

Signed-off-by: David Marchand 
---

Acked-by: Anatoly Burakov 

--
Thanks,
Anatoly


Re: [PATCH v2 5/5] eal/x86: defer power intrinsics variable allocation

2024-12-18 Thread Burakov, Anatoly

On 12/17/2024 9:59 AM, David Marchand wrote:

The lcore variable in this code unit is only used through
rte_power_monitor*() public symbols.

Defer the unconditional lcore variable allocation in those symbols.

Fixes: 18b5049ab4fe ("eal/x86: keep power intrinsics state in lcore variable")
Cc: sta...@dpdk.org

Signed-off-by: David Marchand 
---

Acked-by: Anatoly Burakov 

--
Thanks,
Anatoly


Re: [PATCH v2] app/test: fix stack overflow in lpm6_perf_autotest

2024-12-18 Thread Andre Muezerie
On Fri, Dec 13, 2024 at 09:15:40AM -0800, Stephen Hemminger wrote:
> On Fri, 13 Dec 2024 09:08:22 -0800
> Andre Muezerie  wrote:
> 
> > +   struct rte_ipv6_addr *ip_batch =
> > +   (struct rte_ipv6_addr *)rte_malloc("ip_batch",
> > +   sizeof(struct rte_ipv6_addr) * NUM_IPS_ENTRIES, 0
> 
> Cast is not needed here.
> If you are going to allocate an array, use calloc() or rte_calloc()

Thanks for the comments. I'll update the patch.


[PATCH v3] app/test: fix stack overflow in lpm6_perf_autotest

2024-12-18 Thread Andre Muezerie
Test lpm6_perf_autotest was hitting a stack overflow on Windows
with both MSVC and Clang.

The fix is to move some of the data from the stack to the heap.

Signed-off-by: Andre Muezerie 
---
 app/test/test_lpm6_perf.c | 13 +++--
 1 file changed, 11 insertions(+), 2 deletions(-)

diff --git a/app/test/test_lpm6_perf.c b/app/test/test_lpm6_perf.c
index 1860a99ed6..59df0a958a 100644
--- a/app/test/test_lpm6_perf.c
+++ b/app/test/test_lpm6_perf.c
@@ -10,6 +10,7 @@
 
 #include 
 #include 
+#include 
 #include 
 #include 
 
@@ -117,8 +118,13 @@ test_lpm6_perf(void)
total_time = 0;
count = 0;
 
-   struct rte_ipv6_addr ip_batch[NUM_IPS_ENTRIES];
-   int32_t next_hops[NUM_IPS_ENTRIES];
+   struct rte_ipv6_addr *ip_batch = rte_calloc("ip_batch",
+   NUM_IPS_ENTRIES, sizeof(struct rte_ipv6_addr), 0);
+   TEST_LPM_ASSERT(ip_batch != NULL);
+
+   int32_t *next_hops = rte_calloc("next_hops",
+   NUM_IPS_ENTRIES, sizeof(int32_t), 0);
+   TEST_LPM_ASSERT(next_hops != NULL);
 
for (i = 0; i < NUM_IPS_ENTRIES; i++)
ip_batch[i] = large_ips_table[i].ip;
@@ -153,6 +159,9 @@ test_lpm6_perf(void)
printf("Average LPM Delete: %g cycles\n",
(double)total_time / NUM_ROUTE_ENTRIES);
 
+   rte_free(next_hops);
+   rte_free(ip_batch);
+
rte_lpm6_delete_all(lpm);
rte_lpm6_free(lpm);
 
-- 
2.47.0.vfs.0.3



Re: [DPDK/core Bug 1605] lib/eal build failure with address and undefined behavior sanitizers

2024-12-18 Thread Stephen Hemminger
On Tue, 17 Dec 2024 20:39:12 +
bugzi...@dpdk.org wrote:

> https://bugs.dpdk.org/show_bug.cgi?id=1605
> 
> Bug ID: 1605
>Summary: lib/eal build failure with address and undefined
> behavior sanitizers
>Product: DPDK
>Version: 25.03
>   Hardware: All
> OS: All
> Status: UNCONFIRMED
>   Severity: normal
>   Priority: Normal
>  Component: core
>   Assignee: dev@dpdk.org
>   Reporter: alia...@nvidia.com
>   Target Milestone: ---
> 
> $ meson --werror -Db_sanitize=address,undefined build && ninja -C build
> [..]
> /usr/include/x86_64-linux-gnu/bits/string_fortified.h:29:10: error: writing 4
> bytes into a region of size 0 [-Werror=stringop-overflow=]
> [..]
> 
> OS: Ubuntu 22.04
> Compiler: gcc 11.2.0
> Meson: 0.61.2
> Ninja-build: 1.10.1
> 
> Build passes with "-Db_sanitize=address".
> 


Doing this found some other bugs:
[182/3218] Compiling C object 
drivers/libtmp_rte_bus_fslmc.a.p/bus_fslmc_portal_dpaa2_hw_dpio.c.o
../drivers/bus/fslmc/portal/dpaa2_hw_dpio.c: In function 
‘dpaa2_create_dpio_device’:
../drivers/bus/fslmc/portal/dpaa2_hw_dpio.c:607:1: warning: pointer ‘dpio_dev’ 
used after ‘rte_free’ [-Wuse-after-free]
  607 | }
  | ^
../drivers/bus/fslmc/portal/dpaa2_hw_dpio.c:600:17: note: call to ‘rte_free’ 
here
  600 | rte_free(dpio_dev);
  | ^~


Re: [PATCH v2 2/5] random: defer seeding to EAL init

2024-12-18 Thread Mattias Rönnblom

On 2024-12-18 17:35, Stephen Hemminger wrote:

On Tue, 17 Dec 2024 09:59:49 +0100
David Marchand  wrote:


The RNG is documented as being seeded as part of EAL init.

Move the initialisation (seeding) helper out of a constructor and
call it explicitly from rte_eal_init() as it was done before commit
3f002f069612 ("eal: replace libc-based random generation with LFSR").

This also moves the unconditional lcore variable allocation out of a
constructor.

While at it, mark local symbol rand_state as static.

Fixes: 29c39cd3d54d ("random: keep PRNG state in lcore variable")
Cc: sta...@dpdk.org

Signed-off-by: David Marchand 
Reviewed-by: Mattias Rönnblom 
Acked-by: Anatoly Burakov 


Probably need to add a check to rte_random() so it crashes
if called before initialization, rather than returning an un-random
number which could be a hidden long term bug.


If you do, do it either as a RTE_ASSERT() or an RTE_VERIFY() in the 
(lcore_id == LCORE_ID_ANY) path, since that is what will be taken.


Preferably, you should have as little as possible in rte_rand() fast 
path, because this function is used in packet processing.


That said, the "unrandom" number will always be 0, so it shouldn't go 
unnoticed for too long.




Re: [PATCH] net/virtio: fix Rx checksum calculation

2024-12-18 Thread Maxime Coquelin

Hi,

On 12/18/24 08:34, Wangyunjian(wangyunjian,TongTu) wrote:

-Original Message-
From: Maxime Coquelin [mailto:maxime.coque...@redhat.com]
Sent: Tuesday, December 17, 2024 11:33 PM
To: dev@dpdk.org
Cc: Olivier Matz ; Maxime Gouin
; Maxime Coquelin

Subject: [PATCH] net/virtio: fix Rx checksum calculation

From: Olivier Matz 

If hdr->csum_start is larger than packet length, the len argument passed
to rte_raw_cksum_mbuf() overflows and causes a segmentation fault.

Ignore checksum computation in this case.

CVE-2024-11614

Fixes: ca7036b4af3a ("vhost: fix offload flags in Rx path")

Signed-off-by: Maxime Gouin 
Signed-off-by: Olivier Matz 
Reviewed-by: Maxime Coquelin 
---
  lib/vhost/virtio_net.c | 3 +++
  1 file changed, 3 insertions(+)

diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index d764d4bc6a..69901ab3b5 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -2823,6 +2823,9 @@ vhost_dequeue_offload(struct virtio_net *dev,
struct virtio_net_hdr *hdr,
 */
uint16_t csum = 0, off;

+   if (hdr->csum_start >= rte_pktmbuf_pkt_len(m))
+   return;
+


The hdr->csum_start does two successive reads from user space to read
a variable length data structure. The result overflow if the data structure
changes between the two reads.

We can prevent double fetch issue by using the temporary variable csum_start.


Right, that's a good catch! The exploitation od this issue seem
difficult though.

We may systematically copy the full header, as we only do it for ones
not contiguous in host VA space.

What do you think? Are you willing to contribute a fix?

Thanks,
Maxime


Thanks,
Yunjian


if (rte_raw_cksum_mbuf(m, hdr->csum_start,
rte_pktmbuf_pkt_len(m) - hdr->csum_start, 
&csum) <
0)
return;
--
2.47.0






Re: [PATCH v2 1/5] eal: check lcore variable handle

2024-12-18 Thread Burakov, Anatoly

On 12/17/2024 9:59 AM, David Marchand wrote:

Add an assert to double check the passed handle is not NULL, as it
points at an initialisation/allocation issue prior to accessing this
lcore variable.

Signed-off-by: David Marchand 
---
  lib/eal/include/rte_lcore_var.h | 2 ++
  1 file changed, 2 insertions(+)

Acked-by: Anatoly Burakov 

--
Thanks,
Anatoly


[DPDK/examples Bug 1606] flow_filtering/flow_filtering_mismatch_rule: Some mismatch rule packets match the rule

2024-12-18 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1606

Bug ID: 1606
   Summary: flow_filtering/flow_filtering_mismatch_rule: Some
mismatch rule packets match the rule
   Product: DPDK
   Version: 24.11
  Hardware: x86
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: examples
  Assignee: dev@dpdk.org
  Reporter: linglix.c...@intel.com
  Target Milestone: ---

Environment
===
DPDK version: a4f455560f version: 24.11-rc4
OS: Ubuntu24.04.1 LTS (Noble Numbat)/6.8.0-41-generic
Compiler:  gcc version 13.2.0
Hardware platform: Intel(R) Xeon(R) Gold 6139 CPU @ 2.30GHz
NIC hardware: Ethernet Controller XXV710 for 25GbE SFP28 158b
NIC firmware: 
  FW: 9.50 0x8000f4e1 1.3682.0
  Driver: i40e-2.26.8

Test Setup
Steps to reproduce
==
1. bind ports to dpdk
./usertools/dpdk-devbind.py -b vfio-pci :af:00.0 :af:00.1

2. build flow_filtering
meson configure -Dexamples=flow_filtering x86_64-native-linuxapp-gcc
ninja -C x86_64-native-linuxapp-gcc

3. launch flow_filtering 
x86_64-native-linuxapp-gcc/examples/dpdk-flow_filtering -l 1 -n 4 -a
:af:00.0 -a :af:00.1 -- --non-template

4. Send packets which mismatches the defined rule from tester 
sendp([Ether(dst='3c:fd:fe:cf:33:74')/IP(src='0.0.0.0',
dst='192.169.1.1')/Raw(load=b''),Ether(dst='3c:fd:fe:cf:33:74')/IP(src='0.0.0.0',
dst='193.168.1.1')/Raw(load=b'')],iface="ens2f0",count=1,loop=0,inter=0.01)

5. Check the packets are not received by queue 1

Results: 

EAL: Detected CPU lcores: 72
EAL: Detected NUMA nodes: 2
EAL: Detected static linkage of DPDK
EAL: Multi-process socket /var/run/dpdk/rte/mp_socket
EAL: Selected IOVA mode 'VA'
EAL: 1024 hugepages of size 2097152 reserved, but no mounted hugetlbfs found
for that size
EAL: VFIO support initialized
EAL: Using IOMMU type 1 (Type 1)
I40E_DRIVER: i40e_GLQF_reg_init(): i40e device :af:00.0 changed global
register [0x002689a0]. original: 0x0021, new: 0x0029
:: warn: 2 ports detected, but we use only one: port 0
:: initializing port: 0
:: initializing port: 0 done
Flow created!!:
src=00:00:00:00:00:00 - dst=3C:FD:FE:CF:33:74 - queue=0x1
src=00:00:00:00:00:00 - dst=3C:FD:FE:CF:33:74 - queue=0x1

Expected Result:

src=00:00:00:00:00:00 - dst=3C:FD:FE:CF:33:74 - queue=0x0
src=00:00:00:00:00:00 - dst=3C:FD:FE:CF:33:74 - queue=0x0

bad commit:
commit 16158f34900075f2f30b879bf3708e54e07455f4
Author: Shani Peretz 
Date:   Mon Sep 30 22:53:21 2024 +0300

examples/flow_filtering: introduce use cases snippets

These code snippets demonstrate rule creation using
template and non-template APIs.
They include functions that enable developers to create rules.
The purpose of providing these snippets is to allow developers
to reuse them, thereby saving time and effort during the
implementation of flow rules.

The code snippets are categorized based on their usage and can be copied,
paste and modified to suit any requirements.
The snippets provided here are kept up to date and are being compiled
along with the rest of the examples.

There is a skeleton that demonstrates rule creation
using both template and non template APIs.

Developers can change the functions in the skeleton to the corresponding
snippet functions with the appropriate suffix and create rules using the
snippets themselves for easy testing. Each snippet has the same functions
to implement the actions and patterns for the corresponding feature.

Signed-off-by: Shani Peretz 
Acked-by: Stephen Hemminger 

-- 
You are receiving this mail because:
You are the assignee for the bug.

[PATCH v4 04/15] net/zxdh: port tables unint implementations

2024-12-18 Thread Junlong Wang
delete port tables in host.

Signed-off-by: Junlong Wang 
---
 drivers/net/zxdh/zxdh_ethdev.c |  18 ++
 drivers/net/zxdh/zxdh_msg.h|   1 +
 drivers/net/zxdh/zxdh_np.c | 103 +
 drivers/net/zxdh/zxdh_np.h |   9 +++
 drivers/net/zxdh/zxdh_tables.c |  33 ++-
 drivers/net/zxdh/zxdh_tables.h |   1 +
 6 files changed, 164 insertions(+), 1 deletion(-)

diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index ff44816384..717a1d2b0b 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -887,12 +887,30 @@ zxdh_np_uninit(struct rte_eth_dev *dev)
zxdh_np_dtb_data_res_free(hw);
 }
 
+static int
+zxdh_tables_uninit(struct rte_eth_dev *dev)
+{
+   int ret;
+
+   ret = zxdh_port_attr_uninit(dev);
+   if (ret)
+   PMD_DRV_LOG(ERR, "zxdh_port_attr_uninit failed");
+
+   return ret;
+}
+
 static int
 zxdh_dev_close(struct rte_eth_dev *dev)
 {
struct zxdh_hw *hw = dev->data->dev_private;
int ret = 0;
 
+   ret = zxdh_tables_uninit(dev);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR, "%s :tables uninit %s failed ", __func__, 
dev->device->name);
+   return -1;
+   }
+
zxdh_intr_release(dev);
zxdh_np_uninit(dev);
zxdh_pci_reset(hw);
diff --git a/drivers/net/zxdh/zxdh_msg.h b/drivers/net/zxdh/zxdh_msg.h
index b7b17b8696..613ca71170 100644
--- a/drivers/net/zxdh/zxdh_msg.h
+++ b/drivers/net/zxdh/zxdh_msg.h
@@ -167,6 +167,7 @@ enum pciebar_layout_type {
 enum zxdh_msg_type {
ZXDH_NULL = 0,
ZXDH_VF_PORT_INIT = 1,
+   ZXDH_VF_PORT_UNINIT = 2,
 
ZXDH_MSG_TYPE_END,
 };
diff --git a/drivers/net/zxdh/zxdh_np.c b/drivers/net/zxdh/zxdh_np.c
index db536d96e3..99a7dc11b4 100644
--- a/drivers/net/zxdh/zxdh_np.c
+++ b/drivers/net/zxdh/zxdh_np.c
@@ -25,6 +25,7 @@ ZXDH_RISCV_DTB_MGR 
*p_riscv_dtb_queue_mgr[ZXDH_DEV_CHANNEL_MAX];
 ZXDH_TLB_MGR_T *g_p_dpp_tlb_mgr[ZXDH_DEV_CHANNEL_MAX];
 ZXDH_REG_T g_dpp_reg_info[4];
 ZXDH_DTB_TABLE_T g_dpp_dtb_table_info[4];
+ZXDH_SDT_TBL_DATA_T g_sdt_info[ZXDH_DEV_CHANNEL_MAX][ZXDH_DEV_SDT_ID_MAX];
 
 #define ZXDH_SDT_MGR_PTR_GET()(&g_sdt_mgr)
 #define ZXDH_SDT_SOFT_TBL_GET(id) (g_sdt_mgr.sdt_tbl_array[id])
@@ -1454,3 +1455,105 @@ zxdh_np_dtb_table_entry_write(uint32_t dev_id,
 
return rc;
 }
+
+static uint32_t
+zxdh_np_sdt_tbl_data_get(uint32_t dev_id, uint32_t sdt_no, ZXDH_SDT_TBL_DATA_T 
*p_sdt_data)
+{
+   uint32_t rc   = 0;
+
+   p_sdt_data->data_high32 = g_sdt_info[dev_id][sdt_no].data_high32;
+   p_sdt_data->data_low32  = g_sdt_info[dev_id][sdt_no].data_low32;
+
+   return rc;
+}
+
+int
+zxdh_np_dtb_table_entry_delete(uint32_t dev_id,
+uint32_t queue_id,
+uint32_t entrynum,
+ZXDH_DTB_USER_ENTRY_T *delete_entries)
+{
+   ZXDH_SDT_TBL_DATA_T sdt_tbl = {0};
+   ZXDH_DTB_USER_ENTRY_T *pentry = NULL;
+   ZXDH_DTB_ENTRY_T   dtb_one_entry = {0};
+   uint8_t entry_cmd[ZXDH_DTB_TABLE_CMD_SIZE_BIT / 8] = {0};
+   uint8_t entry_data[ZXDH_ETCAM_WIDTH_MAX / 8] = {0};
+   uint8_t *p_data_buff = NULL;
+   uint8_t *p_data_buff_ex = NULL;
+   uint32_t tbl_type = 0;
+   uint32_t element_id = 0xff;
+   uint32_t one_dtb_len = 0;
+   uint32_t dtb_len = 0;
+   uint32_t entry_index;
+   uint32_t sdt_no;
+   uint32_t addr_offset;
+   uint32_t max_size;
+   uint32_t rc;
+
+   ZXDH_COMM_CHECK_POINT(delete_entries);
+
+   p_data_buff = rte_calloc(NULL, 1, ZXDH_DTB_TABLE_DATA_BUFF_SIZE, 0);
+   ZXDH_COMM_CHECK_POINT(p_data_buff);
+
+   p_data_buff_ex = rte_calloc(NULL, 1, ZXDH_DTB_TABLE_DATA_BUFF_SIZE, 0);
+   ZXDH_COMM_CHECK_POINT_MEMORY_FREE(p_data_buff_ex, p_data_buff);
+
+   dtb_one_entry.cmd = entry_cmd;
+   dtb_one_entry.data = entry_data;
+
+   max_size = (ZXDH_DTB_TABLE_DATA_BUFF_SIZE / 16) - 1;
+
+   for (entry_index = 0; entry_index < entrynum; entry_index++) {
+   pentry = delete_entries + entry_index;
+
+   sdt_no = pentry->sdt_no;
+   rc = zxdh_np_sdt_tbl_data_get(dev_id, sdt_no, &sdt_tbl);
+   switch (tbl_type) {
+   case ZXDH_SDT_TBLT_ERAM:
+   {
+   rc = zxdh_np_dtb_eram_one_entry(dev_id, sdt_no, 
ZXDH_DTB_ITEM_DELETE,
+   pentry->p_entry_data, &one_dtb_len, 
&dtb_one_entry);
+   break;
+   }
+   default:
+   {
+   PMD_DRV_LOG(ERR, "SDT table_type[ %d ] is invalid!", 
tbl_type);
+   rte_free(p_data_buff);
+   rte_free(p_data_buff_ex);
+   return 1;
+   }
+   }
+
+   addr_offset = dtb_len * ZXDH_DTB_LEN_POS_SETP;
+   dtb_len += one_dtb_len;
+   if (dtb_len > max_s

[PATCH v4 08/15] net/zxdh: provided dev simple rx implementations

2024-12-18 Thread Junlong Wang
provided dev simple rx implementations.

Signed-off-by: Junlong Wang 
---
 doc/guides/nics/features/zxdh.ini |   1 +
 doc/guides/nics/zxdh.rst  |   1 +
 drivers/net/zxdh/zxdh_ethdev.c|   1 +
 drivers/net/zxdh/zxdh_rxtx.c  | 313 ++
 drivers/net/zxdh/zxdh_rxtx.h  |   2 +
 5 files changed, 318 insertions(+)

diff --git a/doc/guides/nics/features/zxdh.ini 
b/doc/guides/nics/features/zxdh.ini
index 7b72be5f25..bb44e93fad 100644
--- a/doc/guides/nics/features/zxdh.ini
+++ b/doc/guides/nics/features/zxdh.ini
@@ -9,3 +9,4 @@ x86-64   = Y
 ARMv8= Y
 SR-IOV   = Y
 Multiprocess aware   = Y
+Scattered Rx = Y
diff --git a/doc/guides/nics/zxdh.rst b/doc/guides/nics/zxdh.rst
index eb970a888f..f42db9c1f1 100644
--- a/doc/guides/nics/zxdh.rst
+++ b/doc/guides/nics/zxdh.rst
@@ -20,6 +20,7 @@ Features of the ZXDH PMD are:
 - Multi arch support: x86_64, ARMv8.
 - Multiple queues for TX and RX
 - SR-IOV VF
+- Scattered and gather for TX and RX
 
 
 Driver compilation and testing
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index b6057edeaf..0d63129d8d 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -967,6 +967,7 @@ zxdh_set_rxtx_funcs(struct rte_eth_dev *eth_dev)
}
eth_dev->tx_pkt_prepare = zxdh_xmit_pkts_prepare;
eth_dev->tx_pkt_burst = &zxdh_xmit_pkts_packed;
+   eth_dev->rx_pkt_burst = &zxdh_recv_pkts_packed;
 
return 0;
 }
diff --git a/drivers/net/zxdh/zxdh_rxtx.c b/drivers/net/zxdh/zxdh_rxtx.c
index 10034a0e98..06290d48bb 100644
--- a/drivers/net/zxdh/zxdh_rxtx.c
+++ b/drivers/net/zxdh/zxdh_rxtx.c
@@ -31,6 +31,93 @@
 #define ZXDH_TX_MAX_SEGS  31
 #define ZXDH_RX_MAX_SEGS  31
 
+uint32_t zxdh_outer_l2_type[16] = {
+   0,
+   RTE_PTYPE_L2_ETHER,
+   RTE_PTYPE_L2_ETHER_TIMESYNC,
+   RTE_PTYPE_L2_ETHER_ARP,
+   RTE_PTYPE_L2_ETHER_LLDP,
+   RTE_PTYPE_L2_ETHER_NSH,
+   RTE_PTYPE_L2_ETHER_VLAN,
+   RTE_PTYPE_L2_ETHER_QINQ,
+   RTE_PTYPE_L2_ETHER_PPPOE,
+   RTE_PTYPE_L2_ETHER_FCOE,
+   RTE_PTYPE_L2_ETHER_MPLS,
+};
+
+uint32_t zxdh_outer_l3_type[16] = {
+   0,
+   RTE_PTYPE_L3_IPV4,
+   RTE_PTYPE_L3_IPV4_EXT,
+   RTE_PTYPE_L3_IPV6,
+   RTE_PTYPE_L3_IPV4_EXT_UNKNOWN,
+   RTE_PTYPE_L3_IPV6_EXT,
+   RTE_PTYPE_L3_IPV6_EXT_UNKNOWN,
+};
+
+uint32_t zxdh_outer_l4_type[16] = {
+   0,
+   RTE_PTYPE_L4_TCP,
+   RTE_PTYPE_L4_UDP,
+   RTE_PTYPE_L4_FRAG,
+   RTE_PTYPE_L4_SCTP,
+   RTE_PTYPE_L4_ICMP,
+   RTE_PTYPE_L4_NONFRAG,
+   RTE_PTYPE_L4_IGMP,
+};
+
+uint32_t zxdh_tunnel_type[16] = {
+   0,
+   RTE_PTYPE_TUNNEL_IP,
+   RTE_PTYPE_TUNNEL_GRE,
+   RTE_PTYPE_TUNNEL_VXLAN,
+   RTE_PTYPE_TUNNEL_NVGRE,
+   RTE_PTYPE_TUNNEL_GENEVE,
+   RTE_PTYPE_TUNNEL_GRENAT,
+   RTE_PTYPE_TUNNEL_GTPC,
+   RTE_PTYPE_TUNNEL_GTPU,
+   RTE_PTYPE_TUNNEL_ESP,
+   RTE_PTYPE_TUNNEL_L2TP,
+   RTE_PTYPE_TUNNEL_VXLAN_GPE,
+   RTE_PTYPE_TUNNEL_MPLS_IN_GRE,
+   RTE_PTYPE_TUNNEL_MPLS_IN_UDP,
+};
+
+uint32_t zxdh_inner_l2_type[16] = {
+   0,
+   RTE_PTYPE_INNER_L2_ETHER,
+   0,
+   0,
+   0,
+   0,
+   RTE_PTYPE_INNER_L2_ETHER_VLAN,
+   RTE_PTYPE_INNER_L2_ETHER_QINQ,
+   0,
+   0,
+   0,
+};
+
+uint32_t zxdh_inner_l3_type[16] = {
+   0,
+   RTE_PTYPE_INNER_L3_IPV4,
+   RTE_PTYPE_INNER_L3_IPV4_EXT,
+   RTE_PTYPE_INNER_L3_IPV6,
+   RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN,
+   RTE_PTYPE_INNER_L3_IPV6_EXT,
+   RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN,
+};
+
+uint32_t zxdh_inner_l4_type[16] = {
+   0,
+   RTE_PTYPE_INNER_L4_TCP,
+   RTE_PTYPE_INNER_L4_UDP,
+   RTE_PTYPE_INNER_L4_FRAG,
+   RTE_PTYPE_INNER_L4_SCTP,
+   RTE_PTYPE_INNER_L4_ICMP,
+   0,
+   0,
+};
+
 static void
 zxdh_xmit_cleanup_inorder_packed(struct zxdh_virtqueue *vq, int32_t num)
 {
@@ -394,3 +481,229 @@ uint16_t zxdh_xmit_pkts_prepare(void *tx_queue 
__rte_unused, struct rte_mbuf **t
}
return nb_tx;
 }
+
+static uint16_t zxdh_dequeue_burst_rx_packed(struct zxdh_virtqueue *vq,
+   struct rte_mbuf **rx_pkts,
+   uint32_t *len,
+   uint16_t num)
+{
+   struct zxdh_vring_packed_desc *desc = vq->vq_packed.ring.desc;
+   struct rte_mbuf *cookie = NULL;
+   uint16_t i, used_idx;
+   uint16_t id;
+
+   for (i = 0; i < num; i++) {
+   used_idx = vq->vq_used_cons_idx;
+   /**
+* desc_is_used has a load-acquire or rte_io_rmb inside
+* and wait for used desc in virtqueue.
+*/
+   if (!zxdh_desc_used(&desc[used_idx], vq))
+   return i;
+   len[i] = desc[used_idx].len;
+   id

[PATCH v4 09/15] net/zxdh: link info update, set link up/down

2024-12-18 Thread Junlong Wang
provided link info update, set link up /down,
and link intr.

Signed-off-by: Junlong Wang 
---
 doc/guides/nics/features/zxdh.ini  |   2 +
 doc/guides/nics/zxdh.rst   |   3 +
 drivers/net/zxdh/meson.build   |   1 +
 drivers/net/zxdh/zxdh_ethdev.c |  14 ++-
 drivers/net/zxdh/zxdh_ethdev.h |   2 +
 drivers/net/zxdh/zxdh_ethdev_ops.c | 166 
 drivers/net/zxdh/zxdh_ethdev_ops.h |  14 +++
 drivers/net/zxdh/zxdh_msg.c|  57 ++
 drivers/net/zxdh/zxdh_msg.h|  40 +++
 drivers/net/zxdh/zxdh_np.c | 172 -
 drivers/net/zxdh/zxdh_np.h |  20 
 drivers/net/zxdh/zxdh_tables.c |  15 +++
 drivers/net/zxdh/zxdh_tables.h |   6 +-
 13 files changed, 503 insertions(+), 9 deletions(-)
 create mode 100644 drivers/net/zxdh/zxdh_ethdev_ops.c
 create mode 100644 drivers/net/zxdh/zxdh_ethdev_ops.h

diff --git a/doc/guides/nics/features/zxdh.ini 
b/doc/guides/nics/features/zxdh.ini
index bb44e93fad..7da3aaced1 100644
--- a/doc/guides/nics/features/zxdh.ini
+++ b/doc/guides/nics/features/zxdh.ini
@@ -10,3 +10,5 @@ ARMv8= Y
 SR-IOV   = Y
 Multiprocess aware   = Y
 Scattered Rx = Y
+Link status  = Y
+Link status event= Y
diff --git a/doc/guides/nics/zxdh.rst b/doc/guides/nics/zxdh.rst
index f42db9c1f1..fdbc3b3923 100644
--- a/doc/guides/nics/zxdh.rst
+++ b/doc/guides/nics/zxdh.rst
@@ -21,6 +21,9 @@ Features of the ZXDH PMD are:
 - Multiple queues for TX and RX
 - SR-IOV VF
 - Scattered and gather for TX and RX
+- Link Auto-negotiation
+- Link state information
+- Set Link down or up
 
 
 Driver compilation and testing
diff --git a/drivers/net/zxdh/meson.build b/drivers/net/zxdh/meson.build
index 20b2cf484a..48f8f5e1ee 100644
--- a/drivers/net/zxdh/meson.build
+++ b/drivers/net/zxdh/meson.build
@@ -22,4 +22,5 @@ sources = files(
 'zxdh_np.c',
 'zxdh_tables.c',
 'zxdh_rxtx.c',
+'zxdh_ethdev_ops.c',
 )
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index 0d63129d8d..d3876ec9b3 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -16,6 +16,7 @@
 #include "zxdh_np.h"
 #include "zxdh_tables.h"
 #include "zxdh_rxtx.h"
+#include "zxdh_ethdev_ops.h"
 
 struct zxdh_hw_internal zxdh_hw_internal[RTE_MAX_ETHPORTS];
 struct zxdh_shared_data *zxdh_shared_data;
@@ -105,12 +106,18 @@ static void
 zxdh_devconf_intr_handler(void *param)
 {
struct rte_eth_dev *dev = param;
+   struct zxdh_hw *hw = dev->data->dev_private;
+
+   uint8_t isr = zxdh_pci_isr(hw);
 
if (zxdh_intr_unmask(dev) < 0)
PMD_DRV_LOG(ERR, "interrupt enable failed");
+   if (isr & ZXDH_PCI_ISR_CONFIG) {
+   if (zxdh_dev_link_update(dev, 0) == 0)
+   rte_eth_dev_callback_process(dev, 
RTE_ETH_EVENT_INTR_LSC, NULL);
+   }
 }
 
-
 /* Interrupt handler triggered by NIC for handling specific interrupt. */
 static void
 zxdh_fromriscv_intr_handler(void *param)
@@ -1006,6 +1013,8 @@ zxdh_dev_start(struct rte_eth_dev *dev)
vq = hw->vqs[logic_qidx];
zxdh_queue_notify(vq);
}
+   zxdh_dev_set_link_up(dev);
+
return 0;
 }
 
@@ -1020,6 +1029,9 @@ static const struct eth_dev_ops zxdh_eth_dev_ops = {
.tx_queue_setup  = zxdh_dev_tx_queue_setup,
.rx_queue_intr_enable= zxdh_dev_rx_queue_intr_enable,
.rx_queue_intr_disable   = zxdh_dev_rx_queue_intr_disable,
+   .link_update = zxdh_dev_link_update,
+   .dev_set_link_up = zxdh_dev_set_link_up,
+   .dev_set_link_down   = zxdh_dev_set_link_down,
 };
 
 static int32_t
diff --git a/drivers/net/zxdh/zxdh_ethdev.h b/drivers/net/zxdh/zxdh_ethdev.h
index b1f398b28e..c0b719062c 100644
--- a/drivers/net/zxdh/zxdh_ethdev.h
+++ b/drivers/net/zxdh/zxdh_ethdev.h
@@ -72,6 +72,7 @@ struct zxdh_hw {
uint64_t guest_features;
uint32_t max_queue_pairs;
uint32_t speed;
+   uint32_t speed_mode;
uint32_t notify_off_multiplier;
uint16_t *notify_base;
uint16_t pcie_id;
@@ -93,6 +94,7 @@ struct zxdh_hw {
uint8_t panel_id;
uint8_t has_tx_offload;
uint8_t has_rx_offload;
+   uint8_t admin_status;
 };
 
 struct zxdh_dtb_shared_data {
diff --git a/drivers/net/zxdh/zxdh_ethdev_ops.c 
b/drivers/net/zxdh/zxdh_ethdev_ops.c
new file mode 100644
index 00..5a0af98cc0
--- /dev/null
+++ b/drivers/net/zxdh/zxdh_ethdev_ops.c
@@ -0,0 +1,166 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2024 ZTE Corporation
+ */
+
+#include "zxdh_ethdev.h"
+#include "zxdh_pci.h"
+#include "zxdh_msg.h"
+#include "zxdh_ethdev_ops.h"
+#include "zxdh_tables.h"
+#include "zxdh_logs.h"
+
+static int32_t zxdh_config_port_status(struct rte_eth_dev *dev, uint16_t 
link_status)
+{
+   struct zxdh_hw *hw = dev->data->dev_priva

[PATCH v4 02/15] net/zxdh: zxdh np uninit implementation

2024-12-18 Thread Junlong Wang
(np)network processor release resources in host.

Signed-off-by: Junlong Wang 
---
 drivers/net/zxdh/zxdh_ethdev.c |  48 
 drivers/net/zxdh/zxdh_np.c | 470 +
 drivers/net/zxdh/zxdh_np.h | 107 
 3 files changed, 625 insertions(+)

diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index b8f4415e00..4e114d95da 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -841,6 +841,51 @@ zxdh_dev_configure(struct rte_eth_dev *dev)
return ret;
 }
 
+static void
+zxdh_np_dtb_data_res_free(struct zxdh_hw *hw)
+{
+   struct rte_eth_dev *dev = hw->eth_dev;
+   int ret;
+   int i;
+
+   if (g_dtb_data.init_done && g_dtb_data.bind_device == dev) {
+   ret = zxdh_np_online_uninit(0, dev->data->name, 
g_dtb_data.queueid);
+   if (ret)
+   PMD_DRV_LOG(ERR, "%s dpp_np_online_uninstall failed", 
dev->data->name);
+
+   if (g_dtb_data.dtb_table_conf_mz)
+   rte_memzone_free(g_dtb_data.dtb_table_conf_mz);
+
+   if (g_dtb_data.dtb_table_dump_mz) {
+   rte_memzone_free(g_dtb_data.dtb_table_dump_mz);
+   g_dtb_data.dtb_table_dump_mz = NULL;
+   }
+
+   for (i = 0; i < ZXDH_MAX_BASE_DTB_TABLE_COUNT; i++) {
+   if (g_dtb_data.dtb_table_bulk_dump_mz[i]) {
+   
rte_memzone_free(g_dtb_data.dtb_table_bulk_dump_mz[i]);
+   g_dtb_data.dtb_table_bulk_dump_mz[i] = NULL;
+   }
+   }
+   g_dtb_data.init_done = 0;
+   g_dtb_data.bind_device = NULL;
+   }
+   if (zxdh_shared_data != NULL)
+   zxdh_shared_data->np_init_done = 0;
+}
+
+static void
+zxdh_np_uninit(struct rte_eth_dev *dev)
+{
+   struct zxdh_hw *hw = dev->data->dev_private;
+
+   if (!g_dtb_data.init_done && !g_dtb_data.dev_refcnt)
+   return;
+
+   if (--g_dtb_data.dev_refcnt == 0)
+   zxdh_np_dtb_data_res_free(hw);
+}
+
 static int
 zxdh_dev_close(struct rte_eth_dev *dev)
 {
@@ -848,6 +893,7 @@ zxdh_dev_close(struct rte_eth_dev *dev)
int ret = 0;
 
zxdh_intr_release(dev);
+   zxdh_np_uninit(dev);
zxdh_pci_reset(hw);
 
zxdh_dev_free_mbufs(dev);
@@ -1010,6 +1056,7 @@ zxdh_np_dtb_res_init(struct rte_eth_dev *dev)
return 0;
 
 free_res:
+   zxdh_np_dtb_data_res_free(hw);
rte_free(dpp_ctrl);
return ret;
 }
@@ -1177,6 +1224,7 @@ zxdh_eth_dev_init(struct rte_eth_dev *eth_dev)
 
 err_zxdh_init:
zxdh_intr_release(eth_dev);
+   zxdh_np_uninit(eth_dev);
zxdh_bar_msg_chan_exit();
rte_free(eth_dev->data->mac_addrs);
eth_dev->data->mac_addrs = NULL;
diff --git a/drivers/net/zxdh/zxdh_np.c b/drivers/net/zxdh/zxdh_np.c
index e44d7ff501..28728b0c68 100644
--- a/drivers/net/zxdh/zxdh_np.c
+++ b/drivers/net/zxdh/zxdh_np.c
@@ -18,10 +18,21 @@ static ZXDH_DEV_MGR_T g_dev_mgr;
 static ZXDH_SDT_MGR_T g_sdt_mgr;
 ZXDH_PPU_CLS_BITMAP_T g_ppu_cls_bit_map[ZXDH_DEV_CHANNEL_MAX];
 ZXDH_DTB_MGR_T *p_dpp_dtb_mgr[ZXDH_DEV_CHANNEL_MAX];
+ZXDH_RISCV_DTB_MGR *p_riscv_dtb_queue_mgr[ZXDH_DEV_CHANNEL_MAX];
+ZXDH_TLB_MGR_T *g_p_dpp_tlb_mgr[ZXDH_DEV_CHANNEL_MAX];
+ZXDH_REG_T g_dpp_reg_info[4];
 
 #define ZXDH_SDT_MGR_PTR_GET()(&g_sdt_mgr)
 #define ZXDH_SDT_SOFT_TBL_GET(id) (g_sdt_mgr.sdt_tbl_array[id])
 
+#define ZXDH_COMM_MASK_BIT(_bitnum_)\
+   (0x1U << (_bitnum_))
+
+#define ZXDH_COMM_GET_BIT_MASK(_inttype_, _bitqnt_)\
+   ((_inttype_)(((_bitqnt_) < 32)))
+
+#define ZXDH_REG_DATA_MAX  (128)
+
 #define ZXDH_COMM_CHECK_DEV_POINT(dev_id, point)\
 do {\
if (NULL == (point)) {\
@@ -338,3 +349,462 @@ zxdh_np_host_init(uint32_t dev_id,
 
return 0;
 }
+
+static ZXDH_RISCV_DTB_MGR *
+zxdh_np_riscv_dtb_queue_mgr_get(uint32_t dev_id)
+{
+   if (dev_id >= ZXDH_DEV_CHANNEL_MAX)
+   return NULL;
+   else
+   return p_riscv_dtb_queue_mgr[dev_id];
+}
+
+static uint32_t
+zxdh_np_riscv_dtb_mgr_queue_info_delete(uint32_t dev_id, uint32_t queue_id)
+{
+   ZXDH_RISCV_DTB_MGR *p_riscv_dtb_mgr = NULL;
+
+   p_riscv_dtb_mgr = zxdh_np_riscv_dtb_queue_mgr_get(dev_id);
+   if (p_riscv_dtb_mgr == NULL)
+   return 1;
+
+   p_riscv_dtb_mgr->queue_alloc_count--;
+   p_riscv_dtb_mgr->queue_user_info[queue_id].alloc_flag = 0;
+   p_riscv_dtb_mgr->queue_user_info[queue_id].queue_id = 0xFF;
+   p_riscv_dtb_mgr->queue_user_info[queue_id].vport = 0;
+   memset(p_riscv_dtb_mgr->queue_user_info[queue_id].user_name, 0, 
ZXDH_PORT_NAME_MAX);
+
+   return 0;
+}
+
+static uint32_t
+zxdh_np_dev_get_dev_type(uint32_t dev_id)
+{
+   ZXDH_DEV_MGR_T *p_dev_mgr = NULL;
+   ZXDH_DEV_CFG_T *p_dev_info = NULL;
+
+   p_dev_mgr = &g_dev_mgr;
+   p_dev_info = p_dev_mgr->p_dev_array[de

[PATCH v4 06/15] net/zxdh: dev start/stop ops implementations

2024-12-18 Thread Junlong Wang
dev start/stop implementations, start/stop the rx/tx queues.

Signed-off-by: Junlong Wang 
---
 doc/guides/nics/features/zxdh.ini |  2 +
 doc/guides/nics/zxdh.rst  |  2 +
 drivers/net/zxdh/zxdh_ethdev.c| 61 +
 drivers/net/zxdh/zxdh_pci.c   | 24 
 drivers/net/zxdh/zxdh_pci.h   |  1 +
 drivers/net/zxdh/zxdh_queue.c | 91 +++
 drivers/net/zxdh/zxdh_queue.h | 69 +++
 drivers/net/zxdh/zxdh_rxtx.h  | 14 ++---
 8 files changed, 256 insertions(+), 8 deletions(-)

diff --git a/doc/guides/nics/features/zxdh.ini 
b/doc/guides/nics/features/zxdh.ini
index 05c8091ed7..7b72be5f25 100644
--- a/doc/guides/nics/features/zxdh.ini
+++ b/doc/guides/nics/features/zxdh.ini
@@ -7,3 +7,5 @@
 Linux= Y
 x86-64   = Y
 ARMv8= Y
+SR-IOV   = Y
+Multiprocess aware   = Y
diff --git a/doc/guides/nics/zxdh.rst b/doc/guides/nics/zxdh.rst
index 2144753d75..eb970a888f 100644
--- a/doc/guides/nics/zxdh.rst
+++ b/doc/guides/nics/zxdh.rst
@@ -18,6 +18,8 @@ Features
 Features of the ZXDH PMD are:
 
 - Multi arch support: x86_64, ARMv8.
+- Multiple queues for TX and RX
+- SR-IOV VF
 
 
 Driver compilation and testing
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index 521d7ed433..59ee942bdd 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -899,12 +899,35 @@ zxdh_tables_uninit(struct rte_eth_dev *dev)
return ret;
 }
 
+static int
+zxdh_dev_stop(struct rte_eth_dev *dev)
+{
+   int ret = 0;
+
+   if (dev->data->dev_started == 0)
+   return 0;
+
+   ret = zxdh_intr_disable(dev);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "intr disable failed");
+   return -EINVAL;
+   }
+
+   return 0;
+}
+
 static int
 zxdh_dev_close(struct rte_eth_dev *dev)
 {
struct zxdh_hw *hw = dev->data->dev_private;
int ret = 0;
 
+   ret = zxdh_dev_stop(dev);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR, " stop port %s failed.", dev->device->name);
+   return -1;
+   }
+
ret = zxdh_tables_uninit(dev);
if (ret != 0) {
PMD_DRV_LOG(ERR, "%s :tables uninit %s failed ", __func__, 
dev->device->name);
@@ -928,9 +951,47 @@ zxdh_dev_close(struct rte_eth_dev *dev)
return ret;
 }
 
+static int
+zxdh_dev_start(struct rte_eth_dev *dev)
+{
+   struct zxdh_hw *hw = dev->data->dev_private;
+   struct zxdh_virtqueue *vq;
+   int32_t ret;
+   uint16_t logic_qidx;
+   uint16_t i;
+
+   for (i = 0; i < dev->data->nb_rx_queues; i++) {
+   logic_qidx = 2 * i + ZXDH_RQ_QUEUE_IDX;
+   ret = zxdh_dev_rx_queue_setup_finish(dev, logic_qidx);
+   if (ret < 0)
+   return ret;
+   }
+   ret = zxdh_intr_enable(dev);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "interrupt enable failed");
+   return -EINVAL;
+   }
+
+   for (i = 0; i < dev->data->nb_rx_queues; i++) {
+   logic_qidx = 2 * i + ZXDH_RQ_QUEUE_IDX;
+   vq = hw->vqs[logic_qidx];
+   /* Flush the old packets */
+   zxdh_queue_rxvq_flush(vq);
+   zxdh_queue_notify(vq);
+   }
+   for (i = 0; i < dev->data->nb_tx_queues; i++) {
+   logic_qidx = 2 * i + ZXDH_TQ_QUEUE_IDX;
+   vq = hw->vqs[logic_qidx];
+   zxdh_queue_notify(vq);
+   }
+   return 0;
+}
+
 /* dev_ops for zxdh, bare necessities for basic operation */
 static const struct eth_dev_ops zxdh_eth_dev_ops = {
.dev_configure   = zxdh_dev_configure,
+   .dev_start   = zxdh_dev_start,
+   .dev_stop= zxdh_dev_stop,
.dev_close   = zxdh_dev_close,
.dev_infos_get   = zxdh_dev_infos_get,
.rx_queue_setup  = zxdh_dev_rx_queue_setup,
diff --git a/drivers/net/zxdh/zxdh_pci.c b/drivers/net/zxdh/zxdh_pci.c
index 250e67d560..83164a5c79 100644
--- a/drivers/net/zxdh/zxdh_pci.c
+++ b/drivers/net/zxdh/zxdh_pci.c
@@ -202,6 +202,29 @@ zxdh_del_queue(struct zxdh_hw *hw, struct zxdh_virtqueue 
*vq)
rte_write16(0, &hw->common_cfg->queue_enable);
 }
 
+static void
+zxdh_notify_queue(struct zxdh_hw *hw, struct zxdh_virtqueue *vq)
+{
+   uint32_t notify_data = 0;
+
+   if (!zxdh_pci_with_feature(hw, ZXDH_F_NOTIFICATION_DATA)) {
+   rte_write16(vq->vq_queue_index, vq->notify_addr);
+   return;
+   }
+
+   if (zxdh_pci_with_feature(hw, ZXDH_F_RING_PACKED)) {
+   notify_data = ((uint32_t)(!!(vq->vq_packed.cached_flags &
+   
ZXDH_VRING_PACKED_DESC_F_AVAIL)) << 31) |
+   ((uint32_t)vq->vq_avail_idx << 
16) |
+

[PATCH v4 05/15] net/zxdh: rx/tx queue setup and intr enable

2024-12-18 Thread Junlong Wang
rx/tx queue setup and intr enable implementations.

Signed-off-by: Junlong Wang 
---
 drivers/net/zxdh/zxdh_ethdev.c |   4 +
 drivers/net/zxdh/zxdh_queue.c  | 149 +
 drivers/net/zxdh/zxdh_queue.h  |  33 
 3 files changed, 186 insertions(+)

diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index 717a1d2b0b..521d7ed433 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -933,6 +933,10 @@ static const struct eth_dev_ops zxdh_eth_dev_ops = {
.dev_configure   = zxdh_dev_configure,
.dev_close   = zxdh_dev_close,
.dev_infos_get   = zxdh_dev_infos_get,
+   .rx_queue_setup  = zxdh_dev_rx_queue_setup,
+   .tx_queue_setup  = zxdh_dev_tx_queue_setup,
+   .rx_queue_intr_enable= zxdh_dev_rx_queue_intr_enable,
+   .rx_queue_intr_disable   = zxdh_dev_rx_queue_intr_disable,
 };
 
 static int32_t
diff --git a/drivers/net/zxdh/zxdh_queue.c b/drivers/net/zxdh/zxdh_queue.c
index b4ef90ea36..af21f046ad 100644
--- a/drivers/net/zxdh/zxdh_queue.c
+++ b/drivers/net/zxdh/zxdh_queue.c
@@ -12,6 +12,11 @@
 #include "zxdh_common.h"
 #include "zxdh_msg.h"
 
+#define ZXDH_MBUF_MIN_SIZE   sizeof(struct zxdh_net_hdr_dl)
+#define ZXDH_MBUF_SIZE_4K 4096
+#define ZXDH_RX_FREE_THRESH   32
+#define ZXDH_TX_FREE_THRESH   32
+
 struct rte_mbuf *
 zxdh_queue_detach_unused(struct zxdh_virtqueue *vq)
 {
@@ -125,3 +130,147 @@ zxdh_free_queues(struct rte_eth_dev *dev)
 
return 0;
 }
+
+static int
+zxdh_check_mempool(struct rte_mempool *mp, uint16_t offset, uint16_t 
min_length)
+{
+   uint16_t data_room_size;
+
+   if (mp == NULL)
+   return -EINVAL;
+   data_room_size = rte_pktmbuf_data_room_size(mp);
+   if (data_room_size < offset + min_length) {
+   PMD_RX_LOG(ERR,
+  "%s mbuf_data_room_size %u < %u (%u + %u)",
+  mp->name, data_room_size,
+  offset + min_length, offset, min_length);
+   return -EINVAL;
+   }
+   return 0;
+}
+
+int32_t
+zxdh_dev_rx_queue_setup(struct rte_eth_dev *dev,
+   uint16_t queue_idx,
+   uint16_t nb_desc,
+   uint32_t socket_id __rte_unused,
+   const struct rte_eth_rxconf *rx_conf,
+   struct rte_mempool *mp)
+{
+   struct zxdh_hw *hw = dev->data->dev_private;
+   uint16_t vtpci_logic_qidx = 2 * queue_idx + ZXDH_RQ_QUEUE_IDX;
+   struct zxdh_virtqueue *vq = hw->vqs[vtpci_logic_qidx];
+   int32_t ret = 0;
+
+   if (rx_conf->rx_deferred_start) {
+   PMD_RX_LOG(ERR, "Rx deferred start is not supported");
+   return -EINVAL;
+   }
+   uint16_t rx_free_thresh = rx_conf->rx_free_thresh;
+
+   if (rx_free_thresh == 0)
+   rx_free_thresh = RTE_MIN(vq->vq_nentries / 4, 
ZXDH_RX_FREE_THRESH);
+
+   /* rx_free_thresh must be multiples of four. */
+   if (rx_free_thresh & 0x3) {
+   PMD_RX_LOG(ERR, "(rx_free_thresh=%u port=%u queue=%u)",
+   rx_free_thresh, dev->data->port_id, queue_idx);
+   return -EINVAL;
+   }
+   /* rx_free_thresh must be less than the number of RX entries */
+   if (rx_free_thresh >= vq->vq_nentries) {
+   PMD_RX_LOG(ERR, "RX entries (%u). (rx_free_thresh=%u port=%u 
queue=%u)",
+   vq->vq_nentries, rx_free_thresh, dev->data->port_id, 
queue_idx);
+   return -EINVAL;
+   }
+   vq->vq_free_thresh = rx_free_thresh;
+   nb_desc = ZXDH_QUEUE_DEPTH;
+
+   vq->vq_free_cnt = RTE_MIN(vq->vq_free_cnt, nb_desc);
+   struct zxdh_virtnet_rx *rxvq = &vq->rxq;
+
+   rxvq->queue_id = vtpci_logic_qidx;
+
+   int mbuf_min_size  = ZXDH_MBUF_MIN_SIZE;
+
+   if (rx_conf->offloads & RTE_ETH_RX_OFFLOAD_TCP_LRO)
+   mbuf_min_size = ZXDH_MBUF_SIZE_4K;
+
+   ret = zxdh_check_mempool(mp, RTE_PKTMBUF_HEADROOM, mbuf_min_size);
+   if (ret != 0) {
+   PMD_RX_LOG(ERR,
+   "rxq setup but mpool size too small(<%d) failed", 
mbuf_min_size);
+   return -EINVAL;
+   }
+   rxvq->mpool = mp;
+   if (queue_idx < dev->data->nb_rx_queues)
+   dev->data->rx_queues[queue_idx] = rxvq;
+
+   return 0;
+}
+
+int32_t
+zxdh_dev_tx_queue_setup(struct rte_eth_dev *dev,
+   uint16_t queue_idx,
+   uint16_t nb_desc,
+   uint32_t socket_id __rte_unused,
+   const struct rte_eth_txconf *tx_conf)
+{
+   uint16_t vtpci_logic_qidx = 2 * queue_idx + ZXDH_TQ_QUEUE_IDX;
+   struct zxdh_hw *hw = dev->data->dev_private;
+   struct zxdh_virtqueue *vq = hw->vqs[vtpc

[PATCH v4 00/15] net/zxdh: updated net zxdh driver

2024-12-18 Thread Junlong Wang
V4:
  - resolved ci compile issues.

V3:
  - use rte_zmalloc and rte_calloc to avoid memset.
  - remove unnecessary initialization, which first usage will set.
  - adjust some function which is always return 0, changed to void 
and skip the ASSERTION later.
  - resolved some WARNING:MACRO_ARG_UNUSED issues.
  - resolved some other issues.

V2:
  - resolve code style and github-robot build issue.

V1:
  - updated net zxdh driver
provided insert/delete/get table code funcs.
provided link/mac/vlan/promiscuous/rss/mtu ops.

Junlong Wang (15):
  net/zxdh: zxdh np init implementation
  net/zxdh: zxdh np uninit implementation
  net/zxdh: port tables init implementations
  net/zxdh: port tables unint implementations
  net/zxdh: rx/tx queue setup and intr enable
  net/zxdh: dev start/stop ops implementations
  net/zxdh: provided dev simple tx implementations
  net/zxdh: provided dev simple rx implementations
  net/zxdh: link info update, set link up/down
  net/zxdh: mac set/add/remove ops implementations
  net/zxdh: promisc/allmulti ops implementations
  net/zxdh: vlan filter/ offload ops implementations
  net/zxdh: rss hash config/update, reta update/get
  net/zxdh: basic stats ops implementations
  net/zxdh: mtu update ops implementations

 doc/guides/nics/features/zxdh.ini  |   18 +
 doc/guides/nics/zxdh.rst   |   17 +
 drivers/net/zxdh/meson.build   |4 +
 drivers/net/zxdh/zxdh_common.c |   24 +
 drivers/net/zxdh/zxdh_common.h |1 +
 drivers/net/zxdh/zxdh_ethdev.c |  575 +++-
 drivers/net/zxdh/zxdh_ethdev.h |   40 +
 drivers/net/zxdh/zxdh_ethdev_ops.c | 1595 +
 drivers/net/zxdh/zxdh_ethdev_ops.h |   80 ++
 drivers/net/zxdh/zxdh_msg.c|  166 +++
 drivers/net/zxdh/zxdh_msg.h|  232 
 drivers/net/zxdh/zxdh_np.c | 2060 
 drivers/net/zxdh/zxdh_np.h |  579 
 drivers/net/zxdh/zxdh_pci.c|   26 +-
 drivers/net/zxdh/zxdh_pci.h|9 +-
 drivers/net/zxdh/zxdh_queue.c  |  242 +++-
 drivers/net/zxdh/zxdh_queue.h  |  144 +-
 drivers/net/zxdh/zxdh_rxtx.c   |  804 +++
 drivers/net/zxdh/zxdh_rxtx.h   |   20 +-
 drivers/net/zxdh/zxdh_tables.c |  794 +++
 drivers/net/zxdh/zxdh_tables.h |  231 
 21 files changed, 7615 insertions(+), 46 deletions(-)
 create mode 100644 drivers/net/zxdh/zxdh_ethdev_ops.c
 create mode 100644 drivers/net/zxdh/zxdh_ethdev_ops.h
 create mode 100644 drivers/net/zxdh/zxdh_np.c
 create mode 100644 drivers/net/zxdh/zxdh_np.h
 create mode 100644 drivers/net/zxdh/zxdh_rxtx.c
 create mode 100644 drivers/net/zxdh/zxdh_tables.c
 create mode 100644 drivers/net/zxdh/zxdh_tables.h

-- 
2.27.0

[PATCH v4 11/15] net/zxdh: promisc/allmulti ops implementations

2024-12-18 Thread Junlong Wang
provided promiscuous/allmulticast ops.

Signed-off-by: Junlong Wang 
---
 doc/guides/nics/features/zxdh.ini  |   2 +
 doc/guides/nics/zxdh.rst   |   2 +
 drivers/net/zxdh/zxdh_ethdev.c |  21 ++-
 drivers/net/zxdh/zxdh_ethdev.h |   2 +
 drivers/net/zxdh/zxdh_ethdev_ops.c | 132 +
 drivers/net/zxdh/zxdh_ethdev_ops.h |   4 +
 drivers/net/zxdh/zxdh_msg.h|  10 ++
 drivers/net/zxdh/zxdh_tables.c | 223 +
 drivers/net/zxdh/zxdh_tables.h |  22 +++
 9 files changed, 417 insertions(+), 1 deletion(-)

diff --git a/doc/guides/nics/features/zxdh.ini 
b/doc/guides/nics/features/zxdh.ini
index dc09fe3453..e9b237e102 100644
--- a/doc/guides/nics/features/zxdh.ini
+++ b/doc/guides/nics/features/zxdh.ini
@@ -14,3 +14,5 @@ Link status  = Y
 Link status event= Y
 Unicast MAC filter   = Y
 Multicast MAC filter = Y
+Promiscuous mode = Y
+Allmulticast mode= Y
diff --git a/doc/guides/nics/zxdh.rst b/doc/guides/nics/zxdh.rst
index e0b0776aca..0399df1302 100644
--- a/doc/guides/nics/zxdh.rst
+++ b/doc/guides/nics/zxdh.rst
@@ -26,6 +26,8 @@ Features of the ZXDH PMD are:
 - Set Link down or up
 - Unicast MAC filter
 - Multicast MAC filter
+- Promiscuous mode
+- Multicast mode
 
 
 Driver compilation and testing
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index 85ada87cdc..1d64b877c1 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -901,8 +901,16 @@ zxdh_tables_uninit(struct rte_eth_dev *dev)
int ret;
 
ret = zxdh_port_attr_uninit(dev);
-   if (ret)
+   if (ret) {
PMD_DRV_LOG(ERR, "zxdh_port_attr_uninit failed");
+   return ret;
+   }
+
+   ret = zxdh_promisc_table_uninit(dev);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "uninit promisc_table failed");
+   return ret;
+   }
 
return ret;
 }
@@ -1055,6 +1063,10 @@ static const struct eth_dev_ops zxdh_eth_dev_ops = {
.mac_addr_add= zxdh_dev_mac_addr_add,
.mac_addr_remove = zxdh_dev_mac_addr_remove,
.mac_addr_set= zxdh_dev_mac_addr_set,
+   .promiscuous_enable  = zxdh_dev_promiscuous_enable,
+   .promiscuous_disable = zxdh_dev_promiscuous_disable,
+   .allmulticast_enable = zxdh_dev_allmulticast_enable,
+   .allmulticast_disable= zxdh_dev_allmulticast_disable,
 };
 
 static int32_t
@@ -1306,6 +1318,13 @@ zxdh_tables_init(struct rte_eth_dev *dev)
PMD_DRV_LOG(ERR, " panel table init failed");
return ret;
}
+
+   ret = zxdh_promisc_table_init(dev);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "promisc_table_init failed");
+   return ret;
+   }
+
return ret;
 }
 
diff --git a/drivers/net/zxdh/zxdh_ethdev.h b/drivers/net/zxdh/zxdh_ethdev.h
index 5b95cb1c2a..3cdac5de73 100644
--- a/drivers/net/zxdh/zxdh_ethdev.h
+++ b/drivers/net/zxdh/zxdh_ethdev.h
@@ -98,6 +98,8 @@ struct zxdh_hw {
uint8_t has_tx_offload;
uint8_t has_rx_offload;
uint8_t admin_status;
+   uint8_t promisc_status;
+   uint8_t allmulti_status;
 };
 
 struct zxdh_dtb_shared_data {
diff --git a/drivers/net/zxdh/zxdh_ethdev_ops.c 
b/drivers/net/zxdh/zxdh_ethdev_ops.c
index 751f80e9b4..aed4e6410c 100644
--- a/drivers/net/zxdh/zxdh_ethdev_ops.c
+++ b/drivers/net/zxdh/zxdh_ethdev_ops.c
@@ -397,3 +397,135 @@ void zxdh_dev_mac_addr_remove(struct rte_eth_dev *dev 
__rte_unused, uint32_t ind
}
memset(&dev->data->mac_addrs[index], 0, sizeof(struct rte_ether_addr));
 }
+
+int zxdh_dev_promiscuous_enable(struct rte_eth_dev *dev)
+{
+   struct zxdh_hw *hw  = dev->data->dev_private;
+   struct zxdh_msg_info msg_info = {0};
+   int16_t ret = 0;
+
+   if (hw->promisc_status == 0) {
+   if (hw->is_pf) {
+   ret = zxdh_dev_unicast_table_set(hw, hw->vport.vport, 
true);
+   if (hw->allmulti_status == 0)
+   ret = zxdh_dev_multicast_table_set(hw, 
hw->vport.vport, true);
+
+   } else {
+   struct zxdh_port_promisc_msg *promisc_msg = 
&msg_info.data.port_promisc_msg;
+
+   zxdh_msg_head_build(hw, ZXDH_PORT_PROMISC_SET, 
&msg_info);
+   promisc_msg->mode = ZXDH_PROMISC_MODE;
+   promisc_msg->value = true;
+   if (hw->allmulti_status == 0)
+   promisc_msg->mc_follow = true;
+
+   ret = zxdh_vf_send_msg_to_pf(dev, &msg_info, 
sizeof(msg_info), NULL, 0);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "Failed to send msg: port 0x%x 
msg type %d",
+   hw->vport.vport, 
ZXDH_PROMISC_MODE);
+   return ret;
+  

[PATCH v4 13/15] net/zxdh: rss hash config/update, reta update/get

2024-12-18 Thread Junlong Wang
provided rss hash config/update, reta update/get ops.

Signed-off-by: Junlong Wang 
---
 doc/guides/nics/features/zxdh.ini  |   3 +
 doc/guides/nics/zxdh.rst   |   1 +
 drivers/net/zxdh/zxdh_ethdev.c |  52 
 drivers/net/zxdh/zxdh_ethdev.h |   3 +
 drivers/net/zxdh/zxdh_ethdev_ops.c | 410 +
 drivers/net/zxdh/zxdh_ethdev_ops.h |  26 ++
 drivers/net/zxdh/zxdh_msg.h|  22 ++
 drivers/net/zxdh/zxdh_tables.c |  82 ++
 drivers/net/zxdh/zxdh_tables.h |   7 +
 9 files changed, 606 insertions(+)

diff --git a/doc/guides/nics/features/zxdh.ini 
b/doc/guides/nics/features/zxdh.ini
index 6fb006c2da..415ca547d0 100644
--- a/doc/guides/nics/features/zxdh.ini
+++ b/doc/guides/nics/features/zxdh.ini
@@ -19,3 +19,6 @@ Allmulticast mode= Y
 VLAN filter  = Y
 VLAN offload = Y
 QinQ offload = Y
+RSS hash = Y
+RSS reta update  = Y
+Inner RSS= Y
diff --git a/doc/guides/nics/zxdh.rst b/doc/guides/nics/zxdh.rst
index 3a7585d123..3cc6a1d348 100644
--- a/doc/guides/nics/zxdh.rst
+++ b/doc/guides/nics/zxdh.rst
@@ -31,6 +31,7 @@ Features of the ZXDH PMD are:
 - VLAN filter and VLAN offload
 - VLAN stripping and inserting
 - QINQ stripping and inserting
+- Receive Side Scaling (RSS)
 
 
 Driver compilation and testing
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index cc32b467a9..1349559c9b 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -61,6 +61,9 @@ zxdh_dev_infos_get(struct rte_eth_dev *dev,
dev_info->rx_offload_capa |=  RTE_ETH_RX_OFFLOAD_TCP_LRO;
dev_info->rx_offload_capa |=  RTE_ETH_RX_OFFLOAD_RSS_HASH;
 
+   dev_info->reta_size = RTE_ETH_RSS_RETA_SIZE_256;
+   dev_info->flow_type_rss_offloads = ZXDH_RSS_HF;
+
dev_info->tx_offload_capa = (RTE_ETH_TX_OFFLOAD_MULTI_SEGS);
dev_info->tx_offload_capa |= (RTE_ETH_TX_OFFLOAD_TCP_TSO |
RTE_ETH_TX_OFFLOAD_UDP_TSO);
@@ -784,9 +787,48 @@ zxdh_dev_conf_offload(struct rte_eth_dev *dev)
return ret;
}
 
+   ret = zxdh_rss_configure(dev);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "rss configure failed");
+   return ret;
+   }
+
return 0;
 }
 
+static int
+zxdh_rss_qid_config(struct rte_eth_dev *dev)
+{
+   struct zxdh_hw *hw = dev->data->dev_private;
+   struct zxdh_port_attr_table port_attr = {0};
+   struct zxdh_msg_info msg_info = {0};
+   int ret = 0;
+
+   if (hw->is_pf) {
+   ret = zxdh_get_port_attr(hw->vport.vfid, &port_attr);
+   port_attr.port_base_qid = hw->channel_context[0].ph_chno & 
0xfff;
+
+   ret = zxdh_set_port_attr(hw->vport.vfid, &port_attr);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "PF:%d port_base_qid insert failed", 
hw->vfid);
+   return ret;
+   }
+   } else {
+   struct zxdh_port_attr_set_msg *attr_msg = 
&msg_info.data.port_attr_msg;
+
+   zxdh_msg_head_build(hw, ZXDH_PORT_ATTRS_SET, &msg_info);
+   attr_msg->mode = ZXDH_PORT_BASE_QID_FLAG;
+   attr_msg->value = hw->channel_context[0].ph_chno & 0xfff;
+   ret = zxdh_vf_send_msg_to_pf(dev, &msg_info, sizeof(msg_info), 
NULL, 0);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "Failed to send msg: port 0x%x msg 
type %d ",
+   hw->vport.vport, 
ZXDH_PORT_BASE_QID_FLAG);
+   return ret;
+   }
+   }
+   return ret;
+}
+
 static int32_t
 zxdh_dev_configure(struct rte_eth_dev *dev)
 {
@@ -873,6 +915,12 @@ zxdh_dev_configure(struct rte_eth_dev *dev)
return -1;
}
 
+   ret = zxdh_rss_qid_config(dev);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "Failed to configure base qid!");
+   return -1;
+   }
+
zxdh_pci_reinit_complete(hw);
 
 end:
@@ -1099,6 +1147,10 @@ static const struct eth_dev_ops zxdh_eth_dev_ops = {
.allmulticast_disable= zxdh_dev_allmulticast_disable,
.vlan_filter_set = zxdh_dev_vlan_filter_set,
.vlan_offload_set= zxdh_dev_vlan_offload_set,
+   .reta_update = zxdh_dev_rss_reta_update,
+   .reta_query  = zxdh_dev_rss_reta_query,
+   .rss_hash_update = zxdh_rss_hash_update,
+   .rss_hash_conf_get   = zxdh_rss_hash_conf_get,
 };
 
 static int32_t
diff --git a/drivers/net/zxdh/zxdh_ethdev.h b/drivers/net/zxdh/zxdh_ethdev.h
index 3cdac5de73..2934fa264a 100644
--- a/drivers/net/zxdh/zxdh_ethdev.h
+++ b/drivers/net/zxdh/zxdh_ethdev.h
@@ -82,6 +82,7 @@ struct zxdh_hw {
uint16_t queue_num;
uint16_t mc_num;
uint16_t uc_num;
+   uint16_t *rss_reta;
 
uint8_t *isr;
uint8_t wea

[PATCH v4 10/15] net/zxdh: mac set/add/remove ops implementations

2024-12-18 Thread Junlong Wang
provided mac set/add/remove ops.

Signed-off-by: Junlong Wang 
---
 doc/guides/nics/features/zxdh.ini  |   2 +
 doc/guides/nics/zxdh.rst   |   2 +
 drivers/net/zxdh/zxdh_common.c |  24 +++
 drivers/net/zxdh/zxdh_common.h |   1 +
 drivers/net/zxdh/zxdh_ethdev.c |  32 +++-
 drivers/net/zxdh/zxdh_ethdev.h |   3 +
 drivers/net/zxdh/zxdh_ethdev_ops.c | 233 +
 drivers/net/zxdh/zxdh_ethdev_ops.h |   4 +
 drivers/net/zxdh/zxdh_msg.h|  12 ++
 drivers/net/zxdh/zxdh_np.h |   5 +
 drivers/net/zxdh/zxdh_tables.c | 197 
 drivers/net/zxdh/zxdh_tables.h |  36 +
 12 files changed, 549 insertions(+), 2 deletions(-)

diff --git a/doc/guides/nics/features/zxdh.ini 
b/doc/guides/nics/features/zxdh.ini
index 7da3aaced1..dc09fe3453 100644
--- a/doc/guides/nics/features/zxdh.ini
+++ b/doc/guides/nics/features/zxdh.ini
@@ -12,3 +12,5 @@ Multiprocess aware   = Y
 Scattered Rx = Y
 Link status  = Y
 Link status event= Y
+Unicast MAC filter   = Y
+Multicast MAC filter = Y
diff --git a/doc/guides/nics/zxdh.rst b/doc/guides/nics/zxdh.rst
index fdbc3b3923..e0b0776aca 100644
--- a/doc/guides/nics/zxdh.rst
+++ b/doc/guides/nics/zxdh.rst
@@ -24,6 +24,8 @@ Features of the ZXDH PMD are:
 - Link Auto-negotiation
 - Link state information
 - Set Link down or up
+- Unicast MAC filter
+- Multicast MAC filter
 
 
 Driver compilation and testing
diff --git a/drivers/net/zxdh/zxdh_common.c b/drivers/net/zxdh/zxdh_common.c
index 4f18c97ed7..75883a8897 100644
--- a/drivers/net/zxdh/zxdh_common.c
+++ b/drivers/net/zxdh/zxdh_common.c
@@ -256,6 +256,30 @@ zxdh_panelid_get(struct rte_eth_dev *dev, uint8_t *panelid)
return ret;
 }
 
+static int
+zxdh_get_res_hash_id(struct zxdh_res_para *in, uint8_t *hash_id)
+{
+   uint8_t reps = 0;
+   uint16_t reps_len = 0;
+
+   if (zxdh_get_res_info(in, ZXDH_TBL_FIELD_HASHID, &reps, &reps_len) != 
ZXDH_BAR_MSG_OK)
+   return -1;
+
+   *hash_id = reps;
+   return ZXDH_BAR_MSG_OK;
+}
+
+int32_t
+zxdh_hashidx_get(struct rte_eth_dev *dev, uint8_t *hash_idx)
+{
+   struct zxdh_res_para param;
+
+   zxdh_fill_res_para(dev, ¶m);
+   int32_t ret = zxdh_get_res_hash_id(¶m, hash_idx);
+
+   return ret;
+}
+
 uint32_t
 zxdh_read_bar_reg(struct rte_eth_dev *dev, uint32_t bar, uint32_t reg)
 {
diff --git a/drivers/net/zxdh/zxdh_common.h b/drivers/net/zxdh/zxdh_common.h
index 72c29e1522..826f1fb95d 100644
--- a/drivers/net/zxdh/zxdh_common.h
+++ b/drivers/net/zxdh/zxdh_common.h
@@ -22,6 +22,7 @@ struct zxdh_res_para {
 
 int32_t zxdh_phyport_get(struct rte_eth_dev *dev, uint8_t *phyport);
 int32_t zxdh_panelid_get(struct rte_eth_dev *dev, uint8_t *pannelid);
+int32_t zxdh_hashidx_get(struct rte_eth_dev *dev, uint8_t *hash_idx);
 uint32_t zxdh_read_bar_reg(struct rte_eth_dev *dev, uint32_t bar, uint32_t 
reg);
 void zxdh_write_bar_reg(struct rte_eth_dev *dev, uint32_t bar, uint32_t reg, 
uint32_t val);
 void zxdh_release_lock(struct zxdh_hw *hw);
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index d3876ec9b3..85ada87cdc 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -979,6 +979,23 @@ zxdh_set_rxtx_funcs(struct rte_eth_dev *eth_dev)
return 0;
 }
 
+static int
+zxdh_mac_config(struct rte_eth_dev *eth_dev)
+{
+   struct zxdh_hw *hw = eth_dev->data->dev_private;
+   int ret = 0;
+
+   if (hw->is_pf) {
+   ret = zxdh_set_mac_table(hw->vport.vport,
+   ð_dev->data->mac_addrs[0], 
hw->hash_search_index);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "Failed to add mac: port 0x%x", 
hw->vport.vport);
+   return ret;
+   }
+   }
+   return ret;
+}
+
 static int
 zxdh_dev_start(struct rte_eth_dev *dev)
 {
@@ -1014,6 +1031,9 @@ zxdh_dev_start(struct rte_eth_dev *dev)
zxdh_queue_notify(vq);
}
zxdh_dev_set_link_up(dev);
+   ret = zxdh_mac_config(hw->eth_dev);
+   if (ret)
+   PMD_DRV_LOG(ERR, " mac config failed");
 
return 0;
 }
@@ -1032,6 +1052,9 @@ static const struct eth_dev_ops zxdh_eth_dev_ops = {
.link_update = zxdh_dev_link_update,
.dev_set_link_up = zxdh_dev_set_link_up,
.dev_set_link_down   = zxdh_dev_set_link_down,
+   .mac_addr_add= zxdh_dev_mac_addr_add,
+   .mac_addr_remove = zxdh_dev_mac_addr_remove,
+   .mac_addr_set= zxdh_dev_mac_addr_set,
 };
 
 static int32_t
@@ -1073,15 +1096,20 @@ zxdh_agent_comm(struct rte_eth_dev *eth_dev, struct 
zxdh_hw *hw)
PMD_DRV_LOG(ERR, "Failed to get phyport");
return -1;
}
-   PMD_DRV_LOG(INFO, "Get phyport success: 0x%x", hw->phyport);
+   PMD_DRV_LOG(DEBUG, "Get phyport success: 0x%x", hw

[PATCH v4 07/15] net/zxdh: provided dev simple tx implementations

2024-12-18 Thread Junlong Wang
provided dev simple tx implementations.

Signed-off-by: Junlong Wang 
---
 drivers/net/zxdh/meson.build   |   1 +
 drivers/net/zxdh/zxdh_ethdev.c |  21 ++
 drivers/net/zxdh/zxdh_queue.h  |  26 ++-
 drivers/net/zxdh/zxdh_rxtx.c   | 396 +
 drivers/net/zxdh/zxdh_rxtx.h   |   4 +
 5 files changed, 447 insertions(+), 1 deletion(-)
 create mode 100644 drivers/net/zxdh/zxdh_rxtx.c

diff --git a/drivers/net/zxdh/meson.build b/drivers/net/zxdh/meson.build
index 5b3af87c5b..20b2cf484a 100644
--- a/drivers/net/zxdh/meson.build
+++ b/drivers/net/zxdh/meson.build
@@ -21,4 +21,5 @@ sources = files(
 'zxdh_queue.c',
 'zxdh_np.c',
 'zxdh_tables.c',
+'zxdh_rxtx.c',
 )
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index 59ee942bdd..b6057edeaf 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -15,6 +15,7 @@
 #include "zxdh_queue.h"
 #include "zxdh_np.h"
 #include "zxdh_tables.h"
+#include "zxdh_rxtx.h"
 
 struct zxdh_hw_internal zxdh_hw_internal[RTE_MAX_ETHPORTS];
 struct zxdh_shared_data *zxdh_shared_data;
@@ -951,6 +952,25 @@ zxdh_dev_close(struct rte_eth_dev *dev)
return ret;
 }
 
+static int32_t
+zxdh_set_rxtx_funcs(struct rte_eth_dev *eth_dev)
+{
+   struct zxdh_hw *hw = eth_dev->data->dev_private;
+
+   if (!zxdh_pci_packed_queue(hw)) {
+   PMD_DRV_LOG(ERR, " port %u not support packed queue", 
eth_dev->data->port_id);
+   return -1;
+   }
+   if (!zxdh_pci_with_feature(hw, ZXDH_NET_F_MRG_RXBUF)) {
+   PMD_DRV_LOG(ERR, " port %u not support rx mergeable", 
eth_dev->data->port_id);
+   return -1;
+   }
+   eth_dev->tx_pkt_prepare = zxdh_xmit_pkts_prepare;
+   eth_dev->tx_pkt_burst = &zxdh_xmit_pkts_packed;
+
+   return 0;
+}
+
 static int
 zxdh_dev_start(struct rte_eth_dev *dev)
 {
@@ -966,6 +986,7 @@ zxdh_dev_start(struct rte_eth_dev *dev)
if (ret < 0)
return ret;
}
+   zxdh_set_rxtx_funcs(dev);
ret = zxdh_intr_enable(dev);
if (ret) {
PMD_DRV_LOG(ERR, "interrupt enable failed");
diff --git a/drivers/net/zxdh/zxdh_queue.h b/drivers/net/zxdh/zxdh_queue.h
index 6513aec3f0..9343df81ac 100644
--- a/drivers/net/zxdh/zxdh_queue.h
+++ b/drivers/net/zxdh/zxdh_queue.h
@@ -21,8 +21,15 @@ enum { ZXDH_VTNET_RQ = 0, ZXDH_VTNET_TQ = 1 };
 #define ZXDH_TQ_QUEUE_IDX 1
 #define ZXDH_MAX_TX_INDIRECT  8
 
+/* This marks a buffer as continuing via the next field. */
+#define ZXDH_VRING_DESC_F_NEXT 1
+
 /* This marks a buffer as write-only (otherwise read-only). */
-#define ZXDH_VRING_DESC_F_WRITE   2
+#define ZXDH_VRING_DESC_F_WRITE2
+
+/* This means the buffer contains a list of buffer descriptors. */
+#define ZXDH_VRING_DESC_F_INDIRECT 4
+
 /* This flag means the descriptor was made available by the driver */
 #define ZXDH_VRING_PACKED_DESC_F_AVAIL   (1 << (7))
 #define ZXDH_VRING_PACKED_DESC_F_USED(1 << (15))
@@ -35,11 +42,17 @@ enum { ZXDH_VTNET_RQ = 0, ZXDH_VTNET_TQ = 1 };
 #define ZXDH_RING_EVENT_FLAGS_DISABLE 0x1
 #define ZXDH_RING_EVENT_FLAGS_DESC0x2
 
+#define ZXDH_RING_F_INDIRECT_DESC 28
+
 #define ZXDH_VQ_RING_DESC_CHAIN_END   32768
 #define ZXDH_QUEUE_DEPTH  1024
 
 #define ZXDH_RQ_QUEUE_IDX 0
 #define ZXDH_TQ_QUEUE_IDX 1
+#define ZXDH_TYPE_HDR_SIZEsizeof(struct zxdh_type_hdr)
+#define ZXDH_PI_HDR_SIZE  sizeof(struct zxdh_pi_hdr)
+#define ZXDH_DL_NET_HDR_SIZE  sizeof(struct zxdh_net_hdr_dl)
+#define ZXDH_UL_NET_HDR_SIZE  sizeof(struct zxdh_net_hdr_ul)
 
 /*
  * ring descriptors: 16 bytes.
@@ -355,6 +368,17 @@ static inline void zxdh_queue_notify(struct zxdh_virtqueue 
*vq)
ZXDH_VTPCI_OPS(vq->hw)->notify_queue(vq->hw, vq);
 }
 
+static inline int32_t
+zxdh_queue_kick_prepare_packed(struct zxdh_virtqueue *vq)
+{
+   uint16_t flags = 0;
+
+   zxdh_mb(vq->hw->weak_barriers);
+   flags = vq->vq_packed.ring.device->desc_event_flags;
+
+   return (flags != ZXDH_RING_EVENT_FLAGS_DISABLE);
+}
+
 struct rte_mbuf *zxdh_queue_detach_unused(struct zxdh_virtqueue *vq);
 int32_t zxdh_free_queues(struct rte_eth_dev *dev);
 int32_t zxdh_get_queue_type(uint16_t vtpci_queue_idx);
diff --git a/drivers/net/zxdh/zxdh_rxtx.c b/drivers/net/zxdh/zxdh_rxtx.c
new file mode 100644
index 00..10034a0e98
--- /dev/null
+++ b/drivers/net/zxdh/zxdh_rxtx.c
@@ -0,0 +1,396 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2024 ZTE Corporation
+ */
+
+#include 
+#include 
+
+#include 
+
+#include "zxdh_logs.h"
+#include "zxdh_pci.h"
+#include "zxdh_queue.h"
+
+#define ZXDH_PKT_FORM_CPU 0x20/* 1-cpu 0-np */
+#define ZXDH_NO_IP_FRAGMENT   0x2000  /* ip fragment flag */
+#define ZXDH_NO_IPID_UPDATE

[PATCH v4 12/15] net/zxdh: vlan filter/ offload ops implementations

2024-12-18 Thread Junlong Wang
provided vlan filter, vlan offload ops.

Signed-off-by: Junlong Wang 
---
 doc/guides/nics/features/zxdh.ini  |   3 +
 doc/guides/nics/zxdh.rst   |   3 +
 drivers/net/zxdh/zxdh_ethdev.c |  40 +-
 drivers/net/zxdh/zxdh_ethdev_ops.c | 223 +
 drivers/net/zxdh/zxdh_ethdev_ops.h |   2 +
 drivers/net/zxdh/zxdh_msg.h|  22 +++
 drivers/net/zxdh/zxdh_rxtx.c   |  18 +++
 drivers/net/zxdh/zxdh_tables.c |  99 +
 drivers/net/zxdh/zxdh_tables.h |  10 +-
 9 files changed, 417 insertions(+), 3 deletions(-)

diff --git a/doc/guides/nics/features/zxdh.ini 
b/doc/guides/nics/features/zxdh.ini
index e9b237e102..6fb006c2da 100644
--- a/doc/guides/nics/features/zxdh.ini
+++ b/doc/guides/nics/features/zxdh.ini
@@ -16,3 +16,6 @@ Unicast MAC filter   = Y
 Multicast MAC filter = Y
 Promiscuous mode = Y
 Allmulticast mode= Y
+VLAN filter  = Y
+VLAN offload = Y
+QinQ offload = Y
diff --git a/doc/guides/nics/zxdh.rst b/doc/guides/nics/zxdh.rst
index 0399df1302..3a7585d123 100644
--- a/doc/guides/nics/zxdh.rst
+++ b/doc/guides/nics/zxdh.rst
@@ -28,6 +28,9 @@ Features of the ZXDH PMD are:
 - Multicast MAC filter
 - Promiscuous mode
 - Multicast mode
+- VLAN filter and VLAN offload
+- VLAN stripping and inserting
+- QINQ stripping and inserting
 
 
 Driver compilation and testing
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index 1d64b877c1..cc32b467a9 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -758,6 +758,34 @@ zxdh_alloc_queues(struct rte_eth_dev *dev, uint16_t nr_vq)
return 0;
 }
 
+static int
+zxdh_vlan_offload_configure(struct rte_eth_dev *dev)
+{
+   int ret;
+   int mask = RTE_ETH_VLAN_STRIP_MASK | RTE_ETH_VLAN_FILTER_MASK | 
RTE_ETH_QINQ_STRIP_MASK;
+
+   ret = zxdh_dev_vlan_offload_set(dev, mask);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "vlan offload set error");
+   return -1;
+   }
+
+   return 0;
+}
+
+static int
+zxdh_dev_conf_offload(struct rte_eth_dev *dev)
+{
+   int ret = 0;
+
+   ret = zxdh_vlan_offload_configure(dev);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "zxdh_vlan_offload_configure failed");
+   return ret;
+   }
+
+   return 0;
+}
 
 static int32_t
 zxdh_dev_configure(struct rte_eth_dev *dev)
@@ -815,7 +843,7 @@ zxdh_dev_configure(struct rte_eth_dev *dev)
 
nr_vq = dev->data->nb_rx_queues + dev->data->nb_tx_queues;
if (nr_vq == hw->queue_num)
-   return 0;
+   goto end;
 
PMD_DRV_LOG(DEBUG, "queue changed need reset ");
/* Reset the device although not necessary at startup */
@@ -847,6 +875,8 @@ zxdh_dev_configure(struct rte_eth_dev *dev)
 
zxdh_pci_reinit_complete(hw);
 
+end:
+   zxdh_dev_conf_offload(dev);
return ret;
 }
 
@@ -1067,6 +1097,8 @@ static const struct eth_dev_ops zxdh_eth_dev_ops = {
.promiscuous_disable = zxdh_dev_promiscuous_disable,
.allmulticast_enable = zxdh_dev_allmulticast_enable,
.allmulticast_disable= zxdh_dev_allmulticast_disable,
+   .vlan_filter_set = zxdh_dev_vlan_filter_set,
+   .vlan_offload_set= zxdh_dev_vlan_offload_set,
 };
 
 static int32_t
@@ -1325,6 +1357,12 @@ zxdh_tables_init(struct rte_eth_dev *dev)
return ret;
}
 
+   ret = zxdh_vlan_filter_table_init(dev);
+   if (ret) {
+   PMD_DRV_LOG(ERR, " vlan filter table init failed");
+   return ret;
+   }
+
return ret;
 }
 
diff --git a/drivers/net/zxdh/zxdh_ethdev_ops.c 
b/drivers/net/zxdh/zxdh_ethdev_ops.c
index aed4e6410c..94c5e6dbc8 100644
--- a/drivers/net/zxdh/zxdh_ethdev_ops.c
+++ b/drivers/net/zxdh/zxdh_ethdev_ops.c
@@ -2,6 +2,8 @@
  * Copyright(c) 2024 ZTE Corporation
  */
 
+#include 
+
 #include "zxdh_ethdev.h"
 #include "zxdh_pci.h"
 #include "zxdh_msg.h"
@@ -9,6 +11,8 @@
 #include "zxdh_tables.h"
 #include "zxdh_logs.h"
 
+#define ZXDH_VLAN_FILTER_GROUPS   64
+
 static int32_t zxdh_config_port_status(struct rte_eth_dev *dev, uint16_t 
link_status)
 {
struct zxdh_hw *hw = dev->data->dev_private;
@@ -529,3 +533,222 @@ int zxdh_dev_allmulticast_disable(struct rte_eth_dev *dev)
hw->allmulti_status = 0;
return ret;
 }
+
+int
+zxdh_dev_vlan_filter_set(struct rte_eth_dev *dev, uint16_t vlan_id, int on)
+{
+   struct zxdh_hw *hw = (struct zxdh_hw *)dev->data->dev_private;
+   uint16_t idx = 0;
+   uint16_t bit_idx = 0;
+   uint8_t msg_type = 0;
+   int ret = 0;
+
+   vlan_id &= RTE_VLAN_ID_MASK;
+   if (vlan_id == 0 || vlan_id == RTE_ETHER_MAX_VLAN_ID) {
+   PMD_DRV_LOG(ERR, "vlan id (%d) is reserved", vlan_id);
+   return -EINVAL;
+   }
+
+   if (dev->data->dev_started == 0) {
+   PMD_DRV_LOG(ERR, "vlan_filter dev not start");
+   

[PATCH v4 15/15] net/zxdh: mtu update ops implementations

2024-12-18 Thread Junlong Wang
mtu update ops implementations.

Signed-off-by: Junlong Wang 
---
 doc/guides/nics/features/zxdh.ini  |  1 +
 doc/guides/nics/zxdh.rst   |  2 +
 drivers/net/zxdh/zxdh_ethdev.c |  5 ++
 drivers/net/zxdh/zxdh_ethdev_ops.c | 78 ++
 drivers/net/zxdh/zxdh_ethdev_ops.h |  3 ++
 drivers/net/zxdh/zxdh_tables.c | 42 
 drivers/net/zxdh/zxdh_tables.h |  4 ++
 7 files changed, 135 insertions(+)

diff --git a/doc/guides/nics/features/zxdh.ini 
b/doc/guides/nics/features/zxdh.ini
index 98c141cf95..3561e31666 100644
--- a/doc/guides/nics/features/zxdh.ini
+++ b/doc/guides/nics/features/zxdh.ini
@@ -24,3 +24,4 @@ RSS reta update  = Y
 Inner RSS= Y
 Basic stats  = Y
 Stats per queue  = Y
+MTU update   = Y
diff --git a/doc/guides/nics/zxdh.rst b/doc/guides/nics/zxdh.rst
index c8a52b587c..58e0c49a2e 100644
--- a/doc/guides/nics/zxdh.rst
+++ b/doc/guides/nics/zxdh.rst
@@ -33,6 +33,8 @@ Features of the ZXDH PMD are:
 - QINQ stripping and inserting
 - Receive Side Scaling (RSS)
 - Port hardware statistics
+- MTU update
+- Jumbo frames
 
 
 Driver compilation and testing
diff --git a/drivers/net/zxdh/zxdh_ethdev.c b/drivers/net/zxdh/zxdh_ethdev.c
index a1822e1556..2906e3be6e 100644
--- a/drivers/net/zxdh/zxdh_ethdev.c
+++ b/drivers/net/zxdh/zxdh_ethdev.c
@@ -64,6 +64,10 @@ zxdh_dev_infos_get(struct rte_eth_dev *dev,
dev_info->reta_size = RTE_ETH_RSS_RETA_SIZE_256;
dev_info->flow_type_rss_offloads = ZXDH_RSS_HF;
 
+   dev_info->max_mtu = ZXDH_MAX_RX_PKTLEN - RTE_ETHER_HDR_LEN -
+   RTE_VLAN_HLEN - ZXDH_DL_NET_HDR_SIZE;
+   dev_info->min_mtu = ZXDH_ETHER_MIN_MTU;
+
dev_info->tx_offload_capa = (RTE_ETH_TX_OFFLOAD_MULTI_SEGS);
dev_info->tx_offload_capa |= (RTE_ETH_TX_OFFLOAD_TCP_TSO |
RTE_ETH_TX_OFFLOAD_UDP_TSO);
@@ -1153,6 +1157,7 @@ static const struct eth_dev_ops zxdh_eth_dev_ops = {
.rss_hash_conf_get   = zxdh_rss_hash_conf_get,
.stats_get   = zxdh_dev_stats_get,
.stats_reset = zxdh_dev_stats_reset,
+   .mtu_set = zxdh_dev_mtu_set,
 };
 
 static int32_t
diff --git a/drivers/net/zxdh/zxdh_ethdev_ops.c 
b/drivers/net/zxdh/zxdh_ethdev_ops.c
index 2c10f171aa..8e72a9a6b2 100644
--- a/drivers/net/zxdh/zxdh_ethdev_ops.c
+++ b/drivers/net/zxdh/zxdh_ethdev_ops.c
@@ -13,6 +13,7 @@
 #include "zxdh_logs.h"
 #include "zxdh_rxtx.h"
 #include "zxdh_np.h"
+#include "zxdh_queue.h"
 
 #define ZXDH_VLAN_FILTER_GROUPS   64
 #define ZXDH_INVALID_LOGIC_QID0xU
@@ -1515,3 +1516,80 @@ int zxdh_dev_stats_reset(struct rte_eth_dev *dev)
 
return 0;
 }
+
+int zxdh_dev_mtu_set(struct rte_eth_dev *dev, uint16_t new_mtu)
+{
+   struct zxdh_hw *hw = dev->data->dev_private;
+   struct zxdh_panel_table panel = {0};
+   struct zxdh_port_attr_table vport_att = {0};
+   uint16_t vfid = zxdh_vport_to_vfid(hw->vport);
+   uint16_t max_mtu = 0;
+   int ret = 0;
+
+   max_mtu = ZXDH_MAX_RX_PKTLEN - RTE_ETHER_HDR_LEN - RTE_VLAN_HLEN - 
ZXDH_DL_NET_HDR_SIZE;
+   if (new_mtu < ZXDH_ETHER_MIN_MTU || new_mtu > max_mtu) {
+   PMD_DRV_LOG(ERR, "invalid mtu:%d, range[%d, %d]",
+   new_mtu, ZXDH_ETHER_MIN_MTU, max_mtu);
+   return -EINVAL;
+   }
+
+   if (dev->data->mtu == new_mtu)
+   return 0;
+
+   if (hw->is_pf) {
+   memset(&panel, 0, sizeof(panel));
+   memset(&vport_att, 0, sizeof(vport_att));
+   ret = zxdh_get_panel_attr(dev, &panel);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR, "get_panel_attr ret:%d", ret);
+   return -1;
+   }
+
+   ret = zxdh_get_port_attr(vfid, &vport_att);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR,
+   "[vfid:%d] zxdh_dev_mtu, get vport dpp_ret:%d", 
vfid, ret);
+   return -1;
+   }
+
+   panel.mtu = new_mtu;
+   panel.mtu_enable = 1;
+   ret = zxdh_set_panel_attr(dev, &panel);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR, "set zxdh_dev_mtu failed, ret:%u", 
ret);
+   return ret;
+   }
+
+   vport_att.mtu_enable = 1;
+   vport_att.mtu = new_mtu;
+   ret = zxdh_set_port_attr(vfid, &vport_att);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR,
+   "[vfid:%d] zxdh_dev_mtu, set vport dpp_ret:%d", 
vfid, ret);
+   return ret;
+   }
+   } else {
+   struct zxdh_msg_info msg_info = {0};
+   struct zxdh_port_attr_set_msg *attr_msg = 
&msg_info.data.port_attr_msg;
+
+   zxdh_ms

[PATCH v2] net/nfp: implement the burst mode get operation

2024-12-18 Thread Chaoyong He
Implement the burst mode get operation functions for both Rx and Tx.

Signed-off-by: Chaoyong He 
Reviewed-by: Peng Zhang 
---
v2:
* Replace 'snprintf()' with 'strlcpy()'.
---
 drivers/net/nfp/nfp_ethdev.c|  2 ++
 drivers/net/nfp/nfp_ethdev_vf.c |  2 ++
 drivers/net/nfp/nfp_rxtx.c  | 46 +
 drivers/net/nfp/nfp_rxtx.h  |  4 +++
 4 files changed, 54 insertions(+)

diff --git a/drivers/net/nfp/nfp_ethdev.c b/drivers/net/nfp/nfp_ethdev.c
index f54483822f..df5482f74a 100644
--- a/drivers/net/nfp/nfp_ethdev.c
+++ b/drivers/net/nfp/nfp_ethdev.c
@@ -985,6 +985,8 @@ static const struct eth_dev_ops nfp_net_eth_dev_ops = {
.get_module_eeprom  = nfp_net_get_module_eeprom,
.dev_led_on = nfp_net_led_on,
.dev_led_off= nfp_net_led_off,
+   .rx_burst_mode_get  = nfp_net_rx_burst_mode_get,
+   .tx_burst_mode_get  = nfp_net_tx_burst_mode_get,
 };
 
 static inline void
diff --git a/drivers/net/nfp/nfp_ethdev_vf.c b/drivers/net/nfp/nfp_ethdev_vf.c
index 36b98dc0c2..23fa5b82ad 100644
--- a/drivers/net/nfp/nfp_ethdev_vf.c
+++ b/drivers/net/nfp/nfp_ethdev_vf.c
@@ -232,6 +232,8 @@ static const struct eth_dev_ops nfp_netvf_eth_dev_ops = {
.tx_queue_release   = nfp_net_tx_queue_release,
.rx_queue_intr_enable   = nfp_rx_queue_intr_enable,
.rx_queue_intr_disable  = nfp_rx_queue_intr_disable,
+   .rx_burst_mode_get  = nfp_net_rx_burst_mode_get,
+   .tx_burst_mode_get  = nfp_net_tx_burst_mode_get,
 };
 
 static inline void
diff --git a/drivers/net/nfp/nfp_rxtx.c b/drivers/net/nfp/nfp_rxtx.c
index 35fb637b21..57c0b7351b 100644
--- a/drivers/net/nfp/nfp_rxtx.c
+++ b/drivers/net/nfp/nfp_rxtx.c
@@ -12,6 +12,7 @@
 
 #include "nfd3/nfp_nfd3.h"
 #include "nfdk/nfp_nfdk.h"
+#include "nfdk/nfp_nfdk_vec.h"
 #include "flower/nfp_flower.h"
 
 #include "nfp_ipsec.h"
@@ -893,3 +894,48 @@ nfp_net_recv_pkts_set(struct rte_eth_dev *eth_dev)
else
eth_dev->rx_pkt_burst = nfp_net_recv_pkts;
 }
+
+int
+nfp_net_rx_burst_mode_get(struct rte_eth_dev *eth_dev,
+   uint16_t queue_id __rte_unused,
+   struct rte_eth_burst_mode *mode)
+{
+   eth_rx_burst_t pkt_burst;
+
+   pkt_burst = eth_dev->rx_pkt_burst;
+   if (pkt_burst == nfp_net_recv_pkts) {
+   strlcpy(mode->info, "Scalar",
+   RTE_ETH_BURST_MODE_INFO_SIZE);
+   } else if (pkt_burst == nfp_net_vec_avx2_recv_pkts) {
+   strlcpy(mode->info, "Vector AVX2",
+   RTE_ETH_BURST_MODE_INFO_SIZE);
+   } else {
+   return -EINVAL;
+   }
+
+   return 0;
+}
+
+int
+nfp_net_tx_burst_mode_get(struct rte_eth_dev *eth_dev,
+   uint16_t queue_id __rte_unused,
+   struct rte_eth_burst_mode *mode)
+{
+   eth_tx_burst_t pkt_burst;
+
+   pkt_burst = eth_dev->tx_pkt_burst;
+   if (pkt_burst == nfp_net_nfd3_xmit_pkts) {
+   strlcpy(mode->info, "NFD3 Scalar",
+   RTE_ETH_BURST_MODE_INFO_SIZE);
+   } else if (pkt_burst == nfp_net_nfdk_xmit_pkts) {
+   strlcpy(mode->info, "NFDk Scalar",
+   RTE_ETH_BURST_MODE_INFO_SIZE);
+   } else if (pkt_burst == nfp_net_nfdk_vec_avx2_xmit_pkts) {
+   strlcpy(mode->info, "NFDk Vector AVX2",
+   RTE_ETH_BURST_MODE_INFO_SIZE);
+   } else {
+   return -EINVAL;
+   }
+
+   return 0;
+}
diff --git a/drivers/net/nfp/nfp_rxtx.h b/drivers/net/nfp/nfp_rxtx.h
index c717d97003..48cbd83787 100644
--- a/drivers/net/nfp/nfp_rxtx.h
+++ b/drivers/net/nfp/nfp_rxtx.h
@@ -245,6 +245,10 @@ void nfp_net_tx_queue_info_get(struct rte_eth_dev *dev,
uint16_t queue_id,
struct rte_eth_txq_info *qinfo);
 void nfp_net_recv_pkts_set(struct rte_eth_dev *eth_dev);
+int nfp_net_rx_burst_mode_get(struct rte_eth_dev *eth_dev, uint16_t queue_id,
+   struct rte_eth_burst_mode *mode);
+int nfp_net_tx_burst_mode_get(struct rte_eth_dev *eth_dev, uint16_t queue_id,
+   struct rte_eth_burst_mode *mode);
 void nfp_net_parse_ptype(struct nfp_net_rxq *rxq,
struct nfp_net_rx_desc *rxds,
struct rte_mbuf *mb);
-- 
2.43.5



Re: [PATCH] net/gve: Allocate qpl pages using malloc if memzone allocation fails

2024-12-18 Thread Stephen Hemminger
On Wed, 18 Dec 2024 15:46:35 -0800
Joshua Washington  wrote:

> From: Praveen Kaligineedi 
> 
> Allocating QPL for an RX queue might fail if enough contiguous IOVA
> memory cannot be allocated. However, the only requirement for QPL
> for RX is that each 4K buffer be IOVA contiguous, not the entire
> QPL. Therefore, use malloc to allocate 4K buffers if the allocation
> using memzone fails.
> 
> Signed-off-by: Praveen Kaligineedi 
> Reviewed-by: Joshua Washington 
> ---

Why keep the memzone code? rte_malloc and memzone are both coming from
huge pages. Is there any advantage to memzone for what you are doing?

Better to not have two potential allocation paths to test.


RE: [PATCH v2 2/2] net/bonding: add command to set dedicated queue size

2024-12-18 Thread Chaoyong He
> > On Wed, 4 Dec 2024 06:21:00 +
> > Chaoyong He  wrote:
> >
> > > > The definition of what a "dedicated queue" is a bit confusing.
> > > > If it is only for LACP packets, it should never need to be very big.
> > > > Only under a mis-configuration and DoS kind of flood should there
> > > > ever be many packets.
> > >
> > > Yes, the dedicated queue is only for LACP packets now and it doesn't
> > > need be
> > set very big.
> > >
> > > But if we use a hardware queue as the "dedicated queue", we must
> > > consider the hardware capability. The minimum queue size of some
> > > NICs may be larger than the hardcode dedicated queue size. In this
> > > case, I think it
> > is better to add an interface to set the dedicated queue size.
> >
> > How about using the existing descriptor queue limits api for that?
> > It is reported by info get
> 
> Using existing descriptor queue limits api is good enough for current
> problem(hardware capability), but I think it is not very flexible.
> Now we use a macro as a default value for dedicated queue size, but we can
> replace the macro with queue limit while still retaining the interface for
> modifying queue size.
> What do you think of this?

A gentle ping ~


[PATCH 1/1] vhost: fix a double fetch when dequeue offloading

2024-12-18 Thread Yunjian Wang
The hdr->csum_start does two successive reads from user space to read a
variable length data structure. The result overflow if the data structure
changes between the two reads.

To fix this, we can prevent double fetch issue by copying virtio_hdr to
the temporary variable.

Fixes: 4dc4e33ffa10 ("net/virtio: fix Rx checksum calculation")
Cc: sta...@dpdk.org

Signed-off-by: Yunjian Wang 
---
 lib/vhost/virtio_net.c | 13 -
 1 file changed, 8 insertions(+), 5 deletions(-)

diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index 69901ab3b5..5c40ae7069 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -2914,10 +2914,12 @@ desc_to_mbuf(struct virtio_net *dev, struct 
vhost_virtqueue *vq,
 * in a contiguous virtual area.
 */
copy_vnet_hdr_from_desc(&tmp_hdr, buf_vec);
-   hdr = &tmp_hdr;
} else {
-   hdr = (struct virtio_net_hdr 
*)((uintptr_t)buf_vec[0].buf_addr);
+   rte_memcpy((void *)(uintptr_t)&tmp_hdr,
+   (void *)(uintptr_t)buf_vec[0].buf_addr,
+   sizeof(struct virtio_net_hdr));
}
+   hdr = &tmp_hdr;
}
 
for (vec_idx = 0; vec_idx < nr_vec; vec_idx++) {
@@ -3363,7 +3365,7 @@ virtio_dev_tx_batch_packed(struct virtio_net *dev,
 {
uint16_t avail_idx = vq->last_avail_idx;
uint32_t buf_offset = sizeof(struct virtio_net_hdr_mrg_rxbuf);
-   struct virtio_net_hdr *hdr;
+   struct virtio_net_hdr hdr;
uintptr_t desc_addrs[PACKED_BATCH_SIZE];
uint16_t ids[PACKED_BATCH_SIZE];
uint16_t i;
@@ -3382,8 +3384,9 @@ virtio_dev_tx_batch_packed(struct virtio_net *dev,
 
if (virtio_net_with_host_offload(dev)) {
vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
-   hdr = (struct virtio_net_hdr *)(desc_addrs[i]);
-   vhost_dequeue_offload(dev, hdr, pkts[i], 
legacy_ol_flags);
+   rte_memcpy((void *)(uintptr_t)&hdr,
+   (void *)(uintptr_t)desc_addrs[i], sizeof(struct 
virtio_net_hdr));
+   vhost_dequeue_offload(dev, &hdr, pkts[i], 
legacy_ol_flags);
}
}
 
-- 
2.33.0



[RFC PATCH] eventdev: adapter API to configure multiple Rx queues

2024-12-18 Thread Shijith Thotton
This RFC introduces a new API, rte_event_eth_rx_adapter_queues_add(),
designed to enhance the flexibility of configuring multiple Rx queues
in eventdev Rx adapter.

The existing rte_event_eth_rx_adapter_queue_add() API supports adding
multiple queues by specifying rx_queue_id = -1, but it lacks the ability
to apply specific configurations to each of the added queues.

The proposed API, rte_event_eth_rx_adapter_queues_add, addresses this
limitation by:

- Enabling users to specify an array of rx_queue_id values alongside
  individual configurations for each queue.

- Supporting a nb_rx_queues argument to define the number of queues to
  configure. When set to 0, the API applies a common configuration to
  all queues, similar to the existing rx_queue_id = -1 behavior.

This enhancement allows for more granular control when configuring
multiple Rx queues. Additionally, the API can act as a replacement for
the older API, offering both flexibility and improved functionality.

Signed-off-by: Shijith Thotton 
---
 lib/eventdev/eventdev_pmd.h | 34 +
 lib/eventdev/rte_event_eth_rx_adapter.h | 34 +
 2 files changed, 68 insertions(+)

diff --git a/lib/eventdev/eventdev_pmd.h b/lib/eventdev/eventdev_pmd.h
index 36148f8d86..2e458a9779 100644
--- a/lib/eventdev/eventdev_pmd.h
+++ b/lib/eventdev/eventdev_pmd.h
@@ -25,6 +25,7 @@
 #include 
 
 #include "event_timer_adapter_pmd.h"
+#include "rte_event_eth_rx_adapter.h"
 #include "rte_eventdev.h"
 
 #ifdef __cplusplus
@@ -708,6 +709,37 @@ typedef int (*eventdev_eth_rx_adapter_queue_add_t)(
int32_t rx_queue_id,
const struct rte_event_eth_rx_adapter_queue_conf *queue_conf);
 
+/**
+ * Add ethernet Rx queues to event device. This callback is invoked if
+ * the caps returned from rte_eventdev_eth_rx_adapter_caps_get(, eth_port_id)
+ * has RTE_EVENT_ETH_RX_ADAPTER_CAP_INTERNAL_PORT set.
+ *
+ * @param dev
+ *   Event device pointer
+ *
+ * @param eth_dev
+ *   Ethernet device pointer
+ *
+ * @param rx_queue_id
+ *   Ethernet device receive queue index array
+ *
+ * @param queue_conf
+ *   Additional configuration structure array
+ *
+ * @param nb_rx_queues
+ *   Number of ethernet device receive queues
+ *
+ * @return
+ *   - 0: Success, ethernet receive queues added successfully.
+ *   - <0: Error code returned by the driver function.
+ */
+typedef int (*eventdev_eth_rx_adapter_queues_add_t)(
+   const struct rte_eventdev *dev,
+   const struct rte_eth_dev *eth_dev,
+   int32_t rx_queue_id[],
+   const struct rte_event_eth_rx_adapter_queue_conf queue_conf[],
+   uint16_t nb_rx_queues);
+
 /**
  * Delete ethernet Rx queues from event device. This callback is invoked if
  * the caps returned from eventdev_eth_rx_adapter_caps_get(, eth_port_id)
@@ -1578,6 +1610,8 @@ struct eventdev_ops {
/**< Get ethernet Rx adapter capabilities */
eventdev_eth_rx_adapter_queue_add_t eth_rx_adapter_queue_add;
/**< Add Rx queues to ethernet Rx adapter */
+   eventdev_eth_rx_adapter_queues_add_t eth_rx_adapter_queues_add;
+   /**< Add Rx queues to ethernet Rx adapter */
eventdev_eth_rx_adapter_queue_del_t eth_rx_adapter_queue_del;
/**< Delete Rx queues from ethernet Rx adapter */
eventdev_eth_rx_adapter_queue_conf_get_t eth_rx_adapter_queue_conf_get;
diff --git a/lib/eventdev/rte_event_eth_rx_adapter.h 
b/lib/eventdev/rte_event_eth_rx_adapter.h
index 9237e198a7..9a5c560b67 100644
--- a/lib/eventdev/rte_event_eth_rx_adapter.h
+++ b/lib/eventdev/rte_event_eth_rx_adapter.h
@@ -553,6 +553,40 @@ int rte_event_eth_rx_adapter_queue_add(uint8_t id,
int32_t rx_queue_id,
const struct rte_event_eth_rx_adapter_queue_conf *conf);
 
+/**
+ * Add multiple receive queues to an event adapter.
+ *
+ * @param id
+ *  Adapter identifier.
+ *
+ * @param eth_dev_id
+ *  Port identifier of Ethernet device.
+ *
+ * @param rx_queue_id
+ *  Array of Ethernet device receive queue indices.
+ *  If nb_rx_queues is 0, then rx_queue_id is ignored.
+ *
+ * @param conf
+ *  Array of additional configuration structures of type
+ *  *rte_event_eth_rx_adapter_queue_conf*. conf[i] is used for rx_queue_id[i].
+ *  If nb_rx_queues is 0, then conf[0] is used for all Rx queues.
+ *
+ * @param nb_rx_queues
+ *  Number of receive queues to add.
+ *  If nb_rx_queues is 0, then all Rx queues configured for
+ *  the device are added with the same configuration in conf[0].
+ * @see RTE_EVENT_ETH_RX_ADAPTER_CAP_MULTI_EVENTQ
+ *
+ * @return
+ *  - 0: Success, Receive queues added correctly.
+ *  - <0: Error code on failure.
+ */
+__rte_experimental
+int rte_event_eth_rx_adapter_queues_add(
+   uint8_t id, uint16_t eth_dev_id, int32_t rx_queue_id[],
+   const struct rte_event_eth_rx_adapter_queue_conf conf[],
+   uint16_t nb_rx_queues);
+

[RFC PATCH] eventdev: adapter API to configure multiple Rx queues

2024-12-18 Thread Shijith Thotton
This RFC introduces a new API, rte_event_eth_rx_adapter_queues_add(),
designed to enhance the flexibility of configuring multiple Rx queues
in eventdev Rx adapter.

The existing rte_event_eth_rx_adapter_queue_add() API supports adding
multiple queues by specifying rx_queue_id = -1, but it lacks the ability
to apply specific configurations to each of the added queues.

The proposed API, rte_event_eth_rx_adapter_queues_add, addresses this
limitation by:

- Enabling users to specify an array of rx_queue_id values alongside
  individual configurations for each queue.

- Supporting a nb_rx_queues argument to define the number of queues to
  configure. When set to 0, the API applies a common configuration to
  all queues, similar to the existing rx_queue_id = -1 behavior.

This enhancement allows for more granular control when configuring
multiple Rx queues. Additionally, the API can act as a replacement for
the older API, offering both flexibility and improved functionality.

Signed-off-by: Shijith Thotton 
---
 lib/eventdev/eventdev_pmd.h | 34 +
 lib/eventdev/rte_event_eth_rx_adapter.h | 34 +
 2 files changed, 68 insertions(+)

diff --git a/lib/eventdev/eventdev_pmd.h b/lib/eventdev/eventdev_pmd.h
index 36148f8d86..2e458a9779 100644
--- a/lib/eventdev/eventdev_pmd.h
+++ b/lib/eventdev/eventdev_pmd.h
@@ -25,6 +25,7 @@
 #include 
 
 #include "event_timer_adapter_pmd.h"
+#include "rte_event_eth_rx_adapter.h"
 #include "rte_eventdev.h"
 
 #ifdef __cplusplus
@@ -708,6 +709,37 @@ typedef int (*eventdev_eth_rx_adapter_queue_add_t)(
int32_t rx_queue_id,
const struct rte_event_eth_rx_adapter_queue_conf *queue_conf);
 
+/**
+ * Add ethernet Rx queues to event device. This callback is invoked if
+ * the caps returned from rte_eventdev_eth_rx_adapter_caps_get(, eth_port_id)
+ * has RTE_EVENT_ETH_RX_ADAPTER_CAP_INTERNAL_PORT set.
+ *
+ * @param dev
+ *   Event device pointer
+ *
+ * @param eth_dev
+ *   Ethernet device pointer
+ *
+ * @param rx_queue_id
+ *   Ethernet device receive queue index array
+ *
+ * @param queue_conf
+ *   Additional configuration structure array
+ *
+ * @param nb_rx_queues
+ *   Number of ethernet device receive queues
+ *
+ * @return
+ *   - 0: Success, ethernet receive queues added successfully.
+ *   - <0: Error code returned by the driver function.
+ */
+typedef int (*eventdev_eth_rx_adapter_queues_add_t)(
+   const struct rte_eventdev *dev,
+   const struct rte_eth_dev *eth_dev,
+   int32_t rx_queue_id[],
+   const struct rte_event_eth_rx_adapter_queue_conf queue_conf[],
+   uint16_t nb_rx_queues);
+
 /**
  * Delete ethernet Rx queues from event device. This callback is invoked if
  * the caps returned from eventdev_eth_rx_adapter_caps_get(, eth_port_id)
@@ -1578,6 +1610,8 @@ struct eventdev_ops {
/**< Get ethernet Rx adapter capabilities */
eventdev_eth_rx_adapter_queue_add_t eth_rx_adapter_queue_add;
/**< Add Rx queues to ethernet Rx adapter */
+   eventdev_eth_rx_adapter_queues_add_t eth_rx_adapter_queues_add;
+   /**< Add Rx queues to ethernet Rx adapter */
eventdev_eth_rx_adapter_queue_del_t eth_rx_adapter_queue_del;
/**< Delete Rx queues from ethernet Rx adapter */
eventdev_eth_rx_adapter_queue_conf_get_t eth_rx_adapter_queue_conf_get;
diff --git a/lib/eventdev/rte_event_eth_rx_adapter.h 
b/lib/eventdev/rte_event_eth_rx_adapter.h
index 9237e198a7..9a5c560b67 100644
--- a/lib/eventdev/rte_event_eth_rx_adapter.h
+++ b/lib/eventdev/rte_event_eth_rx_adapter.h
@@ -553,6 +553,40 @@ int rte_event_eth_rx_adapter_queue_add(uint8_t id,
int32_t rx_queue_id,
const struct rte_event_eth_rx_adapter_queue_conf *conf);
 
+/**
+ * Add multiple receive queues to an event adapter.
+ *
+ * @param id
+ *  Adapter identifier.
+ *
+ * @param eth_dev_id
+ *  Port identifier of Ethernet device.
+ *
+ * @param rx_queue_id
+ *  Array of Ethernet device receive queue indices.
+ *  If nb_rx_queues is 0, then rx_queue_id is ignored.
+ *
+ * @param conf
+ *  Array of additional configuration structures of type
+ *  *rte_event_eth_rx_adapter_queue_conf*. conf[i] is used for rx_queue_id[i].
+ *  If nb_rx_queues is 0, then conf[0] is used for all Rx queues.
+ *
+ * @param nb_rx_queues
+ *  Number of receive queues to add.
+ *  If nb_rx_queues is 0, then all Rx queues configured for
+ *  the device are added with the same configuration in conf[0].
+ * @see RTE_EVENT_ETH_RX_ADAPTER_CAP_MULTI_EVENTQ
+ *
+ * @return
+ *  - 0: Success, Receive queues added correctly.
+ *  - <0: Error code on failure.
+ */
+__rte_experimental
+int rte_event_eth_rx_adapter_queues_add(
+   uint8_t id, uint16_t eth_dev_id, int32_t rx_queue_id[],
+   const struct rte_event_eth_rx_adapter_queue_conf conf[],
+   uint16_t nb_rx_queues);
+

RE: [RFC PATCH] eventdev: adapter API to configure multiple Rx queues

2024-12-18 Thread Shijith Thotton
>This RFC introduces a new API, rte_event_eth_rx_adapter_queues_add(),
>designed to enhance the flexibility of configuring multiple Rx queues
>in eventdev Rx adapter.
>
>The existing rte_event_eth_rx_adapter_queue_add() API supports adding
>multiple queues by specifying rx_queue_id = -1, but it lacks the ability
>to apply specific configurations to each of the added queues.
>
>The proposed API, rte_event_eth_rx_adapter_queues_add, addresses this
>limitation by:
>
>- Enabling users to specify an array of rx_queue_id values alongside
>  individual configurations for each queue.
>
>- Supporting a nb_rx_queues argument to define the number of queues to
>  configure. When set to 0, the API applies a common configuration to
>  all queues, similar to the existing rx_queue_id = -1 behavior.
>
>This enhancement allows for more granular control when configuring
>multiple Rx queues. Additionally, the API can act as a replacement for
>the older API, offering both flexibility and improved functionality.
>
>Signed-off-by: Shijith Thotton 
>---
> lib/eventdev/eventdev_pmd.h | 34 +
> lib/eventdev/rte_event_eth_rx_adapter.h | 34
>+
> 2 files changed, 68 insertions(+)
>
>diff --git a/lib/eventdev/eventdev_pmd.h b/lib/eventdev/eventdev_pmd.h
>index 36148f8d86..2e458a9779 100644
>--- a/lib/eventdev/eventdev_pmd.h
>+++ b/lib/eventdev/eventdev_pmd.h
>@@ -25,6 +25,7 @@
> #include 
>
> #include "event_timer_adapter_pmd.h"
>+#include "rte_event_eth_rx_adapter.h"
> #include "rte_eventdev.h"
>
> #ifdef __cplusplus
>@@ -708,6 +709,37 @@ typedef int
>(*eventdev_eth_rx_adapter_queue_add_t)(
>   int32_t rx_queue_id,
>   const struct rte_event_eth_rx_adapter_queue_conf
>*queue_conf);
>
>+/**
>+ * Add ethernet Rx queues to event device. This callback is invoked if
>+ * the caps returned from rte_eventdev_eth_rx_adapter_caps_get(,
>eth_port_id)
>+ * has RTE_EVENT_ETH_RX_ADAPTER_CAP_INTERNAL_PORT set.
>+ *
>+ * @param dev
>+ *   Event device pointer
>+ *
>+ * @param eth_dev
>+ *   Ethernet device pointer
>+ *
>+ * @param rx_queue_id
>+ *   Ethernet device receive queue index array
>+ *
>+ * @param queue_conf
>+ *   Additional configuration structure array
>+ *
>+ * @param nb_rx_queues
>+ *   Number of ethernet device receive queues
>+ *
>+ * @return
>+ *   - 0: Success, ethernet receive queues added successfully.
>+ *   - <0: Error code returned by the driver function.
>+ */
>+typedef int (*eventdev_eth_rx_adapter_queues_add_t)(
>+  const struct rte_eventdev *dev,
>+  const struct rte_eth_dev *eth_dev,
>+  int32_t rx_queue_id[],
>+  const struct rte_event_eth_rx_adapter_queue_conf
>queue_conf[],
>+  uint16_t nb_rx_queues);
>+
> /**
>  * Delete ethernet Rx queues from event device. This callback is invoked if
>  * the caps returned from eventdev_eth_rx_adapter_caps_get(, eth_port_id)
>@@ -1578,6 +1610,8 @@ struct eventdev_ops {
>   /**< Get ethernet Rx adapter capabilities */
>   eventdev_eth_rx_adapter_queue_add_t eth_rx_adapter_queue_add;
>   /**< Add Rx queues to ethernet Rx adapter */
>+  eventdev_eth_rx_adapter_queues_add_t eth_rx_adapter_queues_add;
>+  /**< Add Rx queues to ethernet Rx adapter */
>   eventdev_eth_rx_adapter_queue_del_t eth_rx_adapter_queue_del;
>   /**< Delete Rx queues from ethernet Rx adapter */
>   eventdev_eth_rx_adapter_queue_conf_get_t
>eth_rx_adapter_queue_conf_get;
>diff --git a/lib/eventdev/rte_event_eth_rx_adapter.h
>b/lib/eventdev/rte_event_eth_rx_adapter.h
>index 9237e198a7..9a5c560b67 100644
>--- a/lib/eventdev/rte_event_eth_rx_adapter.h
>+++ b/lib/eventdev/rte_event_eth_rx_adapter.h
>@@ -553,6 +553,40 @@ int rte_event_eth_rx_adapter_queue_add(uint8_t id,
>   int32_t rx_queue_id,
>   const struct rte_event_eth_rx_adapter_queue_conf
>*conf);
>
>+/**
>+ * Add multiple receive queues to an event adapter.
>+ *
>+ * @param id
>+ *  Adapter identifier.
>+ *
>+ * @param eth_dev_id
>+ *  Port identifier of Ethernet device.
>+ *
>+ * @param rx_queue_id
>+ *  Array of Ethernet device receive queue indices.
>+ *  If nb_rx_queues is 0, then rx_queue_id is ignored.
>+ *
>+ * @param conf
>+ *  Array of additional configuration structures of type
>+ *  *rte_event_eth_rx_adapter_queue_conf*. conf[i] is used for
>rx_queue_id[i].
>+ *  If nb_rx_queues is 0, then conf[0] is used for all Rx queues.
>+ *
>+ * @param nb_rx_queues
>+ *  Number of receive queues to add.
>+ *  If nb_rx_queues is 0, then all Rx queues configured for
>+ *  the device are added with the same configuration in conf[0].
>+ * @see RTE_EVENT_ETH_RX_ADAPTER_CAP_MULTI_EVENTQ
>+ *
>+ * @return
>+ *  - 0: Success, Receive queues added correctly.
>+ *  - <0: Error code on failure.
>+ */
>+__rte_experimental
>+int rte_event_eth_rx_adapter_queues_add(
>+  uint8_t id, uint16_t eth_dev_id, int32_t rx_queue_id[],
>+ 

[PATCH] net/gve: Allocate qpl pages using malloc if memzone allocation fails

2024-12-18 Thread Joshua Washington
From: Praveen Kaligineedi 

Allocating QPL for an RX queue might fail if enough contiguous IOVA
memory cannot be allocated. However, the only requirement for QPL
for RX is that each 4K buffer be IOVA contiguous, not the entire
QPL. Therefore, use malloc to allocate 4K buffers if the allocation
using memzone fails.

Signed-off-by: Praveen Kaligineedi 
Reviewed-by: Joshua Washington 
---
 drivers/net/gve/gve_ethdev.c | 105 
++
 drivers/net/gve/gve_ethdev.h |   1 +
 drivers/net/gve/gve_rx.c |   2 +-
 3 files changed, 89 insertions(+), 19 deletions(-)

diff --git a/drivers/net/gve/gve_ethdev.c b/drivers/net/gve/gve_ethdev.c
index db4ebe7..09304ef 100644
--- a/drivers/net/gve/gve_ethdev.c
+++ b/drivers/net/gve/gve_ethdev.c
@@ -22,42 +22,100 @@ gve_write_version(uint8_t *driver_version_register)
writeb('\n', driver_version_register);
 }

+static const struct rte_memzone *
+gve_alloc_using_mz(const char *name, uint32_t num_pages)
+{
+   const struct rte_memzone *mz;
+   mz = rte_memzone_reserve_aligned(name, num_pages * PAGE_SIZE,
+rte_socket_id(),
+RTE_MEMZONE_IOVA_CONTIG, PAGE_SIZE);
+   if (mz == NULL)
+   PMD_DRV_LOG(ERR, "Failed to alloc memzone %s.", name);
+   return mz;
+}
+
+static int
+gve_alloc_using_malloc(void **bufs, uint32_t num_entries)
+{
+   uint32_t i;
+
+   for (i = 0; i < num_entries; i++) {
+   bufs[i] = rte_malloc_socket(NULL, PAGE_SIZE, PAGE_SIZE, 
rte_socket_id());
+   if (bufs[i] == NULL) {
+   PMD_DRV_LOG(ERR, "Failed to malloc");
+   goto free_bufs;
+   }
+   }
+   return 0;
+
+free_bufs:
+   while (i > 0)
+   rte_free(bufs[--i]);
+
+   return -ENOMEM;
+}
+
 static struct gve_queue_page_list *
-gve_alloc_queue_page_list(const char *name, uint32_t num_pages)
+gve_alloc_queue_page_list(const char *name, uint32_t num_pages, bool is_rx)
 {
struct gve_queue_page_list *qpl;
const struct rte_memzone *mz;
-   dma_addr_t page_bus;
uint32_t i;

qpl = rte_zmalloc("qpl struct", sizeof(struct gve_queue_page_list), 0);
if (!qpl)
return NULL;

-   mz = rte_memzone_reserve_aligned(name, num_pages * PAGE_SIZE,
-rte_socket_id(),
-RTE_MEMZONE_IOVA_CONTIG, PAGE_SIZE);
-   if (mz == NULL) {
-   PMD_DRV_LOG(ERR, "Failed to alloc %s.", name);
-   goto free_qpl_struct;
-   }
qpl->page_buses = rte_zmalloc("qpl page buses",
num_pages * sizeof(dma_addr_t), 0);
if (qpl->page_buses == NULL) {
PMD_DRV_LOG(ERR, "Failed to alloc qpl page buses");
-   goto free_qpl_memzone;
+   goto free_qpl_struct;
}
-   page_bus = mz->iova;
-   for (i = 0; i < num_pages; i++) {
-   qpl->page_buses[i] = page_bus;
-   page_bus += PAGE_SIZE;
+
+   qpl->qpl_bufs = rte_zmalloc("qpl bufs",
+   num_pages * sizeof(void *), 0);
+   if (qpl->qpl_bufs == NULL) {
+   PMD_DRV_LOG(ERR, "Failed to alloc qpl bufs");
+   goto free_qpl_page_buses;
}
-   qpl->mz = mz;
+
+   mz = gve_alloc_using_mz(name, num_pages);
+   if (mz) {
+   qpl->mz = mz;
+
+   /* Populate the buffer addresses */
+   for (i = 0; i < num_pages; i++)
+   qpl->qpl_bufs[i] =
+   (void *)((uint64_t)(mz->addr) + i * PAGE_SIZE);
+
+   /* Populate the IOVA addresses */
+   for (i = 0; i < num_pages; i++)
+   qpl->page_buses[i] = mz->iova + i * PAGE_SIZE;
+   } else {
+   /* Allocate 4K size buffers.
+* Cannot use non-contiguous approach for tx fifo.
+*/
+   if (!is_rx)
+   goto free_qpl_page_bufs;
+
+   PMD_DRV_LOG(ERR, "Allocating bufs using malloc for %s ", name);
+   if (gve_alloc_using_malloc(qpl->qpl_bufs, num_pages))
+   goto free_qpl_page_bufs;
+
+   /* Populate the IOVA addresses */
+   for (i = 0; i < num_pages; i++)
+   qpl->page_buses[i] =
+   rte_malloc_virt2iova(qpl->qpl_bufs[i]);
+   }
+
qpl->num_entries = num_pages;
return qpl;

-free_qpl_memzone:
-   rte_memzone_free(qpl->mz);
+free_qpl_page_bufs:
+   rte_free(qpl->qpl_bufs);
+free_qpl_page_buses:
+   rte_free(qpl->page_buses);
 free_qpl_struct:
rte_free(qpl);
return NULL;
@@ -69,7 +127,18 @@ gve_free_queue_page_list(struct gve_queue_page_list *qpl)
if (qpl->mz) {
rte_memzone_free(qpl->mz);
qpl->mz