Re: [dpdk-dev] [PATCH v6 07/10] eal: add lcore init callbacks

2022-12-16 Thread David Marchand
Morten,

On Thu, Dec 15, 2022 at 11:21 AM Morten Brørup  
wrote:
> > > Shouldn't these callbacks be called from the EAL threads too, e.g.
> > from eal_thread_loop()?
> > >
> > > I looks like they are only called from eal_lcore_non_eal_allocate(),
> > which is only called from rte_thread_register().
> >
> > These should be called for already present lcores on callback
> > registration.
> > See rte_lcore_callback_register().
>
> That is completely useless! They need to be called from the thread itself, so 
> they have the correct environment, thread local storage, thread id, and 
> similar context.

If it is broken, please fix it or come with a better implementation.

At the time, existing modules were storing their per lcore/thread
private info using a local array and dereferencing it using
rte_lcore_id() or a criteria of their liking.
The callbacks have to be carefully written to explicitly initialise
per lcore/thread private info.
I thought it was enough, plus nobody objected.


Now, some things to consider for this requirement you want to add.

If we want an application to be able to register callbacks after
rte_eal_init(), we need to wake/interrupt all EAL threads from what
they are doing.
Once EAL threads enter some job the application gave, they won't
reread anything from EAL.

Or we could require that callbacks are registered before
rte_eal_init(), the init() callback could be called as the last step
once all DPDK modules are initialised but before the application calls
rte_eal_remote_launch().

But the problem of needing the break EAL threads from what they are
doing is also present when we want to call the uninit() callback.
And there, I have no good idea.


-- 
David Marchand



[RFC 0/3] Async vhost packed ring optimization

2022-12-16 Thread Cheng Jiang
To improve the performance of async vhost packed ring. We remove the
unnecessary data copy in async vhost packed ring. And add the batch
data path in both enqueue data path and dequeue data path.

Cheng Jiang (3):
  vhost: remove redundant copy for packed shadow used ring
  vhost: add batch enqueue in async vhost packed ring
  vhost: add batch dequeue in async vhost packed ring

 lib/vhost/virtio_net.c | 395 -
 1 file changed, 355 insertions(+), 40 deletions(-)

--
2.35.1



[RFC 1/3] vhost: remove redundant copy for packed shadow used ring

2022-12-16 Thread Cheng Jiang
In the packed ring enqueue data path of the current asynchronous
Vhost design, the shadow used ring is first copied to the sync
shadow used ring, and then it will be moved to the async shadow
used ring for some historical reasons. This is completely unnecessary.
This patch removes redundant copy for the shadow used ring. The async
shadow used ring will be updated directly.

Signed-off-by: Cheng Jiang 
---
 lib/vhost/virtio_net.c | 66 --
 1 file changed, 31 insertions(+), 35 deletions(-)

diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index 4358899718..22f97d4f77 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -572,6 +572,26 @@ vhost_shadow_enqueue_packed(struct vhost_virtqueue *vq,
}
 }
 
+static __rte_always_inline void
+vhost_async_shadow_enqueue_packed(struct vhost_virtqueue *vq,
+  uint32_t *len,
+  uint16_t *id,
+  uint16_t *count,
+  uint16_t num_buffers)
+{
+   uint16_t i;
+   struct vhost_async *async = vq->async;
+
+   for (i = 0; i < num_buffers; i++) {
+   async->buffers_packed[async->buffer_idx_packed].id  = id[i];
+   async->buffers_packed[async->buffer_idx_packed].len = len[i];
+   async->buffers_packed[async->buffer_idx_packed].count = 
count[i];
+   async->buffer_idx_packed++;
+   if (async->buffer_idx_packed >= vq->size)
+   async->buffer_idx_packed -= vq->size;
+   }
+}
+
 static __rte_always_inline void
 vhost_shadow_enqueue_single_packed(struct virtio_net *dev,
   struct vhost_virtqueue *vq,
@@ -1647,23 +1667,6 @@ store_dma_desc_info_split(struct vring_used_elem 
*s_ring, struct vring_used_elem
}
 }
 
-static __rte_always_inline void
-store_dma_desc_info_packed(struct vring_used_elem_packed *s_ring,
-   struct vring_used_elem_packed *d_ring,
-   uint16_t ring_size, uint16_t s_idx, uint16_t d_idx, uint16_t 
count)
-{
-   size_t elem_size = sizeof(struct vring_used_elem_packed);
-
-   if (d_idx + count <= ring_size) {
-   rte_memcpy(d_ring + d_idx, s_ring + s_idx, count * elem_size);
-   } else {
-   uint16_t size = ring_size - d_idx;
-
-   rte_memcpy(d_ring + d_idx, s_ring + s_idx, size * elem_size);
-   rte_memcpy(d_ring, s_ring + s_idx + size, (count - size) * 
elem_size);
-   }
-}
-
 static __rte_noinline uint32_t
 virtio_dev_rx_async_submit_split(struct virtio_net *dev, struct 
vhost_virtqueue *vq,
struct rte_mbuf **pkts, uint32_t count, int16_t dma_id, uint16_t 
vchan_id)
@@ -1822,7 +1825,8 @@ vhost_enqueue_async_packed(struct virtio_net *dev,
if (unlikely(mbuf_to_desc(dev, vq, pkt, buf_vec, nr_vec, *nr_buffers, 
true) < 0))
return -1;
 
-   vhost_shadow_enqueue_packed(vq, buffer_len, buffer_buf_id, 
buffer_desc_count, *nr_buffers);
+   vhost_async_shadow_enqueue_packed(vq, buffer_len, buffer_buf_id,
+   buffer_desc_count, *nr_buffers);
 
return 0;
 }
@@ -1852,6 +1856,7 @@ dma_error_handler_packed(struct vhost_virtqueue *vq, 
uint16_t slot_idx,
 {
uint16_t descs_err = 0;
uint16_t buffers_err = 0;
+   struct vhost_async *async = vq->async;
struct async_inflight_info *pkts_info = vq->async->pkts_info;
 
*pkt_idx -= nr_err;
@@ -1869,7 +1874,10 @@ dma_error_handler_packed(struct vhost_virtqueue *vq, 
uint16_t slot_idx,
vq->avail_wrap_counter ^= 1;
}
 
-   vq->shadow_used_idx -= buffers_err;
+   if (async->buffer_idx_packed >= buffers_err)
+   async->buffer_idx_packed -= buffers_err;
+   else
+   async->buffer_idx_packed = async->buffer_idx_packed + vq->size 
- buffers_err;
 }
 
 static __rte_noinline uint32_t
@@ -1923,23 +1931,11 @@ virtio_dev_rx_async_submit_packed(struct virtio_net 
*dev, struct vhost_virtqueue
dma_error_handler_packed(vq, slot_idx, pkt_err, &pkt_idx);
}
 
-   if (likely(vq->shadow_used_idx)) {
-   /* keep used descriptors. */
-   store_dma_desc_info_packed(vq->shadow_used_packed, 
async->buffers_packed,
-   vq->size, 0, async->buffer_idx_packed,
-   vq->shadow_used_idx);
-
-   async->buffer_idx_packed += vq->shadow_used_idx;
-   if (async->buffer_idx_packed >= vq->size)
-   async->buffer_idx_packed -= vq->size;
-
-   async->pkts_idx += pkt_idx;
-   if (async->pkts_idx >= vq->size)
-   async->pkts_idx -= vq->size;
+   async->pkts_idx += pkt_idx;
+   if (async->pkts_idx >= vq->size)
+   async->pkts_idx -= vq->size;
 
-   vq->shadow_use

[RFC 2/3] vhost: add batch enqueue in async vhost packed ring

2022-12-16 Thread Cheng Jiang
Add batch enqueue function in asynchronous vhost packed ring to
improve the performance. Chained mbufs are not supported, it will
be handled in single enqueue function.

Signed-off-by: Cheng Jiang 
---
 lib/vhost/virtio_net.c | 159 -
 1 file changed, 157 insertions(+), 2 deletions(-)

diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index 22f97d4f77..b87405ba54 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -432,6 +432,24 @@ vhost_flush_enqueue_batch_packed(struct virtio_net *dev,
vq_inc_last_used_packed(vq, PACKED_BATCH_SIZE);
 }
 
+static __rte_always_inline void
+vhost_async_shadow_enqueue_packed_batch(struct vhost_virtqueue *vq,
+uint64_t *lens,
+uint16_t *ids)
+{
+   uint16_t i;
+   struct vhost_async *async = vq->async;
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   async->buffers_packed[async->buffer_idx_packed].id  = ids[i];
+   async->buffers_packed[async->buffer_idx_packed].len = lens[i];
+   async->buffers_packed[async->buffer_idx_packed].count = 1;
+   async->buffer_idx_packed++;
+   if (async->buffer_idx_packed >= vq->size)
+   async->buffer_idx_packed -= vq->size;
+   }
+}
+
 static __rte_always_inline void
 vhost_shadow_dequeue_batch_packed_inorder(struct vhost_virtqueue *vq,
  uint16_t id)
@@ -1451,6 +1469,58 @@ virtio_dev_rx_sync_batch_check(struct virtio_net *dev,
return 0;
 }
 
+static __rte_always_inline int
+virtio_dev_rx_async_batch_check(struct vhost_virtqueue *vq,
+  struct rte_mbuf **pkts,
+  uint64_t *desc_addrs,
+  uint64_t *lens,
+  int16_t dma_id,
+  uint16_t vchan_id)
+{
+   bool wrap_counter = vq->avail_wrap_counter;
+   struct vring_packed_desc *descs = vq->desc_packed;
+   uint16_t avail_idx = vq->last_avail_idx;
+   uint32_t buf_offset = sizeof(struct virtio_net_hdr_mrg_rxbuf);
+   uint16_t i;
+
+   if (unlikely(avail_idx & PACKED_BATCH_MASK))
+   return -1;
+
+   if (unlikely((avail_idx + PACKED_BATCH_SIZE) > vq->size))
+   return -1;
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   if (unlikely(pkts[i]->next != NULL))
+   return -1;
+   if (unlikely(!desc_is_avail(&descs[avail_idx + i],
+   wrap_counter)))
+   return -1;
+   }
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE)
+   lens[i] = descs[avail_idx + i].len;
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   if (unlikely(pkts[i]->pkt_len > (lens[i] - buf_offset)))
+   return -1;
+   }
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE)
+   desc_addrs[i] =  descs[avail_idx + i].addr;
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   if (unlikely(!desc_addrs[i]))
+   return -1;
+   if (unlikely(lens[i] != descs[avail_idx + i].len))
+   return -1;
+   }
+
+   if (rte_dma_burst_capacity(dma_id, vchan_id) < PACKED_BATCH_SIZE)
+   return -1;
+
+   return 0;
+}
+
 static __rte_always_inline void
 virtio_dev_rx_batch_packed_copy(struct virtio_net *dev,
   struct vhost_virtqueue *vq,
@@ -1850,6 +1920,78 @@ virtio_dev_rx_async_packed(struct virtio_net *dev, 
struct vhost_virtqueue *vq,
return 0;
 }
 
+static __rte_always_inline void
+virtio_dev_rx_async_packed_batch_enqueue(struct virtio_net *dev,
+  struct vhost_virtqueue *vq,
+  struct rte_mbuf **pkts,
+  uint64_t *desc_addrs,
+  uint64_t *lens)
+{
+   uint32_t buf_offset = sizeof(struct virtio_net_hdr_mrg_rxbuf);
+   struct virtio_net_hdr_mrg_rxbuf *hdrs[PACKED_BATCH_SIZE];
+   struct vring_packed_desc *descs = vq->desc_packed;
+   struct vhost_async *async = vq->async;
+   uint16_t avail_idx = vq->last_avail_idx;
+   uint32_t mbuf_offset = 0;
+   uint16_t ids[PACKED_BATCH_SIZE];
+   uint64_t mapped_len[PACKED_BATCH_SIZE];
+   void *host_iova[PACKED_BATCH_SIZE];
+   uintptr_t desc;
+   uint16_t i;
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   rte_prefetch0((void *)(uintptr_t)desc_addrs[i]);
+   desc = vhost_iova_to_vva(dev, vq, desc_addrs[i], &lens[i], 
VHOST_ACCESS_RW);
+   hdrs[i] = (struct virtio_net_hdr_mrg_rxbuf *)(uintptr_t)desc;
+   lens[i] = pkts[i]->pkt_len +
+   sizeof(struct virtio_net_hdr_mrg_rxbuf);
+   }
+
+ 

[RFC 3/3] vhost: add batch dequeue in async vhost packed ring

2022-12-16 Thread Cheng Jiang
Add batch dequeue function in asynchronous vhost packed ring to
improve the performance. Chained mbufs are not supported, it will
be handled in single enqueue function.

Signed-off-by: Cheng Jiang 
Signed-off-by: Yuan Wang 
---
 lib/vhost/virtio_net.c | 170 -
 1 file changed, 167 insertions(+), 3 deletions(-)

diff --git a/lib/vhost/virtio_net.c b/lib/vhost/virtio_net.c
index b87405ba54..24307a4ec9 100644
--- a/lib/vhost/virtio_net.c
+++ b/lib/vhost/virtio_net.c
@@ -450,6 +450,23 @@ vhost_async_shadow_enqueue_packed_batch(struct 
vhost_virtqueue *vq,
}
 }
 
+static __rte_always_inline void
+vhost_async_shadow_dequeue_packed_batch(struct vhost_virtqueue *vq, uint16_t 
*ids)
+{
+   uint16_t i;
+   struct vhost_async *async = vq->async;
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   async->buffers_packed[async->buffer_idx_packed].id  = ids[i];
+   async->buffers_packed[async->buffer_idx_packed].len = 0;
+   async->buffers_packed[async->buffer_idx_packed].count = 1;
+
+   async->buffer_idx_packed++;
+   if (async->buffer_idx_packed >= vq->size)
+   async->buffer_idx_packed -= vq->size;
+   }
+}
+
 static __rte_always_inline void
 vhost_shadow_dequeue_batch_packed_inorder(struct vhost_virtqueue *vq,
  uint16_t id)
@@ -3193,6 +3210,80 @@ vhost_reserve_avail_batch_packed(struct virtio_net *dev,
return -1;
 }
 
+static __rte_always_inline int
+vhost_async_tx_batch_packed_check(struct virtio_net *dev,
+struct vhost_virtqueue *vq,
+struct rte_mbuf **pkts,
+uint16_t avail_idx,
+uintptr_t *desc_addrs,
+uint64_t *lens,
+uint16_t *ids,
+int16_t dma_id,
+uint16_t vchan_id)
+{
+   bool wrap = vq->avail_wrap_counter;
+   struct vring_packed_desc *descs = vq->desc_packed;
+   uint64_t buf_lens[PACKED_BATCH_SIZE];
+   uint32_t buf_offset = sizeof(struct virtio_net_hdr_mrg_rxbuf);
+   uint16_t flags, i;
+
+   if (unlikely(avail_idx & PACKED_BATCH_MASK))
+   return -1;
+   if (unlikely((avail_idx + PACKED_BATCH_SIZE) > vq->size))
+   return -1;
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   flags = descs[avail_idx + i].flags;
+   if (unlikely((wrap != !!(flags & VRING_DESC_F_AVAIL)) ||
+(wrap == !!(flags & VRING_DESC_F_USED))  ||
+(flags & PACKED_DESC_SINGLE_DEQUEUE_FLAG)))
+   return -1;
+   }
+
+   rte_atomic_thread_fence(__ATOMIC_ACQUIRE);
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE)
+   lens[i] = descs[avail_idx + i].len;
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   desc_addrs[i] = descs[avail_idx + i].addr;
+   }
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   if (unlikely(!desc_addrs[i]))
+   return -1;
+   if (unlikely((lens[i] != descs[avail_idx + i].len)))
+   return -1;
+   }
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   if (virtio_dev_pktmbuf_prep(dev, pkts[i], lens[i]))
+   goto err;
+   }
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE)
+   buf_lens[i] = pkts[i]->buf_len - pkts[i]->data_off;
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   if (unlikely(buf_lens[i] < (lens[i] - buf_offset)))
+   goto err;
+   }
+
+   vhost_for_each_try_unroll(i, 0, PACKED_BATCH_SIZE) {
+   pkts[i]->pkt_len = lens[i] - buf_offset;
+   pkts[i]->data_len = pkts[i]->pkt_len;
+   ids[i] = descs[avail_idx + i].id;
+   }
+
+   if (rte_dma_burst_capacity(dma_id, vchan_id) < PACKED_BATCH_SIZE)
+   return -1;
+
+   return 0;
+
+err:
+   return -1;
+}
+
 static __rte_always_inline int
 virtio_dev_tx_batch_packed(struct virtio_net *dev,
   struct vhost_virtqueue *vq,
@@ -3769,16 +3860,74 @@ virtio_dev_tx_async_single_packed(struct virtio_net 
*dev,
return err;
 }
 
+static __rte_always_inline int
+virtio_dev_tx_async_packed_batch(struct virtio_net *dev,
+  struct vhost_virtqueue *vq,
+  struct rte_mbuf **pkts, uint16_t slot_idx,
+  uint16_t dma_id, uint16_t vchan_id)
+{
+   uint16_t avail_idx = vq->last_avail_idx;
+   uint32_t buf_offset = sizeof(struct virtio_net_hdr_mrg_rxbuf);
+   struct vhost_async *async = vq->async;
+   struct async_inflig

Re: [PATCH] crypto/qat: add SHA3 plain hash support

2022-12-16 Thread Zhang, Fan

On 12/14/2022 4:26 PM, Ciara Power wrote:


Add support for plain SHA3-224, SHA3-256,
SHA3-384, and SHA3-512 hash support in QAT GEN3.
Add support for SHA3-256 in GEN2.

Signed-off-by: Ciara Power 

---
Tested using testcases from patch not yet merged.
https://patchwork.dpdk.org/project/dpdk/patch/20221109152809.2026484-1-vfia...@marvell.com/
---
  doc/guides/cryptodevs/features/qat.ini   |  4 ++
  doc/guides/cryptodevs/qat.rst|  4 ++
  doc/guides/rel_notes/release_23_03.rst   |  4 ++
  drivers/common/qat/qat_adf/icp_qat_hw.h  |  4 +-
  drivers/crypto/qat/dev/qat_crypto_pmd_gen2.c |  3 +
  drivers/crypto/qat/dev/qat_crypto_pmd_gen3.c | 12 
  drivers/crypto/qat/qat_sym_session.c | 60 
  7 files changed, 89 insertions(+), 2 deletions(-)

Acked-by: Fan Zhang 


Re: questions about crypto_scheduler

2022-12-16 Thread Zhang, Fan

Hi Zhangfei,

The crypto scheduler PMD's session contains "sub-sessions" for all 
works, 1 sub-session per driver ID, to minimize the memory footprint. 
When configuring session, it configures the sub-sessions one by one. 
Since your UADK devices sharing the same driver ID, the crypto scheduler 
will only create one session for them to share.


Hope this answers your question.

Regards,

Fan

On 12/16/2022 6:54 AM, Zhangfei Gao wrote:

Hi, Akhil

Excuse me for the question.

I am testing UADK crypto performance with dpdk-test-crypto-perf, and 
want to use multi-thread or multi-session for better performance, so 
trying to use crypto_scheduler.


CMD like
sudo dpdk-test-crypto-perf -l 1,2 --vdev "crypto_uadk0" --vdev 
"crypto_uadk1" --vdev 
"crypto_scheduler,worker=crypto_uadk0,worker=crypto_uadk1,mode=round-robin" 
-- --devtype crypto_scheduler --optype cipher-only --buffer-sz 8192


Though multi-worker is set but found it is just using one worker.
Since uadk_cryptodev_probe gets the same dev->driver_id = 
uadk_cryptodev_driver_id in multi-probe.


Then scheduler_pmd_sym_session_configure will treat only one worker,
and call rte_cryptodev_sym_session_create only once.
if (configured_sess[j].driver_id == worker->driver_id)
        break

Do I misuse the crypto_scheduler for the multi-thread or multi-session.

Thanks in advance.



Re: [PATCH V7 7/8] test: add test cases for adding hex integer value API

2022-12-16 Thread Bruce Richardson
On Fri, Dec 16, 2022 at 09:54:27AM +0800, Huisong Li wrote:
> Add test cases for adding hexadecimal unsigned integer value API.
> 
> Signed-off-by: Huisong Li 
> Acked-by: Morten Brørup 
> Acked-by: Chengwen Feng 
> ---

The coding style is still a little off for indentation, but otherwise ok.

Acked-by: Bruce Richardson 



RE: [dpdk-dev] [PATCH v6 07/10] eal: add lcore init callbacks

2022-12-16 Thread Morten Brørup
> From: David Marchand [mailto:david.march...@redhat.com]
> Sent: Friday, 16 December 2022 09.09
> 
> Morten,
> 
> On Thu, Dec 15, 2022 at 11:21 AM Morten Brørup
>  wrote:
> > > > Shouldn't these callbacks be called from the EAL threads too,
> e.g.
> > > from eal_thread_loop()?
> > > >
> > > > I looks like they are only called from
> eal_lcore_non_eal_allocate(),
> > > which is only called from rte_thread_register().
> > >
> > > These should be called for already present lcores on callback
> > > registration.
> > > See rte_lcore_callback_register().
> >
> > That is completely useless! They need to be called from the thread
> itself, so they have the correct environment, thread local storage,
> thread id, and similar context.
> 
> If it is broken, please fix it or come with a better implementation.
> 
> At the time, existing modules were storing their per lcore/thread
> private info using a local array and dereferencing it using
> rte_lcore_id() or a criteria of their liking.
> The callbacks have to be carefully written to explicitly initialise
> per lcore/thread private info.
> I thought it was enough, plus nobody objected.

Apparently, you were not alone thinking it was enough. :-)

> 
> 
> Now, some things to consider for this requirement you want to add.
> 
> If we want an application to be able to register callbacks after
> rte_eal_init(), we need to wake/interrupt all EAL threads from what
> they are doing.
> Once EAL threads enter some job the application gave, they won't
> reread anything from EAL.
> 
> Or we could require that callbacks are registered before
> rte_eal_init(), the init() callback could be called as the last step
> once all DPDK modules are initialised but before the application calls
> rte_eal_remote_launch().
> 
> But the problem of needing the break EAL threads from what they are
> doing is also present when we want to call the uninit() callback.
> And there, I have no good idea.

Thank you for sharing your thoughts on this, David. Very valuable insights!

I will try to determine the relevant scope and come up with a rough RFC, hoping 
to meet the December 25th proposal deadline for version 23.03.

-Morten



[PATCH v5 0/4] lcore telemetry improvements

2022-12-16 Thread Robin Jarry
This is a follow up on previous work by Kevin Laatz:

http://patches.dpdk.org/project/dpdk/list/?series=24658&state=*

This series is aimed at allowing DPDK applications to expose their CPU
usage stats in the DPDK telemetry under /eal/lcore/info. This is a much
more basic and naive approach which leaves the cpu cycles accounting
completely up to the application.

For reference, I have implemented a draft patch in OvS to use
rte_lcore_register_usage_cb() and report the already available busy
cycles information.

https://github.com/rjarry/ovs/commit/643e672fe388e348ea7ccbbda6f5a87a066fd919

Changes since v4:

- rte_lcore_usage_cb now takes a pointer to a rte_lcore_usage structure.
  I chose not to include any API version tracking mechanism since the
  unsupported/unused fields can simply be left to zero. This is only
  telemetry after all.

Changes since v3:

- Changed nomenclature from CPU cycles to TSC cycles in the docstring of
  rte_lcore_usage_cb.

Changes since v2:

- Fixed typos in docstrings.
- Used if (xxx != NULL) instead of if (xxx) test convention.
- Guarded against an unlikely race if rte_lcore_dump() is called by
  a thread while another one calls rte_lcore_register_usage_cb(NULL).
- s/utilization/usage/
- Fixed build on Windows.

Changes since v1:

- The cpuset field in telemetry is now a JSON list of CPU ids.
- Applications must now report their raw CPU cycles counts. The busyness
  ratio and rate of change is left to external monitoring tools.
- Renamed show lcores -> dump_lcores in testpmd.

Robin Jarry (4):
  eal: add lcore info in telemetry
  eal: allow applications to report their cpu usage
  testpmd: add dump_lcores command
  testpmd: report lcore usage

 app/test-pmd/5tswap.c |   5 +-
 app/test-pmd/cmdline.c|   3 +
 app/test-pmd/csumonly.c   |   6 +-
 app/test-pmd/flowgen.c|   2 +-
 app/test-pmd/icmpecho.c   |   6 +-
 app/test-pmd/iofwd.c  |   5 +-
 app/test-pmd/macfwd.c |   5 +-
 app/test-pmd/macswap.c|   5 +-
 app/test-pmd/noisy_vnf.c  |   4 +
 app/test-pmd/rxonly.c |   5 +-
 app/test-pmd/shared_rxq_fwd.c |   5 +-
 app/test-pmd/testpmd.c|  39 -
 app/test-pmd/testpmd.h|  14 +++-
 app/test-pmd/txonly.c |   7 +-
 lib/eal/common/eal_common_lcore.c | 129 +-
 lib/eal/include/rte_lcore.h   |  35 
 lib/eal/version.map   |   1 +
 17 files changed, 246 insertions(+), 30 deletions(-)

--
2.38.1



[PATCH v5 1/4] eal: add lcore info in telemetry

2022-12-16 Thread Robin Jarry
Report the same information than rte_lcore_dump() in the telemetry
API into /eal/lcore/list and /eal/lcore/info,ID.

Example:

  --> /eal/lcore/info,3
  {
"/eal/lcore/info": {
  "lcore_id": 3,
  "socket": 0,
  "role": "RTE",
  "cpuset": [
3
  ]
}
  }

Signed-off-by: Robin Jarry 
Acked-by: Morten Brørup 
---
v4 -> v5: No change

 lib/eal/common/eal_common_lcore.c | 96 +++
 1 file changed, 96 insertions(+)

diff --git a/lib/eal/common/eal_common_lcore.c 
b/lib/eal/common/eal_common_lcore.c
index 06c594b0224f..16548977dce8 100644
--- a/lib/eal/common/eal_common_lcore.c
+++ b/lib/eal/common/eal_common_lcore.c
@@ -10,6 +10,9 @@
 #include 
 #include 
 #include 
+#ifndef RTE_EXEC_ENV_WINDOWS
+#include 
+#endif
 
 #include "eal_private.h"
 #include "eal_thread.h"
@@ -456,3 +459,96 @@ rte_lcore_dump(FILE *f)
 {
rte_lcore_iterate(lcore_dump_cb, f);
 }
+
+#ifndef RTE_EXEC_ENV_WINDOWS
+static int
+lcore_telemetry_id_cb(unsigned int lcore_id, void *arg)
+{
+   struct rte_tel_data *d = arg;
+   return rte_tel_data_add_array_int(d, lcore_id);
+}
+
+static int
+handle_lcore_list(const char *cmd __rte_unused,
+   const char *params __rte_unused,
+   struct rte_tel_data *d)
+{
+   int ret = rte_tel_data_start_array(d, RTE_TEL_INT_VAL);
+   if (ret)
+   return ret;
+   return rte_lcore_iterate(lcore_telemetry_id_cb, d);
+}
+
+struct lcore_telemetry_info {
+   unsigned int lcore_id;
+   struct rte_tel_data *d;
+};
+
+static int
+lcore_telemetry_info_cb(unsigned int lcore_id, void *arg)
+{
+   struct lcore_telemetry_info *info = arg;
+   struct rte_config *cfg = rte_eal_get_configuration();
+   struct rte_tel_data *cpuset;
+   const char *role;
+   unsigned int cpu;
+
+   if (info->lcore_id != lcore_id)
+   return 0;
+
+   switch (cfg->lcore_role[lcore_id]) {
+   case ROLE_RTE:
+   role = "RTE";
+   break;
+   case ROLE_SERVICE:
+   role = "SERVICE";
+   break;
+   case ROLE_NON_EAL:
+   role = "NON_EAL";
+   break;
+   default:
+   role = "UNKNOWN";
+   break;
+   }
+   rte_tel_data_start_dict(info->d);
+   rte_tel_data_add_dict_int(info->d, "lcore_id", lcore_id);
+   rte_tel_data_add_dict_int(info->d, "socket", 
rte_lcore_to_socket_id(lcore_id));
+   rte_tel_data_add_dict_string(info->d, "role", role);
+   cpuset = rte_tel_data_alloc();
+   if (!cpuset)
+   return -ENOMEM;
+   rte_tel_data_start_array(cpuset, RTE_TEL_INT_VAL);
+   for (cpu = 0; cpu < CPU_SETSIZE; cpu++)
+   if (CPU_ISSET(cpu, &lcore_config[lcore_id].cpuset))
+   rte_tel_data_add_array_int(cpuset, cpu);
+   rte_tel_data_add_dict_container(info->d, "cpuset", cpuset, 0);
+
+   return 0;
+}
+
+static int
+handle_lcore_info(const char *cmd __rte_unused, const char *params, struct 
rte_tel_data *d)
+{
+   struct lcore_telemetry_info info = { .d = d };
+   char *endptr = NULL;
+   if (params == NULL || strlen(params) == 0)
+   return -EINVAL;
+   errno = 0;
+   info.lcore_id = strtoul(params, &endptr, 10);
+   if (errno)
+   return -errno;
+   if (endptr == params)
+   return -EINVAL;
+   return rte_lcore_iterate(lcore_telemetry_info_cb, &info);
+}
+
+RTE_INIT(lcore_telemetry)
+{
+   rte_telemetry_register_cmd(
+   "/eal/lcore/list", handle_lcore_list,
+   "List of lcore ids. Takes no parameters");
+   rte_telemetry_register_cmd(
+   "/eal/lcore/info", handle_lcore_info,
+   "Returns lcore info. Parameters: int lcore_id");
+}
+#endif /* !RTE_EXEC_ENV_WINDOWS */
-- 
2.38.1



[PATCH v5 3/4] testpmd: add dump_lcores command

2022-12-16 Thread Robin Jarry
Add a simple command that calls rte_lcore_dump().

Signed-off-by: Robin Jarry 
Acked-by: Morten Brørup 
---
v4 -> v5: no change

 app/test-pmd/cmdline.c | 3 +++
 1 file changed, 3 insertions(+)

diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c
index b32dc8bfd445..96474d2ae458 100644
--- a/app/test-pmd/cmdline.c
+++ b/app/test-pmd/cmdline.c
@@ -8345,6 +8345,8 @@ static void cmd_dump_parsed(void *parsed_result,
rte_mempool_list_dump(stdout);
else if (!strcmp(res->dump, "dump_devargs"))
rte_devargs_dump(stdout);
+   else if (!strcmp(res->dump, "dump_lcores"))
+   rte_lcore_dump(stdout);
else if (!strcmp(res->dump, "dump_log_types"))
rte_log_dump(stdout);
 }
@@ -8358,6 +8360,7 @@ static cmdline_parse_token_string_t cmd_dump_dump =
"dump_ring#"
"dump_mempool#"
"dump_devargs#"
+   "dump_lcores#"
"dump_log_types");
 
 static cmdline_parse_inst_t cmd_dump = {
-- 
2.38.1



[PATCH v5 2/4] eal: allow applications to report their cpu usage

2022-12-16 Thread Robin Jarry
Allow applications to register a callback that will be invoked in
rte_lcore_dump() and when requesting lcore info in the telemetry API.

The callback is expected to return the number of TSC cycles that have
passed since application start and the number of these cycles that were
spent doing busy work.

Signed-off-by: Robin Jarry 
Acked-by: Morten Brørup 
---
v4 -> v5:

The callback now takes a pointer to a rte_lcore_usage structure.
I chose not to include any API version tracking mechanism since the
unsupported/unused fields can simply be left to zero. This is only
telemetry after all.

 lib/eal/common/eal_common_lcore.c | 33 ++---
 lib/eal/include/rte_lcore.h   | 35 +++
 lib/eal/version.map   |  1 +
 3 files changed, 66 insertions(+), 3 deletions(-)

diff --git a/lib/eal/common/eal_common_lcore.c 
b/lib/eal/common/eal_common_lcore.c
index 16548977dce8..210636d21d6b 100644
--- a/lib/eal/common/eal_common_lcore.c
+++ b/lib/eal/common/eal_common_lcore.c
@@ -2,6 +2,7 @@
  * Copyright(c) 2010-2014 Intel Corporation
  */
 
+#include 
 #include 
 #include 
 
@@ -422,11 +423,21 @@ rte_lcore_iterate(rte_lcore_iterate_cb cb, void *arg)
return ret;
 }
 
+static rte_lcore_usage_cb lcore_usage_cb;
+
+void
+rte_lcore_register_usage_cb(rte_lcore_usage_cb cb)
+{
+   lcore_usage_cb = cb;
+}
+
 static int
 lcore_dump_cb(unsigned int lcore_id, void *arg)
 {
struct rte_config *cfg = rte_eal_get_configuration();
-   char cpuset[RTE_CPU_AFFINITY_STR_LEN];
+   char cpuset[RTE_CPU_AFFINITY_STR_LEN], usage_str[256];
+   struct rte_lcore_usage usage;
+   rte_lcore_usage_cb usage_cb;
const char *role;
FILE *f = arg;
int ret;
@@ -446,11 +457,19 @@ lcore_dump_cb(unsigned int lcore_id, void *arg)
break;
}
 
+   memset(&usage, 0, sizeof(usage));
+   usage_str[0] = '\0';
+   usage_cb = lcore_usage_cb;
+   if (usage_cb != NULL && usage_cb(lcore_id, &usage) == 0) {
+   snprintf(usage_str, sizeof(usage_str), ", busy cycles 
%"PRIu64"/%"PRIu64,
+   usage.busy_cycles, usage.total_cycles);
+   }
ret = eal_thread_dump_affinity(&lcore_config[lcore_id].cpuset, cpuset,
sizeof(cpuset));
-   fprintf(f, "lcore %u, socket %u, role %s, cpuset %s%s\n", lcore_id,
+   fprintf(f, "lcore %u, socket %u, role %s, cpuset %s%s%s\n", lcore_id,
rte_lcore_to_socket_id(lcore_id), role, cpuset,
-   ret == 0 ? "" : "...");
+   ret == 0 ? "" : "...", usage_str);
+
return 0;
 }
 
@@ -489,7 +508,9 @@ lcore_telemetry_info_cb(unsigned int lcore_id, void *arg)
 {
struct lcore_telemetry_info *info = arg;
struct rte_config *cfg = rte_eal_get_configuration();
+   struct rte_lcore_usage usage;
struct rte_tel_data *cpuset;
+   rte_lcore_usage_cb usage_cb;
const char *role;
unsigned int cpu;
 
@@ -522,6 +543,12 @@ lcore_telemetry_info_cb(unsigned int lcore_id, void *arg)
if (CPU_ISSET(cpu, &lcore_config[lcore_id].cpuset))
rte_tel_data_add_array_int(cpuset, cpu);
rte_tel_data_add_dict_container(info->d, "cpuset", cpuset, 0);
+   memset(&usage, 0, sizeof(usage));
+   usage_cb = lcore_usage_cb;
+   if (usage_cb != NULL && usage_cb(lcore_id, &usage) == 0) {
+   rte_tel_data_add_dict_u64(info->d, "busy_cycles", 
usage.busy_cycles);
+   rte_tel_data_add_dict_u64(info->d, "total_cycles", 
usage.total_cycles);
+   }
 
return 0;
 }
diff --git a/lib/eal/include/rte_lcore.h b/lib/eal/include/rte_lcore.h
index 6938c3fd7b81..a92313577355 100644
--- a/lib/eal/include/rte_lcore.h
+++ b/lib/eal/include/rte_lcore.h
@@ -328,6 +328,41 @@ typedef int (*rte_lcore_iterate_cb)(unsigned int lcore_id, 
void *arg);
 int
 rte_lcore_iterate(rte_lcore_iterate_cb cb, void *arg);
 
+/**
+ * CPU usage statistics.
+ */
+struct rte_lcore_usage {
+   uint64_t total_cycles;
+   /**< The total amount of time since application start, in TSC cycles. */
+   uint64_t busy_cycles;
+   /**< The amount of busy time since application start, in TSC cycles. */
+};
+
+/**
+ * Callback to allow applications to report CPU usage.
+ *
+ * @param [in] lcore_id
+ *   The lcore to consider.
+ * @param [out] usage
+ *   Counters representing this lcore usage. This can never be NULL.
+ * @return
+ *   - 0 if fields in usage were updated successfully. The fields that the
+ *   application does not support should be left to their default value.
+ *   - a negative value if the information is not available or if any error 
occurred.
+ */
+typedef int (*rte_lcore_usage_cb)(unsigned int lcore_id, struct 
rte_lcore_usage *usage);
+
+/**
+ * Register a callback from an application to be called in rte_lcore_dump() and
+ * the /eal/lcore/info telemetry endpoint handler. Applications are expected to
+ * re

[PATCH v5 4/4] testpmd: report lcore usage

2022-12-16 Thread Robin Jarry
Reuse the --record-core-cycles option to account for busy cycles. One
turn of packet_fwd_t is considered "busy" if there was at least one
received or transmitted packet.

Add a new busy_cycles field in struct fwd_stream. Update get_end_cycles
to accept an additional argument for the number of processed packets.
Update fwd_stream.busy_cycles when the number of packets is greater than
zero.

When --record-core-cycles is specified, register a callback with
rte_lcore_register_usage_cb(). In the callback, use the new lcore_id
field in struct fwd_lcore to identify the correct index in fwd_lcores
and return the sum of busy/total cycles of all fwd_streams.

This makes the cycles counters available in rte_lcore_dump() and the
lcore telemetry API:

 testpmd> dump_lcores
 lcore 3, socket 0, role RTE, cpuset 3
 lcore 4, socket 0, role RTE, cpuset 4, busy cycles 1228584096/9239923140
 lcore 5, socket 0, role RTE, cpuset 5, busy cycles 1255661768/9218141538

 --> /eal/lcore/info,4
 {
   "/eal/lcore/info": {
 "lcore_id": 4,
 "socket": 0,
 "role": "RTE",
 "cpuset": [
   4
 ],
 "busy_cycles": 10623340318,
 "total_cycles": 55331167354
   }
 }

Signed-off-by: Robin Jarry 
Acked-by: Morten Brørup 
---
v4 -> v5: Updated to use rte_lcore_usage struct.

 app/test-pmd/5tswap.c |  5 +++--
 app/test-pmd/csumonly.c   |  6 +++---
 app/test-pmd/flowgen.c|  2 +-
 app/test-pmd/icmpecho.c   |  6 +++---
 app/test-pmd/iofwd.c  |  5 +++--
 app/test-pmd/macfwd.c |  5 +++--
 app/test-pmd/macswap.c|  5 +++--
 app/test-pmd/noisy_vnf.c  |  4 
 app/test-pmd/rxonly.c |  5 +++--
 app/test-pmd/shared_rxq_fwd.c |  5 +++--
 app/test-pmd/testpmd.c| 39 ++-
 app/test-pmd/testpmd.h| 14 +
 app/test-pmd/txonly.c |  7 ---
 13 files changed, 81 insertions(+), 27 deletions(-)

diff --git a/app/test-pmd/5tswap.c b/app/test-pmd/5tswap.c
index f041a5e1d530..03225075716c 100644
--- a/app/test-pmd/5tswap.c
+++ b/app/test-pmd/5tswap.c
@@ -116,7 +116,7 @@ pkt_burst_5tuple_swap(struct fwd_stream *fs)
 nb_pkt_per_burst);
inc_rx_burst_stats(fs, nb_rx);
if (unlikely(nb_rx == 0))
-   return;
+   goto end;
 
fs->rx_packets += nb_rx;
txp = &ports[fs->tx_port];
@@ -182,7 +182,8 @@ pkt_burst_5tuple_swap(struct fwd_stream *fs)
rte_pktmbuf_free(pkts_burst[nb_tx]);
} while (++nb_tx < nb_rx);
}
-   get_end_cycles(fs, start_tsc);
+end:
+   get_end_cycles(fs, start_tsc, nb_rx);
 }
 
 static void
diff --git a/app/test-pmd/csumonly.c b/app/test-pmd/csumonly.c
index 1c2459851522..03e141221a56 100644
--- a/app/test-pmd/csumonly.c
+++ b/app/test-pmd/csumonly.c
@@ -868,7 +868,7 @@ pkt_burst_checksum_forward(struct fwd_stream *fs)
 nb_pkt_per_burst);
inc_rx_burst_stats(fs, nb_rx);
if (unlikely(nb_rx == 0))
-   return;
+   goto end;
 
fs->rx_packets += nb_rx;
rx_bad_ip_csum = 0;
@@ -1200,8 +1200,8 @@ pkt_burst_checksum_forward(struct fwd_stream *fs)
rte_pktmbuf_free(tx_pkts_burst[nb_tx]);
} while (++nb_tx < nb_rx);
}
-
-   get_end_cycles(fs, start_tsc);
+end:
+   get_end_cycles(fs, start_tsc, nb_rx);
 }
 
 static void
diff --git a/app/test-pmd/flowgen.c b/app/test-pmd/flowgen.c
index fd6abc0f4124..7b2f0ffdf0f5 100644
--- a/app/test-pmd/flowgen.c
+++ b/app/test-pmd/flowgen.c
@@ -196,7 +196,7 @@ pkt_burst_flow_gen(struct fwd_stream *fs)
 
RTE_PER_LCORE(_next_flow) = next_flow;
 
-   get_end_cycles(fs, start_tsc);
+   get_end_cycles(fs, start_tsc, nb_tx);
 }
 
 static int
diff --git a/app/test-pmd/icmpecho.c b/app/test-pmd/icmpecho.c
index 066f2a3ab79b..2fc9f96dc95f 100644
--- a/app/test-pmd/icmpecho.c
+++ b/app/test-pmd/icmpecho.c
@@ -303,7 +303,7 @@ reply_to_icmp_echo_rqsts(struct fwd_stream *fs)
 nb_pkt_per_burst);
inc_rx_burst_stats(fs, nb_rx);
if (unlikely(nb_rx == 0))
-   return;
+   goto end;
 
fs->rx_packets += nb_rx;
nb_replies = 0;
@@ -508,8 +508,8 @@ reply_to_icmp_echo_rqsts(struct fwd_stream *fs)
} while (++nb_tx < nb_replies);
}
}
-
-   get_end_cycles(fs, start_tsc);
+end:
+   get_end_cycles(fs, start_tsc, nb_rx);
 }
 
 static void
diff --git a/app/test-pmd/iofwd.c b/app/test-pmd/iofwd.c
index 8fafdec548ad..e5a2dbe20c69 100644
--- a/app/test-pmd/iofwd.c
+++ b/app/test-pmd/iofwd.c
@@ -59,7 +59,7 @@ pkt_burst_io_forward(struct fwd_stream *fs)
pkts_burst, nb_pkt_per_burst);
inc_rx_burst_stats(fs, nb_rx);
if (unlikely(nb_rx == 0))
-   return;
+   goto end;
fs->rx_packets += nb_rx;
 
nb_tx = rte_eth_tx

RE: [EXT] questions about crypto_scheduler

2022-12-16 Thread Akhil Goyal


> Hi, Akhil
> 
> Excuse me for the question.
> 
> I am testing UADK crypto performance with dpdk-test-crypto-perf, and
> want to use multi-thread or multi-session for better performance, so
> trying to use crypto_scheduler.
> 
> CMD like
> sudo dpdk-test-crypto-perf -l 1,2 --vdev "crypto_uadk0" --vdev
> "crypto_uadk1" --vdev
> "crypto_scheduler,worker=crypto_uadk0,worker=crypto_uadk1,mode=round-
> robin"
> -- --devtype crypto_scheduler --optype cipher-only --buffer-sz 8192
> 
> Though multi-worker is set but found it is just using one worker.
> Since uadk_cryptodev_probe gets the same dev->driver_id =
> uadk_cryptodev_driver_id in multi-probe.
> 
> Then scheduler_pmd_sym_session_configure will treat only one worker,
> and call rte_cryptodev_sym_session_create only once.
> if (configured_sess[j].driver_id == worker->driver_id)
>          break
> 
> Do I misuse the crypto_scheduler for the multi-thread or multi-session.


++ scheduler maintainers


[PATCH 1/7] common/idpf: add hw statistics

2022-12-16 Thread Mingxia Liu
This patch add hardware packets/bytes statistics.

Signed-off-by: Mingxia Liu 
---
 drivers/common/idpf/idpf_common_device.c   | 17 +
 drivers/common/idpf/idpf_common_device.h   |  5 +-
 drivers/common/idpf/idpf_common_virtchnl.c | 27 +++
 drivers/common/idpf/idpf_common_virtchnl.h |  3 +
 drivers/common/idpf/version.map|  2 +
 drivers/net/idpf/idpf_ethdev.c | 87 ++
 6 files changed, 140 insertions(+), 1 deletion(-)

diff --git a/drivers/common/idpf/idpf_common_device.c 
b/drivers/common/idpf/idpf_common_device.c
index 3580028dce..49ed778831 100644
--- a/drivers/common/idpf/idpf_common_device.c
+++ b/drivers/common/idpf/idpf_common_device.c
@@ -672,4 +672,21 @@ idpf_create_vport_info_init(struct idpf_vport *vport,
return 0;
 }
 
+void
+idpf_update_stats(struct virtchnl2_vport_stats *oes, struct 
virtchnl2_vport_stats *nes)
+{
+   nes->rx_bytes = nes->rx_bytes - oes->rx_bytes;
+   nes->rx_unicast = nes->rx_unicast - oes->rx_unicast;
+   nes->rx_multicast = nes->rx_multicast - oes->rx_multicast;
+   nes->rx_broadcast = nes->rx_broadcast - oes->rx_broadcast;
+   nes->rx_errors = nes->rx_errors - oes->rx_errors;
+   nes->rx_discards = nes->rx_discards - oes->rx_discards;
+   nes->tx_bytes = nes->tx_bytes - oes->tx_bytes;
+   nes->tx_unicast = nes->tx_unicast - oes->tx_unicast;
+   nes->tx_multicast = nes->tx_multicast - oes->tx_multicast;
+   nes->tx_broadcast = nes->tx_broadcast - oes->tx_broadcast;
+   nes->tx_errors = nes->tx_errors - oes->tx_errors;
+   nes->tx_discards = nes->tx_discards - oes->tx_discards;
+}
+
 RTE_LOG_REGISTER_SUFFIX(idpf_common_logtype, common, NOTICE);
diff --git a/drivers/common/idpf/idpf_common_device.h 
b/drivers/common/idpf/idpf_common_device.h
index 6c9a65ae3b..5184dcee9f 100644
--- a/drivers/common/idpf/idpf_common_device.h
+++ b/drivers/common/idpf/idpf_common_device.h
@@ -112,6 +112,8 @@ struct idpf_vport {
bool tx_vec_allowed;
bool rx_use_avx512;
bool tx_use_avx512;
+
+   struct virtchnl2_vport_stats eth_stats_offset;
 };
 
 /* Message type read in virtual channel from PF */
@@ -188,5 +190,6 @@ int idpf_config_irq_unmap(struct idpf_vport *vport, 
uint16_t nb_rx_queues);
 __rte_internal
 int idpf_create_vport_info_init(struct idpf_vport *vport,
struct virtchnl2_create_vport *vport_info);
-
+__rte_internal
+void idpf_update_stats(struct virtchnl2_vport_stats *oes, struct 
virtchnl2_vport_stats *nes);
 #endif /* _IDPF_COMMON_DEVICE_H_ */
diff --git a/drivers/common/idpf/idpf_common_virtchnl.c 
b/drivers/common/idpf/idpf_common_virtchnl.c
index 324214caa1..80351d15de 100644
--- a/drivers/common/idpf/idpf_common_virtchnl.c
+++ b/drivers/common/idpf/idpf_common_virtchnl.c
@@ -217,6 +217,7 @@ idpf_execute_vc_cmd(struct idpf_adapter *adapter, struct 
idpf_cmd_info *args)
case VIRTCHNL2_OP_UNMAP_QUEUE_VECTOR:
case VIRTCHNL2_OP_ALLOC_VECTORS:
case VIRTCHNL2_OP_DEALLOC_VECTORS:
+   case VIRTCHNL2_OP_GET_STATS:
/* for init virtchnl ops, need to poll the response */
err = idpf_read_one_msg(adapter, args->ops, args->out_size, 
args->out_buffer);
clear_cmd(adapter);
@@ -806,6 +807,32 @@ idpf_vc_query_ptype_info(struct idpf_adapter *adapter)
return err;
 }
 
+int
+idpf_query_stats(struct idpf_vport *vport,
+   struct virtchnl2_vport_stats **pstats)
+{
+   struct idpf_adapter *adapter = vport->adapter;
+   struct virtchnl2_vport_stats vport_stats;
+   struct idpf_cmd_info args;
+   int err;
+
+   vport_stats.vport_id = vport->vport_id;
+   args.ops = VIRTCHNL2_OP_GET_STATS;
+   args.in_args = (u8 *)&vport_stats;
+   args.in_args_size = sizeof(vport_stats);
+   args.out_buffer = adapter->mbx_resp;
+   args.out_size = IDPF_DFLT_MBX_BUF_SIZE;
+
+   err = idpf_execute_vc_cmd(adapter, &args);
+   if (err) {
+   DRV_LOG(ERR, "Failed to execute command of 
VIRTCHNL2_OP_GET_STATS");
+   *pstats = NULL;
+   return err;
+   }
+   *pstats = (struct virtchnl2_vport_stats *)args.out_buffer;
+   return 0;
+}
+
 #define IDPF_RX_BUF_STRIDE 64
 int
 idpf_vc_config_rxq(struct idpf_vport *vport, struct idpf_rx_queue *rxq)
diff --git a/drivers/common/idpf/idpf_common_virtchnl.h 
b/drivers/common/idpf/idpf_common_virtchnl.h
index d16b6b66f4..60347fe571 100644
--- a/drivers/common/idpf/idpf_common_virtchnl.h
+++ b/drivers/common/idpf/idpf_common_virtchnl.h
@@ -38,4 +38,7 @@ __rte_internal
 int idpf_vc_config_rxq(struct idpf_vport *vport, struct idpf_rx_queue *rxq);
 __rte_internal
 int idpf_vc_config_txq(struct idpf_vport *vport, struct idpf_tx_queue *txq);
+__rte_internal
+int idpf_query_stats(struct idpf_vport *vport,
+struct virtchnl2_vport_stats **pstats);
 #endif /* _IDPF_COMMON_VIRTCHNL_H_ */
diff --git a/drivers/common/id

[PATCH 0/7] add idpf pmd enhancement features

2022-12-16 Thread Mingxia Liu
This patchset add several enhancement features of idpf pmd. 
Including the following:
- add hw statistics, support stats/xstats ops
- add rss configure/show ops
- add event handle: link status
- add scattered data path for single queue

This patchset is based on the refactor idpf PMD code:
http://patches.dpdk.org/project/dpdk/list/?submitter=410&q=&delegate=&archive=&series=&state=*
http://patches.dpdk.org/project/dpdk/list/?submitter=2083&q=&delegate=&archive=&series=&state=*


Mingxia Liu (7):
  common/idpf: add hw statistics
  common/idpf: add RSS set/get ops
  common/idpf: support single q scatter RX datapath
  common/idpf: add rss_offload hash in singleq rx
  common/idpf: add alarm to support handle vchnl message
  common/idpf: add xstats ops
  common/idpf: update mbuf_alloc_failed multi-thread process

 drivers/common/idpf/idpf_common_device.c  |  17 +
 drivers/common/idpf/idpf_common_device.h  |  11 +-
 drivers/common/idpf/idpf_common_rxtx.c| 158 -
 drivers/common/idpf/idpf_common_rxtx.h|   5 +-
 drivers/common/idpf/idpf_common_rxtx_avx512.c |  12 +-
 drivers/common/idpf/idpf_common_virtchnl.c| 157 -
 drivers/common/idpf/idpf_common_virtchnl.h|  18 +-
 drivers/common/idpf/version.map   |   9 +
 drivers/net/idpf/idpf_ethdev.c| 639 +-
 drivers/net/idpf/idpf_ethdev.h|   7 +-
 drivers/net/idpf/idpf_rxtx.c  |  26 +-
 drivers/net/idpf/idpf_rxtx.h  |   2 +
 12 files changed, 1029 insertions(+), 32 deletions(-)

-- 
2.25.1



[PATCH 2/7] common/idpf: add RSS set/get ops

2022-12-16 Thread Mingxia Liu
Add support for these device ops:
- rss_reta_update
- rss_reta_query
- rss_hash_update
- rss_hash_conf_get

Signed-off-by: Mingxia Liu 
---
 drivers/common/idpf/idpf_common_device.h   |   1 +
 drivers/common/idpf/idpf_common_virtchnl.c | 119 
 drivers/common/idpf/idpf_common_virtchnl.h |  15 +-
 drivers/common/idpf/version.map|   6 +
 drivers/net/idpf/idpf_ethdev.c | 303 +
 drivers/net/idpf/idpf_ethdev.h |   5 +-
 6 files changed, 445 insertions(+), 4 deletions(-)

diff --git a/drivers/common/idpf/idpf_common_device.h 
b/drivers/common/idpf/idpf_common_device.h
index 5184dcee9f..d7d4cd5363 100644
--- a/drivers/common/idpf/idpf_common_device.h
+++ b/drivers/common/idpf/idpf_common_device.h
@@ -95,6 +95,7 @@ struct idpf_vport {
uint32_t *rss_lut;
uint8_t *rss_key;
uint64_t rss_hf;
+   uint64_t last_general_rss_hf;
 
/* MSIX info*/
struct virtchnl2_queue_vector *qv_map; /* queue vector mapping */
diff --git a/drivers/common/idpf/idpf_common_virtchnl.c 
b/drivers/common/idpf/idpf_common_virtchnl.c
index 80351d15de..ae5a983836 100644
--- a/drivers/common/idpf/idpf_common_virtchnl.c
+++ b/drivers/common/idpf/idpf_common_virtchnl.c
@@ -218,6 +218,9 @@ idpf_execute_vc_cmd(struct idpf_adapter *adapter, struct 
idpf_cmd_info *args)
case VIRTCHNL2_OP_ALLOC_VECTORS:
case VIRTCHNL2_OP_DEALLOC_VECTORS:
case VIRTCHNL2_OP_GET_STATS:
+   case VIRTCHNL2_OP_GET_RSS_KEY:
+   case VIRTCHNL2_OP_GET_RSS_HASH:
+   case VIRTCHNL2_OP_GET_RSS_LUT:
/* for init virtchnl ops, need to poll the response */
err = idpf_read_one_msg(adapter, args->ops, args->out_size, 
args->out_buffer);
clear_cmd(adapter);
@@ -448,6 +451,48 @@ idpf_vc_set_rss_key(struct idpf_vport *vport)
return err;
 }
 
+int idpf_vc_get_rss_key(struct idpf_vport *vport)
+{
+   struct idpf_adapter *adapter = vport->adapter;
+   struct virtchnl2_rss_key *rss_key_ret;
+   struct virtchnl2_rss_key rss_key;
+   struct idpf_cmd_info args;
+   int err;
+
+   memset(&rss_key, 0, sizeof(rss_key));
+   rss_key.vport_id = vport->vport_id;
+
+   memset(&args, 0, sizeof(args));
+   args.ops = VIRTCHNL2_OP_GET_RSS_KEY;
+   args.in_args = (uint8_t *)&rss_key;
+   args.in_args_size = sizeof(rss_key);
+   args.out_buffer = adapter->mbx_resp;
+   args.out_size = IDPF_DFLT_MBX_BUF_SIZE;
+
+   err = idpf_execute_vc_cmd(adapter, &args);
+
+   if (!err) {
+   rss_key_ret = (struct virtchnl2_rss_key *)args.out_buffer;
+   if (rss_key_ret->key_len != vport->rss_key_size) {
+   rte_free(vport->rss_key);
+   vport->rss_key = NULL;
+   vport->rss_key_size = RTE_MIN(IDPF_RSS_KEY_LEN,
+ rss_key_ret->key_len);
+   vport->rss_key = rte_zmalloc("rss_key", 
vport->rss_key_size, 0);
+   if (!vport->rss_key) {
+   vport->rss_key_size = 0;
+   DRV_LOG(ERR, "Failed to allocate RSS key");
+   return -ENOMEM;
+   }
+   }
+   rte_memcpy(vport->rss_key, rss_key_ret->key, 
vport->rss_key_size);
+   } else {
+   DRV_LOG(ERR, "Failed to execute command of 
VIRTCHNL2_OP_GET_RSS_KEY");
+   }
+
+   return err;
+}
+
 int
 idpf_vc_set_rss_lut(struct idpf_vport *vport)
 {
@@ -482,6 +527,48 @@ idpf_vc_set_rss_lut(struct idpf_vport *vport)
return err;
 }
 
+int
+idpf_vc_get_rss_lut(struct idpf_vport *vport)
+{
+   struct idpf_adapter *adapter = vport->adapter;
+   struct virtchnl2_rss_lut *rss_lut_ret;
+   struct virtchnl2_rss_lut rss_lut;
+   struct idpf_cmd_info args;
+   int err;
+
+   memset(&rss_lut, 0, sizeof(rss_lut));
+   rss_lut.vport_id = vport->vport_id;
+
+   memset(&args, 0, sizeof(args));
+   args.ops = VIRTCHNL2_OP_GET_RSS_LUT;
+   args.in_args = (uint8_t *)&rss_lut;
+   args.in_args_size = sizeof(rss_lut);
+   args.out_buffer = adapter->mbx_resp;
+   args.out_size = IDPF_DFLT_MBX_BUF_SIZE;
+
+   err = idpf_execute_vc_cmd(adapter, &args);
+
+   if (!err) {
+   rss_lut_ret = (struct virtchnl2_rss_lut *)args.out_buffer;
+   if (rss_lut_ret->lut_entries != vport->rss_lut_size) {
+   rte_free(vport->rss_lut);
+   vport->rss_lut = NULL;
+   vport->rss_lut = rte_zmalloc("rss_lut",
+sizeof(uint32_t) * 
rss_lut_ret->lut_entries, 0);
+   if (vport->rss_lut == NULL) {
+   DRV_LOG(ERR, "Failed to allocate RSS lut");
+   return -ENOMEM;
+   }
+   }
+  

[PATCH 4/7] common/idpf: add rss_offload hash in singleq rx

2022-12-16 Thread Mingxia Liu
This patch add rss valid flag and hash value parsing of rx descriptor.

Signed-off-by: Mingxia Liu 
---
 drivers/common/idpf/idpf_common_rxtx.c | 16 
 1 file changed, 16 insertions(+)

diff --git a/drivers/common/idpf/idpf_common_rxtx.c 
b/drivers/common/idpf/idpf_common_rxtx.c
index dcdf43ca0a..cec99d2951 100644
--- a/drivers/common/idpf/idpf_common_rxtx.c
+++ b/drivers/common/idpf/idpf_common_rxtx.c
@@ -1028,6 +1028,20 @@ idpf_update_rx_tail(struct idpf_rx_queue *rxq, uint16_t 
nb_hold,
rxq->nb_rx_hold = nb_hold;
 }
 
+static inline void
+idpf_singleq_rx_rss_offload(struct rte_mbuf *mb,
+   volatile struct virtchnl2_rx_flex_desc_nic *rx_desc,
+   uint64_t *pkt_flags)
+{
+   uint16_t rx_status0 = rte_le_to_cpu_16(rx_desc->status_error0);
+
+   if (rx_status0 & RTE_BIT32(VIRTCHNL2_RX_FLEX_DESC_STATUS0_RSS_VALID_S)) 
{
+   *pkt_flags |= RTE_MBUF_F_RX_RSS_HASH;
+   mb->hash.rss = rte_le_to_cpu_32(rx_desc->rss_hash);
+   }
+
+}
+
 uint16_t
 idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
   uint16_t nb_pkts)
@@ -1116,6 +1130,7 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
rxm->port = rxq->port_id;
rxm->ol_flags = 0;
pkt_flags = idpf_rxd_to_pkt_flags(rx_status0);
+   idpf_singleq_rx_rss_offload(rxm, &rxd.flex_nic_wb, &pkt_flags);
rxm->packet_type =

ptype_tbl[(uint8_t)(rte_cpu_to_le_16(rxd.flex_nic_wb.ptype_flex_flags0) &
VIRTCHNL2_RX_FLEX_DESC_PTYPE_M)];
@@ -1246,6 +1261,7 @@ idpf_singleq_recv_scatter_pkts(void *rx_queue, struct 
rte_mbuf **rx_pkts,
first_seg->port = rxq->port_id;
first_seg->ol_flags = 0;
pkt_flags = idpf_rxd_to_pkt_flags(rx_status0);
+   idpf_singleq_rx_rss_offload(first_seg, &rxd.flex_nic_wb, 
&pkt_flags);
first_seg->packet_type =

ptype_tbl[(uint8_t)(rte_cpu_to_le_16(rxd.flex_nic_wb.ptype_flex_flags0) &
VIRTCHNL2_RX_FLEX_DESC_PTYPE_M)];
-- 
2.25.1



[PATCH 3/7] common/idpf: support single q scatter RX datapath

2022-12-16 Thread Mingxia Liu
This patch add single q recv scatter rx function.

Signed-off-by: Mingxia Liu 
Signed-off-by: Wenjun Wu 
---
 drivers/common/idpf/idpf_common_rxtx.c | 134 +
 drivers/common/idpf/idpf_common_rxtx.h |   3 +
 drivers/common/idpf/version.map|   1 +
 drivers/net/idpf/idpf_ethdev.c |   3 +-
 drivers/net/idpf/idpf_rxtx.c   |  26 -
 drivers/net/idpf/idpf_rxtx.h   |   2 +
 6 files changed, 166 insertions(+), 3 deletions(-)

diff --git a/drivers/common/idpf/idpf_common_rxtx.c 
b/drivers/common/idpf/idpf_common_rxtx.c
index 7f8311d8f6..dcdf43ca0a 100644
--- a/drivers/common/idpf/idpf_common_rxtx.c
+++ b/drivers/common/idpf/idpf_common_rxtx.c
@@ -1144,6 +1144,140 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
return nb_rx;
 }
 
+uint16_t
+idpf_singleq_recv_scatter_pkts(void *rx_queue, struct rte_mbuf **rx_pkts,
+  uint16_t nb_pkts)
+{
+   struct idpf_rx_queue *rxq = rx_queue;
+   volatile union virtchnl2_rx_desc *rx_ring = rxq->rx_ring;
+   volatile union virtchnl2_rx_desc *rxdp;
+   union virtchnl2_rx_desc rxd;
+   struct idpf_adapter *ad;
+   struct rte_mbuf *first_seg = rxq->pkt_first_seg;
+   struct rte_mbuf *last_seg = rxq->pkt_last_seg;
+   struct rte_mbuf *rxm;
+   struct rte_mbuf *nmb;
+   struct rte_eth_dev *dev;
+   const uint32_t *ptype_tbl = rxq->adapter->ptype_tbl;
+   uint16_t nb_hold = 0, nb_rx = 0;
+   uint16_t rx_id = rxq->rx_tail;
+   uint16_t rx_packet_len;
+   uint16_t rx_status0;
+   uint64_t pkt_flags;
+   uint64_t dma_addr;
+   uint64_t ts_ns;
+
+   ad = rxq->adapter;
+
+   if (unlikely(!rxq) || unlikely(!rxq->q_started))
+   return nb_rx;
+
+   while (nb_rx < nb_pkts) {
+   rxdp = &rx_ring[rx_id];
+   rx_status0 = rte_le_to_cpu_16(rxdp->flex_nic_wb.status_error0);
+
+   /* Check the DD bit first */
+   if (!(rx_status0 & (1 << VIRTCHNL2_RX_FLEX_DESC_STATUS0_DD_S)))
+   break;
+
+   nmb = rte_mbuf_raw_alloc(rxq->mp);
+   if (unlikely(!nmb)) {
+   rxq->rx_stats.mbuf_alloc_failed++;
+   RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u "
+  "queue_id=%u", rxq->port_id, rxq->queue_id);
+   break;
+   }
+
+   rxd = *rxdp;
+
+   nb_hold++;
+   rxm = rxq->sw_ring[rx_id];
+   rxq->sw_ring[rx_id] = nmb;
+   rx_id++;
+   if (unlikely(rx_id == rxq->nb_rx_desc))
+   rx_id = 0;
+
+   /* Prefetch next mbuf */
+   rte_prefetch0(rxq->sw_ring[rx_id]);
+
+   /* When next RX descriptor is on a cache line boundary,
+* prefetch the next 4 RX descriptors and next 8 pointers
+* to mbufs.
+*/
+   if ((rx_id & 0x3) == 0) {
+   rte_prefetch0(&rx_ring[rx_id]);
+   rte_prefetch0(rxq->sw_ring[rx_id]);
+   }
+   dma_addr =
+   rte_cpu_to_le_64(rte_mbuf_data_iova_default(nmb));
+   rxdp->read.hdr_addr = 0;
+   rxdp->read.pkt_addr = dma_addr;
+   rx_packet_len = (rte_cpu_to_le_16(rxd.flex_nic_wb.pkt_len) &
+VIRTCHNL2_RX_FLEX_DESC_PKT_LEN_M);
+   rxm->data_len = rx_packet_len;
+   rxm->data_off = RTE_PKTMBUF_HEADROOM;
+
+   /**
+* If this is the first buffer of the received packet, set the
+* pointer to the first mbuf of the packet and initialize its
+* context. Otherwise, update the total length and the number
+* of segments of the current scattered packet, and update the
+* pointer to the last mbuf of the current packet.
+*/
+   if (!first_seg) {
+   first_seg = rxm;
+   first_seg->nb_segs = 1;
+   first_seg->pkt_len = rx_packet_len;
+   } else {
+   first_seg->pkt_len =
+   (uint16_t)(first_seg->pkt_len +
+  rx_packet_len);
+   first_seg->nb_segs++;
+   last_seg->next = rxm;
+   }
+
+   if (!(rx_status0 & (1 << 
VIRTCHNL2_RX_FLEX_DESC_STATUS0_EOF_S))) {
+   last_seg = rxm;
+   continue;
+   }
+
+   rxm->next = NULL;
+
+   first_seg->port = rxq->port_id;
+   first_seg->ol_flags = 0;
+   pkt_flags = idpf_rxd_to_pkt_flags(rx_status0);
+   first_seg->packet_type =
+   
ptype_tbl[(uint8_t)(rte_cpu_to_le_16(rx

[PATCH 6/7] common/idpf: add xstats ops

2022-12-16 Thread Mingxia Liu
Add support for these device ops:
-idpf_dev_xstats_get
-idpf_dev_xstats_get_names
-idpf_dev_xstats_reset

Signed-off-by: Mingxia Liu 
---
 drivers/net/idpf/idpf_ethdev.c | 80 ++
 1 file changed, 80 insertions(+)

diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 3ffc4cd9a3..97c03118e0 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -80,6 +80,30 @@ static const uint64_t idpf_ipv6_rss = 
RTE_ETH_RSS_NONFRAG_IPV6_UDP |
  RTE_ETH_RSS_NONFRAG_IPV6_OTHER |
  RTE_ETH_RSS_FRAG_IPV6;
 
+struct rte_idpf_xstats_name_off {
+   char name[RTE_ETH_XSTATS_NAME_SIZE];
+   unsigned int offset;
+};
+
+static const struct rte_idpf_xstats_name_off rte_idpf_stats_strings[] = {
+   {"rx_bytes", offsetof(struct virtchnl2_vport_stats, rx_bytes)},
+   {"rx_unicast_packets", offsetof(struct virtchnl2_vport_stats, 
rx_unicast)},
+   {"rx_multicast_packets", offsetof(struct virtchnl2_vport_stats, 
rx_multicast)},
+   {"rx_broadcast_packets", offsetof(struct virtchnl2_vport_stats, 
rx_broadcast)},
+   {"rx_dropped_packets", offsetof(struct virtchnl2_vport_stats, 
rx_discards)},
+   {"rx_errors", offsetof(struct virtchnl2_vport_stats, rx_errors)},
+   {"rx_unknown_protocol_packets", offsetof(struct virtchnl2_vport_stats,
+rx_unknown_protocol)},
+   {"tx_bytes", offsetof(struct virtchnl2_vport_stats, tx_bytes)},
+   {"tx_unicast_packets", offsetof(struct virtchnl2_vport_stats, 
tx_unicast)},
+   {"tx_multicast_packets", offsetof(struct virtchnl2_vport_stats, 
tx_multicast)},
+   {"tx_broadcast_packets", offsetof(struct virtchnl2_vport_stats, 
tx_broadcast)},
+   {"tx_dropped_packets", offsetof(struct virtchnl2_vport_stats, 
tx_discards)},
+   {"tx_error_packets", offsetof(struct virtchnl2_vport_stats, 
tx_errors)}};
+
+#define IDPF_NB_XSTATS (sizeof(rte_idpf_stats_strings) / \
+   sizeof(rte_idpf_stats_strings[0]))
+
 static int
 idpf_dev_link_update(struct rte_eth_dev *dev,
 __rte_unused int wait_to_complete)
@@ -303,6 +327,59 @@ idpf_dev_stats_reset(struct rte_eth_dev *dev)
return 0;
 }
 
+static int idpf_dev_xstats_reset(struct rte_eth_dev *dev)
+{
+   idpf_dev_stats_reset(dev);
+   return 0;
+}
+
+static int idpf_dev_xstats_get(struct rte_eth_dev *dev,
+  struct rte_eth_xstat *xstats, unsigned int n)
+{
+   struct idpf_vport *vport =
+   (struct idpf_vport *)dev->data->dev_private;
+   struct virtchnl2_vport_stats *pstats = NULL;
+   unsigned int i;
+   int ret;
+
+   if (n < IDPF_NB_XSTATS)
+   return IDPF_NB_XSTATS;
+
+   if (!xstats)
+   return 0;
+
+   ret = idpf_query_stats(vport, &pstats);
+   if (ret) {
+   PMD_DRV_LOG(ERR, "Get statistics failed");
+   return 0;
+   }
+
+   idpf_update_stats(&vport->eth_stats_offset, pstats);
+
+   /* loop over xstats array and values from pstats */
+   for (i = 0; i < IDPF_NB_XSTATS; i++) {
+   xstats[i].id = i;
+   xstats[i].value = *(uint64_t *)(((char *)pstats) +
+   rte_idpf_stats_strings[i].offset);
+   }
+   return IDPF_NB_XSTATS;
+}
+
+static int idpf_dev_xstats_get_names(__rte_unused struct rte_eth_dev *dev,
+struct rte_eth_xstat_name *xstats_names,
+__rte_unused unsigned int limit)
+{
+   unsigned int i;
+
+   if (xstats_names)
+   for (i = 0; i < IDPF_NB_XSTATS; i++) {
+   snprintf(xstats_names[i].name,
+sizeof(xstats_names[i].name),
+"%s", rte_idpf_stats_strings[i].name);
+   }
+   return IDPF_NB_XSTATS;
+}
+
 static int idpf_config_rss_hf(struct idpf_vport *vport, uint64_t rss_hf)
 {
uint64_t hena = 0, valid_rss_hf = 0;
@@ -1149,6 +1226,9 @@ static const struct eth_dev_ops idpf_eth_dev_ops = {
.reta_query = idpf_rss_reta_query,
.rss_hash_update= idpf_rss_hash_update,
.rss_hash_conf_get  = idpf_rss_hash_conf_get,
+   .xstats_get = idpf_dev_xstats_get,
+   .xstats_get_names   = idpf_dev_xstats_get_names,
+   .xstats_reset   = idpf_dev_xstats_reset,
 };
 
 static uint16_t
-- 
2.25.1



[PATCH 7/7] common/idpf: update mbuf_alloc_failed multi-thread process

2022-12-16 Thread Mingxia Liu
As the variable mbuf_alloc_failed is operated by more than thread,
change it to type rte_atomic64_t and operated by rte_atomic64_xx()
function, this will avoid multithreading issue.

Signed-off-by: Mingxia Liu 
---
 drivers/common/idpf/idpf_common_rxtx.c| 10 ++
 drivers/common/idpf/idpf_common_rxtx.h|  2 +-
 drivers/common/idpf/idpf_common_rxtx_avx512.c | 12 
 drivers/net/idpf/idpf_ethdev.c|  5 +++--
 4 files changed, 18 insertions(+), 11 deletions(-)

diff --git a/drivers/common/idpf/idpf_common_rxtx.c 
b/drivers/common/idpf/idpf_common_rxtx.c
index cec99d2951..dd8e761834 100644
--- a/drivers/common/idpf/idpf_common_rxtx.c
+++ b/drivers/common/idpf/idpf_common_rxtx.c
@@ -592,7 +592,8 @@ idpf_split_rx_bufq_refill(struct idpf_rx_queue *rx_bufq)
next_avail = 0;
rx_bufq->nb_rx_hold -= delta;
} else {
-   rx_bufq->rx_stats.mbuf_alloc_failed += nb_desc - 
next_avail;
+   rte_atomic64_add(&(rx_bufq->rx_stats.mbuf_alloc_failed),
+nb_desc - next_avail);
RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u 
queue_id=%u",
   rx_bufq->port_id, rx_bufq->queue_id);
return;
@@ -611,7 +612,8 @@ idpf_split_rx_bufq_refill(struct idpf_rx_queue *rx_bufq)
next_avail += nb_refill;
rx_bufq->nb_rx_hold -= nb_refill;
} else {
-   rx_bufq->rx_stats.mbuf_alloc_failed += nb_desc - 
next_avail;
+   rte_atomic64_add(&(rx_bufq->rx_stats.mbuf_alloc_failed),
+nb_desc - next_avail);
RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u 
queue_id=%u",
   rx_bufq->port_id, rx_bufq->queue_id);
}
@@ -1088,7 +1090,7 @@ idpf_singleq_recv_pkts(void *rx_queue, struct rte_mbuf 
**rx_pkts,
 
nmb = rte_mbuf_raw_alloc(rxq->mp);
if (unlikely(nmb == NULL)) {
-   rxq->rx_stats.mbuf_alloc_failed++;
+   rte_atomic64_inc(&(rxq->rx_stats.mbuf_alloc_failed));
RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u "
   "queue_id=%u", rxq->port_id, rxq->queue_id);
break;
@@ -1197,7 +1199,7 @@ idpf_singleq_recv_scatter_pkts(void *rx_queue, struct 
rte_mbuf **rx_pkts,
 
nmb = rte_mbuf_raw_alloc(rxq->mp);
if (unlikely(!nmb)) {
-   rxq->rx_stats.mbuf_alloc_failed++;
+   rte_atomic64_inc(&(rxq->rx_stats.mbuf_alloc_failed));
RX_LOG(DEBUG, "RX mbuf alloc failed port_id=%u "
   "queue_id=%u", rxq->port_id, rxq->queue_id);
break;
diff --git a/drivers/common/idpf/idpf_common_rxtx.h 
b/drivers/common/idpf/idpf_common_rxtx.h
index eee9fdbd9e..0209750187 100644
--- a/drivers/common/idpf/idpf_common_rxtx.h
+++ b/drivers/common/idpf/idpf_common_rxtx.h
@@ -91,7 +91,7 @@
 #define PF_GLTSYN_SHTIME_H_5   (PF_TIMESYNC_BAR4_BASE + 0x13C)
 
 struct idpf_rx_stats {
-   uint64_t mbuf_alloc_failed;
+   rte_atomic64_t mbuf_alloc_failed;
 };
 
 struct idpf_rx_queue {
diff --git a/drivers/common/idpf/idpf_common_rxtx_avx512.c 
b/drivers/common/idpf/idpf_common_rxtx_avx512.c
index 5a91ed610e..1fc110cc94 100644
--- a/drivers/common/idpf/idpf_common_rxtx_avx512.c
+++ b/drivers/common/idpf/idpf_common_rxtx_avx512.c
@@ -38,7 +38,8 @@ idpf_singleq_rearm_common(struct idpf_rx_queue *rxq)
dma_addr0);
}
}
-   rxq->rx_stats.mbuf_alloc_failed += IDPF_RXQ_REARM_THRESH;
+   rte_atomic64_add(&(rxq->rx_stats.mbuf_alloc_failed),
+IDPF_RXQ_REARM_THRESH);
return;
}
struct rte_mbuf *mb0, *mb1, *mb2, *mb3;
@@ -167,7 +168,8 @@ idpf_singleq_rearm(struct idpf_rx_queue *rxq)
 dma_addr0);
}
}
-   rxq->rx_stats.mbuf_alloc_failed += 
IDPF_RXQ_REARM_THRESH;
+   rte_atomic64_add(&(rxq->rx_stats.mbuf_alloc_failed),
+IDPF_RXQ_REARM_THRESH);
return;
}
}
@@ -562,7 +564,8 @@ idpf_splitq_rearm_common(struct idpf_rx_queue *rx_bufq)
dma_addr0);
}
}
-   rx_bufq->rx_stats.mbuf_alloc_failed += IDPF_RXQ_REARM_THRESH;
+   rte_atomic64_add(&(rx_bufq->rx_stats.mbuf_alloc_failed),
+IDPF_RXQ_REARM_THRESH);
return;
 

[PATCH 5/7] common/idpf: add alarm to support handle vchnl message

2022-12-16 Thread Mingxia Liu
Handle virtual channel message.
Refine link status update.

Signed-off-by: Mingxia Liu 
Signed-off-by: Beilei Ling 
---
 drivers/common/idpf/idpf_common_device.h   |   5 +
 drivers/common/idpf/idpf_common_virtchnl.c |  19 ---
 drivers/net/idpf/idpf_ethdev.c | 165 -
 drivers/net/idpf/idpf_ethdev.h |   2 +
 4 files changed, 171 insertions(+), 20 deletions(-)

diff --git a/drivers/common/idpf/idpf_common_device.h 
b/drivers/common/idpf/idpf_common_device.h
index d7d4cd5363..03697510bb 100644
--- a/drivers/common/idpf/idpf_common_device.h
+++ b/drivers/common/idpf/idpf_common_device.h
@@ -115,6 +115,11 @@ struct idpf_vport {
bool tx_use_avx512;
 
struct virtchnl2_vport_stats eth_stats_offset;
+
+   void *dev;
+   /* Event from ipf */
+   bool link_up;
+   uint32_t link_speed;
 };
 
 /* Message type read in virtual channel from PF */
diff --git a/drivers/common/idpf/idpf_common_virtchnl.c 
b/drivers/common/idpf/idpf_common_virtchnl.c
index ae5a983836..c3e7569cc2 100644
--- a/drivers/common/idpf/idpf_common_virtchnl.c
+++ b/drivers/common/idpf/idpf_common_virtchnl.c
@@ -202,25 +202,6 @@ idpf_execute_vc_cmd(struct idpf_adapter *adapter, struct 
idpf_cmd_info *args)
switch (args->ops) {
case VIRTCHNL_OP_VERSION:
case VIRTCHNL2_OP_GET_CAPS:
-   case VIRTCHNL2_OP_CREATE_VPORT:
-   case VIRTCHNL2_OP_DESTROY_VPORT:
-   case VIRTCHNL2_OP_SET_RSS_KEY:
-   case VIRTCHNL2_OP_SET_RSS_LUT:
-   case VIRTCHNL2_OP_SET_RSS_HASH:
-   case VIRTCHNL2_OP_CONFIG_RX_QUEUES:
-   case VIRTCHNL2_OP_CONFIG_TX_QUEUES:
-   case VIRTCHNL2_OP_ENABLE_QUEUES:
-   case VIRTCHNL2_OP_DISABLE_QUEUES:
-   case VIRTCHNL2_OP_ENABLE_VPORT:
-   case VIRTCHNL2_OP_DISABLE_VPORT:
-   case VIRTCHNL2_OP_MAP_QUEUE_VECTOR:
-   case VIRTCHNL2_OP_UNMAP_QUEUE_VECTOR:
-   case VIRTCHNL2_OP_ALLOC_VECTORS:
-   case VIRTCHNL2_OP_DEALLOC_VECTORS:
-   case VIRTCHNL2_OP_GET_STATS:
-   case VIRTCHNL2_OP_GET_RSS_KEY:
-   case VIRTCHNL2_OP_GET_RSS_HASH:
-   case VIRTCHNL2_OP_GET_RSS_LUT:
/* for init virtchnl ops, need to poll the response */
err = idpf_read_one_msg(adapter, args->ops, args->out_size, 
args->out_buffer);
clear_cmd(adapter);
diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c
index 573afcab4f..3ffc4cd9a3 100644
--- a/drivers/net/idpf/idpf_ethdev.c
+++ b/drivers/net/idpf/idpf_ethdev.c
@@ -9,6 +9,7 @@
 #include 
 #include 
 #include 
+#include 
 
 #include "idpf_ethdev.h"
 #include "idpf_rxtx.h"
@@ -83,12 +84,49 @@ static int
 idpf_dev_link_update(struct rte_eth_dev *dev,
 __rte_unused int wait_to_complete)
 {
+   struct idpf_vport *vport = dev->data->dev_private;
struct rte_eth_link new_link;
 
memset(&new_link, 0, sizeof(new_link));
 
-   new_link.link_speed = RTE_ETH_SPEED_NUM_NONE;
+   switch (vport->link_speed) {
+   case 10:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_10M;
+   break;
+   case 100:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_100M;
+   break;
+   case 1000:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_1G;
+   break;
+   case 1:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_10G;
+   break;
+   case 2:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_20G;
+   break;
+   case 25000:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_25G;
+   break;
+   case 4:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_40G;
+   break;
+   case 5:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_50G;
+   break;
+   case 10:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_100G;
+   break;
+   case 20:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_200G;
+   break;
+   default:
+   new_link.link_speed = RTE_ETH_SPEED_NUM_NONE;
+   }
+
new_link.link_duplex = RTE_ETH_LINK_FULL_DUPLEX;
+   new_link.link_status = vport->link_up ? RTE_ETH_LINK_UP :
+   RTE_ETH_LINK_DOWN;
new_link.link_autoneg = !(dev->data->dev_conf.link_speeds &
  RTE_ETH_LINK_SPEED_FIXED);
 
@@ -918,6 +956,127 @@ idpf_parse_devargs(struct rte_pci_device *pci_dev, struct 
idpf_adapter_ext *adap
return ret;
 }
 
+static struct idpf_vport *
+idpf_find_vport(struct idpf_adapter_ext *adapter, uint32_t vport_id)
+{
+   struct idpf_vport *vport = NULL;
+   int i;
+
+   for (i = 0; i < adapter->cur_vport_nb; i++) {
+   vport = adapter->vports[i];
+   if (vport->vport_id != vport_id)
+   continue;
+   else
+   return vport;
+   }
+
+   return vport;
+}
+
+st

RE: [PATCH v5 2/4] eal: allow applications to report their cpu usage

2022-12-16 Thread Morten Brørup
> From: Robin Jarry [mailto:rja...@redhat.com]
> Sent: Friday, 16 December 2022 11.21
> 
> Allow applications to register a callback that will be invoked in
> rte_lcore_dump() and when requesting lcore info in the telemetry API.
> 
> The callback is expected to return the number of TSC cycles that have
> passed since application start and the number of these cycles that were
> spent doing busy work.
> 
> Signed-off-by: Robin Jarry 
> Acked-by: Morten Brørup 
> ---
> v4 -> v5:
> 
> The callback now takes a pointer to a rte_lcore_usage structure.
> I chose not to include any API version tracking mechanism since the
> unsupported/unused fields can simply be left to zero. This is only
> telemetry after all.

ACK to this decision, with a minor clarification to avoid any misinterpretation:

The callback should not modify (i.e. zero out) unsupported/unused fields.

The caller needs to clear the structure before calling the callback - because 
the callback might not use the updated size of the structure, if the 
application was written for an older DPDK version with a smaller structure. I 
can see you already do this. Consider adding a comment about it in the code.

[...]

>  static int
>  lcore_dump_cb(unsigned int lcore_id, void *arg)
>  {
>   struct rte_config *cfg = rte_eal_get_configuration();
> - char cpuset[RTE_CPU_AFFINITY_STR_LEN];
> + char cpuset[RTE_CPU_AFFINITY_STR_LEN], usage_str[256];
> + struct rte_lcore_usage usage;
> + rte_lcore_usage_cb usage_cb;
>   const char *role;
>   FILE *f = arg;
>   int ret;
> @@ -446,11 +457,19 @@ lcore_dump_cb(unsigned int lcore_id, void *arg)
>   break;
>   }
> 
> + memset(&usage, 0, sizeof(usage));

I would move this memset() inside the below if-block.

> + usage_str[0] = '\0';
> + usage_cb = lcore_usage_cb;
> + if (usage_cb != NULL && usage_cb(lcore_id, &usage) == 0) {

Move memset() inside here, and add comment:

+ /* The application's callback may not set all the fields in the structure, so 
clear it here. */
+ memset(&usage, 0, sizeof(usage));

> + snprintf(usage_str, sizeof(usage_str), ", busy cycles
> %"PRIu64"/%"PRIu64,
> + usage.busy_cycles, usage.total_cycles);
> + }

[...]

> @@ -522,6 +543,12 @@ lcore_telemetry_info_cb(unsigned int lcore_id,
> void *arg)
>   if (CPU_ISSET(cpu, &lcore_config[lcore_id].cpuset))
>   rte_tel_data_add_array_int(cpuset, cpu);
>   rte_tel_data_add_dict_container(info->d, "cpuset", cpuset, 0);
> + memset(&usage, 0, sizeof(usage));
> + usage_cb = lcore_usage_cb;
> + if (usage_cb != NULL && usage_cb(lcore_id, &usage) == 0) {

Same comment as above: Move memset() inside here, and add a comment about why 
the structure is cleared here.

> + rte_tel_data_add_dict_u64(info->d, "busy_cycles",
> usage.busy_cycles);
> + rte_tel_data_add_dict_u64(info->d, "total_cycles",
> usage.total_cycles);
> + }
> 
>   return 0;
>  }
> diff --git a/lib/eal/include/rte_lcore.h b/lib/eal/include/rte_lcore.h
> index 6938c3fd7b81..a92313577355 100644
> --- a/lib/eal/include/rte_lcore.h
> +++ b/lib/eal/include/rte_lcore.h
> @@ -328,6 +328,41 @@ typedef int (*rte_lcore_iterate_cb)(unsigned int
> lcore_id, void *arg);
>  int
>  rte_lcore_iterate(rte_lcore_iterate_cb cb, void *arg);
> 
> +/**
> + * CPU usage statistics.
> + */
> +struct rte_lcore_usage {
> + uint64_t total_cycles;
> + /**< The total amount of time since application start, in TSC
> cycles. */
> + uint64_t busy_cycles;
> + /**< The amount of busy time since application start, in TSC
> cycles. */
> +};
> +
> +/**
> + * Callback to allow applications to report CPU usage.
> + *
> + * @param [in] lcore_id
> + *   The lcore to consider.
> + * @param [out] usage
> + *   Counters representing this lcore usage. This can never be NULL.
> + * @return
> + *   - 0 if fields in usage were updated successfully. The fields that
> the
> + *   application does not support should be left to their default
> value.

"should be left to their default value." -> "must not be modified."

> + *   - a negative value if the information is not available or if any
> error occurred.
> + */
> +typedef int (*rte_lcore_usage_cb)(unsigned int lcore_id, struct
> rte_lcore_usage *usage);




The link status of the 82599 NIC cannot be setted to down

2022-12-16 Thread wangyunjian
Hi All,
   It is expected that the ixgbe_dev_set_link_down function is called to make
the link status of the NIC down. However, the invoking is successful, but
link status is still up. Anyone have some good ideas?

dpdk version:DPDK 21.11
firmware-version:0x86d5
NIC:Intel Corporation 82599 10 Gigabit Dual Port Backplane Connection (rev 01)



[PATCH dpdk-kmod] linux/igb_uio: fix build with kernel 5.18

2022-12-16 Thread Ferruh Yigit
In Linux kernel v5.18, "pci-dma-compat.h" wrapper file is removed [1].

Some APIs in that wrapper file were used by igb_uio kernel module and
kernel module build fails after mentioned commit.

Fixed build for v5.18 by replacing APIs in igb_uio.

Replaced APIs are available in Linux v4.4 (minimum Linux kernel version
supported by DPDK), so no Linux version check is needed.

[1]
Commit 7968778914e5 ("PCI: Remove the deprecated "pci-dma-compat.h" API")

Bugzilla ID: 1142

Signed-off-by: Ferruh Yigit 
---
Cc: Daxue Gao 

Cc: Christian Ehrhardt 
This also needs to be backported to v19.11 LTS
---
 linux/igb_uio/igb_uio.c | 8 +---
 1 file changed, 1 insertion(+), 7 deletions(-)

diff --git a/linux/igb_uio/igb_uio.c b/linux/igb_uio/igb_uio.c
index 33e0e0286b69..0045e0fb737a 100644
--- a/linux/igb_uio/igb_uio.c
+++ b/linux/igb_uio/igb_uio.c
@@ -512,18 +512,12 @@ igbuio_pci_probe(struct pci_dev *dev, const struct 
pci_device_id *id)
goto fail_release_iomem;
 
/* set 64-bit DMA mask */
-   err = pci_set_dma_mask(dev,  DMA_BIT_MASK(64));
+   err = dma_set_mask_and_coherent(&dev->dev, DMA_BIT_MASK(64));
if (err != 0) {
dev_err(&dev->dev, "Cannot set DMA mask\n");
goto fail_release_iomem;
}
 
-   err = pci_set_consistent_dma_mask(dev, DMA_BIT_MASK(64));
-   if (err != 0) {
-   dev_err(&dev->dev, "Cannot set consistent DMA mask\n");
-   goto fail_release_iomem;
-   }
-
/* fill uio infos */
udev->info.name = "igb_uio";
udev->info.version = "0.1";
-- 
2.25.1



[PATCH v19.11 LTS] kni: fix build for Suse

2022-12-16 Thread Ferruh Yigit
Wrong macro is used in the patch that detects if '.ndo_tx_timeout'
has 'txqueue' parameter or not, which is causing build error.

Fixed by using correct macro.

Bugzilla ID: 1141
Fixes: d43fa3e198c0 ("kni: fix build for SLES15-SP3 (Make)")
Cc: sta...@dpdk.org

Signed-off-by: Ferruh Yigit 
---
Cc: christian.ehrha...@canonical.com
Cc: Daxue Gao 
---
 kernel/linux/kni/Makefile | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/kernel/linux/kni/Makefile b/kernel/linux/kni/Makefile
index bf0efab96b33..d125783213fb 100644
--- a/kernel/linux/kni/Makefile
+++ b/kernel/linux/kni/Makefile
@@ -22,7 +22,7 @@ MODULE_CFLAGS += -Wall -Werror
 ifdef CONFIG_SUSE_KERNEL
KSRC = /lib/modules/$(shell uname -r)/source
ifneq ($(shell grep -A 1 "ndo_tx_timeout" $(KSRC)/include/linux/netdevice.h 
| grep -o txqueue),)
-  MODULE_CFLAGS += -DHAVE_TX_TIMEOUT_TXQUEUE
+  MODULE_CFLAGS += -DHAVE_ARG_TX_QUEUE
endif
 endif
 
-- 
2.25.1



[Bug 1151] rte_eth_tx_buffer_init does not zeroize field rte_eth_dev_tx_buffer.length

2022-12-16 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1151

Bug ID: 1151
   Summary: rte_eth_tx_buffer_init does not zeroize field
rte_eth_dev_tx_buffer.length
   Product: DPDK
   Version: 22.11
  Hardware: All
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: ethdev
  Assignee: dev@dpdk.org
  Reporter: khiz...@gmail.com
  Target Milestone: ---

This is observed in LTS versions 19.11, 20.11, 21.11, 22.11 earlier I did not
look.
As a result, an unpredictable crash at the start of the program is possible.

-- 
You are receiving this mail because:
You are the assignee for the bug.

Re: [PATCH RESEND v2 00/11] codeql fixes for various subsystems

2022-12-16 Thread Sinan Kaya
On Tue, 2022-12-13 at 12:02 -0500, Sinan Kaya wrote:
> On Tue, 2022-12-13 at 19:59 +0300, Dmitry Kozlyuk wrote:
> > 2022-12-13 11:43 (UTC-0500), Sinan Kaya:
> > > Checking to see if I need to cover any feedback.
> > 
> > There was some feedback to pre-resend v2 that is not visible on
> > patchwork:
> > 
> > http://inbox.dpdk.org/dev/cajfav8yee9awya3ovluqavfw0mzonsqjq-k+2q4wampmyrj...@mail.gmail.com/
> > 
> > 
> > 
> 
> Yes, I sent it again with RESEND tag this time on this series.
> Just checking to see if it went through this time.
> I do see the RESEND series on patchwork.
> 
> 

I was planning to send the fixes in batches hoping this one to
merge first. I guess I'll post the remaning bugfixes using a
different series.



[Bug 1152] net/mlx5 - direct ibverbs dependency when compiling with -Dibverbs_link=dlopen

2022-12-16 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1152

Bug ID: 1152
   Summary: net/mlx5 - direct ibverbs dependency when compiling
with -Dibverbs_link=dlopen
   Product: DPDK
   Version: 21.11
  Hardware: All
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: ethdev
  Assignee: dev@dpdk.org
  Reporter: khiz...@gmail.com
  Target Milestone: ---

This commit
http://git.dpdk.org/dpdk/commit/?id=2235fcda125763fa1a5b3c341b2799de9e44d479
break `dlopen` interacting mode with mlx OFED. 

It introduces direct dependency from mlx ibverbs libs escaping glue level.
For many applications, direct linkage with mlx libs is forbidden.

-- 
You are receiving this mail because:
You are the assignee for the bug.

[Bug 1153] net/i40e: not all tx error counts in NIC statistics

2022-12-16 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1153

Bug ID: 1153
   Summary: net/i40e: not all tx error counts in NIC statistics
   Product: DPDK
   Version: 21.11
  Hardware: All
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: ethdev
  Assignee: dev@dpdk.org
  Reporter: khiz...@gmail.com
  Target Milestone: ---

in i40_ethdev.c, function i40e_read_stats_registers:

```
stats->oerrors  = ns->eth.tx_errors +
  pf->main_vsi->eth_stats.tx_errors;
```

This code does not take into account `ns->tx_dropped_link_down` counter.
Correct oerrors should be:

```
stats->oerrors  = ns->eth.tx_errors +
  pf->main_vsi->eth_stats.tx_errors +
  ns->tx_dropped_link_down
```

-- 
You are receiving this mail because:
You are the assignee for the bug.

[PATCH v2 0/2] Enable the lcores test on Windows instead of skipping it.

2022-12-16 Thread Tyler Retzlaff
Two bugs are fixed to allow this test to build, run & pass.
  * Mark memory configuration complete during rte_eal_init()
  * Use rte thread api to get a proper implementation of thread join.

v2:
  * update commit message to clarify why this test is beneficial on Windows.
  * add missing Fixes tag

Tyler Retzlaff (2):
  eal: add missing call marking memory config complete
  test: enable lcores test on Windows

 app/test/test_lcores.c | 28 +---
 lib/eal/windows/eal.c  |  3 +++
 2 files changed, 16 insertions(+), 15 deletions(-)

-- 
1.8.3.1



[PATCH v2 1/2] eal: add missing call marking memory config complete

2022-12-16 Thread Tyler Retzlaff
Memory configuration was not being marked as completed add the missing
call to rte_eal_init() for Windows.

Allows rte_thread_register to work on Windows and lcores_autotest to be
built and run Windows which also exercises the rte_ctrl_thread_create
API on Windows.

Fixes: 5c307ba2a5b1 ("eal: register non-EAL threads as lcores")

Signed-off-by: Tyler Retzlaff 
---
 lib/eal/windows/eal.c | 3 +++
 1 file changed, 3 insertions(+)

diff --git a/lib/eal/windows/eal.c b/lib/eal/windows/eal.c
index adb929a..56fadc7 100644
--- a/lib/eal/windows/eal.c
+++ b/lib/eal/windows/eal.c
@@ -462,6 +462,9 @@ enum rte_proc_type_t
 */
rte_eal_mp_remote_launch(sync_func, NULL, SKIP_MAIN);
rte_eal_mp_wait_lcore();
+
+   eal_mcfg_complete();
+
return fctret;
 }
 
-- 
1.8.3.1



[PATCH v2 2/2] test: enable lcores test on Windows

2022-12-16 Thread Tyler Retzlaff
Stop using pthread and convert the test to use EAL thread APIs. Because
the EAL thread APIs provide more than just a stub for thread join on
Windows the tests now pass and need not be skipped.

Signed-off-by: Tyler Retzlaff 
---
 app/test/test_lcores.c | 28 +---
 1 file changed, 13 insertions(+), 15 deletions(-)

diff --git a/app/test/test_lcores.c b/app/test/test_lcores.c
index a6bb412..5b43aa5 100644
--- a/app/test/test_lcores.c
+++ b/app/test/test_lcores.c
@@ -8,17 +8,18 @@
 #include 
 #include 
 #include 
+#include 
 
 #include "test.h"
 
 struct thread_context {
enum { Thread_INIT, Thread_ERROR, Thread_DONE } state;
bool lcore_id_any;
-   pthread_t id;
+   rte_thread_t id;
unsigned int *registered_count;
 };
 
-static void *thread_loop(void *arg)
+static uint32_t thread_loop(void *arg)
 {
struct thread_context *t = arg;
unsigned int lcore_id;
@@ -55,7 +56,7 @@ static void *thread_loop(void *arg)
if (t->state != Thread_ERROR)
t->state = Thread_DONE;
 
-   return NULL;
+   return 0;
 }
 
 static int
@@ -77,7 +78,7 @@ static void *thread_loop(void *arg)
t->state = Thread_INIT;
t->registered_count = ®istered_count;
t->lcore_id_any = false;
-   if (pthread_create(&t->id, NULL, thread_loop, t) != 0)
+   if (rte_thread_create(&t->id, NULL, thread_loop, t) != 0)
break;
non_eal_threads_count++;
}
@@ -96,7 +97,7 @@ static void *thread_loop(void *arg)
t->state = Thread_INIT;
t->registered_count = ®istered_count;
t->lcore_id_any = true;
-   if (pthread_create(&t->id, NULL, thread_loop, t) == 0) {
+   if (rte_thread_create(&t->id, NULL, thread_loop, t) == 0) {
non_eal_threads_count++;
printf("non-EAL threads count: %u\n", non_eal_threads_count);
while (__atomic_load_n(®istered_count, __ATOMIC_ACQUIRE) !=
@@ -110,7 +111,7 @@ static void *thread_loop(void *arg)
ret = 0;
for (i = 0; i < non_eal_threads_count; i++) {
t = &thread_contexts[i];
-   pthread_join(t->id, NULL);
+   rte_thread_join(t->id, NULL);
if (t->state != Thread_DONE)
ret = -1;
}
@@ -262,7 +263,7 @@ struct limit_lcore_context {
t->state = Thread_INIT;
t->registered_count = ®istered_count;
t->lcore_id_any = false;
-   if (pthread_create(&t->id, NULL, thread_loop, t) != 0)
+   if (rte_thread_create(&t->id, NULL, thread_loop, t) != 0)
goto cleanup_threads;
non_eal_threads_count++;
while (__atomic_load_n(®istered_count, __ATOMIC_ACQUIRE) !=
@@ -285,7 +286,7 @@ struct limit_lcore_context {
t->state = Thread_INIT;
t->registered_count = ®istered_count;
t->lcore_id_any = true;
-   if (pthread_create(&t->id, NULL, thread_loop, t) != 0)
+   if (rte_thread_create(&t->id, NULL, thread_loop, t) != 0)
goto cleanup_threads;
non_eal_threads_count++;
while (__atomic_load_n(®istered_count, __ATOMIC_ACQUIRE) !=
@@ -309,7 +310,7 @@ struct limit_lcore_context {
ret = 0;
for (i = 0; i < non_eal_threads_count; i++) {
t = &thread_contexts[i];
-   pthread_join(t->id, NULL);
+   rte_thread_join(t->id, NULL);
if (t->state != Thread_DONE)
ret = -1;
}
@@ -330,7 +331,7 @@ struct limit_lcore_context {
__atomic_store_n(®istered_count, 0, __ATOMIC_RELEASE);
for (i = 0; i < non_eal_threads_count; i++) {
t = &thread_contexts[i];
-   pthread_join(t->id, NULL);
+   rte_thread_join(t->id, NULL);
}
 error:
if (handle[1] != NULL)
@@ -361,7 +362,7 @@ static void *ctrl_thread_loop(void *arg)
/* Create one control thread */
t = &ctrl_thread_context;
t->state = Thread_INIT;
-   if (rte_ctrl_thread_create(&t->id, "test_ctrl_threads",
+   if (rte_ctrl_thread_create((pthread_t *)&t->id, "test_ctrl_threads",
NULL, ctrl_thread_loop, t) != 0)
return -1;
 
@@ -369,7 +370,7 @@ static void *ctrl_thread_loop(void *arg)
 * This also acts as the barrier such that the memory operations
 * in control thread are visible to this thread.
 */
-   pthread_join(t->id, NULL);
+   rte_thread_join(t->id, NULL);
 
/* Check if the control thread set the correct state */
if (t->state != Thread_DONE)
@@ -384,9 +385,6 @@ static void *ctrl_thread_loop(void *arg)
unsigned int eal_threads_count = 0;
unsigned int i;
 
-   if (RTE_EXEC_ENV_IS_WINDOWS)
-   return TEST_SKIPPED;
-
for (i = 0; i < RTE_MAX_LCORE; i++) {
if (!rte_lcore_has_role(i,

[PATCH kmods] linux/igb_uio: fix build on Linux 5.18+

2022-12-16 Thread luca . boccassi
From: Luca Boccassi 

Reported-by: Paolo Pisati 
Signed-off-by: Luca Boccassi 
---
Only build-tested.

 linux/igb_uio/igb_uio.c | 8 
 1 file changed, 8 insertions(+)

diff --git a/linux/igb_uio/igb_uio.c b/linux/igb_uio/igb_uio.c
index 33e0e02..3d8a8a8 100644
--- a/linux/igb_uio/igb_uio.c
+++ b/linux/igb_uio/igb_uio.c
@@ -512,13 +512,21 @@ igbuio_pci_probe(struct pci_dev *dev, const struct 
pci_device_id *id)
goto fail_release_iomem;
 
/* set 64-bit DMA mask */
+#if LINUX_VERSION_CODE < KERNEL_VERSION(5, 18, 0)
err = pci_set_dma_mask(dev,  DMA_BIT_MASK(64));
+#else
+   err = dma_set_mask(&dev->dev, DMA_BIT_MASK(64));
+#endif
if (err != 0) {
dev_err(&dev->dev, "Cannot set DMA mask\n");
goto fail_release_iomem;
}
 
+#if LINUX_VERSION_CODE < KERNEL_VERSION(5, 18, 0)
err = pci_set_consistent_dma_mask(dev, DMA_BIT_MASK(64));
+#else
+   err = dma_set_coherent_mask(&dev->dev, DMA_BIT_MASK(64));
+#endif
if (err != 0) {
dev_err(&dev->dev, "Cannot set consistent DMA mask\n");
goto fail_release_iomem;
-- 
2.34.1



Re: [PATCH dpdk-kmod] linux/igb_uio: fix build with kernel 5.18

2022-12-16 Thread Stephen Hemminger
On Fri, 16 Dec 2022 11:57:32 +
Ferruh Yigit  wrote:

> In Linux kernel v5.18, "pci-dma-compat.h" wrapper file is removed [1].
> 
> Some APIs in that wrapper file were used by igb_uio kernel module and
> kernel module build fails after mentioned commit.
> 
> Fixed build for v5.18 by replacing APIs in igb_uio.
> 
> Replaced APIs are available in Linux v4.4 (minimum Linux kernel version
> supported by DPDK), so no Linux version check is needed.
> 

Makes sense, but lets put the correct versions in the commit description.
DPDK should not be trying to support unsupported kernel versions.

5.18 is End of Life now and so the one that matters is really 6.0.

Current DPDK 22 should be supporting last supported LTS which
in January will be 4.14


[PATCH] eventdev/timer: add API to get remaining ticks

2022-12-16 Thread Erik Gabriel Carrillo
Introduce an event timer adapter API which allows users to determine how
many adapter ticks remain until an event timer fires.

Signed-off-by: Erik Gabriel Carrillo 
---
 app/test/test_event_timer_adapter.c| 68 ++
 lib/eventdev/event_timer_adapter_pmd.h |  7 +++
 lib/eventdev/rte_event_timer_adapter.c | 52 
 lib/eventdev/rte_event_timer_adapter.h | 27 ++
 lib/eventdev/version.map   |  3 ++
 5 files changed, 157 insertions(+)

diff --git a/app/test/test_event_timer_adapter.c 
b/app/test/test_event_timer_adapter.c
index 1a440dfd10..6529b14ff9 100644
--- a/app/test/test_event_timer_adapter.c
+++ b/app/test/test_event_timer_adapter.c
@@ -1920,6 +1920,72 @@ adapter_create_max(void)
return TEST_SUCCESS;
 }
 
+static inline int
+test_timer_ticks_remaining(void)
+{
+   uint64_t ticks_remaining = UINT64_MAX;
+   struct rte_event_timer *ev_tim;
+   struct rte_event ev;
+   int ret, i;
+   const struct rte_event_timer tim = {
+   .ev.op = RTE_EVENT_OP_NEW,
+   .ev.queue_id = 0,
+   .ev.sched_type = RTE_SCHED_TYPE_ATOMIC,
+   .ev.priority = RTE_EVENT_DEV_PRIORITY_NORMAL,
+   .ev.event_type =  RTE_EVENT_TYPE_TIMER,
+   .state = RTE_EVENT_TIMER_NOT_ARMED,
+   };
+
+   rte_mempool_get(eventdev_test_mempool, (void **)&ev_tim);
+   *ev_tim = tim;
+   ev_tim->ev.event_ptr = ev_tim;
+#define TEST_TICKS 5
+   ev_tim->timeout_ticks = CALC_TICKS(TEST_TICKS);
+
+   /* Test that unarmed timer returns error */
+   TEST_ASSERT_FAIL(rte_event_timer_ticks_remaining_get(timdev, ev_tim,
+&ticks_remaining),
+"Didn't fail to get ticks for unarmed event timer");
+
+   TEST_ASSERT_EQUAL(rte_event_timer_arm_burst(timdev, &ev_tim, 1), 1,
+ "Failed to arm timer with proper timeout.");
+   TEST_ASSERT_EQUAL(ev_tim->state, RTE_EVENT_TIMER_ARMED,
+ "Improper timer state set expected %d returned %d",
+ RTE_EVENT_TIMER_ARMED, ev_tim->state);
+
+   for (i = 0; i < TEST_TICKS; i++) {
+   ret = rte_event_timer_ticks_remaining_get(timdev, ev_tim,
+ &ticks_remaining);
+   if (ret < 0)
+   return TEST_FAILED;
+
+   TEST_ASSERT_EQUAL((int)ticks_remaining, TEST_TICKS - i,
+ "Expected %d ticks remaining, got %"PRIu64"",
+ TEST_TICKS - i, ticks_remaining);
+
+   rte_delay_ms(100);
+   }
+
+   rte_delay_ms(100);
+
+   TEST_ASSERT_EQUAL(rte_event_dequeue_burst(evdev, 0, &ev, 1, 0), 1,
+ "Armed timer failed to trigger.");
+   TEST_ASSERT_EQUAL(ev_tim->state, RTE_EVENT_TIMER_NOT_ARMED,
+ "Improper timer state set expected %d returned %d",
+ RTE_EVENT_TIMER_NOT_ARMED, ev_tim->state);
+
+   /* Test that timer that fired returns error */
+   TEST_ASSERT_FAIL(rte_event_timer_ticks_remaining_get(timdev, ev_tim,
+  &ticks_remaining),
+"Didn't fail to get ticks for unarmed event timer");
+
+   rte_mempool_put(eventdev_test_mempool, (void *)ev_tim);
+
+#undef TEST_TICKS
+   return TEST_SUCCESS;
+}
+
+
 static struct unit_test_suite event_timer_adptr_functional_testsuite  = {
.suite_name = "event timer functional test suite",
.setup = testsuite_setup,
@@ -1982,6 +2048,8 @@ static struct unit_test_suite 
event_timer_adptr_functional_testsuite  = {
TEST_CASE_ST(timdev_setup_msec, timdev_teardown,
adapter_tick_resolution),
TEST_CASE(adapter_create_max),
+   TEST_CASE_ST(timdev_setup_msec, timdev_teardown,
+   test_timer_ticks_remaining),
TEST_CASES_END() /**< NULL terminate unit test array */
}
 };
diff --git a/lib/eventdev/event_timer_adapter_pmd.h 
b/lib/eventdev/event_timer_adapter_pmd.h
index 189017b5c1..c19ff3576a 100644
--- a/lib/eventdev/event_timer_adapter_pmd.h
+++ b/lib/eventdev/event_timer_adapter_pmd.h
@@ -52,6 +52,11 @@ typedef int (*rte_event_timer_adapter_stats_get_t)(
 typedef int (*rte_event_timer_adapter_stats_reset_t)(
const struct rte_event_timer_adapter *adapter);
 /**< @internal Reset statistics for event timer adapter */
+typedef int (*rte_event_timer_ticks_remaining_get_t)(
+   const struct rte_event_timer_adapter *adapter,
+   const struct rte_event_timer *evtim,
+   uint64_t *ticks_remaining);
+/**< @internal Get remaining ticks for event timer */
 
 /**
  * @internal Structure containing the functions exported by an event timer
@@ -