Hi Yuan,
On 12/16/22 03:00, Yuan Wang wrote:
Vhost asynchronous data-path offloads packet copy from the CPU
to the DMA engine. As a result, large packet copy can be accelerated
by the DMA engine, and vhost can free CPU cycles for higher level
functions.
In this patch, we enable asynchronous data-path for vhostpmd.
Asynchronous data path is enabled per tx/rx queue, and users need
to specify the DMA device used by the tx/rx queue. Each tx/rx queue
only supports to use one DMA device, but one DMA device can be shared
among multiple tx/rx queues of different vhost PMD ports.
Two PMD parameters are added:
- dmas: specify the used DMA device for a tx/rx queue.
(Default: no queues enable asynchronous data path)
- dma-ring-size: DMA ring size.
(Default: 4096).
Here is an example:
--vdev
'eth_vhost0,iface=./s0,dmas=[txq0@0000:00.01.0;rxq0@0000:00.01.1],dma-ring-size=4096'
Signed-off-by: Jiayu Hu <jiayu...@intel.com>
Signed-off-by: Yuan Wang <yuanx.w...@intel.com>
Signed-off-by: Wenwu Ma <wenwux...@intel.com>
---
drivers/net/vhost/meson.build | 1 +
drivers/net/vhost/rte_eth_vhost.c | 512 ++++++++++++++++++++++++++++--
drivers/net/vhost/rte_eth_vhost.h | 15 +
drivers/net/vhost/version.map | 7 +
drivers/net/vhost/vhost_testpmd.c | 67 ++++
5 files changed, 569 insertions(+), 33 deletions(-)
create mode 100644 drivers/net/vhost/vhost_testpmd.c
This RFC is identical to the v5 that you sent for last release, and so
the comments I made on it are still valid.
Is this intentionally re-sent?
Regards,
Maxime