RE: [PATCH] app/pdump: make dpdk-pdump can run multiple simultaneously

2023-05-08 Thread Nole Zhang


> -Original Message-
> From: Stephen Hemminger 
> Sent: Saturday, May 6, 2023 11:48 PM
> To: Chaoyong He 
> Cc: dev@dpdk.org; oss-drivers ; Niklas
> Soderlund ; Nole Zhang
> 
> Subject: Re: [PATCH] app/pdump: make dpdk-pdump can run multiple
> simultaneously
> 
> On Sat,  6 May 2023 16:52:06 +0800
> Chaoyong He  wrote:
> 
> > From: Peng Zhang 
> >
> > The pdump doesn't support start the multiple separate process for
> > different queues or ports to monitor the packets, because it apply the
> > same name for memory space. This commit will use the device_id, port
> > and queue to name the memory space and it will support this function.
> >
> > Adjust the value of RTE_MEMZONE_NAMESIZE to adapt the name length.
> >
> > Signed-off-by: Peng Zhang 
> > Reviewed-by: Chaoyong He 
> > Reviewed-by: Niklas Söderlund 
> 
> FYI - app/dumpcap does support multiple ports already.
> Why not use that?
Yes, you are right, I will abandon this patch, thanks for your reply



DPDK 22.11.2 released

2023-05-08 Thread Xueming Li
Hi all,

Here is a new stable release:
https://fast.dpdk.org/rel/dpdk-22.11.2.tar.xz

The git tree is at:
https://dpdk.org/browse/dpdk-stable/?h=22.11

Xueming Li 

---
 .github/workflows/build.yml|3 +-
 .mailmap   |   17 +
 .travis.yml|3 +-
 MAINTAINERS|2 +-
 VERSION|2 +-
 app/dumpcap/main.c |1 +
 app/test-bbdev/test_bbdev_perf.c   |  124 ++-
 app/test-compress-perf/comp_perf_options.h |6 +-
 app/test-compress-perf/comp_perf_options_parse.c   |8 +-
 app/test-compress-perf/comp_perf_test_common.c |  127 ++-
 app/test-compress-perf/comp_perf_test_cyclecount.c |   71 +-
 app/test-compress-perf/comp_perf_test_throughput.c |   64 +-
 app/test-compress-perf/comp_perf_test_verify.c |   64 +-
 app/test-compress-perf/main.c  |8 +
 app/test-crypto-perf/cperf_ops.c   |   37 +-
 app/test-crypto-perf/cperf_options.h   |1 +
 app/test-crypto-perf/cperf_options_parsing.c   |   16 +
 app/test-crypto-perf/cperf_test_common.c   |8 +-
 app/test-crypto-perf/cperf_test_latency.c  |   27 +-
 app/test-crypto-perf/cperf_test_vector_parsing.c   |1 +
 app/test-crypto-perf/cperf_test_verify.c   |   25 +-
 app/test-flow-perf/main.c  |7 +-
 app/test-pmd/cmdline.c |   29 +-
 app/test-pmd/cmdline_flow.c|   48 +-
 app/test-pmd/csumonly.c|   17 +-
 app/test-pmd/ieee1588fwd.c |2 +-
 app/test-pmd/noisy_vnf.c   |5 +-
 app/test-pmd/testpmd.c |  206 ++--
 app/test-pmd/testpmd.h |1 +
 app/test/packet_burst_generator.c  |   26 +-
 app/test/test_cryptodev.c  |   60 +-
 app/test/test_cryptodev_aes_test_vectors.h |   16 +-
 .../test_cryptodev_security_pdcp_test_vectors.h|8 +-
 app/test/test_mbuf.c   |2 +
 app/test/test_reorder.c|2 +
 config/meson.build |9 +-
 config/rte_config.h|1 -
 devtools/check-git-log.sh  |   17 +-
 devtools/checkpatches.sh   |   30 -
 doc/api/doxy-api-index.md  |1 +
 doc/guides/linux_gsg/enable_func.rst   |4 +-
 doc/guides/linux_gsg/sys_reqs.rst  |2 +-
 doc/guides/nics/bnxt.rst   |1 +
 doc/guides/nics/ice.rst|8 +-
 doc/guides/nics/mana.rst   |8 +-
 doc/guides/nics/mlx5.rst   |3 +
 doc/guides/prog_guide/cryptodev_lib.rst|  163 +--
 doc/guides/prog_guide/event_timer_adapter.rst  |2 +-
 doc/guides/prog_guide/img/cryptodev_sym_sess.svg   |  417 
 doc/guides/prog_guide/rte_flow.rst |   13 +-
 doc/guides/rel_notes/release_22_11.rst |  371 +++
 doc/guides/sample_app_ug/l2_forward_cat.rst|9 +-
 doc/guides/sample_app_ug/l3_forward.rst|3 +-
 doc/guides/sample_app_ug/pipeline.rst  |2 +-
 drivers/baseband/acc/rte_acc100_pmd.c  |   19 +-
 drivers/baseband/acc/rte_acc200_pmd.c  |   61 +-
 drivers/baseband/turbo_sw/meson.build  |   10 +-
 drivers/bus/fslmc/mc/mc_sys.c  |5 +-
 drivers/bus/ifpga/ifpga_bus.c  |3 +-
 drivers/bus/pci/pci_common.c   |3 +-
 drivers/bus/vdev/vdev.c|5 +-
 drivers/common/cnxk/roc_io.h   |   12 +-
 drivers/common/cnxk/roc_mbox.h |2 +-
 drivers/common/cnxk/roc_nix_inl_dev.c  |2 +-
 drivers/common/cnxk/roc_npc.c  |9 +-
 drivers/common/cnxk/roc_npc.h  |   11 +
 drivers/common/cnxk/roc_npc_mcam.c |   58 +-
 drivers/common/cnxk/roc_npc_mcam_dump.c|6 +-
 drivers/common/cnxk/roc_npc_parse.c|  384 +--
 drivers/common/cnxk/roc_npc_priv.h |   20 +-
 drivers/common/cnxk/roc_npc_utils.c|2 +-
 drivers/common/cnxk/roc_se.h   |7 +-
 drivers/common/mlx5/meson.build|9 +-
 drivers/common/mlx5/mlx5_common.h  |8 +-
 drivers/common/mlx5/mlx5_devx_cmds.c   |4 +-
 drivers/common/mlx5/mlx5_prm.h |   16 +-
 drivers/common/mlx5/version.map|1 +
 drivers/common/mlx5/windows/mlx5_w

RE: [PATCH 1/3] net/mlx5: support multi-packet RQ on Windows

2023-05-08 Thread Pier Damouny
> Subject: [PATCH 1/3] net/mlx5: support multi-packet RQ on Windows
> 
> Multi-Packet RQ can further save PCIe bandwidth by posting a single large
> buffer for multiple packets.
> 
> Instead of posting a buffer per a packet, one large buffer is posted to 
> receive
> multiple packets on the buffer.
> 
> Add support for multi-packet RQ on Windows.
> The feature is disabled by default and can by enabled by setting mprq_en=1
> in the PMD specific arguments.
> 
> Signed-off-by: Tal Shnaiderman 
> ---
>  drivers/common/mlx5/mlx5_devx_cmds.c|  3 +++
>  drivers/common/mlx5/mlx5_devx_cmds.h|  2 ++
>  drivers/common/mlx5/windows/mlx5_win_defs.h |  8 +++-
>  drivers/net/mlx5/windows/mlx5_os.c  | 26
> ++
>  4 files changed, 38 insertions(+), 1 deletion(-)
> 
> diff --git a/drivers/common/mlx5/mlx5_devx_cmds.c
> b/drivers/common/mlx5/mlx5_devx_cmds.c
> index d0907fcd49..096bd1d520 100644
> --- a/drivers/common/mlx5/mlx5_devx_cmds.c
> +++ b/drivers/common/mlx5/mlx5_devx_cmds.c
> @@ -1076,6 +1076,9 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
>general_obj_types) &
> 
> MLX5_GENERAL_OBJ_TYPES_CAP_CONN_TRACK_OFFLOAD);
>   attr->rq_delay_drop = MLX5_GET(cmd_hca_cap, hcattr,
> rq_delay_drop);
> + attr->striding_rq = MLX5_GET(cmd_hca_cap, hcattr, striding_rq);
> + attr->ext_stride_num_range =
> + MLX5_GET(cmd_hca_cap, hcattr, ext_stride_num_range);
>   attr->max_flow_counter_15_0 = MLX5_GET(cmd_hca_cap, hcattr,
>   max_flow_counter_15_0);
>   attr->max_flow_counter_31_16 = MLX5_GET(cmd_hca_cap, hcattr,
> diff --git a/drivers/common/mlx5/mlx5_devx_cmds.h
> b/drivers/common/mlx5/mlx5_devx_cmds.h
> index ce173bc36a..9e7992b1c6 100644
> --- a/drivers/common/mlx5/mlx5_devx_cmds.h
> +++ b/drivers/common/mlx5/mlx5_devx_cmds.h
> @@ -282,6 +282,8 @@ struct mlx5_hca_attr {
>   uint32_t crypto_wrapped_import_method:1;
>   uint16_t esw_mgr_vport_id; /* E-Switch Mgr vport ID . */
>   uint16_t max_wqe_sz_sq;
> + uint32_t striding_rq:1;
> + uint32_t ext_stride_num_range:1;
>   uint32_t set_reg_c:8;
>   uint32_t nic_flow_table:1;
>   uint32_t modify_outer_ip_ecn:1;
> diff --git a/drivers/common/mlx5/windows/mlx5_win_defs.h
> b/drivers/common/mlx5/windows/mlx5_win_defs.h
> index 65da820c5e..885114655f 100644
> --- a/drivers/common/mlx5/windows/mlx5_win_defs.h
> +++ b/drivers/common/mlx5/windows/mlx5_win_defs.h
> @@ -270,4 +270,10 @@ enum {
>   MLX5_MATCH_INNER_HEADERS= RTE_BIT32(2),
>  };
> 
> -#endif /* MLX5_WIN_DEFS_H */
> +#define MLX5_MIN_SINGLE_WQE_LOG_NUM_STRIDES 9 #define
> +MLX5_MAX_SINGLE_WQE_LOG_NUM_STRIDES 16 #define
> +MLX5_MIN_SINGLE_STRIDE_LOG_NUM_BYTES 6 #define
> +MLX5_MAX_SINGLE_STRIDE_LOG_NUM_BYTES 13 #define
> +MLX5_EXT_MIN_SINGLE_WQE_LOG_NUM_STRIDES 3 #define
> IB_QPT_RAW_PACKET 8
> +#endif /* __MLX5_WIN_DEFS_H__ */
> diff --git a/drivers/net/mlx5/windows/mlx5_os.c
> b/drivers/net/mlx5/windows/mlx5_os.c
> index f401264b61..0caa8931e4 100644
> --- a/drivers/net/mlx5/windows/mlx5_os.c
> +++ b/drivers/net/mlx5/windows/mlx5_os.c
> @@ -187,6 +187,32 @@ mlx5_os_capabilities_prepare(struct
> mlx5_dev_ctx_shared *sh)
>   if (sh->dev_cap.tso)
>   sh->dev_cap.tso_max_payload_sz = 1 << hca_attr-
> >max_lso_cap;
>   DRV_LOG(DEBUG, "Counters are not supported.");
> + if (hca_attr->striding_rq) {
> + sh->dev_cap.mprq.enabled = 1;
> + sh->dev_cap.mprq.log_min_stride_size =
> + MLX5_MIN_SINGLE_STRIDE_LOG_NUM_BYTES;
> + sh->dev_cap.mprq.log_max_stride_size =
> + MLX5_MAX_SINGLE_STRIDE_LOG_NUM_BYTES;
> + if (hca_attr->ext_stride_num_range)
> + sh->dev_cap.mprq.log_min_stride_num =
> +
>   MLX5_EXT_MIN_SINGLE_WQE_LOG_NUM_STRIDES;
> + else
> + sh->dev_cap.mprq.log_min_stride_num =
> +
>   MLX5_MIN_SINGLE_WQE_LOG_NUM_STRIDES;
> + sh->dev_cap.mprq.log_max_stride_num =
> + MLX5_MAX_SINGLE_WQE_LOG_NUM_STRIDES;
> + DRV_LOG(DEBUG, "\tmin_single_stride_log_num_of_bytes:
> %u",
> + sh->dev_cap.mprq.log_min_stride_size);
> + DRV_LOG(DEBUG,
> "\tmax_single_stride_log_num_of_bytes: %u",
> + sh->dev_cap.mprq.log_max_stride_size);
> + DRV_LOG(DEBUG, "\tmin_single_wqe_log_num_of_strides:
> %u",
> + sh->dev_cap.mprq.log_min_stride_num);
> + DRV_LOG(DEBUG, "\tmax_single_wqe_log_num_of_strides:
> %u",
> + sh->dev_cap.mprq.log_max_stride_num);
> + DRV_LOG(DEBUG, "\tmin_stride_wqe_log_size: %u",
> + sh->dev_cap.mprq.log_min_stride_wqe_size);
> + DRV_LOG(DEBUG, "Device supports Multi-Packet RQ.");
> + }
>   if (hca_attr->rss_ind_tbl_cap) {
>   /*
>* DPDK doesn't support larg

RE: [PATCH 2/3] net/mlx5: support CQE compression on Windows

2023-05-08 Thread Pier Damouny
> Subject: [PATCH 2/3] net/mlx5: support CQE compression on Windows
> 
> CQE Compression reduces PCI overhead by coalescing and compressing
> multiple CQEs into a single merged CQE.
> 
> Add supported for the CQE compression feature on Windows.
> feature is enabled by default unless not supported by the HW or if the
> rxq_cqe_comp_en PMD argument is explicitly disabled.
> 
> Signed-off-by: Tal Shnaiderman 
> ---
>  drivers/common/mlx5/mlx5_devx_cmds.c |  2 ++
> drivers/common/mlx5/mlx5_devx_cmds.h |  1 +
>  drivers/net/mlx5/windows/mlx5_os.c   | 12 
>  3 files changed, 15 insertions(+)
> 
> diff --git a/drivers/common/mlx5/mlx5_devx_cmds.c
> b/drivers/common/mlx5/mlx5_devx_cmds.c
> index 096bd1d520..a31e4995f5 100644
> --- a/drivers/common/mlx5/mlx5_devx_cmds.c
> +++ b/drivers/common/mlx5/mlx5_devx_cmds.c
> @@ -1062,6 +1062,8 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
>   attr->cqe_compression = MLX5_GET(cmd_hca_cap, hcattr,
> cqe_compression);
>   attr->mini_cqe_resp_flow_tag = MLX5_GET(cmd_hca_cap, hcattr,
>   mini_cqe_resp_flow_tag);
> + attr->cqe_compression_128 = MLX5_GET(cmd_hca_cap, hcattr,
> + cqe_compression_128);
>   attr->mini_cqe_resp_l3_l4_tag = MLX5_GET(cmd_hca_cap, hcattr,
>mini_cqe_resp_l3_l4_tag);
>   attr->enhanced_cqe_compression = MLX5_GET(cmd_hca_cap,
> hcattr, diff --git a/drivers/common/mlx5/mlx5_devx_cmds.h
> b/drivers/common/mlx5/mlx5_devx_cmds.h
> index 9e7992b1c6..edcd867c4e 100644
> --- a/drivers/common/mlx5/mlx5_devx_cmds.h
> +++ b/drivers/common/mlx5/mlx5_devx_cmds.h
> @@ -284,6 +284,7 @@ struct mlx5_hca_attr {
>   uint16_t max_wqe_sz_sq;
>   uint32_t striding_rq:1;
>   uint32_t ext_stride_num_range:1;
> + uint32_t cqe_compression_128:1;
>   uint32_t set_reg_c:8;
>   uint32_t nic_flow_table:1;
>   uint32_t modify_outer_ip_ecn:1;
> diff --git a/drivers/net/mlx5/windows/mlx5_os.c
> b/drivers/net/mlx5/windows/mlx5_os.c
> index 0caa8931e4..6527269663 100644
> --- a/drivers/net/mlx5/windows/mlx5_os.c
> +++ b/drivers/net/mlx5/windows/mlx5_os.c
> @@ -237,6 +237,18 @@ mlx5_os_capabilities_prepare(struct
> mlx5_dev_ctx_shared *sh)
>   } else {
>   DRV_LOG(DEBUG, "Tunnel offloading is not supported.");
>   }
> + sh->dev_cap.cqe_comp = 0;
> +#if (RTE_CACHE_LINE_SIZE == 128)
> + if (hca_attr->cqe_compression_128)
> + sh->dev_cap.cqe_comp = 1;
> + DRV_LOG(DEBUG, "Rx CQE 128B compression is %ssupported.",
> + sh->dev_cap.cqe_comp ? "" : "not ");
> +#else
> + if (hca_attr->cqe_compression)
> + sh->dev_cap.cqe_comp = 1;
> + DRV_LOG(DEBUG, "Rx CQE compression is %ssupported.",
> + sh->dev_cap.cqe_comp ? "" : "not ");
> +#endif
>   snprintf(sh->dev_cap.fw_ver, 64, "%x.%x.%04x",
>MLX5_GET(initial_seg, pv_iseg, fw_rev_major),
>MLX5_GET(initial_seg, pv_iseg, fw_rev_minor),
> --
> 2.16.1.windows.4

Tested-by: Pier Damouny  


RE: [PATCH 3/3] net/mlx5: support enhanced multi-packet write on Windows

2023-05-08 Thread Pier Damouny
> Subject: [PATCH 3/3] net/mlx5: support enhanced multi-packet write on
> Windows
> 
> Add support for enhanced multi-packet write on Windows.
> 
> Enhanced multi-packet write allows the Tx burst function to pack up multiple
> packets in a single descriptor session to save PCI bandwidth and improve
> performance.
> 
> The feature can be controlled by the txq_mpw_en PMD argument:
> 
> txq_mpw_en=1 - PMD will first attempt to use "enhanced multi packet
> write"
> if the feature is not supported by the HW the legacy "multi packet write"
> will be used.
> if both are unsupported the multi packet write feature is disabled.
> 
> txq_mpw_en=0 - multi packet write is disabled.
> 
> txq_mpw_en unset(default) - enhanced multi packet write will be activated
> if supported.
> if unsupported the multi packet write feature is disabled.
> 
> Signed-off-by: Tal Shnaiderman 
> ---
>  doc/guides/rel_notes/release_23_07.rst | 33 -
>  drivers/common/mlx5/mlx5_devx_cmds.c   |  6 ++
>  drivers/common/mlx5/mlx5_devx_cmds.h   |  2 ++
>  drivers/net/mlx5/windows/mlx5_os.c |  8 +++-
>  4 files changed, 19 insertions(+), 30 deletions(-)
> 
> diff --git a/doc/guides/rel_notes/release_23_07.rst
> b/doc/guides/rel_notes/release_23_07.rst
> index a9b1293689..d74551414d 100644
> --- a/doc/guides/rel_notes/release_23_07.rst
> +++ b/doc/guides/rel_notes/release_23_07.rst
> @@ -24,36 +24,11 @@ DPDK Release 23.07
>  New Features
>  
> 
> -.. This section should contain new features added in this release.
> -   Sample format:
> +* **Updated NVIDIA mlx5 driver.**
> 
> -   * **Add a title in the past tense with a full stop.**
> -
> - Add a short 1-2 sentence description in the past tense.
> - The description should be enough to allow someone scanning
> - the release notes to understand the new feature.
> -
> - If the feature adds a lot of sub-features you can use a bullet list
> - like this:
> -
> - * Added feature foo to do something.
> - * Enhanced feature bar to do something else.
> -
> - Refer to the previous release notes for examples.
> -
> - Suggested order in release notes items:
> - * Core libs (EAL, mempool, ring, mbuf, buses)
> - * Device abstraction libs and PMDs (ordered alphabetically by vendor
> name)
> -   - ethdev (lib, PMDs)
> -   - cryptodev (lib, PMDs)
> -   - eventdev (lib, PMDs)
> -   - etc
> - * Other libs
> - * Apps, Examples, Tools (if significant)
> -
> - This section is a comment. Do not overwrite or remove it.
> - Also, make sure to start the actual text at the margin.
> - ===
> +  * Added support for multi-packet RQ on Windows.
> +  * Added support for CQE compression on Windows.
> +  * Added support for enhanced multi-packet write on Windows.
> 
> 
>  Removed Items
> diff --git a/drivers/common/mlx5/mlx5_devx_cmds.c
> b/drivers/common/mlx5/mlx5_devx_cmds.c
> index a31e4995f5..b2abc742cf 100644
> --- a/drivers/common/mlx5/mlx5_devx_cmds.c
> +++ b/drivers/common/mlx5/mlx5_devx_cmds.c
> @@ -1298,6 +1298,12 @@ mlx5_devx_cmd_query_hca_attr(void *ctx,
>   attr->rss_ind_tbl_cap = MLX5_GET
> 
>   (per_protocol_networking_offload_caps,
>hcattr, rss_ind_tbl_cap);
> + attr->multi_pkt_send_wqe = MLX5_GET
> +
>   (per_protocol_networking_offload_caps,
> +  hcattr, multi_pkt_send_wqe);
> + attr->enhanced_multi_pkt_send_wqe = MLX5_GET
> +
>   (per_protocol_networking_offload_caps,
> +  hcattr,
> enhanced_multi_pkt_send_wqe);
>   /* Query HCA attribute for ROCE. */
>   if (attr->roce) {
>   hcattr = mlx5_devx_get_hca_cap(ctx, in, out, &rc, diff --git
> a/drivers/common/mlx5/mlx5_devx_cmds.h
> b/drivers/common/mlx5/mlx5_devx_cmds.h
> index edcd867c4e..c8427d2dbb 100644
> --- a/drivers/common/mlx5/mlx5_devx_cmds.h
> +++ b/drivers/common/mlx5/mlx5_devx_cmds.h
> @@ -285,6 +285,8 @@ struct mlx5_hca_attr {
>   uint32_t striding_rq:1;
>   uint32_t ext_stride_num_range:1;
>   uint32_t cqe_compression_128:1;
> + uint32_t multi_pkt_send_wqe:1;
> + uint32_t enhanced_multi_pkt_send_wqe:1;
>   uint32_t set_reg_c:8;
>   uint32_t nic_flow_table:1;
>   uint32_t modify_outer_ip_ecn:1;
> diff --git a/drivers/net/mlx5/windows/mlx5_os.c
> b/drivers/net/mlx5/windows/mlx5_os.c
> index 6527269663..b731bdff06 100644
> --- a/drivers/net/mlx5/windows/mlx5_os.c
> +++ b/drivers/net/mlx5/windows/mlx5_os.c
> @@ -173,7 +173,6 @@ mlx5_os_capabilities_prepare(struct
> mlx5_dev_ctx_shared *sh)
>   sh->dev_cap.max_qp = 1 << hca_attr->log_max_qp;
>   sh->dev_cap.max_qp_wr = 1 << hca_attr->log_max_qp_sz;
>   sh->dev_cap.dv_flow_en = 1;
> - sh->dev_cap.mps = MLX5_MPW_DISABLED;
>   DRV_LOG(DEBUG, "MPW isn't supported.");
>   DRV_LOG(DEBUG, "MPLS over GRE/UDP tunnel offloading 

RE: [PATCH] test/crypto: fix return value from session

2023-05-08 Thread Power, Ciara



> -Original Message-
> From: Tejasree Kondoj 
> Sent: Friday 5 May 2023 17:54
> To: Akhil Goyal ; Fan Zhang
> 
> Cc: Anoob Joseph ; Power, Ciara
> ; Jerin Jacob ; Aakash
> Sasidharan ; Gowrishankar Muthukrishnan
> ; Vidya Sagar Velumuri
> ; dev@dpdk.org; sta...@dpdk.org
> Subject: [PATCH] test/crypto: fix return value from session
> 
> Fixing return value if session failure occurs due to unsupported feature.
> 
> Fixes: eec136f3c54f ("aesni_gcm: add driver for AES-GCM crypto
> operations")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Tejasree Kondoj 
> ---
>  app/test/test_cryptodev.c | 4 ++--
>  1 file changed, 2 insertions(+), 2 deletions(-)
> 

Acked-by: Ciara Power 


[PATCH] net/iavf: fix large vf queues cannot re-use.

2023-05-08 Thread Shiyang He
Multiple vfs created based on one PF. If one of the vf requests a large
number of queues, the queue resources are not released when exiting the
program, resulting in these queues cannot be available to other vfs.

This commit fixes the issue by resetting the number of queues to default
when closing the dev (refer to the kernel driver) .

Fixes: e436cd43835b ("net/iavf: negotiate large VF and request more queues")
Cc: sta...@dpdk.org

Signed-off-by: Shiyang He 
---
 drivers/net/iavf/iavf_ethdev.c | 12 
 1 file changed, 12 insertions(+)

diff --git a/drivers/net/iavf/iavf_ethdev.c b/drivers/net/iavf/iavf_ethdev.c
index f6d68403ce..9c55eaa66e 100644
--- a/drivers/net/iavf/iavf_ethdev.c
+++ b/drivers/net/iavf/iavf_ethdev.c
@@ -2735,6 +2735,18 @@ iavf_dev_close(struct rte_eth_dev *dev)
if (vf->promisc_unicast_enabled || vf->promisc_multicast_enabled)
iavf_config_promisc(adapter, false, false);
 
+   /*
+* Release redundant queue resource when close the dev
+* so that other vfs can re-use the queues.
+*/
+   if (vf->lv_enabled) {
+   ret = iavf_request_queues(dev, IAVF_MAX_NUM_QUEUES_DFLT);
+   if (ret)
+   PMD_DRV_LOG(ERR, "Reset the num of queues failed");
+
+   vf->max_rss_qregion = IAVF_MAX_NUM_QUEUES_DFLT;
+   }
+
iavf_shutdown_adminq(hw);
if (vf->vf_res->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_WB_ON_ITR) {
/* disable uio intr before callback unregister */
-- 
2.37.2



RE: [PATCH v3 5/5] config/arm: add AMD CDX

2023-05-08 Thread Gupta, Nipun


> -Original Message-
> From: Yigit, Ferruh 
> Sent: Thursday, May 4, 2023 8:59 PM
> To: Gupta, Nipun ; dev@dpdk.org;
> tho...@monjalon.net; david.march...@redhat.com
> Cc: Anand, Harpreet ; Agarwal, Nikhil
> 
> Subject: Re: [PATCH v3 5/5] config/arm: add AMD CDX
> 
> On 4/21/2023 3:54 PM, Nipun Gupta wrote:
> > Adding support for AMD CDX devices
> >
> > Signed-off-by: Nipun Gupta 
> > ---
> >  config/arm/arm64_cdx_linux_gcc | 17 +
> >  config/arm/meson.build | 14 ++
> >  2 files changed, 31 insertions(+)
> >  create mode 100644 config/arm/arm64_cdx_linux_gcc
> >
> > diff --git a/config/arm/arm64_cdx_linux_gcc
> b/config/arm/arm64_cdx_linux_gcc
> > new file mode 100644
> > index 00..8e6d619dae
> > --- /dev/null
> > +++ b/config/arm/arm64_cdx_linux_gcc
> > @@ -0,0 +1,17 @@
> > +[binaries]
> > +c = ['ccache', 'aarch64-linux-gnu-gcc']
> > +cpp = ['ccache', 'aarch64-linux-gnu-g++']
> > +ar = 'aarch64-linux-gnu-ar'
> > +as = 'aarch64-linux-gnu-as'
> > +strip = 'aarch64-linux-gnu-strip'
> > +pkgconfig = 'aarch64-linux-gnu-pkg-config'
> > +pcap-config = ''
> > +
> > +[host_machine]
> > +system = 'linux'
> > +cpu_family = 'aarch64'
> > +cpu = 'armv8-a'
> > +endian = 'little'
> > +
> > +[properties]
> > +platform = 'cdx'
> > diff --git a/config/arm/meson.build b/config/arm/meson.build
> > index 5213434ca4..39b8929534 100644
> > --- a/config/arm/meson.build
> > +++ b/config/arm/meson.build
> > @@ -305,6 +305,18 @@ soc_bluefield = {
> >  'numa': false
> >  }
> >
> > +soc_cdx = {
> > +'description': 'AMD CDX',
> > +'implementer': '0x41',
> > +'part_number': '0xd42',
> > +'flags': [
> > +['RTE_MACHINE', '"cdx"'],
> > +['RTE_MAX_LCORE', 16],
> > +['RTE_MAX_NUMA_NODES', 1]
> > +],
> > +'numa': false
> > +}
> 
> Hi Nipun,
> 
> Why we need a new arm platform/config? Is it because of above flags?
> If it can work with default values, I think we can drop this patch.

Hi Ferruh,

CDX driver works with generic ARM compilation too (arm64_armv8_linux_gcc).

The versal platforms supporting CDX have A78 cores, and adding this cdx config
Helps to provide gcc option "march= armv8.4-a" which is for implementer
"0xd42" (ARM cortex A78 cores)., whereas for generic ARM compilation
"march= armv8-a".

Maybe ARM guys can provide more information regarding if there is any impact
on using generic architecture flag (i.e. march= armv8a) on A78 cores.

Thanks,
Nipun


Re: [PATCH v3 5/5] config/arm: add AMD CDX

2023-05-08 Thread Thomas Monjalon
08/05/2023 12:24, Gupta, Nipun:
> From: Yigit, Ferruh 
> > On 4/21/2023 3:54 PM, Nipun Gupta wrote:
> > > Adding support for AMD CDX devices
> > >
> > > Signed-off-by: Nipun Gupta 
> > > ---
> > >  config/arm/arm64_cdx_linux_gcc | 17 +
> > >  config/arm/meson.build | 14 ++
> > >  2 files changed, 31 insertions(+)
> > >  create mode 100644 config/arm/arm64_cdx_linux_gcc
> > >
> > > diff --git a/config/arm/arm64_cdx_linux_gcc
> > b/config/arm/arm64_cdx_linux_gcc
> > > new file mode 100644
> > > index 00..8e6d619dae
> > > --- /dev/null
> > > +++ b/config/arm/arm64_cdx_linux_gcc
> > > @@ -0,0 +1,17 @@
> > > +[binaries]
> > > +c = ['ccache', 'aarch64-linux-gnu-gcc']
> > > +cpp = ['ccache', 'aarch64-linux-gnu-g++']
> > > +ar = 'aarch64-linux-gnu-ar'
> > > +as = 'aarch64-linux-gnu-as'
> > > +strip = 'aarch64-linux-gnu-strip'
> > > +pkgconfig = 'aarch64-linux-gnu-pkg-config'
> > > +pcap-config = ''
> > > +
> > > +[host_machine]
> > > +system = 'linux'
> > > +cpu_family = 'aarch64'
> > > +cpu = 'armv8-a'
> > > +endian = 'little'
> > > +
> > > +[properties]
> > > +platform = 'cdx'
> > > diff --git a/config/arm/meson.build b/config/arm/meson.build
> > > index 5213434ca4..39b8929534 100644
> > > --- a/config/arm/meson.build
> > > +++ b/config/arm/meson.build
> > > @@ -305,6 +305,18 @@ soc_bluefield = {
> > >  'numa': false
> > >  }
> > >
> > > +soc_cdx = {
> > > +'description': 'AMD CDX',
> > > +'implementer': '0x41',
> > > +'part_number': '0xd42',
> > > +'flags': [
> > > +['RTE_MACHINE', '"cdx"'],
> > > +['RTE_MAX_LCORE', 16],
> > > +['RTE_MAX_NUMA_NODES', 1]
> > > +],
> > > +'numa': false
> > > +}
> > 
> > Hi Nipun,
> > 
> > Why we need a new arm platform/config? Is it because of above flags?
> > If it can work with default values, I think we can drop this patch.
> 
> Hi Ferruh,
> 
> CDX driver works with generic ARM compilation too (arm64_armv8_linux_gcc).
> 
> The versal platforms supporting CDX have A78 cores, and adding this cdx config
> Helps to provide gcc option "march= armv8.4-a" which is for implementer
> "0xd42" (ARM cortex A78 cores)., whereas for generic ARM compilation
> "march= armv8-a".
> 
> Maybe ARM guys can provide more information regarding if there is any impact
> on using generic architecture flag (i.e. march= armv8a) on A78 cores.

Please let's skip this patch for now.
You should send it separately later, with perf testing and HW description in 
the commit log.





RE: [PATCH v3 5/5] config/arm: add AMD CDX

2023-05-08 Thread Gupta, Nipun



> -Original Message-
> From: Thomas Monjalon 
> Sent: Monday, May 8, 2023 4:14 PM
> To: Yigit, Ferruh ; Gupta, Nipun
> 
> Cc: dev@dpdk.org; david.march...@redhat.com; Anand, Harpreet
> ; Agarwal, Nikhil 
> Subject: Re: [PATCH v3 5/5] config/arm: add AMD CDX
> 
> 
> 08/05/2023 12:24, Gupta, Nipun:
> > From: Yigit, Ferruh 
> > > On 4/21/2023 3:54 PM, Nipun Gupta wrote:
> > > > Adding support for AMD CDX devices
> > > >
> > > > Signed-off-by: Nipun Gupta 
> > > > ---
> > > >  config/arm/arm64_cdx_linux_gcc | 17 +
> > > >  config/arm/meson.build | 14 ++
> > > >  2 files changed, 31 insertions(+)
> > > >  create mode 100644 config/arm/arm64_cdx_linux_gcc
> > > >
> > > > diff --git a/config/arm/arm64_cdx_linux_gcc
> > > b/config/arm/arm64_cdx_linux_gcc
> > > > new file mode 100644
> > > > index 00..8e6d619dae
> > > > --- /dev/null
> > > > +++ b/config/arm/arm64_cdx_linux_gcc
> > > > @@ -0,0 +1,17 @@
> > > > +[binaries]
> > > > +c = ['ccache', 'aarch64-linux-gnu-gcc']
> > > > +cpp = ['ccache', 'aarch64-linux-gnu-g++']
> > > > +ar = 'aarch64-linux-gnu-ar'
> > > > +as = 'aarch64-linux-gnu-as'
> > > > +strip = 'aarch64-linux-gnu-strip'
> > > > +pkgconfig = 'aarch64-linux-gnu-pkg-config'
> > > > +pcap-config = ''
> > > > +
> > > > +[host_machine]
> > > > +system = 'linux'
> > > > +cpu_family = 'aarch64'
> > > > +cpu = 'armv8-a'
> > > > +endian = 'little'
> > > > +
> > > > +[properties]
> > > > +platform = 'cdx'
> > > > diff --git a/config/arm/meson.build b/config/arm/meson.build
> > > > index 5213434ca4..39b8929534 100644
> > > > --- a/config/arm/meson.build
> > > > +++ b/config/arm/meson.build
> > > > @@ -305,6 +305,18 @@ soc_bluefield = {
> > > >  'numa': false
> > > >  }
> > > >
> > > > +soc_cdx = {
> > > > +'description': 'AMD CDX',
> > > > +'implementer': '0x41',
> > > > +'part_number': '0xd42',
> > > > +'flags': [
> > > > +['RTE_MACHINE', '"cdx"'],
> > > > +['RTE_MAX_LCORE', 16],
> > > > +['RTE_MAX_NUMA_NODES', 1]
> > > > +],
> > > > +'numa': false
> > > > +}
> > >
> > > Hi Nipun,
> > >
> > > Why we need a new arm platform/config? Is it because of above flags?
> > > If it can work with default values, I think we can drop this patch.
> >
> > Hi Ferruh,
> >
> > CDX driver works with generic ARM compilation too
> (arm64_armv8_linux_gcc).
> >
> > The versal platforms supporting CDX have A78 cores, and adding this cdx
> config
> > Helps to provide gcc option "march= armv8.4-a" which is for implementer
> > "0xd42" (ARM cortex A78 cores)., whereas for generic ARM compilation
> > "march= armv8-a".
> >
> > Maybe ARM guys can provide more information regarding if there is any
> impact
> > on using generic architecture flag (i.e. march= armv8a) on A78 cores.
> 
> Please let's skip this patch for now.
> You should send it separately later, with perf testing and HW description in 
> the
> commit log.

Sure, I will skip this and send another spin for CDX bus without this change.

Thanks,
Nipun

> 
> 



[PATCH v4 0/4] Support AMD CDX bus, for FPGA based CDX devices. The CDX

2023-05-08 Thread Nipun Gupta
It uses sysfs interface and the vfio-cdx driver to discover
and initialize the CDX devices.

The CDX bus and VFIO support is available at Xilinx open source tree:
https://github.com/Xilinx/linux-xlnx (drivers/cdx/ and drivers/vfio/cdx)

Linux CDX bus patches has been added into linux next:
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git/tree/drivers/cdx

VFIO patches are also submitted in upstream:
https://www.spinics.net/lists/kvm/msg310623.html

CDX is a Hardware Architecture designed for AMD FPGA devices. It
consists of mechanism for interaction between FPGA, Firmware and 
the APUs (Application CPUs).
Firmware resides on RPU (Realtime CPUs) which interacts with
the FPGA program manager and the APUs. The RPU provides memory-mapped
interface (RPU if) which is used to communicate with APUs.

VFIO CDX driver provides the CDX device resources like MMIO and interrupts
to map to user-space. DPDK CDX bus uses sysfs interface and the vfio-cdx
driver to discover and initialize the CDX devices for user-space
applications.

Changes v3->v4:
- removed platform specific patch (adding config for ARM CDX)
  from this series

Changes v2->v3:
- merged cdx bus compilation enablement in the first patch
- fixed issue reported by check-git-log.sh
- updated release notes
- updated offset to uint64_t instead of off_t in cdx_map_resource

Changes v1->v2:
- Moved file rte_cdx_bus.h to internal bus_cdx_driver.h
  and added this file to deivce_cdx_headers
- Moved cdx.h to private.h
- Removed rte_ prefix from the static symbols in .c files.

Changes RFC->v1:
- Marked few API's as internal which were not required
  to be provided to user.

Nipun Gupta (4):
  bus/cdx: introduce cdx bus
  bus/cdx: add DMA map and unmap support
  bus/cdx: add support for MSI
  bus/cdx: support plug unplug and dev iterator

 MAINTAINERS|   5 +
 doc/guides/rel_notes/release_23_07.rst |   6 +
 drivers/bus/cdx/bus_cdx_driver.h   | 227 
 drivers/bus/cdx/cdx.c  | 693 +
 drivers/bus/cdx/cdx_logs.h |  37 ++
 drivers/bus/cdx/cdx_vfio.c | 615 ++
 drivers/bus/cdx/meson.build|  13 +
 drivers/bus/cdx/private.h  |  49 ++
 drivers/bus/cdx/version.map|  13 +
 drivers/bus/meson.build|   1 +
 lib/eal/common/eal_common_interrupts.c |  21 +
 lib/eal/common/eal_interrupts.h|   1 +
 lib/eal/include/rte_interrupts.h   |  32 ++
 lib/eal/version.map|   2 +
 14 files changed, 1715 insertions(+)
 create mode 100644 drivers/bus/cdx/bus_cdx_driver.h
 create mode 100644 drivers/bus/cdx/cdx.c
 create mode 100644 drivers/bus/cdx/cdx_logs.h
 create mode 100644 drivers/bus/cdx/cdx_vfio.c
 create mode 100644 drivers/bus/cdx/meson.build
 create mode 100644 drivers/bus/cdx/private.h
 create mode 100644 drivers/bus/cdx/version.map

-- 
2.17.1



[PATCH v4 1/4] bus/cdx: introduce cdx bus

2023-05-08 Thread Nipun Gupta
CDX bus supports multiple type of devices, which can be
exposed to user-space via vfio-cdx.

vfio-cdx provides the MMIO IO_MEMORY regions as well as the
DMA interface for the device (IOMMU).

This support aims to enable the DPDK to support the cdx
devices in user-space using VFIO interface.

Signed-off-by: Nipun Gupta 
---
 MAINTAINERS|   5 +
 doc/guides/rel_notes/release_23_07.rst |   6 +
 drivers/bus/cdx/bus_cdx_driver.h   | 201 ++
 drivers/bus/cdx/cdx.c  | 520 +
 drivers/bus/cdx/cdx_logs.h |  37 ++
 drivers/bus/cdx/cdx_vfio.c | 437 +
 drivers/bus/cdx/meson.build|  13 +
 drivers/bus/cdx/private.h  |  49 +++
 drivers/bus/cdx/version.map|  11 +
 drivers/bus/meson.build|   1 +
 10 files changed, 1280 insertions(+)
 create mode 100644 drivers/bus/cdx/bus_cdx_driver.h
 create mode 100644 drivers/bus/cdx/cdx.c
 create mode 100644 drivers/bus/cdx/cdx_logs.h
 create mode 100644 drivers/bus/cdx/cdx_vfio.c
 create mode 100644 drivers/bus/cdx/meson.build
 create mode 100644 drivers/bus/cdx/private.h
 create mode 100644 drivers/bus/cdx/version.map

diff --git a/MAINTAINERS b/MAINTAINERS
index 8df23e5099..1f9b6af9b9 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -569,6 +569,11 @@ M: Parav Pandit 
 M: Xueming Li 
 F: drivers/bus/auxiliary/
 
+CDX bus driver
+M: Nipun Gupta 
+M: Nikhil Agarwal 
+F: drivers/bus/cdx/
+
 Intel FPGA bus
 M: Rosen Xu 
 F: drivers/bus/ifpga/
diff --git a/doc/guides/rel_notes/release_23_07.rst 
b/doc/guides/rel_notes/release_23_07.rst
index a9b1293689..a5f581dc3d 100644
--- a/doc/guides/rel_notes/release_23_07.rst
+++ b/doc/guides/rel_notes/release_23_07.rst
@@ -55,6 +55,12 @@ New Features
  Also, make sure to start the actual text at the margin.
  ===
 
+* **Added CDX bus support.**
+
+  CDX bus driver has been added to support AMD CDX bus, which operates
+  on FPGA based CDX devices. The CDX devices are memory mapped on system
+  bus for embedded CPUs.
+
 
 Removed Items
 -
diff --git a/drivers/bus/cdx/bus_cdx_driver.h b/drivers/bus/cdx/bus_cdx_driver.h
new file mode 100644
index 00..7edcb019eb
--- /dev/null
+++ b/drivers/bus/cdx/bus_cdx_driver.h
@@ -0,0 +1,201 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright (C) 2022-2023, Advanced Micro Devices, Inc.
+ */
+
+#ifndef _BUS_CDX_DRIVER_H_
+#define _BUS_CDX_DRIVER_H_
+
+/**
+ * @file
+ *
+ * CDX bus interface
+ */
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+
+#include 
+#include 
+#include 
+#include 
+#include 
+#include 
+
+#include 
+#include 
+#include 
+#include 
+#include 
+#include 
+
+/* Forward declarations */
+struct rte_cdx_device;
+struct rte_cdx_driver;
+
+#define CDX_MAX_RESOURCE 4
+
+/** List of CDX devices */
+RTE_TAILQ_HEAD(rte_cdx_device_list, rte_cdx_device);
+/** List of CDX drivers */
+RTE_TAILQ_HEAD(rte_cdx_driver_list, rte_cdx_driver);
+
+/* CDX Bus iterators */
+#define FOREACH_DEVICE_ON_CDXBUS(p)\
+   RTE_TAILQ_FOREACH(p, &rte_cdx_bus.device_list, next)
+
+#define FOREACH_DRIVER_ON_CDXBUS(p)\
+   RTE_TAILQ_FOREACH(p, &rte_cdx_bus.driver_list, next)
+
+/** Any CDX device identifier (vendor, device) */
+#define RTE_CDX_ANY_ID (0x)
+
+#define RTE_PMD_REGISTER_CDX_TABLE(name, table) \
+static const char DRV_EXP_TAG(name, cdx_tbl_export)[] __rte_used = \
+RTE_STR(table)
+
+/**
+ * A structure describing an ID for a CDX driver. Each driver provides a
+ * table of these IDs for each device that it supports.
+ */
+struct rte_cdx_id {
+   uint16_t vendor_id; /**< Vendor ID. */
+   uint16_t device_id; /**< Device ID. */
+};
+
+/**
+ * A structure describing a CDX device.
+ */
+struct rte_cdx_device {
+   RTE_TAILQ_ENTRY(rte_cdx_device) next;   /**< Next probed CDX device. */
+   struct rte_device device;   /**< Inherit core device */
+   struct rte_cdx_id id;   /**< CDX ID. */
+   struct rte_mem_resource mem_resource[CDX_MAX_RESOURCE];
+   /**< CDX Memory Resource */
+};
+
+/**
+ * @internal
+ * Helper macro for drivers that need to convert to struct rte_cdx_device.
+ */
+#define RTE_DEV_TO_CDX_DEV(ptr) \
+   container_of(ptr, struct rte_cdx_device, device)
+
+#define RTE_DEV_TO_CDX_DEV_CONST(ptr) \
+   container_of(ptr, const struct rte_cdx_device, device)
+
+#define RTE_ETH_DEV_TO_CDX_DEV(eth_dev)
RTE_DEV_TO_CDX_DEV((eth_dev)->device)
+
+#ifdef __cplusplus
+/** C++ macro used to help building up tables of device IDs */
+#define RTE_CDX_DEVICE(vend, dev)  \
+   (vend), \
+   (dev)
+#else
+/** Macro used to help building up tables of device IDs */
+#define RTE_CDX_DEVICE(vend, dev)  \
+   .vendor_id = (vend),\
+   .device_id

[PATCH v4 2/4] bus/cdx: add DMA map and unmap support

2023-05-08 Thread Nipun Gupta
CDX bus can use VFIO interface for mapping and unmapping
of DMA addresses in the IOMMU. This change adds the callback
support for map and unmap APIs as well as fetching the IOMMU
class.

Signed-off-by: Nipun Gupta 
---
 drivers/bus/cdx/cdx.c | 40 
 1 file changed, 40 insertions(+)

diff --git a/drivers/bus/cdx/cdx.c b/drivers/bus/cdx/cdx.c
index d479daa315..8cc273336e 100644
--- a/drivers/bus/cdx/cdx.c
+++ b/drivers/bus/cdx/cdx.c
@@ -505,12 +505,52 @@ cdx_find_device(const struct rte_device *start, 
rte_dev_cmp_t cmp,
return NULL;
 }
 
+static int
+cdx_dma_map(struct rte_device *dev, void *addr, uint64_t iova, size_t len)
+{
+   struct rte_cdx_device *cdx_dev = RTE_DEV_TO_CDX_DEV(dev);
+
+   if (!cdx_dev) {
+   rte_errno = EINVAL;
+   return -1;
+   }
+
+   return rte_vfio_container_dma_map(RTE_VFIO_DEFAULT_CONTAINER_FD,
+ (uintptr_t)addr, iova, len);
+}
+
+static int
+cdx_dma_unmap(struct rte_device *dev, void *addr, uint64_t iova, size_t len)
+{
+   struct rte_cdx_device *cdx_dev = RTE_DEV_TO_CDX_DEV(dev);
+
+   if (!cdx_dev) {
+   rte_errno = EINVAL;
+   return -1;
+   }
+
+   return rte_vfio_container_dma_unmap(RTE_VFIO_DEFAULT_CONTAINER_FD,
+   (uintptr_t)addr, iova, len);
+}
+
+static enum rte_iova_mode
+cdx_get_iommu_class(void)
+{
+   if (TAILQ_EMPTY(&rte_cdx_bus.device_list))
+   return RTE_IOVA_DC;
+
+   return RTE_IOVA_VA;
+}
+
 struct rte_cdx_bus rte_cdx_bus = {
.bus = {
.scan = cdx_scan,
.probe = cdx_probe,
.find_device = cdx_find_device,
.parse = cdx_parse,
+   .dma_map = cdx_dma_map,
+   .dma_unmap = cdx_dma_unmap,
+   .get_iommu_class = cdx_get_iommu_class,
},
.device_list = TAILQ_HEAD_INITIALIZER(rte_cdx_bus.device_list),
.driver_list = TAILQ_HEAD_INITIALIZER(rte_cdx_bus.driver_list),
-- 
2.17.1



[PATCH v4 4/4] bus/cdx: support plug unplug and dev iterator

2023-05-08 Thread Nipun Gupta
This change adds support for plugging and unplugging
CDX devices on the CDX bus. Also, CDX dev iterator support
has been added for the CDX bus.

Signed-off-by: Nipun Gupta 
---
 drivers/bus/cdx/bus_cdx_driver.h |   1 +
 drivers/bus/cdx/cdx.c| 122 +++
 2 files changed, 123 insertions(+)

diff --git a/drivers/bus/cdx/bus_cdx_driver.h b/drivers/bus/cdx/bus_cdx_driver.h
index fdeaf46664..3d89e7c054 100644
--- a/drivers/bus/cdx/bus_cdx_driver.h
+++ b/drivers/bus/cdx/bus_cdx_driver.h
@@ -69,6 +69,7 @@ struct rte_cdx_id {
 struct rte_cdx_device {
RTE_TAILQ_ENTRY(rte_cdx_device) next;   /**< Next probed CDX device. */
struct rte_device device;   /**< Inherit core device */
+   struct rte_cdx_driver *driver;  /**< CDX driver used in probing 
*/
struct rte_cdx_id id;   /**< CDX ID. */
struct rte_mem_resource mem_resource[CDX_MAX_RESOURCE];
/**< CDX Memory Resource */
diff --git a/drivers/bus/cdx/cdx.c b/drivers/bus/cdx/cdx.c
index 6c9ceaaf7f..0a30b8648a 100644
--- a/drivers/bus/cdx/cdx.c
+++ b/drivers/bus/cdx/cdx.c
@@ -68,6 +68,7 @@
 #include 
 #include 
 #include 
+#include 
 #include 
 #include 
 #include 
@@ -82,6 +83,15 @@
 #define CDX_BUS_NAME   cdx
 #define CDX_DEV_PREFIX "cdx-"
 
+enum cdx_params {
+   RTE_CDX_PARAM_NAME,
+};
+
+static const char * const cdx_params_keys[] = {
+   [RTE_CDX_PARAM_NAME] = "name",
+   NULL,
+};
+
 /**
  * @file
  * CDX probing using Linux sysfs.
@@ -399,6 +409,7 @@ cdx_probe_one_driver(struct rte_cdx_driver *dr,
} else {
dev->device.driver = &dr->driver;
}
+   dev->driver = dr;
 
return ret;
 
@@ -516,6 +527,71 @@ cdx_find_device(const struct rte_device *start, 
rte_dev_cmp_t cmp,
return NULL;
 }
 
+/* Remove a device from CDX bus */
+static void
+cdx_remove_device(struct rte_cdx_device *cdx_dev)
+{
+   TAILQ_REMOVE(&rte_cdx_bus.device_list, cdx_dev, next);
+}
+
+/*
+ * If vendor/device ID match, call the remove() function of the
+ * driver.
+ */
+static int
+cdx_detach_dev(struct rte_cdx_device *dev)
+{
+   struct rte_cdx_driver *dr;
+   int ret = 0;
+
+   if (dev == NULL)
+   return -EINVAL;
+
+   dr = dev->driver;
+
+   CDX_BUS_DEBUG("detach device %s using driver: %s",
+   dev->device.name, dr->driver.name);
+
+   if (dr->remove) {
+   ret = dr->remove(dev);
+   if (ret < 0)
+   return ret;
+   }
+
+   /* clear driver structure */
+   dev->driver = NULL;
+   dev->device.driver = NULL;
+
+   rte_cdx_unmap_device(dev);
+
+   rte_intr_instance_free(dev->intr_handle);
+   dev->intr_handle = NULL;
+
+   return 0;
+}
+
+static int
+cdx_plug(struct rte_device *dev)
+{
+   return cdx_probe_all_drivers(RTE_DEV_TO_CDX_DEV(dev));
+}
+
+static int
+cdx_unplug(struct rte_device *dev)
+{
+   struct rte_cdx_device *cdx_dev;
+   int ret;
+
+   cdx_dev = RTE_DEV_TO_CDX_DEV(dev);
+   ret = cdx_detach_dev(cdx_dev);
+   if (ret == 0) {
+   cdx_remove_device(cdx_dev);
+   rte_devargs_remove(dev->devargs);
+   free(cdx_dev);
+   }
+   return ret;
+}
+
 static int
 cdx_dma_map(struct rte_device *dev, void *addr, uint64_t iova, size_t len)
 {
@@ -553,15 +629,61 @@ cdx_get_iommu_class(void)
return RTE_IOVA_VA;
 }
 
+static int
+cdx_dev_match(const struct rte_device *dev,
+   const void *_kvlist)
+{
+   const struct rte_kvargs *kvlist = _kvlist;
+   const char *key = cdx_params_keys[RTE_CDX_PARAM_NAME];
+   const char *name;
+
+   /* no kvlist arg, all devices match */
+   if (kvlist == NULL)
+   return 0;
+
+   /* if key is present in kvlist and does not match, filter device */
+   name = rte_kvargs_get(kvlist, key);
+   if (name != NULL && strcmp(name, dev->name))
+   return -1;
+
+   return 0;
+}
+
+static void *
+cdx_dev_iterate(const void *start,
+   const char *str,
+   const struct rte_dev_iterator *it __rte_unused)
+{
+   rte_bus_find_device_t find_device;
+   struct rte_kvargs *kvargs = NULL;
+   struct rte_device *dev;
+
+   if (str != NULL) {
+   kvargs = rte_kvargs_parse(str, cdx_params_keys);
+   if (kvargs == NULL) {
+   CDX_BUS_ERR("cannot parse argument list %s", str);
+   rte_errno = EINVAL;
+   return NULL;
+   }
+   }
+   find_device = rte_cdx_bus.bus.find_device;
+   dev = find_device(start, cdx_dev_match, kvargs);
+   rte_kvargs_free(kvargs);
+   return dev;
+}
+
 struct rte_cdx_bus rte_cdx_bus = {
.bus = {
.scan = cdx_scan,
.probe = cdx_probe,
.find_device = cdx_find_device,
+ 

[PATCH v4 3/4] bus/cdx: add support for MSI

2023-05-08 Thread Nipun Gupta
MSI's are exposed to the devices using VFIO (vfio-cdx). This
patch uses the same to add support for MSI for the devices on
the cdx bus.

A couple of API's have been introduced in the EAL interrupt
framework:
- rte_intr_irq_count_set: This API is used to set the total
interrupts on the interrupt handle. This would be provided
by VFIO (irq.count) for VFIO enabled devices.
- rte_intr_irq_count_get: This API returns the total number
interrupts which were set.

Signed-off-by: Nipun Gupta 
---
 drivers/bus/cdx/bus_cdx_driver.h   |  25 
 drivers/bus/cdx/cdx.c  |  11 ++
 drivers/bus/cdx/cdx_vfio.c | 182 -
 drivers/bus/cdx/version.map|   2 +
 lib/eal/common/eal_common_interrupts.c |  21 +++
 lib/eal/common/eal_interrupts.h|   1 +
 lib/eal/include/rte_interrupts.h   |  32 +
 lib/eal/version.map|   2 +
 8 files changed, 274 insertions(+), 2 deletions(-)

diff --git a/drivers/bus/cdx/bus_cdx_driver.h b/drivers/bus/cdx/bus_cdx_driver.h
index 7edcb019eb..fdeaf46664 100644
--- a/drivers/bus/cdx/bus_cdx_driver.h
+++ b/drivers/bus/cdx/bus_cdx_driver.h
@@ -72,6 +72,7 @@ struct rte_cdx_device {
struct rte_cdx_id id;   /**< CDX ID. */
struct rte_mem_resource mem_resource[CDX_MAX_RESOURCE];
/**< CDX Memory Resource */
+   struct rte_intr_handle *intr_handle;/**< Interrupt handle */
 };
 
 /**
@@ -173,6 +174,30 @@ void rte_cdx_unmap_device(struct rte_cdx_device *dev);
 __rte_internal
 void rte_cdx_register(struct rte_cdx_driver *driver);
 
+/**
+ * Enables VFIO Interrupts for CDX bus devices.
+ *
+ * @param intr_handle
+ *   Pointer to the interrupt handle.
+ *
+ *  @return
+ *  0 on success, -1 on error.
+ */
+__rte_internal
+int rte_cdx_vfio_intr_enable(const struct rte_intr_handle *intr_handle);
+
+/**
+ * Disable VFIO Interrupts for CDX bus devices.
+ *
+ * @param intr_handle
+ *   Pointer to the interrupt handle.
+ *
+ *  @return
+ *  0 on success, -1 on error.
+ */
+__rte_internal
+int rte_cdx_vfio_intr_disable(const struct rte_intr_handle *intr_handle);
+
 /**
  * Helper for CDX device registration from driver (eth, crypto, raw) instance
  */
diff --git a/drivers/bus/cdx/cdx.c b/drivers/bus/cdx/cdx.c
index 8cc273336e..6c9ceaaf7f 100644
--- a/drivers/bus/cdx/cdx.c
+++ b/drivers/bus/cdx/cdx.c
@@ -204,6 +204,15 @@ cdx_scan_one(const char *dirname, const char *dev_name)
goto err;
}
 
+   /* Allocate interrupt instance for cdx device */
+   dev->intr_handle =
+   rte_intr_instance_alloc(RTE_INTR_INSTANCE_F_PRIVATE);
+   if (dev->intr_handle == NULL) {
+   CDX_BUS_ERR("Failed to create interrupt instance for %s\n",
+   dev->device.name);
+   return -ENOMEM;
+   }
+
/*
 * Check if device is bound to 'vfio-cdx' driver, so that user-space
 * can gracefully access the device.
@@ -394,6 +403,8 @@ cdx_probe_one_driver(struct rte_cdx_driver *dr,
return ret;
 
 error_probe:
+   rte_intr_instance_free(dev->intr_handle);
+   dev->intr_handle = NULL;
cdx_vfio_unmap_resource(dev);
 error_map_device:
return ret;
diff --git a/drivers/bus/cdx/cdx_vfio.c b/drivers/bus/cdx/cdx_vfio.c
index ae11f589b3..1422b98503 100644
--- a/drivers/bus/cdx/cdx_vfio.c
+++ b/drivers/bus/cdx/cdx_vfio.c
@@ -60,6 +60,10 @@ struct mapped_cdx_resource {
 /** mapped cdx device list */
 TAILQ_HEAD(mapped_cdx_res_list, mapped_cdx_resource);
 
+/* irq set buffer length for MSI interrupts */
+#define MSI_IRQ_SET_BUF_LEN (sizeof(struct vfio_irq_set) + \
+ sizeof(int) * (RTE_MAX_RXTX_INTR_VEC_ID + 1))
+
 static struct rte_tailq_elem cdx_vfio_tailq = {
.name = "VFIO_CDX_RESOURCE_LIST",
 };
@@ -104,6 +108,27 @@ cdx_vfio_unmap_resource_primary(struct rte_cdx_device *dev)
char cdx_addr[PATH_MAX] = {0};
struct mapped_cdx_resource *vfio_res = NULL;
struct mapped_cdx_res_list *vfio_res_list;
+   int ret, vfio_dev_fd;
+
+   if (rte_intr_fd_get(dev->intr_handle) < 0)
+   return -1;
+
+   if (close(rte_intr_fd_get(dev->intr_handle)) < 0) {
+   CDX_BUS_ERR("Error when closing eventfd file descriptor for %s",
+   dev->device.name);
+   return -1;
+   }
+
+   vfio_dev_fd = rte_intr_dev_fd_get(dev->intr_handle);
+   if (vfio_dev_fd < 0)
+   return -1;
+
+   ret = rte_vfio_release_device(rte_cdx_get_sysfs_path(), 
dev->device.name,
+ vfio_dev_fd);
+   if (ret < 0) {
+   CDX_BUS_ERR("Cannot release VFIO device");
+   return ret;
+   }
 
vfio_res_list =
RTE_TAILQ_CAST(cdx_vfio_tailq.head, mapped_cdx_res_list);
@@ -126,6 +151,18 @@ cdx_vfio_unmap_resource_secondary(struct rte_cdx_device 
*dev)

Re: [PATCH v3 5/5] config/arm: add AMD CDX

2023-05-08 Thread Ferruh Yigit
On 5/8/2023 11:24 AM, Gupta, Nipun wrote:
> 
> 
>> -Original Message-
>> From: Yigit, Ferruh 
>> Sent: Thursday, May 4, 2023 8:59 PM
>> To: Gupta, Nipun ; dev@dpdk.org;
>> tho...@monjalon.net; david.march...@redhat.com
>> Cc: Anand, Harpreet ; Agarwal, Nikhil
>> 
>> Subject: Re: [PATCH v3 5/5] config/arm: add AMD CDX
>>
>> On 4/21/2023 3:54 PM, Nipun Gupta wrote:
>>> Adding support for AMD CDX devices
>>>
>>> Signed-off-by: Nipun Gupta 
>>> ---
>>>  config/arm/arm64_cdx_linux_gcc | 17 +
>>>  config/arm/meson.build | 14 ++
>>>  2 files changed, 31 insertions(+)
>>>  create mode 100644 config/arm/arm64_cdx_linux_gcc
>>>
>>> diff --git a/config/arm/arm64_cdx_linux_gcc
>> b/config/arm/arm64_cdx_linux_gcc
>>> new file mode 100644
>>> index 00..8e6d619dae
>>> --- /dev/null
>>> +++ b/config/arm/arm64_cdx_linux_gcc
>>> @@ -0,0 +1,17 @@
>>> +[binaries]
>>> +c = ['ccache', 'aarch64-linux-gnu-gcc']
>>> +cpp = ['ccache', 'aarch64-linux-gnu-g++']
>>> +ar = 'aarch64-linux-gnu-ar'
>>> +as = 'aarch64-linux-gnu-as'
>>> +strip = 'aarch64-linux-gnu-strip'
>>> +pkgconfig = 'aarch64-linux-gnu-pkg-config'
>>> +pcap-config = ''
>>> +
>>> +[host_machine]
>>> +system = 'linux'
>>> +cpu_family = 'aarch64'
>>> +cpu = 'armv8-a'
>>> +endian = 'little'
>>> +
>>> +[properties]
>>> +platform = 'cdx'
>>> diff --git a/config/arm/meson.build b/config/arm/meson.build
>>> index 5213434ca4..39b8929534 100644
>>> --- a/config/arm/meson.build
>>> +++ b/config/arm/meson.build
>>> @@ -305,6 +305,18 @@ soc_bluefield = {
>>>  'numa': false
>>>  }
>>>
>>> +soc_cdx = {
>>> +'description': 'AMD CDX',
>>> +'implementer': '0x41',
>>> +'part_number': '0xd42',
>>> +'flags': [
>>> +['RTE_MACHINE', '"cdx"'],
>>> +['RTE_MAX_LCORE', 16],
>>> +['RTE_MAX_NUMA_NODES', 1]
>>> +],
>>> +'numa': false
>>> +}
>>
>> Hi Nipun,
>>
>> Why we need a new arm platform/config? Is it because of above flags?
>> If it can work with default values, I think we can drop this patch.
> 
> Hi Ferruh,
> 
> CDX driver works with generic ARM compilation too (arm64_armv8_linux_gcc).
> 
> The versal platforms supporting CDX have A78 cores, and adding this cdx config
> Helps to provide gcc option "march= armv8.4-a" which is for implementer
> "0xd42" (ARM cortex A78 cores)., whereas for generic ARM compilation
> "march= armv8-a".
> 
> Maybe ARM guys can provide more information regarding if there is any impact
> on using generic architecture flag (i.e. march= armv8a) on A78 cores.
> 

Hi Honnappa, Ruifeng,

Can you please support on this question, what is the difference of
'march= armv8-a' flag (comparing march= armv8a)?
Should we consider adding an arm config file to support this flag?

Thanks,
ferruh




[RFC PATCH v2 0/3] improve FEC API usage

2023-05-08 Thread Denis Pryazhennikov
The documentation for the FEC API is currently incomplete and contains
inaccuracies in its descriptions of function parameters. 
Specifically, the semantics of the fec_capa parameter in rte_eth_fec_set()
is not well-defined. It does not provide information on what should
be done if only AUTO bit is set or one of the specified FEC modes is 
not supported. Additionally, the fec_capa parameter in rte_eth_fec_get()
implies that more than one FEC mode can be obtained, but it is 
wrong. Furthermore, the behaviour is undefined in 
rte_eth_fec_set() when the fec_capa parameter is zero.

To address these issues, a patch series has been created that updates
the FEC API documentation, renames one of the parameters to improve 
its clarity and adds a check for zero fec_capability.

v2:
* Update documentation for rte_eth_fec_set() to fix review comments.
* Don't rename the fec_capa parameter of rte_eth_fec_get() but
  add a proper description instead.

Denis Pryazhennikov (3):
  ethdev: update documentation for API to set FEC
  ethdev: check that at least one FEC mode is specified
  ethdev: update documentation for API to get FEC

 lib/ethdev/rte_ethdev.c |  5 +
 lib/ethdev/rte_ethdev.h | 16 
 2 files changed, 13 insertions(+), 8 deletions(-)

-- 
2.37.0 (Apple Git-136)



[RFC PATCH v2 2/3] ethdev: check that at least one FEC mode is specified

2023-05-08 Thread Denis Pryazhennikov
The behaviour is undefined in the rte_eth_fec_set() function
when the fec_capa parameter is equal to zero.
Add a check to handle this case.

Signed-off-by: Denis Pryazhennikov 
Acked-by: Ivan Malov 
Acked-by: Viacheslav Galaktionov 
---
 lib/ethdev/rte_ethdev.c | 5 +
 1 file changed, 5 insertions(+)

diff --git a/lib/ethdev/rte_ethdev.c b/lib/ethdev/rte_ethdev.c
index 4d0325568322..d02ee161cf6d 100644
--- a/lib/ethdev/rte_ethdev.c
+++ b/lib/ethdev/rte_ethdev.c
@@ -4751,6 +4751,11 @@ rte_eth_fec_set(uint16_t port_id, uint32_t fec_capa)
RTE_ETH_VALID_PORTID_OR_ERR_RET(port_id, -ENODEV);
dev = &rte_eth_devices[port_id];
 
+   if (fec_capa == 0) {
+   RTE_ETHDEV_LOG(ERR, "At least one FEC mode should be 
specified\n");
+   return -EINVAL;
+   }
+
if (*dev->dev_ops->fec_set == NULL)
return -ENOTSUP;
ret = eth_err(port_id, (*dev->dev_ops->fec_set)(dev, fec_capa));
-- 
2.37.0 (Apple Git-136)



[RFC PATCH v2 1/3] ethdev: update documentation for API to set FEC

2023-05-08 Thread Denis Pryazhennikov
The documentation for the rte_eth_fec_set() is updated
to provide more detailed information about how FEC modes are
handled. It also includes a description of the case when only
the AUTO bit is set.

Signed-off-by: Denis Pryazhennikov 
Acked-by: Ivan Malov 
Acked-by: Viacheslav Galaktionov 
---
 lib/ethdev/rte_ethdev.h | 11 +++
 1 file changed, 7 insertions(+), 4 deletions(-)

diff --git a/lib/ethdev/rte_ethdev.h b/lib/ethdev/rte_ethdev.h
index 99fe9e238b20..777dc521494c 100644
--- a/lib/ethdev/rte_ethdev.h
+++ b/lib/ethdev/rte_ethdev.h
@@ -4227,10 +4227,13 @@ int rte_eth_fec_get(uint16_t port_id, uint32_t 
*fec_capa);
  * @param port_id
  *   The port identifier of the Ethernet device.
  * @param fec_capa
- *   A bitmask of allowed FEC modes. If AUTO bit is set, other
- *   bits specify FEC modes which may be negotiated. If AUTO
- *   bit is clear, specify FEC modes to be used (only one valid
- *   mode per speed may be set).
+ *   A bitmask of allowed FEC modes.
+ *   If only the AUTO bit is set, the decision on which FEC
+ *   mode to use will be made by HW/FW or driver.
+ *   If the AUTO bit is set with some FEC modes, only specified
+ *   FEC modes can be set.
+ *   If AUTO bit is clear, specify FEC mode to be used
+ *   (only one valid mode per speed may be set).
  * @return
  *   - (0) if successful.
  *   - (-EINVAL) if the FEC mode is not valid.
-- 
2.37.0 (Apple Git-136)



[RFC PATCH v2 3/3] ethdev: update documentation for API to get FEC

2023-05-08 Thread Denis Pryazhennikov
The documentation for the rte_eth_fec_get() is updated
to clarify the description for the fec_capa parameter.
The previous description implied that more than one FEC
mode can be obtained.

Signed-off-by: Denis Pryazhennikov 
---
 lib/ethdev/rte_ethdev.h | 5 +
 1 file changed, 1 insertion(+), 4 deletions(-)

diff --git a/lib/ethdev/rte_ethdev.h b/lib/ethdev/rte_ethdev.h
index 777dc521494c..ab9732f21bdf 100644
--- a/lib/ethdev/rte_ethdev.h
+++ b/lib/ethdev/rte_ethdev.h
@@ -4204,10 +4204,7 @@ int rte_eth_fec_get_capability(uint16_t port_id,
  * @param port_id
  *   The port identifier of the Ethernet device.
  * @param fec_capa
- *   A bitmask of enabled FEC modes. If AUTO bit is set, other
- *   bits specify FEC modes which may be negotiated. If AUTO
- *   bit is clear, specify FEC modes to be used (only one valid
- *   mode per speed may be set).
+ *   A bitmask with the current FEC mode.
  * @return
  *   - (0) if successful.
  *   - (-ENOTSUP) if underlying hardware OR driver doesn't support.
-- 
2.37.0 (Apple Git-136)



RE: [PATCH 0/2] vhost: add port mirroring function in the vhost lib

2023-05-08 Thread Jiang, Cheng1
Hi Maxime,

> -Original Message-
> From: Maxime Coquelin 
> Sent: Wednesday, May 3, 2023 5:37 PM
> To: Jiang, Cheng1 ; Xia, Chenbo
> 
> Cc: dev@dpdk.org; Hu, Jiayu ; Ding, Xuan
> ; Ma, WenwuX ; Wang, YuanX
> ; He, Xingguang ; David
> Marchand 
> Subject: Re: [PATCH 0/2] vhost: add port mirroring function in the vhost lib
> 
> Hi Cheng,
> 
> On 4/21/23 03:09, Cheng Jiang wrote:
> > Similar to the port mirroring function on the switch or router, this
> > patch set implements such function on the Vhost lib. When data is sent
> > to a front-end, it will also send the data to its mirror front-end.
> > When data is received from a front-end, it will also send the data to
> > its mirror front-end.
> 
> Why not just keeping mirroring in the switch/router?
> I am really not convinced this is the way to go:
> 1. API is too complex
> 2. It requires async support
> 3. There is too much code duplication, it increases  virtio-net.c by
> 30%, and it is without packed ring support.
> 4. If mirror port is down for any reason, packets to/from the original
> port are dropped.
> 5. It seems to assume negotiated features of the two ports are
> identical, e.g. Virtio-net header length? If so, that's not a
> manageable solution.

Thank you for your feedback.
I concur that placing the mirror function in the Vhost library is not ideal. We 
are currently considering implementing either a mirror Vhost PMD, or adding a 
function to TestPMD to handle this functionality.
Would you please share your thoughts on this plan and let us know which option 
you prefer?

Thanks a lot,
Cheng



> 
> Regards,
> Maxime
> 
> >
> > Cheng Jiang (2):
> >vhost: add ingress API for port mirroring datapath
> >vhost: add egress API for port mirroring datapath
> >
> >   lib/vhost/rte_vhost_async.h |   17 +
> >   lib/vhost/version.map   |3 +
> >   lib/vhost/virtio_net.c  | 1266 +++
> >   3 files changed, 1286 insertions(+)
> >
> > --
> > 2.35.1
> >



RE: [PATCH v7 4/4] eal: add nonnull and access function attributes

2023-05-08 Thread Morten Brørup
> From: Tyler Retzlaff [mailto:roret...@linux.microsoft.com]
> Sent: Monday, 6 February 2023 18.29
> 
> On Mon, Feb 06, 2023 at 05:49:18PM +0100, Morten Brørup wrote:
> > > From: David Marchand [mailto:david.march...@redhat.com]
> > > Sent: Monday, 6 February 2023 17.11
> > >
> > > On Wed, Feb 1, 2023 at 2:16 PM Thomas Monjalon 
> > > wrote:

[...]

> >
> > I'm leaning towards following the existing convention in rte_common.h, and
> embrace Thomas' argument to make them more verbose in order to reduce the risk
> of wrong use. In other words, define these:
> >
> > __rte_nonnull(...)
> > __rte_read_only(ptr_index)
> > __rte_read_only_size(ptr_index, size_index)
> > __rte_write_only(ptr_index)
> > __rte_write_only_size(ptr_index, size_index)
> > __rte_read_write(ptr_index)
> > __rte_read_write_size(ptr_index, size_index)
> > __rte_no_access(ptr_index)
> > __rte_no_access_size(ptr_index, size_index)
> >
> > >
> > >
> > > As for the lock annotations series, if you are not confident with the
> > > form I went with, I don't mind deferring to a later release.
> >
> > The form follows the existing convention in rte_common.h, and I think we
> should stick with it.
> >
> > > Though it adds more work on my pile like rebasing the vhost library.
> > > Additionnally, we lose the opportunity to catch introduction of new
> > > lock issues in the dpdk tree.
> >
> > Conclusion:
> >
> > The names I listed in this email, and what David already has in his lock
> annotation patch, are both in line with an existing convention already
> established in rte_common.h. So unless someone objects very soon, let's go for
> that.

David, Thomas,

FYI:

I am deferring a new version this patch until a later DPDK release, so it 
doesn't get too much in the way of Tyler's MSVC patches.

Stretch goal: I'm considering if these new attributes could somehow also 
support MSVC, but let's not discuss that now!

PS: The other patches in the series are independent of this patch, and can be 
considered individually.

-Morten



RE: [PATCH] cryptodev: clarify error codes returned

2023-05-08 Thread Power, Ciara
Hi Anoob,

> -Original Message-
> From: Anoob Joseph 
> Sent: Monday 24 April 2023 10:19
> To: Akhil Goyal ; Fan Zhang
> 
> Cc: Jerin Jacob ; dev@dpdk.org
> Subject: [PATCH] cryptodev: clarify error codes returned
> 
> When symmetric sessions are created, it may fail due to non-critical errors.
> When PMD doesn't support the specific combination that application
> requested, it can return -ENOTSUP which can be handled so in application.
> The API is already setting rte_errno according to the reason of the failure.
> Clarifying this in the spec to list down possible error codes.
> 
> Fixes: bdce2564dbf7 ("cryptodev: rework session framework")
> 
> Signed-off-by: Anoob Joseph 
> ---
>  lib/cryptodev/rte_cryptodev.h | 7 +--
>  1 file changed, 5 insertions(+), 2 deletions(-)

Acked-by: Ciara Power 


[PATCH v2] ethdev: add Tx queue flow matching item

2023-05-08 Thread kirankumark
From: Kiran Kumar K 

Adding support for Tx queue flow matching item.
This item is valid only for egress rules.
An example use case would be that application can
set different vlan insert rules with different PCP values
based on Tx queue number.

Signed-off-by: Kiran Kumar K 
---
 app/test-pmd/cmdline_flow.c | 28 +
 doc/guides/prog_guide/rte_flow.rst  |  7 ++
 doc/guides/rel_notes/release_23_07.rst  |  5 
 doc/guides/testpmd_app_ug/testpmd_funcs.rst |  4 +++
 lib/ethdev/rte_flow.c   |  1 +
 lib/ethdev/rte_flow.h   | 26 +++
 6 files changed, 71 insertions(+)

diff --git a/app/test-pmd/cmdline_flow.c b/app/test-pmd/cmdline_flow.c
index 58939ec321..a68a6080a8 100644
--- a/app/test-pmd/cmdline_flow.c
+++ b/app/test-pmd/cmdline_flow.c
@@ -496,6 +496,8 @@ enum index {
ITEM_QUOTA_STATE_NAME,
ITEM_AGGR_AFFINITY,
ITEM_AGGR_AFFINITY_VALUE,
+   ITEM_TX_QUEUE,
+   ITEM_TX_QUEUE_VALUE,
 
/* Validate/create actions. */
ACTIONS,
@@ -1452,6 +1454,7 @@ static const enum index next_item[] = {
ITEM_METER,
ITEM_QUOTA,
ITEM_AGGR_AFFINITY,
+   ITEM_TX_QUEUE,
END_SET,
ZERO,
 };
@@ -1953,6 +1956,12 @@ static const enum index item_aggr_affinity[] = {
ZERO,
 };
 
+static const enum index item_tx_queue[] = {
+   ITEM_TX_QUEUE_VALUE,
+   ITEM_NEXT,
+   ZERO,
+};
+
 static const enum index next_action[] = {
ACTION_END,
ACTION_VOID,
@@ -6945,6 +6954,22 @@ static const struct token token_list[] = {
.args = ARGS(ARGS_ENTRY(struct rte_flow_item_aggr_affinity,
affinity)),
},
+   [ITEM_TX_QUEUE] = {
+   .name = "tx_queue",
+   .help = "match on the tx queue of send packet",
+   .priv = PRIV_ITEM(TX_QUEUE,
+ sizeof(struct rte_flow_item_tx_queue)),
+   .next = NEXT(item_tx_queue),
+   .call = parse_vc,
+   },
+   [ITEM_TX_QUEUE_VALUE] = {
+   .name = "tx_queue_value",
+   .help = "tx queue value",
+   .next = NEXT(item_tx_queue, NEXT_ENTRY(COMMON_UNSIGNED),
+item_param),
+   .args = ARGS(ARGS_ENTRY(struct rte_flow_item_tx_queue,
+   tx_queue)),
+   },
 };
 
 /** Remove and return last entry from argument stack. */
@@ -11849,6 +11874,9 @@ flow_item_default_mask(const struct rte_flow_item *item)
case RTE_FLOW_ITEM_TYPE_AGGR_AFFINITY:
mask = &rte_flow_item_aggr_affinity_mask;
break;
+   case RTE_FLOW_ITEM_TYPE_TX_QUEUE:
+   mask = &rte_flow_item_tx_queue_mask;
+   break;
default:
break;
}
diff --git a/doc/guides/prog_guide/rte_flow.rst 
b/doc/guides/prog_guide/rte_flow.rst
index 32fc45516a..ac5c65131f 100644
--- a/doc/guides/prog_guide/rte_flow.rst
+++ b/doc/guides/prog_guide/rte_flow.rst
@@ -1486,6 +1486,13 @@ This item is meant to use the same structure as `Item: 
PORT_REPRESENTOR`_.
 
 See also `Action: REPRESENTED_PORT`_.
 
+Item: ``TX_QUEUE``
+^^^
+
+Matches on the Tx queue of send packet .
+
+- ``tx_queue``: Tx queue.
+
 Item: ``AGGR_AFFINITY``
 ^^^
 
diff --git a/doc/guides/rel_notes/release_23_07.rst 
b/doc/guides/rel_notes/release_23_07.rst
index a9b1293689..bb04d99125 100644
--- a/doc/guides/rel_notes/release_23_07.rst
+++ b/doc/guides/rel_notes/release_23_07.rst
@@ -55,6 +55,11 @@ New Features
  Also, make sure to start the actual text at the margin.
  ===
 
+   * **Added flow matching of tx queue.**
+
+ Added ``RTE_FLOW_ITEM_TYPE_TX_QUEUE`` rte_flow pattern to match tx queue 
of
+ send packet.
+
 
 Removed Items
 -
diff --git a/doc/guides/testpmd_app_ug/testpmd_funcs.rst 
b/doc/guides/testpmd_app_ug/testpmd_funcs.rst
index 8f23847859..29f7dd4428 100644
--- a/doc/guides/testpmd_app_ug/testpmd_funcs.rst
+++ b/doc/guides/testpmd_app_ug/testpmd_funcs.rst
@@ -3779,6 +3779,10 @@ This section lists supported pattern items and their 
attributes, if any.
 
   - ``affinity {value}``: aggregated port (starts from 1).
 
+- ``tx_queue``: match tx queue of send packet.
+
+  - ``tx_queue {value}``: send queue value (starts from 0).
+
 - ``send_to_kernel``: send packets to kernel.
 
 
diff --git a/lib/ethdev/rte_flow.c b/lib/ethdev/rte_flow.c
index 69e6e749f7..f0d7f868fa 100644
--- a/lib/ethdev/rte_flow.c
+++ b/lib/ethdev/rte_flow.c
@@ -164,6 +164,7 @@ static const struct rte_flow_desc_data rte_flow_desc_item[] 
= {
MK_FLOW_ITEM(IPV6_ROUTING_EXT, sizeof(struct 
rte_flow_item_ipv6_routing_ext)),
MK_FLOW_ITEM(QUOTA, sizeof(struct rte_flow_item_quota)),
MK_FLOW_ITEM(AGGR_AFFINITY, sizeof(struct rte_

Re: [PATCH v2 0/3] vhost: add device op to offload the interrupt kick

2023-05-08 Thread Eelco Chaudron



On 5 Apr 2023, at 14:40, Eelco Chaudron wrote:

> This series adds an operation callback which gets called every time the
> library wants to call eventfd_write(). This eventfd_write() call could
> result in a system call, which could potentially block the PMD thread.
>
> The callback function can decide whether it's ok to handle the
> eventfd_write() now or have the newly introduced function,
> rte_vhost_notify_guest(), called at a later time.
>
> This can be used by 3rd party applications, like OVS, to avoid system
> calls being called as part of the PMD threads.


Wondering if anyone had a chance to look at this patchset.

Cheers,

Eelco

> v2: - Used vhost_virtqueue->index to find index for operation.
> - Aligned function name to VDUSE RFC patchset.
> - Added error and offload statistics counter.
> - Mark new API as experimental.
> - Change the virtual queue spin lock to read/write spin lock.
> - Made shared counters atomic.
> - Add versioned rte_vhost_driver_callback_register() for
>   ABI compliance.
>
> Eelco Chaudron (3):
>   vhost: Change vhost_virtqueue access lock to a read/write one.
>   vhost: make the guest_notifications statistic counter atomic.
>   vhost: add device op to offload the interrupt kick
>
>
>  lib/eal/include/generic/rte_rwlock.h | 17 +
>  lib/vhost/meson.build|  2 +
>  lib/vhost/rte_vhost.h| 23 ++-
>  lib/vhost/socket.c   | 72 --
>  lib/vhost/version.map|  9 +++
>  lib/vhost/vhost.c| 92 +---
>  lib/vhost/vhost.h| 70 ++---
>  lib/vhost/vhost_user.c   | 14 ++---
>  lib/vhost/virtio_net.c   | 90 +--
>  9 files changed, 288 insertions(+), 101 deletions(-)



Re: [PATCH] net/ice: fix dcf init rss blocked

2023-05-08 Thread Tyler Retzlaff
On Sat, May 06, 2023 at 02:48:22PM +0800, Kaiwen Deng wrote:
> The i variable type is uint8_t which causes a dead loop,
> changing it to uint32_t will solve this issue.
> 
> Fixes: 3220d865382c ("net/ice: init RSS during DCF start")
> Cc: sta...@dpdk.org
> 
> Signed-off-by: Kaiwen Deng 
> ---
>  drivers/net/ice/ice_dcf.c | 3 ++-
>  1 file changed, 2 insertions(+), 1 deletion(-)
> 
> diff --git a/drivers/net/ice/ice_dcf.c b/drivers/net/ice/ice_dcf.c
> index 1c3d22ae0f..2cf5de0380 100644
> --- a/drivers/net/ice/ice_dcf.c
> +++ b/drivers/net/ice/ice_dcf.c
> @@ -854,7 +854,8 @@ ice_dcf_init_rss(struct ice_dcf_hw *hw)
>  {
>   struct rte_eth_dev *dev = hw->eth_dev;
>   struct rte_eth_rss_conf *rss_conf;
> - uint8_t i, j, nb_q;
> + uint8_t j, nb_q;
> + uint32_t i;

should be size_t

>   int ret;
>  
>   rss_conf = &dev->data->dev_conf.rx_adv_conf.rss_conf;

with or without suggested change, also isn't there a warning on gcc or
clang that can be enabled to detect this?

Acked-by: Tyler Retzlaff 



Re: 21.11.4 patches review and test

2023-05-08 Thread Kevin Traynor

On 05/05/2023 02:42, Xu, HailinX wrote:

-Original Message-
From: Kevin Traynor 
Sent: Thursday, May 4, 2023 6:11 PM
To: Xu, HailinX ; sta...@dpdk.org
Cc: Stokes, Ian ; Mcnamara, John
; Luca Boccassi ; Xu, Qian Q
; Thomas Monjalon ; Peng,
Yuan ; Chen, Zhaoyan ;
dev@dpdk.org
Subject: Re: 21.11.4 patches review and test

On 04/05/2023 03:13, Xu, HailinX wrote:

-Original Message-
From: Kevin Traynor 
Sent: Tuesday, May 2, 2023 5:35 PM
To: Xu, HailinX ; sta...@dpdk.org
Cc: Stokes, Ian ; Mcnamara, John
; Luca Boccassi ; Xu, Qian
Q ; Thomas Monjalon ;

Peng,

Yuan ; Chen, Zhaoyan ;
dev@dpdk.org
Subject: Re: 21.11.4 patches review and test

On 20/04/2023 11:32, Kevin Traynor wrote:

On 20/04/2023 03:40, Xu, HailinX wrote:

-Original Message-
From: Xu, HailinX 
Sent: Thursday, April 13, 2023 2:13 PM
To: Kevin Traynor ; sta...@dpdk.org
Cc: dev@dpdk.org; Abhishek Marathe

;

Ali Alnubani ; Walker, Benjamin
; David Christensen
; Hemant Agrawal

;

Stokes, Ian ; Jerin Jacob
; Mcnamara, John ;
Ju-Hyoung Lee ; Luca Boccassi
; Pei Zhang ; Xu, Qian Q
; Raslan Darawsheh ;

Thomas

Monjalon ; yangh...@redhat.com; Peng, Yuan
; Chen, Zhaoyan 
Subject: RE: 21.11.4 patches review and test


-Original Message-
From: Kevin Traynor 
Sent: Thursday, April 6, 2023 7:38 PM
To: sta...@dpdk.org
Cc: dev@dpdk.org; Abhishek Marathe
; Ali Alnubani
; Walker, Benjamin
; David Christensen
; Hemant Agrawal
; Stokes, Ian ;
Jerin Jacob ; Mcnamara, John
; Ju-Hyoung Lee

;

Kevin Traynor ; Luca Boccassi
; Pei Zhang ; Xu, Qian Q
; Raslan Darawsheh ;

Thomas

Monjalon ; yangh...@redhat.com; Peng, Yuan
; Chen, Zhaoyan 
Subject: 21.11.4 patches review and test

Hi all,

Here is a list of patches targeted for stable release 21.11.4.

The planned date for the final release is 25th April.

Please help with testing and validation of your use cases and
report any issues/results with reply-all to this mail. For the
final release the fixes and reported validations will be added to
the

release notes.


A release candidate tarball can be found at:

https://dpdk.org/browse/dpdk-stable/tag/?id=v21.11.4-rc1

These patches are located at branch 21.11 of dpdk-stable repo:
https://dpdk.org/browse/dpdk-stable/

Thanks.

Kevin


HI All,

Update the test status for Intel part. Till now dpdk21.11.4-rc1
validation test rate is 85%. No critical issue is found.
2 new bugs are found, 1 new issue is under confirming by Intel Dev.
New bugs:   --20.11.8-rc1 also has these two issues
  1.

pvp_qemu_multi_paths_port_restart:perf_pvp_qemu_vector_rx_mac:

performance drop about 23.5% when send small packets
https://bugs.dpdk.org/show_bug.cgi?id=1212-- no fix yet
  2. some of the virtio tests are failing:-- Intel dev is under

investigating

# Basic Intel(R) NIC testing
* Build & CFLAG compile: cover the build test combination with
latest GCC/Clang version and the popular OS revision such as
  Ubuntu20.04, Ubuntu22.04, Fedora35, Fedora37, RHEL8.6,
RHEL8.4, FreeBSD13.1, SUSE15, CentOS7.9, etc.
- All test done. No new dpdk issue is found.
* PF(i40e, ixgbe): test scenarios including
RTE_FLOW/TSO/Jumboframe/checksum offload/VLAN/VXLAN, etc.
- All test done. No new dpdk issue is found.
* VF(i40e, ixgbe): test scenarios including
VF-RTE_FLOW/TSO/Jumboframe/checksum offload/VLAN/VXLAN, etc.
- All test done. No new dpdk issue is found.
* PF/VF(ice): test scenarios including Switch features/Package
Management/Flow Director/Advanced Tx/Advanced

RSS/ACL/DCF/Flexible

Descriptor, etc.
- All test done. No new dpdk issue is found.
* Intel NIC single core/NIC performance: test scenarios including
PF/VF single core performance test, etc.
- All test done. No new dpdk issue is found.
* IPsec: test scenarios including ipsec/ipsec-gw/ipsec library
basic test - QAT&SW/FIB library, etc.
- On going.

# Basic cryptodev and virtio testing
* Virtio: both function and performance test are covered. Such as
PVP/Virtio_loopback/virtio-user loopback/virtio-net VM2VM perf
testing/VMAWARE ESXI 8.0, etc.
- All test done. found bug1.
* Cryptodev:
  *Function test: test scenarios including Cryptodev API
testing/CompressDev ISA-L/QAT/ZLIB PMD Testing/FIPS, etc.
- Execution rate is 90%. found bug2.
  *Performance test: test scenarios including Thoughput
Performance/Cryptodev Latency, etc.
- All test done. No new dpdk issue is found.

Regards,
Xu, Hailin

Update the test status for Intel part. completed dpdk21.11.4-rc1
all

validation. No critical issue is found.


Hi. Thanks for testing.


2 new bugs are found, 1 new issue is under confirming by Intel Dev.
New bugs: --20.11.8-rc1 also has these two issues
  1.

pvp_qemu_multi_paths_port_restart:perf_pvp_qemu_vector_rx_mac:
performance drop about 23.5% when send small packets

https://bugs.dpdk.org/show_bug.cgi?id=1212  --not fix yet, Only
the specified platform exists


Do you know which patch caaused the regression? I'm not fully clear
from the Bz for 20

Re: [PATCH v2] net/gve: add maintainers for GVE

2023-05-08 Thread Joshua Washington
acked-by: Joshua Washington 


Re: [PATCH v2] net/gve: add maintainers for GVE

2023-05-08 Thread Jeroen de Borst
acked-by: Jeroen de Borst 


On Mon, May 8, 2023 at 8:26 AM Joshua Washington  wrote:
>
> acked-by: Joshua Washington 


Re: [PATCH] net/gve: Check whether the driver is compatible with the device presented.

2023-05-08 Thread Ferruh Yigit
On 5/8/2023 7:23 AM, Rushil Gupta wrote:
>>> +#ifndef _GVE_VERSION_H_
>>> +#define _GVE_VERSION_H_
>>> +
>>> +#include 
>>> +
>>> +#define GVE_VERSION_PREFIX "DPDK-"
>>> +#define GVE_VERSION_MAJOR 1
>>> +#define GVE_VERSION_MINOR 0
>>> +#define GVE_VERSION_SUB 0
>>> +
>> Is this GVE base version, or DPDK gve version?
>> Previous version information was "GVE-1.3.0", now it is becoming
>> "DPDK-1.0.0",
>> is this breaking the version link with GVE base version and creating a
>> new versioning scheme for DPDK GVE driver?
>>
>> Btw, documentation still has "v1.3.0. GVE base code", should it be
>> updated as well?
>>
> DPDK-1.0.0 is the DPDK driver version. GVE driver versioning only
> applies for linux kernel Gvnic driver.
> Similarly Windows Gvnic driver has different versioning system.

So creating a new versioning scheme for DPDK driver, I guess that is OK.
Since there is a new version, is it clear how to manage the versioning,
like when to update major version and when to update minor etc? If so
can you please document this?

And base code is still coming from the Linux kernel driver, I wonder if
a mapping between versions of these two is required.
As mentioned "v1.3.0" is still in the documentation, what happens if
base code updated and some relevant update is required in the dpdk
driver code? How to record/manage this?
Some drivers records the base code version in a readme file in the base
folder, maybe same can be done.



RE: [PATCH v3 5/5] config/arm: add AMD CDX

2023-05-08 Thread Honnappa Nagarahalli


> >>
> >> On 4/21/2023 3:54 PM, Nipun Gupta wrote:
> >>> Adding support for AMD CDX devices
> >>>
> >>> Signed-off-by: Nipun Gupta 
> >>> ---
> >>>  config/arm/arm64_cdx_linux_gcc | 17 +
> >>>  config/arm/meson.build | 14 ++
> >>>  2 files changed, 31 insertions(+)
> >>>  create mode 100644 config/arm/arm64_cdx_linux_gcc
> >>>
> >>> diff --git a/config/arm/arm64_cdx_linux_gcc
> >> b/config/arm/arm64_cdx_linux_gcc
> >>> new file mode 100644
> >>> index 00..8e6d619dae
> >>> --- /dev/null
> >>> +++ b/config/arm/arm64_cdx_linux_gcc
> >>> @@ -0,0 +1,17 @@
> >>> +[binaries]
> >>> +c = ['ccache', 'aarch64-linux-gnu-gcc'] cpp = ['ccache',
> >>> +'aarch64-linux-gnu-g++'] ar = 'aarch64-linux-gnu-ar'
> >>> +as = 'aarch64-linux-gnu-as'
> >>> +strip = 'aarch64-linux-gnu-strip'
> >>> +pkgconfig = 'aarch64-linux-gnu-pkg-config'
> >>> +pcap-config = ''
> >>> +
> >>> +[host_machine]
> >>> +system = 'linux'
> >>> +cpu_family = 'aarch64'
> >>> +cpu = 'armv8-a'
> >>> +endian = 'little'
> >>> +
> >>> +[properties]
> >>> +platform = 'cdx'
> >>> diff --git a/config/arm/meson.build b/config/arm/meson.build index
> >>> 5213434ca4..39b8929534 100644
> >>> --- a/config/arm/meson.build
> >>> +++ b/config/arm/meson.build
> >>> @@ -305,6 +305,18 @@ soc_bluefield = {
> >>>  'numa': false
> >>>  }
> >>>
> >>> +soc_cdx = {
> >>> +'description': 'AMD CDX',
> >>> +'implementer': '0x41',
> >>> +'part_number': '0xd42',
> >>> +'flags': [
> >>> +['RTE_MACHINE', '"cdx"'],
> >>> +['RTE_MAX_LCORE', 16],
> >>> +['RTE_MAX_NUMA_NODES', 1]
> >>> +],
> >>> +'numa': false
> >>> +}
> >>
> >> Hi Nipun,
> >>
> >> Why we need a new arm platform/config? Is it because of above flags?
> >> If it can work with default values, I think we can drop this patch.
> >
> > Hi Ferruh,
> >
> > CDX driver works with generic ARM compilation too
> (arm64_armv8_linux_gcc).
> >
> > The versal platforms supporting CDX have A78 cores, and adding this
> > cdx config Helps to provide gcc option "march= armv8.4-a" which is for
> > implementer "0xd42" (ARM cortex A78 cores)., whereas for generic ARM
> > compilation "march= armv8-a".
> >
> > Maybe ARM guys can provide more information regarding if there is any
> > impact on using generic architecture flag (i.e. march= armv8a) on A78 cores.
> >
> 
> Hi Honnappa, Ruifeng,
> 
> Can you please support on this question, what is the difference of 'march=
> armv8-a' flag (comparing march= armv8a)?
I am assuming you meant, 'march=armv8.4-a' vs 'march=armv8-a'.

'march=armv8-a' will compile for the ARM V8.0-a ISA which is the base V8-a 
version of the ISA. This compilation ensures that the binaries can run on any 
ARM V8-a platforms.
'march=armv8.4-a' will compile the binary for ARM V8.4 ISA. The binary will use 
additional (possibly instructions that improve performance) instructions which 
are not supported in ARM V8.0-a. This binary can run only on machines that 
support V8.4-a and higher. But, it will provide optimized binary for the target 
platform.

> Should we consider adding an arm config file to support this flag?
We support the optimized binary through the config files. We should keep these 
config changes.

> 
> Thanks,
> ferruh
> 



Re: [PATCH v2] net/gve: add maintainers for GVE

2023-05-08 Thread Ferruh Yigit
>> On 5/8/2023 5:49 AM, Rushil Gupta wrote:

Moved and accumulated acks down.

>> 
>>> On Thu, May 4, 2023 at 7:19 PM Junfeng Guo  wrote:

 Add maintainers from Google for GVE.

 Signed-off-by: Junfeng Guo 
 Signed-off-by: Jeroen de Borst 
 Signed-off-by: Rushil Gupta 
 Signed-off-by: Joshua Washington > 
>>>
>>> acked-by: Rushil Gupta 
>>> 
>> 
>> acked-by: Joshua Washington 
>>
>
> acked-by: Jeroen de Borst 
>

Applied to dpdk-next-net/main, thanks.


Re: [PATCH v3 5/5] config/arm: add AMD CDX

2023-05-08 Thread Ferruh Yigit
On 5/8/2023 6:16 PM, Honnappa Nagarahalli wrote:
> 
> 

 On 4/21/2023 3:54 PM, Nipun Gupta wrote:
> Adding support for AMD CDX devices
>
> Signed-off-by: Nipun Gupta 
> ---
>  config/arm/arm64_cdx_linux_gcc | 17 +
>  config/arm/meson.build | 14 ++
>  2 files changed, 31 insertions(+)
>  create mode 100644 config/arm/arm64_cdx_linux_gcc
>
> diff --git a/config/arm/arm64_cdx_linux_gcc
 b/config/arm/arm64_cdx_linux_gcc
> new file mode 100644
> index 00..8e6d619dae
> --- /dev/null
> +++ b/config/arm/arm64_cdx_linux_gcc
> @@ -0,0 +1,17 @@
> +[binaries]
> +c = ['ccache', 'aarch64-linux-gnu-gcc'] cpp = ['ccache',
> +'aarch64-linux-gnu-g++'] ar = 'aarch64-linux-gnu-ar'
> +as = 'aarch64-linux-gnu-as'
> +strip = 'aarch64-linux-gnu-strip'
> +pkgconfig = 'aarch64-linux-gnu-pkg-config'
> +pcap-config = ''
> +
> +[host_machine]
> +system = 'linux'
> +cpu_family = 'aarch64'
> +cpu = 'armv8-a'
> +endian = 'little'
> +
> +[properties]
> +platform = 'cdx'
> diff --git a/config/arm/meson.build b/config/arm/meson.build index
> 5213434ca4..39b8929534 100644
> --- a/config/arm/meson.build
> +++ b/config/arm/meson.build
> @@ -305,6 +305,18 @@ soc_bluefield = {
>  'numa': false
>  }
>
> +soc_cdx = {
> +'description': 'AMD CDX',
> +'implementer': '0x41',
> +'part_number': '0xd42',
> +'flags': [
> +['RTE_MACHINE', '"cdx"'],
> +['RTE_MAX_LCORE', 16],
> +['RTE_MAX_NUMA_NODES', 1]
> +],
> +'numa': false
> +}

 Hi Nipun,

 Why we need a new arm platform/config? Is it because of above flags?
 If it can work with default values, I think we can drop this patch.
>>>
>>> Hi Ferruh,
>>>
>>> CDX driver works with generic ARM compilation too
>> (arm64_armv8_linux_gcc).
>>>
>>> The versal platforms supporting CDX have A78 cores, and adding this
>>> cdx config Helps to provide gcc option "march= armv8.4-a" which is for
>>> implementer "0xd42" (ARM cortex A78 cores)., whereas for generic ARM
>>> compilation "march= armv8-a".
>>>
>>> Maybe ARM guys can provide more information regarding if there is any
>>> impact on using generic architecture flag (i.e. march= armv8a) on A78 cores.
>>>
>>
>> Hi Honnappa, Ruifeng,
>>
>> Can you please support on this question, what is the difference of 'march=
>> armv8-a' flag (comparing march= armv8a)?
> I am assuming you meant, 'march=armv8.4-a' vs 'march=armv8-a'.
> 

Yes

> 'march=armv8-a' will compile for the ARM V8.0-a ISA which is the base V8-a 
> version of the ISA. This compilation ensures that the binaries can run on any 
> ARM V8-a platforms.
> 'march=armv8.4-a' will compile the binary for ARM V8.4 ISA. The binary will 
> use additional (possibly instructions that improve performance) instructions 
> which are not supported in ARM V8.0-a. This binary can run only on machines 
> that support V8.4-a and higher. But, it will provide optimized binary for the 
> target platform.
> 
>> Should we consider adding an arm config file to support this flag?
> We support the optimized binary through the config files. We should keep 
> these config changes.
> 

Got it, thanks for the guidance.



[v2] net/gve: check driver compatibility

2023-05-08 Thread Rushil Gupta
Change gve_driver_info fields to report DPDK as OS type and DPDK RTE
version as OS version, reserving driver_version fields for GVE driver
version based on features.

Signed-off-by: Rushil Gupta 
Signed-off-by: Joshua Washington 
Signed-off-by: Junfeng Guo 
Signed-off-by: Jeroen de Borst 
---
 drivers/net/gve/base/gve.h|  3 --
 drivers/net/gve/base/gve_adminq.c | 19 ++
 drivers/net/gve/base/gve_adminq.h | 46 ++-
 drivers/net/gve/base/gve_osdep.h  | 24 
 drivers/net/gve/gve_ethdev.c  | 61 +--
 drivers/net/gve/gve_ethdev.h  |  2 +-
 drivers/net/gve/gve_version.c | 14 +++
 drivers/net/gve/gve_version.h | 25 +
 drivers/net/gve/meson.build   |  7 
 9 files changed, 185 insertions(+), 16 deletions(-)
 create mode 100644 drivers/net/gve/gve_version.c
 create mode 100644 drivers/net/gve/gve_version.h

diff --git a/drivers/net/gve/base/gve.h b/drivers/net/gve/base/gve.h
index 2dc4507acb..89f9654a72 100644
--- a/drivers/net/gve/base/gve.h
+++ b/drivers/net/gve/base/gve.h
@@ -8,9 +8,6 @@
 
 #include "gve_desc.h"
 
-#define GVE_VERSION"1.3.0"
-#define GVE_VERSION_PREFIX "GVE-"
-
 #ifndef GOOGLE_VENDOR_ID
 #define GOOGLE_VENDOR_ID   0x1ae0
 #endif
diff --git a/drivers/net/gve/base/gve_adminq.c 
b/drivers/net/gve/base/gve_adminq.c
index e745b709b2..2e5099a5b0 100644
--- a/drivers/net/gve/base/gve_adminq.c
+++ b/drivers/net/gve/base/gve_adminq.c
@@ -401,6 +401,9 @@ static int gve_adminq_issue_cmd(struct gve_priv *priv,
case GVE_ADMINQ_GET_PTYPE_MAP:
priv->adminq_get_ptype_map_cnt++;
break;
+   case GVE_ADMINQ_VERIFY_DRIVER_COMPATIBILITY:
+   priv->adminq_verify_driver_compatibility_cnt++;
+   break;
default:
PMD_DRV_LOG(ERR, "unknown AQ command opcode %d", opcode);
}
@@ -465,6 +468,22 @@ int gve_adminq_configure_device_resources(struct gve_priv 
*priv,
return gve_adminq_execute_cmd(priv, &cmd);
 }
 
+int gve_adminq_verify_driver_compatibility(struct gve_priv *priv,
+  u64 driver_info_len,
+  dma_addr_t driver_info_addr)
+{
+   union gve_adminq_command cmd;
+
+   memset(&cmd, 0, sizeof(cmd));
+   cmd.opcode = cpu_to_be32(GVE_ADMINQ_VERIFY_DRIVER_COMPATIBILITY);
+   cmd.verify_driver_compatibility = (struct 
gve_adminq_verify_driver_compatibility) {
+   .driver_info_len = cpu_to_be64(driver_info_len),
+   .driver_info_addr = cpu_to_be64(driver_info_addr),
+   };
+
+   return gve_adminq_execute_cmd(priv, &cmd);
+}
+
 int gve_adminq_deconfigure_device_resources(struct gve_priv *priv)
 {
union gve_adminq_command cmd;
diff --git a/drivers/net/gve/base/gve_adminq.h 
b/drivers/net/gve/base/gve_adminq.h
index 05550119de..e30b184913 100644
--- a/drivers/net/gve/base/gve_adminq.h
+++ b/drivers/net/gve/base/gve_adminq.h
@@ -1,6 +1,6 @@
 /* SPDX-License-Identifier: MIT
  * Google Virtual Ethernet (gve) driver
- * Copyright (C) 2015-2022 Google, Inc.
+ * Copyright (C) 2015-2023 Google, Inc.
  */
 
 #ifndef _GVE_ADMINQ_H
@@ -23,6 +23,7 @@ enum gve_adminq_opcodes {
GVE_ADMINQ_REPORT_STATS = 0xC,
GVE_ADMINQ_REPORT_LINK_SPEED= 0xD,
GVE_ADMINQ_GET_PTYPE_MAP= 0xE,
+   GVE_ADMINQ_VERIFY_DRIVER_COMPATIBILITY  = 0xF,
 };
 
 /* Admin queue status codes */
@@ -145,6 +146,44 @@ enum gve_sup_feature_mask {
 };
 
 #define GVE_DEV_OPT_LEN_GQI_RAW_ADDRESSING 0x0
+enum gve_driver_capbility {
+   gve_driver_capability_gqi_qpl = 0,
+   gve_driver_capability_gqi_rda = 1,
+   gve_driver_capability_dqo_qpl = 2, /* reserved for future use */
+   gve_driver_capability_dqo_rda = 3,
+};
+
+#define GVE_CAP1(a) BIT((int)a)
+
+#define GVE_DRIVER_CAPABILITY_FLAGS1 \
+   (GVE_CAP1(gve_driver_capability_gqi_qpl) | \
+GVE_CAP1(gve_driver_capability_gqi_rda) | \
+GVE_CAP1(gve_driver_capability_dqo_rda))
+
+#define GVE_DRIVER_CAPABILITY_FLAGS2 0x0
+#define GVE_DRIVER_CAPABILITY_FLAGS3 0x0
+#define GVE_DRIVER_CAPABILITY_FLAGS4 0x0
+
+struct gve_driver_info {
+   u8 os_type; /* 0x05 = DPDK */
+   u8 driver_major;
+   u8 driver_minor;
+   u8 driver_sub;
+   __be32 os_version_major;
+   __be32 os_version_minor;
+   __be32 os_version_sub;
+   __be64 driver_capability_flags[4];
+   u8 os_version_str1[OS_VERSION_STRLEN];
+   u8 os_version_str2[OS_VERSION_STRLEN];
+};
+
+struct gve_adminq_verify_driver_compatibility {
+   __be64 driver_info_len;
+   __be64 driver_info_addr;
+};
+
+GVE_CHECK_STRUCT_LEN(16,  gve_adminq_verify_driver_compatibility);
+
 
 struct gve_adminq_configure_device_resources {
__be64 counter_array;
@@ -345,6 +384,8 @@ union gve_adminq_command {
struct gve_adminq_report_stats report_stats;
  

[v2] net/gve: check driver compatibility

2023-05-08 Thread Rushil Gupta
Change gve_driver_info fields to report DPDK as OS type and DPDK RTE
version as OS version, reserving driver_version fields for GVE driver
version based on features.

Signed-off-by: Rushil Gupta 
Signed-off-by: Joshua Washington 
Signed-off-by: Junfeng Guo 
Signed-off-by: Jeroen de Borst 
---
 drivers/net/gve/base/gve.h|  3 --
 drivers/net/gve/base/gve_adminq.c | 19 ++
 drivers/net/gve/base/gve_adminq.h | 46 ++-
 drivers/net/gve/base/gve_osdep.h  | 24 
 drivers/net/gve/gve_ethdev.c  | 61 +--
 drivers/net/gve/gve_ethdev.h  |  2 +-
 drivers/net/gve/gve_version.c | 14 +++
 drivers/net/gve/gve_version.h | 25 +
 drivers/net/gve/meson.build   |  7 
 9 files changed, 185 insertions(+), 16 deletions(-)
 create mode 100644 drivers/net/gve/gve_version.c
 create mode 100644 drivers/net/gve/gve_version.h

diff --git a/drivers/net/gve/base/gve.h b/drivers/net/gve/base/gve.h
index 2dc4507acb..89f9654a72 100644
--- a/drivers/net/gve/base/gve.h
+++ b/drivers/net/gve/base/gve.h
@@ -8,9 +8,6 @@
 
 #include "gve_desc.h"
 
-#define GVE_VERSION"1.3.0"
-#define GVE_VERSION_PREFIX "GVE-"
-
 #ifndef GOOGLE_VENDOR_ID
 #define GOOGLE_VENDOR_ID   0x1ae0
 #endif
diff --git a/drivers/net/gve/base/gve_adminq.c 
b/drivers/net/gve/base/gve_adminq.c
index e745b709b2..2e5099a5b0 100644
--- a/drivers/net/gve/base/gve_adminq.c
+++ b/drivers/net/gve/base/gve_adminq.c
@@ -401,6 +401,9 @@ static int gve_adminq_issue_cmd(struct gve_priv *priv,
case GVE_ADMINQ_GET_PTYPE_MAP:
priv->adminq_get_ptype_map_cnt++;
break;
+   case GVE_ADMINQ_VERIFY_DRIVER_COMPATIBILITY:
+   priv->adminq_verify_driver_compatibility_cnt++;
+   break;
default:
PMD_DRV_LOG(ERR, "unknown AQ command opcode %d", opcode);
}
@@ -465,6 +468,22 @@ int gve_adminq_configure_device_resources(struct gve_priv 
*priv,
return gve_adminq_execute_cmd(priv, &cmd);
 }
 
+int gve_adminq_verify_driver_compatibility(struct gve_priv *priv,
+  u64 driver_info_len,
+  dma_addr_t driver_info_addr)
+{
+   union gve_adminq_command cmd;
+
+   memset(&cmd, 0, sizeof(cmd));
+   cmd.opcode = cpu_to_be32(GVE_ADMINQ_VERIFY_DRIVER_COMPATIBILITY);
+   cmd.verify_driver_compatibility = (struct 
gve_adminq_verify_driver_compatibility) {
+   .driver_info_len = cpu_to_be64(driver_info_len),
+   .driver_info_addr = cpu_to_be64(driver_info_addr),
+   };
+
+   return gve_adminq_execute_cmd(priv, &cmd);
+}
+
 int gve_adminq_deconfigure_device_resources(struct gve_priv *priv)
 {
union gve_adminq_command cmd;
diff --git a/drivers/net/gve/base/gve_adminq.h 
b/drivers/net/gve/base/gve_adminq.h
index 05550119de..e30b184913 100644
--- a/drivers/net/gve/base/gve_adminq.h
+++ b/drivers/net/gve/base/gve_adminq.h
@@ -1,6 +1,6 @@
 /* SPDX-License-Identifier: MIT
  * Google Virtual Ethernet (gve) driver
- * Copyright (C) 2015-2022 Google, Inc.
+ * Copyright (C) 2015-2023 Google, Inc.
  */
 
 #ifndef _GVE_ADMINQ_H
@@ -23,6 +23,7 @@ enum gve_adminq_opcodes {
GVE_ADMINQ_REPORT_STATS = 0xC,
GVE_ADMINQ_REPORT_LINK_SPEED= 0xD,
GVE_ADMINQ_GET_PTYPE_MAP= 0xE,
+   GVE_ADMINQ_VERIFY_DRIVER_COMPATIBILITY  = 0xF,
 };
 
 /* Admin queue status codes */
@@ -145,6 +146,44 @@ enum gve_sup_feature_mask {
 };
 
 #define GVE_DEV_OPT_LEN_GQI_RAW_ADDRESSING 0x0
+enum gve_driver_capbility {
+   gve_driver_capability_gqi_qpl = 0,
+   gve_driver_capability_gqi_rda = 1,
+   gve_driver_capability_dqo_qpl = 2, /* reserved for future use */
+   gve_driver_capability_dqo_rda = 3,
+};
+
+#define GVE_CAP1(a) BIT((int)a)
+
+#define GVE_DRIVER_CAPABILITY_FLAGS1 \
+   (GVE_CAP1(gve_driver_capability_gqi_qpl) | \
+GVE_CAP1(gve_driver_capability_gqi_rda) | \
+GVE_CAP1(gve_driver_capability_dqo_rda))
+
+#define GVE_DRIVER_CAPABILITY_FLAGS2 0x0
+#define GVE_DRIVER_CAPABILITY_FLAGS3 0x0
+#define GVE_DRIVER_CAPABILITY_FLAGS4 0x0
+
+struct gve_driver_info {
+   u8 os_type; /* 0x05 = DPDK */
+   u8 driver_major;
+   u8 driver_minor;
+   u8 driver_sub;
+   __be32 os_version_major;
+   __be32 os_version_minor;
+   __be32 os_version_sub;
+   __be64 driver_capability_flags[4];
+   u8 os_version_str1[OS_VERSION_STRLEN];
+   u8 os_version_str2[OS_VERSION_STRLEN];
+};
+
+struct gve_adminq_verify_driver_compatibility {
+   __be64 driver_info_len;
+   __be64 driver_info_addr;
+};
+
+GVE_CHECK_STRUCT_LEN(16,  gve_adminq_verify_driver_compatibility);
+
 
 struct gve_adminq_configure_device_resources {
__be64 counter_array;
@@ -345,6 +384,8 @@ union gve_adminq_command {
struct gve_adminq_report_stats report_stats;
  

Re: [PATCH] net/gve: Check whether the driver is compatible with the device presented.

2023-05-08 Thread Rushil Gupta
Sounds good. I will document the versioning in a separate patch. For
now, the DPDK driver with DQO changes is reported as 1.0.0.

I think there is some confusion here. The base code is not coming from
the linux kernel driver.
Linux kernel driver is in a different repository and it has nothing to
do with the dpdk driver version. Ideally, we'd want the base code of
both drivers to be the same but there are several higher priority
projects that are taking precedence.

Btw I have addressed your comments:
http://patchwork.dpdk.org/project/dpdk/patch/20230508191552.104540-1-rush...@google.com/


On Mon, May 8, 2023 at 9:00 AM Ferruh Yigit  wrote:
>
> On 5/8/2023 7:23 AM, Rushil Gupta wrote:
> >>> +#ifndef _GVE_VERSION_H_
> >>> +#define _GVE_VERSION_H_
> >>> +
> >>> +#include 
> >>> +
> >>> +#define GVE_VERSION_PREFIX "DPDK-"
> >>> +#define GVE_VERSION_MAJOR 1
> >>> +#define GVE_VERSION_MINOR 0
> >>> +#define GVE_VERSION_SUB 0
> >>> +
> >> Is this GVE base version, or DPDK gve version?
> >> Previous version information was "GVE-1.3.0", now it is becoming
> >> "DPDK-1.0.0",
> >> is this breaking the version link with GVE base version and creating a
> >> new versioning scheme for DPDK GVE driver?
> >>
> >> Btw, documentation still has "v1.3.0. GVE base code", should it be
> >> updated as well?
> >>
> > DPDK-1.0.0 is the DPDK driver version. GVE driver versioning only
> > applies for linux kernel Gvnic driver.
> > Similarly Windows Gvnic driver has different versioning system.
>
> So creating a new versioning scheme for DPDK driver, I guess that is OK.
> Since there is a new version, is it clear how to manage the versioning,
> like when to update major version and when to update minor etc? If so
> can you please document this?
>
> And base code is still coming from the Linux kernel driver, I wonder if
> a mapping between versions of these two is required.
> As mentioned "v1.3.0" is still in the documentation, what happens if
> base code updated and some relevant update is required in the dpdk
> driver code? How to record/manage this?
> Some drivers records the base code version in a readme file in the base
> folder, maybe same can be done.
>


[PATCH v5 0/3] add telemetry cmds for ring

2023-05-08 Thread Jie Hai
This patch set supports telemetry cmd to list rings and dump information
of a ring by its name.

v1->v2:
1. Add space after "switch".
2. Fix wrong strlen parameter.

v2->v3:
1. Remove prefix "rte_" for static function.
2. Add Acked-by Konstantin Ananyev for PATCH 1.
3. Introduce functions to return strings instead copy strings.
4. Check pointer to memzone of ring.
5. Remove redundant variable.
6. Hold lock when access ring data.

v3->v4:
1. Update changelog according to reviews of Honnappa Nagarahalli.
2. Add Reviewed-by Honnappa Nagarahalli.
3. Correct grammar in help information.
4. Correct spell warning on "te" reported by checkpatch.pl.
5. Use ring_walk() to query ring info instead of rte_ring_lookup().
6. Fix that type definition the flag field of rte_ring does not match the usage.
7. Use rte_tel_data_add_dict_uint_hex instead of rte_tel_data_add_dict_u64
   for mask and flags.

v4-v5:
1. Add Acked-by Konstantin Ananyev and Chengwen Feng.
2. Add ABI change explanation for commit message of patch 1/3.

Jie Hai (3):
  ring: fix unmatched type definition and usage
  ring: add telemetry cmd to list rings
  ring: add telemetry cmd for ring info

 lib/ring/meson.build |   1 +
 lib/ring/rte_ring.c  | 139 +++
 lib/ring/rte_ring_core.h |   2 +-
 3 files changed, 141 insertions(+), 1 deletion(-)

-- 
2.33.0



[PATCH v5 1/3] ring: fix unmatched type definition and usage

2023-05-08 Thread Jie Hai
Field 'flags' of struct rte_ring is defined as int type. However,
it is used as unsigned int. To ensure consistency, change the
type of flags to unsigned int. Since these two types has the
same byte size, this change is not an ABI change.

Fixes: cc4b218790f6 ("ring: support configurable element size")

Signed-off-by: Jie Hai 
Acked-by: Konstantin Ananyev 
Acked-by: Chengwen Feng 
---
 lib/ring/rte_ring_core.h | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/lib/ring/rte_ring_core.h b/lib/ring/rte_ring_core.h
index 82b237091b71..1c809abeb531 100644
--- a/lib/ring/rte_ring_core.h
+++ b/lib/ring/rte_ring_core.h
@@ -120,7 +120,7 @@ struct rte_ring_hts_headtail {
 struct rte_ring {
char name[RTE_RING_NAMESIZE] __rte_cache_aligned;
/**< Name of the ring. */
-   int flags;   /**< Flags supplied at creation. */
+   uint32_t flags;   /**< Flags supplied at creation. */
const struct rte_memzone *memzone;
/**< Memzone, if any, containing the rte_ring */
uint32_t size;   /**< Size of ring. */
-- 
2.33.0



[PATCH v5 2/3] ring: add telemetry cmd to list rings

2023-05-08 Thread Jie Hai
Add a telemetry command to list the rings used in the system.
An example using this command is shown below:

--> /ring/list
{
  "/ring/list": [
"HT_:7d:00.2",
"MP_mb_pool_0"
  ]
}

Signed-off-by: Jie Hai 
Acked-by: Konstantin Ananyev 
Reviewed-by: Honnappa Nagarahalli 
Acked-by: Huisong Li 
Acked-by: Chengwen Feng 
---
 lib/ring/meson.build |  1 +
 lib/ring/rte_ring.c  | 40 
 2 files changed, 41 insertions(+)

diff --git a/lib/ring/meson.build b/lib/ring/meson.build
index c20685c689ac..7fca958ed7fa 100644
--- a/lib/ring/meson.build
+++ b/lib/ring/meson.build
@@ -18,3 +18,4 @@ indirect_headers += files (
 'rte_ring_rts.h',
 'rte_ring_rts_elem_pvt.h',
 )
+deps += ['telemetry']
diff --git a/lib/ring/rte_ring.c b/lib/ring/rte_ring.c
index 8ed455043dee..0e83d0099363 100644
--- a/lib/ring/rte_ring.c
+++ b/lib/ring/rte_ring.c
@@ -22,6 +22,7 @@
 #include 
 #include 
 #include 
+#include 
 
 #include "rte_ring.h"
 #include "rte_ring_elem.h"
@@ -420,3 +421,42 @@ rte_ring_lookup(const char *name)
 
return r;
 }
+
+static void
+ring_walk(void (*func)(struct rte_ring *, void *), void *arg)
+{
+   struct rte_ring_list *ring_list;
+   struct rte_tailq_entry *tailq_entry;
+
+   ring_list = RTE_TAILQ_CAST(rte_ring_tailq.head, rte_ring_list);
+   rte_mcfg_tailq_read_lock();
+
+   TAILQ_FOREACH(tailq_entry, ring_list, next) {
+   (*func)((struct rte_ring *) tailq_entry->data, arg);
+   }
+
+   rte_mcfg_tailq_read_unlock();
+}
+
+static void
+ring_list_cb(struct rte_ring *r, void *arg)
+{
+   struct rte_tel_data *d = (struct rte_tel_data *)arg;
+
+   rte_tel_data_add_array_string(d, r->name);
+}
+
+static int
+ring_handle_list(const char *cmd __rte_unused,
+   const char *params __rte_unused, struct rte_tel_data *d)
+{
+   rte_tel_data_start_array(d, RTE_TEL_STRING_VAL);
+   ring_walk(ring_list_cb, d);
+   return 0;
+}
+
+RTE_INIT(ring_init_telemetry)
+{
+   rte_telemetry_register_cmd("/ring/list", ring_handle_list,
+   "Returns list of available rings. Takes no parameters");
+}
-- 
2.33.0



[PATCH v5 3/3] ring: add telemetry cmd for ring info

2023-05-08 Thread Jie Hai
This patch supports dump of ring information by its name.
An example using this command is shown below:

--> /ring/info,MP_mb_pool_0
{
  "/ring/info": {
"name": "MP_mb_pool_0",
"socket": 0,
"flags": "0x0",
"producer_type": "MP",
"consumer_type": "MC",
"size": 262144,
"mask": "0x3",
"capacity": 262143,
"used_count": 153197,
"consumer_tail": 2259,
"consumer_head": 2259,
"producer_tail": 155456,
"producer_head": 155456,
"mz_name": "RG_MP_mb_pool_0",
"mz_len": 2097536,
"mz_hugepage_sz": 1073741824,
"mz_socket_id": 0,
"mz_flags": "0x0"
  }
}

Signed-off-by: Jie Hai 
Reviewed-by: Honnappa Nagarahalli 
Acked-by: Konstantin Ananyev 
Acked-by: Huisong Li 
Acked-by: Chengwen Feng 
---
 lib/ring/rte_ring.c | 99 +
 1 file changed, 99 insertions(+)

diff --git a/lib/ring/rte_ring.c b/lib/ring/rte_ring.c
index 0e83d0099363..26c8f2a2e6a2 100644
--- a/lib/ring/rte_ring.c
+++ b/lib/ring/rte_ring.c
@@ -455,8 +455,107 @@ ring_handle_list(const char *cmd __rte_unused,
return 0;
 }
 
+static const char *
+ring_prod_sync_type_to_name(struct rte_ring *r)
+{
+   switch (r->prod.sync_type) {
+   case RTE_RING_SYNC_MT:
+   return "MP";
+   case RTE_RING_SYNC_ST:
+   return "SP";
+   case RTE_RING_SYNC_MT_RTS:
+   return "MP_RTS";
+   case RTE_RING_SYNC_MT_HTS:
+   return "MP_HTS";
+   default:
+   return "Unknown";
+   }
+}
+
+static const char *
+ring_cons_sync_type_to_name(struct rte_ring *r)
+{
+   switch (r->cons.sync_type) {
+   case RTE_RING_SYNC_MT:
+   return "MC";
+   case RTE_RING_SYNC_ST:
+   return "SC";
+   case RTE_RING_SYNC_MT_RTS:
+   return "MC_RTS";
+   case RTE_RING_SYNC_MT_HTS:
+   return "MC_HTS";
+   default:
+   return "Unknown";
+   }
+}
+
+struct ring_info_cb_arg {
+   char *ring_name;
+   struct rte_tel_data *d;
+};
+
+static void
+ring_info_cb(struct rte_ring *r, void *arg)
+{
+   struct ring_info_cb_arg *ring_arg = (struct ring_info_cb_arg *)arg;
+   struct rte_tel_data *d = ring_arg->d;
+   const struct rte_memzone *mz;
+
+   if (strncmp(r->name, ring_arg->ring_name, RTE_RING_NAMESIZE))
+   return;
+
+   rte_tel_data_add_dict_string(d, "name", r->name);
+   rte_tel_data_add_dict_int(d, "socket", r->memzone->socket_id);
+   rte_tel_data_add_dict_uint_hex(d, "flags", r->flags, 0);
+   rte_tel_data_add_dict_string(d, "producer_type",
+   ring_prod_sync_type_to_name(r));
+   rte_tel_data_add_dict_string(d, "consumer_type",
+   ring_cons_sync_type_to_name(r));
+   rte_tel_data_add_dict_uint(d, "size", r->size);
+   rte_tel_data_add_dict_uint_hex(d, "mask", r->mask, 0);
+   rte_tel_data_add_dict_uint(d, "capacity", r->capacity);
+   rte_tel_data_add_dict_uint(d, "used_count", rte_ring_count(r));
+   rte_tel_data_add_dict_uint(d, "consumer_tail", r->cons.tail);
+   rte_tel_data_add_dict_uint(d, "consumer_head", r->cons.head);
+   rte_tel_data_add_dict_uint(d, "producer_tail", r->prod.tail);
+   rte_tel_data_add_dict_uint(d, "producer_head", r->prod.head);
+
+   mz = r->memzone;
+   if (mz == NULL)
+   return;
+   rte_tel_data_add_dict_string(d, "mz_name", mz->name);
+   rte_tel_data_add_dict_uint(d, "mz_len", mz->len);
+   rte_tel_data_add_dict_uint(d, "mz_hugepage_sz", mz->hugepage_sz);
+   rte_tel_data_add_dict_int(d, "mz_socket_id", mz->socket_id);
+   rte_tel_data_add_dict_uint_hex(d, "mz_flags", mz->flags, 0);
+}
+
+static int
+ring_handle_info(const char *cmd __rte_unused, const char *params,
+   struct rte_tel_data *d)
+{
+   char name[RTE_RING_NAMESIZE] = {0};
+   struct ring_info_cb_arg ring_arg;
+
+   if (params == NULL || strlen(params) == 0 ||
+   strlen(params) >= RTE_RING_NAMESIZE)
+   return -EINVAL;
+
+   rte_strlcpy(name, params, RTE_RING_NAMESIZE);
+
+   ring_arg.ring_name = name;
+   ring_arg.d = d;
+
+   rte_tel_data_start_dict(d);
+   ring_walk(ring_info_cb, &ring_arg);
+
+   return 0;
+}
+
 RTE_INIT(ring_init_telemetry)
 {
rte_telemetry_register_cmd("/ring/list", ring_handle_list,
"Returns list of available rings. Takes no parameters");
+   rte_telemetry_register_cmd("/ring/info", ring_handle_info,
+   "Returns ring info. Parameters: ring_name.");
 }
-- 
2.33.0



[PATCH v2] net/ice: fix dcf init rss blocked

2023-05-08 Thread Kaiwen Deng
The i variable type is uint8_t which causes a dead loop,
changing it to size_t will solve this issue.

Signed-off-by: Kaiwen Deng 
---
 drivers/net/ice/ice_dcf.c | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/drivers/net/ice/ice_dcf.c b/drivers/net/ice/ice_dcf.c
index 1c3d22ae0f..3e3004f51b 100644
--- a/drivers/net/ice/ice_dcf.c
+++ b/drivers/net/ice/ice_dcf.c
@@ -854,7 +854,8 @@ ice_dcf_init_rss(struct ice_dcf_hw *hw)
 {
struct rte_eth_dev *dev = hw->eth_dev;
struct rte_eth_rss_conf *rss_conf;
-   uint8_t i, j, nb_q;
+   uint8_t j, nb_q;
+   size_t i;
int ret;
 
rss_conf = &dev->data->dev_conf.rx_adv_conf.rss_conf;
-- 
2.34.1



[PATCH net] net/vmxnet3: fix drop of empty segments in tx

2023-05-08 Thread Ronak Doshi
When empty segments are dropped, some descriptor variable values are
updated in the segment processing loop before it is exited. This can
lead to a wedged queue where all subsequent packets are dropped for
this queue.

Also move the check for empty packet to catch the case of a zero
length packet with multiple segments.

Fixes: d863f19efa4f ("net/vmxnet3: skip empty segments in transmission")
Cc: sta...@dpdk.org
Signed-off-by: Ronak Doshi 
Acked-by: Jochen Behrens 
---
 drivers/net/vmxnet3/vmxnet3_rxtx.c | 33 +++--
 1 file changed, 19 insertions(+), 14 deletions(-)

diff --git a/drivers/net/vmxnet3/vmxnet3_rxtx.c 
b/drivers/net/vmxnet3/vmxnet3_rxtx.c
index a875ffec07..8208cefdff 100644
--- a/drivers/net/vmxnet3/vmxnet3_rxtx.c
+++ b/drivers/net/vmxnet3/vmxnet3_rxtx.c
@@ -412,8 +412,8 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
 
nb_tx = 0;
while (nb_tx < nb_pkts) {
-   Vmxnet3_GenericDesc *gdesc;
-   vmxnet3_buf_info_t *tbi;
+   Vmxnet3_GenericDesc *gdesc = NULL;
+   vmxnet3_buf_info_t *tbi = NULL;
uint32_t first2fill, avail, dw2;
struct rte_mbuf *txm = tx_pkts[nb_tx];
struct rte_mbuf *m_seg = txm;
@@ -457,18 +457,18 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
continue;
}
 
+   /* Skip empty packets */
+   if (unlikely(rte_pktmbuf_pkt_len(txm) == 0)) {
+   txq->stats.drop_total++;
+   rte_pktmbuf_free(txm);
+   nb_tx++;
+   continue;
+   }
+
if (txm->nb_segs == 1 &&
rte_pktmbuf_pkt_len(txm) <= txq->txdata_desc_size) {
struct Vmxnet3_TxDataDesc *tdd;
 
-   /* Skip empty packets */
-   if (unlikely(rte_pktmbuf_pkt_len(txm) == 0)) {
-   txq->stats.drop_total++;
-   rte_pktmbuf_free(txm);
-   nb_tx++;
-   continue;
-   }
-
tdd = (struct Vmxnet3_TxDataDesc *)
((uint8 *)txq->data_ring.base +
 txq->cmd_ring.next2fill *
@@ -481,6 +481,10 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
dw2 = (txq->cmd_ring.gen ^ 0x1) << VMXNET3_TXD_GEN_SHIFT;
first2fill = txq->cmd_ring.next2fill;
do {
+   /* Skip empty segments */
+   if (unlikely(m_seg->data_len == 0))
+   continue;
+
/* Remember the transmit buffer for cleanup */
tbi = txq->cmd_ring.buf_info + txq->cmd_ring.next2fill;
 
@@ -490,10 +494,6 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
 */
gdesc = txq->cmd_ring.base + txq->cmd_ring.next2fill;
 
-   /* Skip empty segments */
-   if (unlikely(m_seg->data_len == 0))
-   continue;
-
if (copy_size) {
uint64 offset =
(uint64)txq->cmd_ring.next2fill *
@@ -514,6 +514,11 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
/* use the right gen for non-SOP desc */
dw2 = txq->cmd_ring.gen << VMXNET3_TXD_GEN_SHIFT;
} while ((m_seg = m_seg->next) != NULL);
+   /* We must have exectuted the complete preceding loop at least
+* once without skipping an empty segment, as we can't have
+* a packet with only empty segments.
+* Thus, tbi and gdesc have been initialized.
+*/
 
/* set the last buf_info for the pkt */
tbi->m = txm;
-- 
2.11.0



[PATCH next] net/vmxnet3: add interrupt handling for reset

2023-05-08 Thread Ronak Doshi
vmxnet3 should call the reset callback when a queue is stopped
by the host. This patch also fixes a logging issue. Currently,
status of only first queue is checked and logged. But status
of all queues should be checked.

Signed-off-by: Ronak Doshi 
Acked-by: Jochen Behrens 
---
 drivers/net/vmxnet3/vmxnet3_ethdev.c | 27 ---
 1 file changed, 20 insertions(+), 7 deletions(-)

diff --git a/drivers/net/vmxnet3/vmxnet3_ethdev.c 
b/drivers/net/vmxnet3/vmxnet3_ethdev.c
index 41073e9798..76e80e3025 100644
--- a/drivers/net/vmxnet3/vmxnet3_ethdev.c
+++ b/drivers/net/vmxnet3/vmxnet3_ethdev.c
@@ -1833,6 +1833,7 @@ vmxnet3_process_events(struct rte_eth_dev *dev)
 {
struct vmxnet3_hw *hw = dev->data->dev_private;
uint32_t events = hw->shared->ecr;
+   int i;
 
if (!events)
return;
@@ -1857,16 +1858,28 @@ vmxnet3_process_events(struct rte_eth_dev *dev)
VMXNET3_WRITE_BAR1_REG(hw, VMXNET3_REG_CMD,
   VMXNET3_CMD_GET_QUEUE_STATUS);
 
-   if (hw->tqd_start->status.stopped)
-   PMD_DRV_LOG(ERR, "tq error 0x%x",
-   hw->tqd_start->status.error);
+   PMD_DRV_LOG(ERR, "queue error event 0x%x for "
+   RTE_ETHER_ADDR_PRT_FMT, events,
+   hw->perm_addr[0], hw->perm_addr[1],
+   hw->perm_addr[2], hw->perm_addr[3],
+   hw->perm_addr[4], hw->perm_addr[5]);
 
-   if (hw->rqd_start->status.stopped)
-   PMD_DRV_LOG(ERR, "rq error 0x%x",
-hw->rqd_start->status.error);
+   for (i = 0; i < hw->num_tx_queues; i++) {
+   if (hw->tqd_start[i].status.stopped)
+   PMD_DRV_LOG(ERR, "tq %d error 0x%x",
+   i, hw->tqd_start[i].status.error);
+   }
+   for (i = 0; i < hw->num_rx_queues; i++) {
+   if (hw->rqd_start[i].status.stopped)
+   PMD_DRV_LOG(ERR, "rq %d error 0x%x",
+   i, hw->rqd_start[i].status.error);
+   }
 
-   /* Reset the device */
/* Have to reset the device */
+   /* Notify the application so that it can reset the device */
+   rte_eth_dev_callback_process(dev,
+RTE_ETH_EVENT_INTR_RESET,
+NULL);
}
 
if (events & VMXNET3_ECR_DIC)
-- 
2.11.0



[PATCH net v2] net/vmxnet3: fix drop of empty segments in tx

2023-05-08 Thread Ronak Doshi
When empty segments are dropped, some descriptor variable values are
updated in the segment processing loop before it is exited. This can
lead to a wedged queue where all subsequent packets are dropped for
this queue.

Also move the check for empty packet to catch the case of a zero
length packet with multiple segments.

Fixes: d863f19efa4f ("net/vmxnet3: skip empty segments in transmission")
Cc: sta...@dpdk.org
Signed-off-by: Ronak Doshi 
Acked-by: Jochen Behrens 
--
Change log
v2: fixed typo
---
 drivers/net/vmxnet3/vmxnet3_rxtx.c | 33 +++--
 1 file changed, 19 insertions(+), 14 deletions(-)

diff --git a/drivers/net/vmxnet3/vmxnet3_rxtx.c 
b/drivers/net/vmxnet3/vmxnet3_rxtx.c
index a875ffec07..f4cade0954 100644
--- a/drivers/net/vmxnet3/vmxnet3_rxtx.c
+++ b/drivers/net/vmxnet3/vmxnet3_rxtx.c
@@ -412,8 +412,8 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts,
 
nb_tx = 0;
while (nb_tx < nb_pkts) {
-   Vmxnet3_GenericDesc *gdesc;
-   vmxnet3_buf_info_t *tbi;
+   Vmxnet3_GenericDesc *gdesc = NULL;
+   vmxnet3_buf_info_t *tbi = NULL;
uint32_t first2fill, avail, dw2;
struct rte_mbuf *txm = tx_pkts[nb_tx];
struct rte_mbuf *m_seg = txm;
@@ -457,18 +457,18 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
continue;
}
 
+   /* Skip empty packets */
+   if (unlikely(rte_pktmbuf_pkt_len(txm) == 0)) {
+   txq->stats.drop_total++;
+   rte_pktmbuf_free(txm);
+   nb_tx++;
+   continue;
+   }
+
if (txm->nb_segs == 1 &&
rte_pktmbuf_pkt_len(txm) <= txq->txdata_desc_size) {
struct Vmxnet3_TxDataDesc *tdd;
 
-   /* Skip empty packets */
-   if (unlikely(rte_pktmbuf_pkt_len(txm) == 0)) {
-   txq->stats.drop_total++;
-   rte_pktmbuf_free(txm);
-   nb_tx++;
-   continue;
-   }
-
tdd = (struct Vmxnet3_TxDataDesc *)
((uint8 *)txq->data_ring.base +
 txq->cmd_ring.next2fill *
@@ -481,6 +481,10 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
dw2 = (txq->cmd_ring.gen ^ 0x1) << VMXNET3_TXD_GEN_SHIFT;
first2fill = txq->cmd_ring.next2fill;
do {
+   /* Skip empty segments */
+   if (unlikely(m_seg->data_len == 0))
+   continue;
+
/* Remember the transmit buffer for cleanup */
tbi = txq->cmd_ring.buf_info + txq->cmd_ring.next2fill;
 
@@ -490,10 +494,6 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
 */
gdesc = txq->cmd_ring.base + txq->cmd_ring.next2fill;
 
-   /* Skip empty segments */
-   if (unlikely(m_seg->data_len == 0))
-   continue;
-
if (copy_size) {
uint64 offset =
(uint64)txq->cmd_ring.next2fill *
@@ -514,6 +514,11 @@ vmxnet3_xmit_pkts(void *tx_queue, struct rte_mbuf 
**tx_pkts,
/* use the right gen for non-SOP desc */
dw2 = txq->cmd_ring.gen << VMXNET3_TXD_GEN_SHIFT;
} while ((m_seg = m_seg->next) != NULL);
+   /* We must have executed the complete preceding loop at least
+* once without skipping an empty segment, as we can't have
+* a packet with only empty segments.
+* Thus, tbi and gdesc have been initialized.
+*/
 
/* set the last buf_info for the pkt */
tbi->m = txm;
-- 
2.11.0



RE: [EXT] [PATCH v2 4/4] app: add testgraph application

2023-05-08 Thread Sunil Kumar Kori
Is there any user guide similar to testpmd ?

> -Original Message-
> From: Vamsi Attunuru 
> Sent: Tuesday, April 25, 2023 6:45 PM
> To: dev@dpdk.org; tho...@monjalon.net; Jerin Jacob Kollanukkaran
> 
> Cc: Vamsi Krishna Attunuru ; Nithin Kumar
> Dabilpuram 
> Subject: [EXT] [PATCH v2 4/4] app: add testgraph application
> 
> External Email
> 
> --
> Patch adds test-graph application to validate graph
> and node libraries.
> 
> Signed-off-by: Vamsi Attunuru 
> ---
>  app/meson.build|1 +
>  app/test-graph/cmdline.c   |  211 +
>  app/test-graph/cmdline_graph.c |  294 +++
>  app/test-graph/cmdline_graph.h |   19 +
>  app/test-graph/meson.build |   14 +
>  app/test-graph/parameters.c|  157 
>  app/test-graph/testgraph.c | 1426 
>  app/test-graph/testgraph.h |   91 ++
>  doc/guides/tools/index.rst |1 +
>  doc/guides/tools/testgraph.rst |  131 +++
>  10 files changed, 2345 insertions(+)
> 
> diff --git a/app/meson.build b/app/meson.build
> index 74d2420f67..6c7b24e604 100644
> --- a/app/meson.build
> +++ b/app/meson.build
> @@ -22,6 +22,7 @@ apps = [
>  'test-eventdev',
>  'test-fib',
>  'test-flow-perf',
> +'test-graph',
>  'test-gpudev',
>  'test-mldev',
>  'test-pipeline',
> diff --git a/app/test-graph/cmdline.c b/app/test-graph/cmdline.c
> new file mode 100644
> index 00..d9474d827a
> --- /dev/null
> +++ b/app/test-graph/cmdline.c
> @@ -0,0 +1,211 @@
> +/* SPDX-License-Identifier: BSD-3-Clause
> + * Copyright(C) 2023 Marvell International Ltd.
> + */

Please check copyright. I think  it should be " Copyright(C) 2023 Marvell." 
Only.
> +
> +#include 
> +
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +
> +#include "cmdline_graph.h"
> +#include "testgraph.h"
> +
> +static struct cmdline *testgraph_cl;
> +static cmdline_parse_ctx_t *main_ctx;
> +
> +/* *** Help command with introduction. *** */
> +struct cmd_help_brief_result {
> + cmdline_fixed_string_t help;
> +};
> +
> +static void
> +cmd_help_brief_parsed(__rte_unused void *parsed_result, struct cmdline
> *cl, __rte_unused void *data)
> +{
> + cmdline_printf(cl,
> +"\n"
> +"Help is available for the following sections:\n\n"
> +"help control: Start and stop 
> graph walk.\n"
> +"help display: Displaying port, 
> stats and
> config "
> +"information.\n"
> +"help config : Configuration 
> information.\n"
> +"help all: All of the above 
> sections.\n\n");
> +}
> +
> +static cmdline_parse_token_string_t cmd_help_brief_help =
> + TOKEN_STRING_INITIALIZER(struct cmd_help_brief_result, help,
> "help");
> +
> +static cmdline_parse_inst_t cmd_help_brief = {
> + .f = cmd_help_brief_parsed,
> + .data = NULL,
> + .help_str = "help: Show help",
> + .tokens = {
> + (void *)&cmd_help_brief_help,
> + NULL,
> + },
> +};
> +
> +/* *** Help command with help sections. *** */
> +struct cmd_help_long_result {
> + cmdline_fixed_string_t help;
> + cmdline_fixed_string_t section;
> +};
> +
> +static void
> +cmd_help_long_parsed(void *parsed_result, struct cmdline *cl,
> __rte_unused void *data)
> +{
> + int show_all = 0;
> + struct cmd_help_long_result *res = parsed_result;
> +
> + if (!strcmp(res->section, "all"))
> + show_all = 1;
> +
> + if (show_all || !strcmp(res->section, "control")) {
> +
> + cmdline_printf(cl, "\n"
> +"Control forwarding:\n"
> +"---\n\n"
> +
> +"start graph_walk\n"
> +" Start graph_walk on worker threads.\n\n"
> +
> +"stop graph_walk\n"
> +" Stop worker threads from running
> graph_walk.\n\n"
> +
> +"quit\n"
> +"Quit to prompt.\n\n");
> + }
> +
> + if (show_all || !strcmp(res->section, "display")) {
> +
> + cmdline_printf(cl,
> +"\n"
> +"Display:\n"
> +"\n\n"
> +
> +"show node_list\n"
> +" Display the list of supported nodes.\n\n"
> +
> +"show graph_stats\n"
> +" Display the node statistics of graph
> cluster.\n\n");
> + }
> +
> + if (show_all || !strcmp(res->section, "config")) {
> + cmdline_printf(cl

[PATCH] net/gve: support queue start and stop operations

2023-05-08 Thread Junfeng Guo
Add support for queue operations for GQI:
 - gve_rx_queue_start
 - gve_tx_queue_start
 - gve_rx_queue_stop
 - gve_tx_queue_stop

Add support for queue operations for DQO:
 - gve_rx_queue_start_dqo
 - gve_tx_queue_start_dqo
 - gve_rx_queue_stop_dqo
 - gve_tx_queue_stop_dqo

Also move the funcs of rxq_mbufs_alloc into the corresponding files.

Signed-off-by: Junfeng Guo 
---
 drivers/net/gve/gve_ethdev.c | 166 +++
 drivers/net/gve/gve_ethdev.h |  36 
 drivers/net/gve/gve_rx.c |  96 ++--
 drivers/net/gve/gve_rx_dqo.c |  97 ++--
 drivers/net/gve/gve_tx.c |  54 ++--
 drivers/net/gve/gve_tx_dqo.c |  54 ++--
 6 files changed, 364 insertions(+), 139 deletions(-)

diff --git a/drivers/net/gve/gve_ethdev.c b/drivers/net/gve/gve_ethdev.c
index 8b6861a24f..1dcb3b3a01 100644
--- a/drivers/net/gve/gve_ethdev.c
+++ b/drivers/net/gve/gve_ethdev.c
@@ -104,81 +104,6 @@ gve_dev_configure(struct rte_eth_dev *dev)
return 0;
 }
 
-static int
-gve_refill_pages(struct gve_rx_queue *rxq)
-{
-   struct rte_mbuf *nmb;
-   uint16_t i;
-   int diag;
-
-   diag = rte_pktmbuf_alloc_bulk(rxq->mpool, &rxq->sw_ring[0], 
rxq->nb_rx_desc);
-   if (diag < 0) {
-   for (i = 0; i < rxq->nb_rx_desc - 1; i++) {
-   nmb = rte_pktmbuf_alloc(rxq->mpool);
-   if (!nmb)
-   break;
-   rxq->sw_ring[i] = nmb;
-   }
-   if (i < rxq->nb_rx_desc - 1)
-   return -ENOMEM;
-   }
-   rxq->nb_avail = 0;
-   rxq->next_avail = rxq->nb_rx_desc - 1;
-
-   for (i = 0; i < rxq->nb_rx_desc; i++) {
-   if (rxq->is_gqi_qpl) {
-   rxq->rx_data_ring[i].addr = rte_cpu_to_be_64(i * 
PAGE_SIZE);
-   } else {
-   if (i == rxq->nb_rx_desc - 1)
-   break;
-   nmb = rxq->sw_ring[i];
-   rxq->rx_data_ring[i].addr = 
rte_cpu_to_be_64(rte_mbuf_data_iova(nmb));
-   }
-   }
-
-   rte_write32(rte_cpu_to_be_32(rxq->next_avail), rxq->qrx_tail);
-
-   return 0;
-}
-
-static int
-gve_refill_dqo(struct gve_rx_queue *rxq)
-{
-   struct rte_mbuf *nmb;
-   uint16_t i;
-   int diag;
-
-   diag = rte_pktmbuf_alloc_bulk(rxq->mpool, &rxq->sw_ring[0], 
rxq->nb_rx_desc);
-   if (diag < 0) {
-   rxq->stats.no_mbufs_bulk++;
-   for (i = 0; i < rxq->nb_rx_desc - 1; i++) {
-   nmb = rte_pktmbuf_alloc(rxq->mpool);
-   if (!nmb)
-   break;
-   rxq->sw_ring[i] = nmb;
-   }
-   if (i < rxq->nb_rx_desc - 1) {
-   rxq->stats.no_mbufs += rxq->nb_rx_desc - 1 - i;
-   return -ENOMEM;
-   }
-   }
-
-   for (i = 0; i < rxq->nb_rx_desc; i++) {
-   if (i == rxq->nb_rx_desc - 1)
-   break;
-   nmb = rxq->sw_ring[i];
-   rxq->rx_ring[i].buf_addr = 
rte_cpu_to_le_64(rte_mbuf_data_iova_default(nmb));
-   rxq->rx_ring[i].buf_id = rte_cpu_to_le_16(i);
-   }
-
-   rxq->nb_rx_hold = 0;
-   rxq->bufq_tail = rxq->nb_rx_desc - 1;
-
-   rte_write32(rxq->bufq_tail, rxq->qrx_tail);
-
-   return 0;
-}
-
 static int
 gve_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete)
 {
@@ -208,65 +133,68 @@ gve_link_update(struct rte_eth_dev *dev, __rte_unused int 
wait_to_complete)
 }
 
 static int
-gve_dev_start(struct rte_eth_dev *dev)
+gve_start_queues(struct rte_eth_dev *dev)
 {
-   uint16_t num_queues = dev->data->nb_tx_queues;
struct gve_priv *priv = dev->data->dev_private;
-   struct gve_tx_queue *txq;
-   struct gve_rx_queue *rxq;
+   uint16_t num_queues;
uint16_t i;
-   int err;
+   int ret;
 
+   num_queues = dev->data->nb_tx_queues;
priv->txqs = (struct gve_tx_queue **)dev->data->tx_queues;
-   err = gve_adminq_create_tx_queues(priv, num_queues);
-   if (err) {
-   PMD_DRV_LOG(ERR, "failed to create %u tx queues.", num_queues);
-   return err;
-   }
-   for (i = 0; i < num_queues; i++) {
-   txq = priv->txqs[i];
-   txq->qtx_tail =
-   &priv->db_bar2[rte_be_to_cpu_32(txq->qres->db_index)];
-   txq->qtx_head =
-   &priv->cnt_array[rte_be_to_cpu_32(txq->qres->counter_index)];
-
-   rte_write32(rte_cpu_to_be_32(GVE_IRQ_MASK), txq->ntfy_addr);
-   }
+   ret = gve_adminq_create_tx_queues(priv, num_queues);
+   if (ret != 0) {
+   PMD_DRV_LOG(ERR, "Failed to create %u tx queues.", num_queues);
+   return ret;
+   }
+   for (i = 0; i < num_queues; i++)
+   if (gve_tx_queu

RE: [EXT] [PATCH v2 4/4] app: add testgraph application

2023-05-08 Thread Vamsi Krishna Attunuru



> -Original Message-
> From: Sunil Kumar Kori 
> Sent: Tuesday, May 9, 2023 8:04 AM
> To: Vamsi Krishna Attunuru ; dev@dpdk.org;
> tho...@monjalon.net; Jerin Jacob Kollanukkaran 
> Cc: Vamsi Krishna Attunuru ; Nithin Kumar
> Dabilpuram 
> Subject: RE: [EXT] [PATCH v2 4/4] app: add testgraph application
> 
> Is there any user guide similar to testpmd ?
> 
   Please refer doc/guides/tools/testgraph.rst for details.

> > -Original Message-
> > From: Vamsi Attunuru 
> > Sent: Tuesday, April 25, 2023 6:45 PM
> > To: dev@dpdk.org; tho...@monjalon.net; Jerin Jacob Kollanukkaran
> > 
> > Cc: Vamsi Krishna Attunuru ; Nithin Kumar
> > Dabilpuram 
> > Subject: [EXT] [PATCH v2 4/4] app: add testgraph application
> >
> > External Email
> >
> > --
> > Patch adds test-graph application to validate graph
> > and node libraries.
> >
> > Signed-off-by: Vamsi Attunuru 
> > ---
> >  app/meson.build|1 +
> >  app/test-graph/cmdline.c   |  211 +
> >  app/test-graph/cmdline_graph.c |  294 +++
> >  app/test-graph/cmdline_graph.h |   19 +
> >  app/test-graph/meson.build |   14 +
> >  app/test-graph/parameters.c|  157 
> >  app/test-graph/testgraph.c | 1426
> 
> >  app/test-graph/testgraph.h |   91 ++
> >  doc/guides/tools/index.rst |1 +
> >  doc/guides/tools/testgraph.rst |  131 +++
> >  10 files changed, 2345 insertions(+)
> >
> > diff --git a/app/meson.build b/app/meson.build
> > index 74d2420f67..6c7b24e604 100644
> > --- a/app/meson.build
> > +++ b/app/meson.build
> > @@ -22,6 +22,7 @@ apps = [
> >  'test-eventdev',
> >  'test-fib',
> >  'test-flow-perf',
> > +'test-graph',
> >  'test-gpudev',
> >  'test-mldev',
> >  'test-pipeline',
> > diff --git a/app/test-graph/cmdline.c b/app/test-graph/cmdline.c
> > new file mode 100644
> > index 00..d9474d827a
> > --- /dev/null
> > +++ b/app/test-graph/cmdline.c
> > @@ -0,0 +1,211 @@
> > +/* SPDX-License-Identifier: BSD-3-Clause
> > + * Copyright(C) 2023 Marvell International Ltd.
> > + */
> 
> Please check copyright. I think  it should be " Copyright(C) 2023 Marvell."
> Only.
> > +
> > +#include 
> > +
> > +#include 
> > +#include 
> > +#include 
> > +#include 
> > +#include 
> > +#include 
> > +#include 
> > +#include 
> > +
> > +#include "cmdline_graph.h"
> > +#include "testgraph.h"
> > +
> > +static struct cmdline *testgraph_cl;
> > +static cmdline_parse_ctx_t *main_ctx;
> > +
> > +/* *** Help command with introduction. *** */
> > +struct cmd_help_brief_result {
> > +   cmdline_fixed_string_t help;
> > +};
> > +
> > +static void
> > +cmd_help_brief_parsed(__rte_unused void *parsed_result, struct
> cmdline
> > *cl, __rte_unused void *data)
> > +{
> > +   cmdline_printf(cl,
> > +  "\n"
> > +  "Help is available for the following sections:\n\n"
> > +  "help control: Start and stop 
> > graph walk.\n"
> > +  "help display: Displaying port, 
> > stats and
> > config "
> > +  "information.\n"
> > +  "help config : Configuration 
> > information.\n"
> > +  "help all: All of the above 
> > sections.\n\n");
> > +}
> > +
> > +static cmdline_parse_token_string_t cmd_help_brief_help =
> > +   TOKEN_STRING_INITIALIZER(struct cmd_help_brief_result, help,
> > "help");
> > +
> > +static cmdline_parse_inst_t cmd_help_brief = {
> > +   .f = cmd_help_brief_parsed,
> > +   .data = NULL,
> > +   .help_str = "help: Show help",
> > +   .tokens = {
> > +   (void *)&cmd_help_brief_help,
> > +   NULL,
> > +   },
> > +};
> > +
> > +/* *** Help command with help sections. *** */
> > +struct cmd_help_long_result {
> > +   cmdline_fixed_string_t help;
> > +   cmdline_fixed_string_t section;
> > +};
> > +
> > +static void
> > +cmd_help_long_parsed(void *parsed_result, struct cmdline *cl,
> > __rte_unused void *data)
> > +{
> > +   int show_all = 0;
> > +   struct cmd_help_long_result *res = parsed_result;
> > +
> > +   if (!strcmp(res->section, "all"))
> > +   show_all = 1;
> > +
> > +   if (show_all || !strcmp(res->section, "control")) {
> > +
> > +   cmdline_printf(cl, "\n"
> > +  "Control forwarding:\n"
> > +  "---\n\n"
> > +
> > +  "start graph_walk\n"
> > +  " Start graph_walk on worker threads.\n\n"
> > +
> > +  "stop graph_walk\n"
> > +  " Stop worker threads from running
> > graph_walk.\n\n"
> > +
> > +  "quit\n"
> > +  "Quit to prompt.\n\n");
> > +   }
> > +
> > + 

RE: [PATCH v3 5/5] config/arm: add AMD CDX

2023-05-08 Thread Gupta, Nipun
[AMD Official Use Only - General]



> -Original Message-
> From: Yigit, Ferruh 
> Sent: Monday, May 8, 2023 11:17 PM
> To: Honnappa Nagarahalli ; Gupta, Nipun
> ; dev@dpdk.org; tho...@monjalon.net;
> david.march...@redhat.com; Ruifeng Wang 
> Cc: Anand, Harpreet ; Agarwal, Nikhil
> ; nd 
> Subject: Re: [PATCH v3 5/5] config/arm: add AMD CDX
> 
> On 5/8/2023 6:16 PM, Honnappa Nagarahalli wrote:
> > 
> >
> 
>  On 4/21/2023 3:54 PM, Nipun Gupta wrote:
> > Adding support for AMD CDX devices
> >
> > Signed-off-by: Nipun Gupta 
> > ---
> >  config/arm/arm64_cdx_linux_gcc | 17 +
> >  config/arm/meson.build | 14 ++
> >  2 files changed, 31 insertions(+)
> >  create mode 100644 config/arm/arm64_cdx_linux_gcc
> >
> > diff --git a/config/arm/arm64_cdx_linux_gcc
>  b/config/arm/arm64_cdx_linux_gcc
> > new file mode 100644
> > index 00..8e6d619dae
> > --- /dev/null
> > +++ b/config/arm/arm64_cdx_linux_gcc
> > @@ -0,0 +1,17 @@
> > +[binaries]
> > +c = ['ccache', 'aarch64-linux-gnu-gcc'] cpp = ['ccache',
> > +'aarch64-linux-gnu-g++'] ar = 'aarch64-linux-gnu-ar'
> > +as = 'aarch64-linux-gnu-as'
> > +strip = 'aarch64-linux-gnu-strip'
> > +pkgconfig = 'aarch64-linux-gnu-pkg-config'
> > +pcap-config = ''
> > +
> > +[host_machine]
> > +system = 'linux'
> > +cpu_family = 'aarch64'
> > +cpu = 'armv8-a'
> > +endian = 'little'
> > +
> > +[properties]
> > +platform = 'cdx'
> > diff --git a/config/arm/meson.build b/config/arm/meson.build index
> > 5213434ca4..39b8929534 100644
> > --- a/config/arm/meson.build
> > +++ b/config/arm/meson.build
> > @@ -305,6 +305,18 @@ soc_bluefield = {
> >  'numa': false
> >  }
> >
> > +soc_cdx = {
> > +'description': 'AMD CDX',
> > +'implementer': '0x41',
> > +'part_number': '0xd42',
> > +'flags': [
> > +['RTE_MACHINE', '"cdx"'],
> > +['RTE_MAX_LCORE', 16],
> > +['RTE_MAX_NUMA_NODES', 1]
> > +],
> > +'numa': false
> > +}
> 
>  Hi Nipun,
> 
>  Why we need a new arm platform/config? Is it because of above flags?
>  If it can work with default values, I think we can drop this patch.
> >>>
> >>> Hi Ferruh,
> >>>
> >>> CDX driver works with generic ARM compilation too
> >> (arm64_armv8_linux_gcc).
> >>>
> >>> The versal platforms supporting CDX have A78 cores, and adding this
> >>> cdx config Helps to provide gcc option "march= armv8.4-a" which is for
> >>> implementer "0xd42" (ARM cortex A78 cores)., whereas for generic
> ARM
> >>> compilation "march= armv8-a".
> >>>
> >>> Maybe ARM guys can provide more information regarding if there is any
> >>> impact on using generic architecture flag (i.e. march= armv8a) on A78
> cores.
> >>>
> >>
> >> Hi Honnappa, Ruifeng,
> >>
> >> Can you please support on this question, what is the difference of
> 'march=
> >> armv8-a' flag (comparing march= armv8a)?
> > I am assuming you meant, 'march=armv8.4-a' vs 'march=armv8-a'.
> >
> 
> Yes
> 
> > 'march=armv8-a' will compile for the ARM V8.0-a ISA which is the base V8-a
> version of the ISA. This compilation ensures that the binaries can run on any
> ARM V8-a platforms.
> > 'march=armv8.4-a' will compile the binary for ARM V8.4 ISA. The binary will
> use additional (possibly instructions that improve performance) instructions
> which are not supported in ARM V8.0-a. This binary can run only on machines
> that support V8.4-a and higher. But, it will provide optimized binary for the
> target platform.
> >
> >> Should we consider adding an arm config file to support this flag?
> > We support the optimized binary through the config files. We should keep
> these config changes.
> >
> 
> Got it, thanks for the guidance.

Thanks Honnappa, Ferruh,
As discussed with Thomas too, we will add this as a separate platform patch,
with more platform details separated from CDX bus (patch sent out v4).


RE: [RFC 17/27] vhost: add control virtqueue support

2023-05-08 Thread Xia, Chenbo
Hi Maxime,

> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 17/27] vhost: add control virtqueue support
> 
> In order to support multi-queue with VDUSE, having
> control queue support in required.

in -> is

> 
> This patch adds control queue implementation, it will be
> used later when adding VDUSE support. Only split ring
> layout is supported for now, packed ring support will be
> added later.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/meson.build   |   1 +
>  lib/vhost/vhost.h   |   2 +
>  lib/vhost/virtio_net_ctrl.c | 282 
>  lib/vhost/virtio_net_ctrl.h |  10 ++
>  4 files changed, 295 insertions(+)
>  create mode 100644 lib/vhost/virtio_net_ctrl.c
>  create mode 100644 lib/vhost/virtio_net_ctrl.h
> 
> diff --git a/lib/vhost/meson.build b/lib/vhost/meson.build
> index 197a51d936..cdcd403df3 100644
> --- a/lib/vhost/meson.build
> +++ b/lib/vhost/meson.build
> @@ -28,6 +28,7 @@ sources = files(
>  'vhost_crypto.c',
>  'vhost_user.c',
>  'virtio_net.c',
> +'virtio_net_ctrl.c',
>  )
>  headers = files(
>  'rte_vdpa.h',
> diff --git a/lib/vhost/vhost.h b/lib/vhost/vhost.h
> index 8f0875b4e2..76663aed24 100644
> --- a/lib/vhost/vhost.h
> +++ b/lib/vhost/vhost.h
> @@ -525,6 +525,8 @@ struct virtio_net {
>   int postcopy_ufd;
>   int postcopy_listening;
> 
> + struct vhost_virtqueue  *cvq;
> +
>   struct rte_vdpa_device *vdpa_dev;
> 
>   /* context data for the external message handlers */
> diff --git a/lib/vhost/virtio_net_ctrl.c b/lib/vhost/virtio_net_ctrl.c
> new file mode 100644
> index 00..16ea63b42f
> --- /dev/null
> +++ b/lib/vhost/virtio_net_ctrl.c
> @@ -0,0 +1,282 @@
> +/* SPDX-License-Identifier: BSD-3-Clause
> + * Copyright (c) 2023 Red Hat, Inc.
> + */
> +
> +#undef RTE_ANNOTATE_LOCKS
> +
> +#include 
> +#include 
> +#include 
> +
> +#include "vhost.h"
> +#include "virtio_net_ctrl.h"
> +
> +struct virtio_net_ctrl {
> + uint8_t class;
> + uint8_t command;
> + uint8_t command_data[];
> +};
> +
> +struct virtio_net_ctrl_elem {
> + struct virtio_net_ctrl *ctrl_req;
> + uint16_t head_idx;
> + uint16_t n_descs;
> + uint8_t *desc_ack;
> +};
> +
> +static int
> +virtio_net_ctrl_pop(struct virtio_net *dev, struct virtio_net_ctrl_elem
> *ctrl_elem)
> +{
> + struct vhost_virtqueue *cvq = dev->cvq;
> + uint16_t avail_idx, desc_idx, n_descs = 0;
> + uint64_t desc_len, desc_addr, desc_iova, data_len = 0;
> + uint8_t *ctrl_req;
> + struct vring_desc *descs;
> +
> + avail_idx = __atomic_load_n(&cvq->avail->idx, __ATOMIC_ACQUIRE);
> + if (avail_idx == cvq->last_avail_idx) {
> + VHOST_LOG_CONFIG(dev->ifname, DEBUG, "Control queue empty\n");
> + return 0;
> + }
> +
> + desc_idx = cvq->avail->ring[cvq->last_avail_idx];
> + if (desc_idx >= cvq->size) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Out of range desc index,
> dropping\n");
> + goto err;
> + }
> +
> + ctrl_elem->head_idx = desc_idx;
> +
> + if (cvq->desc[desc_idx].flags & VRING_DESC_F_INDIRECT) {
> + desc_len = cvq->desc[desc_idx].len;
> + desc_iova = cvq->desc[desc_idx].addr;
> +
> + descs = (struct vring_desc *)(uintptr_t)vhost_iova_to_vva(dev,
> cvq,
> + desc_iova, &desc_len, VHOST_ACCESS_RO);
> + if (!descs || desc_len != cvq->desc[desc_idx].len) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to map ctrl
> indirect descs\n");
> + goto err;
> + }
> +
> + desc_idx = 0;
> + } else {
> + descs = cvq->desc;
> + }
> +
> + while (1) {
> + desc_len = descs[desc_idx].len;
> + desc_iova = descs[desc_idx].addr;
> +
> + n_descs++;
> +
> + if (descs[desc_idx].flags & VRING_DESC_F_WRITE) {
> + if (ctrl_elem->desc_ack) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR,
> + "Unexpected ctrl chain 
> layout\n");
> + goto err;
> + }
> +
> + if (desc_len != sizeof(uint8_t)) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR,
> + "Invalid ack size for ctrl req,
> dropping\n");
> + goto err;
> + }
> +
> + ctrl_elem->desc_ack = (uint8_t
> *)(uintptr_t)vhost_iova_to_vva(dev, cvq,
> +

RE: [RFC 18/27] vhost: add VDUSE device creation and destruction

2023-05-08 Thread Xia, Chenbo
> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 18/27] vhost: add VDUSE device creation and destruction
> 
> This patch adds initial support for VDUSE, which includes
> the device creation and destruction.
> 
> It does not include the virtqueues configuration, so this is
> not functionnal at this point.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/meson.build |   4 +
>  lib/vhost/socket.c|  34 +---
>  lib/vhost/vduse.c | 184 ++
>  lib/vhost/vduse.h |  33 
>  lib/vhost/vhost.h |   2 +
>  5 files changed, 245 insertions(+), 12 deletions(-)
>  create mode 100644 lib/vhost/vduse.c
>  create mode 100644 lib/vhost/vduse.h
> 
> diff --git a/lib/vhost/meson.build b/lib/vhost/meson.build
> index cdcd403df3..a57a15937f 100644
> --- a/lib/vhost/meson.build
> +++ b/lib/vhost/meson.build
> @@ -30,6 +30,10 @@ sources = files(
>  'virtio_net.c',
>  'virtio_net_ctrl.c',
>  )
> +if cc.has_header('linux/vduse.h')
> +sources += files('vduse.c')
> +cflags += '-DVHOST_HAS_VDUSE'
> +endif
>  headers = files(
>  'rte_vdpa.h',
>  'rte_vhost.h',
> diff --git a/lib/vhost/socket.c b/lib/vhost/socket.c
> index e95c3ffeac..a8a1c4cd2b 100644
> --- a/lib/vhost/socket.c
> +++ b/lib/vhost/socket.c
> @@ -18,6 +18,7 @@
>  #include 
> 
>  #include "fd_man.h"
> +#include "vduse.h"
>  #include "vhost.h"
>  #include "vhost_user.h"
> 
> @@ -35,6 +36,7 @@ struct vhost_user_socket {
>   int socket_fd;
>   struct sockaddr_un un;
>   bool is_server;
> + bool is_vduse;
>   bool reconnect;
>   bool iommu_support;
>   bool use_builtin_virtio_net;
> @@ -992,18 +994,21 @@ rte_vhost_driver_register(const char *path, uint64_t
> flags)
>  #endif
>   }
> 
> - if ((flags & RTE_VHOST_USER_CLIENT) != 0) {
> - vsocket->reconnect = !(flags & RTE_VHOST_USER_NO_RECONNECT);
> - if (vsocket->reconnect && reconn_tid == 0) {
> - if (vhost_user_reconnect_init() != 0)
> - goto out_mutex;
> - }
> + if (!strncmp("/dev/vduse/", path, strlen("/dev/vduse/"))) {
> + vsocket->is_vduse = true;
>   } else {
> - vsocket->is_server = true;
> - }
> - ret = create_unix_socket(vsocket);
> - if (ret < 0) {
> - goto out_mutex;
> + if ((flags & RTE_VHOST_USER_CLIENT) != 0) {
> + vsocket->reconnect = !(flags &
> RTE_VHOST_USER_NO_RECONNECT);
> + if (vsocket->reconnect && reconn_tid == 0) {
> + if (vhost_user_reconnect_init() != 0)
> + goto out_mutex;
> + }
> + } else {
> + vsocket->is_server = true;
> + }
> + ret = create_unix_socket(vsocket);
> + if (ret < 0)
> + goto out_mutex;
>   }
> 
>   vhost_user.vsockets[vhost_user.vsocket_cnt++] = vsocket;
> @@ -1068,7 +1073,9 @@ rte_vhost_driver_unregister(const char *path)
>   if (strcmp(vsocket->path, path))
>   continue;
> 
> - if (vsocket->is_server) {
> + if (vsocket->is_vduse) {
> + vduse_device_destroy(path);
> + } else if (vsocket->is_server) {
>   /*
>* If r/wcb is executing, release vhost_user's
>* mutex lock, and try again since the r/wcb
> @@ -1171,6 +1178,9 @@ rte_vhost_driver_start(const char *path)
>   if (!vsocket)
>   return -1;
> 
> + if (vsocket->is_vduse)
> + return vduse_device_create(path);
> +
>   if (fdset_tid == 0) {
>   /**
>* create a pipe which will be waited by poll and notified to
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> new file mode 100644
> index 00..336761c97a
> --- /dev/null
> +++ b/lib/vhost/vduse.c
> @@ -0,0 +1,184 @@
> +/* SPDX-License-Identifier: BSD-3-Clause
> + * Copyright (c) 2023 Red Hat, Inc.
> + */
> +
> +#include 
> +#include 
> +#include 
> +#include 
> +
> +
> +#include 
> +#include 
> +
> +#include 
> +#include 
> +
> +#include 
> +
> +#include "vduse.h"
> +#include "vhost.h"
> +
> +#define VHOST_VDUSE_API_VERSION 0
> +#define VDUSE_CTRL_PATH "/dev/vduse/control"
> +
> +#define VDUSE_NET_SUPPORTED_FEATURES ((1ULL << VIRTIO_NET_F_MRG_RXBUF) |
> \
> + (1ULL << VIRTIO_F_ANY_LAYOUT) | \
> + (1ULL << VIRTIO_F_VERSION_1)   | \
> + (1ULL << VIRTIO_RING_F_INDIRECT_DESC) | 

RE: [RFC 19/27] vhost: add VDUSE callback for IOTLB miss

2023-05-08 Thread Xia, Chenbo
> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 19/27] vhost: add VDUSE callback for IOTLB miss
> 
> This patch implements the VDUSE callback for IOTLB misses,
> where it unmaps the pages from the invalidated IOTLB entry.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/vduse.c | 58 +++
>  1 file changed, 58 insertions(+)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index 336761c97a..f46823f589 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -13,9 +13,11 @@
> 
>  #include 
>  #include 
> +#include 
> 
>  #include 
> 
> +#include "iotlb.h"
>  #include "vduse.h"
>  #include "vhost.h"
> 
> @@ -30,7 +32,63 @@
>   (1ULL << VIRTIO_F_IN_ORDER) | \
>   (1ULL << VIRTIO_F_IOMMU_PLATFORM))
> 
> +static int
> +vduse_iotlb_miss(struct virtio_net *dev, uint64_t iova, uint8_t perm
> __rte_unused)
> +{
> + struct vduse_iotlb_entry entry;
> + uint64_t size, page_size;
> + struct stat stat;
> + void *mmap_addr;
> + int fd, ret;
> +
> + entry.start = iova;
> + entry.last = iova + 1;
> +
> + ret = ioctl(dev->vduse_dev_fd, VDUSE_IOTLB_GET_FD, &entry);
> + if (ret < 0) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to get IOTLB entry
> for 0x%" PRIx64 "\n",
> + iova);
> + return -1;
> + }
> +
> + fd = ret;
> +
> + VHOST_LOG_CONFIG(dev->ifname, DEBUG, "New IOTLB entry:\n");
> + VHOST_LOG_CONFIG(dev->ifname, DEBUG, "\tIOVA: %" PRIx64 " - %"
> PRIx64 "\n",
> + (uint64_t)entry.start, (uint64_t)entry.last);
> + VHOST_LOG_CONFIG(dev->ifname, DEBUG, "\toffset: %" PRIx64 "\n",
> (uint64_t)entry.offset);
> + VHOST_LOG_CONFIG(dev->ifname, DEBUG, "\tfd: %d\n", fd);
> + VHOST_LOG_CONFIG(dev->ifname, DEBUG, "\tperm: %x\n", entry.perm);
> +
> + size = entry.last - entry.start + 1;
> + mmap_addr = mmap(0, size + entry.offset, entry.perm, MAP_SHARED, fd,
> 0);
> + if (!mmap_addr) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR,
> + "Failed to mmap IOTLB entry for 0x%" PRIx64 
> "\n",
> iova);
> + ret = -1;
> + goto close_fd;
> + }
> +
> + ret = fstat(fd, &stat);
> + if (ret < 0) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to get page
> size.\n");
> + munmap(mmap_addr, entry.offset + size);
> + goto close_fd;
> + }
> + page_size = (uint64_t)stat.st_blksize;
> +
> + vhost_user_iotlb_cache_insert(dev, entry.start,
> (uint64_t)(uintptr_t)mmap_addr,
> + entry.offset, size, page_size, entry.perm);
> +
> + ret = 0;
> +close_fd:
> + close(fd);
> +
> + return ret;
> +}
> +
>  static struct vhost_backend_ops vduse_backend_ops = {
> + .iotlb_miss = vduse_iotlb_miss,
>  };
> 
>  int
> --
> 2.39.2

Reviewed-by: Chenbo Xia  


RE: [RFC 20/27] vhost: add VDUSE callback for IOTLB entry removal

2023-05-08 Thread Xia, Chenbo
Hi Maxime,

> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 20/27] vhost: add VDUSE callback for IOTLB entry removal
> 
> This patch implements the VDUSE callback for IOTLB misses,

for IOTLB entry removal? This commit messages seems the same as patch 19.
You may want to change it :)

Thanks,
Chenbo

> where it unmaps the pages from the invalidated IOTLB entry
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/vduse.c | 7 +++
>  1 file changed, 7 insertions(+)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index f46823f589..ff4c9e72f1 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -32,6 +32,12 @@
>   (1ULL << VIRTIO_F_IN_ORDER) | \
>   (1ULL << VIRTIO_F_IOMMU_PLATFORM))
> 
> +static void
> +vduse_iotlb_remove_notify(uint64_t addr, uint64_t offset, uint64_t size)
> +{
> + munmap((void *)(uintptr_t)addr, offset + size);
> +}
> +
>  static int
>  vduse_iotlb_miss(struct virtio_net *dev, uint64_t iova, uint8_t perm
> __rte_unused)
>  {
> @@ -89,6 +95,7 @@ vduse_iotlb_miss(struct virtio_net *dev, uint64_t iova,
> uint8_t perm __rte_unuse
> 
>  static struct vhost_backend_ops vduse_backend_ops = {
>   .iotlb_miss = vduse_iotlb_miss,
> + .iotlb_remove_notify = vduse_iotlb_remove_notify,
>  };
> 
>  int
> --
> 2.39.2



RE: [RFC 22/27] vhost: add VDUSE events handler

2023-05-08 Thread Xia, Chenbo
> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 22/27] vhost: add VDUSE events handler
> 
> This patch makes use of Vhost lib's FD manager to install
> a handler for VDUSE events occurring on the VDUSE device FD.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/vduse.c | 102 ++
>  1 file changed, 102 insertions(+)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index afa8a39498..2a183130d3 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -17,6 +17,7 @@
> 
>  #include 
> 
> +#include "fd_man.h"
>  #include "iotlb.h"
>  #include "vduse.h"
>  #include "vhost.h"
> @@ -32,6 +33,27 @@
>   (1ULL << VIRTIO_F_IN_ORDER) | \
>   (1ULL << VIRTIO_F_IOMMU_PLATFORM))
> 
> +struct vduse {
> + struct fdset fdset;
> +};
> +
> +static struct vduse vduse = {
> + .fdset = {
> + .fd = { [0 ... MAX_FDS - 1] = {-1, NULL, NULL, NULL, 0} },
> + .fd_mutex = PTHREAD_MUTEX_INITIALIZER,
> + .fd_pooling_mutex = PTHREAD_MUTEX_INITIALIZER,
> + .num = 0
> + },
> +};
> +
> +static bool vduse_events_thread;
> +
> +static const char * const vduse_reqs_str[] = {
> + "VDUSE_GET_VQ_STATE",
> + "VDUSE_SET_STATUS",
> + "VDUSE_UPDATE_IOTLB",
> +};
> +
>  static int
>  vduse_inject_irq(struct virtio_net *dev, struct vhost_virtqueue *vq)
>  {
> @@ -105,16 +127,84 @@ static struct vhost_backend_ops vduse_backend_ops =
> {
>   .inject_irq = vduse_inject_irq,
>  };
> 
> +static void
> +vduse_events_handler(int fd, void *arg, int *remove __rte_unused)
> +{
> + struct virtio_net *dev = arg;
> + struct vduse_dev_request req;
> + struct vduse_dev_response resp;
> + int ret;
> +
> + memset(&resp, 0, sizeof(resp));
> +
> + ret = read(fd, &req, sizeof(req));
> + if (ret < 0) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to read
> request: %s\n",
> + strerror(errno));
> + return;
> + } else if (ret < (int)sizeof(req)) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Incomplete to read
> request %d\n", ret);
> + return;
> + }
> +
> + pthread_setcancelstate(PTHREAD_CANCEL_DISABLE, NULL);
> +
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "New request: %s (%u)\n",
> + req.type < RTE_DIM(vduse_reqs_str) ?
> + vduse_reqs_str[req.type] : "Unknown",
> + req.type);
> +
> + switch (req.type) {
> + default:
> + resp.result = VDUSE_REQ_RESULT_FAILED;
> + break;
> + }
> +
> + pthread_setcancelstate(PTHREAD_CANCEL_ENABLE, NULL);
> +
> + resp.request_id = req.request_id;
> +
> + ret = write(dev->vduse_dev_fd, &resp, sizeof(resp));
> + if (ret != sizeof(resp)) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to write
> response %s\n",
> + strerror(errno));
> + return;
> + }
> +}
> +
>  int
>  vduse_device_create(const char *path)
>  {
>   int control_fd, dev_fd, vid, ret;
> + pthread_t fdset_tid;
>   uint32_t i;
>   struct virtio_net *dev;
>   uint64_t ver = VHOST_VDUSE_API_VERSION;
>   struct vduse_dev_config *dev_config = NULL;
>   const char *name = path + strlen("/dev/vduse/");
> 
> + /* If first device, create events dispatcher thread */
> + if (vduse_events_thread == false) {
> + /**
> +  * create a pipe which will be waited by poll and notified to
> +  * rebuild the wait list of poll.
> +  */
> + if (fdset_pipe_init(&vduse.fdset) < 0) {
> + VHOST_LOG_CONFIG(path, ERR, "failed to create pipe for
> vduse fdset\n");
> + return -1;
> + }
> +
> + ret = rte_ctrl_thread_create(&fdset_tid, "vduse-events", NULL,
> + fdset_event_dispatch, &vduse.fdset);
> + if (ret != 0) {
> + VHOST_LOG_CONFIG(path, ERR, "failed to create vduse
> fdset handling thread\n");
> + fdset_pipe_uninit(&vduse.fdset);
> + return -1;
> + }
> +
> + vduse_events_thread = true;
> + }
> +
>   control_fd = open(VDUSE_CTRL_PATH, O_RDWR);
>   if (control_fd < 0) {
>   VHOST_LOG_CONFIG(name, ERR, "Failed to open %s: %s\n",
> @@ -198,6 +288,13 @@ vduse_device_create(const char *path)
>   }
>   }
> 
> + ret = fdset_add(&vduse.fdset, dev->vduse_dev_fd,
> vduse_events_handler, NULL, dev);
> + if (ret) {
> + 

RE: [RFC 21/27] vhost: add VDUSE callback for IRQ injection

2023-05-08 Thread Xia, Chenbo
> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 21/27] vhost: add VDUSE callback for IRQ injection
> 
> This patch implements the VDUSE callback for kicking
> virtqueues.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/vduse.c | 7 +++
>  1 file changed, 7 insertions(+)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index ff4c9e72f1..afa8a39498 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -32,6 +32,12 @@
>   (1ULL << VIRTIO_F_IN_ORDER) | \
>   (1ULL << VIRTIO_F_IOMMU_PLATFORM))
> 
> +static int
> +vduse_inject_irq(struct virtio_net *dev, struct vhost_virtqueue *vq)
> +{
> + return ioctl(dev->vduse_dev_fd, VDUSE_VQ_INJECT_IRQ, &vq->index);
> +}
> +
>  static void
>  vduse_iotlb_remove_notify(uint64_t addr, uint64_t offset, uint64_t size)
>  {
> @@ -96,6 +102,7 @@ vduse_iotlb_miss(struct virtio_net *dev, uint64_t iova,
> uint8_t perm __rte_unuse
>  static struct vhost_backend_ops vduse_backend_ops = {
>   .iotlb_miss = vduse_iotlb_miss,
>   .iotlb_remove_notify = vduse_iotlb_remove_notify,
> + .inject_irq = vduse_inject_irq,
>  };
> 
>  int
> --
> 2.39.2

Reviewed-by: Chenbo Xia  


RE: [RFC 23/27] vhost: add support for virtqueue state get event

2023-05-08 Thread Xia, Chenbo
> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 23/27] vhost: add support for virtqueue state get event
> 
> This patch adds support for VDUSE_GET_VQ_STATE event
> handling, which consists in providing the backend last
> available index for the specified virtqueue.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/vduse.c | 8 
>  1 file changed, 8 insertions(+)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index 2a183130d3..36028b7315 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -133,6 +133,7 @@ vduse_events_handler(int fd, void *arg, int *remove
> __rte_unused)
>   struct virtio_net *dev = arg;
>   struct vduse_dev_request req;
>   struct vduse_dev_response resp;
> + struct vhost_virtqueue *vq;
>   int ret;
> 
>   memset(&resp, 0, sizeof(resp));
> @@ -155,6 +156,13 @@ vduse_events_handler(int fd, void *arg, int *remove
> __rte_unused)
>   req.type);
> 
>   switch (req.type) {
> + case VDUSE_GET_VQ_STATE:
> + vq = dev->virtqueue[req.vq_state.index];
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "\tvq index: %u,
> avail_index: %u\n",
> + req.vq_state.index, vq->last_avail_idx);
> + resp.vq_state.split.avail_index = vq->last_avail_idx;
> + resp.result = VDUSE_REQ_RESULT_OK;
> + break;
>   default:
>   resp.result = VDUSE_REQ_RESULT_FAILED;
>   break;
> --
> 2.39.2

Reviewed-by: Chenbo Xia  


RE: [RFC 24/27] vhost: add support for VDUSE status set event

2023-05-08 Thread Xia, Chenbo
> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 24/27] vhost: add support for VDUSE status set event
> 
> This patch adds support for VDUSE_SET_STATUS event
> handling, which consists in updating the Virtio device
> status set by the Virtio driver.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/vduse.c | 6 ++
>  1 file changed, 6 insertions(+)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index 36028b7315..7d59a5f709 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -163,6 +163,12 @@ vduse_events_handler(int fd, void *arg, int *remove
> __rte_unused)
>   resp.vq_state.split.avail_index = vq->last_avail_idx;
>   resp.result = VDUSE_REQ_RESULT_OK;
>   break;
> + case VDUSE_SET_STATUS:
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "\tnew status: 0x%08x\n",
> + req.s.status);
> + dev->status = req.s.status;
> + resp.result = VDUSE_REQ_RESULT_OK;
> + break;
>   default:
>   resp.result = VDUSE_REQ_RESULT_FAILED;
>   break;
> --
> 2.39.2

Reviewed-by: Chenbo Xia  


RE: [RFC 25/27] vhost: add support for VDUSE IOTLB update event

2023-05-08 Thread Xia, Chenbo
Hi Maxime,

> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 25/27] vhost: add support for VDUSE IOTLB update event
> 
> This patch adds support for VDUSE_UPDATE_IOTLB event
> handling, which consists in invaliding IOTLB entries for
> the range specified in the request.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/vduse.c | 6 ++
>  1 file changed, 6 insertions(+)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index 7d59a5f709..b5b9fa2eb1 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -169,6 +169,12 @@ vduse_events_handler(int fd, void *arg, int *remove
> __rte_unused)
>   dev->status = req.s.status;
>   resp.result = VDUSE_REQ_RESULT_OK;
>   break;
> + case VDUSE_UPDATE_IOTLB:
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "\tIOVA range: %" PRIx64 "
> - %" PRIx64 "\n",
> + (uint64_t)req.iova.start, 
> (uint64_t)req.iova.last);
> + vhost_user_iotlb_cache_remove(dev, req.iova.start,
> + req.iova.last - req.iova.start + 1);
> + break;

We don't need to set the response result here?

Thanks,
Chenbo

>   default:
>   resp.result = VDUSE_REQ_RESULT_FAILED;
>   break;
> --
> 2.39.2



RE: [RFC 26/27] vhost: add VDUSE device startup

2023-05-08 Thread Xia, Chenbo
> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 26/27] vhost: add VDUSE device startup
> 
> This patch adds the device and its virtqueues
> initialization once the Virtio driver has set the DRIVER_OK
> in the Virtio status register.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/vduse.c | 118 ++
>  1 file changed, 118 insertions(+)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index b5b9fa2eb1..1cd04b4872 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -127,6 +127,120 @@ static struct vhost_backend_ops vduse_backend_ops =
> {
>   .inject_irq = vduse_inject_irq,
>  };
> 
> +static void
> +vduse_vring_setup(struct virtio_net *dev, unsigned int index)
> +{
> + struct vhost_virtqueue *vq = dev->virtqueue[index];
> + struct vhost_vring_addr *ra = &vq->ring_addrs;
> + struct vduse_vq_info vq_info;
> + struct vduse_vq_eventfd vq_efd;
> + int ret;
> +
> + vq_info.index = index;
> + ret = ioctl(dev->vduse_dev_fd, VDUSE_VQ_GET_INFO, &vq_info);
> + if (ret) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to get VQ %u
> info: %s\n",
> + index, strerror(errno));
> + return;
> + }
> +
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "VQ %u info:\n", index);
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "\tnum: %u\n", vq_info.num);
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "\tdesc_addr: %llx\n",
> vq_info.desc_addr);
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "\tdriver_addr: %llx\n",
> vq_info.driver_addr);
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "\tdevice_addr: %llx\n",
> vq_info.device_addr);
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "\tavail_idx: %u\n",
> vq_info.split.avail_index);
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "\tready: %u\n", vq_info.ready);
> +
> + vq->last_avail_idx = vq_info.split.avail_index;
> + vq->size = vq_info.num;
> + vq->ready = vq_info.ready;
> + vq->enabled = true;
> + ra->desc_user_addr = vq_info.desc_addr;
> + ra->avail_user_addr = vq_info.driver_addr;
> + ra->used_user_addr = vq_info.device_addr;
> +
> + vq->shadow_used_split = rte_malloc_socket(NULL,
> + vq->size * sizeof(struct vring_used_elem),
> + RTE_CACHE_LINE_SIZE, 0);
> + vq->batch_copy_elems = rte_malloc_socket(NULL,
> + vq->size * sizeof(struct batch_copy_elem),
> + RTE_CACHE_LINE_SIZE, 0);
> +
> + vhost_user_iotlb_rd_lock(vq);
> + if (vring_translate(dev, vq))
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to translate
> vring %d addresses\n",
> + index);
> + vhost_user_iotlb_rd_unlock(vq);
> +
> + vq->kickfd = eventfd(0, EFD_NONBLOCK | EFD_CLOEXEC);
> + if (vq->kickfd < 0) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to init kickfd for
> VQ %u: %s\n",
> + index, strerror(errno));
> + vq->kickfd = VIRTIO_INVALID_EVENTFD;
> + return;
> + }
> +
> + vq_efd.index = index;
> + vq_efd.fd = vq->kickfd;
> +
> + ret = ioctl(dev->vduse_dev_fd, VDUSE_VQ_SETUP_KICKFD, &vq_efd);
> + if (ret) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to setup kickfd for
> VQ %u: %s\n",
> + index, strerror(errno));
> + close(vq->kickfd);
> + vq->kickfd = VIRTIO_UNINITIALIZED_EVENTFD;
> + return;
> + }
> +}
> +
> +static void
> +vduse_device_start(struct virtio_net *dev)
> +{
> + unsigned int i, ret;
> +
> + dev->notify_ops = vhost_driver_callback_get(dev->ifname);
> + if (!dev->notify_ops) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR,
> + "Failed to get callback ops for driver\n");
> + return;
> + }
> +
> + ret = ioctl(dev->vduse_dev_fd, VDUSE_DEV_GET_FEATURES, &dev-
> >features);
> + if (ret) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to get
> features: %s\n",
> + strerror(errno));
> + return;
> + }
> +
> + VHOST_LOG_CONFIG(dev->ifname, INFO, "negotiated Virtio features:
> 0x%" PRIx64 "\n",
> + dev->features);
> +
> + if (dev->features &
> + ((1ULL << VIRTIO_NET_F_MRG_RXBUF) |
> +  (1ULL << VIRTIO_F_VERSION_1) |
> +  (1ULL << VIRTIO_F_RING_PACKED))) {
> + dev->vhost_hlen = sizeof(struct virtio_net_hdr_mrg_rxbuf);
> + } else {
> + dev->vhost_hlen = sizeof(struct virtio_net_hdr);
> + }
> +
> +   

RE: [RFC 27/27] vhost: add multiqueue support to VDUSE

2023-05-08 Thread Xia, Chenbo
> -Original Message-
> From: Maxime Coquelin 
> Sent: Friday, March 31, 2023 11:43 PM
> To: dev@dpdk.org; david.march...@redhat.com; Xia, Chenbo
> ; m...@redhat.com; f...@redhat.com;
> jasow...@redhat.com; Liang, Cunming ; Xie, Yongji
> ; echau...@redhat.com; epere...@redhat.com;
> amore...@redhat.com
> Cc: Maxime Coquelin 
> Subject: [RFC 27/27] vhost: add multiqueue support to VDUSE
> 
> This patch enables control queue support in order to
> support multiqueue.
> 
> Signed-off-by: Maxime Coquelin 
> ---
>  lib/vhost/vduse.c | 69 ++-
>  1 file changed, 63 insertions(+), 6 deletions(-)
> 
> diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c
> index 1cd04b4872..135e78fc35 100644
> --- a/lib/vhost/vduse.c
> +++ b/lib/vhost/vduse.c
> @@ -21,6 +21,7 @@
>  #include "iotlb.h"
>  #include "vduse.h"
>  #include "vhost.h"
> +#include "virtio_net_ctrl.h"
> 
>  #define VHOST_VDUSE_API_VERSION 0
>  #define VDUSE_CTRL_PATH "/dev/vduse/control"
> @@ -31,7 +32,9 @@
>   (1ULL << VIRTIO_RING_F_INDIRECT_DESC) | \
>   (1ULL << VIRTIO_RING_F_EVENT_IDX) | \
>   (1ULL << VIRTIO_F_IN_ORDER) | \
> - (1ULL << VIRTIO_F_IOMMU_PLATFORM))
> + (1ULL << VIRTIO_F_IOMMU_PLATFORM) | \
> + (1ULL << VIRTIO_NET_F_CTRL_VQ) | \
> + (1ULL << VIRTIO_NET_F_MQ))
> 
>  struct vduse {
>   struct fdset fdset;
> @@ -127,6 +130,25 @@ static struct vhost_backend_ops vduse_backend_ops = {
>   .inject_irq = vduse_inject_irq,
>  };
> 
> +static void
> +vduse_control_queue_event(int fd, void *arg, int *remove __rte_unused)
> +{
> + struct virtio_net *dev = arg;
> + uint64_t buf;
> + int ret;
> +
> + ret = read(fd, &buf, sizeof(buf));
> + if (ret < 0) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to read control
> queue event: %s\n",
> + strerror(errno));
> + return;
> + }
> +
> + VHOST_LOG_CONFIG(dev->ifname, DEBUG, "Control queue kicked\n");
> + if (virtio_net_ctrl_handle(dev))
> + VHOST_LOG_CONFIG(dev->ifname, ERR, "Failed to handle ctrl
> request\n");
> +}
> +
>  static void
>  vduse_vring_setup(struct virtio_net *dev, unsigned int index)
>  {
> @@ -192,6 +214,18 @@ vduse_vring_setup(struct virtio_net *dev, unsigned
> int index)
>   vq->kickfd = VIRTIO_UNINITIALIZED_EVENTFD;
>   return;
>   }
> +
> + if (vq == dev->cvq) {
> + vhost_enable_guest_notification(dev, vq, 1);
> + ret = fdset_add(&vduse.fdset, vq->kickfd,
> vduse_control_queue_event, NULL, dev);
> + if (ret) {
> + VHOST_LOG_CONFIG(dev->ifname, ERR,
> + "Failed to setup kickfd handler for
> VQ %u: %s\n",
> + index, strerror(errno));
> + close(vq->kickfd);
> + vq->kickfd = VIRTIO_UNINITIALIZED_EVENTFD;
> + }
> + }
>  }
> 
>  static void
> @@ -236,6 +270,9 @@ vduse_device_start(struct virtio_net *dev)
>   for (i = 0; i < dev->nr_vring; i++) {
>   struct vhost_virtqueue *vq = dev->virtqueue[i];
> 
> + if (vq == dev->cvq)
> + continue;
> +
>   if (dev->notify_ops->vring_state_changed)
>   dev->notify_ops->vring_state_changed(dev->vid, i, vq-
> >enabled);
>   }
> @@ -315,8 +352,9 @@ vduse_device_create(const char *path)
>  {
>   int control_fd, dev_fd, vid, ret;
>   pthread_t fdset_tid;
> - uint32_t i;
> + uint32_t i, max_queue_pairs;
>   struct virtio_net *dev;
> + struct virtio_net_config vnet_config = { 0 };
>   uint64_t ver = VHOST_VDUSE_API_VERSION;
>   struct vduse_dev_config *dev_config = NULL;
>   const char *name = path + strlen("/dev/vduse/");
> @@ -357,22 +395,33 @@ vduse_device_create(const char *path)
>   goto out_ctrl_close;
>   }
> 
> - dev_config = malloc(offsetof(struct vduse_dev_config, config));
> + dev_config = malloc(offsetof(struct vduse_dev_config, config) +
> + sizeof(vnet_config));
>   if (!dev_config) {
>   VHOST_LOG_CONFIG(name, ERR, "Failed to allocate VDUSE
> config\n");
>   ret = -1;
>   goto out_ctrl_close;
>   }
> 
> + ret = rte_vhost_driver_get_queue_num(path, &max_queue_pairs);
> + if (ret < 0) {
> + VHOST_LOG_CONFIG(name, ERR, "Failed to get max queue pairs\n");
> + goto out_free;
> + }
> +
> + VHOST_LOG_CONFIG(path, INFO, "VDUSE max queue pairs: %u\n",
> max_queue_pairs);
> +
> + vnet_config.max_virtqueue_pairs = max_queue_pairs;
>   memset(dev_config, 0, sizeof(struct vduse_dev_config));
> 
>   strncpy(dev_config->name, name, VDUSE_NAME_MAX - 1);
>  

RE: [PATCH v3 5/5] config/arm: add AMD CDX

2023-05-08 Thread Ruifeng Wang
> -Original Message-
> From: Ferruh Yigit 
> Sent: Monday, May 8, 2023 7:27 PM
> To: Gupta, Nipun ; dev@dpdk.org; tho...@monjalon.net;
> david.march...@redhat.com; Honnappa Nagarahalli 
> ; Ruifeng
> Wang 
> Cc: Anand, Harpreet ; Agarwal, Nikhil 
> 
> Subject: Re: [PATCH v3 5/5] config/arm: add AMD CDX
> 
> On 5/8/2023 11:24 AM, Gupta, Nipun wrote:
> >
> >
> >> -Original Message-
> >> From: Yigit, Ferruh 
> >> Sent: Thursday, May 4, 2023 8:59 PM
> >> To: Gupta, Nipun ; dev@dpdk.org;
> >> tho...@monjalon.net; david.march...@redhat.com
> >> Cc: Anand, Harpreet ; Agarwal, Nikhil
> >> 
> >> Subject: Re: [PATCH v3 5/5] config/arm: add AMD CDX
> >>
> >> On 4/21/2023 3:54 PM, Nipun Gupta wrote:
> >>> Adding support for AMD CDX devices
> >>>
> >>> Signed-off-by: Nipun Gupta 
> >>> ---
> >>>  config/arm/arm64_cdx_linux_gcc | 17 +
> >>>  config/arm/meson.build | 14 ++
> >>>  2 files changed, 31 insertions(+)
> >>>  create mode 100644 config/arm/arm64_cdx_linux_gcc
> >>>
> >>> diff --git a/config/arm/arm64_cdx_linux_gcc
> >> b/config/arm/arm64_cdx_linux_gcc
> >>> new file mode 100644
> >>> index 00..8e6d619dae
> >>> --- /dev/null
> >>> +++ b/config/arm/arm64_cdx_linux_gcc
> >>> @@ -0,0 +1,17 @@
> >>> +[binaries]
> >>> +c = ['ccache', 'aarch64-linux-gnu-gcc'] cpp = ['ccache',
> >>> +'aarch64-linux-gnu-g++'] ar = 'aarch64-linux-gnu-ar'
> >>> +as = 'aarch64-linux-gnu-as'
> >>> +strip = 'aarch64-linux-gnu-strip'
> >>> +pkgconfig = 'aarch64-linux-gnu-pkg-config'
> >>> +pcap-config = ''
> >>> +
> >>> +[host_machine]
> >>> +system = 'linux'
> >>> +cpu_family = 'aarch64'
> >>> +cpu = 'armv8-a'
> >>> +endian = 'little'
> >>> +
> >>> +[properties]
> >>> +platform = 'cdx'
> >>> diff --git a/config/arm/meson.build b/config/arm/meson.build index
> >>> 5213434ca4..39b8929534 100644
> >>> --- a/config/arm/meson.build
> >>> +++ b/config/arm/meson.build
> >>> @@ -305,6 +305,18 @@ soc_bluefield = {
> >>>  'numa': false
> >>>  }
> >>>
> >>> +soc_cdx = {
> >>> +'description': 'AMD CDX',
> >>> +'implementer': '0x41',
> >>> +'part_number': '0xd42',
> >>> +'flags': [
> >>> +['RTE_MACHINE', '"cdx"'],
> >>> +['RTE_MAX_LCORE', 16],
> >>> +['RTE_MAX_NUMA_NODES', 1]
> >>> +],
> >>> +'numa': false
> >>> +}
> >>
> >> Hi Nipun,
> >>
> >> Why we need a new arm platform/config? Is it because of above flags?
> >> If it can work with default values, I think we can drop this patch.
> >
> > Hi Ferruh,
> >
> > CDX driver works with generic ARM compilation too (arm64_armv8_linux_gcc).
> >
> > The versal platforms supporting CDX have A78 cores, and adding this
> > cdx config Helps to provide gcc option "march= armv8.4-a" which is for
> > implementer "0xd42" (ARM cortex A78 cores)., whereas for generic ARM
> > compilation "march= armv8-a".
> >
> > Maybe ARM guys can provide more information regarding if there is any
> > impact on using generic architecture flag (i.e. march= armv8a) on A78 cores.
> >
> 
> Hi Honnappa, Ruifeng,
> 
> Can you please support on this question, what is the difference of 'march= 
> armv8-a' flag
> (comparing march= armv8a)?
> Should we consider adding an arm config file to support this flag?

I see there is a new version without change to config file.
FWIW, native build is fine without this change. Because the specific 
(implementer, part number) flags
are already in place. What enabled by this change are options for soc build 
(-Dplatform=cdx) and
cross-build (--cross-file arm64_cdx_linux_gcc).

Regards,
Ruifeng
> 
> Thanks,
> ferruh
> 



[PATCH v6 00/15] graph enhancement for multi-core dispatch

2023-05-08 Thread Zhirun Yan
V6:
Change rte_rdtsc() to rte_rdtsc_precise().
Add union in rte_graph_param to configure models.
Remove memset in fastpath, add RTE_ASSERT for cloned graph.
Update copyright in patch 02.
Update l3fwd-graph node affinity, start from rx core successively.

V5:
Fix CI build issues about dynamically update doc.

V4:
Fix CI build issues about undefined reference of sched apis.
Remove inline for model setting.

V3:
Fix CI build issues about TLS and typo.

V2:
Use git mv to keep git history.
Use TLS for per-thread local storage.
Change model name to mcore dispatch.
Change API with specific mode name.
Split big patch.
Fix CI issues.
Rebase l3fwd-graph example.
Update doc and maintainers files.


Currently, rte_graph supports RTC (Run-To-Completion) model within each
of a single core.
RTC is one of the typical model of packet processing. Others like
Pipeline or Hybrid are lack of support.

The patch set introduces a 'multicore dispatch' model selection which
is a self-reacting scheme according to the core affinity.
The new model enables a cross-core dispatching mechanism which employs a
scheduling work-queue to dispatch streams to other worker cores which
being associated with the destination node. When core flavor of the
destination node is a default 'current', the stream can be continue
executed as normal.

Example:
3-node graph targets 3-core budget

RTC:
Graph: node-0 -> node-1 -> node-2 @Core0.

+ - - - - - - - - - - - - - - - - - - - - - +
'Core #0/1/2'
'   '
' ++ +-+ ++ '
' | Node-0 | --> | Node-1  | --> | Node-2 | '
' ++ +-+ ++ '
'   '
+ - - - - - - - - - - - - - - - - - - - - - +

Dispatch:

Graph topo: node-0 -> Core1; node-1 -> node-2; node-2 -> node-3.
Config graph: node-0 @Core0; node-1/3 @Core1; node-2 @Core2.

.. code-block:: diff

+ - - - - - -+ +- - - - - - - - - - - - - + + - - - - - -+
'  Core #0   ' '  Core #1 ' '  Core #2   '
'' '  ' ''
' ++ ' ' ++++ ' ' ++ '
' | Node-0 | - - - ->| Node-1 || Node-3 |<- - - - | Node-2 | '
' ++ ' ' ++++ ' ' ++ '
'' ' |' '  ^ '
+ - - - - - -+ +- - -|- - - - - - - - - - + + - - -|- - -+
 | |
 + - - - - - - - - - - - - - - - - +


The patch set has been break down as below:

1. Split graph worker into common and default model part.
2. Inline graph node processing to make it reusable.
3. Add set/get APIs to choose worker model.
4. Introduce core affinity API to set the node run on specific worker core.
  (only use in new model)
5. Introduce graph affinity API to bind one graph with specific worker
  core.
6. Introduce graph clone API.
7. Introduce stream moving with scheduler work-queue in patch 8~12.
8. Add stats for new models.
9. Abstract default graph config process and integrate new model into
  example/l3fwd-graph. Add new parameters for model choosing.

We could run with new worker model by this:
./dpdk-l3fwd-graph -l 8,9,10,11 -n 4 -- -p 0x1 --config="(0,0,9)" -P
--model="dispatch"

References:
https://static.sched.com/hosted_files/dpdkuserspace22/a6/graph%20introduce%20remote%20dispatch%20for%20mult-core%20scaling.pdf



Zhirun Yan (15):
  graph: rename rte_graph_work as common
  graph: split graph worker into common and default model
  graph: move node process into inline function
  graph: add get/set graph worker model APIs
  graph: introduce graph node core affinity API
  graph: introduce graph bind unbind API
  graph: introduce graph clone API for other worker core
  graph: add struct for stream moving between cores
  graph: introduce stream moving cross cores
  graph: enable create and destroy graph scheduling workqueue
  graph: introduce graph walk by cross-core dispatch
  graph: enable graph multicore dispatch scheduler model
  graph: add stats for cross-core dispatching
  examples/l3fwd-graph: introduce multicore dispatch worker model
  doc: update multicore dispatch model in graph guides

 MAINTAINERS  |   1 +
 doc/guides/prog_guide/graph_lib.rst  |  59 ++-
 examples/l3fwd-graph/main.c  | 236 +---
 lib/graph/graph.c| 179 +
 lib/graph/graph_debug.c  |   6 +
 lib/graph/graph_populate.c   |   1 +
 lib/graph/graph_private.h|  47 +++
 lib/graph/graph_stats.c  |  74 +++-
 lib/graph/meson.build|   4 +-
 lib/graph/node.c |   1 +
 lib/graph/rte_graph.h|  57 +++
 lib/graph/rte_graph_model_dispatch.c | 190 ++
 lib/graph/rte_graph_model_dispatch.h | 123 ++
 lib/g

[PATCH v6 01/15] graph: rename rte_graph_work as common

2023-05-08 Thread Zhirun Yan
Rename rte_graph_work.h to rte_graph_work_common.h for supporting
multiple graph worker model.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 MAINTAINERS | 1 +
 lib/graph/graph_pcap.c  | 2 +-
 lib/graph/graph_private.h   | 2 +-
 lib/graph/meson.build   | 2 +-
 lib/graph/{rte_graph_worker.h => rte_graph_worker_common.h} | 6 +++---
 5 files changed, 7 insertions(+), 6 deletions(-)
 rename lib/graph/{rte_graph_worker.h => rte_graph_worker_common.h} (99%)

diff --git a/MAINTAINERS b/MAINTAINERS
index 8df23e5099..cc11328242 100644
--- a/MAINTAINERS
+++ b/MAINTAINERS
@@ -1714,6 +1714,7 @@ F: doc/guides/prog_guide/bpf_lib.rst
 Graph - EXPERIMENTAL
 M: Jerin Jacob 
 M: Kiran Kumar K 
+M: Zhirun Yan 
 F: lib/graph/
 F: doc/guides/prog_guide/graph_lib.rst
 F: app/test/test_graph*
diff --git a/lib/graph/graph_pcap.c b/lib/graph/graph_pcap.c
index 6c43330029..8a220370fa 100644
--- a/lib/graph/graph_pcap.c
+++ b/lib/graph/graph_pcap.c
@@ -10,7 +10,7 @@
 #include 
 #include 
 
-#include "rte_graph_worker.h"
+#include "rte_graph_worker_common.h"
 
 #include "graph_pcap_private.h"
 
diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h
index eacdef45f0..307e5f70bc 100644
--- a/lib/graph/graph_private.h
+++ b/lib/graph/graph_private.h
@@ -13,7 +13,7 @@
 #include 
 
 #include "rte_graph.h"
-#include "rte_graph_worker.h"
+#include "rte_graph_worker_common.h"
 
 extern int rte_graph_logtype;
 
diff --git a/lib/graph/meson.build b/lib/graph/meson.build
index 3526d1b5d4..4e2b612ad3 100644
--- a/lib/graph/meson.build
+++ b/lib/graph/meson.build
@@ -16,6 +16,6 @@ sources = files(
 'graph_populate.c',
 'graph_pcap.c',
 )
-headers = files('rte_graph.h', 'rte_graph_worker.h')
+headers = files('rte_graph.h', 'rte_graph_worker_common.h')
 
 deps += ['eal', 'pcapng']
diff --git a/lib/graph/rte_graph_worker.h b/lib/graph/rte_graph_worker_common.h
similarity index 99%
rename from lib/graph/rte_graph_worker.h
rename to lib/graph/rte_graph_worker_common.h
index 438595b15c..0bad2938f3 100644
--- a/lib/graph/rte_graph_worker.h
+++ b/lib/graph/rte_graph_worker_common.h
@@ -2,8 +2,8 @@
  * Copyright(C) 2020 Marvell International Ltd.
  */
 
-#ifndef _RTE_GRAPH_WORKER_H_
-#define _RTE_GRAPH_WORKER_H_
+#ifndef _RTE_GRAPH_WORKER_COMMON_H_
+#define _RTE_GRAPH_WORKER_COMMON_H_
 
 /**
  * @file rte_graph_worker.h
@@ -518,4 +518,4 @@ rte_node_next_stream_move(struct rte_graph *graph, struct 
rte_node *src,
 }
 #endif
 
-#endif /* _RTE_GRAPH_WORKER_H_ */
+#endif /* _RTE_GRAPH_WORKER_COIMMON_H_ */
-- 
2.37.2



[PATCH v6 02/15] graph: split graph worker into common and default model

2023-05-08 Thread Zhirun Yan
To support multiple graph worker model, split graph into common
and default. Naming the current walk function as rte_graph_model_rtc
cause the default model is RTC(Run-to-completion).

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/graph_pcap.c  |  2 +-
 lib/graph/graph_private.h   |  2 +-
 lib/graph/meson.build   |  2 +-
 lib/graph/rte_graph_model_rtc.h | 62 +
 lib/graph/rte_graph_worker.h| 35 
 lib/graph/rte_graph_worker_common.h | 57 --
 6 files changed, 100 insertions(+), 60 deletions(-)
 create mode 100644 lib/graph/rte_graph_model_rtc.h
 create mode 100644 lib/graph/rte_graph_worker.h

diff --git a/lib/graph/graph_pcap.c b/lib/graph/graph_pcap.c
index 8a220370fa..6c43330029 100644
--- a/lib/graph/graph_pcap.c
+++ b/lib/graph/graph_pcap.c
@@ -10,7 +10,7 @@
 #include 
 #include 
 
-#include "rte_graph_worker_common.h"
+#include "rte_graph_worker.h"
 
 #include "graph_pcap_private.h"
 
diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h
index 307e5f70bc..eacdef45f0 100644
--- a/lib/graph/graph_private.h
+++ b/lib/graph/graph_private.h
@@ -13,7 +13,7 @@
 #include 
 
 #include "rte_graph.h"
-#include "rte_graph_worker_common.h"
+#include "rte_graph_worker.h"
 
 extern int rte_graph_logtype;
 
diff --git a/lib/graph/meson.build b/lib/graph/meson.build
index 4e2b612ad3..3526d1b5d4 100644
--- a/lib/graph/meson.build
+++ b/lib/graph/meson.build
@@ -16,6 +16,6 @@ sources = files(
 'graph_populate.c',
 'graph_pcap.c',
 )
-headers = files('rte_graph.h', 'rte_graph_worker_common.h')
+headers = files('rte_graph.h', 'rte_graph_worker.h')
 
 deps += ['eal', 'pcapng']
diff --git a/lib/graph/rte_graph_model_rtc.h b/lib/graph/rte_graph_model_rtc.h
new file mode 100644
index 00..10b359772f
--- /dev/null
+++ b/lib/graph/rte_graph_model_rtc.h
@@ -0,0 +1,62 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2020 Marvell International Ltd.
+ * Copyright(C) 2023 Intel Corporation
+ */
+
+#include "rte_graph_worker_common.h"
+
+/**
+ * Perform graph walk on the circular buffer and invoke the process function
+ * of the nodes and collect the stats.
+ *
+ * @param graph
+ *   Graph pointer returned from rte_graph_lookup function.
+ *
+ * @see rte_graph_lookup()
+ */
+static inline void
+rte_graph_walk_rtc(struct rte_graph *graph)
+{
+   const rte_graph_off_t *cir_start = graph->cir_start;
+   const rte_node_t mask = graph->cir_mask;
+   uint32_t head = graph->head;
+   struct rte_node *node;
+   uint64_t start;
+   uint16_t rc;
+   void **objs;
+
+   /*
+* Walk on the source node(s) ((cir_start - head) -> cir_start) and then
+* on the pending streams (cir_start -> (cir_start + mask) -> cir_start)
+* in a circular buffer fashion.
+*
+*  +-+ <= cir_start - head [number of source nodes]
+*  | |
+*  | ... | <= source nodes
+*  | |
+*  +-+ <= cir_start [head = 0] [tail = 0]
+*  | |
+*  | ... | <= pending streams
+*  | |
+*  +-+ <= cir_start + mask
+*/
+   while (likely(head != graph->tail)) {
+   node = (struct rte_node *)RTE_PTR_ADD(graph, 
cir_start[(int32_t)head++]);
+   RTE_ASSERT(node->fence == RTE_GRAPH_FENCE);
+   objs = node->objs;
+   rte_prefetch0(objs);
+
+   if (rte_graph_has_stats_feature()) {
+   start = rte_rdtsc();
+   rc = node->process(graph, node, objs, node->idx);
+   node->total_cycles += rte_rdtsc() - start;
+   node->total_calls++;
+   node->total_objs += rc;
+   } else {
+   node->process(graph, node, objs, node->idx);
+   }
+   node->idx = 0;
+   head = likely((int32_t)head > 0) ? head & mask : head;
+   }
+   graph->tail = 0;
+}
diff --git a/lib/graph/rte_graph_worker.h b/lib/graph/rte_graph_worker.h
new file mode 100644
index 00..5b58f7bda9
--- /dev/null
+++ b/lib/graph/rte_graph_worker.h
@@ -0,0 +1,35 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2020 Marvell International Ltd.
+ * Copyright(C) 2023 Intel Corporation
+ */
+
+#ifndef _RTE_GRAPH_WORKER_H_
+#define _RTE_GRAPH_WORKER_H_
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+
+#include "rte_graph_model_rtc.h"
+
+/**
+ * Perform graph walk on the circular buffer and invoke the process function
+ * of the nodes and collect the stats.
+ *
+ * @param graph
+ *   Graph pointer returned from rte_graph_lookup function.
+ *
+ * @see rte_graph_lookup()
+ */
+__rte_experimental
+static inline void
+rte_graph_walk(struct rte_graph *graph)
+{
+   rte_graph_walk_rtc(graph)

[PATCH v6 03/15] graph: move node process into inline function

2023-05-08 Thread Zhirun Yan
Node process is a single and reusable block, move the code into an inline
function.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/rte_graph_model_rtc.h | 20 ++---
 lib/graph/rte_graph_worker_common.h | 33 +
 2 files changed, 35 insertions(+), 18 deletions(-)

diff --git a/lib/graph/rte_graph_model_rtc.h b/lib/graph/rte_graph_model_rtc.h
index 10b359772f..4b6236e301 100644
--- a/lib/graph/rte_graph_model_rtc.h
+++ b/lib/graph/rte_graph_model_rtc.h
@@ -21,9 +21,6 @@ rte_graph_walk_rtc(struct rte_graph *graph)
const rte_node_t mask = graph->cir_mask;
uint32_t head = graph->head;
struct rte_node *node;
-   uint64_t start;
-   uint16_t rc;
-   void **objs;
 
/*
 * Walk on the source node(s) ((cir_start - head) -> cir_start) and then
@@ -42,21 +39,8 @@ rte_graph_walk_rtc(struct rte_graph *graph)
 */
while (likely(head != graph->tail)) {
node = (struct rte_node *)RTE_PTR_ADD(graph, 
cir_start[(int32_t)head++]);
-   RTE_ASSERT(node->fence == RTE_GRAPH_FENCE);
-   objs = node->objs;
-   rte_prefetch0(objs);
-
-   if (rte_graph_has_stats_feature()) {
-   start = rte_rdtsc();
-   rc = node->process(graph, node, objs, node->idx);
-   node->total_cycles += rte_rdtsc() - start;
-   node->total_calls++;
-   node->total_objs += rc;
-   } else {
-   node->process(graph, node, objs, node->idx);
-   }
-   node->idx = 0;
-   head = likely((int32_t)head > 0) ? head & mask : head;
+   __rte_node_process(graph, node);
+   head = likely((int32_t)head > 0) ? head & mask : head;
}
graph->tail = 0;
 }
diff --git a/lib/graph/rte_graph_worker_common.h 
b/lib/graph/rte_graph_worker_common.h
index b58f8f6947..e25eabc81f 100644
--- a/lib/graph/rte_graph_worker_common.h
+++ b/lib/graph/rte_graph_worker_common.h
@@ -130,6 +130,39 @@ void __rte_node_stream_alloc_size(struct rte_graph *graph,
 
 /* Fast path helper functions */
 
+/**
+ * @internal
+ *
+ * Enqueue a given node to the tail of the graph reel.
+ *
+ * @param graph
+ *   Pointer Graph object.
+ * @param node
+ *   Pointer to node object to be enqueued.
+ */
+static __rte_always_inline void
+__rte_node_process(struct rte_graph *graph, struct rte_node *node)
+{
+   uint64_t start;
+   uint16_t rc;
+   void **objs;
+
+   RTE_ASSERT(node->fence == RTE_GRAPH_FENCE);
+   objs = node->objs;
+   rte_prefetch0(objs);
+
+   if (rte_graph_has_stats_feature()) {
+   start = rte_rdtsc_precise();
+   rc = node->process(graph, node, objs, node->idx);
+   node->total_cycles += rte_rdtsc_precise() - start;
+   node->total_calls++;
+   node->total_objs += rc;
+   } else {
+   node->process(graph, node, objs, node->idx);
+   }
+   node->idx = 0;
+}
+
 /**
  * @internal
  *
-- 
2.37.2



[PATCH v6 04/15] graph: add get/set graph worker model APIs

2023-05-08 Thread Zhirun Yan
Add new get/set APIs to configure graph worker model which is used to
determine which model will be chosen.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/meson.build   |  1 +
 lib/graph/rte_graph_worker.c| 54 +
 lib/graph/rte_graph_worker_common.h | 19 ++
 lib/graph/version.map   |  3 ++
 4 files changed, 77 insertions(+)
 create mode 100644 lib/graph/rte_graph_worker.c

diff --git a/lib/graph/meson.build b/lib/graph/meson.build
index 3526d1b5d4..9fab8243da 100644
--- a/lib/graph/meson.build
+++ b/lib/graph/meson.build
@@ -15,6 +15,7 @@ sources = files(
 'graph_stats.c',
 'graph_populate.c',
 'graph_pcap.c',
+'rte_graph_worker.c',
 )
 headers = files('rte_graph.h', 'rte_graph_worker.h')
 
diff --git a/lib/graph/rte_graph_worker.c b/lib/graph/rte_graph_worker.c
new file mode 100644
index 00..cabc101262
--- /dev/null
+++ b/lib/graph/rte_graph_worker.c
@@ -0,0 +1,54 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2023 Intel Corporation
+ */
+
+#include "rte_graph_worker_common.h"
+
+RTE_DEFINE_PER_LCORE(enum rte_graph_worker_model, worker_model) = 
RTE_GRAPH_MODEL_DEFAULT;
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: this API may change, or be removed, without prior notice
+ * Set the graph worker model
+ *
+ * @note This function does not perform any locking, and is only safe to call
+ *before graph running.
+ *
+ * @param name
+ *   Name of the graph worker model.
+ *
+ * @return
+ *   0 on success, -1 otherwise.
+ */
+int
+rte_graph_worker_model_set(enum rte_graph_worker_model model)
+{
+   if (model >= RTE_GRAPH_MODEL_LIST_END)
+   goto fail;
+
+   RTE_PER_LCORE(worker_model) = model;
+   return 0;
+
+fail:
+   RTE_PER_LCORE(worker_model) = RTE_GRAPH_MODEL_DEFAULT;
+   return -1;
+}
+
+/**
+ * @warning
+ * @b EXPERIMENTAL: this API may change, or be removed, without prior notice
+ *
+ * Get the graph worker model
+ *
+ * @param name
+ *   Name of the graph worker model.
+ *
+ * @return
+ *   Graph worker model on success.
+ */
+inline
+enum rte_graph_worker_model
+rte_graph_worker_model_get(void)
+{
+   return RTE_PER_LCORE(worker_model);
+}
diff --git a/lib/graph/rte_graph_worker_common.h 
b/lib/graph/rte_graph_worker_common.h
index e25eabc81f..9bde8856ae 100644
--- a/lib/graph/rte_graph_worker_common.h
+++ b/lib/graph/rte_graph_worker_common.h
@@ -19,6 +19,7 @@
 #include 
 #include 
 #include 
+#include 
 #include 
 #include 
 #include 
@@ -95,6 +96,16 @@ struct rte_node {
struct rte_node *nodes[] __rte_cache_min_aligned; /**< Next nodes. */
 } __rte_cache_aligned;
 
+/** Graph worker models */
+enum rte_graph_worker_model {
+   RTE_GRAPH_MODEL_DEFAULT,
+   RTE_GRAPH_MODEL_RTC = RTE_GRAPH_MODEL_DEFAULT,
+   RTE_GRAPH_MODEL_MCORE_DISPATCH,
+   RTE_GRAPH_MODEL_LIST_END
+};
+
+RTE_DECLARE_PER_LCORE(enum rte_graph_worker_model, worker_model);
+
 /**
  * @internal
  *
@@ -490,6 +501,14 @@ rte_node_next_stream_move(struct rte_graph *graph, struct 
rte_node *src,
}
 }
 
+__rte_experimental
+enum rte_graph_worker_model
+rte_graph_worker_model_get(void);
+
+__rte_experimental
+int
+rte_graph_worker_model_set(enum rte_graph_worker_model model);
+
 #ifdef __cplusplus
 }
 #endif
diff --git a/lib/graph/version.map b/lib/graph/version.map
index 13b838752d..eea73ec9ca 100644
--- a/lib/graph/version.map
+++ b/lib/graph/version.map
@@ -43,5 +43,8 @@ EXPERIMENTAL {
rte_node_next_stream_put;
rte_node_next_stream_move;
 
+   rte_graph_worker_model_set;
+   rte_graph_worker_model_get;
+
local: *;
 };
-- 
2.37.2



[PATCH v6 05/15] graph: introduce graph node core affinity API

2023-05-08 Thread Zhirun Yan
Add lcore_id for node to hold affinity core id and impl
rte_graph_model_dispatch_lcore_affinity_set to set node affinity
with specific lcore.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/graph_private.h|  1 +
 lib/graph/meson.build|  1 +
 lib/graph/node.c |  1 +
 lib/graph/rte_graph_model_dispatch.c | 30 +++
 lib/graph/rte_graph_model_dispatch.h | 43 
 lib/graph/version.map|  2 ++
 6 files changed, 78 insertions(+)
 create mode 100644 lib/graph/rte_graph_model_dispatch.c
 create mode 100644 lib/graph/rte_graph_model_dispatch.h

diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h
index eacdef45f0..bd4c576324 100644
--- a/lib/graph/graph_private.h
+++ b/lib/graph/graph_private.h
@@ -51,6 +51,7 @@ struct node {
STAILQ_ENTRY(node) next;  /**< Next node in the list. */
char name[RTE_NODE_NAMESIZE]; /**< Name of the node. */
uint64_t flags;   /**< Node configuration flag. */
+   unsigned int lcore_id;/**< Node runs on the Lcore ID */
rte_node_process_t process;   /**< Node process function. */
rte_node_init_t init; /**< Node init function. */
rte_node_fini_t fini; /**< Node fini function. */
diff --git a/lib/graph/meson.build b/lib/graph/meson.build
index 9fab8243da..c729d984b6 100644
--- a/lib/graph/meson.build
+++ b/lib/graph/meson.build
@@ -16,6 +16,7 @@ sources = files(
 'graph_populate.c',
 'graph_pcap.c',
 'rte_graph_worker.c',
+'rte_graph_model_dispatch.c',
 )
 headers = files('rte_graph.h', 'rte_graph_worker.h')
 
diff --git a/lib/graph/node.c b/lib/graph/node.c
index 149414dcd9..339b4a0da5 100644
--- a/lib/graph/node.c
+++ b/lib/graph/node.c
@@ -100,6 +100,7 @@ __rte_node_register(const struct rte_node_register *reg)
goto free;
}
 
+   node->lcore_id = RTE_MAX_LCORE;
node->id = node_id++;
 
/* Add the node at tail */
diff --git a/lib/graph/rte_graph_model_dispatch.c 
b/lib/graph/rte_graph_model_dispatch.c
new file mode 100644
index 00..3364a76ed4
--- /dev/null
+++ b/lib/graph/rte_graph_model_dispatch.c
@@ -0,0 +1,30 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2023 Intel Corporation
+ */
+
+#include "graph_private.h"
+#include "rte_graph_model_dispatch.h"
+
+int
+rte_graph_model_dispatch_lcore_affinity_set(const char *name, unsigned int 
lcore_id)
+{
+   struct node *node;
+   int ret = -EINVAL;
+
+   if (lcore_id >= RTE_MAX_LCORE)
+   return ret;
+
+   graph_spinlock_lock();
+
+   STAILQ_FOREACH(node, node_list_head_get(), next) {
+   if (strncmp(node->name, name, RTE_NODE_NAMESIZE) == 0) {
+   node->lcore_id = lcore_id;
+   ret = 0;
+   break;
+   }
+   }
+
+   graph_spinlock_unlock();
+
+   return ret;
+}
diff --git a/lib/graph/rte_graph_model_dispatch.h 
b/lib/graph/rte_graph_model_dispatch.h
new file mode 100644
index 00..179624e972
--- /dev/null
+++ b/lib/graph/rte_graph_model_dispatch.h
@@ -0,0 +1,43 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(C) 2023 Intel Corporation
+ */
+
+#ifndef _RTE_GRAPH_MODEL_DISPATCH_H_
+#define _RTE_GRAPH_MODEL_DISPATCH_H_
+
+/**
+ * @file rte_graph_model_dispatch.h
+ *
+ * @warning
+ * @b EXPERIMENTAL:
+ * All functions in this file may be changed or removed without prior notice.
+ *
+ * This API allows to set core affinity with the node.
+ */
+#include "rte_graph_worker_common.h"
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+
+/**
+ * Set lcore affinity with the node.
+ *
+ * @param name
+ *   Valid node name. In the case of the cloned node, the name will be
+ * "parent node name" + "-" + name.
+ * @param lcore_id
+ *   The lcore ID value.
+ *
+ * @return
+ *   0 on success, error otherwise.
+ */
+__rte_experimental
+int rte_graph_model_dispatch_lcore_affinity_set(const char *name,
+   unsigned int lcore_id);
+
+#ifdef __cplusplus
+}
+#endif
+
+#endif /* _RTE_GRAPH_MODEL_DISPATCH_H_ */
diff --git a/lib/graph/version.map b/lib/graph/version.map
index eea73ec9ca..1f090be74e 100644
--- a/lib/graph/version.map
+++ b/lib/graph/version.map
@@ -46,5 +46,7 @@ EXPERIMENTAL {
rte_graph_worker_model_set;
rte_graph_worker_model_get;
 
+   rte_graph_model_dispatch_lcore_affinity_set;
+
local: *;
 };
-- 
2.37.2



[PATCH v6 06/15] graph: introduce graph bind unbind API

2023-05-08 Thread Zhirun Yan
Add lcore_id for graph to hold affinity core id where graph would run on.
Add bind/unbind API to set/unset graph affinity attribute. lcore_id will
be set as MAX by default, it means not enable this attribute.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/graph.c | 59 +++
 lib/graph/graph_private.h |  2 ++
 lib/graph/rte_graph.h | 22 +++
 lib/graph/version.map |  2 ++
 4 files changed, 85 insertions(+)

diff --git a/lib/graph/graph.c b/lib/graph/graph.c
index 5582631b53..b8ef86da45 100644
--- a/lib/graph/graph.c
+++ b/lib/graph/graph.c
@@ -260,6 +260,64 @@ graph_mem_fixup_secondary(struct rte_graph *graph)
return graph_mem_fixup_node_ctx(graph);
 }
 
+static __rte_always_inline bool
+graph_src_node_avail(struct graph *graph)
+{
+   struct graph_node *graph_node;
+
+   STAILQ_FOREACH(graph_node, &graph->node_list, next)
+   if ((graph_node->node->flags & RTE_NODE_SOURCE_F) &&
+   (graph_node->node->lcore_id == RTE_MAX_LCORE ||
+graph->lcore_id == graph_node->node->lcore_id))
+   return true;
+
+   return false;
+}
+
+int
+rte_graph_model_dispatch_core_bind(rte_graph_t id, int lcore)
+{
+   struct graph *graph;
+
+   GRAPH_ID_CHECK(id);
+   if (!rte_lcore_is_enabled(lcore))
+   SET_ERR_JMP(ENOLINK, fail,
+   "lcore %d not enabled\n",
+   lcore);
+
+   STAILQ_FOREACH(graph, &graph_list, next)
+   if (graph->id == id)
+   break;
+
+   graph->lcore_id = lcore;
+   graph->socket = rte_lcore_to_socket_id(lcore);
+
+   /* check the availability of source node */
+   if (!graph_src_node_avail(graph))
+   graph->graph->head = 0;
+
+   return 0;
+
+fail:
+   return -rte_errno;
+}
+
+void
+rte_graph_model_dispatch_core_unbind(rte_graph_t id)
+{
+   struct graph *graph;
+
+   GRAPH_ID_CHECK(id);
+   STAILQ_FOREACH(graph, &graph_list, next)
+   if (graph->id == id)
+   break;
+
+   graph->lcore_id = RTE_MAX_LCORE;
+
+fail:
+   return;
+}
+
 struct rte_graph *
 rte_graph_lookup(const char *name)
 {
@@ -346,6 +404,7 @@ rte_graph_create(const char *name, struct rte_graph_param 
*prm)
graph->src_node_count = src_node_count;
graph->node_count = graph_nodes_count(graph);
graph->id = graph_id;
+   graph->lcore_id = RTE_MAX_LCORE;
graph->num_pkt_to_capture = prm->num_pkt_to_capture;
if (prm->pcap_filename)
rte_strscpy(graph->pcap_filename, prm->pcap_filename, 
RTE_GRAPH_PCAP_FILE_SZ);
diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h
index bd4c576324..f63b339d81 100644
--- a/lib/graph/graph_private.h
+++ b/lib/graph/graph_private.h
@@ -99,6 +99,8 @@ struct graph {
/**< Circular buffer mask for wrap around. */
rte_graph_t id;
/**< Graph identifier. */
+   unsigned int lcore_id;
+   /**< Lcore identifier where the graph prefer to run on. */
size_t mem_sz;
/**< Memory size of the graph. */
int socket;
diff --git a/lib/graph/rte_graph.h b/lib/graph/rte_graph.h
index c9a77297fc..c523809d1f 100644
--- a/lib/graph/rte_graph.h
+++ b/lib/graph/rte_graph.h
@@ -285,6 +285,28 @@ char *rte_graph_id_to_name(rte_graph_t id);
 __rte_experimental
 int rte_graph_export(const char *name, FILE *f);
 
+/**
+ * Bind graph with specific lcore
+ *
+ * @param id
+ *   Graph id to get the pointer of graph object
+ * @param lcore
+ * The lcore where the graph will run on
+ * @return
+ *   0 on success, error otherwise.
+ */
+__rte_experimental
+int rte_graph_model_dispatch_core_bind(rte_graph_t id, int lcore);
+
+/**
+ * Unbind graph with lcore
+ *
+ * @param id
+ * Graph id to get the pointer of graph object
+ */
+__rte_experimental
+void rte_graph_model_dispatch_core_unbind(rte_graph_t id);
+
 /**
  * Get graph object from its name.
  *
diff --git a/lib/graph/version.map b/lib/graph/version.map
index 1f090be74e..7de6f08f59 100644
--- a/lib/graph/version.map
+++ b/lib/graph/version.map
@@ -18,6 +18,8 @@ EXPERIMENTAL {
rte_graph_node_get_by_name;
rte_graph_obj_dump;
rte_graph_walk;
+   rte_graph_model_dispatch_core_bind;
+   rte_graph_model_dispatch_core_unbind;
 
rte_graph_cluster_stats_create;
rte_graph_cluster_stats_destroy;
-- 
2.37.2



[PATCH v6 07/15] graph: introduce graph clone API for other worker core

2023-05-08 Thread Zhirun Yan
This patch adds graph API for supporting to clone the graph object for
a specified worker core. The new graph will also clone all nodes.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/graph.c | 110 ++
 lib/graph/graph_private.h |   2 +
 lib/graph/rte_graph.h |  20 +++
 lib/graph/version.map |   1 +
 4 files changed, 133 insertions(+)

diff --git a/lib/graph/graph.c b/lib/graph/graph.c
index b8ef86da45..2629c79103 100644
--- a/lib/graph/graph.c
+++ b/lib/graph/graph.c
@@ -404,6 +404,7 @@ rte_graph_create(const char *name, struct rte_graph_param 
*prm)
graph->src_node_count = src_node_count;
graph->node_count = graph_nodes_count(graph);
graph->id = graph_id;
+   graph->parent_id = RTE_GRAPH_ID_INVALID;
graph->lcore_id = RTE_MAX_LCORE;
graph->num_pkt_to_capture = prm->num_pkt_to_capture;
if (prm->pcap_filename)
@@ -468,6 +469,115 @@ rte_graph_destroy(rte_graph_t id)
return rc;
 }
 
+static int
+clone_name(struct graph *graph, struct graph *parent_graph, const char *name)
+{
+   ssize_t sz, rc;
+
+#define SZ RTE_GRAPH_NAMESIZE
+   rc = rte_strscpy(graph->name, parent_graph->name, SZ);
+   if (rc < 0)
+   goto fail;
+   sz = rc;
+   rc = rte_strscpy(graph->name + sz, "-", RTE_MAX((int16_t)(SZ - sz), 0));
+   if (rc < 0)
+   goto fail;
+   sz += rc;
+   sz = rte_strscpy(graph->name + sz, name, RTE_MAX((int16_t)(SZ - sz), 
0));
+   if (sz < 0)
+   goto fail;
+
+   return 0;
+fail:
+   rte_errno = E2BIG;
+   return -rte_errno;
+}
+
+static rte_graph_t
+graph_clone(struct graph *parent_graph, const char *name)
+{
+   struct graph_node *graph_node;
+   struct graph *graph;
+
+   graph_spinlock_lock();
+
+   /* Don't allow to clone a node from a cloned graph */
+   if (parent_graph->parent_id != RTE_GRAPH_ID_INVALID)
+   SET_ERR_JMP(EEXIST, fail, "A cloned graph is not allowed to be 
cloned");
+
+   /* Create graph object */
+   graph = calloc(1, sizeof(*graph));
+   if (graph == NULL)
+   SET_ERR_JMP(ENOMEM, fail, "Failed to calloc cloned graph 
object");
+
+   /* Naming ceremony of the new graph. name is node->name + "-" + name */
+   if (clone_name(graph, parent_graph, name))
+   goto free;
+
+   /* Check for existence of duplicate graph */
+   if (rte_graph_from_name(graph->name) != RTE_GRAPH_ID_INVALID)
+   SET_ERR_JMP(EEXIST, free, "Found duplicate graph %s",
+   graph->name);
+
+   /* Clone nodes from parent graph firstly */
+   STAILQ_INIT(&graph->node_list);
+   STAILQ_FOREACH(graph_node, &parent_graph->node_list, next) {
+   if (graph_node_add(graph, graph_node->node))
+   goto graph_cleanup;
+   }
+
+   /* Just update adjacency list of all nodes in the graph */
+   if (graph_adjacency_list_update(graph))
+   goto graph_cleanup;
+
+   /* Initialize the graph object */
+   graph->src_node_count = parent_graph->src_node_count;
+   graph->node_count = parent_graph->node_count;
+   graph->parent_id = parent_graph->id;
+   graph->lcore_id = parent_graph->lcore_id;
+   graph->socket = parent_graph->socket;
+   graph->id = graph_id;
+
+   /* Allocate the Graph fast path memory and populate the data */
+   if (graph_fp_mem_create(graph))
+   goto graph_cleanup;
+
+   /* Call init() of the all the nodes in the graph */
+   if (graph_node_init(graph))
+   goto graph_mem_destroy;
+
+   /* All good, Lets add the graph to the list */
+   graph_id++;
+   STAILQ_INSERT_TAIL(&graph_list, graph, next);
+
+   graph_spinlock_unlock();
+   return graph->id;
+
+graph_mem_destroy:
+   graph_fp_mem_destroy(graph);
+graph_cleanup:
+   graph_cleanup(graph);
+free:
+   free(graph);
+fail:
+   graph_spinlock_unlock();
+   return RTE_GRAPH_ID_INVALID;
+}
+
+rte_graph_t
+rte_graph_clone(rte_graph_t id, const char *name)
+{
+   struct graph *graph;
+
+   GRAPH_ID_CHECK(id);
+   STAILQ_FOREACH(graph, &graph_list, next)
+   if (graph->id == id)
+   return graph_clone(graph, name);
+
+fail:
+   return RTE_GRAPH_ID_INVALID;
+}
+
 rte_graph_t
 rte_graph_from_name(const char *name)
 {
diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h
index f63b339d81..52ca30ed56 100644
--- a/lib/graph/graph_private.h
+++ b/lib/graph/graph_private.h
@@ -99,6 +99,8 @@ struct graph {
/**< Circular buffer mask for wrap around. */
rte_graph_t id;
/**< Graph identifier. */
+   rte_graph_t parent_id;
+   /**< Parent graph identifier. */
unsigned int lcore_id;
/**< Lcore identifier where the graph prefer to run on. *

[PATCH v6 08/15] graph: add struct for stream moving between cores

2023-05-08 Thread Zhirun Yan
Add graph_sched_wq_node to hold graph scheduling workqueue
node.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/graph.c   |  1 +
 lib/graph/graph_populate.c  |  1 +
 lib/graph/graph_private.h   | 12 
 lib/graph/rte_graph_worker_common.h | 21 +
 4 files changed, 35 insertions(+)

diff --git a/lib/graph/graph.c b/lib/graph/graph.c
index 2629c79103..e809aa55b0 100644
--- a/lib/graph/graph.c
+++ b/lib/graph/graph.c
@@ -290,6 +290,7 @@ rte_graph_model_dispatch_core_bind(rte_graph_t id, int 
lcore)
break;
 
graph->lcore_id = lcore;
+   graph->graph->lcore_id = graph->lcore_id;
graph->socket = rte_lcore_to_socket_id(lcore);
 
/* check the availability of source node */
diff --git a/lib/graph/graph_populate.c b/lib/graph/graph_populate.c
index 2c0844ce92..7dcf1420c1 100644
--- a/lib/graph/graph_populate.c
+++ b/lib/graph/graph_populate.c
@@ -89,6 +89,7 @@ graph_nodes_populate(struct graph *_graph)
}
node->id = graph_node->node->id;
node->parent_id = pid;
+   node->lcore_id = graph_node->node->lcore_id;
nb_edges = graph_node->node->nb_edges;
node->nb_edges = nb_edges;
off += sizeof(struct rte_node);
diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h
index 52ca30ed56..02b10ea2b6 100644
--- a/lib/graph/graph_private.h
+++ b/lib/graph/graph_private.h
@@ -61,6 +61,18 @@ struct node {
char next_nodes[][RTE_NODE_NAMESIZE]; /**< Names of next nodes. */
 };
 
+/**
+ * @internal
+ *
+ * Structure that holds the graph scheduling workqueue node stream.
+ * Used for mcore dispatch model.
+ */
+struct graph_sched_wq_node {
+   rte_graph_off_t node_off;
+   uint16_t nb_objs;
+   void *objs[RTE_GRAPH_BURST_SIZE];
+} __rte_cache_aligned;
+
 /**
  * @internal
  *
diff --git a/lib/graph/rte_graph_worker_common.h 
b/lib/graph/rte_graph_worker_common.h
index 9bde8856ae..8e968e2022 100644
--- a/lib/graph/rte_graph_worker_common.h
+++ b/lib/graph/rte_graph_worker_common.h
@@ -30,6 +30,13 @@
 extern "C" {
 #endif
 
+/**
+ * @internal
+ *
+ * Singly-linked list head for graph schedule run-queue.
+ */
+SLIST_HEAD(rte_graph_rq_head, rte_graph);
+
 /**
  * @internal
  *
@@ -41,6 +48,15 @@ struct rte_graph {
uint32_t cir_mask;   /**< Circular buffer wrap around mask. */
rte_node_t nb_nodes; /**< Number of nodes in the graph. */
rte_graph_off_t *cir_start;  /**< Pointer to circular buffer. */
+   /* Graph schedule */
+   struct rte_graph_rq_head *rq __rte_cache_aligned; /* The run-queue */
+   struct rte_graph_rq_head rq_head; /* The head for run-queue list */
+
+   SLIST_ENTRY(rte_graph) rq_next;   /* The next for run-queue list */
+   unsigned int lcore_id;  /**< The graph running Lcore. */
+   struct rte_ring *wq;/**< The work-queue for pending streams. */
+   struct rte_mempool *mp; /**< The mempool for scheduling streams. */
+   /* Graph schedule area */
rte_graph_off_t nodes_start; /**< Offset at which node memory starts. */
rte_graph_t id; /**< Graph identifier. */
int socket; /**< Socket ID where memory is allocated. */
@@ -74,6 +90,11 @@ struct rte_node {
/** Original process function when pcap is enabled. */
rte_node_process_t original_process;
 
+   RTE_STD_C11
+   union {
+   /* Fast schedule area for mcore dispatch model */
+   unsigned int lcore_id;  /**< Node running lcore. */
+   };
/* Fast path area  */
 #define RTE_NODE_CTX_SZ 16
uint8_t ctx[RTE_NODE_CTX_SZ] __rte_cache_aligned; /**< Node Context. */
-- 
2.37.2



[PATCH v6 09/15] graph: introduce stream moving cross cores

2023-05-08 Thread Zhirun Yan
This patch introduces key functions to allow a worker thread to
enable enqueue and move streams of objects to the next nodes over
different cores.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/graph.c|   6 +-
 lib/graph/graph_private.h|  30 +
 lib/graph/meson.build|   2 +-
 lib/graph/rte_graph.h|  15 ++-
 lib/graph/rte_graph_model_dispatch.c | 157 +++
 lib/graph/rte_graph_model_dispatch.h |  37 +++
 lib/graph/version.map|   2 +
 7 files changed, 244 insertions(+), 5 deletions(-)

diff --git a/lib/graph/graph.c b/lib/graph/graph.c
index e809aa55b0..f555844d8f 100644
--- a/lib/graph/graph.c
+++ b/lib/graph/graph.c
@@ -495,7 +495,7 @@ clone_name(struct graph *graph, struct graph *parent_graph, 
const char *name)
 }
 
 static rte_graph_t
-graph_clone(struct graph *parent_graph, const char *name)
+graph_clone(struct graph *parent_graph, const char *name, struct 
rte_graph_param *prm)
 {
struct graph_node *graph_node;
struct graph *graph;
@@ -566,14 +566,14 @@ graph_clone(struct graph *parent_graph, const char *name)
 }
 
 rte_graph_t
-rte_graph_clone(rte_graph_t id, const char *name)
+rte_graph_clone(rte_graph_t id, const char *name, struct rte_graph_param *prm)
 {
struct graph *graph;
 
GRAPH_ID_CHECK(id);
STAILQ_FOREACH(graph, &graph_list, next)
if (graph->id == id)
-   return graph_clone(graph, name);
+   return graph_clone(graph, name, prm);
 
 fail:
return RTE_GRAPH_ID_INVALID;
diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h
index 02b10ea2b6..70347116ba 100644
--- a/lib/graph/graph_private.h
+++ b/lib/graph/graph_private.h
@@ -372,4 +372,34 @@ void graph_dump(FILE *f, struct graph *g);
  */
 void node_dump(FILE *f, struct node *n);
 
+/**
+ * @internal
+ *
+ * Create the graph schedule work queue. And all cloned graphs attached to the
+ * parent graph MUST be destroyed together for fast schedule design limitation.
+ *
+ * @param _graph
+ *   The graph object
+ * @param _parent_graph
+ *   The parent graph object which holds the run-queue head.
+ * @param prm
+ *   Graph parameter, includes model-specific parameters in this graph.
+ *
+ * @return
+ *   - 0: Success.
+ *   - <0: Graph schedule work queue related error.
+ */
+int graph_sched_wq_create(struct graph *_graph, struct graph *_parent_graph,
+  struct rte_graph_param *prm);
+
+/**
+ * @internal
+ *
+ * Destroy the graph schedule work queue.
+ *
+ * @param _graph
+ *   The graph object
+ */
+void graph_sched_wq_destroy(struct graph *_graph);
+
 #endif /* _RTE_GRAPH_PRIVATE_H_ */
diff --git a/lib/graph/meson.build b/lib/graph/meson.build
index c729d984b6..e21affa280 100644
--- a/lib/graph/meson.build
+++ b/lib/graph/meson.build
@@ -20,4 +20,4 @@ sources = files(
 )
 headers = files('rte_graph.h', 'rte_graph_worker.h')
 
-deps += ['eal', 'pcapng']
+deps += ['eal', 'pcapng', 'mempool', 'ring']
diff --git a/lib/graph/rte_graph.h b/lib/graph/rte_graph.h
index 2f86c17de7..0ac764daf8 100644
--- a/lib/graph/rte_graph.h
+++ b/lib/graph/rte_graph.h
@@ -169,6 +169,17 @@ struct rte_graph_param {
bool pcap_enable; /**< Pcap enable. */
uint64_t num_pkt_to_capture; /**< Number of packets to capture. */
char *pcap_filename; /**< Filename in which packets to be captured.*/
+
+   RTE_STD_C11
+   union {
+   struct {
+   uint64_t rsvd[8];
+   } rtc;
+   struct {
+   uint32_t wq_size_max;
+   uint32_t mp_capacity;
+   } dispatch;
+   };
 };
 
 /**
@@ -260,12 +271,14 @@ int rte_graph_destroy(rte_graph_t id);
  *   Name of the new graph. The library prepends the parent graph name to the
  * user-specified name. The final graph name will be,
  * "parent graph name" + "-" + name.
+ * @param prm
+ *   Graph parameter, includes model-specific parameters in this graph.
  *
  * @return
  *   Valid graph id on success, RTE_GRAPH_ID_INVALID otherwise.
  */
 __rte_experimental
-rte_graph_t rte_graph_clone(rte_graph_t id, const char *name);
+rte_graph_t rte_graph_clone(rte_graph_t id, const char *name, struct 
rte_graph_param *prm);
 
 /**
  * Get graph id from graph name.
diff --git a/lib/graph/rte_graph_model_dispatch.c 
b/lib/graph/rte_graph_model_dispatch.c
index 3364a76ed4..4264723485 100644
--- a/lib/graph/rte_graph_model_dispatch.c
+++ b/lib/graph/rte_graph_model_dispatch.c
@@ -5,6 +5,163 @@
 #include "graph_private.h"
 #include "rte_graph_model_dispatch.h"
 
+int
+graph_sched_wq_create(struct graph *_graph, struct graph *_parent_graph,
+  struct rte_graph_param *prm)
+{
+   struct rte_graph *parent_graph = _parent_graph->graph;
+   struct rte_graph *graph = _graph->graph;
+   unsigned int wq_s

[PATCH v6 10/15] graph: enable create and destroy graph scheduling workqueue

2023-05-08 Thread Zhirun Yan
This patch enables to create and destroy scheduling workqueue into
common graph operations.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/graph.c | 9 +
 1 file changed, 9 insertions(+)

diff --git a/lib/graph/graph.c b/lib/graph/graph.c
index f555844d8f..8b42d43193 100644
--- a/lib/graph/graph.c
+++ b/lib/graph/graph.c
@@ -449,6 +449,10 @@ rte_graph_destroy(rte_graph_t id)
while (graph != NULL) {
tmp = STAILQ_NEXT(graph, next);
if (graph->id == id) {
+   /* Destroy the schedule work queue if has */
+   if (rte_graph_worker_model_get() == 
RTE_GRAPH_MODEL_MCORE_DISPATCH)
+   graph_sched_wq_destroy(graph);
+
/* Call fini() of the all the nodes in the graph */
graph_node_fini(graph);
/* Destroy graph fast path memory */
@@ -543,6 +547,11 @@ graph_clone(struct graph *parent_graph, const char *name, 
struct rte_graph_param
if (graph_fp_mem_create(graph))
goto graph_cleanup;
 
+   /* Create the graph schedule work queue */
+   if (rte_graph_worker_model_get() == RTE_GRAPH_MODEL_MCORE_DISPATCH &&
+   graph_sched_wq_create(graph, parent_graph, prm))
+   goto graph_mem_destroy;
+
/* Call init() of the all the nodes in the graph */
if (graph_node_init(graph))
goto graph_mem_destroy;
-- 
2.37.2



[PATCH v6 11/15] graph: introduce graph walk by cross-core dispatch

2023-05-08 Thread Zhirun Yan
This patch introduces the task scheduler mechanism to enable dispatching
tasks to another worker cores. Currently, there is only a local work
queue for one graph to walk. We introduce a scheduler worker queue in
each worker core for dispatching tasks. It will perform the walk on
scheduler work queue first, then handle the local work queue.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/rte_graph_model_dispatch.h | 43 
 1 file changed, 43 insertions(+)

diff --git a/lib/graph/rte_graph_model_dispatch.h 
b/lib/graph/rte_graph_model_dispatch.h
index 18fa7ce0ab..f35cddba31 100644
--- a/lib/graph/rte_graph_model_dispatch.h
+++ b/lib/graph/rte_graph_model_dispatch.h
@@ -73,6 +73,49 @@ __rte_experimental
 int rte_graph_model_dispatch_lcore_affinity_set(const char *name,
unsigned int lcore_id);
 
+/**
+ * Perform graph walk on the circular buffer and invoke the process function
+ * of the nodes and collect the stats.
+ *
+ * @param graph
+ *   Graph pointer returned from rte_graph_lookup function.
+ *
+ * @see rte_graph_lookup()
+ */
+__rte_experimental
+static inline void
+rte_graph_walk_mcore_dispatch(struct rte_graph *graph)
+{
+   const rte_graph_off_t *cir_start = graph->cir_start;
+   const rte_node_t mask = graph->cir_mask;
+   uint32_t head = graph->head;
+   struct rte_node *node;
+
+   RTE_ASSERT(graph->parent_id != RTE_GRAPH_ID_INVALID);
+   if (graph->wq != NULL)
+   __rte_graph_sched_wq_process(graph);
+
+   while (likely(head != graph->tail)) {
+   node = (struct rte_node *)RTE_PTR_ADD(graph, 
cir_start[(int32_t)head++]);
+
+   /* skip the src nodes which not bind with current worker */
+   if ((int32_t)head < 0 && node->lcore_id != graph->lcore_id)
+   continue;
+
+   /* Schedule the node until all task/objs are done */
+   if (node->lcore_id != RTE_MAX_LCORE &&
+   graph->lcore_id != node->lcore_id && graph->rq != NULL &&
+   __rte_graph_sched_node_enqueue(node, graph->rq))
+   continue;
+
+   __rte_node_process(graph, node);
+
+   head = likely((int32_t)head > 0) ? head & mask : head;
+   }
+
+   graph->tail = 0;
+}
+
 #ifdef __cplusplus
 }
 #endif
-- 
2.37.2



[PATCH v6 12/15] graph: enable graph multicore dispatch scheduler model

2023-05-08 Thread Zhirun Yan
This patch enables to chose new scheduler model.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/rte_graph_worker.h | 9 -
 1 file changed, 8 insertions(+), 1 deletion(-)

diff --git a/lib/graph/rte_graph_worker.h b/lib/graph/rte_graph_worker.h
index 5b58f7bda9..2dd27b3949 100644
--- a/lib/graph/rte_graph_worker.h
+++ b/lib/graph/rte_graph_worker.h
@@ -11,6 +11,7 @@ extern "C" {
 #endif
 
 #include "rte_graph_model_rtc.h"
+#include "rte_graph_model_dispatch.h"
 
 /**
  * Perform graph walk on the circular buffer and invoke the process function
@@ -25,7 +26,13 @@ __rte_experimental
 static inline void
 rte_graph_walk(struct rte_graph *graph)
 {
-   rte_graph_walk_rtc(graph);
+   int model = rte_graph_worker_model_get();
+
+   if (model == RTE_GRAPH_MODEL_DEFAULT ||
+   model == RTE_GRAPH_MODEL_RTC)
+   rte_graph_walk_rtc(graph);
+   else if (model == RTE_GRAPH_MODEL_MCORE_DISPATCH)
+   rte_graph_walk_mcore_dispatch(graph);
 }
 
 #ifdef __cplusplus
-- 
2.37.2



[PATCH v6 13/15] graph: add stats for cross-core dispatching

2023-05-08 Thread Zhirun Yan
Add stats for cross-core dispatching scheduler if stats collection is
enabled.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 lib/graph/graph_debug.c  |  6 +++
 lib/graph/graph_stats.c  | 74 +---
 lib/graph/rte_graph.h|  2 +
 lib/graph/rte_graph_model_dispatch.c |  3 ++
 lib/graph/rte_graph_worker_common.h  |  2 +
 5 files changed, 79 insertions(+), 8 deletions(-)

diff --git a/lib/graph/graph_debug.c b/lib/graph/graph_debug.c
index b84412f5dd..7dcf07b080 100644
--- a/lib/graph/graph_debug.c
+++ b/lib/graph/graph_debug.c
@@ -74,6 +74,12 @@ rte_graph_obj_dump(FILE *f, struct rte_graph *g, bool all)
fprintf(f, "   size=%d\n", n->size);
fprintf(f, "   idx=%d\n", n->idx);
fprintf(f, "   total_objs=%" PRId64 "\n", n->total_objs);
+   if (rte_graph_worker_model_get() == 
RTE_GRAPH_MODEL_MCORE_DISPATCH) {
+   fprintf(f, "   total_sched_objs=%" PRId64 "\n",
+   n->total_sched_objs);
+   fprintf(f, "   total_sched_fail=%" PRId64 "\n",
+   n->total_sched_fail);
+   }
fprintf(f, "   total_calls=%" PRId64 "\n", n->total_calls);
for (i = 0; i < n->nb_edges; i++)
fprintf(f, "  edge[%d] <%s>\n", i,
diff --git a/lib/graph/graph_stats.c b/lib/graph/graph_stats.c
index c0140ba922..9ccb358aa2 100644
--- a/lib/graph/graph_stats.c
+++ b/lib/graph/graph_stats.c
@@ -40,13 +40,19 @@ struct rte_graph_cluster_stats {
struct cluster_node clusters[];
 } __rte_cache_aligned;
 
+#define boarder_model_dispatch()   
   \
+   fprintf(f, "+---+---+" \
+  "---+---+---+---+" \
+  "---+---+-" \
+  "--+\n")
+
 #define boarder()  
\
fprintf(f, "+---+---+" \
   "---+---+---+---+-" \
   "--+\n")
 
 static inline void
-print_banner(FILE *f)
+print_banner_default(FILE *f)
 {
boarder();
fprintf(f, "%-32s%-16s%-16s%-16s%-16s%-16s%-16s\n", "|Node", "|calls",
@@ -55,6 +61,27 @@ print_banner(FILE *f)
boarder();
 }
 
+static inline void
+print_banner_dispatch(FILE *f)
+{
+   boarder_model_dispatch();
+   fprintf(f, "%-32s%-16s%-16s%-16s%-16s%-16s%-16s%-16s%-16s\n",
+   "|Node", "|calls",
+   "|objs", "|sched objs", "|sched fail",
+   "|realloc_count", "|objs/call", "|objs/sec(10E6)",
+   "|cycles/call|");
+   boarder_model_dispatch();
+}
+
+static inline void
+print_banner(FILE *f)
+{
+   if (rte_graph_worker_model_get() == RTE_GRAPH_MODEL_MCORE_DISPATCH)
+   print_banner_dispatch(f);
+   else
+   print_banner_default(f);
+}
+
 static inline void
 print_node(FILE *f, const struct rte_graph_cluster_node_stats *stat)
 {
@@ -76,11 +103,21 @@ print_node(FILE *f, const struct 
rte_graph_cluster_node_stats *stat)
objs_per_sec = ts_per_hz ? (objs - prev_objs) / ts_per_hz : 0;
objs_per_sec /= 100;
 
-   fprintf(f,
-   "|%-31s|%-15" PRIu64 "|%-15" PRIu64 "|%-15" PRIu64
-   "|%-15.3f|%-15.6f|%-11.4f|\n",
-   stat->name, calls, objs, stat->realloc_count, objs_per_call,
-   objs_per_sec, cycles_per_call);
+   if (rte_graph_worker_model_get() == RTE_GRAPH_MODEL_MCORE_DISPATCH) {
+   fprintf(f,
+   "|%-31s|%-15" PRIu64 "|%-15" PRIu64 "|%-15" PRIu64
+   "|%-15" PRIu64 "|%-15" PRIu64
+   "|%-15.3f|%-15.6f|%-11.4f|\n",
+   stat->name, calls, objs, stat->sched_objs,
+   stat->sched_fail, stat->realloc_count, objs_per_call,
+   objs_per_sec, cycles_per_call);
+   } else {
+   fprintf(f,
+   "|%-31s|%-15" PRIu64 "|%-15" PRIu64 "|%-15" PRIu64
+   "|%-15.3f|%-15.6f|%-11.4f|\n",
+   stat->name, calls, objs, stat->realloc_count, 
objs_per_call,
+   objs_per_sec, cycles_per_call);
+   }
 }
 
 static int
@@ -88,13 +125,20 @@ graph_cluster_stats_cb(bool is_first, bool is_last, void 
*cookie,
   const struct rte_graph_cluster_node_stats *stat)
 {
FILE *f = cookie;
+   int model;
+
+   model = rte_graph_worker_model_get();
 
if (unlikely(is_first))
print_banner(f);
if (stat->objs)
print_node(f, stat);
- 

[PATCH v6 15/15] doc: update multicore dispatch model in graph guides

2023-05-08 Thread Zhirun Yan
Update graph documentation to introduce new multicore dispatch model.

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 doc/guides/prog_guide/graph_lib.rst | 59 +++--
 1 file changed, 55 insertions(+), 4 deletions(-)

diff --git a/doc/guides/prog_guide/graph_lib.rst 
b/doc/guides/prog_guide/graph_lib.rst
index 1cfdc86433..72e26f3a5a 100644
--- a/doc/guides/prog_guide/graph_lib.rst
+++ b/doc/guides/prog_guide/graph_lib.rst
@@ -189,14 +189,65 @@ In the above example, A graph object will be created with 
ethdev Rx
 node of port 0 and queue 0, all ipv4* nodes in the system,
 and ethdev tx node of all ports.
 
-Multicore graph processing
-~~
-In the current graph library implementation, specifically,
-``rte_graph_walk()`` and ``rte_node_enqueue*`` fast path API functions
+graph model chossing
+
+Currently, there are 2 different walking model. Use
+``rte_graph_worker_model_set()`` to set the walking model.
+
+RTC (Run-To-Completion)
+^^^
+This is the default graph walking model. specifically,
+``rte_graph_walk_rtc()`` and ``rte_node_enqueue*`` fast path API functions
 are designed to work on single-core to have better performance.
 The fast path API works on graph object, So the multi-core graph
 processing strategy would be to create graph object PER WORKER.
 
+Example:
+
+Graph: node-0 -> node-1 -> node-2 @Core0.
+
+.. code-block:: diff
+
++ - - - - - - - - - - - - - - - - - - - - - +
+'  Core #0  '
+'   '
+' ++ +-+ ++ '
+' | Node-0 | --> | Node-1  | --> | Node-2 | '
+' ++ +-+ ++ '
+'   '
++ - - - - - - - - - - - - - - - - - - - - - +
+
+Dispatch model
+^^
+The dispatch model enables a cross-core dispatching mechanism which employs
+a scheduling work-queue to dispatch streams to other worker cores which
+being associated with the destination node.
+
+Use ``rte_graph_model_dispatch_lcore_affinity_set()`` to set lcore affinity
+with the node.
+Each worker core will have a graph repetition. Use ``rte_graph_clone()`` to
+clone graph for each worker and use``rte_graph_model_dispatch_core_bind()``
+to bind graph with the worker core.
+
+Example:
+
+Graph topo: node-0 -> Core1; node-1 -> node-2; node-2 -> node-3.
+Config graph: node-0 @Core0; node-1/3 @Core1; node-2 @Core2.
+
+.. code-block:: diff
+
++ - - - - - -+ +- - - - - - - - - - - - - + + - - - - - -+
+'  Core #0   ' '  Core #1 ' '  Core #2   '
+'' '  ' ''
+' ++ ' ' ++++ ' ' ++ '
+' | Node-0 | - - - ->| Node-1 || Node-3 |<- - - - | Node-2 | '
+' ++ ' ' ++++ ' ' ++ '
+'' ' |' '  ^ '
++ - - - - - -+ +- - -|- - - - - - - - - - + + - - -|- - -+
+ | |
+ + - - - - - - - - - - - - - - - - +
+
+
 In fast path
 
 Typical fast-path code looks like below, where the application
-- 
2.37.2



[PATCH v6 14/15] examples/l3fwd-graph: introduce multicore dispatch worker model

2023-05-08 Thread Zhirun Yan
Add new parameter "model" to choose dispatch or rtc worker model.
And in dispatch model, the node will affinity to worker core successively.

Note:
only support one RX node for remote model in current implementation.

./dpdk-l3fwd-graph  -l 8,9,10,11 -n 4 -- -p 0x1 --config="(0,0,9)" -P
--model="dispatch"

Signed-off-by: Haiyue Wang 
Signed-off-by: Cunming Liang 
Signed-off-by: Zhirun Yan 
---
 examples/l3fwd-graph/main.c | 236 +---
 1 file changed, 194 insertions(+), 42 deletions(-)

diff --git a/examples/l3fwd-graph/main.c b/examples/l3fwd-graph/main.c
index 5feeab4f0f..a91947e940 100644
--- a/examples/l3fwd-graph/main.c
+++ b/examples/l3fwd-graph/main.c
@@ -55,6 +55,9 @@
 
 #define NB_SOCKETS 8
 
+/* Graph module */
+#define WORKER_MODEL_RTC "rtc"
+#define WORKER_MODEL_MCORE_DISPATCH "dispatch"
 /* Static global variables used within this file. */
 static uint16_t nb_rxd = RX_DESC_DEFAULT;
 static uint16_t nb_txd = TX_DESC_DEFAULT;
@@ -88,6 +91,10 @@ struct lcore_rx_queue {
char node_name[RTE_NODE_NAMESIZE];
 };
 
+struct model_conf {
+   enum rte_graph_worker_model model;
+};
+
 /* Lcore conf */
 struct lcore_conf {
uint16_t n_rx_queue;
@@ -153,6 +160,19 @@ static struct ipv4_l3fwd_lpm_route 
ipv4_l3fwd_lpm_route_array[] = {
{RTE_IPV4(198, 18, 6, 0), 24, 6}, {RTE_IPV4(198, 18, 7, 0), 24, 7},
 };
 
+static int
+check_worker_model_params(void)
+{
+   if (rte_graph_worker_model_get() == RTE_GRAPH_MODEL_MCORE_DISPATCH &&
+   nb_lcore_params > 1) {
+   printf("Exceeded max number of lcore params for remote model: 
%hu\n",
+  nb_lcore_params);
+   return -1;
+   }
+
+   return 0;
+}
+
 static int
 check_lcore_params(void)
 {
@@ -276,6 +296,7 @@ print_usage(const char *prgname)
"  --eth-dest=X,MM:MM:MM:MM:MM:MM: Ethernet destination for "
"port X\n"
"  --max-pkt-len PKTLEN: maximum packet length in decimal 
(64-9600)\n"
+   "  --model NAME: walking model name, dispatch or rtc(by 
default)\n"
"  --no-numa: Disable numa awareness\n"
"  --per-port-pool: Use separate buffer pool per port\n"
"  --pcap-enable: Enables pcap capture\n"
@@ -318,6 +339,20 @@ parse_max_pkt_len(const char *pktlen)
return len;
 }
 
+static int
+parse_worker_model(const char *model)
+{
+   if (strcmp(model, WORKER_MODEL_MCORE_DISPATCH) == 0) {
+   rte_graph_worker_model_set(RTE_GRAPH_MODEL_MCORE_DISPATCH);
+   return RTE_GRAPH_MODEL_MCORE_DISPATCH;
+   } else if (strcmp(model, WORKER_MODEL_RTC) == 0)
+   return RTE_GRAPH_MODEL_RTC;
+
+   rte_exit(EXIT_FAILURE, "Invalid worker model: %s", model);
+
+   return RTE_GRAPH_MODEL_LIST_END;
+}
+
 static int
 parse_portmask(const char *portmask)
 {
@@ -434,6 +469,8 @@ static const char short_options[] = "p:" /* portmask */
 #define CMD_LINE_OPT_PCAP_ENABLE   "pcap-enable"
 #define CMD_LINE_OPT_NUM_PKT_CAP   "pcap-num-cap"
 #define CMD_LINE_OPT_PCAP_FILENAME "pcap-file-name"
+#define CMD_LINE_OPT_WORKER_MODEL  "model"
+
 enum {
/* Long options mapped to a short option */
 
@@ -449,6 +486,7 @@ enum {
CMD_LINE_OPT_PARSE_PCAP_ENABLE,
CMD_LINE_OPT_PARSE_NUM_PKT_CAP,
CMD_LINE_OPT_PCAP_FILENAME_CAP,
+   CMD_LINE_OPT_WORKER_MODEL_TYPE,
 };
 
 static const struct option lgopts[] = {
@@ -460,6 +498,7 @@ static const struct option lgopts[] = {
{CMD_LINE_OPT_PCAP_ENABLE, 0, 0, CMD_LINE_OPT_PARSE_PCAP_ENABLE},
{CMD_LINE_OPT_NUM_PKT_CAP, 1, 0, CMD_LINE_OPT_PARSE_NUM_PKT_CAP},
{CMD_LINE_OPT_PCAP_FILENAME, 1, 0, CMD_LINE_OPT_PCAP_FILENAME_CAP},
+   {CMD_LINE_OPT_WORKER_MODEL, 1, 0, CMD_LINE_OPT_WORKER_MODEL_TYPE},
{NULL, 0, 0, 0},
 };
 
@@ -551,6 +590,11 @@ parse_args(int argc, char **argv)
printf("Pcap file name: %s\n", pcap_filename);
break;
 
+   case CMD_LINE_OPT_WORKER_MODEL_TYPE:
+   printf("Use new worker model: %s\n", optarg);
+   parse_worker_model(optarg);
+   break;
+
default:
print_usage(prgname);
return -1;
@@ -726,15 +770,15 @@ print_stats(void)
 static int
 graph_main_loop(void *conf)
 {
+   struct model_conf *mconf = conf;
struct lcore_conf *qconf;
struct rte_graph *graph;
uint32_t lcore_id;
 
-   RTE_SET_USED(conf);
-
lcore_id = rte_lcore_id();
qconf = &lcore_conf[lcore_id];
graph = qconf->graph;
+   rte_graph_worker_model_set(mconf->model);
 
if (!graph) {
RTE_LOG(INFO, L3FWD_GRAPH, "Lcore %u has nothing to do\n",
@@ -788,6 +832,139 @@ config_port_max_pkt_len(struct rte_eth_conf *conf,
return 0;
 }
 
+static void
+graph_config_mcore_dispatch(struct rte_graph_param gra

Re: [PATCH 4/4] app: add testgraph application

2023-05-08 Thread Jerin Jacob
On Fri, Apr 21, 2023 at 11:33 AM Vamsi Attunuru  wrote:
>
> Patch adds test-graph application to validate graph
> and node libraries.
>
> Signed-off-by: Vamsi Attunuru 

+ haiyue.w...@intel.com  cunming.li...@intel.com zhirun@intel.com
for review to align with new multi-core dispatch model.


> ---
>  app/meson.build|1 +
>  app/test-graph/cmdline.c   |  212 ++
>  app/test-graph/cmdline_graph.c |  297 
>  app/test-graph/cmdline_graph.h |   19 +
>  app/test-graph/meson.build |   17 +
>  app/test-graph/parameters.c|  157 
>  app/test-graph/testgraph.c | 1309 
>  app/test-graph/testgraph.h |   92 +++
>  doc/guides/tools/index.rst |1 +
>  doc/guides/tools/testgraph.rst |  131 
>  10 files changed, 2236 insertions(+)
>
> diff --git a/app/meson.build b/app/meson.build
> index 74d2420f67..6c7b24e604 100644
> --- a/app/meson.build
> +++ b/app/meson.build
> @@ -22,6 +22,7 @@ apps = [
>  'test-eventdev',
>  'test-fib',
>  'test-flow-perf',
> +'test-graph',
>  'test-gpudev',
>  'test-mldev',
>  'test-pipeline',
> diff --git a/app/test-graph/cmdline.c b/app/test-graph/cmdline.c
> new file mode 100644
> index 00..a07a8a24f9
> --- /dev/null
> +++ b/app/test-graph/cmdline.c
> @@ -0,0 +1,212 @@
> +/* SPDX-License-Identifier: BSD-3-Clause
> + * Copyright(C) 2023 Marvell International Ltd.
> + */
> +
> +#include 
> +
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +#include 
> +
> +#include "cmdline_graph.h"
> +#include "testgraph.h"
> +
> +static struct cmdline *testgraph_cl;
> +static cmdline_parse_ctx_t *main_ctx;
> +
> +/* *** Help command with introduction. *** */
> +struct cmd_help_brief_result {
> +   cmdline_fixed_string_t help;
> +};
> +
> +static void
> +cmd_help_brief_parsed(__rte_unused void *parsed_result, struct cmdline *cl, 
> __rte_unused void *data)
> +{
> +   cmdline_printf(cl,
> +  "\n"
> +  "Help is available for the following sections:\n\n"
> +  "help control: Start and stop 
> graph walk.\n"
> +  "help display: Displaying 
> port, stats and config "
> +  "information.\n"
> +  "help config : Configuration 
> information.\n"
> +  "help all: All of the 
> above sections.\n\n");
> +}
> +
> +static cmdline_parse_token_string_t cmd_help_brief_help =
> +   TOKEN_STRING_INITIALIZER(struct cmd_help_brief_result, help, "help");
> +
> +static cmdline_parse_inst_t cmd_help_brief = {
> +   .f = cmd_help_brief_parsed,
> +   .data = NULL,
> +   .help_str = "help: Show help",
> +   .tokens = {
> +   (void *)&cmd_help_brief_help,
> +   NULL,
> +   },
> +};
> +
> +/* *** Help command with help sections. *** */
> +struct cmd_help_long_result {
> +   cmdline_fixed_string_t help;
> +   cmdline_fixed_string_t section;
> +};
> +
> +static void
> +cmd_help_long_parsed(void *parsed_result, struct cmdline *cl, __rte_unused 
> void *data)
> +{
> +   int show_all = 0;
> +   struct cmd_help_long_result *res = parsed_result;
> +
> +   if (!strcmp(res->section, "all"))
> +   show_all = 1;
> +
> +   if (show_all || !strcmp(res->section, "control")) {
> +
> +   cmdline_printf(cl, "\n"
> +  "Control forwarding:\n"
> +  "---\n\n"
> +
> +  "start graph_walk\n"
> +  " Start graph_walk on worker threads.\n\n"
> +
> +  "stop graph_walk\n"
> +  " Stop worker threads from running 
> graph_walk.\n\n"
> +
> +  "quit\n"
> +  "Quit to prompt.\n\n");
> +   }
> +
> +   if (show_all || !strcmp(res->section, "display")) {
> +
> +   cmdline_printf(cl,
> +  "\n"
> +  "Display:\n"
> +  "\n\n"
> +
> +  "show node_list\n"
> +  " Display the list of supported nodes.\n\n"
> +
> +  "show graph_stats\n"
> +  " Display the node statistics of graph 
> cluster.\n\n");
> +   }
> +
> +   if (show_all || !strcmp(res->section, "config")) {
> +   cmdline_printf(cl, "\n"
> +  "Configuration:\n"
> +  "--\n"
> +  "set lcore_config 
> (port_id0,rxq0,lcore_idX),..."
> +  

RE: [PATCH v5 1/3] ring: fix unmatched type definition and usage

2023-05-08 Thread Ruifeng Wang
> -Original Message-
> From: Jie Hai 
> Sent: Tuesday, May 9, 2023 9:29 AM
> To: Honnappa Nagarahalli ; Konstantin Ananyev
> ; Ruifeng Wang ; Gavin 
> Hu
> ; Olivier Matz ; Dharmik Jayesh 
> Thakkar
> 
> Cc: dev@dpdk.org; liudongdo...@huawei.com
> Subject: [PATCH v5 1/3] ring: fix unmatched type definition and usage
> 
> Field 'flags' of struct rte_ring is defined as int type. However, it is used 
> as unsigned
> int. To ensure consistency, change the type of flags to unsigned int. Since 
> these two
> types has the same byte size, this change is not an ABI change.
> 
> Fixes: cc4b218790f6 ("ring: support configurable element size")

The change looks good.
However, I think the fix line is not accurate. 
I suppose it fixes af75078fece3 ("first public release").

> 
> Signed-off-by: Jie Hai 
> Acked-by: Konstantin Ananyev 
> Acked-by: Chengwen Feng 
> ---
>  lib/ring/rte_ring_core.h | 2 +-
>  1 file changed, 1 insertion(+), 1 deletion(-)
> 
> diff --git a/lib/ring/rte_ring_core.h b/lib/ring/rte_ring_core.h index
> 82b237091b71..1c809abeb531 100644
> --- a/lib/ring/rte_ring_core.h
> +++ b/lib/ring/rte_ring_core.h
> @@ -120,7 +120,7 @@ struct rte_ring_hts_headtail {  struct rte_ring {
>   char name[RTE_RING_NAMESIZE] __rte_cache_aligned;
>   /**< Name of the ring. */
> - int flags;   /**< Flags supplied at creation. */
> + uint32_t flags;   /**< Flags supplied at creation. */
>   const struct rte_memzone *memzone;
>   /**< Memzone, if any, containing the rte_ring */
>   uint32_t size;   /**< Size of ring. */
> --
> 2.33.0



RE: [PATCH v5 3/3] ring: add telemetry cmd for ring info

2023-05-08 Thread Morten Brørup
> From: Jie Hai [mailto:haij...@huawei.com]
> Sent: Tuesday, 9 May 2023 03.29
> 
> This patch supports dump of ring information by its name.
> An example using this command is shown below:
> 
> --> /ring/info,MP_mb_pool_0
> {
>   "/ring/info": {
> "name": "MP_mb_pool_0",
> "socket": 0,
> "flags": "0x0",
> "producer_type": "MP",
> "consumer_type": "MC",
> "size": 262144,
> "mask": "0x3",
> "capacity": 262143,
> "used_count": 153197,
> "consumer_tail": 2259,
> "consumer_head": 2259,
> "producer_tail": 155456,
> "producer_head": 155456,
> "mz_name": "RG_MP_mb_pool_0",
> "mz_len": 2097536,
> "mz_hugepage_sz": 1073741824,
> "mz_socket_id": 0,
> "mz_flags": "0x0"
>   }
> }
> 
> Signed-off-by: Jie Hai 
> Reviewed-by: Honnappa Nagarahalli 
> Acked-by: Konstantin Ananyev 
> Acked-by: Huisong Li 
> Acked-by: Chengwen Feng 
> ---
>  lib/ring/rte_ring.c | 99 +
>  1 file changed, 99 insertions(+)
> 
> diff --git a/lib/ring/rte_ring.c b/lib/ring/rte_ring.c
> index 0e83d0099363..26c8f2a2e6a2 100644
> --- a/lib/ring/rte_ring.c
> +++ b/lib/ring/rte_ring.c
> @@ -455,8 +455,107 @@ ring_handle_list(const char *cmd __rte_unused,
>   return 0;
>  }
> 
> +static const char *
> +ring_prod_sync_type_to_name(struct rte_ring *r)
> +{
> + switch (r->prod.sync_type) {
> + case RTE_RING_SYNC_MT:
> + return "MP";
> + case RTE_RING_SYNC_ST:
> + return "SP";
> + case RTE_RING_SYNC_MT_RTS:
> + return "MP_RTS";
> + case RTE_RING_SYNC_MT_HTS:
> + return "MP_HTS";
> + default:
> + return "Unknown";
> + }
> +}
> +
> +static const char *
> +ring_cons_sync_type_to_name(struct rte_ring *r)
> +{
> + switch (r->cons.sync_type) {
> + case RTE_RING_SYNC_MT:
> + return "MC";
> + case RTE_RING_SYNC_ST:
> + return "SC";
> + case RTE_RING_SYNC_MT_RTS:
> + return "MC_RTS";
> + case RTE_RING_SYNC_MT_HTS:
> + return "MC_HTS";
> + default:
> + return "Unknown";
> + }
> +}

I considered if these two functions should be replaced by one, returning e.g. 
"MT" instead of "MP"/"MC"; however, the ring flags describe the sync types as 
e.g. "MP" and  "MP RTS", so I eventually came to agree with the approach in the 
patch.

The structure documentation requires that the sync_type field must be in the 
same position, regardless which of the union's types are being used. So this is 
also correct.

Series-Acked-by: Morten Brørup 



RE: [PATCH v4 1/4] bus/cdx: introduce cdx bus

2023-05-08 Thread Xia, Chenbo
> -Original Message-
> From: Nipun Gupta 
> Sent: Monday, May 8, 2023 7:18 PM
> To: dev@dpdk.org; tho...@monjalon.net; david.march...@redhat.com
> Cc: ferruh.yi...@amd.com; harpreet.an...@amd.com; nikhil.agar...@amd.com;
> Nipun Gupta 
> Subject: [PATCH v4 1/4] bus/cdx: introduce cdx bus
> 
> CDX bus supports multiple type of devices, which can be
> exposed to user-space via vfio-cdx.
> 
> vfio-cdx provides the MMIO IO_MEMORY regions as well as the
> DMA interface for the device (IOMMU).
> 
> This support aims to enable the DPDK to support the cdx
> devices in user-space using VFIO interface.
> 
> Signed-off-by: Nipun Gupta 
> ---
>  MAINTAINERS|   5 +
>  doc/guides/rel_notes/release_23_07.rst |   6 +
>  drivers/bus/cdx/bus_cdx_driver.h   | 201 ++
>  drivers/bus/cdx/cdx.c  | 520 +
>  drivers/bus/cdx/cdx_logs.h |  37 ++
>  drivers/bus/cdx/cdx_vfio.c | 437 +
>  drivers/bus/cdx/meson.build|  13 +
>  drivers/bus/cdx/private.h  |  49 +++
>  drivers/bus/cdx/version.map|  11 +
>  drivers/bus/meson.build|   1 +
>  10 files changed, 1280 insertions(+)
>  create mode 100644 drivers/bus/cdx/bus_cdx_driver.h
>  create mode 100644 drivers/bus/cdx/cdx.c
>  create mode 100644 drivers/bus/cdx/cdx_logs.h
>  create mode 100644 drivers/bus/cdx/cdx_vfio.c
>  create mode 100644 drivers/bus/cdx/meson.build
>  create mode 100644 drivers/bus/cdx/private.h
>  create mode 100644 drivers/bus/cdx/version.map
> 

...

> --- /dev/null
> +++ b/drivers/bus/cdx/cdx.c
> @@ -0,0 +1,520 @@
> +/* SPDX-License-Identifier: BSD-3-Clause
> + * Copyright (C) 2022-2023, Advanced Micro Devices, Inc.
> + */
> +
> +/*
> + * Architecture Overview
> + * =
> + * CDX is a Hardware Architecture designed for AMD FPGA devices. It
> + * consists of sophisticated mechanism for interaction between FPGA,
> + * Firmware and the APUs (Application CPUs).
> + *
> + * Firmware resides on RPU (Realtime CPUs) which interacts with
> + * the FPGA program manager and the APUs. The RPU provides memory-mapped
> + * interface (RPU if) which is used to communicate with APUs.
> + *
> + * The diagram below shows an overview of the CDX architecture:
> + *
> + *  +--+
> + *  |   DPDK   |
> + *  |DPDK CDX drivers  |
> + *  | ||
> + *  |  DPDK CDX bus|
> + *  | ||
> + *  +-|+
> + *|
> + *  +-|+
> + *  |Application CPUs (APU)   ||
> + *  | ||
> + *  | VFIO CDX driver  |
> + *  | Linux OS||
> + *  |  Linux CDX bus   |
> + *  | ||
> + *  +-|+
> + *|
> + *|
> + *  +| RPU if |+
> + *  | ||
> + *  | V|
> + *  |  Realtime CPUs (RPU) |
> + *  |  |
> + *  +--+
> + *|
> + *  +-|+
> + *  |  FPGA   ||
> + *  |  +---+   |
> + *  |  |   |   |   |
> + *  | +---++---+   +---+   |
> + *  | | dev 1 || dev 2 |   | dev 3 |   |
> + *  | +---++---+   +---+   |
> + *  +--+
> + *
> + * The RPU firmware extracts the device information from the loaded FPGA
> + * image and implements a mechanism that allows the APU drivers to
> + * enumerate such devices (device personality and resource details) via
> + * a dedicated communication channel.

What is APU? CPU resources that application uses? Then why cpu resources that 
DPDK
Uses are not part of it?

I also notice that there are more and more vfio-XXX devices in DPDK now: 
pci/platform/cdx.
Some vfio related code are very similar. In the future we may need to think 
about make them
a common lib or something.

Thanks,
Chenbo

> + *
> + * VFIO CDX driver provides the CDX device resources like MMIO and
> interrupts
> + * to map to user-space. DPDK CDX bus uses sysfs interface and the vfio-
> cdx
> + * driver to discover and initialize the CDX devices for user-space
> + * applications.
>