[dpdk-dev] [PATCH] examples/ipsec-secgw: fix flow pattern buffer overrun
From: Satheesh Paul This patch fixes the patterns buffer overrun issue reported by Coverity. Coverity issue: 379236 Fixes: 8d0cdaa2d7 ("examples/ipsec-secgw: support mark and security flow action") Cc: sta...@dpdk.org Signed-off-by: Satheesh Paul Reviewed-by: Akhil Goyal --- examples/ipsec-secgw/flow.c | 6 ++ 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/examples/ipsec-secgw/flow.c b/examples/ipsec-secgw/flow.c index aee97b2fc4..9b18be83af 100644 --- a/examples/ipsec-secgw/flow.c +++ b/examples/ipsec-secgw/flow.c @@ -270,7 +270,7 @@ parse_flow_tokens(char **tokens, uint32_t n_tokens, nb_flow_rule++; } -#define MAX_RTE_FLOW_PATTERN (4) +#define MAX_RTE_FLOW_PATTERN (5) #define MAX_RTE_FLOW_ACTIONS (5) static void @@ -334,9 +334,7 @@ flow_init_single(struct flow_rule_entry *rule) pattern[pattern_idx].spec = &rule->ipv4.spec; pattern[pattern_idx].mask = &rule->ipv4.mask; pattern_idx++; - } - - if (rule->is_ipv6) { + } else if (rule->is_ipv6) { pattern[pattern_idx].type = RTE_FLOW_ITEM_TYPE_IPV6; pattern[pattern_idx].spec = &rule->ipv6.spec; pattern[pattern_idx].mask = &rule->ipv6.mask; -- 2.35.3
Re: 19.11.13 patches review and test
On Fri, Jul 29, 2022 at 2:08 PM Jiang, YuX wrote: > > > -Original Message- > > From: christian.ehrha...@canonical.com > > Sent: Monday, July 18, 2022 6:17 PM > > To: sta...@dpdk.org > > Cc: dev@dpdk.org; Abhishek Marathe ; > > Ali Alnubani ; Walker, Benjamin > > ; David Christensen > > ; Hemant Agrawal ; > > Stokes, Ian ; Jerin Jacob ; > > Mcnamara, John ; Ju-Hyoung Lee > > ; Kevin Traynor ; Luca > > Boccassi ; Pei Zhang ; Xu, Qian > > Q ; Raslan Darawsheh ; > > Thomas Monjalon ; Peng, Yuan > > ; Chen, Zhaoyan > > Subject: 19.11.13 patches review and test > > > > Hi all, > > > > Here is a list of patches targeted for stable release 19.11.13. > > > > The planned date for the final release is August 29th. > > > > Please help with testing and validation of your use cases and report any > > issues/results with reply-all to this mail. For the final release the fixes > > and > > reported validations will be added to the release notes. > > > > A release candidate tarball can be found at: > > > > https://dpdk.org/browse/dpdk-stable/tag/?id=v19.11.13-rc1 > > > > These patches are located at branch 19.11 of dpdk-stable repo: > > https://dpdk.org/browse/dpdk-stable/ > > > > Thanks. > > > > Christian Ehrhardt > > > > --- > > Update the test status for Intel part. DPDK19.11.13-rc1 almost test finished, > no critical issue is found, Thank you for the update - that is great to hear. > but find some new bugs on new OS. Yeah we always find those and there will be more over time the older this gets. I already got two fixed (one freebsd and one gcc 11.2) but this kind is just growing without being able to do much :-/ TL;DR: we are glad to know, but only if the build in the established environments 19.11 started in break it is really bad. FYI - there most likely will be an -rc2 later as we have collected many more fixes. I hope that fits somewhere into your time availability - gladly we have a few more weeks. > New report defects: > Bug1: https://bugs.dpdk.org/show_bug.cgi?id=1056 [dpdk-19.11.13-rc1] > unit_tests_eal/link_bonding_rssconf: link_bonding_rssconf_autotest test failed > > This is found in 22.07, but that fix failed to be applied to LTS19.11, > need rebase. > Build-Bug2: [dpdk 19.11.13-rc1] lib/librte_eal meson build error with > gcc12.1 on fedora36, similar as https://bugs.dpdk.org/show_bug.cgi?id=985 > > No fix yet. > Build-Bug3: [dpdk 19.11.13-rc1] drivers/net/ena make build error with > gcc12.1 on fedora36, similar as https://bugs.dpdk.org/show_bug.cgi?id=991 > > No fix yet. > Build-Bug4: [dpdk 19.11.13-rc1] drivers/net/ice/ice_rxtx meson build error > with gcc11.2 on Ubuntu22.04". > > This is found in 22.07, but that fix failed to be applied to LTS19.11, > need rebase, Intel Dev is under investigating. > Build-Bug5: [dpdk 19.11.13-rc1] /lib/eal meson and make build failed with > gcc10.3.0 and clang13.0.0 on FreeBSD13.1/64. > > This is found in 22.07, but that fix failed to be applied to LTS19.11, > need rebase, Intel Dev(Bruce) has send fix. > Build-Bug6: [dpdk 19.11.13-rc1] drivers/net/qede make build error with > clang14 on fedora36/redhat8.6/UB22.04 > > Known issue, https://bugs.dpdk.org/show_bug.cgi?id=912, no fix yet. > > # Basic Intel(R) NIC testing > * Build: cover the build test combination with latest GCC/Clang/ICC version > and the popular OS revision such as Ubuntu20.04&22.04, Fedora36, RHEL8.4, etc. > - All test done. > * PF&VF(i40e, ixgbe): test scenarios including > RTE_FLOW/TSO/Jumboframe/checksum offload/VLAN/VXLAN, etc. > - All test done. Find one new bug: Bug1. > * PF/VF(ice): test scenarios including Switch features/Package > Management/Flow Director/Advanced Tx, etc. > - All test done. No new issue is found. Known bug about [dpdk-19.11.12] > metering_and_policing/ipv4_HASH_table_RFC2698: unable to forward packets > normally. Intel Dev is still investigating. > * Intel NIC single core/NIC performance: test scenarios including PF/VF > single core performance test etc. > - All test done. No big performance drop. > * IPsec: test scenarios including ipsec/ipsec-gw/ipsec library basic test - > QAT&SW/FIB library, etc. > - All test passed. > # Basic cryptodev and virtio testing > * Virtio: both function and performance test are covered. Such as > PVP/Virtio_loopback/virtio-user loopback/virtio-net VM2VM perf testing, etc. > - All test done. No new issue is found. > * Cryptodev: > * Function test: test scenarios including Cryptodev API testing/CompressDev > ISA-L/QAT/ZLIB PMD Testing/ etc. > - All test done. No new issue is found. > * Performance test: test scenarios including Thoughput Performance /Cryptodev > Latency, etc. > - All test done. No performance drop. > > Best regards, > Yu Jiang -- Christian Ehrhardt Senior Staff Engineer, Ubuntu Server Canonical Ltd
Re: 19.11.13 patches review and test
On Wed, Aug 3, 2022 at 6:40 AM YangHang Liu wrote: > > Hi Christian, > > The dpdk 19.11.13-rc1 test result from Red Hat looks good. Awesome, glad to hear that. FYI - there most likely will be an -rc2 later as we have collected many more fixes. I hope that fits somewhere into your time availability - gladly we have a few more weeks. > We tested below 17 scenarios and all got PASS on RHEL8: > > Guest with device assignment(PF) throughput testing(1G hugepage size): PASS > Guest with device assignment(PF) throughput testing(2M hugepage size) : PASS > Guest with device assignment(VF) throughput testing: PASS > PVP (host dpdk testpmd as vswitch) 1Q: throughput testing: PASS > PVP vhost-user 2Q throughput testing: PASS > PVP vhost-user 1Q - cross numa node throughput testing: PASS > Guest with vhost-user 2 queues throughput testing: PASS > vhost-user reconnect with dpdk-client, qemu-server: qemu reconnect: PASS > vhost-user reconnect with dpdk-client, qemu-server: ovs reconnect: PASS > PVP 1Q live migration testing: PASS > PVP 1Q cross numa node live migration testing: PASS > Guest with ovs+dpdk+vhost-user 1Q live migration testing: PASS > Guest with ovs+dpdk+vhost-user 1Q live migration testing (2M): PASS > Guest with ovs+dpdk+vhost-user 2Q live migration testing: PASS > Guest with ovs+dpdk+vhost-user 4Q live migration testing: PASS > Host PF + DPDK testing: PASS > Host VF + DPDK testing: PASS > > > Versions: > > kernel 4.18 > qemu 6.2 > dpdk: git://dpdk.org/dpdk-stable (remotes/origin/19.11) > > # git log > Author: Christian Ehrhardt > Date: Mon Jul 18 11:14:00 2022 +0200 > version: 19.11.13-rc1 > Signed-off-by: Christian Ehrhardt > > NICs: X540-AT2 NIC(ixgbe, 10G) > > > > > Best Regards, > YangHang Liu > > > On Tue, Jul 19, 2022 at 1:33 PM Christian Ehrhardt > wrote: >> >> On Tue, Jul 19, 2022 at 4:43 AM Pei Zhang wrote: >> > >> > cc Yanghang Liu (yangh...@redhat.com). Yanghang is from the RedHat QE team >> > and will finish this dpdk testing with our test cases. >> > >> > Hello Christian, >> > >> > Could you also cc Yanghang in the future once there are new testing >> > requests :)? Thanks. >> >> Hi - Sure we can do that, >> the list is tracked in http://git.dpdk.org/tools/stable-scripts/. >> I've converted your request into a patch against that. >> >> > Best regards, >> > >> > Pei >> > >> > On Mon, Jul 18, 2022 at 6:19 PM wrote: >> >> >> >> Hi all, >> >> >> >> Here is a list of patches targeted for stable release 19.11.13. >> >> >> >> The planned date for the final release is August 29th. >> >> >> >> Please help with testing and validation of your use cases and report >> >> any issues/results with reply-all to this mail. For the final release >> >> the fixes and reported validations will be added to the release notes. >> >> >> >> A release candidate tarball can be found at: >> >> >> >> https://dpdk.org/browse/dpdk-stable/tag/?id=v19.11.13-rc1 >> >> >> >> These patches are located at branch 19.11 of dpdk-stable repo: >> >> https://dpdk.org/browse/dpdk-stable/ >> >> >> >> Thanks. >> >> >> >> Christian Ehrhardt >> >> >> >> --- >> >> Abdullah Ömer Yamaç (1): >> >> examples/distributor: fix distributor on Rx core >> >> >> >> Andy Pei (1): >> >> vdpa/ifc/base: fix null pointer dereference >> >> >> >> Ankur Dwivedi (1): >> >> common/cpt: fix build with GCC 12 >> >> >> >> Chengwen Feng (4): >> >> ethdev: clarify null location case in xstats get >> >> net/hns3: fix xstats get return if xstats is null >> >> net/ipn3ke: fix xstats get return if xstats is null >> >> net/mvpp2: fix xstats get return if xstats is null >> >> >> >> Ciara Power (6): >> >> test/crypto: fix null check for ZUC authentication >> >> test/crypto: fix cipher offset for ZUC >> >> test/crypto: fix authentication IV for ZUC SGL >> >> test/crypto: fix ZUC vector IV format >> >> test/crypto: fix SNOW3G vector IV format >> >> doc: fix grammar and formatting in compressdev guide >> >> >> >> Damodharam Ammepalli (2): >> >> net/bnxt: allow Tx only or Rx only >> >> net/bnxt: cleanup MTU setting >> >> >> >> Dariusz Sosnowski (1): >> >> net/mlx5: fix GTP handling in header modify action >> >> >> >> David Marchand (11): >> >> net/nfp: remove unneeded header inclusion >> >> devtools: fix null test for NUMA systems >> >> app/testpmd: remove useless pointer checks >> >> net/bnxt: fix compatibility with some old firmwares >> >> net/ice: fix build with GCC 12 >> >> vdpa/ifc: fix build with GCC 12 >> >> eal/x86: drop export of internal alignment macro >> >> test/ipsec: fix build with GCC 12 >> >> vhost/crypto: fix build with GCC 12 >> >> vhost/crypto: fix descriptor processing >> >> vhost: add some trailing newline in log messages >> >> >> >> Deepak Khandelwal (1): >> >> mem: skip attaching external memory in secondary process >> >> >> >> Dmitry Kozlyuk (2): >> >> net/mlx5
19.11.13 patches review and test
Hi all, there were three patches close to the deadline that I missed and I considered postponing them to 19.11.14 at first. But in the meantime there arrived 11 more and I think that justfies an -rc2 tag for 19.11.13. We still have almost 4 weeks left - I hope that is ok. Here is the combined list of patches (the same as before plus the new 14) targeted for stable release 19.11.13. The planned date for the final release is August 29th. Please help with testing and validation of your use cases and report any issues/results with reply-all to this mail. For the final release the fixes and reported validations will be added to the release notes. A release candidate tarball can be found at: https://dpdk.org/browse/dpdk-stable/tag/?id=v19.11.13-rc2 These patches are located at branch 19.11 of dpdk-stable repo: https://dpdk.org/browse/dpdk-stable/ Thanks. Christian Ehrhardt --- Abdullah Ömer Yamaç (1): examples/distributor: fix distributor on Rx core Andy Pei (1): vdpa/ifc/base: fix null pointer dereference Ankur Dwivedi (1): common/cpt: fix build with GCC 12 Bruce Richardson (1): eal/freebsd: fix use of newer cpuset macros Chengwen Feng (4): ethdev: clarify null location case in xstats get net/hns3: fix xstats get return if xstats is null net/ipn3ke: fix xstats get return if xstats is null net/mvpp2: fix xstats get return if xstats is null Christian Ehrhardt (1): version: 19.11.13-rc1 Ciara Power (6): test/crypto: fix null check for ZUC authentication test/crypto: fix cipher offset for ZUC test/crypto: fix authentication IV for ZUC SGL test/crypto: fix ZUC vector IV format test/crypto: fix SNOW3G vector IV format doc: fix grammar and formatting in compressdev guide Damodharam Ammepalli (2): net/bnxt: allow Tx only or Rx only net/bnxt: cleanup MTU setting Dariusz Sosnowski (1): net/mlx5: fix GTP handling in header modify action David Marchand (11): net/nfp: remove unneeded header inclusion devtools: fix null test for NUMA systems app/testpmd: remove useless pointer checks net/bnxt: fix compatibility with some old firmwares net/ice: fix build with GCC 12 vdpa/ifc: fix build with GCC 12 eal/x86: drop export of internal alignment macro test/ipsec: fix build with GCC 12 vhost/crypto: fix build with GCC 12 vhost/crypto: fix descriptor processing vhost: add some trailing newline in log messages Deepak Khandelwal (1): mem: skip attaching external memory in secondary process Dmitry Kozlyuk (2): net/mlx5: fix Tx when inlining is impossible ethdev: prohibit polling stopped queue Fidaullah Noonari (1): malloc: fix allocation of almost hugepage size Gagandeep Singh (5): crypto/dpaa_sec: fix digest size crypto/dpaa2_sec: fix buffer pool ID check crypto/dpaa2_sec: fix operation status for simple FD common/dpaax: fix short MAC-I IV calculation for ZUC net/dpaa: fix event queue detach Gowrishankar Muthukrishnan (1): examples/fips_validation: handle empty payload Gregory Etelson (1): Revert "net/mlx5: fix RSS expansion scheme for GRE header" Harman Kalra (1): net/octeontx: fix port close Harry van Haaren (1): service: fix lingering active status Heinrich Schuchardt (2): drivers/crypto: fix warnings for OpenSSL version test/hash: fix out of bound access Hongbo Zheng (1): test: check memory allocation for CRC Huisong Li (9): net/hns3: fix insecure way to query MAC statistics net/bonding: fix RSS inconsistency between ports test/bonding: fix RSS test when disable RSS net/hns3: remove duplicate definition app/testpmd: fix port status of bonding slave device app/testpmd: fix bonding slave devices not released app/testpmd: fix supported RSS offload display net/bonding: fix stopping non-active slaves ethdev: fix RSS update when RSS is disabled Jakub Wysocki (1): crypto/scheduler: fix queue pair in scheduler failover Jeff Daly (1): net/ixgbe: add option for link up check on pin SDP3 Jiri Slaby (1): kni: fix build with Linux 5.18 Joyce Kong (1): net/memif: fix overwriting of head segment Kalesh AP (2): net/bnxt: remove unused macro net/bnxt: recheck FW readiness if in reset process Ke Zhang (6): net/bonding: fix RSS key config with extended key length kni: use dedicated function to set random MAC address kni: use dedicated function to set MAC address net/iavf: fix Rx queue interrupt setting net/iavf: net/iavf: fix mbuf release in multi-process app/testpmd: fix multicast address pool leak Konstantin Ananyev (1): acl: fix rules with 8-byte field size Long Li (2): net/netvsc: fix calculation of checksums based on mbuf flag net/netvsc: fix vmbus device reference in multi-process Luc Pelle
[PATCH 0/3] ethdev: remove deprecated Flow Director configuration
Flow Director configuration fdir_conf was deprecated long time ago and was planned to be removed in 20.11. i40e, ixgbe, qede and txgbe driver maintainers are encouraged to review correspodning patches very carefully. Internal copy of the Flow Director configuration in ixgbe and txgbe drivers is not actually initialized. Should I use configuration from testpmd as the default one? Andrew Rybchenko (3): net/ixgbe: add helper to get Flow Director configuration net/txgbe: add helper to get Flow Director configuration ethdev: remove deprecated Flow Director configuration app/test-pmd/cmdline.c | 343 app/test-pmd/config.c | 100 -- app/test-pmd/parameters.c | 76 - app/test-pmd/testpmd.c | 24 -- app/test-pmd/testpmd.h | 6 - doc/guides/nics/fm10k.rst | 2 - doc/guides/nics/ixgbe.rst | 1 - doc/guides/rel_notes/deprecation.rst| 4 - doc/guides/rel_notes/release_22_11.rst | 4 + doc/guides/testpmd_app_ug/run_app.rst | 27 -- doc/guides/testpmd_app_ug/testpmd_funcs.rst | 38 --- drivers/net/cnxk/cnxk_ethdev.c | 5 - drivers/net/fm10k/fm10k_rxtx_vec.c | 5 - drivers/net/hns3/hns3_rxtx_vec.c| 4 - drivers/net/i40e/i40e_ethdev.c | 31 -- drivers/net/i40e/i40e_fdir.c| 304 + drivers/net/i40e/i40e_rxtx_vec_common.h | 5 - drivers/net/ixgbe/ixgbe_ethdev.c| 2 +- drivers/net/ixgbe/ixgbe_ethdev.h| 4 + drivers/net/ixgbe/ixgbe_fdir.c | 32 +- drivers/net/ixgbe/ixgbe_flow.c | 13 +- drivers/net/ixgbe/ixgbe_rxtx_vec_common.h | 2 +- drivers/net/octeontx/octeontx_ethdev.c | 5 - drivers/net/qede/qede_filter.c | 29 +- drivers/net/sfc/sfc.c | 5 - drivers/net/sfc/sfc_repr.c | 5 - drivers/net/thunderx/nicvf_ethdev.c | 5 - drivers/net/txgbe/txgbe_ethdev.c| 2 +- drivers/net/txgbe/txgbe_ethdev.h| 3 + drivers/net/txgbe/txgbe_fdir.c | 24 +- drivers/net/txgbe/txgbe_flow.c | 13 +- lib/ethdev/ethdev_driver.h | 36 ++ lib/ethdev/rte_ethdev.h | 43 --- 33 files changed, 100 insertions(+), 1102 deletions(-) -- 2.30.2
[PATCH 2/3] net/txgbe: add helper to get Flow Director configuration
Flow Director configuration is deprecated in generic ethdev device configuration and will be removed. However, it is essential for the driver flow API support and cannot be removed completely without loss of flow API support. Prepare to move the configuration into device private data. Signed-off-by: Andrew Rybchenko --- drivers/net/txgbe/txgbe_ethdev.c | 2 +- drivers/net/txgbe/txgbe_ethdev.h | 2 ++ drivers/net/txgbe/txgbe_fdir.c | 24 +++- drivers/net/txgbe/txgbe_flow.c | 2 +- 4 files changed, 15 insertions(+), 15 deletions(-) diff --git a/drivers/net/txgbe/txgbe_ethdev.c b/drivers/net/txgbe/txgbe_ethdev.c index 9dc9948219..0624568bdf 100644 --- a/drivers/net/txgbe/txgbe_ethdev.c +++ b/drivers/net/txgbe/txgbe_ethdev.c @@ -1719,7 +1719,7 @@ txgbe_dev_start(struct rte_eth_dev *dev) txgbe_configure_port(dev); txgbe_configure_dcb(dev); - if (dev->data->dev_conf.fdir_conf.mode != RTE_FDIR_MODE_NONE) { + if (TXGBE_DEV_FDIR_CONF(dev)->mode != RTE_FDIR_MODE_NONE) { err = txgbe_fdir_configure(dev); if (err) goto error; diff --git a/drivers/net/txgbe/txgbe_ethdev.h b/drivers/net/txgbe/txgbe_ethdev.h index 262dbb5e38..e36a2c3d0d 100644 --- a/drivers/net/txgbe/txgbe_ethdev.h +++ b/drivers/net/txgbe/txgbe_ethdev.h @@ -373,6 +373,8 @@ struct txgbe_adapter { #define TXGBE_DEV_ADAPTER(dev) \ ((struct txgbe_adapter *)(dev)->data->dev_private) +#define TXGBE_DEV_FDIR_CONF(dev) (&(dev)->data->dev_conf.fdir_conf) + #define TXGBE_DEV_HW(dev) \ (&((struct txgbe_adapter *)(dev)->data->dev_private)->hw) diff --git a/drivers/net/txgbe/txgbe_fdir.c b/drivers/net/txgbe/txgbe_fdir.c index e303d87176..a198b6781b 100644 --- a/drivers/net/txgbe/txgbe_fdir.c +++ b/drivers/net/txgbe/txgbe_fdir.c @@ -170,7 +170,7 @@ txgbe_fdir_set_input_mask(struct rte_eth_dev *dev) { struct txgbe_hw *hw = TXGBE_DEV_HW(dev); struct txgbe_hw_fdir_info *info = TXGBE_DEV_FDIR(dev); - enum rte_fdir_mode mode = dev->data->dev_conf.fdir_conf.mode; + enum rte_fdir_mode mode = TXGBE_DEV_FDIR_CONF(dev)->mode; /* * mask VM pool and DIPv6 since there are currently not supported * mask FLEX byte, it will be set in flex_conf @@ -232,9 +232,8 @@ txgbe_fdir_set_input_mask(struct rte_eth_dev *dev) static int txgbe_fdir_store_input_mask(struct rte_eth_dev *dev) { - struct rte_eth_fdir_masks *input_mask = - &dev->data->dev_conf.fdir_conf.mask; - enum rte_fdir_mode mode = dev->data->dev_conf.fdir_conf.mode; + struct rte_eth_fdir_masks *input_mask = &TXGBE_DEV_FDIR_CONF(dev)->mask; + enum rte_fdir_mode mode = TXGBE_DEV_FDIR_CONF(dev)->mode; struct txgbe_hw_fdir_info *info = TXGBE_DEV_FDIR(dev); uint16_t dst_ipv6m = 0; uint16_t src_ipv6m = 0; @@ -294,7 +293,7 @@ static int txgbe_set_fdir_flex_conf(struct rte_eth_dev *dev, uint32_t flex) { const struct rte_eth_fdir_flex_conf *conf = - &dev->data->dev_conf.fdir_conf.flex_conf; + &TXGBE_DEV_FDIR_CONF(dev)->flex_conf; struct txgbe_hw *hw = TXGBE_DEV_HW(dev); struct txgbe_hw_fdir_info *info = TXGBE_DEV_FDIR(dev); const struct rte_eth_flex_payload_cfg *flex_cfg; @@ -364,7 +363,7 @@ txgbe_fdir_configure(struct rte_eth_dev *dev) int err; uint32_t fdirctrl, flex, pbsize; int i; - enum rte_fdir_mode mode = dev->data->dev_conf.fdir_conf.mode; + enum rte_fdir_mode mode = TXGBE_DEV_FDIR_CONF(dev)->mode; PMD_INIT_FUNC_TRACE(); @@ -373,8 +372,7 @@ txgbe_fdir_configure(struct rte_eth_dev *dev) mode != RTE_FDIR_MODE_PERFECT) return -ENOSYS; - err = configure_fdir_flags(&dev->data->dev_conf.fdir_conf, - &fdirctrl, &flex); + err = configure_fdir_flags(TXGBE_DEV_FDIR_CONF(dev), &fdirctrl, &flex); if (err) return err; @@ -792,7 +790,7 @@ txgbe_fdir_filter_program(struct rte_eth_dev *dev, bool is_perfect = FALSE; int err; struct txgbe_hw_fdir_info *info = TXGBE_DEV_FDIR(dev); - enum rte_fdir_mode fdir_mode = dev->data->dev_conf.fdir_conf.mode; + enum rte_fdir_mode fdir_mode = TXGBE_DEV_FDIR_CONF(dev)->mode; struct txgbe_fdir_filter *node; if (fdir_mode == RTE_FDIR_MODE_NONE || @@ -809,11 +807,11 @@ txgbe_fdir_filter_program(struct rte_eth_dev *dev, return -ENOTSUP; } fdirhash = atr_compute_perfect_hash(&rule->input, - dev->data->dev_conf.fdir_conf.pballoc); + TXGBE_DEV_FDIR_CONF(dev)->pballoc); fdirhash |= TXGBE_FDIRPIHASH_IDX(rule->soft_id); } else { fdirhash = atr_compute_signature_hash(&rule->input, -
[PATCH 1/3] net/ixgbe: add helper to get Flow Director configuration
Flow Director configuration is deprecated in generic ethdev device configuration and will be removed. However, it is essential for the driver flow API support and cannot be removed completely without loss of flow API support. Prepare to move the configuration into device private data. Signed-off-by: Andrew Rybchenko --- drivers/net/ixgbe/ixgbe_ethdev.c | 2 +- drivers/net/ixgbe/ixgbe_ethdev.h | 3 +++ drivers/net/ixgbe/ixgbe_fdir.c| 32 +++ drivers/net/ixgbe/ixgbe_flow.c| 2 +- drivers/net/ixgbe/ixgbe_rxtx_vec_common.h | 2 +- 5 files changed, 22 insertions(+), 19 deletions(-) diff --git a/drivers/net/ixgbe/ixgbe_ethdev.c b/drivers/net/ixgbe/ixgbe_ethdev.c index f31bbb7895..1dfad0e63d 100644 --- a/drivers/net/ixgbe/ixgbe_ethdev.c +++ b/drivers/net/ixgbe/ixgbe_ethdev.c @@ -2690,7 +2690,7 @@ ixgbe_dev_start(struct rte_eth_dev *dev) /* Configure DCB hw */ ixgbe_configure_dcb(dev); - if (dev->data->dev_conf.fdir_conf.mode != RTE_FDIR_MODE_NONE) { + if (IXGBE_DEV_FDIR_CONF(dev)->mode != RTE_FDIR_MODE_NONE) { err = ixgbe_fdir_configure(dev); if (err) goto error; diff --git a/drivers/net/ixgbe/ixgbe_ethdev.h b/drivers/net/ixgbe/ixgbe_ethdev.h index cc6049a66a..eabefd6343 100644 --- a/drivers/net/ixgbe/ixgbe_ethdev.h +++ b/drivers/net/ixgbe/ixgbe_ethdev.h @@ -522,6 +522,9 @@ struct ixgbe_vf_representor { int ixgbe_vf_representor_init(struct rte_eth_dev *ethdev, void *init_params); int ixgbe_vf_representor_uninit(struct rte_eth_dev *ethdev); +#define IXGBE_DEV_FDIR_CONF(dev) \ + (&(dev)->data->dev_conf.fdir_conf) + #define IXGBE_DEV_PRIVATE_TO_HW(adapter)\ (&((struct ixgbe_adapter *)adapter)->hw) diff --git a/drivers/net/ixgbe/ixgbe_fdir.c b/drivers/net/ixgbe/ixgbe_fdir.c index 834c1b3f51..06d6e2126d 100644 --- a/drivers/net/ixgbe/ixgbe_fdir.c +++ b/drivers/net/ixgbe/ixgbe_fdir.c @@ -317,7 +317,7 @@ fdir_set_input_mask_82599(struct rte_eth_dev *dev) reg = IXGBE_PCI_REG_ADDR(hw, IXGBE_FDIRDIP4M); *reg = ~(info->mask.dst_ipv4_mask); - if (dev->data->dev_conf.fdir_conf.mode == RTE_FDIR_MODE_SIGNATURE) { + if (IXGBE_DEV_FDIR_CONF(dev)->mode == RTE_FDIR_MODE_SIGNATURE) { /* * Store source and destination IPv6 masks (bit reversed) */ @@ -346,7 +346,7 @@ fdir_set_input_mask_x550(struct rte_eth_dev *dev) uint32_t fdirm = IXGBE_FDIRM_POOL | IXGBE_FDIRM_DIPv6 | IXGBE_FDIRM_FLEX; uint32_t fdiripv6m; - enum rte_fdir_mode mode = dev->data->dev_conf.fdir_conf.mode; + enum rte_fdir_mode mode = IXGBE_DEV_FDIR_CONF(dev)->mode; uint16_t mac_mask; PMD_INIT_FUNC_TRACE(); @@ -469,7 +469,7 @@ static int ixgbe_fdir_store_input_mask(struct rte_eth_dev *dev, const struct rte_eth_fdir_masks *input_mask) { - enum rte_fdir_mode mode = dev->data->dev_conf.fdir_conf.mode; + enum rte_fdir_mode mode = IXGBE_DEV_FDIR_CONF(dev)->mode; if (mode >= RTE_FDIR_MODE_SIGNATURE && mode <= RTE_FDIR_MODE_PERFECT) @@ -485,7 +485,7 @@ ixgbe_fdir_store_input_mask(struct rte_eth_dev *dev, int ixgbe_fdir_set_input_mask(struct rte_eth_dev *dev) { - enum rte_fdir_mode mode = dev->data->dev_conf.fdir_conf.mode; + enum rte_fdir_mode mode = IXGBE_DEV_FDIR_CONF(dev)->mode; if (mode >= RTE_FDIR_MODE_SIGNATURE && mode <= RTE_FDIR_MODE_PERFECT) @@ -639,7 +639,7 @@ ixgbe_fdir_configure(struct rte_eth_dev *dev) int err; uint32_t fdirctrl, pbsize; int i; - enum rte_fdir_mode mode = dev->data->dev_conf.fdir_conf.mode; + enum rte_fdir_mode mode = IXGBE_DEV_FDIR_CONF(dev)->mode; PMD_INIT_FUNC_TRACE(); @@ -658,7 +658,7 @@ ixgbe_fdir_configure(struct rte_eth_dev *dev) mode != RTE_FDIR_MODE_PERFECT) return -ENOSYS; - err = configure_fdir_flags(&dev->data->dev_conf.fdir_conf, &fdirctrl); + err = configure_fdir_flags(IXGBE_DEV_FDIR_CONF(dev), &fdirctrl); if (err) return err; @@ -680,13 +680,13 @@ ixgbe_fdir_configure(struct rte_eth_dev *dev) for (i = 1; i < 8; i++) IXGBE_WRITE_REG(hw, IXGBE_RXPBSIZE(i), 0); - err = fdir_set_input_mask(dev, &dev->data->dev_conf.fdir_conf.mask); + err = fdir_set_input_mask(dev, &IXGBE_DEV_FDIR_CONF(dev)->mask); if (err < 0) { PMD_INIT_LOG(ERR, " Error on setting FD mask"); return err; } - err = ixgbe_set_fdir_flex_conf(dev, - &dev->data->dev_conf.fdir_conf.flex_conf, &fdirctrl); + err = ixgbe_set_fdir_flex_conf(dev, &IXGBE_DEV_FDIR_CONF(dev)->flex_conf, + &fdirctrl); if (err < 0) { PMD_INIT_LOG(ERR, " Error on setting FD flexible arguments.");
[PATCH 3/3] ethdev: remove deprecated Flow Director configuration
Remove deprecated fdir_conf from device configuration. Assume that mode is equal to RTE_FDIR_MODE_NONE. Add internal Flow Director configuration copy in ixgbe and txgbe device private data since flow API supports requires it. Initialize mode to the first flow rule mode on the rule validation or creation. Since Flow Director configuration data types are still used by some drivers internally, move it from public API to ethdev driver internal API. Signed-off-by: Andrew Rybchenko --- app/test-pmd/cmdline.c | 343 app/test-pmd/config.c | 100 -- app/test-pmd/parameters.c | 76 - app/test-pmd/testpmd.c | 24 -- app/test-pmd/testpmd.h | 6 - doc/guides/nics/fm10k.rst | 2 - doc/guides/nics/ixgbe.rst | 1 - doc/guides/rel_notes/deprecation.rst| 4 - doc/guides/rel_notes/release_22_11.rst | 4 + doc/guides/testpmd_app_ug/run_app.rst | 27 -- doc/guides/testpmd_app_ug/testpmd_funcs.rst | 38 --- drivers/net/cnxk/cnxk_ethdev.c | 5 - drivers/net/fm10k/fm10k_rxtx_vec.c | 5 - drivers/net/hns3/hns3_rxtx_vec.c| 4 - drivers/net/i40e/i40e_ethdev.c | 31 -- drivers/net/i40e/i40e_fdir.c| 304 + drivers/net/i40e/i40e_rxtx_vec_common.h | 5 - drivers/net/ixgbe/ixgbe_ethdev.h| 3 +- drivers/net/ixgbe/ixgbe_flow.c | 13 +- drivers/net/octeontx/octeontx_ethdev.c | 5 - drivers/net/qede/qede_filter.c | 29 +- drivers/net/sfc/sfc.c | 5 - drivers/net/sfc/sfc_repr.c | 5 - drivers/net/thunderx/nicvf_ethdev.c | 5 - drivers/net/txgbe/txgbe_ethdev.h| 3 +- drivers/net/txgbe/txgbe_flow.c | 13 +- lib/ethdev/ethdev_driver.h | 36 ++ lib/ethdev/rte_ethdev.h | 43 --- 28 files changed, 67 insertions(+), 1072 deletions(-) diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c index b4fe9dfb17..7057c153f6 100644 --- a/app/test-pmd/cmdline.c +++ b/app/test-pmd/cmdline.c @@ -835,25 +835,6 @@ static void cmd_help_long_parsed(void *parsed_result, "filters:\n" "\n\n" - "flow_director_mask (port_id) mode IP vlan (vlan_value)" - " src_mask (ipv4_src) (ipv6_src) (src_port)" - " dst_mask (ipv4_dst) (ipv6_dst) (dst_port)\n" - "Set flow director IP mask.\n\n" - - "flow_director_mask (port_id) mode MAC-VLAN" - " vlan (vlan_value)\n" - "Set flow director MAC-VLAN mask.\n\n" - - "flow_director_mask (port_id) mode Tunnel" - " vlan (vlan_value) mac (mac_value)" - " tunnel-type (tunnel_type_value)" - " tunnel-id (tunnel_id_value)\n" - "Set flow director Tunnel mask.\n\n" - - "flow_director_flex_payload (port_id)" - " (raw|l2|l3|l4) (config)\n" - "Configure flex payload selection.\n\n" - "flow validate {port_id}" " [group {group_id}] [priority {level}]" " [ingress] [egress]" @@ -8981,326 +8962,6 @@ do { \ } \ } while (0) -/* *** deal with flow director mask *** */ -struct cmd_flow_director_mask_result { - cmdline_fixed_string_t flow_director_mask; - portid_t port_id; - cmdline_fixed_string_t mode; - cmdline_fixed_string_t mode_value; - cmdline_fixed_string_t vlan; - uint16_t vlan_mask; - cmdline_fixed_string_t src_mask; - cmdline_ipaddr_t ipv4_src; - cmdline_ipaddr_t ipv6_src; - uint16_t port_src; - cmdline_fixed_string_t dst_mask; - cmdline_ipaddr_t ipv4_dst; - cmdline_ipaddr_t ipv6_dst; - uint16_t port_dst; - cmdline_fixed_string_t mac; - uint8_t mac_addr_byte_mask; - cmdline_fixed_string_t tunnel_id; - uint32_t tunnel_id_mask; - cmdline_fixed_string_t tunnel_type; - uint8_t tunnel_type_mask; -}; - -static void -cmd_flow_director_mask_parsed(void *parsed_result, - __rte_unused struct cmdline *cl, - __rte_unused void *data) -{ - struct cmd_flow_director_mask_result *res = parsed_result; - struct rte_eth_fdir_masks *mask; - struct rte_port *port; - - port = &ports[res->port_id]; - /** Check if the port is not started **/ - if (port->port_status != RTE_PORT_STOPPED) { - fprintf(stderr, "Please stop port %d first\n", res->port_id); - return; - } - - mask = &port->dev_conf.fdir
[PATCH 00/13] add support for idpf PMD in DPDK
This patchset introduced the idpf (Infrastructure Data Path Function) PMD [*EXPERIMENTAL*] in DPDK for Intel Device ID of 0x1452. Junfeng Guo (13): net/idpf/base: introduce base code net/idpf/base: add logs and OS specific implementation net/idpf: support device initialization net/idpf: add queue operations net/idpf: add support to get device information net/idpf: add support to get packet type net/idpf: add support to update link status net/idpf: add basic Rx/Tx datapath net/idpf: add support for RSS net/idpf: add mtu configuration net/idpf: add hw statistics net/idpf: support write back based on ITR expire net/idpf: add AVX512 data path for single queue model drivers/net/idpf/base/iecm_alloc.h| 22 + drivers/net/idpf/base/iecm_common.c | 359 +++ drivers/net/idpf/base/iecm_controlq.c | 662 drivers/net/idpf/base/iecm_controlq.h | 214 ++ drivers/net/idpf/base/iecm_controlq_api.h | 227 ++ drivers/net/idpf/base/iecm_controlq_setup.c | 179 ++ drivers/net/idpf/base/iecm_devids.h | 17 + drivers/net/idpf/base/iecm_lan_pf_regs.h | 134 + drivers/net/idpf/base/iecm_lan_txrx.h | 428 +++ drivers/net/idpf/base/iecm_lan_vf_regs.h | 114 + drivers/net/idpf/base/iecm_osdep.h| 365 +++ drivers/net/idpf/base/iecm_prototype.h| 45 + drivers/net/idpf/base/iecm_type.h | 106 + drivers/net/idpf/base/meson.build | 27 + drivers/net/idpf/base/siov_regs.h | 41 + drivers/net/idpf/base/virtchnl.h | 2743 + drivers/net/idpf/base/virtchnl2.h | 1411 + drivers/net/idpf/base/virtchnl2_lan_desc.h| 603 drivers/net/idpf/base/virtchnl_inline_ipsec.h | 567 drivers/net/idpf/idpf_ethdev.c| 1194 +++ drivers/net/idpf/idpf_ethdev.h| 247 ++ drivers/net/idpf/idpf_logs.h | 38 + drivers/net/idpf/idpf_rxtx.c | 2348 ++ drivers/net/idpf/idpf_rxtx.h | 228 ++ drivers/net/idpf/idpf_rxtx_vec_avx512.c | 917 ++ drivers/net/idpf/idpf_rxtx_vec_common.h | 85 + drivers/net/idpf/idpf_vchnl.c | 1206 drivers/net/idpf/meson.build | 47 + drivers/net/idpf/version.map |3 + drivers/net/meson.build |1 + 30 files changed, 14578 insertions(+) create mode 100644 drivers/net/idpf/base/iecm_alloc.h create mode 100644 drivers/net/idpf/base/iecm_common.c create mode 100644 drivers/net/idpf/base/iecm_controlq.c create mode 100644 drivers/net/idpf/base/iecm_controlq.h create mode 100644 drivers/net/idpf/base/iecm_controlq_api.h create mode 100644 drivers/net/idpf/base/iecm_controlq_setup.c create mode 100644 drivers/net/idpf/base/iecm_devids.h create mode 100644 drivers/net/idpf/base/iecm_lan_pf_regs.h create mode 100644 drivers/net/idpf/base/iecm_lan_txrx.h create mode 100644 drivers/net/idpf/base/iecm_lan_vf_regs.h create mode 100644 drivers/net/idpf/base/iecm_osdep.h create mode 100644 drivers/net/idpf/base/iecm_prototype.h create mode 100644 drivers/net/idpf/base/iecm_type.h create mode 100644 drivers/net/idpf/base/meson.build create mode 100644 drivers/net/idpf/base/siov_regs.h create mode 100644 drivers/net/idpf/base/virtchnl.h create mode 100644 drivers/net/idpf/base/virtchnl2.h create mode 100644 drivers/net/idpf/base/virtchnl2_lan_desc.h create mode 100644 drivers/net/idpf/base/virtchnl_inline_ipsec.h create mode 100644 drivers/net/idpf/idpf_ethdev.c create mode 100644 drivers/net/idpf/idpf_ethdev.h create mode 100644 drivers/net/idpf/idpf_logs.h create mode 100644 drivers/net/idpf/idpf_rxtx.c create mode 100644 drivers/net/idpf/idpf_rxtx.h create mode 100644 drivers/net/idpf/idpf_rxtx_vec_avx512.c create mode 100644 drivers/net/idpf/idpf_rxtx_vec_common.h create mode 100644 drivers/net/idpf/idpf_vchnl.c create mode 100644 drivers/net/idpf/meson.build create mode 100644 drivers/net/idpf/version.map -- 2.25.1
[PATCH 02/13] net/idpf/base: add logs and OS specific implementation
Add PMD logs. Add some MACRO definations and small functions which are specific for DPDK. Signed-off-by: Beilei Xing Signed-off-by: Junfeng Guo --- drivers/net/idpf/base/iecm_osdep.h | 365 + drivers/net/idpf/idpf_logs.h | 38 +++ 2 files changed, 403 insertions(+) create mode 100644 drivers/net/idpf/base/iecm_osdep.h create mode 100644 drivers/net/idpf/idpf_logs.h diff --git a/drivers/net/idpf/base/iecm_osdep.h b/drivers/net/idpf/base/iecm_osdep.h new file mode 100644 index 00..60e21fbc1b --- /dev/null +++ b/drivers/net/idpf/base/iecm_osdep.h @@ -0,0 +1,365 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2001-2022 Intel Corporation + */ + +#ifndef _IECM_OSDEP_H_ +#define _IECM_OSDEP_H_ + +#include +#include +#include +#include +#include +#include +#include + +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#include "../idpf_logs.h" + +#define INLINE inline +#define STATIC static + +typedef uint8_tu8; +typedef int8_t s8; +typedef uint16_t u16; +typedef int16_ts16; +typedef uint32_t u32; +typedef int32_ts32; +typedef uint64_t u64; +typedef uint64_t s64; + +typedef enum iecm_status iecm_status; +typedef struct iecm_lock iecm_lock; + +#define __iomem +#define hw_dbg(hw, S, A...)do {} while (0) +#define upper_32_bits(n) ((u32)(((n) >> 16) >> 16)) +#define lower_32_bits(n) ((u32)(n)) +#define low_16_bits(x) ((x) & 0x) +#define high_16_bits(x)(((x) & 0x) >> 16) + +#ifndef ETH_ADDR_LEN +#define ETH_ADDR_LEN 6 +#endif + +#ifndef __le16 +#define __le16 uint16_t +#endif +#ifndef __le32 +#define __le32 uint32_t +#endif +#ifndef __le64 +#define __le64 uint64_t +#endif +#ifndef __be16 +#define __be16 uint16_t +#endif +#ifndef __be32 +#define __be32 uint32_t +#endif +#ifndef __be64 +#define __be64 uint64_t +#endif + +#ifndef __always_unused +#define __always_unused __attribute__((__unused__)) +#endif +#ifndef __maybe_unused +#define __maybe_unused __attribute__((__unused__)) +#endif +#ifndef __packed +#define __packed __attribute__((packed)) +#endif + +#ifndef BIT_ULL +#define BIT_ULL(a) (1ULL << (a)) +#endif + +#ifndef BIT +#define BIT(a) (1ULL << (a)) +#endif + +#define FALSE 0 +#define TRUE 1 +#define false 0 +#define true 1 + +#define min(a, b) RTE_MIN(a, b) +#define max(a, b) RTE_MAX(a, b) + +#define ARRAY_SIZE(arr) (sizeof(arr) / sizeof(arr[0])) +#define FIELD_SIZEOF(t, f) (sizeof(((t *)0)->(f))) +#define MAKEMASK(m, s) ((m) << (s)) + +#define DEBUGOUT(S) PMD_DRV_LOG_RAW(DEBUG, S) +#define DEBUGOUT2(S, A...) PMD_DRV_LOG_RAW(DEBUG, S, ##A) +#define DEBUGFUNC(F) PMD_DRV_LOG_RAW(DEBUG, F) + +#define iecm_debug(h, m, s, ...) \ + do {\ + if (((m) & (h)->debug_mask))\ + PMD_DRV_LOG_RAW(DEBUG, "iecm %02x.%x " s, \ + (h)->bus.device, (h)->bus.func, \ + ##__VA_ARGS__); \ + } while (0) + +#define iecm_info(hw, fmt, args...) iecm_debug(hw, IECM_DBG_ALL, fmt, ##args) +#define iecm_warn(hw, fmt, args...) iecm_debug(hw, IECM_DBG_ALL, fmt, ##args) +#define iecm_debug_array(hw, type, rowsize, groupsize, buf, len) \ + do {\ + struct iecm_hw *hw_l = hw; \ + u16 len_l = len;\ + u8 *buf_l = buf;\ + int i; \ + for (i = 0; i < len_l; i += 8) \ + iecm_debug(hw_l, type, \ + "0x%04X 0x%016"PRIx64"\n", \ + i, *((u64 *)((buf_l) + i))); \ + } while (0) +#define iecm_snprintf snprintf +#ifndef SNPRINTF +#define SNPRINTF iecm_snprintf +#endif + +#define IECM_PCI_REG(reg) rte_read32(reg) +#define IECM_PCI_REG_ADDR(a, reg) \ + ((volatile uint32_t *)((char *)(a)->hw_addr + (reg))) +#define IECM_PCI_REG64(reg) rte_read64(reg) +#define IECM_PCI_REG_ADDR64(a, reg)\ + ((volatile uint64_t *)((char *)(a)->hw_addr + (reg))) + +#define iecm_wmb() rte_io_wmb() +#define iecm_rmb() rte_io_rmb() +#define iecm_mb() rte_io_mb() + +static inline uint32_t iecm_read_addr(volatile void *addr) +{ + return rte_le_to_cpu_32(IECM_PCI_REG(addr)); +} + +static inline uint64_t iecm_read_addr64(volatile void *addr) +{ + return rte_le_to_cpu_64(IECM_PCI_REG64(addr)); +} + +#define IEC
[PATCH 03/13] net/idpf: support device initialization
Support device init and the following dev ops: - dev_configure - dev_start - dev_stop - dev_close Signed-off-by: Beilei Xing Signed-off-by: Xiaoyun Li Signed-off-by: Xiao Wang Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 651 + drivers/net/idpf/idpf_ethdev.h | 205 +++ drivers/net/idpf/idpf_vchnl.c | 476 drivers/net/idpf/meson.build | 18 + drivers/net/idpf/version.map | 3 + drivers/net/meson.build| 1 + 6 files changed, 1354 insertions(+) create mode 100644 drivers/net/idpf/idpf_ethdev.c create mode 100644 drivers/net/idpf/idpf_ethdev.h create mode 100644 drivers/net/idpf/idpf_vchnl.c create mode 100644 drivers/net/idpf/meson.build create mode 100644 drivers/net/idpf/version.map diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c new file mode 100644 index 00..87c68226dd --- /dev/null +++ b/drivers/net/idpf/idpf_ethdev.c @@ -0,0 +1,651 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2022 Intel Corporation + */ + +#include +#include +#include +#include +#include +#include +#include +#include + +#include "idpf_ethdev.h" + +#define REPRESENTOR"representor" + +struct idpf_adapter *adapter; +uint16_t used_vecs_num; + +static const char * const idpf_valid_args[] = { + REPRESENTOR, + NULL +}; + +static int idpf_dev_configure(struct rte_eth_dev *dev); +static int idpf_dev_start(struct rte_eth_dev *dev); +static int idpf_dev_stop(struct rte_eth_dev *dev); +static int idpf_dev_close(struct rte_eth_dev *dev); + +static const struct eth_dev_ops idpf_eth_dev_ops = { + .dev_configure = idpf_dev_configure, + .dev_start = idpf_dev_start, + .dev_stop = idpf_dev_stop, + .dev_close = idpf_dev_close, +}; + +static int +idpf_init_vport_req_info(struct rte_eth_dev *dev) +{ + struct virtchnl2_create_vport *vport_info; + uint16_t idx = adapter->next_vport_idx; + + if (!adapter->vport_req_info[idx]) { + adapter->vport_req_info[idx] = rte_zmalloc(NULL, + sizeof(struct virtchnl2_create_vport), 0); + if (!adapter->vport_req_info[idx]) { + PMD_INIT_LOG(ERR, "Failed to allocate vport_req_info"); + return -1; + } + } + + vport_info = + (struct virtchnl2_create_vport *)adapter->vport_req_info[idx]; + + vport_info->vport_type = rte_cpu_to_le_16(VIRTCHNL2_VPORT_TYPE_DEFAULT); + + return 0; +} + +static uint16_t +idpf_get_next_vport_idx(struct idpf_vport **vports, uint16_t max_vport_nb, + uint16_t cur_vport_idx) +{ + uint16_t vport_idx; + uint16_t i; + + if (cur_vport_idx < max_vport_nb && !vports[cur_vport_idx + 1]) { + vport_idx = cur_vport_idx + 1; + return vport_idx; + } + + for (i = 0; i < max_vport_nb; i++) { + if (!vports[i]) + break; + } + + if (i == max_vport_nb) + vport_idx = IDPF_INVALID_VPORT_IDX; + else + vport_idx = i; + + return vport_idx; +} + +#ifndef IDPF_RSS_KEY_LEN +#define IDPF_RSS_KEY_LEN 52 +#endif + +static int +idpf_init_vport(struct rte_eth_dev *dev) +{ + uint16_t idx = adapter->next_vport_idx; + struct virtchnl2_create_vport *vport_info = + (struct virtchnl2_create_vport *)adapter->vport_recv_info[idx]; + struct idpf_vport *vport = + (struct idpf_vport *)dev->data->dev_private; + int i; + + vport->adapter = adapter; + vport->vport_id = vport_info->vport_id; + vport->txq_model = vport_info->txq_model; + vport->rxq_model = vport_info->rxq_model; + vport->num_tx_q = vport_info->num_tx_q; + vport->num_tx_complq = vport_info->num_tx_complq; + vport->num_rx_q = vport_info->num_rx_q; + vport->num_rx_bufq = vport_info->num_rx_bufq; + vport->max_mtu = vport_info->max_mtu; + rte_memcpy(vport->default_mac_addr, + vport_info->default_mac_addr, ETH_ALEN); + vport->rss_algorithm = vport_info->rss_algorithm; + vport->rss_key_size = RTE_MIN(IDPF_RSS_KEY_LEN, +vport_info->rss_key_size); + vport->rss_lut_size = vport_info->rss_lut_size; + vport->sw_idx = idx; + + for (i = 0; i < vport_info->chunks.num_chunks; i++) { + if (vport_info->chunks.chunks[i].type == + VIRTCHNL2_QUEUE_TYPE_TX) { + vport->chunks_info.tx_start_qid = + vport_info->chunks.chunks[i].start_queue_id; + vport->chunks_info.tx_qtail_start = + vport_info->chunks.chunks[i].qtail_r
[PATCH 04/13] net/idpf: add queue operations
Add support for queue operations: - rx_queue_start - rx_queue_stop - tx_queue_start - tx_queue_stop - rx_queue_setup - rx_queue_release - tx_queue_setup - tx_queue_release Signed-off-by: Beilei Xing Signed-off-by: Xiaoyun Li Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 53 +- drivers/net/idpf/idpf_ethdev.h |7 + drivers/net/idpf/idpf_rxtx.c | 1264 drivers/net/idpf/idpf_rxtx.h | 182 + drivers/net/idpf/idpf_vchnl.c | 503 + drivers/net/idpf/meson.build |1 + 6 files changed, 2009 insertions(+), 1 deletion(-) create mode 100644 drivers/net/idpf/idpf_rxtx.c create mode 100644 drivers/net/idpf/idpf_rxtx.h diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index 87c68226dd..b302e42a9c 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -12,6 +12,7 @@ #include #include "idpf_ethdev.h" +#include "idpf_rxtx.h" #define REPRESENTOR"representor" @@ -33,10 +34,18 @@ static const struct eth_dev_ops idpf_eth_dev_ops = { .dev_start = idpf_dev_start, .dev_stop = idpf_dev_stop, .dev_close = idpf_dev_close, + .rx_queue_start = idpf_rx_queue_start, + .rx_queue_stop = idpf_rx_queue_stop, + .tx_queue_start = idpf_tx_queue_start, + .tx_queue_stop = idpf_tx_queue_stop, + .rx_queue_setup = idpf_rx_queue_setup, + .rx_queue_release = idpf_dev_rx_queue_release, + .tx_queue_setup = idpf_tx_queue_setup, + .tx_queue_release = idpf_dev_tx_queue_release, }; static int -idpf_init_vport_req_info(struct rte_eth_dev *dev) +idpf_init_vport_req_info(__rte_unused struct rte_eth_dev *dev) { struct virtchnl2_create_vport *vport_info; uint16_t idx = adapter->next_vport_idx; @@ -193,6 +202,39 @@ idpf_dev_configure(struct rte_eth_dev *dev) return ret; } +static int +idpf_start_queues(struct rte_eth_dev *dev) +{ + struct idpf_rx_queue *rxq; + struct idpf_tx_queue *txq; + int err = 0; + int i; + + for (i = 0; i < dev->data->nb_tx_queues; i++) { + txq = dev->data->tx_queues[i]; + if (!txq || txq->tx_deferred_start) + continue; + err = idpf_tx_queue_start(dev, i); + if (err) { + PMD_DRV_LOG(ERR, "Fail to start Tx queue %u", i); + return err; + } + } + + for (i = 0; i < dev->data->nb_rx_queues; i++) { + rxq = dev->data->rx_queues[i]; + if (!rxq || rxq->rx_deferred_start) + continue; + err = idpf_rx_queue_start(dev, i); + if (err) { + PMD_DRV_LOG(ERR, "Fail to start Rx queue %u", i); + return err; + } + } + + return err; +} + static int idpf_dev_start(struct rte_eth_dev *dev) { @@ -203,6 +245,11 @@ idpf_dev_start(struct rte_eth_dev *dev) vport->stopped = 0; + if (idpf_start_queues(dev)) { + PMD_DRV_LOG(ERR, "Failed to start queues"); + goto err_mtu; + } + if (idpf_ena_dis_vport(vport, true)) { PMD_DRV_LOG(ERR, "Failed to enable vport"); goto err_vport; @@ -211,6 +258,8 @@ idpf_dev_start(struct rte_eth_dev *dev) return 0; err_vport: + idpf_stop_queues(dev); +err_mtu: return -1; } @@ -228,6 +277,8 @@ idpf_dev_stop(struct rte_eth_dev *dev) if (idpf_ena_dis_vport(vport, false)) PMD_DRV_LOG(ERR, "disable vport failed"); + idpf_stop_queues(dev); + vport->stopped = 1; dev->data->dev_started = 0; diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h index 501f772fa8..25e0c5cae7 100644 --- a/drivers/net/idpf/idpf_ethdev.h +++ b/drivers/net/idpf/idpf_ethdev.h @@ -200,6 +200,13 @@ int idpf_check_api_version(struct idpf_adapter *adapter); int idpf_get_caps(struct idpf_adapter *adapter); int idpf_create_vport(__rte_unused struct rte_eth_dev *dev); int idpf_destroy_vport(struct idpf_vport *vport); +int idpf_config_rxqs(struct idpf_vport *vport); +int idpf_config_rxq(struct idpf_vport *vport, uint16_t rxq_id); +int idpf_config_txqs(struct idpf_vport *vport); +int idpf_config_txq(struct idpf_vport *vport, uint16_t txq_id); +int idpf_switch_queue(struct idpf_vport *vport, uint16_t qid, + bool rx, bool on); +int idpf_ena_dis_queues(struct idpf_vport *vport, bool enable); int idpf_ena_dis_vport(struct idpf_vport *vport, bool enable); #endif /* _IDPF_ETHDEV_H_ */ diff --git a/drivers/net/idpf/idpf_rxtx.c b/driver
[PATCH 05/13] net/idpf: add support to get device information
Add dev ops dev_infos_get. Signed-off-by: Beilei Xing Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 71 ++ 1 file changed, 71 insertions(+) diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index b302e42a9c..ab991f9181 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -28,6 +28,8 @@ static int idpf_dev_configure(struct rte_eth_dev *dev); static int idpf_dev_start(struct rte_eth_dev *dev); static int idpf_dev_stop(struct rte_eth_dev *dev); static int idpf_dev_close(struct rte_eth_dev *dev); +static int idpf_dev_info_get(struct rte_eth_dev *dev, +struct rte_eth_dev_info *dev_info); static const struct eth_dev_ops idpf_eth_dev_ops = { .dev_configure = idpf_dev_configure, @@ -42,8 +44,77 @@ static const struct eth_dev_ops idpf_eth_dev_ops = { .rx_queue_release = idpf_dev_rx_queue_release, .tx_queue_setup = idpf_tx_queue_setup, .tx_queue_release = idpf_dev_tx_queue_release, + .dev_infos_get = idpf_dev_info_get, }; +static int +idpf_dev_info_get(__rte_unused struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) +{ + dev_info->max_rx_queues = adapter->caps->max_rx_q; + dev_info->max_tx_queues = adapter->caps->max_tx_q; + dev_info->min_rx_bufsize = IDPF_MIN_BUF_SIZE; + dev_info->max_rx_pktlen = IDPF_MAX_FRAME_SIZE; + + dev_info->max_mtu = dev_info->max_rx_pktlen - IDPF_ETH_OVERHEAD; + dev_info->min_mtu = RTE_ETHER_MIN_MTU; + + dev_info->max_mac_addrs = IDPF_NUM_MACADDR_MAX; + dev_info->dev_capa = RTE_ETH_DEV_CAPA_RUNTIME_RX_QUEUE_SETUP | + RTE_ETH_DEV_CAPA_RUNTIME_TX_QUEUE_SETUP; + dev_info->rx_offload_capa = + RTE_ETH_RX_OFFLOAD_VLAN_STRIP | + RTE_ETH_RX_OFFLOAD_QINQ_STRIP | + RTE_ETH_RX_OFFLOAD_IPV4_CKSUM | + RTE_ETH_RX_OFFLOAD_UDP_CKSUM| + RTE_ETH_RX_OFFLOAD_TCP_CKSUM| + RTE_ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM | + RTE_ETH_RX_OFFLOAD_SCATTER | + RTE_ETH_RX_OFFLOAD_VLAN_FILTER | + RTE_ETH_RX_OFFLOAD_RSS_HASH; + + dev_info->tx_offload_capa = + RTE_ETH_TX_OFFLOAD_VLAN_INSERT | + RTE_ETH_TX_OFFLOAD_QINQ_INSERT | + RTE_ETH_TX_OFFLOAD_IPV4_CKSUM | + RTE_ETH_TX_OFFLOAD_UDP_CKSUM| + RTE_ETH_TX_OFFLOAD_TCP_CKSUM| + RTE_ETH_TX_OFFLOAD_SCTP_CKSUM | + RTE_ETH_TX_OFFLOAD_OUTER_IPV4_CKSUM | + RTE_ETH_TX_OFFLOAD_TCP_TSO | + RTE_ETH_TX_OFFLOAD_VXLAN_TNL_TSO| + RTE_ETH_TX_OFFLOAD_GRE_TNL_TSO | + RTE_ETH_TX_OFFLOAD_IPIP_TNL_TSO | + RTE_ETH_TX_OFFLOAD_GENEVE_TNL_TSO | + RTE_ETH_TX_OFFLOAD_MULTI_SEGS | + RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE; + + dev_info->default_rxconf = (struct rte_eth_rxconf) { + .rx_free_thresh = IDPF_DEFAULT_RX_FREE_THRESH, + .rx_drop_en = 0, + .offloads = 0, + }; + + dev_info->default_txconf = (struct rte_eth_txconf) { + .tx_free_thresh = IDPF_DEFAULT_TX_FREE_THRESH, + .tx_rs_thresh = IDPF_DEFAULT_TX_RS_THRESH, + .offloads = 0, + }; + + dev_info->rx_desc_lim = (struct rte_eth_desc_lim) { + .nb_max = IDPF_MAX_RING_DESC, + .nb_min = IDPF_MIN_RING_DESC, + .nb_align = IDPF_ALIGN_RING_DESC, + }; + + dev_info->tx_desc_lim = (struct rte_eth_desc_lim) { + .nb_max = IDPF_MAX_RING_DESC, + .nb_min = IDPF_MIN_RING_DESC, + .nb_align = IDPF_ALIGN_RING_DESC, + }; + + return 0; +} + static int idpf_init_vport_req_info(__rte_unused struct rte_eth_dev *dev) { -- 2.25.1
[PATCH 06/13] net/idpf: add support to get packet type
Add dev ops dev_supported_ptypes_get. Signed-off-by: Beilei Xing Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 3 ++ drivers/net/idpf/idpf_rxtx.c | 51 ++ drivers/net/idpf/idpf_rxtx.h | 3 ++ 3 files changed, 57 insertions(+) diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index ab991f9181..cc70e2e03a 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -32,6 +32,7 @@ static int idpf_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info); static const struct eth_dev_ops idpf_eth_dev_ops = { + .dev_supported_ptypes_get = idpf_dev_supported_ptypes_get, .dev_configure = idpf_dev_configure, .dev_start = idpf_dev_start, .dev_stop = idpf_dev_stop, @@ -488,6 +489,8 @@ idpf_adapter_init(struct rte_eth_dev *dev) if (adapter->initialized) return 0; + idpf_set_default_ptype_table(dev); + hw->hw_addr = (void *)pci_dev->mem_resource[0].addr; hw->hw_addr_len = pci_dev->mem_resource[0].len; hw->back = adapter; diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c index 7d5428b750..eae762ced3 100644 --- a/drivers/net/idpf/idpf_rxtx.c +++ b/drivers/net/idpf/idpf_rxtx.c @@ -8,6 +8,57 @@ #include "idpf_ethdev.h" #include "idpf_rxtx.h" +const uint32_t * +idpf_dev_supported_ptypes_get(struct rte_eth_dev *dev __rte_unused) +{ + static const uint32_t ptypes[] = { + RTE_PTYPE_L2_ETHER, + RTE_PTYPE_L3_IPV4_EXT_UNKNOWN, + RTE_PTYPE_L3_IPV6_EXT_UNKNOWN, + RTE_PTYPE_L4_FRAG, + RTE_PTYPE_L4_NONFRAG, + RTE_PTYPE_L4_UDP, + RTE_PTYPE_L4_TCP, + RTE_PTYPE_L4_SCTP, + RTE_PTYPE_L4_ICMP, + RTE_PTYPE_UNKNOWN + }; + + return ptypes; +} + +static inline uint32_t +idpf_get_default_pkt_type(uint16_t ptype) +{ + static const uint32_t type_table[IDPF_MAX_PKT_TYPE] + __rte_cache_aligned = { + [1] = RTE_PTYPE_L2_ETHER, + [22] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | RTE_PTYPE_L4_FRAG, + [23] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4, + [24] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | RTE_PTYPE_L4_UDP, + [26] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | RTE_PTYPE_L4_TCP, + [27] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | RTE_PTYPE_L4_SCTP, + [28] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | RTE_PTYPE_L4_ICMP, + [88] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | RTE_PTYPE_L4_FRAG, + [89] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6, + [90] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | RTE_PTYPE_L4_UDP, + [92] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | RTE_PTYPE_L4_TCP, + [93] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | RTE_PTYPE_L4_SCTP, + [94] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | RTE_PTYPE_L4_ICMP, + }; + + return type_table[ptype]; +} + +void __rte_cold +idpf_set_default_ptype_table(struct rte_eth_dev *dev __rte_unused) +{ + int i; + + for (i = 0; i < IDPF_MAX_PKT_TYPE; i++) + adapter->ptype_tbl[i] = idpf_get_default_pkt_type(i); +} + static inline int check_rx_thresh(uint16_t nb_desc, uint16_t thresh) { diff --git a/drivers/net/idpf/idpf_rxtx.h b/drivers/net/idpf/idpf_rxtx.h index 3997082b21..5e57995f8c 100644 --- a/drivers/net/idpf/idpf_rxtx.h +++ b/drivers/net/idpf/idpf_rxtx.h @@ -178,5 +178,8 @@ void idpf_dev_tx_queue_release(struct rte_eth_dev *dev, uint16_t qid); void idpf_stop_queues(struct rte_eth_dev *dev); +void idpf_set_default_ptype_table(struct rte_eth_dev *dev); +const uint32_t *idpf_dev_supported_ptypes_get(struct rte_eth_dev *dev); + #endif /* _IDPF_RXTX_H_ */ -- 2.25.1
[PATCH 07/13] net/idpf: add support to update link status
Add dev ops link_update. Signed-off-by: Beilei Xing Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 22 ++ drivers/net/idpf/idpf_ethdev.h | 2 ++ 2 files changed, 24 insertions(+) diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index cc70e2e03a..668e488843 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -31,6 +31,27 @@ static int idpf_dev_close(struct rte_eth_dev *dev); static int idpf_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info); +int +idpf_dev_link_update(struct rte_eth_dev *dev, +__rte_unused int wait_to_complete) +{ + struct idpf_vport *vport = + (struct idpf_vport *)dev->data->dev_private; + struct rte_eth_link new_link; + + memset(&new_link, 0, sizeof(new_link)); + + new_link.link_speed = RTE_ETH_SPEED_NUM_NONE; + + new_link.link_duplex = RTE_ETH_LINK_FULL_DUPLEX; + new_link.link_status = vport->link_up ? RTE_ETH_LINK_UP : + RTE_ETH_LINK_DOWN; + new_link.link_autoneg = !(dev->data->dev_conf.link_speeds & + RTE_ETH_LINK_SPEED_FIXED); + + return rte_eth_linkstatus_set(dev, &new_link); +} + static const struct eth_dev_ops idpf_eth_dev_ops = { .dev_supported_ptypes_get = idpf_dev_supported_ptypes_get, .dev_configure = idpf_dev_configure, @@ -46,6 +67,7 @@ static const struct eth_dev_ops idpf_eth_dev_ops = { .tx_queue_setup = idpf_tx_queue_setup, .tx_queue_release = idpf_dev_tx_queue_release, .dev_infos_get = idpf_dev_info_get, + .link_update= idpf_dev_link_update, }; static int diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h index 25e0c5cae7..130bb7ad2d 100644 --- a/drivers/net/idpf/idpf_ethdev.h +++ b/drivers/net/idpf/idpf_ethdev.h @@ -195,6 +195,8 @@ _atomic_set_cmd(struct idpf_adapter *adapter, enum virtchnl_ops ops) return !ret; } +int idpf_dev_link_update(struct rte_eth_dev *dev, +__rte_unused int wait_to_complete); void idpf_handle_virtchnl_msg(struct rte_eth_dev *dev); int idpf_check_api_version(struct idpf_adapter *adapter); int idpf_get_caps(struct idpf_adapter *adapter); -- 2.25.1
[PATCH 08/13] net/idpf: add basic Rx/Tx datapath
Add basic RX & TX support in split queue mode and single queue mode. Split queue mode is selected by default. Signed-off-by: Beilei Xing Signed-off-by: Xiaoyun Li Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 69 ++- drivers/net/idpf/idpf_rxtx.c | 896 + drivers/net/idpf/idpf_rxtx.h | 35 ++ 3 files changed, 998 insertions(+), 2 deletions(-) diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index 668e488843..55ec24872e 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -14,12 +14,16 @@ #include "idpf_ethdev.h" #include "idpf_rxtx.h" +#define IDPF_TX_SINGLE_Q "tx_single" +#define IDPF_RX_SINGLE_Q "rx_single" #define REPRESENTOR"representor" struct idpf_adapter *adapter; uint16_t used_vecs_num; static const char * const idpf_valid_args[] = { + IDPF_TX_SINGLE_Q, + IDPF_RX_SINGLE_Q, REPRESENTOR, NULL }; @@ -157,6 +161,30 @@ idpf_init_vport_req_info(__rte_unused struct rte_eth_dev *dev) (struct virtchnl2_create_vport *)adapter->vport_req_info[idx]; vport_info->vport_type = rte_cpu_to_le_16(VIRTCHNL2_VPORT_TYPE_DEFAULT); + if (!adapter->txq_model) { + vport_info->txq_model = + rte_cpu_to_le_16(VIRTCHNL2_QUEUE_MODEL_SPLIT); + vport_info->num_tx_q = dev->data->nb_tx_queues; + vport_info->num_tx_complq = + dev->data->nb_tx_queues * IDPF_TX_COMPLQ_PER_GRP; + } else { + vport_info->txq_model = + rte_cpu_to_le_16(VIRTCHNL2_QUEUE_MODEL_SINGLE); + vport_info->num_tx_q = dev->data->nb_tx_queues; + vport_info->num_tx_complq = 0; + } + if (!adapter->rxq_model) { + vport_info->rxq_model = + rte_cpu_to_le_16(VIRTCHNL2_QUEUE_MODEL_SPLIT); + vport_info->num_rx_q = dev->data->nb_rx_queues; + vport_info->num_rx_bufq = + dev->data->nb_rx_queues * IDPF_RX_BUFQ_PER_GRP; + } else { + vport_info->rxq_model = + rte_cpu_to_le_16(VIRTCHNL2_QUEUE_MODEL_SINGLE); + vport_info->num_rx_q = dev->data->nb_rx_queues; + vport_info->num_rx_bufq = 0; + } return 0; } @@ -344,6 +372,9 @@ idpf_dev_start(struct rte_eth_dev *dev) goto err_mtu; } + idpf_set_rx_function(dev); + idpf_set_tx_function(dev); + if (idpf_ena_dis_vport(vport, true)) { PMD_DRV_LOG(ERR, "Failed to enable vport"); goto err_vport; @@ -394,11 +425,31 @@ idpf_dev_close(struct rte_eth_dev *dev) return 0; } +static int +parse_bool(const char *key, const char *value, void *args) +{ + int *i = (int *)args; + char *end; + int num; + + num = strtoul(value, &end, 10); + + if (num != 0 && num != 1) { + PMD_DRV_LOG(WARNING, "invalid value:\"%s\" for key:\"%s\", " + "value must be 0 or 1", + value, key); + return -1; + } + + *i = num; + return 0; +} + static int idpf_parse_devargs(struct rte_eth_dev *dev) { struct rte_devargs *devargs = dev->device->devargs; struct rte_kvargs *kvlist; - int ret = 0; + int ret; if (!devargs) return 0; @@ -409,6 +460,17 @@ static int idpf_parse_devargs(struct rte_eth_dev *dev) return -EINVAL; } + ret = rte_kvargs_process(kvlist, IDPF_TX_SINGLE_Q, &parse_bool, +&adapter->txq_model); + if (ret) + goto bail; + + ret = rte_kvargs_process(kvlist, IDPF_RX_SINGLE_Q, &parse_bool, +&adapter->rxq_model); + if (ret) + goto bail; + +bail: rte_kvargs_free(kvlist); return ret; } @@ -637,8 +699,11 @@ idpf_dev_init(struct rte_eth_dev *dev, __rte_unused void *init_params) /* for secondary processes, we don't initialise any further as primary * has already done this work. */ - if (rte_eal_process_type() != RTE_PROC_PRIMARY) + if (rte_eal_process_type() != RTE_PROC_PRIMARY) { + idpf_set_rx_function(dev); + idpf_set_tx_function(dev); return ret; + } ret = idpf_adapter_init(dev); if (ret) { diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c index eae762ced3..e6040cece1 100644 --- a/drivers/net/idpf/idpf_rxtx.c +++ b/drivers/net/idpf/idpf_rxtx.c @@ -1312,4 +1312,900 @@ idpf_stop_queues(struct rte_eth_dev *dev) } } +#define IDPF_RX_ERR0_QW1 \ + (BIT(VIRTCHNL2_RX_FLEX_DESC_ADV_STATUS0_XSUM_IPE_S) | \ +BIT(VIRTCHNL2_RX_FLEX_DESC_A
[PATCH 09/13] net/idpf: add support for RSS
Add RSS support. Signed-off-by: Beilei Xing Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 108 + drivers/net/idpf/idpf_ethdev.h | 18 ++ drivers/net/idpf/idpf_vchnl.c | 93 3 files changed, 219 insertions(+) diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index 55ec24872e..0b03ec860a 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -85,6 +85,7 @@ idpf_dev_info_get(__rte_unused struct rte_eth_dev *dev, struct rte_eth_dev_info dev_info->max_mtu = dev_info->max_rx_pktlen - IDPF_ETH_OVERHEAD; dev_info->min_mtu = RTE_ETHER_MIN_MTU; + dev_info->flow_type_rss_offloads = IDPF_RSS_OFFLOAD_ALL; dev_info->max_mac_addrs = IDPF_NUM_MACADDR_MAX; dev_info->dev_capa = RTE_ETH_DEV_CAPA_RUNTIME_RX_QUEUE_SETUP | RTE_ETH_DEV_CAPA_RUNTIME_TX_QUEUE_SETUP; @@ -293,6 +294,91 @@ idpf_init_vport(struct rte_eth_dev *dev) return 0; } +static int +idpf_config_rss(struct idpf_vport *vport) +{ + int ret; + + ret = idpf_set_rss_key(vport); + if (ret) { + PMD_INIT_LOG(ERR, "Failed to configure RSS key"); + return ret; + } + + ret = idpf_set_rss_lut(vport); + if (ret) { + PMD_INIT_LOG(ERR, "Failed to configure RSS lut"); + return ret; + } + + ret = idpf_set_rss_hash(vport); + if (ret) { + PMD_INIT_LOG(ERR, "Failed to configure RSS hash"); + return ret; + } + + return ret; +} + +static int +idpf_init_rss(struct idpf_vport *vport) +{ + struct rte_eth_rss_conf *rss_conf; + uint16_t i, nb_q, lut_size; + int ret = 0; + + rss_conf = &vport->dev_data->dev_conf.rx_adv_conf.rss_conf; + nb_q = vport->num_rx_q; + + vport->rss_key = (uint8_t *)rte_zmalloc("rss_key", +vport->rss_key_size, 0); + if (!vport->rss_key) { + PMD_INIT_LOG(ERR, "Failed to allocate RSS key"); + ret = -ENOMEM; + goto err_key; + } + + lut_size = vport->rss_lut_size; + vport->rss_lut = (uint32_t *)rte_zmalloc("rss_lut", + sizeof(uint32_t) * lut_size, 0); + if (!vport->rss_lut) { + PMD_INIT_LOG(ERR, "Failed to allocate RSS lut"); + ret = -ENOMEM; + goto err_lut; + } + + if (!rss_conf->rss_key) { + for (i = 0; i < vport->rss_key_size; i++) + vport->rss_key[i] = (uint8_t)rte_rand(); + } else { + rte_memcpy(vport->rss_key, rss_conf->rss_key, + RTE_MIN(rss_conf->rss_key_len, + vport->rss_key_size)); + } + + for (i = 0; i < lut_size; i++) + vport->rss_lut[i] = i % nb_q; + + vport->rss_hf = IECM_DEFAULT_RSS_HASH_EXPANDED; + + ret = idpf_config_rss(vport); + if (ret) { + PMD_INIT_LOG(ERR, "Failed to configure RSS"); + goto err_cfg; + } + + return ret; + +err_cfg: + rte_free(vport->rss_lut); + vport->rss_lut = NULL; +err_lut: + rte_free(vport->rss_key); + vport->rss_key = NULL; +err_key: + return ret; +} + static int idpf_dev_configure(struct rte_eth_dev *dev) { @@ -300,6 +386,10 @@ idpf_dev_configure(struct rte_eth_dev *dev) (struct idpf_vport *)dev->data->dev_private; int ret = 0; + if (dev->data->dev_conf.rxmode.mq_mode & RTE_ETH_MQ_RX_RSS_FLAG) + dev->data->dev_conf.rxmode.offloads |= + RTE_ETH_RX_OFFLOAD_RSS_HASH; + ret = idpf_init_vport_req_info(dev); if (ret) { PMD_INIT_LOG(ERR, "Failed to init vport req_info."); @@ -321,6 +411,14 @@ idpf_dev_configure(struct rte_eth_dev *dev) rte_ether_addr_copy((struct rte_ether_addr *)vport->default_mac_addr, &dev->data->mac_addrs[0]); + if (adapter->caps->rss_caps) { + ret = idpf_init_rss(vport); + if (ret) { + PMD_INIT_LOG(ERR, "Failed to init rss"); + return ret; + } + } + return ret; } @@ -422,6 +520,16 @@ idpf_dev_close(struct rte_eth_dev *dev) idpf_dev_stop(dev); idpf_destroy_vport(vport); + if (vport->rss_lut) { + rte_free(vport->rss_lut); + vport->rss_lut = NULL; + } + + if (vport->rss_key) { + rte_free(vport->rss_key); + vport->rss_key = NULL; + } + return 0; } diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h index 130bb7ad2d..22096c4185 100644 --- a/drivers/net/idpf/idpf_ethdev.h +++ b/drivers/net/idpf/idpf_ethdev.h @@ -47
[PATCH 10/13] net/idpf: add mtu configuration
Add dev ops mtu_set. Signed-off-by: Beilei Xing Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 21 + 1 file changed, 21 insertions(+) diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index 0b03ec860a..4b8e3c4d5b 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -34,6 +34,7 @@ static int idpf_dev_stop(struct rte_eth_dev *dev); static int idpf_dev_close(struct rte_eth_dev *dev); static int idpf_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info); +static int idpf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu); int idpf_dev_link_update(struct rte_eth_dev *dev, @@ -72,6 +73,7 @@ static const struct eth_dev_ops idpf_eth_dev_ops = { .tx_queue_release = idpf_dev_tx_queue_release, .dev_infos_get = idpf_dev_info_get, .link_update= idpf_dev_link_update, + .mtu_set= idpf_dev_mtu_set, }; static int @@ -143,6 +145,18 @@ idpf_dev_info_get(__rte_unused struct rte_eth_dev *dev, struct rte_eth_dev_info return 0; } +static int +idpf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu __rte_unused) +{ + /* mtu setting is forbidden if port is start */ + if (dev->data->dev_started) { + PMD_DRV_LOG(ERR, "port must be stopped before configuration"); + return -EBUSY; + } + + return 0; +} + static int idpf_init_vport_req_info(__rte_unused struct rte_eth_dev *dev) { @@ -465,6 +479,13 @@ idpf_dev_start(struct rte_eth_dev *dev) vport->stopped = 0; + if (dev->data->mtu > vport->max_mtu) { + PMD_DRV_LOG(ERR, "MTU should be less than %d", vport->max_mtu); + goto err_mtu; + } + + vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD; + if (idpf_start_queues(dev)) { PMD_DRV_LOG(ERR, "Failed to start queues"); goto err_mtu; -- 2.25.1
[PATCH 11/13] net/idpf: add hw statistics
Add hardware packets/bytes statisticsi query and reset. Signed-off-by: Mingxia Liu Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 82 +- drivers/net/idpf/idpf_ethdev.h | 2 + drivers/net/idpf/idpf_vchnl.c | 26 +++ 3 files changed, 109 insertions(+), 1 deletion(-) diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index 4b8e3c4d5b..b934488d0b 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -35,6 +35,9 @@ static int idpf_dev_close(struct rte_eth_dev *dev); static int idpf_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info); static int idpf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu); +static int idpf_dev_stats_get(struct rte_eth_dev *dev, + struct rte_eth_stats *stats); +static int idpf_dev_stats_reset(struct rte_eth_dev *dev); int idpf_dev_link_update(struct rte_eth_dev *dev, @@ -74,6 +77,8 @@ static const struct eth_dev_ops idpf_eth_dev_ops = { .dev_infos_get = idpf_dev_info_get, .link_update= idpf_dev_link_update, .mtu_set= idpf_dev_mtu_set, + .stats_get = idpf_dev_stats_get, + .stats_reset= idpf_dev_stats_reset, }; static int @@ -157,8 +162,79 @@ idpf_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu __rte_unused) return 0; } +static void +idpf_stat_update(uint64_t *offset, uint64_t *stat) +{ + *stat = *stat - *offset; +} + + +static void +idpf_update_stats(struct virtchnl2_vport_stats *oes, struct virtchnl2_vport_stats *nes) +{ + idpf_stat_update(&oes->rx_bytes, &nes->rx_bytes); + idpf_stat_update(&oes->rx_unicast, &nes->rx_unicast); + idpf_stat_update(&oes->rx_multicast, &nes->rx_multicast); + idpf_stat_update(&oes->rx_broadcast, &nes->rx_broadcast); + idpf_stat_update(&oes->rx_discards, &nes->rx_discards); + idpf_stat_update(&oes->tx_bytes, &nes->tx_bytes); + idpf_stat_update(&oes->tx_unicast, &nes->tx_unicast); + idpf_stat_update(&oes->tx_multicast, &nes->tx_multicast); + idpf_stat_update(&oes->tx_broadcast, &nes->tx_broadcast); + idpf_stat_update(&oes->tx_errors, &nes->tx_errors); + idpf_stat_update(&oes->tx_discards, &nes->tx_discards); +} + static int -idpf_init_vport_req_info(__rte_unused struct rte_eth_dev *dev) +idpf_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats) +{ + struct idpf_vport *vport = + (struct idpf_vport *)dev->data->dev_private; + struct virtchnl2_vport_stats *pstats = NULL; + int ret; + + ret = idpf_query_stats(vport, &pstats); + if (ret == 0) { + uint8_t crc_stats_len = (dev->data->dev_conf.rxmode.offloads & +RTE_ETH_RX_OFFLOAD_KEEP_CRC) ? 0 : +RTE_ETHER_CRC_LEN; + idpf_update_stats(&vport->eth_stats_offset, pstats); + stats->ipackets = pstats->rx_unicast + pstats->rx_multicast + + pstats->rx_broadcast - pstats->rx_discards; + stats->opackets = pstats->tx_broadcast + pstats->tx_multicast + + pstats->tx_unicast; + stats->imissed = pstats->rx_discards; + stats->oerrors = pstats->tx_errors + pstats->tx_discards; + stats->ibytes = pstats->rx_bytes; + stats->ibytes -= stats->ipackets * crc_stats_len; + stats->obytes = pstats->tx_bytes; + } else { + PMD_DRV_LOG(ERR, "Get statistics failed"); + } + return ret; +} + + +static int +idpf_dev_stats_reset(struct rte_eth_dev *dev) +{ + struct idpf_vport *vport = + (struct idpf_vport *)dev->data->dev_private; + struct virtchnl2_vport_stats *pstats = NULL; + int ret; + + ret = idpf_query_stats(vport, &pstats); + if (ret != 0) + return ret; + + /* set stats offset base on current values */ + vport->eth_stats_offset = *pstats; + + return 0; +} + +static int +idpf_init_vport_req_info(struct rte_eth_dev *dev) { struct virtchnl2_create_vport *vport_info; uint16_t idx = adapter->next_vport_idx; @@ -499,6 +575,10 @@ idpf_dev_start(struct rte_eth_dev *dev) goto err_vport; } + if (idpf_dev_stats_reset(dev)) { + PMD_DRV_LOG(ERR, "Failed to reset stats"); + goto err_vport; + } return 0; err_vport: diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h index 22096c4185..c0cbf4c3c6 100644 --- a/drivers/net/idpf/idpf_ethdev.h +++ b/drivers/net/idpf/idpf_ethdev.h @@ -227,6 +227,8 @@ int idpf_switch_queue(struct idpf_vport *vport, uint16_t qid,
[PATCH 12/13] net/idpf: support write back based on ITR expire
Force write-backs by setting WB_ON_ITR bit in DYN_CTL register, so that the packets can be received once at a time. Signed-off-by: Beilei Xing Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 117 + drivers/net/idpf/idpf_ethdev.h | 8 +++ drivers/net/idpf/idpf_vchnl.c | 108 ++ 3 files changed, 233 insertions(+) diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index b934488d0b..1e9564a5a9 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -512,6 +512,87 @@ idpf_dev_configure(struct rte_eth_dev *dev) return ret; } +static int +idpf_config_rx_queues_irqs(struct rte_eth_dev *dev) +{ + struct idpf_vport *vport = + (struct idpf_vport *)dev->data->dev_private; + struct virtchnl2_queue_vector *qv_map; + struct iecm_hw *hw = &adapter->hw; + uint32_t dynctl_reg_start; + uint32_t itrn_reg_start; + uint32_t dynctl_val, itrn_val; + uint16_t i; + + qv_map = rte_zmalloc("qv_map", + dev->data->nb_rx_queues * + sizeof(struct virtchnl2_queue_vector), 0); + if (!qv_map) { + PMD_DRV_LOG(ERR, "Failed to allocate %d queue-vector map", + dev->data->nb_rx_queues); + goto qv_map_alloc_err; + } + + /* Rx interrupt disabled, Map interrupt only for writeback */ + + /* The capability flags adapter->caps->other_caps here should be +* compared with bit VIRTCHNL2_CAP_WB_ON_ITR. The if condition should +* be updated when the FW can return correct flag bits. +*/ + if (adapter->caps->other_caps) { + dynctl_reg_start = vport->recv_vectors->vchunks.vchunks->dynctl_reg_start; + itrn_reg_start = vport->recv_vectors->vchunks.vchunks->itrn_reg_start; + dynctl_val = IECM_READ_REG(hw, dynctl_reg_start); + PMD_DRV_LOG(DEBUG, "Value of dynctl_reg_start is 0x%x", dynctl_val); + itrn_val = IECM_READ_REG(hw, itrn_reg_start); + PMD_DRV_LOG(DEBUG, "Value of itrn_reg_start is 0x%x", itrn_val); + /* Force write-backs by setting WB_ON_ITR bit in DYN_CTL +* register. WB_ON_ITR and INTENA are mutually exclusive +* bits. Setting WB_ON_ITR bits means TX and RX Descs +* are writen back based on ITR expiration irrespective +* of INTENA setting. +*/ + /* TBD: need to tune INTERVAL value for better performance. */ + if (itrn_val) + IECM_WRITE_REG(hw, + dynctl_reg_start, + VIRTCHNL2_ITR_IDX_0 << + PF_GLINT_DYN_CTL_ITR_INDX_S | + PF_GLINT_DYN_CTL_WB_ON_ITR_M | + itrn_val << + PF_GLINT_DYN_CTL_INTERVAL_S); + else + IECM_WRITE_REG(hw, + dynctl_reg_start, + VIRTCHNL2_ITR_IDX_0 << + PF_GLINT_DYN_CTL_ITR_INDX_S | + PF_GLINT_DYN_CTL_WB_ON_ITR_M | + IDPF_DFLT_INTERVAL << + PF_GLINT_DYN_CTL_INTERVAL_S); + } + for (i = 0; i < dev->data->nb_rx_queues; i++) { + /* map all queues to the same vector */ + qv_map[i].queue_id = vport->chunks_info.rx_start_qid + i; + qv_map[i].vector_id = + vport->recv_vectors->vchunks.vchunks->start_vector_id; + } + vport->qv_map = qv_map; + + if (idpf_config_irq_map_unmap(vport, true)) { + PMD_DRV_LOG(ERR, "config interrupt mapping failed"); + goto config_irq_map_err; + } + + return 0; + +config_irq_map_err: + rte_free(vport->qv_map); + vport->qv_map = NULL; + +qv_map_alloc_err: + return -1; +} + static int idpf_start_queues(struct rte_eth_dev *dev) { @@ -550,6 +631,9 @@ idpf_dev_start(struct rte_eth_dev *dev) { struct idpf_vport *vport = (struct idpf_vport *)dev->data->dev_private; + uint16_t num_allocated_vectors = + adapter->caps->num_allocated_vectors; + uint16_t req_vecs_num; PMD_INIT_FUNC_TRACE(); @@ -562,6 +646,23 @@ idpf_dev_start(struct rte_eth_dev *dev) vport->max_pkt_len = dev->data->mtu + IDPF_ETH_OVERHEAD; + req_vecs_num = IDPF_DFLT_Q_VEC_NUM; + if (req_vecs_num + used_vecs_num > num_allocated_vectors) { + PMD_DRV_LOG(ERR, "The accumulated request vectors' number should be less than %d", + num_all
[PATCH 13/13] net/idpf: add AVX512 data path for single queue model
Add support of AVX512 vector data path for single queue model. Signed-off-by: Wenjun Wu Signed-off-by: Junfeng Guo --- drivers/net/idpf/idpf_ethdev.c | 5 + drivers/net/idpf/idpf_ethdev.h | 5 + drivers/net/idpf/idpf_rxtx.c| 137 drivers/net/idpf/idpf_rxtx.h| 8 + drivers/net/idpf/idpf_rxtx_vec_avx512.c | 917 drivers/net/idpf/idpf_rxtx_vec_common.h | 85 +++ drivers/net/idpf/meson.build| 28 + 7 files changed, 1185 insertions(+) create mode 100644 drivers/net/idpf/idpf_rxtx_vec_avx512.c create mode 100644 drivers/net/idpf/idpf_rxtx_vec_common.h diff --git a/drivers/net/idpf/idpf_ethdev.c b/drivers/net/idpf/idpf_ethdev.c index 1e9564a5a9..590f1b25aa 100644 --- a/drivers/net/idpf/idpf_ethdev.c +++ b/drivers/net/idpf/idpf_ethdev.c @@ -147,6 +147,11 @@ idpf_dev_info_get(__rte_unused struct rte_eth_dev *dev, struct rte_eth_dev_info .nb_align = IDPF_ALIGN_RING_DESC, }; + dev_info->default_rxportconf.burst_size = IDPF_RX_MAX_BURST; + dev_info->default_txportconf.burst_size = IDPF_TX_MAX_BURST; + dev_info->default_rxportconf.nb_queues = 1; + dev_info->default_txportconf.nb_queues = 1; + return 0; } diff --git a/drivers/net/idpf/idpf_ethdev.h b/drivers/net/idpf/idpf_ethdev.h index 32520c03bb..7ad60f36a8 100644 --- a/drivers/net/idpf/idpf_ethdev.h +++ b/drivers/net/idpf/idpf_ethdev.h @@ -164,6 +164,11 @@ struct idpf_adapter { bool initialized; bool stopped; + + bool rx_vec_allowed; + bool tx_vec_allowed; + bool rx_use_avx512; + bool tx_use_avx512; }; extern struct idpf_adapter *adapter; diff --git a/drivers/net/idpf/idpf_rxtx.c b/drivers/net/idpf/idpf_rxtx.c index e6040cece1..af23857f2a 100644 --- a/drivers/net/idpf/idpf_rxtx.c +++ b/drivers/net/idpf/idpf_rxtx.c @@ -4,9 +4,11 @@ #include #include +#include #include "idpf_ethdev.h" #include "idpf_rxtx.h" +#include "idpf_rxtx_vec_common.h" const uint32_t * idpf_dev_supported_ptypes_get(struct rte_eth_dev *dev __rte_unused) @@ -2177,16 +2179,111 @@ idpf_prep_pkts(__rte_unused void *tx_queue, struct rte_mbuf **tx_pkts, return i; } +static void __rte_cold +release_rxq_mbufs_vec(struct idpf_rx_queue *rxq) +{ + const uint16_t mask = rxq->nb_rx_desc - 1; + uint16_t i; + + if (!rxq->sw_ring || rxq->rxrearm_nb >= rxq->nb_rx_desc) + return; + + /* free all mbufs that are valid in the ring */ + if (rxq->rxrearm_nb == 0) { + for (i = 0; i < rxq->nb_rx_desc; i++) { + if (rxq->sw_ring[i]) + rte_pktmbuf_free_seg(rxq->sw_ring[i]); + } + } else { + for (i = rxq->rx_tail; i != rxq->rxrearm_start; i = (i + 1) & mask) { + if (rxq->sw_ring[i]) + rte_pktmbuf_free_seg(rxq->sw_ring[i]); + } + } + + rxq->rxrearm_nb = rxq->nb_rx_desc; + + /* set all entries to NULL */ + memset(rxq->sw_ring, 0, sizeof(rxq->sw_ring[0]) * rxq->nb_rx_desc); +} + +static const struct idpf_rxq_ops def_singleq_rx_ops_vec = { + .release_mbufs = release_rxq_mbufs_vec, +}; + +static inline int +idpf_singleq_rx_vec_setup_default(struct idpf_rx_queue *rxq) +{ + uintptr_t p; + struct rte_mbuf mb_def = { .buf_addr = 0 }; /* zeroed mbuf */ + + mb_def.nb_segs = 1; + mb_def.data_off = RTE_PKTMBUF_HEADROOM; + mb_def.port = rxq->port_id; + rte_mbuf_refcnt_set(&mb_def, 1); + + /* prevent compiler reordering: rearm_data covers previous fields */ + rte_compiler_barrier(); + p = (uintptr_t)&mb_def.rearm_data; + rxq->mbuf_initializer = *(uint64_t *)p; + return 0; +} + +int __rte_cold +idpf_singleq_rx_vec_setup(struct idpf_rx_queue *rxq) +{ + rxq->ops = &def_singleq_rx_ops_vec; + return idpf_singleq_rx_vec_setup_default(rxq); +} + void idpf_set_rx_function(struct rte_eth_dev *dev) { struct idpf_vport *vport = (struct idpf_vport *)dev->data->dev_private; + struct idpf_adapter *ad = vport->adapter; + struct idpf_rx_queue *rxq; + int i; + +#ifdef RTE_ARCH_X86 + if (rte_eal_process_type() == RTE_PROC_PRIMARY) { + if (idpf_rx_vec_dev_check_default(dev) == IDPF_VECTOR_PATH && + rte_vect_get_max_simd_bitwidth() >= RTE_VECT_SIMD_128) { + ad->rx_vec_allowed = true; + + if (rte_vect_get_max_simd_bitwidth() >= RTE_VECT_SIMD_512) +#ifdef CC_AVX512_SUPPORT + if (rte_cpu_get_flag_enabled(RTE_CPUFLAG_AVX512F) == 1 && + rte_cpu_get_flag_enabled(RTE_CPUFLAG_AVX512BW) == 1) + ad->rx_use_avx512 = true; +#else + PMD_DRV_LOG(NOTICE, +
[dpdk-dev] [RFC PATCH 0/1] mldev: introduce machine learning device library
From: Jerin Jacob Machine learning inference library == Definition of machine learning inference Inference in machine learning is the process of making an output prediction based on new input data using a pre-trained machine learning model. The scope of the RFC would include only inferencing with pre-trained machine learning models, training and building/compiling the ML models is out of scope for this RFC or DPDK mldev API. Use existing machine learning compiler frameworks for model creation. Motivation for the new library -- Multiple semiconductor vendors are offering accelerator products such as DPU (often called Smart-NIC), FPGA, GPU, etc., which have ML inferencing capabilities integrated as part of the product. Use of ML inferencing is increasing in the domain of packet processing for flow classification, intrusion, malware and anomaly detection. Lack of inferencing support through DPDK APIs will involve complexities and increased latency from moving data across frameworks (i.e, dataplane to non dataplane ML frameworks and vice-versa). Having a standardized DPDK APIs for ML inferencing would enable the dataplane solutions to harness the benefit of inline inferencing supported by the hardware. Contents of RFC --- This RFC attempts to define standard APIs for: 1) Discovery of ML capabilities (e.g., device specific features) in a vendor independent fashion 2) Definition of functions to handle ML devices, which includes probing, initialization and termination of the devices. 3) Definition of functions to handle ML models used to perform inference operations. 4) Definition of function to handle quantize and dequantize operations Roadmap --- 1) Address the comments for this RFC. 2) Common code for mldev 3) SW mldev driver based on TVM (https://tvm.apache.org/) 4) HW mldev driver for cn10k 3) Add app/test-mldev application similar to other device class tests Machine learning library framework -- The ML framework is built on the following model: +-+ rte_ml_[en|de]queue_burst() | | | | Machine o--+ ++| | Learning| | | queue ||+--+ | Inference o--+-o|<===o===>|Core 0| | Engine | | | pair 0 | +--+ | o+ | ++ | || | +-+| | ++ ^ | | | queue | +--+ | | +-o|<===>|Core 1| | | | pair 1 | +--+ | | ++ +++| | +-+ || ++ | | Model 0 | || | queue | +--+ | +-+ |+---o|<===>|Core N| | +-+ || pair N | +--+ | | Model 1 | |++ | +-+ | | +-+ |<--- rte_ml_model_load() | | Model .. | |---> rte_ml_model_info() | +-+ |<--- rte_ml_model_start() | +-+ |<--- rte_ml_model_stop() | | Model N | |<--- rte_ml_model_params_update() | +-+ |<--- rte_ml_model_unload() +-+ ML Device: A hardware or software-based implementation of ML device API for running inferences using a pre-trained ML model. ML Model: An ML model is an algorithm trained over a dataset. A model consists of procedure/algorithm and data/pattern required to make predictions on live data. Once the model is created and trained outside of the DPDK scope, the model can be loaded via rte_ml_model_load() and then start it using rte_ml_model_start() API. The rte_ml_model_params_update() can be used to update the model parameters such as weight and bias without unloading the model using rte_ml_model_unload(). ML Inference: ML inference is the process of feeding data to the model via rte_ml_enqueue_burst() API and use rte_ml_dequeue_burst() API to get the calculated outputs/predictions from the started model. In all functions of the ML device API, the ML device is designated by an integer >= 0 named as device identifier *dev_id*. The functions exported by the ML device API to setup a device designated by its device identifier must be invoked in the following order: - rte_ml_dev_configure() - rte_ml_dev_queue_pair_setup() - rte_ml_dev_start() A model is required to run the inference operations with the user specified inputs. Application needs to invoke the ML model API in the following order before queueing inference jobs. - rte_ml_model_load() - rte_ml_model_start() The rte_ml_model_info() API is provi
[dpdk-dev] [RFC PATCH 1/1] mldev: introduce machine learning device library
From: Jerin Jacob Add mldev API specification to standardize and use the machine learning device and inference operations in vendor neutral way. Following operations are abstracted through APIs - ML device capability probe - ML device configuration - ML device queue pair configuration - ML device state management - ML device stat/xstat operations - ML model load/unload/start/stop operations - ML model information probe - ML IO operations to find size for input and output buffers - ML quantize and dequantize operations - ML ops pool creation and free operations - ML device enqueue/dequeue fastpath interference operations Signed-off-by: Jerin Jacob Signed-off-by: Srikanth Yalavarthi --- config/rte_config.h |3 + doc/api/doxy-api-index.md |1 + doc/api/doxy-api.conf.in|1 + doc/guides/prog_guide/index.rst |1 + doc/guides/prog_guide/mldev.rst | 164 + lib/eal/common/eal_common_log.c |1 + lib/eal/include/rte_log.h |1 + lib/meson.build |1 + lib/mldev/meson.build | 12 + lib/mldev/rte_mldev.c |5 + lib/mldev/rte_mldev.h | 1081 +++ lib/mldev/version.map |5 + 12 files changed, 1276 insertions(+) create mode 100644 doc/guides/prog_guide/mldev.rst create mode 100644 lib/mldev/meson.build create mode 100644 lib/mldev/rte_mldev.c create mode 100644 lib/mldev/rte_mldev.h create mode 100644 lib/mldev/version.map diff --git a/config/rte_config.h b/config/rte_config.h index 46549cb062..2adbef3f51 100644 --- a/config/rte_config.h +++ b/config/rte_config.h @@ -81,6 +81,9 @@ /* rawdev defines */ #define RTE_RAWDEV_MAX_DEVS 64 +/* mldev defines */ +#define RTE_MLDEV_MAX_DEVS 64 + /* ip_fragmentation defines */ #define RTE_LIBRTE_IP_FRAG_MAX_FRAG 8 // RTE_LIBRTE_IP_FRAG_TBL_STAT is not set diff --git a/doc/api/doxy-api-index.md b/doc/api/doxy-api-index.md index 186a258be4..d55cca5b97 100644 --- a/doc/api/doxy-api-index.md +++ b/doc/api/doxy-api-index.md @@ -22,6 +22,7 @@ The public API headers are grouped by topics: [compress](@ref rte_comp.h), [regexdev](@ref rte_regexdev.h), [dmadev](@ref rte_dmadev.h), + [mldev](@ref rte_mldev.h), [eventdev](@ref rte_eventdev.h), [event_eth_rx_adapter](@ref rte_event_eth_rx_adapter.h), [event_eth_tx_adapter](@ref rte_event_eth_tx_adapter.h), diff --git a/doc/api/doxy-api.conf.in b/doc/api/doxy-api.conf.in index 608494a7c0..82b28e8b18 100644 --- a/doc/api/doxy-api.conf.in +++ b/doc/api/doxy-api.conf.in @@ -59,6 +59,7 @@ INPUT = @TOPDIR@/doc/api/doxy-api-index.md \ @TOPDIR@/lib/mempool \ @TOPDIR@/lib/meter \ @TOPDIR@/lib/metrics \ + @TOPDIR@/lib/mldev \ @TOPDIR@/lib/node \ @TOPDIR@/lib/net \ @TOPDIR@/lib/pcapng \ diff --git a/doc/guides/prog_guide/index.rst b/doc/guides/prog_guide/index.rst index 8564883018..d7f2a28bdb 100644 --- a/doc/guides/prog_guide/index.rst +++ b/doc/guides/prog_guide/index.rst @@ -30,6 +30,7 @@ Programmer's Guide regexdev dmadev gpudev +mldev rte_security rawdev link_bonding_poll_mode_drv_lib diff --git a/doc/guides/prog_guide/mldev.rst b/doc/guides/prog_guide/mldev.rst new file mode 100644 index 00..2ce8e2f7fe --- /dev/null +++ b/doc/guides/prog_guide/mldev.rst @@ -0,0 +1,164 @@ +.. SPDX-License-Identifier: BSD-3-Clause +Copyright(C) 2022 Marvell International Ltd. + +Machine Learning Device Library +=== + +The MLDEV library provides a Machine Learning device framework for the management and +provisioning of hardware and software ML poll mode drivers, defining APIs which +support a number of ML operations including device handling and inference processing. +The ML model creation and training is outside of the scope of this library. + +Design Principles +- + +The MLDEV library follows the same basic principles as those used in DPDK's +Ethernet Device framework and the Crypto framework. The MLDEV framework provides +a generic Machine Learning device framework which supports both physical (hardware) +and virtual (software) ML devices as well as an ML API to manage and configure ML +devices. The APIs also supports performing ML inference operations through ML poll +mode driver. + + +Device Operations +- + +Device Creation +~~~ + +Physical ML devices are discovered during the PCI probe/enumeration, through the +EAL functions which are executed at DPDK initialization, based on their PCI device +identifier, each unique PCI BDF (bus/bridge, device, function). ML physical devices, +like other physical devices in DPDK can be white-listed or black-listed +using the EAL command line options. + + +Device Identification +~ + +Each devi
[PATCH] net/ring: add umonitor callback
Currently ring pmd does not support ``rte_power_monitor`` api. This patch adds support by adding umonitor callback that is called whenever we enter sleep state and need to check if it is time to wake up. Signed-off-by: Herakliusz Lipiec --- drivers/net/ring/rte_eth_ring.c | 23 +++ 1 file changed, 23 insertions(+) diff --git a/drivers/net/ring/rte_eth_ring.c b/drivers/net/ring/rte_eth_ring.c index cfb81da5fe..1bb6b196cb 100644 --- a/drivers/net/ring/rte_eth_ring.c +++ b/drivers/net/ring/rte_eth_ring.c @@ -284,6 +284,28 @@ eth_dev_close(struct rte_eth_dev *dev) return ret; } +static int ring_monitor_callback(const uint64_t value, + const uint64_t arg[RTE_POWER_MONITOR_OPAQUE_SZ]) +{ + /* Check if the head pointer has changed */ + return value != arg[0]; +} + +static int +eth_get_monitor_addr(void* rx_queue, struct rte_power_monitor_cond *pmc) +{ + struct rte_ring *rng = ((struct ring_queue*) rx_queue)->rng; + + /* Monitor ring head since if head moves +* there are packets to transmit +* */ + pmc->addr = &rng->prod.head; + pmc->size = sizeof(rng->prod.head); + pmc->opaque[0] = rng->prod.head; + pmc->fn = ring_monitor_callback; + return 0; +} + static const struct eth_dev_ops ops = { .dev_close = eth_dev_close, .dev_start = eth_dev_start, @@ -303,6 +325,7 @@ static const struct eth_dev_ops ops = { .promiscuous_disable = eth_promiscuous_disable, .allmulticast_enable = eth_allmulticast_enable, .allmulticast_disable = eth_allmulticast_disable, + .get_monitor_addr = eth_get_monitor_addr, }; static int -- 2.36.1
Re: [PATCH 03/13] net/idpf: support device initialization
On Wed, 3 Aug 2022 19:30:54 +0800 Junfeng Guo wrote: > + > +static int > +idpf_dev_configure(struct rte_eth_dev *dev) > +{ > + struct idpf_vport *vport = > + (struct idpf_vport *)dev->data->dev_private; Cast of void pointer dev_private is unnecessary in C (only in C++ is it needed). > + int ret = 0; Useless initialization, you are setting in next line. > + > + ret = idpf_init_vport_req_info(dev); > + if (ret) {
Re: [PATCH 04/13] net/idpf: add queue operations
On Wed, 3 Aug 2022 19:30:55 +0800 Junfeng Guo wrote: > +static void > +idpf_tx_queue_release(void *txq) > +{ > + struct idpf_tx_queue *q = (struct idpf_tx_queue *)txq; > + > + if (!q) > + return; > + > + if (q->complq) > + rte_free(q->complq); null pointer check before calling rte_free is unnecessary. rte_free(NULL) works. There is even a script used to scan DPDK to fix this. devtools/cocci/nullfree.cocci
Re: [dpdk-dev] [RFC PATCH 0/1] mldev: introduce machine learning device library
On Wed, 3 Aug 2022 18:58:37 +0530 wrote: > Roadmap > --- > 1) Address the comments for this RFC. > 2) Common code for mldev > 3) SW mldev driver based on TVM (https://tvm.apache.org/) Having a SW implementation is important because then it can be covered by tests.
RE: [EXT] Re: [PATCH] mbuf: add mbuf physical address field to dynamic field
Hi Bruce, >> > >> If all devices are configured to run in IOVA mode as VA, physical >> > >> address field of mbuf (buf_iova) won't be used. >> >> Will some of the hardware vendors please comment on this: Has IOVA VA mode >become common over time, or is it still an exotic bleeding edge feature? >> >> If it has become common, we should let DPDK evolve accordingly, and consider >PA (non-VA) mode legacy, treating it as such. Don't get stuck in the past. >> > >IOVA as VA mode is indeed common and we are constantly encouraging users to >switch to using vfio to try and take advantage of this. > >However, in my experience IOVA as PA is still very, very common too. We >cannot drop support for this mode just yet, unfortunately. > >> > >> In such cases, >> > buf_iova >> > >> space is free to use as a dynamic field. So a new dynamic field >> > member >> > >> (dynfield2) is added in mbuf structure to make use of that space. >> > >> >> > >> A new mbuf flag RTE_MBUF_F_DYNFIELD2 is introduced to help identify >> > the >> > >> mbuf that can use dynfield2. >> > >> >> > >> Signed-off-by: Shijith Thotton >> > > >> > > This seems like a complex and potentially error prone way to do this. >> >> Perhaps this optimization should be a compile time option instead? >> > >It could indeed be, and probably could be done very safely, in that we could >mark as disabled all drivers when the mode is enabled for a build. >Thereafter, drivers could be marked as VA-only safe as they are updated as >necessary, i.e. use the build system to enforce that only drivers known to >work with the mode are built when the mode is enabled. > I will prepare a patch to enable VA-only build. >That said, verifying all drivers to work with this mode is a decent effort. >Do we have indications of the perf benefit we would get from doing this for >some real-world app? >
[PATCH 1/4] eventdev/timer: add periodic event timer support
This patch adds support to configure and use periodic event timers in software timer adapter. The structure ``rte_event_timer_adapter_stats`` is extended by adding a new field, ``evtim_drop_count``. This stat represents the number of times an event_timer expiry event is dropped by the event timer adapter. Signed-off-by: Naga Harish K S V --- lib/eventdev/rte_event_timer_adapter.c | 86 ++ lib/eventdev/rte_event_timer_adapter.h | 2 + lib/eventdev/rte_eventdev.c| 6 +- 3 files changed, 67 insertions(+), 27 deletions(-) diff --git a/lib/eventdev/rte_event_timer_adapter.c b/lib/eventdev/rte_event_timer_adapter.c index e0d978d641..f96cb98b5b 100644 --- a/lib/eventdev/rte_event_timer_adapter.c +++ b/lib/eventdev/rte_event_timer_adapter.c @@ -53,6 +53,14 @@ static const struct event_timer_adapter_ops swtim_ops; #define EVTIM_SVC_LOG_DBG(...) (void)0 #endif +static inline enum rte_timer_type +get_event_timer_type(const struct rte_event_timer_adapter *adapter) +{ + return (adapter->data->conf.flags & + RTE_EVENT_TIMER_ADAPTER_F_PERIODIC) ? + PERIODICAL : SINGLE; +} + static int default_port_conf_cb(uint16_t id, uint8_t event_dev_id, uint8_t *event_port_id, void *conf_arg) @@ -195,10 +203,11 @@ rte_event_timer_adapter_create_ext( adapter->data->conf = *conf; /* copy conf structure */ /* Query eventdev PMD for timer adapter capabilities and ops */ - ret = dev->dev_ops->timer_adapter_caps_get(dev, + ret = dev->dev_ops->timer_adapter_caps_get ? + dev->dev_ops->timer_adapter_caps_get(dev, adapter->data->conf.flags, &adapter->data->caps, - &adapter->ops); + &adapter->ops) : 0; if (ret < 0) { rte_errno = -ret; goto free_memzone; @@ -348,10 +357,11 @@ rte_event_timer_adapter_lookup(uint16_t adapter_id) dev = &rte_eventdevs[adapter->data->event_dev_id]; /* Query eventdev PMD for timer adapter capabilities and ops */ - ret = dev->dev_ops->timer_adapter_caps_get(dev, + ret = dev->dev_ops->timer_adapter_caps_get ? + dev->dev_ops->timer_adapter_caps_get(dev, adapter->data->conf.flags, &adapter->data->caps, - &adapter->ops); + &adapter->ops) : 0; if (ret < 0) { rte_errno = EINVAL; return NULL; @@ -612,35 +622,44 @@ swtim_callback(struct rte_timer *tim) uint64_t opaque; int ret; int n_lcores; + enum rte_timer_type type; opaque = evtim->impl_opaque[1]; adapter = (struct rte_event_timer_adapter *)(uintptr_t)opaque; sw = swtim_pmd_priv(adapter); + type = get_event_timer_type(adapter); + + if (unlikely(sw->in_use[lcore].v == 0)) { + sw->in_use[lcore].v = 1; + n_lcores = __atomic_fetch_add(&sw->n_poll_lcores, 1, +__ATOMIC_RELAXED); + __atomic_store_n(&sw->poll_lcores[n_lcores], lcore, + __ATOMIC_RELAXED); + } ret = event_buffer_add(&sw->buffer, &evtim->ev); if (ret < 0) { - /* If event buffer is full, put timer back in list with -* immediate expiry value, so that we process it again on the -* next iteration. -*/ - ret = rte_timer_alt_reset(sw->timer_data_id, tim, 0, SINGLE, - lcore, NULL, evtim); - if (ret < 0) { - EVTIM_LOG_DBG("event buffer full, failed to reset " - "timer with immediate expiry value"); + if (type == SINGLE) { + /* If event buffer is full, put timer back in list with + * immediate expiry value, so that we process it again + * on the next iteration. + */ + ret = rte_timer_alt_reset(sw->timer_data_id, tim, 0, + SINGLE, lcore, NULL, evtim); + if (ret < 0) { + EVTIM_LOG_DBG("event buffer full, failed to " + "reset timer with immediate " + "expiry value"); + } else { + sw->stats.evtim_retry_count++; + EVTIM_LOG_DBG("event buffer full, resettin
[PATCH 4/4] test/event: update periodic event timer tests
This patch updates the software timer adapter tests to configure and use periodic event timers Signed-off-by: Naga Harish K S V --- app/test/test_event_timer_adapter.c | 41 ++--- 1 file changed, 37 insertions(+), 4 deletions(-) diff --git a/app/test/test_event_timer_adapter.c b/app/test/test_event_timer_adapter.c index d6170bb589..654c412836 100644 --- a/app/test/test_event_timer_adapter.c +++ b/app/test/test_event_timer_adapter.c @@ -386,11 +386,22 @@ timdev_setup_msec(void) static int timdev_setup_msec_periodic(void) { + uint32_t caps = 0; + uint64_t max_tmo_ns; + uint64_t flags = RTE_EVENT_TIMER_ADAPTER_F_ADJUST_RES | RTE_EVENT_TIMER_ADAPTER_F_PERIODIC; + TEST_ASSERT_SUCCESS(rte_event_timer_adapter_caps_get(evdev, &caps), + "failed to get adapter capabilities"); + + if (caps & RTE_EVENT_TIMER_ADAPTER_CAP_INTERNAL_PORT) + max_tmo_ns = 0; + else + max_tmo_ns = 180 * NSECPERSEC; + /* Periodic mode with 100 ms resolution */ - return _timdev_setup(0, NSECPERSEC / 10, flags); + return _timdev_setup(max_tmo_ns, NSECPERSEC / 10, flags); } static int @@ -409,7 +420,7 @@ timdev_setup_sec_periodic(void) RTE_EVENT_TIMER_ADAPTER_F_PERIODIC; /* Periodic mode with 1 sec resolution */ - return _timdev_setup(0, NSECPERSEC, flags); + return _timdev_setup(180 * NSECPERSEC, NSECPERSEC, flags); } static int @@ -561,12 +572,23 @@ test_timer_arm(void) static inline int test_timer_arm_periodic(void) { + uint32_t caps = 0; + uint32_t timeout_count = 0; + TEST_ASSERT_SUCCESS(_arm_timers(1, MAX_TIMERS), "Failed to arm timers"); /* With a resolution of 100ms and wait time of 1sec, * there will be 10 * MAX_TIMERS periodic timer triggers. */ - TEST_ASSERT_SUCCESS(_wait_timer_triggers(1, 10 * MAX_TIMERS, 0), + TEST_ASSERT_SUCCESS(rte_event_timer_adapter_caps_get(evdev, &caps), + "failed to get adapter capabilities"); + + if (caps & RTE_EVENT_TIMER_ADAPTER_CAP_INTERNAL_PORT) + timeout_count = 10; + else + timeout_count = 9; + + TEST_ASSERT_SUCCESS(_wait_timer_triggers(1, timeout_count * MAX_TIMERS, 0), "Timer triggered count doesn't match arm count"); return TEST_SUCCESS; } @@ -649,12 +671,23 @@ test_timer_arm_burst(void) static inline int test_timer_arm_burst_periodic(void) { + uint32_t caps = 0; + uint32_t timeout_count = 0; + TEST_ASSERT_SUCCESS(_arm_timers_burst(1, MAX_TIMERS), "Failed to arm timers"); /* With a resolution of 100ms and wait time of 1sec, * there will be 10 * MAX_TIMERS periodic timer triggers. */ - TEST_ASSERT_SUCCESS(_wait_timer_triggers(1, 10 * MAX_TIMERS, 0), + TEST_ASSERT_SUCCESS(rte_event_timer_adapter_caps_get(evdev, &caps), + "failed to get adapter capabilities"); + + if (caps & RTE_EVENT_TIMER_ADAPTER_CAP_INTERNAL_PORT) + timeout_count = 10; + else + timeout_count = 9; + + TEST_ASSERT_SUCCESS(_wait_timer_triggers(1, timeout_count * MAX_TIMERS, 0), "Timer triggered count doesn't match arm count"); return TEST_SUCCESS; -- 2.25.1
[PATCH 2/4] event/sw: report periodic event timer capability
update the software eventdev pmd timer_adapter_caps_get callback function to report the support of periodic event timer capability Signed-off-by: Naga Harish K S V --- drivers/event/sw/sw_evdev.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/drivers/event/sw/sw_evdev.c b/drivers/event/sw/sw_evdev.c index f93313b31b..89c07d30ae 100644 --- a/drivers/event/sw/sw_evdev.c +++ b/drivers/event/sw/sw_evdev.c @@ -564,7 +564,7 @@ sw_timer_adapter_caps_get(const struct rte_eventdev *dev, uint64_t flags, { RTE_SET_USED(dev); RTE_SET_USED(flags); - *caps = 0; + *caps = RTE_EVENT_TIMER_ADAPTER_CAP_PERIODIC; /* Use default SW ops */ *ops = NULL; -- 2.25.1
[PATCH 3/4] timer: fix rte_timer_stop_all
there is a possibility of deadlock in this api, as same spinlock is tried to be acquired in nested manner. This patch removes the acquisition of nested locking. Signed-off-by: Naga Harish K S V --- lib/timer/rte_timer.c | 7 +-- 1 file changed, 1 insertion(+), 6 deletions(-) diff --git a/lib/timer/rte_timer.c b/lib/timer/rte_timer.c index 9994813d0d..cfbc8cb028 100644 --- a/lib/timer/rte_timer.c +++ b/lib/timer/rte_timer.c @@ -987,21 +987,16 @@ rte_timer_stop_all(uint32_t timer_data_id, unsigned int *walk_lcores, walk_lcore = walk_lcores[i]; priv_timer = &timer_data->priv_timer[walk_lcore]; - rte_spinlock_lock(&priv_timer->list_lock); - for (tim = priv_timer->pending_head.sl_next[0]; tim != NULL; tim = next_tim) { next_tim = tim->sl_next[0]; - /* Call timer_stop with lock held */ - __rte_timer_stop(tim, 1, timer_data); + __rte_timer_stop(tim, 0, timer_data); if (f) f(tim, f_arg); } - - rte_spinlock_unlock(&priv_timer->list_lock); } return 0; -- 2.25.1
Re: [PATCH 3/4] timer: fix rte_timer_stop_all
On Wed, 3 Aug 2022 11:26:51 -0500 Naga Harish K S V wrote: > there is a possibility of deadlock in this api, > as same spinlock is tried to be acquired in nested manner. > > This patch removes the acquisition of nested locking. > > Signed-off-by: Naga Harish K S V The wording in this commit message is a little confusing, what is the exact lock conflict? After your patch, there are no longer any callers for __rte_timer_stop() with the local_is_locked flag. Please resubmit and remove all the the local_is_locked from __rte_timer_stop(). It looks like the test suite for rte_timer is not exercising all the exposed API's. That's a problem
[PATCH v2] graph: fix out of bounds access when re-allocate node objs
For __rte_node_enqueue_prologue(), If the number of objs is more than the node->size * 2, the extra objs will write out of bounds memory. It should use __rte_node_stream_alloc_size() to request enough memory. And for rte_node_next_stream_put(), it will re-allocate a small size, when the node free space is small and new objs is less than the current node->size. Some objs pointers behind new size may be lost. And it will cause memory leak. It should request enough size of memory, containing the original objs and new objs at least. Fixes: 40d4f51403ec ("graph: implement fastpath routines") Signed-off-by: Zhirun Yan Signed-off-by: Cunming Liang --- lib/graph/rte_graph_worker.h | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/lib/graph/rte_graph_worker.h b/lib/graph/rte_graph_worker.h index 0c0b9c095a..6dc7461659 100644 --- a/lib/graph/rte_graph_worker.h +++ b/lib/graph/rte_graph_worker.h @@ -224,7 +224,7 @@ __rte_node_enqueue_prologue(struct rte_graph *graph, struct rte_node *node, __rte_node_enqueue_tail_update(graph, node); if (unlikely(node->size < (idx + space))) - __rte_node_stream_alloc(graph, node); + __rte_node_stream_alloc_size(graph, node, node->size + space); } /** @@ -432,7 +432,7 @@ rte_node_next_stream_get(struct rte_graph *graph, struct rte_node *node, uint16_t free_space = node->size - idx; if (unlikely(free_space < nb_objs)) - __rte_node_stream_alloc_size(graph, node, nb_objs); + __rte_node_stream_alloc_size(graph, node, node->size + nb_objs); return &node->objs[idx]; } -- 2.25.1
Re: [RFC] memarea: introduce memory area library
On Thu, Jul 21, 2022 at 10:23 AM Chengwen Feng wrote: > > The memarea library is an allocator of variable-size object. It is a > collection of allocated objects that can be efficiently alloc or free > all at once, the main feature are as follows: > a) it facilitate alloc and free of memory with low overhead. > > b) it provides refcnt feature which could be useful in some scenes. > > c) it supports MT-safe as long as it's specified at creation time. > > d) it's memory source could comes from: > d.1) system API: malloc in C library. > d.2) user provided address: it can be from the rte_malloc API series > or extended memory as long as it is available. > d.3) user provided memarea: it can be from another memarea. > > Note: > a) the memarea is oriented towards the application layer, which could > provides 'region-based memory management' [1] function. > b) the eal library also provide memory zone/heap management, but these > are tied to huge pages management. > > [1] https://en.wikipedia.org/wiki/Region-based_memory_management > Looks like a good feature to add to DPDK . > Signed-off-by: Chengwen Feng > --- > lib/memarea/meson.build | 10 ++ > lib/memarea/rte_memarea.c | 52 ++ > lib/memarea/rte_memarea.h | 205 ++ > lib/memarea/version.map | 16 +++ > lib/meson.build | 1 + > 5 files changed, 284 insertions(+) > create mode 100644 lib/memarea/meson.build > create mode 100644 lib/memarea/rte_memarea.c > create mode 100644 lib/memarea/rte_memarea.h > create mode 100644 lib/memarea/version.map > > diff --git a/lib/memarea/rte_memarea.h b/lib/memarea/rte_memarea.h > new file mode 100644 > index 00..dfbe845595 > --- /dev/null > +++ b/lib/memarea/rte_memarea.h > @@ -0,0 +1,205 @@ > +/* SPDX-License-Identifier: BSD-3-Clause > + * Copyright(c) 2022 HiSilicon Limited > + */ > + > +#ifndef RTE_MEMAREA_H > +#define RTE_MEMAREA_H > + > +/** > + * @file > + * RTE Memarea. > + * > + * A memory area is an allocator of variable-size object. It is identified > + * by its name. > + * > + * The memarea is a collection of allocated objects that can be efficiently > + * alloc or free all at once, the main feature are as follows: > + * a) it facilitate alloc and free of memory with low overhead. > + * b) it provides refcnt feature which could be useful in some scenes. > + * c) it supports MT-safe as long as it's specified at creation time. > + * d) it's memory source could comes from: > + * d.1) system API: malloc in C library. > + * d.2) user provided address: it can be from the rte_malloc API series > + * or extended memory as long as it is available. > + * d.3) user provided memarea: it can be from another memarea. So we can > + * build the following memory management structure: > + * memarea-1 > + * | > + * v > + * > + * | | | > + * v v v > + * memarea-2 memarea-3obj > + * > + */ > + > +#include > +#include > +#include > + > +#include > + > +#ifdef __cplusplus > +extern "C" { > +#endif > + > +#define RTE_MEMAREA_NAMESIZE 64 > + > +/** > + * Memarea memory source. > + */ > +enum rte_memarea_source { > + /** Memory source comes from sys api (e.g. malloc) */ > + RTE_MEMAREA_SOURCE_SYSAPI, > + /** Memory source comes from user-provided address */ > + RTE_MEMAREA_SOURCE_USER_ADDR, > + /** Memory source comes from user-provided memarea */ > + RTE_MEMAREA_SOURCE_USER_MEMAREA, > +}; > + > +struct rte_memarea; > + > +struct rte_memarea_param { > + char name[RTE_MEMAREA_NAMESIZE]; /**< Name of memarea */ > + enum rte_memarea_source source; /**< Memory source of memarea */ > + uint64_t size; /**< Size (byte) of memarea */ > + uint32_t align; /**< Align of allocated object */ > + /** Indicates whether the memarea should be MT-safe */ > + bool mt_safe; > + /** Indicates whether the memarea is visible to multiple process. > +* If the memory source is RTE_MEMAREA_SOURCE_USER_ADDR, this filed > +* depends on user settings and must be set. > +* If the memory source is RTE_MEMAREA_SOURCE_SYSAPI or > +* RTE_MEMAREA_SOURCE_USER_MEMAREA, this filed does not need to be > set. > +*/ > + bool mp_visible; > + /** User provided address, this field is valid only when source > +* is set to RTE_MEMAREA_SOURCE_USER_ADDR. > +*/ > + void *user_addr; > + /** User provided memarea, this field is valid only when source > +* is set to RTE_MEMAREA_SOURCE_MEMAREA. > +*/ > + struct rte_memarea *user_memarea; Above two can be in union as it is based on enum rte_memarea_source. > +}; > + > +struct rte_memarea { > +