[PATCH v8] enhance NUMA affinity heuristic
When a DPDK application is started on only one numa node, memory is allocated for only one socket. When interrupt threads use memory, memory may not be found on the socket where the interrupt thread is currently located, and memory has to be reallocated on the hugepage, this operation will lead to performance degradation. Fixes: 705356f0811f ("eal: simplify control thread creation") Fixes: 770d41bf3309 ("malloc: fix allocation with unknown socket ID") Cc: sta...@dpdk.org Signed-off-by: Kaisen You --- Changes since v7: - Update commet, Changes since v6: - New explanation for easy understanding, Changes since v5: - Add comments to the code, Changes since v4: - mod the patch title, Changes since v3: - add the assignment of socket_id in thread initialization, Changes since v2: - add uncommitted local change and fix compilation, Changes since v1: - accomodate for configurations with main lcore running on multiples physical cores belonging to different numa, --- lib/eal/common/eal_common_thread.c | 2 ++ lib/eal/common/malloc_heap.c | 7 ++- 2 files changed, 8 insertions(+), 1 deletion(-) diff --git a/lib/eal/common/eal_common_thread.c b/lib/eal/common/eal_common_thread.c index 079a385630..95720a61c0 100644 --- a/lib/eal/common/eal_common_thread.c +++ b/lib/eal/common/eal_common_thread.c @@ -252,6 +252,8 @@ static int ctrl_thread_init(void *arg) struct rte_thread_ctrl_params *params = arg; __rte_thread_init(rte_lcore_id(), cpuset); + /* Set control thread socket ID to SOCKET_ID_ANY as +* control threads may be scheduled on any NUMA node. +*/ + RTE_PER_LCORE(_socket_id) = SOCKET_ID_ANY; params->ret = rte_thread_set_affinity_by_id(rte_thread_self(), cpuset); if (params->ret != 0) { __atomic_store_n(¶ms->ctrl_thread_status, diff --git a/lib/eal/common/malloc_heap.c b/lib/eal/common/malloc_heap.c index d25bdc98f9..5de2317827 100644 --- a/lib/eal/common/malloc_heap.c +++ b/lib/eal/common/malloc_heap.c @@ -716,7 +716,12 @@ malloc_get_numa_socket(void) if (conf->socket_mem[socket_id] != 0) return socket_id; } - + /* We couldn't find quickly find a NUMA node where memory +* was available, so fall back to using main lcore socket ID. +*/ + socket_id = rte_lcore_to_socket_id(rte_get_main_lcore()); + /* Main lcore socket ID may be SOCKET_ID_ANY in cases when main +* lcore thread is affinitized to multiple NUMA nodes. +*/ + if (socket_id != (unsigned int)SOCKET_ID_ANY) + return socket_id; + /* Failed to find meaningful socket ID, so just use the first one available */ return rte_socket_id_by_idx(0); } -- 2.25.1
RE: [EXT] [PATCH v3 1/2] cryptodev: support SM3_HMAC,SM4_CFB and SM4_OFB
> Add SM3_HMAC/SM4_CFB/SM4_OFB support in DPDK. > > Signed-off-by: Sunyang Wu > --- > doc/guides/cryptodevs/features/default.ini | 3 +++ > doc/guides/rel_notes/release_23_07.rst | 5 + > lib/cryptodev/rte_crypto_sym.h | 8 +++- > lib/cryptodev/rte_cryptodev.c | 5 - > 4 files changed, 19 insertions(+), 2 deletions(-) > > diff --git a/doc/guides/cryptodevs/features/default.ini > b/doc/guides/cryptodevs/features/default.ini > index 523da0cfa8..8f54d4a2a5 100644 > --- a/doc/guides/cryptodevs/features/default.ini > +++ b/doc/guides/cryptodevs/features/default.ini > @@ -64,6 +64,8 @@ ZUC EEA3 = > SM4 ECB= > SM4 CBC= > SM4 CTR= > +SM4 CFB= > +SM4 OFB= > > ; > ; Supported authentication algorithms of a default crypto driver. > @@ -99,6 +101,7 @@ SHA3_384 HMAC = > SHA3_512= > SHA3_512 HMAC = > SM3 = > +SM3 HMAC= > SHAKE_128 = > SHAKE_256 = > > diff --git a/doc/guides/rel_notes/release_23_07.rst > b/doc/guides/rel_notes/release_23_07.rst > index a9b1293689..405b34c6d2 100644 > --- a/doc/guides/rel_notes/release_23_07.rst > +++ b/doc/guides/rel_notes/release_23_07.rst > @@ -55,6 +55,11 @@ New Features > Also, make sure to start the actual text at the margin. > === > > +* **Added new algorithms to cryptodev.** > + > + * Added symmetric hash algorithm SM3-HMAC. > + * Added symmetric cipher algorithm ShangMi 4 (SM4) in CFB and OFB modes. > + > > Removed Items > - > diff --git a/lib/cryptodev/rte_crypto_sym.h b/lib/cryptodev/rte_crypto_sym.h > index b43174dbec..428603d06e 100644 > --- a/lib/cryptodev/rte_crypto_sym.h > +++ b/lib/cryptodev/rte_crypto_sym.h > @@ -172,8 +172,12 @@ enum rte_crypto_cipher_algorithm { > /**< ShangMi 4 (SM4) algorithm in ECB mode */ > RTE_CRYPTO_CIPHER_SM4_CBC, > /**< ShangMi 4 (SM4) algorithm in CBC mode */ > - RTE_CRYPTO_CIPHER_SM4_CTR > + RTE_CRYPTO_CIPHER_SM4_CTR, > /**< ShangMi 4 (SM4) algorithm in CTR mode */ > + RTE_CRYPTO_CIPHER_SM4_OFB, > + /**< ShangMi 4 (SM4) algorithm in OFB mode */ > + RTE_CRYPTO_CIPHER_SM4_CFB > + /**< ShangMi 4 (SM4) algorithm in CFB mode */ > }; > > /** Cipher algorithm name strings */ > @@ -376,6 +380,8 @@ enum rte_crypto_auth_algorithm { > /**< HMAC using 512 bit SHA3 algorithm. */ > RTE_CRYPTO_AUTH_SM3, > /**< ShangMi 3 (SM3) algorithm */ > + RTE_CRYPTO_AUTH_SM3_HMAC, > + /** < HMAC using ShangMi 3 (SM3) algorithm */ You cannot insert in the middle of enum. This will result in ABI break. http://mails.dpdk.org/archives/test-report/2023-May/400475.html Please move this change to end of enum for this release. You can submit a patch for next release(which is an ABI break release.) to move it back. > > RTE_CRYPTO_AUTH_SHAKE_128, > /**< 128 bit SHAKE algorithm. */ > diff --git a/lib/cryptodev/rte_cryptodev.c b/lib/cryptodev/rte_cryptodev.c > index a96114b2da..4ff7046e97 100644 > --- a/lib/cryptodev/rte_cryptodev.c > +++ b/lib/cryptodev/rte_cryptodev.c > @@ -127,7 +127,9 @@ crypto_cipher_algorithm_strings[] = { > [RTE_CRYPTO_CIPHER_ZUC_EEA3]= "zuc-eea3", > [RTE_CRYPTO_CIPHER_SM4_ECB] = "sm4-ecb", > [RTE_CRYPTO_CIPHER_SM4_CBC] = "sm4-cbc", > - [RTE_CRYPTO_CIPHER_SM4_CTR] = "sm4-ctr" > + [RTE_CRYPTO_CIPHER_SM4_CTR] = "sm4-ctr", > + [RTE_CRYPTO_CIPHER_SM4_CFB] = "sm4-cfb", > + [RTE_CRYPTO_CIPHER_SM4_OFB] = "sm4-ofb" > }; > > /** > @@ -227,6 +229,7 @@ crypto_auth_algorithm_strings[] = { > [RTE_CRYPTO_AUTH_SNOW3G_UIA2] = "snow3g-uia2", > [RTE_CRYPTO_AUTH_ZUC_EIA3] = "zuc-eia3", > [RTE_CRYPTO_AUTH_SM3] = "sm3", > + [RTE_CRYPTO_AUTH_SM3_HMAC] = "sm3-hmac", > > [RTE_CRYPTO_AUTH_SHAKE_128] = "shake-128", > [RTE_CRYPTO_AUTH_SHAKE_256] = "shake-256", > -- > 2.19.0.rc0.windows.1
RE: [v1, 2/3] test/test_cryptodev_asym: add SM2 tests
Hi Akhil, > > If you have the reference from where the vectors are taken it can also be > mentioned. > I generated these test vectors in online. There are no reference test vectors quoted In RFC except for recommended curve parameters, at the end. https://datatracker.ietf.org/doc/html/draft-shen-sm2-ecdsa-02 > > > > Signed-off-by: Gowrishankar Muthukrishnan > > --- > > app/test/test_cryptodev_asym.c | 506 + > > app/test/test_cryptodev_sm2_test_vectors.h | 120 + > > 2 files changed, 626 insertions(+) > > create mode 100644 app/test/test_cryptodev_sm2_test_vectors.h > > > > diff --git a/app/test/test_cryptodev_asym.c > > b/app/test/test_cryptodev_asym.c index 9236817650..bfaeedee27 100644 > > --- a/app/test/test_cryptodev_asym.c > > +++ b/app/test/test_cryptodev_asym.c > > @@ -21,6 +21,7 @@ > > #include "test_cryptodev_ecpm_test_vectors.h" > > #include "test_cryptodev_mod_test_vectors.h" > > #include "test_cryptodev_rsa_test_vectors.h" > > +#include "test_cryptodev_sm2_test_vectors.h" > > #include "test_cryptodev_asym_util.h" > > #include "test.h" > > > > @@ -2196,6 +2197,507 @@ test_ecpm_all_curve(void) > > return overall_status; > > } > > > > +static int > > +test_sm2_sign(void) > > +{ > > + struct crypto_testsuite_params_asym *ts_params = &testsuite_params; > > + struct crypto_testsuite_sm2_params input_params = > > sm2_param_fp256; > > + struct rte_mempool *sess_mpool = ts_params->session_mpool; > > + struct rte_mempool *op_mpool = ts_params->op_mpool; > > + uint8_t dev_id = ts_params->valid_devs[0]; > > + struct rte_crypto_op *result_op = NULL; > > + uint8_t output_buf_r[TEST_DATA_SIZE]; > > + uint8_t output_buf_s[TEST_DATA_SIZE]; > > + struct rte_crypto_asym_xform xform; > > + struct rte_crypto_asym_op *asym_op; > > + struct rte_cryptodev_info dev_info; > > + struct rte_crypto_op *op = NULL; > > + int ret, status = TEST_SUCCESS; > > + void *sess = NULL; > > + > > + rte_cryptodev_info_get(dev_id, &dev_info); > > dev_info is being unused. Not checking for capabilities? It is the same case in other algorithms as well. Shall I collectively address this ? > > + > > + /* Setup crypto op data structure */ > > + op = rte_crypto_op_alloc(op_mpool, > > RTE_CRYPTO_OP_TYPE_ASYMMETRIC); > > + if (op == NULL) { > > + RTE_LOG(ERR, USER1, > > + "line %u FAILED: %s", __LINE__, > > + "Failed to allocate asymmetric crypto " > > + "operation struct\n"); > > + status = TEST_FAILED; > > + goto exit; > > + } > > + asym_op = op->asym; > > + > > + /* Setup asym xform */ > > + xform.next = NULL; > > + xform.xform_type = RTE_CRYPTO_ASYM_XFORM_SM2; > > + xform.sm2.pkey.data = input_params.pkey.data; > > + xform.sm2.pkey.length = input_params.pkey.length; > > + xform.sm2.q.x.data = input_params.pubkey_qx.data; > > + xform.sm2.q.x.length = input_params.pubkey_qx.length; > > + xform.sm2.q.y.data = input_params.pubkey_qy.data; > > + xform.sm2.q.y.length = input_params.pubkey_qy.length; > > + > > + ret = rte_cryptodev_asym_session_create(dev_id, &xform, sess_mpool, > > &sess); > > + if (ret < 0) { > > + RTE_LOG(ERR, USER1, > > + "line %u FAILED: %s", __LINE__, > > + "Session creation failed\n"); > > + status = (ret == -ENOTSUP) ? TEST_SKIPPED : TEST_FAILED; > > + goto exit; > > + } > > + > > + /* Attach asymmetric crypto session to crypto operations */ > > + rte_crypto_op_attach_asym_session(op, sess); > > + > > + /* Compute sign */ > > + > > + /* Populate op with operational details */ > > + op->asym->sm2.op_type = RTE_CRYPTO_ASYM_OP_SIGN; > > + op->asym->sm2.message.data = input_params.message.data; > > + op->asym->sm2.message.length = input_params.message.length; > > + op->asym->sm2.id.data = input_params.id.data; > > + op->asym->sm2.id.length = input_params.id.length; > > + > > + /* Init out buf */ > > + op->asym->sm2.r.data = output_buf_r; > > + op->asym->sm2.s.data = output_buf_s; > > + > > + RTE_LOG(DEBUG, USER1, "Process ASYM operation\n"); > > + > > + /* Process crypto operation */ > > + if (rte_cryptodev_enqueue_burst(dev_id, 0, &op, 1) != 1) { > > + RTE_LOG(ERR, USER1, > > + "line %u FAILED: %s", __LINE__, > > + "Error sending packet for operation\n"); > > + status = TEST_FAILED; > > + goto exit; > > + } > > + > > + while (rte_cryptodev_dequeue_burst(dev_id, 0, &result_op, 1) == 0) > > + rte_pause(); > > Shouldn't this be a finite loop and mark test as failed after some retries? It is the same case in other algorithms as well. Shall I collectively address this ? > > > + > > + if (result_op == NULL) { > > + RTE_LOG(ERR, USER1, > > + "line %u FAILED: %s",
Re: [RFC v2 1/2] dts: add smoke tests
On Thu, May 25, 2023 at 8:03 PM Jeremy Spewock wrote: > > Hey Juraj, > > On Thu, May 25, 2023 at 4:33 AM Juraj Linkeš > wrote: >> >> One more point that doesn't fit elsewhere: >> >> On Wed, May 24, 2023 at 10:45 PM Jeremy Spewock wrote: >> > >> > >> > >> > On Tue, May 23, 2023 at 4:05 AM Juraj Linkeš >> > wrote: >> >> >> >> Hi Jeremy, first, a few general points: >> >> >> >> 1. Send patches to maintainers (Thomas, me, Honnappa, Lijuan and >> >> anyone else involved with DTS or who might be interested) and add the >> >> devlist to cc. >> > >> > >> > Thank you for the tip! I'm still new to sending patches and didn't think >> > to do something like this but I will in the future. >> > >> >> >> >> 2. Run the linter script before submitting. >> > >> > >> > I did forget to run this, I will in the future. >> > >> >> >> >> 3. The use of the various nested objects breaks the current >> >> abstractions. The basic idea is that the test suite developers should >> >> ideally only use the sut/tg node objects and those objects should >> >> delegate logic further to their nested objects. More below. >> >> >> >> I have many comments about the implementation, but I haven't run it >> >> yet. I'm going to do that after this round of comments and I may have >> >> more ideas. >> >> >> >> On Fri, May 12, 2023 at 9:28 PM wrote: >> >> > >> >> > From: Jeremy Spewock >> >> > >> >> > Adds a new test suite for running smoke tests that verify general >> >> > configuration aspects of the system under test. If any of these tests >> >> > fail, the DTS execution terminates as part of a "fail-fast" model. >> >> > >> >> > Signed-off-by: Jeremy Spewock >> >> > --- >> >> > dts/conf.yaml | 9 ++ >> >> > dts/framework/config/__init__.py | 21 + >> >> > dts/framework/config/conf_yaml_schema.json| 32 ++- >> >> > dts/framework/dts.py | 19 +++- >> >> > dts/framework/exception.py| 11 +++ >> >> > dts/framework/remote_session/os_session.py| 6 +- >> >> > .../remote_session/remote/__init__.py | 28 ++ >> >> > dts/framework/test_result.py | 13 ++- >> >> > dts/framework/test_suite.py | 24 - >> >> > dts/framework/testbed_model/__init__.py | 5 + >> >> > .../interactive_apps/__init__.py | 6 ++ >> >> > .../interactive_apps/interactive_command.py | 57 +++ >> >> > .../interactive_apps/testpmd_driver.py| 24 + >> >> > dts/framework/testbed_model/node.py | 2 + >> >> > dts/framework/testbed_model/sut_node.py | 6 ++ >> >> > dts/tests/TestSuite_smoke_tests.py| 94 +++ >> >> > 16 files changed, 348 insertions(+), 9 deletions(-) >> >> > create mode 100644 >> >> > dts/framework/testbed_model/interactive_apps/__init__.py >> >> > create mode 100644 >> >> > dts/framework/testbed_model/interactive_apps/interactive_command.py >> >> > create mode 100644 >> >> > dts/framework/testbed_model/interactive_apps/testpmd_driver.py >> >> >> >> Let's not add any more levels. I don't like even the current hw >> >> subdirectory (which I want to remove in the docs patch) and we don't >> >> need it. I'd also like to move this functionality into remote_session, >> >> as it's handling a type of remote session. >> > >> > >> > I think it makes sense to add a proper wrapper for it, I just didn't >> > create a subclass for it off of remote_session because the idea behind it >> > was only allowing users to interact with the session through the >> > InteractiveScriptHandler/DPDK app handlers. If it were part of the >> > remote_session class it would have to include operations for sending >> > commands the way it is written now. I could do this but it seems like a >> > bigger discussion about whether interactive sessions should create a new >> > SSH session every time or instead just use one existing session and create >> > channels off of it. >> > >> >> I wasn't clear, I meant to move the python modules into the >> remote_session folder. The subclassing would be there only to have a >> common API across these remote sessions (as the >> InteractiveScriptHandler is a kind of a remote session). If at all >> possible, this should be our aim. > > > I see what you mean now, moving them there shouldn't be a problem. > >> >> >> >> >> >> >> >> > create mode 100644 dts/tests/TestSuite_smoke_tests.py >> >> > >> >> > diff --git a/dts/conf.yaml b/dts/conf.yaml >> >> > index a9bd8a3e..042ef954 100644 >> >> > --- a/dts/conf.yaml >> >> > +++ b/dts/conf.yaml >> >> > @@ -10,13 +10,22 @@ executions: >> >> > compiler_wrapper: ccache >> >> > perf: false >> >> > func: true >> >> > +nics: #physical devices to be used for testing >> >> > + - addresses: >> >> > + - ":11:00.0" >> >> > + - ":11:00.1" >> >> > +driver: "i40e" >> >> > +vdevs: #names of virtual devices to be used for
[PATCH v4 00/13] net/cpfl: add hairpin queue support
From: Beilei Xing This patchset adds hairpin queue support. v2 changes: - change hairpin rx queus configuration sequence. - code refine. v3 changes: - Refine the patchset based on the latest code. v4 change: - Remove hairpin rx buffer queue's sw_ring. - Change hairpin rx queus configuration sequence in cpfl_hairpin_bind function. - Refind hairpin queue setup and release. Beilei Xing (13): net/cpfl: refine structures common/idpf: support queue groups add/delete net/cpfl: add haipin queue group during vport init net/cpfl: support hairpin queue capbility get net/cpfl: support hairpin queue setup and release common/idpf: add queue config API net/cpfl: support hairpin queue configuration common/idpf: add switch queue API net/cpfl: support hairpin queue start/stop common/idpf: add irq map config API net/cpfl: enable write back based on ITR expire net/cpfl: support peer ports get net/cpfl: support hairpin bind/unbind drivers/common/idpf/idpf_common_device.c | 75 ++ drivers/common/idpf/idpf_common_device.h | 4 + drivers/common/idpf/idpf_common_virtchnl.c | 138 +++- drivers/common/idpf/idpf_common_virtchnl.h | 18 + drivers/common/idpf/version.map| 6 + drivers/net/cpfl/cpfl_ethdev.c | 605 ++-- drivers/net/cpfl/cpfl_ethdev.h | 35 +- drivers/net/cpfl/cpfl_rxtx.c | 785 +++-- drivers/net/cpfl/cpfl_rxtx.h | 77 ++ drivers/net/cpfl/cpfl_rxtx_vec_common.h| 21 +- 10 files changed, 1645 insertions(+), 119 deletions(-) -- 2.26.2
[PATCH v4 01/13] net/cpfl: refine structures
From: Beilei Xing This patch refines some structures to support hairpin queue, cpfl_rx_queue/cpfl_tx_queue/cpfl_vport. Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/net/cpfl/cpfl_ethdev.c | 85 +++- drivers/net/cpfl/cpfl_ethdev.h | 6 +- drivers/net/cpfl/cpfl_rxtx.c| 175 +--- drivers/net/cpfl/cpfl_rxtx.h| 8 ++ drivers/net/cpfl/cpfl_rxtx_vec_common.h | 17 +-- 5 files changed, 196 insertions(+), 95 deletions(-) diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c index 7528a14d05..e587155db6 100644 --- a/drivers/net/cpfl/cpfl_ethdev.c +++ b/drivers/net/cpfl/cpfl_ethdev.c @@ -124,7 +124,8 @@ static int cpfl_dev_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete) { - struct idpf_vport *vport = dev->data->dev_private; + struct cpfl_vport *cpfl_vport = dev->data->dev_private; + struct idpf_vport *vport = &cpfl_vport->base; struct rte_eth_link new_link; unsigned int i; @@ -156,7 +157,8 @@ cpfl_dev_link_update(struct rte_eth_dev *dev, static int cpfl_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) { - struct idpf_vport *vport = dev->data->dev_private; + struct cpfl_vport *cpfl_vport = dev->data->dev_private; + struct idpf_vport *vport = &cpfl_vport->base; struct idpf_adapter *base = vport->adapter; dev_info->max_rx_queues = base->caps.max_rx_q; @@ -216,7 +218,8 @@ cpfl_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) static int cpfl_dev_mtu_set(struct rte_eth_dev *dev, uint16_t mtu) { - struct idpf_vport *vport = dev->data->dev_private; + struct cpfl_vport *cpfl_vport = dev->data->dev_private; + struct idpf_vport *vport = &cpfl_vport->base; /* mtu setting is forbidden if port is start */ if (dev->data->dev_started) { @@ -256,12 +259,12 @@ static uint64_t cpfl_get_mbuf_alloc_failed_stats(struct rte_eth_dev *dev) { uint64_t mbuf_alloc_failed = 0; - struct idpf_rx_queue *rxq; + struct cpfl_rx_queue *cpfl_rxq; int i = 0; for (i = 0; i < dev->data->nb_rx_queues; i++) { - rxq = dev->data->rx_queues[i]; - mbuf_alloc_failed += __atomic_load_n(&rxq->rx_stats.mbuf_alloc_failed, + cpfl_rxq = dev->data->rx_queues[i]; + mbuf_alloc_failed += __atomic_load_n(&cpfl_rxq->base.rx_stats.mbuf_alloc_failed, __ATOMIC_RELAXED); } @@ -271,8 +274,8 @@ cpfl_get_mbuf_alloc_failed_stats(struct rte_eth_dev *dev) static int cpfl_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats) { - struct idpf_vport *vport = - (struct idpf_vport *)dev->data->dev_private; + struct cpfl_vport *cpfl_vport = dev->data->dev_private; + struct idpf_vport *vport = &cpfl_vport->base; struct virtchnl2_vport_stats *pstats = NULL; int ret; @@ -305,20 +308,20 @@ cpfl_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats) static void cpfl_reset_mbuf_alloc_failed_stats(struct rte_eth_dev *dev) { - struct idpf_rx_queue *rxq; + struct cpfl_rx_queue *cpfl_rxq; int i; for (i = 0; i < dev->data->nb_rx_queues; i++) { - rxq = dev->data->rx_queues[i]; - __atomic_store_n(&rxq->rx_stats.mbuf_alloc_failed, 0, __ATOMIC_RELAXED); + cpfl_rxq = dev->data->rx_queues[i]; + __atomic_store_n(&cpfl_rxq->base.rx_stats.mbuf_alloc_failed, 0, __ATOMIC_RELAXED); } } static int cpfl_dev_stats_reset(struct rte_eth_dev *dev) { - struct idpf_vport *vport = - (struct idpf_vport *)dev->data->dev_private; + struct cpfl_vport *cpfl_vport = dev->data->dev_private; + struct idpf_vport *vport = &cpfl_vport->base; struct virtchnl2_vport_stats *pstats = NULL; int ret; @@ -343,8 +346,8 @@ static int cpfl_dev_xstats_reset(struct rte_eth_dev *dev) static int cpfl_dev_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats, unsigned int n) { - struct idpf_vport *vport = - (struct idpf_vport *)dev->data->dev_private; + struct cpfl_vport *cpfl_vport = dev->data->dev_private; + struct idpf_vport *vport = &cpfl_vport->base; struct virtchnl2_vport_stats *pstats = NULL; unsigned int i; int ret; @@ -459,7 +462,8 @@ cpfl_rss_reta_update(struct rte_eth_dev *dev, struct rte_eth_rss_reta_entry64 *reta_conf, uint16_t reta_size) { - struct idpf_vport *vport = dev->data->dev_private; + struct cpfl_vport *cpfl_vport = dev->data->dev_private; + struct idpf_vport *vport = &cpfl_vport->base; struct idpf_adapter *base = vport->adapter; uin
[PATCH v4 02/13] common/idpf: support queue groups add/delete
From: Beilei Xing This patch adds queue group add/delete virtual channel support. Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/common/idpf/idpf_common_virtchnl.c | 66 ++ drivers/common/idpf/idpf_common_virtchnl.h | 9 +++ drivers/common/idpf/version.map| 2 + 3 files changed, 77 insertions(+) diff --git a/drivers/common/idpf/idpf_common_virtchnl.c b/drivers/common/idpf/idpf_common_virtchnl.c index b713678634..a3fe55c897 100644 --- a/drivers/common/idpf/idpf_common_virtchnl.c +++ b/drivers/common/idpf/idpf_common_virtchnl.c @@ -359,6 +359,72 @@ idpf_vc_vport_destroy(struct idpf_vport *vport) return err; } +int +idpf_vc_queue_grps_add(struct idpf_vport *vport, + struct virtchnl2_add_queue_groups *p2p_queue_grps_info, + uint8_t *p2p_queue_grps_out) +{ + struct idpf_adapter *adapter = vport->adapter; + struct idpf_cmd_info args; + int size, qg_info_size; + int err = -1; + + size = sizeof(*p2p_queue_grps_info) + + (p2p_queue_grps_info->qg_info.num_queue_groups - 1) * + sizeof(struct virtchnl2_queue_group_info); + + memset(&args, 0, sizeof(args)); + args.ops = VIRTCHNL2_OP_ADD_QUEUE_GROUPS; + args.in_args = (uint8_t *)p2p_queue_grps_info; + args.in_args_size = size; + args.out_buffer = adapter->mbx_resp; + args.out_size = IDPF_DFLT_MBX_BUF_SIZE; + + err = idpf_vc_cmd_execute(adapter, &args); + if (err != 0) { + DRV_LOG(ERR, + "Failed to execute command of VIRTCHNL2_OP_ADD_QUEUE_GROUPS"); + return err; + } + + rte_memcpy(p2p_queue_grps_out, args.out_buffer, IDPF_DFLT_MBX_BUF_SIZE); + return 0; +} + +int idpf_vc_queue_grps_del(struct idpf_vport *vport, + uint16_t num_q_grps, + struct virtchnl2_queue_group_id *qg_ids) +{ + struct idpf_adapter *adapter = vport->adapter; + struct virtchnl2_delete_queue_groups *vc_del_q_grps; + struct idpf_cmd_info args; + int size; + int err; + + size = sizeof(*vc_del_q_grps) + + (num_q_grps - 1) * sizeof(struct virtchnl2_queue_group_id); + vc_del_q_grps = rte_zmalloc("vc_del_q_grps", size, 0); + + vc_del_q_grps->vport_id = vport->vport_id; + vc_del_q_grps->num_queue_groups = num_q_grps; + memcpy(vc_del_q_grps->qg_ids, qg_ids, + num_q_grps * sizeof(struct virtchnl2_queue_group_id)); + + memset(&args, 0, sizeof(args)); + args.ops = VIRTCHNL2_OP_DEL_QUEUE_GROUPS; + args.in_args = (uint8_t *)vc_del_q_grps; + args.in_args_size = size; + args.out_buffer = adapter->mbx_resp; + args.out_size = IDPF_DFLT_MBX_BUF_SIZE; + + err = idpf_vc_cmd_execute(adapter, &args); + if (err != 0) + DRV_LOG(ERR, "Failed to execute command of VIRTCHNL2_OP_DEL_QUEUE_GROUPS"); + + rte_free(vc_del_q_grps); + return err; +} + int idpf_vc_rss_key_set(struct idpf_vport *vport) { diff --git a/drivers/common/idpf/idpf_common_virtchnl.h b/drivers/common/idpf/idpf_common_virtchnl.h index c45295290e..58b16e1c5d 100644 --- a/drivers/common/idpf/idpf_common_virtchnl.h +++ b/drivers/common/idpf/idpf_common_virtchnl.h @@ -64,4 +64,13 @@ int idpf_vc_ctlq_recv(struct idpf_ctlq_info *cq, u16 *num_q_msg, __rte_internal int idpf_vc_ctlq_post_rx_buffs(struct idpf_hw *hw, struct idpf_ctlq_info *cq, u16 *buff_count, struct idpf_dma_mem **buffs); +__rte_internal +int idpf_vc_queue_grps_del(struct idpf_vport *vport, + uint16_t num_q_grps, + struct virtchnl2_queue_group_id *qg_ids); +__rte_internal +int +idpf_vc_queue_grps_add(struct idpf_vport *vport, + struct virtchnl2_add_queue_groups *ptp_queue_grps_info, + uint8_t *ptp_queue_grps_out); #endif /* _IDPF_COMMON_VIRTCHNL_H_ */ diff --git a/drivers/common/idpf/version.map b/drivers/common/idpf/version.map index 70334a1b03..01d18f3f3f 100644 --- a/drivers/common/idpf/version.map +++ b/drivers/common/idpf/version.map @@ -43,6 +43,8 @@ INTERNAL { idpf_vc_irq_map_unmap_config; idpf_vc_one_msg_read; idpf_vc_ptype_info_query; + idpf_vc_queue_grps_add; + idpf_vc_queue_grps_del; idpf_vc_queue_switch; idpf_vc_queues_ena_dis; idpf_vc_rss_hash_get; -- 2.26.2
[PATCH v4 03/13] net/cpfl: add haipin queue group during vport init
From: Beilei Xing This patch adds haipin queue group during vport init. Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/net/cpfl/cpfl_ethdev.c | 133 + drivers/net/cpfl/cpfl_ethdev.h | 18 + drivers/net/cpfl/cpfl_rxtx.h | 7 ++ 3 files changed, 158 insertions(+) diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c index e587155db6..c1273a7478 100644 --- a/drivers/net/cpfl/cpfl_ethdev.c +++ b/drivers/net/cpfl/cpfl_ethdev.c @@ -840,6 +840,20 @@ cpfl_dev_stop(struct rte_eth_dev *dev) return 0; } +static int +cpfl_p2p_queue_grps_del(struct idpf_vport *vport) +{ + struct virtchnl2_queue_group_id qg_ids[CPFL_P2P_NB_QUEUE_GRPS] = {0}; + int ret = 0; + + qg_ids[0].queue_group_id = CPFL_P2P_QUEUE_GRP_ID; + qg_ids[0].queue_group_type = VIRTCHNL2_QUEUE_GROUP_P2P; + ret = idpf_vc_queue_grps_del(vport, CPFL_P2P_NB_QUEUE_GRPS, qg_ids); + if (ret) + PMD_DRV_LOG(ERR, "Failed to delete p2p queue groups"); + return ret; +} + static int cpfl_dev_close(struct rte_eth_dev *dev) { @@ -848,7 +862,12 @@ cpfl_dev_close(struct rte_eth_dev *dev) struct cpfl_adapter_ext *adapter = CPFL_ADAPTER_TO_EXT(vport->adapter); cpfl_dev_stop(dev); + + if (!adapter->base.is_rx_singleq && !adapter->base.is_tx_singleq) + cpfl_p2p_queue_grps_del(vport); + idpf_vport_deinit(vport); + rte_free(cpfl_vport->p2p_q_chunks_info); adapter->cur_vports &= ~RTE_BIT32(vport->devarg_id); adapter->cur_vport_nb--; @@ -1284,6 +1303,96 @@ cpfl_vport_idx_alloc(struct cpfl_adapter_ext *adapter) return vport_idx; } +static int +cpfl_p2p_q_grps_add(struct idpf_vport *vport, + struct virtchnl2_add_queue_groups *p2p_queue_grps_info, + uint8_t *p2p_q_vc_out_info) +{ + int ret; + + p2p_queue_grps_info->vport_id = vport->vport_id; + p2p_queue_grps_info->qg_info.num_queue_groups = CPFL_P2P_NB_QUEUE_GRPS; + p2p_queue_grps_info->qg_info.groups[0].num_rx_q = CPFL_MAX_P2P_NB_QUEUES; + p2p_queue_grps_info->qg_info.groups[0].num_rx_bufq = CPFL_P2P_NB_RX_BUFQ; + p2p_queue_grps_info->qg_info.groups[0].num_tx_q = CPFL_MAX_P2P_NB_QUEUES; + p2p_queue_grps_info->qg_info.groups[0].num_tx_complq = CPFL_P2P_NB_TX_COMPLQ; + p2p_queue_grps_info->qg_info.groups[0].qg_id.queue_group_id = CPFL_P2P_QUEUE_GRP_ID; + p2p_queue_grps_info->qg_info.groups[0].qg_id.queue_group_type = VIRTCHNL2_QUEUE_GROUP_P2P; + p2p_queue_grps_info->qg_info.groups[0].rx_q_grp_info.rss_lut_size = 0; + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.tx_tc = 0; + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.priority = 0; + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.is_sp = 0; + p2p_queue_grps_info->qg_info.groups[0].tx_q_grp_info.pir_weight = 0; + + ret = idpf_vc_queue_grps_add(vport, p2p_queue_grps_info, p2p_q_vc_out_info); + if (ret != 0) { + PMD_DRV_LOG(ERR, "Failed to add p2p queue groups."); + return ret; + } + + return ret; +} + +static int +cpfl_p2p_queue_info_init(struct cpfl_vport *cpfl_vport, +struct virtchnl2_add_queue_groups *p2p_q_vc_out_info) +{ + struct p2p_queue_chunks_info *p2p_q_chunks_info = cpfl_vport->p2p_q_chunks_info; + struct virtchnl2_queue_reg_chunks *vc_chunks_out; + int i, type; + + if (p2p_q_vc_out_info->qg_info.groups[0].qg_id.queue_group_type != + VIRTCHNL2_QUEUE_GROUP_P2P) { + PMD_DRV_LOG(ERR, "Add queue group response mismatch."); + return -EINVAL; + } + + vc_chunks_out = &p2p_q_vc_out_info->qg_info.groups[0].chunks; + + for (i = 0; i < vc_chunks_out->num_chunks; i++) { + type = vc_chunks_out->chunks[i].type; + switch (type) { + case VIRTCHNL2_QUEUE_TYPE_TX: + p2p_q_chunks_info->tx_start_qid = + vc_chunks_out->chunks[i].start_queue_id; + p2p_q_chunks_info->tx_qtail_start = + vc_chunks_out->chunks[i].qtail_reg_start; + p2p_q_chunks_info->tx_qtail_spacing = + vc_chunks_out->chunks[i].qtail_reg_spacing; + break; + case VIRTCHNL2_QUEUE_TYPE_RX: + p2p_q_chunks_info->rx_start_qid = + vc_chunks_out->chunks[i].start_queue_id; + p2p_q_chunks_info->rx_qtail_start = + vc_chunks_out->chunks[i].qtail_reg_start; + p2p_q_chunks_info->rx_qtail_spacing = + vc_chunks_out->chunks[i].qtail_reg_spacing; + break; + case VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION:
[PATCH v4 04/13] net/cpfl: support hairpin queue capbility get
From: Beilei Xing This patch adds hairpin_cap_get ops support. Signed-off-by: Xiao Wang Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/net/cpfl/cpfl_ethdev.c | 18 ++ drivers/net/cpfl/cpfl_rxtx.h | 3 +++ 2 files changed, 21 insertions(+) diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c index c1273a7478..40b4515539 100644 --- a/drivers/net/cpfl/cpfl_ethdev.c +++ b/drivers/net/cpfl/cpfl_ethdev.c @@ -154,6 +154,23 @@ cpfl_dev_link_update(struct rte_eth_dev *dev, return rte_eth_linkstatus_set(dev, &new_link); } +static int +cpfl_hairpin_cap_get(struct rte_eth_dev *dev, +struct rte_eth_hairpin_cap *cap) +{ + struct cpfl_vport *cpfl_vport = dev->data->dev_private; + + if (cpfl_vport->p2p_q_chunks_info == NULL) + return -ENOTSUP; + + cap->max_nb_queues = CPFL_MAX_P2P_NB_QUEUES; + cap->max_rx_2_tx = CPFL_MAX_HAIRPINQ_RX_2_TX; + cap->max_tx_2_rx = CPFL_MAX_HAIRPINQ_TX_2_RX; + cap->max_nb_desc = CPFL_MAX_HAIRPINQ_NB_DESC; + + return 0; +} + static int cpfl_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) { @@ -904,6 +921,7 @@ static const struct eth_dev_ops cpfl_eth_dev_ops = { .xstats_get = cpfl_dev_xstats_get, .xstats_get_names = cpfl_dev_xstats_get_names, .xstats_reset = cpfl_dev_xstats_reset, + .hairpin_cap_get= cpfl_hairpin_cap_get, }; static int diff --git a/drivers/net/cpfl/cpfl_rxtx.h b/drivers/net/cpfl/cpfl_rxtx.h index 1fe65778f0..a4a164d462 100644 --- a/drivers/net/cpfl/cpfl_rxtx.h +++ b/drivers/net/cpfl/cpfl_rxtx.h @@ -14,6 +14,9 @@ #define CPFL_MAX_RING_DESC 4096 #define CPFL_DMA_MEM_ALIGN 4096 +#define CPFL_MAX_HAIRPINQ_RX_2_TX 1 +#define CPFL_MAX_HAIRPINQ_TX_2_RX 1 +#define CPFL_MAX_HAIRPINQ_NB_DESC 1024 #define CPFL_MAX_P2P_NB_QUEUES 16 #define CPFL_P2P_NB_RX_BUFQ1 #define CPFL_P2P_NB_TX_COMPLQ 1 -- 2.26.2
[PATCH v4 05/13] net/cpfl: support hairpin queue setup and release
From: Beilei Xing Support hairpin Rx/Tx queue setup and release. Signed-off-by: Xiao Wang Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/net/cpfl/cpfl_ethdev.c | 6 + drivers/net/cpfl/cpfl_ethdev.h | 11 + drivers/net/cpfl/cpfl_rxtx.c| 353 +++- drivers/net/cpfl/cpfl_rxtx.h| 36 +++ drivers/net/cpfl/cpfl_rxtx_vec_common.h | 4 + 5 files changed, 409 insertions(+), 1 deletion(-) diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c index 40b4515539..b17c538ec2 100644 --- a/drivers/net/cpfl/cpfl_ethdev.c +++ b/drivers/net/cpfl/cpfl_ethdev.c @@ -879,6 +879,10 @@ cpfl_dev_close(struct rte_eth_dev *dev) struct cpfl_adapter_ext *adapter = CPFL_ADAPTER_TO_EXT(vport->adapter); cpfl_dev_stop(dev); + if (cpfl_vport->p2p_mp) { + rte_mempool_free(cpfl_vport->p2p_mp); + cpfl_vport->p2p_mp = NULL; + } if (!adapter->base.is_rx_singleq && !adapter->base.is_tx_singleq) cpfl_p2p_queue_grps_del(vport); @@ -922,6 +926,8 @@ static const struct eth_dev_ops cpfl_eth_dev_ops = { .xstats_get_names = cpfl_dev_xstats_get_names, .xstats_reset = cpfl_dev_xstats_reset, .hairpin_cap_get= cpfl_hairpin_cap_get, + .rx_hairpin_queue_setup = cpfl_rx_hairpin_queue_setup, + .tx_hairpin_queue_setup = cpfl_tx_hairpin_queue_setup, }; static int diff --git a/drivers/net/cpfl/cpfl_ethdev.h b/drivers/net/cpfl/cpfl_ethdev.h index 666d46a44a..2e42354f70 100644 --- a/drivers/net/cpfl/cpfl_ethdev.h +++ b/drivers/net/cpfl/cpfl_ethdev.h @@ -89,6 +89,17 @@ struct p2p_queue_chunks_info { struct cpfl_vport { struct idpf_vport base; struct p2p_queue_chunks_info *p2p_q_chunks_info; + + struct rte_mempool *p2p_mp; + + uint16_t nb_data_rxq; + uint16_t nb_data_txq; + uint16_t nb_p2p_rxq; + uint16_t nb_p2p_txq; + + struct idpf_rx_queue *p2p_rx_bufq; + struct idpf_tx_queue *p2p_tx_complq; + bool p2p_manual_bind; }; struct cpfl_adapter_ext { diff --git a/drivers/net/cpfl/cpfl_rxtx.c b/drivers/net/cpfl/cpfl_rxtx.c index 04a51b8d15..9625629a20 100644 --- a/drivers/net/cpfl/cpfl_rxtx.c +++ b/drivers/net/cpfl/cpfl_rxtx.c @@ -10,6 +10,67 @@ #include "cpfl_rxtx.h" #include "cpfl_rxtx_vec_common.h" +static inline void +cpfl_tx_hairpin_descq_reset(struct idpf_tx_queue *txq) +{ + uint32_t i, size; + + if (!txq) { + PMD_DRV_LOG(DEBUG, "Pointer to txq is NULL"); + return; + } + + size = txq->nb_tx_desc * CPFL_P2P_DESC_LEN; + for (i = 0; i < size; i++) + ((volatile char *)txq->desc_ring)[i] = 0; +} + +static inline void +cpfl_tx_hairpin_complq_reset(struct idpf_tx_queue *cq) +{ + uint32_t i, size; + + if (!cq) { + PMD_DRV_LOG(DEBUG, "Pointer to complq is NULL"); + return; + } + + size = cq->nb_tx_desc * CPFL_P2P_DESC_LEN; + for (i = 0; i < size; i++) + ((volatile char *)cq->compl_ring)[i] = 0; +} + +static inline void +cpfl_rx_hairpin_descq_reset(struct idpf_rx_queue *rxq) +{ + uint16_t len; + uint32_t i; + + if (!rxq) + return; + + len = rxq->nb_rx_desc; + for (i = 0; i < len * CPFL_P2P_DESC_LEN; i++) + ((volatile char *)rxq->rx_ring)[i] = 0; +} + +static inline void +cpfl_rx_hairpin_bufq_reset(struct idpf_rx_queue *rxbq) +{ + uint16_t len; + uint32_t i; + + if (!rxbq) + return; + + len = rxbq->nb_rx_desc; + for (i = 0; i < len * CPFL_P2P_DESC_LEN; i++) + ((volatile char *)rxbq->rx_ring)[i] = 0; + + rxbq->bufq1 = NULL; + rxbq->bufq2 = NULL; +} + static uint64_t cpfl_rx_offload_convert(uint64_t offload) { @@ -234,7 +295,10 @@ cpfl_rx_queue_release(void *rxq) /* Split queue */ if (!q->adapter->is_rx_singleq) { - if (q->bufq2) + /* the mz is shared between Tx/Rx hairpin, let Rx_release +* free the buf, q->bufq1->mz and q->mz. +*/ + if (!cpfl_rxq->hairpin_info.hairpin_q && q->bufq2) cpfl_rx_split_bufq_release(q->bufq2); if (q->bufq1) @@ -385,6 +449,7 @@ cpfl_rx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_idx, } } + cpfl_vport->nb_data_rxq++; rxq->q_set = true; dev->data->rx_queues[queue_idx] = cpfl_rxq; @@ -548,6 +613,7 @@ cpfl_tx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_idx, txq->qtx_tail = hw->hw_addr + (vport->chunks_info.tx_qtail_start + queue_idx * vport->chunks_info.tx_qtail_spacing); txq->ops = &def_txq_ops; + cpfl_vport->nb_data_txq++; txq->q_set = true; dev->data->tx_q
[PATCH v4 06/13] common/idpf: add queue config API
From: Beilei Xing This patch supports Rx/Tx queue configuration APIs. Signed-off-by: Beilei Xing --- drivers/common/idpf/idpf_common_virtchnl.c | 70 ++ drivers/common/idpf/idpf_common_virtchnl.h | 6 ++ drivers/common/idpf/version.map| 2 + 3 files changed, 78 insertions(+) diff --git a/drivers/common/idpf/idpf_common_virtchnl.c b/drivers/common/idpf/idpf_common_virtchnl.c index a3fe55c897..211b44a88e 100644 --- a/drivers/common/idpf/idpf_common_virtchnl.c +++ b/drivers/common/idpf/idpf_common_virtchnl.c @@ -1050,6 +1050,41 @@ idpf_vc_rxq_config(struct idpf_vport *vport, struct idpf_rx_queue *rxq) return err; } +int idpf_vc_rxq_config_by_info(struct idpf_vport *vport, struct virtchnl2_rxq_info *rxq_info, + uint16_t num_qs) +{ + struct idpf_adapter *adapter = vport->adapter; + struct virtchnl2_config_rx_queues *vc_rxqs = NULL; + struct idpf_cmd_info args; + int size, err, i; + + size = sizeof(*vc_rxqs) + (num_qs - 1) * + sizeof(struct virtchnl2_rxq_info); + vc_rxqs = rte_zmalloc("cfg_rxqs", size, 0); + if (vc_rxqs == NULL) { + DRV_LOG(ERR, "Failed to allocate virtchnl2_config_rx_queues"); + err = -ENOMEM; + return err; + } + vc_rxqs->vport_id = vport->vport_id; + vc_rxqs->num_qinfo = num_qs; + memcpy(vc_rxqs->qinfo, rxq_info, num_qs * sizeof(struct virtchnl2_rxq_info)); + + memset(&args, 0, sizeof(args)); + args.ops = VIRTCHNL2_OP_CONFIG_RX_QUEUES; + args.in_args = (uint8_t *)vc_rxqs; + args.in_args_size = size; + args.out_buffer = adapter->mbx_resp; + args.out_size = IDPF_DFLT_MBX_BUF_SIZE; + + err = idpf_vc_cmd_execute(adapter, &args); + rte_free(vc_rxqs); + if (err != 0) + DRV_LOG(ERR, "Failed to execute command of VIRTCHNL2_OP_CONFIG_RX_QUEUES"); + + return err; +} + int idpf_vc_txq_config(struct idpf_vport *vport, struct idpf_tx_queue *txq) { @@ -1121,6 +1156,41 @@ idpf_vc_txq_config(struct idpf_vport *vport, struct idpf_tx_queue *txq) return err; } +int +idpf_vc_txq_config_by_info(struct idpf_vport *vport, struct virtchnl2_txq_info *txq_info, + uint16_t num_qs) +{ + struct idpf_adapter *adapter = vport->adapter; + struct virtchnl2_config_tx_queues *vc_txqs = NULL; + struct idpf_cmd_info args; + int size, err; + + size = sizeof(*vc_txqs) + (num_qs - 1) * sizeof(struct virtchnl2_txq_info); + vc_txqs = rte_zmalloc("cfg_txqs", size, 0); + if (vc_txqs == NULL) { + DRV_LOG(ERR, "Failed to allocate virtchnl2_config_tx_queues"); + err = -ENOMEM; + return err; + } + vc_txqs->vport_id = vport->vport_id; + vc_txqs->num_qinfo = num_qs; + memcpy(vc_txqs->qinfo, txq_info, num_qs * sizeof(struct virtchnl2_txq_info)); + + memset(&args, 0, sizeof(args)); + args.ops = VIRTCHNL2_OP_CONFIG_TX_QUEUES; + args.in_args = (uint8_t *)vc_txqs; + args.in_args_size = size; + args.out_buffer = adapter->mbx_resp; + args.out_size = IDPF_DFLT_MBX_BUF_SIZE; + + err = idpf_vc_cmd_execute(adapter, &args); + rte_free(vc_txqs); + if (err != 0) + DRV_LOG(ERR, "Failed to execute command of VIRTCHNL2_OP_CONFIG_TX_QUEUES"); + + return err; +} + int idpf_vc_ctlq_recv(struct idpf_ctlq_info *cq, u16 *num_q_msg, struct idpf_ctlq_msg *q_msg) diff --git a/drivers/common/idpf/idpf_common_virtchnl.h b/drivers/common/idpf/idpf_common_virtchnl.h index 58b16e1c5d..db83761a5e 100644 --- a/drivers/common/idpf/idpf_common_virtchnl.h +++ b/drivers/common/idpf/idpf_common_virtchnl.h @@ -65,6 +65,12 @@ __rte_internal int idpf_vc_ctlq_post_rx_buffs(struct idpf_hw *hw, struct idpf_ctlq_info *cq, u16 *buff_count, struct idpf_dma_mem **buffs); __rte_internal +int idpf_vc_rxq_config_by_info(struct idpf_vport *vport, struct virtchnl2_rxq_info *rxq_info, + uint16_t num_qs); +__rte_internal +int idpf_vc_txq_config_by_info(struct idpf_vport *vport, struct virtchnl2_txq_info *txq_info, + uint16_t num_qs); +__rte_internal int idpf_vc_queue_grps_del(struct idpf_vport *vport, uint16_t num_q_grps, struct virtchnl2_queue_group_id *qg_ids); diff --git a/drivers/common/idpf/version.map b/drivers/common/idpf/version.map index 01d18f3f3f..17e77884ce 100644 --- a/drivers/common/idpf/version.map +++ b/drivers/common/idpf/version.map @@ -54,8 +54,10 @@ INTERNAL { idpf_vc_rss_lut_get; idpf_vc_rss_lut_set; idpf_vc_rxq_config; + idpf_vc_rxq_config_by_info; idpf_vc_stats_query; idpf_vc_txq_config; + idpf_vc_txq_config_by_info; idpf_vc_vectors_alloc; idpf_vc_vect
[PATCH v4 08/13] common/idpf: add switch queue API
From: Beilei Xing This patch adds idpf_vc_ena_dis_one_queue API. Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/common/idpf/idpf_common_virtchnl.c | 2 +- drivers/common/idpf/idpf_common_virtchnl.h | 3 +++ drivers/common/idpf/version.map| 1 + 3 files changed, 5 insertions(+), 1 deletion(-) diff --git a/drivers/common/idpf/idpf_common_virtchnl.c b/drivers/common/idpf/idpf_common_virtchnl.c index 211b44a88e..6455f640da 100644 --- a/drivers/common/idpf/idpf_common_virtchnl.c +++ b/drivers/common/idpf/idpf_common_virtchnl.c @@ -733,7 +733,7 @@ idpf_vc_vectors_dealloc(struct idpf_vport *vport) return err; } -static int +int idpf_vc_ena_dis_one_queue(struct idpf_vport *vport, uint16_t qid, uint32_t type, bool on) { diff --git a/drivers/common/idpf/idpf_common_virtchnl.h b/drivers/common/idpf/idpf_common_virtchnl.h index db83761a5e..9ff5c38c26 100644 --- a/drivers/common/idpf/idpf_common_virtchnl.h +++ b/drivers/common/idpf/idpf_common_virtchnl.h @@ -71,6 +71,9 @@ __rte_internal int idpf_vc_txq_config_by_info(struct idpf_vport *vport, struct virtchnl2_txq_info *txq_info, uint16_t num_qs); __rte_internal +int idpf_vc_ena_dis_one_queue(struct idpf_vport *vport, uint16_t qid, + uint32_t type, bool on); +__rte_internal int idpf_vc_queue_grps_del(struct idpf_vport *vport, uint16_t num_q_grps, struct virtchnl2_queue_group_id *qg_ids); diff --git a/drivers/common/idpf/version.map b/drivers/common/idpf/version.map index 17e77884ce..25624732b0 100644 --- a/drivers/common/idpf/version.map +++ b/drivers/common/idpf/version.map @@ -40,6 +40,7 @@ INTERNAL { idpf_vc_cmd_execute; idpf_vc_ctlq_post_rx_buffs; idpf_vc_ctlq_recv; + idpf_vc_ena_dis_one_queue; idpf_vc_irq_map_unmap_config; idpf_vc_one_msg_read; idpf_vc_ptype_info_query; -- 2.26.2
[PATCH v4 07/13] net/cpfl: support hairpin queue configuration
From: Beilei Xing This patch supports Rx/Tx hairpin queue configuration. Signed-off-by: Xiao Wang Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/net/cpfl/cpfl_ethdev.c | 136 +++-- drivers/net/cpfl/cpfl_rxtx.c | 80 +++ drivers/net/cpfl/cpfl_rxtx.h | 7 ++ 3 files changed, 217 insertions(+), 6 deletions(-) diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c index b17c538ec2..a06def06d0 100644 --- a/drivers/net/cpfl/cpfl_ethdev.c +++ b/drivers/net/cpfl/cpfl_ethdev.c @@ -742,33 +742,157 @@ cpfl_config_rx_queues_irqs(struct rte_eth_dev *dev) return idpf_vport_irq_map_config(vport, nb_rx_queues); } +/* Update hairpin_info for dev's tx hairpin queue */ +static int +cpfl_txq_hairpin_info_update(struct rte_eth_dev *dev, uint16_t rx_port) +{ + struct cpfl_vport *cpfl_tx_vport = dev->data->dev_private; + struct rte_eth_dev *peer_dev = &rte_eth_devices[rx_port]; + struct cpfl_vport *cpfl_rx_vport = peer_dev->data->dev_private; + struct cpfl_txq_hairpin_info *hairpin_info; + struct cpfl_tx_queue *cpfl_txq; + int i; + + for (i = cpfl_tx_vport->nb_data_txq; i < dev->data->nb_tx_queues; i++) { + cpfl_txq = dev->data->tx_queues[i]; + hairpin_info = &cpfl_txq->hairpin_info; + if (hairpin_info->peer_rxp != rx_port) { + PMD_DRV_LOG(ERR, "port %d is not the peer port", rx_port); + return -EINVAL; + } + hairpin_info->peer_rxq_id = + cpfl_hw_qid_get(cpfl_rx_vport->p2p_q_chunks_info->rx_start_qid, + hairpin_info->peer_rxq_id - cpfl_rx_vport->nb_data_rxq); + } + + return 0; +} + +/* Bind Rx hairpin queue's memory zone to peer Tx hairpin queue's memory zone */ +static void +cpfl_rxq_hairpin_mz_bind(struct rte_eth_dev *dev) +{ + struct cpfl_vport *cpfl_rx_vport = dev->data->dev_private; + struct idpf_vport *vport = &cpfl_rx_vport->base; + struct idpf_adapter *adapter = vport->adapter; + struct idpf_hw *hw = &adapter->hw; + struct cpfl_rx_queue *cpfl_rxq; + struct cpfl_tx_queue *cpfl_txq; + struct rte_eth_dev *peer_dev; + const struct rte_memzone *mz; + uint16_t peer_tx_port; + uint16_t peer_tx_qid; + int i; + + for (i = cpfl_rx_vport->nb_data_rxq; i < dev->data->nb_rx_queues; i++) { + cpfl_rxq = dev->data->rx_queues[i]; + peer_tx_port = cpfl_rxq->hairpin_info.peer_txp; + peer_tx_qid = cpfl_rxq->hairpin_info.peer_txq_id; + peer_dev = &rte_eth_devices[peer_tx_port]; + cpfl_txq = peer_dev->data->tx_queues[peer_tx_qid]; + + /* bind rx queue */ + mz = cpfl_txq->base.mz; + cpfl_rxq->base.rx_ring_phys_addr = mz->iova; + cpfl_rxq->base.rx_ring = mz->addr; + cpfl_rxq->base.mz = mz; + + /* bind rx buffer queue */ + mz = cpfl_txq->base.complq->mz; + cpfl_rxq->base.bufq1->rx_ring_phys_addr = mz->iova; + cpfl_rxq->base.bufq1->rx_ring = mz->addr; + cpfl_rxq->base.bufq1->mz = mz; + cpfl_rxq->base.bufq1->qrx_tail = hw->hw_addr + + cpfl_hw_qtail_get(cpfl_rx_vport->p2p_q_chunks_info->rx_buf_qtail_start, + 0, cpfl_rx_vport->p2p_q_chunks_info->rx_buf_qtail_spacing); + } +} + static int cpfl_start_queues(struct rte_eth_dev *dev) { + struct cpfl_vport *cpfl_vport = dev->data->dev_private; + struct idpf_vport *vport = &cpfl_vport->base; struct cpfl_rx_queue *cpfl_rxq; struct cpfl_tx_queue *cpfl_txq; + int update_flag = 0; int err = 0; int i; + /* For normal data queues, configure, init and enale Txq. +* For non-manual bind hairpin queues, configure Txq. +*/ for (i = 0; i < dev->data->nb_tx_queues; i++) { cpfl_txq = dev->data->tx_queues[i]; if (cpfl_txq == NULL || cpfl_txq->base.tx_deferred_start) continue; - err = cpfl_tx_queue_start(dev, i); + if (!cpfl_txq->hairpin_info.hairpin_q) { + err = cpfl_tx_queue_start(dev, i); + if (err != 0) { + PMD_DRV_LOG(ERR, "Fail to start Tx queue %u", i); + return err; + } + } else if (!cpfl_vport->p2p_manual_bind) { + if (update_flag == 0) { + err = cpfl_txq_hairpin_info_update(dev, + cpfl_txq->hairpin_info.peer_rxp); + if (err != 0) { +
[PATCH v4 09/13] net/cpfl: support hairpin queue start/stop
From: Beilei Xing This patch supports Rx/Tx hairpin queue start/stop. Signed-off-by: Xiao Wang Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/net/cpfl/cpfl_ethdev.c | 41 + drivers/net/cpfl/cpfl_rxtx.c | 151 + drivers/net/cpfl/cpfl_rxtx.h | 14 +++ 3 files changed, 188 insertions(+), 18 deletions(-) diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c index a06def06d0..8035878602 100644 --- a/drivers/net/cpfl/cpfl_ethdev.c +++ b/drivers/net/cpfl/cpfl_ethdev.c @@ -896,6 +896,47 @@ cpfl_start_queues(struct rte_eth_dev *dev) } } + /* For non-manual bind hairpin queues, enable Tx queue and Rx queue, +* then enable Tx completion queue and Rx buffer queue. +*/ + for (i = 0; i < dev->data->nb_tx_queues; i++) { + cpfl_txq = dev->data->tx_queues[i]; + if (cpfl_txq->hairpin_info.hairpin_q && !cpfl_vport->p2p_manual_bind) { + err = cpfl_switch_hairpin_rxtx_queue(cpfl_vport, +i - cpfl_vport->nb_data_txq, +false, true); + if (err) + PMD_DRV_LOG(ERR, "Failed to switch hairpin TX queue %u on", + i); + else + cpfl_txq->base.q_started = true; + } + } + + for (i = 0; i < dev->data->nb_rx_queues; i++) { + cpfl_rxq = dev->data->rx_queues[i]; + if (cpfl_rxq->hairpin_info.hairpin_q && !cpfl_vport->p2p_manual_bind) { + err = cpfl_switch_hairpin_rxtx_queue(cpfl_vport, +i - cpfl_vport->nb_data_rxq, +true, true); + if (err) + PMD_DRV_LOG(ERR, "Failed to switch hairpin RX queue %u on", + i); + else + cpfl_rxq->base.q_started = true; + } + } + + if (!cpfl_vport->p2p_manual_bind && + cpfl_vport->p2p_tx_complq != NULL && + cpfl_vport->p2p_rx_bufq != NULL) { + err = cpfl_switch_hairpin_bufq_complq(cpfl_vport, true); + if (err != 0) { + PMD_DRV_LOG(ERR, "Failed to switch hairpin Tx complq and Rx bufq"); + return err; + } + } + return err; } diff --git a/drivers/net/cpfl/cpfl_rxtx.c b/drivers/net/cpfl/cpfl_rxtx.c index 702054d1c5..38c48ad8c7 100644 --- a/drivers/net/cpfl/cpfl_rxtx.c +++ b/drivers/net/cpfl/cpfl_rxtx.c @@ -991,6 +991,81 @@ cpfl_hairpin_txq_config(struct idpf_vport *vport, struct cpfl_tx_queue *cpfl_txq return idpf_vc_txq_config_by_info(vport, txq_info, 1); } +int +cpfl_switch_hairpin_bufq_complq(struct cpfl_vport *cpfl_vport, bool on) +{ + struct idpf_vport *vport = &cpfl_vport->base; + uint32_t type; + int err, queue_id; + + type = VIRTCHNL2_QUEUE_TYPE_TX_COMPLETION; + queue_id = cpfl_vport->p2p_tx_complq->queue_id; + err = idpf_vc_ena_dis_one_queue(vport, queue_id, type, on); + if (err) + return err; + + type = VIRTCHNL2_QUEUE_TYPE_RX_BUFFER; + queue_id = cpfl_vport->p2p_rx_bufq->queue_id; + err = idpf_vc_ena_dis_one_queue(vport, queue_id, type, on); + + return err; +} + +int +cpfl_switch_hairpin_rxtx_queue(struct cpfl_vport *cpfl_vport, uint16_t logic_qid, + bool rx, bool on) +{ + struct idpf_vport *vport = &cpfl_vport->base; + uint32_t type; + int err, queue_id; + + type = rx ? VIRTCHNL2_QUEUE_TYPE_RX : VIRTCHNL2_QUEUE_TYPE_TX; + + if (type == VIRTCHNL2_QUEUE_TYPE_RX) + queue_id = cpfl_hw_qid_get(cpfl_vport->p2p_q_chunks_info->rx_start_qid, logic_qid); + else + queue_id = cpfl_hw_qid_get(cpfl_vport->p2p_q_chunks_info->tx_start_qid, logic_qid); + err = idpf_vc_ena_dis_one_queue(vport, queue_id, type, on); + if (err) + return err; + + return err; +} + +static int +cpfl_alloc_split_p2p_rxq_mbufs(struct idpf_rx_queue *rxq) +{ + volatile struct virtchnl2_p2p_rx_buf_desc *rxd; + struct rte_mbuf *mbuf = NULL; + uint64_t dma_addr; + uint16_t i; + + for (i = 0; i < rxq->nb_rx_desc; i++) { + mbuf = rte_mbuf_raw_alloc(rxq->mp); + if (unlikely(!mbuf)) { + PMD_DRV_LOG(ERR, "Failed to allocate mbuf for RX"); + return -ENOMEM; + } + + rte_mbuf_refcnt_set(mbuf, 1); + mbuf->next = NULL; + mbuf->data_off = RTE_PKTMBUF_H
[PATCH v4 10/13] common/idpf: add irq map config API
From: Beilei Xing This patch supports idpf_vport_irq_map_config_by_qids API. Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/common/idpf/idpf_common_device.c | 75 drivers/common/idpf/idpf_common_device.h | 4 ++ drivers/common/idpf/version.map | 1 + 3 files changed, 80 insertions(+) diff --git a/drivers/common/idpf/idpf_common_device.c b/drivers/common/idpf/idpf_common_device.c index dc47551b17..cc4207a46e 100644 --- a/drivers/common/idpf/idpf_common_device.c +++ b/drivers/common/idpf/idpf_common_device.c @@ -667,6 +667,81 @@ idpf_vport_irq_map_config(struct idpf_vport *vport, uint16_t nb_rx_queues) return ret; } +int +idpf_vport_irq_map_config_by_qids(struct idpf_vport *vport, uint32_t *qids, uint16_t nb_rx_queues) +{ + struct idpf_adapter *adapter = vport->adapter; + struct virtchnl2_queue_vector *qv_map; + struct idpf_hw *hw = &adapter->hw; + uint32_t dynctl_val, itrn_val; + uint32_t dynctl_reg_start; + uint32_t itrn_reg_start; + uint16_t i; + int ret; + + qv_map = rte_zmalloc("qv_map", +nb_rx_queues * +sizeof(struct virtchnl2_queue_vector), 0); + if (qv_map == NULL) { + DRV_LOG(ERR, "Failed to allocate %d queue-vector map", + nb_rx_queues); + ret = -ENOMEM; + goto qv_map_alloc_err; + } + + /* Rx interrupt disabled, Map interrupt only for writeback */ + + /* The capability flags adapter->caps.other_caps should be +* compared with bit VIRTCHNL2_CAP_WB_ON_ITR here. The if +* condition should be updated when the FW can return the +* correct flag bits. +*/ + dynctl_reg_start = + vport->recv_vectors->vchunks.vchunks->dynctl_reg_start; + itrn_reg_start = + vport->recv_vectors->vchunks.vchunks->itrn_reg_start; + dynctl_val = IDPF_READ_REG(hw, dynctl_reg_start); + DRV_LOG(DEBUG, "Value of dynctl_reg_start is 0x%x", dynctl_val); + itrn_val = IDPF_READ_REG(hw, itrn_reg_start); + DRV_LOG(DEBUG, "Value of itrn_reg_start is 0x%x", itrn_val); + /* Force write-backs by setting WB_ON_ITR bit in DYN_CTL +* register. WB_ON_ITR and INTENA are mutually exclusive +* bits. Setting WB_ON_ITR bits means TX and RX Descs +* are written back based on ITR expiration irrespective +* of INTENA setting. +*/ + /* TBD: need to tune INTERVAL value for better performance. */ + itrn_val = (itrn_val == 0) ? IDPF_DFLT_INTERVAL : itrn_val; + dynctl_val = VIRTCHNL2_ITR_IDX_0 << +PF_GLINT_DYN_CTL_ITR_INDX_S | +PF_GLINT_DYN_CTL_WB_ON_ITR_M | +itrn_val << PF_GLINT_DYN_CTL_INTERVAL_S; + IDPF_WRITE_REG(hw, dynctl_reg_start, dynctl_val); + + for (i = 0; i < nb_rx_queues; i++) { + /* map all queues to the same vector */ + qv_map[i].queue_id = qids[i]; + qv_map[i].vector_id = + vport->recv_vectors->vchunks.vchunks->start_vector_id; + } + vport->qv_map = qv_map; + + ret = idpf_vc_irq_map_unmap_config(vport, nb_rx_queues, true); + if (ret != 0) { + DRV_LOG(ERR, "config interrupt mapping failed"); + goto config_irq_map_err; + } + + return 0; + +config_irq_map_err: + rte_free(vport->qv_map); + vport->qv_map = NULL; + +qv_map_alloc_err: + return ret; +} + int idpf_vport_irq_unmap_config(struct idpf_vport *vport, uint16_t nb_rx_queues) { diff --git a/drivers/common/idpf/idpf_common_device.h b/drivers/common/idpf/idpf_common_device.h index 112367dae8..f767ea7cec 100644 --- a/drivers/common/idpf/idpf_common_device.h +++ b/drivers/common/idpf/idpf_common_device.h @@ -200,5 +200,9 @@ int idpf_vport_info_init(struct idpf_vport *vport, struct virtchnl2_create_vport *vport_info); __rte_internal void idpf_vport_stats_update(struct virtchnl2_vport_stats *oes, struct virtchnl2_vport_stats *nes); +__rte_internal +int idpf_vport_irq_map_config_by_qids(struct idpf_vport *vport, + uint32_t *qids, + uint16_t nb_rx_queues); #endif /* _IDPF_COMMON_DEVICE_H_ */ diff --git a/drivers/common/idpf/version.map b/drivers/common/idpf/version.map index 25624732b0..0729f6b912 100644 --- a/drivers/common/idpf/version.map +++ b/drivers/common/idpf/version.map @@ -69,6 +69,7 @@ INTERNAL { idpf_vport_info_init; idpf_vport_init; idpf_vport_irq_map_config; + idpf_vport_irq_map_config_by_qids; idpf_vport_irq_unmap_config; idpf_vport_rss_config; idpf_vport_stats_update; -- 2.26.2
[PATCH v4 11/13] net/cpfl: enable write back based on ITR expire
From: Beilei Xing This patch enables write back based on ITR expire (WR_ON_ITR) for hairpin queues. Signed-off-by: Mingxia Liu Signed-off-by: Beilei Xing --- drivers/net/cpfl/cpfl_ethdev.c | 13 - 1 file changed, 12 insertions(+), 1 deletion(-) diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c index 8035878602..74f33e9814 100644 --- a/drivers/net/cpfl/cpfl_ethdev.c +++ b/drivers/net/cpfl/cpfl_ethdev.c @@ -735,11 +735,22 @@ cpfl_dev_configure(struct rte_eth_dev *dev) static int cpfl_config_rx_queues_irqs(struct rte_eth_dev *dev) { + uint32_t qids[CPFL_MAX_P2P_NB_QUEUES + IDPF_DEFAULT_RXQ_NUM] = {0}; struct cpfl_vport *cpfl_vport = dev->data->dev_private; struct idpf_vport *vport = &cpfl_vport->base; uint16_t nb_rx_queues = dev->data->nb_rx_queues; + struct cpfl_rx_queue *cpfl_rxq; + int i; - return idpf_vport_irq_map_config(vport, nb_rx_queues); + for (i = 0; i < nb_rx_queues; i++) { + cpfl_rxq = dev->data->rx_queues[i]; + if (cpfl_rxq->hairpin_info.hairpin_q) + qids[i] = cpfl_hw_qid_get(cpfl_vport->p2p_q_chunks_info->rx_start_qid, + (i - cpfl_vport->nb_data_rxq)); + else + qids[i] = cpfl_hw_qid_get(vport->chunks_info.rx_start_qid, i); + } + return idpf_vport_irq_map_config_by_qids(vport, qids, nb_rx_queues); } /* Update hairpin_info for dev's tx hairpin queue */ -- 2.26.2
[PATCH v4 12/13] net/cpfl: support peer ports get
From: Beilei Xing This patch supports get hairpin peer ports. Signed-off-by: Xiao Wang Signed-off-by: Beilei Xing --- drivers/net/cpfl/cpfl_ethdev.c | 40 ++ 1 file changed, 40 insertions(+) diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c index 74f33e9814..d6dc1672f1 100644 --- a/drivers/net/cpfl/cpfl_ethdev.c +++ b/drivers/net/cpfl/cpfl_ethdev.c @@ -1075,6 +1075,45 @@ cpfl_dev_close(struct rte_eth_dev *dev) return 0; } +static int +cpfl_hairpin_get_peer_ports(struct rte_eth_dev *dev, uint16_t *peer_ports, + size_t len, uint32_t tx) +{ + struct cpfl_vport *cpfl_vport = + (struct cpfl_vport *)dev->data->dev_private; + struct idpf_tx_queue *txq; + struct idpf_rx_queue *rxq; + struct cpfl_tx_queue *cpfl_txq; + struct cpfl_rx_queue *cpfl_rxq; + int i, j; + + if (len <= 0) + return -EINVAL; + + if (cpfl_vport->p2p_q_chunks_info == NULL) + return -ENOTSUP; + + if (tx > 0) { + for (i = cpfl_vport->nb_data_txq, j = 0; i < dev->data->nb_tx_queues; i++, j++) { + txq = dev->data->tx_queues[i]; + if (txq == NULL) + return -EINVAL; + cpfl_txq = (struct cpfl_tx_queue *)txq; + peer_ports[j] = cpfl_txq->hairpin_info.peer_rxp; + } + } else if (tx == 0) { + for (i = cpfl_vport->nb_data_rxq, j = 0; i < dev->data->nb_rx_queues; i++, j++) { + rxq = dev->data->rx_queues[i]; + if (rxq == NULL) + return -EINVAL; + cpfl_rxq = (struct cpfl_rx_queue *)rxq; + peer_ports[j] = cpfl_rxq->hairpin_info.peer_txp; + } + } + + return j; +} + static const struct eth_dev_ops cpfl_eth_dev_ops = { .dev_configure = cpfl_dev_configure, .dev_close = cpfl_dev_close, @@ -1104,6 +1143,7 @@ static const struct eth_dev_ops cpfl_eth_dev_ops = { .hairpin_cap_get= cpfl_hairpin_cap_get, .rx_hairpin_queue_setup = cpfl_rx_hairpin_queue_setup, .tx_hairpin_queue_setup = cpfl_tx_hairpin_queue_setup, + .hairpin_get_peer_ports = cpfl_hairpin_get_peer_ports, }; static int -- 2.26.2
[PATCH v4 13/13] net/cpfl: support hairpin bind/unbind
From: Beilei Xing This patch supports hairpin_bind/unbind ops. Signed-off-by: Xiao Wang Signed-off-by: Beilei Xing --- drivers/net/cpfl/cpfl_ethdev.c | 137 + drivers/net/cpfl/cpfl_rxtx.c | 28 +++ drivers/net/cpfl/cpfl_rxtx.h | 2 + 3 files changed, 167 insertions(+) diff --git a/drivers/net/cpfl/cpfl_ethdev.c b/drivers/net/cpfl/cpfl_ethdev.c index d6dc1672f1..4b70441e27 100644 --- a/drivers/net/cpfl/cpfl_ethdev.c +++ b/drivers/net/cpfl/cpfl_ethdev.c @@ -1114,6 +1114,141 @@ cpfl_hairpin_get_peer_ports(struct rte_eth_dev *dev, uint16_t *peer_ports, return j; } +static int +cpfl_hairpin_bind(struct rte_eth_dev *dev, uint16_t rx_port) +{ + struct cpfl_vport *cpfl_tx_vport = dev->data->dev_private; + struct idpf_vport *tx_vport = &cpfl_tx_vport->base; + struct cpfl_vport *cpfl_rx_vport; + struct cpfl_tx_queue *cpfl_txq; + struct cpfl_rx_queue *cpfl_rxq; + struct rte_eth_dev *peer_dev; + struct idpf_vport *rx_vport; + int err = 0; + int i; + + err = cpfl_txq_hairpin_info_update(dev, rx_port); + if (err != 0) { + PMD_DRV_LOG(ERR, "Fail to update Tx hairpin queue info."); + return err; + } + + /* configure hairpin queues */ + for (i = cpfl_tx_vport->nb_data_txq; i < dev->data->nb_tx_queues; i++) { + cpfl_txq = dev->data->tx_queues[i]; + err = cpfl_hairpin_txq_config(tx_vport, cpfl_txq); + if (err != 0) { + PMD_DRV_LOG(ERR, "Fail to configure hairpin Tx queue %u", i); + return err; + } + } + + err = cpfl_hairpin_tx_complq_config(cpfl_tx_vport); + if (err != 0) { + PMD_DRV_LOG(ERR, "Fail to config Tx completion queue"); + return err; + } + + peer_dev = &rte_eth_devices[rx_port]; + cpfl_rx_vport = (struct cpfl_vport *)peer_dev->data->dev_private; + rx_vport = &cpfl_rx_vport->base; + cpfl_rxq_hairpin_mz_bind(peer_dev); + + err = cpfl_hairpin_rx_bufq_config(cpfl_rx_vport); + if (err != 0) { + PMD_DRV_LOG(ERR, "Fail to config Rx buffer queue"); + return err; + } + + for (i = cpfl_rx_vport->nb_data_rxq; i < peer_dev->data->nb_rx_queues; i++) { + cpfl_rxq = peer_dev->data->rx_queues[i]; + err = cpfl_hairpin_rxq_config(rx_vport, cpfl_rxq); + if (err != 0) { + PMD_DRV_LOG(ERR, "Fail to configure hairpin Rx queue %u", i); + return err; + } + err = cpfl_rx_queue_init(peer_dev, i); + if (err != 0) { + PMD_DRV_LOG(ERR, "Fail to init hairpin Rx queue %u", i); + return err; + } + } + + /* enable hairpin queues */ + for (i = cpfl_tx_vport->nb_data_txq; i < dev->data->nb_tx_queues; i++) { + cpfl_txq = dev->data->tx_queues[i]; + err = cpfl_switch_hairpin_rxtx_queue(cpfl_tx_vport, +i - cpfl_tx_vport->nb_data_txq, +false, true); + if (err != 0) { + PMD_DRV_LOG(ERR, "Failed to switch hairpin TX queue %u on", + i); + return err; + } + cpfl_txq->base.q_started = true; + } + + err = cpfl_switch_hairpin_complq(cpfl_tx_vport, true); + if (err != 0) { + PMD_DRV_LOG(ERR, "Failed to switch hairpin Tx complq"); + return err; + } + + for (i = cpfl_rx_vport->nb_data_rxq; i < peer_dev->data->nb_rx_queues; i++) { + cpfl_rxq = peer_dev->data->rx_queues[i]; + err = cpfl_switch_hairpin_rxtx_queue(cpfl_rx_vport, +i - cpfl_rx_vport->nb_data_rxq, +true, true); + if (err != 0) { + PMD_DRV_LOG(ERR, "Failed to switch hairpin RX queue %u on", + i); + } + cpfl_rxq->base.q_started = true; + } + + err = cpfl_switch_hairpin_bufq(cpfl_rx_vport, true); + if (err != 0) { + PMD_DRV_LOG(ERR, "Failed to switch hairpin Rx buffer queue"); + return err; + } + + return 0; +} + +static int +cpfl_hairpin_unbind(struct rte_eth_dev *dev, uint16_t rx_port) +{ + struct cpfl_vport *cpfl_tx_vport = dev->data->dev_private; + struct rte_eth_dev *peer_dev = &rte_eth_devices[rx_port]; + struct cpfl_vport *cpfl_rx_vport = peer_dev->data->dev_private; + struct cpfl_tx_queue *cpfl_txq; + struct cpfl_rx_queue *cpfl_rxq; + int i; + + /* disable hairpin queu
RE: [PATCH] common/mlx5: adjust fork call with the new kernel API
> -Original Message- > From: Stephen Hemminger > Sent: Thursday, May 25, 2023 6:28 PM > To: Slava Ovsiienko > Cc: Erez Ferber ; dev@dpdk.org; Matan Azrad > ; Raslan Darawsheh ; > sta...@dpdk.org > Subject: Re: [PATCH] common/mlx5: adjust fork call with the new kernel API > > On Thu, 25 May 2023 08:10:03 + > Slava Ovsiienko wrote: > > > > -Original Message- > > > From: Stephen Hemminger > > > Sent: Wednesday, May 24, 2023 5:50 PM > > > To: Erez Ferber > > > Cc: dev@dpdk.org; Slava Ovsiienko ; Matan > > > Azrad ; Raslan Darawsheh ; > > > sta...@dpdk.org > > > Subject: Re: [PATCH] common/mlx5: adjust fork call with the new > > > kernel API > > > > > > On Wed, 24 May 2023 15:01:40 +0300 > > > wrote: > > > > > > > From: Erez Ferber > > > > > > > > While doing process fork() the operating system remaps all the > > > > parent process's memory to the address space of the child process > > > > and activates the Copy-on-Write mechanics - it duplicates physical > > > > pages once memory writing happens in the child process. Sometimes > > > > memory duplication is not allowed - for example, if the page > > > > contains hardware queue descriptors. To handle similar issues the > > > > rdma-core library should be prepared for forking. > > > > > > > > The ibv_fork_init() prepares the library to track all the related > > > > memory and prevent it from forking using madvise() system API. > > > > This approach allows fork, but not all the memory is forked to the > > > > child process and, application should care not to touch pages > > > > where the parent application allocated the rdma-core objects. > > > > > > > > The newer kernels propose an option of copy-on-fork for DMA pages > > > > and tracking all the memory and disabling it for the forking is no > > > > longer needed. The new API routine ibv_is_fork_initialized() > > > > should be involved to decide if library initialization for forking is > required. > > > > > > > > Fixes: 0e83b8e536 ("net/mlx5: move rdma-core calls to separate > > > > file") > > > > Cc: sta...@dpdk.org > > > > Signed-off-by: Erez Ferber > > > > > Hi, > > > > > I don't think DPDK applications should fork(), and lots other parts > > > of the shared huge pages will break if an application does this. > > > > I agree - application should not, we have the secondary/primary processes > approach. > > Nonetheless, we have the real use case - DPDK application does fork() and > works well. > > Without mlx5 PMD 😊. With mlx5 it ran into some troubles. Now we have > the solution. > > The problem is you are allowing fork(). And many other libraries may break. > Imagine a DPDK library which has some local mutex and hugepages. > If two forked processes use it then the locks won't work and hugepage data > will be corrupted. IMO, this problem is not fork() specific - applications/libraries supporting secondary/primary process approach should be developed with the similar precautions. And, IIRC, fork() is neither disallowed nor discouraged in DPDK documentation. Moreover, some library unit tests ensure fork() works. No wonder users may develop applications using fork(). They do, and app works well, beside mlx5 PMD. Also, the recommendations to re-design with pri/sec approach was given. With best regards, Slava
Re: [PATCH] ethdev: validate reserved fields
On Thu, May 25, 2023 at 01:39:42PM -0700, Stephen Hemminger wrote: > The various reserved fields added to ethdev could not be > safely used for future extensions because they were never > checked on input. Therefore ABI would be broken if these > fields were added in a future DPDK release. > > Fixes: 436b3a6b6e62 ("ethdev: reserve space in main structs for extension") > Cc: tho...@monjalon.net > Signed-off-by: Stephen Hemminger > --- > lib/ethdev/rte_ethdev.c | 41 + > 1 file changed, 41 insertions(+) > Acked-by: Bruce Richardson
RE: [PATCH v17] app/procinfo: display eventdev xstats
> -Original Message- > From: Sevincer, Abdullah > Signed-off-by: Abdullah Sevincer Acked-by: Reshma Pattan
[PATCH] net/igc: support device I225-LMVP
Add support for I225-LMVP device. Signed-off-by: Qiming Yang --- drivers/net/igc/base/igc_api.c | 1 + drivers/net/igc/base/igc_hw.h | 1 + drivers/net/igc/igc_ethdev.c | 1 + 3 files changed, 3 insertions(+) diff --git a/drivers/net/igc/base/igc_api.c b/drivers/net/igc/base/igc_api.c index c9fc9ed4b0..228b8c8f0e 100644 --- a/drivers/net/igc/base/igc_api.c +++ b/drivers/net/igc/base/igc_api.c @@ -883,6 +883,7 @@ s32 igc_set_mac_type(struct igc_hw *hw) mac->type = igc_i211; break; case IGC_DEV_ID_I225_LM: + case IGC_DEV_ID_I225_LMVP: case IGC_DEV_ID_I225_V: case IGC_DEV_ID_I225_K: case IGC_DEV_ID_I225_I: diff --git a/drivers/net/igc/base/igc_hw.h b/drivers/net/igc/base/igc_hw.h index e919a11c02..83b7bcc1f6 100644 --- a/drivers/net/igc/base/igc_hw.h +++ b/drivers/net/igc/base/igc_hw.h @@ -161,6 +161,7 @@ struct igc_hw; #define IGC_DEV_ID_I210_SGMII_FLASHLESS0x15F6 #define IGC_DEV_ID_I211_COPPER 0x1539 #define IGC_DEV_ID_I225_LM 0x15F2 +#define IGC_DEV_ID_I225_LMVP 0x5502 #define IGC_DEV_ID_I225_V 0x15F3 #define IGC_DEV_ID_I225_K 0x3100 #define IGC_DEV_ID_I225_I 0x15F8 diff --git a/drivers/net/igc/igc_ethdev.c b/drivers/net/igc/igc_ethdev.c index fab2ab6d1c..58c4f80927 100644 --- a/drivers/net/igc/igc_ethdev.c +++ b/drivers/net/igc/igc_ethdev.c @@ -107,6 +107,7 @@ static const struct rte_eth_desc_lim tx_desc_lim = { static const struct rte_pci_id pci_id_igc_map[] = { { RTE_PCI_DEVICE(IGC_INTEL_VENDOR_ID, IGC_DEV_ID_I225_LM) }, + { RTE_PCI_DEVICE(IGC_INTEL_VENDOR_ID, IGC_DEV_ID_I225_LMVP) }, { RTE_PCI_DEVICE(IGC_INTEL_VENDOR_ID, IGC_DEV_ID_I225_V) }, { RTE_PCI_DEVICE(IGC_INTEL_VENDOR_ID, IGC_DEV_ID_I225_I) }, { RTE_PCI_DEVICE(IGC_INTEL_VENDOR_ID, IGC_DEV_ID_I225_IT) }, -- 2.25.1
RE: [v1, 2/3] test/test_cryptodev_asym: add SM2 tests
> Hi Akhil, > > > > If you have the reference from where the vectors are taken it can also be > > mentioned. > > > I generated these test vectors in online. There are no reference test vectors > quoted > In RFC except for recommended curve parameters, at the end. > https://datatracker.ietf.org/doc/html/draft-shen-sm2-ecdsa-02 > ok > > > > > > Signed-off-by: Gowrishankar Muthukrishnan > > > --- > > > app/test/test_cryptodev_asym.c | 506 + > > > app/test/test_cryptodev_sm2_test_vectors.h | 120 + > > > 2 files changed, 626 insertions(+) > > > create mode 100644 app/test/test_cryptodev_sm2_test_vectors.h > > > > > > diff --git a/app/test/test_cryptodev_asym.c > > > b/app/test/test_cryptodev_asym.c index 9236817650..bfaeedee27 100644 > > > --- a/app/test/test_cryptodev_asym.c > > > +++ b/app/test/test_cryptodev_asym.c > > > @@ -21,6 +21,7 @@ > > > #include "test_cryptodev_ecpm_test_vectors.h" > > > #include "test_cryptodev_mod_test_vectors.h" > > > #include "test_cryptodev_rsa_test_vectors.h" > > > +#include "test_cryptodev_sm2_test_vectors.h" > > > #include "test_cryptodev_asym_util.h" > > > #include "test.h" > > > > > > @@ -2196,6 +2197,507 @@ test_ecpm_all_curve(void) > > > return overall_status; > > > } > > > > > > +static int > > > +test_sm2_sign(void) > > > +{ > > > + struct crypto_testsuite_params_asym *ts_params = &testsuite_params; > > > + struct crypto_testsuite_sm2_params input_params = > > > sm2_param_fp256; > > > + struct rte_mempool *sess_mpool = ts_params->session_mpool; > > > + struct rte_mempool *op_mpool = ts_params->op_mpool; > > > + uint8_t dev_id = ts_params->valid_devs[0]; > > > + struct rte_crypto_op *result_op = NULL; > > > + uint8_t output_buf_r[TEST_DATA_SIZE]; > > > + uint8_t output_buf_s[TEST_DATA_SIZE]; > > > + struct rte_crypto_asym_xform xform; > > > + struct rte_crypto_asym_op *asym_op; > > > + struct rte_cryptodev_info dev_info; > > > + struct rte_crypto_op *op = NULL; > > > + int ret, status = TEST_SUCCESS; > > > + void *sess = NULL; > > > + > > > + rte_cryptodev_info_get(dev_id, &dev_info); > > > > dev_info is being unused. Not checking for capabilities? > > It is the same case in other algorithms as well. Shall I collectively address > this ? You can remove this dev_info get from here and add capability check in a separate patch for all cases. > > > + > > > + /* Setup crypto op data structure */ > > > + op = rte_crypto_op_alloc(op_mpool, > > > RTE_CRYPTO_OP_TYPE_ASYMMETRIC); > > > + if (op == NULL) { > > > + RTE_LOG(ERR, USER1, > > > + "line %u FAILED: %s", __LINE__, > > > + "Failed to allocate asymmetric crypto " > > > + "operation struct\n"); > > > + status = TEST_FAILED; > > > + goto exit; > > > + } > > > + asym_op = op->asym; > > > + > > > + /* Setup asym xform */ > > > + xform.next = NULL; > > > + xform.xform_type = RTE_CRYPTO_ASYM_XFORM_SM2; > > > + xform.sm2.pkey.data = input_params.pkey.data; > > > + xform.sm2.pkey.length = input_params.pkey.length; > > > + xform.sm2.q.x.data = input_params.pubkey_qx.data; > > > + xform.sm2.q.x.length = input_params.pubkey_qx.length; > > > + xform.sm2.q.y.data = input_params.pubkey_qy.data; > > > + xform.sm2.q.y.length = input_params.pubkey_qy.length; > > > + > > > + ret = rte_cryptodev_asym_session_create(dev_id, &xform, sess_mpool, > > > &sess); > > > + if (ret < 0) { > > > + RTE_LOG(ERR, USER1, > > > + "line %u FAILED: %s", __LINE__, > > > + "Session creation failed\n"); > > > + status = (ret == -ENOTSUP) ? TEST_SKIPPED : TEST_FAILED; > > > + goto exit; > > > + } > > > + > > > + /* Attach asymmetric crypto session to crypto operations */ > > > + rte_crypto_op_attach_asym_session(op, sess); > > > + > > > + /* Compute sign */ > > > + > > > + /* Populate op with operational details */ > > > + op->asym->sm2.op_type = RTE_CRYPTO_ASYM_OP_SIGN; > > > + op->asym->sm2.message.data = input_params.message.data; > > > + op->asym->sm2.message.length = input_params.message.length; > > > + op->asym->sm2.id.data = input_params.id.data; > > > + op->asym->sm2.id.length = input_params.id.length; > > > + > > > + /* Init out buf */ > > > + op->asym->sm2.r.data = output_buf_r; > > > + op->asym->sm2.s.data = output_buf_s; > > > + > > > + RTE_LOG(DEBUG, USER1, "Process ASYM operation\n"); > > > + > > > + /* Process crypto operation */ > > > + if (rte_cryptodev_enqueue_burst(dev_id, 0, &op, 1) != 1) { > > > + RTE_LOG(ERR, USER1, > > > + "line %u FAILED: %s", __LINE__, > > > + "Error sending packet for operation\n"); > > > + status = TEST_FAILED; > > > + goto exit; > > > + } > > > + > > > + while (rte_cryptodev_dequeue_burst(dev_id, 0, &result_op, 1) == 0) > > > + rte_pause(); > > > > Shouldn't this be a finite loop and mark test as failed after some retries?
Re: [PATCH v3 04/28] vhost: add helper of IOTLB entries coredump
On Thu, May 25, 2023 at 6:26 PM Maxime Coquelin wrote: > @@ -149,8 +177,8 @@ vhost_user_iotlb_cache_remove_all(struct virtio_net *dev, > struct vhost_virtqueue > rte_rwlock_write_lock(&vq->iotlb_lock); > > RTE_TAILQ_FOREACH_SAFE(node, &vq->iotlb_list, next, temp_node) { > - mem_set_dump((void *)(uintptr_t)node->uaddr, node->size, > false, > - hua_to_alignment(dev->mem, (void > *)(uintptr_t)node->uaddr)); > + vhost_user_iotlb_set_dump(dev, node); vhost_user_iotlb_clear_dump ? > + > TAILQ_REMOVE(&vq->iotlb_list, node, next); > vhost_user_iotlb_pool_put(vq, node); > } -- David Marchand
[PATCH v4] dmadev: add tracepoints
Add tracepoints at important APIs for tracing support. Signed-off-by: Chengwen Feng Acked-by: Morten Brørup --- v4: Fix asan smoke fail. v3: Address Morten's comment: Move stats_get and vchan_status and to trace_fp.h. v2: Address Morten's comment: Make stats_get as fast-path trace-points. Place fast-path trace-point functions behind in version.map. --- lib/dmadev/meson.build | 2 +- lib/dmadev/rte_dmadev.c | 39 +-- lib/dmadev/rte_dmadev.h | 56 +++--- lib/dmadev/rte_dmadev_trace.h| 118 + lib/dmadev/rte_dmadev_trace_fp.h | 150 +++ lib/dmadev/rte_dmadev_trace_points.c | 59 +++ lib/dmadev/version.map | 10 ++ 7 files changed, 413 insertions(+), 21 deletions(-) create mode 100644 lib/dmadev/rte_dmadev_trace.h create mode 100644 lib/dmadev/rte_dmadev_trace_fp.h create mode 100644 lib/dmadev/rte_dmadev_trace_points.c diff --git a/lib/dmadev/meson.build b/lib/dmadev/meson.build index 2f17587b75..e0d90aea67 100644 --- a/lib/dmadev/meson.build +++ b/lib/dmadev/meson.build @@ -1,7 +1,7 @@ # SPDX-License-Identifier: BSD-3-Clause # Copyright(c) 2021 HiSilicon Limited. -sources = files('rte_dmadev.c') +sources = files('rte_dmadev.c', 'rte_dmadev_trace_points.c') headers = files('rte_dmadev.h') indirect_headers += files('rte_dmadev_core.h') driver_sdk_headers += files('rte_dmadev_pmd.h') diff --git a/lib/dmadev/rte_dmadev.c b/lib/dmadev/rte_dmadev.c index 8c095e1f35..25fa78de8f 100644 --- a/lib/dmadev/rte_dmadev.c +++ b/lib/dmadev/rte_dmadev.c @@ -17,6 +17,7 @@ #include "rte_dmadev.h" #include "rte_dmadev_pmd.h" +#include "rte_dmadev_trace.h" static int16_t dma_devices_max; @@ -434,6 +435,8 @@ rte_dma_info_get(int16_t dev_id, struct rte_dma_info *dev_info) dev_info->numa_node = dev->device->numa_node; dev_info->nb_vchans = dev->data->dev_conf.nb_vchans; + rte_dma_trace_info_get(dev_id, dev_info); + return 0; } @@ -483,6 +486,8 @@ rte_dma_configure(int16_t dev_id, const struct rte_dma_conf *dev_conf) memcpy(&dev->data->dev_conf, dev_conf, sizeof(struct rte_dma_conf)); + rte_dma_trace_configure(dev_id, dev_conf, ret); + return ret; } @@ -509,6 +514,7 @@ rte_dma_start(int16_t dev_id) goto mark_started; ret = (*dev->dev_ops->dev_start)(dev); + rte_dma_trace_start(dev_id, ret); if (ret != 0) return ret; @@ -535,6 +541,7 @@ rte_dma_stop(int16_t dev_id) goto mark_stopped; ret = (*dev->dev_ops->dev_stop)(dev); + rte_dma_trace_stop(dev_id, ret); if (ret != 0) return ret; @@ -565,6 +572,8 @@ rte_dma_close(int16_t dev_id) if (ret == 0) dma_release(dev); + rte_dma_trace_close(dev_id, ret); + return ret; } @@ -655,14 +664,18 @@ rte_dma_vchan_setup(int16_t dev_id, uint16_t vchan, if (*dev->dev_ops->vchan_setup == NULL) return -ENOTSUP; - return (*dev->dev_ops->vchan_setup)(dev, vchan, conf, + ret = (*dev->dev_ops->vchan_setup)(dev, vchan, conf, sizeof(struct rte_dma_vchan_conf)); + rte_dma_trace_vchan_setup(dev_id, vchan, conf, ret); + + return ret; } int rte_dma_stats_get(int16_t dev_id, uint16_t vchan, struct rte_dma_stats *stats) { const struct rte_dma_dev *dev = &rte_dma_devices[dev_id]; + int ret; if (!rte_dma_is_valid(dev_id) || stats == NULL) return -EINVAL; @@ -677,14 +690,18 @@ rte_dma_stats_get(int16_t dev_id, uint16_t vchan, struct rte_dma_stats *stats) if (*dev->dev_ops->stats_get == NULL) return -ENOTSUP; memset(stats, 0, sizeof(struct rte_dma_stats)); - return (*dev->dev_ops->stats_get)(dev, vchan, stats, - sizeof(struct rte_dma_stats)); + ret = (*dev->dev_ops->stats_get)(dev, vchan, stats, +sizeof(struct rte_dma_stats)); + rte_dma_trace_stats_get(dev_id, vchan, stats, ret); + + return ret; } int rte_dma_stats_reset(int16_t dev_id, uint16_t vchan) { struct rte_dma_dev *dev = &rte_dma_devices[dev_id]; + int ret; if (!rte_dma_is_valid(dev_id)) return -EINVAL; @@ -698,13 +715,17 @@ rte_dma_stats_reset(int16_t dev_id, uint16_t vchan) if (*dev->dev_ops->stats_reset == NULL) return -ENOTSUP; - return (*dev->dev_ops->stats_reset)(dev, vchan); + ret = (*dev->dev_ops->stats_reset)(dev, vchan); + rte_dma_trace_stats_reset(dev_id, vchan, ret); + + return ret; } int rte_dma_vchan_status(int16_t dev_id, uint16_t vchan, enum rte_dma_vchan_status *status) { struct rte_dma_dev *dev = &rte_dma_devices[dev_id]; + int ret; if
Re: [PATCH v3 14/28] vhost: add helper for interrupt injection
On Thu, May 25, 2023 at 6:26 PM Maxime Coquelin wrote: > @@ -900,6 +905,24 @@ vhost_need_event(uint16_t event_idx, uint16_t new_idx, > uint16_t old) > return (uint16_t)(new_idx - event_idx - 1) < (uint16_t)(new_idx - > old); > } > > +static __rte_always_inline void > +vhost_vring_inject_irq(struct virtio_net *dev, struct vhost_virtqueue *vq) > +{ > + int ret; > + > + ret = dev->backend_ops->inject_irq(dev, vq); > + if (ret) { No need for ret. > + if (dev->flags & VIRTIO_DEV_STATS_ENABLED) > + vq->stats.guest_notifications_error++; > + return; > + } > + > + if (dev->flags & VIRTIO_DEV_STATS_ENABLED) > + vq->stats.guest_notifications++; > + if (dev->notify_ops->guest_notified) > + dev->notify_ops->guest_notified(dev->vid); > +} > + > static __rte_always_inline void > vhost_vring_call_split(struct virtio_net *dev, struct vhost_virtqueue *vq) > { -- David Marchand
Re: [PATCH v3 32/32] common/cnxk: add check for null auth and anti-replay
On Thu, May 25, 2023 at 3:41 PM Nithin Dabilpuram wrote: > > From: Srujana Challa > > As per IPsec RFC, the anti-replay service can be selected for > an SA only if the integrity service is selected for that SA. > This patch adds the validation check for the same. > > Signed-off-by: Srujana Challa Series applied to dpdk-next-net-mrvl/for-next-net. Thanks > --- > drivers/common/cnxk/cnxk_security.c | 9 + > 1 file changed, 9 insertions(+) > > diff --git a/drivers/common/cnxk/cnxk_security.c > b/drivers/common/cnxk/cnxk_security.c > index 13ca2c7791..a8c3ba90cd 100644 > --- a/drivers/common/cnxk/cnxk_security.c > +++ b/drivers/common/cnxk/cnxk_security.c > @@ -155,6 +155,10 @@ ot_ipsec_sa_common_param_fill(union > roc_ot_ipsec_sa_word2 *w2, > > switch (auth_xfrm->auth.algo) { > case RTE_CRYPTO_AUTH_NULL: > + if (w2->s.dir == ROC_IE_SA_DIR_INBOUND && > ipsec_xfrm->replay_win_sz) { > + plt_err("anti-replay can't be supported with > integrity service disabled"); > + return -EINVAL; > + } > w2->s.auth_type = ROC_IE_OT_SA_AUTH_NULL; > break; > case RTE_CRYPTO_AUTH_SHA1_HMAC: > @@ -1392,6 +1396,11 @@ cnxk_on_ipsec_inb_sa_create(struct > rte_security_ipsec_xform *ipsec, > if (ret) > return ret; > > + if (crypto_xform->type != RTE_CRYPTO_SYM_XFORM_AEAD && > + crypto_xform->auth.algo == RTE_CRYPTO_AUTH_NULL && > ipsec->replay_win_sz) { > + plt_err("anti-replay can't be supported with integrity > service disabled"); > + return -EINVAL; > + } > if (crypto_xform->type == RTE_CRYPTO_SYM_XFORM_AEAD || > auth_xform->auth.algo == RTE_CRYPTO_AUTH_NULL || > auth_xform->auth.algo == RTE_CRYPTO_AUTH_AES_GMAC) { > -- > 2.25.1 >
Re: [PATCH v3 15/28] vhost: add API to set max queue pairs
On Thu, May 25, 2023 at 6:27 PM Maxime Coquelin wrote: > diff --git a/doc/guides/prog_guide/vhost_lib.rst > b/doc/guides/prog_guide/vhost_lib.rst > index e8bb8c9b7b..cd4b109139 100644 > --- a/doc/guides/prog_guide/vhost_lib.rst > +++ b/doc/guides/prog_guide/vhost_lib.rst > @@ -334,6 +334,10 @@ The following is an overview of some key Vhost API > functions: >Clean DMA vChannel finished to use. After this function is called, >the specified DMA vChannel should no longer be used by the Vhost library. > > +* ``rte_vhost_driver_set_max_queue_num(path, max_queue_pairs)`` > + > + Set the maximum number of queue pairs supported by the device. > + > Vhost-user Implementations > -- > > diff --git a/doc/guides/rel_notes/release_23_07.rst > b/doc/guides/rel_notes/release_23_07.rst > index a9b1293689..fa889a5ee7 100644 > --- a/doc/guides/rel_notes/release_23_07.rst > +++ b/doc/guides/rel_notes/release_23_07.rst > @@ -55,6 +55,11 @@ New Features > Also, make sure to start the actual text at the margin. > === > > +* **Added Vhost API to set maximum queue pairs supported **Added Vhost API to set maximum queue pairs supported.** > + > + Introduced ``rte_vhost_driver_set_max_queue_num()`` to be able to limit the > + maximum number of supported queue pairs, required for VDUSE support. > + > > Removed Items > - -- David Marchand
[PATCH v8] enhance NUMA affinity heuristic
When a DPDK application is started on only one numa node, memory is allocated for only one socket. When interrupt threads use memory, memory may not be found on the socket where the interrupt thread is currently located, and memory has to be reallocated on the hugepage, this operation will lead to performance degradation. Fixes: 705356f0811f ("eal: simplify control thread creation") Fixes: 770d41bf3309 ("malloc: fix allocation with unknown socket ID") Cc: sta...@dpdk.org Signed-off-by: Kaisen You --- Changes since v7: - Update commet, Changes since v6: - New explanation for easy understanding, Changes since v5: - Add comments to the code, Changes since v4: - mod the patch title, Changes since v3: - add the assignment of socket_id in thread initialization, Changes since v2: - add uncommitted local change and fix compilation, Changes since v1: - accomodate for configurations with main lcore running on multiples physical cores belonging to different numa, --- lib/eal/common/eal_common_thread.c | 4 lib/eal/common/malloc_heap.c | 11 ++- 2 files changed, 14 insertions(+), 1 deletion(-) diff --git a/lib/eal/common/eal_common_thread.c b/lib/eal/common/eal_common_thread.c index 079a385630..22480aa61f 100644 --- a/lib/eal/common/eal_common_thread.c +++ b/lib/eal/common/eal_common_thread.c @@ -252,6 +252,10 @@ static int ctrl_thread_init(void *arg) struct rte_thread_ctrl_params *params = arg; __rte_thread_init(rte_lcore_id(), cpuset); + /* Set control thread socket ID to SOCKET_ID_ANY as control +* threads may be scheduled on any NUMA node. +*/ + RTE_PER_LCORE(_socket_id) = SOCKET_ID_ANY; params->ret = rte_thread_set_affinity_by_id(rte_thread_self(), cpuset); if (params->ret != 0) { __atomic_store_n(¶ms->ctrl_thread_status, diff --git a/lib/eal/common/malloc_heap.c b/lib/eal/common/malloc_heap.c index d25bdc98f9..d833a71e7a 100644 --- a/lib/eal/common/malloc_heap.c +++ b/lib/eal/common/malloc_heap.c @@ -716,7 +716,16 @@ malloc_get_numa_socket(void) if (conf->socket_mem[socket_id] != 0) return socket_id; } - + /* We couldn't find quickly find a NUMA node where memory was available, +* so fall back to using main lcore socket ID. +*/ + socket_id = rte_lcore_to_socket_id(rte_get_main_lcore()); + /* Main lcore socket ID may be SOCKET_ID_ANY in cases when main lcore +* thread is affinitized to multiple NUMA nodes. +*/ + if (socket_id != (unsigned int)SOCKET_ID_ANY) + return socket_id; + /* Failed to find meaningful socket ID, so just use the first one available */ return rte_socket_id_by_idx(0); } -- 2.25.1
Re: [PATCH] net/qede: fix RSS indirection table initialization
On Thu, May 25, 2023 at 7:24 PM Devendra Singh Rawat wrote: > > RSS indirection table was not updated for VF ports during RSS hash > update, this resulted in device start failure. > This fix sets update rss indirection table in VFPF channel request, > sent for updating RSS hash for VF. > > Fixes: 7ab35bf6b97b ("net/qede: fix RSS") > Cc: sta...@dpdk.org > > Signed-off-by: Devendra Singh Rawat > Signed-off-by: Alok Prasad Applied to dpdk-next-net-mrvl/for-next-net. Thanks > --- > drivers/net/qede/qede_ethdev.c | 1 + > 1 file changed, 1 insertion(+) > > diff --git a/drivers/net/qede/qede_ethdev.c b/drivers/net/qede/qede_ethdev.c > index a4923670d6..22cd470646 100644 > --- a/drivers/net/qede/qede_ethdev.c > +++ b/drivers/net/qede/qede_ethdev.c > @@ -2142,6 +2142,7 @@ int qede_rss_hash_update(struct rte_eth_dev *eth_dev, > rss_params.rss_enable = 1; > } > > + rss_params.update_rss_ind_table = 1; > rss_params.update_rss_config = 1; > /* tbl_size has to be set with capabilities */ > rss_params.rss_table_size_log = 7; > -- > 2.18.2 >
Re: [PATCH v3 18/28] vhost: add VDUSE device creation and destruction
On Thu, May 25, 2023 at 6:27 PM Maxime Coquelin wrote: > diff --git a/lib/vhost/vduse.c b/lib/vhost/vduse.c > new file mode 100644 > index 00..d67818bfb5 > --- /dev/null > +++ b/lib/vhost/vduse.c [snip] > +#define VDUSE_NET_SUPPORTED_FEATURES ((1ULL << VIRTIO_NET_F_MRG_RXBUF) | \ > + (1ULL << VIRTIO_F_ANY_LAYOUT) | \ > + (1ULL << VIRTIO_F_VERSION_1) | \ > + (1ULL << VIRTIO_NET_F_GSO) | \ > + (1ULL << VIRTIO_NET_F_HOST_TSO4) | \ > + (1ULL << VIRTIO_NET_F_HOST_TSO6) | \ > + (1ULL << VIRTIO_NET_F_HOST_UFO) | \ > + (1ULL << VIRTIO_NET_F_HOST_ECN) | \ > + (1ULL << VIRTIO_NET_F_CSUM)| \ > + (1ULL << VIRTIO_NET_F_GUEST_CSUM) | \ > + (1ULL << VIRTIO_NET_F_GUEST_TSO4) | \ > + (1ULL << VIRTIO_NET_F_GUEST_TSO6) | \ > + (1ULL << VIRTIO_NET_F_GUEST_UFO) | \ > + (1ULL << VIRTIO_NET_F_GUEST_ECN) | \ > + (1ULL << VIRTIO_RING_F_INDIRECT_DESC) | \ > + (1ULL << VIRTIO_F_IN_ORDER) | \ > + (1ULL << VIRTIO_F_IOMMU_PLATFORM)) That's a lot of indent/spaces. #define VDUSE_NET_SUPPORTED_FEATURES (\ (1ULL << VIRTIO_NET_F_MRG_RXBUF) | \ (1ULL << VIRTIO_F_ANY_LAYOUT) | \ Plus, can't we use RTE_BIT64? (this could be a cleanup to do on the whole vhost library) -- David Marchand
[PATCH v2 0/3] SM2 crypto algorithm support
This patch series adds SM2 crypto algorithm support, along with tests verified using Openssl. v2: - addressed suggestions in v1. Gowrishankar Muthukrishnan (3): cryptodev: add SM2 asymmetric crypto algorithm test/crypto: add asymmetric SM2 test cases crypto/openssl: add SM2 asymmetric crypto support app/test/test_cryptodev_asym.c | 504 +++ app/test/test_cryptodev_sm2_test_vectors.h | 120 + doc/guides/cryptodevs/features/default.ini | 1 + doc/guides/cryptodevs/features/openssl.ini | 1 + doc/guides/cryptodevs/openssl.rst| 1 + doc/guides/rel_notes/release_23_07.rst | 9 + drivers/crypto/openssl/openssl_pmd_private.h | 7 + drivers/crypto/openssl/rte_openssl_pmd.c | 245 + drivers/crypto/openssl/rte_openssl_pmd_ops.c | 101 lib/cryptodev/rte_crypto_asym.h | 77 +++ lib/cryptodev/rte_cryptodev.c| 1 + 11 files changed, 1067 insertions(+) create mode 100644 app/test/test_cryptodev_sm2_test_vectors.h -- 2.25.1
[PATCH v2 1/3] cryptodev: add SM2 asymmetric crypto algorithm
ShangMi 2 (SM2) is a encryption and digital signature algorithm used in the Chinese National Standard. Added support for asymmetric SM2 in cryptodev along with prime field curve, as referenced in RFC: https://datatracker.ietf.org/doc/html/draft-shen-sm2-ecdsa-02 Signed-off-by: Gowrishankar Muthukrishnan --- doc/guides/cryptodevs/features/default.ini | 1 + doc/guides/rel_notes/release_23_07.rst | 5 ++ lib/cryptodev/rte_crypto_asym.h| 77 ++ lib/cryptodev/rte_cryptodev.c | 1 + 4 files changed, 84 insertions(+) diff --git a/doc/guides/cryptodevs/features/default.ini b/doc/guides/cryptodevs/features/default.ini index 523da0cfa8..a69967bb9e 100644 --- a/doc/guides/cryptodevs/features/default.ini +++ b/doc/guides/cryptodevs/features/default.ini @@ -125,6 +125,7 @@ Diffie-hellman = ECDSA = ECPM= ECDH= +SM2 = ; ; Supported Operating systems of a default crypto driver. diff --git a/doc/guides/rel_notes/release_23_07.rst b/doc/guides/rel_notes/release_23_07.rst index a9b1293689..8b8e69d619 100644 --- a/doc/guides/rel_notes/release_23_07.rst +++ b/doc/guides/rel_notes/release_23_07.rst @@ -55,6 +55,11 @@ New Features Also, make sure to start the actual text at the margin. === +* **Added SM2 asymmetric algorithm in cryptodev.** + + Added support for ShamMi 2 (SM2) asymmetric crypto algorithm + along with prime field curve support. + Removed Items - diff --git a/lib/cryptodev/rte_crypto_asym.h b/lib/cryptodev/rte_crypto_asym.h index 989f38323f..35fa2c0a6d 100644 --- a/lib/cryptodev/rte_crypto_asym.h +++ b/lib/cryptodev/rte_crypto_asym.h @@ -119,6 +119,11 @@ enum rte_crypto_asym_xform_type { /**< Elliptic Curve Point Multiplication */ RTE_CRYPTO_ASYM_XFORM_ECFPM, /**< Elliptic Curve Fixed Point Multiplication */ + RTE_CRYPTO_ASYM_XFORM_SM2, + /**< ShangMi 2 +* Performs Encrypt, Decrypt, Sign and Verify. +* Refer to rte_crypto_asym_op_type. +*/ RTE_CRYPTO_ASYM_XFORM_TYPE_LIST_END /**< End of list */ }; @@ -382,6 +387,20 @@ struct rte_crypto_ec_xform { /**< Pre-defined ec groups */ }; +/** + * Asymmetric SM2 transform data + * + * Structure describing SM2 xform params + * + */ +struct rte_crypto_sm2_xform { + rte_crypto_uint pkey; + /**< Private key of the signer for signature generation. */ + + struct rte_crypto_ec_point q; + /**< Public key of the signer for verification. */ +}; + /** * Operations params for modular operations: * exponentiation and multiplicative inverse @@ -637,9 +656,66 @@ struct rte_crypto_asym_xform { /**< EC xform parameters, used by elliptic curve based * operations. */ + + struct rte_crypto_sm2_xform sm2; + /**< SM2 xform parameters */ }; }; +/** + * SM2 operation params + */ +struct rte_crypto_sm2_op_param { + enum rte_crypto_asym_op_type op_type; + /**< Signature generation or verification */ + + rte_crypto_param message; + /**< +* Pointer to input data +* - to be encrypted for SM2 public encrypt. +* - to be signed for SM2 sign generation. +* - to be authenticated for SM2 sign verification. +* +* Pointer to output data +* - for SM2 private decrypt. +* In this case the underlying array should have been +* allocated with enough memory to hold plaintext output +* (at least encrypted text length). The message.length field +* will be overwritten by the PMD with the decrypted length. +*/ + + rte_crypto_param cipher; + /**< +* Pointer to input data +* - to be decrypted for SM2 private decrypt. +* +* Pointer to output data +* - for SM2 public encrypt. +* In this case the underlying array should have been allocated +* with enough memory to hold ciphertext output (at least X bytes +* for prime field curve of N bytes and for message M bytes, +* where X = (C1 + C2 + C3) and computed based on SM2 RFC as +* C1 (1 + N + N), C2 = M, C3 = N. The cipher.length field will +* be overwritten by the PMD with the encrypted length. +*/ + + rte_crypto_uint id; + /**< The SM2 id used by signer and verifier and is in interval (1, n-1). */ + + rte_crypto_uint r; + /**< r component of elliptic curve signature +* output : for signature generation (of at least N bytes +* where prime field length is N bytes) +* input : for signature verification +*/ + rte_crypto_uint s; + /**< s component of elliptic curve signature +* output : for signature generation (of at leas
[PATCH v2 2/3] test/crypto: add asymmetric SM2 test cases
Added test cases and test vectors for asymmetric SM2 crypto verification. Cases are added for sign/verify/encrypt/decrypt. Signed-off-by: Gowrishankar Muthukrishnan --- app/test/test_cryptodev_asym.c | 504 + app/test/test_cryptodev_sm2_test_vectors.h | 120 + 2 files changed, 624 insertions(+) create mode 100644 app/test/test_cryptodev_sm2_test_vectors.h diff --git a/app/test/test_cryptodev_asym.c b/app/test/test_cryptodev_asym.c index 9236817650..68029c9910 100644 --- a/app/test/test_cryptodev_asym.c +++ b/app/test/test_cryptodev_asym.c @@ -21,6 +21,7 @@ #include "test_cryptodev_ecpm_test_vectors.h" #include "test_cryptodev_mod_test_vectors.h" #include "test_cryptodev_rsa_test_vectors.h" +#include "test_cryptodev_sm2_test_vectors.h" #include "test_cryptodev_asym_util.h" #include "test.h" @@ -2196,6 +2197,505 @@ test_ecpm_all_curve(void) return overall_status; } +static int +test_sm2_sign(void) +{ + struct crypto_testsuite_params_asym *ts_params = &testsuite_params; + struct crypto_testsuite_sm2_params input_params = sm2_param_fp256; + struct rte_mempool *sess_mpool = ts_params->session_mpool; + struct rte_mempool *op_mpool = ts_params->op_mpool; + uint8_t dev_id = ts_params->valid_devs[0]; + struct rte_crypto_op *result_op = NULL; + uint8_t output_buf_r[TEST_DATA_SIZE]; + uint8_t output_buf_s[TEST_DATA_SIZE]; + struct rte_crypto_asym_xform xform; + struct rte_crypto_asym_op *asym_op; + struct rte_crypto_op *op = NULL; + int ret, status = TEST_SUCCESS; + void *sess = NULL; + + /* Setup crypto op data structure */ + op = rte_crypto_op_alloc(op_mpool, RTE_CRYPTO_OP_TYPE_ASYMMETRIC); + if (op == NULL) { + RTE_LOG(ERR, USER1, + "line %u FAILED: %s", __LINE__, + "Failed to allocate asymmetric crypto " + "operation struct\n"); + status = TEST_FAILED; + goto exit; + } + asym_op = op->asym; + + /* Setup asym xform */ + xform.next = NULL; + xform.xform_type = RTE_CRYPTO_ASYM_XFORM_SM2; + xform.sm2.pkey.data = input_params.pkey.data; + xform.sm2.pkey.length = input_params.pkey.length; + xform.sm2.q.x.data = input_params.pubkey_qx.data; + xform.sm2.q.x.length = input_params.pubkey_qx.length; + xform.sm2.q.y.data = input_params.pubkey_qy.data; + xform.sm2.q.y.length = input_params.pubkey_qy.length; + + ret = rte_cryptodev_asym_session_create(dev_id, &xform, sess_mpool, &sess); + if (ret < 0) { + RTE_LOG(ERR, USER1, + "line %u FAILED: %s", __LINE__, + "Session creation failed\n"); + status = (ret == -ENOTSUP) ? TEST_SKIPPED : TEST_FAILED; + goto exit; + } + + /* Attach asymmetric crypto session to crypto operations */ + rte_crypto_op_attach_asym_session(op, sess); + + /* Compute sign */ + + /* Populate op with operational details */ + op->asym->sm2.op_type = RTE_CRYPTO_ASYM_OP_SIGN; + op->asym->sm2.message.data = input_params.message.data; + op->asym->sm2.message.length = input_params.message.length; + op->asym->sm2.id.data = input_params.id.data; + op->asym->sm2.id.length = input_params.id.length; + + /* Init out buf */ + op->asym->sm2.r.data = output_buf_r; + op->asym->sm2.s.data = output_buf_s; + + RTE_LOG(DEBUG, USER1, "Process ASYM operation\n"); + + /* Process crypto operation */ + if (rte_cryptodev_enqueue_burst(dev_id, 0, &op, 1) != 1) { + RTE_LOG(ERR, USER1, + "line %u FAILED: %s", __LINE__, + "Error sending packet for operation\n"); + status = TEST_FAILED; + goto exit; + } + + while (rte_cryptodev_dequeue_burst(dev_id, 0, &result_op, 1) == 0) + rte_pause(); + + if (result_op == NULL) { + RTE_LOG(ERR, USER1, + "line %u FAILED: %s", __LINE__, + "Failed to process asym crypto op\n"); + status = TEST_FAILED; + goto exit; + } + + if (result_op->status != RTE_CRYPTO_OP_STATUS_SUCCESS) { + RTE_LOG(ERR, USER1, + "line %u FAILED: %s", __LINE__, + "Failed to process asym crypto op\n"); + status = TEST_FAILED; + goto exit; + } + + asym_op = result_op->asym; + + debug_hexdump(stdout, "r:", + asym_op->sm2.r.data, asym_op->sm2.r.length); + debug_hexdump(stdout, "s:", + asym_op->sm2.s.data, asym_op->sm2.s.length); + + /* Verify sign (in roundtrip). +
[PATCH v2 3/3] crypto/openssl: add SM2 asymmetric crypto support
Add SM2 asymmetric algorithm support in openssl PMD. Signed-off-by: Gowrishankar Muthukrishnan --- doc/guides/cryptodevs/features/openssl.ini | 1 + doc/guides/cryptodevs/openssl.rst| 1 + doc/guides/rel_notes/release_23_07.rst | 4 + drivers/crypto/openssl/openssl_pmd_private.h | 7 + drivers/crypto/openssl/rte_openssl_pmd.c | 245 +++ drivers/crypto/openssl/rte_openssl_pmd_ops.c | 101 6 files changed, 359 insertions(+) diff --git a/doc/guides/cryptodevs/features/openssl.ini b/doc/guides/cryptodevs/features/openssl.ini index 4b0f9b162e..b64c8ec4a5 100644 --- a/doc/guides/cryptodevs/features/openssl.ini +++ b/doc/guides/cryptodevs/features/openssl.ini @@ -65,6 +65,7 @@ DSA = Y Modular Exponentiation = Y Modular Inversion = Y Diffie-hellman = Y +SM2 = Y ; ; Supported Operating systems of the 'openssl' crypto driver. diff --git a/doc/guides/cryptodevs/openssl.rst b/doc/guides/cryptodevs/openssl.rst index 03041ceda1..ff21d21b23 100644 --- a/doc/guides/cryptodevs/openssl.rst +++ b/doc/guides/cryptodevs/openssl.rst @@ -53,6 +53,7 @@ Supported Asymmetric Crypto algorithms: * ``RTE_CRYPTO_ASYM_XFORM_DH`` * ``RTE_CRYPTO_ASYM_XFORM_MODINV`` * ``RTE_CRYPTO_ASYM_XFORM_MODEX`` +* ``RTE_CRYPTO_ASYM_XFORM_SM2`` Installation diff --git a/doc/guides/rel_notes/release_23_07.rst b/doc/guides/rel_notes/release_23_07.rst index 8b8e69d619..aeebcffb60 100644 --- a/doc/guides/rel_notes/release_23_07.rst +++ b/doc/guides/rel_notes/release_23_07.rst @@ -61,6 +61,10 @@ New Features along with prime field curve support. +* **Updated OpenSSL crypto driver for SM2 support.** + + Added SM2 algorithm support in asymmetric crypto operations. + Removed Items - diff --git a/drivers/crypto/openssl/openssl_pmd_private.h b/drivers/crypto/openssl/openssl_pmd_private.h index ed6841e460..d7990c8333 100644 --- a/drivers/crypto/openssl/openssl_pmd_private.h +++ b/drivers/crypto/openssl/openssl_pmd_private.h @@ -12,6 +12,7 @@ #include #include #include +#include #if (OPENSSL_VERSION_NUMBER >= 0x3000L) #include #include @@ -200,6 +201,12 @@ struct openssl_asym_session { OSSL_PARAM_BLD * param_bld; #endif } s; + struct { +#if (OPENSSL_VERSION_NUMBER >= 0x3000L) + OSSL_PARAM *key_params; + OSSL_PARAM *enc_params; +#endif + } sm2; } u; } __rte_cache_aligned; /** Set and validate OPENSSL crypto session parameters */ diff --git a/drivers/crypto/openssl/rte_openssl_pmd.c b/drivers/crypto/openssl/rte_openssl_pmd.c index 384d262621..25feb88ac3 100644 --- a/drivers/crypto/openssl/rte_openssl_pmd.c +++ b/drivers/crypto/openssl/rte_openssl_pmd.c @@ -13,6 +13,7 @@ #include #include #include +#include #include "openssl_pmd_private.h" #include "compat.h" @@ -2662,6 +2663,234 @@ process_openssl_rsa_op_evp(struct rte_crypto_op *cop, return ret; } + +static int +process_openssl_sm2_op_evp(struct rte_crypto_op *cop, + struct openssl_asym_session *sess) +{ + EVP_PKEY_CTX *kctx = NULL, *sctx = NULL, *cctx = NULL; + OSSL_PARAM *kparams = sess->u.sm2.key_params; + struct rte_crypto_asym_op *op = cop->asym; + EVP_PKEY *pkey = NULL; + int ret = -1; + + cop->status = RTE_CRYPTO_OP_STATUS_ERROR; + + if (!kparams) + return ret; + + switch (op->sm2.op_type) { + case RTE_CRYPTO_ASYM_OP_ENCRYPT: + { + OSSL_PARAM *eparams = sess->u.sm2.enc_params; + size_t output_len; + + kctx = EVP_PKEY_CTX_new_id(EVP_PKEY_SM2, NULL); + if (kctx == NULL || EVP_PKEY_fromdata_init(kctx) <= 0 || + EVP_PKEY_fromdata(kctx, &pkey, EVP_PKEY_KEYPAIR, kparams) <= 0) + goto err_sm2; + + cctx = EVP_PKEY_CTX_new_from_pkey(NULL, pkey, NULL); + if (!cctx) + goto err_sm2; + + if (!EVP_PKEY_encrypt_init(cctx)) + goto err_sm2; + + if (!EVP_PKEY_CTX_set_params(cctx, eparams)) + goto err_sm2; + + if (!EVP_PKEY_encrypt(cctx, op->sm2.cipher.data, &output_len, + op->sm2.message.data, + op->sm2.message.length)) + goto err_sm2; + op->sm2.cipher.length = output_len; + } + break; + case RTE_CRYPTO_ASYM_OP_DECRYPT: + { + OSSL_PARAM *eparams = sess->u.sm2.enc_params; + + kctx = EVP_PKEY_CTX_new_id(EVP_PKEY_SM2, NULL); + if (kctx ==
Re: [PATCH v3 00/28] Add VDUSE support to Vhost library
On Thu, May 25, 2023 at 6:25 PM Maxime Coquelin wrote: > > Note: v2 is identical to v3, it is just a resend because > of an issue when posting v2 breaking the series in patchwork. > > This series introduces a new type of backend, VDUSE, > to the Vhost library. > > VDUSE stands for vDPA device in Userspace, it enables > implementing a Virtio device in userspace and have it > attached to the Kernel vDPA bus. > > Once attached to the vDPA bus, it can be used either by > Kernel Virtio drivers, like virtio-net in our case, via > the virtio-vdpa driver. Doing that, the device is visible > to the Kernel networking stack and is exposed to userspace > as a regular netdev. > > It can also be exposed to userspace thanks to the > vhost-vdpa driver, via a vhost-vdpa chardev that can be > passed to QEMU or Virtio-user PMD. > > While VDUSE support is already available in upstream > Kernel, a couple of patches are required to support > network device type: > > https://gitlab.com/mcoquelin/linux/-/tree/vduse_networking_rfc > > In order to attach the created VDUSE device to the vDPA > bus, a recent iproute2 version containing the vdpa tool is > required. > > Benchmark results: > == > > On this v2, PVP reference benchmark has been run & compared with > Vhost-user. > > When doing macswap forwarding in the worload, no difference is seen. > When doing io forwarding in the workload, we see 4% performance > degradation with VDUSE, comapred to Vhost-user/Virtio-user. It is > explained by the use of the IOTLB layer in the Vhost-library when using > VDUSE, whereas Vhost-user/Virtio-user does not make use of it. > > Usage: > == > > 1. Probe required Kernel modules > # modprobe vdpa > # modprobe vduse > # modprobe virtio-vdpa > > 2. Build (require vduse kernel headers to be available) > # meson build > # ninja -C build > > 3. Create a VDUSE device (vduse0) using Vhost PMD with > testpmd (with 4 queue pairs in this example) > # ./build/app/dpdk-testpmd --no-pci > --vdev=net_vhost0,iface=/dev/vduse/vduse0,queues=4 --log-level=*:9 -- -i > --txq=4 --rxq=4 > > 4. Attach the VDUSE device to the vDPA bus > # vdpa dev add name vduse0 mgmtdev vduse > => The virtio-net netdev shows up (eth0 here) > # ip l show eth0 > 21: eth0: mtu 1500 qdisc mq state UP mode > DEFAULT group default qlen 1000 > link/ether c2:73:ea:a7:68:6d brd ff:ff:ff:ff:ff:ff > > 5. Start/stop traffic in testpmd > testpmd> start > testpmd> show port stats 0 > NIC statistics for port 0 > RX-packets: 11 RX-missed: 0 RX-bytes: 1482 > RX-errors: 0 > RX-nombuf: 0 > TX-packets: 1 TX-errors: 0 TX-bytes: 62 > > Throughput (since last show) > Rx-pps:0 Rx-bps:0 > Tx-pps:0 Tx-bps:0 > > testpmd> stop > > 6. Detach the VDUSE device from the vDPA bus > # vdpa dev del vduse0 > > 7. Quit testpmd > testpmd> quit > > Known issues & remaining work: > == > - Fix issue in FD manager (still polling while FD has been removed) > - Add Netlink support in Vhost library > - Support device reconnection > -> a temporary patch to support reconnection via a tmpfs file is available, > upstream solution would be in-kernel and is being developed. > -> > https://gitlab.com/mcoquelin/dpdk-next-virtio/-/commit/5ad06ce14159a9ce36ee168dd13ef389cec91137 > - Support packed ring > - Provide more performance benchmark results > > Changes in v2/v3: > = > - Fixed mem_set_dump() parameter (patch 4) > - Fixed accidental comment change (patch 7, Chenbo) > - Change from __builtin_ctz to __builtin_ctzll (patch 9, Chenbo) > - move change from patch 12 to 13 (Chenbo) > - Enable locks annotation for control queue (Patch 17) > - Send control queue notification when used descriptors enqueued (Patch 17) > - Lock control queue IOTLB lock (Patch 17) > - Fix error path in virtio_net_ctrl_pop() (Patch 17, Chenbo) > - Set VDUSE dev FD as NONBLOCK (Patch 18) > - Enable more Virtio features (Patch 18) > - Remove calls to pthread_setcancelstate() (Patch 22) > - Add calls to fdset_pipe_notify() when adding and deleting FDs from a set > (Patch 22) > - Use RTE_DIM() to get requests string array size (Patch 22) > - Set reply result for IOTLB update message (Patch 25, Chenbo) > - Fix queues enablement with multiqueue (Patch 26) > - Move kickfd creation for better logging (Patch 26) > - Improve logging (Patch 26) > - Uninstall cvq kickfd in case of handler installation failure (Patch 27) > - Enable CVQ notifications once handler is installed (Patch 27) > - Don't advertise multiqueue and control queue if app only request single > queue pair (Patch 27) > - Add release notes > > Maxime Coquelin (28): > vhost: fix missing guest notif stat increment > vhost: fix invalid call FD handling > vhost: fix IOTLB entries over
Re: [PATCH 01/15] common/cnxk: add ROC MACsec initialization
On Wed, May 24, 2023 at 1:34 AM Akhil Goyal wrote: > > Added ROC init and fini APIs for supporting MACsec. > > Signed-off-by: Ankur Dwivedi > Signed-off-by: Vamsi Attunuru > Signed-off-by: Akhil Goyal > --- > + > +TAILQ_HEAD(roc_mcs_head, roc_mcs); > +/* Local mcs tailq list */ > +static struct roc_mcs_head roc_mcs_head = > TAILQ_HEAD_INITIALIZER(roc_mcs_head); Can we remove this global variable by moving? This is to avoid breaking multiprocess, > +struct roc_mcs * > +roc_mcs_dev_init(uint8_t mcs_idx) > +{ > + struct roc_mcs *mcs; > + struct npa_lf *npa; > + > + if (roc_model_is_cn10kb()) { Use roc_feature_nix_has_macsec() > + mcs = roc_idev_mcs_get(); > + if (mcs) { > + plt_info("Skipping device, mcs device already > probed"); > + mcs->refcount++; > + return mcs; > + } > + } > + > + mcs = plt_zmalloc(sizeof(struct roc_mcs), PLT_CACHE_LINE_SIZE); > + if (!mcs) > + return NULL; > + > + if (roc_model_is_cnf10kb() || roc_model_is_cn10kb()) { Use roc_feature_nix_has_macsec() > + MCS_ERR_PARAM = -900, > + MCS_ERR_HW_NOTSUP = -901, > + MCS_ERR_DEVICE_NOT_FOUND = -902, > +}; > + > +#define MCS_SUPPORT_CHECK > \ > + do { > \ > + if (!(roc_model_is_cnf10kb() || roc_model_is_cn10kb_a0())) > \ Use roc_feature_nix_has_macsec()
[PATCH v4 0/3] Enable iavf Rx Timestamp offload on vector path
Enable timestamp offload with the command '--enable-rx-timestamp', pay attention that getting Rx timestamp offload will drop the performance. --- v4: rework avx2 patch based on offload path --- v3: logging with driver dedicated macro --- v2: fix compile warning and SSE path Zhichao Zeng (3): net/iavf: support Rx timestamp offload on AVX512 net/iavf: support Rx timestamp offload on AVX2 net/iavf: support Rx timestamp offload on SSE drivers/net/iavf/iavf_rxtx.h| 3 +- drivers/net/iavf/iavf_rxtx_vec_avx2.c | 186 +- drivers/net/iavf/iavf_rxtx_vec_avx512.c | 203 +++- drivers/net/iavf/iavf_rxtx_vec_common.h | 3 - drivers/net/iavf/iavf_rxtx_vec_sse.c| 159 ++- 5 files changed, 538 insertions(+), 16 deletions(-) -- 2.34.1
[PATCH v4 1/3] net/iavf: support Rx timestamp offload on AVX512
This patch enables Rx timestamp offload on AVX512 data path. Enable timestamp offload with the command '--enable-rx-timestamp', pay attention that getting Rx timestamp offload will drop the performance. Signed-off-by: Wenjun Wu Signed-off-by: Zhichao Zeng --- v4: rework avx2 patch based on offload path --- v3: logging with driver dedicated macro --- v2: fix compile warning --- drivers/net/iavf/iavf_rxtx.h| 3 +- drivers/net/iavf/iavf_rxtx_vec_avx512.c | 203 +++- drivers/net/iavf/iavf_rxtx_vec_common.h | 3 - 3 files changed, 200 insertions(+), 9 deletions(-) diff --git a/drivers/net/iavf/iavf_rxtx.h b/drivers/net/iavf/iavf_rxtx.h index 547b68f441..0345a6a51d 100644 --- a/drivers/net/iavf/iavf_rxtx.h +++ b/drivers/net/iavf/iavf_rxtx.h @@ -47,7 +47,8 @@ RTE_ETH_RX_OFFLOAD_CHECKSUM |\ RTE_ETH_RX_OFFLOAD_SCTP_CKSUM | \ RTE_ETH_RX_OFFLOAD_VLAN |\ - RTE_ETH_RX_OFFLOAD_RSS_HASH) + RTE_ETH_RX_OFFLOAD_RSS_HASH |\ + RTE_ETH_RX_OFFLOAD_TIMESTAMP) /** * According to the vlan capabilities returned by the driver and FW, the vlan tci diff --git a/drivers/net/iavf/iavf_rxtx_vec_avx512.c b/drivers/net/iavf/iavf_rxtx_vec_avx512.c index 4fe9b97278..f9961e53b8 100644 --- a/drivers/net/iavf/iavf_rxtx_vec_avx512.c +++ b/drivers/net/iavf/iavf_rxtx_vec_avx512.c @@ -16,18 +16,20 @@ /** * If user knows a specific offload is not enabled by APP, * the macro can be commented to save the effort of fast path. - * Currently below 2 features are supported in RX path, + * Currently below 6 features are supported in RX path, * 1, checksum offload * 2, VLAN/QINQ stripping * 3, RSS hash * 4, packet type analysis * 5, flow director ID report + * 6, timestamp offload **/ #define IAVF_RX_CSUM_OFFLOAD #define IAVF_RX_VLAN_OFFLOAD #define IAVF_RX_RSS_OFFLOAD #define IAVF_RX_PTYPE_OFFLOAD #define IAVF_RX_FDIR_OFFLOAD +#define IAVF_RX_TS_OFFLOAD static __rte_always_inline void iavf_rxq_rearm(struct iavf_rx_queue *rxq) @@ -587,9 +589,9 @@ _iavf_recv_raw_pkts_vec_avx512_flex_rxd(struct iavf_rx_queue *rxq, bool offload) { struct iavf_adapter *adapter = rxq->vsi->adapter; - +#ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC uint64_t offloads = adapter->dev_data->dev_conf.rxmode.offloads; - +#endif #ifdef IAVF_RX_PTYPE_OFFLOAD const uint32_t *type_table = adapter->ptype_tbl; #endif @@ -618,6 +620,25 @@ _iavf_recv_raw_pkts_vec_avx512_flex_rxd(struct iavf_rx_queue *rxq, rte_cpu_to_le_32(1 << IAVF_RX_FLEX_DESC_STATUS0_DD_S))) return 0; +#ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC +#ifdef IAVF_RX_TS_OFFLOAD + uint8_t inflection_point = 0; + bool is_tsinit = false; + __m256i hw_low_last = _mm256_set_epi32(0, 0, 0, 0, 0, 0, 0, (uint32_t)rxq->phc_time); + + if (rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) { + uint64_t sw_cur_time = rte_get_timer_cycles() / (rte_get_timer_hz() / 1000); + + if (unlikely(sw_cur_time - rxq->hw_time_update > 4)) { + hw_low_last = _mm256_setzero_si256(); + is_tsinit = 1; + } else { + hw_low_last = _mm256_set_epi32(0, 0, 0, 0, 0, 0, 0, (uint32_t)rxq->phc_time); + } + } +#endif +#endif + /* constants used in processing loop */ const __m512i crc_adjust = _mm512_set_epi32 @@ -1081,12 +1102,13 @@ _iavf_recv_raw_pkts_vec_avx512_flex_rxd(struct iavf_rx_queue *rxq, #ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC if (offload) { -#ifdef IAVF_RX_RSS_OFFLOAD +#if defined(IAVF_RX_RSS_OFFLOAD) || defined(IAVF_RX_TS_OFFLOAD) /** * needs to load 2nd 16B of each desc for RSS hash parsing, * will cause performance drop to get into this context. */ if (offloads & RTE_ETH_RX_OFFLOAD_RSS_HASH || + offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP || rxq->rx_flags & IAVF_RX_FLAGS_VLAN_TAG_LOC_L2TAG2_2) { /* load bottom half of every 32B desc */ const __m128i raw_desc_bh7 = @@ -1138,6 +1160,7 @@ _iavf_recv_raw_pkts_vec_avx512_flex_rxd(struct iavf_rx_queue *rxq, (_mm256_castsi128_si256(raw_desc_bh0), raw_desc_bh1, 1); +#ifdef IAVF_RX_RSS_OFFLOAD if (offloads & RTE_ETH_RX_OFFLOAD_RSS_HASH) { /**
[PATCH v4 2/3] net/iavf: support Rx timestamp offload on AVX2
This patch enables Rx timestamp offload on AVX2 data path. Enable timestamp offload with the command '--enable-rx-timestamp', pay attention that getting Rx timestamp offload will drop the performance. Signed-off-by: Zhichao Zeng --- v4: rework avx2 patch based on offload path --- v3: logging with driver dedicated macro --- v2: fix compile warning --- drivers/net/iavf/iavf_rxtx_vec_avx2.c | 186 +- 1 file changed, 182 insertions(+), 4 deletions(-) diff --git a/drivers/net/iavf/iavf_rxtx_vec_avx2.c b/drivers/net/iavf/iavf_rxtx_vec_avx2.c index 22d4d3a90f..86290c4bbb 100644 --- a/drivers/net/iavf/iavf_rxtx_vec_avx2.c +++ b/drivers/net/iavf/iavf_rxtx_vec_avx2.c @@ -532,7 +532,9 @@ _iavf_recv_raw_pkts_vec_avx2_flex_rxd(struct iavf_rx_queue *rxq, struct iavf_adapter *adapter = rxq->vsi->adapter; +#ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC uint64_t offloads = adapter->dev_data->dev_conf.rxmode.offloads; +#endif const uint32_t *type_table = adapter->ptype_tbl; const __m256i mbuf_init = _mm256_set_epi64x(0, 0, @@ -558,6 +560,21 @@ _iavf_recv_raw_pkts_vec_avx2_flex_rxd(struct iavf_rx_queue *rxq, if (!(rxdp->wb.status_error0 & rte_cpu_to_le_32(1 << IAVF_RX_FLEX_DESC_STATUS0_DD_S))) return 0; +#ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC + bool is_tsinit = false; + uint8_t inflection_point = 0; + __m256i hw_low_last = _mm256_set_epi32(0, 0, 0, 0, 0, 0, 0, rxq->phc_time); + if (rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) { + uint64_t sw_cur_time = rte_get_timer_cycles() / (rte_get_timer_hz() / 1000); + + if (unlikely(sw_cur_time - rxq->hw_time_update > 4)) { + hw_low_last = _mm256_setzero_si256(); + is_tsinit = 1; + } else { + hw_low_last = _mm256_set_epi32(0, 0, 0, 0, 0, 0, 0, rxq->phc_time); + } + } +#endif /* constants used in processing loop */ const __m256i crc_adjust = @@ -967,10 +984,11 @@ _iavf_recv_raw_pkts_vec_avx2_flex_rxd(struct iavf_rx_queue *rxq, if (offload) { #ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC /** -* needs to load 2nd 16B of each desc for RSS hash parsing, +* needs to load 2nd 16B of each desc, * will cause performance drop to get into this context. */ if (offloads & RTE_ETH_RX_OFFLOAD_RSS_HASH || + offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP || rxq->rx_flags & IAVF_RX_FLAGS_VLAN_TAG_LOC_L2TAG2_2) { /* load bottom half of every 32B desc */ const __m128i raw_desc_bh7 = @@ -1053,7 +1071,7 @@ _iavf_recv_raw_pkts_vec_avx2_flex_rxd(struct iavf_rx_queue *rxq, mb4_5 = _mm256_or_si256(mb4_5, rss_hash4_5); mb2_3 = _mm256_or_si256(mb2_3, rss_hash2_3); mb0_1 = _mm256_or_si256(mb0_1, rss_hash0_1); - } + } /* if() on RSS hash parsing */ if (rxq->rx_flags & IAVF_RX_FLAGS_VLAN_TAG_LOC_L2TAG2_2) { /* merge the status/error-1 bits into one register */ @@ -1132,8 +1150,121 @@ _iavf_recv_raw_pkts_vec_avx2_flex_rxd(struct iavf_rx_queue *rxq, mb4_5 = _mm256_or_si256(mb4_5, vlan_tci4_5); mb2_3 = _mm256_or_si256(mb2_3, vlan_tci2_3); mb0_1 = _mm256_or_si256(mb0_1, vlan_tci0_1); - } - } /* if() on RSS hash parsing */ + } /* if() on Vlan parsing */ + + if (offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) { + uint32_t mask = 0x; + __m256i ts; + __m256i ts_low = _mm256_setzero_si256(); + __m256i ts_low1; + __m256i ts_low2; + __m256i max_ret; + __m256i cmp_ret; + uint8_t ret = 0; + uint8_t shift = 8; + __m256i ts_desp_mask = _mm256_set_epi32(mask, 0, 0, 0, mask, 0, 0, 0); + __m256i cmp_mask = _mm256_set1_epi32(mask); + __m256i ts_permute_mask = _mm256_set_epi32(7, 3, 6, 2, 5, 1, 4, 0); + + ts = _m
[PATCH v4 3/3] net/iavf: support Rx timestamp offload on SSE
This patch enables Rx timestamp offload on SSE data path. Enable timestamp offload with the command '--enable-rx-timestamp', pay attention that getting Rx timestamp offload will drop the performance. Signed-off-by: Zhichao Zeng --- v4: rework avx2 patch based on offload path --- v3: logging with driver dedicated macro --- v2: fix compile warning and timestamp error --- drivers/net/iavf/iavf_rxtx_vec_sse.c | 159 ++- 1 file changed, 156 insertions(+), 3 deletions(-) diff --git a/drivers/net/iavf/iavf_rxtx_vec_sse.c b/drivers/net/iavf/iavf_rxtx_vec_sse.c index 3f30be01aa..b754122c51 100644 --- a/drivers/net/iavf/iavf_rxtx_vec_sse.c +++ b/drivers/net/iavf/iavf_rxtx_vec_sse.c @@ -392,6 +392,11 @@ flex_desc_to_olflags_v(struct iavf_rx_queue *rxq, __m128i descs[4], _mm_extract_epi32(fdir_id0_3, 3); } /* if() on fdir_enabled */ +#ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC + if (rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) + flags = _mm_or_si128(flags, _mm_set1_epi32(iavf_timestamp_dynflag)); +#endif + /** * At this point, we have the 4 sets of flags in the low 16-bits * of each 32-bit value in flags. @@ -723,7 +728,9 @@ _recv_raw_pkts_vec_flex_rxd(struct iavf_rx_queue *rxq, int pos; uint64_t var; struct iavf_adapter *adapter = rxq->vsi->adapter; +#ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC uint64_t offloads = adapter->dev_data->dev_conf.rxmode.offloads; +#endif const uint32_t *ptype_tbl = adapter->ptype_tbl; __m128i crc_adjust = _mm_set_epi16 (0, 0, 0, /* ignore non-length fields */ @@ -793,6 +800,24 @@ _recv_raw_pkts_vec_flex_rxd(struct iavf_rx_queue *rxq, rte_cpu_to_le_32(1 << IAVF_RX_FLEX_DESC_STATUS0_DD_S))) return 0; +#ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC + uint8_t inflection_point = 0; + bool is_tsinit = false; + __m128i hw_low_last = _mm_set_epi32(0, 0, 0, (uint32_t)rxq->phc_time); + + if (rxq->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) { + uint64_t sw_cur_time = rte_get_timer_cycles() / (rte_get_timer_hz() / 1000); + + if (unlikely(sw_cur_time - rxq->hw_time_update > 4)) { + hw_low_last = _mm_setzero_si128(); + is_tsinit = 1; + } else { + hw_low_last = _mm_set_epi32(0, 0, 0, (uint32_t)rxq->phc_time); + } + } + +#endif + /** * Compile-time verify the shuffle mask * NOTE: some field positions already verified above, but duplicated @@ -825,7 +850,7 @@ _recv_raw_pkts_vec_flex_rxd(struct iavf_rx_queue *rxq, rxdp += IAVF_VPMD_DESCS_PER_LOOP) { __m128i descs[IAVF_VPMD_DESCS_PER_LOOP]; #ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC - __m128i descs_bh[IAVF_VPMD_DESCS_PER_LOOP]; + __m128i descs_bh[IAVF_VPMD_DESCS_PER_LOOP] = {_mm_setzero_si128()}; #endif __m128i pkt_mb0, pkt_mb1, pkt_mb2, pkt_mb3; __m128i staterr, sterr_tmp1, sterr_tmp2; @@ -895,10 +920,11 @@ _recv_raw_pkts_vec_flex_rxd(struct iavf_rx_queue *rxq, #ifndef RTE_LIBRTE_IAVF_16BYTE_RX_DESC /** -* needs to load 2nd 16B of each desc for RSS hash parsing, +* needs to load 2nd 16B of each desc, * will cause performance drop to get into this context. */ if (offloads & RTE_ETH_RX_OFFLOAD_RSS_HASH || + offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP || rxq->rx_flags & IAVF_RX_FLAGS_VLAN_TAG_LOC_L2TAG2_2) { /* load bottom half of every 32B desc */ descs_bh[3] = _mm_load_si128 @@ -964,7 +990,94 @@ _recv_raw_pkts_vec_flex_rxd(struct iavf_rx_queue *rxq, pkt_mb2 = _mm_or_si128(pkt_mb2, vlan_tci2); pkt_mb1 = _mm_or_si128(pkt_mb1, vlan_tci1); pkt_mb0 = _mm_or_si128(pkt_mb0, vlan_tci0); - } + } /* if() on Vlan parsing */ + + if (offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP) { + uint32_t mask = 0x; + __m128i ts; + __m128i ts_low = _mm_setzero_si128(); + __m128i ts_low1; + __m128i max_ret; + __m128i cmp_ret; + uint8_t ret = 0; + uint8_t shift = 4; + __m128i ts_desp_mask = _mm_set_epi32(mask, 0, 0, 0); + __m128i cmp_mask = _mm_set1_epi32(mask); + + ts = _mm_and_si128(descs_bh[0], ts_desp_mask); + ts_low = _mm_or_si128(ts_low, _mm_srli_si128(ts, 3 * 4)); + ts = _mm_and_si128(descs_bh[1], ts_desp_mask); +
Re: [PATCH v4] lib: set/get max memzone segments
On Thu, May 25, 2023 at 12:26 AM Ophir Munk wrote: > > Currently, the max memzones count constat (RTE_MAX_MEMZONE) is used to > decide how many memzones a DPDK application can have. This value could > technically be changed by manually editing `rte_config.h` before > compilation, but if DPDK is already compiled, that option is not useful. > There are certain use cases that would benefit from making this value > configurable. > > This commit addresses the issue by adding a new API to set the max > number of memzones before EAL initialization (while using the old > constant as default value), as well as an API to get current maximum > number of memzones. > > Signed-off-by: Ophir Munk > Acked-by: Morten Brørup qede maintainers, can you double check the change on your driver please? Thanks. -- David Marchand
RE: [PATCH v6 04/15] graph: add get/set graph worker model APIs
> -Original Message- > From: Jerin Jacob > Sent: Wednesday, May 24, 2023 2:09 PM > To: Yan, Zhirun > Cc: dev@dpdk.org; jer...@marvell.com; kirankum...@marvell.com; > ndabilpu...@marvell.com; step...@networkplumber.org; > pbhagavat...@marvell.com; Liang, Cunming ; Wang, > Haiyue > Subject: Re: [PATCH v6 04/15] graph: add get/set graph worker model APIs > > On Tue, May 9, 2023 at 11:34 AM Zhirun Yan wrote: > > > > Add new get/set APIs to configure graph worker model which is used to > > determine which model will be chosen. > > > > Signed-off-by: Haiyue Wang > > Signed-off-by: Cunming Liang > > Signed-off-by: Zhirun Yan > > --- > > diff --git a/lib/graph/rte_graph_worker.c > > b/lib/graph/rte_graph_worker.c new file mode 100644 index > > 00..cabc101262 > > --- /dev/null > > +++ b/lib/graph/rte_graph_worker.c > > @@ -0,0 +1,54 @@ > > +/* SPDX-License-Identifier: BSD-3-Clause > > + * Copyright(C) 2023 Intel Corporation */ > > + > > +#include "rte_graph_worker_common.h" > > + > > +RTE_DEFINE_PER_LCORE(enum rte_graph_worker_model, worker_model) = > > +RTE_GRAPH_MODEL_DEFAULT; > > + > > +/** > > + * @warning > > + * @b EXPERIMENTAL: this API may change, or be removed, without prior > > +notice > > + * Set the graph worker model > > Just declaring this top of the header file enough to avoid duplicating in > every > functions as all functions in header is experimental. See > lib/graph/rte_graph.h > Got it, I will do in next version. > > > + * > > + * @note This function does not perform any locking, and is only safe to > > call > > + *before graph running. > > + * > > + * @param name > > + * Name of the graph worker model. > > + * > > + * @return > > + * 0 on success, -1 otherwise. > > + */ > > +int > > +rte_graph_worker_model_set(enum rte_graph_worker_model model) { > > + if (model >= RTE_GRAPH_MODEL_LIST_END) > > + goto fail; > > + > > + RTE_PER_LCORE(worker_model) = model; > > Application needs to set this per core . Right? Yes. Each worker needs to know its model. > Are we anticipating a case where one core runs one model and another core > runs with another model? > If not OR it is not practically possible, then, To make application > programmer > life easy, We could loop through all lore and set on all of them instead of > application setting on each one separately. > For current rtc and dispatch models, it is not necessary. To some extent that models are mutually exclusive. For this case: Core 1: A->B->C (RTC) Core 2: A' (DISPATCH) Core 3: B' (DISPATCH) Core 4: C' (DISPATCH) It may change the graph topo, or need some prerequisites like RSS before input node A. BTW, if there are some requirements with more models in future, we could add some attributions for graph, lcore, node. Like taint/affinity for node and model. Then we could allow a node to appeal/repel a set of models. I will change to put the model into struct rte_graph as you suggested in patch 12 for this release. > > > + return 0; > > + > > +fail: > > + RTE_PER_LCORE(worker_model) = RTE_GRAPH_MODEL_DEFAULT; > > + return -1; > > +} > > + > > > +/** Graph worker models */ > > +enum rte_graph_worker_model { > > + RTE_GRAPH_MODEL_DEFAULT, > > Add Doxygen comment > > + RTE_GRAPH_MODEL_RTC = RTE_GRAPH_MODEL_DEFAULT, > > > Add Doxygen comment to explain what this mode does. > > > > + RTE_GRAPH_MODEL_MCORE_DISPATCH, > > Add Doxygen comment to explain what this mode does. > Ok, I will add Doxygen comments for these models. > > + RTE_GRAPH_MODEL_LIST_END > > This can break the ABI if we add one in middle. Please remove this. > See lib/crytodev for > how to handle with _END symbols. Yes, I will remove this.
RE: [PATCH v6 05/15] graph: introduce graph node core affinity API
> -Original Message- > From: Jerin Jacob > Sent: Wednesday, May 24, 2023 2:36 PM > To: Yan, Zhirun > Cc: dev@dpdk.org; jer...@marvell.com; kirankum...@marvell.com; > ndabilpu...@marvell.com; step...@networkplumber.org; > pbhagavat...@marvell.com; Liang, Cunming ; Wang, > Haiyue > Subject: Re: [PATCH v6 05/15] graph: introduce graph node core affinity API > > On Tue, May 9, 2023 at 11:34 AM Zhirun Yan wrote: > > > > Add lcore_id for node to hold affinity core id and impl > > rte_graph_model_dispatch_lcore_affinity_set to set node affinity with > > specific lcore. > > > > Signed-off-by: Haiyue Wang > > Signed-off-by: Cunming Liang > > Signed-off-by: Zhirun Yan > > --- > > lib/graph/graph_private.h| 1 + > > lib/graph/meson.build| 1 + > > lib/graph/node.c | 1 + > > lib/graph/rte_graph_model_dispatch.c | 30 +++ > > lib/graph/rte_graph_model_dispatch.h | 43 > > Could you change all the new model file to prefix _mcore_ like below to align > with enum name. > rte_graph_model_mcore_dispatch.* Yes, I will do in next version. > > > lib/graph/version.map| 2 ++ > > 6 files changed, 78 insertions(+) > > create mode 100644 lib/graph/rte_graph_model_dispatch.c > > create mode 100644 lib/graph/rte_graph_model_dispatch.h > > > > diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h > > index eacdef45f0..bd4c576324 100644 > > --- a/lib/graph/graph_private.h > > +++ b/lib/graph/graph_private.h > > @@ -51,6 +51,7 @@ struct node { > > STAILQ_ENTRY(node) next; /**< Next node in the list. */ > > char name[RTE_NODE_NAMESIZE]; /**< Name of the node. */ > > uint64_t flags; /**< Node configuration flag. */ > > + unsigned int lcore_id;/**< Node runs on the Lcore ID */ > > Could you move to end of existing fast path variables. Also, please extend the > comments for new variable introduced ONLY for dispatch model. Comment > should express it is only for dispatch model. > Got it, will do in next version. > > > rte_node_process_t process; /**< Node process function. */ > > rte_node_init_t init; /**< Node init function. */ > > rte_node_fini_t fini; /**< Node fini function. */ > > diff --git a/lib/graph/meson.build b/lib/graph/meson.build index > > 9fab8243da..c729d984b6 100644 > > --- a/lib/graph/meson.build > > +++ b/lib/graph/meson.build > > @@ -16,6 +16,7 @@ sources = files( > > 'graph_populate.c', > > 'graph_pcap.c', > > 'rte_graph_worker.c', > > +'rte_graph_model_dispatch.c', > > ) > > headers = files('rte_graph.h', 'rte_graph_worker.h') > > > > diff --git a/lib/graph/node.c b/lib/graph/node.c index > > 149414dcd9..339b4a0da5 100644 > > --- a/lib/graph/node.c > > +++ b/lib/graph/node.c > > @@ -100,6 +100,7 @@ __rte_node_register(const struct rte_node_register > *reg) > > goto free; > > } > > > > + node->lcore_id = RTE_MAX_LCORE; > > node->id = node_id++; > > > > /* Add the node at tail */ > > diff --git a/lib/graph/rte_graph_model_dispatch.c > > b/lib/graph/rte_graph_model_dispatch.c > > new file mode 100644 > > index 00..3364a76ed4 > > --- /dev/null > > +++ b/lib/graph/rte_graph_model_dispatch.c > > @@ -0,0 +1,30 @@ > > +/* SPDX-License-Identifier: BSD-3-Clause > > + * Copyright(C) 2023 Intel Corporation */ > > + > > +#include "graph_private.h" > > +#include "rte_graph_model_dispatch.h" > > + > > +int > > +rte_graph_model_dispatch_lcore_affinity_set(const char *name, > > +unsigned int lcore_id) { > > + struct node *node; > > + int ret = -EINVAL; > > > Shouldn't we need to check the is model is dispatch before proceeding? > If so, same comment applicable for all function created solely for new model. > > Also, in case some _exiting_ API not valid for dispatch model, please add the > check for the same. Yes, I will add check for all function if it is only used by new model. Actually, most existing API are valid, I will double check. > > > + > > + if (lcore_id >= RTE_MAX_LCORE) > > + return ret; > > + > > + graph_spinlock_lock(); > > + > > + STAILQ_FOREACH(node, node_list_head_get(), next) { > > + if (strncmp(node->name, name, RTE_NODE_NAMESIZE) == 0) { > > + node->lcore_id = lcore_id; > > + ret = 0; > > + break; > > + } > > + } > > + > > + graph_spinlock_unlock(); > > + > > + return ret; > > +} > > diff --git a/lib/graph/rte_graph_model_dispatch.h > > b/lib/graph/rte_graph_model_dispatch.h > > new file mode 100644 > > index 00..179624e972 > > --- /dev/null > > +++ b/lib/graph/rte_graph_model_dispatch.h > > @@ -0,0 +1,43 @@ > > +/* SPDX-License-Identifier: BSD-3-Clause > > + * Copyright(C) 2023 Intel Corporation */ >
RE: [PATCH v6 06/15] graph: introduce graph bind unbind API
> -Original Message- > From: Jerin Jacob > Sent: Wednesday, May 24, 2023 2:24 PM > To: Yan, Zhirun > Cc: dev@dpdk.org; jer...@marvell.com; kirankum...@marvell.com; > ndabilpu...@marvell.com; step...@networkplumber.org; > pbhagavat...@marvell.com; Liang, Cunming ; Wang, > Haiyue > Subject: Re: [PATCH v6 06/15] graph: introduce graph bind unbind API > > On Tue, May 9, 2023 at 11:34 AM Zhirun Yan wrote: > > > > Add lcore_id for graph to hold affinity core id where graph would run on. > > Add bind/unbind API to set/unset graph affinity attribute. lcore_id > > will be set as MAX by default, it means not enable this attribute. > > > > Signed-off-by: Haiyue Wang > > Signed-off-by: Cunming Liang > > Signed-off-by: Zhirun Yan > > --- > > lib/graph/graph.c | 59 +++ > > lib/graph/graph_private.h | 2 ++ > > lib/graph/rte_graph.h | 22 +++ > > lib/graph/version.map | 2 ++ > > + > > +int > > +rte_graph_model_dispatch_core_bind(rte_graph_t id, int lcore) { > > + struct graph *graph; > > + > > + GRAPH_ID_CHECK(id); > > + if (!rte_lcore_is_enabled(lcore)) > > + SET_ERR_JMP(ENOLINK, fail, > > + "lcore %d not enabled\n", > > \n is already part of it from graph_err() so no need to add it. > Also, DPDK coding standard now supports 100 column as max width, so below > "lcore" can be moved to the same line if there is space. > Ok, I will merge into one line and remove \n. > > > + lcore); > > + > > + STAILQ_FOREACH(graph, &graph_list, next) > > + if (graph->id == id) > > + break; > > + > > + graph->lcore_id = lcore; > > + graph->socket = rte_lcore_to_socket_id(lcore); > > + > > + /* check the availability of source node */ > > + if (!graph_src_node_avail(graph)) > > + graph->graph->head = 0; > > + > > + return 0; > > + > > +fail: > > + return -rte_errno; > > +} > > + > > +void > > +rte_graph_model_dispatch_core_unbind(rte_graph_t id) { > > + struct graph *graph; > > + > > + GRAPH_ID_CHECK(id); > > + STAILQ_FOREACH(graph, &graph_list, next) > > + if (graph->id == id) > > + break; > > + > > + graph->lcore_id = RTE_MAX_LCORE; > > + > > +fail: > > + return; > > +} > > + > > struct rte_graph * > > rte_graph_lookup(const char *name) > > { > > @@ -346,6 +404,7 @@ rte_graph_create(const char *name, struct > rte_graph_param *prm) > > graph->src_node_count = src_node_count; > > graph->node_count = graph_nodes_count(graph); > > graph->id = graph_id; > > + graph->lcore_id = RTE_MAX_LCORE; > > graph->num_pkt_to_capture = prm->num_pkt_to_capture; > > if (prm->pcap_filename) > > rte_strscpy(graph->pcap_filename, prm->pcap_filename, > > RTE_GRAPH_PCAP_FILE_SZ); diff --git a/lib/graph/graph_private.h > > b/lib/graph/graph_private.h index bd4c576324..f63b339d81 100644 > > --- a/lib/graph/graph_private.h > > +++ b/lib/graph/graph_private.h > > @@ -99,6 +99,8 @@ struct graph { > > /**< Circular buffer mask for wrap around. */ > > rte_graph_t id; > > /**< Graph identifier. */ > > + unsigned int lcore_id; > > + /**< Lcore identifier where the graph prefer to run on. */ > > Could you move to end of existing fast path variables. Also, please extend the > comments for new variable introduced ONLY for dispatch model. > Something like " Lcore identifier where the graph prefer to run on." to " > Lcore > identifier where the graph prefer to run on. Used with dispatch model" or so. Yes, I will do in next version. > > > > size_t mem_sz; > > /**< Memory size of the graph. */ > > int socket; > > diff --git a/lib/graph/rte_graph.h b/lib/graph/rte_graph.h index > > c9a77297fc..c523809d1f 100644 > > --- a/lib/graph/rte_graph.h > > +++ b/lib/graph/rte_graph.h > > @@ -285,6 +285,28 @@ char *rte_graph_id_to_name(rte_graph_t id); > > __rte_experimental int rte_graph_export(const char *name, FILE *f); > > > > +/** > > + * Bind graph with specific lcore > > + * > > + * @param id > > + * Graph id to get the pointer of graph object > > + * @param lcore > > + * The lcore where the graph will run on > > + * @return > > + * 0 on success, error otherwise. > > + */ > > +__rte_experimental > > +int rte_graph_model_dispatch_core_bind(rte_graph_t id, int lcore); > > + > > +/** > > + * Unbind graph with lcore > > + * > > + * @param id > > + * Graph id to get the pointer of graph object */ __rte_experimental > > +void rte_graph_model_dispatch_core_unbind(rte_graph_t id); > > + > > /** > > * Get graph object from its name. > > * > > diff --git a/lib/graph/version.map b/lib/graph/version.map index > > 1f090be74e..7de6f08f59 100644 > > --- a/lib/graph/version.map > > +++ b/lib/graph/version.map > > @@ -18,6 +18,8 @@ EX
RE: [PATCH v6 08/15] graph: add struct for stream moving between cores
> -Original Message- > From: Jerin Jacob > Sent: Wednesday, May 24, 2023 3:25 PM > To: Yan, Zhirun > Cc: dev@dpdk.org; jer...@marvell.com; kirankum...@marvell.com; > ndabilpu...@marvell.com; step...@networkplumber.org; > pbhagavat...@marvell.com; Liang, Cunming ; Wang, > Haiyue > Subject: Re: [PATCH v6 08/15] graph: add struct for stream moving between > cores > > On Tue, May 9, 2023 at 11:35 AM Zhirun Yan wrote: > > > > Add graph_sched_wq_node to hold graph scheduling workqueue node. > > > > Signed-off-by: Haiyue Wang > > Signed-off-by: Cunming Liang > > Signed-off-by: Zhirun Yan > > + * @internal > > + * > > + * Structure that holds the graph scheduling workqueue node stream. > > + * Used for mcore dispatch model. > > + */ > > +struct graph_sched_wq_node { > > Could we change to graph_mcore_dispatch_wq_node? Just to make sure this > for mcore dispatch model. Yes.
RE: [PATCH v6 07/15] graph: introduce graph clone API for other worker core
> -Original Message- > From: Jerin Jacob > Sent: Wednesday, May 24, 2023 3:14 PM > To: Yan, Zhirun > Cc: dev@dpdk.org; jer...@marvell.com; kirankum...@marvell.com; > ndabilpu...@marvell.com; step...@networkplumber.org; > pbhagavat...@marvell.com; Liang, Cunming ; Wang, > Haiyue > Subject: Re: [PATCH v6 07/15] graph: introduce graph clone API for other > worker > core > > On Tue, May 9, 2023 at 11:35 AM Zhirun Yan wrote: > > > > This patch adds graph API for supporting to clone the graph object for > > a specified worker core. The new graph will also clone all nodes. > > > > Signed-off-by: Haiyue Wang > > Signed-off-by: Cunming Liang > > Signed-off-by: Zhirun Yan > > --- > > + > > +static rte_graph_t > > +graph_clone(struct graph *parent_graph, const char *name) { > > + struct graph_node *graph_node; > > + struct graph *graph; > > + > > + graph_spinlock_lock(); > > + > > + /* Don't allow to clone a node from a cloned graph */ > > Both clone_name() and graph_clone() kind of duplicate rte_node_clone(), Please > check, Can we have common _private_ function to reuse just the common code > between them rte_graph_clone(), graph_clone() in graph.c and rte_node_clone(), node_clone() in node.c are different. They are specific for node or graph. I think they could remain the same. Only clone_name(struct rte_node_register *reg, struct node *node, const char *name) could be reuse. And I will change the param to clone_name(char *name, char* ori_str, const char *append_str) and put this func into graph_private.h > > > > + > > rte_graph_t > > rte_graph_from_name(const char *name) { diff --git > > a/lib/graph/graph_private.h b/lib/graph/graph_private.h index > > f63b339d81..52ca30ed56 100644 > > --- a/lib/graph/graph_private.h > > +++ b/lib/graph/graph_private.h > > @@ -99,6 +99,8 @@ struct graph { > > /**< Circular buffer mask for wrap around. */ > > rte_graph_t id; > > /**< Graph identifier. */ > > + rte_graph_t parent_id; > > + /**< Parent graph identifier. */ > > unsigned int lcore_id; > > /**< Lcore identifier where the graph prefer to run on. */ > > size_t mem_sz; > > diff --git a/lib/graph/rte_graph.h b/lib/graph/rte_graph.h index > > c523809d1f..2f86c17de7 100644 > > --- a/lib/graph/rte_graph.h > > +++ b/lib/graph/rte_graph.h > > @@ -247,6 +247,26 @@ rte_graph_t rte_graph_create(const char *name, > > struct rte_graph_param *prm); __rte_experimental int > > rte_graph_destroy(rte_graph_t id); > > > > +/** > > + * Clone Graph. > > + * > > + * Clone a graph from static graph (graph created from > > +rte_graph_create). And > > rte_graph_create->rte_graph_create() > > > + * all cloned graphs attached to the parent graph MUST be destroyed > > + together > > > Can we add reference count in the implementation to avoid this limition. If > this > too much for this release, we can try to add next release Yes, I think it's good in next release. Thanks. > > > + * for fast schedule design limitation (stop ALL graph walk firstly). > > + * > > + * @param id > > + * Static graph id to clone from. > > + * @param name > > + * Name of the new graph. The library prepends the parent graph name to > the > > + * user-specified name. The final graph name will be, > > + * "parent graph name" + "-" + name. > > + * > > + * @return > > + * Valid graph id on success, RTE_GRAPH_ID_INVALID otherwise. > > + */ > > +__rte_experimental > > +rte_graph_t rte_graph_clone(rte_graph_t id, const char *name); > > + > > /** > > * Get graph id from graph name. > > * > > diff --git a/lib/graph/version.map b/lib/graph/version.map index > > 7de6f08f59..aaa86f66ed 100644 > > --- a/lib/graph/version.map > > +++ b/lib/graph/version.map > > @@ -7,6 +7,7 @@ EXPERIMENTAL { > > > > rte_graph_create; > > rte_graph_destroy; > > + rte_graph_clone; > > > Found new generic graph API, Please add test case for it at > app/test/test_graph.c. > Yes, I will add test. > > > > rte_graph_dump; > > rte_graph_export; > > rte_graph_from_name; > > -- > > 2.37.2 > >
RE: [PATCH v6 09/15] graph: introduce stream moving cross cores
> -Original Message- > From: Jerin Jacob > Sent: Wednesday, May 24, 2023 4:00 PM > To: Yan, Zhirun > Cc: dev@dpdk.org; jer...@marvell.com; kirankum...@marvell.com; > ndabilpu...@marvell.com; step...@networkplumber.org; > pbhagavat...@marvell.com; Liang, Cunming ; Wang, > Haiyue > Subject: Re: [PATCH v6 09/15] graph: introduce stream moving cross cores > > On Tue, May 9, 2023 at 11:35 AM Zhirun Yan wrote: > > > > This patch introduces key functions to allow a worker thread to enable > > enqueue and move streams of objects to the next nodes over different > > cores. > > different cores-> different cores for mcore dispatch model. > Got it. Thanks. > > > > > Signed-off-by: Haiyue Wang > > Signed-off-by: Cunming Liang > > Signed-off-by: Zhirun Yan > > --- > > lib/graph/graph.c| 6 +- > > lib/graph/graph_private.h| 30 + > > lib/graph/meson.build| 2 +- > > lib/graph/rte_graph.h| 15 ++- > > lib/graph/rte_graph_model_dispatch.c | 157 > > +++ lib/graph/rte_graph_model_dispatch.h | 37 > +++ > > lib/graph/version.map| 2 + > > 7 files changed, 244 insertions(+), 5 deletions(-) > > > > diff --git a/lib/graph/graph.c b/lib/graph/graph.c index > > e809aa55b0..f555844d8f 100644 > > --- a/lib/graph/graph.c > > +++ b/lib/graph/graph.c > > @@ -495,7 +495,7 @@ clone_name(struct graph *graph, struct graph > > *parent_graph, const char *name) } > > > > static rte_graph_t > > -graph_clone(struct graph *parent_graph, const char *name) > > +graph_clone(struct graph *parent_graph, const char *name, struct > > +rte_graph_param *prm) > > { > > struct graph_node *graph_node; > > struct graph *graph; > > @@ -566,14 +566,14 @@ graph_clone(struct graph *parent_graph, const > > char *name) } > > > > --- a/lib/graph/rte_graph.h > > +++ b/lib/graph/rte_graph.h > > @@ -169,6 +169,17 @@ struct rte_graph_param { > > bool pcap_enable; /**< Pcap enable. */ > > uint64_t num_pkt_to_capture; /**< Number of packets to capture. */ > > char *pcap_filename; /**< Filename in which packets to be > > captured.*/ > > + > > + RTE_STD_C11 > > + union { > > + struct { > > + uint64_t rsvd[8]; > > + } rtc; > > + struct { > > + uint32_t wq_size_max; > > + uint32_t mp_capacity; > > Add doxgen comment for all please. > > > + } dispatch; > > + }; > > }; > > > > /** > > @@ -260,12 +271,14 @@ int rte_graph_destroy(rte_graph_t id); > > * Name of the new graph. The library prepends the parent graph name to > the > > * user-specified name. The final graph name will be, > > * "parent graph name" + "-" + name. > > + * @param prm > > + * Graph parameter, includes model-specific parameters in this graph. > > * > > * @return > > * Valid graph id on success, RTE_GRAPH_ID_INVALID otherwise. > > */ > > __rte_experimental > > -rte_graph_t rte_graph_clone(rte_graph_t id, const char *name); > > +rte_graph_t rte_graph_clone(rte_graph_t id, const char *name, struct > > +rte_graph_param *prm); > > > > /** > > +void > > +__rte_graph_sched_wq_process(struct rte_graph *graph) { > > + struct graph_sched_wq_node *wq_node; > > + struct rte_mempool *mp = graph->mp; > > + struct rte_ring *wq = graph->wq; > > + uint16_t idx, free_space; > > + struct rte_node *node; > > + unsigned int i, n; > > + struct graph_sched_wq_node *wq_nodes[32]; > > Use RTE_GRAPH_BURST_SIZE instead of 32, if it is anything do with burst size? > else ignore. No, wq_nodes[32] is just a temporary space to consume the task. I will add a macro WQ_SIZE to define 32. > > > > + > > + n = rte_ring_sc_dequeue_burst_elem(wq, wq_nodes, > sizeof(wq_nodes[0]), > > + RTE_DIM(wq_nodes), NULL); > > + if (n == 0) > > + return; > > + > > + for (i = 0; i < n; i++) { > > + wq_node = wq_nodes[i]; > > + node = RTE_PTR_ADD(graph, wq_node->node_off); > > + RTE_ASSERT(node->fence == RTE_GRAPH_FENCE); > > + idx = node->idx; > > + free_space = node->size - idx; > > + > > + if (unlikely(free_space < wq_node->nb_objs)) > > + __rte_node_stream_alloc_size(graph, node, > > + node->size + wq_node->nb_objs); > > + > > + memmove(&node->objs[idx], wq_node->objs, wq_node->nb_objs * > sizeof(void *)); > > + node->idx = idx + wq_node->nb_objs; > > + > > + __rte_node_process(graph, node); > > + > > + wq_node->nb_objs = 0; > > + node->idx = 0; > > + } > > + > > + rte_mempool_put_bulk(mp, (void **)wq_nodes, n); } > > + > > +/** > > + * @internal > > For both internal function, you can add Doxygen comment
RE: [PATCH v6 13/15] graph: add stats for cross-core dispatching
> -Original Message- > From: Jerin Jacob > Sent: Wednesday, May 24, 2023 4:09 PM > To: Yan, Zhirun > Cc: dev@dpdk.org; jer...@marvell.com; kirankum...@marvell.com; > ndabilpu...@marvell.com; step...@networkplumber.org; > pbhagavat...@marvell.com; Liang, Cunming ; Wang, > Haiyue > Subject: Re: [PATCH v6 13/15] graph: add stats for cross-core dispatching > > On Tue, May 9, 2023 at 11:35 AM Zhirun Yan wrote: > > > > Add stats for cross-core dispatching scheduler if stats collection is > > enabled. > > > > Signed-off-by: Haiyue Wang > > Signed-off-by: Cunming Liang > > Signed-off-by: Zhirun Yan > > > diff --git a/lib/graph/rte_graph.h b/lib/graph/rte_graph.h index > > 0ac764daf8..ee6c970ca4 100644 > > --- a/lib/graph/rte_graph.h > > +++ b/lib/graph/rte_graph.h > > @@ -219,6 +219,8 @@ struct rte_graph_cluster_node_stats { > > uint64_t prev_calls;/**< Previous number of calls. */ > > uint64_t prev_objs; /**< Previous number of processed objs. */ > > uint64_t prev_cycles; /**< Previous number of cycles. */ > > + uint64_t sched_objs;/**< Previous number of scheduled objs. */ > > + uint64_t sched_fail;/**< Previous number of failed schedule > > objs. */ > > Add comment to specify it is for mcore_dispatch model. Also make it as > anonymous union so that later we can add new item for other model. Sure. Will do in next version.
RE: [PATCH v6 15/15] doc: update multicore dispatch model in graph guides
> -Original Message- > From: Jerin Jacob > Sent: Wednesday, May 24, 2023 4:13 PM > To: Yan, Zhirun > Cc: dev@dpdk.org; jer...@marvell.com; kirankum...@marvell.com; > ndabilpu...@marvell.com; step...@networkplumber.org; > pbhagavat...@marvell.com; Liang, Cunming ; Wang, > Haiyue > Subject: Re: [PATCH v6 15/15] doc: update multicore dispatch model in graph > guides > > On Tue, May 9, 2023 at 11:35 AM Zhirun Yan wrote: > > > > Update graph documentation to introduce new multicore dispatch model. > > > > Signed-off-by: Haiyue Wang > > Signed-off-by: Cunming Liang > > Signed-off-by: Zhirun Yan > > --- > > doc/guides/prog_guide/graph_lib.rst | 59 > > +++-- > > 1 file changed, 55 insertions(+), 4 deletions(-) > > > > diff --git a/doc/guides/prog_guide/graph_lib.rst > > b/doc/guides/prog_guide/graph_lib.rst > > index 1cfdc86433..72e26f3a5a 100644 > > --- a/doc/guides/prog_guide/graph_lib.rst > > +++ b/doc/guides/prog_guide/graph_lib.rst > > @@ -189,14 +189,65 @@ In the above example, A graph object will be > > created with ethdev Rx node of port 0 and queue 0, all ipv4* nodes in > > the system, and ethdev tx node of all ports. > > > > -Multicore graph processing > > -~~ > > -In the current graph library implementation, specifically, > > -``rte_graph_walk()`` and ``rte_node_enqueue*`` fast path API > > functions > > +graph model chossing > > Graph models > > > > + > > +Currently, there are 2 different walking model. Use > > model -> models Got it. Thanks.
RE: [PATCH v6 12/15] graph: enable graph multicore dispatch scheduler model
> -Original Message- > From: Jerin Jacob > Sent: Wednesday, May 24, 2023 4:46 PM > To: Yan, Zhirun > Cc: dev@dpdk.org; jer...@marvell.com; kirankum...@marvell.com; > ndabilpu...@marvell.com; step...@networkplumber.org; > pbhagavat...@marvell.com; Liang, Cunming ; Wang, > Haiyue ; mattias.ronnblom > > Subject: Re: [PATCH v6 12/15] graph: enable graph multicore dispatch scheduler > model > > On Tue, May 9, 2023 at 11:35 AM Zhirun Yan wrote: > > > > This patch enables to chose new scheduler model. > > > > Signed-off-by: Haiyue Wang > > Signed-off-by: Cunming Liang > > Signed-off-by: Zhirun Yan > > > rte_graph_walk(struct rte_graph *graph) { > > - rte_graph_walk_rtc(graph); > > + int model = rte_graph_worker_model_get(); > > Any specific to reason to keep model value in LCORE variable , why not in > struct > rte_graph? > It is not specific to this patch. But good to understand as storing in > rte_graph* will avoid cache miss. > Yes, I can put it into rte_graph. > > > + > > + if (model == RTE_GRAPH_MODEL_DEFAULT || > > + model == RTE_GRAPH_MODEL_RTC) > > I think, there can be three ways to do this > > a) Store model in PER_LCORE or struct rte_graph and add runtime check > b) Make it as function pointer for graph_walk > > mcore_dispatch model is reusing all rte_node_enqueue_* functions, so for > NOW only graph walk is different. > But if need to integrate other graph models like eventdev backend(similar > problem trying to solve in > https://patches.dpdk.org/project/dpdk/patch/20230522091628.96236-2- > mattias.ronnb...@ericsson.com/), > I think, we need to change enqueue variants. > Yes, there is no change for rte_node_enqueue_*. I will follow this thread. And may make some contribution after this release. > Both (a) and (b) has little performance impact in "current situation with this > patch" and if we need to add similar check and function pointer for overriding > node_enqueue_ functions it will have major impact. > In order to have NO performance impact and able to overide node_enqueue_ > functions, I think, we can have the following scheme in application and > library. > > In application > #define RTE_GRAPH_MODEL_SELECT RTE_GRAPH_MODEL_RTC #include > > > In library: > #if RTE_GRAPH_MODEL_SELECT == RTE_GRAPH_MODEL_RTC #define > rte_graph_walk rte_graph_walk_rtc #else if RTE_GRAPH_MODEL_SELECT == > RTE_GRAPH_MODEL_MCORE_DISPATCH #define rte_graph_walk > rte_graph_walk_mcore_dispatch > > It is kind of compile time, But application can use function templating by > proving > different values RTE_GRAPH_MODEL_SELECT to make runtime if given > application needs to support all modes at runtime. > > > As an example: > > app_my_graph_processing.h has application code for graph walk and node > processing. > > app_worker_rtc.c > > #define RTE_GRAPH_MODEL_SELECT RTE_GRAPH_MODEL_RTC #include > #include app_my_graph_processing.h > > void app_worker_rtc() > { > while (1) { >rte_graph_walk() > } > } > > app_worker_mcore_dispatch.c > - > > #define RTE_GRAPH_MODEL_SELECT RTE_GRAPH_MODEL_MCORE_DISPATCH > #include #include app_my_graph_processing.h > > void app_worker_mcore_dispatch() > { > while (1) { >rte_graph_walk() > } > } > > in main() > - > > if (command line arg provided worker as rtc) Got it. And we could use rte_graph->model to choose rtc or dispatch for future. Then it could be possible for models coexistence as you said before. > rte_eal_remote_launch(app_worker_rtc) > else > rte_eal_remote_launch(app_worker_mcore_dispatch) > > - > With that note, ending review comment for this series. > > In general patches look good high level, following items need to be sorted in > next version. Then I think, it is good to merge in this release. > > 1) Above points on fixing performance and supporting more graph model > variants > 2) Need to add UT for ALL new APIs in app/test/test_graph.c > 3) Make sure no performance regression with app/test/test_graph_perf.c with > new changes > 4) Addressing existing comments in this series. > > Thanks for great work. Hi Jerin, Thanks for your review. I will fix these in next version.
RE: [PATCH v2] net/ice: support double vlan
> From: Mingjin Ye > Sent: Saturday, May 6, 2023 6:05 PM > To: dev@dpdk.org > Cc: Yang, Qiming ; Zhou, YidingX > ; Ye, MingjinX ; Zhang, Qi Z > > Subject: [PATCH v2] net/ice: support double vlan > > Aligned with kernel driver, optimized for inner and outer VLAN handling in > DPDK, > and implemented double vlan insertion and stripping support. > NIC must work in double vlan mode(DVM), depending on FW/SW. > > 1. Adjust vlan stripping > The vlan stripping enable/disable is applied to the inner vlan. > > 2. Support QinQ stripping > The ice outer vlan strip is enabled/disabled by the mask bit of > `RTE_ETH_RX_OFFLOAD_QINQ_STRIP`, and the user can use "vlan set qinq_strip > on 0" to enable or "vlan setqinq_strip off 0" to disable the ice outer vlan > strip in > testpmd. > > 3. Support outer tag type switching > Implement the ethdev `vlan_tpid_set` api to enable outer tag support to handle > `RTE_ETHER_TYPE_VLAN`` RTE_ETHER_TYPE_QINQ`` RTE_ETHER_TYPE_QINQ1` > outer tag types. > > 4. Support outer port-based vlan insertion Implement port-based outer vlan > insertion. User can use "tx_vlan set pvid 0 45 on" to enable or "tx_vlan set > pvid 0 > 45 off" to disable the outer vlan insertion in testpmd. > > Signed-off-by: Mingjin Ye Tested and passed. There are several points need to be mentioned. 1. Only Scalar path is implemented, needs force-max-simd-bitwidth. 2. Not supported: StripQ. 3. Hardware limitation: Requires different outer and inner VLAN type. It means setting outer TPID to 0x88a8 before double VLAN insert. 4. Undefined behaviour: VLAN insert and TPID configured together is not defined in this new feature. TPID is designed to work with PVID now. Under this implementation changing TPID would not change the TX outer VLAN type of double VLAN insert. Tested-by: Ke Xu > --- > v2: Apply QinQ when initializing vlan offload. > --- > drivers/net/ice/ice_ethdev.c | 422 +-- > drivers/net/ice/ice_ethdev.h | 1 + > 2 files changed, 408 insertions(+), 15 deletions(-) >
Re: [PATCH 02/15] common/cnxk: add MACsec SA configuration
On Wed, May 24, 2023 at 1:34 AM Akhil Goyal wrote: > > Added ROC APIs to allocate/free MACsec resources > and APIs to write SA policy. > > Signed-off-by: Ankur Dwivedi > Signed-off-by: Vamsi Attunuru > Signed-off-by: Akhil Goyal > --- > index 7593c7c890..9266edd9a1 100644 > --- a/drivers/common/cnxk/version.map > +++ b/drivers/common/cnxk/version.map > @@ -133,10 +133,14 @@ INTERNAL { > roc_se_auth_key_set; > roc_se_ciph_key_set; > roc_se_ctx_init; > + roc_mcs_alloc_rsrc; Please change to roc_mcs_rsrc_alloc i.e keeping action as last > roc_mcs_dev_init; > roc_mcs_dev_fini; > roc_mcs_dev_get; > + roc_mcs_free_rsrc; Same as above. > roc_mcs_hw_info_get; > + roc_mcs_sa_policy_read; > + roc_mcs_sa_policy_write; > roc_nix_bpf_alloc; > roc_nix_bpf_config; > roc_nix_bpf_connect; > -- > 2.25.1 >
Re: [PATCH 05/15] common/cnxk: add MACsec PN and LMAC mode configuration
On Wed, May 24, 2023 at 1:34 AM Akhil Goyal wrote: > > Added ROC APIs for setting packet number and LMAC > related configurations. > > Signed-off-by: Ankur Dwivedi > Signed-off-by: Vamsi Attunuru > Signed-off-by: Akhil Goyal > +struct mcs_set_active_lmac { > + struct mbox_msghdr hdr; > + uint32_t __io lmac_bmap; /* bitmap of active lmac per mcs block */ > + uint8_t __io mcs_id; > + uint16_t channel_base; /* MCS channel base */ Gotcha . Found a missing __io > + uint64_t __io rsvd; > +}; > +
Re: [PATCH 09/15] common/cnxk: add MACsec control port configuration
On Wed, May 24, 2023 at 1:35 AM Akhil Goyal wrote: > > Added ROC APIs to configure MACsec control port. > > Signed-off-by: Ankur Dwivedi > Signed-off-by: Vamsi Attunuru > Signed-off-by: Akhil Goyal > --- > __roc_api int roc_mcs_flowid_stats_get(struct roc_mcs *mcs, struct > roc_mcs_stats_req *mcs_req, >struct roc_mcs_flowid_stats *stats); > diff --git a/drivers/common/cnxk/version.map b/drivers/common/cnxk/version.map > index 9ba804a04f..e501cdd4d9 100644 > --- a/drivers/common/cnxk/version.map > +++ b/drivers/common/cnxk/version.map > @@ -134,7 +134,10 @@ INTERNAL { > roc_se_ciph_key_set; > roc_se_ctx_init; > roc_mcs_active_lmac_set; > + roc_mcs_alloc_ctrl_pkt_rule; roc_mcs_ctrl_pkt_rule_alloc > roc_mcs_alloc_rsrc; > + roc_mcs_ctrl_pkt_rule_write; > + roc_mcs_custom_tag_cfg_get; > roc_mcs_dev_init; > roc_mcs_dev_fini; > roc_mcs_dev_get; > @@ -144,6 +147,7 @@ INTERNAL { > roc_mcs_flowid_entry_read; > roc_mcs_flowid_entry_write; > roc_mcs_flowid_stats_get; > + roc_mcs_free_ctrl_pkt_rule; roc_mcs_ctrl_pkt_rule_free > roc_mcs_free_rsrc; > roc_mcs_hw_info_get; > roc_mcs_intr_configure; > -- > 2.25.1 >
[PATCH v4] crypto/qat: default to IPsec MB for computations
Pre and post computations currently use the OpenSSL library by default. This patch changes the default option to Intel IPsec MB library version 1.4 for the required computations. If this version of IPsec is not met it will fallback to use OpenSSL. Added version checks for libipsecmb and libcrypto into meson build. Added directives for detecting IPsec MB or OpenSSL. Signed-off-by: Brian Dooley --- v4: Rebase and fix conflicts v3: Fix checkpatch warnings by replacing return with goto v2: OpenSSL code reintroduced as a fallback feature if Intel IPsec MB 1.4 not available --- doc/guides/cryptodevs/qat.rst| 14 +- drivers/common/qat/meson.build | 30 +- drivers/common/qat/qat_device.c | 1 - drivers/common/qat/qat_device.h | 3 +- drivers/crypto/qat/dev/qat_crypto_pmd_gens.h | 10 + drivers/crypto/qat/qat_sym.c | 9 +- drivers/crypto/qat/qat_sym.h | 27 + drivers/crypto/qat/qat_sym_session.c | 874 ++- drivers/crypto/qat/qat_sym_session.h | 14 + 9 files changed, 542 insertions(+), 440 deletions(-) diff --git a/doc/guides/cryptodevs/qat.rst b/doc/guides/cryptodevs/qat.rst index a4a25711ed..8666ac8b88 100644 --- a/doc/guides/cryptodevs/qat.rst +++ b/doc/guides/cryptodevs/qat.rst @@ -317,18 +317,12 @@ To use this feature the user must set the devarg on process start as a device ad -a 03:01.1,qat_sym_cipher_crc_enable=1 -Running QAT PMD with Intel IPSEC MB library for symmetric precomputes function +Running QAT PMD with Intel IPsec MB library for symmetric precomputes function ~~ -The QAT PMD use Openssl library for partial hash calculation in symmetirc precomputes function by -default, the following parameter is allow QAT PMD switch over to multi-buffer job API if Intel -IPSEC MB library installed on system. - -- qat_ipsec_mb_lib - -To use this feature the user must set the parameter on process start as a device additional parameter:: - - -a 03:01.1,qat_ipsec_mb_lib=1 +The QAT PMD uses Intel IPsec MB library for partial hash calculation in symmetric precomputes function by +default, the minimum required version of IPsec MB library is v1.4. If this version of IPsec is not met it +will fallback to use OpenSSL Device and driver naming diff --git a/drivers/common/qat/meson.build b/drivers/common/qat/meson.build index b84e5b3c6c..9c5eb99acd 100644 --- a/drivers/common/qat/meson.build +++ b/drivers/common/qat/meson.build @@ -28,30 +28,37 @@ if disable_drivers.contains(qat_compress_path) endif libcrypto = dependency('libcrypto', required: false, method: 'pkg-config') -if qat_crypto and not libcrypto.found() -qat_crypto = false -dpdk_drvs_disabled += qat_crypto_path -set_variable(qat_crypto_path.underscorify() + '_disable_reason', -'missing dependency, libcrypto') -endif -IMB_required_ver = '1.2.0' +IMB_required_ver = '1.4.0' IMB_header = '#include' if arch_subdir == 'arm' IMB_header = '#include' endif libipsecmb = cc.find_library('IPSec_MB', required: false) -libcrypto_3 = dependency('libcrypto', required: false, -method: 'pkg-config', version : '>=3.0.0') -if libipsecmb.found() and libcrypto_3.found() +if libipsecmb.found() # version comes with quotes, so we split based on " and take the middle imb_ver = cc.get_define('IMB_VERSION_STR', prefix : IMB_header).split('"')[1] if (imb_ver.version_compare('>=' + IMB_required_ver)) ext_deps += libipsecmb -dpdk_conf.set('RTE_QAT_LIBIPSECMB', true) +elif libcrypto.found() +ext_deps += libcrypto +dpdk_conf.set('RTE_QAT_OPENSSL', true) +else +qat_crypto = false +dpdk_drvs_disabled += qat_crypto_path +set_variable(qat_crypto_path.underscorify() + '_disable_reason', +'missing dependency, libipsecmb or libcrypto') endif +elif libcrypto.found() +ext_deps += libcrypto +dpdk_conf.set('RTE_QAT_OPENSSL', true) +else +qat_crypto = false +dpdk_drvs_disabled += qat_crypto_path +set_variable(qat_crypto_path.underscorify() + '_disable_reason', +'missing dependency, libipsecmb or libcrypto') endif # The driver should not build if both compression and crypto are disabled @@ -103,6 +110,5 @@ if qat_crypto sources += files(join_paths(qat_crypto_relpath, f)) endforeach deps += ['security'] -ext_deps += libcrypto cflags += ['-DBUILD_QAT_SYM', '-DBUILD_QAT_ASYM'] endif diff --git a/drivers/common/qat/qat_device.c b/drivers/common/qat/qat_device.c index 0479175b65..20e56b5cf2 100644 --- a/drivers/common/qat/qat_device.c +++ b/drivers/common/qat/qat_device.c @@ -371,7 +371,6 @@ static int qat_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, struct qat_pci_device *qat_pci_dev; struct qat_dev_hw_spec_funcs *ops_hw; struct qat_dev_cmd_par
Re: [PATCH 1/2] event/cnxk: fix ns to ticks conversion
On Wed, May 24, 2023 at 2:22 PM wrote: > > From: Pavan Nikhilesh > > The number of timeout ticks is dependent on the global dequeue > timeout configured. > > Fixes: 6223ede20361 ("event/cnxk: add event port link and unlink") > > Signed-off-by: Pavan Nikhilesh Series applied to dpdk-next-net-eventdev/for-main. Thanks > --- > drivers/event/cnxk/cnxk_eventdev.c | 4 ++-- > 1 file changed, 2 insertions(+), 2 deletions(-) > > diff --git a/drivers/event/cnxk/cnxk_eventdev.c > b/drivers/event/cnxk/cnxk_eventdev.c > index 99f9cdcd0d..3a05a59c75 100644 > --- a/drivers/event/cnxk/cnxk_eventdev.c > +++ b/drivers/event/cnxk/cnxk_eventdev.c > @@ -323,9 +323,9 @@ int > cnxk_sso_timeout_ticks(struct rte_eventdev *event_dev, uint64_t ns, >uint64_t *tmo_ticks) > { > - RTE_SET_USED(event_dev); > - *tmo_ticks = NSEC2TICK(ns, rte_get_timer_hz()); > + struct cnxk_sso_evdev *dev = cnxk_sso_pmd_priv(event_dev); > > + *tmo_ticks = dev->deq_tmo_ns ? ns / dev->deq_tmo_ns : 0; > return 0; > } > > -- > 2.25.1 >
[PATCH] net/ice: init dvm mode for parser
Double Vlan mode need to be configured for parser Otherwise parser result will not be consistent with hardware. Fixes: 531d2555c8a6 ("net/ice: refactor parser usage") Cc: sta...@dpdk.org Signed-off-by: Qi Zhang --- drivers/net/ice/ice_generic_flow.c | 5 + 1 file changed, 5 insertions(+) diff --git a/drivers/net/ice/ice_generic_flow.c b/drivers/net/ice/ice_generic_flow.c index 86a32f8cb1..ed3075d555 100644 --- a/drivers/net/ice/ice_generic_flow.c +++ b/drivers/net/ice/ice_generic_flow.c @@ -1836,6 +1836,11 @@ ice_flow_init(struct ice_adapter *ad) if (ice_parser_create(&ad->hw, &ad->psr) != ICE_SUCCESS) PMD_INIT_LOG(WARNING, "Failed to initialize DDP parser, raw packet filter will not be supported"); + if (ice_is_dvm_ena(&ad->hw)) + ice_parser_dvm_set(ad->psr, true); + else + ice_parser_dvm_set(ad->psr, false); + RTE_TAILQ_FOREACH_SAFE(engine, &engine_list, node, temp) { if (engine->init == NULL) { PMD_INIT_LOG(ERR, "Invalid engine type (%d)", -- 2.31.1
RE: [PATCH v2] net/ice: support double vlan
> -Original Message- > From: Xu, Ke1 > Sent: Friday, May 26, 2023 6:16 PM > To: Ye, MingjinX ; dev@dpdk.org > Cc: Yang, Qiming ; Zhou, YidingX > ; Ye, MingjinX ; Zhang, Qi > Z > Subject: RE: [PATCH v2] net/ice: support double vlan > > > From: Mingjin Ye > > Sent: Saturday, May 6, 2023 6:05 PM > > To: dev@dpdk.org > > Cc: Yang, Qiming ; Zhou, YidingX > > ; Ye, MingjinX ; Zhang, > > Qi Z > > Subject: [PATCH v2] net/ice: support double vlan > > > > Aligned with kernel driver, optimized for inner and outer VLAN > > handling in DPDK, and implemented double vlan insertion and stripping > support. > > NIC must work in double vlan mode(DVM), depending on FW/SW. > > > > 1. Adjust vlan stripping > > The vlan stripping enable/disable is applied to the inner vlan. > > > > 2. Support QinQ stripping > > The ice outer vlan strip is enabled/disabled by the mask bit of > > `RTE_ETH_RX_OFFLOAD_QINQ_STRIP`, and the user can use "vlan set > > qinq_strip on 0" to enable or "vlan setqinq_strip off 0" to disable > > the ice outer vlan strip in testpmd. > > > > 3. Support outer tag type switching > > Implement the ethdev `vlan_tpid_set` api to enable outer tag support > > to handle `RTE_ETHER_TYPE_VLAN`` RTE_ETHER_TYPE_QINQ`` > > RTE_ETHER_TYPE_QINQ1` outer tag types. > > > > 4. Support outer port-based vlan insertion Implement port-based outer > > vlan insertion. User can use "tx_vlan set pvid 0 45 on" to enable or > > "tx_vlan set pvid 0 > > 45 off" to disable the outer vlan insertion in testpmd. > > > > Signed-off-by: Mingjin Ye > > Tested and passed. > > There are several points need to be mentioned. > 1. Only Scalar path is implemented, needs force-max-simd-bitwidth. > 2. Not supported: StripQ. > 3. Hardware limitation: Requires different outer and inner VLAN type. It > means > setting outer TPID to 0x88a8 before double VLAN insert. > 4. Undefined behaviour: VLAN insert and TPID configured together is not > defined > in this new feature. TPID is designed to work with PVID now. Under this > implementation > changing TPID would not change the TX outer VLAN type of double VLAN > insert. > > Tested-by: Ke Xu Acked-by: Qi Zhang Applied to dpdk-next-net-intel. Thanks Qi
"Thread safety" in rte_flow
Hello Ori, ethdev maintainers, I am a bit puzzled at the RTE_ETH_DEV_FLOW_OPS_THREAD_SAFE checks in rte_flow.c. - The rte_flow.h does not hint at what is being protected. All I can see is a somewhat vague, in lib/ethdev/rte_ethdev.h: /** PMD supports thread-safe flow operations */ #define RTE_ETH_DEV_FLOW_OPS_THREAD_SAFE RTE_BIT32(0) It would be great to have a more detailed description of what this thread safety means. - I don't think many functions of the rte_flow API look at this flag. It seems it is never checked in newly added symbols (this is an impression, I did not enter the details). Could you have a look? Thanks. -- David Marchand
Re: [RFC v2 1/2] dts: add smoke tests
On Fri, May 26, 2023 at 3:36 AM Juraj Linkeš wrote: > On Thu, May 25, 2023 at 8:03 PM Jeremy Spewock > wrote: > > > > Hey Juraj, > > > > On Thu, May 25, 2023 at 4:33 AM Juraj Linkeš > wrote: > >> > >> One more point that doesn't fit elsewhere: > >> > >> On Wed, May 24, 2023 at 10:45 PM Jeremy Spewock > wrote: > >> > > >> > > >> > > >> > On Tue, May 23, 2023 at 4:05 AM Juraj Linkeš > wrote: > >> >> > >> >> Hi Jeremy, first, a few general points: > >> >> > >> >> 1. Send patches to maintainers (Thomas, me, Honnappa, Lijuan and > >> >> anyone else involved with DTS or who might be interested) and add the > >> >> devlist to cc. > >> > > >> > > >> > Thank you for the tip! I'm still new to sending patches and didn't > think to do something like this but I will in the future. > >> > > >> >> > >> >> 2. Run the linter script before submitting. > >> > > >> > > >> > I did forget to run this, I will in the future. > >> > > >> >> > >> >> 3. The use of the various nested objects breaks the current > >> >> abstractions. The basic idea is that the test suite developers should > >> >> ideally only use the sut/tg node objects and those objects should > >> >> delegate logic further to their nested objects. More below. > >> >> > >> >> I have many comments about the implementation, but I haven't run it > >> >> yet. I'm going to do that after this round of comments and I may have > >> >> more ideas. > >> >> > >> >> On Fri, May 12, 2023 at 9:28 PM wrote: > >> >> > > >> >> > From: Jeremy Spewock > >> >> > > >> >> > Adds a new test suite for running smoke tests that verify general > >> >> > configuration aspects of the system under test. If any of these > tests > >> >> > fail, the DTS execution terminates as part of a "fail-fast" model. > >> >> > > >> >> > Signed-off-by: Jeremy Spewock > >> >> > --- > >> >> > dts/conf.yaml | 9 ++ > >> >> > dts/framework/config/__init__.py | 21 + > >> >> > dts/framework/config/conf_yaml_schema.json| 32 ++- > >> >> > dts/framework/dts.py | 19 +++- > >> >> > dts/framework/exception.py| 11 +++ > >> >> > dts/framework/remote_session/os_session.py| 6 +- > >> >> > .../remote_session/remote/__init__.py | 28 ++ > >> >> > dts/framework/test_result.py | 13 ++- > >> >> > dts/framework/test_suite.py | 24 - > >> >> > dts/framework/testbed_model/__init__.py | 5 + > >> >> > .../interactive_apps/__init__.py | 6 ++ > >> >> > .../interactive_apps/interactive_command.py | 57 +++ > >> >> > .../interactive_apps/testpmd_driver.py| 24 + > >> >> > dts/framework/testbed_model/node.py | 2 + > >> >> > dts/framework/testbed_model/sut_node.py | 6 ++ > >> >> > dts/tests/TestSuite_smoke_tests.py| 94 > +++ > >> >> > 16 files changed, 348 insertions(+), 9 deletions(-) > >> >> > create mode 100644 > dts/framework/testbed_model/interactive_apps/__init__.py > >> >> > create mode 100644 > dts/framework/testbed_model/interactive_apps/interactive_command.py > >> >> > create mode 100644 > dts/framework/testbed_model/interactive_apps/testpmd_driver.py > >> >> > >> >> Let's not add any more levels. I don't like even the current hw > >> >> subdirectory (which I want to remove in the docs patch) and we don't > >> >> need it. I'd also like to move this functionality into > remote_session, > >> >> as it's handling a type of remote session. > >> > > >> > > >> > I think it makes sense to add a proper wrapper for it, I just didn't > create a subclass for it off of remote_session because the idea behind it > was only allowing users to interact with the session through the > InteractiveScriptHandler/DPDK app handlers. If it were part of the > remote_session class it would have to include operations for sending > commands the way it is written now. I could do this but it seems like a > bigger discussion about whether interactive sessions should create a new > SSH session every time or instead just use one existing session and create > channels off of it. > >> > > >> > >> I wasn't clear, I meant to move the python modules into the > >> remote_session folder. The subclassing would be there only to have a > >> common API across these remote sessions (as the > >> InteractiveScriptHandler is a kind of a remote session). If at all > >> possible, this should be our aim. > > > > > > I see what you mean now, moving them there shouldn't be a problem. > > > >> > >> > >> >> > >> >> > >> >> > create mode 100644 dts/tests/TestSuite_smoke_tests.py > >> >> > > >> >> > diff --git a/dts/conf.yaml b/dts/conf.yaml > >> >> > index a9bd8a3e..042ef954 100644 > >> >> > --- a/dts/conf.yaml > >> >> > +++ b/dts/conf.yaml > >> >> > @@ -10,13 +10,22 @@ executions: > >> >> > compiler_wrapper: ccache > >> >> > perf: false > >> >> > func: true > >> >> > +nics: #physical devices to be us
[PATCH] common/cnxk: add new APIs for batch operations
Add new APIs for counting and extracting allocated objects from a single cache line in the batch alloc memory. Signed-off-by: Ashwin Sekhar T K --- drivers/common/cnxk/roc_npa.h | 78 ++- 1 file changed, 67 insertions(+), 11 deletions(-) diff --git a/drivers/common/cnxk/roc_npa.h b/drivers/common/cnxk/roc_npa.h index 21608a40d9..493ea01b13 100644 --- a/drivers/common/cnxk/roc_npa.h +++ b/drivers/common/cnxk/roc_npa.h @@ -209,7 +209,6 @@ roc_npa_aura_batch_alloc_issue(uint64_t aura_handle, uint64_t *buf, unsigned int num, const int dis_wait, const int drop) { - unsigned int i; int64_t *addr; uint64_t res; union { @@ -220,10 +219,6 @@ roc_npa_aura_batch_alloc_issue(uint64_t aura_handle, uint64_t *buf, if (num > ROC_CN10K_NPA_BATCH_ALLOC_MAX_PTRS) return -1; - /* Zero first word of every cache line */ - for (i = 0; i < num; i += (ROC_ALIGN / sizeof(uint64_t))) - buf[i] = 0; - addr = (int64_t *)(roc_npa_aura_handle_to_base(aura_handle) + NPA_LF_AURA_BATCH_ALLOC); cmp.u = 0; @@ -240,6 +235,9 @@ roc_npa_aura_batch_alloc_issue(uint64_t aura_handle, uint64_t *buf, return 0; } +/* + * Wait for a batch alloc operation on a cache line to complete. + */ static inline void roc_npa_batch_alloc_wait(uint64_t *cache_line) { @@ -251,6 +249,23 @@ roc_npa_batch_alloc_wait(uint64_t *cache_line) ; } +/* + * Count the number of pointers in a single batch alloc cache line. + */ +static inline unsigned int +roc_npa_aura_batch_alloc_count_line(uint64_t *line, unsigned int wait_us) +{ + struct npa_batch_alloc_status_s *status; + + status = (struct npa_batch_alloc_status_s *)line; + roc_npa_batch_alloc_wait(line, wait_us); + + return status->count; +} + +/* + * Count the number of pointers in a sequence of batch alloc cache lines. + */ static inline unsigned int roc_npa_aura_batch_alloc_count(uint64_t *aligned_buf, unsigned int num, unsigned int do_wait) @@ -276,6 +291,40 @@ roc_npa_aura_batch_alloc_count(uint64_t *aligned_buf, unsigned int num, return count; } +/* + * Extract allocated pointers from a single batch alloc cache line. This api + * only extracts the required number of pointers from the cache line and it + * adjusts the statsus->count so that a subsequent call to this api can + * extract the remaining pointers in the cache line appropriately. + */ +static inline unsigned int +roc_npa_aura_batch_alloc_extract_line(uint64_t *buf, uint64_t *line, + unsigned int num, unsigned int *rem) +{ + struct npa_batch_alloc_status_s *status; + unsigned int avail; + + status = (struct npa_batch_alloc_status_s *)line; + roc_npa_batch_alloc_wait(line, 0); + avail = status->count; + num = avail > num ? num : avail; + if (num) + memcpy(buf, &line[avail - num], num * sizeof(uint64_t)); + avail -= num; + if (avail == 0) { + /* Clear the lowest 7 bits of the first pointer */ + buf[0] &= ~0x7FUL; + status->ccode = 0; + } + status->count = avail; + *rem = avail; + + return num; +} + +/* + * Extract all allocated pointers from a sequence of batch alloc cache lines. + */ static inline unsigned int roc_npa_aura_batch_alloc_extract(uint64_t *buf, uint64_t *aligned_buf, unsigned int num) @@ -327,11 +376,15 @@ roc_npa_aura_op_bulk_free(uint64_t aura_handle, uint64_t const *buf, } } +/* + * Issue a batch alloc operation on a sequence of cache lines, wait for the + * batch alloc to complete and copy the pointers out into the user buffer. + */ static inline unsigned int roc_npa_aura_op_batch_alloc(uint64_t aura_handle, uint64_t *buf, - uint64_t *aligned_buf, unsigned int num, - const int dis_wait, const int drop, - const int partial) + unsigned int num, uint64_t *aligned_buf, + unsigned int aligned_buf_sz, const int dis_wait, + const int drop, const int partial) { unsigned int count, chunk, num_alloc; @@ -341,9 +394,12 @@ roc_npa_aura_op_batch_alloc(uint64_t aura_handle, uint64_t *buf, count = 0; while (num) { - chunk = (num > ROC_CN10K_NPA_BATCH_ALLOC_MAX_PTRS) ? - ROC_CN10K_NPA_BATCH_ALLOC_MAX_PTRS : - num; + /* Make sure that the pointers allocated fit into the cache +* lines reserved. +*/ + chunk = aligned_buf_sz / sizeof(uint64_t); + chunk = PLT_MIN(num, chunk); +
[PATCH 1/2] mempool/cnxk: avoid indefinite wait
Avoid waiting indefinitely when counting batch alloc pointers by adding a wait timeout. Signed-off-by: Ashwin Sekhar T K --- drivers/common/cnxk/roc_npa.h| 15 +-- drivers/mempool/cnxk/cn10k_mempool_ops.c | 3 ++- 2 files changed, 11 insertions(+), 7 deletions(-) diff --git a/drivers/common/cnxk/roc_npa.h b/drivers/common/cnxk/roc_npa.h index 21608a40d9..d3caa71586 100644 --- a/drivers/common/cnxk/roc_npa.h +++ b/drivers/common/cnxk/roc_npa.h @@ -241,19 +241,23 @@ roc_npa_aura_batch_alloc_issue(uint64_t aura_handle, uint64_t *buf, } static inline void -roc_npa_batch_alloc_wait(uint64_t *cache_line) +roc_npa_batch_alloc_wait(uint64_t *cache_line, unsigned int wait_us) { + const uint64_t ticks = (uint64_t)wait_us * plt_tsc_hz() / (uint64_t)1E6; + const uint64_t start = plt_tsc_cycles(); + /* Batch alloc status code is updated in bits [5:6] of the first word * of the 128 byte cache line. */ while (((__atomic_load_n(cache_line, __ATOMIC_RELAXED) >> 5) & 0x3) == ALLOC_CCODE_INVAL) - ; + if (wait_us && (plt_tsc_cycles() - start) >= ticks) + break; } static inline unsigned int roc_npa_aura_batch_alloc_count(uint64_t *aligned_buf, unsigned int num, - unsigned int do_wait) + unsigned int wait_us) { unsigned int count, i; @@ -267,8 +271,7 @@ roc_npa_aura_batch_alloc_count(uint64_t *aligned_buf, unsigned int num, status = (struct npa_batch_alloc_status_s *)&aligned_buf[i]; - if (do_wait) - roc_npa_batch_alloc_wait(&aligned_buf[i]); + roc_npa_batch_alloc_wait(&aligned_buf[i], wait_us); count += status->count; } @@ -293,7 +296,7 @@ roc_npa_aura_batch_alloc_extract(uint64_t *buf, uint64_t *aligned_buf, status = (struct npa_batch_alloc_status_s *)&aligned_buf[i]; - roc_npa_batch_alloc_wait(&aligned_buf[i]); + roc_npa_batch_alloc_wait(&aligned_buf[i], 0); line_count = status->count; diff --git a/drivers/mempool/cnxk/cn10k_mempool_ops.c b/drivers/mempool/cnxk/cn10k_mempool_ops.c index ba826f0f01..ff0015d8de 100644 --- a/drivers/mempool/cnxk/cn10k_mempool_ops.c +++ b/drivers/mempool/cnxk/cn10k_mempool_ops.c @@ -9,6 +9,7 @@ #define BATCH_ALLOC_SZ ROC_CN10K_NPA_BATCH_ALLOC_MAX_PTRS #define BATCH_OP_DATA_TABLE_MZ_NAME "batch_op_data_table_mz" +#define BATCH_ALLOC_WAIT_US 5 enum batch_op_status { BATCH_ALLOC_OP_NOT_ISSUED = 0, @@ -178,7 +179,7 @@ cn10k_mempool_get_count(const struct rte_mempool *mp) if (mem->status == BATCH_ALLOC_OP_ISSUED) count += roc_npa_aura_batch_alloc_count( - mem->objs, BATCH_ALLOC_SZ, 1); + mem->objs, BATCH_ALLOC_SZ, BATCH_ALLOC_WAIT_US); if (mem->status == BATCH_ALLOC_OP_DONE) count += mem->sz; -- 2.25.1
[PATCH 2/2] common/cnxk: add new APIs for batch operations
Add new APIs for counting and extracting allocated objects from a single cache line in the batch alloc memory. Signed-off-by: Ashwin Sekhar T K --- drivers/common/cnxk/roc_npa.h | 78 ++- 1 file changed, 67 insertions(+), 11 deletions(-) diff --git a/drivers/common/cnxk/roc_npa.h b/drivers/common/cnxk/roc_npa.h index d3caa71586..0653531198 100644 --- a/drivers/common/cnxk/roc_npa.h +++ b/drivers/common/cnxk/roc_npa.h @@ -209,7 +209,6 @@ roc_npa_aura_batch_alloc_issue(uint64_t aura_handle, uint64_t *buf, unsigned int num, const int dis_wait, const int drop) { - unsigned int i; int64_t *addr; uint64_t res; union { @@ -220,10 +219,6 @@ roc_npa_aura_batch_alloc_issue(uint64_t aura_handle, uint64_t *buf, if (num > ROC_CN10K_NPA_BATCH_ALLOC_MAX_PTRS) return -1; - /* Zero first word of every cache line */ - for (i = 0; i < num; i += (ROC_ALIGN / sizeof(uint64_t))) - buf[i] = 0; - addr = (int64_t *)(roc_npa_aura_handle_to_base(aura_handle) + NPA_LF_AURA_BATCH_ALLOC); cmp.u = 0; @@ -240,6 +235,9 @@ roc_npa_aura_batch_alloc_issue(uint64_t aura_handle, uint64_t *buf, return 0; } +/* + * Wait for a batch alloc operation on a cache line to complete. + */ static inline void roc_npa_batch_alloc_wait(uint64_t *cache_line, unsigned int wait_us) { @@ -255,6 +253,23 @@ roc_npa_batch_alloc_wait(uint64_t *cache_line, unsigned int wait_us) break; } +/* + * Count the number of pointers in a single batch alloc cache line. + */ +static inline unsigned int +roc_npa_aura_batch_alloc_count_line(uint64_t *line, unsigned int wait_us) +{ + struct npa_batch_alloc_status_s *status; + + status = (struct npa_batch_alloc_status_s *)line; + roc_npa_batch_alloc_wait(line, wait_us); + + return status->count; +} + +/* + * Count the number of pointers in a sequence of batch alloc cache lines. + */ static inline unsigned int roc_npa_aura_batch_alloc_count(uint64_t *aligned_buf, unsigned int num, unsigned int wait_us) @@ -279,6 +294,40 @@ roc_npa_aura_batch_alloc_count(uint64_t *aligned_buf, unsigned int num, return count; } +/* + * Extract allocated pointers from a single batch alloc cache line. This api + * only extracts the required number of pointers from the cache line and it + * adjusts the statsus->count so that a subsequent call to this api can + * extract the remaining pointers in the cache line appropriately. + */ +static inline unsigned int +roc_npa_aura_batch_alloc_extract_line(uint64_t *buf, uint64_t *line, + unsigned int num, unsigned int *rem) +{ + struct npa_batch_alloc_status_s *status; + unsigned int avail; + + status = (struct npa_batch_alloc_status_s *)line; + roc_npa_batch_alloc_wait(line, 0); + avail = status->count; + num = avail > num ? num : avail; + if (num) + memcpy(buf, &line[avail - num], num * sizeof(uint64_t)); + avail -= num; + if (avail == 0) { + /* Clear the lowest 7 bits of the first pointer */ + buf[0] &= ~0x7FUL; + status->ccode = 0; + } + status->count = avail; + *rem = avail; + + return num; +} + +/* + * Extract all allocated pointers from a sequence of batch alloc cache lines. + */ static inline unsigned int roc_npa_aura_batch_alloc_extract(uint64_t *buf, uint64_t *aligned_buf, unsigned int num) @@ -330,11 +379,15 @@ roc_npa_aura_op_bulk_free(uint64_t aura_handle, uint64_t const *buf, } } +/* + * Issue a batch alloc operation on a sequence of cache lines, wait for the + * batch alloc to complete and copy the pointers out into the user buffer. + */ static inline unsigned int roc_npa_aura_op_batch_alloc(uint64_t aura_handle, uint64_t *buf, - uint64_t *aligned_buf, unsigned int num, - const int dis_wait, const int drop, - const int partial) + unsigned int num, uint64_t *aligned_buf, + unsigned int aligned_buf_sz, const int dis_wait, + const int drop, const int partial) { unsigned int count, chunk, num_alloc; @@ -344,9 +397,12 @@ roc_npa_aura_op_batch_alloc(uint64_t aura_handle, uint64_t *buf, count = 0; while (num) { - chunk = (num > ROC_CN10K_NPA_BATCH_ALLOC_MAX_PTRS) ? - ROC_CN10K_NPA_BATCH_ALLOC_MAX_PTRS : - num; + /* Make sure that the pointers allocated fit into the cache +* lines reserved. +*/ + chunk = aligned_buf_sz / sizeof(uint64_
RE: [PATCH 2/4] app/test: refactor mod exp test case
> -Original Message- > From: Kusztal, ArkadiuszX > Sent: Wednesday 17 May 2023 22:19 > To: dev@dpdk.org > Cc: gak...@marvell.com; Ji, Kai ; Power, Ciara > ; Kusztal, ArkadiuszX > Subject: [PATCH 2/4] app/test: refactor mod exp test case > > Refactored modular exponentiation test case. > Added multiple vectors to be checked in KAT tests. > > Signed-off-by: Arek Kusztal > --- Acked-by: Ciara Power
Re: [PATCH v3] eal: fix eal init may failed when too much continuous memsegs under legacy mode
On 5/26/2023 4:41 AM, Fengnan Chang wrote: Under legacy mode, if the number of continuous memsegs greater than RTE_MAX_MEMSEG_PER_LIST, eal init will failed even though another memseg list is empty, because only one memseg list used to check in remap_needed_hugepages. Fix this by make remap_segment return how many segments mapped, remap_segment try to map most contiguous segments it can, if exceed it's capbility, remap_needed_hugepages will continue to map other left pages. For example: hugepage configure: cat /sys/devices/system/node/node*/hugepages/hugepages-2048kB/nr_hugepages 10241 10239 startup log: EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 EAL: Detected memory type: socket_id:1 hugepage_sz:2097152 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 EAL: Creating 4 segment lists: n_segs:8192 socket_id:1 hugepage_sz:2097152 EAL: Requesting 13370 pages of size 2MB from socket 0 EAL: Requesting 7110 pages of size 2MB from socket 1 EAL: Attempting to map 14220M on socket 1 EAL: Allocated 14220M on socket 1 EAL: Attempting to map 26740M on socket 0 EAL: Could not find space for memseg. Please increase 32768 and/or 65536 in configuration. EAL: Couldn't remap hugepage files into memseg lists EAL: FATAL: Cannot init memory EAL: Cannot init memory Signed-off-by: Fengnan Chang Signed-off-by: Lin Li Signed-off-by: Burakov Anatoly Hi, Thanks for taking my suggested implementation on board! --- lib/eal/linux/eal_memory.c | 55 +- 1 file changed, 36 insertions(+), 19 deletions(-) diff --git a/lib/eal/linux/eal_memory.c b/lib/eal/linux/eal_memory.c index 60fc8cc6ca..085defdee5 100644 --- a/lib/eal/linux/eal_memory.c +++ b/lib/eal/linux/eal_memory.c @@ -681,6 +681,7 @@ remap_segment(struct hugepage_file *hugepages, int seg_start, int seg_end) /* find free space in memseg lists */ for (msl_idx = 0; msl_idx < RTE_MAX_MEMSEG_LISTS; msl_idx++) { + int free_len; bool empty; msl = &mcfg->memsegs[msl_idx]; arr = &msl->memseg_arr; @@ -692,24 +693,28 @@ remap_segment(struct hugepage_file *hugepages, int seg_start, int seg_end) /* leave space for a hole if array is not empty */ empty = arr->count == 0; - ms_idx = rte_fbarray_find_next_n_free(arr, 0, - seg_len + (empty ? 0 : 1)); - - /* memseg list is full? */ - if (ms_idx < 0) + /* find start of the biggest contiguous block and its size */ + ms_idx = rte_fbarray_find_biggest_free(arr, 0); + free_len = rte_fbarray_find_contig_free(arr, ms_idx); + if (free_len < 0) continue; Technically, ms_idx can return -1, which should not be passed to rte_fbarray_find_contig_free() because the index value it accepts is unsigned (meaning, -1 will translate to UINT32_MAX). This *would* cause failure at parameter checking, so in practice it's not a bug, but I'm pretty sure code analyzers will complain about it, so the control flow needs to be changed somewhat. Specifically, we should check for `ms_idx < 0` and continue, before using it in `rte_fbarray_find_contig_free()`. - /* leave some space between memsegs, they are not IOVA * contiguous, so they shouldn't be VA contiguous either. */ - if (!empty) + if (!empty) { ms_idx++; + free_len--; + } + printf("free len %d seg_len %d ms_idx %d\n", free_len, seg_len, ms_idx); Debugging leftover? + /* we might not get all of the space we wanted */ + free_len = RTE_MIN(seg_len, free_len); + seg_end = seg_start + free_len; + seg_len = seg_end - seg_start; break; } if (msl_idx == RTE_MAX_MEMSEG_LISTS) { - RTE_LOG(ERR, EAL, "Could not find space for memseg. Please increase %s and/or %s in configuration.\n", - RTE_STR(RTE_MAX_MEMSEG_PER_TYPE), - RTE_STR(RTE_MAX_MEM_MB_PER_TYPE)); + RTE_LOG(ERR, EAL, "Could not find space for memseg. Please increase RTE_MAX_MEMSEG_PER_LIST " + "RTE_MAX_MEMSEG_PER_TYPE and/or RTE_MAX_MEM_MB_PER_TYPE in configuration.\n"); I don't think this change should be part of this patch, as this is fixing a separate issue. If we are to fix it, it should be a separate patch. With the above changes, Reviewed-by: Anatoly Burakov -- Thanks, Anatoly
Re: [PATCH v8] enhance NUMA affinity heuristic
On 5/26/2023 9:45 AM, Kaisen You wrote: When a DPDK application is started on only one numa node, memory is allocated for only one socket. When interrupt threads use memory, memory may not be found on the socket where the interrupt thread is currently located, and memory has to be reallocated on the hugepage, this operation will lead to performance degradation. Fixes: 705356f0811f ("eal: simplify control thread creation") Fixes: 770d41bf3309 ("malloc: fix allocation with unknown socket ID") Cc: sta...@dpdk.org Signed-off-by: Kaisen You --- Changes since v7: - Update commet, Changes since v6: - New explanation for easy understanding, Changes since v5: - Add comments to the code, Changes since v4: - mod the patch title, Changes since v3: - add the assignment of socket_id in thread initialization, Changes since v2: - add uncommitted local change and fix compilation, Changes since v1: - accomodate for configurations with main lcore running on multiples physical cores belonging to different numa, --- lib/eal/common/eal_common_thread.c | 4 lib/eal/common/malloc_heap.c | 11 ++- 2 files changed, 14 insertions(+), 1 deletion(-) diff --git a/lib/eal/common/eal_common_thread.c b/lib/eal/common/eal_common_thread.c index 079a385630..22480aa61f 100644 --- a/lib/eal/common/eal_common_thread.c +++ b/lib/eal/common/eal_common_thread.c @@ -252,6 +252,10 @@ static int ctrl_thread_init(void *arg) struct rte_thread_ctrl_params *params = arg; __rte_thread_init(rte_lcore_id(), cpuset); + /* Set control thread socket ID to SOCKET_ID_ANY as control +* threads may be scheduled on any NUMA node. +*/ + RTE_PER_LCORE(_socket_id) = SOCKET_ID_ANY; params->ret = rte_thread_set_affinity_by_id(rte_thread_self(), cpuset); if (params->ret != 0) { __atomic_store_n(¶ms->ctrl_thread_status, diff --git a/lib/eal/common/malloc_heap.c b/lib/eal/common/malloc_heap.c index d25bdc98f9..d833a71e7a 100644 --- a/lib/eal/common/malloc_heap.c +++ b/lib/eal/common/malloc_heap.c @@ -716,7 +716,16 @@ malloc_get_numa_socket(void) if (conf->socket_mem[socket_id] != 0) return socket_id; } - + /* We couldn't find quickly find a NUMA node where memory was available, typo: `find quickly find`, should probably be `quickly find` Can be fixed on apply. Reviewed-by: Anatoly Burakov +* so fall back to using main lcore socket ID. +*/ + socket_id = rte_lcore_to_socket_id(rte_get_main_lcore()); + /* Main lcore socket ID may be SOCKET_ID_ANY in cases when main lcore +* thread is affinitized to multiple NUMA nodes. +*/ + if (socket_id != (unsigned int)SOCKET_ID_ANY) + return socket_id; + /* Failed to find meaningful socket ID, so just use the first one available */ return rte_socket_id_by_idx(0); } -- Thanks, Anatoly
RE: [PATCH 3/4] app/test: refactor mod inv tests
> -Original Message- > From: Kusztal, ArkadiuszX > Sent: Wednesday 17 May 2023 22:19 > To: dev@dpdk.org > Cc: gak...@marvell.com; Ji, Kai ; Power, Ciara > ; Kusztal, ArkadiuszX > Subject: [PATCH 3/4] app/test: refactor mod inv tests > > Added new modular multiplicative inverse function. > Now it handles changes to the generic setup. > > Signed-off-by: Arek Kusztal > --- Acked-by: Ciara Power
Re: [PATCH v3] eal: fix eal init may failed when too much continuous memsegs under legacy mode
On 5/26/2023 3:41 PM, Burakov, Anatoly wrote: On 5/26/2023 4:41 AM, Fengnan Chang wrote: Under legacy mode, if the number of continuous memsegs greater than RTE_MAX_MEMSEG_PER_LIST, eal init will failed even though another memseg list is empty, because only one memseg list used to check in remap_needed_hugepages. Fix this by make remap_segment return how many segments mapped, remap_segment try to map most contiguous segments it can, if exceed it's capbility, remap_needed_hugepages will continue to map other left pages. For example: hugepage configure: cat /sys/devices/system/node/node*/hugepages/hugepages-2048kB/nr_hugepages 10241 10239 startup log: EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 EAL: Detected memory type: socket_id:1 hugepage_sz:2097152 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 EAL: Creating 4 segment lists: n_segs:8192 socket_id:1 hugepage_sz:2097152 EAL: Requesting 13370 pages of size 2MB from socket 0 EAL: Requesting 7110 pages of size 2MB from socket 1 EAL: Attempting to map 14220M on socket 1 EAL: Allocated 14220M on socket 1 EAL: Attempting to map 26740M on socket 0 EAL: Could not find space for memseg. Please increase 32768 and/or 65536 in configuration. EAL: Couldn't remap hugepage files into memseg lists EAL: FATAL: Cannot init memory EAL: Cannot init memory Signed-off-by: Fengnan Chang Signed-off-by: Lin Li Signed-off-by: Burakov Anatoly Hi, Thanks for taking my suggested implementation on board! --- lib/eal/linux/eal_memory.c | 55 +- 1 file changed, 36 insertions(+), 19 deletions(-) diff --git a/lib/eal/linux/eal_memory.c b/lib/eal/linux/eal_memory.c index 60fc8cc6ca..085defdee5 100644 --- a/lib/eal/linux/eal_memory.c +++ b/lib/eal/linux/eal_memory.c @@ -681,6 +681,7 @@ remap_segment(struct hugepage_file *hugepages, int seg_start, int seg_end) /* find free space in memseg lists */ for (msl_idx = 0; msl_idx < RTE_MAX_MEMSEG_LISTS; msl_idx++) { + int free_len; bool empty; msl = &mcfg->memsegs[msl_idx]; arr = &msl->memseg_arr; @@ -692,24 +693,28 @@ remap_segment(struct hugepage_file *hugepages, int seg_start, int seg_end) /* leave space for a hole if array is not empty */ empty = arr->count == 0; - ms_idx = rte_fbarray_find_next_n_free(arr, 0, - seg_len + (empty ? 0 : 1)); - - /* memseg list is full? */ - if (ms_idx < 0) + /* find start of the biggest contiguous block and its size */ + ms_idx = rte_fbarray_find_biggest_free(arr, 0); + free_len = rte_fbarray_find_contig_free(arr, ms_idx); + if (free_len < 0) continue; Missed this. When we're splitting up segments, we're looking for contiguous free areas that are at least two memsegs long, so if this memseg is full but contains segments that were split up, there will be a bunch of holes in it, and free_len will be 1 (because every hole will be 1 segment long by definition). So, we should not accept anything that is at least two segments long. -- Thanks, Anatoly
Re: [PATCH v3] app/testpmd: expand noisy neighbour forward mode support
On 17/04/2023 19:55, Mike Pattrick wrote: Previously the noisy neighbour vnf simulation would only operate in io mode, forwarding packets as is. However, this limited the usefulness of noisy neighbour simulation. This feature has now been expanded to supporting mac, macswap, and 5tswap modes. To facilitate adding this support, some new header files were added. Hi Mike, I think it makes sense to allow these functionalities to be combined. It does seem like that noisy neighbour shouldn't have been added as a standalone fowarding option in the first place, as it's not mutually exclusive with the other fwding modes. Now it's forcing a user to set a fwding mode within a fwding mode to combine the functionality :/ The approach is to allow the fwding modes to expose their forwarding method so noisy can call them. Did you consider to expose the noisy functionality and allow the other forwarding modes to call that when the --noisy-* flags are set? For example, a user sets fwding mode as normal 'fwd=macfwd' but then if the user adds the --noisy-* params, the macfwd'ing mode calls some functions exposed to add the noise, which seems a bit more intuitive. Also, if someone adds a new fwding mode, they can just call the noisy functions and don't have update noisy fwding mode to add the new mode. I'm not sure if it's better, just throwing it out there as an alternative idea. On the downside, it would break backwards compatibility because previously those --noisy-* params would have had no effect with say macfwd mode, but now they will. So maybe that's enough to prohibit it. In the past, I would have had all the params set and just changed fwdmode to enable/disable noisy vnf. That would behaviour would now be changed with this approach. What do you think? Few comments on the code below. thanks, Kevin. Signed-off-by: Mike Pattrick --- v2: Reverted changes to random memory lookup v3: Refactored entire patch --- app/test-pmd/5tswap.c | 118 +-- app/test-pmd/5tswap.h | 130 ++ app/test-pmd/macfwd.c | 33 +-- app/test-pmd/macfwd.h | 46 + app/test-pmd/noisy_vnf.c | 92 +++--- app/test-pmd/parameters.c | 17 app/test-pmd/testpmd.c| 5 + app/test-pmd/testpmd.h| 8 ++ doc/guides/testpmd_app_ug/run_app.rst | 9 ++ 9 files changed, 299 insertions(+), 159 deletions(-) create mode 100644 app/test-pmd/5tswap.h create mode 100644 app/test-pmd/macfwd.h diff --git a/app/test-pmd/5tswap.c b/app/test-pmd/5tswap.c index ff8c2dcde5..8e8de2557a 100644 --- a/app/test-pmd/5tswap.c +++ b/app/test-pmd/5tswap.c @@ -17,64 +17,8 @@ #include #include -#include "macswap_common.h" #include "testpmd.h" - - -static inline void -swap_mac(struct rte_ether_hdr *eth_hdr) -{ - struct rte_ether_addr addr; - - /* Swap dest and src mac addresses. */ - rte_ether_addr_copy(ð_hdr->dst_addr, &addr); - rte_ether_addr_copy(ð_hdr->src_addr, ð_hdr->dst_addr); - rte_ether_addr_copy(&addr, ð_hdr->src_addr); -} - -static inline void -swap_ipv4(struct rte_ipv4_hdr *ipv4_hdr) -{ - rte_be32_t addr; - - /* Swap dest and src ipv4 addresses. */ - addr = ipv4_hdr->src_addr; - ipv4_hdr->src_addr = ipv4_hdr->dst_addr; - ipv4_hdr->dst_addr = addr; -} - -static inline void -swap_ipv6(struct rte_ipv6_hdr *ipv6_hdr) -{ - uint8_t addr[16]; - - /* Swap dest and src ipv6 addresses. */ - memcpy(&addr, &ipv6_hdr->src_addr, 16); - memcpy(&ipv6_hdr->src_addr, &ipv6_hdr->dst_addr, 16); - memcpy(&ipv6_hdr->dst_addr, &addr, 16); -} - -static inline void -swap_tcp(struct rte_tcp_hdr *tcp_hdr) -{ - rte_be16_t port; - - /* Swap dest and src tcp port. */ - port = tcp_hdr->src_port; - tcp_hdr->src_port = tcp_hdr->dst_port; - tcp_hdr->dst_port = port; -} - -static inline void -swap_udp(struct rte_udp_hdr *udp_hdr) -{ - rte_be16_t port; - - /* Swap dest and src udp port */ - port = udp_hdr->src_port; - udp_hdr->src_port = udp_hdr->dst_port; - udp_hdr->dst_port = port; -} +#include "5tswap.h" /* * 5 tuple swap forwarding mode: Swap the source and the destination of layers @@ -85,22 +29,7 @@ static bool pkt_burst_5tuple_swap(struct fwd_stream *fs) { struct rte_mbuf *pkts_burst[MAX_PKT_BURST]; - struct rte_port *txp; - struct rte_mbuf *mb; - uint16_t next_proto; - uint64_t ol_flags; - uint16_t proto; uint16_t nb_rx; - int i; - union { - struct rte_ether_hdr *eth; - struct rte_vlan_hdr *vlan; - struct rte_ipv4_hdr *ipv4; - struct rte_ipv6_hdr *ipv6; - struct rte_tcp_hdr *tcp; - struct rte_udp_hdr *udp; - uint8_t *byte; -
Re: [PATCH v1] power: support amd-pstate cpufreq driver
On 5/24/2023 7:46 PM, Thomas Monjalon wrote: 12/04/2023 11:52, Sivaprasad Tummala: amd-pstate introduces a new CPU frequency control mechanism for AMD processors using the ACPI Collaborative Performance Power Control feature for a finer grained frequency management. Patch to add support for amd-pstate driver. Signed-off-by: Sivaprasad Tummala --- app/test/test_power.c | 1 + app/test/test_power_cpufreq.c | 5 +- doc/guides/rel_notes/release_23_07.rst | 3 + examples/l3fwd-power/main.c| 1 + lib/power/meson.build | 1 + lib/power/power_amd_pstate_cpufreq.c | 698 + lib/power/power_amd_pstate_cpufreq.h | 219 lib/power/rte_power.c | 26 + lib/power/rte_power.h | 3 +- lib/power/rte_power_pmd_mgmt.c | 6 +- 10 files changed, 958 insertions(+), 5 deletions(-) I'm not comfortable to merge this patch without a word from David Hunt. Given there is 0 review, what do we do? Also, make sure to start the actual text at the margin. === + * **Added amd-pstate driver support to power management library.** + +Added support for amd-pstate driver which works on AMD Zen processors. Looks like the indent is not correct. 'power_pstate_cpufreq.c', +'power_amd_pstate_cpufreq.c', Can you say briefly why AMD has a different pstate? Does it mean power_pstate_cpufreq.c should be renamed power_intel_pstate_cpufreq.c? +++ b/lib/power/power_amd_pstate_cpufreq.c @@ -0,0 +1,698 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2010-2021 Intel Corporation + * Copyright(c) 2021 Arm Limited Why is there copyright for Intel and Arm? Does it mean you copied some code and did not try to keep common code in a common place? Yes, most of it appears to be copied from Intel pstate code, so the copyright would be correct here. I don't know if we can merge these into one, to be honest - while a lot of it is shared (like, a lot), there are some subtle differences in code that might prevent us from sharing the code paths - joys of C. -- Thanks, Anatoly
Re: [PATCH v1] power: support amd-pstate cpufreq driver
On 4/12/2023 10:52 AM, Sivaprasad Tummala wrote: amd-pstate introduces a new CPU frequency control mechanism for AMD processors using the ACPI Collaborative Performance Power Control feature for a finer grained frequency management. Patch to add support for amd-pstate driver. Signed-off-by: Sivaprasad Tummala --- I don't have an AMD machine handy so I can't test this, but code wise, LGTM bar one nitpick below. Acked-by: Anatoly Burakov + } + + ret = read_core_sysfs_u32(f_nom, &nominal_freq); + if (ret < 0) { + RTE_LOG(ERR, POWER, "Failed to read %s\n", + POWER_SYSFILE_NOMINAL_FREQ); + goto out; + } + + power_check_turbo(pi); + + if (scaling_max_freq < scaling_min_freq) It would be nice if this produced an error message of some kind. -- Thanks, Anatoly
Re: [PATCH v3] app/testpmd: expand noisy neighbour forward mode support
On Fri, May 26, 2023 at 11:34 AM Kevin Traynor wrote: > > On 17/04/2023 19:55, Mike Pattrick wrote: > > Previously the noisy neighbour vnf simulation would only operate in io > > mode, forwarding packets as is. However, this limited the usefulness of > > noisy neighbour simulation. > > > > This feature has now been expanded to supporting mac, macswap, and > > 5tswap modes. To facilitate adding this support, some new header files > > were added. > > > > Hi Mike, > > I think it makes sense to allow these functionalities to be combined. It > does seem like that noisy neighbour shouldn't have been added as a > standalone fowarding option in the first place, as it's not mutually > exclusive with the other fwding modes. Now it's forcing a user to set a > fwding mode within a fwding mode to combine the functionality :/ > > The approach is to allow the fwding modes to expose their forwarding > method so noisy can call them. Did you consider to expose the noisy > functionality and allow the other forwarding modes to call that when the > --noisy-* flags are set? Personally I like this approach more, and that was the strategy I took with the previous version. However, that change didn't seem popular at the time. http://patches.dpdk.org/project/dpdk/patch/20230126045516.176917-1-...@redhat.com/#157095 > For example, a user sets fwding mode as normal 'fwd=macfwd' but then if > the user adds the --noisy-* params, the macfwd'ing mode calls some > functions exposed to add the noise, which seems a bit more intuitive. > > Also, if someone adds a new fwding mode, they can just call the noisy > functions and don't have update noisy fwding mode to add the new mode. > > I'm not sure if it's better, just throwing it out there as an > alternative idea. > > On the downside, it would break backwards compatibility because > previously those --noisy-* params would have had no effect with say > macfwd mode, but now they will. So maybe that's enough to prohibit it. > > In the past, I would have had all the params set and just changed > fwdmode to enable/disable noisy vnf. That would behaviour would now be > changed with this approach. > > What do you think? > > Few comments on the code below. > > thanks, > Kevin. > > > Signed-off-by: Mike Pattrick > > --- > > v2: Reverted changes to random memory lookup > > v3: Refactored entire patch > > --- > > app/test-pmd/5tswap.c | 118 +-- > > app/test-pmd/5tswap.h | 130 ++ > > app/test-pmd/macfwd.c | 33 +-- > > app/test-pmd/macfwd.h | 46 + > > app/test-pmd/noisy_vnf.c | 92 +++--- > > app/test-pmd/parameters.c | 17 > > app/test-pmd/testpmd.c| 5 + > > app/test-pmd/testpmd.h| 8 ++ > > doc/guides/testpmd_app_ug/run_app.rst | 9 ++ > > 9 files changed, 299 insertions(+), 159 deletions(-) > > create mode 100644 app/test-pmd/5tswap.h > > create mode 100644 app/test-pmd/macfwd.h > > > > diff --git a/app/test-pmd/5tswap.c b/app/test-pmd/5tswap.c > > index ff8c2dcde5..8e8de2557a 100644 > > --- a/app/test-pmd/5tswap.c > > +++ b/app/test-pmd/5tswap.c > > @@ -17,64 +17,8 @@ > > #include > > #include > > > > -#include "macswap_common.h" > > #include "testpmd.h" > > - > > - > > -static inline void > > -swap_mac(struct rte_ether_hdr *eth_hdr) > > -{ > > - struct rte_ether_addr addr; > > - > > - /* Swap dest and src mac addresses. */ > > - rte_ether_addr_copy(ð_hdr->dst_addr, &addr); > > - rte_ether_addr_copy(ð_hdr->src_addr, ð_hdr->dst_addr); > > - rte_ether_addr_copy(&addr, ð_hdr->src_addr); > > -} > > - > > -static inline void > > -swap_ipv4(struct rte_ipv4_hdr *ipv4_hdr) > > -{ > > - rte_be32_t addr; > > - > > - /* Swap dest and src ipv4 addresses. */ > > - addr = ipv4_hdr->src_addr; > > - ipv4_hdr->src_addr = ipv4_hdr->dst_addr; > > - ipv4_hdr->dst_addr = addr; > > -} > > - > > -static inline void > > -swap_ipv6(struct rte_ipv6_hdr *ipv6_hdr) > > -{ > > - uint8_t addr[16]; > > - > > - /* Swap dest and src ipv6 addresses. */ > > - memcpy(&addr, &ipv6_hdr->src_addr, 16); > > - memcpy(&ipv6_hdr->src_addr, &ipv6_hdr->dst_addr, 16); > > - memcpy(&ipv6_hdr->dst_addr, &addr, 16); > > -} > > - > > -static inline void > > -swap_tcp(struct rte_tcp_hdr *tcp_hdr) > > -{ > > - rte_be16_t port; > > - > > - /* Swap dest and src tcp port. */ > > - port = tcp_hdr->src_port; > > - tcp_hdr->src_port = tcp_hdr->dst_port; > > - tcp_hdr->dst_port = port; > > -} > > - > > -static inline void > > -swap_udp(struct rte_udp_hdr *udp_hdr) > > -{ > > - rte_be16_t port; > > - > > - /* Swap dest and src udp port */ > > - port = udp_hdr->src_port; > > - udp_hdr->src_port = udp_hdr->dst_port; > > - udp_hdr->dst_port = port; > > -} > > +#include "5tswap.h" > > > > /* > >* 5 tuple swap forwarding mode:
[PATCH v4 00/22] lib: add pdcp protocol
Add Packet Data Convergence Protocol (PDCP) processing library. The library is similar to lib_ipsec which provides IPsec processing capabilities in DPDK. PDCP would involve roughly the following operations, 1. Transfer of user plane data 2. Transfer of control plane data 3. Header compression 4. Uplink data compression 5. Ciphering and integrity protection PDCP library provides following control path APIs that is used to configure various PDCP entities, 1. rte_pdcp_entity_establish() 2. rte_pdcp_entity_suspend() 3. rte_pdcp_entity_release() PDCP process is split into 2 parts. One before crypto processing (rte_pdcp_pkt_pre_process()) and one after crypto processing (rte_pdcp_pkt_post_process()). Since cryptodev dequeue can return crypto operations belonging to multiple entities, rte_pdcp_pkt_crypto_group() is added to help grouping crypto operations belonging to same entity. Similar to lib IPsec, lib PDCP would allow application to use same API sequence while leveraging protocol offload features enabled by rte_security library. Lib PDCP would internally change the handles registered for *pre_process* and *post_process* based on features enabled in the entity. Lib PDCP would create the required sessions on the device provided in entity to minimize the application requirements. Also, the crypto_op allocation and free would also be done internally by lib PDCP to allow the library to create crypto ops as required for the input packets. For example, when control PDUs are received, no cryptodev enqueue-dequeue is expected for the same and lib PDCP is expected to handle it differently. Lib PDCP utilizes reorder library for implementing in-order delivery. It utilizes bitmap library for implementing status reports and track the COUNT value of the packets received. To allow application to choose timer implementation of choice, lib PDCP allows application to configure handles that can be used for starting & stopping timers. Upon expiry, application can call corresponding PDCP API(``rte_pdcp_t_reordering_expiry_handle``) for handling the event. Unit tests are added to verify both rte_timer based timers as well as rte_eventdev based timers. PDCP tracks the sequence number of the received packets and during events such as re-establishment, it is required to generate reports and transmit to the peer. This series introduces ``rte_pdcp_control_pdu_create`` for handling control PDU generation. Changes in v4: - Disabled 'annotate locks' with lib PDCP - Enable PDCP autotest only when lib is enabled - Use rwlock instead of spinlock - Avoid per packet checks for thread safety (Stephen) - In DL path, save count determined during pre-process in mbuf and use the same in post-process. Determining count again may lead To errors - Simplified DL path threads to allow more common code between SN 12 & 18 Changes in v3: - Addressed review comments (Akhil) - Addressed build failure in CI (tests with lib eventdev disabled) - Addressed checkpatch issues - Set only positive values to rte_errno (Akhil) Changes in v2: - Added control PDU handling - Added t-Reordering timer - Added in-order delivery - Added status PDU generation - Rebased on top of new features added in reorder library - Split base patch - Increased test coverage - Improved thread safety Changes from RFC - Implementation for all APIs covering basic control plane & user plane packets - Unit test leveraging existing PDCP vectors available in test_cryptodev - Unit test performing both UL & DL operations to verify various protocol features - Updated documentation Sample application sequence: struct rte_mbuf **out_mb, *pkts[MAX_BURST_SIZE]; struct rte_crypto_op *cop[MAX_BURST_SIZE]; struct rte_pdcp_group grp[MAX_BURST_SIZE]; struct rte_pdcp_entity *pdcp_entity; int nb_max_out_mb, ret, nb_grp; /* Create PDCP entity */ pdcp_entity = rte_pdcp_entity_establish(&conf); /** * Allocate buffer for holding mbufs returned during PDCP suspend, * release & post-process APIs. */ /* Max packets that can be cached in entity + burst size */ nb_max_out_mb = pdcp_entity->max_pkt_cache + 1; out_mb = rte_malloc(NULL, nb_max_out_mb * sizeof(uintptr_t), 0); if (out_mb == NULL) { /* Handle error */ } while (1) { /* Receive packet and form mbuf */ /** * Prepare packets for crypto operation. Following operations * would be done, * * Transmitting entity/UL (only data PDUs): * - Perform compression * - Assign sequence number * - Add PDCP header * - Create & prepare crypto_op * - Prepare IV for crypto operation (auth_gen, encrypt) * - Save original PDCP SDU (during PDCP re-establishment, *unconfirmed PDCP SD
[PATCH v4 01/22] net: add PDCP header
From: Volodymyr Fialko Add PDCP protocol header to be used for supporting PDCP protocol processing. Signed-off-by: Anoob Joseph Signed-off-by: Kiran Kumar K Signed-off-by: Volodymyr Fialko Acked-by: Akhil Goyal --- doc/api/doxy-api-index.md | 3 +- lib/net/meson.build | 1 + lib/net/rte_pdcp_hdr.h| 147 ++ 3 files changed, 150 insertions(+), 1 deletion(-) create mode 100644 lib/net/rte_pdcp_hdr.h diff --git a/doc/api/doxy-api-index.md b/doc/api/doxy-api-index.md index c709fd48ad..debbe4134f 100644 --- a/doc/api/doxy-api-index.md +++ b/doc/api/doxy-api-index.md @@ -127,7 +127,8 @@ The public API headers are grouped by topics: [Geneve](@ref rte_geneve.h), [eCPRI](@ref rte_ecpri.h), [L2TPv2](@ref rte_l2tpv2.h), - [PPP](@ref rte_ppp.h) + [PPP](@ref rte_ppp.h), + [PDCP hdr](@ref rte_pdcp_hdr.h) - **QoS**: [metering](@ref rte_meter.h), diff --git a/lib/net/meson.build b/lib/net/meson.build index 379d161ee0..bd56f91c22 100644 --- a/lib/net/meson.build +++ b/lib/net/meson.build @@ -22,6 +22,7 @@ headers = files( 'rte_geneve.h', 'rte_l2tpv2.h', 'rte_ppp.h', +'rte_pdcp_hdr.h', ) sources = files( diff --git a/lib/net/rte_pdcp_hdr.h b/lib/net/rte_pdcp_hdr.h new file mode 100644 index 00..72ae9a66cb --- /dev/null +++ b/lib/net/rte_pdcp_hdr.h @@ -0,0 +1,147 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(C) 2023 Marvell. + */ + +#ifndef RTE_PDCP_HDR_H +#define RTE_PDCP_HDR_H + +/** + * @file + * + * PDCP-related defines + * + * Based on - ETSI TS 138 323 V17.1.0 (2022-08) + * https://www.etsi.org/deliver/etsi_ts/138300_138399/138323/17.01.00_60/ts_138323v170100p.pdf + */ + +#include + +#ifdef __cplusplus +extern "C" { +#endif + +/** + * 4.3.1 + * + * Indicate the maximum supported size of a PDCP Control PDU. + */ +#define RTE_PDCP_CTRL_PDU_SIZE_MAX 9000u + +/** + * 6.3.4 MAC-I + * + * Indicate the size of MAC-I in PDCP PDU. + */ +#define RTE_PDCP_MAC_I_LEN 4 + +/** + * Indicate type of control information included in the corresponding PDCP + * Control PDU. + */ +enum rte_pdcp_ctrl_pdu_type { + RTE_PDCP_CTRL_PDU_TYPE_STATUS_REPORT = 0, + RTE_PDCP_CTRL_PDU_TYPE_ROHC_FEEDBACK = 1, + RTE_PDCP_CTRL_PDU_TYPE_EHC_FEEDBACK = 2, + RTE_PDCP_CRTL_PDU_TYPE_UDC_FEEDBACK = 3, +}; + +/** + * 6.3.7 D/C + * + * This field indicates whether the corresponding PDCP PDU is a + * PDCP Data PDU or a PDCP Control PDU. + */ +enum rte_pdcp_pdu_type { + RTE_PDCP_PDU_TYPE_CTRL = 0, + RTE_PDCP_PDU_TYPE_DATA = 1, +}; + +/** + * 6.2.2.1 Data PDU for SRBs + */ +__extension__ +struct rte_pdcp_cp_data_pdu_sn_12_hdr { +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN + uint8_t sn_11_8 : 4;/**< Sequence number bits 8-11 */ + uint8_t r : 4; /**< Reserved */ +#elif RTE_BYTE_ORDER == RTE_BIG_ENDIAN + uint8_t r : 4; /**< Reserved */ + uint8_t sn_11_8 : 4;/**< Sequence number bits 8-11 */ +#endif + uint8_t sn_7_0; /**< Sequence number bits 0-7 */ +} __rte_packed; + +/** + * 6.2.2.2 Data PDU for DRBs and MRBs with 12 bits PDCP SN + */ +__extension__ +struct rte_pdcp_up_data_pdu_sn_12_hdr { +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN + uint8_t sn_11_8 : 4;/**< Sequence number bits 8-11 */ + uint8_t r : 3; /**< Reserved */ + uint8_t d_c : 1;/**< D/C bit */ +#elif RTE_BYTE_ORDER == RTE_BIG_ENDIAN + uint8_t d_c : 1;/**< D/C bit */ + uint8_t r : 3; /**< Reserved */ + uint8_t sn_11_8 : 4;/**< Sequence number bits 8-11 */ +#endif + uint8_t sn_7_0; /**< Sequence number bits 0-7 */ +} __rte_packed; + +/** + * 6.2.2.3 Data PDU for DRBs and MRBs with 18 bits PDCP SN + */ +__extension__ +struct rte_pdcp_up_data_pdu_sn_18_hdr { +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN + uint8_t sn_17_16 : 2; /**< Sequence number bits 16-17 */ + uint8_t r : 5; /**< Reserved */ + uint8_t d_c : 1;/**< D/C bit */ +#elif RTE_BYTE_ORDER == RTE_BIG_ENDIAN + uint8_t d_c : 1;/**< D/C bit */ + uint8_t r : 5; /**< Reserved */ + uint8_t sn_17_16 : 2; /**< Sequence number bits 16-17 */ +#endif + uint8_t sn_15_8;/**< Sequence number bits 8-15 */ + uint8_t sn_7_0; /**< Sequence number bits 0-7 */ +} __rte_packed; + +/** + * 6.2.3.1 Control PDU for PDCP status report + */ +__extension__ +struct rte_pdcp_up_ctrl_pdu_hdr { +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN + uint8_t r : 4; /**< Reserved */ + uint8_t pdu_type : 3; /**< Control PDU type */ + uint8_t d_c : 1;/**< D/C bit */ +#elif RTE_BYTE_ORDER == RTE_BIG_ENDIAN + uint8_t d_c : 1;/**< D/C bit */ + uint8_t pdu_type : 3; /**< Control PDU type */ + uint8_t r : 4; /**< Reserved */ +#endif + /** +* 6.3.9 FMC +* +* First Missing COUNT. This fiel
[PATCH v4 02/22] lib: add pdcp protocol
Add Packet Data Convergence Protocol (PDCP) processing library. The library is similar to lib_ipsec which provides IPsec processing capabilities in DPDK. PDCP would involve roughly the following options, 1. Transfer of user plane data 2. Transfer of control plane data 3. Header compression 4. Uplink data compression 5. Ciphering and integrity protection PDCP library provides following control path APIs that is used to configure various PDCP entities, 1. rte_pdcp_entity_establish() 2. rte_pdcp_entity_suspend() 3. rte_pdcp_entity_release() Signed-off-by: Anoob Joseph Signed-off-by: Kiran Kumar K Signed-off-by: Volodymyr Fialko --- doc/api/doxy-api-index.md | 3 +- doc/api/doxy-api.conf.in | 1 + lib/meson.build | 1 + lib/pdcp/meson.build | 17 lib/pdcp/pdcp_crypto.c| 21 + lib/pdcp/pdcp_crypto.h| 15 lib/pdcp/pdcp_entity.h| 113 ++ lib/pdcp/pdcp_process.c | 138 +++ lib/pdcp/pdcp_process.h | 13 +++ lib/pdcp/rte_pdcp.c | 141 lib/pdcp/rte_pdcp.h | 167 ++ lib/pdcp/version.map | 10 +++ 12 files changed, 639 insertions(+), 1 deletion(-) create mode 100644 lib/pdcp/meson.build create mode 100644 lib/pdcp/pdcp_crypto.c create mode 100644 lib/pdcp/pdcp_crypto.h create mode 100644 lib/pdcp/pdcp_entity.h create mode 100644 lib/pdcp/pdcp_process.c create mode 100644 lib/pdcp/pdcp_process.h create mode 100644 lib/pdcp/rte_pdcp.c create mode 100644 lib/pdcp/rte_pdcp.h create mode 100644 lib/pdcp/version.map diff --git a/doc/api/doxy-api-index.md b/doc/api/doxy-api-index.md index debbe4134f..cd7a6cae44 100644 --- a/doc/api/doxy-api-index.md +++ b/doc/api/doxy-api-index.md @@ -128,7 +128,8 @@ The public API headers are grouped by topics: [eCPRI](@ref rte_ecpri.h), [L2TPv2](@ref rte_l2tpv2.h), [PPP](@ref rte_ppp.h), - [PDCP hdr](@ref rte_pdcp_hdr.h) + [PDCP hdr](@ref rte_pdcp_hdr.h), + [PDCP](@ref rte_pdcp.h) - **QoS**: [metering](@ref rte_meter.h), diff --git a/doc/api/doxy-api.conf.in b/doc/api/doxy-api.conf.in index d230a19e1f..58789308a9 100644 --- a/doc/api/doxy-api.conf.in +++ b/doc/api/doxy-api.conf.in @@ -62,6 +62,7 @@ INPUT = @TOPDIR@/doc/api/doxy-api-index.md \ @TOPDIR@/lib/net \ @TOPDIR@/lib/pcapng \ @TOPDIR@/lib/pci \ + @TOPDIR@/lib/pdcp \ @TOPDIR@/lib/pdump \ @TOPDIR@/lib/pipeline \ @TOPDIR@/lib/port \ diff --git a/lib/meson.build b/lib/meson.build index dc8aa4ac84..a6a54c196c 100644 --- a/lib/meson.build +++ b/lib/meson.build @@ -64,6 +64,7 @@ libraries = [ 'flow_classify', # flow_classify lib depends on pkt framework table lib 'graph', 'node', +'pdcp', # pdcp lib depends on crypto and security ] optional_libs = [ diff --git a/lib/pdcp/meson.build b/lib/pdcp/meson.build new file mode 100644 index 00..ccaf426240 --- /dev/null +++ b/lib/pdcp/meson.build @@ -0,0 +1,17 @@ +# SPDX-License-Identifier: BSD-3-Clause +# Copyright(C) 2023 Marvell. + +if is_windows +build = false +reason = 'not supported on Windows' +subdir_done() +endif + +sources = files( +'pdcp_crypto.c', +'pdcp_process.c', +'rte_pdcp.c', +) +headers = files('rte_pdcp.h') + +deps += ['mbuf', 'net', 'cryptodev', 'security'] diff --git a/lib/pdcp/pdcp_crypto.c b/lib/pdcp/pdcp_crypto.c new file mode 100644 index 00..755e27ec9e --- /dev/null +++ b/lib/pdcp/pdcp_crypto.c @@ -0,0 +1,21 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(C) 2023 Marvell. + */ + +#include + +#include "pdcp_crypto.h" + +int +pdcp_crypto_sess_create(struct rte_pdcp_entity *entity, const struct rte_pdcp_entity_conf *conf) +{ + RTE_SET_USED(entity); + RTE_SET_USED(conf); + return 0; +} + +void +pdcp_crypto_sess_destroy(struct rte_pdcp_entity *entity) +{ + RTE_SET_USED(entity); +} diff --git a/lib/pdcp/pdcp_crypto.h b/lib/pdcp/pdcp_crypto.h new file mode 100644 index 00..6563331d37 --- /dev/null +++ b/lib/pdcp/pdcp_crypto.h @@ -0,0 +1,15 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(C) 2023 Marvell. + */ + +#ifndef PDCP_CRYPTO_H +#define PDCP_CRYPTO_H + +#include + +int pdcp_crypto_sess_create(struct rte_pdcp_entity *entity, + const struct rte_pdcp_entity_conf *conf); + +void pdcp_crypto_sess_destroy(struct rte_pdcp_entity *entity); + +#endif /* PDCP_CRYPTO_H */ diff --git a/lib/pdcp/pdcp_entity.h b/lib/pdcp/pdcp_entity.h new file mode 100644 index 00..000297588f --- /dev/null +++ b/lib/pdcp/pdcp_entity.h @@ -0,0 +1,113 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(C) 2023 Marvell. + */ + +#ifndef PDCP_ENTITY_H +#define PDCP_ENTITY_H + +#include +#in
[PATCH v4 03/22] pdcp: add pre and post-process
PDCP process is split into 2 parts. One before crypto processing (rte_pdcp_pkt_pre_process()) and one after crypto processing (rte_pdcp_pkt_post_process()). Functionality of pre-process & post-process varies based on the type of entity. Registration of entity specific function pointer allows skipping multiple checks that would come in datapath otherwise. Signed-off-by: Anoob Joseph Signed-off-by: Kiran Kumar K Signed-off-by: Volodymyr Fialko Acked-by: Akhil Goyal --- lib/pdcp/rte_pdcp.h | 97 lib/pdcp/version.map | 3 ++ 2 files changed, 100 insertions(+) diff --git a/lib/pdcp/rte_pdcp.h b/lib/pdcp/rte_pdcp.h index 1f96fdc9a1..46c3c2a416 100644 --- a/lib/pdcp/rte_pdcp.h +++ b/lib/pdcp/rte_pdcp.h @@ -22,6 +22,21 @@ extern "C" { #endif +/* Forward declarations */ +struct rte_pdcp_entity; + +/* PDCP pre-process function based on entity configuration */ +typedef uint16_t (*rte_pdcp_pre_p_t)(const struct rte_pdcp_entity *entity, +struct rte_mbuf *mb[], +struct rte_crypto_op *cop[], +uint16_t num, uint16_t *nb_err); + +/* PDCP post-process function based on entity configuration */ +typedef uint16_t (*rte_pdcp_post_p_t)(const struct rte_pdcp_entity *entity, + struct rte_mbuf *in_mb[], + struct rte_mbuf *out_mb[], + uint16_t num, uint16_t *nb_err); + /** * PDCP entity. * @@ -33,6 +48,10 @@ extern "C" { * plane depending on which radio bearer it is carrying data for. */ struct rte_pdcp_entity { + /** Entity specific pre-process handle. */ + rte_pdcp_pre_p_t pre_process; + /** Entity specific post-process handle. */ + rte_pdcp_post_p_t post_process; /** * PDCP entities may hold packets for purposes of in-order delivery (in * case of receiving PDCP entity) and re-transmission (in case of @@ -160,6 +179,84 @@ int rte_pdcp_entity_suspend(struct rte_pdcp_entity *pdcp_entity, struct rte_mbuf *out_mb[]); +/** + * @warning + * @b EXPERIMENTAL: this API may change without prior notice + * + * For input mbufs and given PDCP entity pre-process the mbufs and prepare + * crypto ops that can be enqueued to the cryptodev associated with given + * session. Only error packets would be moved returned in the input buffer, + * *mb*, and it is the responsibility of the application to free the same. + * + * @param entity + * Pointer to the *rte_pdcp_entity* object the packets belong to. + * @param[in, out] mb + * The address of an array of *num* pointers to *rte_mbuf* structures + * which contain the input packets. Any error packets would be returned in the + * same buffer. + * @param[out] cop + * The address of an array that can hold up to *num* pointers to + * *rte_crypto_op* structures. Crypto ops would be allocated by + * ``rte_pdcp_pkt_pre_process`` API. + * @param num + * The maximum number of packets to process. + * @param[out] nb_err + * Pointer to return the number of error packets returned in *mb* + * @return + * Count of crypto_ops prepared + */ +__rte_experimental +static inline uint16_t +rte_pdcp_pkt_pre_process(const struct rte_pdcp_entity *entity, +struct rte_mbuf *mb[], struct rte_crypto_op *cop[], +uint16_t num, uint16_t *nb_err) +{ + return entity->pre_process(entity, mb, cop, num, nb_err); +} + +/** + * @warning + * @b EXPERIMENTAL: this API may change without prior notice + * + * For input mbufs and given PDCP entity, perform PDCP post-processing of the + * mbufs. + * + * Input mbufs are the ones retrieved from crypto_ops dequeued from cryptodev + * and grouped by *rte_pdcp_pkt_crypto_group()*. + * + * The post-processed packets would be returned in the *out_mb* buffer. + * The resultant mbufs would be grouped into success packets and error packets. + * Error packets would be grouped in the end of the array and it is the + * responsibility of the application to handle the same. + * + * When in-order delivery is enabled, PDCP entity may buffer packets and would + * deliver packets only when all prior packets have been post-processed. That + * would result in returning more/less packets than enqueued. + * + * @param entity + * Pointer to the *rte_pdcp_entity* object the packets belong to. + * @param in_mb + * The address of an array of *num* pointers to *rte_mbuf* structures. + * @param[out] out_mb + * The address of an array of *num* pointers to *rte_mbuf* structures + * to output packets after PDCP post-processing. + * @param num + * The maximum number of packets to process. + * @param[out] nb_err + * The number of error packets returned in *out_mb* buffer. + * @return + * Count of packets returned in *out_mb* buffer. + */ +__rte_experimental +static inline uint16_t +rte_pdcp
[PATCH v4 04/22] pdcp: add packet group
Crypto processing in PDCP is performed asynchronously by rte_cryptodev_enqueue_burst() and rte_cryptodev_dequeue_burst(). Since cryptodev dequeue can return crypto operations belonging to multiple entities, rte_pdcp_pkt_crypto_group() is added to help grouping crypto operations belonging to same entity. Signed-off-by: Anoob Joseph Signed-off-by: Kiran Kumar K Signed-off-by: Volodymyr Fialko --- lib/pdcp/meson.build | 1 + lib/pdcp/rte_pdcp.h | 6 ++ lib/pdcp/rte_pdcp_group.h | 131 ++ lib/pdcp/version.map | 3 + 4 files changed, 141 insertions(+) create mode 100644 lib/pdcp/rte_pdcp_group.h diff --git a/lib/pdcp/meson.build b/lib/pdcp/meson.build index ccaf426240..08679b743a 100644 --- a/lib/pdcp/meson.build +++ b/lib/pdcp/meson.build @@ -13,5 +13,6 @@ sources = files( 'rte_pdcp.c', ) headers = files('rte_pdcp.h') +indirect_headers += files('rte_pdcp_group.h') deps += ['mbuf', 'net', 'cryptodev', 'security'] diff --git a/lib/pdcp/rte_pdcp.h b/lib/pdcp/rte_pdcp.h index 46c3c2a416..b88cad4f64 100644 --- a/lib/pdcp/rte_pdcp.h +++ b/lib/pdcp/rte_pdcp.h @@ -257,6 +257,12 @@ rte_pdcp_pkt_post_process(const struct rte_pdcp_entity *entity, return entity->post_process(entity, in_mb, out_mb, num, nb_err); } +/** + * The header 'rte_pdcp_group.h' depends on defines in 'rte_pdcp.h'. So include + * in the end. + */ +#include + #ifdef __cplusplus } #endif diff --git a/lib/pdcp/rte_pdcp_group.h b/lib/pdcp/rte_pdcp_group.h new file mode 100644 index 00..ece3e8c0ff --- /dev/null +++ b/lib/pdcp/rte_pdcp_group.h @@ -0,0 +1,131 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(C) 2023 Marvell. + */ + +#ifndef RTE_PDCP_GROUP_H +#define RTE_PDCP_GROUP_H + +/** + * @file rte_pdcp_group.h + * + * RTE PDCP grouping support. + * It is not recommended to include this file directly, include + * instead. + * Provides helper functions to process completed crypto-ops and group related + * packets by sessions they belong to. + */ + +#include +#include +#include + +#ifdef __cplusplus +extern "C" { +#endif + +/** + * Group packets belonging to same PDCP entity. + */ +struct rte_pdcp_group { + union { + uint64_t val; + void *ptr; + } id; /**< Grouped by value */ + struct rte_mbuf **m; /**< Start of the group */ + uint32_t cnt; /**< Number of entries in the group */ + int32_t rc; /**< Status code associated with the group */ +}; + +/** + * @warning + * @b EXPERIMENTAL: this API may change without prior notice + * + * Take crypto-op as an input and extract pointer to related PDCP entity. + * @param cop + * The address of an input *rte_crypto_op* structure. + * @return + * The pointer to the related *rte_pdcp_entity* structure. + */ +static inline struct rte_pdcp_entity * +rte_pdcp_en_from_cop(const struct rte_crypto_op *cop) +{ + void *sess = cop->sym[0].session; + + return (struct rte_pdcp_entity *)(uintptr_t) + rte_cryptodev_sym_session_opaque_data_get(sess); +} + +/** + * @warning + * @b EXPERIMENTAL: this API may change without prior notice + * + * Take as input completed crypto ops, extract related mbufs and group them by + * *rte_pdcp_entity* they belong to. Mbuf for which the crypto operation has + * failed would be flagged using *RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED* flag + * in rte_mbuf.ol_flags. The crypto_ops would be freed after the grouping. + * + * Note that application must ensure only crypto-ops prepared by lib_pdcp is + * provided back to @see rte_pdcp_pkt_crypto_group(). + * + * @param cop + * The address of an array of *num* pointers to the input *rte_crypto_op* + * structures. + * @param[out] mb + * The address of an array of *num* pointers to output *rte_mbuf* structures. + * @param[out] grp + * The address of an array of *num* to output *rte_pdcp_group* structures. + * @param num + * The maximum number of crypto-ops to process. + * @return + * Number of filled elements in *grp* array. + * + */ +static inline uint16_t +rte_pdcp_pkt_crypto_group(struct rte_crypto_op *cop[], struct rte_mbuf *mb[], + struct rte_pdcp_group grp[], uint16_t num) +{ + uint32_t i, j = 0, n = 0; + void *ns, *ps = NULL; + struct rte_mbuf *m; + + for (i = 0; i != num; i++) { + m = cop[i]->sym[0].m_src; + ns = cop[i]->sym[0].session; + + m->ol_flags |= RTE_MBUF_F_RX_SEC_OFFLOAD; + if (cop[i]->status != RTE_CRYPTO_OP_STATUS_SUCCESS) + m->ol_flags |= RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED; + + /* Different entity */ + if (ps != ns) { + + /* Finalize open group and start a new one */ + if (ps != NULL) { + grp[n].cnt = mb + j - grp[n].m; + n++; + } +
[PATCH v4 05/22] pdcp: add crypto session create and destroy
Add routines to create & destroy sessions. PDCP lib would take crypto transforms as input and creates the session on the corresponding device after verifying capabilities. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko Acked-by: Akhil Goyal --- lib/pdcp/pdcp_crypto.c | 223 - lib/pdcp/pdcp_crypto.h | 5 + 2 files changed, 225 insertions(+), 3 deletions(-) diff --git a/lib/pdcp/pdcp_crypto.c b/lib/pdcp/pdcp_crypto.c index 755e27ec9e..6d2a85dc7d 100644 --- a/lib/pdcp/pdcp_crypto.c +++ b/lib/pdcp/pdcp_crypto.c @@ -2,20 +2,237 @@ * Copyright(C) 2023 Marvell. */ +#include +#include +#include +#include #include +#include #include "pdcp_crypto.h" +#include "pdcp_entity.h" + +static int +pdcp_crypto_caps_cipher_verify(uint8_t dev_id, const struct rte_crypto_sym_xform *c_xfrm) +{ + const struct rte_cryptodev_symmetric_capability *cap; + struct rte_cryptodev_sym_capability_idx cap_idx; + int ret; + + cap_idx.type = RTE_CRYPTO_SYM_XFORM_CIPHER; + cap_idx.algo.cipher = c_xfrm->cipher.algo; + + cap = rte_cryptodev_sym_capability_get(dev_id, &cap_idx); + if (cap == NULL) + return -1; + + ret = rte_cryptodev_sym_capability_check_cipher(cap, c_xfrm->cipher.key.length, + c_xfrm->cipher.iv.length); + + return ret; +} + +static int +pdcp_crypto_caps_auth_verify(uint8_t dev_id, const struct rte_crypto_sym_xform *a_xfrm) +{ + const struct rte_cryptodev_symmetric_capability *cap; + struct rte_cryptodev_sym_capability_idx cap_idx; + int ret; + + cap_idx.type = RTE_CRYPTO_SYM_XFORM_AUTH; + cap_idx.algo.auth = a_xfrm->auth.algo; + + cap = rte_cryptodev_sym_capability_get(dev_id, &cap_idx); + if (cap == NULL) + return -1; + + ret = rte_cryptodev_sym_capability_check_auth(cap, a_xfrm->auth.key.length, + a_xfrm->auth.digest_length, + a_xfrm->auth.iv.length); + + return ret; +} + +static int +pdcp_crypto_xfrm_validate(const struct rte_pdcp_entity_conf *conf, +const struct rte_crypto_sym_xform *c_xfrm, +const struct rte_crypto_sym_xform *a_xfrm, +bool is_auth_then_cipher) +{ + uint16_t cipher_iv_len, auth_digest_len, auth_iv_len; + int ret; + + /* +* Uplink means PDCP entity is configured for transmit. Downlink means PDCP entity is +* configured for receive. When integrity protection is enabled, PDCP always performs +* digest-encrypted or auth-gen-encrypt for uplink (and decrypt-auth-verify for downlink). +* So for uplink, crypto chain would be auth-cipher while for downlink it would be +* cipher-auth. +* +* When integrity protection is not required, xform would be cipher only. +*/ + + if (c_xfrm == NULL) + return -EINVAL; + + if (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK) { + + /* With UPLINK, if auth is enabled, it should be before cipher */ + if (a_xfrm != NULL && !is_auth_then_cipher) + return -EINVAL; + + /* With UPLINK, cipher operation must be encrypt */ + if (c_xfrm->cipher.op != RTE_CRYPTO_CIPHER_OP_ENCRYPT) + return -EINVAL; + + /* With UPLINK, auth operation (if present) must be generate */ + if (a_xfrm != NULL && a_xfrm->auth.op != RTE_CRYPTO_AUTH_OP_GENERATE) + return -EINVAL; + + } else if (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) { + + /* With DOWNLINK, if auth is enabled, it should be after cipher */ + if (a_xfrm != NULL && is_auth_then_cipher) + return -EINVAL; + + /* With DOWNLINK, cipher operation must be decrypt */ + if (c_xfrm->cipher.op != RTE_CRYPTO_CIPHER_OP_DECRYPT) + return -EINVAL; + + /* With DOWNLINK, auth operation (if present) must be verify */ + if (a_xfrm != NULL && a_xfrm->auth.op != RTE_CRYPTO_AUTH_OP_VERIFY) + return -EINVAL; + + } else { + return -EINVAL; + } + + if ((c_xfrm->cipher.algo != RTE_CRYPTO_CIPHER_NULL) && + (c_xfrm->cipher.algo != RTE_CRYPTO_CIPHER_AES_CTR) && + (c_xfrm->cipher.algo != RTE_CRYPTO_CIPHER_ZUC_EEA3) && + (c_xfrm->cipher.algo != RTE_CRYPTO_CIPHER_SNOW3G_UEA2)) + return -EINVAL; + + if (c_xfrm->cipher.algo == RTE_CRYPTO_CIPHER_NULL) + cipher_iv_len = 0; + else + cipher_iv_len = PDCP_IV_LEN; + + if (cipher_iv_len != c_xfrm->cipher.iv.length) +
[PATCH v4 06/22] pdcp: add pre and post process for UL
Add routines to perform pre & post processing based on the type of entity. To avoid checks in datapath, there are different function pointers registered based on the following, 1. Control plane v/s user plane 2. 12 bit v/s 18 bit SN For control plane only 12 bit SN need to be supported (as per PDCP specification). Signed-off-by: Anoob Joseph Signed-off-by: Kiran Kumar K Signed-off-by: Volodymyr Fialko Acked-by: Akhil Goyal --- lib/pdcp/pdcp_entity.h | 24 +++ lib/pdcp/pdcp_process.c | 334 2 files changed, 358 insertions(+) diff --git a/lib/pdcp/pdcp_entity.h b/lib/pdcp/pdcp_entity.h index 000297588f..23628ebad4 100644 --- a/lib/pdcp/pdcp_entity.h +++ b/lib/pdcp/pdcp_entity.h @@ -92,22 +92,46 @@ pdcp_hdr_size_get(enum rte_security_pdcp_sn_size sn_size) return RTE_ALIGN_MUL_CEIL(sn_size, 8) / 8; } +static inline uint32_t +pdcp_window_size_get(enum rte_security_pdcp_sn_size sn_size) +{ + return 1 << (sn_size - 1); +} + static inline uint32_t pdcp_sn_mask_get(enum rte_security_pdcp_sn_size sn_size) { return (1 << sn_size) - 1; } +static inline uint32_t +pdcp_sn_from_count_get(uint32_t count, enum rte_security_pdcp_sn_size sn_size) +{ + return (count & pdcp_sn_mask_get(sn_size)); +} + static inline uint32_t pdcp_hfn_mask_get(enum rte_security_pdcp_sn_size sn_size) { return ~pdcp_sn_mask_get(sn_size); } +static inline uint32_t +pdcp_hfn_from_count_get(uint32_t count, enum rte_security_pdcp_sn_size sn_size) +{ + return (count & pdcp_hfn_mask_get(sn_size)) >> sn_size; +} + static inline uint32_t pdcp_count_from_hfn_sn_get(uint32_t hfn, uint32_t sn, enum rte_security_pdcp_sn_size sn_size) { return (((hfn << sn_size) & pdcp_hfn_mask_get(sn_size)) | (sn & pdcp_sn_mask_get(sn_size))); } +static inline uint32_t +pdcp_hfn_max(enum rte_security_pdcp_sn_size sn_size) +{ + return (1 << (32 - sn_size)) - 1; +} + #endif /* PDCP_ENTITY_H */ diff --git a/lib/pdcp/pdcp_process.c b/lib/pdcp/pdcp_process.c index 79f5dce5db..9b7de39db6 100644 --- a/lib/pdcp/pdcp_process.c +++ b/lib/pdcp/pdcp_process.c @@ -36,6 +36,336 @@ pdcp_crypto_xfrm_get(const struct rte_pdcp_entity_conf *conf, struct rte_crypto_ return 0; } +static inline void +cop_prepare(const struct entity_priv *en_priv, struct rte_mbuf *mb, struct rte_crypto_op *cop, + uint8_t data_offset, uint32_t count, const bool is_auth) +{ + const struct rte_crypto_op cop_init = { + .type = RTE_CRYPTO_OP_TYPE_SYMMETRIC, + .status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED, + .sess_type = RTE_CRYPTO_OP_WITH_SESSION, + }; + struct rte_crypto_sym_op *op; + uint32_t pkt_len; + + const uint8_t cipher_shift = 3 * en_priv->flags.is_cipher_in_bits; + const uint8_t auth_shift = 3 * en_priv->flags.is_auth_in_bits; + + op = cop->sym; + cop->raw = cop_init.raw; + op->m_src = mb; + op->m_dst = mb; + + /* Set IV */ + en_priv->iv_gen(cop, en_priv, count); + + /* Prepare op */ + pkt_len = rte_pktmbuf_pkt_len(mb); + op->cipher.data.offset = data_offset << cipher_shift; + op->cipher.data.length = (pkt_len - data_offset) << cipher_shift; + + if (is_auth) { + op->auth.data.offset = 0; + op->auth.data.length = (pkt_len - RTE_PDCP_MAC_I_LEN) << auth_shift; + op->auth.digest.data = rte_pktmbuf_mtod_offset(mb, uint8_t *, + (pkt_len - RTE_PDCP_MAC_I_LEN)); + } + + __rte_crypto_sym_op_attach_sym_session(op, en_priv->crypto_sess); +} + +static inline bool +pdcp_pre_process_uplane_sn_12_ul_set_sn(struct entity_priv *en_priv, struct rte_mbuf *mb, + uint32_t *count) +{ + struct rte_pdcp_up_data_pdu_sn_12_hdr *pdu_hdr; + const uint8_t hdr_sz = en_priv->hdr_sz; + uint32_t sn; + + /* Prepend PDU header */ + pdu_hdr = (struct rte_pdcp_up_data_pdu_sn_12_hdr *)rte_pktmbuf_prepend(mb, hdr_sz); + if (unlikely(pdu_hdr == NULL)) + return false; + + /* Update sequence num in the PDU header */ + *count = en_priv->state.tx_next++; + sn = pdcp_sn_from_count_get(*count, RTE_SECURITY_PDCP_SN_SIZE_12); + + pdu_hdr->d_c = RTE_PDCP_PDU_TYPE_DATA; + pdu_hdr->sn_11_8 = ((sn & 0xf00) >> 8); + pdu_hdr->sn_7_0 = (sn & 0xff); + pdu_hdr->r = 0; + return true; +} + +static uint16_t +pdcp_pre_process_uplane_sn_12_ul(const struct rte_pdcp_entity *entity, struct rte_mbuf *in_mb[], +struct rte_crypto_op *cop[], uint16_t num, uint16_t *nb_err_ret) +{ + struct entity_priv *en_priv = entity_priv_get(entity); + uint16_t nb_cop, nb_prep = 0, nb_err = 0; + struct rte_mbuf *mb; + uint32_t count; + uint8_t *mac_i; + int i; + + const uint8_
[PATCH v4 07/22] pdcp: add pre and post process for DL
Add routines to perform pre & post processing for down link entities. Signed-off-by: Anoob Joseph Signed-off-by: Kiran Kumar K Signed-off-by: Volodymyr Fialko --- lib/pdcp/pdcp_entity.h | 2 + lib/pdcp/pdcp_process.c | 383 lib/pdcp/pdcp_process.h | 10 ++ lib/pdcp/rte_pdcp.c | 23 +++ 4 files changed, 418 insertions(+) diff --git a/lib/pdcp/pdcp_entity.h b/lib/pdcp/pdcp_entity.h index 23628ebad4..1d4a43a3bc 100644 --- a/lib/pdcp/pdcp_entity.h +++ b/lib/pdcp/pdcp_entity.h @@ -13,6 +13,8 @@ struct entity_priv; +#define PDCP_HFN_MIN 0 + /* IV generation function based on the entity configuration */ typedef void (*iv_gen_t)(struct rte_crypto_op *cop, const struct entity_priv *en_priv, uint32_t count); diff --git a/lib/pdcp/pdcp_process.c b/lib/pdcp/pdcp_process.c index 9b7de39db6..0737a60df2 100644 --- a/lib/pdcp/pdcp_process.c +++ b/lib/pdcp/pdcp_process.c @@ -333,9 +333,353 @@ pdcp_post_process_ul(const struct rte_pdcp_entity *entity, return nb_success; } +static inline int +pdcp_sn_count_get(const uint32_t rx_deliv, int32_t rsn, uint32_t *count, + const enum rte_security_pdcp_sn_size sn_size) +{ + const uint32_t rx_deliv_sn = pdcp_sn_from_count_get(rx_deliv, sn_size); + const uint32_t window_sz = pdcp_window_size_get(sn_size); + uint32_t rhfn; + + rhfn = pdcp_hfn_from_count_get(rx_deliv, sn_size); + + if (rsn < (int32_t)(rx_deliv_sn - window_sz)) { + if (unlikely(rhfn == pdcp_hfn_max(sn_size))) + return -ERANGE; + rhfn += 1; + } else if ((uint32_t)rsn >= (rx_deliv_sn + window_sz)) { + if (unlikely(rhfn == PDCP_HFN_MIN)) + return -ERANGE; + rhfn -= 1; + } + + *count = pdcp_count_from_hfn_sn_get(rhfn, rsn, sn_size); + + return 0; +} + +static inline uint16_t +pdcp_pre_process_uplane_sn_12_dl_flags(const struct rte_pdcp_entity *entity, + struct rte_mbuf *in_mb[], struct rte_crypto_op *cop[], + uint16_t num, uint16_t *nb_err_ret, + const bool is_integ_protected) +{ + struct entity_priv *en_priv = entity_priv_get(entity); + struct rte_pdcp_up_data_pdu_sn_12_hdr *pdu_hdr; + uint16_t nb_cop, nb_prep = 0, nb_err = 0; + rte_pdcp_dynfield_t *mb_dynfield; + struct rte_mbuf *mb; + int32_t rsn = 0; + uint32_t count; + int i; + + const uint8_t data_offset = en_priv->hdr_sz + en_priv->aad_sz; + + nb_cop = rte_crypto_op_bulk_alloc(en_priv->cop_pool, RTE_CRYPTO_OP_TYPE_SYMMETRIC, cop, + num); + + const uint32_t rx_deliv = en_priv->state.rx_deliv; + + for (i = 0; i < nb_cop; i++) { + mb = in_mb[i]; + pdu_hdr = rte_pktmbuf_mtod(mb, struct rte_pdcp_up_data_pdu_sn_12_hdr *); + + /* Check for PDU type */ + if (likely(pdu_hdr->d_c == RTE_PDCP_PDU_TYPE_DATA)) { + rsn = ((pdu_hdr->sn_11_8 << 8) | (pdu_hdr->sn_7_0)); + } else { + /** NOTE: Control PDU not handled.*/ + in_mb[nb_err++] = mb; + continue; + } + + if (unlikely(pdcp_sn_count_get(rx_deliv, rsn, &count, + RTE_SECURITY_PDCP_SN_SIZE_12))) { + in_mb[nb_err++] = mb; + continue; + } + + cop_prepare(en_priv, mb, cop[nb_prep++], data_offset, count, is_integ_protected); + + mb_dynfield = pdcp_dynfield(mb); + *mb_dynfield = count; + } + + if (unlikely(nb_err)) + rte_mempool_put_bulk(en_priv->cop_pool, (void *)&cop[nb_prep], nb_cop - nb_prep); + + *nb_err_ret = num - nb_prep; + + return nb_prep; +} + +static uint16_t +pdcp_pre_process_uplane_sn_12_dl_ip(const struct rte_pdcp_entity *entity, struct rte_mbuf *mb[], + struct rte_crypto_op *cop[], uint16_t num, uint16_t *nb_err) +{ + return pdcp_pre_process_uplane_sn_12_dl_flags(entity, mb, cop, num, nb_err, true); +} + +static uint16_t +pdcp_pre_process_uplane_sn_12_dl(const struct rte_pdcp_entity *entity, struct rte_mbuf *mb[], +struct rte_crypto_op *cop[], uint16_t num, uint16_t *nb_err) +{ + return pdcp_pre_process_uplane_sn_12_dl_flags(entity, mb, cop, num, nb_err, false); +} + +static inline uint16_t +pdcp_pre_process_uplane_sn_18_dl_flags(const struct rte_pdcp_entity *entity, + struct rte_mbuf *in_mb[], struct rte_crypto_op *cop[], + uint16_t num, uint16_t *nb_err_ret, + const bool is_in
[PATCH v4 08/22] pdcp: add IV generation routines
For PDCP, IV generated has varying formats depending on the ciphering and authentication algorithm used. Add routines to populate IV accordingly. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- lib/pdcp/pdcp_entity.h | 87 lib/pdcp/pdcp_process.c | 284 2 files changed, 371 insertions(+) diff --git a/lib/pdcp/pdcp_entity.h b/lib/pdcp/pdcp_entity.h index 1d4a43a3bc..10a72faae1 100644 --- a/lib/pdcp/pdcp_entity.h +++ b/lib/pdcp/pdcp_entity.h @@ -26,6 +26,89 @@ struct entity_state { uint32_t rx_reord; }; +union auth_iv_partial { + /* For AES-CMAC, there is no IV, but message gets prepended */ + struct { +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN + uint64_t count : 32; + uint64_t zero_38_39 : 2; + uint64_t direction : 1; + uint64_t bearer : 5; + uint64_t zero_40_63 : 24; +#else + uint64_t count : 32; + uint64_t bearer : 5; + uint64_t direction : 1; + uint64_t zero_38_39 : 2; + uint64_t zero_40_63 : 24; +#endif + } aes_cmac; + struct { +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN + uint64_t count : 32; + uint64_t zero_37_39 : 3; + uint64_t bearer : 5; + uint64_t zero_40_63 : 24; + + uint64_t rsvd_65_71 : 7; + uint64_t direction_64 : 1; + uint64_t rsvd_72_111 : 40; + uint64_t rsvd_113_119 : 7; + uint64_t direction_112 : 1; + uint64_t rsvd_120_127 : 8; +#else + uint64_t count : 32; + uint64_t bearer : 5; + uint64_t zero_37_39 : 3; + uint64_t zero_40_63 : 24; + + uint64_t direction_64 : 1; + uint64_t rsvd_65_71 : 7; + uint64_t rsvd_72_111 : 40; + uint64_t direction_112 : 1; + uint64_t rsvd_113_119 : 7; + uint64_t rsvd_120_127 : 8; +#endif + } zs; + uint64_t u64[2]; +}; + +union cipher_iv_partial { + struct { +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN + uint64_t count : 32; + uint64_t zero_38_39 : 2; + uint64_t direction : 1; + uint64_t bearer : 5; + uint64_t zero_40_63 : 24; +#else + uint64_t count : 32; + uint64_t bearer : 5; + uint64_t direction : 1; + uint64_t zero_38_39 : 2; + uint64_t zero_40_63 : 24; +#endif + uint64_t zero_64_127; + } aes_ctr; + struct { +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN + uint64_t count : 32; + uint64_t zero_38_39 : 2; + uint64_t direction : 1; + uint64_t bearer : 5; + uint64_t zero_40_63 : 24; +#else + uint64_t count : 32; + uint64_t bearer : 5; + uint64_t direction : 1; + uint64_t zero_38_39 : 2; + uint64_t zero_40_63 : 24; +#endif + uint64_t rsvd_64_127; + } zs; + uint64_t u64[2]; +}; + /* * Layout of PDCP entity: [rte_pdcp_entity] [entity_priv] [entity_dl/ul] */ @@ -35,6 +118,10 @@ struct entity_priv { struct rte_cryptodev_sym_session *crypto_sess; /** Entity specific IV generation function. */ iv_gen_t iv_gen; + /** Pre-prepared auth IV. */ + union auth_iv_partial auth_iv_part; + /** Pre-prepared cipher IV. */ + union cipher_iv_partial cipher_iv_part; /** Entity state variables. */ struct entity_state state; /** Flags. */ diff --git a/lib/pdcp/pdcp_process.c b/lib/pdcp/pdcp_process.c index 0737a60df2..8c802e735b 100644 --- a/lib/pdcp/pdcp_process.c +++ b/lib/pdcp/pdcp_process.c @@ -13,6 +13,181 @@ #include "pdcp_entity.h" #include "pdcp_process.h" +/* Enum of supported algorithms for ciphering */ +enum pdcp_cipher_algo { + PDCP_CIPHER_ALGO_NULL, + PDCP_CIPHER_ALGO_AES, + PDCP_CIPHER_ALGO_ZUC, + PDCP_CIPHER_ALGO_SNOW3G, + PDCP_CIPHER_ALGO_MAX +}; + +/* Enum of supported algorithms for integrity */ +enum pdcp_auth_algo { + PDCP_AUTH_ALGO_NULL, + PDCP_AUTH_ALGO_AES, + PDCP_AUTH_ALGO_ZUC, + PDCP_AUTH_ALGO_SNOW3G, + PDCP_AUTH_ALGO_MAX +}; + +/* IV generation functions based on type of operation (cipher - auth) */ + +static void +pdcp_iv_gen_null_null(struct rte_crypto_op *cop, const struct entity_priv *en_priv, uint32_t count) +{ + /* No IV required for NULL cipher + NULL auth */ + RTE_SET_USED(cop); + RTE_SET_USED(en_priv); + RTE_SET_USED(count); +} + +static void +pdcp_iv_gen_null_aes_cmac(struct rte_crypto_op *cop, const struct entity_priv *en_priv, + uint32_t count) +{ + struct rte_crypto_sym_op *op = cop->sym; + struct rt
[PATCH v4 09/22] app/test: add lib pdcp tests
Add tests to verify lib PDCP operations. Tests leverage existing PDCP test vectors. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- app/test/meson.build | 5 + app/test/test_cryptodev.h | 3 + app/test/test_pdcp.c | 732 ++ 3 files changed, 740 insertions(+) create mode 100644 app/test/test_pdcp.c diff --git a/app/test/meson.build b/app/test/meson.build index b9b5432496..2894793347 100644 --- a/app/test/meson.build +++ b/app/test/meson.build @@ -435,6 +435,11 @@ if dpdk_conf.has('RTE_HAS_LIBPCAP') endif endif +if dpdk_conf.has('RTE_LIB_PDCP') +test_sources += 'test_pdcp.c' +fast_tests += [['pdcp_autotest', false, true]] +endif + if cc.has_argument('-Wno-format-truncation') cflags += '-Wno-format-truncation' endif diff --git a/app/test/test_cryptodev.h b/app/test/test_cryptodev.h index abd795f54a..89057dba22 100644 --- a/app/test/test_cryptodev.h +++ b/app/test/test_cryptodev.h @@ -4,6 +4,9 @@ #ifndef TEST_CRYPTODEV_H_ #define TEST_CRYPTODEV_H_ +#include +#include + #define HEX_DUMP 0 #define FALSE 0 diff --git a/app/test/test_pdcp.c b/app/test/test_pdcp.c new file mode 100644 index 00..34b759eaef --- /dev/null +++ b/app/test/test_pdcp.c @@ -0,0 +1,732 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(C) 2023 Marvell. + */ + +#include +#include +#include +#include + +#include "test.h" +#include "test_cryptodev.h" +#include "test_cryptodev_security_pdcp_test_vectors.h" + +#define NB_DESC 1024 +#define CDEV_INVALID_ID UINT8_MAX +#define NB_TESTS RTE_DIM(pdcp_test_params) +#define PDCP_IV_LEN 16 + +struct pdcp_testsuite_params { + struct rte_mempool *mbuf_pool; + struct rte_mempool *cop_pool; + struct rte_mempool *sess_pool; + bool cdevs_used[RTE_CRYPTO_MAX_DEVS]; +}; + +static struct pdcp_testsuite_params testsuite_params; + +struct pdcp_test_conf { + struct rte_pdcp_entity_conf entity; + struct rte_crypto_sym_xform c_xfrm; + struct rte_crypto_sym_xform a_xfrm; + bool is_integrity_protected; + uint8_t input[RTE_PDCP_CTRL_PDU_SIZE_MAX]; + uint32_t input_len; + uint8_t output[RTE_PDCP_CTRL_PDU_SIZE_MAX]; + uint32_t output_len; +}; + +static inline int +pdcp_hdr_size_get(enum rte_security_pdcp_sn_size sn_size) +{ + return RTE_ALIGN_MUL_CEIL(sn_size, 8) / 8; +} + +static int +cryptodev_init(int dev_id) +{ + struct pdcp_testsuite_params *ts_params = &testsuite_params; + struct rte_cryptodev_qp_conf qp_conf; + struct rte_cryptodev_info dev_info; + struct rte_cryptodev_config config; + int ret, socket_id; + + /* Check if device was already initialized */ + if (ts_params->cdevs_used[dev_id]) + return 0; + + rte_cryptodev_info_get(dev_id, &dev_info); + + if (dev_info.max_nb_queue_pairs < 1) { + RTE_LOG(ERR, USER1, "Cryptodev doesn't have sufficient queue pairs available\n"); + return -ENODEV; + } + + socket_id = rte_socket_id(); + + memset(&config, 0, sizeof(config)); + config.nb_queue_pairs = 1; + config.socket_id = socket_id; + + ret = rte_cryptodev_configure(dev_id, &config); + if (ret < 0) { + RTE_LOG(ERR, USER1, "Could not configure cryptodev - %d\n", dev_id); + return -ENODEV; + } + + memset(&qp_conf, 0, sizeof(qp_conf)); + qp_conf.nb_descriptors = NB_DESC; + + ret = rte_cryptodev_queue_pair_setup(dev_id, 0, &qp_conf, socket_id); + if (ret < 0) { + RTE_LOG(ERR, USER1, "Could not configure queue pair\n"); + return -ENODEV; + } + + ret = rte_cryptodev_start(dev_id); + if (ret < 0) { + RTE_LOG(ERR, USER1, "Could not start cryptodev\n"); + return -ENODEV; + } + + /* Mark device as initialized */ + ts_params->cdevs_used[dev_id] = true; + + return 0; +} + +static void +cryptodev_fini(int dev_id) +{ + rte_cryptodev_stop(dev_id); +} + +static unsigned int +cryptodev_sess_priv_max_req_get(void) +{ + struct rte_cryptodev_info info; + unsigned int sess_priv_sz; + int i, nb_dev; + void *sec_ctx; + + nb_dev = rte_cryptodev_count(); + + sess_priv_sz = 0; + + for (i = 0; i < nb_dev; i++) { + rte_cryptodev_info_get(i, &info); + sess_priv_sz = RTE_MAX(sess_priv_sz, rte_cryptodev_sym_get_private_session_size(i)); + if (info.feature_flags & RTE_CRYPTODEV_FF_SECURITY) { + sec_ctx = rte_cryptodev_get_sec_ctx(i); + sess_priv_sz = RTE_MAX(sess_priv_sz, + rte_security_session_get_size(sec_ctx)); + } + } + + return sess_priv_sz; +} + +static int +testsuite_setup(void) +{ + struct pdcp_testsuite_params *ts_params
[PATCH v4 10/22] test/pdcp: pdcp HFN tests in combined mode
From: Volodymyr Fialko Add tests to verify HFN/SN behaviour. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- app/test/test_pdcp.c | 302 ++- 1 file changed, 299 insertions(+), 3 deletions(-) diff --git a/app/test/test_pdcp.c b/app/test/test_pdcp.c index 34b759eaef..cfe2ec6aa9 100644 --- a/app/test/test_pdcp.c +++ b/app/test/test_pdcp.c @@ -16,6 +16,9 @@ #define NB_TESTS RTE_DIM(pdcp_test_params) #define PDCP_IV_LEN 16 +/* According to formula(7.2.a Window_Size) */ +#define PDCP_WINDOW_SIZE(sn_size) (1 << (sn_size - 1)) + struct pdcp_testsuite_params { struct rte_mempool *mbuf_pool; struct rte_mempool *cop_pool; @@ -36,12 +39,69 @@ struct pdcp_test_conf { uint32_t output_len; }; +static int create_test_conf_from_index(const int index, struct pdcp_test_conf *conf); + +typedef int (*test_with_conf_t)(struct pdcp_test_conf *conf); + +static int +run_test_foreach_known_vec(test_with_conf_t test, bool stop_on_first_pass) +{ + struct pdcp_test_conf test_conf; + bool all_tests_skipped = true; + uint32_t i; + int ret; + + for (i = 0; i < NB_TESTS; i++) { + create_test_conf_from_index(i, &test_conf); + ret = test(&test_conf); + + if (ret == TEST_FAILED) { + printf("[%03i] - %s - failed\n", i, pdcp_test_params[i].name); + return TEST_FAILED; + } + + if ((ret == TEST_SKIPPED) || (ret == -ENOTSUP)) + continue; + + if (stop_on_first_pass) + return TEST_SUCCESS; + + all_tests_skipped = false; + } + + if (all_tests_skipped) + return TEST_SKIPPED; + + return TEST_SUCCESS; +} + +static int +run_test_with_all_known_vec(const void *args) +{ + test_with_conf_t test = args; + + return run_test_foreach_known_vec(test, false); +} + static inline int pdcp_hdr_size_get(enum rte_security_pdcp_sn_size sn_size) { return RTE_ALIGN_MUL_CEIL(sn_size, 8) / 8; } +static int +pktmbuf_read_into(const struct rte_mbuf *m, void *buf, size_t buf_len) +{ + if (m->pkt_len > buf_len) + return -ENOMEM; + + const void *read = rte_pktmbuf_read(m, 0, m->pkt_len, buf); + if (read != NULL && read != buf) + memcpy(buf, read, m->pkt_len); + + return 0; +} + static int cryptodev_init(int dev_id) { @@ -326,6 +386,21 @@ pdcp_sn_from_raw_get(const void *data, enum rte_security_pdcp_sn_size size) return sn; } +static void +pdcp_sn_to_raw_set(void *data, uint32_t sn, int size) +{ + if (size == RTE_SECURITY_PDCP_SN_SIZE_12) { + struct rte_pdcp_up_data_pdu_sn_12_hdr *pdu_hdr = data; + pdu_hdr->sn_11_8 = ((sn & 0xf00) >> 8); + pdu_hdr->sn_7_0 = (sn & 0xff); + } else if (size == RTE_SECURITY_PDCP_SN_SIZE_18) { + struct rte_pdcp_up_data_pdu_sn_18_hdr *pdu_hdr = data; + pdu_hdr->sn_17_16 = ((sn & 0x3) >> 16); + pdu_hdr->sn_15_8 = ((sn & 0xff00) >> 8); + pdu_hdr->sn_7_0 = (sn & 0xff); + } +} + static int create_test_conf_from_index(const int index, struct pdcp_test_conf *conf) { @@ -648,9 +723,17 @@ test_attempt_single(struct pdcp_test_conf *t_conf) goto mbuf_free; } - ret = pdcp_known_vec_verify(mbuf, t_conf->output, t_conf->output_len); - if (ret) - goto mbuf_free; + /* If expected output provided - verify, else - store for future use */ + if (t_conf->output_len) { + ret = pdcp_known_vec_verify(mbuf, t_conf->output, t_conf->output_len); + if (ret) + goto mbuf_free; + } else { + ret = pktmbuf_read_into(mbuf, t_conf->output, RTE_PDCP_CTRL_PDU_SIZE_MAX); + if (ret) + goto mbuf_free; + t_conf->output_len = mbuf->pkt_len; + } ret = rte_pdcp_entity_suspend(pdcp_entity, out_mb); if (ret) { @@ -667,6 +750,193 @@ test_attempt_single(struct pdcp_test_conf *t_conf) return ret; } +static void +uplink_to_downlink_convert(const struct pdcp_test_conf *ul_cfg, + struct pdcp_test_conf *dl_cfg) +{ + assert(ul_cfg->entity.pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK); + + memcpy(dl_cfg, ul_cfg, sizeof(*dl_cfg)); + dl_cfg->entity.pdcp_xfrm.pkt_dir = RTE_SECURITY_PDCP_DOWNLINK; + dl_cfg->entity.reverse_iv_direction = false; + + if (dl_cfg->is_integrity_protected) { + dl_cfg->entity.crypto_xfrm = &dl_cfg->c_xfrm; + + dl_cfg->c_xfrm.cipher.op = RTE_CRYPTO_CIPHER_OP_DECRYPT; + dl_cfg->c_xfrm.next = &dl_cfg->a_xfrm; + + dl_cfg->a_xfrm.auth.op = RTE_CRYPTO_AUTH_OP_VERIFY; + dl_cfg->a_xfrm.next = NULL;
[PATCH v4 11/22] doc: add PDCP library guide
Add guide for PDCP library. Signed-off-by: Anoob Joseph Signed-off-by: Kiran Kumar K Signed-off-by: Volodymyr Fialko --- .../img/pdcp_functional_overview.svg | 1 + doc/guides/prog_guide/index.rst | 1 + doc/guides/prog_guide/pdcp_lib.rst| 254 ++ 3 files changed, 256 insertions(+) create mode 100644 doc/guides/prog_guide/img/pdcp_functional_overview.svg create mode 100644 doc/guides/prog_guide/pdcp_lib.rst diff --git a/doc/guides/prog_guide/img/pdcp_functional_overview.svg b/doc/guides/prog_guide/img/pdcp_functional_overview.svg new file mode 100644 index 00..287daafc21 --- /dev/null +++ b/doc/guides/prog_guide/img/pdcp_functional_overview.svg @@ -0,0 +1 @@ +http://www.w3.org/2000/svg"; xmlns:xlink="http://www.w3.org/1999/xlink"; overflow="hidden">Radio Interface (Uu/PC5)UE/NG-RAN/UE ANG-RAN/UE/UE BTransmitting PDCP entityReceiving PDCP entityTransmission buffer:SequencenumberingHeader or uplink dataCompressionHeader or uplink dataDecompressionRouting / DuplicationAdd PDCP headerCipheringIntegrity protectionPackets associated to a PDCP SDUPackets not associated to a PDCP SDURemove PDCP HeaderDecipheringIntegrity VerificationReception buffer:ReorderingDuplicate discardingPackets associated to a PDCP SDUPackets not associated to a PDCP SDU \ No newline at end of file diff --git a/doc/guides/prog_guide/index.rst b/doc/guides/prog_guide/index.rst index 87333ee84a..6099ff63cd 100644 --- a/doc/guides/prog_guide/index.rst +++ b/doc/guides/prog_guide/index.rst @@ -77,4 +77,5 @@ Programmer's Guide lto profile_app asan +pdcp_lib glossary diff --git a/doc/guides/prog_guide/pdcp_lib.rst b/doc/guides/prog_guide/pdcp_lib.rst new file mode 100644 index 00..8369c71600 --- /dev/null +++ b/doc/guides/prog_guide/pdcp_lib.rst @@ -0,0 +1,254 @@ +.. SPDX-License-Identifier: BSD-3-Clause +Copyright(C) 2023 Marvell. + +PDCP Protocol Processing Library + + +DPDK provides a library for PDCP protocol processing. The library utilizes +other DPDK libraries such as cryptodev, reorder, etc., to provide the +application with a transparent and high performant PDCP protocol processing +library. + +The library abstracts complete PDCP protocol processing conforming to +``ETSI TS 138 323 V17.1.0 (2022-08)``. +https://www.etsi.org/deliver/etsi_ts/138300_138399/138323/17.01.00_60/ts_138323v170100p.pdf + +PDCP would involve the following operations, + +1. Transfer of user plane data +2. Transfer of control plane data +3. Header compression +4. Uplink data compression +5. Ciphering and integrity protection + +.. _figure_pdcp_functional_overview: + +.. figure:: img/pdcp_functional_overview.* + + PDCP functional overview new + +PDCP library would abstract the protocol offload features of the cryptodev and +would provide a uniform interface and consistent API usage to work with +cryptodev irrespective of the protocol offload features supported. + +PDCP entity API +--- + +PDCP library provides following control path APIs that is used to +configure various PDCP entities, + +1. ``rte_pdcp_entity_establish()`` +2. ``rte_pdcp_entity_suspend()`` +3. ``rte_pdcp_entity_release()`` + +A PDCP entity would translate to one ``rte_cryptodev_sym_session`` or +``rte_security_session`` based on the config. The sessions would be created/ +destroyed while corresponding PDCP entity operations are performed. + +When upper layers request a PDCP entity suspend (``rte_pdcp_entity_suspend()``), +it would result in flushing out of all cached packets. Internal state variables +are updated as described in 5.1.4. + +When upper layers request a PDCP entity release (``rte_pdcp_entity_release()``), +it would result in flushing out of all cached packets and releasing of all +memory associated with the entity. It would internally free any crypto/security +sessions created. All procedures mentioned in 5.1.3 would be performed. + +PDCP PDU (Protocol Data Unit) API +- + +PDCP PDUs can be categorized as, + +1. Control PDU +2. Data PDU + +Control PDUs are used for signalling between entities on either end and can be +one of the following, + +1. PDCP status report +2. ROHC feedback +3. EHC feedback + +Control PDUs are not ciphered or authenticated, and so such packets are not +submitted to cryptodev for processing. + +Data PDUs are regular packets submitted by upper layers for transmission to +other end. Such packets would need to be ciphered and authenticated based on +the entity configuration. + +PDCP packet processing API for data PDU +~~~ + +PDCP processing is split into 2 parts. One before cryptodev processing +(``rte_pdcp_pkt_pre_process()``) and one after cryptodev processing +(``rte_pdcp_pkt_post_process()``). Since cryptodev dequeue can return crypto +operations belonging to multiple entities, ``rte_pdcp_pkt_crypto_group()`` +is added to he
[PATCH v4 12/22] pdcp: add control PDU handling for status report
Add control PDU handling and implement status report generation. Status report generation works only when RX_DELIV = RX_NEXT. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- doc/guides/prog_guide/pdcp_lib.rst | 9 ++ lib/pdcp/meson.build | 2 ++ lib/pdcp/pdcp_cnt.c| 29 ++ lib/pdcp/pdcp_cnt.h| 14 + lib/pdcp/pdcp_ctrl_pdu.c | 46 + lib/pdcp/pdcp_ctrl_pdu.h | 15 ++ lib/pdcp/pdcp_entity.h | 15 -- lib/pdcp/pdcp_process.c| 13 + lib/pdcp/rte_pdcp.c| 47 +- lib/pdcp/rte_pdcp.h| 33 + lib/pdcp/version.map | 2 ++ 11 files changed, 222 insertions(+), 3 deletions(-) create mode 100644 lib/pdcp/pdcp_cnt.c create mode 100644 lib/pdcp/pdcp_cnt.h create mode 100644 lib/pdcp/pdcp_ctrl_pdu.c create mode 100644 lib/pdcp/pdcp_ctrl_pdu.h diff --git a/doc/guides/prog_guide/pdcp_lib.rst b/doc/guides/prog_guide/pdcp_lib.rst index 8369c71600..dcb424bb1d 100644 --- a/doc/guides/prog_guide/pdcp_lib.rst +++ b/doc/guides/prog_guide/pdcp_lib.rst @@ -76,6 +76,15 @@ Data PDUs are regular packets submitted by upper layers for transmission to other end. Such packets would need to be ciphered and authenticated based on the entity configuration. +PDCP packet processing API for control PDU +~~ + +Control PDUs are used in PDCP as a communication channel between transmitting +and receiving entities. When upper layer request for operations such +re-establishment, receiving PDCP entity need to prepare a status report and +send it to the other end. The API ``rte_pdcp_control_pdu_create`` allows +application to request the same. + PDCP packet processing API for data PDU ~~~ diff --git a/lib/pdcp/meson.build b/lib/pdcp/meson.build index 08679b743a..75d476bf6d 100644 --- a/lib/pdcp/meson.build +++ b/lib/pdcp/meson.build @@ -8,7 +8,9 @@ if is_windows endif sources = files( +'pdcp_cnt.c', 'pdcp_crypto.c', +'pdcp_ctrl_pdu.c', 'pdcp_process.c', 'rte_pdcp.c', ) diff --git a/lib/pdcp/pdcp_cnt.c b/lib/pdcp/pdcp_cnt.c new file mode 100644 index 00..c9b952184b --- /dev/null +++ b/lib/pdcp/pdcp_cnt.c @@ -0,0 +1,29 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(C) 2023 Marvell. + */ + +#include + +#include "pdcp_cnt.h" +#include "pdcp_entity.h" + +int +pdcp_cnt_ring_create(struct rte_pdcp_entity *en, const struct rte_pdcp_entity_conf *conf) +{ + struct entity_priv_dl_part *en_priv_dl; + uint32_t window_sz; + + if (en == NULL || conf == NULL) + return -EINVAL; + + if (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK) + return 0; + + en_priv_dl = entity_dl_part_get(en); + window_sz = pdcp_window_size_get(conf->pdcp_xfrm.sn_size); + + RTE_SET_USED(window_sz); + RTE_SET_USED(en_priv_dl); + + return 0; +} diff --git a/lib/pdcp/pdcp_cnt.h b/lib/pdcp/pdcp_cnt.h new file mode 100644 index 00..bbda478b55 --- /dev/null +++ b/lib/pdcp/pdcp_cnt.h @@ -0,0 +1,14 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(C) 2023 Marvell. + */ + +#ifndef PDCP_CNT_H +#define PDCP_CNT_H + +#include + +#include "pdcp_entity.h" + +int pdcp_cnt_ring_create(struct rte_pdcp_entity *en, const struct rte_pdcp_entity_conf *conf); + +#endif /* PDCP_CNT_H */ diff --git a/lib/pdcp/pdcp_ctrl_pdu.c b/lib/pdcp/pdcp_ctrl_pdu.c new file mode 100644 index 00..feb05fd863 --- /dev/null +++ b/lib/pdcp/pdcp_ctrl_pdu.c @@ -0,0 +1,46 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(C) 2023 Marvell. + */ + +#include +#include +#include + +#include "pdcp_ctrl_pdu.h" +#include "pdcp_entity.h" + +static __rte_always_inline void +pdcp_hdr_fill(struct rte_pdcp_up_ctrl_pdu_hdr *pdu_hdr, uint32_t rx_deliv) +{ + pdu_hdr->d_c = RTE_PDCP_PDU_TYPE_CTRL; + pdu_hdr->pdu_type = RTE_PDCP_CTRL_PDU_TYPE_STATUS_REPORT; + pdu_hdr->r = 0; + pdu_hdr->fmc = rte_cpu_to_be_32(rx_deliv); +} + +int +pdcp_ctrl_pdu_status_gen(struct entity_priv *en_priv, struct rte_mbuf *m) +{ + struct rte_pdcp_up_ctrl_pdu_hdr *pdu_hdr; + uint32_t rx_deliv; + int pdu_sz; + + if (!en_priv->flags.is_status_report_required) + return -EINVAL; + + pdu_sz = sizeof(struct rte_pdcp_up_ctrl_pdu_hdr); + + rx_deliv = en_priv->state.rx_deliv; + + /* Zero missing PDUs - status report contains only FMC */ + if (rx_deliv >= en_priv->state.rx_next) { + pdu_hdr = (struct rte_pdcp_up_ctrl_pdu_hdr *)rte_pktmbuf_append(m, pdu_sz); + if (pdu_hdr == NULL) + return -ENOMEM; + pdcp_hdr_fill(pdu_hdr, rx_deliv); + + return 0; + } + +
[PATCH v4 13/22] pdcp: implement t-Reordering and packet buffering
From: Volodymyr Fialko Add in-order delivery of packets in PDCP. Delivery of packets in-order relies on t-Reordering timer. When 'out-of-order delivery' is disabled, PDCP will buffer all received packets that are out of order. The t-Reordering timer determines the time period these packets would be held in the buffer, waiting for any missing packets to arrive. Introduce packet buffering and state variables which indicate status of the timer. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- lib/pdcp/meson.build| 3 +- lib/pdcp/pdcp_entity.h | 19 +++ lib/pdcp/pdcp_process.c | 117 ++-- lib/pdcp/pdcp_reorder.c | 27 ++ lib/pdcp/pdcp_reorder.h | 62 + lib/pdcp/rte_pdcp.c | 53 -- lib/pdcp/rte_pdcp.h | 6 ++- 7 files changed, 252 insertions(+), 35 deletions(-) create mode 100644 lib/pdcp/pdcp_reorder.c create mode 100644 lib/pdcp/pdcp_reorder.h diff --git a/lib/pdcp/meson.build b/lib/pdcp/meson.build index 75d476bf6d..f4f9246bcb 100644 --- a/lib/pdcp/meson.build +++ b/lib/pdcp/meson.build @@ -12,9 +12,10 @@ sources = files( 'pdcp_crypto.c', 'pdcp_ctrl_pdu.c', 'pdcp_process.c', +'pdcp_reorder.c', 'rte_pdcp.c', ) headers = files('rte_pdcp.h') indirect_headers += files('rte_pdcp_group.h') -deps += ['mbuf', 'net', 'cryptodev', 'security'] +deps += ['mbuf', 'net', 'cryptodev', 'security', 'reorder'] diff --git a/lib/pdcp/pdcp_entity.h b/lib/pdcp/pdcp_entity.h index 28691a504b..34341cdc11 100644 --- a/lib/pdcp/pdcp_entity.h +++ b/lib/pdcp/pdcp_entity.h @@ -11,6 +11,8 @@ #include #include +#include "pdcp_reorder.h" + struct entity_priv; #define PDCP_HFN_MIN 0 @@ -109,6 +111,17 @@ union cipher_iv_partial { uint64_t u64[2]; }; +enum timer_state { + TIMER_STOP, + TIMER_RUNNING, + TIMER_EXPIRED, +}; + +struct pdcp_t_reordering { + /** Represent timer state */ + enum timer_state state; +}; + struct pdcp_cnt_bitmap { /** Number of entries that can be stored. */ uint32_t size; @@ -145,6 +158,8 @@ struct entity_priv { uint64_t is_null_auth : 1; /** Is status report required.*/ uint64_t is_status_report_required : 1; + /** Is out-of-order delivery enabled */ + uint64_t is_out_of_order_delivery : 1; } flags; /** Crypto op pool. */ struct rte_mempool *cop_pool; @@ -161,6 +176,10 @@ struct entity_priv { struct entity_priv_dl_part { /** PDCP would need to track the count values that are already received.*/ struct pdcp_cnt_bitmap bitmap; + /** t-Reordering handles */ + struct pdcp_t_reordering t_reorder; + /** Reorder packet buffer */ + struct pdcp_reorder reorder; }; struct entity_priv_ul_part { diff --git a/lib/pdcp/pdcp_process.c b/lib/pdcp/pdcp_process.c index 860ca16596..d910cb7c03 100644 --- a/lib/pdcp/pdcp_process.c +++ b/lib/pdcp/pdcp_process.c @@ -836,25 +836,88 @@ pdcp_packet_strip(struct rte_mbuf *mb, const uint32_t hdr_trim_sz, const bool tr } } -static inline bool +static inline int pdcp_post_process_update_entity_state(const struct rte_pdcp_entity *entity, - const uint32_t count) + const uint32_t count, struct rte_mbuf *mb, + struct rte_mbuf *out_mb[], + const bool trim_mac) { struct entity_priv *en_priv = entity_priv_get(entity); + struct pdcp_t_reordering *t_reorder; + struct pdcp_reorder *reorder; + uint16_t processed = 0; - if (count < en_priv->state.rx_deliv) - return false; + struct entity_priv_dl_part *dl = entity_dl_part_get(entity); + const uint32_t hdr_trim_sz = en_priv->hdr_sz + en_priv->aad_sz; - /* t-Reordering timer is not supported - SDU will be delivered immediately. -* Update RX_DELIV to the COUNT value of the first PDCP SDU which has not -* been delivered to upper layers -*/ - en_priv->state.rx_next = count + 1; + if (count < en_priv->state.rx_deliv) + return -EINVAL; if (count >= en_priv->state.rx_next) en_priv->state.rx_next = count + 1; - return true; + pdcp_packet_strip(mb, hdr_trim_sz, trim_mac); + + if (en_priv->flags.is_out_of_order_delivery) { + out_mb[0] = mb; + en_priv->state.rx_deliv = count + 1; + + return 1; + } + + reorder = &dl->reorder; + t_reorder = &dl->t_reorder; + + if (count == en_priv->state.rx_deliv) { + if (reorder->is_active) { + /* +* This insert used only to increment reorder->min_seqn +* To remove it - min_seqn_s
[PATCH v4 14/22] test/pdcp: add in-order delivery cases
From: Volodymyr Fialko Add test cases to verify behaviour when in-order delivery is enabled and packets arrive in out-of-order. PDCP library is expected to buffer the packets and return packets in-order when the missing packet arrives. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- app/test/test_pdcp.c | 223 +++ 1 file changed, 223 insertions(+) diff --git a/app/test/test_pdcp.c b/app/test/test_pdcp.c index cfe2ec6aa9..24d7826bc2 100644 --- a/app/test/test_pdcp.c +++ b/app/test/test_pdcp.c @@ -16,6 +16,15 @@ #define NB_TESTS RTE_DIM(pdcp_test_params) #define PDCP_IV_LEN 16 +/* Assert that condition is true, or goto the mark */ +#define ASSERT_TRUE_OR_GOTO(cond, mark, ...) do {\ + if (!(cond)) { \ + RTE_LOG(ERR, USER1, "Error at: %s:%d\n", __func__, __LINE__); \ + RTE_LOG(ERR, USER1, __VA_ARGS__); \ + goto mark; \ + } \ +} while (0) + /* According to formula(7.2.a Window_Size) */ #define PDCP_WINDOW_SIZE(sn_size) (1 << (sn_size - 1)) @@ -83,6 +92,38 @@ run_test_with_all_known_vec(const void *args) return run_test_foreach_known_vec(test, false); } +static int +run_test_with_all_known_vec_until_first_pass(const void *args) +{ + test_with_conf_t test = args; + + return run_test_foreach_known_vec(test, true); +} + +static inline uint32_t +pdcp_sn_mask_get(enum rte_security_pdcp_sn_size sn_size) +{ + return (1 << sn_size) - 1; +} + +static inline uint32_t +pdcp_sn_from_count_get(uint32_t count, enum rte_security_pdcp_sn_size sn_size) +{ + return (count & pdcp_sn_mask_get(sn_size)); +} + +static inline uint32_t +pdcp_hfn_mask_get(enum rte_security_pdcp_sn_size sn_size) +{ + return ~pdcp_sn_mask_get(sn_size); +} + +static inline uint32_t +pdcp_hfn_from_count_get(uint32_t count, enum rte_security_pdcp_sn_size sn_size) +{ + return (count & pdcp_hfn_mask_get(sn_size)) >> sn_size; +} + static inline int pdcp_hdr_size_get(enum rte_security_pdcp_sn_size sn_size) { @@ -416,6 +457,7 @@ create_test_conf_from_index(const int index, struct pdcp_test_conf *conf) conf->entity.sess_mpool = ts_params->sess_pool; conf->entity.cop_pool = ts_params->cop_pool; + conf->entity.ctrl_pdu_pool = ts_params->mbuf_pool; conf->entity.pdcp_xfrm.bearer = pdcp_test_bearer[index]; conf->entity.pdcp_xfrm.en_ordering = 0; conf->entity.pdcp_xfrm.remove_duplicates = 0; @@ -868,6 +910,7 @@ test_sn_range_type(enum sn_range_type type, struct pdcp_test_conf *conf) /* Configure Uplink to generate expected, encrypted packet */ pdcp_sn_to_raw_set(conf->input, new_sn, conf->entity.pdcp_xfrm.sn_size); + conf->entity.out_of_order_delivery = true; conf->entity.reverse_iv_direction = true; conf->entity.pdcp_xfrm.hfn = new_hfn; conf->entity.sn = new_sn; @@ -915,6 +958,171 @@ test_sn_minus_outside(struct pdcp_test_conf *t_conf) return test_sn_range_type(SN_RANGE_MINUS_OUTSIDE, t_conf); } +static struct rte_mbuf * +generate_packet_for_dl_with_sn(struct pdcp_test_conf ul_conf, uint32_t count) +{ + enum rte_security_pdcp_sn_size sn_size = ul_conf.entity.pdcp_xfrm.sn_size; + int ret; + + ul_conf.entity.pdcp_xfrm.hfn = pdcp_hfn_from_count_get(count, sn_size); + ul_conf.entity.sn = pdcp_sn_from_count_get(count, sn_size); + ul_conf.entity.out_of_order_delivery = true; + ul_conf.entity.reverse_iv_direction = true; + ul_conf.output_len = 0; + + ret = test_attempt_single(&ul_conf); + if (ret != TEST_SUCCESS) + return NULL; + + return mbuf_from_data_create(ul_conf.output, ul_conf.output_len); +} + +static bool +array_asc_sorted_check(struct rte_mbuf *m[], uint32_t len, enum rte_security_pdcp_sn_size sn_size) +{ + uint32_t i; + + if (len < 2) + return true; + + for (i = 0; i < (len - 1); i++) { + if (pdcp_sn_from_raw_get(rte_pktmbuf_mtod(m[i], void *), sn_size) > + pdcp_sn_from_raw_get(rte_pktmbuf_mtod(m[i + 1], void *), sn_size)) + return false; + } + + return true; +} + +static int +test_reorder_gap_fill(struct pdcp_test_conf *ul_conf) +{ + const enum rte_security_pdcp_sn_size sn_size = ul_conf->entity.pdcp_xfrm.sn_size; + struct rte_mbuf *m0 = NULL, *m1 = NULL, *out_mb[2] = {0}; + uint16_t nb_success = 0, nb_err = 0; + struct rte_pdcp_entity *pdcp_entity; + struct pdcp_test_conf dl_conf; + int ret = TEST_FAILED, nb_out; + uint8_t cdev_id; + + const int start_count = 0; + + if (ul_conf->entity.pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) + return TEST_SKIPPED; + + /* Create configuration for actual testing */ + uplink_to_downlink_convert(ul_conf, &dl_conf); + dl_conf.entity.pdcp_xfrm.hfn = pdcp_hfn_from_count_get(start_count, sn_
[PATCH v4 15/22] pdcp: add timer callback handlers
From: Volodymyr Fialko PDCP has a windowing mechanism which allows only packets that fall in a reception window. The pivot point for this window is RX_REORD which happens to be the first missing or next expected packet. If the missing packet is not received after a specified time, then the RX_REORD state variable needs to be moved up to slide the reception window. PDCP relies on timers for such operations. The timer needs to be armed when PDCP library doesn't receive all packets in-order and starts buffering packets that arrived after a missing packet. The timer needs to be cancelled when a missing packet is received. To avoid dependency on particular timer implementation, PDCP library allows application to register two callbacks, timer_start() and timer_stop() that will be called later by library. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- lib/pdcp/pdcp_entity.h | 2 ++ lib/pdcp/pdcp_process.c | 2 ++ lib/pdcp/rte_pdcp.c | 1 + lib/pdcp/rte_pdcp.h | 47 + 4 files changed, 52 insertions(+) diff --git a/lib/pdcp/pdcp_entity.h b/lib/pdcp/pdcp_entity.h index 34341cdc11..efc74ba9b9 100644 --- a/lib/pdcp/pdcp_entity.h +++ b/lib/pdcp/pdcp_entity.h @@ -120,6 +120,8 @@ enum timer_state { struct pdcp_t_reordering { /** Represent timer state */ enum timer_state state; + /** User defined callback handles */ + struct rte_pdcp_t_reordering handle; }; struct pdcp_cnt_bitmap { diff --git a/lib/pdcp/pdcp_process.c b/lib/pdcp/pdcp_process.c index d910cb7c03..4fbfc1cfb5 100644 --- a/lib/pdcp/pdcp_process.c +++ b/lib/pdcp/pdcp_process.c @@ -901,6 +901,7 @@ pdcp_post_process_update_entity_state(const struct rte_pdcp_entity *entity, if (t_reorder->state == TIMER_RUNNING && en_priv->state.rx_deliv >= en_priv->state.rx_reord) { t_reorder->state = TIMER_STOP; + t_reorder->handle.stop(t_reorder->handle.timer, t_reorder->handle.args); /* Stop reorder buffer, only if it's empty */ if (en_priv->state.rx_deliv == en_priv->state.rx_next) pdcp_reorder_stop(reorder); @@ -915,6 +916,7 @@ pdcp_post_process_update_entity_state(const struct rte_pdcp_entity *entity, en_priv->state.rx_reord = en_priv->state.rx_next; /* Start t-Reordering */ t_reorder->state = TIMER_RUNNING; + t_reorder->handle.start(t_reorder->handle.timer, t_reorder->handle.args); } return processed; diff --git a/lib/pdcp/rte_pdcp.c b/lib/pdcp/rte_pdcp.c index be37ff392c..a0558b99ae 100644 --- a/lib/pdcp/rte_pdcp.c +++ b/lib/pdcp/rte_pdcp.c @@ -56,6 +56,7 @@ pdcp_dl_establish(struct rte_pdcp_entity *entity, const struct rte_pdcp_entity_c struct entity_priv_dl_part *dl = entity_dl_part_get(entity); entity->max_pkt_cache = RTE_MAX(entity->max_pkt_cache, window_size); + dl->t_reorder.handle = conf->t_reordering; return pdcp_reorder_create(&dl->reorder, window_size); } diff --git a/lib/pdcp/rte_pdcp.h b/lib/pdcp/rte_pdcp.h index 980086a93a..05c922819e 100644 --- a/lib/pdcp/rte_pdcp.h +++ b/lib/pdcp/rte_pdcp.h @@ -67,6 +67,51 @@ struct rte_pdcp_entity { uint32_t max_pkt_cache; } __rte_cache_aligned; +/** + * Callback function type for t-Reordering timer start, set during PDCP entity establish. + * This callback is invoked by PDCP library, during t-Reordering timer start event. + * Only one t-Reordering per receiving PDCP entity would be running at a given time. + * + * @see struct rte_pdcp_timer + * @see rte_pdcp_entity_establish() + * + * @param timer + * Pointer to timer. + * @param args + * Pointer to timer arguments. + */ +typedef void (*rte_pdcp_t_reordering_start_cb_t)(void *timer, void *args); + +/** + * Callback function type for t-Reordering timer stop, set during PDCP entity establish. + * This callback will be invoked by PDCP library, during t-Reordering timer stop event. + * + * @see struct rte_pdcp_timer + * @see rte_pdcp_entity_establish() + * + * @param timer + * Pointer to timer. + * @param args + * Pointer to timer arguments. + */ +typedef void (*rte_pdcp_t_reordering_stop_cb_t)(void *timer, void *args); + +/** + * PDCP t-Reordering timer interface + * + * Configuration provided by user, that PDCP library will invoke according to timer behaviour. + */ +struct rte_pdcp_t_reordering { + /** Timer pointer, stored for later use in callback functions */ + void *timer; + /** Timer arguments, stored for later use in callback functions */ + void *args; + /** Timer start callback handle */ + rte_pdcp_t_reordering_start_cb_t start; + /** Timer stop callback handle */ + rte_pdcp_t_reordering_stop_cb_t stop; +}; + /** * PDCP entity configuration to be used for establishing an entity. */ @@ -113,6 +158,8 @@ struct rte_pdcp_entity_conf { bool stat
[PATCH v4 16/22] pdcp: add timer expiry handle
From: Volodymyr Fialko The PDCP protocol requires usage of timers to keep track of how long an out-of-order packet should be buffered while waiting for missing packets. Applications can register a desired timer implementation with the PDCP library. Once the timer expires, the application will be notified, and further handling of the event will be performed in the PDCP library. When the timer expires, the PDCP library will return the cached packets, and PDCP internal state variables (like RX_REORD, RX_DELIV etc) will be updated accordingly. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- doc/guides/prog_guide/pdcp_lib.rst | 30 ++ lib/pdcp/rte_pdcp.c| 49 ++ lib/pdcp/rte_pdcp.h| 31 +++ lib/pdcp/version.map | 2 ++ 4 files changed, 112 insertions(+) diff --git a/doc/guides/prog_guide/pdcp_lib.rst b/doc/guides/prog_guide/pdcp_lib.rst index dcb424bb1d..16deaead15 100644 --- a/doc/guides/prog_guide/pdcp_lib.rst +++ b/doc/guides/prog_guide/pdcp_lib.rst @@ -130,6 +130,36 @@ Supported integrity protection algorithms - SNOW3G-AUTH - ZUC-AUTH +Timers +-- + +PDCP utilizes a reception window mechanism to limit the bits of COUNT value +transmitted in the packet. It utilizes state variables such as RX_REORD, +RX_DELIV to define the window and uses RX_DELIV as the lower pivot point of the +window. + +RX_DELIV would be updated only when packets are received in-order. Any missing +packet would mean RX_DELIV won't be updated. A timer, t-Reordering, helps PDCP +to slide the window if the missing packet is not received in a specified time +duration. + +While starting and stopping the timer need to be done by lib PDCP, application +could register its own timer implementation. This is to make sure application +can choose between timers such as rte_timer and rte_event based timers. Starting +and stopping of timer would happen during pre & post process API. + +When the t-Reordering timer expires, application would receive the expiry event. +To perform the PDCP handling of the expiry event, +``rte_pdcp_t_reordering_expiry_handle`` can be used. Expiry handling would +involve sliding the window by updating state variables and passing the expired +packets to the application. + +.. literalinclude:: ../../../lib/pdcp/rte_pdcp.h + :language: c + :start-after: Structure rte_pdcp_t_reordering 8< + :end-before: >8 End of structure rte_pdcp_t_reordering. + + Sample API usage diff --git a/lib/pdcp/rte_pdcp.c b/lib/pdcp/rte_pdcp.c index a0558b99ae..819c66bd08 100644 --- a/lib/pdcp/rte_pdcp.c +++ b/lib/pdcp/rte_pdcp.c @@ -251,3 +251,52 @@ rte_pdcp_control_pdu_create(struct rte_pdcp_entity *pdcp_entity, return m; } + +uint16_t +rte_pdcp_t_reordering_expiry_handle(const struct rte_pdcp_entity *entity, struct rte_mbuf *out_mb[]) +{ + struct entity_priv_dl_part *dl = entity_dl_part_get(entity); + struct entity_priv *en_priv = entity_priv_get(entity); + uint16_t capacity = entity->max_pkt_cache; + uint16_t nb_out, nb_seq; + + /* 5.2.2.2 Actions when a t-Reordering expires */ + + /* +* - deliver to upper layers in ascending order of the associated COUNT value after +* performing header decompression, if not decompressed before: +*/ + + /* - all stored PDCP SDU(s) with associated COUNT value(s) < RX_REORD; */ + nb_out = pdcp_reorder_up_to_get(&dl->reorder, out_mb, capacity, en_priv->state.rx_reord); + capacity -= nb_out; + out_mb = &out_mb[nb_out]; + + /* +* - all stored PDCP SDU(s) with consecutively associated COUNT value(s) starting from +* RX_REORD; +*/ + nb_seq = pdcp_reorder_get_sequential(&dl->reorder, out_mb, capacity); + nb_out += nb_seq; + + /* +* - update RX_DELIV to the COUNT value of the first PDCP SDU which has not been delivered +* to upper layers, with COUNT value >= RX_REORD; +*/ + en_priv->state.rx_deliv = en_priv->state.rx_reord + nb_seq; + + /* +* - if RX_DELIV < RX_NEXT: +* - update RX_REORD to RX_NEXT; +* - start t-Reordering. +*/ + if (en_priv->state.rx_deliv < en_priv->state.rx_next) { + en_priv->state.rx_reord = en_priv->state.rx_next; + dl->t_reorder.state = TIMER_RUNNING; + dl->t_reorder.handle.start(dl->t_reorder.handle.timer, dl->t_reorder.handle.args); + } else { + dl->t_reorder.state = TIMER_EXPIRED; + } + + return nb_out; +} diff --git a/lib/pdcp/rte_pdcp.h b/lib/pdcp/rte_pdcp.h index 05c922819e..b926b0df29 100644 --- a/lib/pdcp/rte_pdcp.h +++ b/lib/pdcp/rte_pdcp.h @@ -101,6 +101,7 @@ typedef void (*rte_pdcp_t_reordering_stop_cb_t)(void *timer, void *args); * * Configuration provided by user, that PDCP library will invoke according to ti
[PATCH v4 17/22] test/pdcp: add timer expiry cases
From: Volodymyr Fialko Add test cases for handling the expiry with rte_timer and rte_event_timer. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- app/test/test_pdcp.c | 350 +++ 1 file changed, 350 insertions(+) diff --git a/app/test/test_pdcp.c b/app/test/test_pdcp.c index 24d7826bc2..25729b2bdd 100644 --- a/app/test/test_pdcp.c +++ b/app/test/test_pdcp.c @@ -3,15 +3,24 @@ */ #include +#ifdef RTE_LIB_EVENTDEV +#include +#include +#endif /* RTE_LIB_EVENTDEV */ #include #include #include +#include #include "test.h" #include "test_cryptodev.h" #include "test_cryptodev_security_pdcp_test_vectors.h" +#define NSECPERSEC 1E9 #define NB_DESC 1024 +#define TIMER_ADAPTER_ID 0 +#define TEST_EV_QUEUE_ID 0 +#define TEST_EV_PORT_ID 0 #define CDEV_INVALID_ID UINT8_MAX #define NB_TESTS RTE_DIM(pdcp_test_params) #define PDCP_IV_LEN 16 @@ -33,10 +42,21 @@ struct pdcp_testsuite_params { struct rte_mempool *cop_pool; struct rte_mempool *sess_pool; bool cdevs_used[RTE_CRYPTO_MAX_DEVS]; + int evdev; +#ifdef RTE_LIB_EVENTDEV + struct rte_event_timer_adapter *timdev; +#endif /* RTE_LIB_EVENTDEV */ + bool timer_is_running; + uint64_t min_resolution_ns; }; static struct pdcp_testsuite_params testsuite_params; +struct test_rte_timer_args { + int status; + struct rte_pdcp_entity *pdcp_entity; +}; + struct pdcp_test_conf { struct rte_pdcp_entity_conf entity; struct rte_crypto_sym_xform c_xfrm; @@ -124,6 +144,30 @@ pdcp_hfn_from_count_get(uint32_t count, enum rte_security_pdcp_sn_size sn_size) return (count & pdcp_hfn_mask_get(sn_size)) >> sn_size; } +static void +pdcp_timer_start_cb(void *timer, void *args) +{ + bool *is_timer_running = timer; + + RTE_SET_USED(args); + *is_timer_running = true; +} + +static void +pdcp_timer_stop_cb(void *timer, void *args) +{ + bool *is_timer_running = timer; + + RTE_SET_USED(args); + *is_timer_running = false; +} + +static struct rte_pdcp_t_reordering t_reorder_timer = { + .timer = &testsuite_params.timer_is_running, + .start = pdcp_timer_start_cb, + .stop = pdcp_timer_stop_cb, +}; + static inline int pdcp_hdr_size_get(enum rte_security_pdcp_sn_size sn_size) { @@ -462,6 +506,7 @@ create_test_conf_from_index(const int index, struct pdcp_test_conf *conf) conf->entity.pdcp_xfrm.en_ordering = 0; conf->entity.pdcp_xfrm.remove_duplicates = 0; conf->entity.pdcp_xfrm.domain = pdcp_test_params[index].domain; + conf->entity.t_reordering = t_reorder_timer; if (pdcp_test_packet_direction[index] == PDCP_DIR_UPLINK) conf->entity.pdcp_xfrm.pkt_dir = RTE_SECURITY_PDCP_UPLINK; @@ -1048,6 +1093,8 @@ test_reorder_gap_fill(struct pdcp_test_conf *ul_conf) /* Check that packets in correct order */ ASSERT_TRUE_OR_GOTO(array_asc_sorted_check(out_mb, nb_success, sn_size), exit, "Error occurred during packet drain\n"); + ASSERT_TRUE_OR_GOTO(testsuite_params.timer_is_running == false, exit, + "Timer should be stopped after full drain\n"); ret = TEST_SUCCESS; exit: @@ -1123,6 +1170,181 @@ test_reorder_buffer_full_window_size_sn_12(const struct pdcp_test_conf *ul_conf) return ret; } +#ifdef RTE_LIB_EVENTDEV +static void +event_timer_start_cb(void *timer, void *args) +{ + struct rte_event_timer *evtims = args; + int ret = 0; + + ret = rte_event_timer_arm_burst(timer, &evtims, 1); + assert(ret == 1); +} +#endif /* RTE_LIB_EVENTDEV */ + +static int +test_expiry_with_event_timer(const struct pdcp_test_conf *ul_conf) +{ +#ifdef RTE_LIB_EVENTDEV + const enum rte_security_pdcp_sn_size sn_size = ul_conf->entity.pdcp_xfrm.sn_size; + struct rte_mbuf *m1 = NULL, *out_mb[1] = {0}; + uint16_t n = 0, nb_err = 0, nb_try = 5; + struct rte_pdcp_entity *pdcp_entity; + struct pdcp_test_conf dl_conf; + int ret = TEST_FAILED, nb_out; + struct rte_event event; + + const int start_count = 0; + struct rte_event_timer evtim = { + .ev.op = RTE_EVENT_OP_NEW, + .ev.queue_id = TEST_EV_QUEUE_ID, + .ev.sched_type = RTE_SCHED_TYPE_ATOMIC, + .ev.priority = RTE_EVENT_DEV_PRIORITY_NORMAL, + .ev.event_type = RTE_EVENT_TYPE_TIMER, + .state = RTE_EVENT_TIMER_NOT_ARMED, + .timeout_ticks = 1, + }; + + if (ul_conf->entity.pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) + return TEST_SKIPPED; + + /* Create configuration for actual testing */ + uplink_to_downlink_convert(ul_conf, &dl_conf); + dl_conf.entity.pdcp_xfrm.hfn = pdcp_hfn_from_count_get(start_count, sn_size); + dl_conf.entity.sn = pdcp_sn_from_count_get(start_count, sn_size); + dl_conf.entity.t_r
[PATCH v4 18/22] test/pdcp: add timer restart case
From: Volodymyr Fialko Add test to cover the case when t-reordering timer should be restarted on the same packet. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- app/test/test_pdcp.c | 68 1 file changed, 68 insertions(+) diff --git a/app/test/test_pdcp.c b/app/test/test_pdcp.c index 25729b2bdd..82cc25ec7a 100644 --- a/app/test/test_pdcp.c +++ b/app/test/test_pdcp.c @@ -1106,6 +1106,71 @@ test_reorder_gap_fill(struct pdcp_test_conf *ul_conf) return ret; } +static int +test_reorder_gap_in_reorder_buffer(const struct pdcp_test_conf *ul_conf) +{ + const enum rte_security_pdcp_sn_size sn_size = ul_conf->entity.pdcp_xfrm.sn_size; + struct rte_mbuf *m = NULL, *out_mb[2] = {0}; + uint16_t nb_success = 0, nb_err = 0; + struct rte_pdcp_entity *pdcp_entity; + int ret = TEST_FAILED, nb_out, i; + struct pdcp_test_conf dl_conf; + uint8_t cdev_id; + + const int start_count = 0; + + if (ul_conf->entity.pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) + return TEST_SKIPPED; + + /* Create configuration for actual testing */ + uplink_to_downlink_convert(ul_conf, &dl_conf); + dl_conf.entity.pdcp_xfrm.hfn = pdcp_hfn_from_count_get(start_count, sn_size); + dl_conf.entity.sn = pdcp_sn_from_count_get(start_count, sn_size); + pdcp_entity = test_entity_create(&dl_conf, &ret); + if (pdcp_entity == NULL) + return ret; + + cdev_id = dl_conf.entity.dev_id; + + /* Create two gaps [NULL, m1, NULL, m3]*/ + for (i = 0; i < 2; i++) { + m = generate_packet_for_dl_with_sn(*ul_conf, start_count + 2 * i + 1); + ASSERT_TRUE_OR_GOTO(m != NULL, exit, "Could not allocate buffer for packet\n"); + nb_success = test_process_packets(pdcp_entity, cdev_id, &m, 1, out_mb, &nb_err); + ASSERT_TRUE_OR_GOTO(nb_err == 0, exit, "Error occurred during packet process\n"); + ASSERT_TRUE_OR_GOTO(nb_success == 0, exit, "Packet was not buffered as expected\n"); + m = NULL; /* Packet was moved to PDCP lib */ + } + + /* Generate packet to fill the first gap */ + m = generate_packet_for_dl_with_sn(*ul_conf, start_count); + ASSERT_TRUE_OR_GOTO(m != NULL, exit, "Could not allocate buffer for packet\n"); + + /* +* Buffered packets after insert [m0, m1, NULL, m3] +* Only first gap should be filled, timer should be restarted for second gap +*/ + nb_success = test_process_packets(pdcp_entity, cdev_id, &m, 1, out_mb, &nb_err); + ASSERT_TRUE_OR_GOTO(nb_err == 0, exit, "Error occurred during packet process\n"); + ASSERT_TRUE_OR_GOTO(nb_success == 2, exit, + "Packet count mismatch (received: %i, expected: 2)\n", nb_success); + m = NULL; + /* Check that packets in correct order */ + ASSERT_TRUE_OR_GOTO(array_asc_sorted_check(out_mb, nb_success, sn_size), + exit, "Error occurred during packet drain\n"); + ASSERT_TRUE_OR_GOTO(testsuite_params.timer_is_running == true, exit, + "Timer should be restarted after partial drain"); + + + ret = TEST_SUCCESS; +exit: + rte_pktmbuf_free(m); + rte_pktmbuf_free_bulk(out_mb, nb_success); + nb_out = rte_pdcp_entity_release(pdcp_entity, out_mb); + rte_pktmbuf_free_bulk(out_mb, nb_out); + return ret; +} + static int test_reorder_buffer_full_window_size_sn_12(const struct pdcp_test_conf *ul_conf) { @@ -1527,6 +1592,9 @@ static struct unit_test_suite reorder_test_cases = { TEST_CASE_NAMED_WITH_DATA("test_reorder_gap_fill", ut_setup_pdcp, ut_teardown_pdcp, run_test_with_all_known_vec, test_reorder_gap_fill), + TEST_CASE_NAMED_WITH_DATA("test_reorder_gap_in_reorder_buffer", + ut_setup_pdcp, ut_teardown_pdcp, + run_test_with_all_known_vec, test_reorder_gap_in_reorder_buffer), TEST_CASE_NAMED_WITH_DATA("test_reorder_buffer_full_window_size_sn_12", ut_setup_pdcp, ut_teardown_pdcp, run_test_with_all_known_vec_until_first_pass, -- 2.25.1
[PATCH v4 19/22] pdcp: add support for status report
From: Volodymyr Fialko Implement status report generation for PDCP entity. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- lib/pdcp/pdcp_cnt.c | 158 --- lib/pdcp/pdcp_cnt.h | 11 ++- lib/pdcp/pdcp_ctrl_pdu.c | 34 - lib/pdcp/pdcp_ctrl_pdu.h | 3 +- lib/pdcp/pdcp_entity.h | 2 + lib/pdcp/pdcp_process.c | 9 ++- lib/pdcp/pdcp_process.h | 13 lib/pdcp/rte_pdcp.c | 34 ++--- 8 files changed, 236 insertions(+), 28 deletions(-) diff --git a/lib/pdcp/pdcp_cnt.c b/lib/pdcp/pdcp_cnt.c index c9b952184b..af027b00d3 100644 --- a/lib/pdcp/pdcp_cnt.c +++ b/lib/pdcp/pdcp_cnt.c @@ -2,28 +2,164 @@ * Copyright(C) 2023 Marvell. */ +#include #include #include "pdcp_cnt.h" +#include "pdcp_ctrl_pdu.h" #include "pdcp_entity.h" +#define SLAB_BYTE_SIZE (RTE_BITMAP_SLAB_BIT_SIZE / 8) + +uint32_t +pdcp_cnt_bitmap_get_memory_footprint(const struct rte_pdcp_entity_conf *conf) +{ + uint32_t n_bits = pdcp_window_size_get(conf->pdcp_xfrm.sn_size); + + return rte_bitmap_get_memory_footprint(n_bits); +} + int -pdcp_cnt_ring_create(struct rte_pdcp_entity *en, const struct rte_pdcp_entity_conf *conf) +pdcp_cnt_bitmap_create(struct entity_priv_dl_part *dl, void *bitmap_mem, uint32_t window_size) { - struct entity_priv_dl_part *en_priv_dl; - uint32_t window_sz; + uint32_t mem_size = rte_bitmap_get_memory_footprint(window_size); - if (en == NULL || conf == NULL) + dl->bitmap.bmp = rte_bitmap_init(window_size, bitmap_mem, mem_size); + if (dl->bitmap.bmp == NULL) return -EINVAL; - if (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK) - return 0; + dl->bitmap.size = window_size; - en_priv_dl = entity_dl_part_get(en); - window_sz = pdcp_window_size_get(conf->pdcp_xfrm.sn_size); + return 0; +} - RTE_SET_USED(window_sz); - RTE_SET_USED(en_priv_dl); +void +pdcp_cnt_bitmap_set(struct pdcp_cnt_bitmap bitmap, uint32_t count) +{ + rte_bitmap_set(bitmap.bmp, count % bitmap.size); +} - return 0; +bool +pdcp_cnt_bitmap_is_set(struct pdcp_cnt_bitmap bitmap, uint32_t count) +{ + return rte_bitmap_get(bitmap.bmp, count % bitmap.size); +} + +void +pdcp_cnt_bitmap_range_clear(struct pdcp_cnt_bitmap bitmap, uint32_t start, uint32_t stop) +{ + uint32_t i; + + for (i = start; i < stop; i++) + rte_bitmap_clear(bitmap.bmp, i % bitmap.size); +} + +uint16_t +pdcp_cnt_get_bitmap_size(uint32_t pending_bytes) +{ + /* +* Round up bitmap size to slab size to operate only on slabs sizes, instead of individual +* bytes +*/ + return RTE_ALIGN_MUL_CEIL(pending_bytes, SLAB_BYTE_SIZE); +} + +static __rte_always_inline uint64_t +leftover_get(uint64_t slab, uint32_t shift, uint64_t mask) +{ + return (slab & mask) << shift; +} + +void +pdcp_cnt_report_fill(struct pdcp_cnt_bitmap bitmap, struct entity_state state, +uint8_t *data, uint16_t data_len) +{ + uint64_t slab = 0, next_slab = 0, leftover; + uint32_t zeros, report_len, diff; + uint32_t slab_id, next_slab_id; + uint32_t pos = 0, next_pos = 0; + + const uint32_t start_count = state.rx_deliv + 1; + const uint32_t nb_slabs = bitmap.size / RTE_BITMAP_SLAB_BIT_SIZE; + const uint32_t nb_data_slabs = data_len / SLAB_BYTE_SIZE; + const uint32_t start_slab_id = start_count / RTE_BITMAP_SLAB_BIT_SIZE; + const uint32_t stop_slab_id = (start_slab_id + nb_data_slabs) % nb_slabs; + const uint32_t shift = start_count % RTE_BITMAP_SLAB_BIT_SIZE; + const uint32_t leftover_shift = shift ? RTE_BITMAP_SLAB_BIT_SIZE - shift : 0; + const uint8_t *data_end = RTE_PTR_ADD(data, data_len + SLAB_BYTE_SIZE); + + /* NOTE: Mask required to workaround case - when shift is not needed */ + const uint64_t leftover_mask = shift ? ~0 : 0; + + /* NOTE: implement scan init at to set custom position */ + __rte_bitmap_scan_init(bitmap.bmp); + while (true) { + assert(rte_bitmap_scan(bitmap.bmp, &pos, &slab) == 1); + slab_id = pos / RTE_BITMAP_SLAB_BIT_SIZE; + if (slab_id >= start_slab_id) + break; + } + + report_len = nb_data_slabs; + + if (slab_id > start_slab_id) { + /* Zero slabs at beginning */ + zeros = (slab_id - start_slab_id - 1) * SLAB_BYTE_SIZE; + memset(data, 0, zeros); + data = RTE_PTR_ADD(data, zeros); + leftover = leftover_get(slab, leftover_shift, leftover_mask); + memcpy(data, &leftover, SLAB_BYTE_SIZE); + data = RTE_PTR_ADD(data, SLAB_BYTE_SIZE); + report_len -= (slab_id - start_slab_id); + } + + while (report_len) { + rte_bitmap_scan(bitmap.bmp, &next_pos, &next_slab); +
[PATCH v4 20/22] pdcp: allocate reorder buffer alongside with entity
From: Volodymyr Fialko Instead of allocating reorder buffer separately on heap, allocate memory for it together with rest of entity, and then only initialize buffer via `rte_reorder_init()`. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- lib/pdcp/pdcp_cnt.c | 9 +++ lib/pdcp/pdcp_cnt.h | 3 ++- lib/pdcp/pdcp_entity.h | 2 +- lib/pdcp/pdcp_reorder.c | 11 ++-- lib/pdcp/pdcp_reorder.h | 12 ++--- lib/pdcp/rte_pdcp.c | 58 ++--- 6 files changed, 55 insertions(+), 40 deletions(-) diff --git a/lib/pdcp/pdcp_cnt.c b/lib/pdcp/pdcp_cnt.c index af027b00d3..e1d0634b4d 100644 --- a/lib/pdcp/pdcp_cnt.c +++ b/lib/pdcp/pdcp_cnt.c @@ -20,15 +20,14 @@ pdcp_cnt_bitmap_get_memory_footprint(const struct rte_pdcp_entity_conf *conf) } int -pdcp_cnt_bitmap_create(struct entity_priv_dl_part *dl, void *bitmap_mem, uint32_t window_size) +pdcp_cnt_bitmap_create(struct entity_priv_dl_part *dl, uint32_t nb_elem, + void *bitmap_mem, uint32_t mem_size) { - uint32_t mem_size = rte_bitmap_get_memory_footprint(window_size); - - dl->bitmap.bmp = rte_bitmap_init(window_size, bitmap_mem, mem_size); + dl->bitmap.bmp = rte_bitmap_init(nb_elem, bitmap_mem, mem_size); if (dl->bitmap.bmp == NULL) return -EINVAL; - dl->bitmap.size = window_size; + dl->bitmap.size = nb_elem; return 0; } diff --git a/lib/pdcp/pdcp_cnt.h b/lib/pdcp/pdcp_cnt.h index 5941b7a406..87b011f9dc 100644 --- a/lib/pdcp/pdcp_cnt.h +++ b/lib/pdcp/pdcp_cnt.h @@ -10,7 +10,8 @@ #include "pdcp_entity.h" uint32_t pdcp_cnt_bitmap_get_memory_footprint(const struct rte_pdcp_entity_conf *conf); -int pdcp_cnt_bitmap_create(struct entity_priv_dl_part *dl, void *bitmap_mem, uint32_t window_size); +int pdcp_cnt_bitmap_create(struct entity_priv_dl_part *dl, uint32_t nb_elem, + void *bitmap_mem, uint32_t mem_size); void pdcp_cnt_bitmap_set(struct pdcp_cnt_bitmap bitmap, uint32_t count); bool pdcp_cnt_bitmap_is_set(struct pdcp_cnt_bitmap bitmap, uint32_t count); diff --git a/lib/pdcp/pdcp_entity.h b/lib/pdcp/pdcp_entity.h index a9b1428c7a..9f74b5d0e5 100644 --- a/lib/pdcp/pdcp_entity.h +++ b/lib/pdcp/pdcp_entity.h @@ -132,7 +132,7 @@ struct pdcp_cnt_bitmap { }; /* - * Layout of PDCP entity: [rte_pdcp_entity] [entity_priv] [entity_dl/ul] + * Layout of PDCP entity: [rte_pdcp_entity] [entity_priv] [entity_dl/ul] [reorder/bitmap] */ struct entity_priv { diff --git a/lib/pdcp/pdcp_reorder.c b/lib/pdcp/pdcp_reorder.c index 5399f0dc28..bc45f2e19b 100644 --- a/lib/pdcp/pdcp_reorder.c +++ b/lib/pdcp/pdcp_reorder.c @@ -8,20 +8,13 @@ #include "pdcp_reorder.h" int -pdcp_reorder_create(struct pdcp_reorder *reorder, uint32_t window_size) +pdcp_reorder_create(struct pdcp_reorder *reorder, size_t nb_elem, void *mem, size_t mem_size) { - reorder->buf = rte_reorder_create("reorder_buffer", SOCKET_ID_ANY, window_size); + reorder->buf = rte_reorder_init(mem, mem_size, "reorder_buffer", nb_elem); if (reorder->buf == NULL) return -rte_errno; - reorder->window_size = window_size; reorder->is_active = false; return 0; } - -void -pdcp_reorder_destroy(const struct pdcp_reorder *reorder) -{ - rte_reorder_free(reorder->buf); -} diff --git a/lib/pdcp/pdcp_reorder.h b/lib/pdcp/pdcp_reorder.h index 6a2f61d6ae..7e4f079d4b 100644 --- a/lib/pdcp/pdcp_reorder.h +++ b/lib/pdcp/pdcp_reorder.h @@ -9,12 +9,18 @@ struct pdcp_reorder { struct rte_reorder_buffer *buf; - uint32_t window_size; bool is_active; }; -int pdcp_reorder_create(struct pdcp_reorder *reorder, uint32_t window_size); -void pdcp_reorder_destroy(const struct pdcp_reorder *reorder); +int pdcp_reorder_create(struct pdcp_reorder *reorder, size_t nb_elem, void *mem, size_t mem_size); + +/* NOTE: replace with `rte_reorder_memory_footprint_get` after DPDK 23.07 */ +#define SIZE_OF_REORDER_BUFFER (4 * RTE_CACHE_LINE_SIZE) +static inline size_t +pdcp_reorder_memory_footprint_get(size_t nb_elem) +{ + return SIZE_OF_REORDER_BUFFER + (2 * nb_elem * sizeof(struct rte_mbuf *)); +} static inline uint32_t pdcp_reorder_get_sequential(struct pdcp_reorder *reorder, struct rte_mbuf **mbufs, diff --git a/lib/pdcp/rte_pdcp.c b/lib/pdcp/rte_pdcp.c index 9865c620b7..1c6d2466b2 100644 --- a/lib/pdcp/rte_pdcp.c +++ b/lib/pdcp/rte_pdcp.c @@ -14,7 +14,15 @@ #define RTE_PDCP_DYNFIELD_NAME "rte_pdcp_dynfield" -static int bitmap_mem_offset; +struct entity_layout { + size_t bitmap_offset; + size_t bitmap_size; + + size_t reorder_buf_offset; + size_t reorder_buf_size; + + size_t total_size; +}; int rte_pdcp_dynfield_offset = -1; @@ -35,46 +43,54 @@ pdcp_dynfield_register(void) } static int -pdcp_entity_size_get(const struct rte_pdcp_entity_conf *conf) +pdcp_entity_layout_get(const struct rte_pdcp_entity_conf *conf, s
[PATCH v4 21/22] pdcp: add thread safe processing
From: Volodymyr Fialko PDCP state has to be guarded for: - Uplink pre_process: - tx_next atomic increment - Downlink pre_process: - rx_deliv - read - Downlink post_process: - rx_deliv, rx_reorder, rx_next - read/write - bitmask/reorder buffer - read/write When application requires thread safe processing, the state variables need to be updated atomically. Add config option to select this option per entity. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- lib/pdcp/meson.build| 2 + lib/pdcp/pdcp_entity.h | 64 ++ lib/pdcp/pdcp_process.c | 117 +--- lib/pdcp/rte_pdcp.c | 9 lib/pdcp/rte_pdcp.h | 2 + 5 files changed, 175 insertions(+), 19 deletions(-) diff --git a/lib/pdcp/meson.build b/lib/pdcp/meson.build index f4f9246bcb..c4f135778e 100644 --- a/lib/pdcp/meson.build +++ b/lib/pdcp/meson.build @@ -19,3 +19,5 @@ headers = files('rte_pdcp.h') indirect_headers += files('rte_pdcp_group.h') deps += ['mbuf', 'net', 'cryptodev', 'security', 'reorder'] + +annotate_locks = false diff --git a/lib/pdcp/pdcp_entity.h b/lib/pdcp/pdcp_entity.h index 9f74b5d0e5..e88af4c3b3 100644 --- a/lib/pdcp/pdcp_entity.h +++ b/lib/pdcp/pdcp_entity.h @@ -9,6 +9,7 @@ #include #include #include +#include #include #include "pdcp_reorder.h" @@ -162,6 +163,8 @@ struct entity_priv { uint64_t is_status_report_required : 1; /** Is out-of-order delivery enabled */ uint64_t is_out_of_order_delivery : 1; + /** Is thread safety enabled. */ + uint64_t is_thread_safety_enabled : 1; } flags; /** Crypto op pool. */ struct rte_mempool *cop_pool; @@ -182,6 +185,8 @@ struct entity_priv_dl_part { struct pdcp_t_reordering t_reorder; /** Reorder packet buffer */ struct pdcp_reorder reorder; + /* Lock to protect concurrent updates */ + rte_rwlock_t rwl; /** Bitmap memory region */ uint8_t bitmap_mem[0]; }; @@ -257,4 +262,63 @@ pdcp_hfn_max(enum rte_security_pdcp_sn_size sn_size) return (1 << (32 - sn_size)) - 1; } +static inline uint32_t +pdcp_atomic_inc(uint32_t *val, const bool mt_safe) +{ + if (mt_safe) + return __atomic_fetch_add(val, 1, __ATOMIC_RELAXED); + else + return (*val)++; +} + +static inline void +pdcp_lock_init(const struct rte_pdcp_entity *entity) +{ + struct entity_priv_dl_part *dl = entity_dl_part_get(entity); + struct entity_priv *en_priv = entity_priv_get(entity); + + if (en_priv->flags.is_thread_safety_enabled) + rte_rwlock_init(&dl->rwl); +} + +static inline void +pdcp_read_lock(const struct rte_pdcp_entity *entity) +{ + struct entity_priv_dl_part *dl = entity_dl_part_get(entity); + struct entity_priv *en_priv = entity_priv_get(entity); + + if (en_priv->flags.is_thread_safety_enabled) + rte_rwlock_read_lock(&dl->rwl); +} + +static inline void +pdcp_read_unlock(const struct rte_pdcp_entity *entity) +{ + struct entity_priv_dl_part *dl = entity_dl_part_get(entity); + struct entity_priv *en_priv = entity_priv_get(entity); + + if (en_priv->flags.is_thread_safety_enabled) + rte_rwlock_read_unlock(&dl->rwl); +} + +static inline void +pdcp_write_lock(const struct rte_pdcp_entity *entity) +{ + struct entity_priv_dl_part *dl = entity_dl_part_get(entity); + struct entity_priv *en_priv = entity_priv_get(entity); + + if (en_priv->flags.is_thread_safety_enabled) + rte_rwlock_write_lock(&dl->rwl); +} + +static inline void +pdcp_write_unlock(const struct rte_pdcp_entity *entity) +{ + struct entity_priv_dl_part *dl = entity_dl_part_get(entity); + struct entity_priv *en_priv = entity_priv_get(entity); + + if (en_priv->flags.is_thread_safety_enabled) + rte_rwlock_write_unlock(&dl->rwl); +} + #endif /* PDCP_ENTITY_H */ diff --git a/lib/pdcp/pdcp_process.c b/lib/pdcp/pdcp_process.c index c093b78e10..aa13801d5d 100644 --- a/lib/pdcp/pdcp_process.c +++ b/lib/pdcp/pdcp_process.c @@ -357,7 +357,7 @@ cop_prepare(const struct entity_priv *en_priv, struct rte_mbuf *mb, struct rte_c static inline bool pdcp_pre_process_uplane_sn_12_ul_set_sn(struct entity_priv *en_priv, struct rte_mbuf *mb, - uint32_t *count) + uint32_t *count, const bool mt_safe) { struct rte_pdcp_up_data_pdu_sn_12_hdr *pdu_hdr; const uint8_t hdr_sz = en_priv->hdr_sz; @@ -369,7 +369,7 @@ pdcp_pre_process_uplane_sn_12_ul_set_sn(struct entity_priv *en_priv, struct rte_ return false; /* Update sequence num in the PDU header */ - *count = en_priv->state.tx_next++; + *count = pdcp_atomic_inc(&en_priv->state.tx_next, mt_safe); sn = pdcp_sn_from_count_get(*count, RTE_
[PATCH v4 22/22] test/pdcp: add PDCP status report cases
From: Volodymyr Fialko Test PDCP status report generation. Signed-off-by: Anoob Joseph Signed-off-by: Volodymyr Fialko --- app/test/test_pdcp.c | 312 +++ 1 file changed, 312 insertions(+) diff --git a/app/test/test_pdcp.c b/app/test/test_pdcp.c index 82cc25ec7a..423526380f 100644 --- a/app/test/test_pdcp.c +++ b/app/test/test_pdcp.c @@ -2,6 +2,7 @@ * Copyright(C) 2023 Marvell. */ +#include #include #ifdef RTE_LIB_EVENTDEV #include @@ -48,6 +49,9 @@ struct pdcp_testsuite_params { #endif /* RTE_LIB_EVENTDEV */ bool timer_is_running; uint64_t min_resolution_ns; + struct rte_pdcp_up_ctrl_pdu_hdr *status_report; + uint32_t status_report_bitmask_capacity; + uint8_t *ctrl_pdu_buf; }; static struct pdcp_testsuite_params testsuite_params; @@ -168,6 +172,18 @@ static struct rte_pdcp_t_reordering t_reorder_timer = { .stop = pdcp_timer_stop_cb, }; +static inline void +bitmask_set_bit(uint8_t *mask, uint32_t bit) +{ + mask[bit / 8] |= (1 << bit % 8); +} + +static inline bool +bitmask_is_bit_set(const uint8_t *mask, uint32_t bit) +{ + return mask[bit / 8] & (1 << (bit % 8)); +} + static inline int pdcp_hdr_size_get(enum rte_security_pdcp_sn_size sn_size) { @@ -314,6 +330,21 @@ testsuite_setup(void) goto cop_pool_free; } + /* Allocate memory for longest possible status report */ + ts_params->status_report_bitmask_capacity = RTE_PDCP_CTRL_PDU_SIZE_MAX - + sizeof(struct rte_pdcp_up_ctrl_pdu_hdr); + ts_params->status_report = rte_zmalloc(NULL, RTE_PDCP_CTRL_PDU_SIZE_MAX, 0); + if (ts_params->status_report == NULL) { + RTE_LOG(ERR, USER1, "Could not allocate status report\n"); + goto cop_pool_free; + } + + ts_params->ctrl_pdu_buf = rte_zmalloc(NULL, RTE_PDCP_CTRL_PDU_SIZE_MAX, 0); + if (ts_params->ctrl_pdu_buf == NULL) { + RTE_LOG(ERR, USER1, "Could not allocate status report data\n"); + goto cop_pool_free; + } + return 0; cop_pool_free: @@ -322,6 +353,8 @@ testsuite_setup(void) mbuf_pool_free: rte_mempool_free(ts_params->mbuf_pool); ts_params->mbuf_pool = NULL; + rte_free(ts_params->status_report); + rte_free(ts_params->ctrl_pdu_buf); return TEST_FAILED; } @@ -344,6 +377,9 @@ testsuite_teardown(void) rte_mempool_free(ts_params->mbuf_pool); ts_params->mbuf_pool = NULL; + + rte_free(ts_params->status_report); + rte_free(ts_params->ctrl_pdu_buf); } static int @@ -1410,6 +1446,246 @@ test_expiry_with_rte_timer(const struct pdcp_test_conf *ul_conf) return ret; } +static struct rte_pdcp_up_ctrl_pdu_hdr * +pdcp_status_report_init(uint32_t fmc) +{ + struct rte_pdcp_up_ctrl_pdu_hdr *hdr = testsuite_params.status_report; + + hdr->d_c = RTE_PDCP_PDU_TYPE_CTRL; + hdr->pdu_type = RTE_PDCP_CTRL_PDU_TYPE_STATUS_REPORT; + hdr->fmc = rte_cpu_to_be_32(fmc); + hdr->r = 0; + memset(hdr->bitmap, 0, testsuite_params.status_report_bitmask_capacity); + + return hdr; +} + +static uint32_t +pdcp_status_report_len(void) +{ + struct rte_pdcp_up_ctrl_pdu_hdr *hdr = testsuite_params.status_report; + uint32_t i; + + for (i = testsuite_params.status_report_bitmask_capacity; i != 0; i--) { + if (hdr->bitmap[i - 1]) + return i; + } + + return 0; +} + +static int +pdcp_status_report_verify(struct rte_mbuf *status_report, +const struct rte_pdcp_up_ctrl_pdu_hdr *expected_hdr, uint32_t expected_len) +{ + uint32_t received_len = rte_pktmbuf_pkt_len(status_report); + uint8_t *received_buf = testsuite_params.ctrl_pdu_buf; + int ret; + + ret = pktmbuf_read_into(status_report, received_buf, RTE_PDCP_CTRL_PDU_SIZE_MAX); + TEST_ASSERT_SUCCESS(ret, "Failed to copy status report pkt into continuous buffer"); + + debug_hexdump(stdout, "Received:", received_buf, received_len); + debug_hexdump(stdout, "Expected:", expected_hdr, expected_len); + + TEST_ASSERT_EQUAL(expected_len, received_len, + "Mismatch in packet lengths [expected: %d, received: %d]", + expected_len, received_len); + + TEST_ASSERT_BUFFERS_ARE_EQUAL(received_buf, expected_hdr, expected_len, +"Generated packet not as expected"); + + return 0; +} + +static int +test_status_report_gen(const struct pdcp_test_conf *ul_conf, + const struct rte_pdcp_up_ctrl_pdu_hdr *hdr, + uint32_t bitmap_len) +{ + const enum rte_security_pdcp_sn_size sn_size = ul_conf->entity.pdcp_xfrm.sn_size; + struct rte_mbuf *status_report = NULL, **out_mb, *m; + uint16_t nb_success = 0, nb_err = 0; + struct rte_pdcp_entity *pdcp_entity; +
Re: [PATCH v4 21/22] pdcp: add thread safe processing
On Sat, 27 May 2023 02:32:13 +0530 Anoob Joseph wrote: > +static inline uint32_t > +pdcp_atomic_inc(uint32_t *val, const bool mt_safe) > +{ > + if (mt_safe) > + return __atomic_fetch_add(val, 1, __ATOMIC_RELAXED); > + else > + return (*val)++; > +} This is a bad pattern. None of the rest of DPDK does this. Either be thread safe or not.
Re: Generic flow string parser
On Sat, Apr 29, 2023 at 2:49 PM Cliff Burdick wrote: > > So the only things we need are 2 functions, if I understand well: > > > > int rte_flow_to_text(const struct rte_flow*); > > struct rte_flow *rte_flow_from_text(const char *); > > > > Here I assume the output of rte_flow_from_text() would be a created flow, > > meaning it calls rte_flow_create() under the hood. > > Is it what you wish? > > Or should it fill port ID, attributes, patterns and actions? > > > > I think it should follow closely with what "flow_parse" already does: > https://github.com/DPDK/dpdk/blob/d03446724972d2a1bb645ce7f3e64f5ef0203d61/app/test-pmd/cmdline_flow.c#L11304 > > > > Namely, just do the part of populating attributes, patterns, and actions > from a string. It's then up to the user if they want to create, validate, > or do something else with it (like see how it populate> d the structures). > The flow_parse function appears to take an opaque pointer that's specific > to a structure inside of cmdline_flow.c and assign the attributes, actions, > and patterns to members of that result struct. I don't know the reason for > this, but when >calling the function the user doesn't know how many > patterns or actions their string will generate. They would either need to > pass in structures that are allocated larger than needed, or have a > separate API that returns how many actions and patterns are needed for a > string, then they need to allocate the correct size themselves. I'm > assuming it's not ideal to have the library itself do dynamic memory > allocations for the correct size. > >> >> Tom, for what it's worth I'm on a quest to get this to work since I think it's necessary. I'm just hacking through it like you did and I ran into the same "template" keyword error. It's probably worthwhile to fix that anyways. I'm maintaining a set of patches as I go. The general strategy has been to remove testpmd's main function, compile it as a library, and link against that.
Re: [PATCH v3] dmadev: add tracepoints
Hi Thomas, On 2023/5/25 5:12, Thomas Monjalon wrote: > 15/04/2023 02:33, Chengwen Feng: >> Add tracepoints at important APIs for tracing support. >> >> Signed-off-by: Chengwen Feng >> Acked-by: Morten Brørup > > This patch triggers a failure in Intel CI: > asan_smoke | test_rxtx_with_ASan_enable| FAILED > > If you think it is a false positive, > please could you submit it again as v4? V4 already sent, and I observed the V4 fix it in patchwork. Thanks > > > . >
Re: [PATCH V5 2/5] ethdev: fix skip valid port in probing callback
在 2023/5/22 19:04, fengchengwen 写道: On 2023/1/31 11:33, Huisong Li wrote: The event callback in application may use the macro RTE_ETH_FOREACH_DEV to iterate over all enabled ports to do something(like, verifying the port id validity) when receive a probing event. If the ethdev state of a port is not RTE_ETH_DEV_UNUSED, this port will be considered as a valid port. However, this state is set to RTE_ETH_DEV_ATTACHED after pushing probing event. It means that probing callback will skip this port. But this assignment can not move to front of probing notification. See commit be8cd210379a ("ethdev: fix port probing notification") ... struct rte_eth_dev_sriov { diff --git a/lib/ethdev/version.map b/lib/ethdev/version.map index 17201fbe0f..094c2a952e 100644 --- a/lib/ethdev/version.map +++ b/lib/ethdev/version.map @@ -327,4 +327,5 @@ INTERNAL { rte_eth_representor_id_get; rte_eth_switch_domain_alloc; rte_eth_switch_domain_free; + rte_eth_dev_is_used; requires alphabetical order. Thanks, will fix it in v6. }; .