Hi Anoob, Please see inline comments. > Subject: [PATCH 2/5] lib: add pdcp protocol > > Add Packet Data Convergence Protocol (PDCP) processing library. > > The library is similar to lib_ipsec which provides IPsec processing > capabilities in DPDK. > > PDCP would involve roughly the following options, > 1. Transfer of user plane data > 2. Transfer of control plane data > 3. Header compression > 4. Uplink data compression > 5. Ciphering and integrity protection > > PDCP library provides following control path APIs that is used to > configure various PDCP entities, > 1. rte_pdcp_entity_establish() > 2. rte_pdcp_entity_suspend() > 3. rte_pdcp_entity_release() > > PDCP process is split into 2 parts. One before crypto processing > (rte_pdcp_pkt_pre_process()) and one after crypto processing > (rte_pdcp_pkt_post_process()). Since cryptodev dequeue can return crypto > operations belonging to multiple entities, rte_pdcp_pkt_crypto_group() > is added to help grouping crypto operations belonging to same entity. > > Signed-off-by: Anoob Joseph <ano...@marvell.com> > Signed-off-by: Kiran Kumar K <kirankum...@marvell.com> > Signed-off-by: Volodymyr Fialko <vfia...@marvell.com> > --- > doc/api/doxy-api-index.md | 3 +- > doc/api/doxy-api.conf.in | 1 + > lib/meson.build | 1 + > lib/pdcp/meson.build | 8 + > lib/pdcp/pdcp_crypto.c | 240 ++++++++ > lib/pdcp/pdcp_crypto.h | 20 + > lib/pdcp/pdcp_entity.h | 218 +++++++ > lib/pdcp/pdcp_process.c | 1195 +++++++++++++++++++++++++++++++++++++ > lib/pdcp/pdcp_process.h | 13 + > lib/pdcp/rte_pdcp.c | 136 +++++ > lib/pdcp/rte_pdcp.h | 263 ++++++++ > lib/pdcp/rte_pdcp_group.h | 133 +++++ > lib/pdcp/version.map | 13 + > 13 files changed, 2243 insertions(+), 1 deletion(-) > create mode 100644 lib/pdcp/meson.build > create mode 100644 lib/pdcp/pdcp_crypto.c > create mode 100644 lib/pdcp/pdcp_crypto.h > create mode 100644 lib/pdcp/pdcp_entity.h > create mode 100644 lib/pdcp/pdcp_process.c > create mode 100644 lib/pdcp/pdcp_process.h > create mode 100644 lib/pdcp/rte_pdcp.c > create mode 100644 lib/pdcp/rte_pdcp.h > create mode 100644 lib/pdcp/rte_pdcp_group.h > create mode 100644 lib/pdcp/version.map > > diff --git a/doc/api/doxy-api-index.md b/doc/api/doxy-api-index.md > index ae4b107240..6014bee079 100644 > --- a/doc/api/doxy-api-index.md > +++ b/doc/api/doxy-api-index.md > @@ -126,7 +126,8 @@ The public API headers are grouped by topics: > [eCPRI](@ref rte_ecpri.h), > [L2TPv2](@ref rte_l2tpv2.h), > [PPP](@ref rte_ppp.h), > - [PDCP hdr](@ref rte_pdcp_hdr.h) > + [PDCP hdr](@ref rte_pdcp_hdr.h), > + [PDCP](@ref rte_pdcp.h), > > - **QoS**: > [metering](@ref rte_meter.h), > diff --git a/doc/api/doxy-api.conf.in b/doc/api/doxy-api.conf.in > index f0886c3bd1..01314b087e 100644 > --- a/doc/api/doxy-api.conf.in > +++ b/doc/api/doxy-api.conf.in > @@ -61,6 +61,7 @@ INPUT = @TOPDIR@/doc/api/doxy-api- > index.md \ > @TOPDIR@/lib/net \ > @TOPDIR@/lib/pcapng \ > @TOPDIR@/lib/pci \ > + @TOPDIR@/lib/pdcp \ > @TOPDIR@/lib/pdump \ > @TOPDIR@/lib/pipeline \ > @TOPDIR@/lib/port \ > diff --git a/lib/meson.build b/lib/meson.build > index fd55925340..a827006d29 100644 > --- a/lib/meson.build > +++ b/lib/meson.build > @@ -63,6 +63,7 @@ libraries = [ > 'flow_classify', # flow_classify lib depends on pkt framework table > lib > 'graph', > 'node', > + 'pdcp', # pdcp lib depends on crypto and security > ] > > optional_libs = [ > diff --git a/lib/pdcp/meson.build b/lib/pdcp/meson.build > new file mode 100644 > index 0000000000..a7f5a408cf > --- /dev/null > +++ b/lib/pdcp/meson.build > @@ -0,0 +1,8 @@ > +# SPDX-License-Identifier: BSD-3-Clause > +# Copyright(C) 2022 Marvell. > +# Extra # here.
Do we support compilation on Windows as well? Check missing here. > + > +sources = files('pdcp_crypto.c', 'pdcp_process.c', 'rte_pdcp.c') > +headers = files('rte_pdcp.h') Do we need to add the indirect header as well for lib/pdcp/rte_pdcp_group.h? > + > +deps += ['security'] Crypto not needed as dependency? > diff --git a/lib/pdcp/pdcp_crypto.c b/lib/pdcp/pdcp_crypto.c > new file mode 100644 > index 0000000000..7ffb8a07a7 > --- /dev/null > +++ b/lib/pdcp/pdcp_crypto.c > @@ -0,0 +1,240 @@ > +/* SPDX-License-Identifier: BSD-3-Clause > + * Copyright(C) 2022 Marvell. > + */ > + > +#include <rte_crypto.h> > +#include <rte_crypto_sym.h> > +#include <rte_cryptodev.h> > +#include <rte_pdcp.h> > + > +#include "pdcp_crypto.h" > +#include "pdcp_entity.h" > + > +static int > +pdcp_crypto_caps_cipher_verify(uint8_t dev_id, const struct > rte_crypto_sym_xform *c_xfrm) > +{ > + const struct rte_cryptodev_symmetric_capability *cap; > + struct rte_cryptodev_sym_capability_idx cap_idx; > + int ret; > + > + cap_idx.type = RTE_CRYPTO_SYM_XFORM_CIPHER; > + cap_idx.algo.cipher = c_xfrm->cipher.algo; > + > + cap = rte_cryptodev_sym_capability_get(dev_id, &cap_idx); > + if (cap == NULL) > + return -1; > + > + ret = rte_cryptodev_sym_capability_check_cipher(cap, c_xfrm- > >cipher.key.length, > + c_xfrm- > >cipher.iv.length); > + > + return ret; > +} > + > +static int > +pdcp_crypto_caps_auth_verify(uint8_t dev_id, const struct > rte_crypto_sym_xform *a_xfrm) > +{ > + const struct rte_cryptodev_symmetric_capability *cap; > + struct rte_cryptodev_sym_capability_idx cap_idx; > + int ret; > + > + cap_idx.type = RTE_CRYPTO_SYM_XFORM_AUTH; > + cap_idx.algo.auth = a_xfrm->auth.algo; > + > + cap = rte_cryptodev_sym_capability_get(dev_id, &cap_idx); > + if (cap == NULL) > + return -1; > + > + ret = rte_cryptodev_sym_capability_check_auth(cap, a_xfrm- > >auth.key.length, > + a_xfrm- > >auth.digest_length, > + a_xfrm->auth.iv.length); > + > + return ret; > +} > + > +static int > +pdcp_crypto_xfrm_validate(const struct rte_pdcp_entity_conf *conf, > + const struct rte_crypto_sym_xform *c_xfrm, > + const struct rte_crypto_sym_xform *a_xfrm, > + bool is_auth_then_cipher) > +{ > + uint16_t ciph_iv_len, auth_digest_len, auth_iv_len; > + int ret; > + > + /* > + * Uplink means PDCP entity is configured for transmit. Downlink means > PDCP entity is > + * configured for receive. When integrity protection is enabled, PDCP > always performs > + * digest-encrypted or auth-gen-encrypt for uplink (and decrypt-auth- > verify for downlink). > + * So for uplink, crypto chain would be auth-cipher while for downlink > it > would be > + * cipher-auth. > + * > + * When integrity protection is not required, xform would be cipher > only. > + */ > + > + if (c_xfrm == NULL) > + return -EINVAL; > + > + if (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK) { > + > + /* With UPLINK, if auth is enabled, it should be before cipher > */ > + if (a_xfrm != NULL && !is_auth_then_cipher) > + return -EINVAL; > + > + /* With UPLINK, cipher operation must be encrypt */ > + if (c_xfrm->cipher.op != RTE_CRYPTO_CIPHER_OP_ENCRYPT) > + return -EINVAL; > + > + /* With UPLINK, auth operation (if present) must be generate */ > + if (a_xfrm != NULL && a_xfrm->auth.op != > RTE_CRYPTO_AUTH_OP_GENERATE) > + return -EINVAL; > + > + } else if (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) > { > + > + /* With DOWNLINK, if auth is enabled, it should be after cipher > */ > + if (a_xfrm != NULL && is_auth_then_cipher) > + return -EINVAL; > + > + /* With DOWNLINK, cipher operation must be decrypt */ > + if (c_xfrm->cipher.op != RTE_CRYPTO_CIPHER_OP_DECRYPT) > + return -EINVAL; > + > + /* With DOWNLINK, auth operation (if present) must be verify > */ > + if (a_xfrm != NULL && a_xfrm->auth.op != > RTE_CRYPTO_AUTH_OP_VERIFY) > + return -EINVAL; > + > + } else { > + return -EINVAL; > + } > + > + if ((c_xfrm->cipher.algo != RTE_CRYPTO_CIPHER_NULL) && > + (c_xfrm->cipher.algo != RTE_CRYPTO_CIPHER_AES_CTR) && > + (c_xfrm->cipher.algo != RTE_CRYPTO_CIPHER_ZUC_EEA3) && > + (c_xfrm->cipher.algo != RTE_CRYPTO_CIPHER_SNOW3G_UEA2)) > + return -EINVAL; > + > + if (c_xfrm->cipher.algo == RTE_CRYPTO_CIPHER_NULL) > + ciph_iv_len = 0; > + else > + ciph_iv_len = PDCP_IV_LENGTH; > + > + if (ciph_iv_len != c_xfrm->cipher.iv.length) > + return -EINVAL; > + > + if (a_xfrm != NULL) { > + if ((a_xfrm->auth.algo != RTE_CRYPTO_AUTH_NULL) && > + (a_xfrm->auth.algo != RTE_CRYPTO_AUTH_AES_CMAC) && > + (a_xfrm->auth.algo != RTE_CRYPTO_AUTH_ZUC_EIA3) && > + (a_xfrm->auth.algo != RTE_CRYPTO_AUTH_SNOW3G_UIA2)) > + return -EINVAL; > + > + if (a_xfrm->auth.algo == RTE_CRYPTO_AUTH_NULL) > + auth_digest_len = 0; > + else > + auth_digest_len = 4; If we have a macro for IV length, why not for digest also? Moreover, for NULL integrity, digest length is also 4 with all 0s. Refer Annex D.1 in https://www.etsi.org/deliver/etsi_ts/133500_133599/133501/15.04.00_60/ts_133501v150400p.pdf Digest len would be 0 only in case of a_xfrm == NULL > + > + if (auth_digest_len != a_xfrm->auth.digest_length) > + return -EINVAL; > + > + if ((a_xfrm->auth.algo == RTE_CRYPTO_AUTH_ZUC_EIA3) || > + (a_xfrm->auth.algo == RTE_CRYPTO_AUTH_SNOW3G_UIA2)) > + auth_iv_len = PDCP_IV_LENGTH; > + else > + auth_iv_len = 0; > + > + if (a_xfrm->auth.iv.length != auth_iv_len) > + return -EINVAL; > + } > + > + if (!rte_cryptodev_is_valid_dev(conf->dev_id)) > + return -EINVAL; > + > + ret = pdcp_crypto_caps_cipher_verify(conf->dev_id, c_xfrm); > + if (ret) > + return -ENOTSUP; > + > + if (a_xfrm != NULL) { > + ret = pdcp_crypto_caps_auth_verify(conf->dev_id, a_xfrm); > + if (ret) > + return -ENOTSUP; > + } > + > + return 0; > +} > + > +int > +pdcp_crypto_sess_create(struct rte_pdcp_entity *entity, const struct > rte_pdcp_entity_conf *conf) > +{ > + struct rte_crypto_sym_xform *c_xfrm, *a_xfrm; > + struct entity_priv *en_priv; > + bool is_auth_then_cipher; > + int ret; > + > + if (entity == NULL || conf == NULL || conf->crypto_xfrm == NULL) > + return -EINVAL; > + > + en_priv = entity_priv_get(entity); > + > + en_priv->dev_id = conf->dev_id; > + > + if (conf->crypto_xfrm->type == RTE_CRYPTO_SYM_XFORM_CIPHER) { > + c_xfrm = conf->crypto_xfrm; > + a_xfrm = conf->crypto_xfrm->next; > + is_auth_then_cipher = false; > + } else if (conf->crypto_xfrm->type == > RTE_CRYPTO_SYM_XFORM_AUTH) { > + a_xfrm = conf->crypto_xfrm; > + c_xfrm = conf->crypto_xfrm->next; > + is_auth_then_cipher = true; > + } else { > + return -EINVAL; > + } > + > + ret = pdcp_crypto_xfrm_validate(conf, c_xfrm, a_xfrm, > is_auth_then_cipher); > + if (ret) > + return ret; > + > + if (c_xfrm->cipher.algo == RTE_CRYPTO_CIPHER_NULL) > + c_xfrm->cipher.iv.offset = 0; > + else > + c_xfrm->cipher.iv.offset = PDCP_IV_OFFSET; > + > + if (a_xfrm != NULL) { > + if (a_xfrm->auth.algo == RTE_CRYPTO_AUTH_NULL) > + a_xfrm->auth.iv.offset = 0; > + else > + if (c_xfrm->cipher.iv.offset) > + a_xfrm->auth.iv.offset = PDCP_IV_OFFSET + > PDCP_IV_LENGTH; > + else > + a_xfrm->auth.iv.offset = PDCP_IV_OFFSET; > + } > + > + if (conf->sess_mpool == NULL) > + return -EINVAL; > + > + en_priv->crypto_sess = rte_cryptodev_sym_session_create(conf- > >dev_id, conf->crypto_xfrm, > + conf- > >sess_mpool); > + if (en_priv->crypto_sess == NULL) { > + /* API returns positive values as error codes */ > + return -rte_errno; > + } > + > + rte_cryptodev_sym_session_opaque_data_set(en_priv->crypto_sess, > (uint64_t)entity); > + > + return 0; > +} > + > +int > +pdcp_crypto_sess_destroy(struct rte_pdcp_entity *entity) > +{ > + struct entity_priv *en_priv; > + > + if (entity == NULL) > + return -EINVAL; > + > + en_priv = entity_priv_get(entity); > + > + if (en_priv->crypto_sess != NULL) { > + rte_cryptodev_sym_session_free(en_priv->dev_id, en_priv- > >crypto_sess); > + en_priv->crypto_sess = NULL; > + } > + > + return 0; > +} > diff --git a/lib/pdcp/pdcp_crypto.h b/lib/pdcp/pdcp_crypto.h > new file mode 100644 > index 0000000000..dc625b35d0 > --- /dev/null > +++ b/lib/pdcp/pdcp_crypto.h > @@ -0,0 +1,20 @@ > +/* SPDX-License-Identifier: BSD-3-Clause > + * Copyright(C) 2022 Marvell. > + */ > + > +#ifndef _PDCP_CRYPTO_H_ > +#define _PDCP_CRYPTO_H_ > + > +#include <rte_crypto.h> > +#include <rte_crypto_sym.h> > +#include <rte_pdcp.h> > + > +#define PDCP_IV_OFFSET (sizeof(struct rte_crypto_op) + sizeof(struct > rte_crypto_sym_op)) > +#define PDCP_IV_LENGTH 16 > + > +int pdcp_crypto_sess_create(struct rte_pdcp_entity *entity, > + const struct rte_pdcp_entity_conf *conf); > + > +int pdcp_crypto_sess_destroy(struct rte_pdcp_entity *entity); > + > +#endif /* _PDCP_CRYPTO_H_ */ > diff --git a/lib/pdcp/pdcp_entity.h b/lib/pdcp/pdcp_entity.h > new file mode 100644 > index 0000000000..e312fd4a8c > --- /dev/null > +++ b/lib/pdcp/pdcp_entity.h > @@ -0,0 +1,218 @@ > +/* SPDX-License-Identifier: BSD-3-Clause > + * Copyright(C) 2022 Marvell. > + */ > + > +#ifndef _PDCP_ENTITY_H_ > +#define _PDCP_ENTITY_H_ > + > +#include <rte_common.h> > +#include <rte_crypto_sym.h> > +#include <rte_mempool.h> > +#include <rte_pdcp.h> > +#include <rte_security.h> > + > +struct entity_priv; > + > +#define PDCP_PDU_HDR_SIZE_SN_12 (RTE_ALIGN_MUL_CEIL(12, 8) / 8) > +#define PDCP_PDU_HDR_SIZE_SN_18 (RTE_ALIGN_MUL_CEIL(18, 8) / 8) > + > +#define PDCP_GET_SN_12_FROM_COUNT(c) ((c) & 0xfff) > +#define PDCP_GET_SN_18_FROM_COUNT(c) ((c) & 0x3ffff) > + > +#define PDCP_GET_HFN_SN_12_FROM_COUNT(c) (((c) >> 12) & 0xfffff) > +#define PDCP_GET_HFN_SN_18_FROM_COUNT(c) (((c) >> 18) & 0x3fff) > + > +#define PDCP_SET_COUNT_FROM_HFN_SN_12(h, s) ((((h) & 0xfffff) << 12) | ((s) > & 0xfff)) > +#define PDCP_SET_COUNT_FROM_HFN_SN_18(h, s) ((((h) & 0x3fff) << 18) | ((s) > & 0x3ffff)) > + > +#define PDCP_SN_12_WINDOW_SZ 0x800 > +#define PDCP_SN_18_WINDOW_SZ 0x20000 > + > +#define PDCP_SN_12_HFN_MAX ((1 << (32 - 12)) - 1) > +#define PDCP_SN_12_HFN_MIN 0 > +#define PDCP_SN_18_HFN_MAX ((1 << (32 - 18)) - 1) > +#define PDCP_SN_18_HFN_MIN 0 > + Can we have common defines for SN-12 and SN-18 and take SN as parameter? We can have something like this. #define PDCP_PDU_HDR_SIZE(sn_size) (RTE_ALIGN_MUL_CEIL((sn_size), 8) / 8) #define PDCP_GET_SN_FROM_COUNT(c, sn_size) ((c) & ((1<<sn_size)-1)) #define PDCP_GET_HFN_FROM_COUNT(c, sn_size) (((c) >> sn_size) & ((1 << (32 - sn_size)) - 1)) #define PDCP_SET_COUNT_FROM_HFN_SN(h, s, sn_size) ((((h) & ((1 << (32 - sn_size)) - 1)) << sn_size) | ((s) & ((1<<sn_size)-1))) #define PDCP_HFN_MAX(sn_size) ((1 << (32 - (sn_size))) - 1) #define PDCP_HFN_MIN 0 > +/* IV generation function based on the entity configuration */ > +typedef void (*iv_gen_t)(struct rte_crypto_op *cop, const struct entity_priv > *en_priv, > + uint32_t count); > + > +enum pdcp_pdu_type { > + PDCP_PDU_TYPE_CTRL = 0, > + PDCP_PDU_TYPE_DATA = 1, > +}; > + > +enum pdcp_up_ctrl_pdu_type { > + PDCP_UP_CTRL_PDU_TYPE_STATUS_REPORT, > + PDCP_UP_CTRL_PDU_TYPE_ROHC_FEEDBACK, > + PDCP_UP_CTRL_PDU_TYPE_EHC_FEEDBACK, > + PDCP_UP_CRTL_PDU_TYPE_UDC_FEEDBACK > +}; > + > +struct entity_state { > + uint32_t rx_next; > + uint32_t tx_next; > + uint32_t rx_deliv; > + uint32_t rx_reord; > +}; > + > +union auth_iv_partial { > + /* For AES-CMAC, there is no IV, but message gets prepended */ > + struct { > +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN > + uint64_t count : 32; > + uint64_t zero_38_39 : 2; > + uint64_t direction : 1; > + uint64_t bearer : 5; > + uint64_t zero_40_63 : 24; > +#else > + uint64_t count : 32; > + uint64_t bearer : 5; > + uint64_t direction : 1; > + uint64_t zero_38_39 : 2; > + uint64_t zero_40_63 : 24; > +#endif > + } aes_cmac; > + struct { > +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN > + uint64_t count : 32; > + uint64_t zero_37_39 : 3; > + uint64_t bearer : 5; > + uint64_t zero_40_63 : 24; > + > + uint64_t rsvd_65_71 : 7; > + uint64_t direction_64 : 1; > + uint64_t rsvd_72_111 : 40; > + uint64_t rsvd_113_119 : 7; > + uint64_t direction_112 : 1; > + uint64_t rsvd_120_127 : 8; > +#else > + uint64_t count : 32; > + uint64_t bearer : 5; > + uint64_t zero_37_39 : 3; > + uint64_t zero_40_63 : 24; > + > + uint64_t direction_64 : 1; > + uint64_t rsvd_65_71 : 7; > + uint64_t rsvd_72_111 : 40; > + uint64_t direction_112 : 1; > + uint64_t rsvd_113_119 : 7; > + uint64_t rsvd_120_127 : 8; > +#endif > + } zs; > + uint64_t u64[2]; > +}; > + > +union cipher_iv_partial { > + struct { > +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN > + uint64_t count : 32; > + uint64_t zero_38_39 : 2; > + uint64_t direction : 1; > + uint64_t bearer : 5; > + uint64_t zero_40_63 : 24; > + > + uint64_t zero_64_127; > +#else > + uint64_t count : 32; > + uint64_t bearer : 5; > + uint64_t direction : 1; > + uint64_t zero_38_39 : 2; > + uint64_t zero_40_63 : 24; > + > + uint64_t zero_64_127; Can we take zero_64_127 out of #if-else > +#endif > + } aes_ctr; > + struct { > +#if RTE_BYTE_ORDER == RTE_LITTLE_ENDIAN > + uint64_t count : 32; > + uint64_t zero_38_39 : 2; > + uint64_t direction : 1; > + uint64_t bearer : 5; > + uint64_t zero_40_63 : 24; > + > + uint64_t rsvd_64_127; > +#else > + uint64_t count : 32; > + uint64_t bearer : 5; > + uint64_t direction : 1; > + uint64_t zero_38_39 : 2; > + uint64_t zero_40_63 : 24; > + > + uint64_t rsvd_64_127; > +#endif rsvd_64_127 can also be out of #if-else > + } zs; > + uint64_t u64[2]; > +}; > + > +/* > + * Layout of PDCP entity: [rte_pdcp_entity] [entity_priv] [entity_dl/ul] If the layout is fixed, can we have 0 length array in rte_pdcp_entity for entity_priv and in entity_priv for entity_dl/ul. I see that entity_dl/ul are not completely defined. You can define them later when they are supported. > + */ > + Extra line > +struct entity_priv { > + /** Crypto sym session. */ > + struct rte_cryptodev_sym_session *crypto_sess; > + /** Entity specific IV generation function. */ > + iv_gen_t iv_gen; > + /** Pre-prepared auth IV. */ > + union auth_iv_partial auth_iv_part; > + /** Pre-prepared cipher IV. */ > + union cipher_iv_partial cipher_iv_part; > + /** Entity state variables. */ > + struct entity_state state; > + /** Flags. */ > + struct { > + /** PDCP PDU has 4 byte MAC-I. */ > + uint64_t is_authenticated : 1; > + /** Cipher offset & length in bits. */ > + uint64_t is_ciph_in_bits : 1; > + /** Auth offset & length in bits. */ > + uint64_t is_auth_in_bits : 1; > + /** Is UL/transmitting PDCP entity */ > + uint64_t is_ul_entity : 1; > + } flags; > + /** Crypto op pool. */ > + struct rte_mempool *cop_pool; > + /** PDCP header size. */ > + uint8_t hdr_sz; > + /** PDCP AAD size. For AES-CMAC, additional message is prepended for > the operation. */ > + uint8_t aad_sz; > + /** Device ID of the device to be used for offload. */ > + uint8_t dev_id; > +}; > + > +struct entity_priv_dl_part { > + /* TODO - when in-order-delivery is supported, post PDCP packets > would need to cached. */ > + uint8_t dummy; > +}; > + > +struct entity_priv_ul_part { > + /* > + * TODO - when re-establish is supported, both plain & post PDCP > packets would need to be > + * cached. > + */ > + uint8_t dummy; > +}; > + > +static inline struct entity_priv * > +entity_priv_get(const struct rte_pdcp_entity *entity) { > + return RTE_PTR_ADD(entity, sizeof(struct rte_pdcp_entity)); > +} > + > +static inline struct entity_priv_dl_part * > +entity_dl_part_get(const struct rte_pdcp_entity *entity) { > + return RTE_PTR_ADD(entity, sizeof(struct rte_pdcp_entity) + > sizeof(struct entity_priv)); > +} > + > +static inline struct entity_priv_ul_part * > +entity_ul_part_get(const struct rte_pdcp_entity *entity) { > + return RTE_PTR_ADD(entity, sizeof(struct rte_pdcp_entity) + > sizeof(struct entity_priv)); > +} Above inline functions may not be needed also if we have 0 len arrays. > + > +static inline int > +pdcp_hdr_size_get(enum rte_security_pdcp_sn_size sn_size) > +{ > + return RTE_ALIGN_MUL_CEIL(sn_size, 8) / 8; > +} PDCP_PDU_HDR_SIZE is same as this inline function. Can we get away with this one? > + > +#endif /* _PDCP_ENTITY_H_ */ > diff --git a/lib/pdcp/pdcp_process.c b/lib/pdcp/pdcp_process.c > new file mode 100644 > index 0000000000..282cf38ec4 > --- /dev/null > +++ b/lib/pdcp/pdcp_process.c > @@ -0,0 +1,1195 @@ > +/* SPDX-License-Identifier: BSD-3-Clause > + * Copyright(C) 2022 Marvell. > + */ > + > +#include <rte_crypto.h> > +#include <rte_crypto_sym.h> > +#include <rte_cryptodev.h> > +#include <rte_memcpy.h> > +#include <rte_pdcp.h> > +#include <rte_pdcp_hdr.h> > + > +#include "pdcp_crypto.h" > +#include "pdcp_entity.h" > +#include "pdcp_process.h" > + > +#define PDCP_MAC_I_LEN 4 Can you define it at same place where PDCP_IV_LENGTH is defined and use this in xform validate. > + > +/* Enum of supported algorithms for ciphering */ > +enum pdcp_cipher_algo { > + PDCP_CIPHER_ALGO_NULL, > + PDCP_CIPHER_ALGO_AES, > + PDCP_CIPHER_ALGO_ZUC, > + PDCP_CIPHER_ALGO_SNOW3G, > + PDCP_CIPHER_ALGO_MAX > +}; > + > +/* Enum of supported algorithms for integrity */ > +enum pdcp_auth_algo { > + PDCP_AUTH_ALGO_NULL, > + PDCP_AUTH_ALGO_AES, > + PDCP_AUTH_ALGO_ZUC, > + PDCP_AUTH_ALGO_SNOW3G, > + PDCP_AUTH_ALGO_MAX > +}; > + > +/* IV generation functions based on type of operation (cipher - auth) */ > + > +static void > +pdcp_iv_gen_null_null(struct rte_crypto_op *cop, const struct entity_priv > *en_priv, uint32_t count) > +{ > + /* No IV required for NULL cipher + NULL auth */ > + RTE_SET_USED(cop); > + RTE_SET_USED(en_priv); > + RTE_SET_USED(count); > +} > + > +static void > +pdcp_iv_gen_null_aes_cmac(struct rte_crypto_op *cop, const struct > entity_priv *en_priv, > + uint32_t count) > +{ > + struct rte_crypto_sym_op *op = cop->sym; > + struct rte_mbuf *mb = op->m_src; > + uint8_t *m_ptr; > + uint64_t m; > + > + /* AES-CMAC requires message to be prepended with info on count etc > */ > + > + /* Prepend by 8 bytes to add custom message */ > + m_ptr = (uint8_t *)rte_pktmbuf_prepend(mb, 8); > + > + m = en_priv->auth_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + > + rte_memcpy(m_ptr, &m, 8); > +} > + > +static void > +pdcp_iv_gen_null_zs(struct rte_crypto_op *cop, const struct entity_priv > *en_priv, uint32_t count) > +{ > + uint64_t iv_u64[2]; > + uint8_t *iv; > + > + iv = rte_crypto_op_ctod_offset(cop, uint8_t *, PDCP_IV_OFFSET); > + > + iv_u64[0] = en_priv->auth_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + rte_memcpy(iv, &iv_u64[0], 8); > + > + iv_u64[1] = iv_u64[0] ^ en_priv->auth_iv_part.u64[1]; > + rte_memcpy(iv + 8, &iv_u64[1], 8); > +} > + > +static void > +pdcp_iv_gen_aes_ctr_null(struct rte_crypto_op *cop, const struct entity_priv > *en_priv, > + uint32_t count) > +{ > + uint64_t iv_u64[2]; > + uint8_t *iv; > + > + iv = rte_crypto_op_ctod_offset(cop, uint8_t *, PDCP_IV_OFFSET); > + > + iv_u64[0] = en_priv->cipher_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + iv_u64[1] = 0; > + rte_memcpy(iv, iv_u64, 16); > +} > + > +static void > +pdcp_iv_gen_zs_null(struct rte_crypto_op *cop, const struct entity_priv > *en_priv, uint32_t count) > +{ > + uint64_t iv_u64; > + uint8_t *iv; > + > + iv = rte_crypto_op_ctod_offset(cop, uint8_t *, PDCP_IV_OFFSET); > + > + iv_u64 = en_priv->cipher_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + rte_memcpy(iv, &iv_u64, 8); > + rte_memcpy(iv + 8, &iv_u64, 8); > +} > + > +static void > +pdcp_iv_gen_zs_zs(struct rte_crypto_op *cop, const struct entity_priv > *en_priv, uint32_t count) > +{ > + uint64_t iv_u64[2]; > + uint8_t *iv; > + > + iv = rte_crypto_op_ctod_offset(cop, uint8_t *, PDCP_IV_OFFSET); > + > + /* Generating cipher IV */ > + iv_u64[0] = en_priv->cipher_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + rte_memcpy(iv, &iv_u64[0], 8); > + rte_memcpy(iv + 8, &iv_u64[0], 8); > + > + iv += PDCP_IV_LENGTH; > + > + /* Generating auth IV */ > + iv_u64[0] = en_priv->auth_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + rte_memcpy(iv, &iv_u64[0], 8); > + > + iv_u64[1] = iv_u64[0] ^ en_priv->auth_iv_part.u64[1]; > + rte_memcpy(iv + 8, &iv_u64[1], 8); > +} > + > +static void > +pdcp_iv_gen_zs_aes_cmac(struct rte_crypto_op *cop, const struct entity_priv > *en_priv, > + uint32_t count) > +{ > + struct rte_crypto_sym_op *op = cop->sym; > + struct rte_mbuf *mb = op->m_src; > + uint8_t *m_ptr, *iv; > + uint64_t iv_u64[2]; > + uint64_t m; > + > + iv = rte_crypto_op_ctod_offset(cop, uint8_t *, PDCP_IV_OFFSET); > + iv_u64[0] = en_priv->cipher_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + rte_memcpy(iv, &iv_u64[0], 8); > + rte_memcpy(iv + 8, &iv_u64[0], 8); > + > + m_ptr = (uint8_t *)rte_pktmbuf_prepend(mb, 8); > + m = en_priv->auth_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + rte_memcpy(m_ptr, &m, 8); > +} > + > +static void > +pdcp_iv_gen_aes_ctr_aes_cmac(struct rte_crypto_op *cop, const struct > entity_priv *en_priv, > + uint32_t count) > +{ > + struct rte_crypto_sym_op *op = cop->sym; > + struct rte_mbuf *mb = op->m_src; > + uint8_t *m_ptr, *iv; > + uint64_t iv_u64[2]; > + uint64_t m; > + > + iv = rte_crypto_op_ctod_offset(cop, uint8_t *, PDCP_IV_OFFSET); > + > + iv_u64[0] = en_priv->cipher_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + iv_u64[1] = 0; > + rte_memcpy(iv, iv_u64, PDCP_IV_LENGTH); > + > + m_ptr = (uint8_t *)rte_pktmbuf_prepend(mb, 8); > + m = en_priv->auth_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + rte_memcpy(m_ptr, &m, 8); > +} > + > +static void > +pdcp_iv_gen_aes_ctr_zs(struct rte_crypto_op *cop, const struct entity_priv > *en_priv, uint32_t count) > +{ > + uint64_t iv_u64[2]; > + uint8_t *iv; > + > + iv = rte_crypto_op_ctod_offset(cop, uint8_t *, PDCP_IV_OFFSET); > + > + iv_u64[0] = en_priv->cipher_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + iv_u64[1] = 0; > + rte_memcpy(iv, iv_u64, PDCP_IV_LENGTH); > + > + iv += PDCP_IV_LENGTH; > + > + iv_u64[0] = en_priv->auth_iv_part.u64[0] | > ((uint64_t)(rte_cpu_to_be_32(count))); > + rte_memcpy(iv, &iv_u64[0], 8); > + > + iv_u64[1] = iv_u64[0] ^ en_priv->auth_iv_part.u64[1]; > + rte_memcpy(iv + 8, &iv_u64[1], 8); > +} > + > +static int > +pdcp_crypto_xfrm_get(const struct rte_pdcp_entity_conf *conf, struct > rte_crypto_sym_xform **c_xfrm, > + struct rte_crypto_sym_xform **a_xfrm) > +{ > + *c_xfrm = NULL; > + *a_xfrm = NULL; > + > + if (conf->crypto_xfrm == NULL) > + return -EINVAL; > + > + if (conf->crypto_xfrm->type == RTE_CRYPTO_SYM_XFORM_CIPHER) { > + *c_xfrm = conf->crypto_xfrm; > + *a_xfrm = conf->crypto_xfrm->next; > + } else if (conf->crypto_xfrm->type == > RTE_CRYPTO_SYM_XFORM_AUTH) { > + *a_xfrm = conf->crypto_xfrm; > + *c_xfrm = conf->crypto_xfrm->next; > + } else { > + return -EINVAL; > + } > + > + return 0; > +} > + > +static int > +pdcp_iv_gen_func_set(struct rte_pdcp_entity *entity, const struct > rte_pdcp_entity_conf *conf) > +{ > + struct rte_crypto_sym_xform *c_xfrm, *a_xfrm; > + enum rte_security_pdcp_direction direction; > + enum pdcp_cipher_algo ciph_algo; > + enum pdcp_auth_algo auth_algo; > + struct entity_priv *en_priv; > + int ret; > + > + en_priv = entity_priv_get(entity); > + > + direction = conf->pdcp_xfrm.pkt_dir; > + if (conf->reverse_iv_direction) > + direction = !direction; > + > + ret = pdcp_crypto_xfrm_get(conf, &c_xfrm, &a_xfrm); > + if (ret) > + return ret; > + > + if (c_xfrm == NULL) > + return -EINVAL; > + > + memset(&en_priv->auth_iv_part, 0, sizeof(en_priv->auth_iv_part)); > + memset(&en_priv->cipher_iv_part, 0, sizeof(en_priv->cipher_iv_part)); > + > + switch (c_xfrm->cipher.algo) { > + case RTE_CRYPTO_CIPHER_NULL: > + ciph_algo = PDCP_CIPHER_ALGO_NULL; > + break; > + case RTE_CRYPTO_CIPHER_AES_CTR: > + ciph_algo = PDCP_CIPHER_ALGO_AES; > + en_priv->cipher_iv_part.aes_ctr.bearer = conf- > >pdcp_xfrm.bearer; > + en_priv->cipher_iv_part.aes_ctr.direction = direction; > + break; > + case RTE_CRYPTO_CIPHER_SNOW3G_UEA2: > + ciph_algo = PDCP_CIPHER_ALGO_SNOW3G; > + en_priv->cipher_iv_part.zs.bearer = conf->pdcp_xfrm.bearer; > + en_priv->cipher_iv_part.zs.direction = direction; > + break; > + case RTE_CRYPTO_CIPHER_ZUC_EEA3: > + ciph_algo = PDCP_CIPHER_ALGO_ZUC; > + en_priv->cipher_iv_part.zs.bearer = conf->pdcp_xfrm.bearer; > + en_priv->cipher_iv_part.zs.direction = direction; > + break; > + default: > + return -ENOTSUP; > + } > + > + if (a_xfrm != NULL) { > + switch (a_xfrm->auth.algo) { > + case RTE_CRYPTO_AUTH_NULL: > + auth_algo = PDCP_AUTH_ALGO_NULL; > + break; > + case RTE_CRYPTO_AUTH_AES_CMAC: > + auth_algo = PDCP_AUTH_ALGO_AES; > + en_priv->auth_iv_part.aes_cmac.bearer = conf- > >pdcp_xfrm.bearer; > + en_priv->auth_iv_part.aes_cmac.direction = direction; > + break; > + case RTE_CRYPTO_AUTH_SNOW3G_UIA2: > + auth_algo = PDCP_AUTH_ALGO_SNOW3G; > + en_priv->auth_iv_part.zs.bearer = conf- > >pdcp_xfrm.bearer; > + en_priv->auth_iv_part.zs.direction_64 = direction; > + en_priv->auth_iv_part.zs.direction_112 = direction; > + break; > + case RTE_CRYPTO_AUTH_ZUC_EIA3: > + auth_algo = PDCP_AUTH_ALGO_ZUC; > + en_priv->auth_iv_part.zs.bearer = conf- > >pdcp_xfrm.bearer; > + en_priv->auth_iv_part.zs.direction_64 = direction; > + en_priv->auth_iv_part.zs.direction_112 = direction; > + break; > + default: > + return -ENOTSUP; > + } > + } else { > + auth_algo = PDCP_AUTH_ALGO_NULL; > + } > + > + static const iv_gen_t > iv_gen_map[PDCP_CIPHER_ALGO_MAX][PDCP_AUTH_ALGO_MAX] = { > + [PDCP_CIPHER_ALGO_NULL][PDCP_AUTH_ALGO_NULL] = > pdcp_iv_gen_null_null, > + [PDCP_CIPHER_ALGO_NULL][PDCP_AUTH_ALGO_AES] = > pdcp_iv_gen_null_aes_cmac, > + [PDCP_CIPHER_ALGO_NULL][PDCP_AUTH_ALGO_SNOW3G] = > pdcp_iv_gen_null_zs, > + [PDCP_CIPHER_ALGO_NULL][PDCP_AUTH_ALGO_ZUC] = > pdcp_iv_gen_null_zs, > + > + [PDCP_CIPHER_ALGO_AES][PDCP_AUTH_ALGO_NULL] = > pdcp_iv_gen_aes_ctr_null, > + [PDCP_CIPHER_ALGO_AES][PDCP_AUTH_ALGO_AES] = > pdcp_iv_gen_aes_ctr_aes_cmac, > + [PDCP_CIPHER_ALGO_AES][PDCP_AUTH_ALGO_SNOW3G] = > pdcp_iv_gen_aes_ctr_zs, > + [PDCP_CIPHER_ALGO_AES][PDCP_AUTH_ALGO_ZUC] = > pdcp_iv_gen_aes_ctr_zs, > + > + [PDCP_CIPHER_ALGO_SNOW3G][PDCP_AUTH_ALGO_NULL] = > pdcp_iv_gen_zs_null, > + [PDCP_CIPHER_ALGO_SNOW3G][PDCP_AUTH_ALGO_AES] = > pdcp_iv_gen_zs_aes_cmac, > + > [PDCP_CIPHER_ALGO_SNOW3G][PDCP_AUTH_ALGO_SNOW3G] = > pdcp_iv_gen_zs_zs, > + [PDCP_CIPHER_ALGO_SNOW3G][PDCP_AUTH_ALGO_ZUC] = > pdcp_iv_gen_zs_zs, > + > + [PDCP_CIPHER_ALGO_ZUC][PDCP_AUTH_ALGO_NULL] = > pdcp_iv_gen_zs_null, > + [PDCP_CIPHER_ALGO_ZUC][PDCP_AUTH_ALGO_AES] = > pdcp_iv_gen_zs_aes_cmac, > + [PDCP_CIPHER_ALGO_ZUC][PDCP_AUTH_ALGO_SNOW3G] = > pdcp_iv_gen_zs_zs, > + [PDCP_CIPHER_ALGO_ZUC][PDCP_AUTH_ALGO_ZUC] = > pdcp_iv_gen_zs_zs, > + }; > + > + en_priv->iv_gen = iv_gen_map[ciph_algo][auth_algo]; > + > + return 0; > +} > + > +static inline void > +cop_prepare(const struct entity_priv *en_priv, struct rte_mbuf *mb, struct > rte_crypto_op *cop, > + uint8_t data_offset, uint32_t count, const bool is_auth) > +{ > + const struct rte_crypto_op cop_init = { > + .type = RTE_CRYPTO_OP_TYPE_SYMMETRIC, > + .status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED, > + .sess_type = RTE_CRYPTO_OP_WITH_SESSION, > + }; > + struct rte_crypto_sym_op *op; > + uint32_t pkt_len; > + > + const uint8_t ciph_shift = 3 * en_priv->flags.is_ciph_in_bits; > + const uint8_t auth_shift = 3 * en_priv->flags.is_auth_in_bits; > + > + op = cop->sym; > + cop->raw = cop_init.raw; > + op->m_src = mb; > + op->m_dst = mb; > + > + /* Set IV */ > + en_priv->iv_gen(cop, en_priv, count); > + > + /* Prepare op */ > + pkt_len = rte_pktmbuf_pkt_len(mb); > + op->cipher.data.offset = data_offset << ciph_shift; > + op->cipher.data.length = (pkt_len - data_offset) << ciph_shift; > + > + if (is_auth) { > + op->auth.data.offset = 0; > + op->auth.data.length = (pkt_len - PDCP_MAC_I_LEN) << > auth_shift; > + op->auth.digest.data = rte_pktmbuf_mtod_offset(mb, uint8_t > *, > + (pkt_len - > PDCP_MAC_I_LEN)); > + } > + > + __rte_crypto_sym_op_attach_sym_session(op, en_priv->crypto_sess); > +} > + > +static inline bool > +pdcp_pre_process_uplane_sn_12_ul_set_sn(struct entity_priv *en_priv, struct > rte_mbuf *mb, > + uint32_t *count) > +{ > + struct rte_pdcp_up_data_pdu_sn_12_hdr *pdu_hdr; > + const uint8_t hdr_sz = en_priv->hdr_sz; > + uint32_t sn; > + > + /* Prepend PDU header */ > + pdu_hdr = (struct rte_pdcp_up_data_pdu_sn_12_hdr > *)rte_pktmbuf_prepend(mb, hdr_sz); > + if (unlikely(pdu_hdr == NULL)) > + return false; > + > + /* Update sequence num in the PDU header */ > + *count = __atomic_fetch_add(&en_priv->state.tx_next, 1, > __ATOMIC_RELAXED); > + sn = PDCP_GET_SN_12_FROM_COUNT(*count); > + > + pdu_hdr->d_c = PDCP_PDU_TYPE_DATA; > + pdu_hdr->sn_11_8 = ((sn & 0xf00) >> 8); > + pdu_hdr->sn_7_0 = (sn & 0xff); > + pdu_hdr->r = 0; > + return true; > +} > + > +static uint16_t > +pdcp_pre_process_uplane_sn_12_ul(const struct rte_pdcp_entity *entity, > struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t num, > uint16_t *nb_err) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + uint16_t nb_cop; > + uint32_t count; > + int i; > + > + const uint8_t data_offset = en_priv->hdr_sz + en_priv->aad_sz; > + > + nb_cop = rte_crypto_op_bulk_alloc(en_priv->cop_pool, > RTE_CRYPTO_OP_TYPE_SYMMETRIC, cop, > + num); > + > + if (en_priv->flags.is_authenticated) { > + for (i = 0; i < nb_cop; i++) { > + if (unlikely(rte_pktmbuf_append(mb[i], > PDCP_MAC_I_LEN) == NULL)) > + goto cop_free; > + if > (unlikely(!pdcp_pre_process_uplane_sn_12_ul_set_sn(en_priv, mb[i], > + > &count))) > + goto cop_free; > + cop_prepare(en_priv, mb[i], cop[i], data_offset, count, > true); > + } > + } else { > + for (i = 0; i < nb_cop; i++) { > + if > (unlikely(!pdcp_pre_process_uplane_sn_12_ul_set_sn(en_priv, mb[i], > + > &count))) > + goto cop_free; > + cop_prepare(en_priv, mb[i], cop[i], data_offset, count, > false); > + } > + } > + > + *nb_err = num - nb_cop; > + return nb_cop; > +cop_free: > + /* Using mempool API since crypto API is not providing bulk free */ > + rte_mempool_put_bulk(en_priv->cop_pool, (void *)&cop[i], nb_cop - i); > + *nb_err = num - i; > + return i; > +} > + > +static inline bool > +pdcp_pre_process_uplane_sn_18_ul_set_sn(struct entity_priv *en_priv, struct > rte_mbuf *mb, > + uint32_t *count) > +{ > + struct rte_pdcp_up_data_pdu_sn_18_hdr *pdu_hdr; > + const uint8_t hdr_sz = en_priv->hdr_sz; > + uint32_t sn; > + > + /* Prepend PDU header */ > + pdu_hdr = (struct rte_pdcp_up_data_pdu_sn_18_hdr > *)rte_pktmbuf_prepend(mb, hdr_sz); > + if (unlikely(pdu_hdr == NULL)) > + return false; > + > + /* Update sequence num in the PDU header */ > + *count = __atomic_fetch_add(&en_priv->state.tx_next, 1, > __ATOMIC_RELAXED); > + sn = PDCP_GET_SN_18_FROM_COUNT(*count); > + > + pdu_hdr->d_c = PDCP_PDU_TYPE_DATA; > + pdu_hdr->sn_17_16 = ((sn & 0x30000) >> 16); > + pdu_hdr->sn_15_8 = ((sn & 0xff00) >> 8); > + pdu_hdr->sn_7_0 = (sn & 0xff); > + pdu_hdr->r = 0; > + > + return true; > +} > + > +static inline uint16_t > +pdcp_pre_process_uplane_sn_18_ul(const struct rte_pdcp_entity *entity, > struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t num, > uint16_t *nb_err) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + uint16_t nb_cop; > + uint32_t count; > + int i; > + > + const uint8_t data_offset = en_priv->hdr_sz + en_priv->aad_sz; > + > + nb_cop = rte_crypto_op_bulk_alloc(en_priv->cop_pool, > RTE_CRYPTO_OP_TYPE_SYMMETRIC, cop, > + num); > + > + if (en_priv->flags.is_authenticated) { > + for (i = 0; i < nb_cop; i++) { > + if (unlikely(rte_pktmbuf_append(mb[i], > PDCP_MAC_I_LEN) == NULL)) > + goto cop_free; > + if > (unlikely(!pdcp_pre_process_uplane_sn_18_ul_set_sn(en_priv, mb[i], > + > &count))) > + goto cop_free; > + cop_prepare(en_priv, mb[i], cop[i], data_offset, count, > true); > + } > + } else { > + for (i = 0; i < nb_cop; i++) { > + if > (unlikely(!pdcp_pre_process_uplane_sn_18_ul_set_sn(en_priv, mb[i], > + > &count))) > + goto cop_free; > + cop_prepare(en_priv, mb[i], cop[i], data_offset, count, > false); > + } > + } > + > + *nb_err = num - nb_cop; > + return nb_cop; > + > +cop_free: > + /* Using mempool API since crypto API is not providing bulk free */ > + rte_mempool_put_bulk(en_priv->cop_pool, (void *)&cop[i], nb_cop - i); > + *nb_err = num - i; > + return i; > +} > + > +static uint16_t > +pdcp_pre_process_cplane_sn_12_ul(const struct rte_pdcp_entity *entity, > struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t num, > uint16_t *nb_err) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + struct rte_pdcp_cp_data_pdu_sn_12_hdr *pdu_hdr; > + uint32_t count, sn; > + uint16_t nb_cop; > + int i; > + > + const uint8_t hdr_sz = en_priv->hdr_sz; > + const uint8_t data_offset = hdr_sz + en_priv->aad_sz; > + > + nb_cop = rte_crypto_op_bulk_alloc(en_priv->cop_pool, > RTE_CRYPTO_OP_TYPE_SYMMETRIC, cop, > + num); > + > + for (i = 0; i < nb_cop; i++) { > + /* Prepend PDU header */ > + pdu_hdr = (struct rte_pdcp_cp_data_pdu_sn_12_hdr > *)rte_pktmbuf_prepend(mb[i], > + > hdr_sz); > + if (unlikely(pdu_hdr == NULL)) > + goto cop_free; > + if (unlikely(rte_pktmbuf_append(mb[i], PDCP_MAC_I_LEN) == > NULL)) > + goto cop_free; > + > + /* Update sequence number in the PDU header */ > + count = __atomic_fetch_add(&en_priv->state.tx_next, 1, > __ATOMIC_RELAXED); > + sn = PDCP_GET_SN_12_FROM_COUNT(count); > + > + pdu_hdr->sn_11_8 = ((sn & 0xf00) >> 8); > + pdu_hdr->sn_7_0 = (sn & 0xff); > + pdu_hdr->r = 0; > + > + cop_prepare(en_priv, mb[i], cop[i], data_offset, count, true); > + } > + > + *nb_err = num - nb_cop; > + return nb_cop; > + > +cop_free: > + /* Using mempool API since crypto API is not providing bulk free */ > + rte_mempool_put_bulk(en_priv->cop_pool, (void *)&cop[i], nb_cop - i); > + *nb_err = num - i; > + return i; > +} > + > +static uint16_t > +pdcp_post_process_uplane_sn_12_ul(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err_ret) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + const uint32_t hdr_trim_sz = en_priv->aad_sz; > + int i, nb_success = 0, nb_err = 0; > + struct rte_mbuf *err_mb[num]; > + struct rte_mbuf *mb; > + > + for (i = 0; i < num; i++) { > + mb = in_mb[i]; > + if (unlikely(mb->ol_flags & > RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED)) { > + err_mb[nb_err++] = mb; > + continue; > + } > + > + if (hdr_trim_sz) > + rte_pktmbuf_adj(mb, hdr_trim_sz); > + > + out_mb[nb_success++] = mb; > + } > + > + if (unlikely(nb_err != 0)) > + rte_memcpy(&out_mb[nb_success], err_mb, nb_err * > sizeof(struct rte_mbuf *)); > + > + *nb_err_ret = nb_err; > + return nb_success; > +} > + > +static uint16_t > +pdcp_post_process_uplane_sn_18_ul(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err_ret) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + const uint32_t hdr_trim_sz = en_priv->aad_sz; > + int i, nb_success = 0, nb_err = 0; > + struct rte_mbuf *err_mb[num]; > + struct rte_mbuf *mb; > + > + for (i = 0; i < num; i++) { > + mb = in_mb[i]; > + if (unlikely(mb->ol_flags & > RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED)) { > + err_mb[nb_err++] = mb; > + continue; > + } > + > + if (hdr_trim_sz) > + rte_pktmbuf_adj(mb, hdr_trim_sz); > + > + out_mb[nb_success++] = mb; > + } > + > + if (unlikely(nb_err != 0)) > + rte_memcpy(&out_mb[nb_success], err_mb, nb_err * > sizeof(struct rte_mbuf *)); > + > + *nb_err_ret = nb_err; > + return nb_success; > +} > + > +static uint16_t > +pdcp_post_process_cplane_sn_12_ul(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err_ret) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + const uint32_t hdr_trim_sz = en_priv->aad_sz; > + int i, nb_success = 0, nb_err = 0; > + struct rte_mbuf *mb, *err_mb[num]; > + > + for (i = 0; i < num; i++) { > + mb = in_mb[i]; > + if (unlikely(mb->ol_flags & > RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED)) { > + err_mb[nb_err++] = mb; > + continue; > + } > + > + if (hdr_trim_sz) > + rte_pktmbuf_adj(mb, hdr_trim_sz); > + > + out_mb[nb_success++] = mb; > + } > + > + if (unlikely(nb_err != 0)) > + rte_memcpy(&out_mb[nb_success], err_mb, nb_err * > sizeof(struct rte_mbuf *)); > + > + *nb_err_ret = nb_err; > + return nb_success; > +} > + > +static inline int > +pdcp_sn_18_count_get(const struct rte_pdcp_entity *entity, int32_t rsn, > uint32_t *count) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + uint32_t rhfn, rx_deliv; > + > + rx_deliv = __atomic_load_n(&en_priv->state.rx_deliv, > __ATOMIC_RELAXED); > + rhfn = PDCP_GET_HFN_SN_18_FROM_COUNT(rx_deliv); > + > + if (rsn < (int32_t)(PDCP_GET_SN_18_FROM_COUNT(rx_deliv) - > PDCP_SN_18_WINDOW_SZ)) { > + if (unlikely(rhfn == PDCP_SN_18_HFN_MAX)) > + return -ERANGE; > + rhfn += 1; > + } else if ((uint32_t)rsn >= (PDCP_GET_SN_18_FROM_COUNT(rx_deliv) + > PDCP_SN_18_WINDOW_SZ)) { > + if (unlikely(rhfn == PDCP_SN_18_HFN_MIN)) > + return -ERANGE; > + rhfn -= 1; > + } > + > + *count = PDCP_SET_COUNT_FROM_HFN_SN_18(rhfn, rsn); > + > + return 0; > +} > + > +static inline int > +pdcp_sn_12_count_get(const struct rte_pdcp_entity *entity, int32_t rsn, > uint32_t *count) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + uint32_t rhfn, rx_deliv; > + > + rx_deliv = __atomic_load_n(&en_priv->state.rx_deliv, > __ATOMIC_RELAXED); > + rhfn = PDCP_GET_HFN_SN_12_FROM_COUNT(rx_deliv); > + > + if (rsn < (int32_t)(PDCP_GET_SN_12_FROM_COUNT(rx_deliv) - > PDCP_SN_12_WINDOW_SZ)) { > + if (unlikely(rhfn == PDCP_SN_12_HFN_MAX)) > + return -ERANGE; > + rhfn += 1; > + } else if ((uint32_t)rsn >= (PDCP_GET_SN_12_FROM_COUNT(rx_deliv) + > PDCP_SN_12_WINDOW_SZ)) { > + if (unlikely(rhfn == PDCP_SN_12_HFN_MIN)) > + return -ERANGE; > + rhfn -= 1; > + } > + > + *count = PDCP_SET_COUNT_FROM_HFN_SN_12(rhfn, rsn); > + > + return 0; > +} > + > +static inline uint16_t > +pdcp_pre_process_uplane_sn_12_dl_flags(const struct rte_pdcp_entity > *entity, struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t > num, > uint16_t *nb_err, > + const bool is_integ_protected) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + struct rte_pdcp_up_data_pdu_sn_12_hdr *pdu_hdr; > + uint16_t nb_cop; > + int32_t rsn = 0; > + uint32_t count; > + int i; > + > + const uint8_t data_offset = en_priv->hdr_sz + en_priv->aad_sz; > + > + nb_cop = rte_crypto_op_bulk_alloc(en_priv->cop_pool, > RTE_CRYPTO_OP_TYPE_SYMMETRIC, cop, > + num); > + > + for (i = 0; i < nb_cop; i++) { > + > + pdu_hdr = rte_pktmbuf_mtod(mb[i], struct > rte_pdcp_up_data_pdu_sn_12_hdr *); > + > + /* Check for PDU type */ > + if (likely(pdu_hdr->d_c == PDCP_PDU_TYPE_DATA)) > + rsn = ((pdu_hdr->sn_11_8 << 8) | (pdu_hdr->sn_7_0)); > + else > + rte_panic("TODO: Control PDU not handled"); > + > + if (unlikely(pdcp_sn_12_count_get(entity, rsn, &count))) > + break; > + cop_prepare(en_priv, mb[i], cop[i], data_offset, count, > is_integ_protected); > + } > + > + *nb_err = num - nb_cop; > + > + return nb_cop; > +} > + > +static uint16_t > +pdcp_pre_process_uplane_sn_12_dl_ip(const struct rte_pdcp_entity *entity, > struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t num, > uint16_t *nb_err) > +{ > + return pdcp_pre_process_uplane_sn_12_dl_flags(entity, mb, cop, num, > nb_err, true); > +} > + > +static uint16_t > +pdcp_pre_process_uplane_sn_12_dl(const struct rte_pdcp_entity *entity, > struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t num, > uint16_t *nb_err) > +{ > + return pdcp_pre_process_uplane_sn_12_dl_flags(entity, mb, cop, num, > nb_err, false); > +} > + > +static inline uint16_t > +pdcp_pre_process_uplane_sn_18_dl_flags(const struct rte_pdcp_entity > *entity, struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t > num, > uint16_t *nb_err, > + const bool is_integ_protected) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + struct rte_pdcp_up_data_pdu_sn_18_hdr *pdu_hdr; > + uint16_t nb_cop; > + int32_t rsn = 0; > + uint32_t count; > + int i; > + > + const uint8_t data_offset = en_priv->hdr_sz + en_priv->aad_sz; > + nb_cop = rte_crypto_op_bulk_alloc(en_priv->cop_pool, > RTE_CRYPTO_OP_TYPE_SYMMETRIC, cop, > + num); > + > + for (i = 0; i < nb_cop; i++) { > + pdu_hdr = rte_pktmbuf_mtod(mb[i], struct > rte_pdcp_up_data_pdu_sn_18_hdr *); > + > + /* Check for PDU type */ > + if (likely(pdu_hdr->d_c == PDCP_PDU_TYPE_DATA)) > + rsn = ((pdu_hdr->sn_17_16 << 16) | (pdu_hdr->sn_15_8 > << 8) | > + (pdu_hdr->sn_7_0)); > + else > + rte_panic("TODO: Control PDU not handled"); > + > + if (unlikely(pdcp_sn_18_count_get(entity, rsn, &count))) > + break; > + cop_prepare(en_priv, mb[i], cop[i], data_offset, count, > is_integ_protected); > + } > + > + *nb_err = num - nb_cop; > + > + return nb_cop; > +} > + > +static uint16_t > +pdcp_pre_process_uplane_sn_18_dl_ip(const struct rte_pdcp_entity *entity, > struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t num, > uint16_t *nb_err) > +{ > + return pdcp_pre_process_uplane_sn_18_dl_flags(entity, mb, cop, num, > nb_err, true); > +} > + > +static uint16_t > +pdcp_pre_process_uplane_sn_18_dl(const struct rte_pdcp_entity *entity, > struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t num, > uint16_t *nb_err) > +{ > + return pdcp_pre_process_uplane_sn_18_dl_flags(entity, mb, cop, num, > nb_err, false); > +} > + > +static uint16_t > +pdcp_pre_process_cplane_sn_12_dl(const struct rte_pdcp_entity *entity, > struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], uint16_t num, > uint16_t *nb_err) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + struct rte_pdcp_cp_data_pdu_sn_12_hdr *pdu_hdr; > + uint16_t nb_cop; > + uint32_t count; > + int32_t rsn; > + int i; > + > + const uint8_t data_offset = en_priv->hdr_sz + en_priv->aad_sz; > + > + nb_cop = rte_crypto_op_bulk_alloc(en_priv->cop_pool, > RTE_CRYPTO_OP_TYPE_SYMMETRIC, cop, > + num); > + > + for (i = 0; i < nb_cop; i++) { > + pdu_hdr = rte_pktmbuf_mtod(mb[i], struct > rte_pdcp_cp_data_pdu_sn_12_hdr *); > + rsn = ((pdu_hdr->sn_11_8 << 8) | (pdu_hdr->sn_7_0)); > + if (unlikely(pdcp_sn_12_count_get(entity, rsn, &count))) > + break; > + cop_prepare(en_priv, mb[i], cop[i], data_offset, count, true); > + } > + > + *nb_err = num - nb_cop; > + return nb_cop; > +} > + > +static inline bool > +pdcp_post_process_update_entity_state(const struct rte_pdcp_entity *entity, > + const uint32_t count) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + > + if (count < __atomic_load_n(&en_priv->state.rx_deliv, > __ATOMIC_RELAXED)) > + return false; > + > + /* t-Reordering timer is not supported - SDU will be delivered > immediately. > + * Update RX_DELIV to the COUNT value of the first PDCP SDU which > has not > + * been delivered to upper layers > + */ > + __atomic_store_n(&en_priv->state.rx_deliv, (count + 1), > __ATOMIC_RELAXED); > + > + if (count >= __atomic_load_n(&en_priv->state.rx_next, > __ATOMIC_RELAXED)) > + __atomic_store_n(&en_priv->state.rx_next, (count + 1), > __ATOMIC_RELAXED); > + > + return true; > +} > + > +static inline uint16_t > +pdcp_post_process_uplane_sn_12_dl_flags(const struct rte_pdcp_entity > *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err_ret, > + const bool is_integ_protected) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + struct rte_pdcp_up_data_pdu_sn_12_hdr *pdu_hdr; > + int i, nb_success = 0, nb_err = 0, rsn = 0; > + const uint32_t aad_sz = en_priv->aad_sz; > + struct rte_mbuf *err_mb[num]; > + struct rte_mbuf *mb; > + uint32_t count; > + > + const uint32_t hdr_trim_sz = en_priv->hdr_sz + aad_sz; > + > + for (i = 0; i < num; i++) { > + mb = in_mb[i]; > + if (unlikely(mb->ol_flags & > RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED)) > + goto error; > + pdu_hdr = rte_pktmbuf_mtod_offset(mb, struct > rte_pdcp_up_data_pdu_sn_12_hdr *, > + aad_sz); > + > + /* Check for PDU type */ > + if (likely(pdu_hdr->d_c == PDCP_PDU_TYPE_DATA)) > + rsn = ((pdu_hdr->sn_11_8 << 8) | (pdu_hdr->sn_7_0)); > + else > + rte_panic("Control PDU should not be received"); > + > + if (unlikely(pdcp_sn_12_count_get(entity, rsn, &count))) > + goto error; > + > + if (unlikely(!pdcp_post_process_update_entity_state(entity, > count))) > + goto error; > + > + rte_pktmbuf_adj(mb, hdr_trim_sz); > + if (is_integ_protected) > + rte_pktmbuf_trim(mb, PDCP_MAC_I_LEN); > + out_mb[nb_success++] = mb; > + continue; > + > +error: > + err_mb[nb_err++] = mb; > + } > + > + if (unlikely(nb_err != 0)) > + rte_memcpy(&out_mb[nb_success], err_mb, nb_err * > sizeof(struct rte_mbuf *)); > + > + *nb_err_ret = nb_err; > + return nb_success; > +} > + > +static uint16_t > +pdcp_post_process_uplane_sn_12_dl_ip(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err) > +{ > + return pdcp_post_process_uplane_sn_12_dl_flags(entity, in_mb, > out_mb, num, nb_err, true); > +} > + > +static uint16_t > +pdcp_post_process_uplane_sn_12_dl(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err) > +{ > + return pdcp_post_process_uplane_sn_12_dl_flags(entity, in_mb, > out_mb, num, nb_err, false); > +} > + > +static inline uint16_t > +pdcp_post_process_uplane_sn_18_dl_flags(const struct rte_pdcp_entity > *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err_ret, > + const bool is_integ_protected) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + struct rte_pdcp_up_data_pdu_sn_18_hdr *pdu_hdr; > + const uint32_t aad_sz = en_priv->aad_sz; > + int i, nb_success = 0, nb_err = 0; > + struct rte_mbuf *mb, *err_mb[num]; > + int32_t rsn = 0; > + uint32_t count; > + > + const uint32_t hdr_trim_sz = en_priv->hdr_sz + aad_sz; > + > + for (i = 0; i < num; i++) { > + mb = in_mb[i]; > + if (unlikely(mb->ol_flags & > RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED)) > + goto error; > + > + pdu_hdr = rte_pktmbuf_mtod_offset(mb, struct > rte_pdcp_up_data_pdu_sn_18_hdr *, > + aad_sz); > + > + /* Check for PDU type */ > + if (likely(pdu_hdr->d_c == PDCP_PDU_TYPE_DATA)) > + rsn = ((pdu_hdr->sn_17_16 << 16) | (pdu_hdr->sn_15_8 > << 8) | > + (pdu_hdr->sn_7_0)); > + else > + rte_panic("Control PDU should not be received"); > + > + if (unlikely(pdcp_sn_18_count_get(entity, rsn, &count))) > + goto error; > + > + if (unlikely(!pdcp_post_process_update_entity_state(entity, > count))) > + goto error; > + > + rte_pktmbuf_adj(mb, hdr_trim_sz); > + if (is_integ_protected) > + rte_pktmbuf_trim(mb, PDCP_MAC_I_LEN); > + out_mb[nb_success++] = mb; > + continue; > + > +error: > + err_mb[nb_err++] = mb; > + } > + > + if (unlikely(nb_err != 0)) > + rte_memcpy(&out_mb[nb_success], err_mb, nb_err * > sizeof(struct rte_mbuf *)); > + > + *nb_err_ret = nb_err; > + return nb_success; > +} > + > +static uint16_t > +pdcp_post_process_uplane_sn_18_dl_ip(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err) > +{ > + return pdcp_post_process_uplane_sn_18_dl_flags(entity, in_mb, > out_mb, num, nb_err, true); > +} > + > +static uint16_t > +pdcp_post_process_uplane_sn_18_dl(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err) > +{ > + return pdcp_post_process_uplane_sn_18_dl_flags(entity, in_mb, > out_mb, num, nb_err, false); > +} > + > +static uint16_t > +pdcp_post_process_cplane_sn_12_dl(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err_ret) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + struct rte_pdcp_cp_data_pdu_sn_12_hdr *pdu_hdr; > + const uint32_t aad_sz = en_priv->aad_sz; > + int i, nb_success = 0, nb_err = 0; > + struct rte_mbuf *err_mb[num]; > + struct rte_mbuf *mb; > + uint32_t count; > + int32_t rsn; > + > + const uint32_t hdr_trim_sz = en_priv->hdr_sz + aad_sz; > + > + for (i = 0; i < num; i++) { > + mb = in_mb[i]; > + if (unlikely(mb->ol_flags & > RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED)) > + goto error; > + > + pdu_hdr = rte_pktmbuf_mtod_offset(mb, struct > rte_pdcp_cp_data_pdu_sn_12_hdr *, > + aad_sz); > + rsn = ((pdu_hdr->sn_11_8 << 8) | (pdu_hdr->sn_7_0)); > + > + if (unlikely(pdcp_sn_12_count_get(entity, rsn, &count))) > + goto error; > + > + if (unlikely(!pdcp_post_process_update_entity_state(entity, > count))) > + goto error; > + > + rte_pktmbuf_adj(mb, hdr_trim_sz); > + rte_pktmbuf_trim(mb, PDCP_MAC_I_LEN); > + out_mb[nb_success++] = mb; > + continue; > + > +error: > + err_mb[nb_err++] = mb; > + } > + > + if (unlikely(nb_err != 0)) > + rte_memcpy(&out_mb[nb_success], err_mb, nb_err * > sizeof(struct rte_mbuf *)); > + > + *nb_err_ret = nb_err; > + return nb_success; > +} > + > +static int > +pdcp_pre_process_func_set(struct rte_pdcp_entity *entity, const struct > rte_pdcp_entity_conf *conf) > +{ > + struct entity_priv *en_priv = entity_priv_get(entity); > + > + entity->pre_process = NULL; > + entity->post_process = NULL; > + > + if ((conf->pdcp_xfrm.domain == > RTE_SECURITY_PDCP_MODE_CONTROL) && > + (conf->pdcp_xfrm.sn_size == RTE_SECURITY_PDCP_SN_SIZE_12) && > + (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK)) { > + entity->pre_process = pdcp_pre_process_cplane_sn_12_ul; > + entity->post_process = pdcp_post_process_cplane_sn_12_ul; > + } > + > + if ((conf->pdcp_xfrm.domain == > RTE_SECURITY_PDCP_MODE_CONTROL) && > + (conf->pdcp_xfrm.sn_size == RTE_SECURITY_PDCP_SN_SIZE_12) && > + (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK)) { > + entity->pre_process = pdcp_pre_process_cplane_sn_12_dl; > + entity->post_process = pdcp_post_process_cplane_sn_12_dl; > + } > + > + if ((conf->pdcp_xfrm.domain == RTE_SECURITY_PDCP_MODE_DATA) > && > + (conf->pdcp_xfrm.sn_size == RTE_SECURITY_PDCP_SN_SIZE_12) && > + (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK)) { > + entity->pre_process = pdcp_pre_process_uplane_sn_12_ul; > + entity->post_process = pdcp_post_process_uplane_sn_12_ul; > + } > + > + if ((conf->pdcp_xfrm.domain == RTE_SECURITY_PDCP_MODE_DATA) > && > + (conf->pdcp_xfrm.sn_size == RTE_SECURITY_PDCP_SN_SIZE_18) && > + (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK)) { > + entity->pre_process = pdcp_pre_process_uplane_sn_18_ul; > + entity->post_process = pdcp_post_process_uplane_sn_18_ul; > + } > + > + if ((conf->pdcp_xfrm.domain == RTE_SECURITY_PDCP_MODE_DATA) > && > + (conf->pdcp_xfrm.sn_size == RTE_SECURITY_PDCP_SN_SIZE_12) && > + (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) && > + (en_priv->flags.is_authenticated)) { > + entity->pre_process = pdcp_pre_process_uplane_sn_12_dl_ip; > + entity->post_process = > pdcp_post_process_uplane_sn_12_dl_ip; > + } > + > + if ((conf->pdcp_xfrm.domain == RTE_SECURITY_PDCP_MODE_DATA) > && > + (conf->pdcp_xfrm.sn_size == RTE_SECURITY_PDCP_SN_SIZE_12) && > + (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) && > + (!en_priv->flags.is_authenticated)) { > + entity->pre_process = pdcp_pre_process_uplane_sn_12_dl; > + entity->post_process = pdcp_post_process_uplane_sn_12_dl; > + } > + > + if ((conf->pdcp_xfrm.domain == RTE_SECURITY_PDCP_MODE_DATA) > && > + (conf->pdcp_xfrm.sn_size == RTE_SECURITY_PDCP_SN_SIZE_18) && > + (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) && > + (en_priv->flags.is_authenticated)) { > + entity->pre_process = pdcp_pre_process_uplane_sn_18_dl_ip; > + entity->post_process = > pdcp_post_process_uplane_sn_18_dl_ip; > + } > + > + if ((conf->pdcp_xfrm.domain == RTE_SECURITY_PDCP_MODE_DATA) > && > + (conf->pdcp_xfrm.sn_size == RTE_SECURITY_PDCP_SN_SIZE_18) && > + (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) && > + (!en_priv->flags.is_authenticated)) { > + entity->pre_process = pdcp_pre_process_uplane_sn_18_dl; > + entity->post_process = pdcp_post_process_uplane_sn_18_dl; > + } > + > + if (entity->pre_process == NULL || entity->post_process == NULL) > + return -ENOTSUP; > + > + return 0; > +} > + > +static int > +pdcp_entity_priv_populate(struct entity_priv *en_priv, const struct > rte_pdcp_entity_conf *conf) > +{ > + struct rte_crypto_sym_xform *c_xfrm, *a_xfrm; > + int ret; > + > + /** > + * flags.is_authenticated > + * > + * MAC-I would be added in case of control plane packets and when > authentication > + * transform is not NULL. > + */ > + > + if (conf->pdcp_xfrm.domain == > RTE_SECURITY_PDCP_MODE_CONTROL) > + en_priv->flags.is_authenticated = 1; > + > + ret = pdcp_crypto_xfrm_get(conf, &c_xfrm, &a_xfrm); > + if (ret) > + return ret; > + > + if (a_xfrm != NULL) > + en_priv->flags.is_authenticated = 1; > + > + /** > + * flags.is_ciph_in_bits > + * > + * For ZUC & SNOW3G cipher algos, offset & length need to be provided > in bits. > + */ > + > + if ((c_xfrm->cipher.algo == RTE_CRYPTO_CIPHER_SNOW3G_UEA2) || > + (c_xfrm->cipher.algo == RTE_CRYPTO_CIPHER_ZUC_EEA3)) > + en_priv->flags.is_ciph_in_bits = 1; > + > + /** > + * flags.is_auth_in_bits > + * > + * For ZUC & SNOW3G authentication algos, offset & length need to be > provided in bits. > + */ > + > + if (a_xfrm != NULL) { > + if ((a_xfrm->auth.algo == RTE_CRYPTO_AUTH_SNOW3G_UIA2) > || > + (a_xfrm->auth.algo == RTE_CRYPTO_AUTH_ZUC_EIA3)) > + en_priv->flags.is_auth_in_bits = 1; > + } > + > + /** > + * flags.is_ul_entity > + * > + * Indicate whether the entity is UL/transmitting PDCP entity. > + */ > + if (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK) > + en_priv->flags.is_ul_entity = 1; > + > + /** > + * hdr_sz > + * > + * PDCP header size of the entity > + */ > + en_priv->hdr_sz = pdcp_hdr_size_get(conf->pdcp_xfrm.sn_size); > + > + /** > + * aad_sz > + * > + * For AES-CMAC, additional message is prepended for processing. Need > to be trimmed after > + * crypto processing is done. > + */ > + if (a_xfrm != NULL && a_xfrm->auth.algo == > RTE_CRYPTO_AUTH_AES_CMAC) > + en_priv->aad_sz = 8; > + else > + en_priv->aad_sz = 0; > + > + return 0; > +} > + > +int > +pdcp_process_func_set(struct rte_pdcp_entity *entity, const struct > rte_pdcp_entity_conf *conf) > +{ > + struct entity_priv *en_priv; > + int ret; > + > + if (entity == NULL || conf == NULL) > + return -EINVAL; > + > + en_priv = entity_priv_get(entity); > + > + ret = pdcp_iv_gen_func_set(entity, conf); > + if (ret) > + return ret; > + > + ret = pdcp_entity_priv_populate(en_priv, conf); > + if (ret) > + return ret; > + > + ret = pdcp_pre_process_func_set(entity, conf); > + if (ret) > + return ret; > + > + return 0; > +} > diff --git a/lib/pdcp/pdcp_process.h b/lib/pdcp/pdcp_process.h > new file mode 100644 > index 0000000000..c92ab34c40 > --- /dev/null > +++ b/lib/pdcp/pdcp_process.h > @@ -0,0 +1,13 @@ > +/* SPDX-License-Identifier: BSD-3-Clause > + * Copyright(C) 2022 Marvell. > + */ > + > +#ifndef _PDCP_PROCESS_H_ > +#define _PDCP_PROCESS_H_ > + > +#include <rte_pdcp.h> > + > +int > +pdcp_process_func_set(struct rte_pdcp_entity *entity, const struct > rte_pdcp_entity_conf *conf); > + > +#endif /* _PDCP_PROCESS_H_ */ > diff --git a/lib/pdcp/rte_pdcp.c b/lib/pdcp/rte_pdcp.c > new file mode 100644 > index 0000000000..b1533971c2 > --- /dev/null > +++ b/lib/pdcp/rte_pdcp.c > @@ -0,0 +1,136 @@ > +/* SPDX-License-Identifier: BSD-3-Clause > + * Copyright(C) 2022 Marvell. > + */ > + > +#include <rte_pdcp.h> > +#include <rte_malloc.h> > + > +#include "pdcp_crypto.h" > +#include "pdcp_entity.h" > +#include "pdcp_process.h" > + > +static int > +pdcp_entity_size_get(const struct rte_pdcp_entity_conf *conf) > +{ > + int size; > + > + size = sizeof(struct rte_pdcp_entity) + sizeof(struct entity_priv); > + > + if (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_DOWNLINK) > + size += sizeof(struct entity_priv_dl_part); > + else if (conf->pdcp_xfrm.pkt_dir == RTE_SECURITY_PDCP_UPLINK) > + size += sizeof(struct entity_priv_ul_part); > + else > + return -EINVAL; > + > + return RTE_ALIGN_CEIL(size, RTE_CACHE_LINE_SIZE); > +} > + > +struct rte_pdcp_entity * > +rte_pdcp_entity_establish(const struct rte_pdcp_entity_conf *conf) > +{ > + struct rte_pdcp_entity *entity = NULL; > + struct entity_priv *en_priv; > + int ret; > + > + if (conf == NULL || conf->cop_pool == NULL) { > + rte_errno = -EINVAL; > + return NULL; > + } > + > + if (conf->pdcp_xfrm.en_ordering || conf- > >pdcp_xfrm.remove_duplicates || conf->is_slrb || > + conf->en_sec_offload) { > + rte_errno = -ENOTSUP; > + return NULL; > + } > + > + /* > + * 6.3.2 PDCP SN > + * Length: 12 or 18 bits as indicated in table 6.3.2-1. The length of > the > PDCP SN is > + * configured by upper layers (pdcp-SN-SizeUL, pdcp-SN-SizeDL, or sl- > PDCP-SN-Size in > + * TS 38.331 [3]) > + */ > + if ((conf->pdcp_xfrm.sn_size != RTE_SECURITY_PDCP_SN_SIZE_12) && > + (conf->pdcp_xfrm.sn_size != RTE_SECURITY_PDCP_SN_SIZE_18)) { > + rte_errno = -ENOTSUP; > + return NULL; > + } > + > + if (conf->pdcp_xfrm.hfn || conf->pdcp_xfrm.hfn_threshold) { > + rte_errno = -EINVAL; > + return NULL; > + } > + > + entity = rte_zmalloc_socket("pdcp_entity", pdcp_entity_size_get(conf), > + RTE_CACHE_LINE_SIZE, SOCKET_ID_ANY); > + if (entity == NULL) { > + rte_errno = -ENOMEM; > + return NULL; > + } > + > + en_priv = entity_priv_get(entity); > + > + en_priv->state.rx_deliv = conf->count; > + en_priv->state.tx_next = conf->count; > + en_priv->cop_pool = conf->cop_pool; > + > + /* Setup crypto session */ > + ret = pdcp_crypto_sess_create(entity, conf); > + if (ret) > + goto entity_free; > + > + ret = pdcp_process_func_set(entity, conf); > + if (ret) > + goto crypto_sess_destroy; > + > + return entity; > + > +crypto_sess_destroy: > + pdcp_crypto_sess_destroy(entity); > +entity_free: > + rte_free(entity); > + rte_errno = ret; > + return NULL; > +} > + > +int > +rte_pdcp_entity_release(struct rte_pdcp_entity *pdcp_entity, struct rte_mbuf > *out_mb[]) > +{ > + int ret; > + > + if (pdcp_entity == NULL) > + return -EINVAL; > + > + /* Teardown crypto sessions */ > + ret = pdcp_crypto_sess_destroy(pdcp_entity); > + if (ret) > + return ret; > + > + rte_free(pdcp_entity); > + > + RTE_SET_USED(out_mb); > + return 0; > +} > + > +int > +rte_pdcp_entity_suspend(struct rte_pdcp_entity *pdcp_entity, > + struct rte_mbuf *out_mb[]) > +{ > + struct entity_priv *en_priv; > + > + if (pdcp_entity == NULL) > + return -EINVAL; > + > + en_priv = entity_priv_get(pdcp_entity); > + > + if (en_priv->flags.is_ul_entity) { > + en_priv->state.tx_next = 0; > + } else { > + en_priv->state.rx_next = 0; > + en_priv->state.rx_deliv = 0; > + } > + > + RTE_SET_USED(out_mb); > + > + return 0; > +} > diff --git a/lib/pdcp/rte_pdcp.h b/lib/pdcp/rte_pdcp.h > new file mode 100644 > index 0000000000..b6c7f32c05 > --- /dev/null > +++ b/lib/pdcp/rte_pdcp.h > @@ -0,0 +1,263 @@ > +/* SPDX-License-Identifier: BSD-3-Clause > + * Copyright(C) 2022 Marvell. > + */ > + > +#ifndef _RTE_PDCP_H_ > +#define _RTE_PDCP_H_ > + > +/** > + * @file rte_pdcp.h > + * > + * RTE PDCP support. > + * > + * librte_pdcp provides a framework for PDCP protocol processing. > + */ > + > +#include <rte_compat.h> > +#include <rte_common.h> > +#include <rte_errno.h> > +#include <rte_mempool.h> > +#include <rte_security.h> Remove header file which is not needed. I do not see use of rte_errno.h I believe rte_common.h and rte_compat.h are also not needed. > + > +#ifdef __cplusplus > +extern "C" { > +#endif > + > +/* Forward declarations */ > +struct rte_pdcp_entity; > + > +/* PDCP pre-process function based on entity configuration */ > +typedef uint16_t (*rte_pdcp_pre_p_t)(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *mb[], > + struct rte_crypto_op *cop[], > + uint16_t num, uint16_t *nb_err); > + > +/* PDCP post-process function based on entity configuration */ > +typedef uint16_t (*rte_pdcp_post_p_t)(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err); > + > +/** > + * PDCP entity. > + */ > +struct rte_pdcp_entity { > + /** Entity specific pre-process handle. */ > + rte_pdcp_pre_p_t pre_process; > + /** Entity specific post-process handle. */ > + rte_pdcp_post_p_t post_process; > + /** > + * PDCP entities may hold packets for purposes of in-order delivery (in > + * case of receiving PDCP entity) and re-transmission (in case of > + * transmitting PDCP entity). > + * > + * For receiving PDCP entity, it may hold packets when in-order > + * delivery is enabled. The packets would be cached until either a > + * packet that completes the sequence arrives or when discard timer > + * expires. > + * > + * When post-processing of PDCP packet which completes a sequence is > + * done, the API may return more packets than enqueued. Application is > + * expected to provide *rte_pdcp_pkt_post_process()* with *out_mb* > + * which can hold maximum number of packets which may be returned. > + * > + * For transmitting PDCP entity, during re-establishment (5.1.2), > + * entity may be required to perform re-transmission of the buffers > + * after applying new ciphering & integrity algorithms. For performing > + * crypto operation, *rte_pdcp_entity_re_establish()* would return as > + * many crypto_ops as the ones cached. > + */ > + uint16_t max_pkt_cache; > + /** User area for saving application data. */ > + uint64_t user_area[2]; > +} __rte_cache_aligned; > + > +/** > + * PDCP entity configuration to be used for establishing an entity. > + */ > +struct rte_pdcp_entity_conf { > + /** PDCP transform for the entity. */ > + struct rte_security_pdcp_xform pdcp_xfrm; > + /** Crypto transform applicable for the entity. */ > + struct rte_crypto_sym_xform *crypto_xfrm; > + /** Mempool for crypto symmetric session. */ > + struct rte_mempool *sess_mpool; > + /** Crypto op pool.*/ > + struct rte_mempool *cop_pool; > + /** > + * 32 bit count value (HFN + SN) to be used for the first packet. > + * pdcp_xfrm.hfn would be ignored as the HFN would be derived from > this value. > + */ > + uint32_t count; > + /** Indicate whether the PDCP entity belongs to Side Link Radio Bearer. > */ > + bool is_slrb; > + /** Enable security offload on the device specified. */ > + bool en_sec_offload; > + /** Enable non-atomic usage of entity. */ > + bool en_non_atomic; > + /** Device on which security/crypto session need to be created. */ > + uint8_t dev_id; > + /** Reverse direction during IV generation. Can be used to simulate UE > crypto processing.*/ > + bool reverse_iv_direction; > +}; > + > +/** > + * @warning > + * @b EXPERIMENTAL: this API may change without prior notice > + * > + * 5.1.1 PDCP entity establishment > + * > + * Establish PDCP entity based on provided input configuration. > + * > + * @param conf > + * Parameters to be used for initializing PDCP entity object. > + * @return > + * - Valid handle if success > + * - NULL in case of failure. rte_errno will be set to error code > + */ > +__rte_experimental > +struct rte_pdcp_entity * > +rte_pdcp_entity_establish(const struct rte_pdcp_entity_conf *conf); > + > +/** > + * @warning > + * @b EXPERIMENTAL: this API may change without prior notice > + * > + * 5.1.3 PDCP entity release > + * > + * Release PDCP entity. > + * > + * For UL/transmitting PDCP entity, all stored PDCP SDUs would be dropped. > + * For DL/receiving PDCP entity, the stored PDCP SDUs would be returned in > + * *out_mb* buffer. The buffer should be large enough to hold all cached > + * packets in the entity. > + * > + * @param pdcp_entity > + * Pointer to the PDCP entity to be released. > + * @param[out] out_mb > + * The address of an array that can hold up to > *rte_pdcp_entity.max_pkt_cache* > + * pointers to *rte_mbuf* structures. > + * @return > + * - 0: Success and no cached packets to return > + * - >0: Success and the number of packets returned in out_mb > + * - <0: Error code in case of failures > + */ > +__rte_experimental > +int > +rte_pdcp_entity_release(struct rte_pdcp_entity *pdcp_entity, > + struct rte_mbuf *out_mb[]); > + > +/** > + * @warning > + * @b EXPERIMENTAL: this API may change without prior notice > + * > + * 5.1.4 PDCP entity suspend > + * > + * Suspend PDCP entity. > + * > + * For DL/receiving PDCP entity, the stored PDCP SDUs would be returned in > + * *out_mb* buffer. The buffer should be large enough to hold all cached > + * packets in the entity. > + * > + * For UL/transmitting PDCP entity, *out_mb* buffer would be unused. > + * > + * @param pdcp_entity > + * Pointer to the PDCP entity to be suspended. > + * @param[out] out_mb > + * The address of an array that can hold up to > *rte_pdcp_entity.max_pkt_cache* > + * pointers to *rte_mbuf* structures. > + * @return > + * - 0: Success and no cached packets to return > + * - >0: Success and the number of packets returned in out_mb > + * - <0: Error code in case of failures > + */ > +__rte_experimental > +int > +rte_pdcp_entity_suspend(struct rte_pdcp_entity *pdcp_entity, > + struct rte_mbuf *out_mb[]); > + > +/** > + * @warning > + * @b EXPERIMENTAL: this API may change without prior notice > + * > + * For input mbufs and given PDCP entity pre-process the mbufs and prepare > + * crypto ops that can be enqueued to the cryptodev associated with given > + * session. Only error packets would be moved returned in the input buffer, > + * *mb*, and it is the responsibility of the application to free the same. > + * > + * @param entity > + * Pointer to the *rte_pdcp_entity* object the packets belong to. > + * @param[in, out] mb > + * The address of an array of *num* pointers to *rte_mbuf* structures > + * which contain the input packets. Any error packets would be returned in > the > + * same buffer. > + * @param[out] cop > + * The address of an array that can hold up to *num* pointers to > + * *rte_crypto_op* structures. Crypto ops would be allocated by > + * ``rte_pdcp_pkt_pre_process`` API. > + * @param num > + * The maximum number of packets to process. > + * @param[out] nb_err > + * Pointer to return the number of error packets returned in *mb* > + * @return > + * Count of crypto_ops prepared > + */ > +__rte_experimental > +static inline uint16_t > +rte_pdcp_pkt_pre_process(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *mb[], struct rte_crypto_op *cop[], > + uint16_t num, uint16_t *nb_err) > +{ > + return entity->pre_process(entity, mb, cop, num, nb_err); > +} > + > +/** > + * @warning > + * @b EXPERIMENTAL: this API may change without prior notice > + * > + * For input mbufs and given PDCP entity, perform PDCP post-processing of the > + * mbufs. > + * > + * Input mbufs are the ones retrieved from crypto_ops dequeued from > cryptodev > + * and grouped by *rte_pdcp_pkt_crypto_group()*. > + * > + * The post-processed packets would be returned in the *out_mb* buffer. > + * The resultant mbufs would be grouped into success packets and error > packets. > + * Error packets would be grouped in the end of the array and it is the > + * responsibility of the application to handle the same. > + * > + * When in-order delivery is enabled, PDCP entity may buffer packets and > would > + * deliver packets only when all prior packets have been post-processed. That > + * would result in returning more/less packets than enqueued. > + * > + * @param entity > + * Pointer to the *rte_pdcp_entity* object the packets belong to. > + * @param in_mb > + * The address of an array of *num* pointers to *rte_mbuf* structures. > + * @param[out] out_mb > + * The address of an array of *num* pointers to *rte_mbuf* structures > + * to output packets after PDCP post-processing. > + * @param num > + * The maximum number of packets to process. > + * @param[out] nb_err > + * The number of error packets returned in *out_mb* buffer. > + * @return > + * Count of packets returned in *out_mb* buffer. > + */ > +__rte_experimental > +static inline uint16_t > +rte_pdcp_pkt_post_process(const struct rte_pdcp_entity *entity, > + struct rte_mbuf *in_mb[], > + struct rte_mbuf *out_mb[], > + uint16_t num, uint16_t *nb_err) > +{ > + return entity->post_process(entity, in_mb, out_mb, num, nb_err); > +} > + > +#include <rte_pdcp_group.h> > + > +#ifdef __cplusplus > +} > +#endif > + > +#endif /* _RTE_PDCP_H_ */ > diff --git a/lib/pdcp/rte_pdcp_group.h b/lib/pdcp/rte_pdcp_group.h > new file mode 100644 > index 0000000000..2c01c19d4e > --- /dev/null > +++ b/lib/pdcp/rte_pdcp_group.h > @@ -0,0 +1,133 @@ > +/* SPDX-License-Identifier: BSD-3-Clause > + * Copyright(C) 2022 Marvell. > + */ > + > +#ifndef _RTE_PDCP_GROUP_H_ > +#define _RTE_PDCP_GROUP_H_ > + > +/** > + * @file rte_pdcp_group.h > + * > + * RTE PDCP grouping support. > + * It is not recommended to include this file directly, include <rte_pdcp.h> > + * instead. > + * Provides helper functions to process completed crypto-ops and group > related > + * packets by sessions they belong to. > + */ > + > +#include <rte_common.h> > +#include <rte_crypto.h> > +#include <rte_cryptodev.h> > +#include <rte_security.h> Remove header files which are not needed. > + > +#ifdef __cplusplus > +extern "C" { > +#endif > + > +/** > + * Group packets belonging to same PDCP entity. > + */ > +struct rte_pdcp_group { > + union { > + uint64_t val; > + void *ptr; > + } id; /**< Grouped by value */ > + struct rte_mbuf **m; /**< Start of the group */ > + uint32_t cnt; /**< Number of entries in the group */ > + int32_t rc; /**< Status code associated with the group */ > +}; > + > +/** > + * Take crypto-op as an input and extract pointer to related PDCP entity. > + * @param cop > + * The address of an input *rte_crypto_op* structure. > + * @return > + * The pointer to the related *rte_pdcp_entity* structure. > + */ > +static inline struct rte_pdcp_entity * > +rte_pdcp_en_from_cop(const struct rte_crypto_op *cop) > +{ > + void *sess = cop->sym[0].session; > + > + if (cop->sess_type == RTE_CRYPTO_OP_SECURITY_SESSION) { > + return (struct rte_pdcp_entity *)(uintptr_t) > + rte_security_session_opaque_data_get(sess); > + } else if (cop->sess_type == RTE_CRYPTO_OP_WITH_SESSION) { > + return (struct rte_pdcp_entity *)(uintptr_t) > + rte_cryptodev_sym_session_opaque_data_get(sess); > + } This patchset is not supporting security sessions, so it would be better to return NULL for that. Moreover, we can directly call rte_cryptodev_sym_session_opaque_data_get(cop->sym[0].session) >From rte_pdcp_pkt_crypto_group. No need to have a wrapper. > + > + return NULL; > +} > + > +/** > + * Take as input completed crypto ops, extract related mbufs and group them > by > + * *rte_pdcp_entity* they belong to. Mbuf for which the crypto operation has > + * failed would be flagged using *RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED* > flag > + * in rte_mbuf.ol_flags. The crypto_ops would be freed after the grouping. > + * > + * Note that application must ensure only crypto-ops prepared by lib_pdcp is > + * provided back to @see rte_pdcp_pkt_crypto_group(). > + * > + * @param cop > + * The address of an array of *num* pointers to the input *rte_crypto_op* > + * structures. > + * @param[out] mb > + * The address of an array of *num* pointers to output *rte_mbuf* > structures. > + * @param[out] grp > + * The address of an array of *num* to output *rte_pdcp_group* structures. > + * @param num > + * The maximum number of crypto-ops to process. > + * @return > + * Number of filled elements in *grp* array. > + * > + */ > +static inline uint16_t > +rte_pdcp_pkt_crypto_group(struct rte_crypto_op *cop[], struct rte_mbuf > *mb[], > + struct rte_pdcp_group grp[], uint16_t num) > +{ > + uint32_t i, j = 0, n = 0; > + void *ns, *ps = NULL; > + struct rte_mbuf *m; > + > + for (i = 0; i != num; i++) { > + m = cop[i]->sym[0].m_src; > + ns = cop[i]->sym[0].session; > + > + m->ol_flags |= RTE_MBUF_F_RX_SEC_OFFLOAD; > + if (cop[i]->status != RTE_CRYPTO_OP_STATUS_SUCCESS) > + m->ol_flags |= > RTE_MBUF_F_RX_SEC_OFFLOAD_FAILED; > + > + /* Different entity */ > + if (ps != ns) { > + > + /* Finalize open group and start a new one */ > + if (ps != NULL) { > + grp[n].cnt = mb + j - grp[n].m; > + n++; > + } > + > + /* Start new group */ > + grp[n].m = mb + j; > + ps = ns; > + grp[n].id.ptr = rte_pdcp_en_from_cop(cop[i]); > + } > + > + mb[j++] = m; > + rte_crypto_op_free(cop[i]); > + } > + > + /* Finalize last group */ > + if (ps != NULL) { > + grp[n].cnt = mb + j - grp[n].m; > + n++; > + } > + > + return n; > +} > + > +#ifdef __cplusplus > +} > +#endif > + > +#endif /* _RTE_PDCP_GROUP_H_ */ > diff --git a/lib/pdcp/version.map b/lib/pdcp/version.map > new file mode 100644 > index 0000000000..8fa9d5d7cc > --- /dev/null > +++ b/lib/pdcp/version.map > @@ -0,0 +1,13 @@ > +EXPERIMENTAL { > + global: > + > + # added in 22.11 Change to 23.03 > + rte_pdcp_entity_establish; > + rte_pdcp_entity_release; > + rte_pdcp_entity_suspend; > + > + rte_pdcp_pkt_post_process; > + rte_pdcp_pkt_pre_process; > + > + local: *; > +}; > -- > 2.25.1