> 
> Add support for AES_CCM, CHACHA20_POLY1305 and AES_GMAC.

I think it would be good to add new test-cases to examples/ipsec-secgw/test 
harness
to cover these new algs supported.
Apart from that:
Acked-by: Konstantin Ananyev <konstantin.anan...@intel.com>

> 
> Signed-off-by: Declan Doherty <declan.dohe...@intel.com>
> Signed-off-by: Radu Nicolau <radu.nico...@intel.com>
> Signed-off-by: Abhijit Sinha <abhijit.si...@intel.com>
> Signed-off-by: Daniel Martin Buckley <daniel.m.buck...@intel.com>
> Acked-by: Fan Zhang <roy.fan.zh...@intel.com>
> ---
>  lib/ipsec/crypto.h   | 137 +++++++++++++++++++++++++++++++++++++++++++
>  lib/ipsec/esp_inb.c  |  66 ++++++++++++++++++++-
>  lib/ipsec/esp_outb.c |  70 +++++++++++++++++++++-
>  lib/ipsec/sa.c       |  54 +++++++++++++++--
>  lib/ipsec/sa.h       |   6 ++
>  5 files changed, 322 insertions(+), 11 deletions(-)
> 
> diff --git a/lib/ipsec/crypto.h b/lib/ipsec/crypto.h
> index 3d03034590..93d20aaaa0 100644
> --- a/lib/ipsec/crypto.h
> +++ b/lib/ipsec/crypto.h
> @@ -21,6 +21,37 @@ struct aesctr_cnt_blk {
>       uint32_t cnt;
>  } __rte_packed;
> 
> + /*
> +  * CHACHA20-POLY1305 devices have some specific requirements
> +  * for IV and AAD formats.
> +  * Ideally that to be done by the driver itself.
> +  */
> +
> +struct aead_chacha20_poly1305_iv {
> +     uint32_t salt;
> +     uint64_t iv;
> +     uint32_t cnt;
> +} __rte_packed;
> +
> +struct aead_chacha20_poly1305_aad {
> +     uint32_t spi;
> +     /*
> +      * RFC 4106, section 5:
> +      * Two formats of the AAD are defined:
> +      * one for 32-bit sequence numbers, and one for 64-bit ESN.
> +      */
> +     union {
> +             uint32_t u32[2];
> +             uint64_t u64;
> +     } sqn;
> +     uint32_t align0; /* align to 16B boundary */
> +} __rte_packed;
> +
> +struct chacha20_poly1305_esph_iv {
> +     struct rte_esp_hdr esph;
> +     uint64_t iv;
> +} __rte_packed;
> +
>   /*
>    * AES-GCM devices have some specific requirements for IV and AAD formats.
>    * Ideally that to be done by the driver itself.
> @@ -51,6 +82,47 @@ struct gcm_esph_iv {
>       uint64_t iv;
>  } __rte_packed;
> 
> + /*
> +  * AES-CCM devices have some specific requirements for IV and AAD formats.
> +  * Ideally that to be done by the driver itself.
> +  */
> +union aead_ccm_salt {
> +     uint32_t salt;
> +     struct inner {
> +             uint8_t salt8[3];
> +             uint8_t ccm_flags;
> +     } inner;
> +} __rte_packed;
> +
> +
> +struct aead_ccm_iv {
> +     uint8_t ccm_flags;
> +     uint8_t salt[3];
> +     uint64_t iv;
> +     uint32_t cnt;
> +} __rte_packed;
> +
> +struct aead_ccm_aad {
> +     uint8_t padding[18];
> +     uint32_t spi;
> +     /*
> +      * RFC 4309, section 5:
> +      * Two formats of the AAD are defined:
> +      * one for 32-bit sequence numbers, and one for 64-bit ESN.
> +      */
> +     union {
> +             uint32_t u32[2];
> +             uint64_t u64;
> +     } sqn;
> +     uint32_t align0; /* align to 16B boundary */
> +} __rte_packed;
> +
> +struct ccm_esph_iv {
> +     struct rte_esp_hdr esph;
> +     uint64_t iv;
> +} __rte_packed;
> +
> +
>  static inline void
>  aes_ctr_cnt_blk_fill(struct aesctr_cnt_blk *ctr, uint64_t iv, uint32_t nonce)
>  {
> @@ -59,6 +131,16 @@ aes_ctr_cnt_blk_fill(struct aesctr_cnt_blk *ctr, uint64_t 
> iv, uint32_t nonce)
>       ctr->cnt = rte_cpu_to_be_32(1);
>  }
> 
> +static inline void
> +aead_chacha20_poly1305_iv_fill(struct aead_chacha20_poly1305_iv
> +                            *chacha20_poly1305,
> +                            uint64_t iv, uint32_t salt)
> +{
> +     chacha20_poly1305->salt = salt;
> +     chacha20_poly1305->iv = iv;
> +     chacha20_poly1305->cnt = rte_cpu_to_be_32(1);
> +}
> +
>  static inline void
>  aead_gcm_iv_fill(struct aead_gcm_iv *gcm, uint64_t iv, uint32_t salt)
>  {
> @@ -67,6 +149,21 @@ aead_gcm_iv_fill(struct aead_gcm_iv *gcm, uint64_t iv, 
> uint32_t salt)
>       gcm->cnt = rte_cpu_to_be_32(1);
>  }
> 
> +static inline void
> +aead_ccm_iv_fill(struct aead_ccm_iv *ccm, uint64_t iv, uint32_t salt)
> +{
> +     union aead_ccm_salt tsalt;
> +
> +     tsalt.salt = salt;
> +     ccm->ccm_flags = tsalt.inner.ccm_flags;
> +     ccm->salt[0] = tsalt.inner.salt8[0];
> +     ccm->salt[1] = tsalt.inner.salt8[1];
> +     ccm->salt[2] = tsalt.inner.salt8[2];
> +     ccm->iv = iv;
> +     ccm->cnt = rte_cpu_to_be_32(1);
> +}
> +
> +
>  /*
>   * RFC 4106, 5 AAD Construction
>   * spi and sqn should already be converted into network byte order.
> @@ -86,6 +183,25 @@ aead_gcm_aad_fill(struct aead_gcm_aad *aad, rte_be32_t 
> spi, rte_be64_t sqn,
>       aad->align0 = 0;
>  }
> 
> +/*
> + * RFC 4309, 5 AAD Construction
> + * spi and sqn should already be converted into network byte order.
> + * Make sure that not used bytes are zeroed.
> + */
> +static inline void
> +aead_ccm_aad_fill(struct aead_ccm_aad *aad, rte_be32_t spi, rte_be64_t sqn,
> +     int esn)
> +{
> +     aad->spi = spi;
> +     if (esn)
> +             aad->sqn.u64 = sqn;
> +     else {
> +             aad->sqn.u32[0] = sqn_low32(sqn);
> +             aad->sqn.u32[1] = 0;
> +     }
> +     aad->align0 = 0;
> +}
> +
>  static inline void
>  gen_iv(uint64_t iv[IPSEC_MAX_IV_QWORD], rte_be64_t sqn)
>  {
> @@ -93,6 +209,27 @@ gen_iv(uint64_t iv[IPSEC_MAX_IV_QWORD], rte_be64_t sqn)
>       iv[1] = 0;
>  }
> 
> +
> +/*
> + * RFC 7634, 2.1 AAD Construction
> + * spi and sqn should already be converted into network byte order.
> + * Make sure that not used bytes are zeroed.
> + */
> +static inline void
> +aead_chacha20_poly1305_aad_fill(struct aead_chacha20_poly1305_aad *aad,
> +                                     rte_be32_t spi, rte_be64_t sqn,
> +                                     int esn)
> +{
> +     aad->spi = spi;
> +     if (esn)
> +             aad->sqn.u64 = sqn;
> +     else {
> +             aad->sqn.u32[0] = sqn_low32(sqn);
> +             aad->sqn.u32[1] = 0;
> +     }
> +     aad->align0 = 0;
> +}
> +
>  /*
>   * Helper routine to copy IV
>   * Right now we support only algorithms with IV length equals 0/8/16 bytes.
> diff --git a/lib/ipsec/esp_inb.c b/lib/ipsec/esp_inb.c
> index 2b1df6a032..d66c88f05d 100644
> --- a/lib/ipsec/esp_inb.c
> +++ b/lib/ipsec/esp_inb.c
> @@ -63,6 +63,8 @@ inb_cop_prepare(struct rte_crypto_op *cop,
>  {
>       struct rte_crypto_sym_op *sop;
>       struct aead_gcm_iv *gcm;
> +     struct aead_ccm_iv *ccm;
> +     struct aead_chacha20_poly1305_iv *chacha20_poly1305;
>       struct aesctr_cnt_blk *ctr;
>       uint64_t *ivc, *ivp;
>       uint32_t algo;
> @@ -83,6 +85,24 @@ inb_cop_prepare(struct rte_crypto_op *cop,
>                       sa->iv_ofs);
>               aead_gcm_iv_fill(gcm, ivp[0], sa->salt);
>               break;
> +     case ALGO_TYPE_AES_CCM:
> +             sop_aead_prepare(sop, sa, icv, pofs, plen);
> +
> +             /* fill AAD IV (located inside crypto op) */
> +             ccm = rte_crypto_op_ctod_offset(cop, struct aead_ccm_iv *,
> +                     sa->iv_ofs);
> +             aead_ccm_iv_fill(ccm, ivp[0], sa->salt);
> +             break;
> +     case ALGO_TYPE_CHACHA20_POLY1305:
> +             sop_aead_prepare(sop, sa, icv, pofs, plen);
> +
> +             /* fill AAD IV (located inside crypto op) */
> +             chacha20_poly1305 = rte_crypto_op_ctod_offset(cop,
> +                             struct aead_chacha20_poly1305_iv *,
> +                             sa->iv_ofs);
> +             aead_chacha20_poly1305_iv_fill(chacha20_poly1305,
> +                                            ivp[0], sa->salt);
> +             break;
>       case ALGO_TYPE_AES_CBC:
>       case ALGO_TYPE_3DES_CBC:
>               sop_ciph_auth_prepare(sop, sa, icv, pofs, plen);
> @@ -91,6 +111,14 @@ inb_cop_prepare(struct rte_crypto_op *cop,
>               ivc = rte_crypto_op_ctod_offset(cop, uint64_t *, sa->iv_ofs);
>               copy_iv(ivc, ivp, sa->iv_len);
>               break;
> +     case ALGO_TYPE_AES_GMAC:
> +             sop_ciph_auth_prepare(sop, sa, icv, pofs, plen);
> +
> +             /* fill AAD IV (located inside crypto op) */
> +             gcm = rte_crypto_op_ctod_offset(cop, struct aead_gcm_iv *,
> +                     sa->iv_ofs);
> +             aead_gcm_iv_fill(gcm, ivp[0], sa->salt);
> +             break;
>       case ALGO_TYPE_AES_CTR:
>               sop_ciph_auth_prepare(sop, sa, icv, pofs, plen);
> 
> @@ -110,6 +138,8 @@ inb_cpu_crypto_prepare(const struct rte_ipsec_sa *sa, 
> struct rte_mbuf *mb,
>       uint32_t *pofs, uint32_t plen, void *iv)
>  {
>       struct aead_gcm_iv *gcm;
> +     struct aead_ccm_iv *ccm;
> +     struct aead_chacha20_poly1305_iv *chacha20_poly1305;
>       struct aesctr_cnt_blk *ctr;
>       uint64_t *ivp;
>       uint32_t clen;
> @@ -120,9 +150,19 @@ inb_cpu_crypto_prepare(const struct rte_ipsec_sa *sa, 
> struct rte_mbuf *mb,
> 
>       switch (sa->algo_type) {
>       case ALGO_TYPE_AES_GCM:
> +     case ALGO_TYPE_AES_GMAC:
>               gcm = (struct aead_gcm_iv *)iv;
>               aead_gcm_iv_fill(gcm, ivp[0], sa->salt);
>               break;
> +     case ALGO_TYPE_AES_CCM:
> +             ccm = (struct aead_ccm_iv *)iv;
> +             aead_ccm_iv_fill(ccm, ivp[0], sa->salt);
> +             break;
> +     case ALGO_TYPE_CHACHA20_POLY1305:
> +             chacha20_poly1305 = (struct aead_chacha20_poly1305_iv *)iv;
> +             aead_chacha20_poly1305_iv_fill(chacha20_poly1305,
> +                                            ivp[0], sa->salt);
> +             break;
>       case ALGO_TYPE_AES_CBC:
>       case ALGO_TYPE_3DES_CBC:
>               copy_iv(iv, ivp, sa->iv_len);
> @@ -175,6 +215,8 @@ inb_pkt_xprepare(const struct rte_ipsec_sa *sa, 
> rte_be64_t sqc,
>       const union sym_op_data *icv)
>  {
>       struct aead_gcm_aad *aad;
> +     struct aead_ccm_aad *caad;
> +     struct aead_chacha20_poly1305_aad *chacha_aad;
> 
>       /* insert SQN.hi between ESP trailer and ICV */
>       if (sa->sqh_len != 0)
> @@ -184,9 +226,27 @@ inb_pkt_xprepare(const struct rte_ipsec_sa *sa, 
> rte_be64_t sqc,
>        * fill AAD fields, if any (aad fields are placed after icv),
>        * right now we support only one AEAD algorithm: AES-GCM.
>        */
> -     if (sa->aad_len != 0) {
> -             aad = (struct aead_gcm_aad *)(icv->va + sa->icv_len);
> -             aead_gcm_aad_fill(aad, sa->spi, sqc, IS_ESN(sa));
> +     switch (sa->algo_type) {
> +     case ALGO_TYPE_AES_GCM:
> +             if (sa->aad_len != 0) {
> +                     aad = (struct aead_gcm_aad *)(icv->va + sa->icv_len);
> +                     aead_gcm_aad_fill(aad, sa->spi, sqc, IS_ESN(sa));
> +             }
> +             break;
> +     case ALGO_TYPE_AES_CCM:
> +             if (sa->aad_len != 0) {
> +                     caad = (struct aead_ccm_aad *)(icv->va + sa->icv_len);
> +                     aead_ccm_aad_fill(caad, sa->spi, sqc, IS_ESN(sa));
> +             }
> +             break;
> +     case ALGO_TYPE_CHACHA20_POLY1305:
> +             if (sa->aad_len != 0) {
> +                     chacha_aad = (struct aead_chacha20_poly1305_aad *)
> +                         (icv->va + sa->icv_len);
> +                     aead_chacha20_poly1305_aad_fill(chacha_aad,
> +                                             sa->spi, sqc, IS_ESN(sa));
> +             }
> +             break;
>       }
>  }
> 
> diff --git a/lib/ipsec/esp_outb.c b/lib/ipsec/esp_outb.c
> index 1e181cf2ce..a3f77469c3 100644
> --- a/lib/ipsec/esp_outb.c
> +++ b/lib/ipsec/esp_outb.c
> @@ -63,6 +63,8 @@ outb_cop_prepare(struct rte_crypto_op *cop,
>  {
>       struct rte_crypto_sym_op *sop;
>       struct aead_gcm_iv *gcm;
> +     struct aead_ccm_iv *ccm;
> +     struct aead_chacha20_poly1305_iv *chacha20_poly1305;
>       struct aesctr_cnt_blk *ctr;
>       uint32_t algo;
> 
> @@ -80,6 +82,15 @@ outb_cop_prepare(struct rte_crypto_op *cop,
>               /* NULL case */
>               sop_ciph_auth_prepare(sop, sa, icv, hlen, plen);
>               break;
> +     case ALGO_TYPE_AES_GMAC:
> +             /* GMAC case */
> +             sop_ciph_auth_prepare(sop, sa, icv, hlen, plen);
> +
> +             /* fill AAD IV (located inside crypto op) */
> +             gcm = rte_crypto_op_ctod_offset(cop, struct aead_gcm_iv *,
> +                     sa->iv_ofs);
> +             aead_gcm_iv_fill(gcm, ivp[0], sa->salt);
> +             break;
>       case ALGO_TYPE_AES_GCM:
>               /* AEAD (AES_GCM) case */
>               sop_aead_prepare(sop, sa, icv, hlen, plen);
> @@ -89,6 +100,26 @@ outb_cop_prepare(struct rte_crypto_op *cop,
>                       sa->iv_ofs);
>               aead_gcm_iv_fill(gcm, ivp[0], sa->salt);
>               break;
> +     case ALGO_TYPE_AES_CCM:
> +             /* AEAD (AES_CCM) case */
> +             sop_aead_prepare(sop, sa, icv, hlen, plen);
> +
> +             /* fill AAD IV (located inside crypto op) */
> +             ccm = rte_crypto_op_ctod_offset(cop, struct aead_ccm_iv *,
> +                     sa->iv_ofs);
> +             aead_ccm_iv_fill(ccm, ivp[0], sa->salt);
> +             break;
> +     case ALGO_TYPE_CHACHA20_POLY1305:
> +             /* AEAD (CHACHA20_POLY) case */
> +             sop_aead_prepare(sop, sa, icv, hlen, plen);
> +
> +             /* fill AAD IV (located inside crypto op) */
> +             chacha20_poly1305 = rte_crypto_op_ctod_offset(cop,
> +                     struct aead_chacha20_poly1305_iv *,
> +                     sa->iv_ofs);
> +             aead_chacha20_poly1305_iv_fill(chacha20_poly1305,
> +                                            ivp[0], sa->salt);
> +             break;
>       case ALGO_TYPE_AES_CTR:
>               /* Cipher-Auth (AES-CTR *) case */
>               sop_ciph_auth_prepare(sop, sa, icv, hlen, plen);
> @@ -196,7 +227,9 @@ outb_pkt_xprepare(const struct rte_ipsec_sa *sa, 
> rte_be64_t sqc,
>       const union sym_op_data *icv)
>  {
>       uint32_t *psqh;
> -     struct aead_gcm_aad *aad;
> +     struct aead_gcm_aad *gaad;
> +     struct aead_ccm_aad *caad;
> +     struct aead_chacha20_poly1305_aad *chacha20_poly1305_aad;
> 
>       /* insert SQN.hi between ESP trailer and ICV */
>       if (sa->sqh_len != 0) {
> @@ -208,9 +241,29 @@ outb_pkt_xprepare(const struct rte_ipsec_sa *sa, 
> rte_be64_t sqc,
>        * fill IV and AAD fields, if any (aad fields are placed after icv),
>        * right now we support only one AEAD algorithm: AES-GCM .
>        */
> +     switch (sa->algo_type) {
> +     case ALGO_TYPE_AES_GCM:
>       if (sa->aad_len != 0) {
> -             aad = (struct aead_gcm_aad *)(icv->va + sa->icv_len);
> -             aead_gcm_aad_fill(aad, sa->spi, sqc, IS_ESN(sa));
> +             gaad = (struct aead_gcm_aad *)(icv->va + sa->icv_len);
> +             aead_gcm_aad_fill(gaad, sa->spi, sqc, IS_ESN(sa));
> +     }
> +             break;
> +     case ALGO_TYPE_AES_CCM:
> +     if (sa->aad_len != 0) {
> +             caad = (struct aead_ccm_aad *)(icv->va + sa->icv_len);
> +             aead_ccm_aad_fill(caad, sa->spi, sqc, IS_ESN(sa));
> +     }
> +             break;
> +     case ALGO_TYPE_CHACHA20_POLY1305:
> +     if (sa->aad_len != 0) {
> +             chacha20_poly1305_aad = (struct aead_chacha20_poly1305_aad *)
> +                     (icv->va + sa->icv_len);
> +             aead_chacha20_poly1305_aad_fill(chacha20_poly1305_aad,
> +                     sa->spi, sqc, IS_ESN(sa));
> +     }
> +             break;
> +     default:
> +             break;
>       }
>  }
> 
> @@ -418,6 +471,8 @@ outb_cpu_crypto_prepare(const struct rte_ipsec_sa *sa, 
> uint32_t *pofs,
>  {
>       uint64_t *ivp = iv;
>       struct aead_gcm_iv *gcm;
> +     struct aead_ccm_iv *ccm;
> +     struct aead_chacha20_poly1305_iv *chacha20_poly1305;
>       struct aesctr_cnt_blk *ctr;
>       uint32_t clen;
> 
> @@ -426,6 +481,15 @@ outb_cpu_crypto_prepare(const struct rte_ipsec_sa *sa, 
> uint32_t *pofs,
>               gcm = iv;
>               aead_gcm_iv_fill(gcm, ivp[0], sa->salt);
>               break;
> +     case ALGO_TYPE_AES_CCM:
> +             ccm = iv;
> +             aead_ccm_iv_fill(ccm, ivp[0], sa->salt);
> +             break;
> +     case ALGO_TYPE_CHACHA20_POLY1305:
> +             chacha20_poly1305 = iv;
> +             aead_chacha20_poly1305_iv_fill(chacha20_poly1305,
> +                                            ivp[0], sa->salt);
> +             break;
>       case ALGO_TYPE_AES_CTR:
>               ctr = iv;
>               aes_ctr_cnt_blk_fill(ctr, ivp[0], sa->salt);
> diff --git a/lib/ipsec/sa.c b/lib/ipsec/sa.c
> index e59189d215..720e0f365b 100644
> --- a/lib/ipsec/sa.c
> +++ b/lib/ipsec/sa.c
> @@ -47,6 +47,15 @@ fill_crypto_xform(struct crypto_xform *xform, uint64_t 
> type,
>               if (xfn != NULL)
>                       return -EINVAL;
>               xform->aead = &xf->aead;
> +
> +     /* GMAC has only auth */
> +     } else if (xf->type == RTE_CRYPTO_SYM_XFORM_AUTH &&
> +                     xf->auth.algo == RTE_CRYPTO_AUTH_AES_GMAC) {
> +             if (xfn != NULL)
> +                     return -EINVAL;
> +             xform->auth = &xf->auth;
> +             xform->cipher = &xfn->cipher;
> +
>       /*
>        * CIPHER+AUTH xforms are expected in strict order,
>        * depending on SA direction:
> @@ -247,12 +256,13 @@ esp_inb_init(struct rte_ipsec_sa *sa)
>       sa->ctp.cipher.length = sa->icv_len + sa->ctp.cipher.offset;
> 
>       /*
> -      * for AEAD and NULL algorithms we can assume that
> +      * for AEAD algorithms we can assume that
>        * auth and cipher offsets would be equal.
>        */
>       switch (sa->algo_type) {
>       case ALGO_TYPE_AES_GCM:
> -     case ALGO_TYPE_NULL:
> +     case ALGO_TYPE_AES_CCM:
> +     case ALGO_TYPE_CHACHA20_POLY1305:
>               sa->ctp.auth.raw = sa->ctp.cipher.raw;
>               break;
>       default:
> @@ -294,6 +304,8 @@ esp_outb_init(struct rte_ipsec_sa *sa, uint32_t hlen)
> 
>       switch (algo_type) {
>       case ALGO_TYPE_AES_GCM:
> +     case ALGO_TYPE_AES_CCM:
> +     case ALGO_TYPE_CHACHA20_POLY1305:
>       case ALGO_TYPE_AES_CTR:
>       case ALGO_TYPE_NULL:
>               sa->ctp.cipher.offset = hlen + sizeof(struct rte_esp_hdr) +
> @@ -305,15 +317,20 @@ esp_outb_init(struct rte_ipsec_sa *sa, uint32_t hlen)
>               sa->ctp.cipher.offset = hlen + sizeof(struct rte_esp_hdr);
>               sa->ctp.cipher.length = sa->iv_len;
>               break;
> +     case ALGO_TYPE_AES_GMAC:
> +             sa->ctp.cipher.offset = 0;
> +             sa->ctp.cipher.length = 0;
> +             break;
>       }
> 
>       /*
> -      * for AEAD and NULL algorithms we can assume that
> +      * for AEAD algorithms we can assume that
>        * auth and cipher offsets would be equal.
>        */
>       switch (algo_type) {
>       case ALGO_TYPE_AES_GCM:
> -     case ALGO_TYPE_NULL:
> +     case ALGO_TYPE_AES_CCM:
> +     case ALGO_TYPE_CHACHA20_POLY1305:
>               sa->ctp.auth.raw = sa->ctp.cipher.raw;
>               break;
>       default:
> @@ -374,13 +391,39 @@ esp_sa_init(struct rte_ipsec_sa *sa, const struct 
> rte_ipsec_sa_prm *prm,
>                       sa->pad_align = IPSEC_PAD_AES_GCM;
>                       sa->algo_type = ALGO_TYPE_AES_GCM;
>                       break;
> +             case RTE_CRYPTO_AEAD_AES_CCM:
> +                     /* RFC 4309 */
> +                     sa->aad_len = sizeof(struct aead_ccm_aad);
> +                     sa->icv_len = cxf->aead->digest_length;
> +                     sa->iv_ofs = cxf->aead->iv.offset;
> +                     sa->iv_len = sizeof(uint64_t);
> +                     sa->pad_align = IPSEC_PAD_AES_CCM;
> +                     sa->algo_type = ALGO_TYPE_AES_CCM;
> +                     break;
> +             case RTE_CRYPTO_AEAD_CHACHA20_POLY1305:
> +                     /* RFC 7634 & 8439*/
> +                     sa->aad_len = sizeof(struct aead_chacha20_poly1305_aad);
> +                     sa->icv_len = cxf->aead->digest_length;
> +                     sa->iv_ofs = cxf->aead->iv.offset;
> +                     sa->iv_len = sizeof(uint64_t);
> +                     sa->pad_align = IPSEC_PAD_CHACHA20_POLY1305;
> +                     sa->algo_type = ALGO_TYPE_CHACHA20_POLY1305;
> +                     break;
>               default:
>                       return -EINVAL;
>               }
> +     } else if (cxf->auth->algo == RTE_CRYPTO_AUTH_AES_GMAC) {
> +             /* RFC 4543 */
> +             /* AES-GMAC is a special case of auth that needs IV */
> +             sa->pad_align = IPSEC_PAD_AES_GMAC;
> +             sa->iv_len = sizeof(uint64_t);
> +             sa->icv_len = cxf->auth->digest_length;
> +             sa->iv_ofs = cxf->auth->iv.offset;
> +             sa->algo_type = ALGO_TYPE_AES_GMAC;
> +
>       } else {
>               sa->icv_len = cxf->auth->digest_length;
>               sa->iv_ofs = cxf->cipher->iv.offset;
> -             sa->sqh_len = IS_ESN(sa) ? sizeof(uint32_t) : 0;
> 
>               switch (cxf->cipher->algo) {
>               case RTE_CRYPTO_CIPHER_NULL:
> @@ -414,6 +457,7 @@ esp_sa_init(struct rte_ipsec_sa *sa, const struct 
> rte_ipsec_sa_prm *prm,
>               }
>       }
> 
> +     sa->sqh_len = IS_ESN(sa) ? sizeof(uint32_t) : 0;
>       sa->udata = prm->userdata;
>       sa->spi = rte_cpu_to_be_32(prm->ipsec_xform.spi);
>       sa->salt = prm->ipsec_xform.salt;
> diff --git a/lib/ipsec/sa.h b/lib/ipsec/sa.h
> index 1bffe751f5..107ebd1519 100644
> --- a/lib/ipsec/sa.h
> +++ b/lib/ipsec/sa.h
> @@ -19,7 +19,10 @@ enum {
>       IPSEC_PAD_AES_CBC = IPSEC_MAX_IV_SIZE,
>       IPSEC_PAD_AES_CTR = IPSEC_PAD_DEFAULT,
>       IPSEC_PAD_AES_GCM = IPSEC_PAD_DEFAULT,
> +     IPSEC_PAD_AES_CCM = IPSEC_PAD_DEFAULT,
> +     IPSEC_PAD_CHACHA20_POLY1305 = IPSEC_PAD_DEFAULT,
>       IPSEC_PAD_NULL = IPSEC_PAD_DEFAULT,
> +     IPSEC_PAD_AES_GMAC = IPSEC_PAD_DEFAULT,
>  };
> 
>  /* iv sizes for different algorithms */
> @@ -67,6 +70,9 @@ enum sa_algo_type   {
>       ALGO_TYPE_AES_CBC,
>       ALGO_TYPE_AES_CTR,
>       ALGO_TYPE_AES_GCM,
> +     ALGO_TYPE_AES_CCM,
> +     ALGO_TYPE_CHACHA20_POLY1305,
> +     ALGO_TYPE_AES_GMAC,
>       ALGO_TYPE_MAX
>  };
> 
> --
> 2.25.1

Reply via email to