Thanks, PSB.

> -----Original Message-----
> From: Jack Min
> Sent: Monday, July 22, 2019 12:31 PM
> To: Dekel Peled <dek...@mellanox.com>
> Cc: Yongseok Koh <ys...@mellanox.com>; Slava Ovsiienko
> <viachesl...@mellanox.com>; Shahaf Shuler <shah...@mellanox.com>; Ori
> Kam <or...@mellanox.com>; dev@dpdk.org
> Subject: Re: [PATCH] net/mlx5: fix NVGRE matching
> 
> On Thu, 19-07-18, 22:42, Dekel Peled wrote:
> > NVGRE has a GRE header with c_rsvd0_ver value 0x2000 and protocol
> > value 0x6558.
> > These should be matched when item_nvgre is provided.
> >
> > This patch adds validation function of NVGRE item, to validate that
> > the input values, if exist, are as required.
> > It also updates the translate function of NVGRE item, to add the
> > required values, if they were not specified.
> >
> > Original work by Xiaoyu Min <jack...@mellanox.com>
> >
> > Fixes: fc2c498ccb94 ("net/mlx5: add Direct Verbs translate items")
> > Signed-off-by: Dekel Peled <dek...@mellanox.com>
> > ---
> >  drivers/net/mlx5/mlx5_flow.c    | 69
> +++++++++++++++++++++++++++++++++++++++++
> >  drivers/net/mlx5/mlx5_flow.h    | 10 ++++--
> >  drivers/net/mlx5/mlx5_flow_dv.c | 25 +++++++++++++--
> >  drivers/net/mlx5/mlx5_rxtx.h    |  2 +-
> >  4 files changed, 101 insertions(+), 5 deletions(-)
> >
> > diff --git a/drivers/net/mlx5/mlx5_flow.c
> > b/drivers/net/mlx5/mlx5_flow.c index e082cbb..6aca4d6 100644
> > --- a/drivers/net/mlx5/mlx5_flow.c
> > +++ b/drivers/net/mlx5/mlx5_flow.c
> > @@ -298,6 +298,10 @@ struct mlx5_flow_tunnel_info {
> >             .tunnel = MLX5_FLOW_LAYER_MPLS,
> >             .ptype = RTE_PTYPE_TUNNEL_MPLS_IN_GRE,
> >     },
> > +   {
> > +           .tunnel = MLX5_FLOW_LAYER_NVGRE,
> > +           .ptype = RTE_PTYPE_TUNNEL_NVGRE,
> > +   },
> >  };
> >
> >  /**
> > @@ -1323,6 +1327,11 @@ uint32_t mlx5_flow_adjust_priority(struct
> rte_eth_dev *dev, int32_t priority,
> >             return rte_flow_error_set(error, EINVAL,
> >                                       RTE_FLOW_ERROR_TYPE_ITEM,
> item,
> >                                       "L3 cannot follow an L4 layer.");
> > +   else if ((item_flags & MLX5_FLOW_LAYER_NVGRE) &&
> > +             !(item_flags & MLX5_FLOW_LAYER_INNER_L2))
> > +           return rte_flow_error_set(error, EINVAL,
> > +                                     RTE_FLOW_ERROR_TYPE_ITEM,
> item,
> > +                                     "L3 cannot follow an NVGRE layer.");
> >     if (!mask)
> >             mask = &rte_flow_item_ipv4_mask;
> >     else if (mask->hdr.next_proto_id != 0 && @@ -1409,6 +1418,11 @@
> > uint32_t mlx5_flow_adjust_priority(struct rte_eth_dev *dev, int32_t
> priority,
> >             return rte_flow_error_set(error, EINVAL,
> >                                       RTE_FLOW_ERROR_TYPE_ITEM,
> item,
> >                                       "L3 cannot follow an L4 layer.");
> > +   else if ((item_flags & MLX5_FLOW_LAYER_NVGRE) &&
> > +             !(item_flags & MLX5_FLOW_LAYER_INNER_L2))
> > +           return rte_flow_error_set(error, EINVAL,
> > +                                     RTE_FLOW_ERROR_TYPE_ITEM,
> item,
> > +                                     "L3 cannot follow an NVGRE layer.");
> >     if (!mask)
> >             mask = &rte_flow_item_ipv6_mask;
> >     ret = mlx5_flow_item_acceptable(item, (const uint8_t *)mask, @@
> > -1887,6 +1901,61 @@ uint32_t mlx5_flow_adjust_priority(struct
> rte_eth_dev *dev, int32_t priority,
> >                               " update.");
> >  }
> >
> > +/**
> > + * Validate NVGRE item.
> > + *
> > + * @param[in] item
> > + *   Item specification.
> > + * @param[in] item_flags
> > + *   Bit flags to mark detected items.
> > + * @param[in] target_protocol
> > + *   The next protocol in the previous item.
> > + * @param[out] error
> > + *   Pointer to error structure.
> > + *
> > + * @return
> > + *   0 on success, a negative errno value otherwise and rte_errno is set.
> > + */
> > +int
> > +mlx5_flow_validate_item_nvgre(const struct rte_flow_item *item,
> > +                         uint64_t item_flags,
> > +                         uint8_t target_protocol,
> > +                         struct rte_flow_error *error) {
> > +   const struct rte_flow_item_nvgre *mask = item->mask;
> > +   const struct rte_flow_item_nvgre *spec = item->spec;
> > +   int ret;
> > +
> > +   if (target_protocol != 0xff && target_protocol != IPPROTO_GRE)
> > +           return rte_flow_error_set(error, EINVAL,
> > +                                     RTE_FLOW_ERROR_TYPE_ITEM,
> item,
> > +                                     "protocol filtering not compatible"
> > +                                     " with this GRE layer");
> > +   if (item_flags & MLX5_FLOW_LAYER_TUNNEL)
> > +           return rte_flow_error_set(error, ENOTSUP,
> > +                                     RTE_FLOW_ERROR_TYPE_ITEM,
> item,
> > +                                     "multiple tunnel layers not"
> > +                                     " supported");
> > +   if (!(item_flags & MLX5_FLOW_LAYER_OUTER_L3))
> > +           return rte_flow_error_set(error, ENOTSUP,
> > +                                     RTE_FLOW_ERROR_TYPE_ITEM,
> item,
> > +                                     "L3 Layer is missing");
> > +   if (spec && (spec->protocol != RTE_BE16(RTE_ETHER_TYPE_TEB) ||
> > +                spec->c_k_s_rsvd0_ver != RTE_BE16(0x2000)))
> > +           return rte_flow_error_set(error, EINVAL,
> > +                                     RTE_FLOW_ERROR_TYPE_ITEM,
> item,
> > +                                     "wrong values for NVGRE");
> Not necessary to check the spec because the following
> mlx5_flow_item_acceptable only accept matching on .tni field.
> Since there is no meaning allowing the user to match on .protocol and
> .c_k_s_rsvd0_ver.
> What do you think?

I think it is safer to verify that application doesn't insert invalid values.
It covers any future pmd or applications change.

> 
> > +   if (!mask)
> > +           mask = &rte_flow_item_nvgre_mask;
> > +   ret = mlx5_flow_item_acceptable
> > +           (item, (const uint8_t *)mask,
> > +            (const uint8_t *)&rte_flow_item_nvgre_mask,
> > +            sizeof(struct rte_flow_item_nvgre), error);
> > +   if (ret < 0)
> > +           return ret;
> > +   return 0;
> > +}
> > +
> >  static int
> >  flow_null_validate(struct rte_eth_dev *dev __rte_unused,
> >                const struct rte_flow_attr *attr __rte_unused, diff --git
> > a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h index
> > 3f96bec..24da74b 100644
> > --- a/drivers/net/mlx5/mlx5_flow.h
> > +++ b/drivers/net/mlx5/mlx5_flow.h
> > @@ -48,6 +48,7 @@
> >  #define MLX5_FLOW_LAYER_VXLAN_GPE (1u << 13)  #define
> > MLX5_FLOW_LAYER_GRE (1u << 14)  #define MLX5_FLOW_LAYER_MPLS
> (1u <<
> > 15)
> > +/* List of tunnel Layer bits continued below. */
> >
> >  /* General pattern items bits. */
> >  #define MLX5_FLOW_ITEM_METADATA (1u << 16) @@ -58,8 +59,10 @@
> > #define MLX5_FLOW_LAYER_ICMP6 (1u << 19)  #define
> > MLX5_FLOW_LAYER_GRE_KEY (1u << 20)
> >
> > +/* Pattern tunnel Layer bits (continued). */
> >  #define MLX5_FLOW_LAYER_IPIP (1u << 21)  #define
> > MLX5_FLOW_LAYER_IPV6_ENCAP (1u << 22)
> > +#define MLX5_FLOW_LAYER_NVGRE (1u << 23)
> >
> >  /* Outer Masks. */
> >  #define MLX5_FLOW_LAYER_OUTER_L3 \
> > @@ -79,7 +82,7 @@
> >  /* Tunnel Masks. */
> >  #define MLX5_FLOW_LAYER_TUNNEL \
> >     (MLX5_FLOW_LAYER_VXLAN | MLX5_FLOW_LAYER_VXLAN_GPE | \
> > -    MLX5_FLOW_LAYER_GRE | MLX5_FLOW_LAYER_MPLS | \
> > +    MLX5_FLOW_LAYER_GRE | MLX5_FLOW_LAYER_NVGRE |
> MLX5_FLOW_LAYER_MPLS
> > +| \
> >      MLX5_FLOW_LAYER_IPIP | MLX5_FLOW_LAYER_IPV6_ENCAP)
> >
> >  /* Inner Masks. */
> > @@ -518,5 +521,8 @@ int mlx5_flow_validate_item_icmp6(const struct
> rte_flow_item *item,
> >                                uint64_t item_flags,
> >                                uint8_t target_protocol,
> >                                struct rte_flow_error *error);
> > -
> > +int mlx5_flow_validate_item_nvgre(const struct rte_flow_item *item,
> > +                             uint64_t item_flags,
> > +                             uint8_t target_protocol,
> > +                             struct rte_flow_error *error);
> >  #endif /* RTE_PMD_MLX5_FLOW_H_ */
> > diff --git a/drivers/net/mlx5/mlx5_flow_dv.c
> > b/drivers/net/mlx5/mlx5_flow_dv.c index 7240d3b..ab758d4 100644
> > --- a/drivers/net/mlx5/mlx5_flow_dv.c
> > +++ b/drivers/net/mlx5/mlx5_flow_dv.c
> > @@ -2966,7 +2966,6 @@ struct field_modify_info modify_tcp[] = {
> >
> MLX5_FLOW_LAYER_OUTER_L4_UDP;
> >                     break;
> >             case RTE_FLOW_ITEM_TYPE_GRE:
> > -           case RTE_FLOW_ITEM_TYPE_NVGRE:
> >                     ret = mlx5_flow_validate_item_gre(items,
> item_flags,
> >                                                       next_protocol,
> error);
> >                     if (ret < 0)
> > @@ -2974,6 +2973,14 @@ struct field_modify_info modify_tcp[] = {
> >                     gre_item = items;
> >                     last_item = MLX5_FLOW_LAYER_GRE;
> >                     break;
> > +           case RTE_FLOW_ITEM_TYPE_NVGRE:
> > +                   ret = mlx5_flow_validate_item_nvgre(items,
> item_flags,
> > +                                                       next_protocol,
> > +                                                       error);
> > +                   if (ret < 0)
> > +                           return ret;
> > +                   last_item = MLX5_FLOW_LAYER_NVGRE;
> > +                   break;
> >             case RTE_FLOW_ITEM_TYPE_GRE_KEY:
> >                     ret = mlx5_flow_validate_item_gre_key
> >                             (items, item_flags, gre_item, error); @@ -
> 3919,7 +3926,21 @@
> > struct field_modify_info modify_tcp[] = {
> >     int size;
> >     int i;
> >
> > -   flow_dv_translate_item_gre(matcher, key, item, inner);
> > +   /* For NVGRE, GRE header fields must be set with defined values. */
> > +   const struct rte_flow_item_gre gre_spec = {
> > +           .c_rsvd0_ver = RTE_BE16(0x2000),
> > +           .protocol = RTE_BE16(RTE_ETHER_TYPE_TEB)
> > +   };
> > +   const struct rte_flow_item_gre gre_mask = {
> > +           .c_rsvd0_ver = RTE_BE16(UINT16_MAX),
> Well, it should be `RTE_BE16(0xB000)`, which, I think, is more explicit.
> Because our NIC only support matching on C,K,S bits, not else bits in
> c_rsvd0_ver. Our PMD just ignore the other bits.

The spec is specific 0x2000 and with full 0xffff mask to make sure all bits are 
covered.

> 
> > +           .protocol = RTE_BE16(UINT16_MAX),
> > +   };
> > +   const struct rte_flow_item gre_item = {
> > +           .spec = &gre_spec,
> > +           .mask = &gre_mask,
> > +           .last = NULL,
> > +   };
> > +   flow_dv_translate_item_gre(matcher, key, &gre_item, inner);
> >     if (!nvgre_v)
> >             return;
> >     if (!nvgre_m)
> > diff --git a/drivers/net/mlx5/mlx5_rxtx.h
> > b/drivers/net/mlx5/mlx5_rxtx.h index dfa79e2..d732757 100644
> > --- a/drivers/net/mlx5/mlx5_rxtx.h
> > +++ b/drivers/net/mlx5/mlx5_rxtx.h
> > @@ -40,7 +40,7 @@
> >  #include "mlx5_glue.h"
> >
> >  /* Support tunnel matching. */
> > -#define MLX5_FLOW_TUNNEL 5
> > +#define MLX5_FLOW_TUNNEL 6
> >
> >  struct mlx5_rxq_stats {
> >  #ifdef MLX5_PMD_SOFT_COUNTERS
> > --
> > 1.8.3.1
> >

Reply via email to