On Wed, Jul 11, 2018 at 09:22:47AM +0200, Nelio Laranjeiro wrote:
> Signed-off-by: Nelio Laranjeiro <nelio.laranje...@6wind.com>
> ---
>  drivers/net/mlx5/mlx5_flow.c | 682 +++++++++++++++++++++++++++--------
>  1 file changed, 537 insertions(+), 145 deletions(-)
> 
[...]
> @@ -1322,26 +1583,122 @@ mlx5_flow_merge(struct rte_eth_dev *dev, struct 
> rte_flow *flow,
>               struct rte_flow_error *error)
>  {
>       struct rte_flow local_flow = { .layers = 0, };
> -     size_t size = sizeof(*flow) + sizeof(struct ibv_flow_attr);
> -     int remain = (flow_size > size) ? flow_size - size : 0;
> +     size_t size = sizeof(*flow);
> +     union {
> +             struct rte_flow_expand_rss buf;
> +             uint8_t buffer[2048];
> +     } expand_buffer;
> +     struct rte_flow_expand_rss *buf = &expand_buffer.buf;
> +     struct mlx5_flow_verbs *original_verbs = NULL;
> +     size_t original_verbs_size = 0;
> +     uint32_t original_layers = 0;
>       int ret;
> +     uint32_t i;
>  
> -     if (!remain)
> +     if (size > flow_size)
>               flow = &local_flow;
>       ret = mlx5_flow_attributes(dev, attributes, flow, error);
>       if (ret < 0)
>               return ret;
> -     ret = mlx5_flow_items(pattern, flow, remain, error);
> -     if (ret < 0)
> -             return ret;
> -     size += ret;
> -     remain = (flow_size > size) ? flow_size - size : 0;
> -     ret = mlx5_flow_actions(dev, actions, flow, remain, error);
> +     ret = mlx5_flow_actions(dev, actions, &local_flow, 0, error);
>       if (ret < 0)
>               return ret;
> -     size += ret;
> +     if (local_flow.rss.types) {
> +             ret = rte_flow_expand_rss(buf, sizeof(expand_buffer.buffer),
> +                                       pattern, local_flow.rss.types,
> +                                       mlx5_support_expansion,
> +                                       MLX5_EXPANSION_ROOT);
> +             assert(ret > 0 &&
> +                    (unsigned int)ret < sizeof(expand_buffer.buffer));
> +     } else {
> +             buf->entries = 1;
> +             buf->entry[0].pattern = (void *)(uintptr_t)pattern;
> +     }
> +     size += RTE_ALIGN_CEIL(local_flow.rss.queue_num * sizeof(uint16_t),
> +                            sizeof(void *));
>       if (size <= flow_size)
> -             flow->verbs.attr->priority = flow->attributes.priority;
> +             flow->queue = (void *)(flow + 1);
> +     LIST_INIT(&flow->verbs);
> +     flow->layers = 0;
> +     flow->modifier = 0;
> +     flow->fate = 0;
> +     for (i = 0; i != buf->entries; ++i) {
> +             size_t off = size;
> +             size_t off2;
> +             int pattern_start_idx = 0;
> +
> +             flow->layers = original_layers;
> +             size += sizeof(struct ibv_flow_attr) +
> +                     sizeof(struct mlx5_flow_verbs);
> +             off2 = size;
> +             if (size < flow_size) {
> +                     flow->cur_verbs = (void *)((uintptr_t)flow + off);
> +                     flow->cur_verbs->attr = (void *)(flow->cur_verbs + 1);
> +                     flow->cur_verbs->specs =
> +                             (void *)(flow->cur_verbs->attr + 1);
> +             }
> +             /* First iteration convert the pattern into Verbs. */
> +             if (i == 0) {
> +                     /* Actions don't need to be converted several time. */
> +                     ret = mlx5_flow_actions(dev, actions, flow,
> +                                             (size < flow_size) ?
> +                                             flow_size - size : 0,
> +                                             error);
> +                     if (ret < 0)
> +                             return ret;
> +                     size += ret;
> +             } else {
> +                     /*
> +                      * Next iteration means the pattern has already been
> +                      * converted and an expansion is necessary to match
> +                      * the user RSS request.  For that only the expanded
> +                      * items will be converted, the common part with the
> +                      * user pattern are just copied into the next buffer
> +                      * zone.
> +                      */
> +                     const struct rte_flow_item *item = pattern;
> +
> +                     size += original_verbs_size;
> +                     if (size < flow_size) {
> +                             rte_memcpy(flow->cur_verbs->attr,
> +                                        original_verbs->attr,
> +                                        original_verbs_size +
> +                                        sizeof(struct ibv_flow_attr));
> +                             flow->cur_verbs->size = original_verbs_size;
> +                     }
> +                     if (pattern->type == RTE_FLOW_ITEM_TYPE_END) {
> +                             pattern_start_idx++;
> +                     } else {
> +                             for (item = pattern;
> +                                  item->type != RTE_FLOW_ITEM_TYPE_END;
> +                                  ++item)
> +                                     pattern_start_idx++;
> +                     }

Small nit.
Can't this be run once when? As pattern isn't changing, it might be run 
redundant
from i > 1. And how about expanded_pattern_idx instead of pattern_start_idx?

Acked-by: Yongseok Koh <ys...@mellanox.com>

Thanks

> +             }
> +             ret = mlx5_flow_items
> +                     ((const struct rte_flow_item *)
> +                      &buf->entry[i].pattern[pattern_start_idx],
> +                      flow,
> +                      (size < flow_size) ? flow_size - size : 0, error);
> +             if (ret < 0)
> +                     return ret;
> +             size += ret;
> +             if (size <= flow_size) {
> +                     mlx5_flow_adjust_priority(dev, flow);
> +                     LIST_INSERT_HEAD(&flow->verbs, flow->cur_verbs, next);
> +             }
> +             /*
> +              * Keep a pointer of the first verbs conversion and the layers
> +              * it has encountered.
> +              */
> +             if (i == 0) {
> +                     original_verbs = flow->cur_verbs;
> +                     original_verbs_size = size - off2;
> +                     original_layers = flow->layers;
> +             }
> +     }
> +     /* Restore the origin layers in the flow. */
> +     flow->layers = original_layers;
>       return size;
>  }
>  
> @@ -1359,12 +1716,17 @@ mlx5_flow_rxq_mark_set(struct rte_eth_dev *dev, 
> struct rte_flow *flow)
>       struct priv *priv = dev->data->dev_private;
>  
>       if (flow->modifier & (MLX5_FLOW_MOD_FLAG | MLX5_FLOW_MOD_MARK)) {
> -             struct mlx5_rxq_ctrl *rxq_ctrl =
> -                     container_of((*priv->rxqs)[flow->queue],
> -                                  struct mlx5_rxq_ctrl, rxq);
> +             unsigned int i;
> +
> +             for (i = 0; i != flow->rss.queue_num; ++i) {
> +                     int idx = (*flow->queue)[i];
> +                     struct mlx5_rxq_ctrl *rxq_ctrl =
> +                             container_of((*priv->rxqs)[idx],
> +                                          struct mlx5_rxq_ctrl, rxq);
>  
> -             rxq_ctrl->rxq.mark = 1;
> -             rxq_ctrl->flow_mark_n++;
> +                     rxq_ctrl->rxq.mark = 1;
> +                     rxq_ctrl->flow_mark_n++;
> +             }
>       }
>  }
>  
> @@ -1383,12 +1745,17 @@ mlx5_flow_rxq_mark_trim(struct rte_eth_dev *dev, 
> struct rte_flow *flow)
>       struct priv *priv = dev->data->dev_private;
>  
>       if (flow->modifier & (MLX5_FLOW_MOD_FLAG | MLX5_FLOW_MOD_MARK)) {
> -             struct mlx5_rxq_ctrl *rxq_ctrl =
> -                     container_of((*priv->rxqs)[flow->queue],
> -                                  struct mlx5_rxq_ctrl, rxq);
> +             unsigned int i;
> +
> +             for (i = 0; i != flow->rss.queue_num; ++i) {
> +                     int idx = (*flow->queue)[i];
> +                     struct mlx5_rxq_ctrl *rxq_ctrl =
> +                             container_of((*priv->rxqs)[idx],
> +                                          struct mlx5_rxq_ctrl, rxq);
>  
> -             rxq_ctrl->flow_mark_n--;
> -             rxq_ctrl->rxq.mark = !!rxq_ctrl->flow_mark_n;
> +                     rxq_ctrl->flow_mark_n--;
> +                     rxq_ctrl->rxq.mark = !!rxq_ctrl->flow_mark_n;
> +             }
>       }
>  }
>  
> @@ -1449,18 +1816,20 @@ mlx5_flow_validate(struct rte_eth_dev *dev,
>  static void
>  mlx5_flow_remove(struct rte_eth_dev *dev, struct rte_flow *flow)
>  {
> -     if (flow->fate & MLX5_FLOW_FATE_DROP) {
> -             if (flow->verbs.flow) {
> -                     claim_zero(mlx5_glue->destroy_flow(flow->verbs.flow));
> -                     flow->verbs.flow = NULL;
> +     struct mlx5_flow_verbs *verbs;
> +
> +     LIST_FOREACH(verbs, &flow->verbs, next) {
> +             if (verbs->flow) {
> +                     claim_zero(mlx5_glue->destroy_flow(verbs->flow));
> +                     verbs->flow = NULL;
> +             }
> +             if (verbs->hrxq) {
> +                     if (flow->fate & MLX5_FLOW_FATE_DROP)
> +                             mlx5_hrxq_drop_release(dev);
> +                     else
> +                             mlx5_hrxq_release(dev, verbs->hrxq);
> +                     verbs->hrxq = NULL;
>               }
> -     }
> -     if (flow->verbs.hrxq) {
> -             if (flow->fate & MLX5_FLOW_FATE_DROP)
> -                     mlx5_hrxq_drop_release(dev);
> -             else if (flow->fate & MLX5_FLOW_FATE_QUEUE)
> -                     mlx5_hrxq_release(dev, flow->verbs.hrxq);
> -             flow->verbs.hrxq = NULL;
>       }
>  }
>  
> @@ -1481,46 +1850,68 @@ static int
>  mlx5_flow_apply(struct rte_eth_dev *dev, struct rte_flow *flow,
>               struct rte_flow_error *error)
>  {
> -     if (flow->fate & MLX5_FLOW_FATE_DROP) {
> -             flow->verbs.hrxq = mlx5_hrxq_drop_new(dev);
> -             if (!flow->verbs.hrxq)
> -                     return rte_flow_error_set
> -                             (error, errno,
> -                              RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
> -                              NULL,
> -                              "cannot allocate Drop queue");
> -     } else if (flow->fate & MLX5_FLOW_FATE_QUEUE) {
> -             struct mlx5_hrxq *hrxq;
> -
> -             hrxq = mlx5_hrxq_get(dev, rss_hash_default_key,
> -                                  MLX5_RSS_HASH_KEY_LEN, 0,
> -                                  &flow->queue, 1, 0, 0);
> -             if (!hrxq)
> -                     hrxq = mlx5_hrxq_new(dev, rss_hash_default_key,
> -                                          MLX5_RSS_HASH_KEY_LEN, 0,
> -                                          &flow->queue, 1, 0, 0);
> -             if (!hrxq)
> -                     return rte_flow_error_set(error, rte_errno,
> -                                     RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
> -                                     NULL,
> -                                     "cannot create flow");
> -             flow->verbs.hrxq = hrxq;
> -     }
> -     flow->verbs.flow =
> -             mlx5_glue->create_flow(flow->verbs.hrxq->qp, flow->verbs.attr);
> -     if (!flow->verbs.flow) {
> -             if (flow->fate & MLX5_FLOW_FATE_DROP)
> -                     mlx5_hrxq_drop_release(dev);
> -             else
> -                     mlx5_hrxq_release(dev, flow->verbs.hrxq);
> -             flow->verbs.hrxq = NULL;
> -             return rte_flow_error_set(error, errno,
> -                                       RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
> -                                       NULL,
> -                                       "kernel module refuses to create"
> -                                       " flow");
> +     struct mlx5_flow_verbs *verbs;
> +     int err;
> +
> +     LIST_FOREACH(verbs, &flow->verbs, next) {
> +             if (flow->fate & MLX5_FLOW_FATE_DROP) {
> +                     verbs->hrxq = mlx5_hrxq_drop_new(dev);
> +                     if (!verbs->hrxq) {
> +                             rte_flow_error_set
> +                                     (error, errno,
> +                                      RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
> +                                      NULL,
> +                                      "cannot get drop hash queue");
> +                             goto error;
> +                     }
> +             } else {
> +                     struct mlx5_hrxq *hrxq;
> +
> +                     hrxq = mlx5_hrxq_get(dev, flow->key,
> +                                          MLX5_RSS_HASH_KEY_LEN,
> +                                          verbs->hash_fields,
> +                                          (*flow->queue),
> +                                          flow->rss.queue_num, 0, 0);
> +                     if (!hrxq)
> +                             hrxq = mlx5_hrxq_new(dev, flow->key,
> +                                                  MLX5_RSS_HASH_KEY_LEN,
> +                                                  verbs->hash_fields,
> +                                                  (*flow->queue),
> +                                                  flow->rss.queue_num, 0, 0);
> +                     if (!hrxq) {
> +                             rte_flow_error_set
> +                                     (error, rte_errno,
> +                                      RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
> +                                      NULL,
> +                                      "cannot get hash queue");
> +                             goto error;
> +                     }
> +                     verbs->hrxq = hrxq;
> +             }
> +             verbs->flow =
> +                     mlx5_glue->create_flow(verbs->hrxq->qp, verbs->attr);
> +             if (!verbs->flow) {
> +                     rte_flow_error_set(error, errno,
> +                                        RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
> +                                        NULL,
> +                                        "hardware refuses to create flow");
> +                     goto error;
> +             }
>       }
>       return 0;
> +error:
> +     err = rte_errno; /* Save rte_errno before cleanup. */
> +     LIST_FOREACH(verbs, &flow->verbs, next) {
> +             if (verbs->hrxq) {
> +                     if (flow->fate & MLX5_FLOW_FATE_DROP)
> +                             mlx5_hrxq_drop_release(dev);
> +                     else
> +                             mlx5_hrxq_release(dev, verbs->hrxq);
> +                     verbs->hrxq = NULL;
> +             }
> +     }
> +     rte_errno = err; /* Restore rte_errno. */
> +     return -rte_errno;
>  }
>  
>  /**
> @@ -1550,42 +1941,43 @@ mlx5_flow_list_create(struct rte_eth_dev *dev,
>                     const struct rte_flow_action actions[],
>                     struct rte_flow_error *error)
>  {
> -     struct rte_flow *flow;
> -     size_t size;
> +     struct rte_flow *flow = NULL;
> +     size_t size = 0;
>       int ret;
>  
> -     ret = mlx5_flow_merge(dev, NULL, 0, attr, items, actions, error);
> +     ret = mlx5_flow_merge(dev, flow, size, attr, items, actions, error);
>       if (ret < 0)
>               return NULL;
>       size = ret;
> -     flow = rte_zmalloc(__func__, size, 0);
> +     flow = rte_calloc(__func__, 1, size, 0);
>       if (!flow) {
>               rte_flow_error_set(error, ENOMEM,
>                                  RTE_FLOW_ERROR_TYPE_UNSPECIFIED,
>                                  NULL,
> -                                "cannot allocate memory");
> +                                "not enough memory to create flow");
>               return NULL;
>       }
> -     flow->verbs.attr = (struct ibv_flow_attr *)(flow + 1);
> -     flow->verbs.specs = (uint8_t *)(flow->verbs.attr + 1);
>       ret = mlx5_flow_merge(dev, flow, size, attr, items, actions, error);
> -     if (ret < 0)
> -             goto error;
> +     if (ret < 0) {
> +             rte_free(flow);
> +             return NULL;
> +     }
>       assert((size_t)ret == size);
>       if (dev->data->dev_started) {
>               ret = mlx5_flow_apply(dev, flow, error);
> -             if (ret < 0)
> -                     goto error;
> +             if (ret < 0) {
> +                     ret = rte_errno; /* Save rte_errno before cleanup. */
> +                     if (flow) {
> +                             mlx5_flow_remove(dev, flow);
> +                             rte_free(flow);
> +                     }
> +                     rte_errno = ret; /* Restore rte_errno. */
> +                     return NULL;
> +             }
>       }
>       mlx5_flow_rxq_mark_set(dev, flow);
>       TAILQ_INSERT_TAIL(list, flow, next);
>       return flow;
> -error:
> -     ret = rte_errno; /* Save rte_errno before cleanup. */
> -     mlx5_flow_remove(dev, flow);
> -     rte_free(flow);
> -     rte_errno = ret; /* Restore rte_errno. */
> -     return NULL;
>  }
>  
>  /**
> @@ -1745,7 +2137,7 @@ mlx5_ctrl_flow_vlan(struct rte_eth_dev *dev,
>       struct priv *priv = dev->data->dev_private;
>       const struct rte_flow_attr attr = {
>               .ingress = 1,
> -             .priority = priv->config.flow_prio - 1,
> +             .priority = MLX5_FLOW_PRIO_RSVD,
>       };
>       struct rte_flow_item items[] = {
>               {
> -- 
> 2.18.0
> 

Reply via email to