Does it need a lock for flow about below scene. Thread1: flow_list_destroy----flow_list_create Thread2: ---------flow_list_destroy---- Maybe the same flow can be operate at the same time.
When i start mlx5 bond and trigger LSC at the same time. It is possible to assert in mlx5_rx_queue_release func and print "port 4 Rx queue 0 is still used by a flow and cannot be removed". I use dpdk-testpmd to simulate the test. Signed-off-by: Weifeng Li <liweifen...@126.com> --- v2: adjust coding style issue. --- drivers/net/mlx5/linux/mlx5_os.c | 1 + drivers/net/mlx5/mlx5.h | 1 + drivers/net/mlx5/mlx5_flow.c | 13 +++++++++++-- 3 files changed, 13 insertions(+), 2 deletions(-) diff --git a/drivers/net/mlx5/linux/mlx5_os.c b/drivers/net/mlx5/linux/mlx5_os.c index c78d56f..59c074e 100644 --- a/drivers/net/mlx5/linux/mlx5_os.c +++ b/drivers/net/mlx5/linux/mlx5_os.c @@ -1426,6 +1426,7 @@ mlx5_dev_spawn(struct rte_device *dpdk_dev, MLX5_MAX_MAC_ADDRESSES); priv->flows = 0; priv->ctrl_flows = 0; + rte_spinlock_init(&priv->flow_lock); rte_spinlock_init(&priv->flow_list_lock); TAILQ_INIT(&priv->flow_meters); TAILQ_INIT(&priv->flow_meter_profiles); diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h index b43a8c9..860bf2f 100644 --- a/drivers/net/mlx5/mlx5.h +++ b/drivers/net/mlx5/mlx5.h @@ -963,6 +963,7 @@ struct mlx5_priv { unsigned int reta_idx_n; /* RETA index size. */ struct mlx5_drop drop_queue; /* Flow drop queues. */ uint32_t flows; /* RTE Flow rules. */ + rte_spinlock_t flow_lock; uint32_t ctrl_flows; /* Control flow rules. */ rte_spinlock_t flow_list_lock; struct mlx5_obj_ops obj_ops; /* HW objects operations. */ diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c index e9c0ddd..69d8159 100644 --- a/drivers/net/mlx5/mlx5_flow.c +++ b/drivers/net/mlx5/mlx5_flow.c @@ -5577,6 +5577,7 @@ flow_list_create(struct rte_eth_dev *dev, uint32_t *list, external, hairpin_flow, error); if (ret < 0) goto error_before_hairpin_split; + rte_spinlock_lock(&priv->flow_lock); flow = mlx5_ipool_zmalloc(priv->sh->ipool[MLX5_IPOOL_RTE_FLOW], &idx); if (!flow) { rte_errno = ENOMEM; @@ -5598,8 +5599,10 @@ flow_list_create(struct rte_eth_dev *dev, uint32_t *list, memset(rss_desc, 0, offsetof(struct mlx5_flow_rss_desc, queue)); rss = flow_get_rss_action(p_actions_rx); if (rss) { - if (flow_rss_workspace_adjust(wks, rss_desc, rss->queue_num)) + if (flow_rss_workspace_adjust(wks, rss_desc, rss->queue_num)) { + rte_spinlock_unlock(&priv->flow_lock); return 0; + } /* * The following information is required by * mlx5_flow_hashfields_adjust() in advance. @@ -5723,6 +5726,7 @@ flow_list_create(struct rte_eth_dev *dev, uint32_t *list, __atomic_add_fetch(&tunnel->refctn, 1, __ATOMIC_RELAXED); mlx5_free(default_miss_ctx.queue); } + rte_spinlock_unlock(&priv->flow_lock); return idx; error: MLX5_ASSERT(flow); @@ -5738,6 +5742,7 @@ flow_list_create(struct rte_eth_dev *dev, uint32_t *list, wks->flow_nested_idx = 0; error_before_hairpin_split: rte_free(translated_actions); + rte_spinlock_unlock(&priv->flow_lock); return 0; } @@ -5877,11 +5882,14 @@ flow_list_destroy(struct rte_eth_dev *dev, uint32_t *list, uint32_t flow_idx) { struct mlx5_priv *priv = dev->data->dev_private; + rte_spinlock_lock(&priv->flow_lock); struct rte_flow *flow = mlx5_ipool_get(priv->sh->ipool [MLX5_IPOOL_RTE_FLOW], flow_idx); - if (!flow) + if (!flow) { + rte_spinlock_unlock(&priv->flow_lock); return; + } /* * Update RX queue flags only if port is started, otherwise it is * already clean. @@ -5908,6 +5916,7 @@ flow_list_destroy(struct rte_eth_dev *dev, uint32_t *list, } flow_mreg_del_copy_action(dev, flow); mlx5_ipool_free(priv->sh->ipool[MLX5_IPOOL_RTE_FLOW], flow_idx); + rte_spinlock_unlock(&priv->flow_lock); } /** -- 2.9.5