The graph id is determined based on a global variable that is incremented every time a graph is created, and decremented every time a graph is destroyed. This only works if graphs are destroyed in the reverse order in which they have been created.
The following code produces duplicate graph IDs which can lead to use-after-free bugs and other undefined behaviours: a = rte_graph_create(...); // id=0 graph_id=1 b = rte_graph_create(...); // id=1 graph_id=2 rte_graph_destroy(a); // graph_id=1 c = rte_graph_create(...); // id=1 graph_id=2 (duplicate with b) rte_graph_destroy(c); // frees memory still used by b Remove the global counter. Make sure that the graph list is always ordered by increasing graph ids. When creating a new graph, pick a free id which is not allocated. Signed-off-by: Robin Jarry <rja...@redhat.com> --- lib/graph/graph.c | 86 +++++++++++++++++++++++++++++++++++++---------- 1 file changed, 69 insertions(+), 17 deletions(-) diff --git a/lib/graph/graph.c b/lib/graph/graph.c index 147bc6c685c5..50616580d06b 100644 --- a/lib/graph/graph.c +++ b/lib/graph/graph.c @@ -19,11 +19,54 @@ static struct graph_head graph_list = STAILQ_HEAD_INITIALIZER(graph_list); static rte_spinlock_t graph_lock = RTE_SPINLOCK_INITIALIZER; -static rte_graph_t graph_id; - -#define GRAPH_ID_CHECK(id) ID_CHECK(id, graph_id) /* Private functions */ +static struct graph * +graph_from_id(rte_graph_t id) +{ + struct graph *graph; + STAILQ_FOREACH(graph, &graph_list, next) { + if (graph->id == id) + return graph; + } + rte_errno = EINVAL; + return NULL; +} + +static rte_graph_t +graph_next_free_id(void) +{ + struct graph *graph; + rte_graph_t id = 0; + + STAILQ_FOREACH(graph, &graph_list, next) { + if (id < graph->id) + break; + id = graph->id + 1; + } + + return id; +} + +static void +graph_insert_ordered(struct graph *graph) +{ + struct graph *after, *g; + + after = NULL; + STAILQ_FOREACH(g, &graph_list, next) { + if (g->id < graph->id) { + after = g; + break; + } + } + if (after == NULL) { + STAILQ_INSERT_TAIL(&graph_list, graph, next); + } else { + STAILQ_INSERT_AFTER(&graph_list, after, graph, next); + } +} + struct graph_head * graph_list_head_get(void) { @@ -279,7 +322,8 @@ rte_graph_model_mcore_dispatch_core_bind(rte_graph_t id, int lcore) { struct graph *graph; - GRAPH_ID_CHECK(id); + if (graph_from_id(id) == NULL) + goto fail; if (!rte_lcore_is_enabled(lcore)) SET_ERR_JMP(ENOLINK, fail, "lcore %d not enabled", lcore); @@ -309,7 +353,8 @@ rte_graph_model_mcore_dispatch_core_unbind(rte_graph_t id) { struct graph *graph; - GRAPH_ID_CHECK(id); + if (graph_from_id(id) == NULL) + goto fail; STAILQ_FOREACH(graph, &graph_list, next) if (graph->id == id) break; @@ -406,7 +451,7 @@ rte_graph_create(const char *name, struct rte_graph_param *prm) graph->socket = prm->socket_id; graph->src_node_count = src_node_count; graph->node_count = graph_nodes_count(graph); - graph->id = graph_id; + graph->id = graph_next_free_id(); graph->parent_id = RTE_GRAPH_ID_INVALID; graph->lcore_id = RTE_MAX_LCORE; graph->num_pkt_to_capture = prm->num_pkt_to_capture; @@ -422,8 +467,7 @@ rte_graph_create(const char *name, struct rte_graph_param *prm) goto graph_mem_destroy; /* All good, Lets add the graph to the list */ - graph_id++; - STAILQ_INSERT_TAIL(&graph_list, graph, next); + graph_insert_ordered(graph); graph_spinlock_unlock(); return graph->id; @@ -467,7 +511,6 @@ rte_graph_destroy(rte_graph_t id) graph_cleanup(graph); STAILQ_REMOVE(&graph_list, graph, graph, next); free(graph); - graph_id--; goto done; } graph = tmp; @@ -520,7 +563,7 @@ graph_clone(struct graph *parent_graph, const char *name, struct rte_graph_param graph->parent_id = parent_graph->id; graph->lcore_id = parent_graph->lcore_id; graph->socket = parent_graph->socket; - graph->id = graph_id; + graph->id = graph_next_free_id(); /* Allocate the Graph fast path memory and populate the data */ if (graph_fp_mem_create(graph)) @@ -539,8 +582,7 @@ graph_clone(struct graph *parent_graph, const char *name, struct rte_graph_param goto graph_mem_destroy; /* All good, Lets add the graph to the list */ - graph_id++; - STAILQ_INSERT_TAIL(&graph_list, graph, next); + graph_insert_ordered(graph); graph_spinlock_unlock(); return graph->id; @@ -561,7 +603,8 @@ rte_graph_clone(rte_graph_t id, const char *name, struct rte_graph_param *prm) { struct graph *graph; - GRAPH_ID_CHECK(id); + if (graph_from_id(id) == NULL) + goto fail; STAILQ_FOREACH(graph, &graph_list, next) if (graph->id == id) return graph_clone(graph, name, prm); @@ -587,7 +630,8 @@ rte_graph_id_to_name(rte_graph_t id) { struct graph *graph; - GRAPH_ID_CHECK(id); + if (graph_from_id(id) == NULL) + goto fail; STAILQ_FOREACH(graph, &graph_list, next) if (graph->id == id) return graph->name; @@ -604,7 +648,8 @@ rte_graph_node_get(rte_graph_t gid, uint32_t nid) rte_graph_off_t off; rte_node_t count; - GRAPH_ID_CHECK(gid); + if (graph_from_id(gid) == NULL) + goto fail; STAILQ_FOREACH(graph, &graph_list, next) if (graph->id == gid) { rte_graph_foreach_node(count, off, graph->graph, @@ -747,7 +792,8 @@ graph_scan_dump(FILE *f, rte_graph_t id, bool all) struct graph *graph; RTE_VERIFY(f); - GRAPH_ID_CHECK(id); + if (graph_from_id(id) == NULL) + goto fail; STAILQ_FOREACH(graph, &graph_list, next) { if (all == true) { @@ -776,7 +822,13 @@ rte_graph_list_dump(FILE *f) rte_graph_t rte_graph_max_count(void) { - return graph_id; + struct graph *graph; + rte_graph_t count = 0; + + STAILQ_FOREACH(graph, &graph_list, next) + count++; + + return count; } RTE_LOG_REGISTER_DEFAULT(rte_graph_logtype, INFO); -- 2.44.0