From: Maciej Gajdzica <maciejx.t.gajdz...@intel.com> Added statistics for hash ext table.
Signed-off-by: Maciej Gajdzica <maciejx.t.gajdzica at intel.com> --- lib/librte_table/rte_table_hash_ext.c | 44 +++++++++++++++++++++++++++++++++ 1 file changed, 44 insertions(+) diff --git a/lib/librte_table/rte_table_hash_ext.c b/lib/librte_table/rte_table_hash_ext.c index 66e416b..3c12273 100644 --- a/lib/librte_table/rte_table_hash_ext.c +++ b/lib/librte_table/rte_table_hash_ext.c @@ -74,6 +74,20 @@ do \ (bucket)->next = (bucket2)->next; \ while (0) +#ifdef RTE_TABLE_STATS_COLLECT + +#define RTE_TABLE_HASH_EXT_STATS_PKTS_IN_ADD(table, val) \ + table->stats.n_pkts_in += val +#define RTE_TABLE_HASH_EXT_STATS_PKTS_LOOKUP_MISS(table, val) \ + table->stats.n_pkts_lookup_miss += val + +#else + +#define RTE_TABLE_HASH_EXT_STATS_PKTS_IN_ADD(table, val) +#define RTE_TABLE_HASH_EXT_STATS_PKTS_LOOKUP_MISS(table, val) + +#endif + struct grinder { struct bucket *bkt; uint64_t sig; @@ -82,6 +96,8 @@ struct grinder { }; struct rte_table_hash { + struct rte_table_stats stats; + /* Input parameters */ uint32_t key_size; uint32_t entry_size; @@ -440,6 +456,9 @@ static int rte_table_hash_ext_lookup_unoptimized( struct rte_table_hash *t = (struct rte_table_hash *) table; uint64_t pkts_mask_out = 0; + __rte_unused uint32_t n_pkts_in = __builtin_popcountll(pkts_mask); + RTE_TABLE_HASH_EXT_STATS_PKTS_IN_ADD(t, n_pkts_in); + for ( ; pkts_mask; ) { struct bucket *bkt0, *bkt; struct rte_mbuf *pkt; @@ -484,6 +503,7 @@ static int rte_table_hash_ext_lookup_unoptimized( } *lookup_hit_mask = pkts_mask_out; + RTE_TABLE_HASH_EXT_STATS_PKTS_LOOKUP_MISS(t, n_pkts_in - __builtin_popcountll(pkts_mask_out)); return 0; } @@ -861,6 +881,9 @@ static int rte_table_hash_ext_lookup( uint64_t pkts_mask_out = 0, pkts_mask_match_many = 0; int status = 0; + __rte_unused uint32_t n_pkts_in = __builtin_popcountll(pkts_mask); + RTE_TABLE_HASH_EXT_STATS_PKTS_IN_ADD(t, n_pkts_in); + /* Cannot run the pipeline with less than 7 packets */ if (__builtin_popcountll(pkts_mask) < 7) return rte_table_hash_ext_lookup_unoptimized(table, pkts, @@ -973,6 +996,7 @@ static int rte_table_hash_ext_lookup( } *lookup_hit_mask = pkts_mask_out; + RTE_TABLE_HASH_EXT_STATS_PKTS_LOOKUP_MISS(t, n_pkts_in - __builtin_popcountll(pkts_mask_out)); return status; } @@ -990,6 +1014,9 @@ static int rte_table_hash_ext_lookup_dosig( uint64_t pkts_mask_out = 0, pkts_mask_match_many = 0; int status = 0; + __rte_unused uint32_t n_pkts_in = __builtin_popcountll(pkts_mask); + RTE_TABLE_HASH_EXT_STATS_PKTS_IN_ADD(t, n_pkts_in); + /* Cannot run the pipeline with less than 7 packets */ if (__builtin_popcountll(pkts_mask) < 7) return rte_table_hash_ext_lookup_unoptimized(table, pkts, @@ -1102,15 +1129,31 @@ static int rte_table_hash_ext_lookup_dosig( } *lookup_hit_mask = pkts_mask_out; + RTE_TABLE_HASH_EXT_STATS_PKTS_LOOKUP_MISS(t, n_pkts_in - __builtin_popcountll(pkts_mask_out)); return status; } +static int +rte_table_hash_ext_stats_read(void *table, struct rte_table_stats *stats, int clear) +{ + struct rte_table_hash *t = (struct rte_table_hash *) table; + + if (stats != NULL) + memcpy(stats, &t->stats, sizeof(t->stats)); + + if (clear) + memset(&t->stats, 0, sizeof(t->stats)); + + return 0; +} + struct rte_table_ops rte_table_hash_ext_ops = { .f_create = rte_table_hash_ext_create, .f_free = rte_table_hash_ext_free, .f_add = rte_table_hash_ext_entry_add, .f_delete = rte_table_hash_ext_entry_delete, .f_lookup = rte_table_hash_ext_lookup, + .f_stats = rte_table_hash_ext_stats_read, }; struct rte_table_ops rte_table_hash_ext_dosig_ops = { @@ -1119,4 +1162,5 @@ struct rte_table_ops rte_table_hash_ext_dosig_ops = { .f_add = rte_table_hash_ext_entry_add, .f_delete = rte_table_hash_ext_entry_delete, .f_lookup = rte_table_hash_ext_lookup_dosig, + .f_stats = rte_table_hash_ext_stats_read, }; -- 1.7.9.5