[PATCH v1] examples/ipsec-secgw: free the actual mbuf pointer

2024-08-16 Thread Vidya Sagar Velumuri
In case of crypto event vector, the vector points to crypto op structure
in priv area and not actual mbuf.
Extract the mbuf pointer and pass these to rte_mbuf_free to free the
mbufs.

Signed-off-by: Vidya Sagar Velumuri 

diff --git a/examples/ipsec-secgw/ipsec_worker.c 
b/examples/ipsec-secgw/ipsec_worker.c
index c9c43ebd2b..dd14226e81 100644
--- a/examples/ipsec-secgw/ipsec_worker.c
+++ b/examples/ipsec-secgw/ipsec_worker.c
@@ -960,7 +960,18 @@ static void
 ipsec_event_vector_free(struct rte_event *ev)
 {
struct rte_event_vector *vec = ev->vec;
-   rte_pktmbuf_free_bulk(vec->mbufs + vec->elem_offset, vec->nb_elem);
+
+   if (ev->event_type == RTE_EVENT_TYPE_CRYPTODEV_VECTOR) {
+   struct rte_crypto_op *cop;
+   int i;
+
+   for (i = 0; i < vec->nb_elem; i++) {
+   cop = vec->ptrs[i];
+   rte_pktmbuf_free(cop->sym->m_src);
+   }
+   } else {
+   rte_pktmbuf_free_bulk(vec->mbufs + vec->elem_offset, 
vec->nb_elem);
+   }
rte_mempool_put(rte_mempool_from_obj(vec), vec);
 }
 
-- 
2.25.1



Re: [dpdk-dev] [PATCH v13 1/4] devtools: script to track symbols over releases

2024-08-16 Thread Ray Kinsella


Stephen Hemminger  writes:

> On Thu,  9 Sep 2021 14:48:05 +0100
> Ray Kinsella  wrote:
>
>> This script tracks the growth of stable and experimental symbols
>> over releases since v19.11. The script has the ability to
>> count the added symbols between two dpdk releases, and to
>> list experimental symbols present in two dpdk releases
>> (expired symbols).
>> 
>> example usages:
>> 
>> Count symbols added since v19.11
>> $ devtools/symbol-tool.py count-symbols
>> 
>> Count symbols added since v20.11
>> $ devtools/symbol-tool.py count-symbols --releases v20.11,v21.05
>> 
>> List experimental symbols present in v20.11 and v21.05
>> $ devtools/symbol-tool.py list-expired --releases v20.11,v21.05
>> 
>> List experimental symbols in libraries only, present since v19.11
>> $ devtools/symbol-tool.py list-expired --directory lib
>> 
>> Signed-off-by: Ray Kinsella 
>
> Tried it, looks useful. Minor notes:

I think it was mostly lack of interest that this didn't get over the
line when I submitted it orginally.

The flake8 errors as easily fixed, I can do that if you thought it might
be useful?

>   - requires python parsley library which was not required before
>   - minor style format
>
> $ flake8 ./devtools/symbol-tool.py 
> ./devtools/symbol-tool.py:369:11: E275 missing whitespace after keyword
>
> Acked-by: Stephen Hemminger 


-- 
Regards, Ray K


Re: [PATCH] app/testpmd: add L4 port to verbose output

2024-08-16 Thread Luca Vizzarro

On 15/08/2024 16:22, Stephen Hemminger wrote:

The verbose output is already too verbose.
Maybe you would like the simpler format (which does include the port number)
see the network packet dissector patches.


Hi Stephen,

Thank you for the reply you left to Alex's patch. This is actually quite 
helpful.


I've had a look at your patch series and the hexdump is actually quite 
great for our purpose at DTS, as it will allow us to detect a preset 
packet identifier in the payload.


Unfortunately, your patch doesn't solve the problem that Alex is having. 
He's currently writing up a test suite to test the RSS hash keys and the 
respective queues to which the incoming packets are placed into. To 
correctly match the packets to the queue and hash key, he needs a way to 
identify them. We explored some solutions and the L4 port seemed to be 
the most problem-free with the least effort. Using the payload is 
probably best and you've already posted the required changes too. The 
only problem is that your hexdump output does not print the required RSS 
values.


Would you be be keen to add these to your patch? Otherwise, do you have 
any ideas and/or solutions to tackle this specific problem?


Thank you,
Luca Vizzarro


[RFC] table: report victim replace stats in LRU tables

2024-08-16 Thread Kamil Vojanec
LRU caches replace records when requested table bucket is full. There
is, however, no information about this happenning in either the return
value or the statistics. This commit introduces a counter for such
cases.

Signed-off-by: Kamil Vojanec 
---
 lib/table/rte_table.h| 1 +
 lib/table/rte_table_hash_key16.c | 4 
 lib/table/rte_table_hash_key32.c | 4 
 lib/table/rte_table_hash_key8.c  | 4 
 lib/table/rte_table_hash_lru.c   | 4 
 5 files changed, 17 insertions(+)

diff --git a/lib/table/rte_table.h b/lib/table/rte_table.h
index 9a5faf0e32..e097e25868 100644
--- a/lib/table/rte_table.h
+++ b/lib/table/rte_table.h
@@ -33,6 +33,7 @@ struct rte_mbuf;
 struct rte_table_stats {
uint64_t n_pkts_in;
uint64_t n_pkts_lookup_miss;
+   uint64_t n_pkts_insert_victims;
 };
 
 /**
diff --git a/lib/table/rte_table_hash_key16.c b/lib/table/rte_table_hash_key16.c
index 67b77c16a0..f97d9c1266 100644
--- a/lib/table/rte_table_hash_key16.c
+++ b/lib/table/rte_table_hash_key16.c
@@ -27,11 +27,14 @@
table->stats.n_pkts_in += val
 #define RTE_TABLE_HASH_KEY16_STATS_PKTS_LOOKUP_MISS(table, val) \
table->stats.n_pkts_lookup_miss += val
+#define RTE_TABLE_HASH_KEY16_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) \
+   table->stats.n_pkts_insert_victims += val
 
 #else
 
 #define RTE_TABLE_HASH_KEY16_STATS_PKTS_IN_ADD(table, val)
 #define RTE_TABLE_HASH_KEY16_STATS_PKTS_LOOKUP_MISS(table, val)
+#define RTE_TABLE_HASH_KEY16_STATS_PKTS_INSERT_VICTIMS_ADD(table, val)
 
 #endif
 
@@ -304,6 +307,7 @@ rte_table_hash_entry_add_key16_lru(
}
 
/* Bucket full: replace LRU entry */
+   RTE_TABLE_HASH_KEY16_STATS_PKTS_INSERT_VICTIMS_ADD(f, 1);
pos = lru_pos(bucket);
bucket->signature[pos] = signature;
keycpy(&bucket->key[pos], key, f->key_mask);
diff --git a/lib/table/rte_table_hash_key32.c b/lib/table/rte_table_hash_key32.c
index 1aa86c6a49..37ad87bb85 100644
--- a/lib/table/rte_table_hash_key32.c
+++ b/lib/table/rte_table_hash_key32.c
@@ -27,11 +27,14 @@
table->stats.n_pkts_in += val
 #define RTE_TABLE_HASH_KEY32_STATS_PKTS_LOOKUP_MISS(table, val) \
table->stats.n_pkts_lookup_miss += val
+#define RTE_TABLE_HASH_KEY32_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) \
+   table->stats.n_pkts_insert_victims += val
 
 #else
 
 #define RTE_TABLE_HASH_KEY32_STATS_PKTS_IN_ADD(table, val)
 #define RTE_TABLE_HASH_KEY32_STATS_PKTS_LOOKUP_MISS(table, val)
+#define RTE_TABLE_HASH_KEY32_STATS_PKTS_INSERT_VICTIMS_ADD(table, val)
 
 #endif
 
@@ -312,6 +315,7 @@ rte_table_hash_entry_add_key32_lru(
}
 
/* Bucket full: replace LRU entry */
+   RTE_TABLE_HASH_KEY32_STATS_PKTS_INSERT_VICTIMS_ADD(f, 1);
pos = lru_pos(bucket);
bucket->signature[pos] = signature;
keycpy(&bucket->key[pos], key, f->key_mask);
diff --git a/lib/table/rte_table_hash_key8.c b/lib/table/rte_table_hash_key8.c
index c8d72b333d..c5722f3709 100644
--- a/lib/table/rte_table_hash_key8.c
+++ b/lib/table/rte_table_hash_key8.c
@@ -25,11 +25,14 @@
table->stats.n_pkts_in += val
 #define RTE_TABLE_HASH_KEY8_STATS_PKTS_LOOKUP_MISS(table, val) \
table->stats.n_pkts_lookup_miss += val
+#define RTE_TABLE_HASH_KEY8_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) \
+   table->stats.n_pkts_insert_victims += val
 
 #else
 
 #define RTE_TABLE_HASH_KEY8_STATS_PKTS_IN_ADD(table, val)
 #define RTE_TABLE_HASH_KEY8_STATS_PKTS_LOOKUP_MISS(table, val)
+#define RTE_TABLE_HASH_KEY8_STATS_PKTS_INSERT_VICTIMS_ADD(table, val)
 
 #endif
 
@@ -292,6 +295,7 @@ rte_table_hash_entry_add_key8_lru(
}
 
/* Bucket full: replace LRU entry */
+   RTE_TABLE_HASH_KEY8_STATS_PKTS_INSERT_VICTIMS_ADD(f, 1);
pos = lru_pos(bucket);
keycpy(&bucket->key[pos], key, &f->key_mask);
memcpy(&bucket->data[pos * f->entry_size], entry, f->entry_size);
diff --git a/lib/table/rte_table_hash_lru.c b/lib/table/rte_table_hash_lru.c
index 801e48f5ba..bcf9c34f8e 100644
--- a/lib/table/rte_table_hash_lru.c
+++ b/lib/table/rte_table_hash_lru.c
@@ -23,11 +23,14 @@
table->stats.n_pkts_in += val
 #define RTE_TABLE_HASH_LRU_STATS_PKTS_LOOKUP_MISS(table, val) \
table->stats.n_pkts_lookup_miss += val
+#define RTE_TABLE_HASH_LRU_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) \
+   table->stats.n_pkts_insert_victims += val
 
 #else
 
 #define RTE_TABLE_HASH_LRU_STATS_PKTS_IN_ADD(table, val)
 #define RTE_TABLE_HASH_LRU_STATS_PKTS_LOOKUP_MISS(table, val)
+#define RTE_TABLE_HASH_LRU_STATS_PKTS_INSERT_VICTIMS_ADD(table, val)
 
 #endif
 
@@ -340,6 +343,7 @@ rte_table_hash_lru_entry_add(void *table, void *key, void 
*entry,
 
/* Bucket full */
{
+   RTE_TABLE_HASH_LRU_STATS_PKTS_INSERT_VICTIMS_ADD(t, 1);
uint64_t pos = lru_pos(bkt);
uint32_t bkt_key_index = bkt->key_pos[pos];
uint8_t *bkt_key = &t->key_mem[bkt_key_index <<
-- 
2.43.5



Re: Ethdev tracepoints optimization

2024-08-16 Thread Jerin Jacob
On Fri, Aug 16, 2024 at 4:43 AM Adel Belkhiri  wrote:
>
> Hi DPDK Community,
>
> I am currently working on developing performance analyses for applications 
> using the ethdev library. These analyses are being implemented in Trace 
> Compass, an open-source performance analyzer. One of the views I’ve 
> implemented shows the rate of traffic received or sent by an ethernet port, 
> measured in packets per second. However, I've encountered an issue with the 
> lib.ethdev.rx.burst event, which triggers even when no packets are polled, 
> leading to a significant number of irrelevant events in the trace. This 
> becomes problematic as these "empty" events can overwhelm the tracer buffer, 
> potentially causing the loss of more critical events due to their high 
> frequency.
>
> To address this, I've modified the DPDK code in lib/ethdev/rte_ethdev.h to 
> add a conditional statement that only triggers the event when nb_rx > 0. My 
> question to the community is whether there are use cases where an "empty" 
> lib.ethdev.rx.burst event could be useful. If not, would there be interest in 
> submitting a patch with this modification?
>
> Moreover, I am looking to develop an analysis that calculates the throughput 
> (in kb/s, mb/s, etc.) per NIC, utilizing the same events (i.e., 
> lib.ethdev.rx.burst and lib.ethdev.tx.burst). These tracepoints do not 
> provide packet size directly, only a pointer to the packet array. My attempt 
> to use an eBPF program to iterate through that array to access the packet 
> sizes was unsuccessful, as I found no method to export the computed data 
> (e.g., via a custom tracepoint). Does anyone have suggestions or alternative 
> approaches for achieving a throughput measurement?


Call rte_eth_stats_get() in N second interval and find throughput in slow path.



> I would be grateful for any insights or suggestions you might have.
>
> Thank you!
> Adel


[PATCH v2 1/4] usertools/cpu_layout: update coding style

2024-08-16 Thread Anatoly Burakov
Update coding style:

- make it PEP-484 compliant
- address all flake8, mypy etc. warnings
- use f-strings in place of old-style string interpolation
- refactor printing to make the code more readable

Signed-off-by: Anatoly Burakov 
---
 usertools/cpu_layout.py | 162 ++--
 1 file changed, 104 insertions(+), 58 deletions(-)

diff --git a/usertools/cpu_layout.py b/usertools/cpu_layout.py
index 891b9238fa..be86f06938 100755
--- a/usertools/cpu_layout.py
+++ b/usertools/cpu_layout.py
@@ -3,62 +3,108 @@
 # Copyright(c) 2010-2014 Intel Corporation
 # Copyright(c) 2017 Cavium, Inc. All rights reserved.
 
-sockets = []
-cores = []
-core_map = {}
-base_path = "/sys/devices/system/cpu"
-fd = open("{}/kernel_max".format(base_path))
-max_cpus = int(fd.read())
-fd.close()
-for cpu in range(max_cpus + 1):
-try:
-fd = open("{}/cpu{}/topology/core_id".format(base_path, cpu))
-except IOError:
-continue
-core = int(fd.read())
-fd.close()
-fd = open("{}/cpu{}/topology/physical_package_id".format(base_path, cpu))
-socket = int(fd.read())
-fd.close()
-if core not in cores:
-cores.append(core)
-if socket not in sockets:
-sockets.append(socket)
-key = (socket, core)
-if key not in core_map:
-core_map[key] = []
-core_map[key].append(cpu)
-
-print(format("=" * (47 + len(base_path
-print("Core and Socket Information (as reported by '{}')".format(base_path))
-print("{}\n".format("=" * (47 + len(base_path
-print("cores = ", cores)
-print("sockets = ", sockets)
-print("")
-
-max_processor_len = len(str(len(cores) * len(sockets) * 2 - 1))
-max_thread_count = len(list(core_map.values())[0])
-max_core_map_len = (max_processor_len * max_thread_count)  \
-  + len(", ") * (max_thread_count - 1) \
-  + len('[]') + len('Socket ')
-max_core_id_len = len(str(max(cores)))
-
-output = " ".ljust(max_core_id_len + len('Core '))
-for s in sockets:
-output += " Socket %s" % str(s).ljust(max_core_map_len - len('Socket '))
-print(output)
-
-output = " ".ljust(max_core_id_len + len('Core '))
-for s in sockets:
-output += " ".ljust(max_core_map_len)
-output += " "
-print(output)
-
-for c in cores:
-output = "Core %s" % str(c).ljust(max_core_id_len)
-for s in sockets:
-if (s, c) in core_map:
-output += " " + str(core_map[(s, c)]).ljust(max_core_map_len)
+from typing import List, Set, Dict, Tuple
+
+
+def _range_expand(rstr: str) -> List[int]:
+"""Expand a range string into a list of integers."""
+# 0,1-3 => [0, 1-3]
+ranges = rstr.split(",")
+valset: List[int] = []
+for r in ranges:
+# 1-3 => [1, 2, 3]
+if "-" in r:
+start, end = r.split("-")
+valset.extend(range(int(start), int(end) + 1))
 else:
-output += " " * (max_core_map_len + 1)
-print(output)
+valset.append(int(r))
+return valset
+
+
+def _read_sysfs(path: str) -> str:
+with open(path, encoding="utf-8") as fd:
+return fd.read().strip()
+
+
+def _print_row(row: Tuple[str, ...], col_widths: List[int]) -> None:
+first, *rest = row
+w_first, *w_rest = col_widths
+first_end = " " * 4
+rest_end = " " * 10
+
+print(first.ljust(w_first), end=first_end)
+for cell, width in zip(rest, w_rest):
+print(cell.rjust(width), end=rest_end)
+print()
+
+
+def _print_section(heading: str) -> None:
+sep = "=" * len(heading)
+print(sep)
+print(heading)
+print(sep)
+print()
+
+
+def _main() -> None:
+sockets_s: Set[int] = set()
+cores_s: Set[int] = set()
+core_map: Dict[Tuple[int, int], List[int]] = {}
+base_path = "/sys/devices/system/cpu"
+
+cpus = _range_expand(_read_sysfs(f"{base_path}/online"))
+
+for cpu in cpus:
+lcore_base = f"{base_path}/cpu{cpu}"
+core = int(_read_sysfs(f"{lcore_base}/topology/core_id"))
+socket = int(_read_sysfs(f"{lcore_base}/topology/physical_package_id"))
+
+cores_s.add(core)
+sockets_s.add(socket)
+key = (socket, core)
+core_map.setdefault(key, [])
+core_map[key].append(cpu)
+
+cores = sorted(cores_s)
+sockets = sorted(sockets_s)
+
+_print_section("Core and Socket Information "
+   f"(as reported by '{base_path}')")
+
+print("cores = ", cores)
+print("sockets = ", sockets)
+print()
+
+# Core, [Socket, Socket, ...]
+heading_strs = "", *[f"Socket {s}" for s in sockets]
+sep_strs = tuple("-" * len(hstr) for hstr in heading_strs)
+rows: List[Tuple[str, ...]] = []
+
+for c in cores:
+# Core,
+row: Tuple[str, ...] = (f"Core {c}",)
+
+# [lcores, lcores, ...]
+for s in sockets:
+try:
+lcores = core_map[(s, c)]
+row += (str(lcores),)
+except KeyError:
+row += ("",)
+   

[PATCH v2 2/4] usertools/cpu_layout: print out NUMA nodes

2024-08-16 Thread Anatoly Burakov
In traditional NUMA case, NUMA nodes and physical sockets were used
interchangeably, but there are cases where there can be multiple NUMA
nodes per socket, as well as all CPU's being assigned NUMA node 0 even in
cases of multiple sockets. Use sysfs to print out NUMA information.

Signed-off-by: Anatoly Burakov 
---
 usertools/cpu_layout.py | 35 ++-
 1 file changed, 30 insertions(+), 5 deletions(-)

diff --git a/usertools/cpu_layout.py b/usertools/cpu_layout.py
index be86f06938..e43bdbf343 100755
--- a/usertools/cpu_layout.py
+++ b/usertools/cpu_layout.py
@@ -4,6 +4,7 @@
 # Copyright(c) 2017 Cavium, Inc. All rights reserved.
 
 from typing import List, Set, Dict, Tuple
+import glob
 
 
 def _range_expand(rstr: str) -> List[int]:
@@ -26,11 +27,19 @@ def _read_sysfs(path: str) -> str:
 return fd.read().strip()
 
 
+def _read_numa_node(base: str) -> int:
+node_glob = f"{base}/node*"
+node_dirs = glob.glob(node_glob)
+if not node_dirs:
+return 0  # default to node 0
+return int(node_dirs[0].split("node")[1])
+
+
 def _print_row(row: Tuple[str, ...], col_widths: List[int]) -> None:
 first, *rest = row
 w_first, *w_rest = col_widths
 first_end = " " * 4
-rest_end = " " * 10
+rest_end = " " * 4
 
 print(first.ljust(w_first), end=first_end)
 for cell, width in zip(rest, w_rest):
@@ -50,6 +59,7 @@ def _main() -> None:
 sockets_s: Set[int] = set()
 cores_s: Set[int] = set()
 core_map: Dict[Tuple[int, int], List[int]] = {}
+numa_map: Dict[int, int] = {}
 base_path = "/sys/devices/system/cpu"
 
 cpus = _range_expand(_read_sysfs(f"{base_path}/online"))
@@ -58,12 +68,14 @@ def _main() -> None:
 lcore_base = f"{base_path}/cpu{cpu}"
 core = int(_read_sysfs(f"{lcore_base}/topology/core_id"))
 socket = int(_read_sysfs(f"{lcore_base}/topology/physical_package_id"))
+node = _read_numa_node(lcore_base)
 
 cores_s.add(core)
 sockets_s.add(socket)
 key = (socket, core)
 core_map.setdefault(key, [])
 core_map[key].append(cpu)
+numa_map[cpu] = node
 
 cores = sorted(cores_s)
 sockets = sorted(sockets_s)
@@ -73,24 +85,37 @@ def _main() -> None:
 
 print("cores = ", cores)
 print("sockets = ", sockets)
+print("numa = ", sorted(set(numa_map.values(
 print()
 
-# Core, [Socket, Socket, ...]
-heading_strs = "", *[f"Socket {s}" for s in sockets]
+# Core, [NUMA, Socket, NUMA, Socket, ...]
+heading_strs = "", *[v for s in sockets for v in ("", f"Socket {s}")]
 sep_strs = tuple("-" * len(hstr) for hstr in heading_strs)
 rows: List[Tuple[str, ...]] = []
 
+prev_numa = None
 for c in cores:
 # Core,
 row: Tuple[str, ...] = (f"Core {c}",)
 
-# [lcores, lcores, ...]
+# assume NUMA changes symmetrically
+first_lcore = core_map[(0, c)][0]
+cur_numa = numa_map[first_lcore]
+numa_changed = prev_numa != cur_numa
+prev_numa = cur_numa
+
+# [NUMA, lcores, NUMA, lcores, ...]
 for s in sockets:
 try:
 lcores = core_map[(s, c)]
+numa = numa_map[lcores[0]]
+if numa_changed:
+row += (f"NUMA {numa}",)
+else:
+row += ("",)
 row += (str(lcores),)
 except KeyError:
-row += ("",)
+row += ("", "")
 rows += [row]
 
 # find max widths for each column, including header and rows
-- 
2.43.5



[PATCH v2 3/4] usertools/dpdk-hugepages.py: sort by NUMA node

2024-08-16 Thread Anatoly Burakov
Currently, the list of per-NUMA node hugepages is displayed in glob order,
which can be arbitrary. Fix it to sort the glob order.

Signed-off-by: Anatoly Burakov 
---
 usertools/dpdk-hugepages.py | 40 ++---
 1 file changed, 28 insertions(+), 12 deletions(-)

diff --git a/usertools/dpdk-hugepages.py b/usertools/dpdk-hugepages.py
index bf2575ba36..54232ddf22 100755
--- a/usertools/dpdk-hugepages.py
+++ b/usertools/dpdk-hugepages.py
@@ -74,21 +74,37 @@ def set_hugepages(path, reqpages):
  gotpages, reqpages, filename))
 
 
+def get_numa_pages_node(node):
+'''Read list of hugepage reservations on specific NUMA node'''
+hp_path = f'/sys/devices/system/node/node{node}/hugepages'
+if not os.path.exists(hp_path):
+return
+res = []
+for pg_sz_dir in os.listdir(hp_path):
+pg_sz = int(pg_sz_dir[10:-2])
+nr_pages = get_hugepages(f'{hp_path}/{pg_sz_dir}')
+if nr_pages > 0:
+pg_sz_str = fmt_memsize(pg_sz)
+total_sz_str = fmt_memsize(nr_pages * pg_sz)
+res += [(nr_pages, pg_sz_str, total_sz_str)]
+else:
+res += [(0, None, None)]
+return res
+
+
 def show_numa_pages():
 '''Show huge page reservations on Numa system'''
+# get list of NUMA nodes and sort them by integer order
 print('Node Pages Size Total')
-for numa_path in glob.glob('/sys/devices/system/node/node*'):
-node = numa_path[29:]  # slice after /sys/devices/system/node/node
-path = numa_path + '/hugepages'
-if not os.path.exists(path):
-continue
-for hdir in os.listdir(path):
-pages = get_hugepages(path + '/' + hdir)
-if pages > 0:
-kb = int(hdir[10:-2])  # slice out of hugepages-NNNkB
-print('{:<4} {:<5} {:<6} {}'.format(node, pages,
-fmt_memsize(kb),
-fmt_memsize(pages * kb)))
+nodes = sorted(int(node[29:])
+   for node in glob.glob('/sys/devices/system/node/node*'))
+for node in nodes:
+pg_sz_data = get_numa_pages_node(node)
+for nr_pages, pg_sz, total_sz in pg_sz_data:
+if not nr_pages:
+continue
+print('{:<4} {:<5} {:<6} {}'
+  .format(node, nr_pages, pg_sz, total_sz))
 
 
 def show_non_numa_pages():
-- 
2.43.5



[PATCH v2 4/4] usertools/dpdk-devbind: print NUMA node

2024-08-16 Thread Anatoly Burakov
Currently, devbind does not print out any NUMA information, which makes
figuring out which NUMA node device belongs to not trivial. Add printouts
for NUMA information if NUMA support is enabled on the system.

Signed-off-by: Anatoly Burakov 
---
 usertools/dpdk-devbind.py | 27 +++
 1 file changed, 19 insertions(+), 8 deletions(-)

diff --git a/usertools/dpdk-devbind.py b/usertools/dpdk-devbind.py
index b276e8efc8..c0611a501d 100755
--- a/usertools/dpdk-devbind.py
+++ b/usertools/dpdk-devbind.py
@@ -110,6 +110,11 @@
 args = []
 
 
+# check if this system has NUMA support
+def is_numa():
+return os.path.exists('/sys/devices/system/node')
+
+
 # check if a specific kernel module is loaded
 def module_is_loaded(module):
 global loaded_modules
@@ -579,18 +584,24 @@ def show_device_status(devices_type, device_name, 
if_field=False):
 
 # print each category separately, so we can clearly see what's used by DPDK
 if dpdk_drv:
+extra_param = "drv=%(Driver_str)s unused=%(Module_str)s"
+if is_numa():
+extra_param = "numa_node=%(NUMANode)s " + extra_param
 display_devices("%s devices using DPDK-compatible driver" % 
device_name,
-dpdk_drv, "drv=%(Driver_str)s unused=%(Module_str)s")
+dpdk_drv, extra_param)
 if kernel_drv:
-if_text = ""
+extra_param = "drv=%(Driver_str)s unused=%(Module_str)s"
 if if_field:
-if_text = "if=%(Interface)s "
-display_devices("%s devices using kernel driver" % device_name, 
kernel_drv,
-if_text + "drv=%(Driver_str)s "
-"unused=%(Module_str)s %(Active)s")
+extra_param = "if=%(Interface)s " + extra_param
+if is_numa():
+extra_param = "numa_node=%(NUMANode)s " + extra_param
+display_devices("%s devices using kernel driver" % device_name,
+kernel_drv, extra_param)
 if no_drv:
-display_devices("Other %s devices" % device_name, no_drv,
-"unused=%(Module_str)s")
+extra_param = "unused=%(Module_str)s"
+if is_numa():
+extra_param = "numa_node=%(NUMANode)s " + extra_param
+display_devices("Other %s devices" % device_name, no_drv, extra_param)
 
 
 def show_status():
-- 
2.43.5



Re: [PATCH v2 3/4] usertools/dpdk-hugepages.py: sort by NUMA node

2024-08-16 Thread Burakov, Anatoly

On 8/16/2024 2:16 PM, Anatoly Burakov wrote:

Currently, the list of per-NUMA node hugepages is displayed in glob order,
which can be arbitrary. Fix it to sort the glob order.

Signed-off-by: Anatoly Burakov 
---
  usertools/dpdk-hugepages.py | 40 ++---
  1 file changed, 28 insertions(+), 12 deletions(-)

diff --git a/usertools/dpdk-hugepages.py b/usertools/dpdk-hugepages.py
index bf2575ba36..54232ddf22 100755
--- a/usertools/dpdk-hugepages.py
+++ b/usertools/dpdk-hugepages.py
@@ -74,21 +74,37 @@ def set_hugepages(path, reqpages):
   gotpages, reqpages, filename))
  
  
+def get_numa_pages_node(node):

+'''Read list of hugepage reservations on specific NUMA node'''
+hp_path = f'/sys/devices/system/node/node{node}/hugepages'
+if not os.path.exists(hp_path):
+return


Oops, should have been return []

--
Thanks,
Anatoly



[PATCH v1 0/2] dts: port over checksum offload suite

2024-08-16 Thread Dean Marx
Port over checksum hardware offload testing suite from old DTS. The
suite verifies the ability of the PMD to recognize whether an incoming
packet has valid or invalid L4/IP checksum values.

v1: In the original test plan, there were two Tx checksum test cases. I
removed them due to the lack of consistency in testpmd with Tx checksum
flags, either not being displayed during packet transmission or showing
values that did not align with the original test plan. 

Dean Marx (2):
  dts: add csum HW offload to testpmd shell
  dts: checksum offload test suite

 dts/framework/remote_session/testpmd_shell.py | 124 +
 dts/tests/TestSuite_checksum_offload.py   | 255 ++
 2 files changed, 379 insertions(+)
 create mode 100644 dts/tests/TestSuite_checksum_offload.py

-- 
2.44.0



[PATCH v1 1/2] dts: add csum HW offload to testpmd shell

2024-08-16 Thread Dean Marx
add csum_set_hw method to testpmd shell class. Port over
set_verbose and port start/stop from queue start/stop suite.

Signed-off-by: Dean Marx 
---
 dts/framework/remote_session/testpmd_shell.py | 124 ++
 1 file changed, 124 insertions(+)

diff --git a/dts/framework/remote_session/testpmd_shell.py 
b/dts/framework/remote_session/testpmd_shell.py
index 43e9f56517..be7cd16b96 100644
--- a/dts/framework/remote_session/testpmd_shell.py
+++ b/dts/framework/remote_session/testpmd_shell.py
@@ -334,6 +334,32 @@ def make_parser(cls) -> ParserFn:
 )
 
 
+class ChecksumOffloadOptions(Flag):
+"""Flag representing checksum hardware offload layer options."""
+
+#:
+ip = auto()
+#:
+udp = auto()
+#:
+tcp = auto()
+#:
+sctp = auto()
+#:
+outerip = auto()
+#:
+outerudp = auto()
+
+def __str__(self):
+"""String method for use in csum_set_hw."""
+if self == ChecksumOffloadOptions.outerip:
+return "outer-ip"
+elif self == ChecksumOffloadOptions.outerudp:
+return "outer-udp"
+else:
+return f"{self.name}"
+
+
 class DeviceErrorHandlingMode(StrEnum):
 """Enum representing the device error handling mode."""
 
@@ -806,6 +832,104 @@ def show_port_stats(self, port_id: int) -> 
TestPmdPortStats:
 
 return TestPmdPortStats.parse(output)
 
+def port_stop(self, port: int, verify: bool = True):
+"""Stop specified port.
+
+Args:
+port: Specifies the port number to use, must be between 0-32.
+verify: If :data:`True`, the output of the command is scanned
+to ensure specified port is stopped. If not, it is considered
+an error.
+
+Raises:
+InteractiveCommandExecutionError: If `verify` is :data:`True` and 
the port
+is not stopped.
+"""
+port_output = self.send_command(f"port stop {port}")
+if verify:
+if "Done" not in port_output:
+self._logger.debug(f"Failed to stop port {port}: 
\n{port_output}")
+raise InteractiveCommandExecutionError(f"Testpmd failed to 
stop port {port}.")
+
+def port_start(self, port: int, verify: bool = True):
+"""Start specified port.
+
+Args:
+port: Specifies the port number to use, must be between 0-32.
+verify: If :data:`True`, the output of the command is scanned
+to ensure specified port is started. If not, it is considered
+an error.
+
+Raises:
+InteractiveCommandExecutionError: If `verify` is :data:`True` and 
the port
+is not started.
+"""
+port_output = self.send_command(f"port start {port}")
+if verify:
+if "Done" not in port_output:
+self._logger.debug(f"Failed to start port {port}: 
\n{port_output}")
+raise InteractiveCommandExecutionError(f"Testpmd failed to 
start port {port}.")
+
+def csum_set_hw(self, layer: ChecksumOffloadOptions, port_id: int, verify: 
bool = True) -> None:
+"""Enables hardware checksum offloading on the specified layer.
+
+Args:
+layer: The layer that checksum offloading should be enabled on.
+options: tcp, ip, udp, sctp, outer-ip, outer-udp.
+port_id: The port number to enable checksum offloading on, should 
be within 0-32.
+verify: If :data:`True` the output of the command will be scanned 
in an attempt to
+verify that checksum offloading was enabled on the port.
+
+Raises:
+InteractiveCommandExecutionError: If checksum offload is not 
enabled successfully.
+"""
+csum_output = self.send_command(f"csum set {str(layer)} hw {port_id}")
+if verify:
+if "Bad arguments" in csum_output or f"Please stop port {port_id} 
first" in csum_output:
+self._logger.debug(f"Failed to set csum hw mode on port 
{port_id}:\n{csum_output}")
+raise InteractiveCommandExecutionError(
+f"Failed to set csum hw mode on port {port_id}"
+)
+if verify and f"checksum offload is not supported by port {port_id}" 
in csum_output:
+self._logger.debug(f"Checksum {layer} offload is not 
supported:\n{csum_output}")
+
+success = False
+if layer == ChecksumOffloadOptions.outerip:
+if "Outer-Ip checksum offload is hw" in csum_output:
+success = True
+elif layer == ChecksumOffloadOptions.outerudp:
+if "Outer-Udp checksum offload is hw" in csum_output:
+success = True
+else:
+if f"{str(layer).upper} checksum offload is hw" in csum_output:
+success = True
+if not success and verify:
+self._logger.debug(f"Failed to set csum hw mode on port 
{port_id}:\n{

[PATCH v1 2/2] dts: checksum offload test suite

2024-08-16 Thread Dean Marx
test suite for verifying layer 3/4 checksum offload
features on poll mode driver.

Depends-on: patch-143033
("dts: add text parser for testpmd verbose output")
Depends-on: patch-142691
("dts: add send_packets to test suites and rework packet addressing")

Signed-off-by: Dean Marx 
---
 dts/tests/TestSuite_checksum_offload.py | 255 
 1 file changed, 255 insertions(+)
 create mode 100644 dts/tests/TestSuite_checksum_offload.py

diff --git a/dts/tests/TestSuite_checksum_offload.py 
b/dts/tests/TestSuite_checksum_offload.py
new file mode 100644
index 00..b327261ff3
--- /dev/null
+++ b/dts/tests/TestSuite_checksum_offload.py
@@ -0,0 +1,255 @@
+# SPDX-License-Identifier: BSD-3-Clause
+# Copyright(c) 2024 University of New Hampshire
+
+"""DPDK checksum offload testing suite.
+
+This suite verifies L3/L4 checksum offload features of the Poll Mode Driver.
+On the Rx side, IPv4 and UDP/TCP checksum by hardware is checked to ensure
+checksum flags match expected flags. On the Tx side, IPv4/UDP, IPv4/TCP,
+IPv6/UDP, and IPv6/TCP insertion by hardware is checked to checksum flags
+match expected flags.
+
+"""
+
+from typing import List
+
+from scapy.all import Packet  # type: ignore[import-untyped]
+from scapy.layers.inet import IP, TCP, UDP  # type: ignore[import-untyped]
+from scapy.layers.inet6 import IPv6  # type: ignore[import-untyped]
+from scapy.layers.sctp import SCTP  # type: ignore[import-untyped]
+from scapy.layers.l2 import Dot1Q  # type: ignore[import-untyped]
+from scapy.layers.l2 import Ether
+from scapy.packet import Raw  # type: ignore[import-untyped]
+
+from framework.remote_session.testpmd_shell import (
+SimpleForwardingModes,
+TestPmdShell,
+VerboseOLFlag,
+ChecksumOffloadOptions
+)
+from framework.test_suite import TestSuite
+
+
+class TestChecksumOffload(TestSuite):
+"""Checksum offload test suite.
+
+This suite consists of 8 test cases:
+1. Insert checksum on transmit packet
+2. Do not insert checksum on transmit packet
+3. Validate Rx checksum valid flags
+4. Hardware checksum check L4 Rx
+5. Hardware checksum check L3 Rx
+6. Checksum offload with vlan
+
+"""
+
+def set_up_suite(self) -> None:
+"""Set up the test suite.
+
+Setup:
+Verify that at least two port links are created when the
+test run is initialized.
+"""
+self.verify(len(self._port_links) > 1, "Not enough port links.")
+
+def send_packet_and_verify(
+self, packet_list: List[Packet], load: str, should_receive: bool
+) -> None:
+"""Send and verify packet is received on the traffic generator.
+
+Args:
+packet_list: list of Scapy packets to send and verify.
+load: Raw layer load attribute in the sent packet.
+should_receive: Indicates whether the packet should be received
+by the traffic generator.
+"""
+for i in range(0, len(packet_list)):
+received_packets = 
self.send_packet_and_capture(packet=packet_list[i])
+received = any(
+packet.haslayer(Raw) and load in str(packet.load) for packet 
in received_packets
+)
+self.verify(
+received == should_receive,
+f"Packet was {'dropped' if should_receive else 'received'}",
+)
+
+def send_packet_and_verify_checksum(
+self, packet: Packet, goodL4: bool, goodIP: bool, testpmd: TestPmdShell
+) -> None:
+"""Send packet and verify verbose output matches expected output.
+
+Args:
+packet: Scapy packet to send to DUT.
+goodL4: Verifies RTE_MBUF_F_RX_L4_CKSUM_GOOD in verbose output
+if :data:`True`, or RTE_MBUF_F_RX_L4_CKSUM_UNKNOWN if 
:data:`False`.
+goodIP: Verifies RTE_MBUF_F_RX_IP_CKSUM_GOOD in verbose output
+if :data:`True`, or RTE_MBUF_F_RX_IP_CKSUM_UNKNOWN if 
:data:`False`.
+testpmd: Testpmd shell session to analyze verbose output of.
+"""
+testpmd.start()
+self.send_packet_and_capture(packet=packet)
+verbose_output = testpmd.extract_verbose_output(testpmd.stop())
+isL4 = any(
+VerboseOLFlag.RTE_MBUF_F_RX_L4_CKSUM_GOOD in packet.ol_flags
+for index in verbose_output
+for packet in index.packets
+)
+isIP = any(
+VerboseOLFlag.RTE_MBUF_F_RX_IP_CKSUM_GOOD in packet.ol_flags
+for index in verbose_output
+for packet in index.packets
+)
+self.verify(isL4 == goodL4, "Layer 4 checksum flag did not match 
expected checksum flag.")
+self.verify(isIP == goodIP, "IP checksum flag did not match expected 
checksum flag.")
+
+def setup_hw_offload(self, testpmd: TestPmdShell) -> None:
+"""Sets IP, UDP, TCP, and SCTP layers to hardware offload."""
+testpmd.port_stop(port=0)
+t

[PATCH v4 3/5] graph: add stats for node specific errors

2024-08-16 Thread pbhagavatula
From: Pavan Nikhilesh 

Add support for retrieving/printing stats for node specific
errors using rte_graph_cluster_stats_get().

Signed-off-by: Pavan Nikhilesh 
---
 lib/graph/graph_stats.c | 79 -
 lib/graph/rte_graph.h   |  4 +++
 2 files changed, 82 insertions(+), 1 deletion(-)

diff --git a/lib/graph/graph_stats.c b/lib/graph/graph_stats.c
index d71451a17b..1ac0a6797f 100644
--- a/lib/graph/graph_stats.c
+++ b/lib/graph/graph_stats.c
@@ -121,6 +121,25 @@ print_node(FILE *f, const struct 
rte_graph_cluster_node_stats *stat, bool dispat
}
 }
 
+static inline void
+print_err(FILE *f, const struct rte_graph_cluster_node_stats *stat, bool 
dispatch)
+{
+   int i;
+
+   if (dispatch) {
+   for (i = 0; i < stat->node_error_cntrs; i++)
+   fprintf(f,
+   "|\t%-24s|%15s|%-15" PRIu64 
"|%15s|%15s|%15s|%15s|%15s|%11.4s|\n",
+   stat->node_error_desc[i], "", 
stat->node_error_count[i], "", "", "",
+   "", "", "");
+   } else {
+   for (i = 0; i < stat->node_error_cntrs; i++)
+   fprintf(f, "|\t%-24s|%15s|%-15" PRIu64 
"|%15s|%15.3s|%15.6s|%11.4s|\n",
+   stat->node_error_desc[i], "", 
stat->node_error_count[i], "", "", "",
+   "");
+   }
+}
+
 static int
 graph_cluster_stats_cb(bool dispatch, bool is_first, bool is_last, void 
*cookie,
   const struct rte_graph_cluster_node_stats *stat)
@@ -129,8 +148,11 @@ graph_cluster_stats_cb(bool dispatch, bool is_first, bool 
is_last, void *cookie,
 
if (unlikely(is_first))
print_banner(f, dispatch);
-   if (stat->objs)
+   if (stat->objs) {
print_node(f, stat, dispatch);
+   if (stat->node_error_cntrs)
+   print_err(f, stat, dispatch);
+   }
if (unlikely(is_last)) {
if (dispatch)
boarder_model_dispatch();
@@ -203,6 +225,7 @@ stats_mem_populate(struct rte_graph_cluster_stats 
**stats_in,
struct cluster_node *cluster;
struct rte_node *node;
rte_node_t count;
+   uint8_t i;
 
cluster = stats->clusters;
 
@@ -240,6 +263,36 @@ stats_mem_populate(struct rte_graph_cluster_stats 
**stats_in,
SET_ERR_JMP(ENOENT, free, "Failed to find node %s in graph %s",
graph_node->node->name, graph->name);
cluster->nodes[cluster->nb_nodes++] = node;
+   if (graph_node->node->errs) {
+   cluster->stat.node_error_cntrs = 
graph_node->node->errs->nb_errors;
+   cluster->stat.node_error_count = rte_zmalloc_socket(
+   NULL, sizeof(uint64_t) * 
graph_node->node->errs->nb_errors,
+   RTE_CACHE_LINE_SIZE, stats->socket_id);
+   if (cluster->stat.node_error_count == NULL)
+   SET_ERR_JMP(ENOMEM, free, "Failed to allocate memory 
node %s graph %s",
+   graph_node->node->name, graph->name);
+
+   cluster->stat.node_error_desc = rte_zmalloc_socket(
+   NULL, sizeof(RTE_NODE_ERROR_DESC_SIZE) * 
graph_node->node->errs->nb_errors,
+   RTE_CACHE_LINE_SIZE, stats->socket_id);
+   if (cluster->stat.node_error_desc == NULL) {
+   rte_free(cluster->stat.node_error_count);
+   SET_ERR_JMP(ENOMEM, free, "Failed to allocate memory 
node %s graph %s",
+   graph_node->node->name, graph->name);
+   }
+
+   for (i = 0; i < cluster->stat.node_error_cntrs; i++) {
+   if (rte_strscpy(cluster->stat.node_error_desc[i],
+   graph_node->node->errs->err_desc[i],
+   RTE_NODE_ERROR_DESC_SIZE) < 0) {
+   rte_free(cluster->stat.node_error_count);
+   rte_free(cluster->stat.node_error_desc);
+   SET_ERR_JMP(E2BIG, free,
+   "Error description overflow node %s 
graph %s",
+   graph_node->node->name, 
graph->name);
+   }
+   }
+   }
 
stats->sz += stats->cluster_node_size;
stats->max_nodes++;
@@ -388,6 +441,18 @@ rte_graph_cluster_stats_create(const struct 
rte_graph_cluster_stats_param *prm)
 void
 rte_graph_cluster_stats_destroy(struct rte_graph_cluster_stats *stat)
 {
+   struct cluster_node *cluster;
+   rte_node_t count;
+
+   cluster = stat->clusters;
+   for (count = 0; count < stat->max_nodes; count++) {
+   if (cluster->stat.node_error_cntrs) {
+   rte_free(cluster->stat.node_error_count);
+   

[PATCH v4 2/5] graph: add node fastpath error counters

2024-08-16 Thread pbhagavatula
From: Pavan Nikhilesh 

Add node fastpath error counters advertised during
node registration.

Signed-off-by: Pavan Nikhilesh 
---
 lib/graph/graph_populate.c  | 20 ++--
 lib/graph/graph_private.h   |  2 ++
 lib/graph/rte_graph_worker_common.h |  1 +
 3 files changed, 21 insertions(+), 2 deletions(-)

diff --git a/lib/graph/graph_populate.c b/lib/graph/graph_populate.c
index ed596a7711..9c4c919e14 100644
--- a/lib/graph/graph_populate.c
+++ b/lib/graph/graph_populate.c
@@ -39,6 +39,15 @@ graph_fp_mem_calc_size(struct graph *graph)
/* Pointer to next nodes(edges) */
sz += sizeof(struct rte_node *) * graph_node->node->nb_edges;
}
+   sz = RTE_ALIGN(sz, RTE_CACHE_LINE_SIZE);
+   graph->errors_start = sz;
+   /* For 0..N node objects with error counters */
+   STAILQ_FOREACH(graph_node, &graph->node_list, next) {
+   if (graph_node->node->errs == NULL)
+   continue;
+   sz = RTE_ALIGN(sz, RTE_CACHE_LINE_SIZE);
+   sz += sizeof(uint64_t) * graph_node->node->errs->nb_errors;
+   }

graph->mem_sz = sz;
return sz;
@@ -64,6 +73,7 @@ graph_header_popluate(struct graph *_graph)
 static void
 graph_nodes_populate(struct graph *_graph)
 {
+   rte_graph_off_t err_off = _graph->errors_start;
rte_graph_off_t off = _graph->nodes_start;
struct rte_graph *graph = _graph->graph;
struct graph_node *graph_node;
@@ -99,6 +109,12 @@ graph_nodes_populate(struct graph *_graph)
 ->adjacency_list[count]
 ->node->name[0];

+   if (graph_node->node->errs != NULL) {
+   node->err_off = err_off - off;
+   err_off += sizeof(uint64_t) * 
graph_node->node->errs->nb_errors;
+   err_off = RTE_ALIGN(err_off, RTE_CACHE_LINE_SIZE);
+   }
+
off += sizeof(struct rte_node *) * nb_edges;
off = RTE_ALIGN(off, RTE_CACHE_LINE_SIZE);
node->next = off;
@@ -158,7 +174,7 @@ graph_node_nexts_populate(struct graph *_graph)
 }

 static int
-graph_src_nodes_populate(struct graph *_graph)
+graph_src_nodes_offset_populate(struct graph *_graph)
 {
struct rte_graph *graph = _graph->graph;
struct graph_node *graph_node;
@@ -193,7 +209,7 @@ graph_fp_mem_populate(struct graph *graph)
graph_pcap_init(graph);
graph_nodes_populate(graph);
rc = graph_node_nexts_populate(graph);
-   rc |= graph_src_nodes_populate(graph);
+   rc |= graph_src_nodes_offset_populate(graph);

return rc;
 }
diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h
index e663b04d8b..01921b254c 100644
--- a/lib/graph/graph_private.h
+++ b/lib/graph/graph_private.h
@@ -103,6 +103,8 @@ struct graph {
/**< Memzone to store graph data. */
rte_graph_off_t nodes_start;
/**< Node memory start offset in graph reel. */
+   rte_graph_off_t errors_start;
+   /**< Node error memory start offset in graph reel. */
rte_node_t src_node_count;
/**< Number of source nodes in a graph. */
struct rte_graph *graph;
diff --git a/lib/graph/rte_graph_worker_common.h 
b/lib/graph/rte_graph_worker_common.h
index 36d864e2c1..fa59d40f57 100644
--- a/lib/graph/rte_graph_worker_common.h
+++ b/lib/graph/rte_graph_worker_common.h
@@ -110,6 +110,7 @@ struct __rte_cache_aligned rte_node {
uint64_t total_sched_fail; /**< Number of scheduled 
failure. */
} dispatch;
};
+   rte_graph_off_t err_off; /**< Offset to error counters. */
/* Fast path area  */
 #define RTE_NODE_CTX_SZ 16
alignas(RTE_CACHE_LINE_SIZE) uint8_t ctx[RTE_NODE_CTX_SZ]; /**< Node 
Context. */
--
2.25.1



[PATCH v4 4/5] node: add error stats for ip4 lookup node

2024-08-16 Thread pbhagavatula
From: Pavan Nikhilesh 

Add error counters for ip4 LPM lookup failures in
ip4_lookup node.

Signed-off-by: Pavan Nikhilesh 
---
 lib/node/ip4_lookup.c  | 9 +
 lib/node/ip4_lookup_neon.h | 5 +
 lib/node/ip4_lookup_sse.h  | 6 ++
 lib/node/node_private.h| 8 
 4 files changed, 28 insertions(+)

diff --git a/lib/node/ip4_lookup.c b/lib/node/ip4_lookup.c
index 18955971f6..5a7921db75 100644
--- a/lib/node/ip4_lookup.c
+++ b/lib/node/ip4_lookup.c
@@ -86,6 +86,7 @@ ip4_lookup_node_process_scalar(struct rte_graph *graph, 
struct rte_node *node,
rc = rte_lpm_lookup(lpm, rte_be_to_cpu_32(ipv4_hdr->dst_addr),
&next_hop);
next_hop = (rc == 0) ? next_hop : drop_nh;
+   NODE_INCREMENT_ERROR_ID(node, 0, (rc != 0), 1);
 
node_mbuf_priv1(mbuf, dyn)->nh = (uint16_t)next_hop;
next_hop = next_hop >> 16;
@@ -219,11 +220,19 @@ ip4_lookup_node_init(const struct rte_graph *graph, 
struct rte_node *node)
return 0;
 }
 
+static struct rte_node_errors ip4_lookup_errors = {
+   .nb_errors = 1,
+   .err_desc = {
+   [0] = "ip4_lookup_error",
+   },
+};
+
 static struct rte_node_register ip4_lookup_node = {
.process = ip4_lookup_node_process_scalar,
.name = "ip4_lookup",
 
.init = ip4_lookup_node_init,
+   .errs = &ip4_lookup_errors,
 
.nb_edges = RTE_NODE_IP4_LOOKUP_NEXT_PKT_DROP + 1,
.next_nodes = {
diff --git a/lib/node/ip4_lookup_neon.h b/lib/node/ip4_lookup_neon.h
index d5c8da3719..907c7c955a 100644
--- a/lib/node/ip4_lookup_neon.h
+++ b/lib/node/ip4_lookup_neon.h
@@ -116,6 +116,10 @@ ip4_lookup_node_process_vec(struct rte_graph *graph, 
struct rte_node *node,
priv01.u16[4] = result.u16[2];
priv23.u16[0] = result.u16[4];
priv23.u16[4] = result.u16[6];
+   NODE_INCREMENT_ERROR_ID(node, 0, result.u16[1] == (drop_nh >> 
16), 1);
+   NODE_INCREMENT_ERROR_ID(node, 0, result.u16[3] == (drop_nh >> 
16), 1);
+   NODE_INCREMENT_ERROR_ID(node, 0, result.u16[5] == (drop_nh >> 
16), 1);
+   NODE_INCREMENT_ERROR_ID(node, 0, result.u16[7] == (drop_nh >> 
16), 1);
 
node_mbuf_priv1(mbuf0, dyn)->u = priv01.u64[0];
node_mbuf_priv1(mbuf1, dyn)->u = priv01.u64[1];
@@ -202,6 +206,7 @@ ip4_lookup_node_process_vec(struct rte_graph *graph, struct 
rte_node *node,
&next_hop);
next_hop = (rc == 0) ? next_hop : drop_nh;
 
+   NODE_INCREMENT_ERROR_ID(node, 0, (rc != 0), 1);
node_mbuf_priv1(mbuf0, dyn)->nh = (uint16_t)next_hop;
next_hop = next_hop >> 16;
next0 = (uint16_t)next_hop;
diff --git a/lib/node/ip4_lookup_sse.h b/lib/node/ip4_lookup_sse.h
index 74dbf97533..a38131e629 100644
--- a/lib/node/ip4_lookup_sse.h
+++ b/lib/node/ip4_lookup_sse.h
@@ -115,6 +115,11 @@ ip4_lookup_node_process_vec(struct rte_graph *graph, 
struct rte_node *node,
/* Perform LPM lookup to get NH and next node */
rte_lpm_lookupx4(lpm, dip, dst.u32, drop_nh);
 
+   NODE_INCREMENT_ERROR_ID(node, 0, dst.u16[1] == (drop_nh >> 16), 
1);
+   NODE_INCREMENT_ERROR_ID(node, 0, dst.u16[3] == (drop_nh >> 16), 
1);
+   NODE_INCREMENT_ERROR_ID(node, 0, dst.u16[5] == (drop_nh >> 16), 
1);
+   NODE_INCREMENT_ERROR_ID(node, 0, dst.u16[7] == (drop_nh >> 16), 
1);
+
/* Extract next node id and NH */
node_mbuf_priv1(mbuf0, dyn)->nh = dst.u32[0] & 0x;
next0 = (dst.u32[0] >> 16);
@@ -206,6 +211,7 @@ ip4_lookup_node_process_vec(struct rte_graph *graph, struct 
rte_node *node,
rc = rte_lpm_lookup(lpm, rte_be_to_cpu_32(ipv4_hdr->dst_addr),
&next_hop);
next_hop = (rc == 0) ? next_hop : drop_nh;
+   NODE_INCREMENT_ERROR_ID(node, 0, rc != 0, 1);
 
node_mbuf_priv1(mbuf0, dyn)->nh = next_hop & 0x;
next0 = (next_hop >> 16);
diff --git a/lib/node/node_private.h b/lib/node/node_private.h
index 1de7306792..36b2a733db 100644
--- a/lib/node/node_private.h
+++ b/lib/node/node_private.h
@@ -12,6 +12,8 @@
 #include 
 #include 
 
+#include 
+
 extern int rte_node_logtype;
 #define RTE_LOGTYPE_NODE rte_node_logtype
 
@@ -88,4 +90,10 @@ node_mbuf_priv2(struct rte_mbuf *m)
return (struct node_mbuf_priv2 *)rte_mbuf_to_priv(m);
 }
 
+#define NODE_INCREMENT_ERROR_ID(node, id, cond, cnt)   
\
+   {   
   \
+   if (unlikely(rte_graph_has_stats_feature() && (cond)))  
   \
+   ((uint64_t *)RTE_PTR_ADD(node, node->err_off))[id] += 
(cnt);   

[PATCH v4 5/5] node: add error stats for ip4 reassembly node

2024-08-16 Thread pbhagavatula
From: Pavan Nikhilesh 

Add reassembly failure error counter for ip4 reassembly
node.

Signed-off-by: Pavan Nikhilesh 
---
 lib/node/ip4_reassembly.c | 9 +
 1 file changed, 9 insertions(+)

diff --git a/lib/node/ip4_reassembly.c b/lib/node/ip4_reassembly.c
index 04823cc596..ab71ef1331 100644
--- a/lib/node/ip4_reassembly.c
+++ b/lib/node/ip4_reassembly.c
@@ -120,6 +120,7 @@ ip4_reassembly_node_process(struct rte_graph *graph, struct 
rte_node *node, void
rte_node_next_stream_put(graph, node, 
RTE_NODE_IP4_REASSEMBLY_NEXT_PKT_DROP,
 dr->cnt);
idx += dr->cnt;
+   NODE_INCREMENT_ERROR_ID(node, 0, dr->cnt, dr->cnt);
dr->cnt = 0;
}
 
@@ -165,11 +166,19 @@ ip4_reassembly_node_init(const struct rte_graph *graph, 
struct rte_node *node)
return 0;
 }
 
+static struct rte_node_errors ip4_reassembly_errors = {
+   .nb_errors = 1,
+   .err_desc = {
+   [0] = "ip4_reassembly_error",
+   },
+};
+
 static struct rte_node_register ip4_reassembly_node = {
.process = ip4_reassembly_node_process,
.name = "ip4_reassembly",
 
.init = ip4_reassembly_node_init,
+   .errs = &ip4_reassembly_errors,
 
.nb_edges = RTE_NODE_IP4_REASSEMBLY_NEXT_PKT_DROP + 1,
.next_nodes = {
-- 
2.25.1



[DPDK/cryptodev Bug 1523] cryptodev_openssl_asym_autotest unit test fails on OpenSUSE Leap 15.6

2024-08-16 Thread bugzilla
https://bugs.dpdk.org/show_bug.cgi?id=1523

Bug ID: 1523
   Summary: cryptodev_openssl_asym_autotest unit test fails on
OpenSUSE Leap 15.6
   Product: DPDK
   Version: unspecified
  Hardware: All
OS: All
Status: UNCONFIRMED
  Severity: normal
  Priority: Normal
 Component: cryptodev
  Assignee: dev@dpdk.org
  Reporter: cch...@iol.unh.edu
  Target Milestone: ---

Created attachment 285
  --> https://bugs.dpdk.org/attachment.cgi?id=285&action=edit
DPDK 23.11 compile and unit test log

On OpenSUSE Leap 15.6, the cryptodev_openssl_asym_autotest unit test is failing
with DPDK 24.11 and 23.11 and the cryptodev_openssl_autotest unit test is
failing with DPDK 21.11. This fail appeared after our container images were
updated from OpenSUSE Leap 15.5 (which had openssl 1.1.1l) to 15.6 (which has
openssl 3.1.4).

GCC version: gcc (SUSE Linux) 7.5.0
OpenSSL version: 3.1.4

Because I can only add one attachment, I have attached the logs for DPDK 23.11
here and have uploaded the logs for DPDK 21.11 on Google Drive here:
https://drive.google.com/file/d/1tKVYWIN82Mjc0PGGLZ_pNuiqG89uc6JQ/view.

-- 
You are receiving this mail because:
You are the assignee for the bug.

Re: [dpdk-dev] [PATCH v2 00/41] add MLX5 FreeBSD support

2024-08-16 Thread Stephen Hemminger
On Fri,  8 Oct 2021 00:13:09 +0530
Srikanth Kaka  wrote:

> This patch series introduces FreeBSD OS support for MLX5 PMD
> 
> Currently there is an ongoing review of FreeBSD OFED drivers and
> RDMA-core libraries. Their status can be tracked at
> https://reviews.freebsd.org/p/vag.singh_oneconvergence.com
> 
> Only MLX5 SRIOV interface is supported along with the following features:
> 
>   - Hypervisors: KVM, ESXi and Azure/HyperV
>   - Multiple Tx and Rx queues
>   - Support for scattered Tx and Rx frames
>   - RSS
>   - VLAN filtering, stripping and insertion
>   - RX CRC stripping configuration
>   - Promiscuous and Multicast mode
>   - Statistics query including Basic, Extended, and per queue.
>   - Configurable MTU and MAC
>   - Jumbo frame support
>   - Trust mode support
>   - Vxlan
>   - QoS
>   - Flow steering
> 
> Tested on :
> MT27710 Family [ConnectX-4 Lx]
> MT27710 Family [ConnectX-4 Lx Virtual Function] (Azure)
> MT27700 Family [ConnectX-4]
> MT27800 Family [ConnectX-5]
> MT2894 Family  [ConnectX-6 Lx]
> MT28908 Family [ConnectX-6]
> MT2892 Family  [ConnectX-6 Dx]
> 
> v1 : Initial submission
> v2 : Addressed community comments

Not obvious why this patch set never got merged. Most likely because
it required upstream FreeBSD libraries that were not merged at the time.

Is this still of interest?

Has the support for RDMA and IB been merged in upstream FreeBSD?


Does the meson build handle the case where the required dependencies are
not present in the version of FreeBSD being used?

If nothing else, the driver needs to be resubmitted to address 3 years
of changes to mlx5 common code. And will need an entry in MAINTAINERS.