[PATCH v1] examples/ipsec-secgw: free the actual mbuf pointer
In case of crypto event vector, the vector points to crypto op structure in priv area and not actual mbuf. Extract the mbuf pointer and pass these to rte_mbuf_free to free the mbufs. Signed-off-by: Vidya Sagar Velumuri diff --git a/examples/ipsec-secgw/ipsec_worker.c b/examples/ipsec-secgw/ipsec_worker.c index c9c43ebd2b..dd14226e81 100644 --- a/examples/ipsec-secgw/ipsec_worker.c +++ b/examples/ipsec-secgw/ipsec_worker.c @@ -960,7 +960,18 @@ static void ipsec_event_vector_free(struct rte_event *ev) { struct rte_event_vector *vec = ev->vec; - rte_pktmbuf_free_bulk(vec->mbufs + vec->elem_offset, vec->nb_elem); + + if (ev->event_type == RTE_EVENT_TYPE_CRYPTODEV_VECTOR) { + struct rte_crypto_op *cop; + int i; + + for (i = 0; i < vec->nb_elem; i++) { + cop = vec->ptrs[i]; + rte_pktmbuf_free(cop->sym->m_src); + } + } else { + rte_pktmbuf_free_bulk(vec->mbufs + vec->elem_offset, vec->nb_elem); + } rte_mempool_put(rte_mempool_from_obj(vec), vec); } -- 2.25.1
Re: [dpdk-dev] [PATCH v13 1/4] devtools: script to track symbols over releases
Stephen Hemminger writes: > On Thu, 9 Sep 2021 14:48:05 +0100 > Ray Kinsella wrote: > >> This script tracks the growth of stable and experimental symbols >> over releases since v19.11. The script has the ability to >> count the added symbols between two dpdk releases, and to >> list experimental symbols present in two dpdk releases >> (expired symbols). >> >> example usages: >> >> Count symbols added since v19.11 >> $ devtools/symbol-tool.py count-symbols >> >> Count symbols added since v20.11 >> $ devtools/symbol-tool.py count-symbols --releases v20.11,v21.05 >> >> List experimental symbols present in v20.11 and v21.05 >> $ devtools/symbol-tool.py list-expired --releases v20.11,v21.05 >> >> List experimental symbols in libraries only, present since v19.11 >> $ devtools/symbol-tool.py list-expired --directory lib >> >> Signed-off-by: Ray Kinsella > > Tried it, looks useful. Minor notes: I think it was mostly lack of interest that this didn't get over the line when I submitted it orginally. The flake8 errors as easily fixed, I can do that if you thought it might be useful? > - requires python parsley library which was not required before > - minor style format > > $ flake8 ./devtools/symbol-tool.py > ./devtools/symbol-tool.py:369:11: E275 missing whitespace after keyword > > Acked-by: Stephen Hemminger -- Regards, Ray K
Re: [PATCH] app/testpmd: add L4 port to verbose output
On 15/08/2024 16:22, Stephen Hemminger wrote: The verbose output is already too verbose. Maybe you would like the simpler format (which does include the port number) see the network packet dissector patches. Hi Stephen, Thank you for the reply you left to Alex's patch. This is actually quite helpful. I've had a look at your patch series and the hexdump is actually quite great for our purpose at DTS, as it will allow us to detect a preset packet identifier in the payload. Unfortunately, your patch doesn't solve the problem that Alex is having. He's currently writing up a test suite to test the RSS hash keys and the respective queues to which the incoming packets are placed into. To correctly match the packets to the queue and hash key, he needs a way to identify them. We explored some solutions and the L4 port seemed to be the most problem-free with the least effort. Using the payload is probably best and you've already posted the required changes too. The only problem is that your hexdump output does not print the required RSS values. Would you be be keen to add these to your patch? Otherwise, do you have any ideas and/or solutions to tackle this specific problem? Thank you, Luca Vizzarro
[RFC] table: report victim replace stats in LRU tables
LRU caches replace records when requested table bucket is full. There is, however, no information about this happenning in either the return value or the statistics. This commit introduces a counter for such cases. Signed-off-by: Kamil Vojanec --- lib/table/rte_table.h| 1 + lib/table/rte_table_hash_key16.c | 4 lib/table/rte_table_hash_key32.c | 4 lib/table/rte_table_hash_key8.c | 4 lib/table/rte_table_hash_lru.c | 4 5 files changed, 17 insertions(+) diff --git a/lib/table/rte_table.h b/lib/table/rte_table.h index 9a5faf0e32..e097e25868 100644 --- a/lib/table/rte_table.h +++ b/lib/table/rte_table.h @@ -33,6 +33,7 @@ struct rte_mbuf; struct rte_table_stats { uint64_t n_pkts_in; uint64_t n_pkts_lookup_miss; + uint64_t n_pkts_insert_victims; }; /** diff --git a/lib/table/rte_table_hash_key16.c b/lib/table/rte_table_hash_key16.c index 67b77c16a0..f97d9c1266 100644 --- a/lib/table/rte_table_hash_key16.c +++ b/lib/table/rte_table_hash_key16.c @@ -27,11 +27,14 @@ table->stats.n_pkts_in += val #define RTE_TABLE_HASH_KEY16_STATS_PKTS_LOOKUP_MISS(table, val) \ table->stats.n_pkts_lookup_miss += val +#define RTE_TABLE_HASH_KEY16_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) \ + table->stats.n_pkts_insert_victims += val #else #define RTE_TABLE_HASH_KEY16_STATS_PKTS_IN_ADD(table, val) #define RTE_TABLE_HASH_KEY16_STATS_PKTS_LOOKUP_MISS(table, val) +#define RTE_TABLE_HASH_KEY16_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) #endif @@ -304,6 +307,7 @@ rte_table_hash_entry_add_key16_lru( } /* Bucket full: replace LRU entry */ + RTE_TABLE_HASH_KEY16_STATS_PKTS_INSERT_VICTIMS_ADD(f, 1); pos = lru_pos(bucket); bucket->signature[pos] = signature; keycpy(&bucket->key[pos], key, f->key_mask); diff --git a/lib/table/rte_table_hash_key32.c b/lib/table/rte_table_hash_key32.c index 1aa86c6a49..37ad87bb85 100644 --- a/lib/table/rte_table_hash_key32.c +++ b/lib/table/rte_table_hash_key32.c @@ -27,11 +27,14 @@ table->stats.n_pkts_in += val #define RTE_TABLE_HASH_KEY32_STATS_PKTS_LOOKUP_MISS(table, val) \ table->stats.n_pkts_lookup_miss += val +#define RTE_TABLE_HASH_KEY32_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) \ + table->stats.n_pkts_insert_victims += val #else #define RTE_TABLE_HASH_KEY32_STATS_PKTS_IN_ADD(table, val) #define RTE_TABLE_HASH_KEY32_STATS_PKTS_LOOKUP_MISS(table, val) +#define RTE_TABLE_HASH_KEY32_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) #endif @@ -312,6 +315,7 @@ rte_table_hash_entry_add_key32_lru( } /* Bucket full: replace LRU entry */ + RTE_TABLE_HASH_KEY32_STATS_PKTS_INSERT_VICTIMS_ADD(f, 1); pos = lru_pos(bucket); bucket->signature[pos] = signature; keycpy(&bucket->key[pos], key, f->key_mask); diff --git a/lib/table/rte_table_hash_key8.c b/lib/table/rte_table_hash_key8.c index c8d72b333d..c5722f3709 100644 --- a/lib/table/rte_table_hash_key8.c +++ b/lib/table/rte_table_hash_key8.c @@ -25,11 +25,14 @@ table->stats.n_pkts_in += val #define RTE_TABLE_HASH_KEY8_STATS_PKTS_LOOKUP_MISS(table, val) \ table->stats.n_pkts_lookup_miss += val +#define RTE_TABLE_HASH_KEY8_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) \ + table->stats.n_pkts_insert_victims += val #else #define RTE_TABLE_HASH_KEY8_STATS_PKTS_IN_ADD(table, val) #define RTE_TABLE_HASH_KEY8_STATS_PKTS_LOOKUP_MISS(table, val) +#define RTE_TABLE_HASH_KEY8_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) #endif @@ -292,6 +295,7 @@ rte_table_hash_entry_add_key8_lru( } /* Bucket full: replace LRU entry */ + RTE_TABLE_HASH_KEY8_STATS_PKTS_INSERT_VICTIMS_ADD(f, 1); pos = lru_pos(bucket); keycpy(&bucket->key[pos], key, &f->key_mask); memcpy(&bucket->data[pos * f->entry_size], entry, f->entry_size); diff --git a/lib/table/rte_table_hash_lru.c b/lib/table/rte_table_hash_lru.c index 801e48f5ba..bcf9c34f8e 100644 --- a/lib/table/rte_table_hash_lru.c +++ b/lib/table/rte_table_hash_lru.c @@ -23,11 +23,14 @@ table->stats.n_pkts_in += val #define RTE_TABLE_HASH_LRU_STATS_PKTS_LOOKUP_MISS(table, val) \ table->stats.n_pkts_lookup_miss += val +#define RTE_TABLE_HASH_LRU_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) \ + table->stats.n_pkts_insert_victims += val #else #define RTE_TABLE_HASH_LRU_STATS_PKTS_IN_ADD(table, val) #define RTE_TABLE_HASH_LRU_STATS_PKTS_LOOKUP_MISS(table, val) +#define RTE_TABLE_HASH_LRU_STATS_PKTS_INSERT_VICTIMS_ADD(table, val) #endif @@ -340,6 +343,7 @@ rte_table_hash_lru_entry_add(void *table, void *key, void *entry, /* Bucket full */ { + RTE_TABLE_HASH_LRU_STATS_PKTS_INSERT_VICTIMS_ADD(t, 1); uint64_t pos = lru_pos(bkt); uint32_t bkt_key_index = bkt->key_pos[pos]; uint8_t *bkt_key = &t->key_mem[bkt_key_index << -- 2.43.5
Re: Ethdev tracepoints optimization
On Fri, Aug 16, 2024 at 4:43 AM Adel Belkhiri wrote: > > Hi DPDK Community, > > I am currently working on developing performance analyses for applications > using the ethdev library. These analyses are being implemented in Trace > Compass, an open-source performance analyzer. One of the views I’ve > implemented shows the rate of traffic received or sent by an ethernet port, > measured in packets per second. However, I've encountered an issue with the > lib.ethdev.rx.burst event, which triggers even when no packets are polled, > leading to a significant number of irrelevant events in the trace. This > becomes problematic as these "empty" events can overwhelm the tracer buffer, > potentially causing the loss of more critical events due to their high > frequency. > > To address this, I've modified the DPDK code in lib/ethdev/rte_ethdev.h to > add a conditional statement that only triggers the event when nb_rx > 0. My > question to the community is whether there are use cases where an "empty" > lib.ethdev.rx.burst event could be useful. If not, would there be interest in > submitting a patch with this modification? > > Moreover, I am looking to develop an analysis that calculates the throughput > (in kb/s, mb/s, etc.) per NIC, utilizing the same events (i.e., > lib.ethdev.rx.burst and lib.ethdev.tx.burst). These tracepoints do not > provide packet size directly, only a pointer to the packet array. My attempt > to use an eBPF program to iterate through that array to access the packet > sizes was unsuccessful, as I found no method to export the computed data > (e.g., via a custom tracepoint). Does anyone have suggestions or alternative > approaches for achieving a throughput measurement? Call rte_eth_stats_get() in N second interval and find throughput in slow path. > I would be grateful for any insights or suggestions you might have. > > Thank you! > Adel
[PATCH v2 1/4] usertools/cpu_layout: update coding style
Update coding style: - make it PEP-484 compliant - address all flake8, mypy etc. warnings - use f-strings in place of old-style string interpolation - refactor printing to make the code more readable Signed-off-by: Anatoly Burakov --- usertools/cpu_layout.py | 162 ++-- 1 file changed, 104 insertions(+), 58 deletions(-) diff --git a/usertools/cpu_layout.py b/usertools/cpu_layout.py index 891b9238fa..be86f06938 100755 --- a/usertools/cpu_layout.py +++ b/usertools/cpu_layout.py @@ -3,62 +3,108 @@ # Copyright(c) 2010-2014 Intel Corporation # Copyright(c) 2017 Cavium, Inc. All rights reserved. -sockets = [] -cores = [] -core_map = {} -base_path = "/sys/devices/system/cpu" -fd = open("{}/kernel_max".format(base_path)) -max_cpus = int(fd.read()) -fd.close() -for cpu in range(max_cpus + 1): -try: -fd = open("{}/cpu{}/topology/core_id".format(base_path, cpu)) -except IOError: -continue -core = int(fd.read()) -fd.close() -fd = open("{}/cpu{}/topology/physical_package_id".format(base_path, cpu)) -socket = int(fd.read()) -fd.close() -if core not in cores: -cores.append(core) -if socket not in sockets: -sockets.append(socket) -key = (socket, core) -if key not in core_map: -core_map[key] = [] -core_map[key].append(cpu) - -print(format("=" * (47 + len(base_path -print("Core and Socket Information (as reported by '{}')".format(base_path)) -print("{}\n".format("=" * (47 + len(base_path -print("cores = ", cores) -print("sockets = ", sockets) -print("") - -max_processor_len = len(str(len(cores) * len(sockets) * 2 - 1)) -max_thread_count = len(list(core_map.values())[0]) -max_core_map_len = (max_processor_len * max_thread_count) \ - + len(", ") * (max_thread_count - 1) \ - + len('[]') + len('Socket ') -max_core_id_len = len(str(max(cores))) - -output = " ".ljust(max_core_id_len + len('Core ')) -for s in sockets: -output += " Socket %s" % str(s).ljust(max_core_map_len - len('Socket ')) -print(output) - -output = " ".ljust(max_core_id_len + len('Core ')) -for s in sockets: -output += " ".ljust(max_core_map_len) -output += " " -print(output) - -for c in cores: -output = "Core %s" % str(c).ljust(max_core_id_len) -for s in sockets: -if (s, c) in core_map: -output += " " + str(core_map[(s, c)]).ljust(max_core_map_len) +from typing import List, Set, Dict, Tuple + + +def _range_expand(rstr: str) -> List[int]: +"""Expand a range string into a list of integers.""" +# 0,1-3 => [0, 1-3] +ranges = rstr.split(",") +valset: List[int] = [] +for r in ranges: +# 1-3 => [1, 2, 3] +if "-" in r: +start, end = r.split("-") +valset.extend(range(int(start), int(end) + 1)) else: -output += " " * (max_core_map_len + 1) -print(output) +valset.append(int(r)) +return valset + + +def _read_sysfs(path: str) -> str: +with open(path, encoding="utf-8") as fd: +return fd.read().strip() + + +def _print_row(row: Tuple[str, ...], col_widths: List[int]) -> None: +first, *rest = row +w_first, *w_rest = col_widths +first_end = " " * 4 +rest_end = " " * 10 + +print(first.ljust(w_first), end=first_end) +for cell, width in zip(rest, w_rest): +print(cell.rjust(width), end=rest_end) +print() + + +def _print_section(heading: str) -> None: +sep = "=" * len(heading) +print(sep) +print(heading) +print(sep) +print() + + +def _main() -> None: +sockets_s: Set[int] = set() +cores_s: Set[int] = set() +core_map: Dict[Tuple[int, int], List[int]] = {} +base_path = "/sys/devices/system/cpu" + +cpus = _range_expand(_read_sysfs(f"{base_path}/online")) + +for cpu in cpus: +lcore_base = f"{base_path}/cpu{cpu}" +core = int(_read_sysfs(f"{lcore_base}/topology/core_id")) +socket = int(_read_sysfs(f"{lcore_base}/topology/physical_package_id")) + +cores_s.add(core) +sockets_s.add(socket) +key = (socket, core) +core_map.setdefault(key, []) +core_map[key].append(cpu) + +cores = sorted(cores_s) +sockets = sorted(sockets_s) + +_print_section("Core and Socket Information " + f"(as reported by '{base_path}')") + +print("cores = ", cores) +print("sockets = ", sockets) +print() + +# Core, [Socket, Socket, ...] +heading_strs = "", *[f"Socket {s}" for s in sockets] +sep_strs = tuple("-" * len(hstr) for hstr in heading_strs) +rows: List[Tuple[str, ...]] = [] + +for c in cores: +# Core, +row: Tuple[str, ...] = (f"Core {c}",) + +# [lcores, lcores, ...] +for s in sockets: +try: +lcores = core_map[(s, c)] +row += (str(lcores),) +except KeyError: +row += ("",) +
[PATCH v2 2/4] usertools/cpu_layout: print out NUMA nodes
In traditional NUMA case, NUMA nodes and physical sockets were used interchangeably, but there are cases where there can be multiple NUMA nodes per socket, as well as all CPU's being assigned NUMA node 0 even in cases of multiple sockets. Use sysfs to print out NUMA information. Signed-off-by: Anatoly Burakov --- usertools/cpu_layout.py | 35 ++- 1 file changed, 30 insertions(+), 5 deletions(-) diff --git a/usertools/cpu_layout.py b/usertools/cpu_layout.py index be86f06938..e43bdbf343 100755 --- a/usertools/cpu_layout.py +++ b/usertools/cpu_layout.py @@ -4,6 +4,7 @@ # Copyright(c) 2017 Cavium, Inc. All rights reserved. from typing import List, Set, Dict, Tuple +import glob def _range_expand(rstr: str) -> List[int]: @@ -26,11 +27,19 @@ def _read_sysfs(path: str) -> str: return fd.read().strip() +def _read_numa_node(base: str) -> int: +node_glob = f"{base}/node*" +node_dirs = glob.glob(node_glob) +if not node_dirs: +return 0 # default to node 0 +return int(node_dirs[0].split("node")[1]) + + def _print_row(row: Tuple[str, ...], col_widths: List[int]) -> None: first, *rest = row w_first, *w_rest = col_widths first_end = " " * 4 -rest_end = " " * 10 +rest_end = " " * 4 print(first.ljust(w_first), end=first_end) for cell, width in zip(rest, w_rest): @@ -50,6 +59,7 @@ def _main() -> None: sockets_s: Set[int] = set() cores_s: Set[int] = set() core_map: Dict[Tuple[int, int], List[int]] = {} +numa_map: Dict[int, int] = {} base_path = "/sys/devices/system/cpu" cpus = _range_expand(_read_sysfs(f"{base_path}/online")) @@ -58,12 +68,14 @@ def _main() -> None: lcore_base = f"{base_path}/cpu{cpu}" core = int(_read_sysfs(f"{lcore_base}/topology/core_id")) socket = int(_read_sysfs(f"{lcore_base}/topology/physical_package_id")) +node = _read_numa_node(lcore_base) cores_s.add(core) sockets_s.add(socket) key = (socket, core) core_map.setdefault(key, []) core_map[key].append(cpu) +numa_map[cpu] = node cores = sorted(cores_s) sockets = sorted(sockets_s) @@ -73,24 +85,37 @@ def _main() -> None: print("cores = ", cores) print("sockets = ", sockets) +print("numa = ", sorted(set(numa_map.values( print() -# Core, [Socket, Socket, ...] -heading_strs = "", *[f"Socket {s}" for s in sockets] +# Core, [NUMA, Socket, NUMA, Socket, ...] +heading_strs = "", *[v for s in sockets for v in ("", f"Socket {s}")] sep_strs = tuple("-" * len(hstr) for hstr in heading_strs) rows: List[Tuple[str, ...]] = [] +prev_numa = None for c in cores: # Core, row: Tuple[str, ...] = (f"Core {c}",) -# [lcores, lcores, ...] +# assume NUMA changes symmetrically +first_lcore = core_map[(0, c)][0] +cur_numa = numa_map[first_lcore] +numa_changed = prev_numa != cur_numa +prev_numa = cur_numa + +# [NUMA, lcores, NUMA, lcores, ...] for s in sockets: try: lcores = core_map[(s, c)] +numa = numa_map[lcores[0]] +if numa_changed: +row += (f"NUMA {numa}",) +else: +row += ("",) row += (str(lcores),) except KeyError: -row += ("",) +row += ("", "") rows += [row] # find max widths for each column, including header and rows -- 2.43.5
[PATCH v2 3/4] usertools/dpdk-hugepages.py: sort by NUMA node
Currently, the list of per-NUMA node hugepages is displayed in glob order, which can be arbitrary. Fix it to sort the glob order. Signed-off-by: Anatoly Burakov --- usertools/dpdk-hugepages.py | 40 ++--- 1 file changed, 28 insertions(+), 12 deletions(-) diff --git a/usertools/dpdk-hugepages.py b/usertools/dpdk-hugepages.py index bf2575ba36..54232ddf22 100755 --- a/usertools/dpdk-hugepages.py +++ b/usertools/dpdk-hugepages.py @@ -74,21 +74,37 @@ def set_hugepages(path, reqpages): gotpages, reqpages, filename)) +def get_numa_pages_node(node): +'''Read list of hugepage reservations on specific NUMA node''' +hp_path = f'/sys/devices/system/node/node{node}/hugepages' +if not os.path.exists(hp_path): +return +res = [] +for pg_sz_dir in os.listdir(hp_path): +pg_sz = int(pg_sz_dir[10:-2]) +nr_pages = get_hugepages(f'{hp_path}/{pg_sz_dir}') +if nr_pages > 0: +pg_sz_str = fmt_memsize(pg_sz) +total_sz_str = fmt_memsize(nr_pages * pg_sz) +res += [(nr_pages, pg_sz_str, total_sz_str)] +else: +res += [(0, None, None)] +return res + + def show_numa_pages(): '''Show huge page reservations on Numa system''' +# get list of NUMA nodes and sort them by integer order print('Node Pages Size Total') -for numa_path in glob.glob('/sys/devices/system/node/node*'): -node = numa_path[29:] # slice after /sys/devices/system/node/node -path = numa_path + '/hugepages' -if not os.path.exists(path): -continue -for hdir in os.listdir(path): -pages = get_hugepages(path + '/' + hdir) -if pages > 0: -kb = int(hdir[10:-2]) # slice out of hugepages-NNNkB -print('{:<4} {:<5} {:<6} {}'.format(node, pages, -fmt_memsize(kb), -fmt_memsize(pages * kb))) +nodes = sorted(int(node[29:]) + for node in glob.glob('/sys/devices/system/node/node*')) +for node in nodes: +pg_sz_data = get_numa_pages_node(node) +for nr_pages, pg_sz, total_sz in pg_sz_data: +if not nr_pages: +continue +print('{:<4} {:<5} {:<6} {}' + .format(node, nr_pages, pg_sz, total_sz)) def show_non_numa_pages(): -- 2.43.5
[PATCH v2 4/4] usertools/dpdk-devbind: print NUMA node
Currently, devbind does not print out any NUMA information, which makes figuring out which NUMA node device belongs to not trivial. Add printouts for NUMA information if NUMA support is enabled on the system. Signed-off-by: Anatoly Burakov --- usertools/dpdk-devbind.py | 27 +++ 1 file changed, 19 insertions(+), 8 deletions(-) diff --git a/usertools/dpdk-devbind.py b/usertools/dpdk-devbind.py index b276e8efc8..c0611a501d 100755 --- a/usertools/dpdk-devbind.py +++ b/usertools/dpdk-devbind.py @@ -110,6 +110,11 @@ args = [] +# check if this system has NUMA support +def is_numa(): +return os.path.exists('/sys/devices/system/node') + + # check if a specific kernel module is loaded def module_is_loaded(module): global loaded_modules @@ -579,18 +584,24 @@ def show_device_status(devices_type, device_name, if_field=False): # print each category separately, so we can clearly see what's used by DPDK if dpdk_drv: +extra_param = "drv=%(Driver_str)s unused=%(Module_str)s" +if is_numa(): +extra_param = "numa_node=%(NUMANode)s " + extra_param display_devices("%s devices using DPDK-compatible driver" % device_name, -dpdk_drv, "drv=%(Driver_str)s unused=%(Module_str)s") +dpdk_drv, extra_param) if kernel_drv: -if_text = "" +extra_param = "drv=%(Driver_str)s unused=%(Module_str)s" if if_field: -if_text = "if=%(Interface)s " -display_devices("%s devices using kernel driver" % device_name, kernel_drv, -if_text + "drv=%(Driver_str)s " -"unused=%(Module_str)s %(Active)s") +extra_param = "if=%(Interface)s " + extra_param +if is_numa(): +extra_param = "numa_node=%(NUMANode)s " + extra_param +display_devices("%s devices using kernel driver" % device_name, +kernel_drv, extra_param) if no_drv: -display_devices("Other %s devices" % device_name, no_drv, -"unused=%(Module_str)s") +extra_param = "unused=%(Module_str)s" +if is_numa(): +extra_param = "numa_node=%(NUMANode)s " + extra_param +display_devices("Other %s devices" % device_name, no_drv, extra_param) def show_status(): -- 2.43.5
Re: [PATCH v2 3/4] usertools/dpdk-hugepages.py: sort by NUMA node
On 8/16/2024 2:16 PM, Anatoly Burakov wrote: Currently, the list of per-NUMA node hugepages is displayed in glob order, which can be arbitrary. Fix it to sort the glob order. Signed-off-by: Anatoly Burakov --- usertools/dpdk-hugepages.py | 40 ++--- 1 file changed, 28 insertions(+), 12 deletions(-) diff --git a/usertools/dpdk-hugepages.py b/usertools/dpdk-hugepages.py index bf2575ba36..54232ddf22 100755 --- a/usertools/dpdk-hugepages.py +++ b/usertools/dpdk-hugepages.py @@ -74,21 +74,37 @@ def set_hugepages(path, reqpages): gotpages, reqpages, filename)) +def get_numa_pages_node(node): +'''Read list of hugepage reservations on specific NUMA node''' +hp_path = f'/sys/devices/system/node/node{node}/hugepages' +if not os.path.exists(hp_path): +return Oops, should have been return [] -- Thanks, Anatoly
[PATCH v1 0/2] dts: port over checksum offload suite
Port over checksum hardware offload testing suite from old DTS. The suite verifies the ability of the PMD to recognize whether an incoming packet has valid or invalid L4/IP checksum values. v1: In the original test plan, there were two Tx checksum test cases. I removed them due to the lack of consistency in testpmd with Tx checksum flags, either not being displayed during packet transmission or showing values that did not align with the original test plan. Dean Marx (2): dts: add csum HW offload to testpmd shell dts: checksum offload test suite dts/framework/remote_session/testpmd_shell.py | 124 + dts/tests/TestSuite_checksum_offload.py | 255 ++ 2 files changed, 379 insertions(+) create mode 100644 dts/tests/TestSuite_checksum_offload.py -- 2.44.0
[PATCH v1 1/2] dts: add csum HW offload to testpmd shell
add csum_set_hw method to testpmd shell class. Port over set_verbose and port start/stop from queue start/stop suite. Signed-off-by: Dean Marx --- dts/framework/remote_session/testpmd_shell.py | 124 ++ 1 file changed, 124 insertions(+) diff --git a/dts/framework/remote_session/testpmd_shell.py b/dts/framework/remote_session/testpmd_shell.py index 43e9f56517..be7cd16b96 100644 --- a/dts/framework/remote_session/testpmd_shell.py +++ b/dts/framework/remote_session/testpmd_shell.py @@ -334,6 +334,32 @@ def make_parser(cls) -> ParserFn: ) +class ChecksumOffloadOptions(Flag): +"""Flag representing checksum hardware offload layer options.""" + +#: +ip = auto() +#: +udp = auto() +#: +tcp = auto() +#: +sctp = auto() +#: +outerip = auto() +#: +outerudp = auto() + +def __str__(self): +"""String method for use in csum_set_hw.""" +if self == ChecksumOffloadOptions.outerip: +return "outer-ip" +elif self == ChecksumOffloadOptions.outerudp: +return "outer-udp" +else: +return f"{self.name}" + + class DeviceErrorHandlingMode(StrEnum): """Enum representing the device error handling mode.""" @@ -806,6 +832,104 @@ def show_port_stats(self, port_id: int) -> TestPmdPortStats: return TestPmdPortStats.parse(output) +def port_stop(self, port: int, verify: bool = True): +"""Stop specified port. + +Args: +port: Specifies the port number to use, must be between 0-32. +verify: If :data:`True`, the output of the command is scanned +to ensure specified port is stopped. If not, it is considered +an error. + +Raises: +InteractiveCommandExecutionError: If `verify` is :data:`True` and the port +is not stopped. +""" +port_output = self.send_command(f"port stop {port}") +if verify: +if "Done" not in port_output: +self._logger.debug(f"Failed to stop port {port}: \n{port_output}") +raise InteractiveCommandExecutionError(f"Testpmd failed to stop port {port}.") + +def port_start(self, port: int, verify: bool = True): +"""Start specified port. + +Args: +port: Specifies the port number to use, must be between 0-32. +verify: If :data:`True`, the output of the command is scanned +to ensure specified port is started. If not, it is considered +an error. + +Raises: +InteractiveCommandExecutionError: If `verify` is :data:`True` and the port +is not started. +""" +port_output = self.send_command(f"port start {port}") +if verify: +if "Done" not in port_output: +self._logger.debug(f"Failed to start port {port}: \n{port_output}") +raise InteractiveCommandExecutionError(f"Testpmd failed to start port {port}.") + +def csum_set_hw(self, layer: ChecksumOffloadOptions, port_id: int, verify: bool = True) -> None: +"""Enables hardware checksum offloading on the specified layer. + +Args: +layer: The layer that checksum offloading should be enabled on. +options: tcp, ip, udp, sctp, outer-ip, outer-udp. +port_id: The port number to enable checksum offloading on, should be within 0-32. +verify: If :data:`True` the output of the command will be scanned in an attempt to +verify that checksum offloading was enabled on the port. + +Raises: +InteractiveCommandExecutionError: If checksum offload is not enabled successfully. +""" +csum_output = self.send_command(f"csum set {str(layer)} hw {port_id}") +if verify: +if "Bad arguments" in csum_output or f"Please stop port {port_id} first" in csum_output: +self._logger.debug(f"Failed to set csum hw mode on port {port_id}:\n{csum_output}") +raise InteractiveCommandExecutionError( +f"Failed to set csum hw mode on port {port_id}" +) +if verify and f"checksum offload is not supported by port {port_id}" in csum_output: +self._logger.debug(f"Checksum {layer} offload is not supported:\n{csum_output}") + +success = False +if layer == ChecksumOffloadOptions.outerip: +if "Outer-Ip checksum offload is hw" in csum_output: +success = True +elif layer == ChecksumOffloadOptions.outerudp: +if "Outer-Udp checksum offload is hw" in csum_output: +success = True +else: +if f"{str(layer).upper} checksum offload is hw" in csum_output: +success = True +if not success and verify: +self._logger.debug(f"Failed to set csum hw mode on port {port_id}:\n{
[PATCH v1 2/2] dts: checksum offload test suite
test suite for verifying layer 3/4 checksum offload features on poll mode driver. Depends-on: patch-143033 ("dts: add text parser for testpmd verbose output") Depends-on: patch-142691 ("dts: add send_packets to test suites and rework packet addressing") Signed-off-by: Dean Marx --- dts/tests/TestSuite_checksum_offload.py | 255 1 file changed, 255 insertions(+) create mode 100644 dts/tests/TestSuite_checksum_offload.py diff --git a/dts/tests/TestSuite_checksum_offload.py b/dts/tests/TestSuite_checksum_offload.py new file mode 100644 index 00..b327261ff3 --- /dev/null +++ b/dts/tests/TestSuite_checksum_offload.py @@ -0,0 +1,255 @@ +# SPDX-License-Identifier: BSD-3-Clause +# Copyright(c) 2024 University of New Hampshire + +"""DPDK checksum offload testing suite. + +This suite verifies L3/L4 checksum offload features of the Poll Mode Driver. +On the Rx side, IPv4 and UDP/TCP checksum by hardware is checked to ensure +checksum flags match expected flags. On the Tx side, IPv4/UDP, IPv4/TCP, +IPv6/UDP, and IPv6/TCP insertion by hardware is checked to checksum flags +match expected flags. + +""" + +from typing import List + +from scapy.all import Packet # type: ignore[import-untyped] +from scapy.layers.inet import IP, TCP, UDP # type: ignore[import-untyped] +from scapy.layers.inet6 import IPv6 # type: ignore[import-untyped] +from scapy.layers.sctp import SCTP # type: ignore[import-untyped] +from scapy.layers.l2 import Dot1Q # type: ignore[import-untyped] +from scapy.layers.l2 import Ether +from scapy.packet import Raw # type: ignore[import-untyped] + +from framework.remote_session.testpmd_shell import ( +SimpleForwardingModes, +TestPmdShell, +VerboseOLFlag, +ChecksumOffloadOptions +) +from framework.test_suite import TestSuite + + +class TestChecksumOffload(TestSuite): +"""Checksum offload test suite. + +This suite consists of 8 test cases: +1. Insert checksum on transmit packet +2. Do not insert checksum on transmit packet +3. Validate Rx checksum valid flags +4. Hardware checksum check L4 Rx +5. Hardware checksum check L3 Rx +6. Checksum offload with vlan + +""" + +def set_up_suite(self) -> None: +"""Set up the test suite. + +Setup: +Verify that at least two port links are created when the +test run is initialized. +""" +self.verify(len(self._port_links) > 1, "Not enough port links.") + +def send_packet_and_verify( +self, packet_list: List[Packet], load: str, should_receive: bool +) -> None: +"""Send and verify packet is received on the traffic generator. + +Args: +packet_list: list of Scapy packets to send and verify. +load: Raw layer load attribute in the sent packet. +should_receive: Indicates whether the packet should be received +by the traffic generator. +""" +for i in range(0, len(packet_list)): +received_packets = self.send_packet_and_capture(packet=packet_list[i]) +received = any( +packet.haslayer(Raw) and load in str(packet.load) for packet in received_packets +) +self.verify( +received == should_receive, +f"Packet was {'dropped' if should_receive else 'received'}", +) + +def send_packet_and_verify_checksum( +self, packet: Packet, goodL4: bool, goodIP: bool, testpmd: TestPmdShell +) -> None: +"""Send packet and verify verbose output matches expected output. + +Args: +packet: Scapy packet to send to DUT. +goodL4: Verifies RTE_MBUF_F_RX_L4_CKSUM_GOOD in verbose output +if :data:`True`, or RTE_MBUF_F_RX_L4_CKSUM_UNKNOWN if :data:`False`. +goodIP: Verifies RTE_MBUF_F_RX_IP_CKSUM_GOOD in verbose output +if :data:`True`, or RTE_MBUF_F_RX_IP_CKSUM_UNKNOWN if :data:`False`. +testpmd: Testpmd shell session to analyze verbose output of. +""" +testpmd.start() +self.send_packet_and_capture(packet=packet) +verbose_output = testpmd.extract_verbose_output(testpmd.stop()) +isL4 = any( +VerboseOLFlag.RTE_MBUF_F_RX_L4_CKSUM_GOOD in packet.ol_flags +for index in verbose_output +for packet in index.packets +) +isIP = any( +VerboseOLFlag.RTE_MBUF_F_RX_IP_CKSUM_GOOD in packet.ol_flags +for index in verbose_output +for packet in index.packets +) +self.verify(isL4 == goodL4, "Layer 4 checksum flag did not match expected checksum flag.") +self.verify(isIP == goodIP, "IP checksum flag did not match expected checksum flag.") + +def setup_hw_offload(self, testpmd: TestPmdShell) -> None: +"""Sets IP, UDP, TCP, and SCTP layers to hardware offload.""" +testpmd.port_stop(port=0) +t
[PATCH v4 3/5] graph: add stats for node specific errors
From: Pavan Nikhilesh Add support for retrieving/printing stats for node specific errors using rte_graph_cluster_stats_get(). Signed-off-by: Pavan Nikhilesh --- lib/graph/graph_stats.c | 79 - lib/graph/rte_graph.h | 4 +++ 2 files changed, 82 insertions(+), 1 deletion(-) diff --git a/lib/graph/graph_stats.c b/lib/graph/graph_stats.c index d71451a17b..1ac0a6797f 100644 --- a/lib/graph/graph_stats.c +++ b/lib/graph/graph_stats.c @@ -121,6 +121,25 @@ print_node(FILE *f, const struct rte_graph_cluster_node_stats *stat, bool dispat } } +static inline void +print_err(FILE *f, const struct rte_graph_cluster_node_stats *stat, bool dispatch) +{ + int i; + + if (dispatch) { + for (i = 0; i < stat->node_error_cntrs; i++) + fprintf(f, + "|\t%-24s|%15s|%-15" PRIu64 "|%15s|%15s|%15s|%15s|%15s|%11.4s|\n", + stat->node_error_desc[i], "", stat->node_error_count[i], "", "", "", + "", "", ""); + } else { + for (i = 0; i < stat->node_error_cntrs; i++) + fprintf(f, "|\t%-24s|%15s|%-15" PRIu64 "|%15s|%15.3s|%15.6s|%11.4s|\n", + stat->node_error_desc[i], "", stat->node_error_count[i], "", "", "", + ""); + } +} + static int graph_cluster_stats_cb(bool dispatch, bool is_first, bool is_last, void *cookie, const struct rte_graph_cluster_node_stats *stat) @@ -129,8 +148,11 @@ graph_cluster_stats_cb(bool dispatch, bool is_first, bool is_last, void *cookie, if (unlikely(is_first)) print_banner(f, dispatch); - if (stat->objs) + if (stat->objs) { print_node(f, stat, dispatch); + if (stat->node_error_cntrs) + print_err(f, stat, dispatch); + } if (unlikely(is_last)) { if (dispatch) boarder_model_dispatch(); @@ -203,6 +225,7 @@ stats_mem_populate(struct rte_graph_cluster_stats **stats_in, struct cluster_node *cluster; struct rte_node *node; rte_node_t count; + uint8_t i; cluster = stats->clusters; @@ -240,6 +263,36 @@ stats_mem_populate(struct rte_graph_cluster_stats **stats_in, SET_ERR_JMP(ENOENT, free, "Failed to find node %s in graph %s", graph_node->node->name, graph->name); cluster->nodes[cluster->nb_nodes++] = node; + if (graph_node->node->errs) { + cluster->stat.node_error_cntrs = graph_node->node->errs->nb_errors; + cluster->stat.node_error_count = rte_zmalloc_socket( + NULL, sizeof(uint64_t) * graph_node->node->errs->nb_errors, + RTE_CACHE_LINE_SIZE, stats->socket_id); + if (cluster->stat.node_error_count == NULL) + SET_ERR_JMP(ENOMEM, free, "Failed to allocate memory node %s graph %s", + graph_node->node->name, graph->name); + + cluster->stat.node_error_desc = rte_zmalloc_socket( + NULL, sizeof(RTE_NODE_ERROR_DESC_SIZE) * graph_node->node->errs->nb_errors, + RTE_CACHE_LINE_SIZE, stats->socket_id); + if (cluster->stat.node_error_desc == NULL) { + rte_free(cluster->stat.node_error_count); + SET_ERR_JMP(ENOMEM, free, "Failed to allocate memory node %s graph %s", + graph_node->node->name, graph->name); + } + + for (i = 0; i < cluster->stat.node_error_cntrs; i++) { + if (rte_strscpy(cluster->stat.node_error_desc[i], + graph_node->node->errs->err_desc[i], + RTE_NODE_ERROR_DESC_SIZE) < 0) { + rte_free(cluster->stat.node_error_count); + rte_free(cluster->stat.node_error_desc); + SET_ERR_JMP(E2BIG, free, + "Error description overflow node %s graph %s", + graph_node->node->name, graph->name); + } + } + } stats->sz += stats->cluster_node_size; stats->max_nodes++; @@ -388,6 +441,18 @@ rte_graph_cluster_stats_create(const struct rte_graph_cluster_stats_param *prm) void rte_graph_cluster_stats_destroy(struct rte_graph_cluster_stats *stat) { + struct cluster_node *cluster; + rte_node_t count; + + cluster = stat->clusters; + for (count = 0; count < stat->max_nodes; count++) { + if (cluster->stat.node_error_cntrs) { + rte_free(cluster->stat.node_error_count); +
[PATCH v4 2/5] graph: add node fastpath error counters
From: Pavan Nikhilesh Add node fastpath error counters advertised during node registration. Signed-off-by: Pavan Nikhilesh --- lib/graph/graph_populate.c | 20 ++-- lib/graph/graph_private.h | 2 ++ lib/graph/rte_graph_worker_common.h | 1 + 3 files changed, 21 insertions(+), 2 deletions(-) diff --git a/lib/graph/graph_populate.c b/lib/graph/graph_populate.c index ed596a7711..9c4c919e14 100644 --- a/lib/graph/graph_populate.c +++ b/lib/graph/graph_populate.c @@ -39,6 +39,15 @@ graph_fp_mem_calc_size(struct graph *graph) /* Pointer to next nodes(edges) */ sz += sizeof(struct rte_node *) * graph_node->node->nb_edges; } + sz = RTE_ALIGN(sz, RTE_CACHE_LINE_SIZE); + graph->errors_start = sz; + /* For 0..N node objects with error counters */ + STAILQ_FOREACH(graph_node, &graph->node_list, next) { + if (graph_node->node->errs == NULL) + continue; + sz = RTE_ALIGN(sz, RTE_CACHE_LINE_SIZE); + sz += sizeof(uint64_t) * graph_node->node->errs->nb_errors; + } graph->mem_sz = sz; return sz; @@ -64,6 +73,7 @@ graph_header_popluate(struct graph *_graph) static void graph_nodes_populate(struct graph *_graph) { + rte_graph_off_t err_off = _graph->errors_start; rte_graph_off_t off = _graph->nodes_start; struct rte_graph *graph = _graph->graph; struct graph_node *graph_node; @@ -99,6 +109,12 @@ graph_nodes_populate(struct graph *_graph) ->adjacency_list[count] ->node->name[0]; + if (graph_node->node->errs != NULL) { + node->err_off = err_off - off; + err_off += sizeof(uint64_t) * graph_node->node->errs->nb_errors; + err_off = RTE_ALIGN(err_off, RTE_CACHE_LINE_SIZE); + } + off += sizeof(struct rte_node *) * nb_edges; off = RTE_ALIGN(off, RTE_CACHE_LINE_SIZE); node->next = off; @@ -158,7 +174,7 @@ graph_node_nexts_populate(struct graph *_graph) } static int -graph_src_nodes_populate(struct graph *_graph) +graph_src_nodes_offset_populate(struct graph *_graph) { struct rte_graph *graph = _graph->graph; struct graph_node *graph_node; @@ -193,7 +209,7 @@ graph_fp_mem_populate(struct graph *graph) graph_pcap_init(graph); graph_nodes_populate(graph); rc = graph_node_nexts_populate(graph); - rc |= graph_src_nodes_populate(graph); + rc |= graph_src_nodes_offset_populate(graph); return rc; } diff --git a/lib/graph/graph_private.h b/lib/graph/graph_private.h index e663b04d8b..01921b254c 100644 --- a/lib/graph/graph_private.h +++ b/lib/graph/graph_private.h @@ -103,6 +103,8 @@ struct graph { /**< Memzone to store graph data. */ rte_graph_off_t nodes_start; /**< Node memory start offset in graph reel. */ + rte_graph_off_t errors_start; + /**< Node error memory start offset in graph reel. */ rte_node_t src_node_count; /**< Number of source nodes in a graph. */ struct rte_graph *graph; diff --git a/lib/graph/rte_graph_worker_common.h b/lib/graph/rte_graph_worker_common.h index 36d864e2c1..fa59d40f57 100644 --- a/lib/graph/rte_graph_worker_common.h +++ b/lib/graph/rte_graph_worker_common.h @@ -110,6 +110,7 @@ struct __rte_cache_aligned rte_node { uint64_t total_sched_fail; /**< Number of scheduled failure. */ } dispatch; }; + rte_graph_off_t err_off; /**< Offset to error counters. */ /* Fast path area */ #define RTE_NODE_CTX_SZ 16 alignas(RTE_CACHE_LINE_SIZE) uint8_t ctx[RTE_NODE_CTX_SZ]; /**< Node Context. */ -- 2.25.1
[PATCH v4 4/5] node: add error stats for ip4 lookup node
From: Pavan Nikhilesh Add error counters for ip4 LPM lookup failures in ip4_lookup node. Signed-off-by: Pavan Nikhilesh --- lib/node/ip4_lookup.c | 9 + lib/node/ip4_lookup_neon.h | 5 + lib/node/ip4_lookup_sse.h | 6 ++ lib/node/node_private.h| 8 4 files changed, 28 insertions(+) diff --git a/lib/node/ip4_lookup.c b/lib/node/ip4_lookup.c index 18955971f6..5a7921db75 100644 --- a/lib/node/ip4_lookup.c +++ b/lib/node/ip4_lookup.c @@ -86,6 +86,7 @@ ip4_lookup_node_process_scalar(struct rte_graph *graph, struct rte_node *node, rc = rte_lpm_lookup(lpm, rte_be_to_cpu_32(ipv4_hdr->dst_addr), &next_hop); next_hop = (rc == 0) ? next_hop : drop_nh; + NODE_INCREMENT_ERROR_ID(node, 0, (rc != 0), 1); node_mbuf_priv1(mbuf, dyn)->nh = (uint16_t)next_hop; next_hop = next_hop >> 16; @@ -219,11 +220,19 @@ ip4_lookup_node_init(const struct rte_graph *graph, struct rte_node *node) return 0; } +static struct rte_node_errors ip4_lookup_errors = { + .nb_errors = 1, + .err_desc = { + [0] = "ip4_lookup_error", + }, +}; + static struct rte_node_register ip4_lookup_node = { .process = ip4_lookup_node_process_scalar, .name = "ip4_lookup", .init = ip4_lookup_node_init, + .errs = &ip4_lookup_errors, .nb_edges = RTE_NODE_IP4_LOOKUP_NEXT_PKT_DROP + 1, .next_nodes = { diff --git a/lib/node/ip4_lookup_neon.h b/lib/node/ip4_lookup_neon.h index d5c8da3719..907c7c955a 100644 --- a/lib/node/ip4_lookup_neon.h +++ b/lib/node/ip4_lookup_neon.h @@ -116,6 +116,10 @@ ip4_lookup_node_process_vec(struct rte_graph *graph, struct rte_node *node, priv01.u16[4] = result.u16[2]; priv23.u16[0] = result.u16[4]; priv23.u16[4] = result.u16[6]; + NODE_INCREMENT_ERROR_ID(node, 0, result.u16[1] == (drop_nh >> 16), 1); + NODE_INCREMENT_ERROR_ID(node, 0, result.u16[3] == (drop_nh >> 16), 1); + NODE_INCREMENT_ERROR_ID(node, 0, result.u16[5] == (drop_nh >> 16), 1); + NODE_INCREMENT_ERROR_ID(node, 0, result.u16[7] == (drop_nh >> 16), 1); node_mbuf_priv1(mbuf0, dyn)->u = priv01.u64[0]; node_mbuf_priv1(mbuf1, dyn)->u = priv01.u64[1]; @@ -202,6 +206,7 @@ ip4_lookup_node_process_vec(struct rte_graph *graph, struct rte_node *node, &next_hop); next_hop = (rc == 0) ? next_hop : drop_nh; + NODE_INCREMENT_ERROR_ID(node, 0, (rc != 0), 1); node_mbuf_priv1(mbuf0, dyn)->nh = (uint16_t)next_hop; next_hop = next_hop >> 16; next0 = (uint16_t)next_hop; diff --git a/lib/node/ip4_lookup_sse.h b/lib/node/ip4_lookup_sse.h index 74dbf97533..a38131e629 100644 --- a/lib/node/ip4_lookup_sse.h +++ b/lib/node/ip4_lookup_sse.h @@ -115,6 +115,11 @@ ip4_lookup_node_process_vec(struct rte_graph *graph, struct rte_node *node, /* Perform LPM lookup to get NH and next node */ rte_lpm_lookupx4(lpm, dip, dst.u32, drop_nh); + NODE_INCREMENT_ERROR_ID(node, 0, dst.u16[1] == (drop_nh >> 16), 1); + NODE_INCREMENT_ERROR_ID(node, 0, dst.u16[3] == (drop_nh >> 16), 1); + NODE_INCREMENT_ERROR_ID(node, 0, dst.u16[5] == (drop_nh >> 16), 1); + NODE_INCREMENT_ERROR_ID(node, 0, dst.u16[7] == (drop_nh >> 16), 1); + /* Extract next node id and NH */ node_mbuf_priv1(mbuf0, dyn)->nh = dst.u32[0] & 0x; next0 = (dst.u32[0] >> 16); @@ -206,6 +211,7 @@ ip4_lookup_node_process_vec(struct rte_graph *graph, struct rte_node *node, rc = rte_lpm_lookup(lpm, rte_be_to_cpu_32(ipv4_hdr->dst_addr), &next_hop); next_hop = (rc == 0) ? next_hop : drop_nh; + NODE_INCREMENT_ERROR_ID(node, 0, rc != 0, 1); node_mbuf_priv1(mbuf0, dyn)->nh = next_hop & 0x; next0 = (next_hop >> 16); diff --git a/lib/node/node_private.h b/lib/node/node_private.h index 1de7306792..36b2a733db 100644 --- a/lib/node/node_private.h +++ b/lib/node/node_private.h @@ -12,6 +12,8 @@ #include #include +#include + extern int rte_node_logtype; #define RTE_LOGTYPE_NODE rte_node_logtype @@ -88,4 +90,10 @@ node_mbuf_priv2(struct rte_mbuf *m) return (struct node_mbuf_priv2 *)rte_mbuf_to_priv(m); } +#define NODE_INCREMENT_ERROR_ID(node, id, cond, cnt) \ + { \ + if (unlikely(rte_graph_has_stats_feature() && (cond))) \ + ((uint64_t *)RTE_PTR_ADD(node, node->err_off))[id] += (cnt);
[PATCH v4 5/5] node: add error stats for ip4 reassembly node
From: Pavan Nikhilesh Add reassembly failure error counter for ip4 reassembly node. Signed-off-by: Pavan Nikhilesh --- lib/node/ip4_reassembly.c | 9 + 1 file changed, 9 insertions(+) diff --git a/lib/node/ip4_reassembly.c b/lib/node/ip4_reassembly.c index 04823cc596..ab71ef1331 100644 --- a/lib/node/ip4_reassembly.c +++ b/lib/node/ip4_reassembly.c @@ -120,6 +120,7 @@ ip4_reassembly_node_process(struct rte_graph *graph, struct rte_node *node, void rte_node_next_stream_put(graph, node, RTE_NODE_IP4_REASSEMBLY_NEXT_PKT_DROP, dr->cnt); idx += dr->cnt; + NODE_INCREMENT_ERROR_ID(node, 0, dr->cnt, dr->cnt); dr->cnt = 0; } @@ -165,11 +166,19 @@ ip4_reassembly_node_init(const struct rte_graph *graph, struct rte_node *node) return 0; } +static struct rte_node_errors ip4_reassembly_errors = { + .nb_errors = 1, + .err_desc = { + [0] = "ip4_reassembly_error", + }, +}; + static struct rte_node_register ip4_reassembly_node = { .process = ip4_reassembly_node_process, .name = "ip4_reassembly", .init = ip4_reassembly_node_init, + .errs = &ip4_reassembly_errors, .nb_edges = RTE_NODE_IP4_REASSEMBLY_NEXT_PKT_DROP + 1, .next_nodes = { -- 2.25.1
[DPDK/cryptodev Bug 1523] cryptodev_openssl_asym_autotest unit test fails on OpenSUSE Leap 15.6
https://bugs.dpdk.org/show_bug.cgi?id=1523 Bug ID: 1523 Summary: cryptodev_openssl_asym_autotest unit test fails on OpenSUSE Leap 15.6 Product: DPDK Version: unspecified Hardware: All OS: All Status: UNCONFIRMED Severity: normal Priority: Normal Component: cryptodev Assignee: dev@dpdk.org Reporter: cch...@iol.unh.edu Target Milestone: --- Created attachment 285 --> https://bugs.dpdk.org/attachment.cgi?id=285&action=edit DPDK 23.11 compile and unit test log On OpenSUSE Leap 15.6, the cryptodev_openssl_asym_autotest unit test is failing with DPDK 24.11 and 23.11 and the cryptodev_openssl_autotest unit test is failing with DPDK 21.11. This fail appeared after our container images were updated from OpenSUSE Leap 15.5 (which had openssl 1.1.1l) to 15.6 (which has openssl 3.1.4). GCC version: gcc (SUSE Linux) 7.5.0 OpenSSL version: 3.1.4 Because I can only add one attachment, I have attached the logs for DPDK 23.11 here and have uploaded the logs for DPDK 21.11 on Google Drive here: https://drive.google.com/file/d/1tKVYWIN82Mjc0PGGLZ_pNuiqG89uc6JQ/view. -- You are receiving this mail because: You are the assignee for the bug.
Re: [dpdk-dev] [PATCH v2 00/41] add MLX5 FreeBSD support
On Fri, 8 Oct 2021 00:13:09 +0530 Srikanth Kaka wrote: > This patch series introduces FreeBSD OS support for MLX5 PMD > > Currently there is an ongoing review of FreeBSD OFED drivers and > RDMA-core libraries. Their status can be tracked at > https://reviews.freebsd.org/p/vag.singh_oneconvergence.com > > Only MLX5 SRIOV interface is supported along with the following features: > > - Hypervisors: KVM, ESXi and Azure/HyperV > - Multiple Tx and Rx queues > - Support for scattered Tx and Rx frames > - RSS > - VLAN filtering, stripping and insertion > - RX CRC stripping configuration > - Promiscuous and Multicast mode > - Statistics query including Basic, Extended, and per queue. > - Configurable MTU and MAC > - Jumbo frame support > - Trust mode support > - Vxlan > - QoS > - Flow steering > > Tested on : > MT27710 Family [ConnectX-4 Lx] > MT27710 Family [ConnectX-4 Lx Virtual Function] (Azure) > MT27700 Family [ConnectX-4] > MT27800 Family [ConnectX-5] > MT2894 Family [ConnectX-6 Lx] > MT28908 Family [ConnectX-6] > MT2892 Family [ConnectX-6 Dx] > > v1 : Initial submission > v2 : Addressed community comments Not obvious why this patch set never got merged. Most likely because it required upstream FreeBSD libraries that were not merged at the time. Is this still of interest? Has the support for RDMA and IB been merged in upstream FreeBSD? Does the meson build handle the case where the required dependencies are not present in the version of FreeBSD being used? If nothing else, the driver needs to be resubmitted to address 3 years of changes to mlx5 common code. And will need an entry in MAINTAINERS.