This is what we measure on CSIT with Intel Cascade Lake for 1 CPU core: http://csit.fd.io/trending/#eNrVlMFuwyAMhp8mvVSWwC3JduihXd6jYuAu0UjKgFbtnn6kyuRE2ibtUm0HjNBn-8e_EDEdA-0juU2hdkW1K7BqbQ7FarvM29l7wB6Mu4AU4oXQS3NRqtQQbKeh9Wso188gDVBqhlNe0WhHKF4h9Ba6EIZu-DR0s6c0a83EN1cm3wtygQ6kuSKrMkoUJzJf34uzD0F3FNt34pI8EXOT_WEkzVwnXf2EjsNW9S3jt2Y6nHiZT14n09zJQIc_-Pd5l79iGj1IYd5Anw_3eYCs92_en6oXMemQNigQQaxAlkshctoYFtTbkT2CxDnrj6G7fQeq_gAVCBnd - red line is Mellanox Cx5 l2 patch (ie max perf we can sustain for this HW setup) and yields ~42Mpps - back line is Mellanox Cx5 IPv4 forwarding with 20k prefixes and random traffic (should be close to your case I suppose) and yields ~15Mpps - blue line is Intel Columbiaville IPv4 forwarding with 20k prefixes and random traffic (for reference) and yields ~18Mpps
Your results seem to be close to what we see *for a single core*. So now I'm wondering, is your traffic distributed on all the cores? To go further, can you share the output of: ~# vppctl clear run && sleep 1 && vppctl show run While the traffic is going through? That should show us where CPU cycles are spent and how traffic distribution looks like. Best ben > -----Original Message----- > From: vpp-dev@lists.fd.io <vpp-dev@lists.fd.io> On Behalf Of Pawel > Staszewski > Sent: Sunday, September 11, 2022 22:34 > To: vpp-dev@lists.fd.io > Subject: [vpp-dev] mellanox mlx5 + rdma + lcpng + bond - performance > (tuning ? or just FIB/RIB processing limit) > > Hi > > First I want to thanks to all ppl that created RDMA native support in > vpp and also ppl from LCPNG / Linux-CP - it is working and looks stable :) > > > But > > > Was testing some scenarions with rdma+vpp+lcpng+frr BGP with 200k routes > - with mellanox mlx5 2x100G nics - where 24 cores are used for RX queue > - and vpp config has cpu configuration where we are using 25(vpp only) > cores for all workers (controll+24 workers and 8 cores left for frr > software routes and other stuff ) > > We are using isolcpus where last 24 cores are isolated with nohz nocb > etc... > > > Hardware used is : > > Intel 6246R > > 96GB ram > > Mellanox Connect-X 5 2x 100GB Ethernet NIC > > > -------- > > And the problem is that we can reach only like 7Mpps per port (2 ports > are bonded) - so in total 14Mpps for two ports in bonding > > Bonding is done via vppctl > > bond is reporting drops in stats - but this stats shows "no error" > > Example drops from bond: > > vpp# show interface > Name Idx State MTU > (L3/IP4/IP6/MPLS) Counter Count > BondEthernet0 5 up 1500/0/0/0 rx > packets 16140265734 > rx > bytes 14149099692091 > tx > packets 16136936492 > tx > bytes 14148733722122 > drops 3451254 > punt 1326 > tx-error 1 > > > And show errors: > > vpp# show errors > Count Node Reason Severity > 16 ip6-icmp-input neighbor solicitations for > unknown error > 952172 bond-input no > error error > 1 null-node blackholed > packets error > 2 lacp-input good lacp packets -- > cache hit error > 24 arp-reply ARP replies > sent info > 9 ip4-input ip4 ttl <= > 1 error > 32 ip6-icmp-input neighbor solicitations for > unknown error > 1 ip6-icmp-input neighbor > advertisements sent info > 9 ip4-icmp-error hop limit exceeded > response sent info > 1772997 bond-input no > error error > 2 bond-input pass through (CDP, LLDP, > slow proto error > 140 snap-input unknown oui/snap > protocol error > 180 ethernet-input no > error error > 1 ethernet-input unknown ethernet > type error > 61 ethernet-input unknown > vlan error > 1 null-node blackholed > packets error > 12 ip4-input ip4 ttl <= > 1 error > 1 ip6-icmp-input neighbor solicitations > from source error > 30 ip6-icmp-input neighbor solicitations for > unknown error > 12 ip4-icmp-error hop limit exceeded > response sent info > 1853434 bond-input no > error error > 1 ethernet-input unknown > vlan error > 1 ip4-input ip4 ttl <= > 1 error > 1 ip4-local ip4 source lookup > miss error > 30 ip6-icmp-input neighbor solicitations for > unknown error > 1 ip6-icmp-input neighbor > advertisements sent info > 1 ip4-icmp-error hop limit exceeded > response sent info > 1864191 bond-input no > error error > 1 null-node blackholed > packets error > 5 ip4-input ip4 ttl <= > 1 error > 35 ip6-icmp-input neighbor solicitations for > unknown error > 5 ip4-icmp-error hop limit exceeded > response sent info > 1743015 bond-input no > error error > 3 ethernet-input unknown > vlan error > 1 null-node blackholed > packets error > 4 ip4-input ip4 ttl <= > 1 error > 1 ip6-icmp-input neighbor solicitations > from source error > 35 ip6-icmp-input neighbor solicitations for > unknown error > 4 ip4-icmp-error hop limit exceeded > response sent info > 1917745 bond-input no > error error > 2 ip4-input ip4 ttl <= > 1 error > 18 ip6-icmp-input neighbor solicitations for > unknown error > 2 ip4-icmp-error hop limit exceeded > response sent info > 886249 bond-input no > error error > 1 ethernet-input unknown > vlan error > 7 ip4-input ip4 ttl <= > 1 error > 1 ip6-icmp-input valid > packets info > 14 ip6-icmp-input neighbor solicitations for > unknown error > 1 ip6-icmp-input router advertisements > received info > > > This counter : > > 1743015 bond-input no > error error > > Is wild :) and it increasing really fast - and also simple icmp ping to > this host or from this host is reporting froms from 2 to 5% - when both > 100Gnics are receiving and transmitting 7Mpps - per port ( so in total > 14Mpps RX and 14Mpps TX) > > Basically it looks like something cant handle more - and probabbly > FIB/RIB processing is the case here - because without forwarding and > frr+routes we was able to reach like 40Mpps per 100G port > > > Bonded nics: (m0rdma0+m1rdma1 - created by lcp from two physical mlx > ports as RDMA pairs) - create commands: > > create int rdma host-if ens1f1np1 name m1rdma1 num-rx-queues 12 > create int rdma host-if ens1f0np0 name m0rdma0 num-rx-queues 12 > > create bond mode lacp load-balance l34 > bond add BondEthernet0 m0rdma0 > bond add BondEthernet0 m1rdma1 > > > Nics attached via rdma: > > m0rdma0 4 up m0rdma0 > Link speed: 100 Gbps > RX Queues: > queue thread mode > 0 vpp_wk_13 (14) polling > 1 vpp_wk_14 (15) polling > 2 vpp_wk_15 (16) polling > 3 vpp_wk_16 (17) polling > 4 vpp_wk_17 (18) polling > 5 vpp_wk_18 (19) polling > 6 vpp_wk_19 (20) polling > 7 vpp_wk_20 (21) polling > 8 vpp_wk_21 (22) polling > 9 vpp_wk_22 (23) polling > 10 vpp_wk_23 (24) polling > 11 vpp_wk_0 (1) polling > Ethernet address 98:03:9b:67:f6:1e > netdev ens1f0np0 pci-addr 0000:b3:00.0 > product name: CX516A - ConnectX-5 QSFP28 > part number: MCX516A-CCAT > revision: B2 > serial number: MT2043K02404 > flags: admin-up mlx5dv > rss: ipv4-tcp ipv6-tcp > mlx5: version 0 > device flags: cqe-v1 mpw-allowed enhanced-mpw cqe-128b-comp > cqe-128b-pad > m1rdma1 3 up m1rdma1 > Link speed: 100 Gbps > RX Queues: > queue thread mode > 0 vpp_wk_1 (2) polling > 1 vpp_wk_2 (3) polling > 2 vpp_wk_3 (4) polling > 3 vpp_wk_4 (5) polling > 4 vpp_wk_5 (6) polling > 5 vpp_wk_6 (7) polling > 6 vpp_wk_7 (8) polling > 7 vpp_wk_8 (9) polling > 8 vpp_wk_9 (10) polling > 9 vpp_wk_10 (11) polling > 10 vpp_wk_11 (12) polling > 11 vpp_wk_12 (13) polling > Ethernet address 98:03:9b:67:f6:1e > netdev ens1f1np1 pci-addr 0000:b3:00.1 > product name: CX516A - ConnectX-5 QSFP28 > part number: MCX516A-CCAT > revision: B2 > serial number: MT2043K02404 > flags: admin-up mlx5dv > rss: ipv4-tcp ipv6-tcp > mlx5: version 0 > device flags: cqe-v1 mpw-allowed enhanced-mpw cqe-128b-comp > cqe-128b-pad > > .... > > .... > > Same NIC same Queue threads numbers (12 per card from 24 available cores > that are isolcpus - at kernel boot) > > > vpp# show interface rx-placement > Thread 1 (vpp_wk_0): > node rdma-input: > m0rdma0 queue 11 (polling) > node virtio-input: > tap1 queue 0 (polling) > Thread 2 (vpp_wk_1): > node rdma-input: > m1rdma1 queue 0 (polling) > node virtio-input: > tap5 queue 0 (polling) > Thread 3 (vpp_wk_2): > node rdma-input: > m1rdma1 queue 1 (polling) > Thread 4 (vpp_wk_3): > m1rdma1 queue 2 (polling) > Thread 5 (vpp_wk_4): > m1rdma1 queue 3 (polling) > Thread 6 (vpp_wk_5): > m1rdma1 queue 4 (polling) > Thread 7 (vpp_wk_6): > m1rdma1 queue 5 (polling) > Thread 8 (vpp_wk_7): > m1rdma1 queue 6 (polling) > Thread 9 (vpp_wk_8): > m1rdma1 queue 7 (polling) > Thread 10 (vpp_wk_9): > m1rdma1 queue 8 (polling) > Thread 11 (vpp_wk_10): > m1rdma1 queue 9 (polling) > Thread 12 (vpp_wk_11): > m1rdma1 queue 10 (polling) > Thread 13 (vpp_wk_12): > m1rdma1 queue 11 (polling) > Thread 14 (vpp_wk_13): > m0rdma0 queue 0 (polling) > Thread 15 (vpp_wk_14): > m0rdma0 queue 1 (polling) > Thread 16 (vpp_wk_15): > m0rdma0 queue 2 (polling) > Thread 17 (vpp_wk_16): > m0rdma0 queue 3 (polling) > Thread 18 (vpp_wk_17): > m0rdma0 queue 4 (polling) > Thread 19 (vpp_wk_18): > m0rdma0 queue 5 (polling) > Thread 20 (vpp_wk_19): > m0rdma0 queue 6 (polling) > Thread 21 (vpp_wk_20): > m0rdma0 queue 7 (polling) > Thread 22 (vpp_wk_21): > m0rdma0 queue 8 (polling) > Thread 23 (vpp_wk_22): > m0rdma0 queue 9 (polling) > Thread 24 (vpp_wk_23): > m0rdma0 queue 10 (polling) > > Below also some stats about memory used: > > vpp# show memory > show memory: Need one of api-segment, stats-segment, main-heap, > numa-heaps or map > vpp# show memory main-heap > Thread 0 vpp_main > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.37M, free: 1.64G, trimmable: 1.62G > > Thread 1 vpp_wk_0 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.37M, free: 1.64G, trimmable: 1.62G > > Thread 2 vpp_wk_1 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 3 vpp_wk_2 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 4 vpp_wk_3 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 5 vpp_wk_4 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 6 vpp_wk_5 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 7 vpp_wk_6 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 8 vpp_wk_7 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 9 vpp_wk_8 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 10 vpp_wk_9 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 11 vpp_wk_10 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 12 vpp_wk_11 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 13 vpp_wk_12 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 14 vpp_wk_13 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 15 vpp_wk_14 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 16 vpp_wk_15 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 17 vpp_wk_16 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 18 vpp_wk_17 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 19 vpp_wk_18 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 20 vpp_wk_19 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 21 vpp_wk_20 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 22 vpp_wk_21 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 23 vpp_wk_22 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 24 vpp_wk_23 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 25 vpp_wk_24 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 26 vpp_wk_25 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.38M, free: 1.64G, trimmable: 1.62G > > Thread 27 vpp_wk_26 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.39M, free: 1.64G, trimmable: 1.62G > vpp# show memory main-heap > Thread 0 vpp_main > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 1 vpp_wk_0 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 2 vpp_wk_1 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 3 vpp_wk_2 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 4 vpp_wk_3 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 5 vpp_wk_4 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 6 vpp_wk_5 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 7 vpp_wk_6 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 8 vpp_wk_7 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > vpp# show memory main-heap > Thread 0 vpp_main > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 1 vpp_wk_0 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 2 vpp_wk_1 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 3 vpp_wk_2 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 4 vpp_wk_3 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 5 vpp_wk_4 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 6 vpp_wk_5 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 7 vpp_wk_6 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > page stats: page-size 1G, total 2, mapped 2, not-mapped 0 > numa 0: 2 pages, 2g bytes > total: 1.99G, used: 369.58M, free: 1.64G, trimmable: 1.62G > > Thread 8 vpp_wk_7 > base 0x7f64c0000000, size 2g, locked, unmap-on-destroy, name 'main > heap' > vpp# show memory ? > show memory show memory > [api-segment][stats-segment][verbose] > [numa-heaps][map] > vpp# show memory > show memory: Need one of api-segment, stats-segment, main-heap, > numa-heaps or map > vpp# show memory numa-heaps > Numa 0 uses the main heap... > vpp# show memory map > > StartAddr size FD PageSz Pages Numa0 NotMap Name > 00007f64c0000000 2g 1G 2 2 0 main heap > 00007f655b663000 2m 4K 512 9 503 thread stack: > thread 0 > 00007f6440000000 1g 4 1G 1 1 0 stat segment > 00007f655bc9a000 32k 4K 8 2 6 process stack: > wg-timer-manager > 00007f655b642000 128k 4K 32 2 30 process stack: > vrrp-periodic-process > 00007f655bc7a000 32k 4K 8 1 7 process stack: > prom-scraper-process > 00007f655b639000 32k 4K 8 2 6 process stack: > nsh-md2-ioam-export-process > 00007f655b630000 32k 4K 8 2 6 process stack: > nat44-ei-ha-process > 00007f655b627000 32k 4K 8 2 6 process stack: > memif-process > 00007f655b61e000 32k 4K 8 2 6 process stack: > lldp-process > 00007f655b5fd000 128k 4K 32 3 29 process stack: > linux-cp-netlink-process > 00007f655b5f4000 32k 4K 8 2 6 process stack: > udp-ping-process > 00007f655b5eb000 32k 4K 8 2 6 process stack: > vxlan-gpe-ioam-export-process > 00007f655b5e2000 32k 4K 8 2 6 process stack: > ioam-export-process > 00007f655b5d9000 32k 4K 8 2 6 process stack: > ikev2-manager-process > 00007f655b5d0000 32k 4K 8 2 6 process stack: > igmp-timer-process > 00007f655b5c7000 32k 4K 8 1 7 process stack: > http-timer-process > 00007f655b5be000 32k 4K 8 2 6 process stack: > flowprobe-timer-process > 00007f655b59d000 128k 4K 32 2 30 process stack: > dpdk-process > 00007f655b57c000 128k 4K 32 2 30 process stack: > admin-up-down-process > 00007f655b573000 32k 4K 8 2 6 process stack: > send-dhcp6-pd-client-message-process > 00007f655b56a000 32k 4K 8 2 6 process stack: > dhcp6-pd-client-cp-process > 00007f655b561000 32k 4K 8 2 6 process stack: > dhcp6-client-cp-process > 00007f655b558000 32k 4K 8 2 6 process stack: > send-dhcp6-client-message-process > 00007f655b54f000 32k 4K 8 2 6 process stack: > dhcp6-pd-reply-publisher-process > 00007f655b546000 32k 4K 8 2 6 process stack: > dhcp6-reply-publisher-process > 00007f655b535000 64k 4K 16 2 14 process stack: > dhcp-client-process > 00007f655b52c000 32k 4K 8 2 6 process stack: > cnat-scanner-process > 00007f655b523000 32k 4K 8 2 6 process stack: > avf-process > 00007f655b51a000 32k 4K 8 2 6 process stack: > acl-plugin-fa-cleaner-process > 00007f655b4d9000 256k 4K 64 3 61 process stack: > api-rx-from-ring > 00007f655b4d0000 32k 4K 8 2 6 process stack: > rd-cp-process > 00007f655b4c7000 32k 4K 8 2 6 process stack: > ip6-ra-process > 00007f655b4be000 32k 4K 8 2 6 process stack: > ip6-rs-process > 00007f655b4b5000 32k 4K 8 2 6 process stack: > ip6-mld-process > 00007f655b4ac000 32k 4K 8 2 6 process stack: > fib-walk > 00007f655b4a3000 32k 4K 8 1 7 process stack: > session-queue-process > 00007f655b49a000 32k 4K 8 2 6 process stack: > vhost-user-process > 00007f655b491000 32k 4K 8 2 6 process stack: > vhost-user-send-interrupt-process > 00007f655b488000 32k 4K 8 2 6 process stack: > flow-report-process > 00007f655b47f000 32k 4K 8 2 6 process stack: > bfd-process > 00007f655b476000 32k 4K 8 2 6 process stack: > ip-neighbor-event > 00007f655b46d000 32k 4K 8 2 6 process stack: > ip6-neighbor-age-process > 00007f655b464000 32k 4K 8 2 6 process stack: > ip4-neighbor-age-process > 00007f655b45b000 32k 4K 8 2 6 process stack: > ip6-sv-reassembly-expire-walk > 00007f655b452000 32k 4K 8 2 6 process stack: > ip6-full-reassembly-expire-walk > 00007f655b449000 32k 4K 8 2 6 process stack: > ip4-sv-reassembly-expire-walk > 00007f655b440000 32k 4K 8 2 6 process stack: > ip4-full-reassembly-expire-walk > 00007f655b437000 32k 4K 8 2 6 process stack: > bond-process > 00007f655b42e000 32k 4K 8 2 6 process stack: > l2fib-mac-age-scanner-process > 00007f655b425000 32k 4K 8 2 6 process stack: > l2-arp-term-publisher > 00007f655b41c000 32k 4K 8 2 6 process stack: > vpe-link-state-process > 00007f655b3db000 256k 4K 64 5 59 process stack: > startup-config-process > 00007f655b3d2000 32k 4K 8 2 6 process stack: > statseg-collector-process > 00007f655a5a8000 2m 4K 512 7 505 thread stack: > thread 1 > 00007f655a3a7000 2m 4K 512 10 502 thread stack: > thread 2 > 00007f655a1a6000 2m 4K 512 10 502 thread stack: > thread 3 > 00007f6559fa5000 2m 4K 512 10 502 thread stack: > thread 4 > 00007f6559da4000 2m 4K 512 10 502 thread stack: > thread 5 > 00007f6559ba3000 2m 4K 512 10 502 thread stack: > thread 6 > 00007f65599a2000 2m 4K 512 10 502 thread stack: > thread 7 > 00007f65597a1000 2m 4K 512 11 501 thread stack: > thread 8 > 00007f65595a0000 2m 4K 512 10 502 thread stack: > thread 9 > 00007f655939f000 2m 4K 512 10 502 thread stack: > thread 10 > 00007f655919e000 2m 4K 512 10 502 thread stack: > thread 11 > 00007f6558f9d000 2m 4K 512 10 502 thread stack: > thread 12 > 00007f6558d9c000 2m 4K 512 10 502 thread stack: > thread 13 > 00007f6558b9b000 2m 4K 512 10 502 thread stack: > thread 14 > 00007f655899a000 2m 4K 512 10 502 thread stack: > thread 15 > 00007f6558799000 2m 4K 512 10 502 thread stack: > thread 16 > 00007f6558598000 2m 4K 512 10 502 thread stack: > thread 17 > 00007f6558397000 2m 4K 512 10 502 thread stack: > thread 18 > 00007f6558196000 2m 4K 512 10 502 thread stack: > thread 19 > 00007f6553e00000 2m 4K 512 10 502 thread stack: > thread 20 > 00007f6553bff000 2m 4K 512 10 502 thread stack: > thread 21 > 00007f65539fe000 2m 4K 512 10 502 thread stack: > thread 22 > 00007f65537fd000 2m 4K 512 10 502 thread stack: > thread 23 > 00007f65535fc000 2m 4K 512 10 502 thread stack: > thread 24 > 00007f65533fb000 2m 4K 512 10 502 thread stack: > thread 25 > 00007f65531fa000 2m 4K 512 10 502 thread stack: > thread 26 > 00007f6552ff9000 2m 4K 512 10 502 thread stack: > thread 27 > 00007f6552f98000 64k 4K 16 2 14 process stack: > lacp-process > 00007f6552f57000 256k 4K 64 3 61 process stack: > unix-cli-local:0 > 00007f6552f46000 64k 4K 16 2 14 process stack: > unix-cli-new-session > >
-=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#21868): https://lists.fd.io/g/vpp-dev/message/21868 Mute This Topic: https://lists.fd.io/mt/93619900/21656 Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/leave/1480452/21656/631435203/xyzzy [arch...@mail-archive.com] -=-=-=-=-=-=-=-=-=-=-=-