Hi

Yes - probabbly this tap interface is just only logical local only for ip_input/ip_output (local) traffic

I will share show run after I will have another test window in my lab :)

Currently just playing with some settings and checking what I can show for debug when I will test it again - because now it is a little hard to debug when there is no info how much load traffic is doing to cpu's (for example cpu cycles etc... something like perf top)



On 9/15/22 10:30 AM, Benoit Ganne (bganne) via lists.fd.io wrote:
Well, it depends how your traffic is flowing.
 From the output you shared, the rdma interface does have 8 rxq in poll-mode, 1 
per worker using 8 workers (side note: that means you're not going to use more 
than 8 cores to process packets from that interface, even if you dedicated 24 
cores to VPP). So if traffic is forwarded from/to this interfaces only, it 
should be using 8 cores and you should reach performance much higher than 
14Mpps.
The tap interface is created by lcp, and as far as I understand it, it is used as a 
"representor" for Linux but it should not be used for most traffic apart maybe 
some control-plane traffic. The fact that is uses only 1rxq (and hence 1 core) shouldn't 
be an issue.
If you want us to help on performance issue, please share the output of:
   ~# vppctl clear run && sleep 1 && vppctl show run
While the traffic is going through? That should show us where CPU cycles are 
spent and how traffic distribution looks like.

Best
ben

-----Original Message-----
From: vpp-dev@lists.fd.io <vpp-dev@lists.fd.io> On Behalf Of Pawel
Staszewski
Sent: Thursday, September 15, 2022 0:03
To: vpp-dev@lists.fd.io
Subject: Re: [vpp-dev] mellanox mlx5 + rdma + lcpng + bond - performance
(tuning ? or just FIB/RIB processing limit)

Hi




So hmm...

maybee there is some problem with rx/tx queue creation because after this
commands:

create interface rdma host-if enp59s0f0 name enp59s0f0-rdma rx-queue-size
4096 tx-queue-size 512 num-rx-queues 8 no-multi-seg
lcp create enp59s0f0-rdma host-if e0-rdma






When i do show for tap interface:Interface: tap3 (ifindex 5)
   name "e0-rdma"
   host-mtu-size "9000"
   host-mac-addr: 02:fe:ad:1c:47:da
   host-carrier-up: 1
   vhost-fds 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75
   tap-fds 58
   gso-enabled 0
   csum-enabled 0
   packet-coalesce 0
   packet-buffering 0
   Mac Address: 02:fe:57:90:b6:d7
   Device instance: 1
   flags 0x1
     admin-up (0)
   features 0x110008000
     VIRTIO_NET_F_MRG_RXBUF (15)
     VIRTIO_RING_F_INDIRECT_DESC (28)
     VIRTIO_F_VERSION_1 (32)
   remote-features 0x33d008000
     VIRTIO_NET_F_MRG_RXBUF (15)
     VIRTIO_F_NOTIFY_ON_EMPTY (24)
     VHOST_F_LOG_ALL (26)
     VIRTIO_F_ANY_LAYOUT (27)
     VIRTIO_RING_F_INDIRECT_DESC (28)
     VIRTIO_RING_F_EVENT_IDX (29)
     VIRTIO_F_VERSION_1 (32)
     VIRTIO_F_IOMMU_PLATFORM (33)
   Number of RX Virtqueue  1
   Number of TX Virtqueue  17






hmm

   Number of RX Virtqueue  1


Number of TX Virtqueue  17




Why ?




When

vpp# show hardware-interfaces enp59s0f0-rdma
               Name                Idx   Link  Hardware
enp59s0f0-rdma                     3     up   enp59s0f0-rdma
   Link speed: 100 Gbps
   RX Queues:
     queue thread         mode
     0     vpp_wk_1 (2)   polling
     1     vpp_wk_2 (3)   polling
     2     vpp_wk_3 (4)   polling
     3     vpp_wk_4 (5)   polling
     4     vpp_wk_5 (6)   polling
     5     vpp_wk_6 (7)   polling
     6     vpp_wk_7 (8)   polling
     7     vpp_wk_8 (9)   polling
   Ethernet address 0c:42:a1:a9:6c:a6
   netdev enp59s0f0 pci-addr 0000:3b:00.0
   product name: CX516A - ConnectX-5 QSFP28
   part number: MCX516A-CCAT
   revision: AC
   serial number: MT2024K11216
   flags: admin-up link-up mlx5dv
   rss: ipv4-tcp ipv6-tcp
   mlx5: version 0
     device flags:  cqe-v1 mpw-allowed enhanced-mpw cqe-128b-comp cqe-128b-
pad
vpp#







remember this is not dpdk - but I need to create rdma interfaces first

But it looks like there is some problems with tap interface queue assign











On 9/12/22 12:20 PM, Benoit Ganne (bganne) via lists.fd.io wrote:


        This is what we measure on CSIT with Intel Cascade Lake for 1 CPU
core:
        http://csit.fd.io/trending/#eNrVlMFuwyAMhp8mvVSWwC3JduihXd6jYuAu0Uj
KgFbtnn6kyuRE2ibtUm0HjNBn-8e_EDEdA-
0juU2hdkW1K7BqbQ7FarvM29l7wB6Mu4AU4oXQS3NRqtQQbKeh9Wso188gDVBqhlNe0WhHKF4h
9Ba6EIZu-
DR0s6c0a83EN1cm3wtygQ6kuSKrMkoUJzJf34uzD0F3FNt34pI8EXOT_WEkzVwnXf2EjsNW9S3
jt2Y6nHiZT14n09zJQIc_-
Pd5l79iGj1IYd5Anw_3eYCs92_en6oXMemQNigQQaxAlkshctoYFtTbkT2CxDnrj6G7fQeq_gA
VCBnd
         - red line is Mellanox Cx5 l2 patch (ie max perf we can sustain
for this HW setup) and yields ~42Mpps
         - back line is Mellanox Cx5 IPv4 forwarding with 20k prefixes and
random traffic (should be close to your case I suppose) and yields ~15Mpps
         - blue line is Intel Columbiaville IPv4 forwarding with 20k
prefixes and random traffic (for reference) and yields ~18Mpps

        Your results seem to be close to what we see *for a single core*.
So now I'm wondering, is your traffic distributed on all the cores?
        To go further, can you share the output of:
          ~# vppctl clear run && sleep 1 && vppctl show run
        While the traffic is going through? That should show us where CPU
cycles are spent and how traffic distribution looks like.

        Best
        ben


                -----Original Message-----
                From: vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>
<vpp-dev@lists.fd.io> <mailto:vpp-dev@lists.fd.io>  On Behalf Of Pawel
                Staszewski
                Sent: Sunday, September 11, 2022 22:34
                To: vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>
                Subject: [vpp-dev] mellanox mlx5 + rdma + lcpng + bond -
performance
                (tuning ? or just FIB/RIB processing limit)

                Hi

                First I want to thanks to all ppl that created RDMA native
support in
                vpp and also ppl from LCPNG / Linux-CP - it is working and
looks stable :)


                But


                Was testing some scenarions with rdma+vpp+lcpng+frr BGP with
200k routes
                - with mellanox mlx5 2x100G nics - where 24 cores are used
for RX queue
                - and vpp config has cpu configuration where we are using
25(vpp only)
                cores for all workers (controll+24 workers and 8 cores left
for frr
                software routes and other stuff )

                We are using isolcpus where last 24 cores are isolated with
nohz nocb
                etc...


                Hardware used is :

                Intel 6246R

                96GB ram

                Mellanox Connect-X 5 2x 100GB Ethernet NIC


                --------

                And the problem is that we can reach only like 7Mpps per
port (2 ports
                are bonded) - so in total 14Mpps for two ports in bonding

                Bonding is done via vppctl

                bond is reporting drops in stats - but this stats shows "no
error"

                Example drops from bond:

                vpp# show interface
                               Name               Idx    State  MTU
                (L3/IP4/IP6/MPLS)     Counter          Count
                BondEthernet0                     5      up 1500/0/0/0
rx
                packets           16140265734

rx
                bytes          14149099692091

tx
                packets           16136936492

tx
                bytes          14148733722122
                drops                    3451254
                punt                        1326
                tx-error                       1


                And show errors:

                vpp# show errors
                    Count                  Node Reason
Severity
                         16           ip6-icmp-input          neighbor
solicitations for
                unknown    error
                     952172             bond-input
no
                error                error
                          1             null-node
blackholed
                packets           error
                          2             lacp-input               good lacp
packets --
                cache hit     error
                         24             arp-reply                       ARP
replies
                sent            info
                          9             ip4-input
ip4 ttl <=
                1              error
                         32           ip6-icmp-input          neighbor
solicitations for
                unknown    error
                          1           ip6-icmp-input              neighbor
                advertisements sent      info
                          9           ip4-icmp-error            hop limit
exceeded
                response sent    info
                    1772997             bond-input
no
                error                error
                          2             bond-input            pass through
(CDP, LLDP,
                slow proto   error
                        140             snap-input                 unknown
oui/snap
                protocol        error
                        180           ethernet-input
no
                error                error
                          1           ethernet-input                 unknown
ethernet
                type          error
                         61           ethernet-input
unknown
                vlan              error
                          1             null-node
blackholed
                packets           error
                         12             ip4-input
ip4 ttl <=
                1              error
                          1           ip6-icmp-input          neighbor
solicitations
                from source    error
                         30           ip6-icmp-input          neighbor
solicitations for
                unknown    error
                         12           ip4-icmp-error            hop limit
exceeded
                response sent    info
                    1853434             bond-input
no
                error                error
                          1           ethernet-input
unknown
                vlan              error
                          1             ip4-input
ip4 ttl <=
                1              error
                          1             ip4-local                    ip4
source lookup
                miss         error
                         30           ip6-icmp-input          neighbor
solicitations for
                unknown    error
                          1           ip6-icmp-input              neighbor
                advertisements sent      info
                          1           ip4-icmp-error            hop limit
exceeded
                response sent    info
                    1864191             bond-input
no
                error                error
                          1             null-node
blackholed
                packets           error
                          5             ip4-input
ip4 ttl <=
                1              error
                         35           ip6-icmp-input          neighbor
solicitations for
                unknown    error
                          5           ip4-icmp-error            hop limit
exceeded
                response sent    info
                    1743015             bond-input
no
                error                error
                          3           ethernet-input
unknown
                vlan              error
                          1             null-node
blackholed
                packets           error
                          4             ip4-input
ip4 ttl <=
                1              error
                          1           ip6-icmp-input          neighbor
solicitations
                from source    error
                         35           ip6-icmp-input          neighbor
solicitations for
                unknown    error
                          4           ip4-icmp-error            hop limit
exceeded
                response sent    info
                    1917745             bond-input
no
                error                error
                          2             ip4-input
ip4 ttl <=
                1              error
                         18           ip6-icmp-input          neighbor
solicitations for
                unknown    error
                          2           ip4-icmp-error            hop limit
exceeded
                response sent    info
                     886249             bond-input
no
                error                error
                          1           ethernet-input
unknown
                vlan              error
                          7             ip4-input
ip4 ttl <=
                1              error
                          1           ip6-icmp-input
valid
                packets              info
                         14           ip6-icmp-input          neighbor
solicitations for
                unknown    error
                          1           ip6-icmp-input             router
advertisements
                received     info


                This counter :

                    1743015             bond-input
no
                error                error

                Is wild :) and it increasing really fast - and also simple
icmp ping to
                this host or from this host is reporting froms from 2 to 5%
- when both
                100Gnics are receiving and transmitting 7Mpps - per port (
so in total
                14Mpps RX and 14Mpps TX)

                Basically it looks like something cant handle more - and
probabbly
                FIB/RIB processing is the case here - because without
forwarding  and
                frr+routes we was able to reach like 40Mpps per 100G port


                Bonded nics: (m0rdma0+m1rdma1 - created by lcp from two
physical mlx
                ports as RDMA pairs) - create commands:

                create int rdma host-if ens1f1np1 name m1rdma1 num-rx-queues
12
                create int rdma host-if ens1f0np0 name m0rdma0 num-rx-queues
12

                create bond mode lacp load-balance l34
                bond add BondEthernet0 m0rdma0
                bond add BondEthernet0 m1rdma1


                Nics attached via rdma:

                m0rdma0                            4     up   m0rdma0
                   Link speed: 100 Gbps
                   RX Queues:
                     queue thread         mode
                     0     vpp_wk_13 (14) polling
                     1     vpp_wk_14 (15) polling
                     2     vpp_wk_15 (16) polling
                     3     vpp_wk_16 (17) polling
                     4     vpp_wk_17 (18) polling
                     5     vpp_wk_18 (19) polling
                     6     vpp_wk_19 (20) polling
                     7     vpp_wk_20 (21) polling
                     8     vpp_wk_21 (22) polling
                     9     vpp_wk_22 (23) polling
                     10    vpp_wk_23 (24) polling
                     11    vpp_wk_0 (1)   polling
                   Ethernet address 98:03:9b:67:f6:1e
                   netdev ens1f0np0 pci-addr 0000:b3:00.0
                   product name: CX516A - ConnectX-5 QSFP28
                   part number: MCX516A-CCAT
                   revision: B2
                   serial number: MT2043K02404
                   flags: admin-up mlx5dv
                   rss: ipv4-tcp ipv6-tcp
                   mlx5: version 0
                     device flags:  cqe-v1 mpw-allowed enhanced-mpw cqe-
128b-comp
                cqe-128b-pad
                m1rdma1                            3     up   m1rdma1
                   Link speed: 100 Gbps
                   RX Queues:
                     queue thread         mode
                     0     vpp_wk_1 (2)   polling
                     1     vpp_wk_2 (3)   polling
                     2     vpp_wk_3 (4)   polling
                     3     vpp_wk_4 (5)   polling
                     4     vpp_wk_5 (6)   polling
                     5     vpp_wk_6 (7)   polling
                     6     vpp_wk_7 (8)   polling
                     7     vpp_wk_8 (9)   polling
                     8     vpp_wk_9 (10)  polling
                     9     vpp_wk_10 (11) polling
                     10    vpp_wk_11 (12) polling
                     11    vpp_wk_12 (13) polling
                   Ethernet address 98:03:9b:67:f6:1e
                   netdev ens1f1np1 pci-addr 0000:b3:00.1
                   product name: CX516A - ConnectX-5 QSFP28
                   part number: MCX516A-CCAT
                   revision: B2
                   serial number: MT2043K02404
                   flags: admin-up mlx5dv
                   rss: ipv4-tcp ipv6-tcp
                   mlx5: version 0
                     device flags:  cqe-v1 mpw-allowed enhanced-mpw cqe-
128b-comp
                cqe-128b-pad

                ....

                ....

                Same NIC same Queue threads numbers (12 per card from 24
available cores
                that are isolcpus - at kernel boot)


                vpp# show interface rx-placement
                Thread 1 (vpp_wk_0):
                  node rdma-input:
                     m0rdma0 queue 11 (polling)
                  node virtio-input:
                     tap1 queue 0 (polling)
                Thread 2 (vpp_wk_1):
                  node rdma-input:
                     m1rdma1 queue 0 (polling)
                  node virtio-input:
                     tap5 queue 0 (polling)
                Thread 3 (vpp_wk_2):
                  node rdma-input:
                     m1rdma1 queue 1 (polling)
                Thread 4 (vpp_wk_3):
                     m1rdma1 queue 2 (polling)
                Thread 5 (vpp_wk_4):
                     m1rdma1 queue 3 (polling)
                Thread 6 (vpp_wk_5):
                     m1rdma1 queue 4 (polling)
                Thread 7 (vpp_wk_6):
                     m1rdma1 queue 5 (polling)
                Thread 8 (vpp_wk_7):
                     m1rdma1 queue 6 (polling)
                Thread 9 (vpp_wk_8):
                     m1rdma1 queue 7 (polling)
                Thread 10 (vpp_wk_9):
                     m1rdma1 queue 8 (polling)
                Thread 11 (vpp_wk_10):
                     m1rdma1 queue 9 (polling)
                Thread 12 (vpp_wk_11):
                     m1rdma1 queue 10 (polling)
                Thread 13 (vpp_wk_12):
                     m1rdma1 queue 11 (polling)
                Thread 14 (vpp_wk_13):
                     m0rdma0 queue 0 (polling)
                Thread 15 (vpp_wk_14):
                     m0rdma0 queue 1 (polling)
                Thread 16 (vpp_wk_15):
                     m0rdma0 queue 2 (polling)
                Thread 17 (vpp_wk_16):
                     m0rdma0 queue 3 (polling)
                Thread 18 (vpp_wk_17):
                     m0rdma0 queue 4 (polling)
                Thread 19 (vpp_wk_18):
                     m0rdma0 queue 5 (polling)
                Thread 20 (vpp_wk_19):
                     m0rdma0 queue 6 (polling)
                Thread 21 (vpp_wk_20):
                     m0rdma0 queue 7 (polling)
                Thread 22 (vpp_wk_21):
                     m0rdma0 queue 8 (polling)
                Thread 23 (vpp_wk_22):
                     m0rdma0 queue 9 (polling)
                Thread 24 (vpp_wk_23):
                     m0rdma0 queue 10 (polling)

                Below also some stats about memory used:

                vpp# show memory
                show memory: Need one of api-segment, stats-segment, main-
heap,
                numa-heaps or map
                vpp# show memory main-heap
                Thread 0 vpp_main
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.37M, free: 1.64G, trimmable:
1.62G

                Thread 1 vpp_wk_0
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.37M, free: 1.64G, trimmable:
1.62G

                Thread 2 vpp_wk_1
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 3 vpp_wk_2
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 4 vpp_wk_3
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 5 vpp_wk_4
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 6 vpp_wk_5
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 7 vpp_wk_6
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 8 vpp_wk_7
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 9 vpp_wk_8
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 10 vpp_wk_9
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 11 vpp_wk_10
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 12 vpp_wk_11
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 13 vpp_wk_12
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 14 vpp_wk_13
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 15 vpp_wk_14
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 16 vpp_wk_15
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 17 vpp_wk_16
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 18 vpp_wk_17
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 19 vpp_wk_18
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 20 vpp_wk_19
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 21 vpp_wk_20
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 22 vpp_wk_21
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 23 vpp_wk_22
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 24 vpp_wk_23
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 25 vpp_wk_24
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 26 vpp_wk_25
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.38M, free: 1.64G, trimmable:
1.62G

                Thread 27 vpp_wk_26
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.39M, free: 1.64G, trimmable:
1.62G
                vpp# show memory main-heap
                Thread 0 vpp_main
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 1 vpp_wk_0
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 2 vpp_wk_1
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 3 vpp_wk_2
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 4 vpp_wk_3
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 5 vpp_wk_4
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 6 vpp_wk_5
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 7 vpp_wk_6
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 8 vpp_wk_7
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                vpp# show memory main-heap
                Thread 0 vpp_main
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 1 vpp_wk_0
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 2 vpp_wk_1
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 3 vpp_wk_2
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 4 vpp_wk_3
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 5 vpp_wk_4
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 6 vpp_wk_5
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 7 vpp_wk_6
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                     page stats: page-size 1G, total 2, mapped 2, not-mapped
0
                       numa 0: 2 pages, 2g bytes
                     total: 1.99G, used: 369.58M, free: 1.64G, trimmable:
1.62G

                Thread 8 vpp_wk_7
                   base 0x7f64c0000000, size 2g, locked, unmap-on-destroy,
name 'main
                heap'
                vpp# show memory ?
                   show memory                              show memory
                [api-segment][stats-segment][verbose]
                             [numa-heaps][map]
                vpp# show memory
                show memory: Need one of api-segment, stats-segment, main-
heap,
                numa-heaps or map
                vpp# show memory numa-heaps
                Numa 0 uses the main heap...
                vpp# show memory map

                StartAddr          size   FD PageSz  Pages Numa0 NotMap Name
                00007f64c0000000     2g          1G      2     2      0 main
heap
                00007f655b663000     2m          4K    512     9    503
thread stack:
                thread 0
                00007f6440000000     1g    4     1G      1     1      0 stat
segment
                00007f655bc9a000    32k          4K      8     2      6
process stack:
                wg-timer-manager
                00007f655b642000   128k          4K     32     2     30
process stack:
                vrrp-periodic-process
                00007f655bc7a000    32k          4K      8     1      7
process stack:
                prom-scraper-process
                00007f655b639000    32k          4K      8     2      6
process stack:
                nsh-md2-ioam-export-process
                00007f655b630000    32k          4K      8     2      6
process stack:
                nat44-ei-ha-process
                00007f655b627000    32k          4K      8     2      6
process stack:
                memif-process
                00007f655b61e000    32k          4K      8     2      6
process stack:
                lldp-process
                00007f655b5fd000   128k          4K     32     3     29
process stack:
                linux-cp-netlink-process
                00007f655b5f4000    32k          4K      8     2      6
process stack:
                udp-ping-process
                00007f655b5eb000    32k          4K      8     2      6
process stack:
                vxlan-gpe-ioam-export-process
                00007f655b5e2000    32k          4K      8     2      6
process stack:
                ioam-export-process
                00007f655b5d9000    32k          4K      8     2      6
process stack:
                ikev2-manager-process
                00007f655b5d0000    32k          4K      8     2      6
process stack:
                igmp-timer-process
                00007f655b5c7000    32k          4K      8     1      7
process stack:
                http-timer-process
                00007f655b5be000    32k          4K      8     2      6
process stack:
                flowprobe-timer-process
                00007f655b59d000   128k          4K     32     2     30
process stack:
                dpdk-process
                00007f655b57c000   128k          4K     32     2     30
process stack:
                admin-up-down-process
                00007f655b573000    32k          4K      8     2      6
process stack:
                send-dhcp6-pd-client-message-process
                00007f655b56a000    32k          4K      8     2      6
process stack:
                dhcp6-pd-client-cp-process
                00007f655b561000    32k          4K      8     2      6
process stack:
                dhcp6-client-cp-process
                00007f655b558000    32k          4K      8     2      6
process stack:
                send-dhcp6-client-message-process
                00007f655b54f000    32k          4K      8     2      6
process stack:
                dhcp6-pd-reply-publisher-process
                00007f655b546000    32k          4K      8     2      6
process stack:
                dhcp6-reply-publisher-process
                00007f655b535000    64k          4K     16     2     14
process stack:
                dhcp-client-process
                00007f655b52c000    32k          4K      8     2      6
process stack:
                cnat-scanner-process
                00007f655b523000    32k          4K      8     2      6
process stack:
                avf-process
                00007f655b51a000    32k          4K      8     2      6
process stack:
                acl-plugin-fa-cleaner-process
                00007f655b4d9000   256k          4K     64     3     61
process stack:
                api-rx-from-ring
                00007f655b4d0000    32k          4K      8     2      6
process stack:
                rd-cp-process
                00007f655b4c7000    32k          4K      8     2      6
process stack:
                ip6-ra-process
                00007f655b4be000    32k          4K      8     2      6
process stack:
                ip6-rs-process
                00007f655b4b5000    32k          4K      8     2      6
process stack:
                ip6-mld-process
                00007f655b4ac000    32k          4K      8     2      6
process stack:
                fib-walk
                00007f655b4a3000    32k          4K      8     1      7
process stack:
                session-queue-process
                00007f655b49a000    32k          4K      8     2      6
process stack:
                vhost-user-process
                00007f655b491000    32k          4K      8     2      6
process stack:
                vhost-user-send-interrupt-process
                00007f655b488000    32k          4K      8     2      6
process stack:
                flow-report-process
                00007f655b47f000    32k          4K      8     2      6
process stack:
                bfd-process
                00007f655b476000    32k          4K      8     2      6
process stack:
                ip-neighbor-event
                00007f655b46d000    32k          4K      8     2      6
process stack:
                ip6-neighbor-age-process
                00007f655b464000    32k          4K      8     2      6
process stack:
                ip4-neighbor-age-process
                00007f655b45b000    32k          4K      8     2      6
process stack:
                ip6-sv-reassembly-expire-walk
                00007f655b452000    32k          4K      8     2      6
process stack:
                ip6-full-reassembly-expire-walk
                00007f655b449000    32k          4K      8     2      6
process stack:
                ip4-sv-reassembly-expire-walk
                00007f655b440000    32k          4K      8     2      6
process stack:
                ip4-full-reassembly-expire-walk
                00007f655b437000    32k          4K      8     2      6
process stack:
                bond-process
                00007f655b42e000    32k          4K      8     2      6
process stack:
                l2fib-mac-age-scanner-process
                00007f655b425000    32k          4K      8     2      6
process stack:
                l2-arp-term-publisher
                00007f655b41c000    32k          4K      8     2      6
process stack:
                vpe-link-state-process
                00007f655b3db000   256k          4K     64     5     59
process stack:
                startup-config-process
                00007f655b3d2000    32k          4K      8     2      6
process stack:
                statseg-collector-process
                00007f655a5a8000     2m          4K    512     7    505
thread stack:
                thread 1
                00007f655a3a7000     2m          4K    512    10    502
thread stack:
                thread 2
                00007f655a1a6000     2m          4K    512    10    502
thread stack:
                thread 3
                00007f6559fa5000     2m          4K    512    10    502
thread stack:
                thread 4
                00007f6559da4000     2m          4K    512    10    502
thread stack:
                thread 5
                00007f6559ba3000     2m          4K    512    10    502
thread stack:
                thread 6
                00007f65599a2000     2m          4K    512    10    502
thread stack:
                thread 7
                00007f65597a1000     2m          4K    512    11    501
thread stack:
                thread 8
                00007f65595a0000     2m          4K    512    10    502
thread stack:
                thread 9
                00007f655939f000     2m          4K    512    10    502
thread stack:
                thread 10
                00007f655919e000     2m          4K    512    10    502
thread stack:
                thread 11
                00007f6558f9d000     2m          4K    512    10    502
thread stack:
                thread 12
                00007f6558d9c000     2m          4K    512    10    502
thread stack:
                thread 13
                00007f6558b9b000     2m          4K    512    10    502
thread stack:
                thread 14
                00007f655899a000     2m          4K    512    10    502
thread stack:
                thread 15
                00007f6558799000     2m          4K    512    10    502
thread stack:
                thread 16
                00007f6558598000     2m          4K    512    10    502
thread stack:
                thread 17
                00007f6558397000     2m          4K    512    10    502
thread stack:
                thread 18
                00007f6558196000     2m          4K    512    10    502
thread stack:
                thread 19
                00007f6553e00000     2m          4K    512    10    502
thread stack:
                thread 20
                00007f6553bff000     2m          4K    512    10    502
thread stack:
                thread 21
                00007f65539fe000     2m          4K    512    10    502
thread stack:
                thread 22
                00007f65537fd000     2m          4K    512    10    502
thread stack:
                thread 23
                00007f65535fc000     2m          4K    512    10    502
thread stack:
                thread 24
                00007f65533fb000     2m          4K    512    10    502
thread stack:
                thread 25
                00007f65531fa000     2m          4K    512    10    502
thread stack:
                thread 26
                00007f6552ff9000     2m          4K    512    10    502
thread stack:
                thread 27
                00007f6552f98000    64k          4K     16     2     14
process stack:
                lacp-process
                00007f6552f57000   256k          4K     64     3     61
process stack:
                unix-cli-local:0
                00007f6552f46000    64k          4K     16     2     14
process stack:
                unix-cli-new-session










--
Paweł Staszewski
CEO at ITCare

Cell:    +48 609911040
pstaszew...@itcare.pl <mailto:pstaszew...@itcare.pl>

This message (including any attachments) contains confidential information
intended
for a specific individual and purpose, and is protected by law.  If you
are not the
intended recipient, you should delete this message.  Any disclosure,
copying, or
distribution of this message, or the taking of any action based on it, is
strictly
prohibited.

--------------------------------------------------------------------------
-----------

Wiadomość ta jest przeznaczona jedynie dla osoby lub podmiotu będącego jej
adresatem i
może zawierać poufne lub uprzywilejowane informacje. Zakazane jest
przeglądanie,
przesyłanie, rozpowszechnianie lub inne wykorzystywanie tych informacji,
jak również
podejmowanie działań na ich podstawie, przez osoby lub podmioty inne niż
zamierzony
adresat. Jeśli otrzymali Państwo tę wiadomość przez pomyłkę, prosimy o
poinformowanie
nadawcy i usunięcie jej z komputera.



-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.
View/Reply Online (#21892): https://lists.fd.io/g/vpp-dev/message/21892
Mute This Topic: https://lists.fd.io/mt/93619900/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/leave/1480452/21656/631435203/xyzzy 
[arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to