The "show hardw" is as follow, the statistics may be different from yesterday.
vpp# show hardware-interfaces Name Idx Link Hardware VirtualFunctionEthernet5/0/2 1 up VirtualFunctionEthernet5/0/2 Ethernet address 72:62:8a:40:43:12 Cavium ThunderX carrier up full duplex speed 10000 mtu 9190 flags: admin-up pmd maybe-multiseg rx queues 2, rx desc 1024, tx queues 2, tx desc 1024 cpu socket 0 tx frames ok 268302 tx bytes ok 74319654 rx frames ok 4000000 rx bytes ok 688000000 extended stats: rx good packets 4000000 tx good packets 268302 rx good bytes 688000000 tx good bytes 74319654 rx q0packets 3999976 rx q0bytes 687995872 rx q1packets 24 rx q1bytes 4128 tx q0packets 12 tx q0bytes 3324 tx q1packets 268290 tx q1bytes 74316330 VirtualFunctionEthernet5/0/3 2 down VirtualFunctionEthernet5/0/3 Ethernet address 2a:f2:d5:47:67:f1 Cavium ThunderX carrier down flags: pmd maybe-multiseg rx queues 2, rx desc 1024, tx queues 2, tx desc 1024 cpu socket 0 local0 0 down local0 local mik...@yeah.net From: Damjan Marion Date: 2018-12-18 20:38 To: mik...@yeah.net CC: vpp-dev@lists.fd.io Subject: Re: [vpp-dev] dpdk-input : serious load imbalance What kind of nic do you have? Can you capture "show hardw" ? -- Damjan On 18 Dec 2018, at 04:03, mik...@yeah.net wrote: Hi, I configured 2 worker thread and 2 dpdk rx-queuein startup.conf . Then I forged 400w packets and send them to a single dpdk interface . It turns out that the second thread received only 24 packets .I test it for several times and the results are almost the same. Why did this happen ? Here is some config and "show": VPP : 18.07 startup.conf: cpu { main-core 1 corelist-workers 2,3 } dpdk { dev default { num-rx-queues 2 num-tx-queues 2 } } packets: these pkts share the same src mac , dst mac and ipv4 body , only ipv4 src ip and dst ip are different from each other. <Catch.jpg> ------------------------------------------------------------------ # sh runtime Thread 1 vpp_wk_0 (lcore 2) Time 69.8, average vectors/node 1.02, last 128 main loops 0.00 per node 0.00 vector rates in 5.7911e4, out 5.0225e2, drop 5.7783e4, punt 0.0000e0 Name State Calls Vectors Suspends Clocks Vectors/Call dpdk-input polling 104246554 3999630 0 7.84e2 .04 --------------- Thread 2 vpp_wk_1 (lcore 3) Time 69.8, average vectors/node 1.00, last 128 main loops 0.00 per node 0.00 vector rates in 5.1557e-1, out 1.7186e-1, drop 5.1557e-1, punt 0.0000e0 Name State Calls Vectors Suspends Clocks Vectors/Call dpdk-input polling 132390000 24 0 1.59e8 0.00 ----------------------------------------------------------------- # show interface rx-placement Thread 1 (vpp_wk_0): node dpdk-input: VirtualFunctionEthernet5/0/2 queue 0 (polling) VirtualFunctionEthernet5/0/3 queue 0 (polling) Thread 2 (vpp_wk_1): node dpdk-input: VirtualFunctionEthernet5/0/2 queue 1 (polling) VirtualFunctionEthernet5/0/3 queue 1 (polling) ----------------------------------------------------------------- vpp# show interface Name Idx State MTU (L3/IP4/IP6/MPLS) Counter Count VirtualFunctionEthernet5/0/2 1 up 9000/0/0/0 rx packets 3999654 rx bytes 687940488 tx packets 35082 tx bytes 9647550 rx-miss 346 VirtualFunctionEthernet5/0/3 2 down 9000/0/0/0 local0 0 down 0/0/0/0 Thanks in advance. Mikado mik...@yeah.net -=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#11675): https://lists.fd.io/g/vpp-dev/message/11675 Mute This Topic: https://lists.fd.io/mt/28791566/675642 Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [dmar...@me.com] -=-=-=-=-=-=-=-=-=-=-=-
-=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#11713): https://lists.fd.io/g/vpp-dev/message/11713 Mute This Topic: https://lists.fd.io/mt/28791566/21656 Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com] -=-=-=-=-=-=-=-=-=-=-=-