ok , thanks anyway.


mik...@yeah.net
 
From: Damjan Marion
Date: 2018-12-19 16:04
To: mik...@yeah.net
CC: vpp-dev
Subject: Re: [vpp-dev] dpdk-input : serious load imbalance

Yeah, I did not expect change, but at least we know that RSS is confiugred on 
this card:

    rss avail:         ipv4 ipv4-tcp ipv4-udp ipv6 ipv6-tcp ipv6-udp port 
                       vxlan geneve nvgre 
    rss active:        ipv4 ipv4-tcp ipv4-udp ipv6 ipv6-tcp ipv6-udp 


I'm affraid this is not VPP issue, you will need to talk with Cavium guys and 
ask them why RSS doesn't work properly...

-- 
Damjan

On 19 Dec 2018, at 09:00, mik...@yeah.net wrote:

The result is the same.  The 2rd queue only received 24 packets as it is in 
18.07.

version : vpp v18.10-19~ga8e3001-dirty 
vpp# show hardware-interfaces 
              Name                Idx   Link  Hardware
VirtualFunctionEthernet5/0/2       1     up   VirtualFunctionEthernet5/0/2
  Ethernet address 12:90:b1:65:2d:19
  Cavium ThunderX
    carrier up full duplex speed 10000 mtu 9190 
    flags: admin-up pmd maybe-multiseg
    rx: queues 2 (max 96), desc 1024 (min 0 max 65535 align 1)
    tx: queues 2 (max 96), desc 1024 (min 0 max 65535 align 1)
    pci: device 177d:a034 subsystem 177d:a234 address 0000:05:00.02 numa 0
    module: unknown
    max rx packet len: 9204
    promiscuous: unicast off all-multicast off
    vlan offload: strip off filter off qinq off
    rx offload avail:  vlan-strip ipv4-cksum udp-cksum tcp-cksum jumbo-frame 
                       crc-strip scatter 
    rx offload active: jumbo-frame crc-strip scatter 
    tx offload avail:  ipv4-cksum udp-cksum tcp-cksum outer-ipv4-cksum 
    tx offload active: 
    rss avail:         ipv4 ipv4-tcp ipv4-udp ipv6 ipv6-tcp ipv6-udp port 
                       vxlan geneve nvgre 
    rss active:        ipv4 ipv4-tcp ipv4-udp ipv6 ipv6-tcp ipv6-udp 
    tx burst function: nicvf_xmit_pkts_multiseg
    rx burst function: nicvf_recv_pkts_multiseg_no_offload

    tx frames ok                                       11064
    tx bytes ok                                      3042600
    rx frames ok                                     3995658
    rx bytes ok                                    687253292
    rx missed                                           4342
    extended stats:
      rx good packets                                3995658
      tx good packets                                  11064
      rx good bytes                                687253292
      tx good bytes                                  3042600
      rx missed errors                                  4342
      rx q0packets                                   3995634
      rx q0bytes                                   687249164
      rx q1packets                                        24
      rx q1bytes                                        4128
      tx q0packets                                        12
      tx q0bytes                                        3300
      tx q1packets                                     11052
      tx q1bytes                                     3039300




mik...@yeah.net
 
From: Damjan Marion
Date: 2018-12-19 15:46
To: mik...@yeah.net
CC: vpp-dev
Subject: Re: [vpp-dev] dpdk-input : serious load imbalance
Can you try to capture "show hardw" with 18.10 ?

Looks like ThunderX is not acting as PCI device so part of the output is 
suppressed in 18.07 and we changed that behaviour in 18.10.

I'm looking for someething like:

    rss avail:         ipv4 ipv4-tcp ipv4-udp ipv6 ipv6-tcp ipv6-udp ipv6-tcp-ex
                       ipv6-udp-ex ipv6-ex ipv6-tcp-ex ipv6-udp-ex
    rss active:        none


-- 
Damjan

On 19 Dec 2018, at 08:26, mik...@yeah.net wrote:

vpp v18.07.1-10~gc548f5d-dirty 



mik...@yeah.net
 
From: Damjan Marion
Date: 2018-12-19 15:21
To: mik...@yeah.net
CC: vpp-dev
Subject: Re: [vpp-dev] dpdk-input : serious load imbalance

What version of VPP do you use.
I'm missing some outputs in "show hardware"...

-- 
Damjan

On 19 Dec 2018, at 02:19, mik...@yeah.net wrote:

The "show hardw" is as follow, the statistics may be different from yesterday.

vpp# show hardware-interfaces 
Name Idx Link Hardware
VirtualFunctionEthernet5/0/2 1 up VirtualFunctionEthernet5/0/2
Ethernet address 72:62:8a:40:43:12
Cavium ThunderX
carrier up full duplex speed 10000 mtu 9190 
flags: admin-up pmd maybe-multiseg
rx queues 2, rx desc 1024, tx queues 2, tx desc 1024
cpu socket 0

tx frames ok 268302
tx bytes ok 74319654
rx frames ok 4000000
rx bytes ok 688000000
extended stats:
rx good packets 4000000
tx good packets 268302
rx good bytes 688000000
tx good bytes 74319654
rx q0packets 3999976
rx q0bytes 687995872
rx q1packets 24
rx q1bytes 4128
tx q0packets 12
tx q0bytes 3324
tx q1packets 268290
tx q1bytes 74316330
VirtualFunctionEthernet5/0/3 2 down VirtualFunctionEthernet5/0/3
Ethernet address 2a:f2:d5:47:67:f1
Cavium ThunderX
carrier down 
flags: pmd maybe-multiseg
rx queues 2, rx desc 1024, tx queues 2, tx desc 1024
cpu socket 0

local0 0 down local0
local



mik...@yeah.net
 
From: Damjan Marion
Date: 2018-12-18 20:38
To: mik...@yeah.net
CC: vpp-dev@lists.fd.io
Subject: Re: [vpp-dev] dpdk-input : serious load imbalance

What kind of nic do you have? Can you capture "show hardw" ?

-- 
Damjan

On 18 Dec 2018, at 04:03, mik...@yeah.net wrote:

Hi,
   I configured 2 worker thread and 2 dpdk rx-queuein startup.conf . Then I 
forged 400w packets and send them to a single dpdk interface . It turns out 
that the second thread received only 24 packets .I test it for several times 
and the results are almost the same. Why did this happen ?

   Here is some config and "show":
VPP : 18.07
startup.conf:
cpu {
        main-core 1
        corelist-workers 2,3
}

dpdk {
         dev default {
                 num-rx-queues 2
                 num-tx-queues 2
         }
}

packets:
these pkts share the same src mac  , dst mac  and ipv4 body , only ipv4 src ip 
and dst ip are different from each other.
<Catch.jpg>

------------------------------------------------------------------
# sh runtime
Thread 1 vpp_wk_0 (lcore 2)
Time 69.8, average vectors/node 1.02, last 128 main loops 0.00 per node 0.00
  vector rates in 5.7911e4, out 5.0225e2, drop 5.7783e4, punt 0.0000e0
             Name                 State         Calls          Vectors        
Suspends         Clocks       Vectors/Call  
dpdk-input                       polling         104246554         3999630      
         0          7.84e2             .04
---------------
Thread 2 vpp_wk_1 (lcore 3)
Time 69.8, average vectors/node 1.00, last 128 main loops 0.00 per node 0.00
  vector rates in 5.1557e-1, out 1.7186e-1, drop 5.1557e-1, punt 0.0000e0
             Name                 State         Calls          Vectors        
Suspends         Clocks       Vectors/Call  
dpdk-input                       polling         132390000              24      
         0          1.59e8            0.00
-----------------------------------------------------------------
# show interface rx-placement 
Thread 1 (vpp_wk_0):
  node dpdk-input:
    VirtualFunctionEthernet5/0/2 queue 0 (polling)
    VirtualFunctionEthernet5/0/3 queue 0 (polling)
Thread 2 (vpp_wk_1):
  node dpdk-input:
    VirtualFunctionEthernet5/0/2 queue 1 (polling)
    VirtualFunctionEthernet5/0/3 queue 1 (polling)
-----------------------------------------------------------------
vpp# show interface 
              Name               Idx    State  MTU (L3/IP4/IP6/MPLS)     
Counter          Count     
VirtualFunctionEthernet5/0/2      1      up          9000/0/0/0     rx packets  
             3999654
                                                                                
                   rx bytes               687940488
                                                                                
                   tx packets                 35082
                                                                                
                   tx bytes                 9647550
                                                                                
                   rx-miss                      346
VirtualFunctionEthernet5/0/3      2     down         9000/0/0/0     
local0                            0     down          0/0/0/0       


Thanks in advance.
Mikado


mik...@yeah.net
-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#11675): https://lists.fd.io/g/vpp-dev/message/11675
Mute This Topic: https://lists.fd.io/mt/28791566/675642
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [dmar...@me.com]
-=-=-=-=-=-=-=-=-=-=-=-

-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#11716): https://lists.fd.io/g/vpp-dev/message/11716
Mute This Topic: https://lists.fd.io/mt/28791566/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to