Have you tried to use native AVF driver instead?

> On 4 Sep 2019, at 12:42, Miroslav Kováč <miroslav.ko...@pantheon.tech> wrote:
> 
> Hello,
> 
> We are trying a setup with intel 25 GB card XXV710 and sr-iov. We need sr-iov 
> to sort packets based on vlan in between the VFs. We are using trex on one 
> machine to generate packets and multiple VPPs (each in docker container, 
> using one VF) on another one. Trex machine contains the exact same hardware. 
> 
> Each VF contains one vlan with spoof checking off and trust on and specific 
> MAC address. For example ->
> 
> vf 0 MAC ba:dc:0f:fe:ed:00, vlan 1537, spoof checking off, link-state auto, 
> trust on
> 
> 
> We are generating packets with VF destination MACs with the corresponding 
> VLAN. When sending packets to 3 VFs trex shows 35 million tx-packets and Dpdk 
> stats on the trex machine show that 35 million were in fact sent out:
> 
> ##### DPDK Statistics port0 #####
> {
>     "tx_good_bytes": 2142835740,
>     "tx_good_packets": 35713929,
>     "tx_size_64_packets": 35713929,
>     "tx_unicast_packets": 35713929
> }
> 
> rate= '96%'; pktSize=       64; frameLoss%=51.31%; bytesReceived/s=    
> 1112966528.00; totalReceived=   17390102; totalSent=   35713929; frameLoss=   
> 18323827; bytesReceived=    1112966528; targetDuration=1.0
> 
> 
> However VPP shows only 33 million rx-packets:
> 
> VirtualFunctionEthernet17/a/0     2      up          9000/0/0/0     
> rx packets               5718196
> rx bytes               343091760
> rx-miss                  5572089     
> 
> VirtualFunctionEthernet17/a/1     2      up          9000/0/0/0     
> rx packets               5831396
> rx bytes               349883760
> rx-miss                  5459089
> 
> VirtualFunctionEthernet17/a/2     2      up          9000/0/0/0     
> rx packets               5840512
> rx bytes               350430720
> rx-miss                  5449466
> 
> Sum of rx packets and rx-miss is 33,870,748. About 2 million is missing.
> 
> 
> Even when I check VFs stats I see only 33 million to come (out of which 9.9 
> million are rx-missed):
> 
> root@protonet:/home/protonet# for f in $(ls 
> /sys/class/net/enp23s0f1/device/sriov/*/stats/rx_packets); do echo "$f: $(cat 
> $f)"; done | grep -v ' 0$'
> 
> /sys/class/net/enp23s0f1/device/sriov/0/stats/rx_packets: 11290290
> /sys/class/net/enp23s0f1/device/sriov/1/stats/rx_packets: 11290485
> /sys/class/net/enp23s0f1/device/sriov/2/stats/rx_packets: 11289978
> 
> 
> When increasing the number of VFs the number of rx-packets on VPP is actually 
> decreasing. Up to 6 or 7 VFs I still receive somewhere around 28-33 million 
> packets, but when I use 8 VFs all the sudden it drops to 16 million packets 
> (no rx-miss any more). The same goes with trunk mode:
> 
> VirtualFunctionEthernet17/a/0     2      up          9000/0/0/0     
> rx packets               1959110
> rx bytes               117546600
> 
> VirtualFunctionEthernet17/a/1     2      up          9000/0/0/0     
> rx packets               1959181
> rx bytes               117550860
> 
> VirtualFunctionEthernet17/a/2     2      up          9000/0/0/0     
> rx packets               1956242
> rx bytes               117374520
> .
> .
> .
> Approximately the same amount of packets for each VPP instance which is 2 
> million packets * 8 = 16 million packets out of 35 million sent. Almost 20 
> million are gone
> 
> 
> We are using vfio-pci driver.
> 
> The strange thing is that when I use only PF, no sr-iov VFs are on and I try 
> the same vpp setup I can see all 35 million packets to come across. 
> 
> We have also tested this with X710 10GB intel card and we have received 
> similar results.
> 
> Regards,
> Miroslav Kovac
> -=-=-=-=-=-=-=-=-=-=-=-
> Links: You receive all messages sent to this group.
> 
> View/Reply Online (#13894): https://lists.fd.io/g/vpp-dev/message/13894 
> <https://lists.fd.io/g/vpp-dev/message/13894>
> Mute This Topic: https://lists.fd.io/mt/33136868/675642 
> <https://lists.fd.io/mt/33136868/675642>
> Group Owner: vpp-dev+ow...@lists.fd.io <mailto:vpp-dev+ow...@lists.fd.io>
> Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub 
> <https://lists.fd.io/g/vpp-dev/unsub>  [dmar...@me.com 
> <mailto:dmar...@me.com>]
> -=-=-=-=-=-=-=-=-=-=-=-

-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.

View/Reply Online (#13895): https://lists.fd.io/g/vpp-dev/message/13895
Mute This Topic: https://lists.fd.io/mt/33136890/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub  [arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to