Karl, Thanks for all answers and clarifications!
~23Mpps that’s indeed your NIC - we’re observing same in CSIT.
One more point re host setup - I read it’s CPU power management disabled,
CPU TurboBoost disabled, and all memory channels populated?
I looks so, but wanted to recheck as not listed on your opening slide :)

-Maciek

> On 15 Feb 2017, at 21:55, Karl Rister <kris...@redhat.com> wrote:
> 
> On 02/15/2017 03:28 PM, Maciek Konstantynowicz (mkonstan) wrote:
>> Thomas, many thanks for sending this.
>> 
>> Few comments and questions after reading the slides:
>> 
>> 1. s3 clarification - host and data plane thread setup - vswitch pmd
>> (data plane) thread placement
>>    a. "1PMD/core (4 core)” - HT (SMT) disabled, 4 phy cores used for
>> vswitch, each with data plane thread.
>>    b. “2PMD/core (2 core)” - HT (SMT) enabled, 2 phy cores, 4 logical
>> cores used for vswitch, each with data plane thread.
>>    c. in both cases each data plane thread handling a single interface
>> - 2* physical, 2* vhost => 4 threads, all busy.
> 
> Correct.
> 
>>    d. in both cases frames are dropped by vswitch or in vring due to
>> vswitch not keeping up - IOW testpmd in kvm guest is not DUT.
> 
> That is the intent.
> 
>> 2. s3 question - vswitch setup - it is unclear what is the forwarding
>> mode of each vswitch, as only srcIp changed in flows
>>    a. flow or MAC learning mode?
> 
> In OVS we program flow rules that pass bidirectional traffic between a
> physical and vhost port pair.
> 
>>    b. port to port crossconnect?
> 
> In VPP we are using xconnect.
> 
>> 3. s3 comment - host and data plane thread setup
>>    a. “2PMD/core (2 core)” case - thread placement may yield different
>> results 
>>        - physical interface threads as siblings vs. 
>>        - physical and virtual interface threads as siblings.
> 
> In both OVS and VPP a physical interface thread is paired with a virtual
> interface thread on the same core.
> 
>>    b. "1PMD/core (4 core)” - one would expect these to be much higher
>> than “2PMD/core (2 core)”
>>        - speculation: possibly due to "instruction load" imbalance
>> between threads.
>>        - two types of thread with different "instruction load":
>> phy->vhost vs. vhost->phy
>>        - "instruction load" = instr/pkt, instr/cycle (IPC efficiency).
>> 4. s4 comment - results look as expected for vpp
>> 5. s5 question - unclear why throughput doubled 
>>    a. e.g. for vpp from "11.16 Mpps" to "22.03 Mpps"
>>    b. if only queues increased, and cpu resources did not, or have they?
>> 6. s6 question - similar to point 5. - unclear cpu and thread reasources.
> 
> Queues and cores increase together.  In the host single queue used 4 PMD
> threads on 2 core, two queue uses 8 PMD threads on 4 cores, and three
> queue uses 12 PMD threads on 6 cores.  In the guest we used 2, 4, and 6
> cores in testpmd without using sibling hyperthreads in order to avoid
> bottlenecks in the guest.
> 
>> 7. s7 comment - anomaly for 3q (virtio multi-queue) for (srcMAc,dstMAC)
>>    a. could be due to flow hashing inefficiency.
> 
> That was our thinking and where we were going to look first.
> 
> I think I have tracked the three queue issue to using too many mbufs for
> multi-queue as addressed in this commit:
> 
> https://github.com/vpp-dev/vpp/commit/a06dfb39c6bee3fbfd702c10e1e1416b98e65455
> 
> I originally used the suggestion of 131072 from this page:
> 
> https://wiki.fd.io/view/VPP/Command-line_Arguments#.22dpdk.22_parameters
> 
> I'm now testing with 3 queue and 32768 mbufs and getting in excess of 23
> Mpps across all the flow configurations except the one with the hashing
> issue.  For our hardware configuration we believe this is hardware
> limited and could potentially go faster (as mentioned on slide 6).
> 
>> 
>> -Maciek
>> 
>>> On 15 Feb 2017, at 17:34, Thomas F Herbert <therb...@redhat.com
>>> <mailto:therb...@redhat.com>> wrote:
>>> 
>>> Here are test results on VPP 17.01 compared with OVS/DPDK 2.6/1611
>>> performed by Karl Rister of Red Hat.
>>> 
>>> This is PVP testing with 1, 2 and 3 queues. It is an interesting
>>> comparison with the CSIT results. Of particular interest is the drop
>>> off on the 3 queue results.
>>> 
>>> --TFH
>>> 
>>> 
>>> -- 
>>> *Thomas F Herbert*
>>> SDN Group
>>> Office of Technology
>>> *Red Hat*
>>> <vpp-17.01_vs_ovs-2.6.pdf>_______________________________________________
>>> vpp-dev mailing list
>>> vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>
>>> https://lists.fd.io/mailman/listinfo/vpp-dev
>> 
> 
> 
> -- 
> Karl Rister <kris...@redhat.com>

_______________________________________________
vpp-dev mailing list
vpp-dev@lists.fd.io
https://lists.fd.io/mailman/listinfo/vpp-dev

Reply via email to