On 02/16/2017 04:25 AM, Maciek Konstantynowicz (mkonstan) wrote:
> Karl, Thanks for all answers and clarifications!
> ~23Mpps that’s indeed your NIC - we’re observing same in CSIT.
> One more point re host setup - I read it’s CPU power management disabled,
> CPU TurboBoost disabled, and all memory channels populated?
> I looks so, but wanted to recheck as not listed on your opening slide :)

Correct on all counts.

> 
> -Maciek
> 
>> On 15 Feb 2017, at 21:55, Karl Rister <kris...@redhat.com> wrote:
>>
>> On 02/15/2017 03:28 PM, Maciek Konstantynowicz (mkonstan) wrote:
>>> Thomas, many thanks for sending this.
>>>
>>> Few comments and questions after reading the slides:
>>>
>>> 1. s3 clarification - host and data plane thread setup - vswitch pmd
>>> (data plane) thread placement
>>>    a. "1PMD/core (4 core)” - HT (SMT) disabled, 4 phy cores used for
>>> vswitch, each with data plane thread.
>>>    b. “2PMD/core (2 core)” - HT (SMT) enabled, 2 phy cores, 4 logical
>>> cores used for vswitch, each with data plane thread.
>>>    c. in both cases each data plane thread handling a single interface
>>> - 2* physical, 2* vhost => 4 threads, all busy.
>>
>> Correct.
>>
>>>    d. in both cases frames are dropped by vswitch or in vring due to
>>> vswitch not keeping up - IOW testpmd in kvm guest is not DUT.
>>
>> That is the intent.
>>
>>> 2. s3 question - vswitch setup - it is unclear what is the forwarding
>>> mode of each vswitch, as only srcIp changed in flows
>>>    a. flow or MAC learning mode?
>>
>> In OVS we program flow rules that pass bidirectional traffic between a
>> physical and vhost port pair.
>>
>>>    b. port to port crossconnect?
>>
>> In VPP we are using xconnect.
>>
>>> 3. s3 comment - host and data plane thread setup
>>>    a. “2PMD/core (2 core)” case - thread placement may yield different
>>> results 
>>>        - physical interface threads as siblings vs. 
>>>        - physical and virtual interface threads as siblings.
>>
>> In both OVS and VPP a physical interface thread is paired with a virtual
>> interface thread on the same core.
>>
>>>    b. "1PMD/core (4 core)” - one would expect these to be much higher
>>> than “2PMD/core (2 core)”
>>>        - speculation: possibly due to "instruction load" imbalance
>>> between threads.
>>>        - two types of thread with different "instruction load":
>>> phy->vhost vs. vhost->phy
>>>        - "instruction load" = instr/pkt, instr/cycle (IPC efficiency).
>>> 4. s4 comment - results look as expected for vpp
>>> 5. s5 question - unclear why throughput doubled 
>>>    a. e.g. for vpp from "11.16 Mpps" to "22.03 Mpps"
>>>    b. if only queues increased, and cpu resources did not, or have they?
>>> 6. s6 question - similar to point 5. - unclear cpu and thread reasources.
>>
>> Queues and cores increase together.  In the host single queue used 4 PMD
>> threads on 2 core, two queue uses 8 PMD threads on 4 cores, and three
>> queue uses 12 PMD threads on 6 cores.  In the guest we used 2, 4, and 6
>> cores in testpmd without using sibling hyperthreads in order to avoid
>> bottlenecks in the guest.
>>
>>> 7. s7 comment - anomaly for 3q (virtio multi-queue) for (srcMAc,dstMAC)
>>>    a. could be due to flow hashing inefficiency.
>>
>> That was our thinking and where we were going to look first.
>>
>> I think I have tracked the three queue issue to using too many mbufs for
>> multi-queue as addressed in this commit:
>>
>> https://github.com/vpp-dev/vpp/commit/a06dfb39c6bee3fbfd702c10e1e1416b98e65455
>>
>> I originally used the suggestion of 131072 from this page:
>>
>> https://wiki.fd.io/view/VPP/Command-line_Arguments#.22dpdk.22_parameters
>>
>> I'm now testing with 3 queue and 32768 mbufs and getting in excess of 23
>> Mpps across all the flow configurations except the one with the hashing
>> issue.  For our hardware configuration we believe this is hardware
>> limited and could potentially go faster (as mentioned on slide 6).
>>
>>>
>>> -Maciek
>>>
>>>> On 15 Feb 2017, at 17:34, Thomas F Herbert <therb...@redhat.com
>>>> <mailto:therb...@redhat.com>> wrote:
>>>>
>>>> Here are test results on VPP 17.01 compared with OVS/DPDK 2.6/1611
>>>> performed by Karl Rister of Red Hat.
>>>>
>>>> This is PVP testing with 1, 2 and 3 queues. It is an interesting
>>>> comparison with the CSIT results. Of particular interest is the drop
>>>> off on the 3 queue results.
>>>>
>>>> --TFH
>>>>
>>>>
>>>> -- 
>>>> *Thomas F Herbert*
>>>> SDN Group
>>>> Office of Technology
>>>> *Red Hat*
>>>> <vpp-17.01_vs_ovs-2.6.pdf>_______________________________________________
>>>> vpp-dev mailing list
>>>> vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>
>>>> https://lists.fd.io/mailman/listinfo/vpp-dev
>>>
>>
>>
>> -- 
>> Karl Rister <kris...@redhat.com>
> 


-- 
Karl Rister <kris...@redhat.com>
_______________________________________________
vpp-dev mailing list
vpp-dev@lists.fd.io
https://lists.fd.io/mailman/listinfo/vpp-dev

Reply via email to