On 02/16/2017 04:25 AM, Maciek Konstantynowicz (mkonstan) wrote: > Karl, Thanks for all answers and clarifications! > ~23Mpps that’s indeed your NIC - we’re observing same in CSIT. > One more point re host setup - I read it’s CPU power management disabled, > CPU TurboBoost disabled, and all memory channels populated? > I looks so, but wanted to recheck as not listed on your opening slide :)
Correct on all counts. > > -Maciek > >> On 15 Feb 2017, at 21:55, Karl Rister <kris...@redhat.com> wrote: >> >> On 02/15/2017 03:28 PM, Maciek Konstantynowicz (mkonstan) wrote: >>> Thomas, many thanks for sending this. >>> >>> Few comments and questions after reading the slides: >>> >>> 1. s3 clarification - host and data plane thread setup - vswitch pmd >>> (data plane) thread placement >>> a. "1PMD/core (4 core)” - HT (SMT) disabled, 4 phy cores used for >>> vswitch, each with data plane thread. >>> b. “2PMD/core (2 core)” - HT (SMT) enabled, 2 phy cores, 4 logical >>> cores used for vswitch, each with data plane thread. >>> c. in both cases each data plane thread handling a single interface >>> - 2* physical, 2* vhost => 4 threads, all busy. >> >> Correct. >> >>> d. in both cases frames are dropped by vswitch or in vring due to >>> vswitch not keeping up - IOW testpmd in kvm guest is not DUT. >> >> That is the intent. >> >>> 2. s3 question - vswitch setup - it is unclear what is the forwarding >>> mode of each vswitch, as only srcIp changed in flows >>> a. flow or MAC learning mode? >> >> In OVS we program flow rules that pass bidirectional traffic between a >> physical and vhost port pair. >> >>> b. port to port crossconnect? >> >> In VPP we are using xconnect. >> >>> 3. s3 comment - host and data plane thread setup >>> a. “2PMD/core (2 core)” case - thread placement may yield different >>> results >>> - physical interface threads as siblings vs. >>> - physical and virtual interface threads as siblings. >> >> In both OVS and VPP a physical interface thread is paired with a virtual >> interface thread on the same core. >> >>> b. "1PMD/core (4 core)” - one would expect these to be much higher >>> than “2PMD/core (2 core)” >>> - speculation: possibly due to "instruction load" imbalance >>> between threads. >>> - two types of thread with different "instruction load": >>> phy->vhost vs. vhost->phy >>> - "instruction load" = instr/pkt, instr/cycle (IPC efficiency). >>> 4. s4 comment - results look as expected for vpp >>> 5. s5 question - unclear why throughput doubled >>> a. e.g. for vpp from "11.16 Mpps" to "22.03 Mpps" >>> b. if only queues increased, and cpu resources did not, or have they? >>> 6. s6 question - similar to point 5. - unclear cpu and thread reasources. >> >> Queues and cores increase together. In the host single queue used 4 PMD >> threads on 2 core, two queue uses 8 PMD threads on 4 cores, and three >> queue uses 12 PMD threads on 6 cores. In the guest we used 2, 4, and 6 >> cores in testpmd without using sibling hyperthreads in order to avoid >> bottlenecks in the guest. >> >>> 7. s7 comment - anomaly for 3q (virtio multi-queue) for (srcMAc,dstMAC) >>> a. could be due to flow hashing inefficiency. >> >> That was our thinking and where we were going to look first. >> >> I think I have tracked the three queue issue to using too many mbufs for >> multi-queue as addressed in this commit: >> >> https://github.com/vpp-dev/vpp/commit/a06dfb39c6bee3fbfd702c10e1e1416b98e65455 >> >> I originally used the suggestion of 131072 from this page: >> >> https://wiki.fd.io/view/VPP/Command-line_Arguments#.22dpdk.22_parameters >> >> I'm now testing with 3 queue and 32768 mbufs and getting in excess of 23 >> Mpps across all the flow configurations except the one with the hashing >> issue. For our hardware configuration we believe this is hardware >> limited and could potentially go faster (as mentioned on slide 6). >> >>> >>> -Maciek >>> >>>> On 15 Feb 2017, at 17:34, Thomas F Herbert <therb...@redhat.com >>>> <mailto:therb...@redhat.com>> wrote: >>>> >>>> Here are test results on VPP 17.01 compared with OVS/DPDK 2.6/1611 >>>> performed by Karl Rister of Red Hat. >>>> >>>> This is PVP testing with 1, 2 and 3 queues. It is an interesting >>>> comparison with the CSIT results. Of particular interest is the drop >>>> off on the 3 queue results. >>>> >>>> --TFH >>>> >>>> >>>> -- >>>> *Thomas F Herbert* >>>> SDN Group >>>> Office of Technology >>>> *Red Hat* >>>> <vpp-17.01_vs_ovs-2.6.pdf>_______________________________________________ >>>> vpp-dev mailing list >>>> vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io> >>>> https://lists.fd.io/mailman/listinfo/vpp-dev >>> >> >> >> -- >> Karl Rister <kris...@redhat.com> > -- Karl Rister <kris...@redhat.com> _______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io https://lists.fd.io/mailman/listinfo/vpp-dev