On 10/25/2016 08:03 AM, Thomas F Herbert wrote: > Pierre, Thanks. > > > On 10/25/2016 03:14 AM, Pierre Pfister (ppfister) wrote: >> Hello, >> >> For now the multi-queue patch is still stuck in gerrit because >> jenkin's qemu is using an old buggy version... > Is this true for all CSIT test platforms including the one running Centos? >> I made some measurements on >> vhost: FD.io_mini-summit_916_Vhost_Performance_and_Optimization.pptx >> <https://wiki.fd.io/images/c/cc/FD.io_mini-summit_916_Vhost_Performance_and_Optimization.pptx> > Thanks! >> >> I see you try different combinations with and without mergeable >> descriptors. >> Do you do the same with 'indirect descriptors' ? They are supported by >> VPP since september or so. >> The issue with these zillions ways a buffer may be forwarded is that >> we only know what mode is enabled or disabled, but you never know >> exactly what is happening for real. > Yes, I think the test should be re-run on 16.12-rc0 to pick up latest > patches mentioned in Pierre's presentation. Also, I see you merged a > patch that fixes branch prediction.
I installed 16.12-rc0~247_g9c2964c~b1272 and am re-running the same configurations before changing anything. Once the baselines are done I'll try the indirect descriptors. >> >> Using indirect descriptors, I got VPP doing 0% loss 10Mpps (5Mpps each >> way). And the setup was stricter than yours as VPP had only 2 threads >> on the same core. > Is this with multi-queue. In your opinion, Pierre, do you think upstream > dpdk merge will improve perf? These numbers are all single queue for now. >> >> You may also want to try 'chrt -r' on your working processes. This >> improves scheduling real-time properties. All PMD threads in the host and guest are running with SCHED_FIFO priority 95 on isolated CPU threads. If/when we get to packing multiple PMD threads on the same CPU thread I'll probably have to experiment with SCHED_RR. >> >> Thanks, >> >> - Pierre >> >> >> >> >>> Le 25 oct. 2016 à 06:36, Jerome Tollet (jtollet) <jtol...@cisco.com >>> <mailto:jtol...@cisco.com>> a écrit : >>> >>> + Pierre Pfister (ppfister) who ran a lot of benchmarks for VPP/vhostuser >>> >>> *De : *<<mailto:vpp-dev-boun...@lists.fd.io>vpp-dev-boun...@lists.fd.io> >>> au nom de Thomas F Herbert >>> <<mailto:therb...@redhat.com>therb...@redhat.com> >>> *Date : *lundi 24 octobre 2016 à 21:32 >>> *À : *"<mailto:kris...@redhat.com>kris...@redhat.com" >>> <kris...@redhat.com <mailto:kris...@redhat.com>>, Andrew Theurer >>> <atheu...@redhat.com <mailto:atheu...@redhat.com>>, Franck Baudin >>> <fbau...@redhat.com <mailto:fbau...@redhat.com>>, Rashid Khan >>> <rk...@redhat.com <mailto:rk...@redhat.com>>, Bill Michalowski >>> <bmich...@redhat.com <mailto:bmich...@redhat.com>>, Billy McFall >>> <bmcf...@redhat.com <mailto:bmcf...@redhat.com>>, Douglas Shakshober >>> <dsh...@redhat.com <mailto:dsh...@redhat.com>> >>> *Cc : *vpp-dev <vpp-dev@lists.fd.io <mailto:vpp-dev@lists.fd.io>>, >>> "Damjan Marion (damarion)" >>> <<mailto:damar...@cisco.com>damar...@cisco.com> >>> *Objet : *Re: [vpp-dev] updated ovs vs. vpp results for 0.002% and 0% >>> loss >>> >>> >>> +Maciek Konstantynowicz CSIT (mkonstan) >>> >>> +vpp-dev >>> >>> +Damjan Marion (damarion) >>> >>> Karl, Thanks! >>> >>> Your results seem close to consistent with VPP's CSIT testing for >>> vhost for 16.09 but for broader visibility, I am including some >>> people on the VPP team, Damjan who is working on multi-queue etc. (I >>> see that there were some perf related patches merged in vhost that >>> might help since 16.09.) and Maciek who works in the CSIT project and >>> has done the testing of VPP. >>> >>> I want to open up the discussion WRT to the following: >>> >>> 1, Optimizing for maximum vhost perf with vpp including vhost-user >>> multi-queue. >>> >>> 2. Comparision with CSIT results for vhost. Following are two links >>> for CSIT >>> >>> 3. Statistics: >>> >>> 4. Tuning suggestions: >>> >>> Following are some CSIT results: >>> >>> compiled 16.09 results for >>> vhost-user: >>> <https://wiki.fd.io/view/CSIT/VPP-16.09_Test_Report#VM_vhost-user_Throughput_Measurements>https://wiki.fd.io/view/CSIT/VPP-16.09_Test_Report#VM_vhost-user_Throughput_Measurements >>> >>> Latest CSIT output from top of master, 16.12-rc0 >>> >>> https://jenkins.fd.io/view/csit/job/csit-vpp-verify-perf-master-nightly-all/1085/console >>> >>> --Tom >>> >>> On 10/21/2016 04:06 PM, Karl Rister wrote: >>> >>> Hi All >>> >>> >>> >>> Below are updated performance results for OVS and VPP on our new >>> >>> Broadwell testbed. I've tried to include all the relevant details, let >>> >>> me know if I have forgotten anything of interest to you. >>> >>> >>> >>> Karl >>> >>> >>> >>> >>> >>> >>> >>> Processor: Intel(R) Xeon(R) CPU E5-2690 v4 @ 2.60GHz (Broadwell) >>> >>> Environment: RT + Hyperthreading (see [1] for details on KVM-RT) >>> >>> Kernel: 3.10.0-510.rt56.415.el7.x86_64 >>> >>> Tuned: 2.7.1-3.el7 >>> >>> >>> >>> /proc/cmdline: >>> >>> <...> default_hugepagesz=1G iommu=pt intel_iommu=on isolcpus=4-55 >>> >>> nohz=on nohz_full=4-55 rcu_nocbs=4-55 intel_pstate=disable nosoftlockup >>> >>> >>> >>> Versions: >>> >>> - OVS: openvswitch-2.5.0-10.git20160727.el7fdb + BZ fix [2] >>> >>> - VPP: v16.09 >>> >>> >>> >>> NUMA node 0 CPU sibling pairs: >>> >>> - (0,28)(2,30)(4,32)(6,34)(8,36)(10,38)(12,40)(14,42)(16,44)(18,46) >>> >>> (20,48)(22,50)(24,52)(26,54) >>> >>> >>> >>> Host PMD Assignment: >>> >>> - dpdk0 = CPU 6 >>> >>> - vhost-user1 = CPU 34 >>> >>> - dpdk1 = CPU 8 >>> >>> - vhost-user2 = CPU 36 >>> >>> >>> >>> Guest CPU Assignment: >>> >>> - Emulator = CPU 20 >>> >>> - VCPU 0 (Housekeeping) = CPU 22 >>> >>> - VCPU 1 (PMD) = CPU 24 >>> >>> - VCPU 2 (PMD) = CPU 26 >>> >>> >>> >>> Configuration Details: >>> >>> - OVS: custom OpenFlow rules direct packets similarly to VPP L2 xconnect >>> >>> - VPP: L2 xconnect >>> >>> - DPDK v16.07.0 testpmd in guest >>> >>> - SCHED_FIFO priority 95 applied to all PMD threads (OVS/VPP/testpmd) >>> >>> - SCHED_FIFO priority 1 applied to Guest VCPUs used for PMDs >>> >>> >>> >>> Test Parameters: >>> >>> - 64B packet size >>> >>> - L2 forwarding test >>> >>> - All tests are bidirectional PVP (physical<->virtual<->physical) >>> >>> - Packets enter on a NIC port and are forwarded to the guest >>> >>> - Inside the guests, received packets are sent out the opposite >>> >>> direction >>> >>> - Binary search starting at line rate (14.88 Mpps each way) >>> >>> - 10 Minute Search Duration >>> >>> - 2 Hour Validation Duration follows passing run for 10 Minute Search >>> >>> - If validation fails, search continues >>> >>> >>> >>> Mergeable Buffers Disabled: >>> >>> - OVS: >>> >>> - 0.002% Loss: 11.5216 Mpps bidirectional (5.7608 Mpps each way) >>> >>> - 0% Loss: 11.5216 Mpps bidirectional (5.7608 Mpps each way) >>> >>> - VPP: >>> >>> - 0.002% Loss: 7.5537 Mpps bidirectional (3.7769 Mpps each way)Andre >>> >>> Fredette <afred...@redhat.com> <mailto:afred...@redhat.com> >>> >>> - 0% Loss: 5.2971 Mpps bidirectional (2.6486 Mpps each way) >>> >>> >>> >>> Mergeable Buffers Enabled: >>> >>> - OVS: >>> >>> - 0.002% Loss: 6.5626 Mpps bidirectional (3.2813 Mpps each way) >>> >>> - 0% Loss: 6.3622 Mpps bidirectional (3.1811 Mpps each way) >>> >>> - VPP: >>> >>> - 0.002% Loss: 7.8134 Mpps bidirectional (3.9067 Mpps each way) >>> >>> - 0% Loss: 5.1029 Mpps bidirectional (2.5515 Mpps each way) >>> >>> >>> >>> Mergeable Buffers Disabled + VPP no-multi-seg: >>> >>> - VPP: >>> >>> - 0.002% Loss: 8.0654 Mpps bidirectional (4.0327 Mpps each way) >>> >>> - 0% Loss: 5.6442 Mpps bidirectional (2.8221 Mpps each way) >>> >>> >>> >>> The details of these results (including latency metrics and links to the >>> >>> raw data) are available at [3]. >>> >>> >>> >>> [1]: https://virt-wiki.lab.eng.brq.redhat.com/KVM/RealTime >>> >>> [2]: https://bugzilla.redhat.com/show_bug.cgi?id=1344787 >>> >>> [3]: >>> >>> >>> https://docs.google.com/a/redhat.com/spreadsheets/d/1K6zDVgZYPJL-7EsIYMBIZCn65NAkVL_GtkBrAnAdXao/edit?usp=sharing >>> >>> >>> >>> >>> -- >>> *Thomas F Herbert* >>> SDN Group >>> Office of Technology >>> *Red Hat* >> > > -- > *Thomas F Herbert* > SDN Group > Office of Technology > *Red Hat* -- Karl Rister <kris...@redhat.com> _______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io https://lists.fd.io/mailman/listinfo/vpp-dev