HI All, After configuring hugepages of size 2048KiB at the host,I am not seeing any improvement in the performance.Still after pumping more than 4Mpps i am seeing this counter getting incremented rx_no_dma_resources: 93356 at the host.
output of following commands are mentioned below-: vpp# show interface Name Idx State Counter Count VirtualFunctionEthernet0/a/0 1 up rx packets 177953727 rx bytes 11389034442 tx packets 177952680 tx bytes 16371646028 drops 1052 ip4 177952663 local0 0 down vpp# show run Thread 0 vpp_main (lcore 0) Time 393.8, average vectors/node 1.00, last 128 main loops 0.00 per node 0.00 vector rates in 0.0000e0, out 1.0158e-2, drop 2.5394e-3, punt 0.0000e0 Name State Calls Vectors Suspends Clocks Vectors/Call VirtualFunctionEthernet0/a/0-o active 4 4 0 2.69e3 1.00 VirtualFunctionEthernet0/a/0-t active 4 4 0 4.72e3 1.00 acl-plugin-fa-cleaner-process event wait 0 0 1 1.64e4 0.00 admin-up-down-process event wait 0 0 1 7.39e3 0.00 api-rx-from-ring any wait 0 0 29 2.64e5 0.00 bfd-process event wait 0 0 1 6.42e3 0.00 cdp-process any wait 0 0 41 1.28e5 0.00 dhcp-client-process any wait 0 0 4 5.04e3 0.00 dpdk-ipsec-process done 1 0 0 2.89e5 0.00 dpdk-process any wait 0 0 132 6.89e4 0.00 error-drop active 1 1 0 2.35e3 1.00 fib-walk any wait 0 0 197 8.02e3 0.00 flow-report-process any wait 0 0 1 7.02e2 0.00 flowprobe-timer-process any wait 0 0 1 4.88e3 0.00 ikev2-manager-process any wait 0 0 394 4.81e3 0.00 ioam-export-process any wait 0 0 1 1.18e3 0.00 ip4-glean active 1 1 0 1.56e4 1.00 ip4-lookup active 4 4 0 6.53e3 1.00 ip4-rewrite active 3 3 0 3.02e3 1.00 ip6-icmp-neighbor-discovery-ev any wait 0 0 394 3.78e3 0.00 l2fib-mac-age-scanner-process event wait 0 0 1 1.22e3 0.00 lisp-retry-service any wait 0 0 197 8.37e3 0.00 lldp-process event wait 0 0 1 4.15e6 0.00 memif-process event wait 0 0 1 1.74e4 0.00 nat64-expire-walk done 1 0 0 1.01e4 0.00 send-garp-na-process event wait 0 0 1 1.29e3 0.00 snat-det-expire-walk done 1 0 0 1.05e3 0.00 startup-config-process done 1 0 1 5.97e3 0.00 udp-ping-process any wait 0 0 1 1.05e4 0.00 unix-cli-sockaddr family 1 active 0 0 298 7.72e6 0.00 unix-epoll-input polling 655279 0 0 1.56e6 0.00 vhost-user-process any wait 0 0 1 1.33e3 0.00 vhost-user-send-interrupt-proc any wait 0 0 1 1.06e3 0.00 vpe-link-state-process event wait 0 0 2 5.78e3 0.00 vpe-oam-process any wait 0 0 193 5.68e3 0.00 vpe-route-resolver-process any wait 0 0 4 7.33e3 0.00 vxlan-gpe-ioam-export-process any wait 0 0 1 2.31e3 0.00 --------------- Thread 1 vpp_wk_0 (lcore 1) Time 393.8, average vectors/node 17.54, last 128 main loops 0.00 per node 0.00 vector rates in 4.5190e5, out 4.5190e5, drop 3.3698e0, punt 0.0000e0 Name State Calls Vectors Suspends Clocks Vectors/Call VirtualFunctionEthernet0/a/0-o active 10146115 177952676 0 1.70e1 17.54 VirtualFunctionEthernet0/a/0-t active 10146115 177952676 0 7.86e1 17.54 arp-input active 1337 1340 0 2.58e3 1.00 dpdk-input polling 6678250972 177954003 0 3.26e3 .03 error-drop active 1324 1327 0 8.68e2 1.00 ethernet-input active 1337 1340 0 1.87e3 1.00 interface-output active 16 16 0 8.06e2 1.00 ip4-icmp-echo-reply active 3 3 0 3.01e4 1.00 ip4-icmp-echo-request active 2 2 0 3.80e3 1.00 ip4-icmp-error active 10146097 177952658 0 1.28e2 17.54 ip4-icmp-input active 5 5 0 2.59e3 1.00 ip4-input-no-checksum active 10146102 177952663 0 4.05e1 17.54 ip4-load-balance active 2 2 0 3.67e3 1.00 ip4-local active 10146102 177952663 0 5.54e1 17.54 ip4-lookup active 20292199 355905321 0 4.37e1 17.54 ip4-rewrite active 10146099 177952660 0 3.23e1 17.54 ip4-udp-lookup active 10146097 177952658 0 3.95e1 17.54 Thanks, Rahul On Fri, Aug 11, 2017 at 1:45 AM, Kinsella, Ray <ray.kinse...@intel.com> wrote: > Hi Rahul, > > So there a few additional hoops you have to jump through to get good > performance. The biggest of which is making sure that > > 1. The host is setup to make hugepages available. > 2. QEMU is pointed at the hugepages. > 3. The guest is setup to make hugepages available. > > The following documented from DPDK covers it in some detail. > > http://dpdk.readthedocs.io/en/v16.04/sample_app_ug/vhost.html > > Ray K > > > On 09/08/2017 23:33, Rahul Negi wrote: > >> Hi All, >> I am working on an use case where i want to measure the Maximum PPS >> handled by vpp in sriov configuration.I have created a virtual machine >> on a host having specifications as follows:- >> 1.RHEL 7.3 installed >> 2.Intel X540 10 gig NIC attached >> >> I have created a virtual function from one of the interface of 10 gig >> NIC(for e.g ens3f0) and attached it to my Virtual Machine.I have >> installed ubuntu 16.04 on my Vm. My vpp version running on vm is 17.10. >> >> vpp# show version verbose >> Version: v17.10-rc0~86-g7d4a22c >> Compiled by: root >> Compile host: ubuntu >> Compile date: Wed Jul 26 18:56:51 EDT 2017 >> Compile location: /root/vpp/vpp >> Compiler: GCC 5.4.0 20160609 >> Current PID: 5006 >> >> Currently My Vm is with 4vcpu and vpp cpu model is as follow:- >> 1.one main thread >> 2.Three worker threads >> >> I am not able to get more than 4Mpps with this configuration of vpp.When >> i am pumping more than 4Mpps traffic to my VM. I can see this >> counter rx_no_dma_resources: 5628104 getting incremented at the host.As >> there is no eth stats available in sriov configuration for virtual >> functions attached to VM. >> >> Guest vcpus are pin to host physical cpus. >> >> So 4Mpps is the expected number that we can get in this configuration? >> >> Thanks, >> Rahul >> >> >> _______________________________________________ >> vpp-dev mailing list >> vpp-dev@lists.fd.io >> https://lists.fd.io/mailman/listinfo/vpp-dev >> >>
_______________________________________________ vpp-dev mailing list vpp-dev@lists.fd.io https://lists.fd.io/mailman/listinfo/vpp-dev