I think i found solution, it improve little bit performance on tapdd55b834-f8 interface default txqueuelen size was 500, I have changed it to 10000
On Thu, Dec 17, 2015 at 12:07 PM, Satish Patel <satish....@gmail.com> wrote: > On host (compute) machine tap device showing lots of TX packet loss. > > tapdd55b834-f8: flags=4163<UP,BROADCAST,RUNNING,MULTICAST> mtu 1500 > inet6 fe80::fc16:3eff:fecd:2aa4 prefixlen 64 scopeid 0x20<link> > ether fe:16:3e:cd:2a:a4 txqueuelen 500 (Ethernet) > RX packets 1060731031 bytes 298573599135 (278.0 GiB) > RX errors 0 dropped 0 overruns 0 frame 0 > TX packets 503372157 bytes 81345389820 (75.7 GiB) > TX errors 0 dropped 171333348 overruns 0 carrier 0 collisions 0 > > On Thu, Dec 17, 2015 at 11:45 AM, Satish Patel <satish....@gmail.com> wrote: >> I just ran iperf and here is the results, bi-directional >> >> [ ID] Interval Transfer Bandwidth >> [ 5] 0.0-10.0 sec 1.04 GBytes 895 Mbits/sec >> [ 4] 0.0-10.0 sec 1.09 GBytes 935 Mbits/sec >> >> Look like when guest is under load, it can't handle traffic and start >> dropping packets. >> >> >> >> On Thu, Dec 17, 2015 at 11:36 AM, Vahric Muhtaryan <vah...@doruk.net.tr> >> wrote: >>> Maybe this also could be reference >>> >>> http://42.62.73.30/wordpress/wp-content/uploads/2013/10/Neutron-performance >>> -testing.pdf >>> >>> >>> On 17/12/15 18:32, "Satish Patel" <satish....@gmail.com> wrote: >>> >>>>We have our own custom application running on guest VM. i am just >>>>checking performance of VM. but i found if load go little up machine >>>>start breaking ping. >>>> >>>>I didn't run iperf test yet. which i am going to run in a min. >>>> >>>>On Thu, Dec 17, 2015 at 11:28 AM, Vahric Muhtaryan <vah...@doruk.net.tr> >>>>wrote: >>>>> Okay Satish , >>>>> >>>>> Im not expert and do not want to divert you wrong direction. >>>>> I want to make your test on my server to cross check , can you pls >>>>>explain >>>>> how and with what tool and to where you are making this test. >>>>> >>>>> Regards >>>>> VM >>>>> >>>>> On 17/12/15 18:11, "Satish Patel" <satish....@gmail.com> wrote: >>>>> >>>>>>Thanks for reply, >>>>>> >>>>>>Following is my guest xml file. >>>>>> >>>>>>I am using Openstack JUNO and it use OVS. >>>>>> >>>>>><interface type='bridge'> >>>>>> <mac address='fa:16:3e:cd:2a:a4'/> >>>>>> <source bridge='qbrdd55b834-f8'/> >>>>>> <target dev='tapdd55b834-f8'/> >>>>>> <model type='virtio'/> >>>>>> <alias name='net0'/> >>>>>> <address type='pci' domain='0x0000' bus='0x00' slot='0x03' >>>>>>function='0x0'/> >>>>>> </interface> >>>>>> >>>>>> >>>>>>Following is TOP command on guest machine, at this point i am getting >>>>>>ping breaks >>>>>> >>>>>>top - 16:10:30 up 20:46, 1 user, load average: 4.63, 4.41, 3.60 >>>>>>Tasks: 165 total, 6 running, 159 sleeping, 0 stopped, 0 zombie >>>>>>%Cpu0 : 15.1 us, 12.3 sy, 0.0 ni, 60.9 id, 0.0 wa, 0.0 hi, 0.3 si, >>>>>>11.4 st >>>>>>%Cpu1 : 22.9 us, 17.2 sy, 0.0 ni, 51.4 id, 0.0 wa, 0.0 hi, 0.3 si, >>>>>>8.2 st >>>>>>%Cpu2 : 28.8 us, 22.4 sy, 0.0 ni, 47.5 id, 0.0 wa, 0.0 hi, 1.0 si, >>>>>>0.3 st >>>>>>%Cpu3 : 16.6 us, 15.0 sy, 0.0 ni, 66.4 id, 0.0 wa, 0.0 hi, 0.3 si, >>>>>>1.7 st >>>>>>%Cpu4 : 9.8 us, 11.8 sy, 0.0 ni, 0.0 id, 75.4 wa, 0.0 hi, 0.3 si, >>>>>>2.6 st >>>>>>%Cpu5 : 7.6 us, 6.1 sy, 0.0 ni, 81.4 id, 0.0 wa, 0.0 hi, 4.2 si, >>>>>>0.8 st >>>>>>%Cpu6 : 8.1 us, 7.4 sy, 0.0 ni, 83.0 id, 0.0 wa, 0.0 hi, 1.4 si, >>>>>>0.0 st >>>>>>%Cpu7 : 17.8 us, 17.8 sy, 0.0 ni, 64.1 id, 0.0 wa, 0.0 hi, 0.3 si, >>>>>>0.0 st >>>>>>KiB Mem : 8175332 total, 4630124 free, 653284 used, 2891924 >>>>>>buff/cache >>>>>>KiB Swap: 0 total, 0 free, 0 used. 7131540 avail >>>>>>Mem >>>>>> >>>>>> >>>>>>On Thu, Dec 17, 2015 at 11:06 AM, Vahric Muhtaryan <vah...@doruk.net.tr> >>>>>>wrote: >>>>>>> Hello , >>>>>>> >>>>>>> Cool ! Todays I am testing to but not such way just only KVM. >>>>>>> >>>>>>> Wonder 2/3 things ; >>>>>>> >>>>>>> 1) Are you using virtio? >>>>>>> <model type='virtio¹/> >>>>>>> >>>>>>> >>>>>>> 2) is your vm saturating %100 of cpu inside guest ? >>>>>>> >>>>>>> 3) Using OVS or bridge ? >>>>>>> >>>>>>> Regards >>>>>>> Vahric Muhtaryan >>>>>>> >>>>>>> On 17/12/15 17:41, "Satish Patel" <satish....@gmail.com> wrote: >>>>>>> >>>>>>>>I am doing some testing on Openstack VM network performance. following >>>>>>>>i am doing >>>>>>>> >>>>>>>>Comute node: 8 CPU / 32GB mem >>>>>>>>VM instance: 8vCPU / 8GB mem >>>>>>>> >>>>>>>>when i am running heavy load application, immediate my ping started >>>>>>>>breaking for VM.. but at same time if i ping Compute Host, its working >>>>>>>>fine without any packet loss. >>>>>>>> >>>>>>>>I have disabled GSO/TSO setting on eth0 but still same result. i have >>>>>>>>notice when it reach 250mbps my ping start breaking... >>>>>>>> >>>>>>>>_______________________________________________ >>>>>>>>Mailing list: >>>>>>>>http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>>>>>>Post to : openstack@lists.openstack.org >>>>>>>>Unsubscribe : >>>>>>>>http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>>>>> >>>>>>> >>>>> >>>>> >>> >>> _______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack