Liping, Last 2 days i am running test with hping3 and found following behavior, if you noticed my result UDP doing very bad if i increase number of queue, do you know why ?
UDP: If i set "ethtool -L eth0 combined 1" then UDP pps rate is 100kpps if i set "ethtool -L eth0 combined 8" then UDP pps rate is 40kpps TCP: If i set "ethtool -L eth0 combined 1" then UDP pps rate is ~150kpps If i set "ethtool -L eth0 combined 1" then UDP pps rate is ~150kpps On Mon, Sep 17, 2018 at 8:33 AM Satish Patel <satish....@gmail.com> wrote: > > Thanks Liping, > > I will try to reach out or open new thread to get sriov info. > > By the way what version of openstack you guys using and what hardware > specially NIC. Just trying to see if it's hardware related. > > I'm running kernel 3.10.x do you think it's not something related kernel. > > Sent from my iPhone > > On Sep 17, 2018, at 1:27 AM, Liping Mao (limao) <li...@cisco.com> wrote: > > >> Question: I have br-vlan interface mapp with bond0 to run my VM (VLAN > > > > traffic), so do i need to do anything in bond0 to enable VF/PF > > > > function? Just confused because currently my VM nic map with compute > > > > node br-vlan bridge. > > > > > > > > I had not actually used SRIOV in my env~ maybe others could help. > > > > > > > > Thanks, > > > > Liping Mao > > > > > > > > 在 2018/9/17 11:48,“Satish Patel”<satish....@gmail.com> 写入: > > > > > > > > Thanks Liping, > > > > > > > > I will check bug for tx/rx queue size and see if i can make it work > > > > but look like my 10G NIC support SR-IOV so i am trying that path > > > > because it will be better for long run. > > > > > > > > I have deploy my cloud using openstack-ansible so now i need to figure > > > > out how do i wire that up with openstack-ansible deployment, here is > > > > the article [1] > > > > > > > > Question: I have br-vlan interface mapp with bond0 to run my VM (VLAN > > > > traffic), so do i need to do anything in bond0 to enable VF/PF > > > > function? Just confused because currently my VM nic map with compute > > > > node br-vlan bridge. > > > > > > > > [root@compute-65 ~]# lspci -nn | grep -i ethernet > > > > 03:00.0 Ethernet controller [0200]: Broadcom Limited NetXtreme II > > > > BCM57810 10 Gigabit Ethernet [14e4:168e] (rev 10) > > > > 03:00.1 Ethernet controller [0200]: Broadcom Limited NetXtreme II > > > > BCM57810 10 Gigabit Ethernet [14e4:168e] (rev 10) > > > > 03:01.0 Ethernet controller [0200]: Broadcom Limited NetXtreme II > > > > BCM57810 10 Gigabit Ethernet Virtual Function [14e4:16af] > > > > 03:01.1 Ethernet controller [0200]: Broadcom Limited NetXtreme II > > > > BCM57810 10 Gigabit Ethernet Virtual Function [14e4:16af] > > > > 03:01.2 Ethernet controller [0200]: Broadcom Limited NetXtreme II > > > > BCM57810 10 Gigabit Ethernet Virtual Function [14e4:16af] > > > > 03:01.3 Ethernet controller [0200]: Broadcom Limited NetXtreme II > > > > BCM57810 10 Gigabit Ethernet Virtual Function [14e4:16af] > > > > 03:01.4 Ethernet controller [0200]: Broadcom Limited NetXtreme II > > > > BCM57810 10 Gigabit Ethernet Virtual Function [14e4:16af] > > > > 03:01.5 Ethernet controller [0200]: Broadcom Limited NetXtreme II > > > > BCM57810 10 Gigabit Ethernet Virtual Function [14e4:16af] > > > > 03:01.6 Ethernet controller [0200]: Broadcom Limited NetXtreme II > > > > BCM57810 10 Gigabit Ethernet Virtual Function [14e4:16af] > > > > > > > > > > > > [1] > > https://docs.openstack.org/openstack-ansible-os_neutron/latest/configure-network-services.html > > > >> On Sun, Sep 16, 2018 at 7:06 PM Liping Mao (limao) <li...@cisco.com> > >> wrote: > >> > >> > > > >> Hi Satish, > > > >> > > > >> > > > >> > > > >> > > > >> > > > >> There are hard limitations in nova's code, I did not actually used more > >> thant 8 queues: > > > >> > > > >> def _get_max_tap_queues(self): > > > >> > > > >> # NOTE(kengo.sakai): In kernels prior to 3.0, > > > >> > > > >> # multiple queues on a tap interface is not supported. > > > >> > > > >> # In kernels 3.x, the number of queues on a tap interface > > > >> > > > >> # is limited to 8. From 4.0, the number is 256. > > > >> > > > >> # See: https://bugs.launchpad.net/nova/+bug/1570631 > > > >> > > > >> kernel_version = int(os.uname()[2].split(".")[0]) > > > >> > > > >> if kernel_version <= 2: > > > >> > > > >> return 1 > > > >> > > > >> elif kernel_version == 3: > > > >> > > > >> return 8 > > > >> > > > >> elif kernel_version == 4: > > > >> > > > >> return 256 > > > >> > > > >> else: > > > >> > > > >> return None > > > >> > > > >> > > > >> > > > >>> I am currently playing with those setting and trying to generate > > > >> > > > >> traffic with hping3 tools, do you have any tool to test traffic > > > >> > > > >> performance for specially udp style small packets. > > > >> > > > >> > > > >> > > > >> Hping3 is good enough to reproduce it, we have app level test tool, but > >> that is not your case. > > > >> > > > >> > > > >> > > > >> > > > >> > > > >>> Here i am trying to increase rx_queue_size & tx_queue_size but its not > > > >> > > > >> working somehow. I have tired following. > > > >> > > > >> > > > >> > > > >> Since you are not rocky code, it should only works in qemu.conf, maybe > >> check if this bug[1] affect you. > > > >> > > > >> > > > >> > > > >> > > > >> > > > >>> Is there a way i can automate this last task to update queue number > > > >> > > > >> action after reboot vm :) otherwise i can use cloud-init to make sure > > > >> > > > >> all VM build with same config. > > > >> > > > >> > > > >> > > > >> Cloud-init or rc.local could be the place to do that. > > > >> > > > >> > > > >> > > > >> [1] https://bugzilla.redhat.com/show_bug.cgi?id=1541960 > > > >> > > > >> > > > >> > > > >> Regards, > > > >> > > > >> Liping Mao > > > >> > > > >> > > > >> > > > >> 在 2018/9/17 04:09,“Satish Patel”<satish....@gmail.com> 写入: > > > >> > > > >> > > > >> > > > >> Update on my last email. > > > >> > > > >> > > > >> > > > >> I am able to achieve 150kpps with queue=8 and my goal is to do 300kpps > > > >> > > > >> because some of voice application using 300kps. > > > >> > > > >> > > > >> > > > >> Here i am trying to increase rx_queue_size & tx_queue_size but its not > > > >> > > > >> working somehow. I have tired following. > > > >> > > > >> > > > >> > > > >> 1. add rx/tx size in /etc/nova/nova.conf in libvirt section - (didn't > >> work) > > > >> > > > >> 2. add /etc/libvirtd/qemu.conf - (didn't work) > > > >> > > > >> > > > >> > > > >> I have try to edit virsh edit <XML> file but somehow my changes not > > > >> > > > >> getting reflected, i did virsh define <XML> after change and hard > > > >> > > > >> reboot guest but no luck.. how do i edit that option in xml if i want > > > >> > > > >> to do that? > > > >> > > > >>> On Sun, Sep 16, 2018 at 1:41 PM Satish Patel <satish....@gmail.com> > >>> wrote: > >> > >> > > > >>> > > > >> > > > >>> I successful reproduce this error with hping3 tool and look like > > > >> > > > >>> multiqueue is our solution :) but i have few question you may have > > > >> > > > >>> answer of that. > > > >> > > > >>> > > > >> > > > >>> 1. I have created two instance (vm1.example.com & vm2.example.com) > > > >> > > > >>> > > > >> > > > >>> 2. I have flood traffic from vm1 using "hping3 vm2.example.com > > > >> > > > >>> --flood" and i have noticed drops on tap interface. ( This is without > > > >> > > > >>> multiqueue) > > > >> > > > >>> > > > >> > > > >>> 3. Enable multiqueue in image and run same test and again got packet > > > >> > > > >>> drops on tap interface ( I didn't update queue on vm2 guest, so > > > >> > > > >>> definitely i was expecting packet drops) > > > >> > > > >>> > > > >> > > > >>> 4. Now i have try to update vm2 queue using ethtool and i got > > > >> > > > >>> following error, I have 15vCPU and i was trying to add 15 queue > > > >> > > > >>> > > > >> > > > >>> [root@bar-mq ~]# ethtool -L eth0 combined 15 > > > >> > > > >>> Cannot set device channel parameters: Invalid argument > > > >> > > > >>> > > > >> > > > >>> Then i have tried 8 queue which works. > > > >> > > > >>> > > > >> > > > >>> [root@bar-mq ~]# ethtool -L eth0 combined 8 > > > >> > > > >>> combined unmodified, ignoring > > > >> > > > >>> no channel parameters changed, aborting > > > >> > > > >>> current values: tx 0 rx 0 other 0 combined 8 > > > >> > > > >>> > > > >> > > > >>> Now i am not seeing any packet drops on tap interface, I have measure > > > >> > > > >>> PPS and i was able to get 160kpps without packet drops. > > > >> > > > >>> > > > >> > > > >>> Question: > > > >> > > > >>> > > > >> > > > >>> 1. why i am not able to add 15 queue? ( is this NIC or driver > >>> limitation?) > > > >> > > > >>> 2. how do i automate "ethtool -L eth0 combined 8" command in instance > > > >> > > > >>> so i don't need to tell my customer to do this manually? > > > >> > > > >>>> On Sun, Sep 16, 2018 at 11:53 AM Satish Patel <satish....@gmail.com> > >>>> wrote: > >> > >> > > > >>>> > > > >> > > > >>>> Hi Liping, > > > >> > > > >>>> > > > >> > > > >>>>>> I think multi queue feature should help.(be careful to make sure the > >>>>>> ethtool update queue number action also did after reboot the vm). > > > >> > > > >>>> > > > >> > > > >>>> Is there a way i can automate this last task to update queue number > > > >> > > > >>>> action after reboot vm :) otherwise i can use cloud-init to make sure > > > >> > > > >>>> all VM build with same config. > > > >> > > > >>>>> On Sun, Sep 16, 2018 at 11:51 AM Satish Patel <satish....@gmail.com> > >>>>> wrote: > >> > >> > > > >>>>> > > > >> > > > >>>>> I am currently playing with those setting and trying to generate > > > >> > > > >>>>> traffic with hping3 tools, do you have any tool to test traffic > > > >> > > > >>>>> performance for specially udp style small packets. > > > >> > > > >>>>> > > > >> > > > >>>>> I am going to share all my result and see what do you feel because i > > > >> > > > >>>>> have noticed you went through this pain :) I will try every single > > > >> > > > >>>>> option which you suggested to make sure we are good before i move > > > >> > > > >>>>> forward to production. > > > >> > > > >>>>>> On Sun, Sep 16, 2018 at 11:25 AM Liping Mao (limao) <li...@cisco.com> > >>>>>> wrote: > >> > >> > > > >>>>>> > > > >> > > > >>>>>> I think multi queue feature should help.(be careful to make sure the > >>>>>> ethtool update queue number action also did after reboot the vm). > > > >> > > > >>>>>> > > > >> > > > >>>>>> Numa cpu pin and queue length will be a plus in my exp. You may need > >>>>>> yo have performance test in your situatuon,in my case cpu numa helpped > >>>>>> the app get very stable 720p/1080p transcoding performance. Not sure > >>>>>> if your app get benifit. > > > >> > > > >>>>>> > > > >> > > > >>>>>> You are not using L3,this will let you avoid a lot of performance > >>>>>> issue. And since only two instance with 80kpps packets,so in your > >>>>>> case,HW interface should not be bottleneck too. And your Nexus 5k/7k > >>>>>> will not be bottleneck for sure ;-) > > > >> > > > >>>>>> > > > >> > > > >>>>>> > > > >> > > > >>>>>> Thanks, > > > >> > > > >>>>>> Liping Mao > > > >> > > > >>>>>> > > > >> > > > >>>>>>>> 在 2018年9月16日,23:09,Satish Patel <satish....@gmail.com> 写道: > >> > >> > > > >>>>>>> > > > >> > > > >>>>>>> Thanks Liping, > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> I am using libvertd 3.9.0 version so look like i am eligible take > > > >> > > > >>>>>>> advantage of that feature. phew! > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> [root@compute-47 ~]# libvirtd -V > > > >> > > > >>>>>>> libvirtd (libvirt) 3.9.0 > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> Let me tell you how i am running instance on my openstack, my compute > > > >> > > > >>>>>>> has 32 core / 32G memory and i have created two instance on compute > > > >> > > > >>>>>>> node 15vcpu and 14G memory ( two instance using 30 vcpu core, i have > > > >> > > > >>>>>>> kept 2 core for compute node). on compute node i disabled overcommit > > > >> > > > >>>>>>> using ratio (1.0) > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> I didn't configure NUMA yet because i wasn't aware of this feature, as > > > >> > > > >>>>>>> per your last post do you think numa will help to fix this issue? > > > >> > > > >>>>>>> following is my numa view > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> [root@compute-47 ~]# numactl --hardware > > > >> > > > >>>>>>> available: 2 nodes (0-1) > > > >> > > > >>>>>>> node 0 cpus: 0 1 2 3 4 5 6 7 16 17 18 19 20 21 22 23 > > > >> > > > >>>>>>> node 0 size: 16349 MB > > > >> > > > >>>>>>> node 0 free: 133 MB > > > >> > > > >>>>>>> node 1 cpus: 8 9 10 11 12 13 14 15 24 25 26 27 28 29 30 31 > > > >> > > > >>>>>>> node 1 size: 16383 MB > > > >> > > > >>>>>>> node 1 free: 317 MB > > > >> > > > >>>>>>> node distances: > > > >> > > > >>>>>>> node 0 1 > > > >> > > > >>>>>>> 0: 10 20 > > > >> > > > >>>>>>> 1: 20 10 > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> I am not using any L3 router, i am using provide VLAN network and > > > >> > > > >>>>>>> using Cisco Nexus switch for my L3 function so i am not seeing any > > > >> > > > >>>>>>> bottleneck there. > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> This is the 10G NIC i have on all my compute node, dual 10G port with > > > >> > > > >>>>>>> bonding (20G) > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> 03:00.0 Ethernet controller: Broadcom Limited NetXtreme II BCM57810 10 > > > >> > > > >>>>>>> Gigabit Ethernet (rev 10) > > > >> > > > >>>>>>> 03:00.1 Ethernet controller: Broadcom Limited NetXtreme II BCM57810 10 > > > >> > > > >>>>>>> Gigabit Ethernet (rev 10) > > > >> > > > >>>>>>> > > > >> > > > >>>>>>> > > > >> > > > >>>>>>>>> On Sun, Sep 16, 2018 at 10:50 AM Liping Mao (limao) > >>>>>>>>> <li...@cisco.com> wrote: > >> > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> It is still possible to update rx and tx queues length if your qemu > >>>>>>>> and libvirt version is higher than the version recorded in [3]. (You > >>>>>>>> should possible to update directly in libvirt configuration if my > >>>>>>>> memory is correct) > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> We also have some similar use case which run audio/vedio serivcs. > >>>>>>>> They are CPU consuming and have UDP small packets. Another possible > >>>>>>>> tunning is using CPU pin for the vm. you can use numa awared cpu > >>>>>>>> feature to get stable cpu performance ,vm network dropped packets > >>>>>>>> sometimes because of the vm cpu is too busy,with numa cpu it works > >>>>>>>> better performance,our way is similar with [a]. You need to create > >>>>>>>> flavor with special metadata and dedicated Host Agg for numa awared > >>>>>>>> VMs. Dedicated CPU is very good for media service. It makes the CPU > >>>>>>>> performance stable. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Another packet loss case we get is because of vm kernel, some of our > >>>>>>>> app are using 32bit OS, that cause memory issue, when traffic larger > >>>>>>>> then 50kpps, it dropped a lot,sometimes,it even crash. In this case, > >>>>>>>> 32bit os can actually use very limited memory, we have to add swap > >>>>>>>> for the vm. Hope your app is using 64 bit OS. Because 32 bit could > >>>>>>>> cause tons of trouble. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> BTW,if you are using vrouter on L3, you’d better to move provider > >>>>>>>> network(no vrouter). I did not tried DVR, but if you are running > >>>>>>>> without DVR, the L3 node will be bottleneck very quick. Especially > >>>>>>>> default iptables conntrack is 65535, you will reach to it and drop > >>>>>>>> packet on L3, even after you tun that value, it still hard to more > >>>>>>>> that 1Mpps for your network node. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> If your App more than 200kpps per compute node, you may be better > >>>>>>>> also have a look your physical network driver tx/rx configuration. > >>>>>>>> Most of the HW default value for tx/rx queues number and length are > >>>>>>>> very poor,you may start to get packet on eth interface on physical > >>>>>>>> host when rx queue is full. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> [a]https://redhatstackblog.redhat.com/2015/05/05/cpu-pinning-and-numa-topology-awareness-in-openstack-compute/ > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Regards, > > > >> > > > >>>>>>>> Liping Mao > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>>> 在 2018年9月16日,21:18,Satish Patel <satish....@gmail.com> 写道: > >> > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Hi Liping, > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Thank you for your reply, > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> We notice packet drops during high load, I did try txqueue and > >>>>>>>> didn't help so I believe I am going to try miltiqueue. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> For SRIOV I have to look if I have support in my nic. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> We are using queens so I think queue size option not possible :( > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> We are using voip application and traffic is udp so our pps rate is > >>>>>>>> 60k to 80k per vm instance. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> I will share my result as soon as I try multiqueue. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Sent from my iPhone > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>>> On Sep 16, 2018, at 2:27 AM, Liping Mao (limao) <li...@cisco.com> > >>>>>>>>> wrote: > >> > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Hi Satish, > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Did your packet loss happen always or it only happened when heavy > >>>>>>>> load? > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> AFAIK, if you do not tun anything, the vm tap can process about > >>>>>>>> 50kpps before the tap device start to drop packets. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> If it happened in heavy load, couple of things you can try: > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> 1) increase tap queue length, usually the default value is 500, you > >>>>>>>> can try larger. (seems like you already tried) > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> 2) Try to use virtio multi queues feature , see [1]. Virtio use one > >>>>>>>> queue for rx/tx in vm, with this feature you can get more queues. > >>>>>>>> You can check > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> 3) In rock version, you can use [2] to increase virtio queue size, > >>>>>>>> the default queues size is 256/512, you may increase it to 1024, > >>>>>>>> this would help to increase pps of the tap device. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> If all these things can not get your network performance > >>>>>>>> requirement, you may need to move to use dpdk / sriov stuff to get > >>>>>>>> more vm performance. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> I did not actually used them in our env, you may refer to [3] > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> [1] > >>>>>>>> https://specs.openstack.org/openstack/nova-specs/specs/liberty/implemented/libvirt-virtiomq.html > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> [2] > >>>>>>>> https://specs.openstack.org/openstack/nova-specs/specs/rocky/implemented/libvirt-virtio-set-queue-sizes.html > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> [3] > >>>>>>>> https://docs.openstack.org/ocata/networking-guide/config-sriov.html > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Regards, > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Liping Mao > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> 在 2018/9/16 13:07,“Satish Patel”<satish....@gmail.com> 写入: > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> [root@compute-33 ~]# ifconfig tap5af7f525-5f | grep -i drop > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> RX errors 0 dropped 0 overruns 0 frame 0 > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> TX errors 0 dropped 2528788837 overruns 0 carrier 0 > >>>>>>>> collisions 0 > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Noticed tap interface dropping TX packets and even after increasing > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> txqueue from 1000 to 10000 nothing changed, still getting packet > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> drops. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>>> On Sat, Sep 15, 2018 at 4:22 PM Satish Patel <satish....@gmail.com> > >>>>>>>>> wrote: > >> > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Folks, > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> I need some advice or suggestion to find out what is going on with my > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> network, we have notice high packet loss on openstack instance and > >>>>>>>> not > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> sure what is going on, same time if i check on host machine and it > >>>>>>>> has > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> zero packet loss.. this is what i did for test... > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> ping 8.8.8.8 > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> from instance: 50% packet loss > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> from compute host: 0% packet loss > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> I have disabled TSO/GSO/SG setting on physical compute node but still > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> getting packet loss. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> We have 10G NIC on our network, look like something related to tap > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> interface setting.. > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> _______________________________________________ > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Mailing list: > >>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Post to : openstack@lists.openstack.org > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> Unsubscribe : > >>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >>>>>>>> > > > >> > > > >> > > > >> > > > >> > > > > > > > > _______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack