Hi Kaushal, Sorry about the troubles you're having with ixgbe. I was a bit confused by what exactly you're seeing so let me ask you a few questions. When you say " I see that out of 16 Rx Queues, only *One Queue* is dropping all the packets." is that queue getting the majority of the traffic or the same load as the other queues? I would be interesting to see how your interrupt were laid out (cat /proc/interrupt | grep <ethX>)
As far as why MPC incrementing this can happen due to two reasons. First we ran out of space in descriptor ring for the packet, this doesn't seem to be your case as you checked to make sure the ring was empty. Next we could be running on of bandwidth on the PCIe bus. This might be possible if all your traffic was going to just one queue, but it would be worthwhile to see what the connectivity looks like anyway (lspci -vvv) It is also worth noting that your driver is quite old 2.0.38.3. I would suggest at least trying the latest driver from source forge (currently 3.15.1). https://sourceforge.net/projects/e1000/files/ixgbe%20stable/ I would also like to see if anything is being logged while this occurs (dmesg). Thanks, -Don Skidmore <[email protected]> > -----Original Message----- > From: Kaushal Bhandankar [mailto:[email protected]] > Sent: Sunday, June 16, 2013 12:32 PM > To: [email protected] > Subject: Re: [E1000-devel] Intel 82599 Driver dropping packets Queue > Receive Descriptor Head > > -bash-3.2$ /home/service/ethtool_x64 -i po0_0 > driver: tun > version: 1.6 > firmware-version: N/A > bus-info: tap > supports-statistics: no > supports-test: no > supports-eeprom-access: no > supports-register-dump: no > > > -bash-3.2# modinfo ixgbe > filename: /lib/modules/2.6.29.1/kernel/drivers/net/ixgbe/ixgbe.ko > version: 2.0.38.3-NAPI > license: GPL > description: Intel(R) 10 Gigabit PCI Express Network Driver > author: Intel Corporation, <[email protected]> > srcversion: CF158600678C3ADC41F341A > alias: pci:v00008086d000010FBsv*sd*bc*sc*i* > alias: pci:v00008086d000010FCsv*sd*bc*sc*i* > alias: pci:v00008086d000010F7sv*sd*bc*sc*i* > alias: pci:v00008086d000010DBsv*sd*bc*sc*i* > alias: pci:v00008086d000010F4sv*sd*bc*sc*i* > alias: pci:v00008086d000010E1sv*sd*bc*sc*i* > alias: pci:v00008086d000010F1sv*sd*bc*sc*i* > alias: pci:v00008086d000010ECsv*sd*bc*sc*i* > alias: pci:v00008086d000010DDsv*sd*bc*sc*i* > alias: pci:v00008086d0000150Bsv*sd*bc*sc*i* > alias: pci:v00008086d000010C8sv*sd*bc*sc*i* > alias: pci:v00008086d000010C7sv*sd*bc*sc*i* > alias: pci:v00008086d000010C6sv*sd*bc*sc*i* > alias: pci:v00008086d00001508sv*sd*bc*sc*i* > alias: pci:v00008086d000010B6sv*sd*bc*sc*i* > depends: > vermagic: 2.6.29.1 SMP mod_unload > parm: InterruptType:Change Interrupt Mode (0=Legacy, 1=MSI, > 2=MSI-X), default 2 (array of int) > parm: MQ:Disable or enable Multiple Queues, default 1 (array of > int) > parm: DCA:Disable or enable Direct Cache Access, 0=disabled, > 1=descriptor only, 2=descriptor and data (array of int) > parm: RSS:Number of Receive-Side Scaling Descriptor Queues, > default 1=number of cpus (array of int) > parm: InterruptThrottleRate:Maximum interrupts per second, per > vector, (956-488281), default 8000 (array of int) > parm: LLIPort:Low Latency Interrupt TCP Port (0-65535) (array of > int) > parm: LLIPush:Low Latency Interrupt on TCP Push flag (0,1) (array > of int) > parm: LLISize:Low Latency Interrupt on Packet Size (0-1500) > (array of int) > parm: LLIEType:Low Latency Interrupt Ethernet Protocol Type > (array of int) > parm: LLIVLANP:Low Latency Interrupt on VLAN priority threshold > (array of int) > parm: RxBufferMode:0=1 descriptor per packet, > 1=use packet split, multiple descriptors per jumbo > frame > 2 (default)=use 1buf mode for 1500 mtu, packet split > for jumbo > (array of int) > parm: FdirMode:Flow Director filtering modes: > 0 = Filtering off > 1 = Signature Hashing filters (SW ATR) > 2 = Perfect Filters (array of int) > parm: FdirPballoc:Flow Director packet buffer allocation level: > 0 = 8k hash filters or 2k perfect filters > 1 = 16k hash filters or 4k perfect filters > 2 = 32k hash filters or 8k perfect filters (array of > int) > parm: AtrSampleRate:Software ATR Tx packet sample rate (array of > int) > parm: DoubleVlan:Disable or enable double Vlan support, default 0 > (array of int) > parm: InnerVlanMode:Disable or enable Inner Vlan stripping, > default 0 (array of int) > > > > -bash-3.2# modprobe -l ixgbe > /lib/modules/2.6.29.1/kernel/drivers/net/ixgbe/ixgbe.ko > > -bash-3.2# lsmod > Module Size Used by > cidmodcap 4208 16 > cpp_base 845040 8 > tipc 118392 2 > rebootkom 2468 0 > nf_conntrack_ipv4 13376 1 > nf_defrag_ipv4 1976 1 nf_conntrack_ipv4 > xt_state 2232 1 > nf_conntrack 60000 2 nf_conntrack_ipv4,xt_state > iptable_filter 2888 1 > ip_tables 15848 1 iptable_filter > x_tables 18208 2 xt_state,ip_tables > igb 76876 0 > e1000e 113256 0 > ixgbe 168116 0 > ihm 6252 2 > cids_shared 579704 0 > linux_user_bde 15624 0 > linux_kernel_bde 28816 1 linux_user_bde > -bash-3.2# > > > cat /proc/cpuinfo = 16 cores > > > > > On Sun, Jun 16, 2013 at 10:04 AM, Kaushal Bhandankar > <[email protected]>wrote: > > > Hi, > > I am using Intel 82599 driver in my product. I see that out of 16 Rx > > Queues, only *One Queue* is dropping all the packets. When I did a > > test *to en-queue packets only to the problematic Queue*, I found that > > the > > > > -> Rx Missed Packet Count ( mpc ) is 0 Good Packet Received Count ( > > -> gprc ) in increased for received packets - > > L2 filtering is pass. > > > > What may be the reason for this behavior ? > > > > For debug, I am also printing the "Queue Receive Descriptor Head" and > > "Queue Receive Descriptor Tail" to get information about how many > > descriptors are in-use. My question is, in further tests, if I find > > that No Descriptor is in-use, however still packets are getting > > dropped -- Can I infer Hardware Failure from it ? > > > > Regards, > > Kaushal > > ------------------------------------------------------------------------------ This SF.net email is sponsored by Windows: Build for Windows Store. http://p.sf.net/sfu/windows-dev2dev _______________________________________________ E1000-devel mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/e1000-devel To learn more about Intel® Ethernet, visit http://communities.intel.com/community/wired
