On 4/4/2016 9:50 PM, Alexei Starovoitov wrote:
On Mon, Apr 04, 2016 at 08:22:03AM -0700, Eric Dumazet wrote:
A single flow is able to use 40Gbit on those 40Gbit NIC, so there is not
a single 10GB trunk used for a given flow.
This 14Mpps thing seems to be a queue limitation on mlx4.
yeah, could be queueing related. Multiple cpus can send ~30Mpps of the same 64
byte packet,
but mlx4 can only receive 14.5Mpps. Odd.
Or (and other mellanox guys), what is really going on inside 40G nic?
Hi Alexei,
Not that I know everything that goes inside there, and not that if I
knew it all I could have posted that here (I heard HWs sometimes have
IP)... but, anyway, as for your questions:
ConnectX3 40Gbs NIC can receive > 10Gbs packet-worthy (14.5M) in single
ring and Mellanox
100Gbs NICs can receive > 25Gbs packet-worthy (37.5M) in single ring,
people that use DPDK (...) even see this numbers and AFAIU we now
attempt to see that in the kernel with XDP :)
I realize that we might have some issues in the mlx4 driver reporting on
HW drops. Eran (cc-ed) and Co are looking on that.
In parallel to doing so, I would suggest you to do some experiments that
might shed some more light, if on the TX side you do
$ ./pktgen_sample03_burst_single_flow.sh -i $DEV -d $IP -m $MAC -t 4
On the RX side, skip RSS and force the packets that match that traffic
pattern to go to (say) ring (==action) 0
$ ethtool -U $DEV flow-type ip4 dst-mac $MAC dst-ip $IP action 0 loc 0
to go back to RSS remove the rule
$ ethtool -U $DEV delete action 0
FWIW (not that I see how it helps you now), you can do HW drop on the RX
side with ring -1
$ ethtool -U $DEV flow-type ip4 dst-mac $MAC dst-ip $IP action -1 loc 0
Or.