On Sun, 4 Nov 2018 01:24:03 +0100 Paweł Staszewski <pstaszew...@itcare.pl> wrote:
> And today again after allpy patch for page allocator - reached again > 64/64 Gbit/s > > with only 50-60% cpu load Great. > today no slowpath hit for netwoking :) > > But again dropped pckt at 64GbitRX and 64TX .... > And as it should not be pcie express limit -i think something more is Well, this does sounds like a PCIe bandwidth limit to me. See the PCIe BW here: https://en.wikipedia.org/wiki/PCI_Express You likely have PCIe v3, where 1-lane have 984.6 MBytes/s or 7.87 Gbit/s Thus, x16-lanes have 15.75 GBytes or 126 Gbit/s. It does say "in each direction", but you are also forwarding this RX->TX on both (dual) ports NIC that is sharing the same PCIe slot. > going on there - and hard to catch - cause perf top doestn chenged > besides there is no queued slowpath hit now > > I ordered now also intel cards to compare - but 3 weeks eta > Faster - cause 3 days - i will have mellanox connectx 5 - so can > separate traffic to two different x16 pcie busses I do think you need to separate traffic to two different x16 PCIe slots. I have found that the ConnectX-5 is significantly faster packet-per-sec performance than ConnectX-4, but that is not your use-case (max BW). I've not tested these NICs for maximum _bidirectional_ bandwidth limits, I've only made sure I can do 100G unidirectional, which can hit some funny motherboard memory limits (remember to equip motherboard with 4 RAM blocks for full memory BW). -- Best regards, Jesper Dangaard Brouer MSc.CS, Principal Kernel Engineer at Red Hat LinkedIn: http://www.linkedin.com/in/brouer