Thanks Michael,
Thanks for all of the information. Regarding your final paragraph, you
mentioned that the 64 Gb/s could be handled on one 100 Gb link.  However,
that seems at odds with the following statement in the UHD manual in the X410
section about FPGA types
<https://files.ettus.com/manual/page_usrp_x4xx.html#x4xx_updating_fpga_types>


   - CG_400: 400 MHz analog bandwidth streaming per channel between the
   X4x0 and an external host computer. The current implementation requires
   dual 100 GbE connections for 4 full-duplex channels or a single 100 GbE
   connection for 2 full-duplex channels.

Do you think that this statement in the UHD manual is a mistake?  This is
the statement that made me think that I needed two 100Gb links even though
the 4 channels at 500 MS/s is aggregate 64Gb/s.  If only one link is truly
needed, then I can feel more confident purchasing an E810.
Rob



On Wed, Mar 26, 2025 at 3:53 PM Michael Dickens <michael.dick...@ettus.com>
wrote:

> Hey Rob! Great questions. Here's way too much information taken from
> internal notes I have on the subject, to help you process all of this :)
> {{{
> E810 QCDA2 provides 100 Gb aggregate between both ports. Dual port to USRP
> is not recommended since UHD doesn't "know" this limitation.
>
> E810 2QCAD2 provides 2 bifurcated 100 Gb links, so can do 200 Gb
> aggregate. I -think- one has to tell BIOS / OS about this bifurcation to
> get the NIC fully working. I don't have one to test out.
>
> There are now newer Intel E82* NICs. I don't know their capabilities.
>
> Any of the Intel E8* NICs can be configured in various ways, the most
> relevant for USRPs are:
> * 2x1x100 : 2 ports, each hosting 1 virtual link at 100 Gb
> * 100 : 1 port with a single virtual link at 100 Gb
> * 8x10 (formerly 2x4x10 : 2 ports, each hosting 4 virtual link at 10 Gb
> each
> {{{
> $ sudo ./epct64e -get -nic 1
> Ethernet Port Configuration Tool
> EPCT version: v1.42.24.04
> Copyright 2019 - 2024 Intel Corporation.
>
> Available Port Options:
> ==========================================================================
>         Port                             Quad 0           Quad 1
> Option  Option (Gbps)                    L0  L1  L2  L3   L4  L5  L6  L7
> ======= =============================    ================ ================
>         2x1x100                       -> 100   -   -   -  100   -   -   -
>         2x50                          ->  50   -  50   -    -   -   -   -
>         4x25                          ->  25  25  25  25    -   -   -   -
>         2x2x25                        ->  25  25   -   -   25  25   -   -
> Active  8x10                          ->  10  10  10  10   10  10  10  10
>         100                           -> 100   -   -   -    -   -   -   -
> }}}
>
> FWIW: We're had a number of customers with E810 CQDA2 issues recently. My
> current belief is that the NIC (NVM) and OS drivers do not play nicely
> together & hence updating both to the latest is needed to get everything
> working properly.
>
> Intel E8* NICs used the ICE driver, which is in active development & works
> pretty well overall. ICE drivers -do not- work seamlessly with DPDK unlike
> Mellanox ones. It's easy to create a script to do the driver binding & link
> stuff both down and up, but this can be very confusing for people not used
> to taking down a link and rebinding the driver & then the reverse to get it
> back working in the system again.
>
> The Mellanox drivers & hardware use a little less CPU time than the Intel
> ones, so a little better single-core performance — which helps when using
> DPDK and doing max data throughput.
>
> Yes, 500 GS/s on 4 channels (2 GS/s aggregate) is 64 Gb/s and thus well
> within the capabilities of a single 100 Gb port on either NIC ... That's
> fine for an X410. For an X440 we double that to 4 GS/s aggregate, which
> clearly requires 2x 100 Gb links. For this use-case the Mellanox NICs are
> the way to go.
> }}}
>
> On Tue, Mar 25, 2025 at 3:53 PM Rob Kossler via USRP-users <
> usrp-users@lists.ettus.com> wrote:
>
>> Hi,
>> I am in the process of purchasing a 100Gb NIC for use with the X410 and
>> have seen documentation and previous posts indicating that the ConnectX
>> NICs are preferred. But I did note in the DPDK knowledge base article that
>> the Intel E810 could also work.  I prefer the E810 because it seems to be
>> less expensive and can be configured for 4x10Gb, but I don't want to create
>> a headache for myself.  Let me know if you have had success or issues with
>> the E810 using a 100Gb link (or two 100Gb links) to the X410.
>>
>> I am also confused about the E810 which comes in a couple of 100Gb
>> models: CQDA2 and 2CQDA2, where they both have two 100Gb QSFP28 ports, but
>> the former can only handle aggregate 100Gb whereas the latter can handle
>> aggregate 200Gb.  My confusion is "why does it matter for the X410?".  With
>> 4 channels at 500 MS/s, the aggregate bit rate is only 64Gb/s so why does
>> it matter if the E810 CQDA2 only supports aggregate 100Gb?  It seems to me
>> that either model supports the maximum rate of the X410.
>>
>> Thanks.
>> Rob
>> _______________________________________________
>> USRP-users mailing list -- usrp-users@lists.ettus.com
>> To unsubscribe send an email to usrp-users-le...@lists.ettus.com
>>
>
_______________________________________________
USRP-users mailing list -- usrp-users@lists.ettus.com
To unsubscribe send an email to usrp-users-le...@lists.ettus.com

Reply via email to