> -----Original Message----- > From: Wiles, Keith > Sent: Tuesday, January 24, 2017 2:49 PM > To: Ananyev, Konstantin <konstantin.anan...@intel.com> > Cc: Stephen Hemminger <step...@networkplumber.org>; Hu, Jiayu > <jiayu...@intel.com>; dev@dpdk.org; Kinsella, Ray > <ray.kinse...@intel.com>; Gilmore, Walter E <walter.e.gilm...@intel.com>; > Venkatesan, Venky <venky.venkate...@intel.com>; > yuanhan....@linux.intel.com > Subject: Re: [dpdk-dev] [RFC] Add GRO support in DPDK > > > > On Jan 24, 2017, at 3:33 AM, Ananyev, Konstantin > > <konstantin.anan...@intel.com> wrote: > > > > > > > >> -----Original Message----- > >> From: Wiles, Keith > >> Sent: Tuesday, January 24, 2017 5:26 AM > >> To: Ananyev, Konstantin <konstantin.anan...@intel.com> > >> Cc: Stephen Hemminger <step...@networkplumber.org>; Hu, Jiayu > >> <jiayu...@intel.com>; dev@dpdk.org; Kinsella, Ray > >> <ray.kinse...@intel.com>; Gilmore, Walter E <walter.e.gilm...@intel.com>; > >> Venkatesan, Venky <venky.venkate...@intel.com>; > >> yuanhan....@linux.intel.com > >> Subject: Re: [dpdk-dev] [RFC] Add GRO support in DPDK > >> > >> > >>> On Jan 23, 2017, at 6:43 PM, Ananyev, Konstantin > >>> <konstantin.anan...@intel.com> wrote: > >>> > >>> > >>> > >>>> -----Original Message----- > >>>> From: Wiles, Keith > >>>> Sent: Monday, January 23, 2017 9:53 PM > >>>> To: Stephen Hemminger <step...@networkplumber.org> > >>>> Cc: Hu, Jiayu <jiayu...@intel.com>; dev@dpdk.org; Kinsella, Ray > >>>> <ray.kinse...@intel.com>; Ananyev, Konstantin > >>>> <konstantin.anan...@intel.com>; Gilmore, Walter E > >>>> <walter.e.gilm...@intel.com>; Venkatesan, Venky > >> <venky.venkate...@intel.com>; > >>>> yuanhan....@linux.intel.com > >>>> Subject: Re: [dpdk-dev] [RFC] Add GRO support in DPDK > >>>> > >>>> > >>>>> On Jan 23, 2017, at 10:15 AM, Stephen Hemminger > >>>>> <step...@networkplumber.org> wrote: > >>>>> > >>>>> On Mon, 23 Jan 2017 21:03:12 +0800 > >>>>> Jiayu Hu <jiayu...@intel.com> wrote: > >>>>> > >>>>>> With the support of hardware segmentation techniques in DPDK, the > >>>>>> networking stack overheads of send-side of applications, which directly > >>>>>> leverage DPDK, have been greatly reduced. But for receive-side, > >>>>>> numbers of > >>>>>> segmented packets seriously burden the networking stack of > >>>>>> applications. > >>>>>> Generic Receive Offload (GRO) is a widely used method to solve the > >>>>>> receive-side issue, which gains performance by reducing the amount of > >>>>>> packets processed by the networking stack. But currently, DPDK doesn't > >>>>>> support GRO. Therefore, we propose to add GRO support in DPDK, and this > >>>>>> RFC is used to explain the basic DPDK GRO design. > >>>>>> > >>>>>> DPDK GRO is a SW-based packets assembly library, which provides GRO > >>>>>> abilities for numbers of protocols. In DPDK GRO, packets are merged > >>>>>> before returning to applications and after receiving from drivers. > >>>>>> > >>>>>> In DPDK, GRO is a capability of NIC drivers. That support GRO or not > >>>>>> and > >>>>>> what GRO types are supported are up to NIC drivers. Different drivers > >>>>>> may > >>>>>> support different GRO types. By default, drivers enable all supported > >>>>>> GRO > >>>>>> types. For applications, they can inquire the supported GRO types by > >>>>>> each driver, and can control what GRO types are applied. For example, > >>>>>> ixgbe supports TCP and UDP GRO, but the application just needs TCP GRO. > >>>>>> The application can disable ixgbe UDP GRO. > >>>>>> > >>>>>> To support GRO, a driver should provide a way to tell applications what > >>>>>> GRO types are supported, and provides a GRO function, which is in > >>>>>> charge > >>>>>> of assembling packets. Since different drivers may support different > >>>>>> GRO > >>>>>> types, their GRO functions may be different. For applications, they > >>>>>> don't > >>>>>> need extra operations to enable GRO. But if there are some GRO types > >>>>>> that > >>>>>> are not needed, applications can use an API, like > >>>>>> rte_eth_gro_disable_protocols, to disable them. Besides, they can > >>>>>> re-enable the disabled ones. > >>>>>> > >>>>>> The GRO function processes numbers of packets at a time. In each > >>>>>> invocation, what GRO types are applied depends on applications, and the > >>>>>> amount of packets to merge depends on the networking status and > >>>>>> applications. Specifically, applications determine the maximum number > >>>>>> of > >>>>>> packets to be processed by the GRO function, but how many packets are > >>>>>> actually processed depends on if there are available packets to > >>>>>> receive. > >>>>>> For example, the receive-side application asks the GRO function to > >>>>>> process 64 packets, but the sender only sends 40 packets. At this time, > >>>>>> the GRO function returns after processing 40 packets. To reassemble the > >>>>>> given packets, the GRO function performs an "assembly procedure" on > >>>>>> each > >>>>>> packet. We use an example to demonstrate this procedure. Supposing the > >>>>>> GRO function is going to process packetX, it will do the following two > >>>>>> things: > >>>>>> a. Find a L4 assembly function according to the packet type of > >>>>>> packetX. A L4 assembly function is in charge of merging packets > >>>>>> of a > >>>>>> specific type. For example, TCPv4 assembly function merges > >>>>>> packets > >>>>>> whose L3 IPv4 and L4 is TCP. Each L4 assembly function has a > >>>>>> packet > >>>>>> array, which keeps the packets that are unable to assemble. > >>>>>> Initially, the packet array is empty; > >>>>>> b. The L4 assembly function traverses own packet array to find a > >>>>>> mergeable packet (comparing Ethernet, IP and L4 header fields). > >>>>>> If > >>>>>> finds, merges it and packetX via chaining them together; if > >>>>>> doesn't, > >>>>>> allocates a new array element to store packetX and updates > >>>>>> element > >>>>>> number of the array. > >>>>>> After performing the assembly procedure to all packets, the GRO > >>>>>> function > >>>>>> combines the results of all packet arrays, and returns these packets to > >>>>>> applications. > >>>>>> > >>>>>> There are lots of ways to implement the above design in DPDK. One of > >>>>>> the > >>>>>> ways is: > >>>>>> a. Drivers tell applications what GRO types are supported via > >>>>>> dev->dev_ops->dev_infos_get; > >>>>>> b. When initialize, drivers register own GRO function as a RX > >>>>>> callback, which is invoked inside rte_eth_rx_burst. The name of > >>>>>> the > >>>>>> GRO function should be like xxx_gro_receive (e.g. > >>>>>> ixgbe_gro_receive). > >>>>>> Currently, the RX callback can only process the packets > >>>>>> returned by > >>>>>> dev->rx_pkt_burst each time, and the maximum packet number > >>>>>> dev->rx_pkt_burst returns is determined by each driver, which > >>>>>> can't > >>>>>> be interfered by applications. Therefore, to implement the > >>>>>> above GRO > >>>>>> design, we have to modify current RX implementation to make > >>>>>> driver > >>>>>> return packets as many as possible until the packet number > >>>>>> meets the > >>>>>> demand of applications or there are not available packets to > >>>>>> receive. > >>>>>> This modification is also proposed in patch: > >>>>>> http://dpdk.org/ml/archives/dev/2017-January/055887.html; > >>>>>> c. The GRO types to apply and the maximum number of packets to > >>>>>> merge > >>>>>> are passed by resetting RX callback parameters. It can be > >>>>>> achieved by > >>>>>> invoking rte_eth_rx_callback; > >>>>>> d. Simply, we can just store packet addresses into the packet > >>>>>> array. > >>>>>> To check one element, we need to fetch the packet via its > >>>>>> address. > >>>>>> However, this simple design is not efficient enough. Since > >>>>>> whenever > >>>>>> checking one packet, one pointer dereference is generated. And a > >>>>>> pointer dereference always causes a cache line miss. A better > >>>>>> way is > >>>>>> to store some rules in each array element. The rules must be the > >>>>>> prerequisites of merging two packets, like the sequence number > >>>>>> of TCP > >>>>>> packets. We first compare the rules, then retrieve the packet > >>>>>> if the > >>>>>> rules match. If storing the rules causes the packet array > >>>>>> structure > >>>>>> is cache-unfriendly, we can store a fixed-length signature of > >>>>>> the > >>>>>> rules instead. For example, the signature can be calculated by > >>>>>> performing XOR operation on IP addresses. Both design can avoid > >>>>>> unnecessary pointer dereferences. > >>>>> > >>>>> > >>>>> Since DPDK does burst mode already, GRO is a lot less relevant. > >>>>> GRO in Linux was invented because there is no burst mode in the receive > >>>>> API. > >>>>> > >>>>> If you look at VPP in FD.io you will see they already do aggregration > >>>>> and > >>>>> steering at the higher level in the stack. > >>>>> > >>>>> The point of GRO is that it is generic, no driver changes are necessary. > >>>>> Your proposal would add a lot of overhead, and cause drivers to have to > >>>>> be aware of higher level flows. > >>>> > >>>> NACK > >>>> > >>>> The design is not super clear to me here and we need to understand the > >>>> impact to DPDK, performance and the application. I would > like > >> to > >>>> have a clean transparent design to the application and as little impact > >>>> on performance as possible. > >>>> > >>>> Let discuss this as I am not sure my previous concerns were addressed in > >>>> this RFC. > >>>> > >>> > >>> I would agree that design looks overcomplicated and strange: > >>> If GRO can (and supposed to be) done fully in SW, why do we need to > >>> modify PMDs at all, > >>> why it can't be just a standalone DPDK library that user can use on > >>> his/her convenience? > >>> I'd suggest to start with some simple and most widespread case (TCP?) and > >>> try to implement > >>> a library for it first: something similar to what we have for ip > >>> reassembly. > >> > >> The reason this should not be a library the application calls is to allow > >> for a transparent design for HW and SW support of this feature. > Using > >> the SW version the application should not need to understand (other then > >> performance) that GRO is being done for this port. > >> > > > > Why is that? > > Let say we have ip reassembly library that is called explicitly by the > > application. > > I think for L4 grouping we can do the same. > > After all it is a pure SW feature, so to me it makes sense to allow > > application to decide > > when/where to call it. > > Again it would allow people to develop/use it without any modifications in > > current PMDs. > > I guess I did not make it clear, we need to support HW and this SW version > transparently just as we handle other features in HW/SW under a > generic API for DPDK.
Ok, I probably wasn't very clear too. What I meant: Let's try to implement GRO (in SW) as a standalone DPDK library, with clean & simple interface and see how fast and useful it would be. We can refer to it as step 1. When (if) we'll have step 1 in place, then we can start thinking about adding combined HW/SW solution for it (step 2). I think at that stage it would be much clearer: is there any point in it at all, and if yes, how it should be done: -changes at rte_ethedev or on PMD layers or both - would changes at rte_ethdev API be needed and if yes what particular, etc. >From my perspective, without step 1 in place, there is no much point in >approaching step 2. BTW, any particular HW you have in mind? Currently, as I can see LRO (HW) is supported only by ixgbe and probably by viritual PMDs (virtio/vmxent3). Though even for ixgbe there are plenty of limitations: SRIOV mode should be off, HW CRC stropping should be off, etc. So my guess, right now step 1 is much more useful and feasible. > > > > >> As I was told the Linux kernel hides this features and make it transparent. > > > > Yes, but DPDK does a lot things in a different way. > > So it doesn't look like a compelling reason for me :) > > Just looking at different options here and it is a compelling reason to me as > it enforces the design can be transparent to the application. > Having the application in a NFV deciding on hw or sw or both is not a good > place to put that logic IMO. Actually could you provide an example of linux NIC driver, that uses HW offloads (and which) to implement GRO? I presume some might use HW generated hashes, but apart from that, when HW performs actual packet grouping? >From what I've seen Intel ones rely SW implementation for that. But I am not a linux/GRO expert, so feel free to correct me here. Konstantin