From: Saeed Mahameed > Sent: 05 February 2017 11:24 > On Thu, Feb 2, 2017 at 4:47 PM, Daniel Jurgens <dani...@mellanox.com> wrote: > > On 2/1/2017 5:12 AM, David Laight wrote: > >> From: Saeed Mahameed > >>> Sent: 31 January 2017 20:59 > >>> From: Daniel Jurgens <dani...@mellanox.com> > >>> > >>> There is a hardware feature that will pad the start or end of a DMA to > >>> be cache line aligned to avoid RMWs on the last cache line. The default > >>> cache line size setting for this feature is 64B. This change configures > >>> the hardware to use 128B alignment on systems with 128B cache lines. > >> What guarantees that the extra bytes are actually inside the receive skb's > >> head and tail room? > >> > >> David > >> > >> > > The hardware won't over write the length of the posted buffer. This > > feature is already enabled and > defaults to 64B stride, this patch just configures it properly for 128B cache > line sizes. > > > Right, and next patch will make sure RX stride is aligned to 128B in > case 128B cacheline size configured into the HW.
Doesn't that mean these patches are in the wrong order? > > Thanks for reviewing it. Don't assume I've done anything other than look for obvious fubars/ David