On Thu, Feb 2, 2017 at 4:47 PM, Daniel Jurgens <dani...@mellanox.com> wrote: > On 2/1/2017 5:12 AM, David Laight wrote: >> From: Saeed Mahameed >>> Sent: 31 January 2017 20:59 >>> From: Daniel Jurgens <dani...@mellanox.com> >>> >>> There is a hardware feature that will pad the start or end of a DMA to >>> be cache line aligned to avoid RMWs on the last cache line. The default >>> cache line size setting for this feature is 64B. This change configures >>> the hardware to use 128B alignment on systems with 128B cache lines. >> What guarantees that the extra bytes are actually inside the receive skb's >> head and tail room? >> >> David >> >> > The hardware won't over write the length of the posted buffer. This feature > is already enabled and defaults to 64B stride, this patch just configures it > properly for 128B cache line sizes. > Right, and next patch will make sure RX stride is aligned to 128B in case 128B cacheline size configured into the HW.
> Thanks for reviewing it. > > Dan >