On Thu, Sep 26, 2024, 1:22 PM Brian E Carpenter <brian.e.carpen...@gmail.com>
wrote:

> On 27-Sep-24 05:56, Templin (US), Fred L wrote:
> > Hi Tom,
> >
> > I would like to gently suggest a new terminology. Rather than calling
> them "the multi-segment buffers managed by GSO and GRO", can we begin
> calling them "parcel buffers" or simply "parcels"? Not suggesting this in a
> self-serving manner - I just think it is a more concise yet more
> descriptive terminology.
>
> But that isn't the same thing. RFC2675 jumbograms are single datagrams.
> They were originally intended for use over HIPPI, i.e. internally to data
> centres as they existed 25 years ago, so the usage that Tom reported seems
> close to what they were designed for.
>
> Tom, is there a full description of this usage?
>

Hi Brain,

Here's Eric Dumazet's preso from Netdev 0x15. There's also work by Redhat
and Cilium on the web. I believe this was integrated in Linux 6.3.

https://netdevconf.info/0x15/session.html?BIG-TCP

Tom


> Regards
>     Brian
>
> >
> > Thank you - Fred
> >
> >> -----Original Message-----
> >> From: Tom Herbert <tom=40herbertland....@dmarc.ietf.org>
> >> Sent: Thursday, September 26, 2024 10:15 AM
> >> To: Tim Chown <tim.ch...@jisc.ac.uk>
> >> Cc: Paul Vixie <p...@redbarn.org>; Templin (US), Fred L <
> fred.l.temp...@boeing.com>; Internet Area <Int-area@ietf.org>; IPv6 List
> >> <i...@ietf.org>
> >> Subject: Re: [Int-area] Re: IP Parcels and Advanced Jumbos (AJs)
> >>
> >> On Thu, Sep 26, 2024 at 9:03 AM Tim Chown
> >> <Tim.Chown=40jisc.ac...@dmarc.ietf.org> wrote:
> >>>
> >>> Hi,
> >>>
> >>>
> >>>
> >>> From: Paul Vixie <paul=40redbarn....@dmarc.ietf.org>
> >>> Date: Tuesday, 24 September 2024 at 20:59
> >>> To: Templin (US), Fred L <Fred.L.Templin=40boeing....@dmarc.ietf.org>,
> Internet Area <Int-area@ietf.org>, IPv6 List <i...@ietf.org>
> >>> Subject: [Int-area] Re: IP Parcels and Advanced Jumbos (AJs)
> >>>
> >>> Something like this is long needed and will become badly needed. Every
> 10X of speed increase since 10mbit/sec has gone straight to PPS,
> >> whereas the speed increase from 3mbit/sec to 10mbit/sec was shared
> between PPS and MTU.
> >>>
> >>>
> >>>
> >>> If every 10X has been shared between PPS and MTU, say sqrt(10) for
> each, our MTU would be well over 64K by now and our PPS wouldn't
> >> require dedicated NPU hardware to source, sink, and ferry those packets
> at link saturation levels.
> >>>
> >>>
> >>>
> >>> Every attempt at PMTUD so far has failed but that's not an excuse to
> stop trying.
> >>>
> >>>
> >>>
> >>> I think that depends on the deployment scenario and environment.  In
> R&E networking the adoption of 9000 MTU for large scale wide
> >> area data transfers has grown, in particular by dozens of sites
> worldwide that take part in the CERN experiments. CERN did a site survey
> >> recently, for which I could dig out the results.
> >>>
> >>>
> >>>
> >>> The sites running 9000 MTU are interoperating with the sites still at
> 1500, which is an indication that PMTUD is working well enough. The
> >> large majority of CERN traffic is IPv6, so for that there’s no
> fragmentation on path.
> >>
> >> Tim,
> >>
> >> That's also happening in some datacenters. I believe Google is using a
> >> 9K MTU internally as it makes zero copy on hosts feasible (two 4K
> >> pages per packet). Interestingly, there's also increasing use of
> >> RFC2675 jumbograms, they're not sent on the wire but used internally
> >> for GSO and GRO for greater than 64K packets.
> >>
> >> Tom
> >>>
> >>>
> >>>
> >>> The use case is somewhat constrained in that it’s only the parts of
> the campus with the storage, the campus paths to the edge, and the
> >> intervening R&E backbones that need to be configured. But with correct
> ICMPv6 filtering, it seems robust.
> >>>
> >>>
> >>>
> >>> Best wishes,
> >>>
> >>> Tim
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> Thanks for driving this Fred.
> >>>
> >>>
> >>>
> >>> p vixie
> >>>
> >>>
> >>>
> >>> On Sep 24, 2024 14:39, "Templin (US), Fred L" <Fred.L.Templin=
> 40boeing....@dmarc.ietf.org> wrote:
> >>>
> >>> It has been a while since I have posted about this, and there are some
> updates to highlight.
> >>>
> >>> See below for the IPv6 and IPv4 versions of “IP Parcels and Advanced
> Jumbos (AJs)”:
> >>>
> >>>
> >>>
> >>> https://datatracker.ietf.org/doc/draft-templin-6man-parcels2/
> >>>
> >>> https://datatracker.ietf.org/doc/draft-templin-intarea-parcels2/
> >>>
> >>>
> >>>
> >>> The documents acknowledge that parcels are analogous to Generic
> Segment/Receive Offload
> >>>
> >>> (GSO/GRO) but taken to the ultimate aspiration of encapsulating
> multi-segment buffers in
> >>>
> >>> {TCP/UDP}/IP headers for transmission over parcel-capable network
> paths. They further give
> >>>
> >>> a name to the multi-segment buffers used by GSO/GRO, suggesting that
> they be called
> >>>
> >>> “parcel buffers” or simply “parcels”.
> >>>
> >>>
> >>>
> >>> AJs are simply single-segment parcels that can range in size from very
> small to very large.
> >>>
> >>> They differ from ordinary jumbograms in several important ways, most
> notably in terms
> >>>
> >>> of integrity verification and error correction. They also suggest a
> new link service model
> >>>
> >>> that defers integrity checks to the end systems where bad data can be
> discarded while
> >>>
> >>> good data can be accepted as an end-to-end function, reducing
> retransmissions.
> >>>
> >>>
> >>>
> >>> Together, these documents cover all possible packet sizes and
> configurations that may
> >>>
> >>> be necessary both in the near term and for the foreseeable future for
> Internetworking
> >>>
> >>> performance maximization . Comments on the list(s) are welcome.
> >>>
> >>>
> >>>
> >>> Fred Templin
> >>>
> >>> _______________________________________________
> >>> Int-area mailing list -- int-area@ietf.org
> >>> To unsubscribe send an email to int-area-le...@ietf.org
> >
> > _______________________________________________
> > Int-area mailing list -- int-area@ietf.org
> > To unsubscribe send an email to int-area-le...@ietf.org
>
_______________________________________________
Int-area mailing list -- int-area@ietf.org
To unsubscribe send an email to int-area-le...@ietf.org

Reply via email to