On Fri, Dec 13, 2024 at 7:32 PM Dave Taht <dave.t...@gmail.com> wrote:
> A couple cake notes below...

Hey Dave, thanks for replying and for all your hard work on AQM and latency.


> On Fri, Dec 13, 2024 at 3:50 PM <sro...@ronan-online.com> wrote:
> > Do you have an explanation for the question he asked? I am sure it would be 
> > of interest to many here.

Sean noted live video uses a custom protocol called Sye that estimates
available throughput. My understanding is this estimation can be too
low as compared to TCP when the network is slow or congested.

That is, in cases where TCP would buffer and then be able to move a
large amount of data at the expense of latency, Sye may instead push
much less data, sacrificing quality for latency.

IMO amazon streaming should fall back to TCP in these conditions,
perhaps noting to the customer their stream is no longer "live" but is
slightly delayed to improve video quality. Ideally a network that
could push on average 10mbit/s consistently over TCP could also push
10mbit/s consistently over a custom UDP protocol, but when this is not
the case (due to any number of bizarre real-world conditions), the
system should detect this and reset. Giving the customer a delayed but
high-quality nearly-live stream would (again, IMO) be a better
experience than a live but poor-quality video.

Of course I could probably have achieved myself this by simply
rewinding the live stream, but I was blinded to this option at the
time by my surprise and amazement at how poor the video quality was.


> 2) Applying a 200Mbit inbound limit to a gig fiber connection is
> overkill. Worst case it should be at 50%, and we generally recommend
> 85%.

The reasoning for 200mbit is it's about 50% of best-case real-world
802.11 performance across a house. The goal is to keep buffers in the
APs as empty as possible. I'd rather enforce this on the APs, but
lacking that ability I do it for all traffic from the router to the
rest of the network.


> If you have demand for less bandwidth that your fair share, you
> experience near zero latency and no loss for your flow. At 200Mbit,
> assuming nat mode was on, your amazon flow (probably running below
> 20mbit) should have seen no congestion or loss at all while another
> machine merrily downloaded at 180, dropping the occasional packet to
> keep it under control.

You're absolutely right. It's very possible the issue I experienced
was due to slowness at the wireless network level, and not Linux
traffic shaping.


> > >>> The live stream appears to use UDP on a non-standard port (not 443).
> > >>> Does anyone know what amazon has done to cause their congestion
> > >>> control algorithms to yield so much bandwidth and not fight for their
> > >>> fair share?
>
> Anyway, this is a good question, if that was the observed behavior. A
> packet capture showing it not recovering after a loss or delay would
> be interesting.

My guess is that since Sye prioritizes live data over throughput, it
will essentially by design deliver poor quality in situations where
bandwidth is limited and TCP streams are vying to use as much of it as
they can. This unfortunately describes a lot of home networks using
wifi in real-world conditions.

-- Dan

Reply via email to