On Fri, Dec 13, 2024 at 7:32 PM Dave Taht <dave.t...@gmail.com> wrote: > A couple cake notes below...
Hey Dave, thanks for replying and for all your hard work on AQM and latency. > On Fri, Dec 13, 2024 at 3:50 PM <sro...@ronan-online.com> wrote: > > Do you have an explanation for the question he asked? I am sure it would be > > of interest to many here. Sean noted live video uses a custom protocol called Sye that estimates available throughput. My understanding is this estimation can be too low as compared to TCP when the network is slow or congested. That is, in cases where TCP would buffer and then be able to move a large amount of data at the expense of latency, Sye may instead push much less data, sacrificing quality for latency. IMO amazon streaming should fall back to TCP in these conditions, perhaps noting to the customer their stream is no longer "live" but is slightly delayed to improve video quality. Ideally a network that could push on average 10mbit/s consistently over TCP could also push 10mbit/s consistently over a custom UDP protocol, but when this is not the case (due to any number of bizarre real-world conditions), the system should detect this and reset. Giving the customer a delayed but high-quality nearly-live stream would (again, IMO) be a better experience than a live but poor-quality video. Of course I could probably have achieved myself this by simply rewinding the live stream, but I was blinded to this option at the time by my surprise and amazement at how poor the video quality was. > 2) Applying a 200Mbit inbound limit to a gig fiber connection is > overkill. Worst case it should be at 50%, and we generally recommend > 85%. The reasoning for 200mbit is it's about 50% of best-case real-world 802.11 performance across a house. The goal is to keep buffers in the APs as empty as possible. I'd rather enforce this on the APs, but lacking that ability I do it for all traffic from the router to the rest of the network. > If you have demand for less bandwidth that your fair share, you > experience near zero latency and no loss for your flow. At 200Mbit, > assuming nat mode was on, your amazon flow (probably running below > 20mbit) should have seen no congestion or loss at all while another > machine merrily downloaded at 180, dropping the occasional packet to > keep it under control. You're absolutely right. It's very possible the issue I experienced was due to slowness at the wireless network level, and not Linux traffic shaping. > > >>> The live stream appears to use UDP on a non-standard port (not 443). > > >>> Does anyone know what amazon has done to cause their congestion > > >>> control algorithms to yield so much bandwidth and not fight for their > > >>> fair share? > > Anyway, this is a good question, if that was the observed behavior. A > packet capture showing it not recovering after a loss or delay would > be interesting. My guess is that since Sye prioritizes live data over throughput, it will essentially by design deliver poor quality in situations where bandwidth is limited and TCP streams are vying to use as much of it as they can. This unfortunately describes a lot of home networks using wifi in real-world conditions. -- Dan