On 2/28/2012 5:57 PM, Nicholas Weaver wrote:
> But since if you could coalesce you could do parallel, this savings doesn't 
> help latency much at all: just the transit time for 50B out and 50B back.  If 
> anything, parallel will be BETTER on the latency since batching would 
> probably require a coalesced response, while parallel is just that, parallel, 
> so if the first record is more useful, it can be acted on immediately.

parallel (what i called "blast") is great solution for things that don't
run at scale. but the lock-step serialization that we get from the MX
approach (where the A/AAAA RRset may be in the additional data section
and so we have to wait for the first answer before we ask the second or
third question) has a feedback loop effect on rate limiting. it's not as
good as tcp windowing but it does tend to avoid WRED in core and edge
router egress buffers.

all i'm saying is, we have to be careful about too much parallelism
since UDP unlike TCP has no windowing of its own.

_______________________________________________
DNSOP mailing list
DNSOP@ietf.org
https://www.ietf.org/mailman/listinfo/dnsop

Reply via email to