On Fri, May 15, 2009 at 4:30 PM, AchipA <attila.cs...@gmail.com> wrote:
> I did a fair amount of testing on a couple of systems and the numbers
> say that in real life, it's quite a significant gain except for some
> very specific cases. I'd say the article you quoted concludes the
> same. Even if the latency does not improve for ONE client, the next
> one will benefit as the *server's* network throughput increases.

I agree. Even the size or time is the same in:
uncompressed data (server) ---> (cliente) read data
and
uncompressed data ---> compress (server) ---> (cliente) uncompress --> read data

For client it can change anything (in the worst case: if
compression/uncompression time is high so 2 times above are equal) but
it for server it can "save" bandwidth with "lost" of processing time.
As processors are cheaper than traffic, it is a good thing.

-- 
 Álvaro Justen
 Peta5 - Telecomunicações e Software Livre
 21 3021-6001 / 9898-0141
 http://www.peta5.com.br/

--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"web2py Web Framework" group.
To post to this group, send email to web2py@googlegroups.com
To unsubscribe from this group, send email to 
web2py+unsubscr...@googlegroups.com
For more options, visit this group at 
http://groups.google.com/group/web2py?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to