Using rar to compress and chunk up the file at the same time could be a good option, especially if this can be done on the client (for example using a java-applet as a file download/upload manager). Chunking on the client would be the only way to go over the 10MB limit as well.
GAE app development seems to be an exercise in being as efficient as you can, steering clear of restrictions, limits and time-outs. I like the idea of running rar server-side but I'd worry about hitting limits on individual page requests. For my needs I'm not so much bothered about the compression part - it's the chunking that would be great to have. I'm hopeful since this has been done for Java on GAE: http://code.google.com/p/google-file-service/wiki/GoogleFile Thanks, - Alex On Feb 22, 3:36 pm, Thadeus Burgess <thade...@thadeusb.com> wrote: > Well, with RAR files you can split them up into 1MB files. So you get > the uploaded file, rar it in .9MB splits, upload to a table that > stores "chunks". > > Now when you need the file, pull the "chunks" out of your table, and > unrar them. I know that this works on the unix command line, not sure > if the unrar lib in python supports multiple > > files in memory. > > However this will be very expensive to do, if you have a 10MB file it > will most likely take at "least" double the ram to perform this > operation. > > -Thadeus > > > > On Mon, Feb 22, 2010 at 9:29 AM, what_ho <a...@viovi.com> wrote: > > > Has anyone tried splitting and joining one file across several blobs > > in GAE to allow uploads up to the 10M -- You received this message because you are subscribed to the Google Groups "web2py-users" group. To post to this group, send email to web...@googlegroups.com. To unsubscribe from this group, send email to web2py+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/web2py?hl=en.