Well, mine isn't completely done yet (the tar.gz is not correct and some other things) but the myfaces-trinidad-site job has a working principal for downloading as a cron job. It pulls the sites source into a staging directory and packages it up during the build. It then stores the tar.gz as an artifact so it's always available and only updated on a successful build.
I hope to have it done and generating a valid archive soon, but feel free to take a look at it if you like. I figured a gzipped archive was convenient enough to wget. Scott On Mar 3, 2011, at 6:34 AM, Jukka Zitting <jukka.zitt...@gmail.com> wrote: > Hi, > > On Thu, Mar 3, 2011 at 2:30 PM, Ulrich Stärk <u...@apache.org> wrote: >> The only thing that seems reasonable apart from having an rsync daemon is >> doing a recursive wget on >> the generated site every hour or so. This will however fetch all 4000 files >> and I don't know how >> much load it puts on Jenkins. Is this OK? > > Doesn't sound like a very good approach. > > The Apache CMS setup running on Buildbot simply commits the generated > site to svn from where it's picked up by the public web servers. I > guess a similar setup could be done also with Maven site builds > running in Hudson. > > BR, > > Jukka Zitting