I myself don't use puppet to pull big files. Maybe you use puppet with the default Webrick HTTP frontend. You may test another frontend like mongrel or passenger : http://reductivelabs.com/trac/puppet/wiki/UsingMongrel http://reductivelabs.com/trac/puppet/wiki/UsingPassenger
For my use, I use an HTTP server and a custom curl definition. But for slow connections, it didn't resolve the timeout problem. define common::archive::tar-gz($source, $target) { exec {"$name unpack": command => "curl ${source} | tar -xzf - -C ${target} && touch ${name}", creates => $name } } But the more elegant solution would be to package hadoop. 2009/7/18 Fernando Padilla <f...@alum.mit.edu> > > Hi. I'm a beginner, but I have a basic puppet setup working. I am > doing a manual tarball installation and it seems to be hanging then > eventually timing out on just downloading the file: > > file { "/opt/hadoop-0.20.0.tar.gz": > source => "puppet:///hadoop020/hadoop-0.20.0.tar.gz" > } > > I have another module that does the same things and works, my only guess > is the size of the tarball: > > modules/hadoop020/files/hadoop-0.20.0.tar.gz - 41M > modules/zookeeper320/files/zookeeper-3.2.0.tar.gz - 12M > > Any ideas or suggestions to speed up file transfers?? > > If I manually scp the file, it takes only 30seconds (between office and > ec2), why would it take so long and eventually timeout inside the colo ( > ec2)? > > > --~--~---------~--~----~------------~-------~--~----~ You received this message because you are subscribed to the Google Groups "Puppet Users" group. To post to this group, send email to puppet-users@googlegroups.com To unsubscribe from this group, send email to puppet-users+unsubscr...@googlegroups.com For more options, visit this group at http://groups.google.com/group/puppet-users?hl=en -~----------~----~----~----~------~----~------~--~---