I myself don't use puppet to pull big files.
Maybe you use puppet with the default Webrick HTTP frontend. You may
test another frontend like mongrel or passenger :
http://reductivelabs.com/trac/puppet/wiki/UsingMongrel
http://reductivelabs.com/trac/puppet/wiki/UsingPassenger

For my use, I use an HTTP server and a custom curl definition. But for
slow connections, it didn't resolve the timeout problem.

define common::archive::tar-gz($source, $target) {
  exec {"$name unpack":
    command => "curl ${source} | tar -xzf - -C ${target} && touch ${name}",
    creates => $name
  }
}

But the more elegant solution would be to package hadoop.

2009/7/18 Fernando Padilla <f...@alum.mit.edu>
>
> Hi.  I'm a beginner, but I have a basic puppet setup working.  I am
> doing a manual tarball installation and it seems to be hanging then
> eventually timing out on just downloading the file:
>
>     file { "/opt/hadoop-0.20.0.tar.gz":
>        source => "puppet:///hadoop020/hadoop-0.20.0.tar.gz"
>     }
>
> I have another module that does the same things and works, my only guess
> is the size of the tarball:
>
> modules/hadoop020/files/hadoop-0.20.0.tar.gz - 41M
> modules/zookeeper320/files/zookeeper-3.2.0.tar.gz - 12M
>
> Any ideas or suggestions to speed up file transfers??
>
> If I manually scp the file, it takes only 30seconds (between office and
> ec2), why would it take so long and eventually timeout inside the colo (
> ec2)?
>
> >

--~--~---------~--~----~------------~-------~--~----~
You received this message because you are subscribed to the Google Groups 
"Puppet Users" group.
To post to this group, send email to puppet-users@googlegroups.com
To unsubscribe from this group, send email to 
puppet-users+unsubscr...@googlegroups.com
For more options, visit this group at 
http://groups.google.com/group/puppet-users?hl=en
-~----------~----~----~----~------~----~------~--~---

Reply via email to