We are migrating our scripts to r3. Thr lineage is in spark-ec2 would be
happy to migrate those too.
Having trouble with ganglia setup currently :)
Regards
Mayur
On 31 May 2014 09:07, "Patrick Wendell" <pwend...@gmail.com> wrote:

> Hi Jeremy,
>
> That's interesting, I don't think anyone has ever reported an issue
> running these scripts due to Python incompatibility, but they may require
> Python 2.7+. I regularly run them from the AWS Ubuntu 12.04 AMI... that
> might be a good place to start. But if there is a straightforward way to
> make them compatible with 2.6 we should do that.
>
> For r3.large, we can add that to the script. It's a newer type. Any
> interest in contributing this?
>
> - Patrick
>
> On May 30, 2014 5:08 AM, "Jeremy Lee" <unorthodox.engine...@gmail.com>
> wrote:
>
>>
>> Hi there! I'm relatively new to the list, so sorry if this is a repeat:
>>
>> I just wanted to mention there are still problems with the EC2 scripts.
>> Basically, they don't work.
>>
>> First, if you run the scripts on Amazon's own suggested version of linux,
>> they break because amazon installs Python2.6.9, and the scripts use a
>> couple of Python2.7 commands. I have to "sudo yum install python27", and
>> then edit the spark-ec2 shell script to use that specific version.
>> Annoying, but minor.
>>
>> (the base "python" command isn't upgraded to 2.7 on many systems,
>> apparently because it would break yum)
>>
>> The second minor problem is that the script doesn't know about the
>> "r3.large" servers... also easily fixed by adding to the spark_ec2.py
>> script. Minor,
>>
>> The big problem is that after the EC2 cluster is provisioned, installed,
>> set up, and everything, it fails to start up the webserver on the master.
>> Here's the tail of the log:
>>
>> Starting GANGLIA gmond:                                    [  OK  ]
>> Shutting down GANGLIA gmond:                               [FAILED]
>> Starting GANGLIA gmond:                                    [  OK  ]
>> Connection to ec2-54-183-82-48.us-west-1.compute.amazonaws.com closed.
>> Shutting down GANGLIA gmond:                               [FAILED]
>> Starting GANGLIA gmond:                                    [  OK  ]
>> Connection to ec2-54-183-82-24.us-west-1.compute.amazonaws.com closed.
>> Shutting down GANGLIA gmetad:                              [FAILED]
>> Starting GANGLIA gmetad:                                   [  OK  ]
>> Stopping httpd:                                            [FAILED]
>> Starting httpd: httpd: Syntax error on line 153 of
>> /etc/httpd/conf/httpd.conf: Cannot load modules/mod_authn_alias.so into
>> server: /etc/httpd/modules/mod_authn_alias.so: cannot open shared object
>> file: No such file or directory
>>                                                            [FAILED]
>>
>> Basically, the AMI you have chosen does not seem to have a "full" install
>> of apache, and is missing several modules that are referred to in the
>> httpd.conf file that is installed. The full list of missing modules is:
>>
>> authn_alias_module modules/mod_authn_alias.so
>> authn_default_module modules/mod_authn_default.so
>> authz_default_module modules/mod_authz_default.so
>> ldap_module modules/mod_ldap.so
>> authnz_ldap_module modules/mod_authnz_ldap.so
>> disk_cache_module modules/mod_disk_cache.so
>>
>> Alas, even if these modules are commented out, the server still fails to
>> start.
>>
>> root@ip-172-31-11-193 ~]$ service httpd start
>> Starting httpd: AH00534: httpd: Configuration error: No MPM loaded.
>>
>> That means Spark 1.0.0 clusters on EC2 are Dead-On-Arrival when run
>> according to the instructions. Sorry.
>>
>> Any suggestions on how to proceed? I'll keep trying to fix the webserver,
>> but (a) changes to httpd.conf get blown away by "resume", and (b) anything
>> I do has to be redone every time I provision another cluster. Ugh.
>>
>> --
>> Jeremy Lee  BCompSci(Hons)
>>   The Unorthodox Engineers
>>
>
>
>
>
>

Reply via email to