hi,

thanks for your answer!

I have few more:

1) the file /root/spark/conf/slaves , has the full DNS names of servers (
ec2-52-26-7-137.us-west-2.compute.amazonaws.com), did you add there the
internal ip?
2) You call to start-all. Isn't it too aggressive? Let's say I have 20
slaves up, and I want to add one more, why should we stop the entire
cluster for this?

thanks, nizan

On Thu, May 28, 2015 at 10:19 AM, Akhil Das <ak...@sigmoidanalytics.com>
wrote:

> I do this way:
>
> - Launch a new instance by clicking on the slave instance and choose *launch
> more like this *
> *- *Once its launched, ssh into it and add the master public key to
> .ssh/authorized_keys
> - Add the slaves internal IP to the master's conf/slaves file
> - do sbin/start-all.sh and it will show up along with other slaves.
>
>
>
> Thanks
> Best Regards
>
> On Thu, May 28, 2015 at 12:29 PM, nizang <ni...@windward.eu> wrote:
>
>> hi,
>>
>> I'm working on spark standalone system on ec2, and I'm having problems on
>> resizing the cluster (meaning - adding or removing slaves).
>>
>> In the basic ec2 scripts
>> (http://spark.apache.org/docs/latest/ec2-scripts.html), there's only
>> script
>> for lunching the cluster, not adding slaves to it. On the spark-standalone
>> page
>> (
>> http://spark.apache.org/docs/latest/spark-standalone.html#cluster-launch-scripts
>> ),
>> I can see only options for stopping and starting slaves, not adding them.
>>
>> What I try to do now (as a bad workaround...), is the following:
>>
>> 1) Go to the ec2 UI, create image from the current slave
>> 2) Lunch new instance based on this image
>> 3) Copy the public DNS of this slave
>> 4) SSH to the master, and edit the file
>> "/root/spark-ec2/ec2-variables.sh",
>> and add the DNS to the "export SLAVES" variables
>> 5) Running the script /root/spark-ec2/setup.sh
>>
>> After doing the above steps, I can see the new slave in the UI (8080 port)
>> of the master. However, this solution is bad for many reasons:
>>
>> 1) It requires many manual steps
>> 2) It requires stopping and starting the cluster
>> 3) There's no auto-detection in case slave stopped
>>
>> and many other reasons...
>>
>> oes anybody have another idea on how to add/remove slaves for standalone
>> on
>> a simple and safe way?
>>
>> thanks, nizan
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/Adding-slaves-on-spark-standalone-on-ec2-tp23064.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
>> For additional commands, e-mail: user-h...@spark.apache.org
>>
>>
>

Reply via email to