Thanks Arush.
My scenario is that In standalone mode, if I have one worker, when I start
spark-shell, there will be one executor launched. But if I have 2 workers,
there will be 2 executors launched, so I am wondering the mechanism of
executor allocation.
Is it possible to specify how many executors I want in the code ?

On Tue, May 26, 2015 at 5:57 PM, Arush Kharbanda <ar...@sigmoidanalytics.com
> wrote:

> I believe you would be restricted by the number of cores you have in your
> cluster. Having a worker running without a core is useless.
>
> On Tue, May 26, 2015 at 3:04 PM, canan chen <ccn...@gmail.com> wrote:
>
>> In spark standalone mode, there will be one executor per worker. I am
>> wondering how many executor can I acquire when I submit app ? Is it greedy
>> mode (as many as I can acquire )?
>>
>
>
>
> --
>
> [image: Sigmoid Analytics] <http://htmlsig.com/www.sigmoidanalytics.com>
>
> *Arush Kharbanda* || Technical Teamlead
>
> ar...@sigmoidanalytics.com || www.sigmoidanalytics.com
>

Reply via email to