Re: Spark concurrency question

2015-02-08 Thread Sean Owen
On Sun, Feb 8, 2015 at 10:26 PM, java8964 wrote: > standalone one box environment, if I want to use all 48G memory allocated to > worker for my application, I should ask 48G memory for the executor in the > spark shell, right? Because 48G is too big for a JVM heap in normal case, I > can and shoul

RE: Spark concurrency question

2015-02-08 Thread java8964
t one partition (or one task for it) will be run by one executor? Is that one executor will run the task sequentially, but job concurrency comes from that multi executors could run synchronous, right? Thanks > From: so...@cloudera.com > Date: Sun, 8 Feb 2015 20:53:15 + > Subjec

Re: Spark concurrency question

2015-02-08 Thread Sean Owen
I think I have this right: You will run one executor per application per worker. Generally there is one worker per machine, and it manages all of the machine's resources. So if you want one app to use this whole machine you need to ask for 48G and 24 cores. That's better than splitting up the reso

Spark concurrency question

2015-02-08 Thread java8964
Hi, I have some questions about how the spark run the job concurrently. For example, if I setup the Spark on one standalone test box, which has 24 core and 64G memory. I setup the Worker memory to 48G, and Executor memory to 4G, and using spark-shell to run some jobs. Here is something confusing