[mailto:so...@cloudera.com]
Sent: Thursday, February 26, 2015 2:11 AM
To: Judy Nash
Cc: user@spark.apache.org
Subject: Re: spark standalone with multiple executors in one work node
--num-executors is the total number of executors. In YARN there is not quite
the same notion of a Spark worker. Of
--num-executors is the total number of executors. In YARN there is not
quite the same notion of a Spark worker. Of course, one worker has an
executor for each running app, so yes, but you mean for one app? it's
possible, though not usual, to run multiple executors for one app on
one worker. This ma
My understanding is if you run multi applications on the work node, then each
application will have an executorbackend process and an executor as well.
bit1...@163.com
From: Judy Nash
Date: 2015-02-26 09:58
To: user@spark.apache.org
Subject: spark standalone with multiple executors in one wor