In YARN cluster mode, there is no Spark master, since YARN is your resource manager. Yes you could force your AM somehow to run on the same node as the RM, but why -- what do think is faster about that?
On Tue, Mar 10, 2015 at 10:06 AM, Harika <matha.har...@gmail.com> wrote: > Hi all, > > I have Spark cluster setup on YARN with 4 nodes(1 master and 3 slaves). When > I run an application, YARN chooses, at random, one Application Master from > among the slaves. This means that my final computation is being carried > only on two slaves. This decreases the performance of the cluster. > > 1. Is this the correct way of configuration? What is the architecture of > Spark on YARN? > 2. Is there a way in which I can run Spark master, YARN application master > and resource manager on a single node?(so that I can use three other nodes > for the computation) > > Thanks > Harika > > > > > > -- > View this message in context: > http://apache-spark-user-list.1001560.n3.nabble.com/Spark-on-YARN-architecture-tp21986.html > Sent from the Apache Spark User List mailing list archive at Nabble.com. > > --------------------------------------------------------------------- > To unsubscribe, e-mail: user-unsubscr...@spark.apache.org > For additional commands, e-mail: user-h...@spark.apache.org > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org