Hello Pavel: Thanks for the pointers.
For standalone cluster manager: I understand that I just have to start several masters with a subset of slaves attached. Then each master will listen on a different pair of <hostname,port>, allowing me to spark-submit to any of these pairs depending on the subset of machines I want to use. For Mesos: I haven't used Mesos much. Any references or documentation I can use to set this up? Best Regards 2017-02-07 11:36 GMT+01:00 Pavel Plotnikov <pavel.plotni...@team.wrike.com>: > Hi, Alvaro > You can create different clusters using standalone cluster manager, and > than manage subset of machines through submitting application on different > masters. Or you can use Mesos attributes to mark subset of workers and > specify it in spark.mesos.constraints > > > On Tue, Feb 7, 2017 at 1:21 PM Alvaro Brandon <alvarobran...@gmail.com> > wrote: > >> Hello all: >> >> I have the following scenario. >> - I have a cluster of 50 machines with Hadoop and Spark installed on >> them. >> - I want to launch one Spark application through spark submit. However I >> want this application to run on only a subset of these machines, >> disregarding data locality. (e.g. 10 machines) >> >> Is this possible?. Is there any option in the standalone scheduler, YARN >> or Mesos that allows such thing?. >> >> >>