> This is not independent programmatic way of running of Spark job on Yarn cluster.
The example I created simply demonstrates how to wire up the classpath so that spark submit can be called programmatically. For my use case, I wanted to hold open a connection so I could send tasks to the executors on demand. If you were to submit this via yarn-cluster mode, it would only require any extra files be placed on the executors, if needed. On Wed, Jun 17, 2015 at 9:01 PM, Elkhan Dadashov <elkhan8...@gmail.com> wrote: > This is not independent programmatic way of running of Spark job on Yarn > cluster. > > That example demonstrates running on *Yarn-client* mode, also will be > dependent of Jetty. Users writing Spark programs do not want to depend on > that. > > I found this SparkLauncher class introduced in Spark 1.4 version ( > https://github.com/apache/spark/tree/master/launcher) which allows > running Spark jobs in programmatic way. > > SparkLauncher exists in Java and Scala APIs, but I could not find in > Python API. > > Did not try it yet, but seems promising. > > Example: > > import org.apache.spark.launcher.SparkLauncher; > > public class MyLauncher { > > public static void main(String[] args) throws Exception { > > Process spark = new SparkLauncher() > > .setAppResource("/my/app.jar") > > .setMainClass("my.spark.app.Main") > > .setMaster("local") > > .setConf(SparkLauncher.DRIVER_MEMORY, "2g") > > .launch(); > > spark.waitFor(); > > } > > } > > } > > > > On Wed, Jun 17, 2015 at 5:51 PM, Corey Nolet <cjno...@gmail.com> wrote: > >> An example of being able to do this is provided in the Spark Jetty Server >> project [1] >> >> [1] https://github.com/calrissian/spark-jetty-server >> >> On Wed, Jun 17, 2015 at 8:29 PM, Elkhan Dadashov <elkhan8...@gmail.com> >> wrote: >> >>> Hi all, >>> >>> Is there any way running Spark job in programmatic way on Yarn cluster >>> without using spark-submit script ? >>> >>> I cannot include Spark jars on my Java application (due o dependency >>> conflict and other reasons), so I'll be shipping Spark assembly uber jar >>> (spark-assembly-1.3.1-hadoop2.3.0.jar) to Yarn cluster, and then execute >>> job (Python or Java) on Yarn-cluster. >>> >>> So is there any way running Spark job implemented in python file/Java >>> class without calling it through spark-submit script ? >>> >>> Thanks. >>> >>> >>> >> > > > -- > > Best regards, > Elkhan Dadashov >