Hi!

I'm trying to submit Spark applications to Mesos using the 'cluster' deploy mode. I'm using Marathon as the container orchestration platform and launching the Mesos Cluster Dispatcher through it. I'm using Spark 1.6.1 with Scala 2.11.

I'm able to successfully communicate with the cluster dispatcher, but executors that run on a node different from the one where the driver is started fail to run. I believe that they're assuming the same (local) classpath as the driver, and thus failing to run. Here's what I see from the failed tasks on the executor nodes:

sh -c ' "/tmp/mesos/slaves/3b945207-1aac-497d-b2fb-9671d3d0646b-S1/frameworks/b925b271-ce42-48c9-8474-9e19304c2d20-0002/executors/driver-20160418101719-0001/runs/39272646-d305-40a6-9431-5efcacc67b80/spark-1.6.1-bin-2.4.0/bin/spark-class" org.apache.spark.executor.CoarseGrainedExecutorBackend --driver-url spark://CoarseGrainedScheduler@<driver_ip>:35600 --executor-id 00706e76-f790-465d-b4cd-edd3934ddb6a-S1 --hostname <executor_hostname> --cores 5 --app-id 00706e76-f790-465d-b4cd-edd3934ddb6a-0048'
Command exited with status 127 (pid: 4118)

When the executors are started in the same node where the driver is, the application runs ok.

Here's how I'm submitting the applications:

bin/spark-submit --master mesos://<mesos_cluster_dispatcher_hostname>:7077 --deploy-mode cluster --conf "spark.executor.uri=<uri_to_spark>" --executor-memory 5G --total-executor-cores 5 --supervise --class org.apache.spark.examples.SparkPi <uri_to_spark_example> 10

Does anyone have any ideas on what I might be doing wrong?


        Joao

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to