Dear all, I have spark1.0.0 and mesos0.18.1. After setting in mesos and spark and starting the mesos cluster, I try to run the pyspark job by the command below:
spark-submit /path/to/my_pyspark_job.py --master mesos://192.168.0.21:5050 It occurs error below: 14/07/29 18:40:49 INFO server.Server: jetty-8.y.z-SNAPSHOT 14/07/29 18:40:49 INFO server.AbstractConnector: Started SelectChannelConnector@0.0.0.0:4041 14/07/29 18:40:49 INFO ui.SparkUI: Started SparkUI at http://CentOS-19:4041 14/07/29 18:40:49 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 14/07/29 18:40:50 INFO scheduler.EventLoggingListener: Logging events to /tmp/spark-events/my_test.py-1406630449771 14/07/29 18:40:50 INFO util.Utils: Copying /home/daijia/deal_three_word/my_test.py to /tmp/spark-4365b01d-b57a-4abb-b39c-cb57b83a28ce/my_test.py 14/07/29 18:40:50 INFO spark.SparkContext: Added file file:/home/daijia/deal_three_word/my_test.py at http://192.168.3.91:51188/files/my_test.py with timestamp 1406630450333 I0729 18:40:50.440551 15033 sched.cpp:121] Version: 0.18.1 I0729 18:40:50.442450 15035 sched.cpp:217] New master detected at master@192.168.3.91:5050 I0729 18:40:50.442570 15035 sched.cpp:225] No credentials provided. Attempting to register without authentication I0729 18:40:50.443234 15036 sched.cpp:391] Framework registered with 20140729-174911-1526966464-5050-13758-0006 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Registered as framework ID 20140729-174911-1526966464-5050-13758-0006 14/07/29 18:40:50 INFO spark.SparkContext: Starting job: count at /home/daijia/deal_three_word/my_test.py:27 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Mesos task 0 is now TASK_LOST 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Mesos task 1 is now TASK_LOST 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Mesos task 3 is now TASK_LOST 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Blacklisting Mesos slave value: "20140729-163345-1526966464-5050-10913-0" due to too many failures; is Spark installed on it? 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Mesos task 2 is now TASK_LOST 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Blacklisting Mesos slave value: "20140729-163345-1526966464-5050-10913-2" due to too many failures; is Spark installed on it? 14/07/29 18:40:50 INFO scheduler.DAGScheduler: Got job 0 (count at /home/daijia/deal_three_word/my_test.py:27) with 2 output partitions (allowLocal=false) 14/07/29 18:40:50 INFO scheduler.DAGScheduler: Final stage: Stage 0(count at /home/daijia/deal_three_word/my_test.py:27) 14/07/29 18:40:50 INFO scheduler.DAGScheduler: Parents of final stage: List() 14/07/29 18:40:50 INFO scheduler.DAGScheduler: Missing parents: List() 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Mesos task 4 is now TASK_LOST 14/07/29 18:40:50 INFO scheduler.DAGScheduler: Submitting Stage 0 (PythonRDD[1] at RDD at PythonRDD.scala:37), which has no missing parents 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Mesos task 5 is now TASK_LOST 14/07/29 18:40:50 INFO mesos.CoarseMesosSchedulerBackend: Blacklisting Mesos slave value: "20140729-163345-1526966464-5050-10913-1" due to too many failures; is Spark installed on it? 14/07/29 18:40:50 INFO scheduler.DAGScheduler: Submitting 2 missing tasks from Stage 0 (PythonRDD[1] at RDD at PythonRDD.scala:37) 14/07/29 18:40:50 INFO scheduler.TaskSchedulerImpl: Adding task set 0.0 with 2 tasks 14/07/29 18:41:05 WARN scheduler.TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory 14/07/29 18:41:20 WARN scheduler.TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory 14/07/29 18:41:20 WARN scheduler.TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory It just repeats the last message. Here is my python scirpt: #!/usr/bin/env python #coding=utf-8 from pyspark import SparkContext sc = SparkContext() temp = [] for index in range(1000): temp.append(index) sc.parallelize(temp).count() So, the running command is right? Or some other reasons lead to the problem. Thanks in advance, Daijia -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/How-to-submit-Pyspark-job-in-mesos-tp10905.html Sent from the Apache Spark User List mailing list archive at Nabble.com.