Can you go through the Mesos UI and look at the driver/executor log from steer file and see what the problem is?
Tim > On Mar 1, 2016, at 8:05 AM, Ashish Soni <asoni.le...@gmail.com> wrote: > > Not sure what is the issue but i am getting below error when i try to run > spark PI example > > Blacklisting Mesos slave value: "5345asdasdasdkas234234asdasdasdasd" > due to too many failures; is Spark installed on it? > WARN TaskSchedulerImpl: Initial job has not accepted any resources; check > your cluster UI to ensure that workers are registered and have sufficient > resources > >> On Mon, Feb 29, 2016 at 1:39 PM, Sathish Kumaran Vairavelu >> <vsathishkuma...@gmail.com> wrote: >> May be the Mesos executor couldn't find spark image or the constraints are >> not satisfied. Check your Mesos UI if you see Spark application in the >> Frameworks tab >> >>> On Mon, Feb 29, 2016 at 12:23 PM Ashish Soni <asoni.le...@gmail.com> wrote: >>> What is the Best practice , I have everything running as docker container >>> in single host ( mesos and marathon also as docker container ) and >>> everything comes up fine but when i try to launch the spark shell i get >>> below error >>> >>> >>> SQL context available as sqlContext. >>> >>> scala> val data = sc.parallelize(1 to 100) >>> data: org.apache.spark.rdd.RDD[Int] = ParallelCollectionRDD[0] at >>> parallelize at <console>:27 >>> >>> scala> data.count >>> [Stage 0:> (0 + 0) >>> / 2]16/02/29 18:21:12 WARN TaskSchedulerImpl: Initial job has not accepted >>> any resources; check your cluster UI to ensure that workers are registered >>> and have sufficient resources >>> 16/02/29 18:21:27 WARN TaskSchedulerImpl: Initial job has not accepted any >>> resources; check your cluster UI to ensure that workers are registered and >>> have sufficient resources >>> >>> >>> >>>> On Mon, Feb 29, 2016 at 12:04 PM, Tim Chen <t...@mesosphere.io> wrote: >>>> No you don't have to run Mesos in docker containers to run Spark in docker >>>> containers. >>>> >>>> Once you have Mesos cluster running you can then specfiy the Spark >>>> configurations in your Spark job (i.e: >>>> spark.mesos.executor.docker.image=mesosphere/spark:1.6) and Mesos will >>>> automatically launch docker containers for you. >>>> >>>> Tim >>>> >>>>> On Mon, Feb 29, 2016 at 7:36 AM, Ashish Soni <asoni.le...@gmail.com> >>>>> wrote: >>>>> Yes i read that and not much details here. >>>>> >>>>> Is it true that we need to have spark installed on each mesos docker >>>>> container ( master and slave ) ... >>>>> >>>>> Ashish >>>>> >>>>>> On Fri, Feb 26, 2016 at 2:14 PM, Tim Chen <t...@mesosphere.io> wrote: >>>>>> https://spark.apache.org/docs/latest/running-on-mesos.html should be the >>>>>> best source, what problems were you running into? >>>>>> >>>>>> Tim >>>>>> >>>>>>> On Fri, Feb 26, 2016 at 11:06 AM, Yin Yang <yy201...@gmail.com> wrote: >>>>>>> Have you read this ? >>>>>>> https://spark.apache.org/docs/latest/running-on-mesos.html >>>>>>> >>>>>>>> On Fri, Feb 26, 2016 at 11:03 AM, Ashish Soni <asoni.le...@gmail.com> >>>>>>>> wrote: >>>>>>>> Hi All , >>>>>>>> >>>>>>>> Is there any proper documentation as how to run spark on mesos , I am >>>>>>>> trying from the last few days and not able to make it work. >>>>>>>> >>>>>>>> Please help >>>>>>>> >>>>>>>> Ashish >