Can you go through the Mesos UI and look at the driver/executor log from steer 
file and see what the problem is?

Tim

> On Mar 1, 2016, at 8:05 AM, Ashish Soni <asoni.le...@gmail.com> wrote:
> 
> Not sure what is the issue but i am getting below error  when i try to run 
> spark PI example
> 
> Blacklisting Mesos slave value: "5345asdasdasdkas234234asdasdasdasd"
>    due to too many failures; is Spark installed on it?
>     WARN TaskSchedulerImpl: Initial job has not accepted any resources; check 
> your cluster UI to ensure that workers are registered and have sufficient 
> resources
> 
>> On Mon, Feb 29, 2016 at 1:39 PM, Sathish Kumaran Vairavelu 
>> <vsathishkuma...@gmail.com> wrote:
>> May be the Mesos executor couldn't find spark image or the constraints are 
>> not satisfied. Check your Mesos UI if you see Spark application in the 
>> Frameworks tab
>> 
>>> On Mon, Feb 29, 2016 at 12:23 PM Ashish Soni <asoni.le...@gmail.com> wrote:
>>> What is the Best practice , I have everything running as docker container 
>>> in single host ( mesos and marathon also as docker container )  and 
>>> everything comes up fine but when i try to launch the spark shell i get 
>>> below error
>>> 
>>> 
>>> SQL context available as sqlContext.
>>> 
>>> scala> val data = sc.parallelize(1 to 100)
>>> data: org.apache.spark.rdd.RDD[Int] = ParallelCollectionRDD[0] at 
>>> parallelize at <console>:27
>>> 
>>> scala> data.count
>>> [Stage 0:>                                                          (0 + 0) 
>>> / 2]16/02/29 18:21:12 WARN TaskSchedulerImpl: Initial job has not accepted 
>>> any resources; check your cluster UI to ensure that workers are registered 
>>> and have sufficient resources
>>> 16/02/29 18:21:27 WARN TaskSchedulerImpl: Initial job has not accepted any 
>>> resources; check your cluster UI to ensure that workers are registered and 
>>> have sufficient resources
>>>  
>>> 
>>> 
>>>> On Mon, Feb 29, 2016 at 12:04 PM, Tim Chen <t...@mesosphere.io> wrote:
>>>> No you don't have to run Mesos in docker containers to run Spark in docker 
>>>> containers.
>>>> 
>>>> Once you have Mesos cluster running you can then specfiy the Spark 
>>>> configurations in your Spark job (i.e: 
>>>> spark.mesos.executor.docker.image=mesosphere/spark:1.6) and Mesos will 
>>>> automatically launch docker containers for you.
>>>> 
>>>> Tim
>>>> 
>>>>> On Mon, Feb 29, 2016 at 7:36 AM, Ashish Soni <asoni.le...@gmail.com> 
>>>>> wrote:
>>>>> Yes i read that and not much details here.
>>>>> 
>>>>> Is it true that we need to have spark installed on each mesos docker 
>>>>> container ( master and slave ) ...
>>>>> 
>>>>> Ashish
>>>>> 
>>>>>> On Fri, Feb 26, 2016 at 2:14 PM, Tim Chen <t...@mesosphere.io> wrote:
>>>>>> https://spark.apache.org/docs/latest/running-on-mesos.html should be the 
>>>>>> best source, what problems were you running into?
>>>>>> 
>>>>>> Tim
>>>>>> 
>>>>>>> On Fri, Feb 26, 2016 at 11:06 AM, Yin Yang <yy201...@gmail.com> wrote:
>>>>>>> Have you read this ?
>>>>>>> https://spark.apache.org/docs/latest/running-on-mesos.html
>>>>>>> 
>>>>>>>> On Fri, Feb 26, 2016 at 11:03 AM, Ashish Soni <asoni.le...@gmail.com> 
>>>>>>>> wrote:
>>>>>>>> Hi All , 
>>>>>>>> 
>>>>>>>> Is there any proper documentation as how to run spark on mesos , I am 
>>>>>>>> trying from the last few days and not able to make it work.
>>>>>>>> 
>>>>>>>> Please help
>>>>>>>> 
>>>>>>>> Ashish
> 

Reply via email to