I believe the issue may be using binaries with CDH. What I need to do is
build source with the hadoop, spark and switches for yarn, right?

Thanks,

Abhi

On Wed, Feb 17, 2016 at 9:02 PM, Jeff Zhang <zjf...@gmail.com> wrote:

> According the log, spark interpreter is started in yarn-client as
> application_1455038611898_0015. Could you check the yarn application of
> this app ?
>
> On Thu, Feb 18, 2016 at 1:26 AM, Abhi Basu <9000r...@gmail.com> wrote:
>
>> Additional info, installed Zeppelin 0.56 using binaries on cdh5.1 /spark
>> 1.5.0
>>
>> Any help is appreciated.
>>
>> Thanks,
>>
>> Abhi
>>
>> On Wed, Feb 17, 2016 at 9:07 AM, Abhi Basu <9000r...@gmail.com> wrote:
>>
>>> Logs attached. Am I supposed to edit the spark location in the zeppelin
>>> config file? All I have changed is the hadoop conf folder.
>>>
>>> Thanks,
>>>
>>> Abhi
>>>
>>> On Tue, Feb 16, 2016 at 5:29 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>
>>>> Can you check zeppelin log to confirm whether it is running in
>>>> yarn-client mode ? I suspect it is still in local mode. Spark require
>>>> python version of driver and executor to be the same. In your case it
>>>> should fail if driver is python2.7 while executor is python 2.6
>>>>
>>>> On Wed, Feb 17, 2016 at 9:03 AM, Abhi Basu <9000r...@gmail.com> wrote:
>>>>
>>>>> I have a 6 node cluster and 1 edge node to access. The edge node has
>>>>> Python 2.7 + NLTK + other libraries + hadoop client and Zeppelin 
>>>>> installed.
>>>>> All hadoop nodes have Python 2.6 and no other additional libraries.
>>>>>
>>>>> Running Zeppelin and my python code (with NLTK) is running under
>>>>> pyspark interpreter fine. It must be running locally as I have not
>>>>> distributed the python libraries to the other nodes yet. I dont see any
>>>>> errors in my Yarn logs either.
>>>>>
>>>>> This is my interpreter setup. Can you please tell  me how this is
>>>>> working?
>>>>>
>>>>> Also, if it is working locally, how to distribute over multiple nodes?
>>>>>
>>>>>
>>>>> Thanks,
>>>>>
>>>>> Abhi
>>>>>
>>>>> spark %spark (default), %pyspark, %sql, %dep edit  restart  remove
>>>>> Properties
>>>>> namevalue
>>>>> args
>>>>> master yarn-client
>>>>> spark.app.name Zeppelin-App
>>>>> spark.cores.max 4
>>>>> spark.executor.memory 1024m
>>>>> zeppelin.dep.additionalRemoteRepository spark-packages,
>>>>> http://dl.bintray.com/spark-packages/maven,false;
>>>>> zeppelin.dep.localrepo local-repo
>>>>> zeppelin.pyspark.python /usr/local/bin/python2.7
>>>>> zeppelin.spark.concurrentSQL true
>>>>> zeppelin.spark.maxResult 1000
>>>>> zeppelin.spark.useHiveContext true
>>>>>
>>>>>
>>>>> --
>>>>> Abhi Basu
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Best Regards
>>>>
>>>> Jeff Zhang
>>>>
>>>
>>>
>>>
>>> --
>>> Abhi Basu
>>>
>>
>>
>>
>> --
>> Abhi Basu
>>
>
>
>
> --
> Best Regards
>
> Jeff Zhang
>



-- 
Abhi Basu

Reply via email to