Thanks for your reply,
I have tried setting it within zeppelin-env.sh but it doesn't work any
better.

Thanks

On Wed, Nov 2, 2016 at 2:13 AM, Hyung Sung Shim <hss...@nflabs.com> wrote:

> Hello.
> You should set the HADOOP_CONF_DIR to /usr/local/lib/hadoop/etc/hadoop/
> in the conf/zeppelin-env.sh.
> Thanks.
> 2016년 11월 2일 (수) 오전 5:07, Benoit Hanotte <benoit.h...@gmail.com>님이 작성:
>
>> Hello,
>>
>> I'd like to use zeppelin on my local computer and use it to run spark
>> executors on a distant yarn cluster since I can't easily install zeppelin
>> on the cluster gateway.
>>
>> I installed the correct hadoop version (2.6), and compiled zeppelin (from
>> the master branch) as following:
>>
>> *mvn clean package -DskipTests -Phadoop-2.6
>> -Dhadoop.version=2.6.0-cdh5.5.0 -Pyarn -Pspark-2.0 -Pscala-2.11*
>>
>> I also set HADOOP_HOME_DIR to /usr/local/lib/hadoop where my hadoop is
>> installed (I also tried with /usr/local/lib/hadoop/etc/hadoop/ where the
>> conf files such as yarn-site.xml are). I set yarn.resourcemanager.hostname
>> to the resource manager of the cluster (I copied the value from the config
>> file on the cluster) but when I start a spark command it still tries to
>> connect to 0.0.0.0:8032 as one can see in the logs:
>>
>> *INFO [2016-11-01 20:48:26,581] ({pool-2-thread-2}
>> Client.java[handleConnectionFailure]:862) - Retrying connect to server:
>> 0.0.0.0/0.0.0.0:8032 <http://0.0.0.0/0.0.0.0:8032>. Already tried 9
>> time(s); retry policy is RetryUpToMaximumCountWithFixedSleep(maxRetries=10,
>> sleepTime=1000 MILLISECONDS)*
>>
>> Am I missing something something? Is there any additional parameters to
>> set?
>>
>> Thanks!
>>
>> Benoit
>>
>>
>>

Reply via email to