Hi,

You can set SPARK_HOME in your interpreter tab instead of setting in
zeppelin-env.sh. Remove SPARK_HOME in your zeppelin-env.sh and set it up
interpreter tabs with two different spark interpreters.

On Tue, Oct 24, 2017 at 11:39 AM, Jeff Zhang <zjf...@gmail.com> wrote:

>
> Take a look at spark submit command
>
> Refer this link for details
> https://community.hortonworks.com/articles/70658/how-to-
> diagnose-zeppelin.html
>
>
>
> Meier, Alexander <alexander.me...@t-systems-dmc.com>于2017年10月24日周二
> 上午2:54写道:
>
>> Hi there
>>
>> Any ideas why the new interpreter is not using the specified SPARK_HOME?
>>
>> Best regards
>> Alex Meier
>>
>>
>> On 18 Oct 2017, at 11:14, Meier, Alexander <Alexander.Meier@t-systems-
>> dmc.com> wrote:
>>
>> Hi there
>>
>> We have a CDH 5.9 Cluster where zeppelin 0.7.1is running fine with spark
>> 1.6, impala, hive, hbase and apache phoenix. As we have both, Spark 1.6 and
>> Spark 2 distributed and activated on the cluster, we now also want to use a
>> spark2 interpreter in zeppelin.
>>
>> I created a new interpreter %spark2 in zeppelin belonging to the spark
>> interpreter group.
>> Also I added an interpreter property “SPARK_HOME” with
>> value /opt/cloudera/parcels/SPARK2-2.1.0.cloudera1-1.cdh5.
>> 7.0.p0.120904/lib/spark2
>> (I have also tried to set SPARK_HOME to 
>> /opt/cloudera/parcels/SPARK2/lib/spark2
>> with the same result.)
>>
>> When testing the interpreter:
>>
>> %spark2
>> sc.version
>> spark.version
>>
>> we get
>>
>> res0: String = 1.6.0
>> <console>:28: error: not found: value spark
>>               spark.version
>>
>>
>> Why is the interpreter still using Spark 1.6? I thought the interpreter
>> property would override the SPARK_HOME setting in zeppelin-env.sh
>>
>> Any ideas?
>>
>> btw: we also use impersonation and run the interpreter instantiated per
>> user in isolated mode, but that  as "just in case” info.
>>
>> Best regards
>> Alex
>>
>>
>>
>>


-- 
이종열, Jongyoul Lee, 李宗烈
http://madeng.net

Reply via email to