You need to specify SPARK_HOME which is where spark installed.

Alex Dzhagriev <dzh...@gmail.com>于2018年10月20日周六 上午3:12写道:

> Hello,
>
> I have a remote Spark cluster and I'm trying to use it by setting the
> spark interpreter property:
>
> master spark://spark-cluster-master:7077, however I'm getting the
> following error:
>
> java.lang.RuntimeException: SPARK_HOME is not specified in
> interpreter-setting for non-local mode, if you specify it in
> zeppelin-env.sh, please move that into interpreter setting
>
> version: Docker Image 0.8.0
>
> Thanks, Alex.
>

Reply via email to