I have the same configuration but when I run the program the process never
returns. status remain on running
log say SEND >> PROGRESS
forever....

On Sun, Jun 14, 2015 at 2:34 AM MrAsanjar . <afsan...@gmail.com> wrote:

> hi
> I had a similar issue, try these:
> 1) add following settings to zeppelin-env.sh ( it must be added there at
> this time)
> export MASTER=yarn-client
> export HADOOP_CONF_DIR=/etc/hadoop/conf
> export
> PYTHONPATH=/usr/lib/spark/python:/usr/lib/spark/python/lib/py4j-0.8.2.1-src.zip
> export SPARK_YARN_USER_ENV="PYTHONPATH=${PYTHONPATH}"
> export SPARK_HOME=/usr/lib/spark
> 2) Make sure /tmp has 1777 access permission
> 3) restart the daemon "zeppelin-daemon.sh restart"
>
> On Thu, Jun 11, 2015 at 2:13 PM, Marcel Hofmann <mhofmann....@gmail.com>
> wrote:
>
>> Hey everybody,
>>
>> I'm currently testing Zeppelin, but unfortunatley, I can't really get it
>> up and running. The example notebook is running just fine, and everything
>> works there, but a simple pyspark statement like:
>>
>> %pyspark
>> list = range(1,4)
>> print(list)
>>
>> will not execute. Looking at the interpretor-spark logs the only thing
>> happening is:
>>
>>  INFO [2015-06-11 12:05:21,664] ({pool-2-thread-3}
>> SchedulerFactory.java[jobStarted]:132) - Job
>> remoteInterpretJob_1434049521664 started by scheduler interpreter_836706052
>>  INFO [2015-06-11 12:05:21,667] ({pool-2-thread-3}
>> PySparkInterpreter.java[createPythonScript]:123) - File
>> /tmp/zeppelin_pyspark.py created
>>
>> I have set the following env-variables:
>> export SPARK_HOME=/home/biadmin/spark-1.3.0/spark-1.3.0-bin-hadoop2.4
>> export PYSPARK_PYTHON=python
>> export
>> PYTHONPATH=$SPARK_HOME/python:$SPARK_HOME/python/lib/py4j-0.8.2.1-src.zip
>>
>> Thanks for any advice!
>>
>> Marcel
>>
>>
>

Reply via email to