Hi Stefano, Can you tell me which spark you are using? If you have set cdh spark as SPARK_HOME, could you confirm that you installed spark-pyspark package?
Let me know I On Wed, Sep 28, 2016 at 12:06 AM Stefano Ortolani <ostef...@gmail.com> wrote: > Hi, > > I have been using Zeppelin for quite a while without issues. > Lately, I have been trying to configure pyspark, but I can't seem to make > it work. > Using the local pyspark works perfectly, but then, regardless of the > PYTHONPATH I specify in zeppelin-env.sh, any usage of pyspark results in: > > Error from python worker: > /usr/bin/python: No module named pyspark > PYTHONPATH was: > > /usr/lib/spark/python/lib/pyspark.zip:/usr/lib/spark/python/lib/py4j-0.9-src.zip:/usr/lib/spark/lib/spark-assembly-1.6.1-hadoop2.6.0.jar > > > Now, that pyspark.zip, is actually not there (my distribution have a > tar.gz inside /usr/lib/spark/lib), but no matter what I set, the PYTHONPATH > does not change. > > Any idea? > > Regards, > Stefano >