@moon, any thoughts? Thanks
> On Dec 8, 2015, at 5:11 PM, Fengdong Yu <fengdo...@everstring.com> wrote: > > My notebook: > > %pyspark > from pyspark.sql.types import BooleanType > sqlContext.udf.register("is_empty",lambda x : True if not x else False, > BooleanType()) > sqlContext.sql("select is_empty(job_functions), job_functions from test_table > limit 10").show() > > > It can import modules, but throw exceptions after job submitted. > > > > >> On Dec 8, 2015, at 5:04 PM, Amjad ALSHABANI <ashshab...@gmail.com >> <mailto:ashshab...@gmail.com>> wrote: >> >> Hello, >> >> Are you sure that you ve installed the module pyspark. >> >> Please check your spark installation directory if you could see the python >> sub-directory >> Amjad >> >> On Dec 8, 2015 9:55 AM, "Fengdong Yu" <fengdo...@everstring.com >> <mailto:fengdo...@everstring.com>> wrote: >> Hi >> >> I am using Zeppelin-0.5.5 with Spark 1.5.2 >> >> It cannot find pyspark module. >> >> >> Error from python worker: >> /usr/local/bin/python: No module named pyspark >> PYTHONPATH was: >> >> >> >> I’ve configured pyspark in zeppelin-env.sh: >> >> export >> PYTHONPATH=$SPARK_HOME/python:$SPARK_HOME/python/lib/py4j-0.8.2.1-src.zip:$SPARK_HOME/python/lib/pyspark.zip >> >> >> any others I skipped? Thanks >> >> >> >