Hi Jeff, I've tried the release from http://zeppelin.apache.org/download.html, both in a docker and without a docker. They both have the same issue as previously described.
Can I somehow set spark.master to "local[*]" in zeppelin, perhaps using some environment variable? When is the next Zeppelin 0.9.0 docker image planned to be released? Best Regards, Patrik Iselind On Sun, May 10, 2020 at 9:26 AM Jeff Zhang <zjf...@gmail.com> wrote: > Hi Patric, > > Do you mind to try the 0.9.0-preview, it might be an issue of docker > container. > > http://zeppelin.apache.org/download.html > > > > Patrik Iselind <patrik....@gmail.com> 于2020年5月10日周日 上午2:30写道: > >> Hello Jeff, >> >> Thank you for looking into this for me. >> >> Using the latest pushed docker image for 0.9.0 (image ID 92890adfadfb, >> built 6 weeks ago), I still see the same issue. My image has the >> digest "apache/zeppelin@sha256 >> :0691909f6884319d366f5d3a5add8802738d6240a83b2e53e980caeb6c658092". >> >> If it's not on the tip of master, could you guys please release a newer >> 0.9.0 image? >> >> Best Regards, >> Patrik Iselind >> >> >> On Sat, May 9, 2020 at 4:03 PM Jeff Zhang <zjf...@gmail.com> wrote: >> >>> This might be a bug of 0.8, I tried that in 0.9 (master branch), it >>> works for me. >>> >>> print(sc.master) >>> print(sc.defaultParallelism) >>> >>> --- >>> local[*] 8 >>> >>> >>> Patrik Iselind <patrik....@gmail.com> 于2020年5月9日周六 下午8:34写道: >>> >>>> Hi, >>>> >>>> First comes some background, then I have some questions. >>>> >>>> *Background* >>>> I'm trying out Zeppelin 0.8.2 based on the Docker image. My Docker file >>>> looks like this: >>>> >>>> ```Dockerfile >>>> FROM apache/zeppelin:0.8.2 >>>> >>>> >>>> # Install Java and some tools >>>> RUN apt-get -y update &&\ >>>> DEBIAN_FRONTEND=noninteractive \ >>>> apt -y install vim python3-pip >>>> >>>> RUN python3 -m pip install -U pyspark >>>> >>>> ENV PYSPARK_PYTHON python3 >>>> ENV PYSPARK_DRIVER_PYTHON python3 >>>> ``` >>>> >>>> When I start a section like so >>>> >>>> ```Zeppelin paragraph >>>> %pyspark >>>> >>>> print(sc) >>>> print() >>>> print(dir(sc)) >>>> print() >>>> print(sc.master) >>>> print() >>>> print(sc.defaultParallelism) >>>> ``` >>>> >>>> I get the following output >>>> >>>> ```output >>>> <SparkContext master=local appName=Zeppelin> ['PACKAGE_EXTENSIONS', >>>> '__class__', '__delattr__', '__dict__', '__dir__', '__doc__', '__enter__', >>>> '__eq__', '__exit__', '__format__', '__ge__', '__getattribute__', >>>> '__getnewargs__', '__gt__', '__hash__', '__init__', '__le__', '__lt__', >>>> '__module__', '__ne__', '__new__', '__reduce__', '__reduce_ex__', >>>> '__repr__', '__setattr__', '__sizeof__', '__str__', '__subclasshook__', >>>> '__weakref__', '_accumulatorServer', '_active_spark_context', '_batchSize', >>>> '_callsite', '_checkpointFile', '_conf', '_dictToJavaMap', '_do_init', >>>> '_ensure_initialized', '_gateway', '_getJavaStorageLevel', >>>> '_initialize_context', '_javaAccumulator', '_jsc', '_jvm', '_lock', >>>> '_next_accum_id', '_pickled_broadcast_vars', '_python_includes', >>>> '_repr_html_', '_temp_dir', '_unbatched_serializer', 'accumulator', >>>> 'addFile', 'addPyFile', 'appName', 'applicationId', 'binaryFiles', >>>> 'binaryRecords', 'broadcast', 'cancelAllJobs', 'cancelJobGroup', >>>> 'defaultMinPartitions', 'defaultParallelism', 'dump_profiles', 'emptyRDD', >>>> 'environment', 'getConf', 'getLocalProperty', 'getOrCreate', 'hadoopFile', >>>> 'hadoopRDD', 'master', 'newAPIHadoopFile', 'newAPIHadoopRDD', >>>> 'parallelize', 'pickleFile', 'profiler_collector', 'pythonExec', >>>> 'pythonVer', 'range', 'runJob', 'sequenceFile', 'serializer', >>>> 'setCheckpointDir', 'setJobGroup', 'setLocalProperty', 'setLogLevel', >>>> 'setSystemProperty', 'show_profiles', 'sparkHome', 'sparkUser', >>>> 'startTime', 'statusTracker', 'stop', 'textFile', 'uiWebUrl', 'union', >>>> 'version', 'wholeTextFiles'] local 1 >>>> ``` >>>> >>>> This even though the "master" property in the interpretter is set to >>>> "local[*]". I'd like to use all cores on my machine. To do that I have to >>>> explicitly create the "spark.master" property in the spark >>>> interpretter with the value "local[*]", then I get >>>> >>>> ```new output >>>> <SparkContext master=local[*] appName=Zeppelin> ['PACKAGE_EXTENSIONS', >>>> '__class__', '__delattr__', '__dict__', '__dir__', '__doc__', '__enter__', >>>> '__eq__', '__exit__', '__format__', '__ge__', '__getattribute__', >>>> '__getnewargs__', '__gt__', '__hash__', '__init__', '__le__', '__lt__', >>>> '__module__', '__ne__', '__new__', '__reduce__', '__reduce_ex__', >>>> '__repr__', '__setattr__', '__sizeof__', '__str__', '__subclasshook__', >>>> '__weakref__', '_accumulatorServer', '_active_spark_context', '_batchSize', >>>> '_callsite', '_checkpointFile', '_conf', '_dictToJavaMap', '_do_init', >>>> '_ensure_initialized', '_gateway', '_getJavaStorageLevel', >>>> '_initialize_context', '_javaAccumulator', '_jsc', '_jvm', '_lock', >>>> '_next_accum_id', '_pickled_broadcast_vars', '_python_includes', >>>> '_repr_html_', '_temp_dir', '_unbatched_serializer', 'accumulator', >>>> 'addFile', 'addPyFile', 'appName', 'applicationId', 'binaryFiles', >>>> 'binaryRecords', 'broadcast', 'cancelAllJobs', 'cancelJobGroup', >>>> 'defaultMinPartitions', 'defaultParallelism', 'dump_profiles', 'emptyRDD', >>>> 'environment', 'getConf', 'getLocalProperty', 'getOrCreate', 'hadoopFile', >>>> 'hadoopRDD', 'master', 'newAPIHadoopFile', 'newAPIHadoopRDD', >>>> 'parallelize', 'pickleFile', 'profiler_collector', 'pythonExec', >>>> 'pythonVer', 'range', 'runJob', 'sequenceFile', 'serializer', >>>> 'setCheckpointDir', 'setJobGroup', 'setLocalProperty', 'setLogLevel', >>>> 'setSystemProperty', 'show_profiles', 'sparkHome', 'sparkUser', >>>> 'startTime', 'statusTracker', 'stop', 'textFile', 'uiWebUrl', 'union', >>>> 'version', 'wholeTextFiles'] local[*] 8 >>>> ``` >>>> This is what I want. >>>> >>>> *The Questions* >>>> >>>> - Why is the "master" property not used in the created SparkContext? >>>> - How do I add the spark.master property to the docker image? >>>> >>>> >>>> Any hint or support you can provide would be greatly appreciated. >>>> >>>> Yours Sincerely, >>>> Patrik Iselind >>>> >>> >>> >>> -- >>> Best Regards >>> >>> Jeff Zhang >>> >> > > -- > Best Regards > > Jeff Zhang >