Hyung, thx for your help. I opened these:

https://issues.apache.org/jira/browse/ZEPPELIN-1096 (this scala problem)
https://issues.apache.org/jira/browse/ZEPPELIN-1097 (similar looking python
problem)

LMK if I can provide more info or help in some way.

On Fri, Jul 1, 2016 at 5:08 AM, Hyung Sung Shim <hss...@nflabs.com> wrote:

> Hi Jonathan.
> Unfortunately I got same error in my test bed.
> Do you mind create an jira issue for this?
>
> 2016-07-01 3:19 GMT+09:00 Jonathan Esterhazy <jonathan.esterh...@gmail.com
> >:
>
>> I added this and still get the same exception. The same property is also
>> set in spark-defaults.conf.
>>
>> After that didn't work, I also tried adding --conf
>> spark.authenticate=true and --conf
>> spark.authenticate.enableSaslEncryption=true, to match the other related
>> settings in spark-defaults.conf. Still get the same classnotfoundexception.
>>
>>
>> On Thu, Jun 30, 2016 at 10:45 AM, Hyung Sung Shim <hss...@nflabs.com>
>> wrote:
>>
>>> Please add the *export SPARK_SUBMIT_OPTIONS="--conf
>>> spark.authenticate.secret=secret"* in conf/zeppelin-env.sh and restart
>>> zeppelin and retry your code.
>>>
>>>
>>> 2016-06-30 23:34 GMT+09:00 Jonathan Esterhazy <
>>> jonathan.esterh...@gmail.com>:
>>>
>>>> yes it does. i only see this problem in zeppelin.
>>>>
>>>> On Thu, Jun 30, 2016 at 7:05 AM, Hyung Sung Shim <hss...@nflabs.com>
>>>> wrote:
>>>>
>>>>> Hi Jonathan.
>>>>> It's not easy to build the test environments but I am working on this.
>>>>> I have question for you.
>>>>> Does your code working well on spark-shell in the spark.authenticate
>>>>> mode?
>>>>>
>>>>> 2016-06-30 22:47 GMT+09:00 Jonathan Esterhazy <
>>>>> jonathan.esterh...@gmail.com>:
>>>>>
>>>>>> Hyung, did you have any luck w/ zeppelin + spark authentication? I'm
>>>>>> quite stumped.
>>>>>>
>>>>>> thx.
>>>>>>
>>>>>> On Tue, Jun 28, 2016 at 9:11 PM, Hyung Sung Shim <hss...@nflabs.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Thank you.
>>>>>>> Let me try.
>>>>>>>
>>>>>>> 2016-06-28 22:18 GMT+09:00 Jonathan Esterhazy <
>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>
>>>>>>>> Hyung,
>>>>>>>>
>>>>>>>> Yes, here they are.
>>>>>>>>
>>>>>>>> zeppelin-env.sh:
>>>>>>>>
>>>>>>>> export ZEPPELIN_PORT=8890
>>>>>>>> export ZEPPELIN_CONF_DIR=/etc/zeppelin/conf
>>>>>>>> export ZEPPELIN_LOG_DIR=/var/log/zeppelin
>>>>>>>> export ZEPPELIN_PID_DIR=/var/run/zeppelin
>>>>>>>> export ZEPPELIN_PID=$ZEPPELIN_PID_DIR/zeppelin.pid
>>>>>>>> export ZEPPELIN_NOTEBOOK_DIR=/var/lib/zeppelin/notebook
>>>>>>>> export ZEPPELIN_WAR_TEMPDIR=/var/run/zeppelin/webapps
>>>>>>>> export MASTER=yarn-client
>>>>>>>> export SPARK_HOME=/usr/lib/spark
>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf
>>>>>>>> export
>>>>>>>> CLASSPATH=":/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf:/usr/share/aws/emr/emrfs/lib/*
>>>>>>>> :/usr/share/aws/emr/emrfs/auxlib/*"
>>>>>>>> export JAVA_HOME=/usr/lib/jvm/java-1.8.0
>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_BUCKET=mybucket
>>>>>>>> export ZEPPELIN_NOTEBOOK_S3_USER=zeppelin
>>>>>>>> export
>>>>>>>> ZEPPELIN_NOTEBOOK_STORAGE=org.apache.zeppelin.notebook.repo.S3NotebookRepo
>>>>>>>>
>>>>>>>> spark-defaults.conf:
>>>>>>>>
>>>>>>>> spark.master                     yarn
>>>>>>>> spark.driver.extraClassPath
>>>>>>>>  
>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>> spark.driver.extraLibraryPath
>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>> spark.executor.extraClassPath
>>>>>>>>  
>>>>>>>> /etc/hadoop/conf:/etc/hive/conf:/usr/lib/hadoop-lzo/lib/*:/usr/lib/hadoop/hadoop-aws.jar:/usr/share/aws/aws-java-sdk/*:/usr/share/aws/emr/emrfs/conf
>>>>>>>> :/usr/share/aws/emr/emrfs/lib/*:/usr/share/aws/emr/emrfs/auxlib/*
>>>>>>>> spark.executor.extraLibraryPath
>>>>>>>>  /usr/lib/hadoop/lib/native:/usr/lib/hadoop-lzo/lib/native
>>>>>>>> spark.eventLog.enabled           true
>>>>>>>> spark.eventLog.dir               hdfs:///var/log/spark/apps
>>>>>>>> spark.history.fs.logDirectory    hdfs:///var/log/spark/apps
>>>>>>>> spark.yarn.historyServer.address ip-172-30-54-30.ec2.internal:18080
>>>>>>>> spark.history.ui.port            18080
>>>>>>>> spark.shuffle.service.enabled    true
>>>>>>>> spark.driver.extraJavaOptions
>>>>>>>>  -Dlog4j.configuration=file:///etc/spark/conf/log4j.properties
>>>>>>>> -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70
>>>>>>>> -XX:MaxHeapFreeRatio=70
>>>>>>>> -XX:+CMSClassUnloadingEnabled -XX:MaxPermSize=512M
>>>>>>>> -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>> spark.dynamicAllocation.enabled  true
>>>>>>>> spark.executor.extraJavaOptions  -verbose:gc -XX:+PrintGCDetails
>>>>>>>> -XX:+PrintGCDateStamps -XX:+UseConcMarkSweepGC
>>>>>>>> -XX:CMSInitiatingOccupancyFraction=70 -XX:MaxHeapFreeRatio=70 -XX:+CM
>>>>>>>> SClassUnloadingEnabled -XX:OnOutOfMemoryError='kill -9 %p'
>>>>>>>> spark.executor.memory            8640m
>>>>>>>> spark.executor.cores             7
>>>>>>>> spark.authenticate.enableSaslEncryption true
>>>>>>>> spark.driver.memory              1g
>>>>>>>> spark.network.sasl.serverAlwaysEncrypt true
>>>>>>>> spark.driver.cores               1
>>>>>>>> spark.ssl.protocol               TLSv1.2
>>>>>>>> spark.ssl.keyStorePassword       password
>>>>>>>> spark.yarn.maxAppAttempts        1
>>>>>>>> spark.ssl.keyStore               /etc/emr/security/keystore.jks
>>>>>>>> spark.authenticate               true
>>>>>>>> spark.ssl.keyPassword            password
>>>>>>>> spark.ssl.enabled                true
>>>>>>>> spark.ssl.enabledAlgorithms      TLS_RSA_WITH_AES_256_CBC_SHA
>>>>>>>> spark.ssl.trustStore             /etc/emr/security/truststore.jks
>>>>>>>> spark.authenticate.secret        secret
>>>>>>>> spark.ssl.trustStorePassword     password
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Jun 27, 2016 at 7:33 PM, Hyung Sung Shim <hss...@nflabs.com
>>>>>>>> > wrote:
>>>>>>>>
>>>>>>>>> Hi.
>>>>>>>>> Could you share your conf/zeppelin-env.sh and spark-defaults.conf ?
>>>>>>>>>
>>>>>>>>> 2016-06-28 8:52 GMT+09:00 Jonathan Esterhazy <
>>>>>>>>> jonathan.esterh...@gmail.com>:
>>>>>>>>>
>>>>>>>>>> I am having trouble using zeppelin in a spark cluster that has
>>>>>>>>>> spark node authentication turned on (e.g. with 
>>>>>>>>>> spark.authenticate=true,
>>>>>>>>>> spark.authenticate.secret=...)
>>>>>>>>>>
>>>>>>>>>> Notebook code that calls built-in spark functions (or other
>>>>>>>>>> things on executor classpath) work fine, but functions defined in the
>>>>>>>>>> notebook (anonymous or named) throw ClassNotFoundExceptions when 
>>>>>>>>>> called
>>>>>>>>>> from an executor.
>>>>>>>>>>
>>>>>>>>>> For example, this code works:
>>>>>>>>>>
>>>>>>>>>> val rdd = sc.textFile("hdfs://my-text-file")
>>>>>>>>>> rdd.take(1).foreach(println)
>>>>>>>>>>
>>>>>>>>>> rdd.saveAsTextFile("hdfs:///my-other-text-file")
>>>>>>>>>>
>>>>>>>>>> but code like this...
>>>>>>>>>>
>>>>>>>>>> rdd.filter(_.contains("my data"))
>>>>>>>>>>
>>>>>>>>>> fails with
>>>>>>>>>>
>>>>>>>>>> Caused by: java.lang.ClassNotFoundException:
>>>>>>>>>> $iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1
>>>>>>>>>> at
>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:84)
>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>         ...
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> I get the same kind of error if the filter function is defined as
>>>>>>>>>> a named function in the notebook, or as a member of singleton object
>>>>>>>>>> defined in the notebook.
>>>>>>>>>>
>>>>>>>>>> When I look at the executor's log output, I see this error:
>>>>>>>>>>
>>>>>>>>>> 16/06/27 21:36:23 ERROR repl.ExecutorClassLoader: Failed to check
>>>>>>>>>> existence of class
>>>>>>>>>> $line31.$read$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$iwC$$anonfun$1 on 
>>>>>>>>>> REPL
>>>>>>>>>> class server at https://172.30.54.30:34980
>>>>>>>>>> java.lang.NullPointerException
>>>>>>>>>> at
>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.getClassFileInputStreamFromHttpServer(ExecutorClassLoader.scala:113)
>>>>>>>>>> at
>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClassLocally(ExecutorClassLoader.scala:146)
>>>>>>>>>> at
>>>>>>>>>> org.apache.spark.repl.ExecutorClassLoader.findClass(ExecutorClassLoader.scala:76)
>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>>>>>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>>>>>>>>>> at java.lang.Class.forName0(Native Method)
>>>>>>>>>> at java.lang.Class.forName(Class.java:348)
>>>>>>>>>>
>>>>>>>>>>         ...
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> If I disable spark authentication, everything works as expected.
>>>>>>>>>> I am running zeppelin 0.5.6 on spark 1.6.1 with yarn.
>>>>>>>>>>
>>>>>>>>>> Has anyone been able to get zeppelin working with spark
>>>>>>>>>> authentication?
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to