AFAIK, kerberos should not related here. ZEPPELIN-1175 just remove
ZEPPELIN_HOME/conf from the classpath of interpreter process. I guess you
put hive-site.xml under ZEPPELIN_HOME/conf, can you try to put it
under SPARK_HOME/conf ?

On Thu, Sep 1, 2016 at 10:01 AM, Pradeep Reddy <pradeepreddy.a...@gmail.com>
wrote:

> I just tried enabling kerberos on 0.6.1 and its now able to talk to my
> hive metastore, I see all my databases and tables. However the moment I
> takeout kerberos and run zeppelin 0.6.1 in local mode by resetting the
> "master" & removing spark variables, i see just one default database.
>
> I can live with this, as enabling kerberos integration is able to make
> zeppelin talk to my hive metastore.
>
> Thanks,
> Pradeep
>
> On Wed, Aug 31, 2016 at 8:36 PM, Pradeep Reddy <
> pradeepreddy.a...@gmail.com> wrote:
>
>> No luck, even after copying the hive-site.xml in
>> interpreter/spark/dep....I also tried, downloading 0.6.1 src  vs 0.5.6. for
>> the same steps I followed, 0.5.6 is able to talk to my hive metastore,
>> where as the other recent builds are not, they just show one "default"
>> database.
>>
>> Thanks,
>> Pradeep
>>
>> On Wed, Aug 31, 2016 at 12:48 PM, Jongyoul Lee <jongy...@gmail.com>
>> wrote:
>>
>>> I think it's related to https://issues.apache.org/j
>>> ira/browse/ZEPPELIN-1175 which remove some class path when Zeppelin
>>> launches interpreter. Could you please check your hive-site.xml is included
>>> in your interpreter process? It looks like a configuration issue because
>>> you can see the default database. If it doesn't exists, you should copy
>>> your xml into interpreter/spark/dep/
>>>
>>> Regards,
>>> JL
>>>
>>> On Wed, Aug 31, 2016 at 9:52 PM, Pradeep Reddy <
>>> pradeepreddy.a...@gmail.com> wrote:
>>>
>>>> Hi Jongyoul- I followed the exact same steps for compiling and setting
>>>> up the new build from source as 0.5.6 (only difference is, I acquired the
>>>> source for latest build using "git clone")
>>>>
>>>> hive-site.xml was copied to conf directory. But, the spark interpreter
>>>> is not talking to the hive metastore. Both the 0.5.6 & the latest builds
>>>> are running in the same machine. In 0.5.6 when i run the below command, I
>>>> see 116 databases listed, as per my expectations and I'm able to run my
>>>> notebooks built on those databases.
>>>>
>>>> [image: Inline image 1]
>>>>
>>>> Thanks,
>>>> Pradeep
>>>>
>>>>
>>>> On Wed, Aug 31, 2016 at 2:52 AM, Jongyoul Lee <jongy...@gmail.com>
>>>> wrote:
>>>>
>>>>> Hello,
>>>>>
>>>>> Do you copy your hive-site.xml in a proper position?
>>>>>
>>>>> On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy <
>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>
>>>>>> nothing obvious. I will stick to 0.5.6 build, until the latest builds
>>>>>> stabilize.
>>>>>>
>>>>>> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>>>>
>>>>>>> Then I guess maybe you are connecting to different database. Why not
>>>>>>> using  'z.show(sql("databases"))' to display the databases ? Then
>>>>>>> you will get a hint what's going on.
>>>>>>>
>>>>>>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy <
>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Yes...I didn't wish to show the names of the databases that we have
>>>>>>>> in our data lake on that screen shot. so thats why I chose to display 
>>>>>>>> the
>>>>>>>> count. The latest zeppelin build just shows 1 count which is "default"
>>>>>>>> database.
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Pradeep
>>>>>>>>
>>>>>>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zjf...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> 116 is the databases count number. Do you expect a list of
>>>>>>>>> database ? then you need to use 'z.show(sql("databases"))'
>>>>>>>>>
>>>>>>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <
>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Here it is Jeff
>>>>>>>>>>
>>>>>>>>>> [image: Inline image 1]
>>>>>>>>>>
>>>>>>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi Pradeep,
>>>>>>>>>>>
>>>>>>>>>>> I don't see the databases on your screenshot (second one for
>>>>>>>>>>> 0.5.6). I think the output is correct.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy <
>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin
>>>>>>>>>>>> build. It seems like Kerberos not working & spark not able to talk 
>>>>>>>>>>>> to the
>>>>>>>>>>>> shared hive meta store are defects in the current build.
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy <
>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi Jeff-
>>>>>>>>>>>>>
>>>>>>>>>>>>> I switched to local mode now, I'm able to summon the implicit
>>>>>>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my 
>>>>>>>>>>>>> databases &
>>>>>>>>>>>>> tables, just shows 1 database "default".
>>>>>>>>>>>>>
>>>>>>>>>>>>> Zeppelin Latest Build
>>>>>>>>>>>>>
>>>>>>>>>>>>> [image: Inline image 3]
>>>>>>>>>>>>>
>>>>>>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show
>>>>>>>>>>>>> my databases and tables.
>>>>>>>>>>>>>
>>>>>>>>>>>>> [image: Inline image 4]
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com>
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> > the spark interpreter is not showing my tables &
>>>>>>>>>>>>>> databases, may be its running in an isolated mode... I'm just 
>>>>>>>>>>>>>> getting empty
>>>>>>>>>>>>>> list, so I attempted to do kerberos authentication to workaround 
>>>>>>>>>>>>>> that
>>>>>>>>>>>>>> issue, and bumped into this road block.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> kerberos would not help here, actually I think it would make
>>>>>>>>>>>>>> the problem more complicated.  You need to first check the log 
>>>>>>>>>>>>>> why you get
>>>>>>>>>>>>>> empty list.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <
>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Jeff- I was also successfully able to run spark shell, after
>>>>>>>>>>>>>>> running kdestroy, with the below command and was able to get to 
>>>>>>>>>>>>>>> my hive
>>>>>>>>>>>>>>> tables.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab
>>>>>>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode
>>>>>>>>>>>>>>> client --master yarn --queue <QUEUE_NAME>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <
>>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but
>>>>>>>>>>>>>>>> when I migrated from 0.5.6 to this version, the spark 
>>>>>>>>>>>>>>>> interpreter is not
>>>>>>>>>>>>>>>> showing my tables & databases, may be its running in an 
>>>>>>>>>>>>>>>> isolated mode...
>>>>>>>>>>>>>>>> I'm just getting empty list, so I attempted to do kerberos 
>>>>>>>>>>>>>>>> authentication
>>>>>>>>>>>>>>>> to workaround that issue, and bumped into this road block.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Below is the configuration, I also tested my keytab file
>>>>>>>>>>>>>>>> and its working fine.
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> *Kerberos test:*
>>>>>>>>>>>>>>>> $ kdestroy
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>>> *klist: No credentials cache found (ticket cache
>>>>>>>>>>>>>>>> FILE:/tmp/krb5cc_12027)*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
>>>>>>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027*
>>>>>>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab*
>>>>>>>>>>>>>>>> *Authenticated to Kerberos v5*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> $ klist
>>>>>>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027*
>>>>>>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> *Valid starting     Expires            Service principal*
>>>>>>>>>>>>>>>> *08/30/16 20:25:19  08/31/16 06:25:19
>>>>>>>>>>>>>>>>  krbtgt/<DOMAIN1>@<DOMAIN1>*
>>>>>>>>>>>>>>>> *        renew until 08/31/16 20:25:19*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> *zeppelin-env.sh*
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
>>>>>>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce
>>>>>>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark
>>>>>>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master
>>>>>>>>>>>>>>>> yarn --num-executors 2 --executor-memory 2g --queue 
>>>>>>>>>>>>>>>> <QUEUE_NAME>"
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> *Interpreter.json (Spark interpreter config)*
>>>>>>>>>>>>>>>> "2BUTFVN89": {
>>>>>>>>>>>>>>>>       "id": "2BUTFVN89",
>>>>>>>>>>>>>>>>       "name": "spark",
>>>>>>>>>>>>>>>>       "group": "spark",
>>>>>>>>>>>>>>>>       "properties": {
>>>>>>>>>>>>>>>>         "spark.cores.max": "",
>>>>>>>>>>>>>>>>         "zeppelin.spark.printREPLOutput": "true",
>>>>>>>>>>>>>>>>         "master": "yarn-client",
>>>>>>>>>>>>>>>>         "zeppelin.spark.maxResult": "1000",
>>>>>>>>>>>>>>>>         "zeppelin.dep.localrepo": "local-repo",
>>>>>>>>>>>>>>>>         "spark.app.name": "Zeppelin",
>>>>>>>>>>>>>>>>         "spark.executor.memory": "",
>>>>>>>>>>>>>>>>         "zeppelin.spark.importImplicit": "true",
>>>>>>>>>>>>>>>>         "zeppelin.spark.sql.stacktrace": "true",
>>>>>>>>>>>>>>>>         "zeppelin.spark.useHiveContext": "true",
>>>>>>>>>>>>>>>>         "zeppelin.interpreter.localRepo":
>>>>>>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
>>>>>>>>>>>>>>>>         "zeppelin.spark.concurrentSQL": "false",
>>>>>>>>>>>>>>>>         "args": "",
>>>>>>>>>>>>>>>>         "zeppelin.pyspark.python": "python",
>>>>>>>>>>>>>>>>         "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.
>>>>>>>>>>>>>>>> x.alla.keytab",
>>>>>>>>>>>>>>>>         "spark.yarn.principal": "pradeep.x.alla",
>>>>>>>>>>>>>>>>         "zeppelin.dep.additionalRemoteRepository":
>>>>>>>>>>>>>>>> "spark-packages,http://dl.bint
>>>>>>>>>>>>>>>> ray.com/spark-packages/maven,false;"
>>>>>>>>>>>>>>>>       },
>>>>>>>>>>>>>>>>       "status": "READY",
>>>>>>>>>>>>>>>>       "interpreterGroup": [
>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>           "name": "spark",
>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>>> rkInterpreter",
>>>>>>>>>>>>>>>>           "defaultInterpreter": true
>>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>           "name": "sql",
>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>>>>>>> rkSqlInterpreter",
>>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>           "name": "dep",
>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Dep
>>>>>>>>>>>>>>>> Interpreter",
>>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>>         },
>>>>>>>>>>>>>>>>         {
>>>>>>>>>>>>>>>>           "name": "pyspark",
>>>>>>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.PyS
>>>>>>>>>>>>>>>> parkInterpreter",
>>>>>>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>>>>>>         }
>>>>>>>>>>>>>>>>       ],
>>>>>>>>>>>>>>>>       "dependencies": [],
>>>>>>>>>>>>>>>>       "option": {
>>>>>>>>>>>>>>>>         "remote": true,
>>>>>>>>>>>>>>>>         "port": -1,
>>>>>>>>>>>>>>>>         "perNoteSession": false,
>>>>>>>>>>>>>>>>         "perNoteProcess": false,
>>>>>>>>>>>>>>>>         "isExistingProcess": false,
>>>>>>>>>>>>>>>>         "setPermission": false,
>>>>>>>>>>>>>>>>         "users": []
>>>>>>>>>>>>>>>>       }
>>>>>>>>>>>>>>>>     }
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <
>>>>>>>>>>>>>>>> zjf...@gmail.com> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you
>>>>>>>>>>>>>>>>> mind to share your configuration ? Or you can first try to 
>>>>>>>>>>>>>>>>> run spark-shell
>>>>>>>>>>>>>>>>> using spark.yarn.keytab & spark.yarn.principle to verify them.
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
>>>>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured
>>>>>>>>>>>>>>>>>> kerberos authentication. For Kerberos I added 
>>>>>>>>>>>>>>>>>> "spark.yarn.keytab" &
>>>>>>>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client". 
>>>>>>>>>>>>>>>>>>  But I keep
>>>>>>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the 
>>>>>>>>>>>>>>>>>> notebook
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> 3536728 started by scheduler
>>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>>> Job.java[run]:189) - Job failed
>>>>>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L
>>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L
>>>>>>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>>>>>>> DEFAULT
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>>> parseRules(KerberosName.java:321)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>>>>>>> hentication.util.KerberosName.
>>>>>>>>>>>>>>>>>> setRules(KerberosName.java:386)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Had
>>>>>>>>>>>>>>>>>> oopKerberosName.setConfigurati
>>>>>>>>>>>>>>>>>> on(HadoopKerberosName.java:75)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.initialize(U
>>>>>>>>>>>>>>>>>> serGroupInformation.java:227)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.ensureInitia
>>>>>>>>>>>>>>>>>> lized(UserGroupInformation.java:214)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.isAuthentica
>>>>>>>>>>>>>>>>>> tionMethodEnabled(UserGroupInformation.java:275)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.isSecurityEn
>>>>>>>>>>>>>>>>>> abled(UserGroupInformation.java:269)
>>>>>>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>>>>>>> rGroupInformation.loginUserFro
>>>>>>>>>>>>>>>>>> mKeytab(UserGroupInformation.java:820)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.interpre
>>>>>>>>>>>>>>>>>> t(LazyOpenInterpreter.java:93)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>>>>>>> r.remote.RemoteInterpreterServ
>>>>>>>>>>>>>>>>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>>> Job.run(Job.java:176)
>>>>>>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.Executors
>>>>>>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>>>>>>>>>>>>> k.run(FutureTask.java:266)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu
>>>>>>>>>>>>>>>>>> tureTask.run(ScheduledThreadPoolExecutor.java:293)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>>>>>>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>>>>>>>>>>>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>>>>>>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job
>>>>>>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler
>>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>> Pradeep
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Best Regards
>>>>>>>>>>>
>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Best Regards
>>>>>>>>>
>>>>>>>>> Jeff Zhang
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best Regards
>>>>>>>
>>>>>>> Jeff Zhang
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> 이종열, Jongyoul Lee, 李宗烈
>>>>> http://madeng.net
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> 이종열, Jongyoul Lee, 李宗烈
>>> http://madeng.net
>>>
>>
>>
>


-- 
Best Regards

Jeff Zhang

Reply via email to