Jeff- Thanks!! I figured the issue, I didn't need to copy the hive-site.xml to spark_home/conf.
All I needed to do was set the SPARK_HOME environment variable in "zeppelin-env.sh". That has made the local mode to work as well. export SPARK_HOME=/opt/cloudera/parcels/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark In 0.5.6, even though I didn't set the SPARK_HOME in "zeppelin-env.sh", the embedded spark was able to bootstrap itself and talk to my hive metastore. Thanks & Regards, Pradeep On Wed, Aug 31, 2016 at 9:09 PM, Jeff Zhang <zjf...@gmail.com> wrote: > AFAIK, kerberos should not related here. ZEPPELIN-1175 just remove > ZEPPELIN_HOME/conf from the classpath of interpreter process. I guess you > put hive-site.xml under ZEPPELIN_HOME/conf, can you try to put it > under SPARK_HOME/conf ? > > On Thu, Sep 1, 2016 at 10:01 AM, Pradeep Reddy < > pradeepreddy.a...@gmail.com> wrote: > >> I just tried enabling kerberos on 0.6.1 and its now able to talk to my >> hive metastore, I see all my databases and tables. However the moment I >> takeout kerberos and run zeppelin 0.6.1 in local mode by resetting the >> "master" & removing spark variables, i see just one default database. >> >> I can live with this, as enabling kerberos integration is able to make >> zeppelin talk to my hive metastore. >> >> Thanks, >> Pradeep >> >> On Wed, Aug 31, 2016 at 8:36 PM, Pradeep Reddy < >> pradeepreddy.a...@gmail.com> wrote: >> >>> No luck, even after copying the hive-site.xml in >>> interpreter/spark/dep....I also tried, downloading 0.6.1 src vs 0.5.6. for >>> the same steps I followed, 0.5.6 is able to talk to my hive metastore, >>> where as the other recent builds are not, they just show one "default" >>> database. >>> >>> Thanks, >>> Pradeep >>> >>> On Wed, Aug 31, 2016 at 12:48 PM, Jongyoul Lee <jongy...@gmail.com> >>> wrote: >>> >>>> I think it's related to https://issues.apache.org/j >>>> ira/browse/ZEPPELIN-1175 which remove some class path when Zeppelin >>>> launches interpreter. Could you please check your hive-site.xml is included >>>> in your interpreter process? It looks like a configuration issue because >>>> you can see the default database. If it doesn't exists, you should copy >>>> your xml into interpreter/spark/dep/ >>>> >>>> Regards, >>>> JL >>>> >>>> On Wed, Aug 31, 2016 at 9:52 PM, Pradeep Reddy < >>>> pradeepreddy.a...@gmail.com> wrote: >>>> >>>>> Hi Jongyoul- I followed the exact same steps for compiling and setting >>>>> up the new build from source as 0.5.6 (only difference is, I acquired the >>>>> source for latest build using "git clone") >>>>> >>>>> hive-site.xml was copied to conf directory. But, the spark interpreter >>>>> is not talking to the hive metastore. Both the 0.5.6 & the latest builds >>>>> are running in the same machine. In 0.5.6 when i run the below command, I >>>>> see 116 databases listed, as per my expectations and I'm able to run my >>>>> notebooks built on those databases. >>>>> >>>>> [image: Inline image 1] >>>>> >>>>> Thanks, >>>>> Pradeep >>>>> >>>>> >>>>> On Wed, Aug 31, 2016 at 2:52 AM, Jongyoul Lee <jongy...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hello, >>>>>> >>>>>> Do you copy your hive-site.xml in a proper position? >>>>>> >>>>>> On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy < >>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>> >>>>>>> nothing obvious. I will stick to 0.5.6 build, until the latest >>>>>>> builds stabilize. >>>>>>> >>>>>>> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zjf...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Then I guess maybe you are connecting to different database. Why >>>>>>>> not using 'z.show(sql("databases"))' to display the databases ? >>>>>>>> Then you will get a hint what's going on. >>>>>>>> >>>>>>>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy < >>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Yes...I didn't wish to show the names of the databases that we >>>>>>>>> have in our data lake on that screen shot. so thats why I chose to >>>>>>>>> display >>>>>>>>> the count. The latest zeppelin build just shows 1 count which is >>>>>>>>> "default" >>>>>>>>> database. >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Pradeep >>>>>>>>> >>>>>>>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zjf...@gmail.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> 116 is the databases count number. Do you expect a list of >>>>>>>>>> database ? then you need to use 'z.show(sql("databases"))' >>>>>>>>>> >>>>>>>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy < >>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> Here it is Jeff >>>>>>>>>>> >>>>>>>>>>> [image: Inline image 1] >>>>>>>>>>> >>>>>>>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi Pradeep, >>>>>>>>>>>> >>>>>>>>>>>> I don't see the databases on your screenshot (second one for >>>>>>>>>>>> 0.5.6). I think the output is correct. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy < >>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin >>>>>>>>>>>>> build. It seems like Kerberos not working & spark not able to >>>>>>>>>>>>> talk to the >>>>>>>>>>>>> shared hive meta store are defects in the current build. >>>>>>>>>>>>> >>>>>>>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy < >>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi Jeff- >>>>>>>>>>>>>> >>>>>>>>>>>>>> I switched to local mode now, I'm able to summon the implicit >>>>>>>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my >>>>>>>>>>>>>> databases & >>>>>>>>>>>>>> tables, just shows 1 database "default". >>>>>>>>>>>>>> >>>>>>>>>>>>>> Zeppelin Latest Build >>>>>>>>>>>>>> >>>>>>>>>>>>>> [image: Inline image 3] >>>>>>>>>>>>>> >>>>>>>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show >>>>>>>>>>>>>> my databases and tables. >>>>>>>>>>>>>> >>>>>>>>>>>>>> [image: Inline image 4] >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com >>>>>>>>>>>>>> > wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> > the spark interpreter is not showing my tables & >>>>>>>>>>>>>>> databases, may be its running in an isolated mode... I'm just >>>>>>>>>>>>>>> getting empty >>>>>>>>>>>>>>> list, so I attempted to do kerberos authentication to >>>>>>>>>>>>>>> workaround that >>>>>>>>>>>>>>> issue, and bumped into this road block. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> kerberos would not help here, actually I think it would make >>>>>>>>>>>>>>> the problem more complicated. You need to first check the log >>>>>>>>>>>>>>> why you get >>>>>>>>>>>>>>> empty list. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy < >>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Jeff- I was also successfully able to run spark shell, >>>>>>>>>>>>>>>> after running kdestroy, with the below command and was able to >>>>>>>>>>>>>>>> get to my >>>>>>>>>>>>>>>> hive tables. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> spark-shell --conf >>>>>>>>>>>>>>>> spark.yarn.keytab=$HOME/pradeep.x.alla.keytab >>>>>>>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode >>>>>>>>>>>>>>>> client --master yarn --queue <QUEUE_NAME> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy < >>>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, >>>>>>>>>>>>>>>>> but when I migrated from 0.5.6 to this version, the spark >>>>>>>>>>>>>>>>> interpreter is >>>>>>>>>>>>>>>>> not showing my tables & databases, may be its running in an >>>>>>>>>>>>>>>>> isolated >>>>>>>>>>>>>>>>> mode... I'm just getting empty list, so I attempted to do >>>>>>>>>>>>>>>>> kerberos >>>>>>>>>>>>>>>>> authentication to workaround that issue, and bumped into this >>>>>>>>>>>>>>>>> road block. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Below is the configuration, I also tested my keytab file >>>>>>>>>>>>>>>>> and its working fine. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> *Kerberos test:* >>>>>>>>>>>>>>>>> $ kdestroy >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> $ klist >>>>>>>>>>>>>>>>> *klist: No credentials cache found (ticket cache >>>>>>>>>>>>>>>>> FILE:/tmp/krb5cc_12027)* >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla >>>>>>>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027* >>>>>>>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>* >>>>>>>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab* >>>>>>>>>>>>>>>>> *Authenticated to Kerberos v5* >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> $ klist >>>>>>>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027* >>>>>>>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>* >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> *Valid starting Expires Service principal* >>>>>>>>>>>>>>>>> *08/30/16 20:25:19 08/31/16 06:25:19 >>>>>>>>>>>>>>>>> krbtgt/<DOMAIN1>@<DOMAIN1>* >>>>>>>>>>>>>>>>> * renew until 08/31/16 20:25:19* >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> *zeppelin-env.sh* >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf >>>>>>>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce >>>>>>>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark >>>>>>>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client >>>>>>>>>>>>>>>>> --master yarn --num-executors 2 --executor-memory 2g --queue >>>>>>>>>>>>>>>>> <QUEUE_NAME>" >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> *Interpreter.json (Spark interpreter config)* >>>>>>>>>>>>>>>>> "2BUTFVN89": { >>>>>>>>>>>>>>>>> "id": "2BUTFVN89", >>>>>>>>>>>>>>>>> "name": "spark", >>>>>>>>>>>>>>>>> "group": "spark", >>>>>>>>>>>>>>>>> "properties": { >>>>>>>>>>>>>>>>> "spark.cores.max": "", >>>>>>>>>>>>>>>>> "zeppelin.spark.printREPLOutput": "true", >>>>>>>>>>>>>>>>> "master": "yarn-client", >>>>>>>>>>>>>>>>> "zeppelin.spark.maxResult": "1000", >>>>>>>>>>>>>>>>> "zeppelin.dep.localrepo": "local-repo", >>>>>>>>>>>>>>>>> "spark.app.name": "Zeppelin", >>>>>>>>>>>>>>>>> "spark.executor.memory": "", >>>>>>>>>>>>>>>>> "zeppelin.spark.importImplicit": "true", >>>>>>>>>>>>>>>>> "zeppelin.spark.sql.stacktrace": "true", >>>>>>>>>>>>>>>>> "zeppelin.spark.useHiveContext": "true", >>>>>>>>>>>>>>>>> "zeppelin.interpreter.localRepo": >>>>>>>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89", >>>>>>>>>>>>>>>>> "zeppelin.spark.concurrentSQL": "false", >>>>>>>>>>>>>>>>> "args": "", >>>>>>>>>>>>>>>>> "zeppelin.pyspark.python": "python", >>>>>>>>>>>>>>>>> "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep. >>>>>>>>>>>>>>>>> x.alla.keytab", >>>>>>>>>>>>>>>>> "spark.yarn.principal": "pradeep.x.alla", >>>>>>>>>>>>>>>>> "zeppelin.dep.additionalRemoteRepository": >>>>>>>>>>>>>>>>> "spark-packages,http://dl.bint >>>>>>>>>>>>>>>>> ray.com/spark-packages/maven,false;" >>>>>>>>>>>>>>>>> }, >>>>>>>>>>>>>>>>> "status": "READY", >>>>>>>>>>>>>>>>> "interpreterGroup": [ >>>>>>>>>>>>>>>>> { >>>>>>>>>>>>>>>>> "name": "spark", >>>>>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.Spa >>>>>>>>>>>>>>>>> rkInterpreter", >>>>>>>>>>>>>>>>> "defaultInterpreter": true >>>>>>>>>>>>>>>>> }, >>>>>>>>>>>>>>>>> { >>>>>>>>>>>>>>>>> "name": "sql", >>>>>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.Spa >>>>>>>>>>>>>>>>> rkSqlInterpreter", >>>>>>>>>>>>>>>>> "defaultInterpreter": false >>>>>>>>>>>>>>>>> }, >>>>>>>>>>>>>>>>> { >>>>>>>>>>>>>>>>> "name": "dep", >>>>>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.Dep >>>>>>>>>>>>>>>>> Interpreter", >>>>>>>>>>>>>>>>> "defaultInterpreter": false >>>>>>>>>>>>>>>>> }, >>>>>>>>>>>>>>>>> { >>>>>>>>>>>>>>>>> "name": "pyspark", >>>>>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.PyS >>>>>>>>>>>>>>>>> parkInterpreter", >>>>>>>>>>>>>>>>> "defaultInterpreter": false >>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>>> ], >>>>>>>>>>>>>>>>> "dependencies": [], >>>>>>>>>>>>>>>>> "option": { >>>>>>>>>>>>>>>>> "remote": true, >>>>>>>>>>>>>>>>> "port": -1, >>>>>>>>>>>>>>>>> "perNoteSession": false, >>>>>>>>>>>>>>>>> "perNoteProcess": false, >>>>>>>>>>>>>>>>> "isExistingProcess": false, >>>>>>>>>>>>>>>>> "setPermission": false, >>>>>>>>>>>>>>>>> "users": [] >>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang < >>>>>>>>>>>>>>>>> zjf...@gmail.com> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you >>>>>>>>>>>>>>>>>> mind to share your configuration ? Or you can first try to >>>>>>>>>>>>>>>>>> run spark-shell >>>>>>>>>>>>>>>>>> using spark.yarn.keytab & spark.yarn.principle to verify >>>>>>>>>>>>>>>>>> them. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy < >>>>>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured >>>>>>>>>>>>>>>>>>> kerberos authentication. For Kerberos I added >>>>>>>>>>>>>>>>>>> "spark.yarn.keytab" & >>>>>>>>>>>>>>>>>>> "spark.yarn.principal" and also set master to >>>>>>>>>>>>>>>>>>> "yarn-client". But I keep >>>>>>>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the >>>>>>>>>>>>>>>>>>> notebook >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> 3536728 started by scheduler >>>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091 >>>>>>>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2} >>>>>>>>>>>>>>>>>>> Job.java[run]:189) - Job failed >>>>>>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L >>>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>. >>>>>>>>>>>>>>>>>>> COM\E$)s/@\Q<DOMAIN1>\E$//L >>>>>>>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>>>>>>>>>>>>> DEFAULT >>>>>>>>>>>>>>>>>>> at org.apache.hadoop.security.aut >>>>>>>>>>>>>>>>>>> hentication.util.KerberosName. >>>>>>>>>>>>>>>>>>> parseRules(KerberosName.java:321) >>>>>>>>>>>>>>>>>>> at org.apache.hadoop.security.aut >>>>>>>>>>>>>>>>>>> hentication.util.KerberosName. >>>>>>>>>>>>>>>>>>> setRules(KerberosName.java:386) >>>>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Had >>>>>>>>>>>>>>>>>>> oopKerberosName.setConfigurati >>>>>>>>>>>>>>>>>>> on(HadoopKerberosName.java:75) >>>>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>>>> rGroupInformation.initialize(U >>>>>>>>>>>>>>>>>>> serGroupInformation.java:227) >>>>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>>>> rGroupInformation.ensureInitia >>>>>>>>>>>>>>>>>>> lized(UserGroupInformation.java:214) >>>>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>>>> rGroupInformation.isAuthentica >>>>>>>>>>>>>>>>>>> tionMethodEnabled(UserGroupInformation.java:275) >>>>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>>>> rGroupInformation.isSecurityEn >>>>>>>>>>>>>>>>>>> abled(UserGroupInformation.java:269) >>>>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>>>> rGroupInformation.loginUserFro >>>>>>>>>>>>>>>>>>> mKeytab(UserGroupInformation.java:820) >>>>>>>>>>>>>>>>>>> at org.apache.zeppelin.spark.Spar >>>>>>>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539) >>>>>>>>>>>>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) >>>>>>>>>>>>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.interpre >>>>>>>>>>>>>>>>>>> t(LazyOpenInterpreter.java:93) >>>>>>>>>>>>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>>>>>>>>>>>> r.remote.RemoteInterpreterServ >>>>>>>>>>>>>>>>>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383) >>>>>>>>>>>>>>>>>>> at org.apache.zeppelin.scheduler. >>>>>>>>>>>>>>>>>>> Job.run(Job.java:176) >>>>>>>>>>>>>>>>>>> at org.apache.zeppelin.scheduler. >>>>>>>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139) >>>>>>>>>>>>>>>>>>> at java.util.concurrent.Executors >>>>>>>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511) >>>>>>>>>>>>>>>>>>> at java.util.concurrent.FutureTas >>>>>>>>>>>>>>>>>>> k.run(FutureTask.java:266) >>>>>>>>>>>>>>>>>>> at java.util.concurrent.Scheduled >>>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu >>>>>>>>>>>>>>>>>>> tureTask.access$201(ScheduledT >>>>>>>>>>>>>>>>>>> hreadPoolExecutor.java:180) >>>>>>>>>>>>>>>>>>> at java.util.concurrent.Scheduled >>>>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu >>>>>>>>>>>>>>>>>>> tureTask.run(ScheduledThreadPoolExecutor.java:293) >>>>>>>>>>>>>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142) >>>>>>>>>>>>>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617) >>>>>>>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>>>>>>>>>>>>>>> INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2} >>>>>>>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job >>>>>>>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler >>>>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>>> Pradeep >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> Best Regards >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Jeff Zhang >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Best Regards >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Jeff Zhang >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Best Regards >>>>>>>>>>>> >>>>>>>>>>>> Jeff Zhang >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Best Regards >>>>>>>>>> >>>>>>>>>> Jeff Zhang >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Best Regards >>>>>>>> >>>>>>>> Jeff Zhang >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> 이종열, Jongyoul Lee, 李宗烈 >>>>>> http://madeng.net >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> 이종열, Jongyoul Lee, 李宗烈 >>>> http://madeng.net >>>> >>> >>> >> > > > -- > Best Regards > > Jeff Zhang >