No luck, even after copying the hive-site.xml in interpreter/spark/dep....I also tried, downloading 0.6.1 src vs 0.5.6. for the same steps I followed, 0.5.6 is able to talk to my hive metastore, where as the other recent builds are not, they just show one "default" database.
Thanks, Pradeep On Wed, Aug 31, 2016 at 12:48 PM, Jongyoul Lee <jongy...@gmail.com> wrote: > I think it's related to https://issues.apache.org/ > jira/browse/ZEPPELIN-1175 which remove some class path when Zeppelin > launches interpreter. Could you please check your hive-site.xml is included > in your interpreter process? It looks like a configuration issue because > you can see the default database. If it doesn't exists, you should copy > your xml into interpreter/spark/dep/ > > Regards, > JL > > On Wed, Aug 31, 2016 at 9:52 PM, Pradeep Reddy < > pradeepreddy.a...@gmail.com> wrote: > >> Hi Jongyoul- I followed the exact same steps for compiling and setting up >> the new build from source as 0.5.6 (only difference is, I acquired the >> source for latest build using "git clone") >> >> hive-site.xml was copied to conf directory. But, the spark interpreter is >> not talking to the hive metastore. Both the 0.5.6 & the latest builds are >> running in the same machine. In 0.5.6 when i run the below command, I see >> 116 databases listed, as per my expectations and I'm able to run my >> notebooks built on those databases. >> >> [image: Inline image 1] >> >> Thanks, >> Pradeep >> >> >> On Wed, Aug 31, 2016 at 2:52 AM, Jongyoul Lee <jongy...@gmail.com> wrote: >> >>> Hello, >>> >>> Do you copy your hive-site.xml in a proper position? >>> >>> On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy < >>> pradeepreddy.a...@gmail.com> wrote: >>> >>>> nothing obvious. I will stick to 0.5.6 build, until the latest builds >>>> stabilize. >>>> >>>> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zjf...@gmail.com> wrote: >>>> >>>>> Then I guess maybe you are connecting to different database. Why not >>>>> using 'z.show(sql("databases"))' to display the databases ? Then you >>>>> will get a hint what's going on. >>>>> >>>>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy < >>>>> pradeepreddy.a...@gmail.com> wrote: >>>>> >>>>>> Yes...I didn't wish to show the names of the databases that we have >>>>>> in our data lake on that screen shot. so thats why I chose to display the >>>>>> count. The latest zeppelin build just shows 1 count which is "default" >>>>>> database. >>>>>> >>>>>> Thanks, >>>>>> Pradeep >>>>>> >>>>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zjf...@gmail.com> wrote: >>>>>> >>>>>>> 116 is the databases count number. Do you expect a list of database >>>>>>> ? then you need to use 'z.show(sql("databases"))' >>>>>>> >>>>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy < >>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>> >>>>>>>> Here it is Jeff >>>>>>>> >>>>>>>> [image: Inline image 1] >>>>>>>> >>>>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi Pradeep, >>>>>>>>> >>>>>>>>> I don't see the databases on your screenshot (second one for >>>>>>>>> 0.5.6). I think the output is correct. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy < >>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin >>>>>>>>>> build. It seems like Kerberos not working & spark not able to talk >>>>>>>>>> to the >>>>>>>>>> shared hive meta store are defects in the current build. >>>>>>>>>> >>>>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy < >>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Jeff- >>>>>>>>>>> >>>>>>>>>>> I switched to local mode now, I'm able to summon the implicit >>>>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my databases & >>>>>>>>>>> tables, just shows 1 database "default". >>>>>>>>>>> >>>>>>>>>>> Zeppelin Latest Build >>>>>>>>>>> >>>>>>>>>>> [image: Inline image 3] >>>>>>>>>>> >>>>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show my >>>>>>>>>>> databases and tables. >>>>>>>>>>> >>>>>>>>>>> [image: Inline image 4] >>>>>>>>>>> >>>>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> > the spark interpreter is not showing my tables & databases, >>>>>>>>>>>> may be its running in an isolated mode... I'm just getting empty >>>>>>>>>>>> list, so I >>>>>>>>>>>> attempted to do kerberos authentication to workaround that issue, >>>>>>>>>>>> and >>>>>>>>>>>> bumped into this road block. >>>>>>>>>>>> >>>>>>>>>>>> kerberos would not help here, actually I think it would make >>>>>>>>>>>> the problem more complicated. You need to first check the log why >>>>>>>>>>>> you get >>>>>>>>>>>> empty list. >>>>>>>>>>>> >>>>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy < >>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Jeff- I was also successfully able to run spark shell, after >>>>>>>>>>>>> running kdestroy, with the below command and was able to get to >>>>>>>>>>>>> my hive >>>>>>>>>>>>> tables. >>>>>>>>>>>>> >>>>>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab >>>>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode >>>>>>>>>>>>> client --master yarn --queue <QUEUE_NAME> >>>>>>>>>>>>> >>>>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy < >>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but >>>>>>>>>>>>>> when I migrated from 0.5.6 to this version, the spark >>>>>>>>>>>>>> interpreter is not >>>>>>>>>>>>>> showing my tables & databases, may be its running in an isolated >>>>>>>>>>>>>> mode... >>>>>>>>>>>>>> I'm just getting empty list, so I attempted to do kerberos >>>>>>>>>>>>>> authentication >>>>>>>>>>>>>> to workaround that issue, and bumped into this road block. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Below is the configuration, I also tested my keytab file and >>>>>>>>>>>>>> its working fine. >>>>>>>>>>>>>> >>>>>>>>>>>>>> *Kerberos test:* >>>>>>>>>>>>>> $ kdestroy >>>>>>>>>>>>>> >>>>>>>>>>>>>> $ klist >>>>>>>>>>>>>> *klist: No credentials cache found (ticket cache >>>>>>>>>>>>>> FILE:/tmp/krb5cc_12027)* >>>>>>>>>>>>>> >>>>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla >>>>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027* >>>>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>* >>>>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab* >>>>>>>>>>>>>> *Authenticated to Kerberos v5* >>>>>>>>>>>>>> >>>>>>>>>>>>>> $ klist >>>>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027* >>>>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>* >>>>>>>>>>>>>> >>>>>>>>>>>>>> *Valid starting Expires Service principal* >>>>>>>>>>>>>> *08/30/16 20:25:19 08/31/16 06:25:19 >>>>>>>>>>>>>> krbtgt/<DOMAIN1>@<DOMAIN1>* >>>>>>>>>>>>>> * renew until 08/31/16 20:25:19* >>>>>>>>>>>>>> >>>>>>>>>>>>>> *zeppelin-env.sh* >>>>>>>>>>>>>> >>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf >>>>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce >>>>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark >>>>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master >>>>>>>>>>>>>> yarn --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>" >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> *Interpreter.json (Spark interpreter config)* >>>>>>>>>>>>>> "2BUTFVN89": { >>>>>>>>>>>>>> "id": "2BUTFVN89", >>>>>>>>>>>>>> "name": "spark", >>>>>>>>>>>>>> "group": "spark", >>>>>>>>>>>>>> "properties": { >>>>>>>>>>>>>> "spark.cores.max": "", >>>>>>>>>>>>>> "zeppelin.spark.printREPLOutput": "true", >>>>>>>>>>>>>> "master": "yarn-client", >>>>>>>>>>>>>> "zeppelin.spark.maxResult": "1000", >>>>>>>>>>>>>> "zeppelin.dep.localrepo": "local-repo", >>>>>>>>>>>>>> "spark.app.name": "Zeppelin", >>>>>>>>>>>>>> "spark.executor.memory": "", >>>>>>>>>>>>>> "zeppelin.spark.importImplicit": "true", >>>>>>>>>>>>>> "zeppelin.spark.sql.stacktrace": "true", >>>>>>>>>>>>>> "zeppelin.spark.useHiveContext": "true", >>>>>>>>>>>>>> "zeppelin.interpreter.localRepo": >>>>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89", >>>>>>>>>>>>>> "zeppelin.spark.concurrentSQL": "false", >>>>>>>>>>>>>> "args": "", >>>>>>>>>>>>>> "zeppelin.pyspark.python": "python", >>>>>>>>>>>>>> "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep. >>>>>>>>>>>>>> x.alla.keytab", >>>>>>>>>>>>>> "spark.yarn.principal": "pradeep.x.alla", >>>>>>>>>>>>>> "zeppelin.dep.additionalRemoteRepository": >>>>>>>>>>>>>> "spark-packages,http://dl.bintray.com/spark-packages/maven,f >>>>>>>>>>>>>> alse;" >>>>>>>>>>>>>> }, >>>>>>>>>>>>>> "status": "READY", >>>>>>>>>>>>>> "interpreterGroup": [ >>>>>>>>>>>>>> { >>>>>>>>>>>>>> "name": "spark", >>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.Spa >>>>>>>>>>>>>> rkInterpreter", >>>>>>>>>>>>>> "defaultInterpreter": true >>>>>>>>>>>>>> }, >>>>>>>>>>>>>> { >>>>>>>>>>>>>> "name": "sql", >>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.Spa >>>>>>>>>>>>>> rkSqlInterpreter", >>>>>>>>>>>>>> "defaultInterpreter": false >>>>>>>>>>>>>> }, >>>>>>>>>>>>>> { >>>>>>>>>>>>>> "name": "dep", >>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.Dep >>>>>>>>>>>>>> Interpreter", >>>>>>>>>>>>>> "defaultInterpreter": false >>>>>>>>>>>>>> }, >>>>>>>>>>>>>> { >>>>>>>>>>>>>> "name": "pyspark", >>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.PyS >>>>>>>>>>>>>> parkInterpreter", >>>>>>>>>>>>>> "defaultInterpreter": false >>>>>>>>>>>>>> } >>>>>>>>>>>>>> ], >>>>>>>>>>>>>> "dependencies": [], >>>>>>>>>>>>>> "option": { >>>>>>>>>>>>>> "remote": true, >>>>>>>>>>>>>> "port": -1, >>>>>>>>>>>>>> "perNoteSession": false, >>>>>>>>>>>>>> "perNoteProcess": false, >>>>>>>>>>>>>> "isExistingProcess": false, >>>>>>>>>>>>>> "setPermission": false, >>>>>>>>>>>>>> "users": [] >>>>>>>>>>>>>> } >>>>>>>>>>>>>> } >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com >>>>>>>>>>>>>> > wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you mind >>>>>>>>>>>>>>> to share your configuration ? Or you can first try to run >>>>>>>>>>>>>>> spark-shell using >>>>>>>>>>>>>>> spark.yarn.keytab & spark.yarn.principle to verify them. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy < >>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured >>>>>>>>>>>>>>>> kerberos authentication. For Kerberos I added >>>>>>>>>>>>>>>> "spark.yarn.keytab" & >>>>>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client". >>>>>>>>>>>>>>>> But I keep >>>>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the >>>>>>>>>>>>>>>> notebook >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar >>>>>>>>>>>>>>>> kInterpreter335845091 >>>>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2} >>>>>>>>>>>>>>>> Job.java[run]:189) - Job failed >>>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L >>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L >>>>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>>>>>>>>>> DEFAULT >>>>>>>>>>>>>>>> at org.apache.hadoop.security.aut >>>>>>>>>>>>>>>> hentication.util.KerberosName. >>>>>>>>>>>>>>>> parseRules(KerberosName.java:321) >>>>>>>>>>>>>>>> at org.apache.hadoop.security.aut >>>>>>>>>>>>>>>> hentication.util.KerberosName. >>>>>>>>>>>>>>>> setRules(KerberosName.java:386) >>>>>>>>>>>>>>>> at org.apache.hadoop.security.Had >>>>>>>>>>>>>>>> oopKerberosName.setConfigurati >>>>>>>>>>>>>>>> on(HadoopKerberosName.java:75) >>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>> rGroupInformation.initialize(UserGroupInformation.java:227) >>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>> rGroupInformation.ensureInitia >>>>>>>>>>>>>>>> lized(UserGroupInformation.java:214) >>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>> rGroupInformation.isAuthentica >>>>>>>>>>>>>>>> tionMethodEnabled(UserGroupInformation.java:275) >>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>> rGroupInformation.isSecurityEn >>>>>>>>>>>>>>>> abled(UserGroupInformation.java:269) >>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>> rGroupInformation.loginUserFro >>>>>>>>>>>>>>>> mKeytab(UserGroupInformation.java:820) >>>>>>>>>>>>>>>> at org.apache.zeppelin.spark.Spar >>>>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539) >>>>>>>>>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) >>>>>>>>>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>>>>>>>>> r.LazyOpenInterpreter.interpre >>>>>>>>>>>>>>>> t(LazyOpenInterpreter.java:93) >>>>>>>>>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>>>>>>>>> r.remote.RemoteInterpreterServ >>>>>>>>>>>>>>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383) >>>>>>>>>>>>>>>> at org.apache.zeppelin.scheduler. >>>>>>>>>>>>>>>> Job.run(Job.java:176) >>>>>>>>>>>>>>>> at org.apache.zeppelin.scheduler. >>>>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139) >>>>>>>>>>>>>>>> at java.util.concurrent.Executors >>>>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511) >>>>>>>>>>>>>>>> at java.util.concurrent.FutureTas >>>>>>>>>>>>>>>> k.run(FutureTask.java:266) >>>>>>>>>>>>>>>> at java.util.concurrent.Scheduled >>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu >>>>>>>>>>>>>>>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180) >>>>>>>>>>>>>>>> at java.util.concurrent.Scheduled >>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu >>>>>>>>>>>>>>>> tureTask.run(ScheduledThreadPoolExecutor.java:293) >>>>>>>>>>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142) >>>>>>>>>>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617) >>>>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>>>>>>>>>>>> INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2} >>>>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job >>>>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler >>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091 >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>> Pradeep >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Best Regards >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Jeff Zhang >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Best Regards >>>>>>>>>>>> >>>>>>>>>>>> Jeff Zhang >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Best Regards >>>>>>>>> >>>>>>>>> Jeff Zhang >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Best Regards >>>>>>> >>>>>>> Jeff Zhang >>>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> Best Regards >>>>> >>>>> Jeff Zhang >>>>> >>>> >>>> >>> >>> >>> -- >>> 이종열, Jongyoul Lee, 李宗烈 >>> http://madeng.net >>> >> >> > > > -- > 이종열, Jongyoul Lee, 李宗烈 > http://madeng.net >