I just tried enabling kerberos on 0.6.1 and its now able to talk to my hive metastore, I see all my databases and tables. However the moment I takeout kerberos and run zeppelin 0.6.1 in local mode by resetting the "master" & removing spark variables, i see just one default database.
I can live with this, as enabling kerberos integration is able to make zeppelin talk to my hive metastore. Thanks, Pradeep On Wed, Aug 31, 2016 at 8:36 PM, Pradeep Reddy <pradeepreddy.a...@gmail.com> wrote: > No luck, even after copying the hive-site.xml in > interpreter/spark/dep....I also tried, downloading 0.6.1 src vs 0.5.6. for > the same steps I followed, 0.5.6 is able to talk to my hive metastore, > where as the other recent builds are not, they just show one "default" > database. > > Thanks, > Pradeep > > On Wed, Aug 31, 2016 at 12:48 PM, Jongyoul Lee <jongy...@gmail.com> wrote: > >> I think it's related to https://issues.apache.org/j >> ira/browse/ZEPPELIN-1175 which remove some class path when Zeppelin >> launches interpreter. Could you please check your hive-site.xml is included >> in your interpreter process? It looks like a configuration issue because >> you can see the default database. If it doesn't exists, you should copy >> your xml into interpreter/spark/dep/ >> >> Regards, >> JL >> >> On Wed, Aug 31, 2016 at 9:52 PM, Pradeep Reddy < >> pradeepreddy.a...@gmail.com> wrote: >> >>> Hi Jongyoul- I followed the exact same steps for compiling and setting >>> up the new build from source as 0.5.6 (only difference is, I acquired the >>> source for latest build using "git clone") >>> >>> hive-site.xml was copied to conf directory. But, the spark interpreter >>> is not talking to the hive metastore. Both the 0.5.6 & the latest builds >>> are running in the same machine. In 0.5.6 when i run the below command, I >>> see 116 databases listed, as per my expectations and I'm able to run my >>> notebooks built on those databases. >>> >>> [image: Inline image 1] >>> >>> Thanks, >>> Pradeep >>> >>> >>> On Wed, Aug 31, 2016 at 2:52 AM, Jongyoul Lee <jongy...@gmail.com> >>> wrote: >>> >>>> Hello, >>>> >>>> Do you copy your hive-site.xml in a proper position? >>>> >>>> On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy < >>>> pradeepreddy.a...@gmail.com> wrote: >>>> >>>>> nothing obvious. I will stick to 0.5.6 build, until the latest builds >>>>> stabilize. >>>>> >>>>> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zjf...@gmail.com> wrote: >>>>> >>>>>> Then I guess maybe you are connecting to different database. Why not >>>>>> using 'z.show(sql("databases"))' to display the databases ? Then >>>>>> you will get a hint what's going on. >>>>>> >>>>>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy < >>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>> >>>>>>> Yes...I didn't wish to show the names of the databases that we have >>>>>>> in our data lake on that screen shot. so thats why I chose to display >>>>>>> the >>>>>>> count. The latest zeppelin build just shows 1 count which is "default" >>>>>>> database. >>>>>>> >>>>>>> Thanks, >>>>>>> Pradeep >>>>>>> >>>>>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zjf...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> 116 is the databases count number. Do you expect a list of database >>>>>>>> ? then you need to use 'z.show(sql("databases"))' >>>>>>>> >>>>>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy < >>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Here it is Jeff >>>>>>>>> >>>>>>>>> [image: Inline image 1] >>>>>>>>> >>>>>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi Pradeep, >>>>>>>>>> >>>>>>>>>> I don't see the databases on your screenshot (second one for >>>>>>>>>> 0.5.6). I think the output is correct. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy < >>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin >>>>>>>>>>> build. It seems like Kerberos not working & spark not able to talk >>>>>>>>>>> to the >>>>>>>>>>> shared hive meta store are defects in the current build. >>>>>>>>>>> >>>>>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy < >>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi Jeff- >>>>>>>>>>>> >>>>>>>>>>>> I switched to local mode now, I'm able to summon the implicit >>>>>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my databases >>>>>>>>>>>> & >>>>>>>>>>>> tables, just shows 1 database "default". >>>>>>>>>>>> >>>>>>>>>>>> Zeppelin Latest Build >>>>>>>>>>>> >>>>>>>>>>>> [image: Inline image 3] >>>>>>>>>>>> >>>>>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show my >>>>>>>>>>>> databases and tables. >>>>>>>>>>>> >>>>>>>>>>>> [image: Inline image 4] >>>>>>>>>>>> >>>>>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com> >>>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> > the spark interpreter is not showing my tables & databases, >>>>>>>>>>>>> may be its running in an isolated mode... I'm just getting empty >>>>>>>>>>>>> list, so I >>>>>>>>>>>>> attempted to do kerberos authentication to workaround that issue, >>>>>>>>>>>>> and >>>>>>>>>>>>> bumped into this road block. >>>>>>>>>>>>> >>>>>>>>>>>>> kerberos would not help here, actually I think it would make >>>>>>>>>>>>> the problem more complicated. You need to first check the log >>>>>>>>>>>>> why you get >>>>>>>>>>>>> empty list. >>>>>>>>>>>>> >>>>>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy < >>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Jeff- I was also successfully able to run spark shell, after >>>>>>>>>>>>>> running kdestroy, with the below command and was able to get to >>>>>>>>>>>>>> my hive >>>>>>>>>>>>>> tables. >>>>>>>>>>>>>> >>>>>>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab >>>>>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode >>>>>>>>>>>>>> client --master yarn --queue <QUEUE_NAME> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy < >>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but >>>>>>>>>>>>>>> when I migrated from 0.5.6 to this version, the spark >>>>>>>>>>>>>>> interpreter is not >>>>>>>>>>>>>>> showing my tables & databases, may be its running in an >>>>>>>>>>>>>>> isolated mode... >>>>>>>>>>>>>>> I'm just getting empty list, so I attempted to do kerberos >>>>>>>>>>>>>>> authentication >>>>>>>>>>>>>>> to workaround that issue, and bumped into this road block. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Below is the configuration, I also tested my keytab file and >>>>>>>>>>>>>>> its working fine. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> *Kerberos test:* >>>>>>>>>>>>>>> $ kdestroy >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> $ klist >>>>>>>>>>>>>>> *klist: No credentials cache found (ticket cache >>>>>>>>>>>>>>> FILE:/tmp/krb5cc_12027)* >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla >>>>>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027* >>>>>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>* >>>>>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab* >>>>>>>>>>>>>>> *Authenticated to Kerberos v5* >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> $ klist >>>>>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027* >>>>>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>* >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> *Valid starting Expires Service principal* >>>>>>>>>>>>>>> *08/30/16 20:25:19 08/31/16 06:25:19 >>>>>>>>>>>>>>> krbtgt/<DOMAIN1>@<DOMAIN1>* >>>>>>>>>>>>>>> * renew until 08/31/16 20:25:19* >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> *zeppelin-env.sh* >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf >>>>>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce >>>>>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark >>>>>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master >>>>>>>>>>>>>>> yarn --num-executors 2 --executor-memory 2g --queue >>>>>>>>>>>>>>> <QUEUE_NAME>" >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> *Interpreter.json (Spark interpreter config)* >>>>>>>>>>>>>>> "2BUTFVN89": { >>>>>>>>>>>>>>> "id": "2BUTFVN89", >>>>>>>>>>>>>>> "name": "spark", >>>>>>>>>>>>>>> "group": "spark", >>>>>>>>>>>>>>> "properties": { >>>>>>>>>>>>>>> "spark.cores.max": "", >>>>>>>>>>>>>>> "zeppelin.spark.printREPLOutput": "true", >>>>>>>>>>>>>>> "master": "yarn-client", >>>>>>>>>>>>>>> "zeppelin.spark.maxResult": "1000", >>>>>>>>>>>>>>> "zeppelin.dep.localrepo": "local-repo", >>>>>>>>>>>>>>> "spark.app.name": "Zeppelin", >>>>>>>>>>>>>>> "spark.executor.memory": "", >>>>>>>>>>>>>>> "zeppelin.spark.importImplicit": "true", >>>>>>>>>>>>>>> "zeppelin.spark.sql.stacktrace": "true", >>>>>>>>>>>>>>> "zeppelin.spark.useHiveContext": "true", >>>>>>>>>>>>>>> "zeppelin.interpreter.localRepo": >>>>>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89", >>>>>>>>>>>>>>> "zeppelin.spark.concurrentSQL": "false", >>>>>>>>>>>>>>> "args": "", >>>>>>>>>>>>>>> "zeppelin.pyspark.python": "python", >>>>>>>>>>>>>>> "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep. >>>>>>>>>>>>>>> x.alla.keytab", >>>>>>>>>>>>>>> "spark.yarn.principal": "pradeep.x.alla", >>>>>>>>>>>>>>> "zeppelin.dep.additionalRemoteRepository": >>>>>>>>>>>>>>> "spark-packages,http://dl.bintray.com/spark-packages/maven,f >>>>>>>>>>>>>>> alse;" >>>>>>>>>>>>>>> }, >>>>>>>>>>>>>>> "status": "READY", >>>>>>>>>>>>>>> "interpreterGroup": [ >>>>>>>>>>>>>>> { >>>>>>>>>>>>>>> "name": "spark", >>>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.Spa >>>>>>>>>>>>>>> rkInterpreter", >>>>>>>>>>>>>>> "defaultInterpreter": true >>>>>>>>>>>>>>> }, >>>>>>>>>>>>>>> { >>>>>>>>>>>>>>> "name": "sql", >>>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.Spa >>>>>>>>>>>>>>> rkSqlInterpreter", >>>>>>>>>>>>>>> "defaultInterpreter": false >>>>>>>>>>>>>>> }, >>>>>>>>>>>>>>> { >>>>>>>>>>>>>>> "name": "dep", >>>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.Dep >>>>>>>>>>>>>>> Interpreter", >>>>>>>>>>>>>>> "defaultInterpreter": false >>>>>>>>>>>>>>> }, >>>>>>>>>>>>>>> { >>>>>>>>>>>>>>> "name": "pyspark", >>>>>>>>>>>>>>> "class": "org.apache.zeppelin.spark.PyS >>>>>>>>>>>>>>> parkInterpreter", >>>>>>>>>>>>>>> "defaultInterpreter": false >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> ], >>>>>>>>>>>>>>> "dependencies": [], >>>>>>>>>>>>>>> "option": { >>>>>>>>>>>>>>> "remote": true, >>>>>>>>>>>>>>> "port": -1, >>>>>>>>>>>>>>> "perNoteSession": false, >>>>>>>>>>>>>>> "perNoteProcess": false, >>>>>>>>>>>>>>> "isExistingProcess": false, >>>>>>>>>>>>>>> "setPermission": false, >>>>>>>>>>>>>>> "users": [] >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> } >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang < >>>>>>>>>>>>>>> zjf...@gmail.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you >>>>>>>>>>>>>>>> mind to share your configuration ? Or you can first try to run >>>>>>>>>>>>>>>> spark-shell >>>>>>>>>>>>>>>> using spark.yarn.keytab & spark.yarn.principle to verify them. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy < >>>>>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured >>>>>>>>>>>>>>>>> kerberos authentication. For Kerberos I added >>>>>>>>>>>>>>>>> "spark.yarn.keytab" & >>>>>>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client". >>>>>>>>>>>>>>>>> But I keep >>>>>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the >>>>>>>>>>>>>>>>> notebook >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar >>>>>>>>>>>>>>>>> kInterpreter335845091 >>>>>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2} >>>>>>>>>>>>>>>>> Job.java[run]:189) - Job failed >>>>>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L >>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L >>>>>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>>>>>>>>>>> DEFAULT >>>>>>>>>>>>>>>>> at org.apache.hadoop.security.aut >>>>>>>>>>>>>>>>> hentication.util.KerberosName. >>>>>>>>>>>>>>>>> parseRules(KerberosName.java:321) >>>>>>>>>>>>>>>>> at org.apache.hadoop.security.aut >>>>>>>>>>>>>>>>> hentication.util.KerberosName. >>>>>>>>>>>>>>>>> setRules(KerberosName.java:386) >>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Had >>>>>>>>>>>>>>>>> oopKerberosName.setConfigurati >>>>>>>>>>>>>>>>> on(HadoopKerberosName.java:75) >>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>> rGroupInformation.initialize(U >>>>>>>>>>>>>>>>> serGroupInformation.java:227) >>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>> rGroupInformation.ensureInitia >>>>>>>>>>>>>>>>> lized(UserGroupInformation.java:214) >>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>> rGroupInformation.isAuthentica >>>>>>>>>>>>>>>>> tionMethodEnabled(UserGroupInformation.java:275) >>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>> rGroupInformation.isSecurityEn >>>>>>>>>>>>>>>>> abled(UserGroupInformation.java:269) >>>>>>>>>>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>>>>>>>>>> rGroupInformation.loginUserFro >>>>>>>>>>>>>>>>> mKeytab(UserGroupInformation.java:820) >>>>>>>>>>>>>>>>> at org.apache.zeppelin.spark.Spar >>>>>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539) >>>>>>>>>>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) >>>>>>>>>>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>>>>>>>>>> r.LazyOpenInterpreter.interpre >>>>>>>>>>>>>>>>> t(LazyOpenInterpreter.java:93) >>>>>>>>>>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>>>>>>>>>> r.remote.RemoteInterpreterServ >>>>>>>>>>>>>>>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383) >>>>>>>>>>>>>>>>> at org.apache.zeppelin.scheduler. >>>>>>>>>>>>>>>>> Job.run(Job.java:176) >>>>>>>>>>>>>>>>> at org.apache.zeppelin.scheduler. >>>>>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139) >>>>>>>>>>>>>>>>> at java.util.concurrent.Executors >>>>>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511) >>>>>>>>>>>>>>>>> at java.util.concurrent.FutureTas >>>>>>>>>>>>>>>>> k.run(FutureTask.java:266) >>>>>>>>>>>>>>>>> at java.util.concurrent.Scheduled >>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu >>>>>>>>>>>>>>>>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180) >>>>>>>>>>>>>>>>> at java.util.concurrent.Scheduled >>>>>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFu >>>>>>>>>>>>>>>>> tureTask.run(ScheduledThreadPoolExecutor.java:293) >>>>>>>>>>>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142) >>>>>>>>>>>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617) >>>>>>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>>>>>>>>>>>>> INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2} >>>>>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job >>>>>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler >>>>>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091 >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>> Pradeep >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> Best Regards >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Jeff Zhang >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Best Regards >>>>>>>>>>>>> >>>>>>>>>>>>> Jeff Zhang >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Best Regards >>>>>>>>>> >>>>>>>>>> Jeff Zhang >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Best Regards >>>>>>>> >>>>>>>> Jeff Zhang >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Best Regards >>>>>> >>>>>> Jeff Zhang >>>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> 이종열, Jongyoul Lee, 李宗烈 >>>> http://madeng.net >>>> >>> >>> >> >> >> -- >> 이종열, Jongyoul Lee, 李宗烈 >> http://madeng.net >> > >