116 is the databases count number. Do you expect a list of database ? then you need to use 'z.show(sql("databases"))'
On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <pradeepreddy.a...@gmail.com> wrote: > Here it is Jeff > > [image: Inline image 1] > > On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com> wrote: > >> Hi Pradeep, >> >> I don't see the databases on your screenshot (second one for 0.5.6). I >> think the output is correct. >> >> >> >> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy < >> pradeepreddy.a...@gmail.com> wrote: >> >>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin build. It >>> seems like Kerberos not working & spark not able to talk to the shared hive >>> meta store are defects in the current build. >>> >>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy < >>> pradeepreddy.a...@gmail.com> wrote: >>> >>>> Hi Jeff- >>>> >>>> I switched to local mode now, I'm able to summon the implicit objects >>>> like sc, sqlContext etc., but it doesn't show my databases & tables, just >>>> shows 1 database "default". >>>> >>>> Zeppelin Latest Build >>>> >>>> [image: Inline image 3] >>>> >>>> Zeppelin 0.5.6, running on the same machine, is able to show my >>>> databases and tables. >>>> >>>> [image: Inline image 4] >>>> >>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com> wrote: >>>> >>>>> > the spark interpreter is not showing my tables & databases, may be >>>>> its running in an isolated mode... I'm just getting empty list, so I >>>>> attempted to do kerberos authentication to workaround that issue, and >>>>> bumped into this road block. >>>>> >>>>> kerberos would not help here, actually I think it would make the >>>>> problem more complicated. You need to first check the log why you get >>>>> empty list. >>>>> >>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy < >>>>> pradeepreddy.a...@gmail.com> wrote: >>>>> >>>>>> Jeff- I was also successfully able to run spark shell, after running >>>>>> kdestroy, with the below command and was able to get to my hive tables. >>>>>> >>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab >>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode client >>>>>> --master yarn --queue <QUEUE_NAME> >>>>>> >>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy < >>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>> >>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but when I >>>>>>> migrated from 0.5.6 to this version, the spark interpreter is not >>>>>>> showing >>>>>>> my tables & databases, may be its running in an isolated mode... I'm >>>>>>> just >>>>>>> getting empty list, so I attempted to do kerberos authentication to >>>>>>> workaround that issue, and bumped into this road block. >>>>>>> >>>>>>> Below is the configuration, I also tested my keytab file and its >>>>>>> working fine. >>>>>>> >>>>>>> *Kerberos test:* >>>>>>> $ kdestroy >>>>>>> >>>>>>> $ klist >>>>>>> *klist: No credentials cache found (ticket cache >>>>>>> FILE:/tmp/krb5cc_12027)* >>>>>>> >>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla >>>>>>> *Using default cache: /tmp/krb5cc_12027* >>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>* >>>>>>> *Using keytab: pradeep_x_alla.keytab* >>>>>>> *Authenticated to Kerberos v5* >>>>>>> >>>>>>> $ klist >>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027* >>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>* >>>>>>> >>>>>>> *Valid starting Expires Service principal* >>>>>>> *08/30/16 20:25:19 08/31/16 06:25:19 krbtgt/<DOMAIN1>@<DOMAIN1>* >>>>>>> * renew until 08/31/16 20:25:19* >>>>>>> >>>>>>> *zeppelin-env.sh* >>>>>>> >>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf >>>>>>> export SPARK_HOME=/opt/cloudera/parcels/CDH-5.5.1-1.cdh5.5.1.p0.11/ >>>>>>> lib/spark >>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master yarn >>>>>>> --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>" >>>>>>> >>>>>>> >>>>>>> *Interpreter.json (Spark interpreter config)* >>>>>>> "2BUTFVN89": { >>>>>>> "id": "2BUTFVN89", >>>>>>> "name": "spark", >>>>>>> "group": "spark", >>>>>>> "properties": { >>>>>>> "spark.cores.max": "", >>>>>>> "zeppelin.spark.printREPLOutput": "true", >>>>>>> "master": "yarn-client", >>>>>>> "zeppelin.spark.maxResult": "1000", >>>>>>> "zeppelin.dep.localrepo": "local-repo", >>>>>>> "spark.app.name": "Zeppelin", >>>>>>> "spark.executor.memory": "", >>>>>>> "zeppelin.spark.importImplicit": "true", >>>>>>> "zeppelin.spark.sql.stacktrace": "true", >>>>>>> "zeppelin.spark.useHiveContext": "true", >>>>>>> "zeppelin.interpreter.localRepo": >>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89", >>>>>>> "zeppelin.spark.concurrentSQL": "false", >>>>>>> "args": "", >>>>>>> "zeppelin.pyspark.python": "python", >>>>>>> "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep. >>>>>>> x.alla.keytab", >>>>>>> "spark.yarn.principal": "pradeep.x.alla", >>>>>>> "zeppelin.dep.additionalRemoteRepository": "spark-packages, >>>>>>> http://dl.bintray.com/spark-packages/maven,false;" >>>>>>> }, >>>>>>> "status": "READY", >>>>>>> "interpreterGroup": [ >>>>>>> { >>>>>>> "name": "spark", >>>>>>> "class": "org.apache.zeppelin.spark.SparkInterpreter", >>>>>>> "defaultInterpreter": true >>>>>>> }, >>>>>>> { >>>>>>> "name": "sql", >>>>>>> "class": "org.apache.zeppelin.spark.SparkSqlInterpreter", >>>>>>> "defaultInterpreter": false >>>>>>> }, >>>>>>> { >>>>>>> "name": "dep", >>>>>>> "class": "org.apache.zeppelin.spark.DepInterpreter", >>>>>>> "defaultInterpreter": false >>>>>>> }, >>>>>>> { >>>>>>> "name": "pyspark", >>>>>>> "class": "org.apache.zeppelin.spark.PySparkInterpreter", >>>>>>> "defaultInterpreter": false >>>>>>> } >>>>>>> ], >>>>>>> "dependencies": [], >>>>>>> "option": { >>>>>>> "remote": true, >>>>>>> "port": -1, >>>>>>> "perNoteSession": false, >>>>>>> "perNoteProcess": false, >>>>>>> "isExistingProcess": false, >>>>>>> "setPermission": false, >>>>>>> "users": [] >>>>>>> } >>>>>>> } >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> It looks like your kerberos configuration issue. Do you mind to >>>>>>>> share your configuration ? Or you can first try to run spark-shell >>>>>>>> using >>>>>>>> spark.yarn.keytab & spark.yarn.principle to verify them. >>>>>>>> >>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy < >>>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Hi- I recently built zeppelin from source and configured kerberos >>>>>>>>> authentication. For Kerberos I added "spark.yarn.keytab" & >>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client". But I >>>>>>>>> keep >>>>>>>>> getting this error whenever I use spark interpreter in the notebook >>>>>>>>> >>>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar >>>>>>>>> kInterpreter335845091 >>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2} >>>>>>>>> Job.java[run]:189) - Job failed >>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L >>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L >>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>>> DEFAULT >>>>>>>>> at org.apache.hadoop.security.aut >>>>>>>>> hentication.util.KerberosName.parseRules(KerberosName.java:321) >>>>>>>>> at org.apache.hadoop.security.aut >>>>>>>>> hentication.util.KerberosName.setRules(KerberosName.java:386) >>>>>>>>> at org.apache.hadoop.security.Had >>>>>>>>> oopKerberosName.setConfiguration(HadoopKerberosName.java:75) >>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>> rGroupInformation.initialize(UserGroupInformation.java:227) >>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>> rGroupInformation.ensureInitialized(UserGroupInformation.java:214) >>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>> rGroupInformation.isAuthenticationMethodEnabled(UserGroupInf >>>>>>>>> ormation.java:275) >>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>> rGroupInformation.isSecurityEnabled(UserGroupInformation.java:269) >>>>>>>>> at org.apache.hadoop.security.Use >>>>>>>>> rGroupInformation.loginUserFromKeytab(UserGroupInformation.j >>>>>>>>> ava:820) >>>>>>>>> at org.apache.zeppelin.spark.Spar >>>>>>>>> kInterpreter.open(SparkInterpreter.java:539) >>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) >>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>> r.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93) >>>>>>>>> at org.apache.zeppelin.interprete >>>>>>>>> r.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteI >>>>>>>>> nterpreterServer.java:383) >>>>>>>>> at org.apache.zeppelin.scheduler.Job.run(Job.java:176) >>>>>>>>> at org.apache.zeppelin.scheduler. >>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139) >>>>>>>>> at java.util.concurrent.Executors >>>>>>>>> $RunnableAdapter.call(Executors.java:511) >>>>>>>>> at java.util.concurrent.FutureTas >>>>>>>>> k.run(FutureTask.java:266) >>>>>>>>> at java.util.concurrent.Scheduled >>>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledT >>>>>>>>> hreadPoolExecutor.java:180) >>>>>>>>> at java.util.concurrent.Scheduled >>>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPo >>>>>>>>> olExecutor.java:293) >>>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142) >>>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617) >>>>>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>>>>> INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2} >>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job >>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler >>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091 >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Pradeep >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Best Regards >>>>>>>> >>>>>>>> Jeff Zhang >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> Best Regards >>>>> >>>>> Jeff Zhang >>>>> >>>> >>>> >>> >> >> >> -- >> Best Regards >> >> Jeff Zhang >> > > -- Best Regards Jeff Zhang