Here it is Jeff [image: Inline image 1]
On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com> wrote: > Hi Pradeep, > > I don't see the databases on your screenshot (second one for 0.5.6). I > think the output is correct. > > > > On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy < > pradeepreddy.a...@gmail.com> wrote: > >> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin build. It >> seems like Kerberos not working & spark not able to talk to the shared hive >> meta store are defects in the current build. >> >> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy < >> pradeepreddy.a...@gmail.com> wrote: >> >>> Hi Jeff- >>> >>> I switched to local mode now, I'm able to summon the implicit objects >>> like sc, sqlContext etc., but it doesn't show my databases & tables, just >>> shows 1 database "default". >>> >>> Zeppelin Latest Build >>> >>> [image: Inline image 3] >>> >>> Zeppelin 0.5.6, running on the same machine, is able to show my >>> databases and tables. >>> >>> [image: Inline image 4] >>> >>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com> wrote: >>> >>>> > the spark interpreter is not showing my tables & databases, may be >>>> its running in an isolated mode... I'm just getting empty list, so I >>>> attempted to do kerberos authentication to workaround that issue, and >>>> bumped into this road block. >>>> >>>> kerberos would not help here, actually I think it would make the >>>> problem more complicated. You need to first check the log why you get >>>> empty list. >>>> >>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy < >>>> pradeepreddy.a...@gmail.com> wrote: >>>> >>>>> Jeff- I was also successfully able to run spark shell, after running >>>>> kdestroy, with the below command and was able to get to my hive tables. >>>>> >>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab >>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode client >>>>> --master yarn --queue <QUEUE_NAME> >>>>> >>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy < >>>>> pradeepreddy.a...@gmail.com> wrote: >>>>> >>>>>> Thanks Jeff..I have always used zeppelin in local mode, but when I >>>>>> migrated from 0.5.6 to this version, the spark interpreter is not showing >>>>>> my tables & databases, may be its running in an isolated mode... I'm just >>>>>> getting empty list, so I attempted to do kerberos authentication to >>>>>> workaround that issue, and bumped into this road block. >>>>>> >>>>>> Below is the configuration, I also tested my keytab file and its >>>>>> working fine. >>>>>> >>>>>> *Kerberos test:* >>>>>> $ kdestroy >>>>>> >>>>>> $ klist >>>>>> *klist: No credentials cache found (ticket cache >>>>>> FILE:/tmp/krb5cc_12027)* >>>>>> >>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla >>>>>> *Using default cache: /tmp/krb5cc_12027* >>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>* >>>>>> *Using keytab: pradeep_x_alla.keytab* >>>>>> *Authenticated to Kerberos v5* >>>>>> >>>>>> $ klist >>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027* >>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>* >>>>>> >>>>>> *Valid starting Expires Service principal* >>>>>> *08/30/16 20:25:19 08/31/16 06:25:19 krbtgt/<DOMAIN1>@<DOMAIN1>* >>>>>> * renew until 08/31/16 20:25:19* >>>>>> >>>>>> *zeppelin-env.sh* >>>>>> >>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf >>>>>> export SPARK_HOME=/opt/cloudera/parcels/CDH-5.5.1-1.cdh5.5.1.p0.11/ >>>>>> lib/spark >>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master yarn >>>>>> --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>" >>>>>> >>>>>> >>>>>> *Interpreter.json (Spark interpreter config)* >>>>>> "2BUTFVN89": { >>>>>> "id": "2BUTFVN89", >>>>>> "name": "spark", >>>>>> "group": "spark", >>>>>> "properties": { >>>>>> "spark.cores.max": "", >>>>>> "zeppelin.spark.printREPLOutput": "true", >>>>>> "master": "yarn-client", >>>>>> "zeppelin.spark.maxResult": "1000", >>>>>> "zeppelin.dep.localrepo": "local-repo", >>>>>> "spark.app.name": "Zeppelin", >>>>>> "spark.executor.memory": "", >>>>>> "zeppelin.spark.importImplicit": "true", >>>>>> "zeppelin.spark.sql.stacktrace": "true", >>>>>> "zeppelin.spark.useHiveContext": "true", >>>>>> "zeppelin.interpreter.localRepo": >>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89", >>>>>> "zeppelin.spark.concurrentSQL": "false", >>>>>> "args": "", >>>>>> "zeppelin.pyspark.python": "python", >>>>>> "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep. >>>>>> x.alla.keytab", >>>>>> "spark.yarn.principal": "pradeep.x.alla", >>>>>> "zeppelin.dep.additionalRemoteRepository": "spark-packages, >>>>>> http://dl.bintray.com/spark-packages/maven,false;" >>>>>> }, >>>>>> "status": "READY", >>>>>> "interpreterGroup": [ >>>>>> { >>>>>> "name": "spark", >>>>>> "class": "org.apache.zeppelin.spark.SparkInterpreter", >>>>>> "defaultInterpreter": true >>>>>> }, >>>>>> { >>>>>> "name": "sql", >>>>>> "class": "org.apache.zeppelin.spark.SparkSqlInterpreter", >>>>>> "defaultInterpreter": false >>>>>> }, >>>>>> { >>>>>> "name": "dep", >>>>>> "class": "org.apache.zeppelin.spark.DepInterpreter", >>>>>> "defaultInterpreter": false >>>>>> }, >>>>>> { >>>>>> "name": "pyspark", >>>>>> "class": "org.apache.zeppelin.spark.PySparkInterpreter", >>>>>> "defaultInterpreter": false >>>>>> } >>>>>> ], >>>>>> "dependencies": [], >>>>>> "option": { >>>>>> "remote": true, >>>>>> "port": -1, >>>>>> "perNoteSession": false, >>>>>> "perNoteProcess": false, >>>>>> "isExistingProcess": false, >>>>>> "setPermission": false, >>>>>> "users": [] >>>>>> } >>>>>> } >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com> wrote: >>>>>> >>>>>>> It looks like your kerberos configuration issue. Do you mind to >>>>>>> share your configuration ? Or you can first try to run spark-shell using >>>>>>> spark.yarn.keytab & spark.yarn.principle to verify them. >>>>>>> >>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy < >>>>>>> pradeepreddy.a...@gmail.com> wrote: >>>>>>> >>>>>>>> Hi- I recently built zeppelin from source and configured kerberos >>>>>>>> authentication. For Kerberos I added "spark.yarn.keytab" & >>>>>>>> "spark.yarn.principal" and also set master to "yarn-client". But I >>>>>>>> keep >>>>>>>> getting this error whenever I use spark interpreter in the notebook >>>>>>>> >>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar >>>>>>>> kInterpreter335845091 >>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2} >>>>>>>> Job.java[run]:189) - Job failed >>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L >>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L >>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>>>>> DEFAULT >>>>>>>> at org.apache.hadoop.security.aut >>>>>>>> hentication.util.KerberosName.parseRules(KerberosName.java:321) >>>>>>>> at org.apache.hadoop.security.aut >>>>>>>> hentication.util.KerberosName.setRules(KerberosName.java:386) >>>>>>>> at org.apache.hadoop.security.Had >>>>>>>> oopKerberosName.setConfiguration(HadoopKerberosName.java:75) >>>>>>>> at org.apache.hadoop.security.Use >>>>>>>> rGroupInformation.initialize(UserGroupInformation.java:227) >>>>>>>> at org.apache.hadoop.security.Use >>>>>>>> rGroupInformation.ensureInitialized(UserGroupInformation.java:214) >>>>>>>> at org.apache.hadoop.security.Use >>>>>>>> rGroupInformation.isAuthenticationMethodEnabled(UserGroupInf >>>>>>>> ormation.java:275) >>>>>>>> at org.apache.hadoop.security.Use >>>>>>>> rGroupInformation.isSecurityEnabled(UserGroupInformation.java:269) >>>>>>>> at org.apache.hadoop.security.Use >>>>>>>> rGroupInformation.loginUserFromKeytab(UserGroupInformation.j >>>>>>>> ava:820) >>>>>>>> at org.apache.zeppelin.spark.Spar >>>>>>>> kInterpreter.open(SparkInterpreter.java:539) >>>>>>>> at org.apache.zeppelin.interprete >>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) >>>>>>>> at org.apache.zeppelin.interprete >>>>>>>> r.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93) >>>>>>>> at org.apache.zeppelin.interprete >>>>>>>> r.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteI >>>>>>>> nterpreterServer.java:383) >>>>>>>> at org.apache.zeppelin.scheduler.Job.run(Job.java:176) >>>>>>>> at org.apache.zeppelin.scheduler. >>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139) >>>>>>>> at java.util.concurrent.Executors >>>>>>>> $RunnableAdapter.call(Executors.java:511) >>>>>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>>>>>>> at java.util.concurrent.Scheduled >>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledT >>>>>>>> hreadPoolExecutor.java:180) >>>>>>>> at java.util.concurrent.Scheduled >>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPo >>>>>>>> olExecutor.java:293) >>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142) >>>>>>>> at java.util.concurrent.ThreadPoo >>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617) >>>>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>>>> INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2} >>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job >>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler >>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091 >>>>>>>> >>>>>>>> Thanks, >>>>>>>> Pradeep >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Best Regards >>>>>>> >>>>>>> Jeff Zhang >>>>>>> >>>>>> >>>>>> >>>>> >>>> >>>> >>>> -- >>>> Best Regards >>>> >>>> Jeff Zhang >>>> >>> >>> >> > > > -- > Best Regards > > Jeff Zhang >