Jeff- I was also successfully able to run spark shell, after running kdestroy, with the below command and was able to get to my hive tables.
spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab --conf spark.yarn.principal=pradeep.x.alla --deploy-mode client --master yarn --queue <QUEUE_NAME> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <pradeepreddy.a...@gmail.com> wrote: > Thanks Jeff..I have always used zeppelin in local mode, but when I > migrated from 0.5.6 to this version, the spark interpreter is not showing > my tables & databases, may be its running in an isolated mode... I'm just > getting empty list, so I attempted to do kerberos authentication to > workaround that issue, and bumped into this road block. > > Below is the configuration, I also tested my keytab file and its working > fine. > > *Kerberos test:* > $ kdestroy > > $ klist > *klist: No credentials cache found (ticket cache FILE:/tmp/krb5cc_12027)* > > $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla > *Using default cache: /tmp/krb5cc_12027* > *Using principal: pradeep.x.alla@<DOMAIN1>* > *Using keytab: pradeep_x_alla.keytab* > *Authenticated to Kerberos v5* > > $ klist > *Ticket cache: FILE:/tmp/krb5cc_12027* > *Default principal: pradeep.x.alla@<DOMAIN1>* > > *Valid starting Expires Service principal* > *08/30/16 20:25:19 08/31/16 06:25:19 krbtgt/<DOMAIN1>@<DOMAIN1>* > * renew until 08/31/16 20:25:19* > > *zeppelin-env.sh* > > export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf > export SPARK_HOME=/opt/cloudera/parcels/CDH-5.5.1-1.cdh5.5.1. > p0.11/lib/spark > export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master yarn > --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>" > > > *Interpreter.json (Spark interpreter config)* > "2BUTFVN89": { > "id": "2BUTFVN89", > "name": "spark", > "group": "spark", > "properties": { > "spark.cores.max": "", > "zeppelin.spark.printREPLOutput": "true", > "master": "yarn-client", > "zeppelin.spark.maxResult": "1000", > "zeppelin.dep.localrepo": "local-repo", > "spark.app.name": "Zeppelin", > "spark.executor.memory": "", > "zeppelin.spark.importImplicit": "true", > "zeppelin.spark.sql.stacktrace": "true", > "zeppelin.spark.useHiveContext": "true", > "zeppelin.interpreter.localRepo": "/home/pradeep.x.alla/ > zeppelin/local-repo/2BUTFVN89", > "zeppelin.spark.concurrentSQL": "false", > "args": "", > "zeppelin.pyspark.python": "python", > "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.x.alla.keytab", > "spark.yarn.principal": "pradeep.x.alla", > "zeppelin.dep.additionalRemoteRepository": "spark-packages, > http://dl.bintray.com/spark-packages/maven,false;" > }, > "status": "READY", > "interpreterGroup": [ > { > "name": "spark", > "class": "org.apache.zeppelin.spark.SparkInterpreter", > "defaultInterpreter": true > }, > { > "name": "sql", > "class": "org.apache.zeppelin.spark.SparkSqlInterpreter", > "defaultInterpreter": false > }, > { > "name": "dep", > "class": "org.apache.zeppelin.spark.DepInterpreter", > "defaultInterpreter": false > }, > { > "name": "pyspark", > "class": "org.apache.zeppelin.spark.PySparkInterpreter", > "defaultInterpreter": false > } > ], > "dependencies": [], > "option": { > "remote": true, > "port": -1, > "perNoteSession": false, > "perNoteProcess": false, > "isExistingProcess": false, > "setPermission": false, > "users": [] > } > } > > > > > > > > > > > On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com> wrote: > >> It looks like your kerberos configuration issue. Do you mind to share >> your configuration ? Or you can first try to run spark-shell using >> spark.yarn.keytab & spark.yarn.principle to verify them. >> >> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy < >> pradeepreddy.a...@gmail.com> wrote: >> >>> Hi- I recently built zeppelin from source and configured kerberos >>> authentication. For Kerberos I added "spark.yarn.keytab" & >>> "spark.yarn.principal" and also set master to "yarn-client". But I keep >>> getting this error whenever I use spark interpreter in the notebook >>> >>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar >>> kInterpreter335845091 >>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2} Job.java[run]:189) - >>> Job failed >>> java.lang.IllegalArgumentException: Invalid rule: L >>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L >>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>> DEFAULT >>> at org.apache.hadoop.security.authentication.util.KerberosName. >>> parseRules(KerberosName.java:321) >>> at org.apache.hadoop.security.authentication.util.KerberosName. >>> setRules(KerberosName.java:386) >>> at org.apache.hadoop.security.HadoopKerberosName.setConfigurati >>> on(HadoopKerberosName.java:75) >>> at org.apache.hadoop.security.UserGroupInformation.initialize(U >>> serGroupInformation.java:227) >>> at org.apache.hadoop.security.UserGroupInformation.ensureInitia >>> lized(UserGroupInformation.java:214) >>> at org.apache.hadoop.security.UserGroupInformation.isAuthentica >>> tionMethodEnabled(UserGroupInformation.java:275) >>> at org.apache.hadoop.security.UserGroupInformation.isSecurityEn >>> abled(UserGroupInformation.java:269) >>> at org.apache.hadoop.security.UserGroupInformation.loginUserFro >>> mKeytab(UserGroupInformation.java:820) >>> at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpr >>> eter.java:539) >>> at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(Laz >>> yOpenInterpreter.java:69) >>> at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpre >>> t(LazyOpenInterpreter.java:93) >>> at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServ >>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383) >>> at org.apache.zeppelin.scheduler.Job.run(Job.java:176) >>> at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOSchedu >>> ler.java:139) >>> at java.util.concurrent.Executors$RunnableAdapter.call(Executor >>> s.java:511) >>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>> at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFu >>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180) >>> at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFu >>> tureTask.run(ScheduledThreadPoolExecutor.java:293) >>> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool >>> Executor.java:1142) >>> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo >>> lExecutor.java:617) >>> at java.lang.Thread.run(Thread.java:745) >>> INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2} >>> SchedulerFactory.java[jobFinished]:137) - Job >>> remoteInterpretJob_1472593536728 finished by scheduler >>> org.apache.zeppelin.spark.SparkInterpreter335845091 >>> >>> Thanks, >>> Pradeep >>> >>> >> >> >> -- >> Best Regards >> >> Jeff Zhang >> > >