> the spark interpreter is not showing my tables & databases, may be its running in an isolated mode... I'm just getting empty list, so I attempted to do kerberos authentication to workaround that issue, and bumped into this road block.
kerberos would not help here, actually I think it would make the problem more complicated. You need to first check the log why you get empty list. On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <pradeepreddy.a...@gmail.com> wrote: > Jeff- I was also successfully able to run spark shell, after running > kdestroy, with the below command and was able to get to my hive tables. > > spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab --conf > spark.yarn.principal=pradeep.x.alla --deploy-mode client --master yarn > --queue <QUEUE_NAME> > > On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy < > pradeepreddy.a...@gmail.com> wrote: > >> Thanks Jeff..I have always used zeppelin in local mode, but when I >> migrated from 0.5.6 to this version, the spark interpreter is not showing >> my tables & databases, may be its running in an isolated mode... I'm just >> getting empty list, so I attempted to do kerberos authentication to >> workaround that issue, and bumped into this road block. >> >> Below is the configuration, I also tested my keytab file and its working >> fine. >> >> *Kerberos test:* >> $ kdestroy >> >> $ klist >> *klist: No credentials cache found (ticket cache FILE:/tmp/krb5cc_12027)* >> >> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla >> *Using default cache: /tmp/krb5cc_12027* >> *Using principal: pradeep.x.alla@<DOMAIN1>* >> *Using keytab: pradeep_x_alla.keytab* >> *Authenticated to Kerberos v5* >> >> $ klist >> *Ticket cache: FILE:/tmp/krb5cc_12027* >> *Default principal: pradeep.x.alla@<DOMAIN1>* >> >> *Valid starting Expires Service principal* >> *08/30/16 20:25:19 08/31/16 06:25:19 krbtgt/<DOMAIN1>@<DOMAIN1>* >> * renew until 08/31/16 20:25:19* >> >> *zeppelin-env.sh* >> >> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf >> export SPARK_HOME=/opt/cloudera/parcels/CDH-5.5.1-1.cdh5.5.1.p0.11/ >> lib/spark >> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master yarn >> --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>" >> >> >> *Interpreter.json (Spark interpreter config)* >> "2BUTFVN89": { >> "id": "2BUTFVN89", >> "name": "spark", >> "group": "spark", >> "properties": { >> "spark.cores.max": "", >> "zeppelin.spark.printREPLOutput": "true", >> "master": "yarn-client", >> "zeppelin.spark.maxResult": "1000", >> "zeppelin.dep.localrepo": "local-repo", >> "spark.app.name": "Zeppelin", >> "spark.executor.memory": "", >> "zeppelin.spark.importImplicit": "true", >> "zeppelin.spark.sql.stacktrace": "true", >> "zeppelin.spark.useHiveContext": "true", >> "zeppelin.interpreter.localRepo": "/home/pradeep.x.alla/zeppelin >> /local-repo/2BUTFVN89", >> "zeppelin.spark.concurrentSQL": "false", >> "args": "", >> "zeppelin.pyspark.python": "python", >> "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep. >> x.alla.keytab", >> "spark.yarn.principal": "pradeep.x.alla", >> "zeppelin.dep.additionalRemoteRepository": "spark-packages, >> http://dl.bintray.com/spark-packages/maven,false;" >> }, >> "status": "READY", >> "interpreterGroup": [ >> { >> "name": "spark", >> "class": "org.apache.zeppelin.spark.SparkInterpreter", >> "defaultInterpreter": true >> }, >> { >> "name": "sql", >> "class": "org.apache.zeppelin.spark.SparkSqlInterpreter", >> "defaultInterpreter": false >> }, >> { >> "name": "dep", >> "class": "org.apache.zeppelin.spark.DepInterpreter", >> "defaultInterpreter": false >> }, >> { >> "name": "pyspark", >> "class": "org.apache.zeppelin.spark.PySparkInterpreter", >> "defaultInterpreter": false >> } >> ], >> "dependencies": [], >> "option": { >> "remote": true, >> "port": -1, >> "perNoteSession": false, >> "perNoteProcess": false, >> "isExistingProcess": false, >> "setPermission": false, >> "users": [] >> } >> } >> >> >> >> >> >> >> >> >> >> >> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com> wrote: >> >>> It looks like your kerberos configuration issue. Do you mind to share >>> your configuration ? Or you can first try to run spark-shell using >>> spark.yarn.keytab & spark.yarn.principle to verify them. >>> >>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy < >>> pradeepreddy.a...@gmail.com> wrote: >>> >>>> Hi- I recently built zeppelin from source and configured kerberos >>>> authentication. For Kerberos I added "spark.yarn.keytab" & >>>> "spark.yarn.principal" and also set master to "yarn-client". But I keep >>>> getting this error whenever I use spark interpreter in the notebook >>>> >>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar >>>> kInterpreter335845091 >>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2} Job.java[run]:189) - >>>> Job failed >>>> java.lang.IllegalArgumentException: Invalid rule: L >>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L >>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>> DEFAULT >>>> at org.apache.hadoop.security.authentication.util.KerberosName. >>>> parseRules(KerberosName.java:321) >>>> at org.apache.hadoop.security.authentication.util.KerberosName. >>>> setRules(KerberosName.java:386) >>>> at org.apache.hadoop.security.HadoopKerberosName.setConfigurati >>>> on(HadoopKerberosName.java:75) >>>> at org.apache.hadoop.security.UserGroupInformation.initialize(U >>>> serGroupInformation.java:227) >>>> at org.apache.hadoop.security.UserGroupInformation.ensureInitia >>>> lized(UserGroupInformation.java:214) >>>> at org.apache.hadoop.security.UserGroupInformation.isAuthentica >>>> tionMethodEnabled(UserGroupInformation.java:275) >>>> at org.apache.hadoop.security.UserGroupInformation.isSecurityEn >>>> abled(UserGroupInformation.java:269) >>>> at org.apache.hadoop.security.UserGroupInformation.loginUserFro >>>> mKeytab(UserGroupInformation.java:820) >>>> at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpr >>>> eter.java:539) >>>> at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(Laz >>>> yOpenInterpreter.java:69) >>>> at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpre >>>> t(LazyOpenInterpreter.java:93) >>>> at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServ >>>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383) >>>> at org.apache.zeppelin.scheduler.Job.run(Job.java:176) >>>> at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOSchedu >>>> ler.java:139) >>>> at java.util.concurrent.Executors$RunnableAdapter.call(Executor >>>> s.java:511) >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>>> at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFu >>>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180) >>>> at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFu >>>> tureTask.run(ScheduledThreadPoolExecutor.java:293) >>>> at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool >>>> Executor.java:1142) >>>> at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo >>>> lExecutor.java:617) >>>> at java.lang.Thread.run(Thread.java:745) >>>> INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2} >>>> SchedulerFactory.java[jobFinished]:137) - Job >>>> remoteInterpretJob_1472593536728 finished by scheduler >>>> org.apache.zeppelin.spark.SparkInterpreter335845091 >>>> >>>> Thanks, >>>> Pradeep >>>> >>>> >>> >>> >>> -- >>> Best Regards >>> >>> Jeff Zhang >>> >> >> > -- Best Regards Jeff Zhang