Hi Jeff- I switched to local mode now, I'm able to summon the implicit objects like sc, sqlContext etc., but it doesn't show my databases & tables, just shows 1 database "default".
Zeppelin Latest Build [image: Inline image 3] Zeppelin 0.5.6, running on the same machine, is able to show my databases and tables. [image: Inline image 4] On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com> wrote: > > the spark interpreter is not showing my tables & databases, may be its > running in an isolated mode... I'm just getting empty list, so I attempted > to do kerberos authentication to workaround that issue, and bumped into > this road block. > > kerberos would not help here, actually I think it would make the problem > more complicated. You need to first check the log why you get empty list. > > On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy < > pradeepreddy.a...@gmail.com> wrote: > >> Jeff- I was also successfully able to run spark shell, after running >> kdestroy, with the below command and was able to get to my hive tables. >> >> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab --conf >> spark.yarn.principal=pradeep.x.alla --deploy-mode client --master yarn >> --queue <QUEUE_NAME> >> >> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy < >> pradeepreddy.a...@gmail.com> wrote: >> >>> Thanks Jeff..I have always used zeppelin in local mode, but when I >>> migrated from 0.5.6 to this version, the spark interpreter is not showing >>> my tables & databases, may be its running in an isolated mode... I'm just >>> getting empty list, so I attempted to do kerberos authentication to >>> workaround that issue, and bumped into this road block. >>> >>> Below is the configuration, I also tested my keytab file and its working >>> fine. >>> >>> *Kerberos test:* >>> $ kdestroy >>> >>> $ klist >>> *klist: No credentials cache found (ticket cache FILE:/tmp/krb5cc_12027)* >>> >>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla >>> *Using default cache: /tmp/krb5cc_12027* >>> *Using principal: pradeep.x.alla@<DOMAIN1>* >>> *Using keytab: pradeep_x_alla.keytab* >>> *Authenticated to Kerberos v5* >>> >>> $ klist >>> *Ticket cache: FILE:/tmp/krb5cc_12027* >>> *Default principal: pradeep.x.alla@<DOMAIN1>* >>> >>> *Valid starting Expires Service principal* >>> *08/30/16 20:25:19 08/31/16 06:25:19 krbtgt/<DOMAIN1>@<DOMAIN1>* >>> * renew until 08/31/16 20:25:19* >>> >>> *zeppelin-env.sh* >>> >>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf >>> export SPARK_HOME=/opt/cloudera/parcels/CDH-5.5.1-1.cdh5.5.1.p0.11/ >>> lib/spark >>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master yarn >>> --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>" >>> >>> >>> *Interpreter.json (Spark interpreter config)* >>> "2BUTFVN89": { >>> "id": "2BUTFVN89", >>> "name": "spark", >>> "group": "spark", >>> "properties": { >>> "spark.cores.max": "", >>> "zeppelin.spark.printREPLOutput": "true", >>> "master": "yarn-client", >>> "zeppelin.spark.maxResult": "1000", >>> "zeppelin.dep.localrepo": "local-repo", >>> "spark.app.name": "Zeppelin", >>> "spark.executor.memory": "", >>> "zeppelin.spark.importImplicit": "true", >>> "zeppelin.spark.sql.stacktrace": "true", >>> "zeppelin.spark.useHiveContext": "true", >>> "zeppelin.interpreter.localRepo": "/home/pradeep.x.alla/zeppelin >>> /local-repo/2BUTFVN89", >>> "zeppelin.spark.concurrentSQL": "false", >>> "args": "", >>> "zeppelin.pyspark.python": "python", >>> "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep. >>> x.alla.keytab", >>> "spark.yarn.principal": "pradeep.x.alla", >>> "zeppelin.dep.additionalRemoteRepository": "spark-packages, >>> http://dl.bintray.com/spark-packages/maven,false;" >>> }, >>> "status": "READY", >>> "interpreterGroup": [ >>> { >>> "name": "spark", >>> "class": "org.apache.zeppelin.spark.SparkInterpreter", >>> "defaultInterpreter": true >>> }, >>> { >>> "name": "sql", >>> "class": "org.apache.zeppelin.spark.SparkSqlInterpreter", >>> "defaultInterpreter": false >>> }, >>> { >>> "name": "dep", >>> "class": "org.apache.zeppelin.spark.DepInterpreter", >>> "defaultInterpreter": false >>> }, >>> { >>> "name": "pyspark", >>> "class": "org.apache.zeppelin.spark.PySparkInterpreter", >>> "defaultInterpreter": false >>> } >>> ], >>> "dependencies": [], >>> "option": { >>> "remote": true, >>> "port": -1, >>> "perNoteSession": false, >>> "perNoteProcess": false, >>> "isExistingProcess": false, >>> "setPermission": false, >>> "users": [] >>> } >>> } >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com> wrote: >>> >>>> It looks like your kerberos configuration issue. Do you mind to share >>>> your configuration ? Or you can first try to run spark-shell using >>>> spark.yarn.keytab & spark.yarn.principle to verify them. >>>> >>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy < >>>> pradeepreddy.a...@gmail.com> wrote: >>>> >>>>> Hi- I recently built zeppelin from source and configured kerberos >>>>> authentication. For Kerberos I added "spark.yarn.keytab" & >>>>> "spark.yarn.principal" and also set master to "yarn-client". But I keep >>>>> getting this error whenever I use spark interpreter in the notebook >>>>> >>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar >>>>> kInterpreter335845091 >>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2} Job.java[run]:189) >>>>> - Job failed >>>>> java.lang.IllegalArgumentException: Invalid rule: L >>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L >>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L >>>>> DEFAULT >>>>> at org.apache.hadoop.security.aut >>>>> hentication.util.KerberosName.parseRules(KerberosName.java:321) >>>>> at org.apache.hadoop.security.aut >>>>> hentication.util.KerberosName.setRules(KerberosName.java:386) >>>>> at org.apache.hadoop.security.Had >>>>> oopKerberosName.setConfiguration(HadoopKerberosName.java:75) >>>>> at org.apache.hadoop.security.Use >>>>> rGroupInformation.initialize(UserGroupInformation.java:227) >>>>> at org.apache.hadoop.security.Use >>>>> rGroupInformation.ensureInitialized(UserGroupInformation.java:214) >>>>> at org.apache.hadoop.security.Use >>>>> rGroupInformation.isAuthenticationMethodEnabled(UserGroupInf >>>>> ormation.java:275) >>>>> at org.apache.hadoop.security.Use >>>>> rGroupInformation.isSecurityEnabled(UserGroupInformation.java:269) >>>>> at org.apache.hadoop.security.Use >>>>> rGroupInformation.loginUserFromKeytab(UserGroupInformation.java:820) >>>>> at org.apache.zeppelin.spark.Spar >>>>> kInterpreter.open(SparkInterpreter.java:539) >>>>> at org.apache.zeppelin.interprete >>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69) >>>>> at org.apache.zeppelin.interprete >>>>> r.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93) >>>>> at org.apache.zeppelin.interprete >>>>> r.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteI >>>>> nterpreterServer.java:383) >>>>> at org.apache.zeppelin.scheduler.Job.run(Job.java:176) >>>>> at org.apache.zeppelin.scheduler. >>>>> FIFOScheduler$1.run(FIFOScheduler.java:139) >>>>> at java.util.concurrent.Executors >>>>> $RunnableAdapter.call(Executors.java:511) >>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) >>>>> at java.util.concurrent.Scheduled >>>>> ThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledT >>>>> hreadPoolExecutor.java:180) >>>>> at java.util.concurrent.Scheduled >>>>> ThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPo >>>>> olExecutor.java:293) >>>>> at java.util.concurrent.ThreadPoo >>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142) >>>>> at java.util.concurrent.ThreadPoo >>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617) >>>>> at java.lang.Thread.run(Thread.java:745) >>>>> INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2} >>>>> SchedulerFactory.java[jobFinished]:137) - Job >>>>> remoteInterpretJob_1472593536728 finished by scheduler >>>>> org.apache.zeppelin.spark.SparkInterpreter335845091 >>>>> >>>>> Thanks, >>>>> Pradeep >>>>> >>>>> >>>> >>>> >>>> -- >>>> Best Regards >>>> >>>> Jeff Zhang >>>> >>> >>> >> > > > -- > Best Regards > > Jeff Zhang >