116 is the databases count number. Do you expect a list of database ? then
you need to use 'z.show(sql("databases"))'

On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <pradeepreddy.a...@gmail.com>
wrote:

> Here it is Jeff
>
> [image: Inline image 1]
>
> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>
>> Hi Pradeep,
>>
>> I don't see the databases on your screenshot (second one for 0.5.6). I
>> think the output is correct.
>>
>>
>>
>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy <
>> pradeepreddy.a...@gmail.com> wrote:
>>
>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin build. It
>>> seems like Kerberos not working & spark not able to talk to the shared hive
>>> meta store are defects in the current build.
>>>
>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy <
>>> pradeepreddy.a...@gmail.com> wrote:
>>>
>>>> Hi Jeff-
>>>>
>>>> I switched to local mode now, I'm able to summon the implicit objects
>>>> like sc, sqlContext etc., but it doesn't show my databases & tables, just
>>>> shows 1 database "default".
>>>>
>>>> Zeppelin Latest Build
>>>>
>>>> [image: Inline image 3]
>>>>
>>>> Zeppelin 0.5.6, running on the same machine, is able to show my
>>>> databases and tables.
>>>>
>>>> [image: Inline image 4]
>>>>
>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>>
>>>>> > the spark interpreter is not showing my tables & databases, may be
>>>>> its running in an isolated mode... I'm just getting empty list, so I
>>>>> attempted to do kerberos authentication to workaround that issue, and
>>>>> bumped into this road block.
>>>>>
>>>>> kerberos would not help here, actually I think it would make the
>>>>> problem more complicated.  You need to first check the log why you get
>>>>> empty list.
>>>>>
>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <
>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>
>>>>>> Jeff- I was also successfully able to run spark shell, after running
>>>>>> kdestroy, with the below command and was able to get to my hive tables.
>>>>>>
>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab
>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode client
>>>>>> --master yarn --queue <QUEUE_NAME>
>>>>>>
>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <
>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>
>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but when I
>>>>>>> migrated from 0.5.6 to this version, the spark interpreter is not 
>>>>>>> showing
>>>>>>> my tables & databases, may be its running in an isolated mode... I'm 
>>>>>>> just
>>>>>>> getting empty list, so I attempted to do kerberos authentication to
>>>>>>> workaround that issue, and bumped into this road block.
>>>>>>>
>>>>>>> Below is the configuration, I also tested my keytab file and its
>>>>>>> working fine.
>>>>>>>
>>>>>>> *Kerberos test:*
>>>>>>> $ kdestroy
>>>>>>>
>>>>>>> $ klist
>>>>>>> *klist: No credentials cache found (ticket cache
>>>>>>> FILE:/tmp/krb5cc_12027)*
>>>>>>>
>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
>>>>>>> *Using default cache: /tmp/krb5cc_12027*
>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>> *Using keytab: pradeep_x_alla.keytab*
>>>>>>> *Authenticated to Kerberos v5*
>>>>>>>
>>>>>>> $ klist
>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027*
>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>
>>>>>>> *Valid starting     Expires            Service principal*
>>>>>>> *08/30/16 20:25:19  08/31/16 06:25:19  krbtgt/<DOMAIN1>@<DOMAIN1>*
>>>>>>> *        renew until 08/31/16 20:25:19*
>>>>>>>
>>>>>>> *zeppelin-env.sh*
>>>>>>>
>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
>>>>>>> export SPARK_HOME=/opt/cloudera/parcels/CDH-5.5.1-1.cdh5.5.1.p0.11/
>>>>>>> lib/spark
>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master yarn
>>>>>>> --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>"
>>>>>>>
>>>>>>>
>>>>>>> *Interpreter.json (Spark interpreter config)*
>>>>>>> "2BUTFVN89": {
>>>>>>>       "id": "2BUTFVN89",
>>>>>>>       "name": "spark",
>>>>>>>       "group": "spark",
>>>>>>>       "properties": {
>>>>>>>         "spark.cores.max": "",
>>>>>>>         "zeppelin.spark.printREPLOutput": "true",
>>>>>>>         "master": "yarn-client",
>>>>>>>         "zeppelin.spark.maxResult": "1000",
>>>>>>>         "zeppelin.dep.localrepo": "local-repo",
>>>>>>>         "spark.app.name": "Zeppelin",
>>>>>>>         "spark.executor.memory": "",
>>>>>>>         "zeppelin.spark.importImplicit": "true",
>>>>>>>         "zeppelin.spark.sql.stacktrace": "true",
>>>>>>>         "zeppelin.spark.useHiveContext": "true",
>>>>>>>         "zeppelin.interpreter.localRepo":
>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
>>>>>>>         "zeppelin.spark.concurrentSQL": "false",
>>>>>>>         "args": "",
>>>>>>>         "zeppelin.pyspark.python": "python",
>>>>>>>         "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.
>>>>>>> x.alla.keytab",
>>>>>>>         "spark.yarn.principal": "pradeep.x.alla",
>>>>>>>         "zeppelin.dep.additionalRemoteRepository": "spark-packages,
>>>>>>> http://dl.bintray.com/spark-packages/maven,false;";
>>>>>>>       },
>>>>>>>       "status": "READY",
>>>>>>>       "interpreterGroup": [
>>>>>>>         {
>>>>>>>           "name": "spark",
>>>>>>>           "class": "org.apache.zeppelin.spark.SparkInterpreter",
>>>>>>>           "defaultInterpreter": true
>>>>>>>         },
>>>>>>>         {
>>>>>>>           "name": "sql",
>>>>>>>           "class": "org.apache.zeppelin.spark.SparkSqlInterpreter",
>>>>>>>           "defaultInterpreter": false
>>>>>>>         },
>>>>>>>         {
>>>>>>>           "name": "dep",
>>>>>>>           "class": "org.apache.zeppelin.spark.DepInterpreter",
>>>>>>>           "defaultInterpreter": false
>>>>>>>         },
>>>>>>>         {
>>>>>>>           "name": "pyspark",
>>>>>>>           "class": "org.apache.zeppelin.spark.PySparkInterpreter",
>>>>>>>           "defaultInterpreter": false
>>>>>>>         }
>>>>>>>       ],
>>>>>>>       "dependencies": [],
>>>>>>>       "option": {
>>>>>>>         "remote": true,
>>>>>>>         "port": -1,
>>>>>>>         "perNoteSession": false,
>>>>>>>         "perNoteProcess": false,
>>>>>>>         "isExistingProcess": false,
>>>>>>>         "setPermission": false,
>>>>>>>         "users": []
>>>>>>>       }
>>>>>>>     }
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> It looks like your kerberos configuration issue. Do you mind to
>>>>>>>> share your configuration ? Or you can first try to run spark-shell 
>>>>>>>> using
>>>>>>>> spark.yarn.keytab & spark.yarn.principle to verify them.
>>>>>>>>
>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Hi- I recently built zeppelin from source and configured kerberos
>>>>>>>>> authentication. For Kerberos I added "spark.yarn.keytab" &
>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client".  But I 
>>>>>>>>> keep
>>>>>>>>> getting this error whenever I use spark interpreter in the notebook
>>>>>>>>>
>>>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar
>>>>>>>>> kInterpreter335845091
>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2}
>>>>>>>>> Job.java[run]:189) - Job failed
>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L
>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L
>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>> DEFAULT
>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>> hentication.util.KerberosName.parseRules(KerberosName.java:321)
>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>> hentication.util.KerberosName.setRules(KerberosName.java:386)
>>>>>>>>>         at org.apache.hadoop.security.Had
>>>>>>>>> oopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>> rGroupInformation.initialize(UserGroupInformation.java:227)
>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>> rGroupInformation.ensureInitialized(UserGroupInformation.java:214)
>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>> rGroupInformation.isAuthenticationMethodEnabled(UserGroupInf
>>>>>>>>> ormation.java:275)
>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>> rGroupInformation.isSecurityEnabled(UserGroupInformation.java:269)
>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>> rGroupInformation.loginUserFromKeytab(UserGroupInformation.j
>>>>>>>>> ava:820)
>>>>>>>>>         at org.apache.zeppelin.spark.Spar
>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539)
>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>> r.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93)
>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>> r.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteI
>>>>>>>>> nterpreterServer.java:383)
>>>>>>>>>         at org.apache.zeppelin.scheduler.Job.run(Job.java:176)
>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139)
>>>>>>>>>         at java.util.concurrent.Executors
>>>>>>>>> $RunnableAdapter.call(Executors.java:511)
>>>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>>>> k.run(FutureTask.java:266)
>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledT
>>>>>>>>> hreadPoolExecutor.java:180)
>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPo
>>>>>>>>> olExecutor.java:293)
>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job
>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler
>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Pradeep
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best Regards
>>>>>>>>
>>>>>>>> Jeff Zhang
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Best Regards
>>>>>
>>>>> Jeff Zhang
>>>>>
>>>>
>>>>
>>>
>>
>>
>> --
>> Best Regards
>>
>> Jeff Zhang
>>
>
>


-- 
Best Regards

Jeff Zhang

Reply via email to