nothing obvious. I will stick to 0.5.6 build, until the latest builds
stabilize.

On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zjf...@gmail.com> wrote:

> Then I guess maybe you are connecting to different database. Why not using
>  'z.show(sql("databases"))' to display the databases ? Then you will get
> a hint what's going on.
>
> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy <
> pradeepreddy.a...@gmail.com> wrote:
>
>> Yes...I didn't wish to show the names of the databases that we have in
>> our data lake on that screen shot. so thats why I chose to display the
>> count. The latest zeppelin build just shows 1 count which is "default"
>> database.
>>
>> Thanks,
>> Pradeep
>>
>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>>
>>> 116 is the databases count number. Do you expect a list of database ?
>>> then you need to use 'z.show(sql("databases"))'
>>>
>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <
>>> pradeepreddy.a...@gmail.com> wrote:
>>>
>>>> Here it is Jeff
>>>>
>>>> [image: Inline image 1]
>>>>
>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>>
>>>>> Hi Pradeep,
>>>>>
>>>>> I don't see the databases on your screenshot (second one for 0.5.6). I
>>>>> think the output is correct.
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy <
>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>
>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin build. It
>>>>>> seems like Kerberos not working & spark not able to talk to the shared 
>>>>>> hive
>>>>>> meta store are defects in the current build.
>>>>>>
>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy <
>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>
>>>>>>> Hi Jeff-
>>>>>>>
>>>>>>> I switched to local mode now, I'm able to summon the implicit
>>>>>>> objects like sc, sqlContext etc., but it doesn't show my databases &
>>>>>>> tables, just shows 1 database "default".
>>>>>>>
>>>>>>> Zeppelin Latest Build
>>>>>>>
>>>>>>> [image: Inline image 3]
>>>>>>>
>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show my
>>>>>>> databases and tables.
>>>>>>>
>>>>>>> [image: Inline image 4]
>>>>>>>
>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> > the spark interpreter is not showing my tables & databases, may
>>>>>>>> be its running in an isolated mode... I'm just getting empty list, so I
>>>>>>>> attempted to do kerberos authentication to workaround that issue, and
>>>>>>>> bumped into this road block.
>>>>>>>>
>>>>>>>> kerberos would not help here, actually I think it would make the
>>>>>>>> problem more complicated.  You need to first check the log why you get
>>>>>>>> empty list.
>>>>>>>>
>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <
>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Jeff- I was also successfully able to run spark shell, after
>>>>>>>>> running kdestroy, with the below command and was able to get to my 
>>>>>>>>> hive
>>>>>>>>> tables.
>>>>>>>>>
>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab
>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode client
>>>>>>>>> --master yarn --queue <QUEUE_NAME>
>>>>>>>>>
>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <
>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but when
>>>>>>>>>> I migrated from 0.5.6 to this version, the spark interpreter is not 
>>>>>>>>>> showing
>>>>>>>>>> my tables & databases, may be its running in an isolated mode... I'm 
>>>>>>>>>> just
>>>>>>>>>> getting empty list, so I attempted to do kerberos authentication to
>>>>>>>>>> workaround that issue, and bumped into this road block.
>>>>>>>>>>
>>>>>>>>>> Below is the configuration, I also tested my keytab file and its
>>>>>>>>>> working fine.
>>>>>>>>>>
>>>>>>>>>> *Kerberos test:*
>>>>>>>>>> $ kdestroy
>>>>>>>>>>
>>>>>>>>>> $ klist
>>>>>>>>>> *klist: No credentials cache found (ticket cache
>>>>>>>>>> FILE:/tmp/krb5cc_12027)*
>>>>>>>>>>
>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027*
>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab*
>>>>>>>>>> *Authenticated to Kerberos v5*
>>>>>>>>>>
>>>>>>>>>> $ klist
>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027*
>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>
>>>>>>>>>> *Valid starting     Expires            Service principal*
>>>>>>>>>> *08/30/16 20:25:19  08/31/16 06:25:19  krbtgt/<DOMAIN1>@<DOMAIN1>*
>>>>>>>>>> *        renew until 08/31/16 20:25:19*
>>>>>>>>>>
>>>>>>>>>> *zeppelin-env.sh*
>>>>>>>>>>
>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce
>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark
>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master yarn
>>>>>>>>>> --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>"
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> *Interpreter.json (Spark interpreter config)*
>>>>>>>>>> "2BUTFVN89": {
>>>>>>>>>>       "id": "2BUTFVN89",
>>>>>>>>>>       "name": "spark",
>>>>>>>>>>       "group": "spark",
>>>>>>>>>>       "properties": {
>>>>>>>>>>         "spark.cores.max": "",
>>>>>>>>>>         "zeppelin.spark.printREPLOutput": "true",
>>>>>>>>>>         "master": "yarn-client",
>>>>>>>>>>         "zeppelin.spark.maxResult": "1000",
>>>>>>>>>>         "zeppelin.dep.localrepo": "local-repo",
>>>>>>>>>>         "spark.app.name": "Zeppelin",
>>>>>>>>>>         "spark.executor.memory": "",
>>>>>>>>>>         "zeppelin.spark.importImplicit": "true",
>>>>>>>>>>         "zeppelin.spark.sql.stacktrace": "true",
>>>>>>>>>>         "zeppelin.spark.useHiveContext": "true",
>>>>>>>>>>         "zeppelin.interpreter.localRepo":
>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
>>>>>>>>>>         "zeppelin.spark.concurrentSQL": "false",
>>>>>>>>>>         "args": "",
>>>>>>>>>>         "zeppelin.pyspark.python": "python",
>>>>>>>>>>         "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.
>>>>>>>>>> x.alla.keytab",
>>>>>>>>>>         "spark.yarn.principal": "pradeep.x.alla",
>>>>>>>>>>         "zeppelin.dep.additionalRemoteRepository":
>>>>>>>>>> "spark-packages,http://dl.bintray.com/spark-packages/maven,false
>>>>>>>>>> ;"
>>>>>>>>>>       },
>>>>>>>>>>       "status": "READY",
>>>>>>>>>>       "interpreterGroup": [
>>>>>>>>>>         {
>>>>>>>>>>           "name": "spark",
>>>>>>>>>>           "class": "org.apache.zeppelin.spark.SparkInterpreter",
>>>>>>>>>>           "defaultInterpreter": true
>>>>>>>>>>         },
>>>>>>>>>>         {
>>>>>>>>>>           "name": "sql",
>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>> rkSqlInterpreter",
>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>         },
>>>>>>>>>>         {
>>>>>>>>>>           "name": "dep",
>>>>>>>>>>           "class": "org.apache.zeppelin.spark.DepInterpreter",
>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>         },
>>>>>>>>>>         {
>>>>>>>>>>           "name": "pyspark",
>>>>>>>>>>           "class": "org.apache.zeppelin.spark.PyS
>>>>>>>>>> parkInterpreter",
>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>         }
>>>>>>>>>>       ],
>>>>>>>>>>       "dependencies": [],
>>>>>>>>>>       "option": {
>>>>>>>>>>         "remote": true,
>>>>>>>>>>         "port": -1,
>>>>>>>>>>         "perNoteSession": false,
>>>>>>>>>>         "perNoteProcess": false,
>>>>>>>>>>         "isExistingProcess": false,
>>>>>>>>>>         "setPermission": false,
>>>>>>>>>>         "users": []
>>>>>>>>>>       }
>>>>>>>>>>     }
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> It looks like your kerberos configuration issue. Do you mind to
>>>>>>>>>>> share your configuration ? Or you can first try to run spark-shell 
>>>>>>>>>>> using
>>>>>>>>>>> spark.yarn.keytab & spark.yarn.principle to verify them.
>>>>>>>>>>>
>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured
>>>>>>>>>>>> kerberos authentication. For Kerberos I added "spark.yarn.keytab" &
>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client".  But 
>>>>>>>>>>>> I keep
>>>>>>>>>>>> getting this error whenever I use spark interpreter in the notebook
>>>>>>>>>>>>
>>>>>>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar
>>>>>>>>>>>> kInterpreter335845091
>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2}
>>>>>>>>>>>> Job.java[run]:189) - Job failed
>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L
>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L
>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>> DEFAULT
>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>> hentication.util.KerberosName.parseRules(KerberosName.java:321)
>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>> hentication.util.KerberosName.setRules(KerberosName.java:386)
>>>>>>>>>>>>         at org.apache.hadoop.security.Had
>>>>>>>>>>>> oopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>> rGroupInformation.initialize(UserGroupInformation.java:227)
>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>> rGroupInformation.ensureInitialized(UserGroupInformation.jav
>>>>>>>>>>>> a:214)
>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>> rGroupInformation.isAuthenticationMethodEnabled(UserGroupInf
>>>>>>>>>>>> ormation.java:275)
>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>> rGroupInformation.isSecurityEnabled(UserGroupInformation.jav
>>>>>>>>>>>> a:269)
>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>> rGroupInformation.loginUserFromKeytab(UserGroupInformation.j
>>>>>>>>>>>> ava:820)
>>>>>>>>>>>>         at org.apache.zeppelin.spark.Spar
>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539)
>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>> r.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93)
>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>> r.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteI
>>>>>>>>>>>> nterpreterServer.java:383)
>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.Job.run(Job.java:176)
>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139)
>>>>>>>>>>>>         at java.util.concurrent.Executors
>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511)
>>>>>>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>>>>>>> k.run(FutureTask.java:266)
>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledT
>>>>>>>>>>>> hreadPoolExecutor.java:180)
>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPo
>>>>>>>>>>>> olExecutor.java:293)
>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>>>>>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job
>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler
>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks,
>>>>>>>>>>>> Pradeep
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Best Regards
>>>>>>>>>>>
>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Best Regards
>>>>>>>>
>>>>>>>> Jeff Zhang
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Best Regards
>>>>>
>>>>> Jeff Zhang
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Best Regards
>>>
>>> Jeff Zhang
>>>
>>
>>
>
>
> --
> Best Regards
>
> Jeff Zhang
>

Reply via email to