Hello,

Do you copy your hive-site.xml in a proper position?

On Wed, Aug 31, 2016 at 3:52 PM, Pradeep Reddy <pradeepreddy.a...@gmail.com>
wrote:

> nothing obvious. I will stick to 0.5.6 build, until the latest builds
> stabilize.
>
> On Wed, Aug 31, 2016 at 1:39 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>
>> Then I guess maybe you are connecting to different database. Why not
>> using  'z.show(sql("databases"))' to display the databases ? Then you
>> will get a hint what's going on.
>>
>> On Wed, Aug 31, 2016 at 2:36 PM, Pradeep Reddy <
>> pradeepreddy.a...@gmail.com> wrote:
>>
>>> Yes...I didn't wish to show the names of the databases that we have in
>>> our data lake on that screen shot. so thats why I chose to display the
>>> count. The latest zeppelin build just shows 1 count which is "default"
>>> database.
>>>
>>> Thanks,
>>> Pradeep
>>>
>>> On Wed, Aug 31, 2016 at 1:33 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>
>>>> 116 is the databases count number. Do you expect a list of database ?
>>>> then you need to use 'z.show(sql("databases"))'
>>>>
>>>> On Wed, Aug 31, 2016 at 2:26 PM, Pradeep Reddy <
>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>
>>>>> Here it is Jeff
>>>>>
>>>>> [image: Inline image 1]
>>>>>
>>>>> On Wed, Aug 31, 2016 at 1:24 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>>>
>>>>>> Hi Pradeep,
>>>>>>
>>>>>> I don't see the databases on your screenshot (second one for 0.5.6).
>>>>>> I think the output is correct.
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, Aug 31, 2016 at 12:55 PM, Pradeep Reddy <
>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>
>>>>>>> Hi Jeff- I was able to make Kerberos work in 0.5.6 zeppelin build.
>>>>>>> It seems like Kerberos not working & spark not able to talk to the 
>>>>>>> shared
>>>>>>> hive meta store are defects in the current build.
>>>>>>>
>>>>>>> On Tue, Aug 30, 2016 at 11:09 PM, Pradeep Reddy <
>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi Jeff-
>>>>>>>>
>>>>>>>> I switched to local mode now, I'm able to summon the implicit
>>>>>>>> objects like sc, sqlContext etc., but it doesn't show my databases &
>>>>>>>> tables, just shows 1 database "default".
>>>>>>>>
>>>>>>>> Zeppelin Latest Build
>>>>>>>>
>>>>>>>> [image: Inline image 3]
>>>>>>>>
>>>>>>>> Zeppelin 0.5.6, running on the same machine, is able to show my
>>>>>>>> databases and tables.
>>>>>>>>
>>>>>>>> [image: Inline image 4]
>>>>>>>>
>>>>>>>> On Tue, Aug 30, 2016 at 8:20 PM, Jeff Zhang <zjf...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> > the spark interpreter is not showing my tables & databases, may
>>>>>>>>> be its running in an isolated mode... I'm just getting empty list, so 
>>>>>>>>> I
>>>>>>>>> attempted to do kerberos authentication to workaround that issue, and
>>>>>>>>> bumped into this road block.
>>>>>>>>>
>>>>>>>>> kerberos would not help here, actually I think it would make the
>>>>>>>>> problem more complicated.  You need to first check the log why you get
>>>>>>>>> empty list.
>>>>>>>>>
>>>>>>>>> On Wed, Aug 31, 2016 at 8:56 AM, Pradeep Reddy <
>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Jeff- I was also successfully able to run spark shell, after
>>>>>>>>>> running kdestroy, with the below command and was able to get to my 
>>>>>>>>>> hive
>>>>>>>>>> tables.
>>>>>>>>>>
>>>>>>>>>> spark-shell --conf spark.yarn.keytab=$HOME/pradeep.x.alla.keytab
>>>>>>>>>> --conf spark.yarn.principal=pradeep.x.alla --deploy-mode client
>>>>>>>>>> --master yarn --queue <QUEUE_NAME>
>>>>>>>>>>
>>>>>>>>>> On Tue, Aug 30, 2016 at 7:34 PM, Pradeep Reddy <
>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Thanks Jeff..I have always used zeppelin in local mode, but when
>>>>>>>>>>> I migrated from 0.5.6 to this version, the spark interpreter is not 
>>>>>>>>>>> showing
>>>>>>>>>>> my tables & databases, may be its running in an isolated mode... 
>>>>>>>>>>> I'm just
>>>>>>>>>>> getting empty list, so I attempted to do kerberos authentication to
>>>>>>>>>>> workaround that issue, and bumped into this road block.
>>>>>>>>>>>
>>>>>>>>>>> Below is the configuration, I also tested my keytab file and its
>>>>>>>>>>> working fine.
>>>>>>>>>>>
>>>>>>>>>>> *Kerberos test:*
>>>>>>>>>>> $ kdestroy
>>>>>>>>>>>
>>>>>>>>>>> $ klist
>>>>>>>>>>> *klist: No credentials cache found (ticket cache
>>>>>>>>>>> FILE:/tmp/krb5cc_12027)*
>>>>>>>>>>>
>>>>>>>>>>> $ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
>>>>>>>>>>> *Using default cache: /tmp/krb5cc_12027*
>>>>>>>>>>> *Using principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>> *Using keytab: pradeep_x_alla.keytab*
>>>>>>>>>>> *Authenticated to Kerberos v5*
>>>>>>>>>>>
>>>>>>>>>>> $ klist
>>>>>>>>>>> *Ticket cache: FILE:/tmp/krb5cc_12027*
>>>>>>>>>>> *Default principal: pradeep.x.alla@<DOMAIN1>*
>>>>>>>>>>>
>>>>>>>>>>> *Valid starting     Expires            Service principal*
>>>>>>>>>>> *08/30/16 20:25:19  08/31/16 06:25:19
>>>>>>>>>>>  krbtgt/<DOMAIN1>@<DOMAIN1>*
>>>>>>>>>>> *        renew until 08/31/16 20:25:19*
>>>>>>>>>>>
>>>>>>>>>>> *zeppelin-env.sh*
>>>>>>>>>>>
>>>>>>>>>>> export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
>>>>>>>>>>> export SPARK_HOME=/opt/cloudera/parce
>>>>>>>>>>> ls/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark
>>>>>>>>>>> export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master yarn
>>>>>>>>>>> --num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>"
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> *Interpreter.json (Spark interpreter config)*
>>>>>>>>>>> "2BUTFVN89": {
>>>>>>>>>>>       "id": "2BUTFVN89",
>>>>>>>>>>>       "name": "spark",
>>>>>>>>>>>       "group": "spark",
>>>>>>>>>>>       "properties": {
>>>>>>>>>>>         "spark.cores.max": "",
>>>>>>>>>>>         "zeppelin.spark.printREPLOutput": "true",
>>>>>>>>>>>         "master": "yarn-client",
>>>>>>>>>>>         "zeppelin.spark.maxResult": "1000",
>>>>>>>>>>>         "zeppelin.dep.localrepo": "local-repo",
>>>>>>>>>>>         "spark.app.name": "Zeppelin",
>>>>>>>>>>>         "spark.executor.memory": "",
>>>>>>>>>>>         "zeppelin.spark.importImplicit": "true",
>>>>>>>>>>>         "zeppelin.spark.sql.stacktrace": "true",
>>>>>>>>>>>         "zeppelin.spark.useHiveContext": "true",
>>>>>>>>>>>         "zeppelin.interpreter.localRepo":
>>>>>>>>>>> "/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
>>>>>>>>>>>         "zeppelin.spark.concurrentSQL": "false",
>>>>>>>>>>>         "args": "",
>>>>>>>>>>>         "zeppelin.pyspark.python": "python",
>>>>>>>>>>>         "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.
>>>>>>>>>>> x.alla.keytab",
>>>>>>>>>>>         "spark.yarn.principal": "pradeep.x.alla",
>>>>>>>>>>>         "zeppelin.dep.additionalRemoteRepository":
>>>>>>>>>>> "spark-packages,http://dl.bintray.com/spark-packages/maven,false
>>>>>>>>>>> ;"
>>>>>>>>>>>       },
>>>>>>>>>>>       "status": "READY",
>>>>>>>>>>>       "interpreterGroup": [
>>>>>>>>>>>         {
>>>>>>>>>>>           "name": "spark",
>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.SparkInterpreter",
>>>>>>>>>>>           "defaultInterpreter": true
>>>>>>>>>>>         },
>>>>>>>>>>>         {
>>>>>>>>>>>           "name": "sql",
>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.Spa
>>>>>>>>>>> rkSqlInterpreter",
>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>         },
>>>>>>>>>>>         {
>>>>>>>>>>>           "name": "dep",
>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.DepInterpreter",
>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>         },
>>>>>>>>>>>         {
>>>>>>>>>>>           "name": "pyspark",
>>>>>>>>>>>           "class": "org.apache.zeppelin.spark.PyS
>>>>>>>>>>> parkInterpreter",
>>>>>>>>>>>           "defaultInterpreter": false
>>>>>>>>>>>         }
>>>>>>>>>>>       ],
>>>>>>>>>>>       "dependencies": [],
>>>>>>>>>>>       "option": {
>>>>>>>>>>>         "remote": true,
>>>>>>>>>>>         "port": -1,
>>>>>>>>>>>         "perNoteSession": false,
>>>>>>>>>>>         "perNoteProcess": false,
>>>>>>>>>>>         "isExistingProcess": false,
>>>>>>>>>>>         "setPermission": false,
>>>>>>>>>>>         "users": []
>>>>>>>>>>>       }
>>>>>>>>>>>     }
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com>
>>>>>>>>>>> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> It looks like your kerberos configuration issue. Do you mind to
>>>>>>>>>>>> share your configuration ? Or you can first try to run spark-shell 
>>>>>>>>>>>> using
>>>>>>>>>>>> spark.yarn.keytab & spark.yarn.principle to verify them.
>>>>>>>>>>>>
>>>>>>>>>>>> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
>>>>>>>>>>>> pradeepreddy.a...@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> Hi- I recently built zeppelin from source and configured
>>>>>>>>>>>>> kerberos authentication. For Kerberos I added "spark.yarn.keytab" 
>>>>>>>>>>>>> &
>>>>>>>>>>>>> "spark.yarn.principal" and also set master to "yarn-client".  But 
>>>>>>>>>>>>> I keep
>>>>>>>>>>>>> getting this error whenever I use spark interpreter in the 
>>>>>>>>>>>>> notebook
>>>>>>>>>>>>>
>>>>>>>>>>>>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>> kInterpreter335845091
>>>>>>>>>>>>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2}
>>>>>>>>>>>>> Job.java[run]:189) - Job failed
>>>>>>>>>>>>> java.lang.IllegalArgumentException: Invalid rule: L
>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L
>>>>>>>>>>>>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>>>>>>>>>>>>> DEFAULT
>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>> hentication.util.KerberosName.parseRules(KerberosName.java:3
>>>>>>>>>>>>> 21)
>>>>>>>>>>>>>         at org.apache.hadoop.security.aut
>>>>>>>>>>>>> hentication.util.KerberosName.setRules(KerberosName.java:386)
>>>>>>>>>>>>>         at org.apache.hadoop.security.Had
>>>>>>>>>>>>> oopKerberosName.setConfiguration(HadoopKerberosName.java:75)
>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>> rGroupInformation.initialize(UserGroupInformation.java:227)
>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>> rGroupInformation.ensureInitialized(UserGroupInformation.jav
>>>>>>>>>>>>> a:214)
>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>> rGroupInformation.isAuthenticationMethodEnabled(UserGroupInf
>>>>>>>>>>>>> ormation.java:275)
>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>> rGroupInformation.isSecurityEnabled(UserGroupInformation.jav
>>>>>>>>>>>>> a:269)
>>>>>>>>>>>>>         at org.apache.hadoop.security.Use
>>>>>>>>>>>>> rGroupInformation.loginUserFromKeytab(UserGroupInformation.j
>>>>>>>>>>>>> ava:820)
>>>>>>>>>>>>>         at org.apache.zeppelin.spark.Spar
>>>>>>>>>>>>> kInterpreter.open(SparkInterpreter.java:539)
>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>> r.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>> r.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:93)
>>>>>>>>>>>>>         at org.apache.zeppelin.interprete
>>>>>>>>>>>>> r.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteI
>>>>>>>>>>>>> nterpreterServer.java:383)
>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.Job.run(Job.java:176)
>>>>>>>>>>>>>         at org.apache.zeppelin.scheduler.
>>>>>>>>>>>>> FIFOScheduler$1.run(FIFOScheduler.java:139)
>>>>>>>>>>>>>         at java.util.concurrent.Executors
>>>>>>>>>>>>> $RunnableAdapter.call(Executors.java:511)
>>>>>>>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>>>>>>>> k.run(FutureTask.java:266)
>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledT
>>>>>>>>>>>>> hreadPoolExecutor.java:180)
>>>>>>>>>>>>>         at java.util.concurrent.Scheduled
>>>>>>>>>>>>> ThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPo
>>>>>>>>>>>>> olExecutor.java:293)
>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>>>>>>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>>>>>>>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>>>>>>>>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>>>>>>>>>>>>> SchedulerFactory.java[jobFinished]:137) - Job
>>>>>>>>>>>>> remoteInterpretJob_1472593536728 finished by scheduler
>>>>>>>>>>>>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>> Pradeep
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --
>>>>>>>>>>>> Best Regards
>>>>>>>>>>>>
>>>>>>>>>>>> Jeff Zhang
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Best Regards
>>>>>>>>>
>>>>>>>>> Jeff Zhang
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best Regards
>>>>>>
>>>>>> Jeff Zhang
>>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>> --
>>>> Best Regards
>>>>
>>>> Jeff Zhang
>>>>
>>>
>>>
>>
>>
>> --
>> Best Regards
>>
>> Jeff Zhang
>>
>
>


-- 
이종열, Jongyoul Lee, 李宗烈
http://madeng.net

Reply via email to