Thanks Jeff..I have always used zeppelin in local mode, but when I migrated
from 0.5.6 to this version, the spark interpreter is not showing my tables
& databases, may be its running in an isolated mode... I'm just getting
empty list, so I attempted to do kerberos authentication to workaround that
issue, and bumped into this road block.

Below is the configuration, I also tested my keytab file and its working
fine.

*Kerberos test:*
$ kdestroy

$ klist
*klist: No credentials cache found (ticket cache FILE:/tmp/krb5cc_12027)*

$ kinit -kt pradeep_x_alla.keytab -V pradeep.x.alla
*Using default cache: /tmp/krb5cc_12027*
*Using principal: pradeep.x.alla@<DOMAIN1>*
*Using keytab: pradeep_x_alla.keytab*
*Authenticated to Kerberos v5*

$ klist
*Ticket cache: FILE:/tmp/krb5cc_12027*
*Default principal: pradeep.x.alla@<DOMAIN1>*

*Valid starting     Expires            Service principal*
*08/30/16 20:25:19  08/31/16 06:25:19  krbtgt/<DOMAIN1>@<DOMAIN1>*
*        renew until 08/31/16 20:25:19*

*zeppelin-env.sh*

export HADOOP_CONF_DIR=/etc/hadoop/conf:/etc/hive/conf
export SPARK_HOME=/opt/cloudera/parcels/CDH-5.5.1-1.cdh5.5.1.p0.11/lib/spark
export SPARK_SUBMIT_OPTIONS="--deploy-mode client --master yarn
--num-executors 2 --executor-memory 2g --queue <QUEUE_NAME>"


*Interpreter.json (Spark interpreter config)*
"2BUTFVN89": {
      "id": "2BUTFVN89",
      "name": "spark",
      "group": "spark",
      "properties": {
        "spark.cores.max": "",
        "zeppelin.spark.printREPLOutput": "true",
        "master": "yarn-client",
        "zeppelin.spark.maxResult": "1000",
        "zeppelin.dep.localrepo": "local-repo",
        "spark.app.name": "Zeppelin",
        "spark.executor.memory": "",
        "zeppelin.spark.importImplicit": "true",
        "zeppelin.spark.sql.stacktrace": "true",
        "zeppelin.spark.useHiveContext": "true",
        "zeppelin.interpreter.localRepo":
"/home/pradeep.x.alla/zeppelin/local-repo/2BUTFVN89",
        "zeppelin.spark.concurrentSQL": "false",
        "args": "",
        "zeppelin.pyspark.python": "python",
        "spark.yarn.keytab": "/home/pradeep.x.alla/pradeep.x.alla.keytab",
        "spark.yarn.principal": "pradeep.x.alla",
        "zeppelin.dep.additionalRemoteRepository": "spark-packages,
http://dl.bintray.com/spark-packages/maven,false;";
      },
      "status": "READY",
      "interpreterGroup": [
        {
          "name": "spark",
          "class": "org.apache.zeppelin.spark.SparkInterpreter",
          "defaultInterpreter": true
        },
        {
          "name": "sql",
          "class": "org.apache.zeppelin.spark.SparkSqlInterpreter",
          "defaultInterpreter": false
        },
        {
          "name": "dep",
          "class": "org.apache.zeppelin.spark.DepInterpreter",
          "defaultInterpreter": false
        },
        {
          "name": "pyspark",
          "class": "org.apache.zeppelin.spark.PySparkInterpreter",
          "defaultInterpreter": false
        }
      ],
      "dependencies": [],
      "option": {
        "remote": true,
        "port": -1,
        "perNoteSession": false,
        "perNoteProcess": false,
        "isExistingProcess": false,
        "setPermission": false,
        "users": []
      }
    }










On Tue, Aug 30, 2016 at 6:52 PM, Jeff Zhang <zjf...@gmail.com> wrote:

> It looks like your kerberos configuration issue. Do you mind to share your
> configuration ? Or you can first try to run spark-shell using
> spark.yarn.keytab & spark.yarn.principle to verify them.
>
> On Wed, Aug 31, 2016 at 6:12 AM, Pradeep Reddy <
> pradeepreddy.a...@gmail.com> wrote:
>
>> Hi- I recently built zeppelin from source and configured kerberos
>> authentication. For Kerberos I added "spark.yarn.keytab" &
>> "spark.yarn.principal" and also set master to "yarn-client".  But I keep
>> getting this error whenever I use spark interpreter in the notebook
>>
>> 3536728 started by scheduler org.apache.zeppelin.spark.Spar
>> kInterpreter335845091
>> ERROR [2016-08-30 17:45:37,237] ({pool-2-thread-2} Job.java[run]:189) -
>> Job failed
>> java.lang.IllegalArgumentException: Invalid rule: L
>> RULE:[2:$1@$0](.*@\Q<DOMAIN1>.COM\E$)s/@\Q<DOMAIN1>\E$//L
>> RULE:[1:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>> RULE:[2:$1@$0](.*@\Q<DOMAIN2>\E$)s/@\Q<DOMAIN2>\E$//L
>> DEFAULT
>>         at org.apache.hadoop.security.authentication.util.KerberosName.
>> parseRules(KerberosName.java:321)
>>         at org.apache.hadoop.security.authentication.util.KerberosName.
>> setRules(KerberosName.java:386)
>>         at org.apache.hadoop.security.HadoopKerberosName.setConfigurati
>> on(HadoopKerberosName.java:75)
>>         at org.apache.hadoop.security.UserGroupInformation.initialize(U
>> serGroupInformation.java:227)
>>         at org.apache.hadoop.security.UserGroupInformation.ensureInitia
>> lized(UserGroupInformation.java:214)
>>         at org.apache.hadoop.security.UserGroupInformation.isAuthentica
>> tionMethodEnabled(UserGroupInformation.java:275)
>>         at org.apache.hadoop.security.UserGroupInformation.isSecurityEn
>> abled(UserGroupInformation.java:269)
>>         at org.apache.hadoop.security.UserGroupInformation.loginUserFro
>> mKeytab(UserGroupInformation.java:820)
>>         at org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpr
>> eter.java:539)
>>         at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(Laz
>> yOpenInterpreter.java:69)
>>         at org.apache.zeppelin.interpreter.LazyOpenInterpreter.
>> interpret(LazyOpenInterpreter.java:93)
>>         at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServ
>> er$InterpretJob.jobRun(RemoteInterpreterServer.java:383)
>>         at org.apache.zeppelin.scheduler.Job.run(Job.java:176)
>>         at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOSchedu
>> ler.java:139)
>>         at java.util.concurrent.Executors$RunnableAdapter.call(
>> Executors.java:511)
>>         at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>>         at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFu
>> tureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>>         at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFu
>> tureTask.run(ScheduledThreadPoolExecutor.java:293)
>>         at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPool
>> Executor.java:1142)
>>         at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoo
>> lExecutor.java:617)
>>         at java.lang.Thread.run(Thread.java:745)
>>  INFO [2016-08-30 17:45:37,247] ({pool-2-thread-2}
>> SchedulerFactory.java[jobFinished]:137) - Job
>> remoteInterpretJob_1472593536728 finished by scheduler
>> org.apache.zeppelin.spark.SparkInterpreter335845091
>>
>> Thanks,
>> Pradeep
>>
>>
>
>
> --
> Best Regards
>
> Jeff Zhang
>

Reply via email to