Thanks a lot,

That worked great!

Thanks,
Lars


On Tue, Jun 3, 2014 at 12:17 PM, Yin Huai <huaiyin....@gmail.com> wrote:

> Hello Lars,
>
> Can you check the value of "hive.security.authenticator.manager" in
> hive-site.xml? I guess the value is
> "org.apache.hadoop.hive.ql.security.ProxyUserAuthenticator". This class was
> introduced in hive 0.13, but Spark SQL is based on hive 0.12 right now. Can
> you change the value of
> "hive.security.authenticator.manager" to
> "org.apache.hadoop.hive.ql.security.HadoopDefaultAuthenticator" and try
> again?
>
> Thanks,
>
> Yin
>
>
> On Tue, Jun 3, 2014 at 10:44 AM, Lars Selsaas <
> lars.sels...@thinkbiganalytics.com> wrote:
>
>> Hi,
>>
>> I've installed Spark 1.0.0 on a HDP 2.1
>>
>> I moved the hive-site.xml file into the conf directory for Spark in an
>> attempt to connect Spark with my existing Hive.
>>
>> Below is the full log from me starting Spark till I get the error. It
>> seems to be building the assembly with hive so that part should be good. Is
>> the Hive version HDP 2.1 not support or any other hints to what I'm doing
>> wrong?
>>
>>
>>  [root@sandbox bin]# ./spark-shell
>>
>> Spark assembly has been built with Hive, including Datanucleus jars on
>> classpath
>>
>> 14/06/03 10:30:21 INFO SecurityManager: Using Spark's default log4j
>> profile: org/apache/spark/log4j-defaults.properties
>>
>> 14/06/03 10:30:21 INFO SecurityManager: Changing view acls to: root
>>
>> 14/06/03 10:30:21 INFO SecurityManager: SecurityManager: authentication
>> disabled; ui acls disabled; users with view permissions: Set(root)
>>
>> 14/06/03 10:30:21 INFO HttpServer: Starting HTTP Server
>>
>> Welcome to
>>
>>       ____              __
>>
>>      / __/__  ___ _____/ /__
>>
>>     _\ \/ _ \/ _ `/ __/  '_/
>>
>>    /___/ .__/\_,_/_/ /_/\_\   version 1.0.0
>>
>>       /_/
>>
>>
>> Using Scala version 2.10.4 (Java HotSpot(TM) 64-Bit Server VM, Java
>> 1.7.0_45)
>>
>> Type in expressions to have them evaluated.
>>
>> Type :help for more information.
>>
>> 14/06/03 10:30:28 INFO SecurityManager: Changing view acls to: root
>>
>> 14/06/03 10:30:28 INFO SecurityManager: SecurityManager: authentication
>> disabled; ui acls disabled; users with view permissions: Set(root)
>>
>> 14/06/03 10:30:29 INFO Slf4jLogger: Slf4jLogger started
>>
>> 14/06/03 10:30:29 INFO Remoting: Starting remoting
>>
>> 14/06/03 10:30:29 INFO Remoting: Remoting started; listening on addresses
>> :[akka.tcp://sp...@sandbox.hortonworks.com:58176]
>>
>> 14/06/03 10:30:29 INFO Remoting: Remoting now listens on addresses:
>> [akka.tcp://sp...@sandbox.hortonworks.com:58176]
>>
>> 14/06/03 10:30:29 INFO SparkEnv: Registering MapOutputTracker
>>
>> 14/06/03 10:30:29 INFO SparkEnv: Registering BlockManagerMaster
>>
>> 14/06/03 10:30:29 INFO DiskBlockManager: Created local directory at
>> /tmp/spark-local-20140603103029-75e9
>>
>> 14/06/03 10:30:29 INFO MemoryStore: MemoryStore started with capacity
>> 294.9 MB.
>>
>> 14/06/03 10:30:29 INFO ConnectionManager: Bound socket to port 33645 with
>> id = ConnectionManagerId(sandbox.hortonworks.com,33645)
>>
>> 14/06/03 10:30:29 INFO BlockManagerMaster: Trying to register BlockManager
>>
>> 14/06/03 10:30:29 INFO BlockManagerInfo: Registering block manager
>> sandbox.hortonworks.com:33645 with 294.9 MB RAM
>>
>> 14/06/03 10:30:29 INFO BlockManagerMaster: Registered BlockManager
>>
>> 14/06/03 10:30:29 INFO HttpServer: Starting HTTP Server
>>
>> 14/06/03 10:30:29 INFO HttpBroadcast: Broadcast server started at
>> http://10.0.2.15:43286
>>
>> 14/06/03 10:30:29 INFO HttpFileServer: HTTP File server directory is
>> /tmp/spark-fc0d099f-16ca-48e3-8c49-6f809bf67449
>>
>> 14/06/03 10:30:29 INFO HttpServer: Starting HTTP Server
>>
>> 14/06/03 10:30:30 INFO SparkUI: Started SparkUI at
>> http://sandbox.hortonworks.com:4040
>>
>> 14/06/03 10:30:30 WARN NativeCodeLoader: Unable to load native-hadoop
>> library for your platform... using builtin-java classes where applicable
>>
>> 14/06/03 10:30:31 INFO Executor: Using REPL class URI:
>> http://10.0.2.15:34588
>>
>> 14/06/03 10:30:31 INFO SparkILoop: Created spark context..
>>
>> Spark context available as sc.
>>
>>
>> scala> val hiveContext = new org.apache.spark.sql.hive.HiveContext(sc)
>>
>> 14/06/03 10:30:53 INFO deprecation: mapred.input.dir.recursive is
>> deprecated. Instead, use mapreduce.input.fileinputformat.input.dir.recursive
>>
>> 14/06/03 10:30:53 INFO deprecation: mapred.max.split.size is deprecated.
>> Instead, use mapreduce.input.fileinputformat.split.maxsize
>>
>> 14/06/03 10:30:53 INFO deprecation: mapred.min.split.size is deprecated.
>> Instead, use mapreduce.input.fileinputformat.split.minsize
>>
>> 14/06/03 10:30:53 INFO deprecation: mapred.min.split.size.per.rack is
>> deprecated. Instead, use
>> mapreduce.input.fileinputformat.split.minsize.per.rack
>>
>> 14/06/03 10:30:53 INFO deprecation: mapred.min.split.size.per.node is
>> deprecated. Instead, use
>> mapreduce.input.fileinputformat.split.minsize.per.node
>>
>> 14/06/03 10:30:53 INFO deprecation: mapred.reduce.tasks is deprecated.
>> Instead, use mapreduce.job.reduces
>>
>> 14/06/03 10:30:53 INFO deprecation:
>> mapred.reduce.tasks.speculative.execution is deprecated. Instead, use
>> mapreduce.reduce.speculative
>>
>> java.lang.RuntimeException:
>> org.apache.hadoop.hive.ql.metadata.HiveException:
>> java.lang.ClassNotFoundException:
>> org.apache.hadoop.hive.ql.security.ProxyUserAuthenticator
>>
>> at
>> org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:286)
>>
>> at org.apache.spark.sql.hive.HiveContext.<init>(HiveContext.scala:166)
>>
>> at $iwC$$iwC$$iwC$$iwC.<init>(<console>:12)
>>
>> at $iwC$$iwC$$iwC.<init>(<console>:17)
>>
>> at $iwC$$iwC.<init>(<console>:19)
>>
>> at $iwC.<init>(<console>:21)
>>
>> at <init>(<console>:23)
>>
>> at .<init>(<console>:27)
>>
>> at .<clinit>(<console>)
>>
>> at .<init>(<console>:7)
>>
>> at .<clinit>(<console>)
>>
>> at $print(<console>)
>>
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>
>> at java.lang.reflect.Method.invoke(Method.java:606)
>>
>> at
>> org.apache.spark.repl.SparkIMain$ReadEvalPrint.call(SparkIMain.scala:788)
>>
>> at
>> org.apache.spark.repl.SparkIMain$Request.loadAndRun(SparkIMain.scala:1056)
>>
>> at org.apache.spark.repl.SparkIMain.loadAndRunReq$1(SparkIMain.scala:614)
>>
>> at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:645)
>>
>> at org.apache.spark.repl.SparkIMain.interpret(SparkIMain.scala:609)
>>
>> at
>> org.apache.spark.repl.SparkILoop.reallyInterpret$1(SparkILoop.scala:796)
>>
>> at
>> org.apache.spark.repl.SparkILoop.interpretStartingWith(SparkILoop.scala:841)
>>
>> at org.apache.spark.repl.SparkILoop.command(SparkILoop.scala:753)
>>
>> at org.apache.spark.repl.SparkILoop.processLine$1(SparkILoop.scala:601)
>>
>> at org.apache.spark.repl.SparkILoop.innerLoop$1(SparkILoop.scala:608)
>>
>> at org.apache.spark.repl.SparkILoop.loop(SparkILoop.scala:611)
>>
>> at
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply$mcZ$sp(SparkILoop.scala:936)
>>
>> at
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884)
>>
>> at
>> org.apache.spark.repl.SparkILoop$$anonfun$process$1.apply(SparkILoop.scala:884)
>>
>> at
>> scala.tools.nsc.util.ScalaClassLoader$.savingContextLoader(ScalaClassLoader.scala:135)
>>
>> at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:884)
>>
>> at org.apache.spark.repl.SparkILoop.process(SparkILoop.scala:982)
>>
>> at org.apache.spark.repl.Main$.main(Main.scala:31)
>>
>> at org.apache.spark.repl.Main.main(Main.scala)
>>
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
>>
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>
>> at java.lang.reflect.Method.invoke(Method.java:606)
>>
>> at org.apache.spark.deploy.SparkSubmit$.launch(SparkSubmit.scala:292)
>>
>> at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:55)
>>
>> at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
>>
>> Caused by: org.apache.hadoop.hive.ql.metadata.HiveException:
>> java.lang.ClassNotFoundException:
>> org.apache.hadoop.hive.ql.security.ProxyUserAuthenticator
>>
>> at
>> org.apache.hadoop.hive.ql.metadata.HiveUtils.getAuthenticator(HiveUtils.java:368)
>>
>> at
>> org.apache.hadoop.hive.ql.session.SessionState.start(SessionState.java:278)
>>
>> ... 41 more
>>
>> Caused by: java.lang.ClassNotFoundException:
>> org.apache.hadoop.hive.ql.security.ProxyUserAuthenticator
>>
>> at
>> scala.tools.nsc.interpreter.AbstractFileClassLoader.findClass(AbstractFileClassLoader.scala:83)
>>
>> at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
>>
>> at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
>>
>> at java.lang.Class.forName0(Native Method)
>>
>> at java.lang.Class.forName(Class.java:270)
>>
>> at
>> org.apache.hadoop.hive.ql.metadata.HiveUtils.getAuthenticator(HiveUtils.java:361)
>>
>> ... 42 more
>>
>> --
>>
>> Lars Selsaas
>>
>> Data Engineer
>>
>> Think Big Analytics <http://thinkbiganalytics.com>
>>
>> lars.sels...@thinkbiganalytics.com
>>
>> 650-537-5321
>>
>>
>


-- 

Lars Selsaas

Data Engineer

Think Big Analytics <http://thinkbiganalytics.com>

lars.sels...@thinkbiganalytics.com

650-537-5321

Reply via email to