[ https://issues.apache.org/jira/browse/HIVE-9934?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14365037#comment-14365037 ]
Chao commented on HIVE-9934: ---------------------------- Found this in log: {noformat} 2015-03-17 04:33:32,728 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - 2015-03-17 04:33:32,725 INFO [pool-1-thread-1] client.RemoteDriver (RemoteDriver.java:call(371)) - Failed to run job 681ccfbe-bf9f-491c-a2e7-ad513f62d1dc 2015-03-17 04:33:32,728 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - java.util.concurrent.ExecutionException: Exception thrown by job 2015-03-17 04:33:32,728 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.JavaFutureActionWrapper.getImpl(FutureAction.scala:311) 2015-03-17 04:33:32,728 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.JavaFutureActionWrapper.get(FutureAction.scala:316) 2015-03-17 04:33:32,728 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:364) 2015-03-17 04:33:32,728 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.hive.spark.client.RemoteDriver$JobWrapper.call(RemoteDriver.java:317) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at java.util.concurrent.FutureTask.run(FutureTask.java:262) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at java.lang.Thread.run(Thread.java:744) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - Caused by: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3, ip-10-182-56-7.ec2.internal): java.io.FileNotFoundException: http://10.182.56.7:34690/jars/hive-exec-1.2.0-SNAPSHOT.jar 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at sun.net.www.protocol.http.HttpURLConnection.getInputStream(HttpURLConnection.java:1624) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:452) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.util.Utils$.fetchFile(Utils.scala:383) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$6.apply(Executor.scala:350) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$6.apply(Executor.scala:347) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:772) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:98) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:226) 2015-03-17 04:33:32,729 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.collection.mutable.HashMap.foreach(HashMap.scala:98) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:771) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:347) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:177) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at java.lang.Thread.run(Thread.java:744) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - Driver stacktrace: 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1214) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1203) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1202) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1202) 2015-03-17 04:33:32,730 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:696) 2015-03-17 04:33:32,731 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:696) 2015-03-17 04:33:32,731 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at scala.Option.foreach(Option.scala:236) 2015-03-17 04:33:32,731 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:696) 2015-03-17 04:33:32,731 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.scheduler.DAGSchedulerEventProcessActor$$anonfun$receive$2.applyOrElse(DAGScheduler.scala:1420) 2015-03-17 04:33:32,731 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at akka.actor.Actor$class.aroundReceive(Actor.scala:465) 2015-03-17 04:33:32,731 INFO [stdout-redir-1]: client.SparkClientImpl (SparkClientImpl.java:run(537)) - at org.apache.spark.scheduler.DAGSchedulerEventProcessActor.aroundReceive(DAGScheduler.scala:137 {noformat} I don't think this is relevant to my patch. > Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to > degrade the authentication mechanism to "none", allowing authentication > without password > ---------------------------------------------------------------------------------------------------------------------------------------------------------------------- > > Key: HIVE-9934 > URL: https://issues.apache.org/jira/browse/HIVE-9934 > Project: Hive > Issue Type: Bug > Components: Security > Affects Versions: 1.1.0 > Reporter: Chao > Assignee: Chao > Attachments: HIVE-9934.1.patch, HIVE-9934.2.patch, HIVE-9934.3.patch > > > Vulnerability in LdapAuthenticationProviderImpl enables HiveServer2 client to > degrade the authentication mechanism to "none", allowing authentication > without password. > See: http://docs.oracle.com/javase/jndi/tutorial/ldap/security/simple.html > “If you supply an empty string, an empty byte/char array, or null to the > Context.SECURITY_CREDENTIALS environment property, then the authentication > mechanism will be "none". This is because the LDAP requires the password to > be nonempty for simple authentication. The protocol automatically converts > the authentication to "none" if a password is not supplied.” > > Since the LdapAuthenticationProviderImpl.Authenticate method is relying on a > NamingException being thrown during creation of initial context, it does not > fail when the context result is an “unauthenticated” positive response from > the LDAP server. The end result is, one can authenticate with HiveServer2 > using the LdapAuthenticationProviderImpl with only a user name and an empty > password. -- This message was sent by Atlassian JIRA (v6.3.4#6332)