Right, it's not currently dependence in Spark. If we already mention it, is it possible to make it part of current dependence, but only for Hadoop profiles 2.4 and up? This will solve a lot of headache to those who use Spark + OpenStack Swift and need every time to manually edit pom.xml to add dependence of it.
From: Ted Yu <yuzhih...@gmail.com> To: Josh Rosen <joshro...@databricks.com> Cc: Steve Loughran <ste...@hortonworks.com>, Gil Vernik/Haifa/IBM@IBMIL, Dev <dev@spark.apache.org> Date: 15/07/2015 18:28 Subject: Re: problems with build of latest the master If I understand correctly, hadoop-openstack is not currently dependence in Spark. On Jul 15, 2015, at 8:21 AM, Josh Rosen <joshro...@databricks.com> wrote: We may be able to fix this from the Spark side by adding appropriate exclusions in our Hadoop dependencies, right? If possible, I think that we should do this. On Wed, Jul 15, 2015 at 7:10 AM, Ted Yu <yuzhih...@gmail.com> wrote: I attached a patch for HADOOP-12235 BTW openstack was not mentioned in the first email from Gil. My email and Gil's second email were sent around the same moment. Cheers On Wed, Jul 15, 2015 at 2:06 AM, Steve Loughran <ste...@hortonworks.com> wrote: On 14 Jul 2015, at 12:22, Ted Yu <yuzhih...@gmail.com> wrote: Looking at Jenkins, master branch compiles. Can you try the following command ? mvn -Phive -Phadoop-2.6 -DskipTests clean package What version of Java are you using ? Ted, Giles has stuck in hadoop-openstack, it's that which is creating the problem Giles, I don't know why hadoop-openstack has a mockito dependency as it should be test time only Looking at the POM it's tag in hadoop-2.7 tis scoped to compile, which <dependency> <groupId>org.mockito</groupId> <artifactId>mockito-all</artifactId> <scope>compile</scope> </dependency> it should be "provided", shouldn't it? Created https://issues.apache.org/jira/browse/HADOOP-12235 : if someone supplies a patch I'll get it in. -steve