You're right Patrick! 

Just had a chat with sbt-pack creator and indeed dependencies with classifiers 
are ignored to avoid problems with dirty cache...

Should be fixed in next version of the plugin.

Cheers

Pierre 

Message sent from a mobile device - excuse typos and abbreviations 

> Le 1 juin 2014 à 20:04, Patrick Wendell <pwend...@gmail.com> a écrit :
> 
> https://github.com/apache/spark/blob/master/project/SparkBuild.scala#L350
> 
>> On Sun, Jun 1, 2014 at 11:03 AM, Patrick Wendell <pwend...@gmail.com> wrote:
>> One potential issue here is that mesos is using classifiers now to
>> publish there jars. It might be that sbt-pack has trouble with
>> dependencies that are published using classifiers. I'm pretty sure
>> mesos is the only dependency in Spark that is using classifiers, so
>> that's why I mention it.
>> 
>> On Sun, Jun 1, 2014 at 2:34 AM, Pierre B
>> <pierre.borckm...@realimpactanalytics.com> wrote:
>>> Hi all!
>>> 
>>> We'be been using the sbt-pack sbt plugin
>>> (https://github.com/xerial/sbt-pack) for building our standalone Spark
>>> application for a while now. Until version 1.0.0, that worked nicely.
>>> 
>>> For those who don't know the sbt-pack plugin, it basically copies all the
>>> dependencies JARs from your local ivy/maven cache to a your target folder
>>> (in target/pack/lib), and creates launch scripts (in target/pack/bin) for
>>> your application (notably setting all these jars on the classpath).
>>> 
>>> Now, since Spark 1.0.0 was released, we are encountering a weird error where
>>> running our project with "sbt run" is fine but running our app with the
>>> launch scripts generated by sbt-pack fails.
>>> 
>>> After a (quite painful) investigation, it turns out some JARs are NOT copied
>>> from the local ivy2 cache to the lib folder. I noticed that all the missing
>>> jars contain "shaded" in their file name (but all not all jars with such
>>> name are missing).
>>> One of the missing JARs is explicitly from the Spark definition
>>> (SparkBuild.scala, line 350): ``mesos-0.18.1-shaded-protobuf.jar``.
>>> 
>>> This file is clearly present in my local ivy cache, but is not copied by
>>> sbt-pack.
>>> 
>>> Is there an evident reason for that?
>>> 
>>> I don't know much about the shading mechanism, maybe I'm missing something
>>> here?
>>> 
>>> 
>>> Any help would be appreciated!
>>> 
>>> Cheers
>>> 
>>> Pierre
>>> 
>>> 
>>> 
>>> --
>>> View this message in context: 
>>> http://apache-spark-user-list.1001560.n3.nabble.com/Using-sbt-pack-with-Spark-1-0-0-tp6649.html
>>> Sent from the Apache Spark User List mailing list archive at Nabble.com.

Reply via email to