Have you looked at http://www.scala-sbt.org/0.13/tutorial/Multi-Project.html
?

Cheers

On Wed, Apr 29, 2015 at 2:45 PM, Dan Dong <dongda...@gmail.com> wrote:

> Hi,
>   Following the Quick Start guide:
> https://spark.apache.org/docs/latest/quick-start.html
>
> I could compile and run a Spark program successfully, now my question is
> how to
> compile multiple programs with sbt in a bunch. E.g, two programs as:
>
>
> ./src
> ./src/main
> ./src/main/scala
> ./src/main/scala/SimpleApp_A.scala
> ./src/main/scala/SimpleApp_B.scala
>
> Hopefully with "sbt package", I will get two .jar files for each of the
> source program, then I can run them separately in Spark. I tried to create
> two .sbt files for each program, but found only one .jar file is created.
>
> ./simpleA.sbt
> name := "Simple Project A"
> version := "1.0"
> scalaVersion := "2.10.4"
> libraryDependencies += "org.apache.spark" %% "spark-core" % "1.3.1"
>
> ./simpleB.sbt
> name := "Simple Project B"
> version := "1.0"
> scalaVersion := "2.10.4"
> libraryDependencies += "org.apache.spark" %% "spark-core" % "1.3.1"
>
>   Does anybody know how to do it?
>
> Cheers,
> Dan
>
>

Reply via email to