1. Yes, they trigger the build processes accordingly. I just built it with
spark 1.3.0 and Hadoop 2.6.0
2. Local will build Zeppelin with Spark to trigger a local context with
local threads. With cluster mode, it will expect the master, hadoop conf
and spark assembly configurations to be specified in the zeppelin-env.sh

I've just started using Zeppelin myself, so experts please validate my
claim.

On Mon, Mar 23, 2015 at 10:48 AM, RJ Nowling <[email protected]> wrote:

> Hi all,
>
> I wanted to clarify how the Maven profiles work, especially with reference
> to cluster vs local mode.
>
> The documentation [1] specifies that local mode can be compiled like so:
>
> $ mvn install -DskipTests
>
> and cluster mode like so:
>
> $ mvn install -DskipTests -Dspark.version=1.1.0 -Dhadoop.version=2.2.0
>
> I may want to specify different Spark or Hadoop versions for local mode,
> though.  In the pom.xml file, I saw that there are different Spark and
> Hadoop profiles specified.
>
> Questions:
> 1. Are the Maven profiles mutually exclusive?  Can I specify -Pspark-1.3
> -Phadoop-2.4 ?
>
> 2. What differentiates the local vs cluster mode in the build?
>
> Thanks,
> RJ
>
> [1] https://zeppelin.incubator.apache.org/docs/install/install.html
>
>

Reply via email to