Thanks Junaid,

And I have found how to check the Spark version installed , a simple
sc.version !

Victor

[image: datastax_logo.png] <http://www.datastax.com/>

Victor Coustenoble

Solutions Engineer | +33 6 70 23 68 82 | victor.cousteno...@datastax.com


[image: linkedin.png] <https://www.linkedin.com/in/victorcoustenoble> [image:
twitter.png] <https://twitter.com/vizanalytics> [image: g+.png]
<https://plus.google.com/+Datastax/about>
<http://feeds.feedburner.com/datastax> <https://github.com/datastax/>


<http://www.datastax.com/gartner-magic-quadrant-odbms>

On Tue, Jan 12, 2016 at 9:01 AM, Junaid Shaikh J <
junaid.j.sha...@ericsson.com> wrote:

> Answers to your questions:
>
> You do not explicitly need to install spark and hadoop before building
> Zeppelin. It can be embedded, and while building Zeppelin, you can specify
> spark and hadoop versions you need. Yes SPARK-HOME can be used to point to
> the external spark and hadoop installation.
>
> To build you may specify any spark and hadoop versions. For example:
>
> mvn clean package -Pspark-1.6 -Phadoop-2.4 -Pyarn -Ppyspark
>
>
> For Cassandra integration, build using the option: -Pcassandra-spark-*xx*
>
> /Junaid
>
>
> On 12 Jan 2016, at 00:45, Victor Coustenoble <
> victor.cousteno...@datastax.com> wrote:
>
> Few questions on build options:
>
> - Spark and Hadoop are needed to build Zeppelin for client binaries
> library, right ? and with SPARK_HOME set, another client library version
> can be used, right ?
> - If I don't specify any options, are there default Spark and Hadoop
> versions embedded ? I don't find it in pom file
> - Possible to check Spark client library version used in a notebook ?
> - If I only specify the cassandra-spark option, I get at the same time the
> corresponding Spark client library version ?
>
> Thanks
> Victor
>
>
>

Reply via email to