Looks like you didn't specify sparkr profile when building.

Cheers

On Sat, Sep 19, 2015 at 12:30 PM, Devl Devel <devl.developm...@gmail.com>
wrote:

> Hi All,
>
> I've built spark 1.5.0 with hadoop 2.6 with a fresh download :
>
> build/mvn  -Phadoop-2.6 -Dhadoop.version=2.6.0 -DskipTests clean package
>
> I try to run SparkR it launches the normal R without the spark addons:
>
> ./bin/sparkR --master local[*]
> Picked up JAVA_TOOL_OPTIONS: -javaagent:/usr/share/java/jayatanaag.jar
>
> R version 3.1.2 (2014-10-31) -- "Pumpkin Helmet"
> Copyright (C) 2014 The R Foundation for Statistical Computing
> Platform: x86_64-pc-linux-gnu (64-bit)
>
> R is free software and comes with ABSOLUTELY NO WARRANTY.
> You are welcome to redistribute it under certain conditions.
> Type 'license()' or 'licence()' for distribution details.
>
>   Natural language support but running in an English locale
>
> R is a collaborative project with many contributors.
> Type 'contributors()' for more information and
> 'citation()' on how to cite R or R packages in publications.
>
> Type 'demo()' for some demos, 'help()' for on-line help, or
> 'help.start()' for an HTML browser interface to help.
> Type 'q()' to quit R.
>
> >
>
> With no "Welcome to SparkR"
>
> also
>
> > sc <- sparkR.init()
> Error: could not find function "sparkR.init"
> > sqlContext <- sparkRSQL.init(sc)
> Error: could not find function "sparkRSQL.init"
> >
>
> Spark-shell and other components are fine. Using scala 2.10.6 and Java
> 1.8_45, Ubuntu 15.0.4. Please can anyone give me any pointers? Is there a
> spark maven profile I need to enable?
>
> Thanks
> Devl
>

Reply via email to