Hi All, I've built spark 1.5.0 with hadoop 2.6 with a fresh download :
build/mvn -Phadoop-2.6 -Dhadoop.version=2.6.0 -DskipTests clean package I try to run SparkR it launches the normal R without the spark addons: ./bin/sparkR --master local[*] Picked up JAVA_TOOL_OPTIONS: -javaagent:/usr/share/java/jayatanaag.jar R version 3.1.2 (2014-10-31) -- "Pumpkin Helmet" Copyright (C) 2014 The R Foundation for Statistical Computing Platform: x86_64-pc-linux-gnu (64-bit) R is free software and comes with ABSOLUTELY NO WARRANTY. You are welcome to redistribute it under certain conditions. Type 'license()' or 'licence()' for distribution details. Natural language support but running in an English locale R is a collaborative project with many contributors. Type 'contributors()' for more information and 'citation()' on how to cite R or R packages in publications. Type 'demo()' for some demos, 'help()' for on-line help, or 'help.start()' for an HTML browser interface to help. Type 'q()' to quit R. > With no "Welcome to SparkR" also > sc <- sparkR.init() Error: could not find function "sparkR.init" > sqlContext <- sparkRSQL.init(sc) Error: could not find function "sparkRSQL.init" > Spark-shell and other components are fine. Using scala 2.10.6 and Java 1.8_45, Ubuntu 15.0.4. Please can anyone give me any pointers? Is there a spark maven profile I need to enable? Thanks Devl