Can you check if the file `~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit` exists ? The error message seems to indicate it is trying to pick up Spark from that location and can't seem to find Spark installed there.
Thanks Shivaram On Thu, Aug 20, 2015 at 3:30 PM, Deborah Siegel <deborah.sie...@gmail.com> wrote: > Hello, > > I have previously successfully run SparkR in RStudio, with: > >>Sys.setenv(SPARK_HOME="~/software/spark-1.4.1-bin-hadoop2.4") >>.libPaths(c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib"), .libPaths())) >>library(SparkR) >>sc <- sparkR.init(master="local[2]",appName="SparkR-example") > > > Then I tried putting some of it into an .Rprofile. It seemed to work to load > the paths and SparkR, but I got an error when trying to create the sc. I > then removed my .Rprofile, as well as .rstudio-desktop. However, I still > cannot create the sc. Here is the error > >> sc <- sparkR.init(master="local[2]",appName="SparkR-example") > Launching java with spark-submit command > ~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit sparkr-shell > /var/folders/p7/k1bpgmx93yd6pjq7dzf35gk80000gn/T//RtmpOitA28/backend_port23377046db > sh: ~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit: No such file or > directory > Error in sparkR.init(master = "local[2]", appName = "SparkR-example") : > JVM is not ready after 10 seconds > > I suspected there was an incomplete process or something. I checked for any > running R or Java processes and there were none. Has someone seen this type > of error? I have the same error in both RStudio and in R shell (but not > sparkR wrapper). > > Thanks, > Deb > > --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org