Can you check if the file
`~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit` exists ? The
error message seems to indicate it is trying to pick up Spark from
that location and can't seem to find Spark installed there.

Thanks
Shivaram

On Thu, Aug 20, 2015 at 3:30 PM, Deborah Siegel
<deborah.sie...@gmail.com> wrote:
> Hello,
>
> I have previously successfully run SparkR in RStudio, with:
>
>>Sys.setenv(SPARK_HOME="~/software/spark-1.4.1-bin-hadoop2.4")
>>.libPaths(c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib"), .libPaths()))
>>library(SparkR)
>>sc <- sparkR.init(master="local[2]",appName="SparkR-example")
>
>
> Then I tried putting some of it into an .Rprofile. It seemed to work to load
> the paths and SparkR, but I got an error when trying to create the sc. I
> then removed my .Rprofile, as well as .rstudio-desktop. However, I still
> cannot create the sc. Here is the error
>
>> sc <- sparkR.init(master="local[2]",appName="SparkR-example")
> Launching java with spark-submit command
> ~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit   sparkr-shell
> /var/folders/p7/k1bpgmx93yd6pjq7dzf35gk80000gn/T//RtmpOitA28/backend_port23377046db
> sh: ~/software/spark-1.4.1-bin-hadoop2.4/bin/spark-submit: No such file or
> directory
> Error in sparkR.init(master = "local[2]", appName = "SparkR-example") :
> JVM is not ready after 10 seconds
>
> I suspected there was an incomplete process or something. I checked for any
> running R or Java processes and there were none. Has someone seen this type
> of error? I have the same error in both RStudio and in R shell (but not
> sparkR wrapper).
>
> Thanks,
> Deb
>
>

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to