Hello, I have been trying to use the SparkR interpreter on Spark 1.6.1 and run into a problem. The SparkR interpreter works well in stand alone configuration with master set to local[*] and SPARK_HOME undefined.
With master set to yarn-client (with hadoop 2.6.4) and SPARK_HOME pointing to the installed location of Spark 1.6.1, the spark and pyspark interpreters work well. However the R/SparkR interpreters always fail with "sparkr is not responding" error. This is with the master branch of the code from GitHub and here is the full output: org.apache.zeppelin.interpreter.InterpreterException: sparkr is not responding R version 3.2.2 (2015-08-14) -- "Fire Safety" Copyright (C) 2015 The R Foundation for Statistical Computing Platform: x86_64-w64-mingw32/x64 (64-bit) R is free software and comes with ABSOLUTELY NO WARRANTY. You are welcome to redistribute it under certain conditions. Type 'license()' or 'licence()' for distribution details. Natural language support but running in an English locale R is a collaborative project with many contributors. Type 'contributors()' for more information and 'citation()' on how to cite R or R packages in publications. Type 'demo()' for some demos, 'help()' for on-line help, or 'help.start()' for an HTML browser interface to help. Type 'q()' to quit R. > # > # Licensed to the Apache Software Foundation (ASF) under one > # or more contributor license agreements. See the NOTICE file > # distributed with this work for additional information > # regarding copyright ownership. The ASF licenses this file > # to you under the Apache License, Version 2.0 (the > # "License"); you may not use this file except in compliance > # with the License. You may obtain a copy of the License at > # > # http://www.apache.org/licenses/LICENSE-2.0 > # > # Unless required by applicable law or agreed to in writing, software > # distributed under the License is distributed on an "AS IS" BASIS, > # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. > # See the License for the specific language governing permissions and > # limitations under the License. > # > > args <- commandArgs(trailingOnly = TRUE) > > hashCode <- as.integer(args[1]) > port <- as.integer(args[2]) > libPath <- args[3] > rm(args) > print(paste("hashCode ", toString(hashCode))) [1] "hashCode 1850484525" > print(paste("Port ", toString(port))) [1] "Port 56762" > print(paste("LibPath ", libPath)) [1] "LibPath c:\\spark-1.6.1-bin-hadoop2.6\\bin\\../R/lib" > > .libPaths(c(file.path(libPath), .libPaths())) > library(SparkR) Attaching package: 'SparkR' The following objects are masked from 'package:stats': cov, filter, lag, na.omit, predict, sd, var The following objects are masked from 'package:base': colnames, colnames<-, intersect, rank, rbind, sample, subset, summary, table, transform > > > SparkR:::connectBackend("localhost", port) description class mode text "->localhost:56762" "sockconn" "wb" "binary" opened can read can write "opened" "yes" "yes" > > # scStartTime is needed by R/pkg/R/sparkR.R > assign(".scStartTime", as.integer(Sys.time()), envir = SparkR:::.sparkREnv) > > # getZeppelinR > .zeppelinR = SparkR:::callJStatic("org.apache.zeppelin.spark.ZeppelinR", "getZeppelinR", hashCode) Error in invokeJava(isStatic = TRUE, className, methodName, ...) : Calls: <Anonymous> -> invokeJava Execution halted at org.apache.zeppelin.spark.ZeppelinR.waitForRScriptInitialized(ZeppelinR.java:312) at org.apache.zeppelin.spark.ZeppelinR.request(ZeppelinR.java:250) at org.apache.zeppelin.spark.ZeppelinR.eval(ZeppelinR.java:198) at org.apache.zeppelin.spark.ZeppelinR.open(ZeppelinR.java:187) at org.apache.zeppelin.spark.SparkRInterpreter.open(SparkRInterpreter.java:104) at org.apache.zeppelin.interpreter.ClassloaderInterpreter.open(ClassloaderInterpreter.java:74) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:68) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:92) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:345) at org.apache.zeppelin.scheduler.Job.run(Job.java:176) at org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)