Zeppelin includes two R interpreter implementations. One used to activated by -Psparkr the other -Pr. Since https://github.com/apache/zeppelin/pull/2215, -Psparkr is activated by default. And if you're trying to use sparkR, -Psparkr (activated by default in master branch) is implementation you might be more interested.
So you can just try use with %spark.r prefix. Let me know if it works for you. Thanks, moon On Wed, Apr 26, 2017 at 12:11 AM Ruslan Dautkhanov <dautkha...@gmail.com> wrote: > Hi moon soo Lee, > > Cloudera's Spark doesn't have $SPARK_HOME/bin/sparkR > Would Zeppelin still enable its sparkR interpreter then? > > Built Zeppelin using > > $ mvn clean package -DskipTests -Pspark-2.1 -Ppyspark >> -Dhadoop.version=2.6.0-cdh5.10.1 -Phadoop-2.6 -Pyarn *-Pr* -Pvendor-repo >> -Pscala-2.10 -pl '!...,!...' -e > > > . . . >> [INFO] Zeppelin: *R Interpreter* ............................ SUCCESS >> [01:01 min] >> [INFO] >> ------------------------------------------------------------------------ >> [INFO] BUILD SUCCESS >> [INFO] >> ------------------------------------------------------------------------ >> [INFO] Total time: 11:28 min > > > None of the R-related interpreters show up nevertheless. > > This is including latest Zeppelin snapshot and was the same on previous > releases of Zeppelin. > So something is missing on our side. > > R and R packages mentioned in > http://zeppelin.apache.org/docs/0.8.0-SNAPSHOT/interpreter/r.html > are installed on the servers that runs Zeppelin (and Spark driver as it is > yarn-client). > > I guess either above build options are wrong or there is another > dependency I missed. > conf/zeppelin-site.xml has R related interpreters mentioned - [1] but none > of them > show up once Zeppelin starts up. > > Any ideas? > > > Thank you, > Ruslan > > > [1] > > <property> >> <name>zeppelin.interpreters</name> >> >> <value>org.apache.zeppelin.spark.PySparkInterpreter,org.apache.zeppelin.spark.SparkInterpreter,org.apache.zeppelin. >> *rinterpreter.RRepl*,org.apache.zeppelin.rinterpreter.*KnitR* >> ,org.apache.zeppelin.spark.*SparkRInterpreter* >> ,org.apache.zeppelin.spark.SparkSqlInterpreter,org.apache.zeppelin.spark.DepInterpreter,org.apache.zeppelin.markdown.Markdown,org.apache.zeppelin.angular.AngularInterpreter,org.apache.zeppelin.shell.ShellInterpreter,org.apache.zeppelin.file.HDFSFileInterpreter,org.apache.zeppelin.flink.FlinkInterpreter,,org.apache.zeppelin.python.PythonInterpreter,org.apache.zeppelin.lens.LensInterpreter,org.apache.zeppelin.ignite.IgniteInterpreter,org.apache.zeppelin.ignite.IgniteSqlInterpreter,org.apache.zeppelin.cassandra.CassandraInterpreter,org.apache.zeppelin.geode.GeodeOqlInterpreter,org.apache.zeppelin.postgresql.PostgreSqlInterpreter,org.apache.zeppelin.jdbc.JDBCInterpreter,org.apache.zeppelin.kylin.KylinInterpreter,org.apache.zeppelin.elasticsearch.ElasticsearchInterpreter,org.apache.zeppelin.scalding.ScaldingInterpreter,org.apache.zeppelin.alluxio.AlluxioInterpreter,org.apache.zeppelin.hbase.HbaseInterpreter,org.apache.zeppelin.livy.LivySparkInterpreter,org.apache.zeppelin.livy.LivyPySparkInterpreter,org.apache.zeppelin.livy.LivySparkRInterpreter,org.apache.zeppelin.livy.LivySparkSQLInterpreter,org.apache.zeppelin.bigquery.BigQueryInterpreter</value> >> <description>Comma separated interpreter configurations. First >> interpreter become a default</description> >> </property> > > > > > > -- > Ruslan Dautkhanov > > On Sun, Mar 19, 2017 at 1:07 PM, moon soo Lee <m...@apache.org> wrote: > >> Easiest way to figure out what your environment needs is, >> >> 1. run SPARK_HOME/bin/sparkR in your shell and make sure it works in the >> same host where Zeppelin going to run. >> 2. try use %spark.r in Zeppelin with SPARK_HOME configured. Normally it >> should work when 1) works without problem, otherwise take a look error >> message and error log to get more informations. >> >> Thanks, >> moon >> >> >> On Sat, Mar 18, 2017 at 8:47 PM Shanmukha Sreenivas Potti < >> shanmu...@utexas.edu> wrote: >> >> I'm not 100% sure as I haven't set it up but it looks like I'm using >>> Zeppelin preconfigured with Spark and I've also taken a snapshot of the >>> Spark Interpreter configuration that I have access to/using in Zeppelin. >>> This interpreter comes with SQL and Python integration and I'm figuring out >>> how do I get to use R. >>> >>> On Sat, Mar 18, 2017 at 8:06 PM, moon soo Lee <m...@apache.org> wrote: >>> >>> AFAIK, Amazon EMR service has an option that launches Zeppelin >>> (preconfigured) with Spark. Do you use Zeppelin provided by EMR or are you >>> setting up Zeppelin separately? >>> >>> Thanks, >>> moon >>> >>> On Sat, Mar 18, 2017 at 4:13 PM Shanmukha Sreenivas Potti < >>> shanmu...@utexas.edu> wrote: >>> >>> >>> Hi Moon, >>> >>> Thanks for responding. Exporting Spark_home is exactly where I have a >>> problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS >>> account on cloud. I'm not the master account holder for that AWS account >>> but I'm guessing I'm a client account with limited access probably. Can I >>> still do it? >>> >>> If yes, can you explain where and how should I do that shell scripting >>> to export the variable? Can I do this in the notebook itself by starting >>> the paragraph with sh% or do I need to do something else? >>> If you can share any video that would be great. I would like to let you >>> know that I'm a novice user just getting to explore Big Data. >>> >>> Sharing more info for better context. >>> >>> Here's my AWS account detail type: >>> assumed-role/ConduitAccessClientRole-DO-NOT-DELETE/shan >>> >>> Spark Interpreter config in Zeppelin: >>> [image: image.png] >>> >>> Thanks for your help. >>> >>> Shan >>> >>> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee <m...@apache.org> wrote: >>> >>> If you don't have spark cluster, then you don't need to do 2). >>> After 1) %spark.r interpreter should work. >>> >>> If you do have spark cluster, export SPARK_HOME env variable in >>> conf/zeppelin-env.sh, that should be enough make it work. >>> >>> Hope this helps. >>> >>> Thanks, >>> moon >>> >>> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti < >>> shanmu...@utexas.edu> wrote: >>> >>> Hello Group! >>> >>> I'm trying to leverage various R functions in Zeppelin but am having >>> challenges in figuring out how to configure the Spark interpreter/ >>> SPARK_HOME variable. >>> >>> I'm going by this >>> <https://zeppelin.apache.org/docs/0.6.0/interpreter/r.html> documentation >>> for now, and specifically have issues with the following steps: >>> >>> 1. >>> >>> To run R code and visualize plots in Apache Zeppelin, you will need >>> R on your master node (or your dev laptop). >>> >>> For Centos: yum install R R-devel libcurl-devel openssl-devel For >>> Ubuntu: apt-get install r-base >>> >>> How do I figure out the master node and install the R interpreter? >>> Novice user here. >>> >>> >>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment >>> variable must be set. The best way to do this is by editing >>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able >>> to interface with Spark. You should also copy >>> conf/zeppelin-site.xml.template to conf/zeppelin-site.xml. That will ensure >>> that Zeppelin sees the R Interpreter the first time it starts up. >>> >>> No idea as to how to do step 2 either. >>> >>> Appreciate your help. If there is a video that you can point me to that >>> talks about these steps, that would be fantabulous. >>> >>> Thanks! Shan >>> >>> -- >>> Shan S. Potti, >>> >>> >>> >>> >>> -- >>> Shan S. Potti, >>> 737-333-1952 <(737)%20333-1952> >>> https://www.linkedin.com/in/shanmukhasreenivas >>> >>> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee <m...@apache.org> wrote: >>> >>> If you don't have spark cluster, then you don't need to do 2). >>> After 1) %spark.r interpreter should work. >>> >>> If you do have spark cluster, export SPARK_HOME env variable in >>> conf/zeppelin-env.sh, that should be enough make it work. >>> >>> Hope this helps. >>> >>> Thanks, >>> moon >>> >>> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti < >>> shanmu...@utexas.edu> wrote: >>> >>> Hello Group! >>> >>> I'm trying to leverage various R functions in Zeppelin but am having >>> challenges in figuring out how to configure the Spark interpreter/ >>> SPARK_HOME variable. >>> >>> I'm going by this >>> <https://zeppelin.apache.org/docs/0.6.0/interpreter/r.html> >>> documentation for now, and specifically have issues with the following >>> steps: >>> >>> 1. >>> >>> To run R code and visualize plots in Apache Zeppelin, you will need >>> R on your master node (or your dev laptop). >>> >>> For Centos: yum install R R-devel libcurl-devel openssl-devel For >>> Ubuntu: apt-get install r-base >>> >>> How do I figure out the master node and install the R interpreter? >>> Novice user here. >>> >>> >>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment >>> variable must be set. The best way to do this is by editing >>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able >>> to interface with Spark. You should also copy >>> conf/zeppelin-site.xml.template to conf/zeppelin-site.xml. That will ensure >>> that Zeppelin sees the R Interpreter the first time it starts up. >>> >>> No idea as to how to do step 2 either. >>> >>> Appreciate your help. If there is a video that you can point me to that >>> talks about these steps, that would be fantabulous. >>> >>> Thanks! Shan >>> >>> -- >>> Shan S. Potti, >>> >>> >>> >>> >>> -- >>> Shan S. Potti, >>> 737-333-1952 <(737)%20333-1952> >>> https://www.linkedin.com/in/shanmukhasreenivas >>> >>> >>> >>> >>> -- >>> Shan S. Potti, >>> 737-333-1952 <(737)%20333-1952> >>> https://www.linkedin.com/in/shanmukhasreenivas >>> >> >