Hi moon soo Lee, Cloudera's Spark doesn't have $SPARK_HOME/bin/sparkR Would Zeppelin still enable its sparkR interpreter then?
Built Zeppelin using $ mvn clean package -DskipTests -Pspark-2.1 -Ppyspark > -Dhadoop.version=2.6.0-cdh5.10.1 -Phadoop-2.6 -Pyarn *-Pr* -Pvendor-repo > -Pscala-2.10 -pl '!...,!...' -e . . . > [INFO] Zeppelin: *R Interpreter* ............................ SUCCESS > [01:01 min] > [INFO] > ------------------------------------------------------------------------ > [INFO] BUILD SUCCESS > [INFO] > ------------------------------------------------------------------------ > [INFO] Total time: 11:28 min None of the R-related interpreters show up nevertheless. This is including latest Zeppelin snapshot and was the same on previous releases of Zeppelin. So something is missing on our side. R and R packages mentioned in http://zeppelin.apache.org/docs/0.8.0-SNAPSHOT/interpreter/r.html are installed on the servers that runs Zeppelin (and Spark driver as it is yarn-client). I guess either above build options are wrong or there is another dependency I missed. conf/zeppelin-site.xml has R related interpreters mentioned - [1] but none of them show up once Zeppelin starts up. Any ideas? Thank you, Ruslan [1] <property> > <name>zeppelin.interpreters</name> > > <value>org.apache.zeppelin.spark.PySparkInterpreter,org.apache.zeppelin.spark.SparkInterpreter,org.apache.zeppelin. > *rinterpreter.RRepl*,org.apache.zeppelin.rinterpreter.*KnitR* > ,org.apache.zeppelin.spark.*SparkRInterpreter* > ,org.apache.zeppelin.spark.SparkSqlInterpreter,org.apache.zeppelin.spark.DepInterpreter,org.apache.zeppelin.markdown.Markdown,org.apache.zeppelin.angular.AngularInterpreter,org.apache.zeppelin.shell.ShellInterpreter,org.apache.zeppelin.file.HDFSFileInterpreter,org.apache.zeppelin.flink.FlinkInterpreter,,org.apache.zeppelin.python.PythonInterpreter,org.apache.zeppelin.lens.LensInterpreter,org.apache.zeppelin.ignite.IgniteInterpreter,org.apache.zeppelin.ignite.IgniteSqlInterpreter,org.apache.zeppelin.cassandra.CassandraInterpreter,org.apache.zeppelin.geode.GeodeOqlInterpreter,org.apache.zeppelin.postgresql.PostgreSqlInterpreter,org.apache.zeppelin.jdbc.JDBCInterpreter,org.apache.zeppelin.kylin.KylinInterpreter,org.apache.zeppelin.elasticsearch.ElasticsearchInterpreter,org.apache.zeppelin.scalding.ScaldingInterpreter,org.apache.zeppelin.alluxio.AlluxioInterpreter,org.apache.zeppelin.hbase.HbaseInterpreter,org.apache.zeppelin.livy.LivySparkInterpreter,org.apache.zeppelin.livy.LivyPySparkInterpreter,org.apache.zeppelin.livy.LivySparkRInterpreter,org.apache.zeppelin.livy.LivySparkSQLInterpreter,org.apache.zeppelin.bigquery.BigQueryInterpreter</value> > <description>Comma separated interpreter configurations. First > interpreter become a default</description> > </property> -- Ruslan Dautkhanov On Sun, Mar 19, 2017 at 1:07 PM, moon soo Lee <m...@apache.org> wrote: > Easiest way to figure out what your environment needs is, > > 1. run SPARK_HOME/bin/sparkR in your shell and make sure it works in the > same host where Zeppelin going to run. > 2. try use %spark.r in Zeppelin with SPARK_HOME configured. Normally it > should work when 1) works without problem, otherwise take a look error > message and error log to get more informations. > > Thanks, > moon > > > On Sat, Mar 18, 2017 at 8:47 PM Shanmukha Sreenivas Potti < > shanmu...@utexas.edu> wrote: > > I'm not 100% sure as I haven't set it up but it looks like I'm using >> Zeppelin preconfigured with Spark and I've also taken a snapshot of the >> Spark Interpreter configuration that I have access to/using in Zeppelin. >> This interpreter comes with SQL and Python integration and I'm figuring out >> how do I get to use R. >> >> On Sat, Mar 18, 2017 at 8:06 PM, moon soo Lee <m...@apache.org> wrote: >> >> AFAIK, Amazon EMR service has an option that launches Zeppelin >> (preconfigured) with Spark. Do you use Zeppelin provided by EMR or are you >> setting up Zeppelin separately? >> >> Thanks, >> moon >> >> On Sat, Mar 18, 2017 at 4:13 PM Shanmukha Sreenivas Potti < >> shanmu...@utexas.edu> wrote: >> >> >> Hi Moon, >> >> Thanks for responding. Exporting Spark_home is exactly where I have a >> problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS >> account on cloud. I'm not the master account holder for that AWS account >> but I'm guessing I'm a client account with limited access probably. Can I >> still do it? >> >> If yes, can you explain where and how should I do that shell scripting to >> export the variable? Can I do this in the notebook itself by starting the >> paragraph with sh% or do I need to do something else? >> If you can share any video that would be great. I would like to let you >> know that I'm a novice user just getting to explore Big Data. >> >> Sharing more info for better context. >> >> Here's my AWS account detail type: >> assumed-role/ConduitAccessClientRole-DO-NOT-DELETE/shan >> >> Spark Interpreter config in Zeppelin: >> [image: image.png] >> >> Thanks for your help. >> >> Shan >> >> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee <m...@apache.org> wrote: >> >> If you don't have spark cluster, then you don't need to do 2). >> After 1) %spark.r interpreter should work. >> >> If you do have spark cluster, export SPARK_HOME env variable in >> conf/zeppelin-env.sh, that should be enough make it work. >> >> Hope this helps. >> >> Thanks, >> moon >> >> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti < >> shanmu...@utexas.edu> wrote: >> >> Hello Group! >> >> I'm trying to leverage various R functions in Zeppelin but am having >> challenges in figuring out how to configure the Spark interpreter/ >> SPARK_HOME variable. >> >> I'm going by this >> <https://zeppelin.apache.org/docs/0.6.0/interpreter/r.html> documentation >> for now, and specifically have issues with the following steps: >> >> 1. >> >> To run R code and visualize plots in Apache Zeppelin, you will need R >> on your master node (or your dev laptop). >> >> For Centos: yum install R R-devel libcurl-devel openssl-devel For >> Ubuntu: apt-get install r-base >> >> How do I figure out the master node and install the R interpreter? Novice >> user here. >> >> >> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment >> variable must be set. The best way to do this is by editing >> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able >> to interface with Spark. You should also copy conf/zeppelin-site.xml.template >> to conf/zeppelin-site.xml. That will ensure that Zeppelin sees the R >> Interpreter the first time it starts up. >> >> No idea as to how to do step 2 either. >> >> Appreciate your help. If there is a video that you can point me to that >> talks about these steps, that would be fantabulous. >> >> Thanks! Shan >> >> -- >> Shan S. Potti, >> >> >> >> >> -- >> Shan S. Potti, >> 737-333-1952 <(737)%20333-1952> >> https://www.linkedin.com/in/shanmukhasreenivas >> >> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee <m...@apache.org> wrote: >> >> If you don't have spark cluster, then you don't need to do 2). >> After 1) %spark.r interpreter should work. >> >> If you do have spark cluster, export SPARK_HOME env variable in >> conf/zeppelin-env.sh, that should be enough make it work. >> >> Hope this helps. >> >> Thanks, >> moon >> >> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti < >> shanmu...@utexas.edu> wrote: >> >> Hello Group! >> >> I'm trying to leverage various R functions in Zeppelin but am having >> challenges in figuring out how to configure the Spark interpreter/ >> SPARK_HOME variable. >> >> I'm going by this >> <https://zeppelin.apache.org/docs/0.6.0/interpreter/r.html> >> documentation for now, and specifically have issues with the following >> steps: >> >> 1. >> >> To run R code and visualize plots in Apache Zeppelin, you will need R >> on your master node (or your dev laptop). >> >> For Centos: yum install R R-devel libcurl-devel openssl-devel For >> Ubuntu: apt-get install r-base >> >> How do I figure out the master node and install the R interpreter? Novice >> user here. >> >> >> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment >> variable must be set. The best way to do this is by editing >> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able >> to interface with Spark. You should also copy conf/zeppelin-site.xml.template >> to conf/zeppelin-site.xml. That will ensure that Zeppelin sees the R >> Interpreter the first time it starts up. >> >> No idea as to how to do step 2 either. >> >> Appreciate your help. If there is a video that you can point me to that >> talks about these steps, that would be fantabulous. >> >> Thanks! Shan >> >> -- >> Shan S. Potti, >> >> >> >> >> -- >> Shan S. Potti, >> 737-333-1952 <(737)%20333-1952> >> https://www.linkedin.com/in/shanmukhasreenivas >> >> >> >> >> -- >> Shan S. Potti, >> 737-333-1952 <(737)%20333-1952> >> https://www.linkedin.com/in/shanmukhasreenivas >> >