Zeppelin includes two R interpreter implementations.

One used to activated by -Psparkr the other -Pr.
Since https://github.com/apache/zeppelin/pull/2215, -Psparkr is activated
by default. And if you're trying to use sparkR, -Psparkr (activated by
default in master branch) is implementation you might be more interested.

So you can just try use with %spark.r prefix.
Let me know if it works for you.

Thanks,
moon

On Wed, Apr 26, 2017 at 12:11 AM Ruslan Dautkhanov <dautkha...@gmail.com>
wrote:

> Hi moon soo Lee,
>
> Cloudera's Spark doesn't have $SPARK_HOME/bin/sparkR
> Would Zeppelin still enable its sparkR interpreter then?
>
> Built Zeppelin using
>
> $ mvn clean package -DskipTests -Pspark-2.1 -Ppyspark
>> -Dhadoop.version=2.6.0-cdh5.10.1 -Phadoop-2.6 -Pyarn *-Pr* -Pvendor-repo
>> -Pscala-2.10 -pl '!...,!...' -e
>
>
> . . .
>> [INFO] Zeppelin: *R Interpreter* ............................ SUCCESS
>> [01:01 min]
>> [INFO]
>> ------------------------------------------------------------------------
>> [INFO] BUILD SUCCESS
>> [INFO]
>> ------------------------------------------------------------------------
>> [INFO] Total time: 11:28 min
>
>
> None of the R-related interpreters show up nevertheless.
>
> This is including latest Zeppelin snapshot and was the same on previous
> releases of Zeppelin.
> So something is missing on our side.
>
> R and R packages mentioned in
> http://zeppelin.apache.org/docs/0.8.0-SNAPSHOT/interpreter/r.html
> are installed on the servers that runs Zeppelin (and Spark driver as it is
> yarn-client).
>
> I guess either above build options are wrong or there is another
> dependency I missed.
> conf/zeppelin-site.xml has R related interpreters mentioned - [1] but none
> of them
> show up once Zeppelin starts up.
>
> Any ideas?
>
>
> Thank you,
> Ruslan
>
>
> [1]
>
> <property>
>>   <name>zeppelin.interpreters</name>
>>
>> <value>org.apache.zeppelin.spark.PySparkInterpreter,org.apache.zeppelin.spark.SparkInterpreter,org.apache.zeppelin.
>> *rinterpreter.RRepl*,org.apache.zeppelin.rinterpreter.*KnitR*
>> ,org.apache.zeppelin.spark.*SparkRInterpreter*
>> ,org.apache.zeppelin.spark.SparkSqlInterpreter,org.apache.zeppelin.spark.DepInterpreter,org.apache.zeppelin.markdown.Markdown,org.apache.zeppelin.angular.AngularInterpreter,org.apache.zeppelin.shell.ShellInterpreter,org.apache.zeppelin.file.HDFSFileInterpreter,org.apache.zeppelin.flink.FlinkInterpreter,,org.apache.zeppelin.python.PythonInterpreter,org.apache.zeppelin.lens.LensInterpreter,org.apache.zeppelin.ignite.IgniteInterpreter,org.apache.zeppelin.ignite.IgniteSqlInterpreter,org.apache.zeppelin.cassandra.CassandraInterpreter,org.apache.zeppelin.geode.GeodeOqlInterpreter,org.apache.zeppelin.postgresql.PostgreSqlInterpreter,org.apache.zeppelin.jdbc.JDBCInterpreter,org.apache.zeppelin.kylin.KylinInterpreter,org.apache.zeppelin.elasticsearch.ElasticsearchInterpreter,org.apache.zeppelin.scalding.ScaldingInterpreter,org.apache.zeppelin.alluxio.AlluxioInterpreter,org.apache.zeppelin.hbase.HbaseInterpreter,org.apache.zeppelin.livy.LivySparkInterpreter,org.apache.zeppelin.livy.LivyPySparkInterpreter,org.apache.zeppelin.livy.LivySparkRInterpreter,org.apache.zeppelin.livy.LivySparkSQLInterpreter,org.apache.zeppelin.bigquery.BigQueryInterpreter</value>
>>   <description>Comma separated interpreter configurations. First
>> interpreter become a default</description>
>> </property>
>
>
>
>
>
> --
> Ruslan Dautkhanov
>
> On Sun, Mar 19, 2017 at 1:07 PM, moon soo Lee <m...@apache.org> wrote:
>
>> Easiest way to figure out what your environment needs is,
>>
>> 1. run SPARK_HOME/bin/sparkR in your shell and make sure it works in the
>> same host where Zeppelin going to run.
>> 2. try use %spark.r in Zeppelin with SPARK_HOME configured. Normally it
>> should work when 1) works without problem, otherwise take a look error
>> message and error log to get more informations.
>>
>> Thanks,
>> moon
>>
>>
>> On Sat, Mar 18, 2017 at 8:47 PM Shanmukha Sreenivas Potti <
>> shanmu...@utexas.edu> wrote:
>>
>> I'm not 100% sure as I haven't set it up but it looks like I'm using
>>> Zeppelin preconfigured with Spark and I've also taken a snapshot of the
>>> Spark Interpreter configuration that I have access to/using in Zeppelin.
>>> This interpreter comes with SQL and Python integration and I'm figuring out
>>> how do I get to use R.
>>>
>>> On Sat, Mar 18, 2017 at 8:06 PM, moon soo Lee <m...@apache.org> wrote:
>>>
>>> AFAIK, Amazon EMR service has an option that launches Zeppelin
>>> (preconfigured) with Spark. Do you use Zeppelin provided by EMR or are you
>>> setting up Zeppelin separately?
>>>
>>> Thanks,
>>> moon
>>>
>>> On Sat, Mar 18, 2017 at 4:13 PM Shanmukha Sreenivas Potti <
>>> shanmu...@utexas.edu> wrote:
>>>
>>> ​​
>>> Hi Moon,
>>>
>>> Thanks for responding. Exporting Spark_home is exactly where I have a
>>> problem. I'm using Zeppelin notebook with Spark on EMR clusters from an AWS
>>> account on cloud. I'm not the master account holder for that AWS account
>>> but I'm guessing I'm a client account with limited access probably. Can I
>>> still do it?
>>>
>>> If yes, can you explain where and how should I do that shell scripting
>>> to export the variable? Can I do this in the notebook itself by starting
>>> the paragraph with sh% or do I need to do something else?
>>> If you can share any video that would be great. I would like to let you
>>> know that I'm a novice user just getting to explore Big Data.
>>>
>>> Sharing more info for better context.
>>>
>>> Here's my AWS account detail type:
>>> assumed-role/ConduitAccessClientRole-DO-NOT-DELETE/shan
>>>
>>> Spark Interpreter config in Zeppelin:
>>> [image: image.png]
>>>
>>> Thanks for your help.
>>>
>>> Shan
>>>
>>> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee <m...@apache.org> wrote:
>>>
>>> If you don't have spark cluster, then you don't need to do 2).
>>> After 1) %spark.r interpreter should work.
>>>
>>> If you do have spark cluster, export SPARK_HOME env variable in
>>> conf/zeppelin-env.sh, that should be enough make it work.
>>>
>>> Hope this helps.
>>>
>>> Thanks,
>>> moon
>>>
>>> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
>>> shanmu...@utexas.edu> wrote:
>>>
>>> Hello Group!
>>>
>>> I'm trying to leverage various R functions in Zeppelin but am having
>>> challenges in figuring out how to configure the Spark interpreter/
>>> SPARK_HOME variable.
>>>
>>> I'm going by this
>>> <https://zeppelin.apache.org/docs/0.6.0/interpreter/r.html> documentation
>>> for now, and specifically have issues with the following steps:
>>>
>>>    1.
>>>
>>>    To run R code and visualize plots in Apache Zeppelin, you will need
>>>    R on your master node (or your dev laptop).
>>>
>>>    For Centos: yum install R R-devel libcurl-devel openssl-devel For
>>>    Ubuntu: apt-get install r-base
>>>
>>> How do I figure out the master node and install the R interpreter?
>>> Novice user here.
>>>
>>>
>>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
>>> variable must be set. The best way to do this is by editing
>>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
>>> to interface with Spark. You should also copy
>>> conf/zeppelin-site.xml.template to conf/zeppelin-site.xml. That will ensure
>>> that Zeppelin sees the R Interpreter the first time it starts up.
>>>
>>> No idea as to how to do step 2 either.
>>>
>>> Appreciate your help. If there is a video that you can point me to that
>>> talks about these steps, that would be fantabulous.
>>>
>>> Thanks! Shan
>>>
>>> --
>>> Shan S. Potti,
>>>
>>>
>>>
>>>
>>> --
>>> Shan S. Potti,
>>> 737-333-1952 <(737)%20333-1952>
>>> https://www.linkedin.com/in/shanmukhasreenivas
>>>
>>> On Sat, Mar 18, 2017 at 8:39 AM, moon soo Lee <m...@apache.org> wrote:
>>>
>>> If you don't have spark cluster, then you don't need to do 2).
>>> After 1) %spark.r interpreter should work.
>>>
>>> If you do have spark cluster, export SPARK_HOME env variable in
>>> conf/zeppelin-env.sh, that should be enough make it work.
>>>
>>> Hope this helps.
>>>
>>> Thanks,
>>> moon
>>>
>>> On Fri, Mar 17, 2017 at 2:41 PM Shanmukha Sreenivas Potti <
>>> shanmu...@utexas.edu> wrote:
>>>
>>> Hello Group!
>>>
>>> I'm trying to leverage various R functions in Zeppelin but am having
>>> challenges in figuring out how to configure the Spark interpreter/
>>> SPARK_HOME variable.
>>>
>>> I'm going by this
>>> <https://zeppelin.apache.org/docs/0.6.0/interpreter/r.html>
>>> documentation for now, and specifically have issues with the following
>>> steps:
>>>
>>>    1.
>>>
>>>    To run R code and visualize plots in Apache Zeppelin, you will need
>>>    R on your master node (or your dev laptop).
>>>
>>>    For Centos: yum install R R-devel libcurl-devel openssl-devel For
>>>    Ubuntu: apt-get install r-base
>>>
>>> How do I figure out the master node and install the R interpreter?
>>> Novice user here.
>>>
>>>
>>> 2. To run Zeppelin with the R Interpreter, the SPARK_HOME environment
>>> variable must be set. The best way to do this is by editing
>>> conf/zeppelin-env.sh. If it is not set, the R Interpreter will not be able
>>> to interface with Spark. You should also copy
>>> conf/zeppelin-site.xml.template to conf/zeppelin-site.xml. That will ensure
>>> that Zeppelin sees the R Interpreter the first time it starts up.
>>>
>>> No idea as to how to do step 2 either.
>>>
>>> Appreciate your help. If there is a video that you can point me to that
>>> talks about these steps, that would be fantabulous.
>>>
>>> Thanks! Shan
>>>
>>> --
>>> Shan S. Potti,
>>>
>>>
>>>
>>>
>>> --
>>> Shan S. Potti,
>>> 737-333-1952 <(737)%20333-1952>
>>> https://www.linkedin.com/in/shanmukhasreenivas
>>>
>>>
>>>
>>>
>>> --
>>> Shan S. Potti,
>>> 737-333-1952 <(737)%20333-1952>
>>> https://www.linkedin.com/in/shanmukhasreenivas
>>>
>>
>

Reply via email to