This problem is fixed by restarting R from R studio. Now see

16/01/22 08:08:38 INFO HiveMetaStore: No user is added in admin role,
since config is empty16/01/22 08:08:38 ERROR RBackendHandler: <init>
on org.apache.spark.sql.hive.HiveContext failedError in
value[[3L]](cond) : Spark SQL is not built with Hive support


 in rstudio while running same code and hive-site.xml is present in the .
It works in sparkR shell.

Any ideas?

On Fri, Jan 22, 2016 at 4:35 PM, Sandeep Khurana <sand...@infoworks.io>
wrote:

> Hello
>
> I installed spark in a folder. I start bin/sparkR on console. Then I
> execute below command and all work fine. I can see the data as well.
>
> hivecontext <<- sparkRHive.init(sc) ;
> df <- loadDF(hivecontext, "/someHdfsPath", "orc")
> showDF(df)
>
>
> But when I give same to rstudio, it throws the error mentioned below
>
> rstudio code
> ============
> Sys.setenv(SPARK_HOME="/home/myname/spark-1.6.0-bin-hadoop2.6")
> .libPaths(c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib"), .libPaths()))
> library(SparkR)
>
> sc <- sparkR.init(master="local")
> hivecontext <<- sparkRHive.init(sc) ;
> df <- loadDF(hivecontext, "/someHdfsPath", "orc")
> print("showing df now")
> showDF(df)
>
> Error thrown from rstudio
> ===================
>
> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more 
> info.Using Spark's default log4j profile: 
> org/apache/spark/log4j-defaults.properties16/01/22 06:00:12 ERROR 
> RBackendHandler: createSparkContext on org.apache.spark.api.r.RRDD 
> failedError in invokeJava(isStatic = TRUE, className, methodName, ...) :
>
>
>
>  What is different in rstudio than sparkR shell ? Should I change any
> setting to make it work in rstudio ?
>
>
>


-- 
Architect
Infoworks.io
http://Infoworks.io

Reply via email to