This problem is fixed by restarting R from R studio. Now see 16/01/22 08:08:38 INFO HiveMetaStore: No user is added in admin role, since config is empty16/01/22 08:08:38 ERROR RBackendHandler: <init> on org.apache.spark.sql.hive.HiveContext failedError in value[[3L]](cond) : Spark SQL is not built with Hive support
in rstudio while running same code and hive-site.xml is present in the . It works in sparkR shell. Any ideas? On Fri, Jan 22, 2016 at 4:35 PM, Sandeep Khurana <sand...@infoworks.io> wrote: > Hello > > I installed spark in a folder. I start bin/sparkR on console. Then I > execute below command and all work fine. I can see the data as well. > > hivecontext <<- sparkRHive.init(sc) ; > df <- loadDF(hivecontext, "/someHdfsPath", "orc") > showDF(df) > > > But when I give same to rstudio, it throws the error mentioned below > > rstudio code > ============ > Sys.setenv(SPARK_HOME="/home/myname/spark-1.6.0-bin-hadoop2.6") > .libPaths(c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib"), .libPaths())) > library(SparkR) > > sc <- sparkR.init(master="local") > hivecontext <<- sparkRHive.init(sc) ; > df <- loadDF(hivecontext, "/someHdfsPath", "orc") > print("showing df now") > showDF(df) > > Error thrown from rstudio > =================== > > log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more > info.Using Spark's default log4j profile: > org/apache/spark/log4j-defaults.properties16/01/22 06:00:12 ERROR > RBackendHandler: createSparkContext on org.apache.spark.api.r.RRDD > failedError in invokeJava(isStatic = TRUE, className, methodName, ...) : > > > > What is different in rstudio than sparkR shell ? Should I change any > setting to make it work in rstudio ? > > > -- Architect Infoworks.io http://Infoworks.io