Resolved this issue after reinstalling r, rstudio. Had issues with earlier installation. On Jan 22, 2016 6:48 PM, "Sandeep Khurana" <sand...@infoworks.io> wrote:
> This problem is fixed by restarting R from R studio. Now see > > 16/01/22 08:08:38 INFO HiveMetaStore: No user is added in admin role, since > config is empty16/01/22 08:08:38 ERROR RBackendHandler: <init> on > org.apache.spark.sql.hive.HiveContext failedError in value[[3L]](cond) : > Spark SQL is not built with Hive support > > > in rstudio while running same code and hive-site.xml is present in the . > It works in sparkR shell. > > Any ideas? > > On Fri, Jan 22, 2016 at 4:35 PM, Sandeep Khurana <sand...@infoworks.io> > wrote: > >> Hello >> >> I installed spark in a folder. I start bin/sparkR on console. Then I >> execute below command and all work fine. I can see the data as well. >> >> hivecontext <<- sparkRHive.init(sc) ; >> df <- loadDF(hivecontext, "/someHdfsPath", "orc") >> showDF(df) >> >> >> But when I give same to rstudio, it throws the error mentioned below >> >> rstudio code >> ============ >> Sys.setenv(SPARK_HOME="/home/myname/spark-1.6.0-bin-hadoop2.6") >> .libPaths(c(file.path(Sys.getenv("SPARK_HOME"), "R", "lib"), .libPaths())) >> library(SparkR) >> >> sc <- sparkR.init(master="local") >> hivecontext <<- sparkRHive.init(sc) ; >> df <- loadDF(hivecontext, "/someHdfsPath", "orc") >> print("showing df now") >> showDF(df) >> >> Error thrown from rstudio >> =================== >> >> log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for >> more info.Using Spark's default log4j profile: >> org/apache/spark/log4j-defaults.properties16/01/22 06:00:12 ERROR >> RBackendHandler: createSparkContext on org.apache.spark.api.r.RRDD >> failedError in invokeJava(isStatic = TRUE, className, methodName, ...) : >> >> >> >> What is different in rstudio than sparkR shell ? Should I change any >> setting to make it work in rstudio ? >> >> >> > > > -- > Architect > Infoworks.io > http://Infoworks.io >