Thanks Deepak. I tried this as well. I created a hivecontext with "hivecontext <<- sparkRHive.init(sc) " .
When I tried to read hive table from this , results <- sql(hivecontext, "FROM test SELECT id") I get below error, Error in callJMethod(sqlContext, "sql", sqlQuery) : Invalid jobj 2. If SparkR was restarted, Spark operations need to be re-executed. Not sure what is causing this? Any leads or ideas? I am using rstudio. On Tue, Jan 5, 2016 at 5:35 PM, Deepak Sharma <deepakmc...@gmail.com> wrote: > Hi Sandeep > I am not sure if ORC can be read directly in R. > But there can be a workaround .First create hive table on top of ORC files > and then access hive table in R. > > Thanks > Deepak > > On Tue, Jan 5, 2016 at 4:57 PM, Sandeep Khurana <sand...@infoworks.io> > wrote: > >> Hello >> >> I need to read an ORC files in hdfs in R using spark. I am not able to >> find a package to do that. >> >> Can anyone help with documentation or example for this purpose? >> >> -- >> Architect >> Infoworks.io >> http://Infoworks.io >> > > > > -- > Thanks > Deepak > www.bigdatabig.com > www.keosha.net > -- Architect Infoworks.io http://Infoworks.io