Thanks Deepak.

I tried this as well. I created a hivecontext   with  "hivecontext <<-
sparkRHive.init(sc) "  .

When I tried to read hive table from this ,

results <- sql(hivecontext, "FROM test SELECT id")

I get below error,

Error in callJMethod(sqlContext, "sql", sqlQuery) :
  Invalid jobj 2. If SparkR was restarted, Spark operations need to be
re-executed.


Not sure what is causing this? Any leads or ideas? I am using rstudio.



On Tue, Jan 5, 2016 at 5:35 PM, Deepak Sharma <deepakmc...@gmail.com> wrote:

> Hi Sandeep
> I am not sure if ORC can be read directly in R.
> But there can be a workaround .First create hive table on top of ORC files
> and then access hive table in R.
>
> Thanks
> Deepak
>
> On Tue, Jan 5, 2016 at 4:57 PM, Sandeep Khurana <sand...@infoworks.io>
> wrote:
>
>> Hello
>>
>> I need to read an ORC files in hdfs in R using spark. I am not able to
>> find a package to do that.
>>
>> Can anyone help with documentation or example for this purpose?
>>
>> --
>> Architect
>> Infoworks.io
>> http://Infoworks.io
>>
>
>
>
> --
> Thanks
> Deepak
> www.bigdatabig.com
> www.keosha.net
>



-- 
Architect
Infoworks.io
http://Infoworks.io

Reply via email to