This is solved. Used Writable instead of LongWritable or NullWritable in
Mapper input key type.

Thanks
Suraj Nayak
On 19-Mar-2015 9:48 PM, "Suraj Nayak" <snay...@gmail.com> wrote:

> Is this related to https://issues.apache.org/jira/browse/HIVE-4329 ? Is
> there a workaround?
>
> On Thu, Mar 19, 2015 at 9:47 PM, Suraj Nayak <snay...@gmail.com> wrote:
>
>> Hi All,
>>
>> I was successfully able to integrate HCatMultipleInputs with the patch
>> for the tables created with TEXTFILE. But I get error when I read table
>> created with ORC file. The error is below :
>>
>> 15/03/19 10:51:32 INFO mapreduce.Job: Task Id :
>> attempt_1425012118520_9756_m_000000_0, Status : FAILED
>> Error: java.lang.ClassCastException: org.apache.hadoop.io.NullWritable
>> cannot be cast to org.apache.hadoop.io.LongWritable
>>     at com.abccompany.mapreduce.MyMapper.map(MyMapper.java:15)
>>     at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:145)
>>     at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)
>>     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:340)
>>     at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:167)
>>     at java.security.AccessController.doPrivileged(Native Method)
>>     at javax.security.auth.Subject.doAs(Subject.java:415)
>>     at
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1557)
>>     at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:162)
>>
>>
>> Can anyone help?
>>
>> Thanks in advance!
>>
>> On Wed, Mar 18, 2015 at 11:00 PM, Suraj Nayak <snay...@gmail.com> wrote:
>>
>>> Hi All,
>>>
>>> https://issues.apache.org/jira/browse/HIVE-4997 patch helped!
>>>
>>>
>>> On Tue, Mar 17, 2015 at 1:05 AM, Suraj Nayak <snay...@gmail.com> wrote:
>>>
>>>> Hi,
>>>>
>>>> I tried reading data via HCatalog for 1 Hive table in MapReduce using
>>>> something similar to
>>>> https://cwiki.apache.org/confluence/display/Hive/HCatalog+InputOutput#HCatalogInputOutput-RunningMapReducewithHCatalog.
>>>> I was able to read successfully.
>>>>
>>>> Now am trying to read 2 tables, as the requirement is to join 2 tables.
>>>> I did not find API similar to *FileInputFormat.addInputPaths* in
>>>> *HCatInputFormat*. What is the equivalent of the same in HCat ?
>>>>
>>>> I had performed join using FilesInputFormat in HDFS(by getting split
>>>> information in mapper). This article(
>>>> http://www.codingjunkie.com/mapreduce-reduce-joins) helped me code
>>>> join. <http://www.codingjunkie.com/mapreduce-reduce-joins/> Can
>>>> someone suggest how I can perform join operation using HCatalog ?
>>>>
>>>> Briefly, the aim is to
>>>>
>>>>    - Read 2 tables (almost similar schema)
>>>>    - If key exists in both the table send it to same reducer.
>>>>    - Do some processing on the records in reducer.
>>>>    - Save the output into file/Hive table.
>>>>
>>>> *P.S : The reason for using MapReduce to perform join is because of
>>>> complex requirement which can't be solved via Hive/Pig directly. *
>>>>
>>>> Any help will be greatly appreciated :)
>>>>
>>>> --
>>>> Thanks
>>>> Suraj Nayak M
>>>>
>>>
>>>
>>>
>>> --
>>> Thanks
>>> Suraj Nayak M
>>>
>>
>>
>>
>> --
>> Thanks
>> Suraj Nayak M
>>
>
>
>
> --
> Thanks
> Suraj Nayak M
>

Reply via email to