Hive likely wishes to format the data differently then Hadoop does.
Hive re-uses what it can. I would diff the two .java files and find
out for yourself :)


On Tue, Jun 12, 2012 at 5:20 PM,  <richin.j...@nokia.com> wrote:
> Hi Edward,
>
> Sorry, If I was not clear. My question is around difference between 
> DoubleWritable in hadoop and hive, other writables from hadoop works fine in 
> hive.
> Hive.serde types are limited to Double, Byte, Short and Timestamp.
>
> I am using hive 0.8
>
> Richin
>
> -----Original Message-----
> From: ext Edward Capriolo [mailto:edlinuxg...@gmail.com]
> Sent: Tuesday, June 12, 2012 5:12 PM
> To: user@hive.apache.org
> Subject: Re: hadoop.io.DoubleWritable v/s hive.serde2.io.DoubleWritable
>
> If you use Double or double hive will automatically convert. I would always 
> recommend the hive.serde types.
>
> Edward
>
> On Tue, Jun 12, 2012 at 4:56 PM,  <richin.j...@nokia.com> wrote:
>> Hi Guys,
>>
>>
>>
>> I am writing a UDF in hive to convert a double value to string, so the
>> evaluate method of my UDF class looks like
>>
>>
>>
>> import org.apache.hadoop.hive.ql.exec.UDF;
>>
>> import org.apache.hadoop.io.Text;
>>
>> //import org.apache.hadoop.io.DoubleWritable; - does not work
>>
>> import org.apache.hadoop.hive.serde2.io.DoubleWritable;
>>
>>
>>
>> public Text evaluate(DoubleWritable d){
>>
>>
>>
>> }
>>
>>
>>
>> When I looked at different UDF examples the Hadoop Writables seem to
>> work fine in case of Text and IntWritable but for DoubleWritable I was
>> getting an error. I figured out looking at couple of examples online
>> that I should use the DoubleWritable from hive.serde2 package instead of 
>> hadoop.io package.
>>
>>
>>
>> Can someone please explain why this special case for DoubleWritable?
>>
>>
>>
>> Thanks,
>>
>> Richin

Reply via email to