Hi Edward,

Sorry, If I was not clear. My question is around difference between 
DoubleWritable in hadoop and hive, other writables from hadoop works fine in 
hive.
Hive.serde types are limited to Double, Byte, Short and Timestamp.

I am using hive 0.8

Richin

-----Original Message-----
From: ext Edward Capriolo [mailto:edlinuxg...@gmail.com] 
Sent: Tuesday, June 12, 2012 5:12 PM
To: user@hive.apache.org
Subject: Re: hadoop.io.DoubleWritable v/s hive.serde2.io.DoubleWritable

If you use Double or double hive will automatically convert. I would always 
recommend the hive.serde types.

Edward

On Tue, Jun 12, 2012 at 4:56 PM,  <richin.j...@nokia.com> wrote:
> Hi Guys,
>
>
>
> I am writing a UDF in hive to convert a double value to string, so the 
> evaluate method of my UDF class looks like
>
>
>
> import org.apache.hadoop.hive.ql.exec.UDF;
>
> import org.apache.hadoop.io.Text;
>
> //import org.apache.hadoop.io.DoubleWritable; - does not work
>
> import org.apache.hadoop.hive.serde2.io.DoubleWritable;
>
>
>
> public Text evaluate(DoubleWritable d){
>
>
>
> }
>
>
>
> When I looked at different UDF examples the Hadoop Writables seem to 
> work fine in case of Text and IntWritable but for DoubleWritable I was 
> getting an error. I figured out looking at couple of examples online 
> that I should use the DoubleWritable from hive.serde2 package instead of 
> hadoop.io package.
>
>
>
> Can someone please explain why this special case for DoubleWritable?
>
>
>
> Thanks,
>
> Richin

Reply via email to