Thank you for the reply. I fixed it by using the stored as clause in my hive 
query. Everything is good.

From: Roberto Congiu <roberto.con...@openx.com<mailto:roberto.con...@openx.com>>
Reply-To: "user@hive.apache.org<mailto:user@hive.apache.org>" 
<user@hive.apache.org<mailto:user@hive.apache.org>>
Date: Monday, July 29, 2013 11:33 AM
To: "user@hive.apache.org<mailto:user@hive.apache.org>" 
<user@hive.apache.org<mailto:user@hive.apache.org>>
Subject: Re: BytesWritable

Hi Josh,
it doesn't have to. However, your table does not specify any input format, 
which makes Hive use the default HiveIgnoreKeyTextOutputFormat (see 
http://grepcode.com/file/repository.cloudera.com/content/repositories/releases/org.apache.hadoop.hive/hive-exec/0.7.1-cdh3u1/org/apache/hadoop/hive/ql/io/HiveIgnoreKeyTextOutputFormat.java).

What's the bigger picture, meaning, what is the actual format you want to write 
down to disk ?

If you're storing it in text format, your SerDe's serialize method has to 
return either a Text or a BytesWritable, as you can see in the code linked 
earlier. It looks like it's returning a DocumentWritable object, which 
HiveIgnoreKeyTextOutputFormat does not know how to handle.

R.


On Mon, Jul 29, 2013 at 5:42 AM, Clum, Josh 
<josh.c...@merck.com<mailto:josh.c...@merck.com>> wrote:
Hi,

I wrote a custom serde that stores takes data and stores it in objects called 
DocumentWritable that inherits from Writable. When I try to run code to insert 
into the table that uses my serde, I get an error saying that DocumentWritable 
cannot be cast to BytesWritable. Does my DocumentWritable need to inherit from 
org.apache.hadoop.io.BytesWritable?

I'm creating a table using a custom serde that I wrote:

CREATE TABLE avro_test (
    NOTEBOOK_ID STRING,
   …
    DLC_XID STRING
  )
ROW FORMAT SERDE 'com.merck.huron.hive.LuceneSerde';

When I try to insert data using a select statement I am storing it into a class 
I have called DocumentWritable:

INSERT OVERWRITE TABLE avro_test SELECT * FROM avro

DocumentWritable inherits from Writable. When I do the select and my serde code 
gets executed I get the following error:

Error: java.lang.RuntimeException: 
org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while 
processing row 
{"notebook_id":"1","stuff_id":"1","person_id":"0","person_role_id":"1","reference":"NO
 
DOC","lastname":"OMITTED","firstinitial":"TBD","source_id":null,"source_name":"MRL
 
SYNTHESIS","notebook":null,"pg":null,"firstname":null,"middlename":null,"site":null,"mail":null,"dlc_xid":null}
at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:161)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:429)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:158)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1441)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:153)
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error 
while processing row 
{"notebook_id":"1","stuff_id":"1","person_id":"0","person_role_id":"1","reference":"NO
 
DOC","lastname":"OMITTED","firstinitial":"TBD","source_id":null,"source_name":"MRL
 
SYNTHESIS","notebook":null,"pg":null,"firstname":null,"middlename":null,"site":null,"mail":null,"dlc_xid":null}
at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:565)
at org.apache.hadoop.hive.ql.exec.ExecMapper.map(ExecMapper.java:143)
... 8 more
Caused by: java.lang.ClassCastException: 
com.comp.proj.indexing.DocumentWritable cannot be cast to 
org.apache.hadoop.io.BytesWritable
at 
org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat$1.write(HiveIgnoreKeyTextOutputFormat.java:90)
at 
org.apache.hadoop.hive.ql.exec.FileSinkOperator.processOp(FileSinkOperator.java:606)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:474)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:800)
at 
org.apache.hadoop.hive.ql.exec.SelectOperator.processOp(SelectOperator.java:84)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:474)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:800)
at 
org.apache.hadoop.hive.ql.exec.TableScanOperator.processOp(TableScanOperator.java:83)
at org.apache.hadoop.hive.ql.exec.Operator.process(Operator.java:474)
at org.apache.hadoop.hive.ql.exec.Operator.forward(Operator.java:800)
at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:546)
... 9 more


Notice:  This e-mail message, together with any attachments, contains
information of Merck & Co., Inc. (One Merck Drive, Whitehouse Station,
New Jersey, USA 08889), and/or its affiliates Direct contact information
for affiliates is available at
http://www.merck.com/contact/contacts.html) that may be confidential,
proprietary copyrighted and/or legally privileged. It is intended solely
for the use of the individual or entity named on this message. If you are
not the intended recipient, and have received this message in error,
please notify us immediately by reply e-mail and then delete it from
your system.



--
----------------------------------------------------------
Good judgement comes with experience.
Experience comes with bad judgement.
----------------------------------------------------------
Roberto Congiu - Data Engineer - OpenX
tel: +1 626 466 1141
Notice:  This e-mail message, together with any attachments, contains
information of Merck & Co., Inc. (One Merck Drive, Whitehouse Station,
New Jersey, USA 08889), and/or its affiliates Direct contact information
for affiliates is available at 
http://www.merck.com/contact/contacts.html) that may be confidential,
proprietary copyrighted and/or legally privileged. It is intended solely
for the use of the individual or entity named on this message. If you are
not the intended recipient, and have received this message in error,
please notify us immediately by reply e-mail and then delete it from 
your system.

Reply via email to