What version of protobuf are you using? Are you compiling hive from source?

Thanks
Prasanth Jayachandran

On Dec 16, 2013, at 4:30 PM, Bryan Jeffrey <bryan.jeff...@gmail.com> wrote:

> Hello.
> 
> Running the following version of Hadoop: hadoop-2.2.0
> Running the following version of Hive: hive-0.12.0
> 
> I have a simple test system setup with (2) datanodes/node manager and (1) 
> namenode/resource manager.  Hive is running on the namenode, and contacting a 
> MySQL database for metastore.
> 
> I have created a small table 'from_text' as follows:
> 
> [server:10001] hive> describe from_text;
> foo                     int                     None
> bar                     int                     None
> boo                     string                  None
> 
> 
> [server:10001] hive> select * from from_text;
> 1       2       Hello
> 2       3       World
> 
> I go to insert the data into my Orc table, 'orc_table':
> 
> [server:10001] hive> describe orc_test;
> foo                     int                     from deserializer
> bar                     int                     from deserializer
> boo                     string                  from deserializer
> 
> 
> The job runs, but fails to complete with the following errors (see below).  
> This seems to be the exact example covered in the example here: 
> 
> http://hortonworks.com/blog/orcfile-in-hdp-2-better-compression-better-performance/
> 
> I took a few minutes to recompile the protbuf library as several other 
> problems mentioned that Hive 0.12 did not have the protobuf library updated. 
> That did not remedy the problem.  Any ideas?
> 
> 
> [server:10001] hive> insert into table orc_test select * from from_text;
> [Hive Error]: Query returned non-zero code: 2, cause: FAILED: Execution 
> Error, return code 2 from org.apache.hadoop.hive.ql.exec.mr.MapRedTask
> 
> 
> Diagnostic Messages for this Task:
> Error: java.lang.RuntimeException: Hive Runtime Error while closing operators
>         at 
> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:240)
>         at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:61)
>         at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:429)
>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:341)
>         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:162)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:396)
>         at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
>         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:157)
> Caused by: java.lang.UnsupportedOperationException: This is supposed to be 
> overridden by subclasses.
>         at 
> com.google.protobuf.GeneratedMessage.getUnknownFields(GeneratedMessage.java:180)
>         at 
> org.apache.hadoop.hive.ql.io.orc.OrcProto$ColumnStatistics.getSerializedSize(OrcProto.java:3046)
>         at 
> com.google.protobuf.CodedOutputStream.computeMessageSizeNoTag(CodedOutputStream.java:749)
>         at 
> com.google.protobuf.CodedOutputStream.computeMessageSize(CodedOutputStream.java:530)
>         at 
> org.apache.hadoop.hive.ql.io.orc.OrcProto$RowIndexEntry.getSerializedSize(OrcProto.java:4129)
>         at 
> com.google.protobuf.CodedOutputStream.computeMessageSizeNoTag(CodedOutputStream.java:749)
>         at 
> com.google.protobuf.CodedOutputStream.computeMessageSize(CodedOutputStream.java:530)
>         at 
> org.apache.hadoop.hive.ql.io.orc.OrcProto$RowIndex.getSerializedSize(OrcProto.java:4641)
>         at 
> com.google.protobuf.AbstractMessageLite.writeTo(AbstractMessageLite.java:75)
>         at 
> org.apache.hadoop.hive.ql.io.orc.WriterImpl$TreeWriter.writeStripe(WriterImpl.java:548)
>         at 
> org.apache.hadoop.hive.ql.io.orc.WriterImpl$StructTreeWriter.writeStripe(WriterImpl.java:1328)
>         at 
> org.apache.hadoop.hive.ql.io.orc.WriterImpl.flushStripe(WriterImpl.java:1699)
>         at 
> org.apache.hadoop.hive.ql.io.orc.WriterImpl.close(WriterImpl.java:1868)
>         at 
> org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat$OrcRecordWriter.close(OrcOutputFormat.java:95)
>         at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator$FSPaths.closeWriters(FileSinkOperator.java:181)
>         at 
> org.apache.hadoop.hive.ql.exec.FileSinkOperator.closeOp(FileSinkOperator.java:866)
>         at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:596)
>         at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>         at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>         at org.apache.hadoop.hive.ql.exec.Operator.close(Operator.java:613)
>         at 
> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.close(ExecMapper.java:207)
>         ... 8 more


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.

Reply via email to