Hi, I am using Hive 0.12 with Hadoop 2.2 and trying to insert data in a new ORC table with an INSERT SELECT statement from a TEXT file based table and I am running into the following error (I have trimmed some of the data showed in the error):
Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: Hive Runtime Error while processing row {"id":"1932685422","ad_id":"7325801318", .... , "account_id":"6875965212"} at org.apache.hadoop.hive.ql.exec.MapOperator.process(MapOperator.java:544) at org.apache.hadoop.hive.ql.exec.mr.ExecMapper.map(ExecMapper.java:157) ... 8 more Caused by: java.lang.ArrayIndexOutOfBoundsException: 26 at org.apache.hadoop.hive.ql.io.orc.RunLengthIntegerWriterV2.preparePatchedBlob(RunLengthIntegerWriterV2.java:593) at org.apache.hadoop.hive.ql.io.orc.RunLengthIntegerWriterV2.determineEncoding(RunLengthIntegerWriterV2.java:541) at org.apache.hadoop.hive.ql.io.orc.RunLengthIntegerWriterV2.write(RunLengthIntegerWriterV2.java:797) at org.apache.hadoop.hive.ql.io.orc.WriterImpl$IntegerTreeWriter.write(WriterImpl.java:744) ... That error is produced when the ad_id column in the destination table has the type BIGINT. When I change the column type to STRING the insert works fine. >From what I see that value is no nearly big enough to cause any overflow issues in a BIGINT. Is this a known bug or do I have to do anything in particular for this to work? Thanks, Juan.