[ 
https://issues.apache.org/jira/browse/HIVE-8909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14220487#comment-14220487
 ] 

Brock Noland commented on HIVE-8909:
------------------------------------

Not sure which test this is from:

{noformat}
Caused by: parquet.io.ParquetDecodingException: Can not read value at 0 in 
block 0 in file 
pfile:/Users/noland/workspaces/hive-apache/hive/itests/qtest/target/warehouse/parquet_jointable2/000000_0
  at 
parquet.hadoop.InternalParquetRecordReader.nextKeyValue(InternalParquetRecordReader.java:213)
  at 
parquet.hadoop.ParquetRecordReader.nextKeyValue(ParquetRecordReader.java:204)
  at 
org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper.<init>(ParquetRecordReaderWrapper.java:102)
  at 
org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper.<init>(ParquetRecordReaderWrapper.java:71)
  at 
org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat.getRecordReader(MapredParquetInputFormat.java:71)
  at 
org.apache.hadoop.hive.ql.io.CombineHiveRecordReader.<init>(CombineHiveRecordReader.java:65)
  ... 16 more
Caused by: java.lang.ArrayIndexOutOfBoundsException: 2
  at 
org.apache.hadoop.hive.ql.io.parquet.convert.HiveStructConverter.set(HiveStructConverter.java:96)
  at 
org.apache.hadoop.hive.ql.io.parquet.convert.ETypeConverter$BinaryConverter.addBinary(ETypeConverter.java:219)
  at 
parquet.column.impl.ColumnReaderImpl$2$6.writeValue(ColumnReaderImpl.java:306)
  at 
parquet.column.impl.ColumnReaderImpl.writeCurrentValueToConverter(ColumnReaderImpl.java:353)
  at 
parquet.io.RecordReaderImplementation.read(RecordReaderImplementation.java:402)
  at 
parquet.hadoop.InternalParquetRecordReader.nextKeyValue(InternalParquetRecordReader.java:194)
  ... 21 more
{noformat}

All failed parquet tests with the patch:
{noformat}
  <testcase name="testCliDriver_parquet_array_null_element" 
classname="org.apache.hadoop.hive.cli.TestCliDriver" time="4.945">
  <testcase name="testCliDriver_parquet_create" 
classname="org.apache.hadoop.hive.cli.TestCliDriver" time="4.416">
  <testcase name="testCliDriver_parquet_decimal" 
classname="org.apache.hadoop.hive.cli.TestCliDriver" time="5.478">
  <testcase name="testCliDriver_parquet_join" 
classname="org.apache.hadoop.hive.cli.TestCliDriver" time="8.928">
  <testcase name="testCliDriver_parquet_types" 
classname="org.apache.hadoop.hive.cli.TestCliDriver" time="4.094">
{noformat}

> Hive doesn't correctly read Parquet nested types
> ------------------------------------------------
>
>                 Key: HIVE-8909
>                 URL: https://issues.apache.org/jira/browse/HIVE-8909
>             Project: Hive
>          Issue Type: Bug
>    Affects Versions: 0.13.1
>            Reporter: Ryan Blue
>            Assignee: Ryan Blue
>         Attachments: HIVE-8909-1.patch, HIVE-8909-2.patch, HIVE-8909.2.patch, 
> HIVE-8909.3.patch, parquet-test-data.tar.gz
>
>
> Parquet's Avro and Thrift object models don't produce the same parquet type 
> representation for lists and maps that Hive does. In the Parquet community, 
> we've defined what should be written and backward-compatibility rules for 
> existing data written by parquet-avro and parquet-thrift in PARQUET-113. We 
> need to implement those rules in the Hive Converter classes.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to