[ https://issues.apache.org/jira/browse/FLINK-31811?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17712390#comment-17712390 ]
Krzysztof Chmielewski commented on FLINK-31811: ----------------------------------------------- I think this might be a duplicate of https://issues.apache.org/jira/browse/FLINK-31202 that manifest in SQL API. P.S. Are you sure that this was working in 1.15.1? I know that writing "simple" complex types like Array of Intigers or Strings bBut not sure about this one -> Arrays Of Map. > Unsupported complex data type for Flink SQL > ------------------------------------------- > > Key: FLINK-31811 > URL: https://issues.apache.org/jira/browse/FLINK-31811 > Project: Flink > Issue Type: Bug > Components: Connectors / FileSystem > Affects Versions: 1.16.1 > Reporter: jirawech.s > Priority: Major > Fix For: 1.16.2 > > > I found this issue when I tried to write data on local filesystem using Flink > SQL > {code:java} > 19:51:32,966 INFO org.apache.flink.runtime.executiongraph.ExecutionGraph > [] - compact-operator (1/4) > (4f2a09b638c786f74262c675d248afd9_80fe6c4f32f605d447b391cdb16cc1ff_0_4) > switched from RUNNING to FAILED on 69ed2306-371b-4bfc-a98e-bf75fb41748f @ > localhost (dataPort=-1). > java.lang.IndexOutOfBoundsException: Index: 1, Size: 1 > at java.util.ArrayList.rangeCheck(ArrayList.java:659) ~[?:1.8.0_301] > at java.util.ArrayList.get(ArrayList.java:435) ~[?:1.8.0_301] > at org.apache.parquet.schema.GroupType.getType(GroupType.java:216) > ~[parquet-column-1.12.2.jar:1.12.2] > at > org.apache.flink.formats.parquet.vector.ParquetSplitReaderUtil.createWritableColumnVector(ParquetSplitReaderUtil.java:523) > ~[flink-parquet-1.16.1.jar:1.16.1] > at > org.apache.flink.formats.parquet.vector.ParquetSplitReaderUtil.createWritableColumnVector(ParquetSplitReaderUtil.java:503) > ~[flink-parquet-1.16.1.jar:1.16.1] > at > org.apache.flink.formats.parquet.ParquetVectorizedInputFormat.createWritableVectors(ParquetVectorizedInputFormat.java:281) > ~[flink-parquet-1.16.1.jar:1.16.1] > at > org.apache.flink.formats.parquet.ParquetVectorizedInputFormat.createReaderBatch(ParquetVectorizedInputFormat.java:270) > ~[flink-parquet-1.16.1.jar:1.16.1] > at > org.apache.flink.formats.parquet.ParquetVectorizedInputFormat.createPoolOfBatches(ParquetVectorizedInputFormat.java:260) > ~[flink-parquet-1.16.1.jar:1.16.1] > {code} > What i tried to do is writing complex data type to parquet file > Here is the schema of sink table. The problematic data type is > ARRAY<MAP<STRING,STRING>> > {code:java} > CREATE TEMPORARY TABLE local_table ( > `user_id` STRING, `order_id` STRING, `amount` INT, `restaurant_id` STRING, > `experiment` ARRAY<MAP<STRING,STRING>>, `dt` STRING > ) PARTITIONED BY (`dt`) WITH ( > 'connector'='filesystem', > 'path'='file:///tmp/test_hadoop_write', > 'format'='parquet', > 'auto-compaction'='true', > 'sink.partition-commit.policy.kind'='success-file' > ) {code} > PS. It is used to work in Flink version 1.15.1 > -- This message was sent by Atlassian Jira (v8.20.10#820010)