Hi Rommel, Hi Thomas, Apache Parquet was bumped from 1.10.0 to 1.11.1 for Flink 1.12 in FLINK-19137 [1]. The error you're seeing looks like some dependency issue where you have a version other than 1.11.1 of org.apache.parquet:parquet-column:jar on your classpath?
Matthias [1] https://issues.apache.org/jira/browse/FLINK-19137 On Wed, Jun 23, 2021 at 1:50 AM Rommel Holmes <rommelhol...@gmail.com> wrote: > To give more information > > parquet-avro version 1.10.0 with Flink 1.11.2 and it was running fine. > > now Flink 1.12.1, the error msg shows up. > > Thank you for help. > > Rommel > > > > > > On Tue, Jun 22, 2021 at 2:41 PM Thomas Wang <w...@datability.io> wrote: > >> Hi, >> >> We recently upgraded our Flink version from 1.11.2 to 1.12.1 and one of >> our jobs that used to run ok, now sees the following error. This error >> doesn't seem to be related to any user code. Can someone help me take a >> look? >> >> Thanks. >> >> Thomas >> >> java.lang.NoSuchMethodError: >> org.apache.parquet.column.ParquetProperties.getColumnIndexTruncateLength()I >> at org.apache.parquet.hadoop.ParquetWriter.<init>(ParquetWriter.java:282) >> ~[?:?] >> at >> org.apache.parquet.hadoop.ParquetWriter$Builder.build(ParquetWriter.java:564) >> ~[?:?] >> at >> org.apache.flink.formats.parquet.avro.ParquetAvroWriters.createAvroParquetWriter(ParquetAvroWriters.java:90) >> ~[?:?] >> at >> org.apache.flink.formats.parquet.avro.ParquetAvroWriters.lambda$forGenericRecord$abd75386$1(ParquetAvroWriters.java:65) >> ~[?:?] >> at >> org.apache.flink.formats.parquet.ParquetWriterFactory.create(ParquetWriterFactory.java:56) >> ~[?:?] >> at >> org.apache.flink.streaming.api.functions.sink.filesystem.BulkBucketWriter.openNew(BulkBucketWriter.java:75) >> ~[flink-table-blink_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.functions.sink.filesystem.OutputStreamBasedPartFileWriter$OutputStreamBasedBucketWriter.openNewInProgressFile(OutputStreamBasedPartFileWriter.java:90) >> ~[flink-table-blink_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.functions.sink.filesystem.BulkBucketWriter.openNewInProgressFile(BulkBucketWriter.java:36) >> ~[flink-table-blink_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.functions.sink.filesystem.Bucket.rollPartFile(Bucket.java:243) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.functions.sink.filesystem.Bucket.write(Bucket.java:220) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.onElement(Buckets.java:305) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSinkHelper.onElement(StreamingFileSinkHelper.java:103) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.invoke(StreamingFileSink.java:492) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.StreamSink.processElement(StreamSink.java:54) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:71) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:46) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:26) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.BroadcastingOutputCollector.collect(BroadcastingOutputCollector.java:75) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.BroadcastingOutputCollector.collect(BroadcastingOutputCollector.java:32) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:38) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:71) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:46) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:26) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.operators.TimestampsAndWatermarksOperator.processElement(TimestampsAndWatermarksOperator.java:104) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:71) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:46) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:26) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollectWithTimestamp(StreamSourceContexts.java:322) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collectWithTimestamp(StreamSourceContexts.java:426) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.connectors.kafka.internals.AbstractFetcher.emitRecordsWithTimestamps(AbstractFetcher.java:365) >> ~[?:?] >> at >> org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.partitionConsumerRecordsHandler(KafkaFetcher.java:183) >> ~[?:?] >> at >> org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.runFetchLoop(KafkaFetcher.java:142) >> ~[?:?] >> at >> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:826) >> ~[?:?] >> at >> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:110) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:66) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> at >> org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:241) >> ~[flink-dist_2.12-1.12.1.jar:1.12.1] >> > > > -- > Yours > Rommel > ************************************* > I waited patiently for the LORD; > he turned to me and heard my cry. > He lifted me out of the slimy pit, > out of the mud and mire; > he set my feet on a rock > and gave me a firm place to stand. > ************************************* >