Hi Rommel, Hi Thomas,
Apache Parquet was bumped from 1.10.0 to 1.11.1 for Flink 1.12 in
FLINK-19137 [1]. The error you're seeing looks like some dependency issue
where you have a version other than 1.11.1
of org.apache.parquet:parquet-column:jar on your classpath?

Matthias

[1] https://issues.apache.org/jira/browse/FLINK-19137

On Wed, Jun 23, 2021 at 1:50 AM Rommel Holmes <rommelhol...@gmail.com>
wrote:

> To give more information
>
> parquet-avro version 1.10.0 with Flink 1.11.2 and it was running fine.
>
> now Flink 1.12.1, the error msg shows up.
>
> Thank you for help.
>
> Rommel
>
>
>
>
>
> On Tue, Jun 22, 2021 at 2:41 PM Thomas Wang <w...@datability.io> wrote:
>
>> Hi,
>>
>> We recently upgraded our Flink version from 1.11.2 to 1.12.1 and one of
>> our jobs that used to run ok, now sees the following error. This error
>> doesn't seem to be related to any user code. Can someone help me take a
>> look?
>>
>> Thanks.
>>
>> Thomas
>>
>> java.lang.NoSuchMethodError:
>> org.apache.parquet.column.ParquetProperties.getColumnIndexTruncateLength()I
>> at org.apache.parquet.hadoop.ParquetWriter.<init>(ParquetWriter.java:282)
>> ~[?:?]
>> at
>> org.apache.parquet.hadoop.ParquetWriter$Builder.build(ParquetWriter.java:564)
>> ~[?:?]
>> at
>> org.apache.flink.formats.parquet.avro.ParquetAvroWriters.createAvroParquetWriter(ParquetAvroWriters.java:90)
>> ~[?:?]
>> at
>> org.apache.flink.formats.parquet.avro.ParquetAvroWriters.lambda$forGenericRecord$abd75386$1(ParquetAvroWriters.java:65)
>> ~[?:?]
>> at
>> org.apache.flink.formats.parquet.ParquetWriterFactory.create(ParquetWriterFactory.java:56)
>> ~[?:?]
>> at
>> org.apache.flink.streaming.api.functions.sink.filesystem.BulkBucketWriter.openNew(BulkBucketWriter.java:75)
>> ~[flink-table-blink_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.functions.sink.filesystem.OutputStreamBasedPartFileWriter$OutputStreamBasedBucketWriter.openNewInProgressFile(OutputStreamBasedPartFileWriter.java:90)
>> ~[flink-table-blink_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.functions.sink.filesystem.BulkBucketWriter.openNewInProgressFile(BulkBucketWriter.java:36)
>> ~[flink-table-blink_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.functions.sink.filesystem.Bucket.rollPartFile(Bucket.java:243)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.functions.sink.filesystem.Bucket.write(Bucket.java:220)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.onElement(Buckets.java:305)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSinkHelper.onElement(StreamingFileSinkHelper.java:103)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.invoke(StreamingFileSink.java:492)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.StreamSink.processElement(StreamSink.java:54)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:71)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:46)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:26)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.BroadcastingOutputCollector.collect(BroadcastingOutputCollector.java:75)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.BroadcastingOutputCollector.collect(BroadcastingOutputCollector.java:32)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.StreamMap.processElement(StreamMap.java:38)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:71)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:46)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:26)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.operators.TimestampsAndWatermarksOperator.processElement(TimestampsAndWatermarksOperator.java:104)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.pushToOperator(CopyingChainingOutput.java:71)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:46)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.CopyingChainingOutput.collect(CopyingChainingOutput.java:26)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:50)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.CountingOutput.collect(CountingOutput.java:28)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.StreamSourceContexts$ManualWatermarkContext.processAndCollectWithTimestamp(StreamSourceContexts.java:322)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.StreamSourceContexts$WatermarkContext.collectWithTimestamp(StreamSourceContexts.java:426)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.connectors.kafka.internals.AbstractFetcher.emitRecordsWithTimestamps(AbstractFetcher.java:365)
>> ~[?:?]
>> at
>> org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.partitionConsumerRecordsHandler(KafkaFetcher.java:183)
>> ~[?:?]
>> at
>> org.apache.flink.streaming.connectors.kafka.internals.KafkaFetcher.runFetchLoop(KafkaFetcher.java:142)
>> ~[?:?]
>> at
>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumerBase.run(FlinkKafkaConsumerBase.java:826)
>> ~[?:?]
>> at
>> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:110)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:66)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>> at
>> org.apache.flink.streaming.runtime.tasks.SourceStreamTask$LegacySourceFunctionThread.run(SourceStreamTask.java:241)
>> ~[flink-dist_2.12-1.12.1.jar:1.12.1]
>>
>
>
> --
>      Yours
>      Rommel
> *************************************
>   I  waited patiently for the LORD;
>    he turned to me and heard my cry.
>  He lifted me out of the slimy pit,
>    out of the mud and mire;
> he set my feet on a rock
>    and gave me a firm place to stand.
> *************************************
>

Reply via email to