Sure.
That's how the ddl table looks like:
CREATETABLEtablea (
`a` BIGINT,
`b` BIGINT,
`c` BIGINT
)
COMMENT ''
WITH(
'auto-compaction'='false',
'connector'='filesystem',
'format'='parquet',
'parquet.block.size'='134217728',
'parquet.compression'='SNAPPY',
'parquet.dictionary.page.size'='1048576',
'parquet.enable.dictionary'='true',
'parquet.page.size'='1048576',
'parquet.writer.max-padding'='2097152',
'path'='s3a://test/test’,
'sink.partition-commit.delay'='1 h',
'sink.partition-commit.policy.kind'='success-file',
'sink.partition-commit.success-file.name
<http://sink.partition-commit.success-file.name>'='_SUCCESS',
'sink.partition-commit.trigger'='process-time',
'sink.rolling-policy.check-interval'='20 min',
'sink.rolling-policy.file-size'='128MB',
'sink.rolling-policy.rollover-interval'='2 h'
);
When a change the connector to a blackhole it immediately works without
errors. I have the redacted the names and paths.
Thanks,
Natu
On Thu, Jul 22, 2021 at 2:24 PM Timo Walther <twal...@apache.org
<mailto:twal...@apache.org>> wrote:
Maybe you can share also which connector/format you are using? What is
the DDL?
Regards,
Timo
On 22.07.21 14:11, Natu Lauchande wrote:
> Hey Timo,
>
> Thanks for the reply.
>
> No custom file as we are using Flink SQL and submitting the job
directly
> through the SQL Editor UI. We are using Flink 1.13.1 as the
supported
> flink version. No custom code all through Flink SQL on UI no jars.
>
> Thanks,
> Natu
>
> On Thu, Jul 22, 2021 at 2:08 PM Timo Walther <twal...@apache.org
<mailto:twal...@apache.org>
> <mailto:twal...@apache.org <mailto:twal...@apache.org>>> wrote:
>
> Hi Natu,
>
> Ververica Platform 2.5 has updated the bundled Hadoop version
but this
> should not result in a NoClassDefFoundError exception. How
are you
> submitting your SQL jobs? You don't use Ververica's SQL
service but
> have
> built a regular JAR file, right? If this is the case, can you
share
> your
> pom.xml file with us? The Flink version stays constant at 1.12?
>
> Regards,
> Timo
>
> On 22.07.21 12:22, Natu Lauchande wrote:
> > Good day Flink community,
> >
> > Apache Flink/Ververica Community Edition - Question
> >
> >
> > I am having an issue with my Flink SQL jobs since updating
> from Flink
> > 1.12/Ververica 2.4 to Ververica 2.5 . For all the jobs
running on
> > parquet and S3 i am getting the following error continuously:
> >
> > INITIALIZING to FAILED on 10.243.3.0:42337-2a3224 @
> > 10-243-3-0.flink-metrics.vvp-jobs.svc.cluster.local
(dataPort=39309).
> >
> > java.lang.NoClassDefFoundError:
org/apache/hadoop/conf/Configuration
> >
> > at java.lang.Class.getDeclaredConstructors0(Native Method)
> ~[?:1.8.0_292]
> >
> > at
java.lang.Class.privateGetDeclaredConstructors(Class.java:2671)
> > ~[?:1.8.0_292]
> >
> > at java.lang.Class.getDeclaredConstructors(Class.java:2020)
> ~[?:1.8.0_292]
> >
> > *....*
> >
> > at
java.io.ObjectInputStream.readObject(ObjectInputStream.java:461)
> > ~[?:1.8.0_292]
> >
> > at
> >
>
org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:615)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:600)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:587)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.util.InstantiationUtil.readObjectFromConfig(InstantiationUtil.java:541)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperatorFactory(StreamConfig.java:322)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperator(OperatorChain.java:653)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:626)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:616)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:616)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:616)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.OperatorChain.<init>(OperatorChain.java:181)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.StreamTask.executeRestore(StreamTask.java:548)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.StreamTask.runWithCleanUpOnFail(StreamTask.java:647)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:537)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:759)
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
org.apache.flink.runtime.taskmanager.Task.run(Task.java:566)
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_292]
> >
> > Caused by: java.lang.ClassNotFoundException:
> > org.apache.hadoop.conf.Configuration
> >
> > at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
> ~[?:1.8.0_292]
> >
> > at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
> ~[?:1.8.0_292]
> >
> > at
> >
>
org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:64)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at
> >
>
org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:48)
>
> > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >
> > at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
> ~[?:1.8.0_292]
> >
> > ... 57 more
> >
> > 2021-07-22 09:38:43,095 DEBUG
> > org.apache.flink.runtime.scheduler.SharedSlot[] - Remove
logical
> slot
> > (SlotRequestId{4297879e795d0516e36a7c26ccc795b2}) for
execution
> vertex
> > (id cbc357ccb763df2852fee8c4fc7d55f2_0) from the physical slot
> > (SlotRequestId{df7c49a6610b56f26aea214c05bcd9ed})
> >
> > 2021-07-22 09:38:43,096 DEBUG
> > org.apache.flink.runtime.scheduler.SharedSlot[] - Release
shared
> slot
> > externally (SlotRequestId{df7c49a6610b56f26aea214c05bcd9ed})
> >
> >
> > Everything works well when i roll back to Ververica v2.4,
has anyone
> > experienced this error before.
> >
> > Thanks,
> >
> > Natu
> >
>