Sure. That's how the ddl table looks like:
CREATE TABLE tablea ( `a` BIGINT, `b` BIGINT, `c` BIGINT ) COMMENT '' WITH ( 'auto-compaction' = 'false', 'connector' = 'filesystem', 'format' = 'parquet', 'parquet.block.size' = '134217728', 'parquet.compression' = 'SNAPPY', 'parquet.dictionary.page.size' = '1048576', 'parquet.enable.dictionary' = 'true', 'parquet.page.size' = '1048576', 'parquet.writer.max-padding' = '2097152', 'path' = 's3a://test/test’, 'sink.partition-commit.delay' = '1 h', 'sink.partition-commit.policy.kind' = 'success-file', 'sink.partition-commit.success-file.name' = '_SUCCESS', 'sink.partition-commit.trigger' = 'process-time', 'sink.rolling-policy.check-interval' = '20 min', 'sink.rolling-policy.file-size' = '128MB', 'sink.rolling-policy.rollover-interval' = '2 h' ); When a change the connector to a blackhole it immediately works without errors. I have the redacted the names and paths. Thanks, Natu On Thu, Jul 22, 2021 at 2:24 PM Timo Walther <twal...@apache.org> wrote: > Maybe you can share also which connector/format you are using? What is > the DDL? > > Regards, > Timo > > > On 22.07.21 14:11, Natu Lauchande wrote: > > Hey Timo, > > > > Thanks for the reply. > > > > No custom file as we are using Flink SQL and submitting the job directly > > through the SQL Editor UI. We are using Flink 1.13.1 as the supported > > flink version. No custom code all through Flink SQL on UI no jars. > > > > Thanks, > > Natu > > > > On Thu, Jul 22, 2021 at 2:08 PM Timo Walther <twal...@apache.org > > <mailto:twal...@apache.org>> wrote: > > > > Hi Natu, > > > > Ververica Platform 2.5 has updated the bundled Hadoop version but > this > > should not result in a NoClassDefFoundError exception. How are you > > submitting your SQL jobs? You don't use Ververica's SQL service but > > have > > built a regular JAR file, right? If this is the case, can you share > > your > > pom.xml file with us? The Flink version stays constant at 1.12? > > > > Regards, > > Timo > > > > On 22.07.21 12:22, Natu Lauchande wrote: > > > Good day Flink community, > > > > > > Apache Flink/Ververica Community Edition - Question > > > > > > > > > I am having an issue with my Flink SQL jobs since updating > > from Flink > > > 1.12/Ververica 2.4 to Ververica 2.5 . For all the jobs running on > > > parquet and S3 i am getting the following error continuously: > > > > > > INITIALIZING to FAILED on 10.243.3.0:42337-2a3224 @ > > > 10-243-3-0.flink-metrics.vvp-jobs.svc.cluster.local > (dataPort=39309). > > > > > > java.lang.NoClassDefFoundError: > org/apache/hadoop/conf/Configuration > > > > > > at java.lang.Class.getDeclaredConstructors0(Native Method) > > ~[?:1.8.0_292] > > > > > > at java.lang.Class.privateGetDeclaredConstructors(Class.java:2671) > > > ~[?:1.8.0_292] > > > > > > at java.lang.Class.getDeclaredConstructors(Class.java:2020) > > ~[?:1.8.0_292] > > > > > > *....* > > > > > > at > java.io.ObjectInputStream.readObject(ObjectInputStream.java:461) > > > ~[?:1.8.0_292] > > > > > > at > > > > > > > org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:615) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:600) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:587) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.util.InstantiationUtil.readObjectFromConfig(InstantiationUtil.java:541) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperatorFactory(StreamConfig.java:322) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperator(OperatorChain.java:653) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:626) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:616) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:616) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:616) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.OperatorChain.<init>(OperatorChain.java:181) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.StreamTask.executeRestore(StreamTask.java:548) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.StreamTask.runWithCleanUpOnFail(StreamTask.java:647) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:537) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:759) > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:566) > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_292] > > > > > > Caused by: java.lang.ClassNotFoundException: > > > org.apache.hadoop.conf.Configuration > > > > > > at java.net.URLClassLoader.findClass(URLClassLoader.java:382) > > ~[?:1.8.0_292] > > > > > > at java.lang.ClassLoader.loadClass(ClassLoader.java:418) > > ~[?:1.8.0_292] > > > > > > at > > > > > > > org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:64) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at > > > > > > > org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:48) > > > > > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[] > > > > > > at java.lang.ClassLoader.loadClass(ClassLoader.java:351) > > ~[?:1.8.0_292] > > > > > > ... 57 more > > > > > > 2021-07-22 09:38:43,095 DEBUG > > > org.apache.flink.runtime.scheduler.SharedSlot[] - Remove logical > > slot > > > (SlotRequestId{4297879e795d0516e36a7c26ccc795b2}) for execution > > vertex > > > (id cbc357ccb763df2852fee8c4fc7d55f2_0) from the physical slot > > > (SlotRequestId{df7c49a6610b56f26aea214c05bcd9ed}) > > > > > > 2021-07-22 09:38:43,096 DEBUG > > > org.apache.flink.runtime.scheduler.SharedSlot[] - Release shared > > slot > > > externally (SlotRequestId{df7c49a6610b56f26aea214c05bcd9ed}) > > > > > > > > > Everything works well when i roll back to Ververica v2.4, has > anyone > > > experienced this error before. > > > > > > Thanks, > > > > > > Natu > > > > > > >