Sure.

That's how the ddl table looks like:

CREATE TABLE tablea (

  `a` BIGINT,

  `b` BIGINT,

  `c` BIGINT

)

COMMENT ''

WITH (

  'auto-compaction' = 'false',

  'connector' = 'filesystem',

  'format' = 'parquet',

  'parquet.block.size' = '134217728',

  'parquet.compression' = 'SNAPPY',

  'parquet.dictionary.page.size' = '1048576',

  'parquet.enable.dictionary' = 'true',

  'parquet.page.size' = '1048576',

  'parquet.writer.max-padding' = '2097152',

  'path' = 's3a://test/test’,

  'sink.partition-commit.delay' = '1 h',

  'sink.partition-commit.policy.kind' = 'success-file',

  'sink.partition-commit.success-file.name' = '_SUCCESS',

  'sink.partition-commit.trigger' = 'process-time',

  'sink.rolling-policy.check-interval' = '20 min',

  'sink.rolling-policy.file-size' = '128MB',

  'sink.rolling-policy.rollover-interval' = '2 h'

);



When a change the connector to a blackhole it immediately works without
errors. I have the redacted the names and paths.



Thanks,
Natu

On Thu, Jul 22, 2021 at 2:24 PM Timo Walther <twal...@apache.org> wrote:

> Maybe you can share also which connector/format you are using? What is
> the DDL?
>
> Regards,
> Timo
>
>
> On 22.07.21 14:11, Natu Lauchande wrote:
> > Hey Timo,
> >
> > Thanks for the reply.
> >
> > No custom file as we are using Flink SQL and submitting the job directly
> > through the SQL Editor UI. We are using Flink 1.13.1 as the supported
> > flink version. No custom code all through Flink SQL on UI no jars.
> >
> > Thanks,
> > Natu
> >
> > On Thu, Jul 22, 2021 at 2:08 PM Timo Walther <twal...@apache.org
> > <mailto:twal...@apache.org>> wrote:
> >
> >     Hi Natu,
> >
> >     Ververica Platform 2.5 has updated the bundled Hadoop version but
> this
> >     should not result in a NoClassDefFoundError exception. How are you
> >     submitting your SQL jobs? You don't use Ververica's SQL service but
> >     have
> >     built a regular JAR file, right? If this is the case, can you share
> >     your
> >     pom.xml file with us? The Flink version stays constant at 1.12?
> >
> >     Regards,
> >     Timo
> >
> >     On 22.07.21 12:22, Natu Lauchande wrote:
> >      > Good day Flink community,
> >      >
> >      > Apache Flink/Ververica Community Edition - Question
> >      >
> >      >
> >      > I am having an issue with my Flink SQL jobs since updating
> >     from Flink
> >      > 1.12/Ververica 2.4 to Ververica 2.5 . For all the jobs running on
> >      > parquet and S3 i am getting the following error continuously:
> >      >
> >      > INITIALIZING to FAILED on 10.243.3.0:42337-2a3224 @
> >      > 10-243-3-0.flink-metrics.vvp-jobs.svc.cluster.local
> (dataPort=39309).
> >      >
> >      > java.lang.NoClassDefFoundError:
> org/apache/hadoop/conf/Configuration
> >      >
> >      > at java.lang.Class.getDeclaredConstructors0(Native Method)
> >     ~[?:1.8.0_292]
> >      >
> >      > at java.lang.Class.privateGetDeclaredConstructors(Class.java:2671)
> >      > ~[?:1.8.0_292]
> >      >
> >      > at java.lang.Class.getDeclaredConstructors(Class.java:2020)
> >     ~[?:1.8.0_292]
> >      >
> >      > *....*
> >      >
> >      > at
> java.io.ObjectInputStream.readObject(ObjectInputStream.java:461)
> >      > ~[?:1.8.0_292]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:615)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:600)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:587)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.util.InstantiationUtil.readObjectFromConfig(InstantiationUtil.java:541)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperatorFactory(StreamConfig.java:322)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperator(OperatorChain.java:653)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:626)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:616)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:616)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOperatorChain(OperatorChain.java:616)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.createOutputCollector(OperatorChain.java:566)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.OperatorChain.<init>(OperatorChain.java:181)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.StreamTask.executeRestore(StreamTask.java:548)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.StreamTask.runWithCleanUpOnFail(StreamTask.java:647)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:537)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:759)
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:566)
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at java.lang.Thread.run(Thread.java:748) ~[?:1.8.0_292]
> >      >
> >      > Caused by: java.lang.ClassNotFoundException:
> >      > org.apache.hadoop.conf.Configuration
> >      >
> >      > at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
> >     ~[?:1.8.0_292]
> >      >
> >      > at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
> >     ~[?:1.8.0_292]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:64)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at
> >      >
> >
>  
> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:48)
> >
> >      > ~[flink-dist_2.12-1.13.1-stream1.jar:1.13.1-stream1[]
> >      >
> >      > at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
> >     ~[?:1.8.0_292]
> >      >
> >      > ... 57 more
> >      >
> >      > 2021-07-22 09:38:43,095 DEBUG
> >      > org.apache.flink.runtime.scheduler.SharedSlot[] - Remove logical
> >     slot
> >      > (SlotRequestId{4297879e795d0516e36a7c26ccc795b2}) for execution
> >     vertex
> >      > (id cbc357ccb763df2852fee8c4fc7d55f2_0) from the physical slot
> >      > (SlotRequestId{df7c49a6610b56f26aea214c05bcd9ed})
> >      >
> >      > 2021-07-22 09:38:43,096 DEBUG
> >      > org.apache.flink.runtime.scheduler.SharedSlot[] - Release shared
> >     slot
> >      > externally (SlotRequestId{df7c49a6610b56f26aea214c05bcd9ed})
> >      >
> >      >
> >      > Everything works well when i roll back to Ververica v2.4, has
> anyone
> >      > experienced this error before.
> >      >
> >      > Thanks,
> >      >
> >      > Natu
> >      >
> >
>
>

Reply via email to