<project xmlns="http://maven.apache.org/POM/4.0.0"; xmlns:xsi="
http://www.w3.org/2001/XMLSchema-instance";
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0
http://maven.apache.org/xsd/maven-4.0.0.xsd";>
<modelVersion>4.0.0</modelVersion>

<groupId>boars_head</groupId>
<artifactId>w_build_d</artifactId>
<version>0.1</version>
<packaging>jar</packaging>

<name>Cql Flink DataStream Java</name>
<url>https://flink.apache.org</url>

<properties>
<project.build.sourceEncoding>UTF-8</project.build.sourceEncoding>
<flink.version>1.14.0</flink.version>
<target.java.version>1.8</target.java.version>
<scala.binary.version>2.11</scala.binary.version>
<maven.compiler.source>${target.java.version}</maven.compiler.source>
<maven.compiler.target>${target.java.version}</maven.compiler.target>
<log4j.version>2.14.1</log4j.version>
<clickhouse-native-jdbc.version>2.6.2</clickhouse-native-jdbc.version>
</properties>

<repositories>
<repository>
<id>apache.snapshots</id>
<name>Apache Development Snapshot Repository</name>
<url>https://repository.apache.org/content/repositories/snapshots/</url>
<releases>
<enabled>false</enabled>
</releases>
<snapshots>
<enabled>true</enabled>
</snapshots>
</repository>
<repository>
<id>central</id>
<name>Central Repository</name>
<url>https://repo1.maven.org/maven2/</url>
<releases>
<enabled>false</enabled>
</releases>
<snapshots>
<enabled>true</enabled>
</snapshots>
</repository>
</repositories>

<dependencies>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-walkthrough-common_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
</dependency>

<!-- This dependency is provided, because it should not be packaged into
the JAR file. -->
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-streaming-java_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
<scope>provided</scope>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-table-api-java-bridge_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
<scope>provided</scope>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-clients_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
<scope>provided</scope>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-table-planner_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
<scope>provided</scope>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-streaming-scala_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
<scope>provided</scope>
</dependency>

<!-- Add connector dependencies here. They must be in the default scope
(compile). -->
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-connector-kafka_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-avro-confluent-registry</artifactId>
<version>${flink.version}</version>
</dependency>
<!-- <dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-sql-avro-confluent-registry</artifactId>
<version>${flink.version}</version>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-sql-connector-kafka</artifactId>
<version>${flink.version}</version>
</dependency> -->
<dependency>
<groupId>com.github.housepower</groupId>
<artifactId>clickhouse-native-jdbc-shaded</artifactId>
<version>${clickhouse-native-jdbc.version}</version>
</dependency>
<dependency>
<groupId>org.apache.flink</groupId>
<artifactId>flink-connector-jdbc_${scala.binary.version}</artifactId>
<version>${flink.version}</version>
</dependency>
<!-- Add logging framework, to produce console output when running in the
IDE. -->
<!-- These dependencies are excluded from the application JAR by default.
-->
<dependency>
<groupId>org.apache.logging.log4j</groupId>
<artifactId>log4j-slf4j-impl</artifactId>
<version>${log4j.version}</version>
<scope>runtime</scope>
</dependency>
<dependency>
<groupId>org.apache.logging.log4j</groupId>
<artifactId>log4j-api</artifactId>
<version>${log4j.version}</version>
<scope>runtime</scope>
</dependency>
<dependency>
<groupId>org.apache.logging.log4j</groupId>
<artifactId>log4j-core</artifactId>
<version>${log4j.version}</version>
<scope>runtime</scope>
</dependency>
</dependencies>

<build>
<plugins>

<!-- Java Compiler -->
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>3.1</version>
<configuration>
<source>${target.java.version}</source>
<target>${target.java.version}</target>
</configuration>
</plugin>

<!-- We use the maven-shade plugin to create a fat jar that contains all
necessary dependencies. -->
<!-- Change the value of <mainClass>...</mainClass> if your program entry
point changes. -->
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-shade-plugin</artifactId>
<version>3.0.0</version>
<executions>
<!-- Run shade goal on package phase -->
<execution>
<phase>package</phase>
<goals>
<goal>shade</goal>
</goals>
<configuration>
<artifactSet>
<excludes>
<exclude>org.apache.flink:flink-shaded-force-shading</exclude>
<exclude>com.google.code.findbugs:jsr305</exclude>
<exclude>org.slf4j:*</exclude>
<exclude>org.apache.logging.log4j:*</exclude>
</excludes>
</artifactSet>
<filters>
<filter>
<!-- Do not copy the signatures in the META-INF folder.
Otherwise, this might cause SecurityExceptions when using the JAR. -->
<artifact>*:*</artifact>
<excludes>
<exclude>META-INF/*.SF</exclude>
<exclude>META-INF/*.DSA</exclude>
<exclude>META-INF/*.RSA</exclude>
</excludes>
</filter>
</filters>
<transformers>
<transformer implementation=
"org.apache.maven.plugins.shade.resource.ManifestResourceTransformer">
<mainClass>carrera.build.BuildFlinkJob</mainClass>
</transformer>
</transformers>
</configuration>
</execution>
</executions>
</plugin>
</plugins>

<pluginManagement>
<plugins>

<!-- This improves the out-of-the-box experience in Eclipse by resolving
some warnings. -->
<plugin>
<groupId>org.eclipse.m2e</groupId>
<artifactId>lifecycle-mapping</artifactId>
<version>1.0.0</version>
<configuration>
<lifecycleMappingMetadata>
<pluginExecutions>
<pluginExecution>
<pluginExecutionFilter>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-shade-plugin</artifactId>
<versionRange>[3.0.0,)</versionRange>
<goals>
<goal>shade</goal>
</goals>
</pluginExecutionFilter>
<action>
<ignore/>
</action>
</pluginExecution>
<pluginExecution>
<pluginExecutionFilter>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<versionRange>[3.1,)</versionRange>
<goals>
<goal>testCompile</goal>
<goal>compile</goal>
</goals>
</pluginExecutionFilter>
<action>
<ignore/>
</action>
</pluginExecution>
</pluginExecutions>
</lifecycleMappingMetadata>
</configuration>
</plugin>
</plugins>
</pluginManagement>
</build>
</project>


Above is the POM.

Here is the entry in our docker-compose where we drop the jars into the
opt/flink/lib folder:

volumes:
-
'./connectors/flink-sql-avro-confluent-registry-1.14.0.jar:/opt/flink/lib/flink-sql-avro-confluent-registry-1.14.0.jar'
-
'./connectors/flink-sql-connector-kafka_2.11-1.14.0.jar:/opt/flink/lib/flink-sql-connector-kafka_2.11-1.14.0.jar'

Thank you


Jason Politis
Solutions Architect, Carrera Group
carrera.io
| jpoli...@carrera.io <kpatter...@carrera.io>
<http://us.linkedin.com/in/jasonpolitis>


On Mon, Jan 10, 2022 at 9:31 PM Caizhi Weng <tsreape...@gmail.com> wrote:

> Hi!
>
> Could you share your pom.xml file of your user project? Did you include
> the flink-avro dependency? Also did you add the avro format jar to the lib
> directory of your Flink distribution?
>
> Jason Politis <jpoli...@carrera.io> 于2022年1月11日周二 08:42写道:
>
>> Good evening all,
>>
>> I'm working on a project for a client.  We are trying to execute Flink
>> SQL using Table API in java.
>> We are going to pull their data from oracle -> debezium -> kafka -> flink.
>>
>>
>> Here is a sample of our java code:
>>
>> package carrera.build;
>>
>> import org.apache.flink.table.api.EnvironmentSettings;
>> import org.apache.flink.table.api.Table;
>> import org.apache.flink.table.api.TableEnvironment;
>>
>> public class BuildFlinkJob {
>>     public static void main(String[] args) throws Exception {
>>         EnvironmentSettings settings = EnvironmentSettings.inStreamingMode();
>>         TableEnvironment tEnv = TableEnvironment.create(settings);
>>
>>         tEnv.executeSql(
>>                 "CREATE TABLE BUILDS (\n" +
>>                 "    `PARTITION` INT METADATA FROM 'partition',\n" +
>>                 "    `OFFSET` BIGINT METADATA FROM 'offset',\n" +
>>                 "    BUILD_ID DOUBLE,\n" +
>>                 "    BUILD_NAME STRING,\n" +
>>                 "    FACILITY_NUMBER STRING,\n" +
>>                 "    START_DATE TIMESTAMP(2),\n" +
>>                 "    END_DATE TIMESTAMP(2),\n" +
>>                 "    RETAILERDIVISION_NAME STRING,\n" +
>>                 "    UPC STRING,\n" +
>>                 "    BUILD_INSTRUCTIONS STRING,\n" +
>>                 "    WORK_INSTRUCTIONS STRING,\n" +
>>                 "    IMAGE_FILE_PATH STRING\n" +
>>                 ") WITH (\n" +
>>                 "    'connector' = 'kafka',\n" +
>>                 "    'topic' = 
>> 'clients-name.OBIANEW_SDS_EBS_12_1_3.BUILDS',\n" +
>>                 "    'properties.bootstrap.servers' = 'broker:29092',\n" +
>>                 "    'properties.group.id' = 'builds',\n" +
>>                 "    'format' = 'debezium-avro-confluent',\n" +
>>                 "    'debezium-avro-confluent.url' = 
>> 'http://schema-registry:8081',\n" +
>>                 "    'scan.startup.mode' = 'earliest-offset'\n" +
>>                 ")"
>>         );
>>
>>         tEnv.executeSql(
>>                 "CREATE TABLE WC_FK_BUILD_D (\n" +
>>                         "    ROW_WID BIGINT,\n" +
>>                         "    BUILD_ID DOUBLE,\n" +
>>                         "    BUILD_NAME STRING,\n" +
>>                         "    FACILITY_NUMBER STRING,\n" +
>>                         "    START_DATE TIMESTAMP(0),\n" +
>>                         "    END_DATE TIMESTAMP(0),\n" +
>>                         "    DIVISION STRING,\n" +
>>                         "    UPC STRING,\n" +
>>                         "    EFFECTIVE_TO_DT TIMESTAMP(0),\n" +
>>                         "    DELETE_FLG STRING,\n" +
>>                         "    INTEGRATION_ID STRING,\n" +
>>                         "    X_CUSTOM STRING,\n" +
>>                         "    PRIMARY KEY (BUILD_ID) NOT ENFORCED\n" +
>>                         ") WITH (\n" +
>>                         "    'connector' = 'upsert-kafka',\n" +
>>                         "    'topic' = 'WC_FK_BUILD_D',\n" +
>>                         "    'properties.bootstrap.servers' = 
>> 'broker:29092',\n" +
>>                         "    'key.format' = 'avro-confluent',\n" +
>>                         "    'key.avro-confluent.url' = 
>> 'http://schema-registry:8081',\n" +
>>                         "    'value.format' = 'avro-confluent',\n" +
>>                         "    'value.avro-confluent.url' = 
>> 'http://schema-registry:8081'\n" +
>>                         ")"
>>         );
>>
>>         Table mapped = tEnv.sqlQuery(
>>                         "SELECT \n" +
>>                         "    CAST((CAST((`PARTITION` + 1) as STRING) || '0' 
>> || CAST(`OFFSET` as STRING)) as BIGINT),\n" +
>>                         "    BUILD_ID,\n" +
>>                         "    BUILD_NAME,\n" +
>>                         "    FACILITY_NUMBER,\n" +
>>                         "    START_DATE,\n" +
>>                         "    END_DATE,\n" +
>>                         "    RETAILERDIVISION_NAME as DIVISION,\n" +
>>                         "    UPC,\n" +
>>                         "    TIMESTAMP '3714-01-01 00:00:00' as 
>> EFFECTIVE_TO_DT,\n" +
>>                         "    'N' as DELETE_FLG,\n" +
>>                         "    CAST(BUILD_ID as STRING) as INTEGRATION_ID,\n" +
>>                         "    '0' as X_CUSTOM\n" +
>>                         "FROM BUILDS"
>>         );
>>
>>         mapped.executeInsert("WC_FK_BUILD_D");
>>     }
>> }
>>
>> These queries work perfectly fine directly in flink SQL client, but when
>> trying to submit our jar as a job, we get this error:
>>
>> 2022-01-10 19:14:56
>> org.apache.flink.runtime.JobException: Recovery is suppressed by
>> NoRestartBackoffTimeStrategy
>> at
>> org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:138)
>> at
>> org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:82)
>> at
>> org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:228)
>> at
>> org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:218)
>> at
>> org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:209)
>> at
>> org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:679)
>> at
>> org.apache.flink.runtime.scheduler.SchedulerNG.updateTaskExecutionState(SchedulerNG.java:79)
>> at
>> org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:444)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>> at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:498)
>> at
>> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.lambda$handleRpcInvocation$1(AkkaRpcActor.java:316)
>> at
>> org.apache.flink.runtime.concurrent.akka.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:83)
>> at
>> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:314)
>> at
>> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:217)
>> at
>> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:78)
>> at
>> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:163)
>> at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:24)
>> at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:20)
>> at scala.PartialFunction.applyOrElse(PartialFunction.scala:123)
>> at scala.PartialFunction.applyOrElse$(PartialFunction.scala:122)
>> at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:20)
>> at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171)
>> at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172)
>> at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172)
>> at akka.actor.Actor.aroundReceive(Actor.scala:537)
>> at akka.actor.Actor.aroundReceive$(Actor.scala:535)
>> at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:220)
>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:580)
>> at akka.actor.ActorCell.invoke(ActorCell.scala:548)
>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:270)
>> at akka.dispatch.Mailbox.run(Mailbox.scala:231)
>> at akka.dispatch.Mailbox.exec(Mailbox.scala:243)
>> at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
>> at
>> java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056)
>> at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692)
>> at
>> java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175)
>> Caused by: org.apache.flink.streaming.runtime.tasks.StreamTaskException:
>> Cannot load user class:
>> org.apache.flink.avro.shaded.org.apache.avro.generic.GenericRecord
>> ClassLoader info: URL ClassLoader:
>>     file:
>> '/tmp/blobStore-32c720eb-f7f8-4d84-8917-4a8e77f51168/job_d6329af81456ed1fa20e6fc2b33cd2d0/blob_p-05a10e6e13eba54f187e8185b129d68cd93e6fff-3a9dfa80abb21f665796e1fd9e7e4f03'
>> (valid JAR)
>> Class not resolvable through given classloader.
>> at
>> org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperatorFactory(StreamConfig.java:338)
>> at
>> org.apache.flink.streaming.runtime.tasks.OperatorChain.<init>(OperatorChain.java:155)
>> at
>> org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.<init>(RegularOperatorChain.java:63)
>> at
>> org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:666)
>> at
>> org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:654)
>> at
>> org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958)
>> at
>> org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:927)
>> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766)
>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:575)
>> at java.lang.Thread.run(Thread.java:748)
>> Caused by: java.lang.ClassNotFoundException:
>> org.apache.flink.avro.shaded.org.apache.avro.generic.GenericRecord
>> at java.net.URLClassLoader.findClass(URLClassLoader.java:387)
>> at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
>> at
>> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:64)
>> at
>> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65)
>> at
>> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:48)
>> at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
>> at
>> org.apache.flink.runtime.execution.librarycache.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:172)
>> at java.lang.Class.forName0(Native Method)
>> at java.lang.Class.forName(Class.java:348)
>> at
>> org.apache.flink.util.InstantiationUtil$ClassLoaderObjectInputStream.resolveClass(InstantiationUtil.java:78)
>> at java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1986)
>> at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1850)
>> at java.io.ObjectInputStream.readClass(ObjectInputStream.java:1813)
>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1638)
>> at
>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405)
>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329)
>> at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187)
>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667)
>> at
>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405)
>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329)
>> at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187)
>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667)
>> at
>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405)
>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329)
>> at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187)
>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667)
>> at
>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405)
>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329)
>> at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187)
>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667)
>> at
>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405)
>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329)
>> at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187)
>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667)
>> at
>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405)
>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329)
>> at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187)
>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667)
>> at
>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405)
>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329)
>> at
>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187)
>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667)
>> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:503)
>> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:461)
>> at
>> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:617)
>> at
>> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:602)
>> at
>> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:589)
>> at
>> org.apache.flink.util.InstantiationUtil.readObjectFromConfig(InstantiationUtil.java:543)
>> at
>> org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperatorFactory(StreamConfig.java:324)
>> ... 9 more
>>
>> Any help, feedback, or advice would be greatly appreciated.
>>
>> Thank you
>>
>>
>> Jason Politis
>> Solutions Architect, Carrera Group
>> carrera.io
>> | jpoli...@carrera.io <kpatter...@carrera.io>
>> <http://us.linkedin.com/in/jasonpolitis>
>>
>

Reply via email to