I realize those 2 specific ones are commented out, but I believe they are only used in flink-sql, and there's currently an issue when I include them in the pom. I might not have the repos correct.
Thank you Jason Politis Solutions Architect, Carrera Group carrera.io | jpoli...@carrera.io <kpatter...@carrera.io> <http://us.linkedin.com/in/jasonpolitis> On Tue, Jan 11, 2022 at 8:13 AM Jason Politis <jpoli...@carrera.io> wrote: > <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi=" > http://www.w3.org/2001/XMLSchema-instance" > xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 > http://maven.apache.org/xsd/maven-4.0.0.xsd"> > <modelVersion>4.0.0</modelVersion> > > <groupId>boars_head</groupId> > <artifactId>w_build_d</artifactId> > <version>0.1</version> > <packaging>jar</packaging> > > <name>Cql Flink DataStream Java</name> > <url>https://flink.apache.org</url> > > <properties> > <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> > <flink.version>1.14.0</flink.version> > <target.java.version>1.8</target.java.version> > <scala.binary.version>2.11</scala.binary.version> > <maven.compiler.source>${target.java.version}</maven.compiler.source> > <maven.compiler.target>${target.java.version}</maven.compiler.target> > <log4j.version>2.14.1</log4j.version> > <clickhouse-native-jdbc.version>2.6.2</clickhouse-native-jdbc.version> > </properties> > > <repositories> > <repository> > <id>apache.snapshots</id> > <name>Apache Development Snapshot Repository</name> > <url>https://repository.apache.org/content/repositories/snapshots/</url> > <releases> > <enabled>false</enabled> > </releases> > <snapshots> > <enabled>true</enabled> > </snapshots> > </repository> > <repository> > <id>central</id> > <name>Central Repository</name> > <url>https://repo1.maven.org/maven2/</url> > <releases> > <enabled>false</enabled> > </releases> > <snapshots> > <enabled>true</enabled> > </snapshots> > </repository> > </repositories> > > <dependencies> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-walkthrough-common_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > </dependency> > > <!-- This dependency is provided, because it should not be packaged into > the JAR file. --> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-streaming-java_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > <scope>provided</scope> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-table-api-java-bridge_${scala.binary.version}</ > artifactId> > <version>${flink.version}</version> > <scope>provided</scope> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-clients_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > <scope>provided</scope> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-table-planner_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > <scope>provided</scope> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-streaming-scala_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > <scope>provided</scope> > </dependency> > > <!-- Add connector dependencies here. They must be in the default scope > (compile). --> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-connector-kafka_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-avro-confluent-registry</artifactId> > <version>${flink.version}</version> > </dependency> > <!-- <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-sql-avro-confluent-registry</artifactId> > <version>${flink.version}</version> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-sql-connector-kafka</artifactId> > <version>${flink.version}</version> > </dependency> --> > <dependency> > <groupId>com.github.housepower</groupId> > <artifactId>clickhouse-native-jdbc-shaded</artifactId> > <version>${clickhouse-native-jdbc.version}</version> > </dependency> > <dependency> > <groupId>org.apache.flink</groupId> > <artifactId>flink-connector-jdbc_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > </dependency> > <!-- Add logging framework, to produce console output when running in the > IDE. --> > <!-- These dependencies are excluded from the application JAR by default. > --> > <dependency> > <groupId>org.apache.logging.log4j</groupId> > <artifactId>log4j-slf4j-impl</artifactId> > <version>${log4j.version}</version> > <scope>runtime</scope> > </dependency> > <dependency> > <groupId>org.apache.logging.log4j</groupId> > <artifactId>log4j-api</artifactId> > <version>${log4j.version}</version> > <scope>runtime</scope> > </dependency> > <dependency> > <groupId>org.apache.logging.log4j</groupId> > <artifactId>log4j-core</artifactId> > <version>${log4j.version}</version> > <scope>runtime</scope> > </dependency> > </dependencies> > > <build> > <plugins> > > <!-- Java Compiler --> > <plugin> > <groupId>org.apache.maven.plugins</groupId> > <artifactId>maven-compiler-plugin</artifactId> > <version>3.1</version> > <configuration> > <source>${target.java.version}</source> > <target>${target.java.version}</target> > </configuration> > </plugin> > > <!-- We use the maven-shade plugin to create a fat jar that contains all > necessary dependencies. --> > <!-- Change the value of <mainClass>...</mainClass> if your program entry > point changes. --> > <plugin> > <groupId>org.apache.maven.plugins</groupId> > <artifactId>maven-shade-plugin</artifactId> > <version>3.0.0</version> > <executions> > <!-- Run shade goal on package phase --> > <execution> > <phase>package</phase> > <goals> > <goal>shade</goal> > </goals> > <configuration> > <artifactSet> > <excludes> > <exclude>org.apache.flink:flink-shaded-force-shading</exclude> > <exclude>com.google.code.findbugs:jsr305</exclude> > <exclude>org.slf4j:*</exclude> > <exclude>org.apache.logging.log4j:*</exclude> > </excludes> > </artifactSet> > <filters> > <filter> > <!-- Do not copy the signatures in the META-INF folder. > Otherwise, this might cause SecurityExceptions when using the JAR. --> > <artifact>*:*</artifact> > <excludes> > <exclude>META-INF/*.SF</exclude> > <exclude>META-INF/*.DSA</exclude> > <exclude>META-INF/*.RSA</exclude> > </excludes> > </filter> > </filters> > <transformers> > <transformer implementation= > "org.apache.maven.plugins.shade.resource.ManifestResourceTransformer"> > <mainClass>carrera.build.BuildFlinkJob</mainClass> > </transformer> > </transformers> > </configuration> > </execution> > </executions> > </plugin> > </plugins> > > <pluginManagement> > <plugins> > > <!-- This improves the out-of-the-box experience in Eclipse by resolving > some warnings. --> > <plugin> > <groupId>org.eclipse.m2e</groupId> > <artifactId>lifecycle-mapping</artifactId> > <version>1.0.0</version> > <configuration> > <lifecycleMappingMetadata> > <pluginExecutions> > <pluginExecution> > <pluginExecutionFilter> > <groupId>org.apache.maven.plugins</groupId> > <artifactId>maven-shade-plugin</artifactId> > <versionRange>[3.0.0,)</versionRange> > <goals> > <goal>shade</goal> > </goals> > </pluginExecutionFilter> > <action> > <ignore/> > </action> > </pluginExecution> > <pluginExecution> > <pluginExecutionFilter> > <groupId>org.apache.maven.plugins</groupId> > <artifactId>maven-compiler-plugin</artifactId> > <versionRange>[3.1,)</versionRange> > <goals> > <goal>testCompile</goal> > <goal>compile</goal> > </goals> > </pluginExecutionFilter> > <action> > <ignore/> > </action> > </pluginExecution> > </pluginExecutions> > </lifecycleMappingMetadata> > </configuration> > </plugin> > </plugins> > </pluginManagement> > </build> > </project> > > > Above is the POM. > > Here is the entry in our docker-compose where we drop the jars into the > opt/flink/lib folder: > > volumes: > - > './connectors/flink-sql-avro-confluent-registry-1.14.0.jar:/opt/flink/lib/flink-sql-avro-confluent-registry-1.14.0.jar' > - > './connectors/flink-sql-connector-kafka_2.11-1.14.0.jar:/opt/flink/lib/flink-sql-connector-kafka_2.11-1.14.0.jar' > > Thank you > > > Jason Politis > Solutions Architect, Carrera Group > carrera.io > | jpoli...@carrera.io <kpatter...@carrera.io> > <http://us.linkedin.com/in/jasonpolitis> > > > On Mon, Jan 10, 2022 at 9:31 PM Caizhi Weng <tsreape...@gmail.com> wrote: > >> Hi! >> >> Could you share your pom.xml file of your user project? Did you include >> the flink-avro dependency? Also did you add the avro format jar to the lib >> directory of your Flink distribution? >> >> Jason Politis <jpoli...@carrera.io> 于2022年1月11日周二 08:42写道: >> >>> Good evening all, >>> >>> I'm working on a project for a client. We are trying to execute Flink >>> SQL using Table API in java. >>> We are going to pull their data from oracle -> debezium -> kafka -> >>> flink. >>> >>> >>> Here is a sample of our java code: >>> >>> package carrera.build; >>> >>> import org.apache.flink.table.api.EnvironmentSettings; >>> import org.apache.flink.table.api.Table; >>> import org.apache.flink.table.api.TableEnvironment; >>> >>> public class BuildFlinkJob { >>> public static void main(String[] args) throws Exception { >>> EnvironmentSettings settings = >>> EnvironmentSettings.inStreamingMode(); >>> TableEnvironment tEnv = TableEnvironment.create(settings); >>> >>> tEnv.executeSql( >>> "CREATE TABLE BUILDS (\n" + >>> " `PARTITION` INT METADATA FROM 'partition',\n" + >>> " `OFFSET` BIGINT METADATA FROM 'offset',\n" + >>> " BUILD_ID DOUBLE,\n" + >>> " BUILD_NAME STRING,\n" + >>> " FACILITY_NUMBER STRING,\n" + >>> " START_DATE TIMESTAMP(2),\n" + >>> " END_DATE TIMESTAMP(2),\n" + >>> " RETAILERDIVISION_NAME STRING,\n" + >>> " UPC STRING,\n" + >>> " BUILD_INSTRUCTIONS STRING,\n" + >>> " WORK_INSTRUCTIONS STRING,\n" + >>> " IMAGE_FILE_PATH STRING\n" + >>> ") WITH (\n" + >>> " 'connector' = 'kafka',\n" + >>> " 'topic' = >>> 'clients-name.OBIANEW_SDS_EBS_12_1_3.BUILDS',\n" + >>> " 'properties.bootstrap.servers' = 'broker:29092',\n" + >>> " 'properties.group.id' = 'builds',\n" + >>> " 'format' = 'debezium-avro-confluent',\n" + >>> " 'debezium-avro-confluent.url' = >>> 'http://schema-registry:8081',\n" + >>> " 'scan.startup.mode' = 'earliest-offset'\n" + >>> ")" >>> ); >>> >>> tEnv.executeSql( >>> "CREATE TABLE WC_FK_BUILD_D (\n" + >>> " ROW_WID BIGINT,\n" + >>> " BUILD_ID DOUBLE,\n" + >>> " BUILD_NAME STRING,\n" + >>> " FACILITY_NUMBER STRING,\n" + >>> " START_DATE TIMESTAMP(0),\n" + >>> " END_DATE TIMESTAMP(0),\n" + >>> " DIVISION STRING,\n" + >>> " UPC STRING,\n" + >>> " EFFECTIVE_TO_DT TIMESTAMP(0),\n" + >>> " DELETE_FLG STRING,\n" + >>> " INTEGRATION_ID STRING,\n" + >>> " X_CUSTOM STRING,\n" + >>> " PRIMARY KEY (BUILD_ID) NOT ENFORCED\n" + >>> ") WITH (\n" + >>> " 'connector' = 'upsert-kafka',\n" + >>> " 'topic' = 'WC_FK_BUILD_D',\n" + >>> " 'properties.bootstrap.servers' = >>> 'broker:29092',\n" + >>> " 'key.format' = 'avro-confluent',\n" + >>> " 'key.avro-confluent.url' = >>> 'http://schema-registry:8081',\n" + >>> " 'value.format' = 'avro-confluent',\n" + >>> " 'value.avro-confluent.url' = >>> 'http://schema-registry:8081'\n" + >>> ")" >>> ); >>> >>> Table mapped = tEnv.sqlQuery( >>> "SELECT \n" + >>> " CAST((CAST((`PARTITION` + 1) as STRING) || '0' >>> || CAST(`OFFSET` as STRING)) as BIGINT),\n" + >>> " BUILD_ID,\n" + >>> " BUILD_NAME,\n" + >>> " FACILITY_NUMBER,\n" + >>> " START_DATE,\n" + >>> " END_DATE,\n" + >>> " RETAILERDIVISION_NAME as DIVISION,\n" + >>> " UPC,\n" + >>> " TIMESTAMP '3714-01-01 00:00:00' as >>> EFFECTIVE_TO_DT,\n" + >>> " 'N' as DELETE_FLG,\n" + >>> " CAST(BUILD_ID as STRING) as INTEGRATION_ID,\n" >>> + >>> " '0' as X_CUSTOM\n" + >>> "FROM BUILDS" >>> ); >>> >>> mapped.executeInsert("WC_FK_BUILD_D"); >>> } >>> } >>> >>> These queries work perfectly fine directly in flink SQL client, but when >>> trying to submit our jar as a job, we get this error: >>> >>> 2022-01-10 19:14:56 >>> org.apache.flink.runtime.JobException: Recovery is suppressed by >>> NoRestartBackoffTimeStrategy >>> at >>> org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.handleFailure(ExecutionFailureHandler.java:138) >>> at >>> org.apache.flink.runtime.executiongraph.failover.flip1.ExecutionFailureHandler.getFailureHandlingResult(ExecutionFailureHandler.java:82) >>> at >>> org.apache.flink.runtime.scheduler.DefaultScheduler.handleTaskFailure(DefaultScheduler.java:228) >>> at >>> org.apache.flink.runtime.scheduler.DefaultScheduler.maybeHandleTaskFailure(DefaultScheduler.java:218) >>> at >>> org.apache.flink.runtime.scheduler.DefaultScheduler.updateTaskExecutionStateInternal(DefaultScheduler.java:209) >>> at >>> org.apache.flink.runtime.scheduler.SchedulerBase.updateTaskExecutionState(SchedulerBase.java:679) >>> at >>> org.apache.flink.runtime.scheduler.SchedulerNG.updateTaskExecutionState(SchedulerNG.java:79) >>> at >>> org.apache.flink.runtime.jobmaster.JobMaster.updateTaskExecutionState(JobMaster.java:444) >>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> at java.lang.reflect.Method.invoke(Method.java:498) >>> at >>> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.lambda$handleRpcInvocation$1(AkkaRpcActor.java:316) >>> at >>> org.apache.flink.runtime.concurrent.akka.ClassLoadingUtils.runWithContextClassLoader(ClassLoadingUtils.java:83) >>> at >>> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcInvocation(AkkaRpcActor.java:314) >>> at >>> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleRpcMessage(AkkaRpcActor.java:217) >>> at >>> org.apache.flink.runtime.rpc.akka.FencedAkkaRpcActor.handleRpcMessage(FencedAkkaRpcActor.java:78) >>> at >>> org.apache.flink.runtime.rpc.akka.AkkaRpcActor.handleMessage(AkkaRpcActor.java:163) >>> at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:24) >>> at akka.japi.pf.UnitCaseStatement.apply(CaseStatements.scala:20) >>> at scala.PartialFunction.applyOrElse(PartialFunction.scala:123) >>> at scala.PartialFunction.applyOrElse$(PartialFunction.scala:122) >>> at akka.japi.pf.UnitCaseStatement.applyOrElse(CaseStatements.scala:20) >>> at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:171) >>> at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172) >>> at scala.PartialFunction$OrElse.applyOrElse(PartialFunction.scala:172) >>> at akka.actor.Actor.aroundReceive(Actor.scala:537) >>> at akka.actor.Actor.aroundReceive$(Actor.scala:535) >>> at akka.actor.AbstractActor.aroundReceive(AbstractActor.scala:220) >>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:580) >>> at akka.actor.ActorCell.invoke(ActorCell.scala:548) >>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:270) >>> at akka.dispatch.Mailbox.run(Mailbox.scala:231) >>> at akka.dispatch.Mailbox.exec(Mailbox.scala:243) >>> at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289) >>> at >>> java.util.concurrent.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1056) >>> at java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1692) >>> at >>> java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:175) >>> Caused by: org.apache.flink.streaming.runtime.tasks.StreamTaskException: >>> Cannot load user class: >>> org.apache.flink.avro.shaded.org.apache.avro.generic.GenericRecord >>> ClassLoader info: URL ClassLoader: >>> file: >>> '/tmp/blobStore-32c720eb-f7f8-4d84-8917-4a8e77f51168/job_d6329af81456ed1fa20e6fc2b33cd2d0/blob_p-05a10e6e13eba54f187e8185b129d68cd93e6fff-3a9dfa80abb21f665796e1fd9e7e4f03' >>> (valid JAR) >>> Class not resolvable through given classloader. >>> at >>> org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperatorFactory(StreamConfig.java:338) >>> at >>> org.apache.flink.streaming.runtime.tasks.OperatorChain.<init>(OperatorChain.java:155) >>> at >>> org.apache.flink.streaming.runtime.tasks.RegularOperatorChain.<init>(RegularOperatorChain.java:63) >>> at >>> org.apache.flink.streaming.runtime.tasks.StreamTask.restoreInternal(StreamTask.java:666) >>> at >>> org.apache.flink.streaming.runtime.tasks.StreamTask.restore(StreamTask.java:654) >>> at >>> org.apache.flink.runtime.taskmanager.Task.runWithSystemExitMonitoring(Task.java:958) >>> at >>> org.apache.flink.runtime.taskmanager.Task.restoreAndInvoke(Task.java:927) >>> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:766) >>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:575) >>> at java.lang.Thread.run(Thread.java:748) >>> Caused by: java.lang.ClassNotFoundException: >>> org.apache.flink.avro.shaded.org.apache.avro.generic.GenericRecord >>> at java.net.URLClassLoader.findClass(URLClassLoader.java:387) >>> at java.lang.ClassLoader.loadClass(ClassLoader.java:418) >>> at >>> org.apache.flink.util.FlinkUserCodeClassLoader.loadClassWithoutExceptionHandling(FlinkUserCodeClassLoader.java:64) >>> at >>> org.apache.flink.util.ChildFirstClassLoader.loadClassWithoutExceptionHandling(ChildFirstClassLoader.java:65) >>> at >>> org.apache.flink.util.FlinkUserCodeClassLoader.loadClass(FlinkUserCodeClassLoader.java:48) >>> at java.lang.ClassLoader.loadClass(ClassLoader.java:351) >>> at >>> org.apache.flink.runtime.execution.librarycache.FlinkUserCodeClassLoaders$SafetyNetWrapperClassLoader.loadClass(FlinkUserCodeClassLoaders.java:172) >>> at java.lang.Class.forName0(Native Method) >>> at java.lang.Class.forName(Class.java:348) >>> at >>> org.apache.flink.util.InstantiationUtil$ClassLoaderObjectInputStream.resolveClass(InstantiationUtil.java:78) >>> at >>> java.io.ObjectInputStream.readNonProxyDesc(ObjectInputStream.java:1986) >>> at java.io.ObjectInputStream.readClassDesc(ObjectInputStream.java:1850) >>> at java.io.ObjectInputStream.readClass(ObjectInputStream.java:1813) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1638) >>> at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405) >>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667) >>> at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405) >>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667) >>> at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405) >>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667) >>> at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405) >>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667) >>> at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405) >>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667) >>> at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405) >>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667) >>> at >>> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2405) >>> at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:2329) >>> at >>> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:2187) >>> at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1667) >>> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:503) >>> at java.io.ObjectInputStream.readObject(ObjectInputStream.java:461) >>> at >>> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:617) >>> at >>> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:602) >>> at >>> org.apache.flink.util.InstantiationUtil.deserializeObject(InstantiationUtil.java:589) >>> at >>> org.apache.flink.util.InstantiationUtil.readObjectFromConfig(InstantiationUtil.java:543) >>> at >>> org.apache.flink.streaming.api.graph.StreamConfig.getStreamOperatorFactory(StreamConfig.java:324) >>> ... 9 more >>> >>> Any help, feedback, or advice would be greatly appreciated. >>> >>> Thank you >>> >>> >>> Jason Politis >>> Solutions Architect, Carrera Group >>> carrera.io >>> | jpoli...@carrera.io <kpatter...@carrera.io> >>> <http://us.linkedin.com/in/jasonpolitis> >>> >>