Hi Avi and Victor, I just opened this ticket on JIRA: https://issues.apache.org/jira/browse/FLINK-13586 (I hadn't seen these e-mails). Backward compatibility is broken between 1.8.0 and 1.8.1 if you use Kafka connectors.
Can you upgrade your flink-connector-kafka dependency to 1.8.1 ? It won't deploy on a 1.8.0 server any more, if that's a concern for you. Gaël On Mon, Aug 5, 2019 at 4:37 PM Wong Victor <jiasheng.w...@outlook.com> wrote: > Hi Avi: > > > > It seems you are submitting your job with an older Flink version (< 1.8), > please check your flink-dist version. > > > > Regards, > > Victor > > > > *From: *Avi Levi <avi.l...@bluevoyant.com> > *Date: *Monday, August 5, 2019 at 9:11 PM > *To: *user <user@flink.apache.org> > *Subject: *getting an exception > > > > Hi, > > I'm using Flink 1.8.1. our code is mostly using Scala. > > When I try to submit my job (on my local machine ) it crashes with the > error below (BTW on the IDE it runs perfectly). > > Any assistance would be appreciated. > > Thanks > > Avi > > 2019-08-05 12:58:03.783 [Flink-DispatcherRestEndpoint-thread-3] ERROR > org.apache.flink.runtime.webmonitor.handlers.JarRunHandler - Unhandled > exception. > > org.apache.flink.client.program.ProgramInvocationException: The program > caused an error: > > at > org.apache.flink.client.program.OptimizerPlanEnvironment.getOptimizedPlan(OptimizerPlanEnvironment.java:93) > > at > org.apache.flink.client.program.PackagedProgramUtils.createJobGraph(PackagedProgramUtils.java:80) > > at > org.apache.flink.runtime.webmonitor.handlers.utils.JarHandlerUtils$JarHandlerContext.toJobGraph(JarHandlerUtils.java:126) > > at > org.apache.flink.runtime.webmonitor.handlers.JarRunHandler.lambda$getJobGraphAsync$6(JarRunHandler.java:142) > > at > java.util.concurrent.CompletableFuture$AsyncSupply.run(CompletableFuture.java:1590) > > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) > > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) > > at java.lang.Thread.run(Thread.java:748) > > Caused by: java.lang.NoSuchMethodError: > org.apache.flink.api.java.ClosureCleaner.clean(Ljava/lang/Object;Lorg/apache/flink/api/common/ExecutionConfig$ClosureCleanerLevel;Z)V > > at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer011.<init>(FlinkKafkaProducer011.java:494) > > at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer011.<init>(FlinkKafkaProducer011.java:448) > > at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaProducer011.<init>(FlinkKafkaProducer011.java:383) > > at > com.bluevoyant.commons.queueHandlers.KeyedKafkaProducerImpl.producer(KafkaImpl.scala:18) > > at > com.bluevoyant.commons.queueHandlers.KeyedKafkaProducerImpl.producer$(KafkaImpl.scala:18) > > at > com.bluevoyant.lookalike.analytic.queueHandlers.QueueHandlerImpl$.producer(QueueHandlerImpl.scala:13) > > at > com.bluevoyant.lookalike.analytic.StreamingJob$.delayedEndpoint$com$bluevoyant$lookalike$analytic$StreamingJob$1(StreamingJob.scala:42) > > at > com.bluevoyant.lookalike.analytic.StreamingJob$delayedInit$body.apply(StreamingJob.scala:14) > > at scala.Function0.apply$mcV$sp(Function0.scala:34) > > at scala.Function0.apply$mcV$sp$(Function0.scala:34) > > at > scala.runtime.AbstractFunction0.apply$mcV$sp(AbstractFunction0.scala:12) > > at scala.App.$anonfun$main$1$adapted(App.scala:76) > > at scala.collection.immutable.List.foreach(List.scala:388) > > at scala.App.main(App.scala:76) > > at scala.App.main$(App.scala:74) > > at > com.bluevoyant.lookalike.analytic.StreamingJob$.main(StreamingJob.scala:14) > > at > com.bluevoyant.lookalike.analytic.StreamingJob.main(StreamingJob.scala) > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > > at java.lang.reflect.Method.invoke(Method.java:498) > > at > org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:529) > > at > org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:421) > > at > org.apache.flink.client.program.OptimizerPlanEnvironment.getOptimizedPlan(OptimizerPlanEnvironment.java:83) > > ... 7 common frames omitted > > > -- Gaël Renoux Senior R&D Engineer, DataDome M +33 6 76 89 16 52 <+33+6+76+89+16+52> E gael.ren...@datadome.co <gael.ren...@datadome.co> W www.datadome.co <http://www.datadome.co?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <https://www.facebook.com/datadome/?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <https://fr.linkedin.com/company/datadome?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> <https://twitter.com/data_dome?utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature> [image: Read DataDome reviews on G2] <https://www.g2.com/products/datadome/reviews?utm_source=review-widget&utm_source=WiseStamp&utm_medium=email&utm_term=&utm_content=&utm_campaign=signature>