It is a known issue reported multiple times that if you are in an early jdk 1.8.x version, upgrade the bugfix version and the issue will vanish.
I don't ever have a log on jm side when this issue reported so I'm sorry unable to explain more... Best, tison. Yang Wang <danrtsey...@gmail.com> 于2020年1月22日周三 上午10:46写道: > The "web.timeout" will be used for all web monitor asynchronous > operations, including the > "DispatcherGateway.submitJob" in the "JobSubmitHandler". > So when you increase the timeout, does it still could not work? > > Best, > Yang > > satya brat <bratsatya...@gmail.com> 于2020年1月21日周二 下午8:57写道: > >> How does web.timeout help hear?? The issue is with respect to aka >> dispatched timing out. The job is submitted to the task managers but the >> response doesn't reach the client. >> >> On Tue, Jan 21, 2020 at 12:34 PM Yang Wang <danrtsey...@gmail.com> wrote: >> >>> Hi satya, >>> >>> Maybe the job has been submitted to Dispatcher successfully and the >>> internal submitting job takes >>> too long time(more than 10s). So it failed with timeout. Could you >>> please set the `web.timeout: 30000` >>> and run again? >>> >>> >>> >>> Best, >>> Yang >>> >>> satya brat <bratsatya...@gmail.com> 于2020年1月20日周一 下午4:34写道: >>> >>>> I am using standalone cluster of Flink with 1 jobManager and n >>>> taskManagers. When I try to submit a job via command line, the job >>>> submission fails with error message as >>>> org.apache.flink.client.program.ProgramInvocationException: Could not >>>> submit job (JobID: f839aefee74aa4483ce8f8fd2e49b69e). >>>> >>>> On jobManager instance, everything works fine till the job is switched >>>> from DEPLOYING to RUNNING. Post that, once akka-timeut expires, I see the >>>> following stacktrace >>>> >>>> akka.pattern.AskTimeoutException: Ask timed out on >>>> [Actor[akka://flink/user/dispatcher#-177004106]] after [100000 ms]. >>>> Sender[null] sent message of type >>>> "org.apache.flink.runtime.rpc.messages.LocalFencedMessage". >>>> at >>>> akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:604) >>>> at akka.actor.Scheduler$$anon$4.run(Scheduler.scala:126) >>>> at >>>> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601) >>>> at >>>> scala.concurrent.BatchingExecutor$class.execute(BatchingExecutor.scala:109) >>>> at >>>> scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:599) >>>> at >>>> akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:329) >>>> at >>>> akka.actor.LightArrayRevolverScheduler$$anon$4.executeBucket$1(LightArrayRevolverScheduler.scala:280) >>>> at >>>> akka.actor.LightArrayRevolverScheduler$$anon$4.nextTick(LightArrayRevolverScheduler.scala:284) >>>> at >>>> akka.actor.LightArrayRevolverScheduler$$anon$4.run(LightArrayRevolverScheduler.scala:236) >>>> at java.lang.Thread.run(Thread.java:745) >>>> >>>> I went through the flink code on github and all the steps required to >>>> execute a job seems to be running fine. However, when jobManager has to >>>> give job submission ack to flink client that triggered the job, the >>>> jobSubmitHandler times out on the akka dispatcher that according to my >>>> understanding takes care of communicating with the job client. >>>> >>>> The Flink job consists for 1 Source (kafka), 2 operators and 1 >>>> sink(Custom Sink). Following link shows the jobManager logs: >>>> https://pastebin.com/raw/3GaTtNrG >>>> >>>> Once the dispatcher times out, all other Flink UI calls also timeout >>>> with same exception. >>>> >>>> Following are the flink client logs that is used to submit job via >>>> command line. >>>> >>>> 2019-09-28 19:34:21,321 INFO org.apache.flink.client.cli.CliFrontend >>>> - >>>> -------------------------------------------------------------------------------- >>>> 2019-09-28 19:34:21,322 INFO org.apache.flink.client.cli.CliFrontend >>>> - Starting Command Line Client (Version: 1.8.0, >>>> Rev:<unknown>, Date:<unknown>) >>>> 2019-09-28 19:34:21,322 INFO org.apache.flink.client.cli.CliFrontend >>>> - OS current user: root >>>> 2019-09-28 19:34:21,322 INFO org.apache.flink.client.cli.CliFrontend >>>> - Current Hadoop/Kerberos user: <no hadoop dependency >>>> found> >>>> 2019-09-28 19:34:21,322 INFO org.apache.flink.client.cli.CliFrontend >>>> - JVM: Java HotSpot(TM) 64-Bit Server VM - Oracle >>>> Corporation - 1.8/25.5-b02 >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - Maximum heap size: 2677 MiBytes >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - JAVA_HOME: (not set) >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - No Hadoop Dependency available >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - JVM Options: >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - >>>> -Dlog.file=/var/lib/fulfillment-stream-processor/flink-executables/flink-executables/log/flink-root-client-fulfillment-stream-processor-flink-task-manager-2-8047357.log >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - >>>> -Dlog4j.configuration=file:/var/lib/fulfillment-stream-processor/flink-executables/flink-executables/conf/log4j-cli.properties >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - >>>> -Dlogback.configurationFile=file:/var/lib/fulfillment-stream-processor/flink-executables/flink-executables/conf/logback.xml >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - Program Arguments: >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - run >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - -d >>>> 2019-09-28 19:34:21,323 INFO org.apache.flink.client.cli.CliFrontend >>>> - -c >>>> 2019-09-28 19:34:21,324 INFO org.apache.flink.client.cli.CliFrontend >>>> - /home/fse/flink-kafka-relayer-0.2.jar >>>> 2019-09-28 19:34:21,324 INFO org.apache.flink.client.cli.CliFrontend >>>> - Classpath: >>>> /var/lib/fulfillment-stream-processor/flink-executables/flink-executables/lib/log4j-1.2.17.jar:/var/lib/fulfillment-stream-processor/flink-executables/flink-executables/lib/slf4j-log4j12-1.7.15.jar:/var/lib/fulfillment-stream-processor/flink-executables/flink-executables/lib/flink-dist_2.11-1.8.0.jar::: >>>> 2019-09-28 19:34:21,324 INFO org.apache.flink.client.cli.CliFrontend >>>> - >>>> -------------------------------------------------------------------------------- >>>> 2019-09-28 19:34:21,328 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: jobmanager.rpc.address, <job-manager-ip> >>>> 2019-09-28 19:34:21,328 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: jobmanager.rpc.port, 6123 >>>> 2019-09-28 19:34:21,328 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: jobmanager.heap.size, 1024m >>>> 2019-09-28 19:34:21,329 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: taskmanager.heap.size, 1024m >>>> 2019-09-28 19:34:21,329 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: taskmanager.numberOfTaskSlots, 4 >>>> 2019-09-28 19:34:21,329 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: parallelism.default, 1 >>>> 2019-09-28 19:34:21,329 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: metrics.reporter.jmx.class, >>>> org.apache.flink.metrics.jmx.JMXReporter >>>> 2019-09-28 19:34:21,329 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: metrics.reporter.jmx.port, 8789 >>>> 2019-09-28 19:34:21,333 WARN org.apache.flink.client.cli.CliFrontend >>>> - Could not load CLI class >>>> org.apache.flink.yarn.cli.FlinkYarnSessionCli. >>>> java.lang.NoClassDefFoundError: >>>> org/apache/hadoop/yarn/exceptions/YarnException >>>> at java.lang.Class.forName0(Native Method) >>>> at java.lang.Class.forName(Class.java:259) >>>> at >>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1230) >>>> at >>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1190) >>>> at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1115) >>>> Caused by: java.lang.ClassNotFoundException: >>>> org.apache.hadoop.yarn.exceptions.YarnException >>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:372) >>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:361) >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:360) >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:424) >>>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:357) >>>> ... 5 more >>>> 2019-09-28 19:34:21,343 INFO org.apache.flink.core.fs.FileSystem >>>> - Hadoop is not in the classpath/dependencies. The >>>> extended set of supported File Systems via Hadoop is not available. >>>> 2019-09-28 19:34:21,545 INFO >>>> org.apache.flink.runtime.security.modules.HadoopModuleFactory - Cannot >>>> create Hadoop Security Module because Hadoop cannot be found in the >>>> Classpath. >>>> 2019-09-28 19:34:21,560 INFO >>>> org.apache.flink.runtime.security.SecurityUtils - Cannot >>>> install HadoopSecurityContext because Hadoop cannot be found in the >>>> Classpath. >>>> 2019-09-28 19:34:21,561 INFO org.apache.flink.client.cli.CliFrontend >>>> - Running 'run' command. >>>> 2019-09-28 19:34:21,566 INFO org.apache.flink.client.cli.CliFrontend >>>> - Building program from JAR file >>>> 2019-09-28 19:34:21,744 INFO org.apache.flink.configuration.Configuration >>>> - Config uses fallback configuration key >>>> 'jobmanager.rpc.address' instead of key 'rest.address' >>>> 2019-09-28 19:34:21,896 INFO org.apache.flink.runtime.rest.RestClient >>>> - Rest client endpoint started. >>>> 2019-09-28 19:34:21,898 INFO org.apache.flink.client.cli.CliFrontend >>>> - Starting execution of program >>>> 2019-09-28 19:34:21,898 INFO >>>> org.apache.flink.client.program.rest.RestClusterClient - Starting >>>> program in interactive mode (detached: true) >>>> 2019-09-28 19:34:22,594 WARN >>>> org.apache.flink.streaming.api.environment.StreamContextEnvironment - Job >>>> was executed in detached mode, the results will be available on completion. >>>> 2019-09-28 19:34:22,632 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: jobmanager.rpc.address, <job-manager-ip> >>>> 2019-09-28 19:34:22,632 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: jobmanager.rpc.port, 6123 >>>> 2019-09-28 19:34:22,632 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: jobmanager.heap.size, 1024m >>>> 2019-09-28 19:34:22,633 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: taskmanager.heap.size, 1024m >>>> 2019-09-28 19:34:22,633 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: taskmanager.numberOfTaskSlots, 4 >>>> 2019-09-28 19:34:22,633 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: parallelism.default, 1 >>>> 2019-09-28 19:34:22,633 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: metrics.reporter.jmx.class, >>>> org.apache.flink.metrics.jmx.JMXReporter >>>> 2019-09-28 19:34:22,633 INFO >>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>> configuration property: metrics.reporter.jmx.port, 8789 >>>> 2019-09-28 19:34:22,635 INFO >>>> org.apache.flink.client.program.rest.RestClusterClient - Submitting >>>> job f839aefee74aa4483ce8f8fd2e49b69e (detached: true). >>>> 2019-09-28 19:36:04,341 INFO org.apache.flink.runtime.rest.RestClient >>>> - Shutting down rest endpoint. >>>> 2019-09-28 19:36:04,343 INFO org.apache.flink.runtime.rest.RestClient >>>> - Rest endpoint shutdown complete. >>>> 2019-09-28 19:36:04,343 ERROR org.apache.flink.client.cli.CliFrontend >>>> - Error while running the command. >>>> org.apache.flink.client.program.ProgramInvocationException: Could not >>>> submit job (JobID: f839aefee74aa4483ce8f8fd2e49b69e) >>>> at >>>> org.apache.flink.client.program.rest.RestClusterClient.submitJob(RestClusterClient.java:250) >>>> at >>>> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:483) >>>> at >>>> org.apache.flink.client.program.DetachedEnvironment.finalizeExecute(DetachedEnvironment.java:77) >>>> at >>>> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:429) >>>> at >>>> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:813) >>>> at >>>> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:287) >>>> at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:213) >>>> at >>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1050) >>>> at >>>> org.apache.flink.client.cli.CliFrontend.lambda$main$11(CliFrontend.java:1126) >>>> at >>>> org.apache.flink.client.cli.CliFrontend$$Lambda$5/1971851377.call(Unknown >>>> Source) >>>> at >>>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30) >>>> at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1126) >>>> Caused by: org.apache.flink.runtime.client.JobSubmissionException: Failed >>>> to submit JobGraph. >>>> at >>>> org.apache.flink.client.program.rest.RestClusterClient.lambda$submitJob$8(RestClusterClient.java:388) >>>> at >>>> org.apache.flink.client.program.rest.RestClusterClient$$Lambda$17/788892554.apply(Unknown >>>> Source) >>>> at >>>> java.util.concurrent.CompletableFuture$ExceptionCompletion.run(CompletableFuture.java:1246) >>>> at >>>> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:193) >>>> at >>>> java.util.concurrent.CompletableFuture.internalComplete(CompletableFuture.java:210) >>>> at >>>> java.util.concurrent.CompletableFuture$ThenApply.run(CompletableFuture.java:723) >>>> at >>>> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:193) >>>> at >>>> java.util.concurrent.CompletableFuture.internalComplete(CompletableFuture.java:210) >>>> at >>>> java.util.concurrent.CompletableFuture$ThenCopy.run(CompletableFuture.java:1333) >>>> at >>>> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:193) >>>> at >>>> java.util.concurrent.CompletableFuture.completeExceptionally(CompletableFuture.java:2361) >>>> at >>>> org.apache.flink.runtime.concurrent.FutureUtils.lambda$retryOperationWithDelay$5(FutureUtils.java:207) >>>> at >>>> org.apache.flink.runtime.concurrent.FutureUtils$$Lambda$34/1092254958.accept(Unknown >>>> Source) >>>> at >>>> java.util.concurrent.CompletableFuture$WhenCompleteCompletion.run(CompletableFuture.java:1298) >>>> at >>>> java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:193) >>>> at >>>> java.util.concurrent.CompletableFuture.internalComplete(CompletableFuture.java:210) >>>> at >>>> java.util.concurrent.CompletableFuture$AsyncCompose.exec(CompletableFuture.java:626) >>>> at >>>> java.util.concurrent.CompletableFuture$Async.run(CompletableFuture.java:428) >>>> at >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) >>>> at >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) >>>> at java.lang.Thread.run(Thread.java:745) >>>> Caused by: org.apache.flink.runtime.rest.util.RestClientException: >>>> [Internal server error., <Exception on server side: >>>> akka.pattern.AskTimeoutException: Ask timed out on >>>> [Actor[akka://flink/user/dispatcher#-177004106]] after [100000 ms]. >>>> Sender[null] sent message of type >>>> "org.apache.flink.runtime.rpc.messages.LocalFencedMessage". >>>> at >>>> akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:604) >>>> at akka.actor.Scheduler$$anon$4.run(Scheduler.scala:126) >>>> at >>>> scala.concurrent.Future$InternalCallbackExecutor$.unbatchedExecute(Future.scala:601) >>>> at >>>> scala.concurrent.BatchingExecutor$class.execute(BatchingExecutor.scala:109) >>>> at >>>> scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:599) >>>> at >>>> akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(LightArrayRevolverScheduler.scala:329) >>>> at >>>> akka.actor.LightArrayRevolverScheduler$$anon$4.executeBucket$1(LightArrayRevolverScheduler.scala:280) >>>> at >>>> akka.actor.LightArrayRevolverScheduler$$anon$4.nextTick(LightArrayRevolverScheduler.scala:284) >>>> at >>>> akka.actor.LightArrayRevolverScheduler$$anon$4.run(LightArrayRevolverScheduler.scala:236) >>>> at java.lang.Thread.run(Thread.java:745) >>>> >>>> End of exception on server side>] >>>> at >>>> org.apache.flink.runtime.rest.RestClient.parseResponse(RestClient.java:389) >>>> at >>>> org.apache.flink.runtime.rest.RestClient.lambda$submitRequest$3(RestClient.java:373) >>>> at >>>> org.apache.flink.runtime.rest.RestClient$$Lambda$33/1155836850.apply(Unknown >>>> Source) >>>> at >>>> java.util.concurrent.CompletableFuture$AsyncCompose.exec(CompletableFuture.java:604) >>>> ... 4 more >>>> >>>> I have turned on debug logs for flink, akka and kafka but not able to >>>> figure out what is going wrong. I have very basic understanding of akka >>>> because of which not able to figure out what is going wrong. Can someone >>>> help me with that?? I am running flink 1.8.0. >>>> >>>