Hi Hanjing, I think the problem is that the Storm compatibility layer only works with legacy mode at the moment. Please set `mode: legacy` in your flink-conf.yaml. I hope this will resolve the problems.
Cheers, Till On Tue, Sep 11, 2018 at 7:10 AM jing <hanjingz...@163.com> wrote: > Hi vino, > Thank you very much. > I'll try more tests. > > Hanjing > > <https://maas.mail.163.com/dashi-web-extend/html/proSignature.html?ftlId=1&name=Hanjing&uid=hanjingzuzu%40163.com&iconUrl=http%3A%2F%2Fmail-online.nosdn.127.net%2Fsmda6015df3a52ec22402a83bdad785acb.jpg&items=%5B%22%22%2C%22%22%2C%22%22%2C%22%22%2C%22%22%5D> > 签名由 网易邮箱大师 <https://mail.163.com/dashi/dlpro.html?from=mail81> 定制 > On 9/11/2018 11:51,vino yang<yanghua1...@gmail.com> > <yanghua1...@gmail.com> wrote: > > Hi Hanjing, > > Flink does not currently support TaskManager HA and only supports > JobManager HA. > In the Standalone environment, once the JobManager triggers a failover, it > will also cause cancel and restart for all jobs. > > Thanks, vino. > > jing <hanjingz...@163.com> 于2018年9月11日周二 上午11:12写道: > >> Hi vino, >> Thanks a lot. >> Besides, I'm also confused about taskmanager's HA. >> There're 2 taskmangaer in my cluster, only one job A worked on >> taskmanager A. If taskmangaer A crashed, what happend about my job. >> I tried, my job failed, taskmanger B does not take over job A. >> Is this right? >> >> Hanjing >> >> <https://maas.mail.163.com/dashi-web-extend/html/proSignature.html?ftlId=1&name=Hanjing&uid=hanjingzuzu%40163.com&iconUrl=http%3A%2F%2Fmail-online.nosdn.127.net%2Fsmda6015df3a52ec22402a83bdad785acb.jpg&items=%5B%22%22%2C%22%22%2C%22%22%2C%22%22%2C%22%22%5D> >> 签名由 网易邮箱大师 <https://mail.163.com/dashi/dlpro.html?from=mail81> 定制 >> On 9/11/2018 10:59,vino yang<yanghua1...@gmail.com> >> <yanghua1...@gmail.com> wrote: >> >> Oh, I thought the flink job could not be submitted. I don't know why the >> storm's example could not be submitted. Because I have never used it. >> >> Maybe Till, Chesnay or Gary can help you. Ping them for you. >> >> Thanks, vino. >> >> jing <hanjingz...@163.com> 于2018年9月11日周二 上午10:26写道: >> >>> Hi vino, >>> My job mangaer log is as below. I can submit regular flink job to this >>> jobmanger, it worked. But the flink-storm example doesn's work. >>> Thanks. >>> Hanjing >>> >>> 2018-09-11 18:22:48,937 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> -------------------------------------------------------------------------------- >>> 2018-09-11 18:22:48,938 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Starting >>> StandaloneSessionClusterEntrypoint (Version: 1.6.0, Rev:ff472b4, >>> Date:07.08.2018 @ 13:31:13 UTC) >>> 2018-09-11 18:22:48,938 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - OS current >>> user: hadoop3 >>> 2018-09-11 18:22:49,143 WARN org.apache.hadoop.util.NativeCodeLoader >>> - Unable to load native-hadoop library for your platform... >>> using builtin-java classes where applicable >>> 2018-09-11 18:22:49,186 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Current >>> Hadoop/Kerberos user: hadoop3 >>> 2018-09-11 18:22:49,186 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - JVM: Java >>> HotSpot(TM) 64-Bit Server VM - Oracle Corporation - 1.8/25.172-b11 >>> 2018-09-11 18:22:49,186 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Maximum >>> heap size: 981 MiBytes >>> 2018-09-11 18:22:49,186 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - JAVA_HOME: >>> /usr/java/jdk1.8.0_172-amd64 >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Hadoop >>> version: 2.7.5 >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - JVM >>> Options: >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> -Xms1024m >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> -Xmx1024m >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> -Dlog.file=/home/hadoop3/zh/flink-1.6.0/log/flink-hadoop3-standalonesession-0-p-a36-72.log >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> -Dlog4j.configuration=file:/home/hadoop3/zh/flink-1.6.0/conf/log4j.properties >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> -Dlogback.configurationFile=file:/home/hadoop3/zh/flink-1.6.0/conf/logback.xml >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Program >>> Arguments: >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> --configDir >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> /home/hadoop3/zh/flink-1.6.0/conf >>> 2018-09-11 18:22:49,188 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> --executionMode >>> 2018-09-11 18:22:49,189 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - cluster >>> 2018-09-11 18:22:49,189 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Classpath: >>> /home/hadoop3/zh/flink-1.6.0/lib/flink-python_2.11-1.6.0.jar:/home/hadoop3/zh/flink-1.6.0/lib/flink-shaded-hadoop2-uber-1.6.0.jar:/home/hadoop3/zh/flink-1.6.0/lib/log4j-1.2.17.jar:/home/hadoop3/zh/flink-1.6.0/lib/slf4j-log4j12-1.7.7.jar:/home/hadoop3/zh/flink-1.6.0/lib/flink-dist_2.11-1.6.0.jar::: >>> 2018-09-11 18:22:49,189 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> -------------------------------------------------------------------------------- >>> 2018-09-11 18:22:49,189 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Registered >>> UNIX signal handlers for [TERM, HUP, INT] >>> 2018-09-11 18:22:49,197 INFO >>> org.apache.flink.configuration.GlobalConfiguration - Loading >>> configuration property: jobmanager.rpc.address, p-a36-72 >>> 2018-09-11 18:22:49,197 INFO >>> org.apache.flink.configuration.GlobalConfiguration - Loading >>> configuration property: jobmanager.rpc.port, 6123 >>> 2018-09-11 18:22:49,197 INFO >>> org.apache.flink.configuration.GlobalConfiguration - Loading >>> configuration property: jobmanager.heap.size, 1024m >>> 2018-09-11 18:22:49,197 INFO >>> org.apache.flink.configuration.GlobalConfiguration - Loading >>> configuration property: taskmanager.heap.size, 10240m >>> 2018-09-11 18:22:49,197 INFO >>> org.apache.flink.configuration.GlobalConfiguration - Loading >>> configuration property: taskmanager.numberOfTaskSlots, 16 >>> 2018-09-11 18:22:49,197 INFO >>> org.apache.flink.configuration.GlobalConfiguration - Loading >>> configuration property: parallelism.default, 2 >>> 2018-09-11 18:22:49,198 INFO >>> org.apache.flink.configuration.GlobalConfiguration - Loading >>> configuration property: rest.port, 8081 >>> 2018-09-11 18:22:49,207 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Starting >>> StandaloneSessionClusterEntrypoint. >>> 2018-09-11 18:22:49,207 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Install >>> default filesystem. >>> 2018-09-11 18:22:49,214 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Install >>> security context. >>> 2018-09-11 18:22:49,237 INFO >>> org.apache.flink.runtime.security.modules.HadoopModule - Hadoop user >>> set to hadoop3 (auth:SIMPLE) >>> 2018-09-11 18:22:49,247 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - >>> Initializing cluster services. >>> 2018-09-11 18:22:49,253 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Trying to >>> start actor system at p-a36-72:6123 >>> 2018-09-11 18:22:49,576 INFO akka.event.slf4j.Slf4jLogger >>> - Slf4jLogger started >>> 2018-09-11 18:22:49,611 INFO akka.remote.Remoting >>> - Starting remoting >>> 2018-09-11 18:22:49,718 INFO akka.remote.Remoting >>> - Remoting started; listening on addresses >>> :[akka.tcp://flink@p-a36-72:6123] >>> 2018-09-11 18:22:49,722 INFO >>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint - Actor >>> system started at akka.tcp://flink@p-a36-72:6123 >>> 2018-09-11 18:22:49,732 WARN org.apache.flink.configuration.Configuration >>> - Config uses deprecated configuration key >>> 'jobmanager.rpc.address' instead of proper key 'rest.address' >>> 2018-09-11 18:22:49,737 INFO org.apache.flink.runtime.blob.BlobServer >>> - Created BLOB server storage directory >>> /tmp/blobStore-62c16996-0f38-43ae-9e40-ac4206329d93 >>> 2018-09-11 18:22:49,739 INFO org.apache.flink.runtime.blob.BlobServer >>> - Started BLOB server at 0.0.0.0:3706 - max concurrent >>> requests: 50 - max backlog: 1000 >>> 2018-09-11 18:22:49,749 INFO >>> org.apache.flink.runtime.metrics.MetricRegistryImpl - No metrics >>> reporter configured, no metrics will be exposed/reported. >>> 2018-09-11 18:22:49,751 INFO >>> org.apache.flink.runtime.dispatcher.FileArchivedExecutionGraphStore - >>> Initializing FileArchivedExecutionGraphStore: Storage directory >>> /tmp/executionGraphStore-fecb7e34-9d33-4af2-a623-ee96d8572800, expiration >>> time 3600000, maximum cache size 52428800 bytes. >>> 2018-09-11 18:22:49,766 INFO >>> org.apache.flink.runtime.blob.TransientBlobCache - Created >>> BLOB cache storage directory >>> /tmp/blobStore-c1d1946d-9e19-40b1-800d-42598900e253 >>> 2018-09-11 18:22:49,771 WARN org.apache.flink.configuration.Configuration >>> - Config uses deprecated configuration key >>> 'jobmanager.rpc.address' instead of proper key 'rest.address' >>> 2018-09-11 18:22:49,772 WARN >>> org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Upload >>> directory >>> /tmp/flink-web-c1cc0dde-0f4b-458e-84ba-841f405f3c78/flink-web-upload does >>> not exist, or has been deleted externally. Previously uploaded files are no >>> longer available. >>> 2018-09-11 18:22:49,772 INFO >>> org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Created >>> directory >>> /tmp/flink-web-c1cc0dde-0f4b-458e-84ba-841f405f3c78/flink-web-upload for >>> file uploads. >>> 2018-09-11 18:22:49,774 INFO >>> org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Starting >>> rest endpoint. >>> 2018-09-11 18:22:49,905 INFO >>> org.apache.flink.runtime.webmonitor.WebMonitorUtils - Determined >>> location of main cluster component log file: >>> /home/hadoop3/zh/flink-1.6.0/log/flink-hadoop3-standalonesession-0-p-a36-72.log >>> 2018-09-11 18:22:49,905 INFO >>> org.apache.flink.runtime.webmonitor.WebMonitorUtils - Determined >>> location of main cluster component stdout file: >>> /home/hadoop3/zh/flink-1.6.0/log/flink-hadoop3-standalonesession-0-p-a36-72.out >>> 2018-09-11 18:22:49,997 INFO >>> org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Rest >>> endpoint listening at p-a36-72:8081 >>> 2018-09-11 18:22:49,997 INFO >>> org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - >>> http://p-a36-72:8081 was granted leadership with >>> leaderSessionID=00000000-0000-0000-0000-000000000000 >>> 2018-09-11 18:22:49,997 INFO >>> org.apache.flink.runtime.dispatcher.DispatcherRestEndpoint - Web >>> frontend listening at http://p-a36-72:8081. >>> 2018-09-11 18:22:50,004 INFO >>> org.apache.flink.runtime.rpc.akka.AkkaRpcService - Starting >>> RPC endpoint for >>> org.apache.flink.runtime.resourcemanager.StandaloneResourceManager at >>> akka://flink/user/resourcemanager . >>> 2018-09-11 18:22:50,045 INFO >>> org.apache.flink.runtime.rpc.akka.AkkaRpcService - Starting >>> RPC endpoint for org.apache.flink.runtime.dispatcher.StandaloneDispatcher >>> at akka://flink/user/dispatcher . >>> 2018-09-11 18:22:50,055 INFO >>> org.apache.flink.runtime.resourcemanager.StandaloneResourceManager - >>> ResourceManager akka.tcp://flink@p-a36-72:6123/user/resourcemanager was >>> granted leadership with fencing token 00000000000000000000000000000000 >>> 2018-09-11 18:22:50,055 INFO >>> org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager - >>> Starting the SlotManager. >>> 2018-09-11 18:22:50,064 INFO >>> org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Dispatcher >>> akka.tcp://flink@p-a36-72:6123/user/dispatcher was granted leadership with >>> fencing token 00000000-0000-0000-0000-000000000000 >>> 2018-09-11 18:22:50,064 INFO >>> org.apache.flink.runtime.dispatcher.StandaloneDispatcher - Recovering >>> all persisted jobs. >>> 2018-09-11 18:22:55,316 INFO >>> org.apache.flink.runtime.resourcemanager.slotmanager.SlotManager - >>> Registering TaskManager c82e0b779fe62e3b3e6efda75c97cd18 under >>> c19ad7a1b58afdfc79b0fbbf08d43653 at the SlotManager. >>> >>> >>> Hanjing >>> >>> <https://maas.mail.163.com/dashi-web-extend/html/proSignature.html?ftlId=1&name=Hanjing&uid=hanjingzuzu%40163.com&iconUrl=http%3A%2F%2Fmail-online.nosdn.127.net%2Fsmda6015df3a52ec22402a83bdad785acb.jpg&items=%5B%22%22%2C%22%22%2C%22%22%2C%22%22%2C%22%22%5D> >>> 签名由 网易邮箱大师 <https://mail.163.com/dashi/dlpro.html?from=mail81> 定制 >>> On 9/11/2018 10:14,vino yang<yanghua1...@gmail.com> >>> <yanghua1...@gmail.com> wrote: >>> >>> Hi Hanjing, >>> >>> Is your JobManager working properly? Can you share your JobManager log? >>> >>> Thanks, vino. >>> >>> jing <hanjingz...@163.com> 于2018年9月11日周二 上午10:06写道: >>> >>>> Hi vino, >>>> >>>> I tried change "localhost" to the real IP. But still throw >>>> exception as below. JobManager configuration is as below. >>>> >>>> >>>> >>>> Thanks. >>>> >>>> Hanjing >>>> >>>> ------------------------------------ >>>> >>>> flink-conf.yaml: >>>> >>>> jobmanager.rpc.address: 170.0.0.46 >>>> >>>> >>>> >>>> # The RPC port where the JobManager is reachable. >>>> >>>> >>>> >>>> jobmanager.rpc.port: 6123 >>>> >>>> >>>> >>>> >>>> >>>> # The heap size for the JobManager JVM >>>> >>>> >>>> >>>> jobmanager.heap.size: 1024m >>>> >>>> >>>> >>>> >>>> >>>> # The heap size for the TaskManager JVM >>>> >>>> >>>> >>>> taskmanager.heap.size: 10240m >>>> >>>> >>>> >>>> >>>> >>>> # The number of task slots that each TaskManager offers. Each slot runs >>>> one parallel pipeline. >>>> >>>> >>>> >>>> taskmanager.numberOfTaskSlots: 16 >>>> >>>> >>>> >>>> # The parallelism used for programs that did not specify and other >>>> parallelism. >>>> >>>> >>>> >>>> parallelism.default: 2 >>>> >>>> >>>> >>>> >>>> >>>> Excepiton log: >>>> >>>> >>>> >>>> Starting execution of program >>>> >>>> >>>> >>>> ------------------------------------------------------------ >>>> >>>> The program finished with the following exception: >>>> >>>> >>>> >>>> org.apache.flink.client.program.ProgramInvocationException: The main >>>> method caused an error. >>>> >>>> at >>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:546) >>>> >>>> at >>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:421) >>>> >>>> at >>>> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:426) >>>> >>>> at >>>> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:804) >>>> >>>> at >>>> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:280) >>>> >>>> at >>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:215) >>>> >>>> at >>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1044) >>>> >>>> at >>>> org.apache.flink.client.cli.CliFrontend.lambda$main$11(CliFrontend.java:1120) >>>> >>>> at java.security.AccessController.doPrivileged(Native Method) >>>> >>>> at javax.security.auth.Subject.doAs(Subject.java:422) >>>> >>>> at >>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836) >>>> >>>> at >>>> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41) >>>> >>>> at >>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1120) >>>> >>>> Caused by: java.lang.RuntimeException: Could not connect to Flink >>>> JobManager with address 170.0.0.46:6123 >>>> >>>> at >>>> org.apache.flink.storm.api.FlinkClient.getTopologyJobId(FlinkClient.java:304) >>>> >>>> at >>>> org.apache.flink.storm.api.FlinkSubmitter.submitTopology(FlinkSubmitter.java:107) >>>> >>>> at >>>> org.apache.flink.storm.wordcount.WordCountRemoteBySubmitter.main(WordCountRemoteBySubmitter.java:75) >>>> >>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>>> >>>> at >>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>> >>>> at >>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>> >>>> at java.lang.reflect.Method.invoke(Method.java:498) >>>> >>>> at >>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:529) >>>> >>>> ... 12 more >>>> >>>> Caused by: java.io.IOException: Actor at akka.tcp:// >>>> flink@170.0.0.46:6123/user/jobmanager not reachable. Please make sure >>>> that the actor is running and its port is reachable. >>>> >>>> at >>>> org.apache.flink.runtime.akka.AkkaUtils$.getActorRef(AkkaUtils.scala:547) >>>> >>>> at >>>> org.apache.flink.runtime.akka.AkkaUtils.getActorRef(AkkaUtils.scala) >>>> >>>> at >>>> org.apache.flink.storm.api.FlinkClient.getJobManager(FlinkClient.java:339) >>>> >>>> at >>>> org.apache.flink.storm.api.FlinkClient.getTopologyJobId(FlinkClient.java:278) >>>> >>>> ... 19 more >>>> >>>> Caused by: akka.actor.ActorNotFound: Actor not found for: >>>> ActorSelection[Anchor(akka.tcp://flink@170.0.0.46:6123/), >>>> Path(/user/jobmanager)] >>>> >>>> at >>>> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:68) >>>> >>>> at >>>> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:66) >>>> >>>> at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36) >>>> >>>> at >>>> akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55) >>>> >>>> at >>>> akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:73) >>>> >>>> at >>>> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:76) >>>> >>>> at >>>> akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:120) >>>> >>>> at >>>> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:75) >>>> >>>> at >>>> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44) >>>> >>>> at >>>> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252) >>>> >>>> at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:534) >>>> >>>> at >>>> akka.remote.DefaultMessageDispatcher.dispatch(Endpoint.scala:97) >>>> >>>> at >>>> akka.remote.EndpointReader$$anonfun$receive$2.applyOrElse(Endpoint.scala:982) >>>> >>>> at akka.actor.Actor$class.aroundReceive(Actor.scala:502) >>>> >>>> at akka.remote.EndpointActor.aroundReceive(Endpoint.scala:446) >>>> >>>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526) >>>> >>>> at akka.actor.ActorCell.invoke(ActorCell.scala:495) >>>> >>>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257) >>>> >>>> at akka.dispatch.Mailbox.run(Mailbox.scala:224) >>>> >>>> at akka.dispatch.Mailbox.exec(Mailbox.scala:234) >>>> >>>> at >>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >>>> >>>> at >>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >>>> >>>> at >>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >>>> >>>> at >>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >>>> >>>> >>>> On 9/10/2018 20:17,vino yang<yanghua1...@gmail.com> >>>> <yanghua1...@gmail.com> wrote: >>>> >>>> Hi Hanjing, >>>> >>>> OK, I mean you change the "localhost" to the real IP. >>>> >>>> Try it. >>>> >>>> Thanks, vino. >>>> >>>> jing <hanjingz...@163.com> 于2018年9月10日周一 下午8:07写道: >>>> >>>>> Hi vino, >>>>> jonmanager rpc address value is setted by localhost. >>>>> hadoop3@p-a36-72 is the node host the jobmanager jvm. >>>>> >>>>> Thanks. >>>>> Hanjing >>>>> >>>>> >>>>> >>>>> jing >>>>> 邮箱hanjingz...@163.com >>>>> >>>>> <https://maas.mail.163.com/dashi-web-extend/html/proSignature.html?ftlId=1&name=jing&uid=hanjingzuzu%40163.com&iconUrl=https%3A%2F%2Fmail-online.nosdn.127.net%2Fqiyelogo%2FdefaultAvatar.png&items=%5B%22%E9%82%AE%E7%AE%B1hanjingzuzu%40163.com%22%5D> >>>>> >>>>> 签名由 网易邮箱大师 <https://mail.163.com/dashi/dlpro.html?from=mail88> 定制 >>>>> >>>>> On 09/10/2018 19:25, vino yang <yanghua1...@gmail.com> wrote: >>>>> Hi Hanjing, >>>>> >>>>> I mean this configuration key.[1] >>>>> >>>>> What's more, Is the "hadoop3@p-a36-72" also the node which host >>>>> JobManager's jvm process? >>>>> >>>>> Thanks, vino. >>>>> >>>>> [1]: >>>>> https://ci.apache.org/projects/flink/flink-docs-release-1.5/ops/config.html#jobmanager-rpc-address >>>>> >>>>> jing <hanjingz...@163.com> 于2018年9月10日周一 下午6:57写道: >>>>> >>>>>> Hi vino, >>>>>> I commit the job on the jvm code with the command below. >>>>>> hadoop3@p-a36-72 flink-1.6.0]$ ./bin/flink run >>>>>> WordCount-StormTopology.jar input output >>>>>> >>>>>> And I'm a new user, which configuation name should be set. All the >>>>>> configuations are the default setting now. >>>>>> >>>>>> Thanks. >>>>>> Hanjing >>>>>> >>>>>> jing >>>>>> 邮箱hanjingz...@163.com >>>>>> >>>>>> <https://maas.mail.163.com/dashi-web-extend/html/proSignature.html?ftlId=1&name=jing&uid=hanjingzuzu%40163.com&iconUrl=https%3A%2F%2Fmail-online.nosdn.127.net%2Fqiyelogo%2FdefaultAvatar.png&items=%5B%22%E9%82%AE%E7%AE%B1hanjingzuzu%40163.com%22%5D> >>>>>> >>>>>> 签名由 网易邮箱大师 <https://mail.163.com/dashi/dlpro.html?from=mail88> 定制 >>>>>> >>>>>> On 09/10/2018 15:49, vino yang <yanghua1...@gmail.com> wrote: >>>>>> Hi Hanjing, >>>>>> >>>>>> Did you perform a CLI commit on the JM node? Is the address bound to >>>>>> "localhost" in the Flink JM configuration? >>>>>> >>>>>> Thanks, vino. >>>>>> >>>>>> jing <hanjingz...@163.com> 于2018年9月10日周一 上午11:00写道: >>>>>> >>>>>>> Hello, >>>>>>> >>>>>>> I’m trying to run flink-storm-example on standalone >>>>>>> clusters. But there’s some exception I can’t sovle. Could anyone >>>>>>> please help me with trouble. >>>>>>> >>>>>>> flink-storm-example version: 1.60 >>>>>>> >>>>>>> flink version: 1.60 >>>>>>> >>>>>>> The log below is the Exception. My job manager status is as >>>>>>> the picture. >>>>>>> >>>>>>> I’v tried to changed the IP address and port, but it doesn’t’ >>>>>>> work. >>>>>>> >>>>>>> >>>>>>> >>>>>>> Thanks a lot. >>>>>>> >>>>>>> ------------------------------------------- >>>>>>> >>>>>>> [hadoop3@p-a36-72 flink-1.6.0]$ ./bin/flink run >>>>>>> WordCount-StormTopology.jar input output >>>>>>> >>>>>>> Starting execution of program >>>>>>> >>>>>>> >>>>>>> ------------------------------------------------------------ >>>>>>> >>>>>>> The program finished with the following exception: >>>>>>> >>>>>>> >>>>>>> org.apache.flink.client.program.ProgramInvocationException: The main >>>>>>> method caused an error. >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:546) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:421) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.program.ClusterClient.run(ClusterClient.java:426) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:804) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.runProgram(CliFrontend.java:280) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:215) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:1044) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$11(CliFrontend.java:1120) >>>>>>> >>>>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>>>> >>>>>>> at javax.security.auth.Subject.doAs(Subject.java:422) >>>>>>> >>>>>>> at >>>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1754) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:1120) >>>>>>> >>>>>>> Caused by: java.lang.RuntimeException: Could not connect to Flink >>>>>>> JobManager with address localhost:6123 >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.storm.api.FlinkClient.getTopologyJobId(FlinkClient.java:304) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.storm.api.FlinkSubmitter.submitTopology(FlinkSubmitter.java:107) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.storm.wordcount.WordCountRemoteBySubmitter.main(WordCountRemoteBySubmitter.java:75) >>>>>>> >>>>>>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native >>>>>>> Method) >>>>>>> >>>>>>> at >>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>>>>>> >>>>>>> at >>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>>>>>> >>>>>>> at java.lang.reflect.Method.invoke(Method.java:498) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:529) >>>>>>> >>>>>>> ... 12 more >>>>>>> >>>>>>> Caused by: java.io.IOException: Actor at >>>>>>> akka.tcp://flink@localhost:6123/user/jobmanager >>>>>>> not reachable. Please make sure that the actor is running and its port >>>>>>> is >>>>>>> reachable. >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.runtime.akka.AkkaUtils$.getActorRef(AkkaUtils.scala:547) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.runtime.akka.AkkaUtils.getActorRef(AkkaUtils.scala) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.storm.api.FlinkClient.getJobManager(FlinkClient.java:339) >>>>>>> >>>>>>> at >>>>>>> org.apache.flink.storm.api.FlinkClient.getTopologyJobId(FlinkClient.java:278) >>>>>>> >>>>>>> ... 19 more >>>>>>> >>>>>>> Caused by: akka.actor.ActorNotFound: Actor not found for: >>>>>>> ActorSelection[Anchor(akka.tcp://flink@localhost:6123/), >>>>>>> Path(/user/jobmanager)] >>>>>>> >>>>>>> at >>>>>>> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:68) >>>>>>> >>>>>>> at >>>>>>> akka.actor.ActorSelection$$anonfun$resolveOne$1.apply(ActorSelection.scala:66) >>>>>>> >>>>>>> at >>>>>>> scala.concurrent.impl.CallbackRunnable.run(Promise.scala:36) >>>>>>> >>>>>>> at >>>>>>> akka.dispatch.BatchingExecutor$AbstractBatch.processBatch(BatchingExecutor.scala:55) >>>>>>> >>>>>>> at >>>>>>> akka.dispatch.BatchingExecutor$Batch.run(BatchingExecutor.scala:73) >>>>>>> >>>>>>> at >>>>>>> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.unbatchedExecute(Future.scala:76) >>>>>>> >>>>>>> at >>>>>>> akka.dispatch.BatchingExecutor$class.execute(BatchingExecutor.scala:120) >>>>>>> >>>>>>> at >>>>>>> akka.dispatch.ExecutionContexts$sameThreadExecutionContext$.execute(Future.scala:75) >>>>>>> >>>>>>> at >>>>>>> scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:44) >>>>>>> >>>>>>> at >>>>>>> scala.concurrent.impl.Promise$DefaultPromise.tryComplete(Promise.scala:252) >>>>>>> >>>>>>> at akka.pattern.PromiseActorRef.$bang(AskSupport.scala:534) >>>>>>> >>>>>>> at >>>>>>> akka.remote.DefaultMessageDispatcher.dispatch(Endpoint.scala:97) >>>>>>> >>>>>>> at >>>>>>> akka.remote.EndpointReader$$anonfun$receive$2.applyOrElse(Endpoint.scala:982) >>>>>>> >>>>>>> at akka.actor.Actor$class.aroundReceive(Actor.scala:502) >>>>>>> >>>>>>> at >>>>>>> akka.remote.EndpointActor.aroundReceive(Endpoint.scala:446) >>>>>>> >>>>>>> at akka.actor.ActorCell.receiveMessage(ActorCell.scala:526) >>>>>>> >>>>>>> at akka.actor.ActorCell.invoke(ActorCell.scala:495) >>>>>>> >>>>>>> at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:257) >>>>>>> >>>>>>> at akka.dispatch.Mailbox.run(Mailbox.scala:224) >>>>>>> >>>>>>> at akka.dispatch.Mailbox.exec(Mailbox.scala:234) >>>>>>> >>>>>>> at >>>>>>> scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) >>>>>>> >>>>>>> at >>>>>>> scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) >>>>>>> >>>>>>> at >>>>>>> scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) >>>>>>> >>>>>>> at >>>>>>> scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> Hanjing >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>