Hi Robert, thanks for the quick response. Here is the jobmanager-main.log:
PS: I’m subscribed now. 11:09:16,144 INFO org.apache.flink.yarn.ApplicationMaster$ - YARN daemon runs as hadoop setting user to execute Flink ApplicationMaster/JobManager to hadoop 11:09:16,199 INFO org.apache.flink.yarn.Utils - Found YARN_CONF_DIR, adding it to configuration 11:09:16,298 INFO org.apache.flink.yarn.ApplicationMaster$ - Start job manager for yarn 11:09:16,298 INFO org.apache.flink.yarn.ApplicationMaster$ - Config path: /data/3/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/container_1420727594991_0068_01_000001. 11:09:17,179 INFO akka.event.slf4j.Slf4jLogger - Slf4jLogger started 11:09:17,271 INFO Remoting - Starting remoting 11:09:17,556 INFO Remoting - Remoting started; listening on addresses :[akka.tcp://fl...@cloud-29.dima.tu-berlin.de:42643] 11:09:17,569 INFO org.apache.flink.yarn.ApplicationMaster$ - Start job manager actor. 11:09:17,570 INFO org.apache.flink.yarn.ApplicationMaster$ - Generate configuration file for application master. 11:09:17,579 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Starting job manager at akka://flink/user/jobmanager. 11:09:17,590 INFO org.apache.flink.yarn.ApplicationMaster$ - Start yarn session on job manager. 11:09:17,591 INFO org.apache.flink.yarn.ApplicationMaster$ - Application Master properly initiated. Await termination of actor system. 11:09:17,601 INFO org.apache.flink.runtime.blob.BlobServer - Started BLOB server on port 35605 11:09:17,616 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Started job manager. Waiting for incoming messages. 11:09:17,621 INFO org.apache.flink.runtime.jobmanager.web.WebInfoServer - Setting up web info server, using web-root directoryjar:file:/data/4/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/filecache/11/flink-dist-0.9-SNAPSHOT-yarn-uberjar.jar!/web-docs-infoserver. 11:09:17,621 INFO org.apache.flink.runtime.jobmanager.web.WebInfoServer - Web info server will display information about flink job-manager on localhost, port 8081. 11:09:17,733 INFO org.apache.flink.runtime.jobmanager.web.WebInfoServer - Starting web info server for JobManager on port 8081 11:09:17,734 INFO org.eclipse.jetty.util.log - jetty-0.9-SNAPSHOT 11:09:17,770 INFO org.eclipse.jetty.util.log - Started SelectChannelConnector@0.0.0.0:8081 11:09:17,784 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Start yarn session. 11:09:17,784 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Requesting 2 task managers. 11:09:18,414 INFO org.apache.hadoop.yarn.client.RMProxy - Connecting to ResourceManager at cloud-11/130.149.21.15:8030 11:09:18,423 INFO org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy - yarn.client.max-nodemanagers-proxies : 500 11:09:18,425 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Registering ApplicationMaster with tracking url http://cloud-29.dima.tu-berlin.de:8081. 11:09:18,787 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Requesting TaskManager container 0. 11:09:18,794 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Requesting TaskManager container 1. 11:09:18,903 INFO org.apache.flink.yarn.Utils - Copying from file:/data/3/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/container_1420727594991_0068_01_000001/flink-conf-modified.yaml to hdfs://cloud-11.dima.tu-berlin.de:60010/user/hadoop/.flink/application_1420727594991_0068/flink-conf-modified.yaml 11:09:19,280 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Prepared local resource for modified yaml: resource { scheme: "hdfs" host: "cloud-11.dima.tu-berlin.de" port: 60010 file: "/user/hadoop/.flink/application_1420727594991_0068/flink-conf-modified.yaml" } size: 3121 timestamp: 1422439759181 type: FILE visibility: APPLICATION 11:09:19,286 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Create container launch context. 11:09:19,300 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Starting TM with command=$JAVA_HOME/bin/java -Xmx819m -Dlog.file="<LOG_DIR>/taskmanager.log" -Dlogback.configurationFile=file:logback.xml -Dlog4j.configuration=file:log4j.properties org.apache.flink.yarn.appMaster.YarnTaskManagerRunner --configDir . 1> <LOG_DIR>/taskmanager-stdout.log 2> <LOG_DIR>/taskmanager-stderr.log 11:09:19,619 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Got new container for TM container_1420727594991_0068_01_000002 on host cloud-26.dima.tu-berlin.de 11:09:19,621 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Launching container #1. 11:09:19,622 INFO org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy - Opening proxy : cloud-26.dima.tu-berlin.de:8045 11:09:19,828 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Got new container for TM container_1420727594991_0068_01_000003 on host cloud-31.dima.tu-berlin.de 11:09:19,829 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Launching container #2. 11:09:19,831 INFO org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy - Opening proxy : cloud-31.dima.tu-berlin.de:8045 11:09:25,748 INFO org.apache.flink.runtime.instance.InstanceManager - Registered TaskManager at akka.tcp://fl...@cloud-26.dima.tu-berlin.de:51449/user/taskmanager as ca731dbbfe941a63077c9c148254607c. Current number of registered hosts is 1. 11:09:25,845 INFO org.apache.flink.runtime.instance.InstanceManager - Registered TaskManager at akka.tcp://fl...@cloud-31.dima.tu-berlin.de:43200/user/taskmanager as 7bf3aa638e00682239c9d8460d142191. Current number of registered hosts is 2. 11:11:13,837 WARN akka.remote.ReliableDeliverySupervisor - Association with remote system [akka.tcp://CliFrontendActorSystem@130.149.21.15:50292] has failed, address is now gated for [5000] ms. Reason is: [Disassociated]. 11:14:30,892 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Received job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 11:14:28 CET 2015). 11:14:30,997 INFO org.apache.hadoop.conf.Configuration.deprecation - job.end.retry.interval is deprecated. Instead, use mapreduce.job.end-notification.retry.interval 11:14:30,998 INFO org.apache.hadoop.conf.Configuration.deprecation - io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 11:14:30,999 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker.retiredjobs.cache.size is deprecated. Instead, use mapreduce.jobtracker.retiredjobs.cache.size 11:14:31,000 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.task.profile.reduces is deprecated. Instead, use mapreduce.task.profile.reduces 11:14:31,000 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.reuse.jvm.num.tasks is deprecated. Instead, use mapreduce.job.jvm.numtasks 11:14:31,001 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.reduce.tasks.speculative.execution is deprecated. Instead, use mapreduce.reduce.speculative 11:14:31,003 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.task.tracker.http.address is deprecated. Instead, use mapreduce.tasktracker.http.address 11:14:31,004 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.system.dir is deprecated. Instead, use mapreduce.jobtracker.system.dir 11:14:31,005 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.task.tracker.report.address is deprecated. Instead, use mapreduce.tasktracker.report.address 11:14:31,006 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.healthChecker.interval is deprecated. Instead, use mapreduce.tasktracker.healthchecker.interval 11:14:31,007 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.child.tmp is deprecated. Instead, use mapreduce.task.tmp.dir 11:14:31,008 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.tasktracker.taskmemorymanager.monitoring-interval is deprecated. Instead, use mapreduce.tasktracker.taskmemorymanager.monitoringinterval 11:14:31,008 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.shuffle.connect.timeout is deprecated. Instead, use mapreduce.reduce.shuffle.connect.timeout 11:14:31,008 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.speculative.execution.speculativeCap is deprecated. Instead, use mapreduce.job.speculative.speculativecap 11:14:31,009 INFO org.apache.hadoop.conf.Configuration.deprecation - io.sort.spill.percent is deprecated. Instead, use mapreduce.map.sort.spill.percent 11:14:31,010 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.shuffle.input.buffer.percent is deprecated. Instead, use mapreduce.reduce.shuffle.input.buffer.percent 11:14:31,010 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.skip.map.max.skip.records is deprecated. Instead, use mapreduce.map.skip.maxrecords 11:14:31,011 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.task.profile.maps is deprecated. Instead, use mapreduce.task.profile.maps 11:14:31,012 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.local.dir is deprecated. Instead, use mapreduce.cluster.local.dir 11:14:31,013 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.merge.recordsBeforeProgress is deprecated. Instead, use mapreduce.task.merge.progress.records 11:14:31,013 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker.http.address is deprecated. Instead, use mapreduce.jobtracker.http.address 11:14:31,013 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.userlog.retain.hours is deprecated. Instead, use mapreduce.job.userlog.retain.hours 11:14:31,014 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.compress.map.output is deprecated. Instead, use mapreduce.map.output.compress 11:14:31,014 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.speculative.execution.slowNodeThreshold is deprecated. Instead, use mapreduce.job.speculative.slownodethreshold 11:14:31,015 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.tasktracker.reduce.tasks.maximum is deprecated. Instead, use mapreduce.tasktracker.reduce.tasks.maximum 11:14:31,017 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.reduce.child.log.level is deprecated. Instead, use mapreduce.reduce.log.level 11:14:31,018 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.jobtracker.restart.recover is deprecated. Instead, use mapreduce.jobtracker.restart.recover 11:14:31,020 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.inmem.merge.threshold is deprecated. Instead, use mapreduce.reduce.merge.inmem.threshold 11:14:31,021 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.reduce.tasks is deprecated. Instead, use mapreduce.job.reduces 11:14:31,023 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.acls.enabled is deprecated. Instead, use mapreduce.cluster.acls.enabled 11:14:31,023 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.tasktracker.dns.nameserver is deprecated. Instead, use mapreduce.tasktracker.dns.nameserver 11:14:31,024 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.submit.replication is deprecated. Instead, use mapreduce.client.submit.file.replication 11:14:31,024 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.map.tasks.speculative.execution is deprecated. Instead, use mapreduce.map.speculative 11:14:31,025 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.map.child.log.level is deprecated. Instead, use mapreduce.map.log.level 11:14:31,026 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.map.max.attempts is deprecated. Instead, use mapreduce.map.maxattempts 11:14:31,026 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.shuffle.merge.percent is deprecated. Instead, use mapreduce.reduce.shuffle.merge.percent 11:14:31,026 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker.jobhistory.lru.cache.size is deprecated. Instead, use mapreduce.jobtracker.jobhistory.lru.cache.size 11:14:31,029 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.queue.name is deprecated. Instead, use mapreduce.job.queuename 11:14:31,030 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker.persist.jobstatus.hours is deprecated. Instead, use mapreduce.jobtracker.persist.jobstatus.hours 11:14:31,031 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.healthChecker.script.timeout is deprecated. Instead, use mapreduce.tasktracker.healthchecker.script.timeout 11:14:31,031 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.tasktracker.map.tasks.maximum is deprecated. Instead, use mapreduce.tasktracker.map.tasks.maximum 11:14:31,032 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.reduce.markreset.buffer.percent is deprecated. Instead, use mapreduce.reduce.markreset.buffer.percent 11:14:31,032 INFO org.apache.hadoop.conf.Configuration.deprecation - jobclient.completion.poll.interval is deprecated. Instead, use mapreduce.client.completion.pollinterval 11:14:31,032 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker.persist.jobstatus.dir is deprecated. Instead, use mapreduce.jobtracker.persist.jobstatus.dir 11:14:31,033 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.reduce.slowstart.completed.maps is deprecated. Instead, use mapreduce.job.reduce.slowstart.completedmaps 11:14:31,034 INFO org.apache.hadoop.conf.Configuration.deprecation - io.sort.mb is deprecated. Instead, use mapreduce.task.io.sort.mb 11:14:31,034 INFO org.apache.hadoop.conf.Configuration.deprecation - dfs.umaskmode is deprecated. Instead, use fs.permissions.umask-mode 11:14:31,037 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.jobtracker.instrumentation is deprecated. Instead, use mapreduce.jobtracker.instrumentation 11:14:31,038 INFO org.apache.hadoop.conf.Configuration.deprecation - topology.node.switch.mapping.impl is deprecated. Instead, use net.topology.node.switch.mapping.impl 11:14:31,039 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.output.compression.type is deprecated. Instead, use mapreduce.output.fileoutputformat.compress.type 11:14:31,039 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.skip.attempts.to.start.skipping is deprecated. Instead, use mapreduce.task.skip.start.attempts 11:14:31,040 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.task.tracker.task-controller is deprecated. Instead, use mapreduce.tasktracker.taskcontroller 11:14:31,041 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.userlog.limit.kb is deprecated. Instead, use mapreduce.task.userlog.limit.kb 11:14:31,043 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.committer.job.setup.cleanup.needed is deprecated. Instead, use mapreduce.job.committer.setup.cleanup.needed 11:14:31,043 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.map.tasks is deprecated. Instead, use mapreduce.job.maps 11:14:31,044 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.local.dir.minspacekill is deprecated. Instead, use mapreduce.tasktracker.local.dir.minspacekill 11:14:31,044 INFO org.apache.hadoop.conf.Configuration.deprecation - mapreduce.jobtracker.split.metainfo.maxsize is deprecated. Instead, use mapreduce.job.split.metainfo.maxsize 11:14:31,045 INFO org.apache.hadoop.conf.Configuration.deprecation - jobclient.progress.monitor.poll.interval is deprecated. Instead, use mapreduce.client.progressmonitor.pollinterval 11:14:31,046 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.min.split.size is deprecated. Instead, use mapreduce.input.fileinputformat.split.minsize 11:14:31,047 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.output.compression.codec is deprecated. Instead, use mapreduce.output.fileoutputformat.compress.codec 11:14:31,048 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.task.profile is deprecated. Instead, use mapreduce.task.profile 11:14:31,048 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.reduce.parallel.copies is deprecated. Instead, use mapreduce.reduce.shuffle.parallelcopies 11:14:31,049 INFO org.apache.hadoop.conf.Configuration.deprecation - io.sort.factor is deprecated. Instead, use mapreduce.task.io.sort.factor 11:14:31,050 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.input.dir is deprecated. Instead, use mapreduce.input.fileinputformat.inputdir 11:14:31,050 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.task.timeout is deprecated. Instead, use mapreduce.task.timeout 11:14:31,052 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.heartbeats.in.second is deprecated. Instead, use mapreduce.jobtracker.heartbeats.in.second 11:14:31,053 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.task.cache.levels is deprecated. Instead, use mapreduce.jobtracker.taskcache.levels 11:14:31,053 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.map.output.compression.codec is deprecated. Instead, use mapreduce.map.output.compress.codec 11:14:31,054 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.tasktracker.dns.interface is deprecated. Instead, use mapreduce.tasktracker.dns.interface 11:14:31,055 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.max.tracker.failures is deprecated. Instead, use mapreduce.job.maxtaskfailures.per.tracker 11:14:31,056 INFO org.apache.hadoop.conf.Configuration.deprecation - dfs.df.interval is deprecated. Instead, use fs.df.interval 11:14:31,056 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.tasktracker.tasks.sleeptime-before-sigkill is deprecated. Instead, use mapreduce.tasktracker.tasks.sleeptimebeforesigkill 11:14:31,057 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.max.tracker.blacklists is deprecated. Instead, use mapreduce.jobtracker.tasktracker.maxblacklists 11:14:31,058 INFO org.apache.hadoop.conf.Configuration.deprecation - jobclient.output.filter is deprecated. Instead, use mapreduce.client.output.filter 11:14:31,059 INFO org.apache.hadoop.conf.Configuration.deprecation - job.end.retry.attempts is deprecated. Instead, use mapreduce.job.end-notification.retry.attempts 11:14:31,059 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.jobtracker.taskScheduler is deprecated. Instead, use mapreduce.jobtracker.taskscheduler 11:14:31,060 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.speculative.execution.slowTaskThreshold is deprecated. Instead, use mapreduce.job.speculative.slowtaskthreshold 11:14:31,061 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.tasktracker.indexcache.mb is deprecated. Instead, use mapreduce.tasktracker.indexcache.mb 11:14:31,061 INFO org.apache.hadoop.conf.Configuration.deprecation - tasktracker.http.threads is deprecated. Instead, use mapreduce.tasktracker.http.threads 11:14:31,062 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker.handler.count is deprecated. Instead, use mapreduce.jobtracker.handler.count 11:14:31,062 INFO org.apache.hadoop.conf.Configuration.deprecation - keep.failed.task.files is deprecated. Instead, use mapreduce.task.files.preserve.failedtasks 11:14:31,062 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.output.compress is deprecated. Instead, use mapreduce.output.fileoutputformat.compress 11:14:31,063 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.jobtracker.job.history.block.size is deprecated. Instead, use mapreduce.jobtracker.jobhistory.block.size 11:14:31,063 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.skip.reduce.max.skip.groups is deprecated. Instead, use mapreduce.reduce.skip.maxgroups 11:14:31,066 INFO org.apache.hadoop.conf.Configuration.deprecation - topology.script.number.args is deprecated. Instead, use net.topology.script.number.args 11:14:31,066 INFO org.apache.hadoop.conf.Configuration.deprecation - fs.default.name is deprecated. Instead, use fs.defaultFS 11:14:31,067 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.local.dir.minspacestart is deprecated. Instead, use mapreduce.tasktracker.local.dir.minspacestart 11:14:31,067 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.jobtracker.maxtasks.per.job is deprecated. Instead, use mapreduce.jobtracker.maxtasks.perjob 11:14:31,068 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.reduce.max.attempts is deprecated. Instead, use mapreduce.reduce.maxattempts 11:14:31,068 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address 11:14:31,069 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.shuffle.read.timeout is deprecated. Instead, use mapreduce.reduce.shuffle.read.timeout 11:14:31,069 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.tasktracker.instrumentation is deprecated. Instead, use mapreduce.tasktracker.instrumentation 11:14:31,070 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.tasktracker.expiry.interval is deprecated. Instead, use mapreduce.jobtracker.expire.trackers.interval 11:14:31,071 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.tracker.persist.jobstatus.active is deprecated. Instead, use mapreduce.jobtracker.persist.jobstatus.active 11:14:31,074 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.temp.dir is deprecated. Instead, use mapreduce.cluster.temp.dir 11:14:31,075 INFO org.apache.hadoop.conf.Configuration.deprecation - hadoop.native.lib is deprecated. Instead, use io.native.lib.available 11:14:31,075 INFO org.apache.hadoop.conf.Configuration.deprecation - mapreduce.job.counters.limit is deprecated. Instead, use mapreduce.job.counters.max 11:14:31,076 INFO org.apache.hadoop.conf.Configuration.deprecation - mapred.job.reduce.input.buffer.percent is deprecated. Instead, use mapreduce.reduce.input.buffer.percent 11:14:31,149 INFO org.apache.hadoop.mapreduce.lib.input.FileInputFormat - Total input paths to process : 1 11:14:31,177 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Scheduling job Flink Java Job at Wed Jan 28 11:14:28 CET 2015. 11:14:31,188 INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying CHAIN DataSource (at createInput(ExecutionEnvironment.java:426) (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap (FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-26 11:14:31,202 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 11:14:28 CET 2015) changed to RUNNING. 11:14:32,525 INFO org.apache.flink.api.common.io.LocatableInputSplitAssigner - Assigning remote split to host cloud-26 11:14:32,603 WARN akka.remote.ReliableDeliverySupervisor - Association with remote system [akka.tcp://fl...@cloud-26.dima.tu-berlin.de:51449] has failed, address is now gated for [5000] ms. Reason is: [Disassociated]. 11:16:12,574 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 11:14:28 CET 2015) changed to FAILINGjava.lang.RuntimeException: Requesting the next InputSplit failed. at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63) at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355) at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154) at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204) at java.lang.Thread.run(Thread.java:745) Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds] at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) at scala.concurrent.Await$.result(package.scala:107) at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265) at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala) at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56) ... 4 more . 11:16:12,584 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 11:14:28 CET 2015) changed to FAILEDjava.lang.RuntimeException: Requesting the next InputSplit failed. at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63) at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355) at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154) at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204) at java.lang.Thread.run(Thread.java:745) Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds] at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) at scala.concurrent.Await$.result(package.scala:107) at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265) at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala) at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56) ... 4 more . 11:17:53,055 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Received job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 11:17:50 CET 2015). 11:17:53,258 INFO org.apache.hadoop.conf.Configuration.deprecation - io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 11:17:53,312 INFO org.apache.hadoop.conf.Configuration.deprecation - fs.default.name is deprecated. Instead, use fs.defaultFS 11:17:53,320 INFO org.apache.hadoop.conf.Configuration.deprecation - mapreduce.job.counters.limit is deprecated. Instead, use mapreduce.job.counters.max 11:17:53,347 INFO org.apache.hadoop.mapreduce.lib.input.FileInputFormat - Total input paths to process : 1 11:17:53,352 INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying CHAIN DataSource (at createInput(ExecutionEnvironment.java:426) (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap (FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-31 11:17:53,353 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Scheduling job Flink Java Job at Wed Jan 28 11:17:50 CET 2015. 11:17:53,354 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 11:17:50 CET 2015) changed to RUNNING. 11:17:54,646 INFO org.apache.flink.api.common.io.LocatableInputSplitAssigner - Assigning remote split to host cloud-31 11:17:54,715 WARN akka.remote.ReliableDeliverySupervisor - Association with remote system [akka.tcp://fl...@cloud-31.dima.tu-berlin.de:43200] has failed, address is now gated for [5000] ms. Reason is: [Disassociated]. 11:19:34,684 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 11:17:50 CET 2015) changed to FAILINGjava.lang.RuntimeException: Requesting the next InputSplit failed. at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63) at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355) at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154) at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204) at java.lang.Thread.run(Thread.java:745) Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds] at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) at scala.concurrent.Await$.result(package.scala:107) at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265) at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala) at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56) ... 4 more . 11:19:34,697 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 11:17:50 CET 2015) changed to FAILEDjava.lang.RuntimeException: Requesting the next InputSplit failed. at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63) at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355) at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154) at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204) at java.lang.Thread.run(Thread.java:745) Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds] at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) at scala.concurrent.Await$.result(package.scala:107) at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265) at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala) at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56) ... 4 more . 11:42:52,441 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Received job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 11:42:50 CET 2015). 11:42:52,650 INFO org.apache.hadoop.conf.Configuration.deprecation - io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum 11:42:52,705 INFO org.apache.hadoop.conf.Configuration.deprecation - fs.default.name is deprecated. Instead, use fs.defaultFS 11:42:52,713 INFO org.apache.hadoop.conf.Configuration.deprecation - mapreduce.job.counters.limit is deprecated. Instead, use mapreduce.job.counters.max 11:42:52,740 INFO org.apache.hadoop.mapreduce.lib.input.FileInputFormat - Total input paths to process : 1 11:42:52,747 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Scheduling job Flink Java Job at Wed Jan 28 11:42:50 CET 2015. 11:42:52,747 INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - Deploying CHAIN DataSource (at createInput(ExecutionEnvironment.java:426) (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap (FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-26 11:42:52,750 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 11:42:50 CET 2015) changed to RUNNING. 11:42:52,956 INFO org.apache.flink.api.common.io.LocatableInputSplitAssigner - Assigning remote split to host cloud-26 11:42:52,983 WARN akka.remote.ReliableDeliverySupervisor - Association with remote system [akka.tcp://fl...@cloud-26.dima.tu-berlin.de:51449] has failed, address is now gated for [5000] ms. Reason is: [Disassociated]. 11:44:32,976 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 11:42:50 CET 2015) changed to FAILINGjava.lang.RuntimeException: Requesting the next InputSplit failed. at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63) at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355) at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154) at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204) at java.lang.Thread.run(Thread.java:745) Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds] at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) at scala.concurrent.Await$.result(package.scala:107) at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265) at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala) at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56) ... 4 more . 11:44:32,984 INFO org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1 - Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 11:42:50 CET 2015) changed to FAILEDjava.lang.RuntimeException: Requesting the next InputSplit failed. at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63) at org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355) at org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154) at org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204) at java.lang.Thread.run(Thread.java:745) Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 seconds] at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) at scala.concurrent.Await$.result(package.scala:107) at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265) at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala) at org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56) ... 4 more . On 28 Jan 2015, at 12:08, Robert Metzger <rmetz...@apache.org> wrote: > Hi, > > it seems that you are not subscribed to our mailing list, so I had to > manually accept your mail. Would be good if you could subscribe. > > Can you send us also the log output of the JobManager? > If your YARN cluster has log aggregation activated, you can retrieve the > logs of a stopped YARN session using: > yarn logs -applicationId <AppId> > > watch out for the jobmanager-main.log or so file. > > I suspect that there has been an exception on the JobManager. > > Best, > Robert > > > > On Wed, Jan 28, 2015 at 12:01 PM, Bruecke, Christoph < > christoph.brue...@campus.tu-berlin.de> wrote: > >> Hi, >> >> I have written a job that reads a SequenceFile from HDFS using the >> Hadoop-Compatibility add-on. Doing so results in a TimeoutException. I’m >> using flink-0.9-SNAPSHOT with PR 342 ( >> https://github.com/apache/flink/pull/342 ). Furthermore I’m running flink >> on yarn with two TM using flink-yarn-0.9-SNAPSHOT/bin/yarn-session.sh -n 2. >> >> Is this a bug or is there something wrong with the configuration? >> >> 01/28/2015 11:42:52 Job execution switched to status RUNNING. >> 01/28/2015 11:42:52 CHAIN DataSource (at >> createInput(ExecutionEnvironment.java:426) >> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> >> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to SCHEDULED >> 01/28/2015 11:42:52 CHAIN DataSource (at >> createInput(ExecutionEnvironment.java:426) >> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> >> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to DEPLOYING >> 01/28/2015 11:42:52 CHAIN DataSource (at >> createInput(ExecutionEnvironment.java:426) >> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> >> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to RUNNING >> 01/28/2015 11:44:32 CHAIN DataSource (at >> createInput(ExecutionEnvironment.java:426) >> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> >> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to FAILED >> java.lang.RuntimeException: Requesting the next InputSplit failed. >> at >> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63) >> at >> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355) >> at >> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154) >> at >> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204) >> at java.lang.Thread.run(Thread.java:745) >> Caused by: java.util.concurrent.TimeoutException: Futures timed out after >> [100 seconds] >> at >> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) >> at >> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) >> at >> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) >> at >> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) >> at scala.concurrent.Await$.result(package.scala:107) >> at >> org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265) >> at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala) >> at >> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56) >> ... 4 more >> >> 01/28/2015 11:44:32 Job execution switched to status FAILING. >> 01/28/2015 11:44:32 GroupReduce (GroupReduce at >> main(ThiaziParser.java:40))(1/1) switched to CANCELED >> 01/28/2015 11:44:32 DataSink(TextOutputFormat (hdfs:// >> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/authors) >> - UTF-8)(1/1) switched to CANCELED >> 01/28/2015 11:44:32 CHAIN GroupReduce (GroupReduce at >> main(ThiaziParser.java:74)) -> Filter (Filter at >> main(ThiaziParser.java:97))(1/1) switched to CANCELED >> 01/28/2015 11:44:32 DataSink(TextOutputFormat (hdfs:// >> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/posts) - >> UTF-8)(1/1) switched to CANCELED >> 01/28/2015 11:44:32 CHAIN FlatMap (FlatMap at >> main(ThiaziParser.java:126)) -> Combine(SUM(1), at >> main(ThiaziParser.java:140)(1/1) switched to CANCELED >> 01/28/2015 11:44:32 Reduce (SUM(1), at >> main(ThiaziParser.java:140)(1/1) switched to CANCELED >> 01/28/2015 11:44:32 DataSink(CsvOutputFormat (path: hdfs:// >> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/wordcount, >> delimiter: ,))(1/1) switched to CANCELED >> 01/28/2015 11:44:32 GroupReduce (GroupReduce at >> main(ThiaziParser.java:106))(1/1) switched to CANCELED >> 01/28/2015 11:44:32 DataSink(TextOutputFormat (hdfs:// >> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/threads) >> - UTF-8)(1/1) switched to CANCELED >> 01/28/2015 11:44:32 Job execution switched to status FAILED. >> Error: The program execution failed: java.lang.RuntimeException: >> Requesting the next InputSplit failed. >> at >> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63) >> at >> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355) >> at >> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154) >> at >> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204) >> at java.lang.Thread.run(Thread.java:745) >> Caused by: java.util.concurrent.TimeoutException: Futures timed out after >> [100 seconds] >> at >> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) >> at >> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) >> at >> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107) >> at >> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) >> at scala.concurrent.Await$.result(package.scala:107) >> at >> org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265) >> at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala) >> at >> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56) >> ... 4 more >> >> >>