I might add that the error only occurs when running with the RemoteExecutor 
regardless of the number of TM. Starting the job in IntelliJ with the 
LocalExecutor with dop 1 works just fine.

Best,
Christoph

On 28 Jan 2015, at 12:17, Bruecke, Christoph 
<christoph.brue...@campus.tu-berlin.de> wrote:

> Hi Robert,
> 
> thanks for the quick response. Here is the jobmanager-main.log:
> 
> PS: I’m subscribed now.
> 
> 11:09:16,144 INFO  org.apache.flink.yarn.ApplicationMaster$                   
>    - YARN daemon runs as hadoop setting user to execute Flink 
> ApplicationMaster/JobManager to hadoop
> 11:09:16,199 INFO  org.apache.flink.yarn.Utils                                
>    - Found YARN_CONF_DIR, adding it to configuration
> 11:09:16,298 INFO  org.apache.flink.yarn.ApplicationMaster$                   
>    - Start job manager for yarn
> 11:09:16,298 INFO  org.apache.flink.yarn.ApplicationMaster$                   
>    - Config path: 
> /data/3/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/container_1420727594991_0068_01_000001.
> 11:09:17,179 INFO  akka.event.slf4j.Slf4jLogger                               
>    - Slf4jLogger started
> 11:09:17,271 INFO  Remoting                                                   
>    - Starting remoting
> 11:09:17,556 INFO  Remoting                                                   
>    - Remoting started; listening on addresses 
> :[akka.tcp://fl...@cloud-29.dima.tu-berlin.de:42643]
> 11:09:17,569 INFO  org.apache.flink.yarn.ApplicationMaster$                   
>    - Start job manager actor.
> 11:09:17,570 INFO  org.apache.flink.yarn.ApplicationMaster$                   
>    - Generate configuration file for application master.
> 11:09:17,579 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Starting job manager at akka://flink/user/jobmanager.
> 11:09:17,590 INFO  org.apache.flink.yarn.ApplicationMaster$                   
>    - Start yarn session on job manager.
> 11:09:17,591 INFO  org.apache.flink.yarn.ApplicationMaster$                   
>    - Application Master properly initiated. Await termination of actor system.
> 11:09:17,601 INFO  org.apache.flink.runtime.blob.BlobServer                   
>    - Started BLOB server on port 35605
> 11:09:17,616 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Started job manager. Waiting for incoming messages.
> 11:09:17,621 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer      
>    - Setting up web info server, using web-root 
> directoryjar:file:/data/4/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/filecache/11/flink-dist-0.9-SNAPSHOT-yarn-uberjar.jar!/web-docs-infoserver.
> 11:09:17,621 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer      
>    - Web info server will display information about flink job-manager on 
> localhost, port 8081.
> 11:09:17,733 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer      
>    - Starting web info server for JobManager on port 8081
> 11:09:17,734 INFO  org.eclipse.jetty.util.log                                 
>    - jetty-0.9-SNAPSHOT
> 11:09:17,770 INFO  org.eclipse.jetty.util.log                                 
>    - Started SelectChannelConnector@0.0.0.0:8081
> 11:09:17,784 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Start yarn session.
> 11:09:17,784 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Requesting 2 task managers.
> 11:09:18,414 INFO  org.apache.hadoop.yarn.client.RMProxy                      
>    - Connecting to ResourceManager at cloud-11/130.149.21.15:8030
> 11:09:18,423 INFO  
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy  - 
> yarn.client.max-nodemanagers-proxies : 500
> 11:09:18,425 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Registering ApplicationMaster with tracking url 
> http://cloud-29.dima.tu-berlin.de:8081.
> 11:09:18,787 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Requesting TaskManager container 0.
> 11:09:18,794 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Requesting TaskManager container 1.
> 11:09:18,903 INFO  org.apache.flink.yarn.Utils                                
>    - Copying from 
> file:/data/3/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/container_1420727594991_0068_01_000001/flink-conf-modified.yaml
>  to 
> hdfs://cloud-11.dima.tu-berlin.de:60010/user/hadoop/.flink/application_1420727594991_0068/flink-conf-modified.yaml
> 11:09:19,280 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Prepared local resource for modified yaml: resource { scheme: "hdfs" host: 
> "cloud-11.dima.tu-berlin.de" port: 60010 file: 
> "/user/hadoop/.flink/application_1420727594991_0068/flink-conf-modified.yaml" 
> } size: 3121 timestamp: 1422439759181 type: FILE visibility: APPLICATION
> 11:09:19,286 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Create container launch context.
> 11:09:19,300 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Starting TM with command=$JAVA_HOME/bin/java -Xmx819m  
> -Dlog.file="<LOG_DIR>/taskmanager.log" 
> -Dlogback.configurationFile=file:logback.xml 
> -Dlog4j.configuration=file:log4j.properties 
> org.apache.flink.yarn.appMaster.YarnTaskManagerRunner --configDir . 1> 
> <LOG_DIR>/taskmanager-stdout.log 2> <LOG_DIR>/taskmanager-stderr.log
> 11:09:19,619 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Got new container for TM container_1420727594991_0068_01_000002 on host 
> cloud-26.dima.tu-berlin.de
> 11:09:19,621 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Launching container #1.
> 11:09:19,622 INFO  
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy  - 
> Opening proxy : cloud-26.dima.tu-berlin.de:8045
> 11:09:19,828 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Got new container for TM container_1420727594991_0068_01_000003 on host 
> cloud-31.dima.tu-berlin.de
> 11:09:19,829 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Launching container #2.
> 11:09:19,831 INFO  
> org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy  - 
> Opening proxy : cloud-31.dima.tu-berlin.de:8045
> 11:09:25,748 INFO  org.apache.flink.runtime.instance.InstanceManager          
>    - Registered TaskManager at 
> akka.tcp://fl...@cloud-26.dima.tu-berlin.de:51449/user/taskmanager as 
> ca731dbbfe941a63077c9c148254607c. Current number of registered hosts is 1.
> 11:09:25,845 INFO  org.apache.flink.runtime.instance.InstanceManager          
>    - Registered TaskManager at 
> akka.tcp://fl...@cloud-31.dima.tu-berlin.de:43200/user/taskmanager as 
> 7bf3aa638e00682239c9d8460d142191. Current number of registered hosts is 2.
> 11:11:13,837 WARN  akka.remote.ReliableDeliverySupervisor                     
>    - Association with remote system 
> [akka.tcp://CliFrontendActorSystem@130.149.21.15:50292] has failed, address 
> is now gated for [5000] ms. Reason is: [Disassociated].
> 11:14:30,892 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Received job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 
> 11:14:28 CET 2015).
> 11:14:30,997 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - job.end.retry.interval is deprecated. Instead, use 
> mapreduce.job.end-notification.retry.interval
> 11:14:30,998 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum
> 11:14:30,999 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.tracker.retiredjobs.cache.size is deprecated. Instead, use 
> mapreduce.jobtracker.retiredjobs.cache.size
> 11:14:31,000 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.task.profile.reduces is deprecated. Instead, use 
> mapreduce.task.profile.reduces
> 11:14:31,000 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.reuse.jvm.num.tasks is deprecated. Instead, use 
> mapreduce.job.jvm.numtasks
> 11:14:31,001 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.reduce.tasks.speculative.execution is deprecated. Instead, use 
> mapreduce.reduce.speculative
> 11:14:31,003 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.task.tracker.http.address is deprecated. Instead, use 
> mapreduce.tasktracker.http.address
> 11:14:31,004 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.system.dir is deprecated. Instead, use 
> mapreduce.jobtracker.system.dir
> 11:14:31,005 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.task.tracker.report.address is deprecated. Instead, use 
> mapreduce.tasktracker.report.address
> 11:14:31,006 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.healthChecker.interval is deprecated. Instead, use 
> mapreduce.tasktracker.healthchecker.interval
> 11:14:31,007 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.child.tmp is deprecated. Instead, use mapreduce.task.tmp.dir
> 11:14:31,008 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.tasktracker.taskmemorymanager.monitoring-interval is deprecated. 
> Instead, use mapreduce.tasktracker.taskmemorymanager.monitoringinterval
> 11:14:31,008 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.shuffle.connect.timeout is deprecated. Instead, use 
> mapreduce.reduce.shuffle.connect.timeout
> 11:14:31,008 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.speculative.execution.speculativeCap is deprecated. Instead, use 
> mapreduce.job.speculative.speculativecap
> 11:14:31,009 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - io.sort.spill.percent is deprecated. Instead, use 
> mapreduce.map.sort.spill.percent
> 11:14:31,010 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.shuffle.input.buffer.percent is deprecated. Instead, use 
> mapreduce.reduce.shuffle.input.buffer.percent
> 11:14:31,010 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.skip.map.max.skip.records is deprecated. Instead, use 
> mapreduce.map.skip.maxrecords
> 11:14:31,011 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.task.profile.maps is deprecated. Instead, use 
> mapreduce.task.profile.maps
> 11:14:31,012 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.local.dir is deprecated. Instead, use mapreduce.cluster.local.dir
> 11:14:31,013 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.merge.recordsBeforeProgress is deprecated. Instead, use 
> mapreduce.task.merge.progress.records
> 11:14:31,013 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.tracker.http.address is deprecated. Instead, use 
> mapreduce.jobtracker.http.address
> 11:14:31,013 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.userlog.retain.hours is deprecated. Instead, use 
> mapreduce.job.userlog.retain.hours
> 11:14:31,014 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.compress.map.output is deprecated. Instead, use 
> mapreduce.map.output.compress
> 11:14:31,014 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.speculative.execution.slowNodeThreshold is deprecated. Instead, 
> use mapreduce.job.speculative.slownodethreshold
> 11:14:31,015 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.tasktracker.reduce.tasks.maximum is deprecated. Instead, use 
> mapreduce.tasktracker.reduce.tasks.maximum
> 11:14:31,017 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.reduce.child.log.level is deprecated. Instead, use 
> mapreduce.reduce.log.level
> 11:14:31,018 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.jobtracker.restart.recover is deprecated. Instead, use 
> mapreduce.jobtracker.restart.recover
> 11:14:31,020 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.inmem.merge.threshold is deprecated. Instead, use 
> mapreduce.reduce.merge.inmem.threshold
> 11:14:31,021 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.reduce.tasks is deprecated. Instead, use mapreduce.job.reduces
> 11:14:31,023 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.acls.enabled is deprecated. Instead, use 
> mapreduce.cluster.acls.enabled
> 11:14:31,023 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.tasktracker.dns.nameserver is deprecated. Instead, use 
> mapreduce.tasktracker.dns.nameserver
> 11:14:31,024 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.submit.replication is deprecated. Instead, use 
> mapreduce.client.submit.file.replication
> 11:14:31,024 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.map.tasks.speculative.execution is deprecated. Instead, use 
> mapreduce.map.speculative
> 11:14:31,025 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.map.child.log.level is deprecated. Instead, use 
> mapreduce.map.log.level
> 11:14:31,026 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.map.max.attempts is deprecated. Instead, use 
> mapreduce.map.maxattempts
> 11:14:31,026 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.shuffle.merge.percent is deprecated. Instead, use 
> mapreduce.reduce.shuffle.merge.percent
> 11:14:31,026 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.tracker.jobhistory.lru.cache.size is deprecated. Instead, use 
> mapreduce.jobtracker.jobhistory.lru.cache.size
> 11:14:31,029 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.queue.name is deprecated. Instead, use mapreduce.job.queuename
> 11:14:31,030 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.tracker.persist.jobstatus.hours is deprecated. Instead, use 
> mapreduce.jobtracker.persist.jobstatus.hours
> 11:14:31,031 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.healthChecker.script.timeout is deprecated. Instead, use 
> mapreduce.tasktracker.healthchecker.script.timeout
> 11:14:31,031 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.tasktracker.map.tasks.maximum is deprecated. Instead, use 
> mapreduce.tasktracker.map.tasks.maximum
> 11:14:31,032 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.reduce.markreset.buffer.percent is deprecated. Instead, use 
> mapreduce.reduce.markreset.buffer.percent
> 11:14:31,032 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - jobclient.completion.poll.interval is deprecated. Instead, use 
> mapreduce.client.completion.pollinterval
> 11:14:31,032 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.tracker.persist.jobstatus.dir is deprecated. Instead, use 
> mapreduce.jobtracker.persist.jobstatus.dir
> 11:14:31,033 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.reduce.slowstart.completed.maps is deprecated. Instead, use 
> mapreduce.job.reduce.slowstart.completedmaps
> 11:14:31,034 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - io.sort.mb is deprecated. Instead, use mapreduce.task.io.sort.mb
> 11:14:31,034 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - dfs.umaskmode is deprecated. Instead, use fs.permissions.umask-mode
> 11:14:31,037 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.jobtracker.instrumentation is deprecated. Instead, use 
> mapreduce.jobtracker.instrumentation
> 11:14:31,038 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - topology.node.switch.mapping.impl is deprecated. Instead, use 
> net.topology.node.switch.mapping.impl
> 11:14:31,039 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.output.compression.type is deprecated. Instead, use 
> mapreduce.output.fileoutputformat.compress.type
> 11:14:31,039 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.skip.attempts.to.start.skipping is deprecated. Instead, use 
> mapreduce.task.skip.start.attempts
> 11:14:31,040 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.task.tracker.task-controller is deprecated. Instead, use 
> mapreduce.tasktracker.taskcontroller
> 11:14:31,041 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.userlog.limit.kb is deprecated. Instead, use 
> mapreduce.task.userlog.limit.kb
> 11:14:31,043 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.committer.job.setup.cleanup.needed is deprecated. Instead, use 
> mapreduce.job.committer.setup.cleanup.needed
> 11:14:31,043 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.map.tasks is deprecated. Instead, use mapreduce.job.maps
> 11:14:31,044 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.local.dir.minspacekill is deprecated. Instead, use 
> mapreduce.tasktracker.local.dir.minspacekill
> 11:14:31,044 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapreduce.jobtracker.split.metainfo.maxsize is deprecated. Instead, use 
> mapreduce.job.split.metainfo.maxsize
> 11:14:31,045 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - jobclient.progress.monitor.poll.interval is deprecated. Instead, use 
> mapreduce.client.progressmonitor.pollinterval
> 11:14:31,046 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.min.split.size is deprecated. Instead, use 
> mapreduce.input.fileinputformat.split.minsize
> 11:14:31,047 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.output.compression.codec is deprecated. Instead, use 
> mapreduce.output.fileoutputformat.compress.codec
> 11:14:31,048 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.task.profile is deprecated. Instead, use mapreduce.task.profile
> 11:14:31,048 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.reduce.parallel.copies is deprecated. Instead, use 
> mapreduce.reduce.shuffle.parallelcopies
> 11:14:31,049 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - io.sort.factor is deprecated. Instead, use mapreduce.task.io.sort.factor
> 11:14:31,050 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.input.dir is deprecated. Instead, use 
> mapreduce.input.fileinputformat.inputdir
> 11:14:31,050 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.task.timeout is deprecated. Instead, use mapreduce.task.timeout
> 11:14:31,052 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.heartbeats.in.second is deprecated. Instead, use 
> mapreduce.jobtracker.heartbeats.in.second
> 11:14:31,053 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.task.cache.levels is deprecated. Instead, use 
> mapreduce.jobtracker.taskcache.levels
> 11:14:31,053 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.map.output.compression.codec is deprecated. Instead, use 
> mapreduce.map.output.compress.codec
> 11:14:31,054 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.tasktracker.dns.interface is deprecated. Instead, use 
> mapreduce.tasktracker.dns.interface
> 11:14:31,055 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.max.tracker.failures is deprecated. Instead, use 
> mapreduce.job.maxtaskfailures.per.tracker
> 11:14:31,056 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - dfs.df.interval is deprecated. Instead, use fs.df.interval
> 11:14:31,056 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.tasktracker.tasks.sleeptime-before-sigkill is deprecated. 
> Instead, use mapreduce.tasktracker.tasks.sleeptimebeforesigkill
> 11:14:31,057 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.max.tracker.blacklists is deprecated. Instead, use 
> mapreduce.jobtracker.tasktracker.maxblacklists
> 11:14:31,058 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - jobclient.output.filter is deprecated. Instead, use 
> mapreduce.client.output.filter
> 11:14:31,059 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - job.end.retry.attempts is deprecated. Instead, use 
> mapreduce.job.end-notification.retry.attempts
> 11:14:31,059 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.jobtracker.taskScheduler is deprecated. Instead, use 
> mapreduce.jobtracker.taskscheduler
> 11:14:31,060 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.speculative.execution.slowTaskThreshold is deprecated. Instead, 
> use mapreduce.job.speculative.slowtaskthreshold
> 11:14:31,061 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.tasktracker.indexcache.mb is deprecated. Instead, use 
> mapreduce.tasktracker.indexcache.mb
> 11:14:31,061 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - tasktracker.http.threads is deprecated. Instead, use 
> mapreduce.tasktracker.http.threads
> 11:14:31,062 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.tracker.handler.count is deprecated. Instead, use 
> mapreduce.jobtracker.handler.count
> 11:14:31,062 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - keep.failed.task.files is deprecated. Instead, use 
> mapreduce.task.files.preserve.failedtasks
> 11:14:31,062 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.output.compress is deprecated. Instead, use 
> mapreduce.output.fileoutputformat.compress
> 11:14:31,063 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.jobtracker.job.history.block.size is deprecated. Instead, use 
> mapreduce.jobtracker.jobhistory.block.size
> 11:14:31,063 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.skip.reduce.max.skip.groups is deprecated. Instead, use 
> mapreduce.reduce.skip.maxgroups
> 11:14:31,066 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - topology.script.number.args is deprecated. Instead, use 
> net.topology.script.number.args
> 11:14:31,066 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - fs.default.name is deprecated. Instead, use fs.defaultFS
> 11:14:31,067 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.local.dir.minspacestart is deprecated. Instead, use 
> mapreduce.tasktracker.local.dir.minspacestart
> 11:14:31,067 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.jobtracker.maxtasks.per.job is deprecated. Instead, use 
> mapreduce.jobtracker.maxtasks.perjob
> 11:14:31,068 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.reduce.max.attempts is deprecated. Instead, use 
> mapreduce.reduce.maxattempts
> 11:14:31,068 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.tracker is deprecated. Instead, use 
> mapreduce.jobtracker.address
> 11:14:31,069 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.shuffle.read.timeout is deprecated. Instead, use 
> mapreduce.reduce.shuffle.read.timeout
> 11:14:31,069 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.tasktracker.instrumentation is deprecated. Instead, use 
> mapreduce.tasktracker.instrumentation
> 11:14:31,070 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.tasktracker.expiry.interval is deprecated. Instead, use 
> mapreduce.jobtracker.expire.trackers.interval
> 11:14:31,071 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.tracker.persist.jobstatus.active is deprecated. Instead, use 
> mapreduce.jobtracker.persist.jobstatus.active
> 11:14:31,074 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.temp.dir is deprecated. Instead, use mapreduce.cluster.temp.dir
> 11:14:31,075 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - hadoop.native.lib is deprecated. Instead, use io.native.lib.available
> 11:14:31,075 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapreduce.job.counters.limit is deprecated. Instead, use 
> mapreduce.job.counters.max
> 11:14:31,076 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapred.job.reduce.input.buffer.percent is deprecated. Instead, use 
> mapreduce.reduce.input.buffer.percent
> 11:14:31,149 INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat      
>    - Total input paths to process : 1
> 11:14:31,177 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Scheduling job Flink Java Job at Wed Jan 28 11:14:28 CET 2015.
> 11:14:31,188 INFO  org.apache.flink.runtime.executiongraph.ExecutionGraph     
>    - Deploying CHAIN DataSource (at 
> createInput(ExecutionEnvironment.java:426) 
> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> 
> FlatMap (FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-26
> 11:14:31,202 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 
> 11:14:28 CET 2015) changed to RUNNING.
> 11:14:32,525 INFO  org.apache.flink.api.common.io.LocatableInputSplitAssigner 
>    - Assigning remote split to host cloud-26
> 11:14:32,603 WARN  akka.remote.ReliableDeliverySupervisor                     
>    - Association with remote system 
> [akka.tcp://fl...@cloud-26.dima.tu-berlin.de:51449] has failed, address is 
> now gated for [5000] ms. Reason is: [Disassociated].
> 11:16:12,574 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 
> 11:14:28 CET 2015) changed to FAILINGjava.lang.RuntimeException: Requesting 
> the next InputSplit failed.
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>       at 
> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>       at java.lang.Thread.run(Thread.java:745)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>       at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>       at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>       at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>       at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>       at scala.concurrent.Await$.result(package.scala:107)
>       at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>       at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>       ... 4 more
> .
> 11:16:12,584 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 
> 11:14:28 CET 2015) changed to FAILEDjava.lang.RuntimeException: Requesting 
> the next InputSplit failed.
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>       at 
> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>       at java.lang.Thread.run(Thread.java:745)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>       at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>       at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>       at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>       at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>       at scala.concurrent.Await$.result(package.scala:107)
>       at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>       at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>       ... 4 more
> .
> 11:17:53,055 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Received job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 
> 11:17:50 CET 2015).
> 11:17:53,258 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum
> 11:17:53,312 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - fs.default.name is deprecated. Instead, use fs.defaultFS
> 11:17:53,320 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapreduce.job.counters.limit is deprecated. Instead, use 
> mapreduce.job.counters.max
> 11:17:53,347 INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat      
>    - Total input paths to process : 1
> 11:17:53,352 INFO  org.apache.flink.runtime.executiongraph.ExecutionGraph     
>    - Deploying CHAIN DataSource (at 
> createInput(ExecutionEnvironment.java:426) 
> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> 
> FlatMap (FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-31
> 11:17:53,353 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Scheduling job Flink Java Job at Wed Jan 28 11:17:50 CET 2015.
> 11:17:53,354 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 
> 11:17:50 CET 2015) changed to RUNNING.
> 11:17:54,646 INFO  org.apache.flink.api.common.io.LocatableInputSplitAssigner 
>    - Assigning remote split to host cloud-31
> 11:17:54,715 WARN  akka.remote.ReliableDeliverySupervisor                     
>    - Association with remote system 
> [akka.tcp://fl...@cloud-31.dima.tu-berlin.de:43200] has failed, address is 
> now gated for [5000] ms. Reason is: [Disassociated].
> 11:19:34,684 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 
> 11:17:50 CET 2015) changed to FAILINGjava.lang.RuntimeException: Requesting 
> the next InputSplit failed.
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>       at 
> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>       at java.lang.Thread.run(Thread.java:745)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>       at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>       at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>       at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>       at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>       at scala.concurrent.Await$.result(package.scala:107)
>       at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>       at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>       ... 4 more
> .
> 11:19:34,697 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 
> 11:17:50 CET 2015) changed to FAILEDjava.lang.RuntimeException: Requesting 
> the next InputSplit failed.
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>       at 
> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>       at java.lang.Thread.run(Thread.java:745)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>       at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>       at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>       at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>       at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>       at scala.concurrent.Await$.result(package.scala:107)
>       at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>       at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>       ... 4 more
> .
> 11:42:52,441 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Received job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 
> 11:42:50 CET 2015).
> 11:42:52,650 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum
> 11:42:52,705 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - fs.default.name is deprecated. Instead, use fs.defaultFS
> 11:42:52,713 INFO  org.apache.hadoop.conf.Configuration.deprecation           
>    - mapreduce.job.counters.limit is deprecated. Instead, use 
> mapreduce.job.counters.max
> 11:42:52,740 INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat      
>    - Total input paths to process : 1
> 11:42:52,747 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Scheduling job Flink Java Job at Wed Jan 28 11:42:50 CET 2015.
> 11:42:52,747 INFO  org.apache.flink.runtime.executiongraph.ExecutionGraph     
>    - Deploying CHAIN DataSource (at 
> createInput(ExecutionEnvironment.java:426) 
> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> 
> FlatMap (FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-26
> 11:42:52,750 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 
> 11:42:50 CET 2015) changed to RUNNING.
> 11:42:52,956 INFO  org.apache.flink.api.common.io.LocatableInputSplitAssigner 
>    - Assigning remote split to host cloud-26
> 11:42:52,983 WARN  akka.remote.ReliableDeliverySupervisor                     
>    - Association with remote system 
> [akka.tcp://fl...@cloud-26.dima.tu-berlin.de:51449] has failed, address is 
> now gated for [5000] ms. Reason is: [Disassociated].
> 11:44:32,976 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 
> 11:42:50 CET 2015) changed to FAILINGjava.lang.RuntimeException: Requesting 
> the next InputSplit failed.
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>       at 
> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>       at java.lang.Thread.run(Thread.java:745)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>       at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>       at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>       at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>       at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>       at scala.concurrent.Await$.result(package.scala:107)
>       at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>       at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>       ... 4 more
> .
> 11:44:32,984 INFO  
> org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
> Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 
> 11:42:50 CET 2015) changed to FAILEDjava.lang.RuntimeException: Requesting 
> the next InputSplit failed.
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>       at 
> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>       at 
> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>       at java.lang.Thread.run(Thread.java:745)
> Caused by: java.util.concurrent.TimeoutException: Futures timed out after 
> [100 seconds]
>       at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>       at 
> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>       at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>       at 
> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>       at scala.concurrent.Await$.result(package.scala:107)
>       at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>       at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>       at 
> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>       ... 4 more
> .
> 
> On 28 Jan 2015, at 12:08, Robert Metzger <rmetz...@apache.org> wrote:
> 
>> Hi,
>> 
>> it seems that you are not subscribed to our mailing list, so I had to
>> manually accept your mail. Would be good if you could subscribe.
>> 
>> Can you send us also the log output of the JobManager?
>> If your YARN cluster has log aggregation activated, you can retrieve the
>> logs of a stopped YARN session using:
>> yarn logs -applicationId <AppId>
>> 
>> watch out for the jobmanager-main.log or so file.
>> 
>> I suspect that there has been an exception on the JobManager.
>> 
>> Best,
>> Robert
>> 
>> 
>> 
>> On Wed, Jan 28, 2015 at 12:01 PM, Bruecke, Christoph <
>> christoph.brue...@campus.tu-berlin.de> wrote:
>> 
>>> Hi,
>>> 
>>> I have written a job that reads a SequenceFile from HDFS using the
>>> Hadoop-Compatibility add-on. Doing so results in a TimeoutException. I’m
>>> using flink-0.9-SNAPSHOT with PR 342 (
>>> https://github.com/apache/flink/pull/342 ). Furthermore I’m running flink
>>> on yarn with two TM using flink-yarn-0.9-SNAPSHOT/bin/yarn-session.sh -n 2.
>>> 
>>> Is this a bug or is there something wrong with the configuration?
>>> 
>>> 01/28/2015 11:42:52     Job execution switched to status RUNNING.
>>> 01/28/2015 11:42:52     CHAIN DataSource (at
>>> createInput(ExecutionEnvironment.java:426)
>>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to SCHEDULED
>>> 01/28/2015 11:42:52     CHAIN DataSource (at
>>> createInput(ExecutionEnvironment.java:426)
>>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to DEPLOYING
>>> 01/28/2015 11:42:52     CHAIN DataSource (at
>>> createInput(ExecutionEnvironment.java:426)
>>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to RUNNING
>>> 01/28/2015 11:44:32     CHAIN DataSource (at
>>> createInput(ExecutionEnvironment.java:426)
>>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to FAILED
>>> java.lang.RuntimeException: Requesting the next InputSplit failed.
>>>       at
>>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>>>       at
>>> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>>>       at
>>> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>>>       at
>>> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>>>       at java.lang.Thread.run(Thread.java:745)
>>> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
>>> [100 seconds]
>>>       at
>>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>>       at
>>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>>       at
>>> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>>       at
>>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>>       at scala.concurrent.Await$.result(package.scala:107)
>>>       at
>>> org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>>>       at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>>>       at
>>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>>>       ... 4 more
>>> 
>>> 01/28/2015 11:44:32     Job execution switched to status FAILING.
>>> 01/28/2015 11:44:32     GroupReduce (GroupReduce at
>>> main(ThiaziParser.java:40))(1/1) switched to CANCELED
>>> 01/28/2015 11:44:32     DataSink(TextOutputFormat (hdfs://
>>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/authors)
>>> - UTF-8)(1/1) switched to CANCELED
>>> 01/28/2015 11:44:32     CHAIN GroupReduce (GroupReduce at
>>> main(ThiaziParser.java:74)) -> Filter (Filter at
>>> main(ThiaziParser.java:97))(1/1) switched to CANCELED
>>> 01/28/2015 11:44:32     DataSink(TextOutputFormat (hdfs://
>>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/posts) -
>>> UTF-8)(1/1) switched to CANCELED
>>> 01/28/2015 11:44:32     CHAIN FlatMap (FlatMap at
>>> main(ThiaziParser.java:126)) -> Combine(SUM(1), at
>>> main(ThiaziParser.java:140)(1/1) switched to CANCELED
>>> 01/28/2015 11:44:32     Reduce (SUM(1), at
>>> main(ThiaziParser.java:140)(1/1) switched to CANCELED
>>> 01/28/2015 11:44:32     DataSink(CsvOutputFormat (path: hdfs://
>>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/wordcount,
>>> delimiter: ,))(1/1) switched to CANCELED
>>> 01/28/2015 11:44:32     GroupReduce (GroupReduce at
>>> main(ThiaziParser.java:106))(1/1) switched to CANCELED
>>> 01/28/2015 11:44:32     DataSink(TextOutputFormat (hdfs://
>>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/threads)
>>> - UTF-8)(1/1) switched to CANCELED
>>> 01/28/2015 11:44:32     Job execution switched to status FAILED.
>>> Error: The program execution failed: java.lang.RuntimeException:
>>> Requesting the next InputSplit failed.
>>>       at
>>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>>>       at
>>> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>>>       at
>>> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>>>       at
>>> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>>>       at java.lang.Thread.run(Thread.java:745)
>>> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
>>> [100 seconds]
>>>       at
>>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>>       at
>>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>>       at
>>> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>>       at
>>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>>       at scala.concurrent.Await$.result(package.scala:107)
>>>       at
>>> org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>>>       at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>>>       at
>>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>>>       ... 4 more
>>> 
>>> 
>>> 
> 

Reply via email to