Hi Robert,

thanks for the quick response. Here is the jobmanager-main.log:

PS: I’m subscribed now.

11:09:16,144 INFO  org.apache.flink.yarn.ApplicationMaster$                     
 - YARN daemon runs as hadoop setting user to execute Flink 
ApplicationMaster/JobManager to hadoop
11:09:16,199 INFO  org.apache.flink.yarn.Utils                                  
 - Found YARN_CONF_DIR, adding it to configuration
11:09:16,298 INFO  org.apache.flink.yarn.ApplicationMaster$                     
 - Start job manager for yarn
11:09:16,298 INFO  org.apache.flink.yarn.ApplicationMaster$                     
 - Config path: 
/data/3/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/container_1420727594991_0068_01_000001.
11:09:17,179 INFO  akka.event.slf4j.Slf4jLogger                                 
 - Slf4jLogger started
11:09:17,271 INFO  Remoting                                                     
 - Starting remoting
11:09:17,556 INFO  Remoting                                                     
 - Remoting started; listening on addresses 
:[akka.tcp://fl...@cloud-29.dima.tu-berlin.de:42643]
11:09:17,569 INFO  org.apache.flink.yarn.ApplicationMaster$                     
 - Start job manager actor.
11:09:17,570 INFO  org.apache.flink.yarn.ApplicationMaster$                     
 - Generate configuration file for application master.
11:09:17,579 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Starting job manager at akka://flink/user/jobmanager.
11:09:17,590 INFO  org.apache.flink.yarn.ApplicationMaster$                     
 - Start yarn session on job manager.
11:09:17,591 INFO  org.apache.flink.yarn.ApplicationMaster$                     
 - Application Master properly initiated. Await termination of actor system.
11:09:17,601 INFO  org.apache.flink.runtime.blob.BlobServer                     
 - Started BLOB server on port 35605
11:09:17,616 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Started job manager. Waiting for incoming messages.
11:09:17,621 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer        
 - Setting up web info server, using web-root 
directoryjar:file:/data/4/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/filecache/11/flink-dist-0.9-SNAPSHOT-yarn-uberjar.jar!/web-docs-infoserver.
11:09:17,621 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer        
 - Web info server will display information about flink job-manager on 
localhost, port 8081.
11:09:17,733 INFO  org.apache.flink.runtime.jobmanager.web.WebInfoServer        
 - Starting web info server for JobManager on port 8081
11:09:17,734 INFO  org.eclipse.jetty.util.log                                   
 - jetty-0.9-SNAPSHOT
11:09:17,770 INFO  org.eclipse.jetty.util.log                                   
 - Started SelectChannelConnector@0.0.0.0:8081
11:09:17,784 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Start yarn session.
11:09:17,784 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Requesting 2 task managers.
11:09:18,414 INFO  org.apache.hadoop.yarn.client.RMProxy                        
 - Connecting to ResourceManager at cloud-11/130.149.21.15:8030
11:09:18,423 INFO  
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy  - 
yarn.client.max-nodemanagers-proxies : 500
11:09:18,425 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Registering ApplicationMaster with tracking url 
http://cloud-29.dima.tu-berlin.de:8081.
11:09:18,787 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Requesting TaskManager container 0.
11:09:18,794 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Requesting TaskManager container 1.
11:09:18,903 INFO  org.apache.flink.yarn.Utils                                  
 - Copying from 
file:/data/3/hadoop-2.4/yarn-local-dir/usercache/hadoop/appcache/application_1420727594991_0068/container_1420727594991_0068_01_000001/flink-conf-modified.yaml
 to 
hdfs://cloud-11.dima.tu-berlin.de:60010/user/hadoop/.flink/application_1420727594991_0068/flink-conf-modified.yaml
11:09:19,280 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Prepared local resource for modified yaml: resource { scheme: "hdfs" host: 
"cloud-11.dima.tu-berlin.de" port: 60010 file: 
"/user/hadoop/.flink/application_1420727594991_0068/flink-conf-modified.yaml" } 
size: 3121 timestamp: 1422439759181 type: FILE visibility: APPLICATION
11:09:19,286 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Create container launch context.
11:09:19,300 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Starting TM with command=$JAVA_HOME/bin/java -Xmx819m  
-Dlog.file="<LOG_DIR>/taskmanager.log" 
-Dlogback.configurationFile=file:logback.xml 
-Dlog4j.configuration=file:log4j.properties 
org.apache.flink.yarn.appMaster.YarnTaskManagerRunner --configDir . 1> 
<LOG_DIR>/taskmanager-stdout.log 2> <LOG_DIR>/taskmanager-stderr.log
11:09:19,619 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Got new container for TM container_1420727594991_0068_01_000002 on host 
cloud-26.dima.tu-berlin.de
11:09:19,621 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Launching container #1.
11:09:19,622 INFO  
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy  - 
Opening proxy : cloud-26.dima.tu-berlin.de:8045
11:09:19,828 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Got new container for TM container_1420727594991_0068_01_000003 on host 
cloud-31.dima.tu-berlin.de
11:09:19,829 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Launching container #2.
11:09:19,831 INFO  
org.apache.hadoop.yarn.client.api.impl.ContainerManagementProtocolProxy  - 
Opening proxy : cloud-31.dima.tu-berlin.de:8045
11:09:25,748 INFO  org.apache.flink.runtime.instance.InstanceManager            
 - Registered TaskManager at 
akka.tcp://fl...@cloud-26.dima.tu-berlin.de:51449/user/taskmanager as 
ca731dbbfe941a63077c9c148254607c. Current number of registered hosts is 1.
11:09:25,845 INFO  org.apache.flink.runtime.instance.InstanceManager            
 - Registered TaskManager at 
akka.tcp://fl...@cloud-31.dima.tu-berlin.de:43200/user/taskmanager as 
7bf3aa638e00682239c9d8460d142191. Current number of registered hosts is 2.
11:11:13,837 WARN  akka.remote.ReliableDeliverySupervisor                       
 - Association with remote system 
[akka.tcp://CliFrontendActorSystem@130.149.21.15:50292] has failed, address is 
now gated for [5000] ms. Reason is: [Disassociated].
11:14:30,892 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Received job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 
11:14:28 CET 2015).
11:14:30,997 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - job.end.retry.interval is deprecated. Instead, use 
mapreduce.job.end-notification.retry.interval
11:14:30,998 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum
11:14:30,999 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.tracker.retiredjobs.cache.size is deprecated. Instead, use 
mapreduce.jobtracker.retiredjobs.cache.size
11:14:31,000 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.task.profile.reduces is deprecated. Instead, use 
mapreduce.task.profile.reduces
11:14:31,000 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.reuse.jvm.num.tasks is deprecated. Instead, use 
mapreduce.job.jvm.numtasks
11:14:31,001 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.reduce.tasks.speculative.execution is deprecated. Instead, use 
mapreduce.reduce.speculative
11:14:31,003 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.task.tracker.http.address is deprecated. Instead, use 
mapreduce.tasktracker.http.address
11:14:31,004 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.system.dir is deprecated. Instead, use mapreduce.jobtracker.system.dir
11:14:31,005 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.task.tracker.report.address is deprecated. Instead, use 
mapreduce.tasktracker.report.address
11:14:31,006 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.healthChecker.interval is deprecated. Instead, use 
mapreduce.tasktracker.healthchecker.interval
11:14:31,007 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.child.tmp is deprecated. Instead, use mapreduce.task.tmp.dir
11:14:31,008 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.tasktracker.taskmemorymanager.monitoring-interval is deprecated. 
Instead, use mapreduce.tasktracker.taskmemorymanager.monitoringinterval
11:14:31,008 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.shuffle.connect.timeout is deprecated. Instead, use 
mapreduce.reduce.shuffle.connect.timeout
11:14:31,008 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.speculative.execution.speculativeCap is deprecated. Instead, use 
mapreduce.job.speculative.speculativecap
11:14:31,009 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - io.sort.spill.percent is deprecated. Instead, use 
mapreduce.map.sort.spill.percent
11:14:31,010 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.shuffle.input.buffer.percent is deprecated. Instead, use 
mapreduce.reduce.shuffle.input.buffer.percent
11:14:31,010 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.skip.map.max.skip.records is deprecated. Instead, use 
mapreduce.map.skip.maxrecords
11:14:31,011 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.task.profile.maps is deprecated. Instead, use 
mapreduce.task.profile.maps
11:14:31,012 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.local.dir is deprecated. Instead, use mapreduce.cluster.local.dir
11:14:31,013 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.merge.recordsBeforeProgress is deprecated. Instead, use 
mapreduce.task.merge.progress.records
11:14:31,013 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.tracker.http.address is deprecated. Instead, use 
mapreduce.jobtracker.http.address
11:14:31,013 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.userlog.retain.hours is deprecated. Instead, use 
mapreduce.job.userlog.retain.hours
11:14:31,014 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.compress.map.output is deprecated. Instead, use 
mapreduce.map.output.compress
11:14:31,014 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.speculative.execution.slowNodeThreshold is deprecated. Instead, use 
mapreduce.job.speculative.slownodethreshold
11:14:31,015 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.tasktracker.reduce.tasks.maximum is deprecated. Instead, use 
mapreduce.tasktracker.reduce.tasks.maximum
11:14:31,017 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.reduce.child.log.level is deprecated. Instead, use 
mapreduce.reduce.log.level
11:14:31,018 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.jobtracker.restart.recover is deprecated. Instead, use 
mapreduce.jobtracker.restart.recover
11:14:31,020 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.inmem.merge.threshold is deprecated. Instead, use 
mapreduce.reduce.merge.inmem.threshold
11:14:31,021 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.reduce.tasks is deprecated. Instead, use mapreduce.job.reduces
11:14:31,023 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.acls.enabled is deprecated. Instead, use 
mapreduce.cluster.acls.enabled
11:14:31,023 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.tasktracker.dns.nameserver is deprecated. Instead, use 
mapreduce.tasktracker.dns.nameserver
11:14:31,024 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.submit.replication is deprecated. Instead, use 
mapreduce.client.submit.file.replication
11:14:31,024 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.map.tasks.speculative.execution is deprecated. Instead, use 
mapreduce.map.speculative
11:14:31,025 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.map.child.log.level is deprecated. Instead, use 
mapreduce.map.log.level
11:14:31,026 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.map.max.attempts is deprecated. Instead, use mapreduce.map.maxattempts
11:14:31,026 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.shuffle.merge.percent is deprecated. Instead, use 
mapreduce.reduce.shuffle.merge.percent
11:14:31,026 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.tracker.jobhistory.lru.cache.size is deprecated. Instead, use 
mapreduce.jobtracker.jobhistory.lru.cache.size
11:14:31,029 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.queue.name is deprecated. Instead, use mapreduce.job.queuename
11:14:31,030 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.tracker.persist.jobstatus.hours is deprecated. Instead, use 
mapreduce.jobtracker.persist.jobstatus.hours
11:14:31,031 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.healthChecker.script.timeout is deprecated. Instead, use 
mapreduce.tasktracker.healthchecker.script.timeout
11:14:31,031 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.tasktracker.map.tasks.maximum is deprecated. Instead, use 
mapreduce.tasktracker.map.tasks.maximum
11:14:31,032 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.reduce.markreset.buffer.percent is deprecated. Instead, use 
mapreduce.reduce.markreset.buffer.percent
11:14:31,032 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - jobclient.completion.poll.interval is deprecated. Instead, use 
mapreduce.client.completion.pollinterval
11:14:31,032 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.tracker.persist.jobstatus.dir is deprecated. Instead, use 
mapreduce.jobtracker.persist.jobstatus.dir
11:14:31,033 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.reduce.slowstart.completed.maps is deprecated. Instead, use 
mapreduce.job.reduce.slowstart.completedmaps
11:14:31,034 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - io.sort.mb is deprecated. Instead, use mapreduce.task.io.sort.mb
11:14:31,034 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - dfs.umaskmode is deprecated. Instead, use fs.permissions.umask-mode
11:14:31,037 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.jobtracker.instrumentation is deprecated. Instead, use 
mapreduce.jobtracker.instrumentation
11:14:31,038 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - topology.node.switch.mapping.impl is deprecated. Instead, use 
net.topology.node.switch.mapping.impl
11:14:31,039 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.output.compression.type is deprecated. Instead, use 
mapreduce.output.fileoutputformat.compress.type
11:14:31,039 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.skip.attempts.to.start.skipping is deprecated. Instead, use 
mapreduce.task.skip.start.attempts
11:14:31,040 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.task.tracker.task-controller is deprecated. Instead, use 
mapreduce.tasktracker.taskcontroller
11:14:31,041 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.userlog.limit.kb is deprecated. Instead, use 
mapreduce.task.userlog.limit.kb
11:14:31,043 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.committer.job.setup.cleanup.needed is deprecated. Instead, use 
mapreduce.job.committer.setup.cleanup.needed
11:14:31,043 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.map.tasks is deprecated. Instead, use mapreduce.job.maps
11:14:31,044 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.local.dir.minspacekill is deprecated. Instead, use 
mapreduce.tasktracker.local.dir.minspacekill
11:14:31,044 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapreduce.jobtracker.split.metainfo.maxsize is deprecated. Instead, use 
mapreduce.job.split.metainfo.maxsize
11:14:31,045 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - jobclient.progress.monitor.poll.interval is deprecated. Instead, use 
mapreduce.client.progressmonitor.pollinterval
11:14:31,046 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.min.split.size is deprecated. Instead, use 
mapreduce.input.fileinputformat.split.minsize
11:14:31,047 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.output.compression.codec is deprecated. Instead, use 
mapreduce.output.fileoutputformat.compress.codec
11:14:31,048 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.task.profile is deprecated. Instead, use mapreduce.task.profile
11:14:31,048 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.reduce.parallel.copies is deprecated. Instead, use 
mapreduce.reduce.shuffle.parallelcopies
11:14:31,049 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - io.sort.factor is deprecated. Instead, use mapreduce.task.io.sort.factor
11:14:31,050 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.input.dir is deprecated. Instead, use 
mapreduce.input.fileinputformat.inputdir
11:14:31,050 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.task.timeout is deprecated. Instead, use mapreduce.task.timeout
11:14:31,052 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.heartbeats.in.second is deprecated. Instead, use 
mapreduce.jobtracker.heartbeats.in.second
11:14:31,053 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.task.cache.levels is deprecated. Instead, use 
mapreduce.jobtracker.taskcache.levels
11:14:31,053 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.map.output.compression.codec is deprecated. Instead, use 
mapreduce.map.output.compress.codec
11:14:31,054 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.tasktracker.dns.interface is deprecated. Instead, use 
mapreduce.tasktracker.dns.interface
11:14:31,055 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.max.tracker.failures is deprecated. Instead, use 
mapreduce.job.maxtaskfailures.per.tracker
11:14:31,056 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - dfs.df.interval is deprecated. Instead, use fs.df.interval
11:14:31,056 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.tasktracker.tasks.sleeptime-before-sigkill is deprecated. Instead, 
use mapreduce.tasktracker.tasks.sleeptimebeforesigkill
11:14:31,057 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.max.tracker.blacklists is deprecated. Instead, use 
mapreduce.jobtracker.tasktracker.maxblacklists
11:14:31,058 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - jobclient.output.filter is deprecated. Instead, use 
mapreduce.client.output.filter
11:14:31,059 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - job.end.retry.attempts is deprecated. Instead, use 
mapreduce.job.end-notification.retry.attempts
11:14:31,059 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.jobtracker.taskScheduler is deprecated. Instead, use 
mapreduce.jobtracker.taskscheduler
11:14:31,060 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.speculative.execution.slowTaskThreshold is deprecated. Instead, use 
mapreduce.job.speculative.slowtaskthreshold
11:14:31,061 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.tasktracker.indexcache.mb is deprecated. Instead, use 
mapreduce.tasktracker.indexcache.mb
11:14:31,061 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - tasktracker.http.threads is deprecated. Instead, use 
mapreduce.tasktracker.http.threads
11:14:31,062 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.tracker.handler.count is deprecated. Instead, use 
mapreduce.jobtracker.handler.count
11:14:31,062 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - keep.failed.task.files is deprecated. Instead, use 
mapreduce.task.files.preserve.failedtasks
11:14:31,062 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.output.compress is deprecated. Instead, use 
mapreduce.output.fileoutputformat.compress
11:14:31,063 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.jobtracker.job.history.block.size is deprecated. Instead, use 
mapreduce.jobtracker.jobhistory.block.size
11:14:31,063 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.skip.reduce.max.skip.groups is deprecated. Instead, use 
mapreduce.reduce.skip.maxgroups
11:14:31,066 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - topology.script.number.args is deprecated. Instead, use 
net.topology.script.number.args
11:14:31,066 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - fs.default.name is deprecated. Instead, use fs.defaultFS
11:14:31,067 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.local.dir.minspacestart is deprecated. Instead, use 
mapreduce.tasktracker.local.dir.minspacestart
11:14:31,067 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.jobtracker.maxtasks.per.job is deprecated. Instead, use 
mapreduce.jobtracker.maxtasks.perjob
11:14:31,068 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.reduce.max.attempts is deprecated. Instead, use 
mapreduce.reduce.maxattempts
11:14:31,068 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.tracker is deprecated. Instead, use mapreduce.jobtracker.address
11:14:31,069 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.shuffle.read.timeout is deprecated. Instead, use 
mapreduce.reduce.shuffle.read.timeout
11:14:31,069 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.tasktracker.instrumentation is deprecated. Instead, use 
mapreduce.tasktracker.instrumentation
11:14:31,070 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.tasktracker.expiry.interval is deprecated. Instead, use 
mapreduce.jobtracker.expire.trackers.interval
11:14:31,071 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.tracker.persist.jobstatus.active is deprecated. Instead, use 
mapreduce.jobtracker.persist.jobstatus.active
11:14:31,074 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.temp.dir is deprecated. Instead, use mapreduce.cluster.temp.dir
11:14:31,075 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - hadoop.native.lib is deprecated. Instead, use io.native.lib.available
11:14:31,075 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapreduce.job.counters.limit is deprecated. Instead, use 
mapreduce.job.counters.max
11:14:31,076 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapred.job.reduce.input.buffer.percent is deprecated. Instead, use 
mapreduce.reduce.input.buffer.percent
11:14:31,149 INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat        
 - Total input paths to process : 1
11:14:31,177 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Scheduling job Flink Java Job at Wed Jan 28 11:14:28 CET 2015.
11:14:31,188 INFO  org.apache.flink.runtime.executiongraph.ExecutionGraph       
 - Deploying CHAIN DataSource (at createInput(ExecutionEnvironment.java:426) 
(org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap 
(FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-26
11:14:31,202 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 
11:14:28 CET 2015) changed to RUNNING.
11:14:32,525 INFO  org.apache.flink.api.common.io.LocatableInputSplitAssigner   
 - Assigning remote split to host cloud-26
11:14:32,603 WARN  akka.remote.ReliableDeliverySupervisor                       
 - Association with remote system 
[akka.tcp://fl...@cloud-26.dima.tu-berlin.de:51449] has failed, address is now 
gated for [5000] ms. Reason is: [Disassociated].
11:16:12,574 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 
11:14:28 CET 2015) changed to FAILINGjava.lang.RuntimeException: Requesting the 
next InputSplit failed.
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
        at 
org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
        at 
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
        at 
org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 
seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
        at 
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
        at scala.concurrent.Await$.result(package.scala:107)
        at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
        ... 4 more
.
11:16:12,584 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Status of job c979ea0b2e358fb988f5a2a30ab48786 (Flink Java Job at Wed Jan 28 
11:14:28 CET 2015) changed to FAILEDjava.lang.RuntimeException: Requesting the 
next InputSplit failed.
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
        at 
org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
        at 
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
        at 
org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 
seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
        at 
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
        at scala.concurrent.Await$.result(package.scala:107)
        at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
        ... 4 more
.
11:17:53,055 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Received job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 
11:17:50 CET 2015).
11:17:53,258 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum
11:17:53,312 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - fs.default.name is deprecated. Instead, use fs.defaultFS
11:17:53,320 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapreduce.job.counters.limit is deprecated. Instead, use 
mapreduce.job.counters.max
11:17:53,347 INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat        
 - Total input paths to process : 1
11:17:53,352 INFO  org.apache.flink.runtime.executiongraph.ExecutionGraph       
 - Deploying CHAIN DataSource (at createInput(ExecutionEnvironment.java:426) 
(org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap 
(FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-31
11:17:53,353 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Scheduling job Flink Java Job at Wed Jan 28 11:17:50 CET 2015.
11:17:53,354 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 
11:17:50 CET 2015) changed to RUNNING.
11:17:54,646 INFO  org.apache.flink.api.common.io.LocatableInputSplitAssigner   
 - Assigning remote split to host cloud-31
11:17:54,715 WARN  akka.remote.ReliableDeliverySupervisor                       
 - Association with remote system 
[akka.tcp://fl...@cloud-31.dima.tu-berlin.de:43200] has failed, address is now 
gated for [5000] ms. Reason is: [Disassociated].
11:19:34,684 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 
11:17:50 CET 2015) changed to FAILINGjava.lang.RuntimeException: Requesting the 
next InputSplit failed.
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
        at 
org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
        at 
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
        at 
org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 
seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
        at 
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
        at scala.concurrent.Await$.result(package.scala:107)
        at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
        ... 4 more
.
11:19:34,697 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Status of job fe8cebf0874c7f67ac03454de8555fb9 (Flink Java Job at Wed Jan 28 
11:17:50 CET 2015) changed to FAILEDjava.lang.RuntimeException: Requesting the 
next InputSplit failed.
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
        at 
org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
        at 
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
        at 
org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 
seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
        at 
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
        at scala.concurrent.Await$.result(package.scala:107)
        at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
        ... 4 more
.
11:42:52,441 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Received job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 
11:42:50 CET 2015).
11:42:52,650 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - io.bytes.per.checksum is deprecated. Instead, use dfs.bytes-per-checksum
11:42:52,705 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - fs.default.name is deprecated. Instead, use fs.defaultFS
11:42:52,713 INFO  org.apache.hadoop.conf.Configuration.deprecation             
 - mapreduce.job.counters.limit is deprecated. Instead, use 
mapreduce.job.counters.max
11:42:52,740 INFO  org.apache.hadoop.mapreduce.lib.input.FileInputFormat        
 - Total input paths to process : 1
11:42:52,747 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Scheduling job Flink Java Job at Wed Jan 28 11:42:50 CET 2015.
11:42:52,747 INFO  org.apache.flink.runtime.executiongraph.ExecutionGraph       
 - Deploying CHAIN DataSource (at createInput(ExecutionEnvironment.java:426) 
(org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap 
(FlatMap at main(ThiaziParser.java:37)) (1/1) (attempt #0) to cloud-26
11:42:52,750 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 
11:42:50 CET 2015) changed to RUNNING.
11:42:52,956 INFO  org.apache.flink.api.common.io.LocatableInputSplitAssigner   
 - Assigning remote split to host cloud-26
11:42:52,983 WARN  akka.remote.ReliableDeliverySupervisor                       
 - Association with remote system 
[akka.tcp://fl...@cloud-26.dima.tu-berlin.de:51449] has failed, address is now 
gated for [5000] ms. Reason is: [Disassociated].
11:44:32,976 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 
11:42:50 CET 2015) changed to FAILINGjava.lang.RuntimeException: Requesting the 
next InputSplit failed.
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
        at 
org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
        at 
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
        at 
org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 
seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
        at 
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
        at scala.concurrent.Await$.result(package.scala:107)
        at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
        ... 4 more
.
11:44:32,984 INFO  
org.apache.flink.yarn.ApplicationMaster$$anonfun$startJobManager$2$$anon$1  - 
Status of job b63a600c2edd93b790f2e984ca987464 (Flink Java Job at Wed Jan 28 
11:42:50 CET 2015) changed to FAILEDjava.lang.RuntimeException: Requesting the 
next InputSplit failed.
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
        at 
org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
        at 
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
        at 
org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100 
seconds]
        at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
        at 
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
        at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
        at 
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
        at scala.concurrent.Await$.result(package.scala:107)
        at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
        at 
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
        ... 4 more
.

On 28 Jan 2015, at 12:08, Robert Metzger <rmetz...@apache.org> wrote:

> Hi,
> 
> it seems that you are not subscribed to our mailing list, so I had to
> manually accept your mail. Would be good if you could subscribe.
> 
> Can you send us also the log output of the JobManager?
> If your YARN cluster has log aggregation activated, you can retrieve the
> logs of a stopped YARN session using:
> yarn logs -applicationId <AppId>
> 
> watch out for the jobmanager-main.log or so file.
> 
> I suspect that there has been an exception on the JobManager.
> 
> Best,
> Robert
> 
> 
> 
> On Wed, Jan 28, 2015 at 12:01 PM, Bruecke, Christoph <
> christoph.brue...@campus.tu-berlin.de> wrote:
> 
>> Hi,
>> 
>> I have written a job that reads a SequenceFile from HDFS using the
>> Hadoop-Compatibility add-on. Doing so results in a TimeoutException. I’m
>> using flink-0.9-SNAPSHOT with PR 342 (
>> https://github.com/apache/flink/pull/342 ). Furthermore I’m running flink
>> on yarn with two TM using flink-yarn-0.9-SNAPSHOT/bin/yarn-session.sh -n 2.
>> 
>> Is this a bug or is there something wrong with the configuration?
>> 
>> 01/28/2015 11:42:52     Job execution switched to status RUNNING.
>> 01/28/2015 11:42:52     CHAIN DataSource (at
>> createInput(ExecutionEnvironment.java:426)
>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to SCHEDULED
>> 01/28/2015 11:42:52     CHAIN DataSource (at
>> createInput(ExecutionEnvironment.java:426)
>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to DEPLOYING
>> 01/28/2015 11:42:52     CHAIN DataSource (at
>> createInput(ExecutionEnvironment.java:426)
>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to RUNNING
>> 01/28/2015 11:44:32     CHAIN DataSource (at
>> createInput(ExecutionEnvironment.java:426)
>> (org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) ->
>> FlatMap (FlatMap at main(ThiaziParser.java:37))(1/1) switched to FAILED
>> java.lang.RuntimeException: Requesting the next InputSplit failed.
>>        at
>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>>        at
>> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>>        at
>> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>>        at
>> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>>        at java.lang.Thread.run(Thread.java:745)
>> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
>> [100 seconds]
>>        at
>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>        at
>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>        at
>> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>        at
>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>        at scala.concurrent.Await$.result(package.scala:107)
>>        at
>> org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>>        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>>        at
>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>>        ... 4 more
>> 
>> 01/28/2015 11:44:32     Job execution switched to status FAILING.
>> 01/28/2015 11:44:32     GroupReduce (GroupReduce at
>> main(ThiaziParser.java:40))(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     DataSink(TextOutputFormat (hdfs://
>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/authors)
>> - UTF-8)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     CHAIN GroupReduce (GroupReduce at
>> main(ThiaziParser.java:74)) -> Filter (Filter at
>> main(ThiaziParser.java:97))(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     DataSink(TextOutputFormat (hdfs://
>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/posts) -
>> UTF-8)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     CHAIN FlatMap (FlatMap at
>> main(ThiaziParser.java:126)) -> Combine(SUM(1), at
>> main(ThiaziParser.java:140)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     Reduce (SUM(1), at
>> main(ThiaziParser.java:140)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     DataSink(CsvOutputFormat (path: hdfs://
>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/wordcount,
>> delimiter: ,))(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     GroupReduce (GroupReduce at
>> main(ThiaziParser.java:106))(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     DataSink(TextOutputFormat (hdfs://
>> cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/threads)
>> - UTF-8)(1/1) switched to CANCELED
>> 01/28/2015 11:44:32     Job execution switched to status FAILED.
>> Error: The program execution failed: java.lang.RuntimeException:
>> Requesting the next InputSplit failed.
>>        at
>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
>>        at
>> org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
>>        at
>> org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
>>        at
>> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
>>        at java.lang.Thread.run(Thread.java:745)
>> Caused by: java.util.concurrent.TimeoutException: Futures timed out after
>> [100 seconds]
>>        at
>> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
>>        at
>> scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
>>        at
>> scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
>>        at
>> scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
>>        at scala.concurrent.Await$.result(package.scala:107)
>>        at
>> org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
>>        at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
>>        at
>> org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
>>        ... 4 more
>> 
>> 
>> 

Reply via email to