Hi,
I have written a job that reads a SequenceFile from HDFS using the
Hadoop-Compatibility add-on. Doing so results in a TimeoutException. I’m using
flink-0.9-SNAPSHOT with PR 342 ( https://github.com/apache/flink/pull/342 ).
Furthermore I’m running flink on yarn with two TM using
flink-yarn-0.9-SNAPSHOT/bin/yarn-session.sh -n 2.
Is this a bug or is there something wrong with the configuration?
01/28/2015 11:42:52 Job execution switched to status RUNNING.
01/28/2015 11:42:52 CHAIN DataSource (at
createInput(ExecutionEnvironment.java:426)
(org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap
(FlatMap at main(ThiaziParser.java:37))(1/1) switched to SCHEDULED
01/28/2015 11:42:52 CHAIN DataSource (at
createInput(ExecutionEnvironment.java:426)
(org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap
(FlatMap at main(ThiaziParser.java:37))(1/1) switched to DEPLOYING
01/28/2015 11:42:52 CHAIN DataSource (at
createInput(ExecutionEnvironment.java:426)
(org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap
(FlatMap at main(ThiaziParser.java:37))(1/1) switched to RUNNING
01/28/2015 11:44:32 CHAIN DataSource (at
createInput(ExecutionEnvironment.java:426)
(org.apache.flink.hadoopcompatibility.mapreduce.HadoopInputFormat)) -> FlatMap
(FlatMap at main(ThiaziParser.java:37))(1/1) switched to FAILED
java.lang.RuntimeException: Requesting the next InputSplit failed.
at
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
at
org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
at
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
at
org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100
seconds]
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
at
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
at scala.concurrent.Await$.result(package.scala:107)
at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
at
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
... 4 more
01/28/2015 11:44:32 Job execution switched to status FAILING.
01/28/2015 11:44:32 GroupReduce (GroupReduce at
main(ThiaziParser.java:40))(1/1) switched to CANCELED
01/28/2015 11:44:32 DataSink(TextOutputFormat
(hdfs://cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/authors)
- UTF-8)(1/1) switched to CANCELED
01/28/2015 11:44:32 CHAIN GroupReduce (GroupReduce at
main(ThiaziParser.java:74)) -> Filter (Filter at
main(ThiaziParser.java:97))(1/1) switched to CANCELED
01/28/2015 11:44:32 DataSink(TextOutputFormat
(hdfs://cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/posts)
- UTF-8)(1/1) switched to CANCELED
01/28/2015 11:44:32 CHAIN FlatMap (FlatMap at main(ThiaziParser.java:126))
-> Combine(SUM(1), at main(ThiaziParser.java:140)(1/1) switched to CANCELED
01/28/2015 11:44:32 Reduce (SUM(1), at main(ThiaziParser.java:140)(1/1)
switched to CANCELED
01/28/2015 11:44:32 DataSink(CsvOutputFormat (path:
hdfs://cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/wordcount,
delimiter: ,))(1/1) switched to CANCELED
01/28/2015 11:44:32 GroupReduce (GroupReduce at
main(ThiaziParser.java:106))(1/1) switched to CANCELED
01/28/2015 11:44:32 DataSink(TextOutputFormat
(hdfs://cloud-11.dima.tu-berlin.de:60010/user/cbruecke/output/thiazi-seq/threads)
- UTF-8)(1/1) switched to CANCELED
01/28/2015 11:44:32 Job execution switched to status FAILED.
Error: The program execution failed: java.lang.RuntimeException: Requesting the
next InputSplit failed.
at
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:63)
at
org.apache.flink.runtime.operators.DataSourceTask$1.hasNext(DataSourceTask.java:355)
at
org.apache.flink.runtime.operators.DataSourceTask.invoke(DataSourceTask.java:154)
at
org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:204)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.util.concurrent.TimeoutException: Futures timed out after [100
seconds]
at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219)
at
scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223)
at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:107)
at
scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53)
at scala.concurrent.Await$.result(package.scala:107)
at org.apache.flink.runtime.akka.AkkaUtils$.ask(AkkaUtils.scala:265)
at org.apache.flink.runtime.akka.AkkaUtils.ask(AkkaUtils.scala)
at
org.apache.flink.runtime.taskmanager.TaskInputSplitProvider.getNextInputSplit(TaskInputSplitProvider.java:56)
... 4 more