Hi Ufuk, so the memory available per node is 48294 megabytes per node, but I reserve 28 by flink conf file. taskmanager.heap.mb = 28672 taskmanager.memory.fraction = 0.7 taskmanager.network.numberOfBuffers = 32768 taskmanager.network.bufferSizeInBytes = 16384
Anyway Follows what I found in log files. *Follows the taskmanager log (task manager that seems failed)* 2016-06-29 11:31:55,673 INFO org.apache.flink.runtime.taskmanager.Task - CHAIN Reduce (Reduce at dima.tu.berlin.benchmark.fli nk.mlr.solver.sGradientDescentL2.createInitialWeightsVector(sGradientDescentL2.scala:43)) -> Map (Map at dima.tu.berlin.benchmark.flink.mlr.solver .sGradientDescentL2.createInitialVector(sGradientDescentL2.scala:69)) (1/1) switched to FAILED with exception. java.lang.IllegalStateException: Received unexpected partition state null for partition request. This is a bug. at org.apache.flink.runtime.taskmanager.Task.onPartitionStateUpdate(Task.java:994) at org.apache.flink.runtime.taskmanager.TaskManager.org $apache$flink$runtime$taskmanager$TaskManager$$handleTaskMessage(TaskManager.scala: 468) at org.apache.flink.runtime.taskmanager.TaskManager$$anonfun$handleMessage$1.applyOrElse(TaskManager.scala:265) at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) at org.apache.flink.runtime.LeaderSessionMessageFilter$$anonfun$receive$1.applyOrElse(LeaderSessionMessageFilter.scala:36) at scala.runtime.AbstractPartialFunction$mcVL$sp.apply$mcVL$sp(AbstractPartialFunction.scala:33) at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:33) at scala.runtime.AbstractPartialFunction$mcVL$sp.apply(AbstractPartialFunction.scala:25) at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:33) at org.apache.flink.runtime.LogMessages$$anon$1.apply(LogMessages.scala:28) at scala.PartialFunction$class.applyOrElse(PartialFunction.scala:118) at org.apache.flink.runtime.LogMessages$$anon$1.applyOrElse(LogMessages.scala:28) at akka.actor.Actor$class.aroundReceive(Actor.scala:465) at org.apache.flink.runtime.taskmanager.TaskManager.aroundReceive(TaskManager.scala:119) at akka.actor.ActorCell.receiveMessage(ActorCell.scala:516) at akka.actor.ActorCell.invoke(ActorCell.scala:487) at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:254) at akka.dispatch.Mailbox.run(Mailbox.scala:221) at akka.dispatch.Mailbox.exec(Mailbox.scala:231) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.pollAndExecAll(ForkJoinPool.java:1253) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1346) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) *Follows the jobmanager log* 2016-06-29 11:31:34,683 INFO org.apache.flink.runtime.executiongraph.ExecutionGraph - CHAIN Reduce (Reduce at dima.tu.berlin.benchmark.fli nk.mlr.solver.sGradientDescentL2.createInitialWeightsVector(sGradientDescentL2.scala:43)) -> Map (Map at dima.tu.berlin.benchmark.flink.mlr.solver .sGradientDescentL2.createInitialVector(sGradientDescentL2.scala:69)) (1/1) (8c2d9a0a0520c2c18e07bad0a97a3911) switched from DEPLOYING to FAILED 2016-06-29 11:31:34,694 INFO org.apache.flink.runtime.jobmanager.JobManager - Status of job 71542654d427e8d0e7e01c538abe1acf (peel -bundle-flink) changed to FAILING. java.lang.Exception: Cannot deploy task CHAIN Reduce (Reduce at dima.tu.berlin.benchmark.flink.mlr.solver.sGradientDescentL2.createInitialWeightsV ector(sGradientDescentL2.scala:43)) -> Map (Map at dima.tu.berlin.benchmark.flink.mlr.solver.sGradientDescentL2.createInitialVector(sGradientDesce ntL2.scala:69)) (1/1) (8c2d9a0a0520c2c18e07bad0a97a3911) - TaskManager (c0b308245dfc5da6d759fb5bc1bc5ad0 @ cloud-12 - 16 slots - URL: akka.tcp://f link@130.149.21.16:6122/user/taskmanager) not responding after a timeout of 10000 milliseconds at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:387) at akka.dispatch.OnComplete.internal(Future.scala:246) at akka.dispatch.OnComplete.internal(Future.scala:244) at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174) at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171) at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) Caused by: akka.pattern.AskTimeoutException: Ask timed out on [Actor[akka.tcp://flink@130.149.21.16:6122/user/taskmanager#1824295872]] after [1000 0 ms] at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333) at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117) at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694) at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691) at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467) at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419) at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423) at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375) at java.lang.Thread.run(Thread.java:745) *Follows the client-{$runtime.hostname}.log* 2016-06-29 11:31:34,687 INFO org.apache.flink.runtime.client.JobClientActor - 06/29/2016 11:31:34 CHAIN Reduce (Reduce at di ma.tu.berlin.benchmark.flink.mlr.solver.sGradientDescentL2.createInitialWeightsVector(sGradientDescentL2.scala:43)) -> Map (Map at dima.tu.berlin. benchmark.flink.mlr.solver.sGradientDescentL2.createInitialVector(sGradientDescentL2.scala:69))(1/1) switched to FAILED java.lang.Exception: Cannot deploy task CHAIN Reduce (Reduce at dima.tu.berlin.benchmark.flink.mlr.solver.sGradientDescentL2.createInitialWeightsV ector(sGradientDescentL2.scala:43)) -> Map (Map at dima.tu.berlin.benchmark.flink.mlr.solver.sGradientDescentL2.createInitialVector(sGradientDesce ntL2.scala:69)) (1/1) (8c2d9a0a0520c2c18e07bad0a97a3911) - TaskManager (c0b308245dfc5da6d759fb5bc1bc5ad0 @ cloud-12 - 16 slots - URL: akka.tcp://f link@130.149.21.16:6122/user/taskmanager) not responding after a timeout of 10000 milliseconds at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:387) at akka.dispatch.OnComplete.internal(Future.scala:246) at akka.dispatch.OnComplete.internal(Future.scala:244) at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174) at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171) at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) Caused by: akka.pattern.AskTimeoutException: Ask timed out on [Actor[akka.tcp://flink@130.149.21.16:6122/user/taskmanager#1824295872]] after [1000 0 ms] at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333) at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117) at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694) at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691) at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467) at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419) at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423) at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375) at java.lang.Thread.run(Thread.java:745) 2016-06-29 11:31:34,709 INFO org.apache.flink.runtime.client.JobClientActor - 06/29/2016 11:31:34 Job execution switched to status FAILING. java.lang.Exception: Cannot deploy task CHAIN Reduce (Reduce at dima.tu.berlin.benchmark.flink.mlr.solver.sGradientDescentL2.createInitialWeightsV ector(sGradientDescentL2.scala:43)) -> Map (Map at dima.tu.berlin.benchmark.flink.mlr.solver.sGradientDescentL2.createInitialVector(sGradientDesce ntL2.scala:69)) (1/1) (8c2d9a0a0520c2c18e07bad0a97a3911) - TaskManager (c0b308245dfc5da6d759fb5bc1bc5ad0 @ cloud-12 - 16 slots - URL: akka.tcp://f link@130.149.21.16:6122/user/taskmanager) not responding after a timeout of 10000 milliseconds at org.apache.flink.runtime.executiongraph.Execution$2.onComplete(Execution.java:387) at akka.dispatch.OnComplete.internal(Future.scala:246) at akka.dispatch.OnComplete.internal(Future.scala:244) at akka.dispatch.japi$CallbackBridge.apply(Future.scala:174) at akka.dispatch.japi$CallbackBridge.apply(Future.scala:171) at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32) at scala.concurrent.impl.ExecutionContextImpl$$anon$3.exec(ExecutionContextImpl.scala:107) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107) Caused by: akka.pattern.AskTimeoutException: Ask timed out on [Actor[akka.tcp://flink@130.149.21.16:6122/user/taskmanager#1824295872]] after [10000 ms] at akka.pattern.PromiseActorRef$$anonfun$1.apply$mcV$sp(AskSupport.scala:333) at akka.actor.Scheduler$$anon$7.run(Scheduler.scala:117) at scala.concurrent.Future$InternalCallbackExecutor$.scala$concurrent$Future$InternalCallbackExecutor$$unbatchedExecute(Future.scala:694) at scala.concurrent.Future$InternalCallbackExecutor$.execute(Future.scala:691) at akka.actor.LightArrayRevolverScheduler$TaskHolder.executeTask(Scheduler.scala:467) at akka.actor.LightArrayRevolverScheduler$$anon$8.executeBucket$1(Scheduler.scala:419) at akka.actor.LightArrayRevolverScheduler$$anon$8.nextTick(Scheduler.scala:423) at akka.actor.LightArrayRevolverScheduler$$anon$8.run(Scheduler.scala:375) at java.lang.Thread.run(Thread.java:745) Really appreciating your help here. :) Cheers, Andrea 2016-06-29 13:48 GMT+02:00 Ufuk Celebi <u...@apache.org>: > Hey Andrea! Sorry for the bad user experience. > > Regarding the network buffers: you should be able to run it after > increasing the number of network buffers, just account for it when > specifying the heap size etc. You currently allocate 32768 * 16384 > bytes = 512 MB for them. If you have a very long pipeline and high > parallelism, you should increase it accordingly. How much memory do > you have on your machines? > > Regarding the IllegalStateException: I suspect that this is **not** > the root failure cause. The null ExecutionState can only happen, if > the producer task (from which data is requested) failed during the > request. The error message is confusing and I opened a JIRA to fix it: > https://issues.apache.org/jira/browse/FLINK-4131. Can you please check > your complete logs to see what the root cause might be, e.g. why did > the producer fail? > > > On Wed, Jun 29, 2016 at 12:19 PM, ANDREA SPINA > <74...@studenti.unimore.it> wrote: > > Hi everyone, > > > > I am running some Flink experiments with Peel benchmark > > http://peel-framework.org/ and I am struggling with exceptions: the > > environment is a 25-nodes cluster, 16 cores per nodes. The dataset is > ~80GiB > > and is located on Hdfs 2.7.1. Flink version is 1.0.3. > > > > At the beginning I tried with 400 as degree of parallelism but not enough > > numberOfBuffers was raised so I changed the parallelism to 200. Flink > > configuration follows: > > > > jobmanager.rpc.address = ${runtime.hostname} > > akka.log.lifecycle.events = ON > > akka.ask.timeout = 300s > > jobmanager.rpc.port = 6002 > > jobmanager.heap.mb = 1024 > > jobmanager.web.port = 6004 > > taskmanager.heap.mb = 28672 > > taskmanager.memory.fraction = 0.7 > > taskmanager.network.numberOfBuffers = 32768 > > taskmanager.network.bufferSizeInBytes = 16384 > > taskmanager.tmp.dirs = > > > "/data/1/peel/flink/tmp:/data/2/peel/flink/tmp:/data/3/peel/flink/tmp:/data/4/peel/flink/tmp" > > taskmanager.debug.memory.startLogThread = true > > > > With a parallelism of 200 the following exception will raise from a node > of > > the cluster: > > > > 2016-06-29 11:31:55,673 INFO org.apache.flink.runtime.taskmanager.Task > > - CHAIN Reduce (Reduce at dima.tu.berlin.benchmark.fli > > > nk.mlr.solver.sGradientDescentL2.createInitialWeightsVector(sGradientDescentL2.scala:43)) > > -> Map (Map at dima.tu.berlin.benchmark.flink.mlr.solver > > .sGradientDescentL2.createInitialVector(sGradientDescentL2.scala:69)) > (1/1) > > switched to FAILED with exception. > > java.lang.IllegalStateException: Received unexpected partition state null > > for partition request. This is a bug. > > at > > > org.apache.flink.runtime.taskmanager.Task.onPartitionStateUpdate(Task.java:994) > > > > > > The reduce code is: > > > > 43 val dimensionsDS = data.map(_.vector.size).reduce((a, b) => b) > > > > The map code is: > > > > 68 def createInitialVector(dimensionDS: DataSet[Int]): DataSet[Vector] > = { > > 69 dimensionDS.map { > > 70 dimension => > > 71 val values = DenseVector(Array.fill(dimension)(0.0)) > > 72 values > > 73 } > > 74 } > > > > I can't figure out a solution, thank you for your help. > > > > Andrea > > > > -- > > Andrea Spina > > N.Tessera: 74598 > > MAT: 89369 > > Ingegneria Informatica [LM] (D.M. 270) > -- *Andrea Spina* N.Tessera: *74598* MAT: *89369* *Ingegneria Informatica* *[LM] *(D.M. 270)