Re: Memory segment error

2015-03-30 Thread Fabian Hueske
Robert or Stephan know the Travis setup quite well. They might know, if we can give a bit more than 80MB. But at some point there will be a hard limit. Once we have dynamic memory management (most of) such problems should be solved. 2015-03-30 23:46 GMT+02:00 Andra Lungu : > Oh! In that case, who

Re: Memory segment error

2015-03-30 Thread Andra Lungu
Oh! In that case, who should I refer to? :D [It's kind of ugly to split this kind of test. I mean if a person is counting the degrees, then that's the result that should be tested - at least in my opinion] In any case, thanks for the help :) On Mon, Mar 30, 2015 at 11:37 PM, Fabian Hueske wrote:

Re: Memory segment error

2015-03-30 Thread Fabian Hueske
Well, each combiner, reducer, join, coGroup, and solutionset needs a share of memory (maps & filters don't). In your case it was pretty much at the edge, the hash joins require 33 buffers and got 32. So one memory-consuming operator less might fix it. I did not look in detail at the other job, but

Re: Memory segment error

2015-03-30 Thread Andra Lungu
Hi Fabian, I'll see what I can do :). I am just a bit shocked. If this set of coGroups and joins was too much for a test case, how come the following worked? https://github.com/andralungu/flink/commit/f60b022de056ac259459b68eee6ff0ae9993f0f8 400 lines of complex computations :) And I have an eve

Re: Memory segment error

2015-03-30 Thread Fabian Hueske
Hi Andra, I found the cause for the exception. Your test case is simply too complex for our testing environment. We restrict the TM memory for testcases to 80MB in order to execute multiple tests in parallel on Travis. I counted the memory consumers in your job and got: - 2 Combine - 4 GroupReduc

Re: Memory segment error

2015-03-30 Thread Andra Lungu
Sure, It was in the first mail but that was sent a while ago :) This is the code: https://github.com/andralungu/gelly-partitioning/tree/alphaSplit I also added the log4j file in case it helps! The error is totally reproducible. 2 out of 2 people got the same. Steps to reproduce: 1). Clone the co

Re: Memory segment error

2015-03-30 Thread Fabian Hueske
Hmm, that is really weird. Can you point me to a branch in your repository and the test case that gives the error? Then I have a look at it and try to figure out what's going wrong. Cheers, Fabian 2015-03-30 10:43 GMT+02:00 Andra Lungu : > Hello, > > I went on and did some further debugging on

Re: Memory segment error

2015-03-30 Thread Andra Lungu
Hello, I went on and did some further debugging on this issue. Even though the exception said that the problem comes from here: 4837 [Join(Join at* weighEdges(NodeSplitting.java:117)*) (1/4)] ERROR org.apache.flink.runtime.operators.RegularPactTask - Error in task code: Join(Join at weighEdges(No

Re: Memory segment error

2015-03-26 Thread Andra Lungu
Opps! Sorry! Did not know the mailing list does not support attachments :) https://gist.github.com/andralungu/fba36d77f79189daa183 On Fri, Mar 27, 2015 at 12:02 AM, Andra Lungu wrote: > Hi Fabian, > > I uploaded a file with my execution plan. > > On Thu, Mar 26, 2015 at 11:50 PM, Fabian Hueske

Re: Memory segment error

2015-03-26 Thread Andra Lungu
Hi Fabian, I uploaded a file with my execution plan. On Thu, Mar 26, 2015 at 11:50 PM, Fabian Hueske wrote: > Hi Andra, > > the error is independent of the size of the data set. A HashTable needs at > least 33 memory pages to operate. > Since you have 820MB of managed memory and the size of a m

Re: Memory segment error

2015-03-26 Thread Fabian Hueske
Hi Andra, the error is independent of the size of the data set. A HashTable needs at least 33 memory pages to operate. Since you have 820MB of managed memory and the size of a memory page is 32KB, there should be more than 25k pages available. Can you post the execution plan of the program you ex

Re: Memory segment error

2015-03-26 Thread Andra Lungu
For 20 edges and 5 nodes, that should be more thank enough. On Thu, Mar 26, 2015 at 11:24 PM, Andra Lungu wrote: > Sure, > > 3470 [main] INFO org.apache.flink.runtime.taskmanager.TaskManager - > Using 820 MB for Flink managed memory. > > On Thu, Mar 26, 2015 at 4:48 PM, Robert Metzger > wrote

Re: Memory segment error

2015-03-26 Thread Andra Lungu
Sure, 3470 [main] INFO org.apache.flink.runtime.taskmanager.TaskManager - Using 820 MB for Flink managed memory. On Thu, Mar 26, 2015 at 4:48 PM, Robert Metzger wrote: > Hi, > > during startup, Flink will log something like: > 16:48:09,669 INFO org.apache.flink.runtime.taskmanager.TaskManage

Re: Memory segment error

2015-03-26 Thread Robert Metzger
Hi, during startup, Flink will log something like: 16:48:09,669 INFO org.apache.flink.runtime.taskmanager.TaskManager - Using 1193 MB for Flink managed memory. Can you tell us how much memory Flink is managing in your case? On Thu, Mar 26, 2015 at 4:46 PM, Andra Lungu wrote: > Hello ev

Re: Memory segment error when migrating functional code from Flink 0.9 to 0.8

2015-02-09 Thread Andra Lungu
Hi Stephan, I tried the solution with DeltaIteration#setSolutionSetUnManaged(), unfortunatelly the error is still there... Even when I try to run it with just one iteration... Also, I am not sure that the job can be broken into subparts in this particular case. Any other suggestions would be appr

Re: Memory segment error when migrating functional code from Flink 0.9 to 0.8

2015-02-09 Thread Stephan Ewen
This is actually a problem of the number of memory segments available to the hash table for the solution set. For complex pipelines, memory currently gets too fragmented. There are two workarounds, until we do the dynamic memory management, or break it into shorter pipelines: Break the job up int

Re: Memory segment error when migrating functional code from Flink 0.9 to 0.8

2015-02-09 Thread Till Rohrmann
Hi Andra, have you tried increasing the number of network buffers in your cluster? You can control by the configuration value: taskmanager.network.numberOfBuffers: #numberBuffers Greets, Till On Mon, Feb 9, 2015 at 9:56 AM, Andra Lungu wrote: > Hello everyone, > > I am implementing a graph a