Opps! Sorry! Did not know the mailing list does not support attachments :) https://gist.github.com/andralungu/fba36d77f79189daa183
On Fri, Mar 27, 2015 at 12:02 AM, Andra Lungu <lungu.an...@gmail.com> wrote: > Hi Fabian, > > I uploaded a file with my execution plan. > > On Thu, Mar 26, 2015 at 11:50 PM, Fabian Hueske <fhue...@gmail.com> wrote: > >> Hi Andra, >> >> the error is independent of the size of the data set. A HashTable needs at >> least 33 memory pages to operate. >> Since you have 820MB of managed memory and the size of a memory page is >> 32KB, there should be more than 25k pages available. >> >> Can you post the execution plan of the program you execute ( >> ExecutionEnvironment.getExecutionPlan() )? >> >> Best, Fabian >> >> 2015-03-26 23:31 GMT+01:00 Andra Lungu <lungu.an...@gmail.com>: >> >> > For 20 edges and 5 nodes, that should be more thank enough. >> > >> > On Thu, Mar 26, 2015 at 11:24 PM, Andra Lungu <lungu.an...@gmail.com> >> > wrote: >> > >> > > Sure, >> > > >> > > 3470 [main] INFO org.apache.flink.runtime.taskmanager.TaskManager - >> > > Using 820 MB for Flink managed memory. >> > > >> > > On Thu, Mar 26, 2015 at 4:48 PM, Robert Metzger <rmetz...@apache.org> >> > > wrote: >> > > >> > >> Hi, >> > >> >> > >> during startup, Flink will log something like: >> > >> 16:48:09,669 INFO org.apache.flink.runtime.taskmanager.TaskManager >> > >> - Using 1193 MB for Flink managed memory. >> > >> >> > >> Can you tell us how much memory Flink is managing in your case? >> > >> >> > >> >> > >> >> > >> On Thu, Mar 26, 2015 at 4:46 PM, Andra Lungu <lungu.an...@gmail.com> >> > >> wrote: >> > >> >> > >> > Hello everyone, >> > >> > >> > >> > I guess I need to revive this old discussion: >> > >> > >> > >> > >> > >> >> > >> http://apache-flink-incubator-mailing-list-archive.1008284.n3.nabble.com/Memory-segment-error-when-migrating-functional-code-from-Flink-0-9-to-0-8-td3687.html >> > >> > >> > >> > At that point, the fix was to kindly ask Alex to make his project >> work >> > >> with >> > >> > 0.9. >> > >> > >> > >> > Now, I am not that lucky! >> > >> > >> > >> > This is the code: >> > >> > https://github.com/andralungu/gelly-partitioning/tree/alphaSplit >> > >> > >> > >> > The main program(NodeSplitting) is working nicely, I get the >> correct >> > >> > result. But if you run the test, you will see that collection >> works >> > and >> > >> > cluster fails miserably with this exception: >> > >> > >> > >> > Caused by: java.lang.Exception: The data preparation for task >> > >> 'Join(Join at >> > >> > weighEdges(NodeSplitting.java:112)) >> > (04e172e761148a65783a4363406e08c0)' >> > >> , >> > >> > caused an error: Too few memory segments provided. Hash Join needs >> at >> > >> least >> > >> > 33 memory segments. >> > >> > at >> > >> > >> > >> > >> > >> >> > >> org.apache.flink.runtime.operators.RegularPactTask.run(RegularPactTask.java:471) >> > >> > at >> > >> > >> > >> > >> > >> >> > >> org.apache.flink.runtime.operators.RegularPactTask.invoke(RegularPactTask.java:362) >> > >> > at >> > >> > >> > >> > >> > >> >> > >> org.apache.flink.runtime.execution.RuntimeEnvironment.run(RuntimeEnvironment.java:209) >> > >> > at java.lang.Thread.run(Thread.java:745) >> > >> > Caused by: java.lang.IllegalArgumentException: Too few memory >> segments >> > >> > provided. Hash Join needs at least 33 memory segments. >> > >> > >> > >> > I am running locally, from IntelliJ, on a tiny graph. >> > >> > $ cat /proc/meminfo >> > >> > MemTotal: 11405696 kB >> > >> > MemFree: 5586012 kB >> > >> > Buffers: 178100 kB >> > >> > >> > >> > I am sure I did not run out of memory... >> > >> > >> > >> > Any thoughts on this? >> > >> > >> > >> > Thanks! >> > >> > Andra >> > >> > >> > >> >> > > >> > > >> > >> > >