AW: Writing Intermediates to disk

2016-06-19 Thread Paschek, Robert
user@flink.apache.org Betreff: Re: Writing Intermediates to disk I do not know if I understand completely, but I would create a new DataSet based on filtering the condition and then persist this DataSet. So : DataSet ds2 = DataSet1.filter(Condition) 2ds.output(...) On Mon, May 9, 2016 at 11:09 AM,

Re: Writing Intermediates to disk

2016-05-09 Thread Vikram Saxena
I do not know if I understand completely, but I would create a new DataSet based on filtering the condition and then persist this DataSet. So : DataSet ds2 = DataSet1.filter(Condition) 2ds.output(...) On Mon, May 9, 2016 at 11:09 AM, Ufuk Celebi wrote: > Flink has support for spillable in

Re: Writing Intermediates to disk

2016-05-09 Thread Ufuk Celebi
Flink has support for spillable intermediate results. Currently they are only set if necessary to avoid pipeline deadlocks. You can force this via env.getConfig().setExecutionMode(ExecutionMode.BATCH); This will write shuffles to disk, but you don't get the fine-grained control you probably need

Writing Intermediates to disk

2016-05-05 Thread Paschek, Robert
Hi Mailing List, I want to write and read intermediates to/from disk. The following foo- codesnippet may illustrate my intention: public void mapPartition(Iterable tuples, Collector out) { for (T tuple : tuples) { if (Condition)