Thanks. Yes, I got that. Cheers
On Wed, Jul 22, 2015 at 2:46 PM, Maximilian Michels <m...@apache.org> wrote: > I mentioned that. @Max: you should only try it out if you want to > experiment/work with the changes. > > On Wed, Jul 22, 2015 at 2:20 PM, Stephan Ewen <se...@apache.org> wrote: > >> The two pull requests do not go all the way, unfortunately. They cover >> only the runtime, the API integration part is missing still, >> unfortunately... >> >> On Mon, Jul 20, 2015 at 5:53 PM, Maximilian Michels <m...@apache.org> >> wrote: >> >>> You could do that but you might run into merge conflicts. Also keep in >>> mind that it is work in progress :) >>> >>> On Mon, Jul 20, 2015 at 4:15 PM, Maximilian Alber < >>> alber.maximil...@gmail.com> wrote: >>> >>>> Thanks! >>>> >>>> Ok, cool. If I would like to test it, I just need to merge those two >>>> pull requests into my current branch? >>>> >>>> Cheers, >>>> Max >>>> >>>> On Mon, Jul 20, 2015 at 4:02 PM, Maximilian Michels <m...@apache.org> >>>> wrote: >>>> >>>>> Now that makes more sense :) I thought by "nested iterations" you >>>>> meant iterations in Flink that can be nested, i.e. starting an iteration >>>>> inside an iteration. >>>>> >>>>> The caching/pinning of intermediate results is still a work in >>>>> progress in Flink. It is actually in a state where it could be merged but >>>>> some pending pull requests got delayed because priorities changed a bit. >>>>> >>>>> Essentially, we need to merge these two pull requests: >>>>> >>>>> https://github.com/apache/flink/pull/858 >>>>> This introduces a session management which allows to keep the >>>>> ExecutionGraph for the session. >>>>> >>>>> https://github.com/apache/flink/pull/640 >>>>> Implements the actual backtracking and caching of the results. >>>>> >>>>> Once these are in, we can change the Java/Scala API to support >>>>> backtracking. I don't exactly know how Spark's API does it but, >>>>> essentially >>>>> it should work then by just creating new operations on an existing DataSet >>>>> and submit to the cluster again. >>>>> >>>>> Cheers, >>>>> Max >>>>> >>>>> On Mon, Jul 20, 2015 at 3:31 PM, Maximilian Alber < >>>>> alber.maximil...@gmail.com> wrote: >>>>> >>>>>> Oh sorry, my fault. When I wrote it, I had iterations in mind. >>>>>> >>>>>> What I actually wanted to say, how "resuming from intermediate >>>>>> results" will work with (non-nested) "non-Flink" iterations? And with >>>>>> iterations I mean something like this: >>>>>> >>>>>> while(...): >>>>>> - change params >>>>>> - submit to cluster >>>>>> >>>>>> where the executed Flink-program is more or less the same at each >>>>>> iterations. But with changing input sets, which are reused between >>>>>> different loop iterations. >>>>>> >>>>>> I might got something wrong, because in our group we mentioned >>>>>> caching a lá Spark for Flink and someone came up that "pinning" will do >>>>>> that. Is that somewhat right? >>>>>> >>>>>> Thanks and Cheers, >>>>>> Max >>>>>> >>>>>> On Mon, Jul 20, 2015 at 1:06 PM, Maximilian Michels <m...@apache.org> >>>>>> wrote: >>>>>> >>>>>>> "So it is up to debate how the support for resuming from >>>>>>> intermediate results will look like." -> What's the current state of >>>>>>> that >>>>>>> debate? >>>>>>> >>>>>>> Since there is no support for nested iterations that I know of, the >>>>>>> debate how intermediate results are integrated has not started yet. >>>>>>> >>>>>>> >>>>>>>> "Intermediate results are not produced within the iterations >>>>>>>> cycles." -> Ok, if there are none, what does it have to do with that >>>>>>>> debate? :-) >>>>>>>> >>>>>>> >>>>>>> I was referring to the existing support for intermediate results >>>>>>> within iterations. If we were to implement nested iterations, this could >>>>>>> (possibly) change. This is all very theoretical because there are no >>>>>>> plans >>>>>>> to support nested iterations. >>>>>>> >>>>>>> Hope this clarifies. Otherwise, please restate your question because >>>>>>> I might have misunderstood. >>>>>>> >>>>>>> Cheers, >>>>>>> Max >>>>>>> >>>>>>> >>>>>>> On Mon, Jul 20, 2015 at 12:11 PM, Maximilian Alber < >>>>>>> alber.maximil...@gmail.com> wrote: >>>>>>> >>>>>>>> Thanks for the answer! But I need some clarification: >>>>>>>> >>>>>>>> "So it is up to debate how the support for resuming from >>>>>>>> intermediate results will look like." -> What's the current state of >>>>>>>> that >>>>>>>> debate? >>>>>>>> "Intermediate results are not produced within the iterations >>>>>>>> cycles." -> Ok, if there are none, what does it have to do with that >>>>>>>> debate? :-) >>>>>>>> >>>>>>>> Cheers, >>>>>>>> Max >>>>>>>> >>>>>>>> On Mon, Jul 20, 2015 at 10:50 AM, Maximilian Michels < >>>>>>>> m...@apache.org> wrote: >>>>>>>> >>>>>>>>> Hi Max, >>>>>>>>> >>>>>>>>> You are right, there is no support for nested iterations yet. As >>>>>>>>> far as I know, there are no concrete plans to add support for it. So >>>>>>>>> it is >>>>>>>>> up to debate how the support for resuming from intermediate results >>>>>>>>> will >>>>>>>>> look like. Intermediate results are not produced within the iterations >>>>>>>>> cycles. Same would be true for nested iterations. So the behavior for >>>>>>>>> resuming from intermediate results should be alike for nested >>>>>>>>> iterations. >>>>>>>>> >>>>>>>>> Cheers, >>>>>>>>> Max >>>>>>>>> >>>>>>>>> On Fri, Jul 17, 2015 at 4:26 PM, Maximilian Alber < >>>>>>>>> alber.maximil...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hi Flinksters, >>>>>>>>>> >>>>>>>>>> as far as I know, there is still no support for nested iterations >>>>>>>>>> planned. Am I right? >>>>>>>>>> >>>>>>>>>> So my question is how such use cases should be handled in the >>>>>>>>>> future. >>>>>>>>>> More specific: when pinning/caching will be available, you >>>>>>>>>> suggest to use that feature and program in "Spark" style? Or is >>>>>>>>>> there some >>>>>>>>>> other, more flexible, mechanism planned for loops? >>>>>>>>>> >>>>>>>>>> Cheers, >>>>>>>>>> Max >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >