Ah ok. And by distributed you mean that the element that starts the window can be processed on a different machine than the element that finishes the window?
On Wed, Jun 3, 2015 at 12:11 PM, Gyula Fóra <gyula.f...@gmail.com> wrote: > This is not connected to the current implementation. So lets not talk about > that. > > This is about theoretical limits to support distributed delta policies > which has far reaching implications for the windowing policies one can > implement in a prallel way. > > But you are welcome to throw in any constructive ideas :) > On Wed, Jun 3, 2015 at 11:49 AM Aljoscha Krettek <aljos...@apache.org> > wrote: > >> Part of the reason for my question is this: >> https://issues.apache.org/jira/browse/FLINK-1967. Especially my latest >> comment there. If we want this, I think we have to overhaul the >> windowing system anyways and then it doesn't make sense to explore >> complicated workarounds for the current system. >> >> On Wed, Jun 3, 2015 at 11:07 AM, Gyula Fóra <gyula.f...@gmail.com> wrote: >> > There are simple ways of implementing it in a non-distributed or >> > inconsistent fashion. >> > On Wed, Jun 3, 2015 at 8:55 AM Aljoscha Krettek <aljos...@apache.org> >> wrote: >> > >> >> This already sounds awfully complicated. Is there no other way to >> >> implement the delta windows? >> >> >> >> On Wed, Jun 3, 2015 at 7:52 AM, Gyula Fóra <gyula.f...@gmail.com> >> wrote: >> >> > Hi Ufuk, >> >> > >> >> > In the concrete use case I have in mind I only want to send events to >> >> > another subtask of the same task vertex. >> >> > >> >> > Specifically: if we want to do distributed delta based windows we >> need to >> >> > send after every trigger the element that has triggered the current >> >> window. >> >> > So practically I want to broadcast some event regularly to all >> subtasks >> >> of >> >> > the same operator. >> >> > >> >> > In this case the operators would wait until they receive this event >> so we >> >> > need to make sure that this event sending is not blocked by the actual >> >> > records. >> >> > >> >> > Gyula >> >> > >> >> > On Tuesday, June 2, 2015, Ufuk Celebi <u...@apache.org> wrote: >> >> > >> >> >> >> >> >> On 02 Jun 2015, at 22:45, Gyula Fóra <gyf...@apache.org >> <javascript:;>> >> >> >> wrote: >> >> >> > I am wondering, what is the suggested way to send some events >> >> directly to >> >> >> > another parallel instance in a flink job? For example from one >> mapper >> >> to >> >> >> > another mapper (of the same operator). >> >> >> > >> >> >> > Do we have any internal support for this? The first thing that we >> >> thought >> >> >> > of is iterations but that is clearly an overkill. >> >> >> >> >> >> There is no support for this at the moment. Any parallel instance? >> Or a >> >> >> subtask instance of the same task? >> >> >> >> >> >> Can you provide more input on the use case? It is certainly possible >> to >> >> >> add support for this. >> >> >> >> >> >> If the events don't need to be inline with the records, we can easily >> >> >> setup the TaskEventDispatcher as a separate actor (or extend the task >> >> >> manager) to process both backwards flowing events and in general any >> >> events >> >> >> that don't need to be inline with the records. The task deployment >> >> >> descriptors need to be extended with the extra parallel instance >> >> >> information. >> >> >> >> >> >> – Ufuk >> >> >>