Hi Ufuk, In the concrete use case I have in mind I only want to send events to another subtask of the same task vertex.
Specifically: if we want to do distributed delta based windows we need to send after every trigger the element that has triggered the current window. So practically I want to broadcast some event regularly to all subtasks of the same operator. In this case the operators would wait until they receive this event so we need to make sure that this event sending is not blocked by the actual records. Gyula On Tuesday, June 2, 2015, Ufuk Celebi <u...@apache.org> wrote: > > On 02 Jun 2015, at 22:45, Gyula Fóra <gyf...@apache.org <javascript:;>> > wrote: > > I am wondering, what is the suggested way to send some events directly to > > another parallel instance in a flink job? For example from one mapper to > > another mapper (of the same operator). > > > > Do we have any internal support for this? The first thing that we thought > > of is iterations but that is clearly an overkill. > > There is no support for this at the moment. Any parallel instance? Or a > subtask instance of the same task? > > Can you provide more input on the use case? It is certainly possible to > add support for this. > > If the events don't need to be inline with the records, we can easily > setup the TaskEventDispatcher as a separate actor (or extend the task > manager) to process both backwards flowing events and in general any events > that don't need to be inline with the records. The task deployment > descriptors need to be extended with the extra parallel instance > information. > > – Ufuk