Hi Yuta,

you got most things right :-)

3) sources (such as Kafka connectors) are also considered operators and
start immediately because they are sources.
4) All other operators start when they need to process their first record.
Operators are never GC'd (unless a job was cancelled), so the setup cost is
a one time thing that only happens when the job is started.

Best, Fabian

2017-09-15 12:43 GMT+02:00 Yuta Morisawa <yu-moris...@kddi-research.jp>:

> Hi Fabian,
>
> Thank you for your description.
>
> This is my understanding.
> 1, At the exact time execute() method called, Flink creates JobGraph,
> submit it to JobManager, deploy tasks to TaskManagers and DOES NOT execute
> each operators.
> 2, Operators are executed when they needed.
> 3, Sources(kafka-connectors) starts before operators.
> 4, The first time operators are called or after GC removes operators'
> instance, a kind of initialization occurs, such as classloading,
> instantiation, memory allocation and so on. It may costs much time.
>
> If there is any misunderstanding, please comment it.
> If not, my question is solved.
>
> Regards.
> Yuta
>
> On 2017/09/15 17:05, Fabian Hueske wrote:
>
>> Hi Yuta,
>>
>> when the execute() method is called, the a so-called JobGraph is
>> constructed from all operators that have been added before by calling
>> map(), keyBy() and so on.
>> The JobGraph is then submitted to the JobManager which is the master
>> process in Flink. Based on the JobGraph, the master deploys tasks to the
>> worker processes (TaskManagers).
>> These are the tasks that do the actual processing and they are
>> subsequently started as I explained before, i.e., the source task starts
>> consuming from Kafka before subsequent tasks have been started.
>>
>> So, there is quite a lot happening when you call execute() including
>> network communication and task deployment.
>>
>> Hope this helps,
>> Fabian
>>
>> 2017-09-15 4:25 GMT+02:00 Yuta Morisawa <yu-moris...@kddi-research.jp
>> <mailto:yu-moris...@kddi-research.jp>>:
>>
>>
>>     Hi, Fabian
>>
>>     > If I understand you correctly, the problem is only for the first
>> events
>>     > that are processed.
>>     Yes. More Precisely, first 300 kafka-messages.
>>
>>     > AFAIK, Flink lazily instantiates its operators which means that a
>> source
>>     > task starts to consume records from Kafka before the subsequent
>> tasks
>>     > have been started.
>>     That's a great indication. It describe well the affair.
>>     But, according to the document, it says "The operations are actually
>>     executed when the execution is explicitly triggered by an execute()
>>     call on the execution environment.".
>>     What does it mean?
>>     AFAIK, common Flink programs invoke execute() in main().
>>     Every operators start at this time? I think maybe no.
>>
>>     - Flink Document
>>
>>     https://ci.apache.org/projects/flink/flink-docs-release-1.3/
>> dev/api_concepts.html#lazy-evaluation
>>     <https://ci.apache.org/projects/flink/flink-docs-release-1.3
>> /dev/api_concepts.html#lazy-evaluation>
>>
>>
>>     > Not sure if or what can be done about this behavior.
>>     > I'll loop in Till who knows more about the lifecycle of tasks.
>>     Thank you very much for your kindness.
>>
>>     Regards, Yuta
>>
>>     On 2017/09/14 19:32, Fabian Hueske wrote:
>>
>>         Hi,
>>
>>         If I understand you correctly, the problem is only for the first
>>         events that are processed.
>>
>>         AFAIK, Flink lazily instantiates its operators which means that
>>         a source task starts to consume records from Kafka before the
>>         subsequent tasks have been started.
>>         That's why the latency of the first records is higher.
>>
>>         Not sure if or what can be done about this behavior.
>>         I'll loop in Till who knows more about the lifecycle of tasks.
>>
>>         Best, Fabian
>>
>>
>>         2017-09-12 11:02 GMT+02:00 Yuta Morisawa
>>         <yu-moris...@kddi-research.jp
>>         <mailto:yu-moris...@kddi-research.jp>
>>         <mailto:yu-moris...@kddi-research.jp
>>         <mailto:yu-moris...@kddi-research.jp>>>:
>>
>>              Hi,
>>
>>              I am worrying about the delay of the Streaming API.
>>              My application is that it gets data from kafka-connectors and
>>              process them, then push data to kafka-producers.
>>              The problem is that the app suffers a long delay when the
>>         first data
>>              come in the cluster.
>>              It takes about 1000ms to process data (I measure the time
>> with
>>              kafka-timestamp). On the other hand, it works well after
>>         2-3 seconds
>>              first data come in (the delay is about 200ms).
>>
>>              The application is so delay sensitive that I want to solve
>>         this problem.
>>              Now, I think this is a matter of JVM but I have no idea to
>>              investigate it.
>>              Is there any way to avoid this delay?
>>
>>
>>
>>              Thank you for your attention
>>              Yuta
>>
>>
>>
>>

Reply via email to