Thanks Biao, just want to not reinvent the wheel :)

> On Jul 22, 2019, at 4:29 PM, Biao Liu <mmyy1...@gmail.com> wrote:
> 
> Hi Andy,
> 
> As far as I know, Flink does not support feature like that. 
> 
> I would suggest recording and calculating the time in user code.
> For example, add a timestamp field (maybe an array) in your record with 
> printing a timestamp on in by each processing.
> 
> 
> Andy Hoang <a...@parcelperform.com <mailto:a...@parcelperform.com>> 
> 于2019年7月22日周一 下午4:49写道:
> Hi guys,
> 
> I’m trying to write elk log for flink, this help us to store/calculate 
> processing time of a group of operators for business auditing.
> 
> I read about process_function and Debugging Windows & Event Time in docs. 
> They’re focus on “keyed” events and monitoring using web/metric, where I want 
> to, for example:
> 
> val stream = env.add_source(kafka_source).map(doX).map(doY).filter(Z)
> Async.unordedWait(stream, calApiBlah).add_source(kafka_source)
> 
> Track total time from `doY` till before add_source operator(). This kind of 
> time can be processing time,
> or I want to track total time from source to sink of a stream event  (maybe 
> disgesting time?),
> So the target is to retrieve out total time of a set of chosen operators
> 
> I think flink should have supported it somehow, but its just me haven’t found 
> it in the docs yet.
> 
> Thanks,
> 
> Andy,
> 
> 

Reply via email to