Hi, Yes there are plans to support early results and control the result latency / result completeness trade off. However, I cannot say when these features will be available.
Best, Fabian 2018-03-19 8:14 GMT+01:00 LiYue <liyue2...@gmail.com>: > Hi , > > Is there any plan to adding these features to flink SQL ? > > Thanks > LiYue > tig.jd.com > > > > 在 2018年3月14日,上午7:48,Fabian Hueske <fhue...@apache.org> 写道: > > Hi, > > Chesnay is right. > SQL and Table API do not support early window results and no allowed > lateness to update results with late arriving data. > If you need such features, you should use the DataStream API. > > Best, Fabian > > > 2018-03-13 12:10 GMT+01:00 Chesnay Schepler <ches...@apache.org>: > >> I don't think you can specify custom triggers when using purer SQL, but >> maybe Fabian or Timo know a SQL way of implementing your goal. >> >> >> On 12.03.2018 13:16, 李玥 wrote: >> >> Hi Chirag, >> Thank for your reply! >> I found a provided ContinuousEventTimeTrigger should be worked in my >> situation. >> Most examples are based on Table API like >> ‘ds.keyBy(0).window().trigger(MyTrigger.of())…’, >> But how to apply the trigger to a pure Flink SQL Application ? >> >> >> >> >> >> >> 在 2018年3月12日,下午5:20,Chirag Dewan <chirag.dewa...@yahoo.in> 写道: >> >> Hi LiYue, >> >> This should help : Apache Flink 1.5-SNAPSHOT Documentation: Windows >> <https://ci.apache.org/projects/flink/flink-docs-master/dev/stream/operators/windows.html#triggers> >> >> >> Apache Flink 1.5-SNAPSHOT Documentation: Windows >> <https://ci.apache.org/projects/flink/flink-docs-master/dev/stream/operators/windows.html#triggers> >> >> >> >> So basically you need to register a processing time trigger at every 10 >> minutes and on callback, you can FIRE the window result like this: >> >> @Override >> public TriggerResult onProcessingTime(long time, TimeWindow window, >> TriggerContext ctx) throws Exception { >> // schedule next timer >> ctx.registerProcessingTimeTimer(System.currentTimeMillis() + >> 1000L); >> return TriggerResult.FIRE; >> } >> >> >> I hope it helps. >> >> Chirag >> >> On Monday, 12 March, 2018, 2:10:25 PM IST, 李玥 <liyue2...@gmail.com> >> wrote: >> >> >> Hi,team >> I’m working on a event-time based aggregation application with flink >> SQL. Is there any way to keep sinking partial aggregation result BEFORE >> time window closed? >> For example, My SQL: >> select … >> from my_table >> GROUP BY TUMBLE(`timestamp`, INTERVAL '1’ DAY),other_column; >> Usually, Flink sink agg result after time-window closed, Is there any way >> to keep sinking TODAY’s partial aggregation result every 10 miniutes so we >> can see today’s performance on my chart. >> >> Thanks! >> LiYue >> >> >> >> > >