Hi ,

Is there any plan to adding these features to flink SQL ?

Thanks 
LiYue
tig.jd.com



> 在 2018年3月14日,上午7:48,Fabian Hueske <fhue...@apache.org> 写道:
> 
> Hi,
> 
> Chesnay is right. 
> SQL and Table API do not support early window results and no allowed lateness 
> to update results with late arriving data.
> If you need such features, you should use the DataStream API.
> 
> Best, Fabian
> 
> 
> 2018-03-13 12:10 GMT+01:00 Chesnay Schepler <ches...@apache.org 
> <mailto:ches...@apache.org>>:
> I don't think you can specify custom triggers when using purer SQL, but maybe 
> Fabian or Timo know a SQL way of implementing your goal.
> 
> 
> On 12.03.2018 13:16, 李玥 wrote:
>> Hi Chirag,
>> Thank for your reply!
>> I found a provided ContinuousEventTimeTrigger should be worked in my 
>> situation. 
>> Most examples are based on Table API like 
>> ‘ds.keyBy(0).window().trigger(MyTrigger.of())…’, But how to apply the 
>> trigger to a pure Flink SQL Application ? 
>> 
>> 
>> 
>> 
>> 
>> 
>>> 在 2018年3月12日,下午5:20,Chirag Dewan <chirag.dewa...@yahoo.in 
>>> <mailto:chirag.dewa...@yahoo.in>> 写道:
>>> 
>>> Hi LiYue,
>>> 
>>> This should help : Apache Flink 1.5-SNAPSHOT Documentation: Windows 
>>> <https://ci.apache.org/projects/flink/flink-docs-master/dev/stream/operators/windows.html#triggers>
>>> 
>>> 
>>> Apache Flink 1.5-SNAPSHOT Documentation: Windows
>>> 
>>>  
>>> <https://ci.apache.org/projects/flink/flink-docs-master/dev/stream/operators/windows.html#triggers>
>>> 
>>> 
>>> 
>>> So basically you need to register a processing time trigger at every 10 
>>> minutes and on callback, you can FIRE the window result like this:
>>> 
>>>   @Override
>>>     public TriggerResult onProcessingTime(long time, TimeWindow window, 
>>> TriggerContext ctx) throws Exception {
>>>       // schedule next timer
>>>       ctx.registerProcessingTimeTimer(System.currentTimeMillis() + 1000L);
>>>       return TriggerResult.FIRE;
>>>     }
>>> 
>>> 
>>> I hope it helps.
>>> 
>>> Chirag
>>> 
>>> On Monday, 12 March, 2018, 2:10:25 PM IST, 李玥 <liyue2...@gmail.com 
>>> <mailto:liyue2...@gmail.com>> wrote:
>>> 
>>> 
>>> Hi,team
>>>     I’m working on a event-time based aggregation application with flink 
>>> SQL.  Is there any way to keep sinking partial aggregation result BEFORE 
>>> time window closed?
>>> For example, My SQL:
>>>     select …
>>>     from my_table
>>>     GROUP BY TUMBLE(`timestamp`, INTERVAL '1’ DAY),other_column;
>>> Usually, Flink sink agg result after time-window closed, Is there any way 
>>> to keep sinking TODAY’s partial aggregation result every 10 miniutes so we 
>>> can see today’s performance on my chart.
>>> 
>>> Thanks!
>>> LiYue
>>> 
>> 
> 
> 

Reply via email to