Ok. Attached is the PPT of what am attempting to achieve w.r.t. time Hope I am all set to achieve the three bullets mentioned in attached slide to create reports with KafkaSource and KafkaBuilder approach.
If you have any additional tips to share please do so after going through the slide attached (example for live dashboards use case) Kind regards Vinay -----Original Message----- From: Aljoscha Krettek <aljos...@apache.org> Sent: 13 January 2021 14:06 To: user@flink.apache.org Cc: RAICHUR, VINAY <vinay.raic...@t-systems.com> Subject: Re: Flink to get historical data from kafka between timespan t1 & t2 On 2021/01/13 07:58, vinay.raic...@t-systems.com wrote: >Not sure about your proposal regarding Point 3: >* firstly how is it ensured that the stream is closed? If I understand >the doc correctly the stream will be established starting with the >latest timestamp (hmm... is it not a standard behaviour?) and will >never finish (UNBOUNDED), On the first question of standard behaviour: the default is to start from the group offsets that are available in Kafka. This uses the configured consumer group. I think it's better to be explicit, though, and specify sth like `EARLIEST` or `LATEST`, etc. And yes, the stream will start but never stop with this version of the Kafka connector. Only when you use the new `KafkaSource` can you also specify an end timestamp that will make the Kafka source shut down eventually. >* secondly it is still not clear how to get the latest event at a >given time point in the past? You are referring to getting a single record, correct? I don't think this is possible with Flink. All you can do is get a stream from Kafka that is potentially bounded by a start timestamp and/or end timestamp. Best, Aljoscha
Positioning_Use_Cases_TrackingData_Past_Now.pptx
Description: Positioning_Use_Cases_TrackingData_Past_Now.pptx