Hi Guillermo, 1) Yes in your case, the streams are really a "changelog" stream, hence you should create the stream as KTable, and do KTable-KTable join.
2) Could elaborate about "achieving this"? What behavior do require in the application logic? Guozhang On Thu, Apr 14, 2016 at 1:30 AM, Guillermo Lammers Corral < guillermo.lammers.cor...@tecsisa.com> wrote: > Hi, > > I am a newbie to Kafka Streams and I am using it trying to solve a > particular use case. Let me explain. > > I have two sources of data both like that: > > Key (string) > DateTime (hourly granularity) > Value > > I need to join the two sources by key and date (hour of day) to obtain: > > Key (string) > DateTime (hourly granularity) > ValueSource1 > ValueSource2 > > I think that first I'd need to push the messages in Kafka topics with the > date as part of the key because I'll group by key taking into account the > date. So maybe the key must be a new string like key_timestamp. But, of > course, it is not the main problem, is just an additional explanation. > > Ok, so data are in topics, here we go! > > - Multiple records allows per key but only the latest value for a record > key will be considered. I should use two KTable with some join strategy, > right? > > - Data of both sources could arrive at any time. What can I do to achieve > this? > > Thanks in advance. > -- -- Guozhang