Hi Guillermo,

1) Yes in your case, the streams are really a "changelog" stream, hence you
should create the stream as KTable, and do KTable-KTable join.

2) Could elaborate about "achieving this"? What behavior do require in the
application logic?


Guozhang


On Thu, Apr 14, 2016 at 1:30 AM, Guillermo Lammers Corral <
guillermo.lammers.cor...@tecsisa.com> wrote:

> Hi,
>
> I am a newbie to Kafka Streams and I am using it trying to solve a
> particular use case. Let me explain.
>
> I have two sources of data both like that:
>
> Key (string)
> DateTime (hourly granularity)
> Value
>
> I need to join the two sources by key and date (hour of day) to obtain:
>
> Key (string)
> DateTime (hourly granularity)
> ValueSource1
> ValueSource2
>
> I think that first I'd need to push the messages in Kafka topics with the
> date as part of the key because I'll group by key taking into account the
> date. So maybe the key must be a new string like key_timestamp. But, of
> course, it is not the main problem, is just an additional explanation.
>
> Ok, so data are in topics, here we go!
>
> - Multiple records allows per key but only the latest value for a record
> key will be considered. I should use two KTable with some join strategy,
> right?
>
> - Data of both sources could arrive at any time. What can I do to achieve
> this?
>
> Thanks in advance.
>



-- 
-- Guozhang

Reply via email to