Thanks Damian,

Does setting log.retention.hours have anything to do with compacted
topics?  Meaning would a topic not compact now for 90 days? I am thinking
all the internal topics that streams creates in the flow.  Having recovery
through 90 days of logs would take a good while I'd imagine.

Thanks for clarifying that the until() does in fact set properties against
the internal topics created.  That makes sense.

On Tue, May 2, 2017 at 11:44 AM, Damian Guy <damian....@gmail.com> wrote:

> Hi Garret,
>
>
> > I was running into data loss when segments are deleted faster than
> > downstream can process.  My knee jerk reaction was to set the broker
> > configs log.retention.hours=2160 and log.segment.delete.delay.ms=
> 21600000
> > and that made it go away, but I do not think this is right?
> >
> >
> I think setting log.retention.hours to 2160 is correct (not sure about
> log.segment.delete.delay.ms) as segment retention is based on the record
> timestamps. So if you have 90 day old data you want to process then you
> should set it to at least 90 days.
>
>
> > For examples sake, assume a source topic 'feed', assume a stream to
> > calculate min/max/avg to start with, using windows of 1 minute and 5
> > minutes.  I wish to use the interactive queries against the window
> stores,
> > and I wish to retain 90 days of window data to query.
> >
> So I need advice for configuration of kafka, the 'feed' topic, the store
> > topics, and the stores themselves.
> >
> >
> When you create the Windows as part of the streams app you should specify
> them something like so: TimeWindows.of(1minute).until(90days) - in this
> way
> the stores and underling changelog topics will be configured with the
> correct retention periods.
>
> Thanks,
> Damian
>

Reply via email to