Do we have the right default?

-Jay


On Wed, Feb 5, 2014 at 2:04 PM, Joel Koshy <jjkosh...@gmail.com> wrote:

>
> > topics are all caught up, but I have one high volume topic (around
> > 40K msgs/sec) that is taking much longer.  I just took a few samples
> > of Replica-MaxLag to see how long it would take to catch up.
> > Currently, it is behind about 12.5 million messages and is catching
> > up at a rate of about 1600 msgs/sec.  At that rate, it'll take
> > around 9 days before the replica is caught up to the leader.
> >
> > Is there any way to speed this up?
>
> During the period your high-volume topic is under-replicated you can
> temporarily try one or both of the following:
> - Increasing num.replica.fetchers (defaults is one)
> - If you don't have too many topic-partitions you can also increase
>   replica.fetch.max.bytes.
>
> > Or, alternatively, I don't actually care about this topic's
> > history.  It is a new topic, and I know that it doesn't yet have any
> > consumers.  I'd be fine with instructing both brokers to drop
> > old logs and just start from the top of the log.  I could do this by
> > manually deleting the topic (kafka data files and in zookeeper), but
> > to do so properly with 0.8.0 I think I'd have to shut down the
> > whole cluster, correct?  I'd rather not do this, as another
> > topic does have a consumer and I don't want to lose messages for
> > it.
>
> Right - or you could do a rolling bounce and change the retention
> settings (http://kafka.apache.org/documentation.html#brokerconfigs) of
> that topic to something low so it gets expired and then do another
> rolling bounce to remove the override.
>
> --
> Joel
>

Reply via email to