[ https://issues.apache.org/jira/browse/KAFKA-2477?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14952127#comment-14952127 ]
Chinmay Soman edited comment on KAFKA-2477 at 10/11/15 1:49 AM: ---------------------------------------------------------------- [~ewencp] I'm attaching the screenshot of Max lag observed for the different brokers which describes the behaviour. Also here's the pertinent log: ======== [2015-10-10 22:17:17,759] 3171793337 [kafka-request-handler-0] WARN kafka.server.ReplicaManager - [Replica Manager on Broker 70]: Error when processing fetch request for partition [...topic...,62] offset 91963211 from follower with correlation id 176614372. Possible cause: Request for offset 91963211 but we only have log segments in the range 55923986 to 91963210. [2015-10-10 22:17:17,759] 3171793337 [kafka-request-handler-4] WARN kafka.server.ReplicaManager - [Replica Manager on Broker 70]: Error when processing fetch request for partition [...topic...,62] offset 91963211 from follower with correlation id 152788081. Possible cause: Request for offset 91963211 but we only have log segments in the range 55923986 to 91963210. [2015-10-10 22:17:20,256] 3171795834 [kafka-scheduler-4] INFO kafka.cluster.Partition - Partition [...topic...,62] on broker 70: Shrinking ISR for partition [hp.event.user.driver_app.experiment,62] from 70,69,71 to 70 ======== was (Author: cpsoman): Max lag observed for the different brokers > Replicas spuriously deleting all segments in partition > ------------------------------------------------------ > > Key: KAFKA-2477 > URL: https://issues.apache.org/jira/browse/KAFKA-2477 > Project: Kafka > Issue Type: Bug > Affects Versions: 0.8.2.1 > Reporter: HÃ¥kon Hitland > Assignee: Jiangjie Qin > Fix For: 0.9.0.0 > > Attachments: Screen Shot 2015-10-10 at 6.54.44 PM.png, kafka_log.txt, > kafka_log_trace.txt > > > We're seeing some strange behaviour in brokers: a replica will sometimes > schedule all segments in a partition for deletion, and then immediately start > replicating them back, triggering our check for under-replicating topics. > This happens on average a couple of times a week, for different brokers and > topics. > We have per-topic retention.ms and retention.bytes configuration, the topics > where we've seen this happen are hitting the size limit. -- This message was sent by Atlassian JIRA (v6.3.4#6332)