Franco Bonazza created KAFKA-6933: ------------------------------------- Summary: Broker reports Corrupted index warnings apparently infinitely Key: KAFKA-6933 URL: https://issues.apache.org/jira/browse/KAFKA-6933 Project: Kafka Issue Type: Bug Affects Versions: 1.0.1 Reporter: Franco Bonazza
I'm running into a situation where the server logs show continuously the following snippet: {noformat} [2018-05-23 10:58:56,590] INFO Loading producer state from offset 20601420 for partition transaction_r10_updates-6 with message format version 2 (kafka.log.Log) [2018-05-23 10:58:56,592] INFO Loading producer state from snapshot file '/data/0/kafka-logs/transaction_r10_updates-6/00000000000020601420.snapshot' for partition transaction_r10_u pdates-6 (kafka.log.ProducerStateManager) [2018-05-23 10:58:56,593] INFO Completed load of log transaction_r10_updates-6 with 74 log segments, log start offset 0 and log end offset 20601420 in 5823 ms (kafka.log.Log) [2018-05-23 10:58:58,761] WARN Found a corrupted index file due to requirement failed: Corrupt index found, index file (/data/0/kafka-logs/transaction_r10_updates-15/00000000000020544956.index) has non-zero size but the last offset is 20544956 which is no larger than the base offset 20544956.}. deleting /data/0/kafka-logs/transaction_r10_updates-15/00000000000020544956.timeindex, /data/0/kafka-logs/transaction_r10_updates-15/00000000000020544956.index, and /data/0/kafka-logs/transaction_r10_updates-15/00000000000020544956.txnindex and rebuilding index... (kafka.log.Log) [2018-05-23 10:58:58,763] INFO Loading producer state from snapshot file '/data/0/kafka-logs/transaction_r10_updates-15/00000000000020544956.snapshot' for partition transaction_r10_updates-15 (kafka.log.ProducerStateManager) [2018-05-23 10:59:02,202] INFO Recovering unflushed segment 20544956 in log transaction_r10_updates-15. (kafka.log.Log){noformat} The set up is the following, Broker is 1.0.1 There are mirrors from another cluster using client 0.10.2.1 There are kafka streams and other custom consumer / producers using 1.0.0 client. While is doing this the JVM of the broker is up but it doesn't respond so it's impossible to produce, consume or run any commands. If I delete all the index files the WARN turns into an ERROR, which takes a long time (1 day last time I tried) but eventually it goes into a healthy state, then I start the producers and things are still healthy, but when I start the consumers it quickly goes into the original WARN loop, which seems infinite. I couldn't find any references to the problem, it seems to be at least mis-reporting the issue, and perhaps it's not infinite? I let it loop over the WARN for over a day and it never moved past that, and if there was something really wrong with the state maybe it should be reported. The log cleaner log showed a few "too many files open" when it originally happened but ulimit has always been set to unlimited so I'm not sure what that error means. -- This message was sent by Atlassian JIRA (v7.6.3#76005)