hudeqi created KAFKA-15086: ------------------------------ Summary: The unreasonable segment size setting of the internal topics in MM2 may cause the worker startup time to be too long Key: KAFKA-15086 URL: https://issues.apache.org/jira/browse/KAFKA-15086 Project: Kafka Issue Type: Improvement Components: mirrormaker Affects Versions: 3.4.1 Reporter: hudeqi Assignee: hudeqi
As the config 'segment.bytes' for topics related MM2(such as offset.storage.topic, config.storage.topic,status.storage.topic), if following the default configuration of the broker or set it larger, then when the MM cluster runs many and complicated tasks, especially the log volume of the topic 'offset.storage.topic' is very large, it will affect the restart speed of the MM workers. After investigation, the reason is that a consumer needs to be started to read the data of ‘offset.storage.topic’ at startup. Although this topic is set to compact, if the 'segment size' is set to a large value, such as the default value of 1G, then this topic may have tens of gigabytes of data that cannot be compacted and has to be read from the earliest (because the active segment cannot be cleaned), which will consume a lot of time (in our online environment, we found that this topic stores 13G of data, it took nearly half an hour for all the data to be consumed), which caused the worker to be unable to start and execute tasks for a long time. Of course, the number of consumer threads can also be adjusted, but I think it may be easier to reduce the 'segment size', for example, refer to the default value of __consumer_offsets: 100MB -- This message was sent by Atlassian Jira (v8.20.10#820010)