Thank you, Ran, yes, I understand about the segment size impact - however, if you look at the configuration of those topics (see below) - you will see they have exactly the same config, with the same segment.bytes setting (to 1G, default), and all other settings are also the same - except for the retention: 1 topic has retention 4 hrs:
segment.bytes=1073741824 sensitive=false synonyms={STATIC_BROKER_CONFIG:log.segment.bytes=1073741824, DEFAULT_CONFIG:log.segment.bytes=1073741824} retention.ms=14400000 sensitive=false synonyms={DYNAMIC_TOPIC_CONFIG:retention.ms=14400000} 2nd topic has retention 720 hrs (30 days), default: segment.bytes=1073741824 sensitive=false synonyms={STATIC_BROKER_CONFIG:log.segment.bytes=1073741824, DEFAULT_CONFIG:log.segment.bytes=1073741824} retention.ms=2592000000 sensitive=false synonyms={} yet, 1st topic is not being cleaned up, but the 2-nd topic is being cleaned up every 5 min, due to " Found deletable segments with base offsets [11755700] due to retention time 2592000000ms breach " And I can see that the size of the segments never reaches 1G for the 2-nd topic either ... thank you, Marina Sent with [ProtonMail](https://protonmail.com) Secure Email. ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐ On Tuesday, May 25, 2021 4:11 PM, Ran Lupovich <ranlupov...@gmail.com> wrote: > Sorry I did not see all the info at first, what do you mean by topic getting > cleaned, you have setting to check retention every 5 minutes, the data that > getting "cleaned" is the older data which is 30 days old... am I missing > something? > > בתאריך יום ג׳, 25 במאי 2021, 23:04, מאת Ran Lupovich <ranlupov...@gmail.com>: > >> By the segment size you are "delete" after 1 giga bytes is full , per >> partition, you need to remmber the retention is done when segments closed , >> per partition >> >> בתאריך יום ג׳, 25 במאי 2021, 22:59, מאת Ran Lupovich >> <ranlupov...@gmail.com>: >> >>> Have you checked the segment size? Did you decribe the topic >>> configuration?maybe you created it with some settings you dont remember >>> >>> בתאריך יום ג׳, 25 במאי 2021, 19:51, מאת Marina Popova >>> <ppine7...@protonmail.com.invalid>: >>> >>>> Any idea what is wrong here? I have restarted Kafka brokers a few times, >>>> and all other Confluent services like KSQL - but I see exactly the same >>>> behavior - one topic gets its logs cleaned up every 5 minutes, while the >>>> other one - does not get cleaned up at all .... >>>> >>>> Is there anything else I could check, apart from what I already did - see >>>> the post below - to troubleshoot this? >>>> >>>> thank you! >>>> Marina >>>> >>>> Sent with ProtonMail Secure Email. >>>> >>>> ‐‐‐‐‐‐‐ Original Message ‐‐‐‐‐‐‐ >>>> On Thursday, May 20, 2021 2:10 PM, Marina Popova >>>> <ppin...@protonmail.com.INVALID> wrote: >>>> >>>>> Hi, I have posted this question on SO: >>>>> https://stackoverflow.com/questions/67625641/kafka-segments-are-deleted-too-often-or-not-at-all >>>>> but wanted to re-post here as well in case someone spots the issue right >>>>> away .... >>>>> >>>>> Thank you for your help! >>>>> >>>>> > > > > > >>>>> >>>>> We have two topics on our Kafka cluster that exhibit weird (wrong) >>>>> behavior related to retention configuration. >>>>> >>>>> One topic, tracking.ap.client.traffic, has retention set explicitly to >>>>> "retention.ms=14400000" (4 hrs) - but it is not cleaned up , grows in >>>>> size, and caused 2 out of 3 kafka brokers to run out of disk space. >>>>> (details about config and log below) >>>>> >>>>> Second topic, tracking.ap.client.traffic.keyed, is created in KSQL as a >>>>> stream topic: >>>>> >>>>> CREATE STREAM AP_CLIENT_TRAFFIC_KEYED >>>>> WITH (KAFKA_TOPIC='tracking.ap.client.traffic.keyed', >>>>> TIMESTAMP='ACTIVITY_DATE', VALUE_FORMAT='JSON', PARTITIONS=6) >>>>> AS SELECT >>>>> ... >>>>> >>>>> >>>>> >>>>> its retention is set to the default broker value, which is 720 hrs : >>>>> >>>>> cat /etc/kafka/server.properties | grep retention >>>>> log.retention.hours=720 >>>>> # A size-based retention policy for logs. Segments are pruned from the >>>>> log unless the remaining >>>>> # segments drop below log.retention.bytes. Functions independently of >>>>> log.retention.hours. >>>>> #log.retention.bytes=1073741824 >>>>> # to the retention policies >>>>> log.retention.check.interval.ms=300000 >>>>> >>>>> >>>>> This topic, though, gets cleaned up every 5 min or so - according to the >>>>> logs.... >>>>> The log entry says the segment is marked for deletion "due to retention >>>>> time 2592000000ms breach (kafka.log.Log)" - but how can that be true if >>>>> this is happening every 5 min?? >>>>> >>>>> No size-based retention is set for any of the topics. >>>>> >>>>> Two questions: >>>>> >>>>> 1. why is the first topic not being cleaned p? >>>>> 2. why is the second topic being cleaned up so often? >>>>> >>>>> Below are the details about logs and full config of both topics: >>>>> >>>>> log entries for tracking.ap.client.traffic.keyed-2 topic/partition - show >>>>> that this partition is getting cleaned too often: >>>>> >>>>> >>>>> [2021-05-19 21:35:05,822] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Incrementing log start offset to 11755700 (kafka.log.Log) >>>>> [2021-05-19 21:36:05,822] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Deleting segment 11753910 (kafka.log.Log) >>>>> [2021-05-19 21:36:05,825] INFO Deleted log >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011753910.log.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:36:05,827] INFO Deleted offset index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011753910.index.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:36:05,829] INFO Deleted time index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011753910.timeindex.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:40:05,838] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] Found >>>>> deletable segments with base offsets [11755700] due to retention time >>>>> 2592000000ms breach (kafka.log.Log) >>>>> [2021-05-19 21:40:05,843] INFO [ProducerStateManager >>>>> partition=tracking.ap.client.traffic.keyed-2] Writing producer snapshot >>>>> at offset 11757417 (kafka.log.ProducerStateManager) >>>>> [2021-05-19 21:40:05,845] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Rolled new log segment at offset 11757417 in 7 ms. (kafka.log.Log) >>>>> [2021-05-19 21:40:05,845] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Scheduling log segment [baseOffset 11755700, size 936249] for deletion. >>>>> (kafka.log.Log) >>>>> [2021-05-19 21:40:05,847] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Incrementing log start offset to 11757417 (kafka.log.Log) >>>>> [2021-05-19 21:41:05,848] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Deleting segment 11755700 (kafka.log.Log) >>>>> [2021-05-19 21:41:05,850] INFO Deleted log >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011755700.log.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:41:05,852] INFO Deleted offset index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011755700.index.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:41:05,853] INFO Deleted time index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011755700.timeindex.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:45:05,834] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] Found >>>>> deletable segments with base offsets [11757417] due to retention time >>>>> 2592000000ms breach (kafka.log.Log) >>>>> [2021-05-19 21:45:05,837] INFO [ProducerStateManager >>>>> partition=tracking.ap.client.traffic.keyed-2] Writing producer snapshot >>>>> at offset 11759201 (kafka.log.ProducerStateManager) >>>>> [2021-05-19 21:45:05,839] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Rolled new log segment at offset 11759201 in 6 ms. (kafka.log.Log) >>>>> [2021-05-19 21:45:05,839] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Scheduling log segment [baseOffset 11757417, size 977444] for deletion. >>>>> (kafka.log.Log) >>>>> [2021-05-19 21:45:05,842] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Incrementing log start offset to 11759201 (kafka.log.Log) >>>>> [2021-05-19 21:46:05,842] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Deleting segment 11757417 (kafka.log.Log) >>>>> [2021-05-19 21:46:05,844] INFO Deleted log >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011757417.log.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:46:05,845] INFO Deleted offset index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011757417.index.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:46:05,847] INFO Deleted time index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011757417.timeindex.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:50:05,832] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] Found >>>>> deletable segments with base offsets [11759201] due to retention time >>>>> 2592000000ms breach (kafka.log.Log) >>>>> [2021-05-19 21:50:05,837] INFO [ProducerStateManager >>>>> partition=tracking.ap.client.traffic.keyed-2] Writing producer snapshot >>>>> at offset 11760980 (kafka.log.ProducerStateManager) >>>>> [2021-05-19 21:50:05,839] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Rolled new log segment at offset 11760980 in 7 ms. (kafka.log.Log) >>>>> [2021-05-19 21:50:05,839] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Scheduling log segment [baseOffset 11759201, size 991447] for deletion. >>>>> (kafka.log.Log) >>>>> [2021-05-19 21:50:05,844] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Incrementing log start offset to 11760980 (kafka.log.Log) >>>>> [2021-05-19 21:51:05,844] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Deleting segment 11759201 (kafka.log.Log) >>>>> [2021-05-19 21:51:05,846] INFO Deleted log >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011759201.log.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:51:05,848] INFO Deleted offset index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011759201.index.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:51:05,849] INFO Deleted time index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011759201.timeindex.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:55:05,828] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] Found >>>>> deletable segments with base offsets [11760980] due to retention time >>>>> 2592000000ms breach (kafka.log.Log) >>>>> [2021-05-19 21:55:05,833] INFO [ProducerStateManager >>>>> partition=tracking.ap.client.traffic.keyed-2] Writing producer snapshot >>>>> at offset 11762941 (kafka.log.ProducerStateManager) >>>>> [2021-05-19 21:55:05,835] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Rolled new log segment at offset 11762941 in 7 ms. (kafka.log.Log) >>>>> [2021-05-19 21:55:05,835] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Scheduling log segment [baseOffset 11760980, size 1079204] for deletion. >>>>> (kafka.log.Log) >>>>> [2021-05-19 21:55:05,838] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Incrementing log start offset to 11762941 (kafka.log.Log) >>>>> [2021-05-19 21:56:05,838] INFO [Log >>>>> partition=tracking.ap.client.traffic.keyed-2, dir=/apps/kafka-data] >>>>> Deleting segment 11760980 (kafka.log.Log) >>>>> [2021-05-19 21:56:05,840] INFO Deleted log >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011760980.log.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:56:05,842] INFO Deleted offset index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011760980.index.deleted. >>>>> (kafka.log.LogSegment) >>>>> [2021-05-19 21:56:05,843] INFO Deleted time index >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2/00000000000011760980.timeindex.deleted. >>>>> (kafka.log.LogSegment) >>>>> >>>>> >>>>> There areno such entries for log cleaning for the first topic, >>>>> tracking.ap.client.traffic, which indicates it is never even checked for >>>>> hitting retention ?? >>>>> >>>>> Full config for tracking.ap.client.traffic.keyed topic: (that is cleaned >>>>> up too often): >>>>> >>>>> mac-lt-mpopova:confluent-6.1.0 mpopova$ ./bin/kafka-configs >>>>> --bootstrap-server mybroker1:9092 --entity-type topics --entity-name >>>>> tracking.ap.client.traffic.keyed --describe --all >>>>> >>>>> All configs for topic tracking.ap.client.traffic.keyed are: >>>>> compression.type=producer sensitive=false >>>>> synonyms={DEFAULT_CONFIG:compression.type=producer} >>>>> leader.replication.throttled.replicas= sensitive=false synonyms={} >>>>> message.downconversion.enable=true sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.message.downconversion.enable=true} >>>>> min.insync.replicas=2 sensitive=false >>>>> synonyms={STATIC_BROKER_CONFIG:min.insync.replicas=2, >>>>> DEFAULT_CONFIG:min.insync.replicas=1} >>>>> segment.jitter.ms=0 sensitive=false synonyms={} >>>>> cleanup.policy=delete sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.cleanup.policy=delete} >>>>> flush.ms=9223372036854775807 sensitive=false synonyms={} >>>>> follower.replication.throttled.replicas= sensitive=false synonyms={} >>>>> segment.bytes=1073741824 sensitive=false >>>>> synonyms={STATIC_BROKER_CONFIG:log.segment.bytes=1073741824, >>>>> DEFAULT_CONFIG:log.segment.bytes=1073741824} >>>>> retention.ms=2592000000 sensitive=false synonyms={} >>>>> flush.messages=9223372036854775807 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.flush.interval.messages=9223372036854775807} >>>>> message.format.version=2.1-IV2 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.message.format.version=2.1-IV2} >>>>> file.delete.delay.ms=60000 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.segment.delete.delay.ms=60000} >>>>> max.message.bytes=1000012 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:message.max.bytes=1000012} >>>>> min.compaction.lag.ms=0 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.cleaner.min.compaction.lag.ms=0} >>>>> message.timestamp.type=CreateTime sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.message.timestamp.type=CreateTime} >>>>> preallocate=false sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.preallocate=false} >>>>> min.cleanable.dirty.ratio=0.5 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.cleaner.min.cleanable.ratio=0.5} >>>>> index.interval.bytes=4096 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.index.interval.bytes=4096} >>>>> unclean.leader.election.enable=false sensitive=false >>>>> synonyms={STATIC_BROKER_CONFIG:unclean.leader.election.enable=false, >>>>> DEFAULT_CONFIG:unclean.leader.election.enable=false} >>>>> retention.bytes=-1 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.retention.bytes=-1} >>>>> delete.retention.ms=86400000 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.cleaner.delete.retention.ms=86400000} >>>>> segment.ms=604800000 sensitive=false synonyms={} >>>>> message.timestamp.difference.max.ms=9223372036854775807 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.message.timestamp.difference.max.ms=9223372036854775807} >>>>> segment.index.bytes=10485760 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.index.size.max.bytes=10485760} >>>>> >>>>> >>>>> Full config fortracking.ap.client.traffic topic: (not being cleaned up at >>>>> all): >>>>> >>>>> mac-lt-mpopova:confluent-6.1.0 mpopova$ ./bin/kafka-configs >>>>> --bootstrap-server mybroker1:9092 --entity-type topics --entity-name >>>>> tracking.ap.client.traffic --describe --all >>>>> >>>>> All configs for topic tracking.ap.client.traffic are: >>>>> compression.type=producer sensitive=false >>>>> synonyms={DEFAULT_CONFIG:compression.type=producer} >>>>> leader.replication.throttled.replicas= sensitive=false synonyms={} >>>>> message.downconversion.enable=true sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.message.downconversion.enable=true} >>>>> min.insync.replicas=2 sensitive=false >>>>> synonyms={STATIC_BROKER_CONFIG:min.insync.replicas=2, >>>>> DEFAULT_CONFIG:min.insync.replicas=1} >>>>> segment.jitter.ms=0 sensitive=false synonyms={} >>>>> cleanup.policy=delete sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.cleanup.policy=delete} >>>>> flush.ms=9223372036854775807 sensitive=false synonyms={} >>>>> follower.replication.throttled.replicas= sensitive=false synonyms={} >>>>> segment.bytes=1073741824 sensitive=false >>>>> synonyms={STATIC_BROKER_CONFIG:log.segment.bytes=1073741824, >>>>> DEFAULT_CONFIG:log.segment.bytes=1073741824} >>>>> retention.ms=14400000 sensitive=false >>>>> synonyms={DYNAMIC_TOPIC_CONFIG:retention.ms=14400000} >>>>> flush.messages=9223372036854775807 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.flush.interval.messages=9223372036854775807} >>>>> message.format.version=2.1-IV2 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.message.format.version=2.1-IV2} >>>>> file.delete.delay.ms=60000 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.segment.delete.delay.ms=60000} >>>>> max.message.bytes=1000012 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:message.max.bytes=1000012} >>>>> min.compaction.lag.ms=0 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.cleaner.min.compaction.lag.ms=0} >>>>> message.timestamp.type=CreateTime sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.message.timestamp.type=CreateTime} >>>>> preallocate=false sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.preallocate=false} >>>>> min.cleanable.dirty.ratio=0.5 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.cleaner.min.cleanable.ratio=0.5} >>>>> index.interval.bytes=4096 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.index.interval.bytes=4096} >>>>> unclean.leader.election.enable=false sensitive=false >>>>> synonyms={STATIC_BROKER_CONFIG:unclean.leader.election.enable=false, >>>>> DEFAULT_CONFIG:unclean.leader.election.enable=false} >>>>> retention.bytes=-1 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.retention.bytes=-1} >>>>> delete.retention.ms=86400000 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.cleaner.delete.retention.ms=86400000} >>>>> segment.ms=604800000 sensitive=false synonyms={} >>>>> message.timestamp.difference.max.ms=9223372036854775807 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.message.timestamp.difference.max.ms=9223372036854775807} >>>>> segment.index.bytes=10485760 sensitive=false >>>>> synonyms={DEFAULT_CONFIG:log.index.size.max.bytes=10485760} >>>>> >>>>> >>>>> Here is an example directory content of the logs for both topics: >>>>> >>>>> ls -la /apps/kafka-data/tracking.ap.client.traffic-* >>>>> >>>>> /apps/kafka-data/tracking.ap.client.traffic-1: >>>>> total 412640 >>>>> drwxr-xr-x 2 cp-kafka confluent 141 May 17 15:01 . >>>>> drwxr-xr-x 290 cp-kafka confluent 20480 May 19 22:48 .. >>>>> -rw-r--r-- 1 cp-kafka confluent 10485760 May 19 22:48 >>>>> 00000000000000000000.index >>>>> -rw-r--r-- 1 cp-kafka confluent 420599482 May 19 22:48 >>>>> 00000000000000000000.log >>>>> -rw-r--r-- 1 cp-kafka confluent 10485756 May 19 22:48 >>>>> 00000000000000000000.timeindex >>>>> -rw-r--r-- 1 cp-kafka confluent 8 May 17 15:01 leader-epoch-checkpoint >>>>> >>>>> /apps/kafka-data/tracking.ap.client.traffic-11: >>>>> total 412844 >>>>> drwxr-xr-x 2 cp-kafka confluent 141 May 17 11:25 . >>>>> drwxr-xr-x 290 cp-kafka confluent 20480 May 19 22:48 .. >>>>> -rw-r--r-- 1 cp-kafka confluent 10485760 May 19 22:48 >>>>> 00000000000000000000.index >>>>> -rw-r--r-- 1 cp-kafka confluent 420788365 May 19 22:48 >>>>> 00000000000000000000.log >>>>> -rw-r--r-- 1 cp-kafka confluent 10485756 May 19 22:48 >>>>> 00000000000000000000.timeindex >>>>> -rw-r--r-- 1 cp-kafka confluent 8 May 17 11:25 leader-epoch-checkpoint >>>>> >>>>> /apps/kafka-data/tracking.ap.client.traffic-13: >>>>> total 412660 >>>>> drwxr-xr-x 2 cp-kafka confluent 141 May 17 15:01 . >>>>> drwxr-xr-x 290 cp-kafka confluent 20480 May 19 22:48 .. >>>>> -rw-r--r-- 1 cp-kafka confluent 10485760 May 19 22:48 >>>>> 00000000000000000000.index >>>>> -rw-r--r-- 1 cp-kafka confluent 420606307 May 19 22:48 >>>>> 00000000000000000000.log >>>>> -rw-r--r-- 1 cp-kafka confluent 10485756 May 19 22:48 >>>>> 00000000000000000000.timeindex >>>>> -rw-r--r-- 1 cp-kafka confluent 8 May 17 15:01 leader-epoch-checkpoint >>>>> >>>>> /apps/kafka-data/tracking.ap.client.traffic-14: >>>>> total 412920 >>>>> drwxr-xr-x 2 cp-kafka confluent 141 May 17 11:25 . >>>>> drwxr-xr-x 290 cp-kafka confluent 20480 May 19 22:48 .. >>>>> -rw-r--r-- 1 cp-kafka confluent 10485760 May 19 22:48 >>>>> 00000000000000000000.index >>>>> -rw-r--r-- 1 cp-kafka confluent 420862000 May 19 22:48 >>>>> 00000000000000000000.log >>>>> -rw-r--r-- 1 cp-kafka confluent 10485756 May 19 22:48 >>>>> 00000000000000000000.timeindex >>>>> -rw-r--r-- 1 cp-kafka confluent 8 May 17 11:25 leader-epoch-checkpoint >>>>> >>>>> >>>>> ls -la /apps/kafka-data/tracking.ap.client.traffic.keyed* >>>>> >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-1: >>>>> total 288 >>>>> drwxr-xr-x 2 cp-kafka confluent 141 May 19 22:47 . >>>>> drwxr-xr-x 290 cp-kafka confluent 20480 May 19 22:48 .. >>>>> -rw-r--r-- 1 cp-kafka confluent 10485760 May 19 22:48 >>>>> 00000000000004255431.index >>>>> -rw-r--r-- 1 cp-kafka confluent 254355 May 19 22:48 >>>>> 00000000000004255431.log >>>>> -rw-r--r-- 1 cp-kafka confluent 10485756 May 19 22:48 >>>>> 00000000000004255431.timeindex >>>>> -rw-r--r-- 1 cp-kafka confluent 14 May 19 22:47 leader-epoch-checkpoint >>>>> >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-2: >>>>> total 816 >>>>> drwxr-xr-x 2 cp-kafka confluent 178 May 19 22:46 . >>>>> drwxr-xr-x 290 cp-kafka confluent 20480 May 19 22:48 .. >>>>> -rw-r--r-- 1 cp-kafka confluent 10485760 May 19 22:48 >>>>> 00000000000011782064.index >>>>> -rw-r--r-- 1 cp-kafka confluent 787175 May 19 22:48 >>>>> 00000000000011782064.log >>>>> -rw-r--r-- 1 cp-kafka confluent 10 May 19 22:45 >>>>> 00000000000011782064.snapshot >>>>> -rw-r--r-- 1 cp-kafka confluent 10485756 May 19 22:48 >>>>> 00000000000011782064.timeindex >>>>> -rw-r--r-- 1 cp-kafka confluent 15 May 19 22:45 leader-epoch-checkpoint >>>>> >>>>> /apps/kafka-data/tracking.ap.client.traffic.keyed-3: >>>>> total 1812 >>>>> drwxr-xr-x 2 cp-kafka confluent 141 May 19 22:47 . >>>>> drwxr-xr-x 290 cp-kafka confluent 20480 May 19 22:48 .. >>>>> -rw-r--r-- 1 cp-kafka confluent 10485760 May 19 22:48 >>>>> 00000000000028393394.index >>>>> -rw-r--r-- 1 cp-kafka confluent 1814528 May 19 22:48 >>>>> 00000000000028393394.log >>>>> -rw-r--r-- 1 cp-kafka confluent 10485756 May 19 22:48 >>>>> 00000000000028393394.timeindex >>>>> -rw-r--r-- 1 cp-kafka confluent 15 May 19 22:47 leader-epoch-checkpoint >>>>> >>>>> >>>>> Thank you! >>>>> >>>>> Sent with ProtonMail Secure Email.