huxihx created KAFKA-6425:
-----------------------------

             Summary: Calculating cleanBytes in LogToClean might not be correct
                 Key: KAFKA-6425
                 URL: https://issues.apache.org/jira/browse/KAFKA-6425
             Project: Kafka
          Issue Type: Bug
          Components: core
    Affects Versions: 1.0.0
            Reporter: huxihx


In class `LogToClean`, the calculation for `cleanBytes` is as below:
{code:java}
val cleanBytes = log.logSegments(-1, firstDirtyOffset).map(_.size.toLong).sum
{code}

Most of the time, the `firstDirtyOffset` is the base offset of active segment 
which works pretty well with log.logSegments, so we can calculate the 
cleanBytes by safely summing up the sizes of all log segments whose base offset 
is less than `firstDirtyOffset`.

However, things changed after `firstUnstableOffset` was introduced. Users could 
indirectly change this offset to a non-base offset(changing log start offset 
for instance). In this case, it's not correct to sum up the total size for a 
log segment. Instead, we should only sum up the bytes between the base offset 
and `firstUnstableOffset`.

Let me show an example:
Say I have three log segments, shown as below:
0L       -->  log segment1, size: 1000Bytes
1234L -->  log segment2, size: 1000Bytes
4567L --> active log segment, current size: 500Bytes

Based on the current code, if `firstUnstableOffset` is deliberately set to 
2000L(this could be possible, since it's lower bounded by the log start offset 
and user could explicitly change LSO), then `cleanBytes` is calculated as 
2000Bytes which is wrong. The expected value should be 1000 + (bytes between 
offset 1234L and 2000L) 

[~junrao] [~ijuma] Do all of these make sense?






--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

Reply via email to