Hi All: I have updated the KIP to be more specific on the motivation based on the comments.Please review as you can. Appreciated. If no more review to follow I will submit the KIP for vote. Thank you!
On 3 January 2025 22:36:06 GMT, De Gao <d...@live.co.uk> wrote: >Thanks for the review. >This is an interesting idea. Indeed this will significantly reduce the data >need to be copied. But this may need to take the TTL time to get the new >replica join the ISRs. Also we need to consider how to handle the partitions >that will only do purge by data size. > >On 2 January 2025 18:11:49 GMT, Kamal Chandraprakash ><kamal.chandraprak...@gmail.com> wrote: >>Hi Deo, >> >>Thanks for the KIP! >> >>"However the limit of messages in a single partition replica is very big. >>This could lead to very big partitions (~TBs). Moving those partitions are >>very time consuming and have a big impact on system performance." >> >>One way to do faster rebalance is to have a latest-offset replica build >>strategy when expanding the replicas for a partition >>and ensure that the expanded replica does not serve as a leader until the >>data in the older nodes expires by retention time/size. >>Currently, Kafka supports only the earliest-offset strategy during >>reassignment. And, this strategy will only work for topics >>with cleanup policy set to "delete". >> >>-- >>Kamal >> >>On Thu, Jan 2, 2025 at 10:23 PM David Arthur <mum...@gmail.com> wrote: >> >>> Hey De Gao, thanks for the KIP! >>> >>> As you’re probably aware, a Partition is a logical construct in Kafka. A >>> broker hosts a partition which is composed of physical log segments. Only >>> the active segment is being written to and the others are immutable. The >>> concept of a Chunk sounds quite similar to our log segments. >>> >>> From what I can tell reading the KIP, the main difference is that a Chunk >>> can have its own assignment and therefore be replicated across different >>> brokers. >>> >>> > Horizontal scalability: the data was distributed more evenly to brokers >>> in cluster. Also achieving a more flexible resource allocation. >>> >>> I think this is only true in cases where we have a small number of >>> partitions with a large amount of data. I have certainly seen cases where a >>> small number of partitions can cause trouble with balancing the cluster. >>> >>> The idea of shuffling around older data in order to spread out the load is >>> interesting. It does seem like it would increase the complexity of the >>> client a bit when it comes to consuming the old data. Usually the client >>> can just read from a single replica from the beginning of the log to the >>> end. With this proposal, the client would need to hop around between >>> replicas as it crossed the chunk boundaries. >>> >>> > Better load balancing: The read of partition data, especially early data >>> can be distributed to more nodes other than just leader nodes. >>> >>> As you know, this is already possible with KIP-392. I guess the idea with >>> the chunks is that clients would be reading older data from less busy >>> brokers (i.e., brokers which are not the leader, or perhaps not even a >>> follower of the active chunk). I’m not sure this would always result in >>> better load balancing. It seems a bit situational. >>> >>> > Increased fault tolerance: failure of leader node will not impact read >>> older data. >>> >>> I don’t think this proposal changes the fault tolerance. A failure of a >>> leader results in a failover to a follower. If a client is consuming using >>> KIP-392, a leader failure will not affect the consumption (besides updating >>> the clients metadata). >>> >>> -- >>> >>> I guess I'm missing a key point here. What problem is this trying to solve? >>> Is it a solution for the "single partition" problem? (i.e., a topic with >>> one partition and a lot of data) >>> >>> Thanks! >>> David A >>> >>> On Tue, Dec 31, 2024 at 3:24 PM De Gao <d...@live.co.uk> wrote: >>> >>> > Thanks for the comments. I have updated the proposal to compare with >>> > tiered storage and fetch from replica. Please check. >>> > >>> > Thanks. >>> > >>> > On 11 December 2024 08:51:43 GMT, David Jacot >>> <dja...@confluent.io.INVALID> >>> > wrote: >>> > >Hi, >>> > > >>> > >Thanks for the KIP. The community is pretty busy with the Apache Kafka >>> 4.0 >>> > >release so I suppose that no one really had the time to engage in >>> > reviewing >>> > >the KIP yet. Sorry for this! >>> > > >>> > >I just read the motivation section. I think that it is an interesting >>> > idea. >>> > >However, I wonder if this is still needed now that we have tier storage >>> in >>> > >place. One of the big selling points of tier storage was that clusters >>> > >don't have to replicate tiered data anymore. Could you perhaps extend >>> the >>> > >motivation of the KIP to include tier storage in the reflexion? >>> > > >>> > >Best, >>> > >David >>> > > >>> > >On Tue, Dec 10, 2024 at 10:46 PM De Gao <d...@live.co.uk> wrote: >>> > > >>> > >> Hi All: >>> > >> >>> > >> There were no discussion in the past week. Just want to double check >>> if >>> > I >>> > >> missed anything? >>> > >> What should be the expectations on KIP discussion? >>> > >> >>> > >> Thank you! >>> > >> >>> > >> De Gao >>> > >> >>> > >> On 1 December 2024 19:36:37 GMT, De Gao <d...@live.co.uk> wrote: >>> > >> >Hi All: >>> > >> > >>> > >> >I would like to start the discussion of KIP-1114 Introducing Chunk in >>> > >> Partition. >>> > >> > >>> > >> > >>> > >> >>> > >>> https://cwiki.apache.org/confluence/display/KAFKA/KIP-1114%3A+Introducing+Chunk+in+Partition >>> > >> >This KIP is complicated so I expect discussion will take longer time. >>> > >> > >>> > >> >Thank you in advance. >>> > >> > >>> > >> >De Gao >>> > >> >>> > >>> >>> >>> -- >>> David Arthur >>>