Hi Artem,

Thanks for your explanation and update to the KIP.
Some comments:

5. In the description for `enable.adaptive.partitioning`, the `false` case,
you said:
> the producer will try to distribute messages uniformly.
I think we should describe the possible skewing distribution. Otherwise,
user might be confused about why adaptive partitioning is important.

6. In the description for `partition.availability.timeout.ms`, I think we
should mention in the last sentence about if `enable.adaptive.partitioning`
is disabled this logic is also disabled.

7. Similar thoughts as Ismael, I think we should have a POC and test to
prove that this adaptive partitioning algorithm can have better uniform
partitioning, compared with original sticky one.

Thank you.
Luke

On Fri, Mar 4, 2022 at 9:22 PM Ismael Juma <ism...@juma.me.uk> wrote:

> Regarding `3`, we should only deprecate it if we're sure the new approach
> handles all cases better. Are we confident about that for both of the
> previous partitioners?
>
> Ismael
>
> On Fri, Mar 4, 2022 at 1:37 AM David Jacot <dja...@confluent.io.invalid>
> wrote:
>
> > Hi Artem,
> >
> > Thanks for the KIP! I have a few comments:
> >
> > 1. In the preamble of the proposed change section, there is still a
> > mention of the
> > -1 approach. My understanding is that we have moved away from it now.
> >
> > 2. I am a bit concerned by the trick suggested about the
> > DefaultPartitioner and
> > the UniformStickyPartitioner. I do agree that implementing the logic in
> the
> > producer itself is a good thing. However, it is weird from a user
> > perspective
> > that he can set a class as partitioner that is not used in the end. I
> > think that
> > this will be confusing for our users. Have we considered changing the
> > default
> > value of partitioner.class to null to indicate that the new built-in
> > partitioner
> > must be used? By default, the built-in partitioner would be used unless
> the
> > user explicitly specify one. The downside is that the new default
> behavior
> > would not work if the user explicitly specify the partitioner but we
> could
> > mitigate this with my next point.
> >
> > 3. Related to the previous point, I think that we could deprecate both
> the
> > DefaultPartitioner and the UniformStickyPartitioner. I would also add a
> > warning if one of them is explicitly provided by the user to inform them
> > that they should switch to the new built-in one. I am pretty sure that
> most
> > of the folks use the default configuration anyway.
> >
> > 4. It would be great if we could explain why the -1 way was rejected. At
> > the moment, the rejected alternative only explain the idea but does not
> > say why we rejected it.
> >
> > Best,
> > David
> >
> > On Fri, Mar 4, 2022 at 6:03 AM Artem Livshits
> > <alivsh...@confluent.io.invalid> wrote:
> > >
> > > Hi Jun,
> > >
> > > 2. Removed the option from the KIP.  Now the sticky partitioning
> > threshold
> > > is hardcoded to batch.size.
> > >
> > > 20. Added the corresponding wording to the KIP.
> > >
> > > -Artem
> > >
> > > On Thu, Mar 3, 2022 at 10:52 AM Jun Rao <j...@confluent.io.invalid>
> > wrote:
> > >
> > > > Hi, Artem,
> > > >
> > > > Thanks for the reply.
> > > >
> > > > 1. Sounds good.
> > > >
> > > > 2. If we don't expect users to change it, we probably could just
> leave
> > out
> > > > the new config. In general, it's easy to add a new config, but hard
> to
> > > > remove an existing config.
> > > >
> > > > 20. The two new configs enable.adaptive.partitioning and
> > > > partition.availability.timeout.ms only apply to the two built-in
> > > > partitioners DefaultPartitioner and UniformStickyPartitioner, right?
> It
> > > > would be useful to document that in the KIP.
> > > >
> > > > Thanks,
> > > >
> > > > Jun
> > > >
> > > > On Thu, Mar 3, 2022 at 9:47 AM Artem Livshits
> > > > <alivsh...@confluent.io.invalid> wrote:
> > > >
> > > > > Hi Jun,
> > > > >
> > > > > Thank you for the suggestions.
> > > > >
> > > > > 1. As we discussed offline, we can hardcode the logic for
> > > > > DefaultPartitioner and UniformStickyPartitioner in the
> KafkaProducer
> > > > (i.e.
> > > > > the DefaultPartitioner.partition won't get called, instead
> > KafkaProducer
> > > > > would check if the partitioner is an instance of DefaultPartitioner
> > and
> > > > > then run the actual partitioning logic itself).  Then the change to
> > the
> > > > > Partitioner wouldn't be required.  I'll update the KIP to reflect
> > that.
> > > > >
> > > > > 2. I don't expect users to change this too often, as changing it
> > would
> > > > > require a bit of studying of the production patterns.  As a general
> > > > > principle, if I can think of a model that requires a deviation from
> > > > > default, I tend to add a configuration option.  It could be that
> > it'll
> > > > > never get used in practice, but I cannot prove that.  I'm ok with
> > > > removing
> > > > > the option, let me know what you think.
> > > > >
> > > > > -Artem
> > > > >
> > > > > On Mon, Feb 28, 2022 at 2:06 PM Jun Rao <j...@confluent.io.invalid>
> > > > wrote:
> > > > >
> > > > > > Hi, Artem,
> > > > > >
> > > > > > Thanks for the reply. A few more comments.
> > > > > >
> > > > > > 1. Since we control the implementation and the usage of
> > > > > DefaultPartitioner,
> > > > > > another way is to instantiate the DefaultPartitioner with a
> special
> > > > > > constructor, which allows it to have more access to internal
> > > > information.
> > > > > > Then we could just change the behavior of  DefaultPartitioner
> such
> > that
> > > > > it
> > > > > > can use the internal infoamtion when choosing the partition. This
> > seems
> > > > > > more intuitive than having DefaultPartitioner return -1
> partition.
> > > > > >
> > > > > > 2. I guess partitioner.sticky.batch.size is introduced because
> the
> > > > > > effective batch size could be less than batch.size and we want to
> > align
> > > > > > partition switching with the effective batch size. How would a
> user
> > > > know
> > > > > > the effective batch size to set partitioner.sticky.batch.size
> > properly?
> > > > > If
> > > > > > the user somehow knows the effective batch size, does setting
> > > > batch.size
> > > > > to
> > > > > > the effective batch size achieve the same result?
> > > > > >
> > > > > > 4. Thanks for the explanation. Makes sense to me.
> > > > > >
> > > > > > Thanks,
> > > > > >
> > > > > > Jun
> > > > > >
> > > > > > Thanks,
> > > > > >
> > > > > > Jun
> > > > > >
> > > > > > On Fri, Feb 25, 2022 at 8:26 PM Artem Livshits
> > > > > > <alivsh...@confluent.io.invalid> wrote:
> > > > > >
> > > > > > > Hi Jun,
> > > > > > >
> > > > > > > 1. Updated the KIP to add a couple paragraphs about
> > implementation
> > > > > > > necessities in the Proposed Changes section.
> > > > > > >
> > > > > > > 2. Sorry if my reply was confusing, what I meant to say (and I
> > > > > elaborated
> > > > > > > on that in point #3) is that there could be patterns for which
> > 16KB
> > > > > > > wouldn't be the most effective setting, thus it would be good
> to
> > make
> > > > > it
> > > > > > > configurable.
> > > > > > >
> > > > > > > 4. We could use broker readiness timeout.  But I'm not sure it
> > would
> > > > > > > correctly model the broker load.  The problem is that latency
> is
> > not
> > > > an
> > > > > > > accurate measure of throughput, we could have 2 brokers that
> have
> > > > equal
> > > > > > > throughput but one has higher latency (so it takes larger
> batches
> > > > less
> > > > > > > frequently, but still takes the same load).  Latency-based
> logic
> > is
> > > > > > likely
> > > > > > > to send less data to the broker with higher latency.  Using the
> > queue
> > > > > > size
> > > > > > > would adapt to throughput, regardless of latency (which could
> be
> > > > just a
> > > > > > > result of deployment topology), so that's the model chosen in
> the
> > > > > > > proposal.  The partition.availability.timeout.ms logic
> > approaches
> > > > the
> > > > > > > model
> > > > > > > from a slightly different angle, say we have a requirement to
> > deliver
> > > > > > > messages via brokers that have a certain latency, then
> > > > > > > partition.availability.timeout.ms could be used to tune that.
> > > > Latency
> > > > > > is
> > > > > > > a
> > > > > > > much more volatile metric than throughput (latency depends on
> > > > external
> > > > > > > load, on capacity, on deployment topology, on jitter in
> network,
> > on
> > > > > > jitter
> > > > > > > in disk, etc.) and I think it would be best to leave
> > latency-based
> > > > > > > thresholds configurable to tune for the environment.
> > > > > > >
> > > > > > > -Artem
> > > > > > >
> > > > > > > On Wed, Feb 23, 2022 at 11:14 AM Jun Rao
> > <j...@confluent.io.invalid>
> > > > > > wrote:
> > > > > > >
> > > > > > > > Hi, Artem,
> > > > > > > >
> > > > > > > > Thanks for the reply. A few more comments.
> > > > > > > >
> > > > > > > > 1. Perhaps you could elaborate a bit more on how the producer
> > > > > > determines
> > > > > > > > the partition if the partitioner returns -1. This will help
> > > > > understand
> > > > > > > why
> > > > > > > > encapsulating that logic as a partitioner is not clean.
> > > > > > > >
> > > > > > > > 2. I am not sure that I understand this part. If 15.5KB is
> more
> > > > > > > efficient,
> > > > > > > > could we just set batch.size to 15.5KB?
> > > > > > > >
> > > > > > > > 4. Yes, we could add a switch (or a variant of the
> > partitioner) for
> > > > > > > > enabling this behavior. Also, choosing partitions based on
> > broker
> > > > > > > readiness
> > > > > > > > can be made in a smoother way. For example, we could track
> the
> > last
> > > > > > time
> > > > > > > a
> > > > > > > > broker has drained any batches from the accumulator. We can
> > then
> > > > > select
> > > > > > > > partitions from brokers proportionally to the inverse of that
> > time.
> > > > > > This
> > > > > > > > seems smoother than a cutoff based on a
> > > > > > > partition.availability.timeout.ms
> > > > > > > >  threshold.
> > > > > > > >
> > > > > > > > Thanks,
> > > > > > > >
> > > > > > > > Jun
> > > > > > > >
> > > > > > > > On Fri, Feb 18, 2022 at 5:14 PM Artem Livshits
> > > > > > > > <alivsh...@confluent.io.invalid> wrote:
> > > > > > > >
> > > > > > > > > Hello Luke, Jun,
> > > > > > > > >
> > > > > > > > > Thank you for your feedback.  I've added the Rejected
> > Alternative
> > > > > > > section
> > > > > > > > > that may clarify some of the questions w.r.t. returning -1.
> > > > > > > > >
> > > > > > > > > 1. I've elaborated on the -1 in the KIP.  The problem is
> > that a
> > > > > > > > significant
> > > > > > > > > part of the logic needs to be in the producer (because it
> now
> > > > uses
> > > > > > > > > information about brokers that only the producer knows), so
> > > > > > > encapsulation
> > > > > > > > > of the logic within the default partitioner isn't as clean.
> > > >  I've
> > > > > > > added
> > > > > > > > > the Rejected Alternative section that documents an attempt
> to
> > > > keep
> > > > > > the
> > > > > > > > > encapsulation by providing new callbacks to the
> partitioner.
> > > > > > > > >
> > > > > > > > > 2. The meaning of the partitioner.sticky.batch.size is
> > explained
> > > > in
> > > > > > the
> > > > > > > > > Uniform Sticky Batch Size section.  Basically, we track the
> > > > amount
> > > > > of
> > > > > > > > bytes
> > > > > > > > > produced to the partition and if it exceeds
> > > > > > > partitioner.sticky.batch.size
> > > > > > > > > then we switch to the next partition.  As far as the reason
> > to
> > > > make
> > > > > > it
> > > > > > > > > different from batch.size, I think Luke answered this with
> > the
> > > > > > question
> > > > > > > > #3
> > > > > > > > > -- what if the load pattern is such that 15.5KB would be
> more
> > > > > > efficient
> > > > > > > > > than 16KB?
> > > > > > > > >
> > > > > > > > > 3. I think it's hard to have one size that would fit all
> > > > patterns.
> > > > > > > E.g.
> > > > > > > > if
> > > > > > > > > the load pattern is such that there is linger and the app
> > fills
> > > > the
> > > > > > > batch
> > > > > > > > > before linger expires, then having 16KB would most likely
> > > > > synchronize
> > > > > > > > > batching and partition switching, so each partition would
> > get a
> > > > > full
> > > > > > > > > batch.  If load pattern is such that there are a few
> > non-complete
> > > > > > > batches
> > > > > > > > > go out before a larger batch starts to fill, then it may
> > actually
> > > > > be
> > > > > > > > > beneficial to make slightly larger (e.g. linger=0, first
> few
> > > > > records
> > > > > > go
> > > > > > > > in
> > > > > > > > > the first batch, then next few records go to second batch,
> > and so
> > > > > on,
> > > > > > > > until
> > > > > > > > > 5 in-flight, then larger batch would form while waiting for
> > > > broker
> > > > > to
> > > > > > > > > respond, but the partition switch would happen before the
> > larger
> > > > > > batch
> > > > > > > is
> > > > > > > > > full).
> > > > > > > > >
> > > > > > > > > 4. There are a couple of reasons for introducing
> > > > > > > > > partition.availability.timeout.ms.  Luke's an Jun's
> > questions
> > > > are
> > > > > > > > slightly
> > > > > > > > > different, so I'm going to separate replies.
> > > > > > > > > (Luke) Is the queue size a good enough signal?  I think
> it's
> > a
> > > > good
> > > > > > > > default
> > > > > > > > > signal as it tries to preserve general fairness and not
> > overreact
> > > > > on
> > > > > > > the
> > > > > > > > > broker's state at each moment in time.  But because it's
> > smooth,
> > > > it
> > > > > > may
> > > > > > > > not
> > > > > > > > > be reactive enough to instantaneous latency jumps.  For
> > > > > > > latency-sensitive
> > > > > > > > > workloads, it may be desirable to react faster when a
> broker
> > > > > becomes
> > > > > > > > > unresponsive (but that may make the distribution really
> > choppy),
> > > > so
> > > > > > > > > partition.availability.timeout.ms provides an opportunity
> to
> > > > tune
> > > > > > > > > adaptiveness.
> > > > > > > > >
> > > > > > > > > (Jun) Can we just not assign partitions to brokers that are
> > not
> > > > > > ready?
> > > > > > > > > Switching partitions purely based on current broker
> readiness
> > > > > > > information
> > > > > > > > > can really skew workload I think (or at least I couldn't
> > build a
> > > > > > model
> > > > > > > > that
> > > > > > > > > proves that over time it's going to be generally fair), I
> > feel
> > > > that
> > > > > > the
> > > > > > > > > algorithm should try to be fair in general and use smoother
> > > > signals
> > > > > > by
> > > > > > > > > default (e.g. a broker with choppier latency may get much
> > less
> > > > load
> > > > > > > even
> > > > > > > > > though it can handle throughput, it then may potentially
> skew
> > > > > > > > consumption),
> > > > > > > > > note that the queue-size-based logic uses probabilities (so
> > we
> > > > > don't
> > > > > > > > fully
> > > > > > > > > remove brokers, just make it less likely) and relative info
> > > > rather
> > > > > > > than a
> > > > > > > > > threshold (so if all brokers are heavily, but equally
> loaded,
> > > > they
> > > > > > will
> > > > > > > > get
> > > > > > > > > equal distribution, rather than get removed because they
> > exceed
> > > > > some
> > > > > > > > > threshold).  So at the very least, I would like this logic
> > to be
> > > > > > turned
> > > > > > > > off
> > > > > > > > > by default as it's hard to predict what it could do with
> > > > different
> > > > > > > > patterns
> > > > > > > > > (which means that there would need to be some
> > configuration).  We
> > > > > > could
> > > > > > > > > just not use brokers that are not ready, but again, I think
> > that
> > > > > it's
> > > > > > > > good
> > > > > > > > > to try to be fair under normal circumstances, so if
> normally
> > > > > brokers
> > > > > > > can
> > > > > > > > > respond under some partition.availability.timeout.ms
> > threshold
> > > > and
> > > > > > the
> > > > > > > > > application works well with those latencies, then we could
> > > > > distribute
> > > > > > > > data
> > > > > > > > > equally between brokers that don't exceed the latencies.
> The
> > > > > value,
> > > > > > of
> > > > > > > > > course, would depend on the environment and app
> requirements,
> > > > hence
> > > > > > > it's
> > > > > > > > > configurable.
> > > > > > > > >
> > > > > > > > > 10. Added a statement at the beginning of the proposed
> > changes.
> > > > > > > > >
> > > > > > > > > -Artem
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > On Thu, Feb 17, 2022 at 3:46 PM Jun Rao
> > <j...@confluent.io.invalid
> > > > >
> > > > > > > > wrote:
> > > > > > > > >
> > > > > > > > > > Hi, Artem,
> > > > > > > > > >
> > > > > > > > > > Thanks for the KIP. A few comments below.
> > > > > > > > > >
> > > > > > > > > > 1. I agree with Luke that having the partitioner
> returning
> > -1
> > > > is
> > > > > > kind
> > > > > > > > of
> > > > > > > > > > weird. Could we just change the implementation of
> > > > > > DefaultPartitioner
> > > > > > > to
> > > > > > > > > the
> > > > > > > > > > new behavior?
> > > > > > > > > >
> > > > > > > > > > 2. partitioner.sticky.batch.size: Similar question to
> > Luke. I
> > > > am
> > > > > > not
> > > > > > > > sure
> > > > > > > > > > why we want to introduce this new configuration. Could we
> > just
> > > > > use
> > > > > > > the
> > > > > > > > > > existing batch.size?
> > > > > > > > > >
> > > > > > > > > > 4. I also agree with Luke that it's not clear why we need
> > > > > > > > > > partition.availability.timeout.ms. The KIP says the
> broker
> > > > > "would
> > > > > > > not
> > > > > > > > be
> > > > > > > > > > chosen until the broker is able to accept the next ready
> > batch
> > > > > from
> > > > > > > the
> > > > > > > > > > partition". If we are keeping track of this, could we
> just
> > > > avoid
> > > > > > > > > assigning
> > > > > > > > > > records to partitions whose leader is not able to accept
> > the
> > > > next
> > > > > > > > batch?
> > > > > > > > > If
> > > > > > > > > > we do that, perhaps we don't need
> > > > > > partition.availability.timeout.ms.
> > > > > > > > > >
> > > > > > > > > > 10. Currently, partitioner.class defaults to
> > > > DefaultPartitioner,
> > > > > > > which
> > > > > > > > > uses
> > > > > > > > > > StickyPartitioner when the key is specified. Since this
> KIP
> > > > > > improves
> > > > > > > > upon
> > > > > > > > > > StickyPartitioner, it would be useful to make the new
> > behavior
> > > > > the
> > > > > > > > > default
> > > > > > > > > > and document that in the KIP.
> > > > > > > > > >
> > > > > > > > > > Thanks,
> > > > > > > > > >
> > > > > > > > > > Jun
> > > > > > > > > >
> > > > > > > > > >
> > > > > > > > > > On Wed, Feb 16, 2022 at 7:30 PM Luke Chen <
> > show...@gmail.com>
> > > > > > wrote:
> > > > > > > > > >
> > > > > > > > > > > Hi Artem,
> > > > > > > > > > >
> > > > > > > > > > > Also, one more thing I think you need to know.
> > > > > > > > > > > As this bug KAFKA-7572 <
> > > > > > > > > https://issues.apache.org/jira/browse/KAFKA-7572
> > > > > > > > > > >
> > > > > > > > > > > mentioned, sometimes the custom partitioner would
> return
> > > > > negative
> > > > > > > > > > partition
> > > > > > > > > > > id accidentally.
> > > > > > > > > > > If it returned -1, how could you know if it is expected
> > or
> > > > not
> > > > > > > > > expected?
> > > > > > > > > > >
> > > > > > > > > > > Thanks.
> > > > > > > > > > > Luke
> > > > > > > > > > >
> > > > > > > > > > > On Wed, Feb 16, 2022 at 3:28 PM Luke Chen <
> > show...@gmail.com
> > > > >
> > > > > > > wrote:
> > > > > > > > > > >
> > > > > > > > > > > > Hi Artem,
> > > > > > > > > > > >
> > > > > > > > > > > > Thanks for the update. I have some questions about
> it:
> > > > > > > > > > > >
> > > > > > > > > > > > 1. Could you explain why you need the `partitioner`
> > return
> > > > > -1?
> > > > > > In
> > > > > > > > > which
> > > > > > > > > > > > case we need it? And how it is used in your KIP?
> > > > > > > > > > > > 2. What does the "partitioner.sticky.batch.size"
> mean?
> > In
> > > > the
> > > > > > > > > > > > "Configuration" part, you didn't explain it. And
> > default to
> > > > > 0,
> > > > > > I
> > > > > > > > > guess
> > > > > > > > > > > it's
> > > > > > > > > > > > the same as current behavior for backward
> > compatibility,
> > > > > right?
> > > > > > > You
> > > > > > > > > > > should
> > > > > > > > > > > > mention it.
> > > > > > > > > > > > 3. I'm thinking we can have a threshold to the
> > > > > > > > > > > > "partitioner.sticky.batch.size". Let's say, we
> already
> > > > > > accumulate
> > > > > > > > > > 15.5KB
> > > > > > > > > > > in
> > > > > > > > > > > > partition1, and sent. So when next batch created, in
> > your
> > > > > > current
> > > > > > > > > > design,
> > > > > > > > > > > > we still stick to partition1, until 16KB reached, and
> > then
> > > > we
> > > > > > > > create
> > > > > > > > > a
> > > > > > > > > > > new
> > > > > > > > > > > > batch to change to next partition, ex: partition2.
> But
> > I
> > > > > think
> > > > > > if
> > > > > > > > we
> > > > > > > > > > set
> > > > > > > > > > > a
> > > > > > > > > > > > threshold to 95% (for example), we can know previous
> > 15.5KB
> > > > > > > already
> > > > > > > > > > > exceeds
> > > > > > > > > > > > the threshold so that we can directly create new
> batch
> > for
> > > > > next
> > > > > > > > > > records.
> > > > > > > > > > > > This way should be able to make it more efficient.
> > WDYT?
> > > > > > > > > > > > 4. I think the improved queuing logic should be good
> > > > enough.
> > > > > I
> > > > > > > > can't
> > > > > > > > > > get
> > > > > > > > > > > > the benefit of having `
> > partition.availability.timeout.ms`
> > > > > > > config.
> > > > > > > > In
> > > > > > > > > > > > short, you want to make the partitioner take the
> broker
> > > > load
> > > > > > into
> > > > > > > > > > > > consideration. We can just improve that in the
> queuing
> > > > logic
> > > > > > (and
> > > > > > > > you
> > > > > > > > > > > > already did it). Why should we add the config? Could
> > you
> > > > use
> > > > > > some
> > > > > > > > > > > examples
> > > > > > > > > > > > to explain why we need it.
> > > > > > > > > > > >
> > > > > > > > > > > > Thank you.
> > > > > > > > > > > > Luke
> > > > > > > > > > > >
> > > > > > > > > > > > On Wed, Feb 16, 2022 at 8:57 AM Artem Livshits
> > > > > > > > > > > > <alivsh...@confluent.io.invalid> wrote:
> > > > > > > > > > > >
> > > > > > > > > > > >> Hello,
> > > > > > > > > > > >>
> > > > > > > > > > > >> Please add your comments about the KIP.  If there
> are
> > no
> > > > > > > > > > considerations,
> > > > > > > > > > > >> I'll put it up for vote in the next few days.
> > > > > > > > > > > >>
> > > > > > > > > > > >> -Artem
> > > > > > > > > > > >>
> > > > > > > > > > > >> On Mon, Feb 7, 2022 at 6:01 PM Artem Livshits <
> > > > > > > > > alivsh...@confluent.io
> > > > > > > > > > >
> > > > > > > > > > > >> wrote:
> > > > > > > > > > > >>
> > > > > > > > > > > >> > Hello,
> > > > > > > > > > > >> >
> > > > > > > > > > > >> > After trying a few prototypes, I've made some
> > changes to
> > > > > the
> > > > > > > > > public
> > > > > > > > > > > >> > interface.  Please see the updated document
> > > > > > > > > > > >> >
> > > > > > > > > > > >>
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> >
> https://cwiki.apache.org/confluence/display/KAFKA/KIP-794%3A+Strictly+Uniform+Sticky+Partitioner
> > > > > > > > > > > >> > .
> > > > > > > > > > > >> >
> > > > > > > > > > > >> > -Artem
> > > > > > > > > > > >> >
> > > > > > > > > > > >> > On Thu, Nov 4, 2021 at 10:37 AM Artem Livshits <
> > > > > > > > > > > alivsh...@confluent.io>
> > > > > > > > > > > >> > wrote:
> > > > > > > > > > > >> >
> > > > > > > > > > > >> >> Hello,
> > > > > > > > > > > >> >>
> > > > > > > > > > > >> >> This is the discussion thread for
> > > > > > > > > > > >> >>
> > > > > > > > > > > >>
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> >
> https://cwiki.apache.org/confluence/display/KAFKA/KIP-794%3A+Strictly+Uniform+Sticky+Partitioner
> > > > > > > > > > > >> >> .
> > > > > > > > > > > >> >>
> > > > > > > > > > > >> >> The proposal is a bug fix for
> > > > > > > > > > > >> >>
> https://issues.apache.org/jira/browse/KAFKA-10888,
> > but
> > > > > it
> > > > > > > does
> > > > > > > > > > > >> include a
> > > > > > > > > > > >> >> client config change, therefore we have a KIP to
> > > > discuss.
> > > > > > > > > > > >> >>
> > > > > > > > > > > >> >> -Artem
> > > > > > > > > > > >> >>
> > > > > > > > > > > >> >
> > > > > > > > > > > >>
> > > > > > > > > > > >
> > > > > > > > > > >
> > > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > >
> > > > > >
> > > > >
> > > >
> >
>

Reply via email to