Re: partition reassignment stuck

2015-04-21 Thread Wesley Chow
There is only one broker that thinks it's the controller right now. The double controller situation happened earlier this morning. Do the other brokers have to be bounced after the controller situation is fixed? I did not do that for all brokers. Wes On Apr 21, 2015 8:25 PM, "Jiangjie Qin" wrot

leader election rate

2015-04-24 Thread Wesley Chow
Looking at the output from the jmx stats from our Kafka cluster, I see a more or less constant leader election rate of around 2.5 from our controller. Is this expected, or does this mean that leaders are shifting around constantly? If they are shifting, how should I go about debugging, and what tr

Re: HDD or SSD or EBS for kafka brokers in Amazon EC2

2015-06-16 Thread Wesley Chow
A call with Amazon confirmed instability for d2 and c4 instances triggered by lots of network activity. They fixed the problem and have since rolled it out. We've been running Kafka with d2's for a little while now and so far so good. Wes On Tue, Jun 2, 2015 at 1:39 PM, Wes Chow wrote: > > We

Re: HDD or SSD or EBS for kafka brokers in Amazon EC2

2015-06-16 Thread Wesley Chow
Should not matter. We're running 12.04. Wes On Jun 16, 2015 12:18 PM, "Henry Cai" wrote: > Does it still matter whether we are using Ubuntu 14 or 12? > > On Tue, Jun 16, 2015 at 8:44 AM, Wesley Chow wrote: > > > > > A call with Amazon confirmed instabil

drop in fetch-follower requests

2015-10-23 Thread Wesley Chow
We deployed a Samza process that effectively doubled our fetch rates in our cluster and noticed a significant drop fetch-follower requests. What is the difference between fetch-consumer and fetch-follower? [image: Inline image 1] We noticed that some topics moved along undisturbed, sending in the

is a topic compressed?

2017-09-20 Thread Wesley Chow
I have a producer configured to snappy compress data sent to a cluster. Is there some way to verify that the data indeed is being compressed? If I peek at the .log files on the broker, I can read some plain text amongst binary. Similarly, tcpdump shows plain text readable data. I do not know if thi

Re: is a topic compressed?

2017-09-26 Thread Wesley Chow
sed with snappy. > > > > Another try would be to change compression to other values (or remove > > compression) and compare the tcp dump with the one you already have. > > > > > > Vincent. > > > > On Wed, Sep 20, 2017 at 11:58 PM, Wesley Chow wrote: >

Re: Hash partition of key with skew

2016-05-03 Thread Wesley Chow
I’ve come up with a couple solutions since we too have a power law distribution. However, we have not put anything into practice. Fixed Slicing One simple thing to do is to take each key and slice it into some fixed number of partitions. So your function might be: (hash(key) % num) + (hash(key

Re: Hash partition of key with skew

2016-05-03 Thread Wesley Chow
Connect with us: Twitter I LinkedIn I Facebook I YouTube > > > -Original Message- > From: Wesley Chow [mailto:w...@chartbeat.com <mailto:w...@chartbeat.com>] > Sent: Tuesday, May 03, 2016 9:51 AM > To: users@kafka.apache.org <mailto:users@kafka.apache.org> > Subject: Re: Ha

Re: Hash partition of key with skew

2016-05-03 Thread Wesley Chow
> > Upload to S3 is partitioned by the "key" field. I.e, one folder per key. It > does offset management to make sure offset commit is in sync with S3 upload. We do this in several spots and I wish we had built our system in such a way that we could just open source it. I’m sure many people have

Re: Hash partition of key with skew

2016-05-04 Thread Wesley Chow
put custom partitioner into production. > Would you mind sharing how you worked around this currently? > > Srikanth > > On Tue, May 3, 2016 at 5:43 PM, Wesley Chow wrote: > >>> >>> Upload to S3 is partitioned by the "key" field. I.e, one folder per ke

compacted log TTLs

2016-05-12 Thread Wesley Chow
Are there any thoughts on supporting TTLs on keys in compacted logs? In other words, some way to set on a per-key basis a time to auto-delete. Wes

Re: compacted log TTLs

2016-05-12 Thread Wesley Chow
period it > will be removed. You could run a process that reprocesses the log and sends > a NULL to keys you want to purge based on some custom logic. > > On Thu, May 12, 2016 at 2:01 PM, Wesley Chow wrote: > >> Are there any thoughts on supporting TTLs on keys in compacted log

Re: compacted log TTLs

2016-05-13 Thread Wesley Chow
? Is your data >> source expiring records and hence you want to auto "delete" the >> corresponding Kafka records as well? >> >> Guozhang >> >>> On Thu, May 12, 2016 at 2:35 PM, Wesley Chow wrote: >>> >>> Right, I’m trying to a

high rate of small packets between brokers

2016-07-05 Thread Wesley Chow
I’ve been investigating some possible network performance issues we’re having with our Kafka brokers, and noticed that traffic sent between brokers tends to show frequent bursts of very small packets: 16:09:52.299863 IP stream02.chartbeat.net.9092 > stream03.chartbeat.net.39399: Flags [P.], seq

compacted log key limits

2016-09-15 Thread Wesley Chow
Is there any guidance on a maximum number different keys in a compacted log? Such total numbers, or "keys need to fit in memory, message data does not", etc. Is it unreasonable to expect tens or hundreds of millions of keys in a single topic to be handled gracefully? Thanks, Wes