RE: Reg Broker information.

2014-06-26 Thread Balasubramanian Jayaraman
Hi, I used the command (bin/zookeeper-shell.sh 0.0.0.0:2181 get /brokers/ids/0) to get the broker metadata information. How do I get the list of brokers attached to a zookeeper ? Is my question clear ? Thanks Bala -Original Message- From: Neha Narkhede [mailto:neha.narkh...@gmail.com]

Re: Kafka 0.8's VerifyConsumerRebalance reports an error

2014-06-26 Thread Yury Ruchin
I have set log level to DEBUG and saw something strange in the output. For each topic partition, I see the following pattern: [2014-06-26 16:00:24,467] ERROR No owner for partition [,0] (kafka.tools.VerifyConsumerRebalance$) ... [2014-06-26 16:00:24,469] INFO Owner of partition [,0] is _-1403049

Re: Reg Broker information.

2014-06-26 Thread Yury Ruchin
In ZK shell the following command: ls /brokers/ids will give you a list like this: [0, 1, 2, 3, 4] where items are broker ids you can further use to issue "get" request to ZK: get /brokers/ids/ 2014-06-26 12:37 GMT+04:00 Balasubramanian Jayaraman < balasubramanian.jayara...@autodesk.com>:

do apps with producers have to be restarted if cluster goes down and comes back up?

2014-06-26 Thread S Ahmed
Hi, A few questions on timing related issues when certain parts of kafka go down. 1. If zookeeper goes down, then I bring it back online, do I have to restart the brokers? 2. If the brokers go down, producers will be erroring out. When the brokers are back online, do I have to restart the proc

Re: Kafka 0.8's VerifyConsumerRebalance reports an error

2014-06-26 Thread Neha Narkhede
This is a bug in the tool. Please file a bug and attach these error/info logs to it. Thanks, Neha On Thu, Jun 26, 2014 at 5:24 AM, Yury Ruchin wrote: > I have set log level to DEBUG and saw something strange in the output. For > each topic partition, I see the following pattern: > > [2014-06-2

Re: Kafka connection loss with high volume of messages

2014-06-26 Thread Ahmed H.
I dug some more and it seems like before these errors show up, I see a few Zookeeper warnings, followed by Kafka errors. 11:57:26,897 INFO [org.I0Itec.zkclient.ZkClient] (pool-5-thread-1-EventThread) zookeeper state changed (Disconnected) 11:57:26,897 INFO [org.I0Itec.zkclient.ZkClient] (clojure

Re: do apps with producers have to be restarted if cluster goes down and comes back up?

2014-06-26 Thread Neha Narkhede
1. If zookeeper goes down, then I bring it back online, do I have to restart the brokers? You shouldn't have to restart the brokers if the broker side config " zookeeper.connection.timeout.ms" is tuned such that zookeeper is brought back online within the configured timeout. 2. If the brokers g

Re: Blacklisting Brokers

2014-06-26 Thread Neha Narkhede
If ssh fails on the broker machine, I'm not why the broker would be in a state where it maintains it's registration in zookeeper? If so, it will automatically be marked dead and will not get elected as the leader. Have you seen a case where such a broker became the leader? Could you elaborate more

Re: Kafka connection loss with high volume of messages

2014-06-26 Thread Neha Narkhede
11:57:26,897 INFO [org.I0Itec.zkclient.ZkClient] (pool-5-thread-1-EventThread) zookeeper state changed (Disconnected) 11:57:26,897 INFO [org.I0Itec.zkclient.ZkClient] (clojure-agent-send-off-pool- 6-EventThread) zookeeper state changed (Disconnected) I wonder why your consumer disconnects from z

Re: Blacklisting Brokers

2014-06-26 Thread Lung, Paul
Yes I have seen it elected as the leader. The broker process is still up, but due to some unknown hardware failure, it¹s not working 100%. It is, however, able to participate in the cluster, and got itself elected as the leader. I¹m not sure what the hardware problem is though, since I couldn¹t log

Re: Unable to delete topic - kafka 0.8.0

2014-06-26 Thread Guozhang Wang
Hello Virendra, Did you have any producer/consumer clients running during the whole process? Guozhang On Wed, Jun 25, 2014 at 11:53 PM, Virendra Pratap Singh < vpsi...@yahoo-inc.com.invalid> wrote: > I am aware of lack of programmatic way of deleting topics in kafka 0.8.0. > So using the sledg

Re: Unable to delete topic - kafka 0.8.0

2014-06-26 Thread Neha Narkhede
Firstly, I'm not sure this process of deleting topics completely works, especially in corner cases. Having said that, for #3, you should just be deleting /brokers/topics/. If producers are sending data to the Kafka cluster, it will recreate the topics once the brokers come up. Thanks, Neha On We

Re: Blacklisting Brokers

2014-06-26 Thread Neha Narkhede
Could you file a JIRA for this? We may have to think through this carefully. Black listing should be done in a way that maintains the replication factor of the partition. Un blacklisting would then either increase the replication factor or require carefully removing the replicas on that broker. Th

Re: Intercept broker operation in Kafka

2014-06-26 Thread Neha Narkhede
Most of these are available via JMX and others can be available via zookeeper. I'm not sure why/how you would monitor "messages being deleted by the broker". In general, monitoring via JMX is preferable to scraping logs. Thanks, Neha On Mon, Jun 23, 2014 at 11:51 PM, ravi singh wrote: > Primar

Re: Unable to delete topic - kafka 0.8.0

2014-06-26 Thread Virendra Pratap Singh
No producers were active. The only producer in our pipeline is a storm topology and had made it a point that the whole cluster was down and no topology running. What I did was nothing short of doing a fresh kafka cluster setup (barring the fact that I didn¹t wiped the zookeepers as its the same set

Re: Unable to delete topic - kafka 0.8.0

2014-06-26 Thread Virendra Pratap Singh
I am 100% sure nothing was running. However I am not sure of consumers. Would that make any difference? I thought producer message write request could only cause kafka to initiate topic creation, not consumer read. Would consumer read request even succeed if the topic metadata is non-existent in zo

Re: Monitoring Producers at Large Scale

2014-06-26 Thread Bhavesh Mistry
Hi All, Thanks for all your responses. JMX metrics are there and we do pull the metrics, but I would like to capture the logs from Kafka lib as well especially WARN, FATAL and ERROR etc to debug the issue. To do this, we intercept Log4j logging and send it to Kafka Log Topics, but I realize

Question on message content, compression, multiple messages per kafka message?

2014-06-26 Thread Bert Corderman
We are in the process of engineering a system that will be using kafka. The legacy system is using the local file system and a database as the queue. In terms of scale we process about 35 billion events per day contained in 15 million files. I am looking for feedback on a design decision we ar

Scalability question?

2014-06-26 Thread Zack Payton
Hi there, There have been some internal debates here about how far we can scale Kafka. Ideally, we'd be able to make it scale to 90 billion events a day. I've seen somewhere that linked scaled it up to 40 billion events a day. Has anyone seen a hard plateau in terms of scalability? Does anyone

Re: Scalability question?

2014-06-26 Thread Jay Kreps
I think currently we do a little over 200 billion events per day at LinkedIn, though we are not actually the largest Kafka user any more. On the whole scaling the volume of messages is actually not that hard in Kafka. Data is partitioned, and partitions don't really communicate with each other, so

Question about kafka-consumer-perf-test.sh

2014-06-26 Thread Baran Nohutçuoğlu
Hi, I’m having trouble understanding the results from running kafka-consumer-perf-test. For low number of messages, I see very low throughput in terms of messages / second. Here is a table of results: fetch.size data.consumed.in.MB MB.sec data.consumed.in.nMsg nMsg.sec 1048576 0.0003 0 1 0.014

Re: Question on message content, compression, multiple messages per kafka message?

2014-06-26 Thread Neha Narkhede
Using a single Kafka message to contain an application snapshot has the upside of getting atomicity for free. Either the snapshot will be written as a whole to Kafka or not. This is poor man's transactionality. Care needs to be taken to ensure that the message is not too large since that might caus

RE: Experiences with larger message sizes

2014-06-26 Thread Luke Forehand
I have used 50MB message size and it is not a great idea. First of all you need to make sure you have these settings in sync: message.max.bytes replica.fetch.max.bytes fetch.message.max.bytes I had not set the replica fetch setting and didn't realize one of my partitions was not replicating aft

Re: Experiences with larger message sizes

2014-06-26 Thread Bert Corderman
Thanks for the details Luke. At what point would you consider a message too big? Are you using compression? Bert On Thursday, June 26, 2014, Luke Forehand < luke.foreh...@networkedinsights.com> wrote: > I have used 50MB message size and it is not a great idea. First of all > you need to make

Re: Question on message content, compression, multiple messages per kafka message?

2014-06-26 Thread Bert Corderman
What would you consider being a message that is “too large” In April I ran a bunch of tests which I outlined in the following thread http://grokbase.com/t/kafka/users/145g8k62rf/performance-testing-data-to-share It includes a google doc link with all the results (its easiest to download in

Apache Kafka NYC Users Group!

2014-06-26 Thread Joe Stein
Hi folks, I just started a new Meetup specifically for Apache Kafka in NYC (everyone is welcome of course) http://www.meetup.com/Apache-Kafka-NYC/ For the last couple of years we have been piggy backing talks and the community with other NYC meetup groups (Storm, Cassandra, Hadoop, etc) and figure

message stuck, possible problem setting fetch.message.max.bytes

2014-06-26 Thread Louis Clark
Hi, I'm trying to stream large message with Kafka into Spark. Generally this has been working nicely, but I found one message (5.1MB in size) which is clogging my pipeline up. I have these settings in server.properties: fetch.message.max.bytes=10485760 replica.fetch.max.bytes=10485760 message.max

Re: message stuck, possible problem setting fetch.message.max.bytes

2014-06-26 Thread Guozhang Wang
Hi Louis, What are your consumer's config properties? Guozhang On Thu, Jun 26, 2014 at 5:54 PM, Louis Clark wrote: > Hi, I'm trying to stream large message with Kafka into Spark. Generally > this has been working nicely, but I found one message (5.1MB in size) which > is clogging my pipeline

Re: message stuck, possible problem setting fetch.message.max.bytes

2014-06-26 Thread Louis Clark
in the consumer.properties file, I've got (default?): zookeeper.connect=127.0.0.1:2181 zookeeper.connection.timeout.ms=100 group.id=test-consumer-group thanks, -Louis On Thu, Jun 26, 2014 at 6:04 PM, Guozhang Wang wrote: > Hi Louis, > > What are your consumer's config properties? > > G

Re: Experiences with larger message sizes

2014-06-26 Thread Denny Lee
Yes,  thanks very much Luke - this is very helpful for my plans.  I was under the same impression but it’s always good to have verification, eh?!  On June 26, 2014 at 4:48:03 PM, Bert Corderman (bertc...@gmail.com) wrote: Thanks for the details Luke. At what point would you consider a messag