Re: JAVA HEAP settings for KAFKA in production

2014-05-05 Thread Todd Palino
I apologize for taking a couple days to jump in on this. We¹re currently running JDK 1.7 u51, and we¹ve switched over to the G1 collector. If you do this (and I highly recommend it), make sure you¹re on u51. We tried out u21 in testing, but we had a number of problems with the GC implementation in

Re: Log Retention in Kafka

2014-05-05 Thread Jun Rao
Dynamically changing configs per topic is only available in 0.8.1. You can try the latest release 0.8.1.1. Thanks, Jun On Mon, May 5, 2014 at 12:28 PM, Kashyap Mhaisekar wrote: > Thanks Jun. I am using kafka_2.8.0-0.8.0-beta1.jar. I dont see this script > kafka-topic.sh in the bin folder. Is t

QOS on Producer Side

2014-05-05 Thread Bhavesh Mistry
We are using Kafka for operation metrics and we do not want to loose any data at all if there is issue with Network or all brokers needs to be rebooted or operation reason down time while all producers are still producing data on front end side. We use async to publish messages and we are using Ka

Re: Review for the new consumer APIs

2014-05-05 Thread Bhavesh Mistry
Hi Neha, How will new Consumer help us with implementing following use case? We have heartbeat as one of topics and all application servers publish metric to this topic. We have to meet near real-time consume SLA (less than 30 seconds). 1) We would like to find out what is latest message per

Re: Log Retention in Kafka

2014-05-05 Thread Kashyap Mhaisekar
Thanks Jun. I am using kafka_2.8.0-0.8.0-beta1.jar. I dont see this script kafka-topic.sh in the bin folder. Is there way to do this in the version mentioned? --Kashyap On Mon, May 5, 2014 at 11:18 AM, Jun Rao wrote: > See http://kafka.apache.org/documentation.html#basic_ops_modify_topic > > T

One Way Kafka (or seperate ports for producers and consumers)

2014-05-05 Thread Rick Richardson
Hey list, I would like to use Kafka as an ingest point for sensitive data in a large production network. We would like to set it up in such a way that anyone that wants can publish data through the brokers, but we want to restrict access for consumers to just a small subset of machines. I'm sure

using Maven to build Java clients for Kafka, looking for POM

2014-05-05 Thread David Novogrodsky
All, Does anyone have a POM to build Java consumers and producers for Kafka? Is there an archetype for Maven? David Novogrodsky david.novogrod...@gmail.com http://www.linkedin.com/in/davidnovogrodsky

Re: 0.7.1 Will there be any log info under poor network which packet may lose?

2014-05-05 Thread Jay Kreps
TCP will attempt to resend until the packets are successfully delivered or a timeout occurs. So a packet loss should not lead to an error. The error you see in your log is the queue of unsent events in the client backing up to the point it hits its configured memory limit and then dropping events w

0.7.1 Will there be any log info under poor network which packet may lose?

2014-05-05 Thread Mingmin Liu
Our network packet loss between the producer and kafka broker can reach 2~5% when the traffic load is high, but in producer's log file, we see no related error logs. Is this expected or there should be some warnings or errors in the log file? PS. we do see some "Event queue is full of unsent mess

Re: Log Retention in Kafka

2014-05-05 Thread Jun Rao
See http://kafka.apache.org/documentation.html#basic_ops_modify_topic Thanks, Jun On Sun, May 4, 2014 at 10:11 PM, Kashyap Mhaisekar wrote: > Is there a way to do this at runtime using some available scripts in > kafka/bin? If so, any pointers on which script? > > Regards, > Kashyap > > > On T

Re: log.retention.size

2014-05-05 Thread Jun Rao
Yes, your understanding is correct. A global knob that controls aggregate log size may make sense. What would be the expected behavior when that limit is reached? Would you reduce the retention uniformly across all topics? Then, it just means that some of the logs may not be retained as long as you

Re: Consumer's Group.id and autooffset.reset

2014-05-05 Thread Guozhang Wang
Hi Chris, The reason that it does not read from the beginning after the first time is probably due to offset commit. Could you check if you set the config auto.commit.enable to true? Setting it to false would solve your problem that you do not need to change to a new group.id everytime. Guozhan

Consumer's Group.id and autooffset.reset

2014-05-05 Thread Chris Helck
Would some one explain the interplay between the group.id and autooffset.reset config parameters? I wrote a simple Consumer that had group.id = "XYZZY" autooffset.reset = smallest My intention was to write something similar to Kafka's ConsoleConsumer; that each time my Consumer ran, it