Re: InvalidMessageException: Message is corrupt, Consumer stuck

2015-08-04 Thread Gwen Shapira
The high level consumer stores its state in ZooKeeper. Theoretically, you should be able to go into ZooKeeper, find the consumer-group, topic and partition, and increment the offset past the "corrupt" point. On Tue, Aug 4, 2015 at 10:23 PM, Henry Cai wrote: > Hi, > > We are using the Kafka high-

InvalidMessageException: Message is corrupt, Consumer stuck

2015-08-04 Thread Henry Cai
Hi, We are using the Kafka high-level consumer 8.1.1, somehow we got a corrupted message in the topic. We are not sure the root cause of this, but the problem we are having now is the HL consumer is stuck in that position: kafka.message.InvalidMessageException: Message is corrupt (stored crc = 5

Re: kafka.common.FailedToSendMessageException: Failed to send messages after 3 tries.

2015-08-04 Thread David Li
you are right. after I run kafka in my localhost directly, it works just fine after further google, i found that need to set two parameters below if the kafka is running on some other machines #advertised.host.name= #advertised.port= more precisely, if the kafka is running within a docker conta

Kafka Broker server cannot be connected by telnet from other Kafka Brokers

2015-08-04 Thread Qi Xu
Hi Everyone, We're trying the deploy the Kafka behind the network balancer and we have created the port map for each Kafka brokers under that network balancer--we only have one public IP and the Kafka clients are in other system and thus cannot access the brokers via internal IP directly. So for e

Kafka vs RabbitMQ latency

2015-08-04 Thread Yuheng Du
Hi guys, I was reading a paper today in which the latency of kafka and rabbitmq is compared: http://downloads.hindawi.com/journals/js/2015/468047.pdf To my surprise, kafka has shown some large variations of latency as the number of records per second increases. So I am curious about why is that.

Re: Decomissioning a broker

2015-08-04 Thread Grant Henke
Thats correct. Thanks for catching that. On Tue, Aug 4, 2015 at 3:27 PM, Andrew Otto wrote: > Thanks! > > > In fact if you use a "Controlled Shutdown" migrating the replicas and > > leaders should happen for you as well. > > Just to clarify, controlled shutdown will only move the leaders to othe

Re: message filterin or "selector"

2015-08-04 Thread Gwen Shapira
The way Kafka is currently implemented is that Kafka is not aware of the content of messages, so there is no Selector logic available. The way to go is to implement the Selector in your client - i.e. your consume() loop will get all messages but will throw away those that don't fit your pattern.

Re: Consumer that consumes only local partition?

2015-08-04 Thread Hawin Jiang
Hi Robert Here is the kafka benchmark for your reference. if you want to use Flink, Storm, Samza or Spark, the performance will be going down. 821,557 records/sec(78.3 MB/sec) https://engineering.linkedin.com/kafka/benchmarking-apache-kafka-2-million-writes-second-three-cheap-machines Best

Re: Got conflicted ephemeral node exception for several hours

2015-08-04 Thread Jaikiran Pai
I am on Kafka 0.8.2.1 (Java 8) and have happened to run into this same issue where the KafkaServer (broker) goes into a indefinite while loop writing out this message: [2015-08-04 15:45:12,350] INFO conflict in /brokers/ids/0 data: {"jmx_port":-1,"timestamp":"1438661432074","host":"foo-bar",

Re: Access control in kafka

2015-08-04 Thread Parth Brahmbhatt
If this (https://cwiki.apache.org/confluence/display/KAFKA/KIP-11+-+Authorization+I nterface) is what you need then watch for https://reviews.apache.org/r/34492/ to get committed to trunk. Thanks Parth On 8/4/15, 1:57 PM, "Alvaro Gareppe" wrote: >Can someone point me to documentation about ac

message filterin or "selector"

2015-08-04 Thread Alvaro Gareppe
The is way to implement a "selector" logic in kafka (similar to JMS selectors) So, allow to consume a message if only the message contains certain header or content ? I'm evaluating to migrate from ActiveMQ to kafka and I'm using the selector logic widely in the application -- Ing. Alvaro Garep

Access control in kafka

2015-08-04 Thread Alvaro Gareppe
Can someone point me to documentation about access control in kafka. There is something implemented in the current or plan for future versions ? I need something that allows me to define what users are allowed to connect to certain topic, and of course user management. Thank you guys in advance!

Re: Decomissioning a broker

2015-08-04 Thread Andrew Otto
Thanks! > In fact if you use a "Controlled Shutdown" migrating the replicas and > leaders should happen for you as well. Just to clarify, controlled shutdown will only move the leaders to other replicas, right? It won’t actually migrate any replicas elsewhere. -Ao > On Aug 4, 2015, at 13:00,

Re: How to read in batch using HighLevel Consumer?

2015-08-04 Thread Gwen Shapira
To add some internals, the high level consumer actually does read entire batches from Kafka. It just exposes them to the user in an event loop, because its a very natural API. Users can then batch events the way they prefer. So if you are worried about batches being more efficient than single even

Re: How to read in batch using HighLevel Consumer?

2015-08-04 Thread shahab
Thanks a lot Shaminder for clarification and thanks Raja for pointing me to the example. best, /shahab On Tue, Aug 4, 2015 at 6:06 PM, Rajasekar Elango wrote: > Here is an example on what sharninder suggested > > http://ingest.tips/2014/10/12/kafka-high-level-consumer-frequently-missing-pieces/

Re: Consumer that consumes only local partition?

2015-08-04 Thread Robert Metzger
Sorry for the very late reply ... The performance issue was not caused by network latency. I had a job like this: FlinkKafkaConsumer --> someSimpleOperation --> FlinkKafkaProducer. I thought that our FlinkKafkaConsumer is slow, but actually our FlinkKafkaProducer was using the old producer API of

Re: Checkpointing with custom metadata

2015-08-04 Thread Jason Gustafson
I couldn't find a jira for this, so I added KAFKA-2403. -Jason On Tue, Aug 4, 2015 at 9:36 AM, Jay Kreps wrote: > Hey James, > > You are right the intended use of that was to have a way to capture some > very small metadata about your state at the time of offset commit in an > atomic way. > > T

Re: Decomissioning a broker

2015-08-04 Thread Grant Henke
The broker will actually unregister itself from zookeeper. The brokers id path uses ephemeral nodes so they are automatically destroyed on shutdown. In fact if you use a "Controlled Shutdown" migrating the replicas and leaders should happen for you as well. Though, manual reassignment may be prefer

Re: Kafka Zookeeper Issues

2015-08-04 Thread Grant Henke
The /brokers/ids nodes are ephemeral nodes that only exists while the brokers maintain a session to zookeeper. There is more information on Kafka's Zookeeper usage here: - http://kafka.apache.org/documentation.html - look for "Broker Node Registry" - https://cwiki.apache.org/confluence/

Re: Checkpointing with custom metadata

2015-08-04 Thread Jay Kreps
Hey James, You are right the intended use of that was to have a way to capture some very small metadata about your state at the time of offset commit in an atomic way. That field isn't exposed but we do need to add it to the new consumer api (I think just no one has done it yet. -Jay On Mon, Au

Re: Lead Broker from kafka.message.MessageAndMetadata

2015-08-04 Thread Grant Henke
Hi Sreeni, Using the SimpleConsumer you can send a TopicMetadataRequest for a topic and the TopicMetadataResponse will contain TopicMetadata for each topic requested (or all) which contains PartitionMetadata for all all partitions. The PartitionMetadata contains the leader, replicas, and isr. Is

Re: new consumer api?

2015-08-04 Thread Jason Gustafson
Hey Simon, The new consumer has the ability to forego group management and assign partitions directly. Once assigned, you can seek to any offset you want. -Jason On Tue, Aug 4, 2015 at 5:08 AM, Simon Cooper < simon.coo...@featurespace.co.uk> wrote: > Reading on the consumer docs, there's no men

Re: How to read in batch using HighLevel Consumer?

2015-08-04 Thread Rajasekar Elango
Here is an example on what sharninder suggested http://ingest.tips/2014/10/12/kafka-high-level-consumer-frequently-missing-pieces/ Thanks, Raja. On Tue, Aug 4, 2015 at 12:01 PM, Sharninder wrote: > You can't. Kafka is essentially a queue, so you always read messages one > by one. What you can d

Re: How to read in batch using HighLevel Consumer?

2015-08-04 Thread Sharninder
You can't. Kafka is essentially a queue, so you always read messages one by one. What you can do is disable auto offset commit, read 100 messages, process them and then manually commit offset. -- Sharninder > On 04-Aug-2015, at 9:07 pm, shahab wrote: > > Hi, > > While we the producer can pu

How to read in batch using HighLevel Consumer?

2015-08-04 Thread shahab
Hi, While we the producer can put data as batch in kafka server, I couldn't find any API (or any document) saying how we can fetch data as batch from Kafka ? Even when data is placed as batch in kafka server, still using High Level consumer I can only read one by one, and I can not specify. for e

Get last snapshot from compacted topic

2015-08-04 Thread Aki
I'd like to save a snapshot of a processing node's state in a compacted kafka topic. A large number of nodes would save their snapshots in the same partition. What is an efficient way for a (restarted) node to find the offset of its latest snapshot? Using just Kafka (no database, local file, e

RE: new consumer api?

2015-08-04 Thread Simon Cooper
Reading on the consumer docs, there's no mention of a relatively simple consumer that doesn't need groups, coordinators, commits, anything like that - just read and poll from specified offsets of specific topic partitions - but automatically deals with leadership changes and connection losses (s

Re: 0.8.3 ETA?

2015-08-04 Thread Stevo Slavić
Thanks Jun for heads up! On Mon, Aug 3, 2015 at 7:17 PM, Jun Rao wrote: > Hi, Stevo, > > Yes, we are still iterating on the new consumer a bit and are waiting for > some of the security jiras to be committed. So now, we are shooting for > releasing 0.8.3 in Oct (just updated > https://cwiki.apac

Re: kafka.common.FailedToSendMessageException: Failed to send messages after 3 tries.

2015-08-04 Thread Jilin Xie
Some suggestions: Check the existence of the topic. Check the firewall of the broker... Try telnet or something to make sure it's available. Try run the producer on the broker machine. Since you get this error, this code is functioning. I think it's some configuration an

Re: New Consumer API and Range Consumption with Fail-over

2015-08-04 Thread Bhavesh Mistry
Hi Jason and Kafka Dev Team, First of all thanks for responding and I think you got expected behavior correctly. The use-case is offset range consumption. We store each minute highest offset for each topic per partition. So if we need to reload or re-consume data from yesterday per say 8AM