Re: Kafka replication error

2018-12-09 Thread Tony Liu
yes, Kafka provides those metrics for monitoring the pending requests and idle request handler, check their documents, you should get it, otherwise, let me know. from the operations you did: 1. Restarting service is not help should be expected, since that broker is already being a bottlenec

Re: Kafka replication error

2018-12-09 Thread Suman B N
Thanks Tony. How can I check the number of pending requests and the idle percentage of the requests handler on node 6? Are there any metrics for those? Controller logs confirm that the controller is not able to send updateMetadataRequest() to that particular node acting as the follower. Whereas t

Re: Kafka replication error

2018-12-08 Thread Tony Liu
you mentioned: 1. broker disconnection error, normally the case I have ever seen is when some broker is busy and can not response connection quickly to other replicas. 2. partitions under-replicated, normally pint to some broker may have a performance issue. 3. 90% under-replicat

Re: Kafka replication error

2018-12-08 Thread Suman B N
Still hoping for some help here. On Fri, Dec 7, 2018 at 12:24 AM Suman B N wrote: > Guys, > Another observation is 90% of under-replicated partitions have the same > node as the follower. > > *Any help in here is very much appreciated. We have very less time to > stabilize kafka. Thanks a lot in

Re: Kafka replication error

2018-12-06 Thread Suman B N
Guys, Another observation is 90% of under-replicated partitions have the same node as the follower. *Any help in here is very much appreciated. We have very less time to stabilize kafka. Thanks a lot in advance.* -Suman On Thu, Dec 6, 2018 at 9:08 PM Suman B N wrote: > +users > > On Thu, Dec 6

Re: Kafka replication error

2018-12-06 Thread Suman B N
+users On Thu, Dec 6, 2018 at 9:01 PM Suman B N wrote: > Team, > > We are observing ISR shrink and expand very frequently. In the logs of the > follower, below errors are observed: > > [2018-12-06 20:00:42,709] WARN [ReplicaFetcherThread-2-15], Error in fetch > kafka.server.ReplicaFetcherThread$

Re: Kafka replication

2018-03-26 Thread Christophe Schmitz
Hi Romain, Zookeeper needs to achieve quorum to function. This means that a strict majority of nodes need to be alive. If your box1 dies, you lose 2 zookeeper services, with only one left. Your system is already offline. So you would need to spread your zookeeper services across 3 physical nodes.

Re: Kafka Replication Factor

2018-01-17 Thread Brett Rann
if RF=2 and min.insync.replicas=1 (the default) then you shouldn't have offline partitions if 1 of 3 brokers is down. I'd first double check your topic config (and broker defaults) for the one that went offline to verify RF/Min. Be sure to check each partition they can be different! ( kafka-topics

Re: Kafka Replication Factor

2018-01-17 Thread Svante Karlsson
whats your config for min.insync.replicas? 2018-01-17 13:37 GMT+01:00 Sameer Kumar : > Hi, > > I have a cluster of 3 Kafka brokers, and replication factor is 2. This > means I can tolerate failure of 1 node without data loss. > > Recently, one of my node crashed and some of my partitions went off

Re: Kafka Replication modes

2017-08-15 Thread Sameer Kumar
Got some info on this, I think this is mostly controlled by acks. The default value of “1” requires an explicit acknowledgement from the partition leader that the write succeeded. The strongest guarantee that Kafka provides is with “acks=all”, which guarantees that not only did the partition leade

Re: Kafka Replication modes

2017-08-10 Thread Sameer Kumar
please refer to the following doc. https://cwiki.apache.org/confluence/display/KAFKA/Kafka+Replication It talks about both the modes, I am not sure of which one is followed by Kafka. Also, is there a property to switch between the two. -Sameer. On Wed, Aug 9, 2017 at 3:24 PM, Manikumar wrote:

Re: Kafka Replication modes

2017-08-09 Thread Manikumar
Not sure about, what you mean by Asynchronous and Synchronous replication. details about replication are here: http://kafka.apache.org/documentation/#replication Kafka producers can choose whether they wait for the message to be acknowledged by 0,1 or all (-1) replicas by using "acks" config prope

Re: Kafka Replication

2016-12-27 Thread hans
Yes. Either using a Change Data Capture product like Oracle GoldenGate Connector for Kafka or JDBC Source & Sink Kafka Connectors like the one included with Confluent Open Source. -hans > On Dec 27, 2016, at 11:47 AM, Julious.Campbell > wrote: > > > Support > > Can Kafka handle data repl

Re: Kafka Replication Factor

2016-07-24 Thread Stevo Slavić
Hello Avi, Yes, replication factor of 1 means just one lead replica gets assigned for the topic partitions. To achieve HA of messages there are more things to watch out, not only replication factor - metadata/ZK HA, producer acks, min-in-sync replicas, rack awareness for replica assignment to nam

Re: Kafka replication factor mode

2014-12-08 Thread Jun Rao
Kafka tolerates 2f failures with 2f+1 replicas by default. What error are you seeing? Thanks, Jun On Fri, Dec 5, 2014 at 10:01 AM, Ankit Jain wrote: > Hi All, > > I have two nodes kafka cluster and using replication factor 2. If both the > node is running, then i am able to push data, but if a

Re: Kafka replication throttling

2014-05-21 Thread Marcos Juarez Lopez
Thanks for your response Jun. JIRA has been filed (see link below). Please let me know if I should add more details/context: https://issues.apache.org/jira/browse/KAFKA-1464 Thanks, Marcos Juarez On Wed, May 21, 2014 at 8:40 AM, Jun Rao wrote: > We don't have such throttling right now. Co

Re: Kafka replication throttling

2014-05-21 Thread Jun Rao
We don't have such throttling right now. Could you file a jira? Thanks, Jun On Tue, May 20, 2014 at 10:29 PM, Marcos Juarez Lopez wrote: > Hi, > > We have several Kafka clusters in production, and we've had to reassign > replication a few times now in production. Some of our topic/partitions

Re: Kafka Replication Behavior

2014-04-28 Thread Hanish Bansal
Thanks Bert for response. Are you saying that if i set replication factor to 3 then two nodes can go down, in that case only 3rd down can serve the data? In that case(replication factor 3 ) also if some partitions are in synchronization process and meanwhile a node goes down then we may face this

Re: Kafka Replication Behavior

2014-04-28 Thread Bert Corderman
Only a single broker needs to be online for data to be available. In your example partition 2 and 3 had copies of data on broker 0 and 1. When those two brokers went down your data was unavailable. To withstand two brokers going offline you would want to change your replication factor to 3. O

Re: Kafka Replication Behavior

2014-04-28 Thread Hanish Bansal
After restarting it start working fine. What i understand from this scenario is: There must be atleast 2 brokers in isr (In Sync Replica) list of each topic partition before any node goes down which has replica, otherwise topic may go to unstable state. Or In other words, If topic partitions are

Re: Kafka Replication Behavior

2014-04-28 Thread Hanish Bansal
After restarting it start working fine. What i understand from this scenario is: On Mon, Apr 28, 2014 at 8:30 PM, Jun Rao wrote: > Partition 2 and 3 are not accessible since both replicas are down. Have you > restarted the 2 failed brokers? > > Thanks, > > Jun > > > On Sun, Apr 27, 2014 at 11

Re: Kafka Replication Behavior

2014-04-28 Thread Jun Rao
Partition 2 and 3 are not accessible since both replicas are down. Have you restarted the 2 failed brokers? Thanks, Jun On Sun, Apr 27, 2014 at 11:58 PM, Hanish Bansal < hanish.bansal.agar...@gmail.com> wrote: > Hi All, > > I have 3 nodes kafka cluster with below configuration: > > Replication

Re: Kafka replication presentation at ApacheCon

2013-02-28 Thread Jun Rao
Right, everything has to go through the leader. If the producer waits for ack=committed, it will get an exception when sending m3. So the producer is aware of that m3 may not be committed by the broker. The producer can choose to resend m3 if it wants to be sure that m3 is safely stored at the brok

Re: Kafka replication presentation at ApacheCon

2013-02-28 Thread S Ahmed
Excellent thanks! BTW, in the slides, it shows the the message 'm3' is lost. I guess the leader is the single point of failure then when a producer sends a message, meaning it can never bypass the leader and write to the followers in case of leader failure right? On Thu, Feb 28, 2013 at 8:35 AM

Re: Kafka replication presentation at ApacheCon

2013-02-28 Thread Matan Safriel
Thanks Jun, great slides! On Wed, Feb 27, 2013 at 11:52 PM, Jun Rao wrote: > Hi, > > I gave a talk on Kafka replication at ApacheCon yesterday. The slides can > be found at > http://www.slideshare.net/junrao/kafka-replication-apachecon2013(it's > also added to Kafka wiki). I will share the link

Re: kafka replication blog

2013-02-11 Thread Michal Haris
Thanks Jun, makes sense. On Feb 8, 2013 4:00 PM, "Jun Rao" wrote: > That's right. If you are partitioning by key, that means you insist a > message has to go to a certain partition, whether it's available or not. > So, if a partition is not available, we will drop the message for the > partition

Re: kafka replication blog

2013-02-08 Thread Michal Haris
Thanks, makes sense. On Feb 8, 2013 4:00 PM, "Jun Rao" wrote: > That's right. If you are partitioning by key, that means you insist a > message has to go to a certain partition, whether it's available or not. > So, if a partition is not available, we will drop the message for the > partition in t

Re: kafka replication blog

2013-02-08 Thread Jun Rao
That's right. If you are partitioning by key, that means you insist a message has to go to a certain partition, whether it's available or not. So, if a partition is not available, we will drop the message for the partition in the async mode and consistently throw an exception to the caller in the s

Re: kafka replication blog

2013-02-08 Thread Michal Haris
So if the produces are partitioning by key we have to have replication if we dont want messages to get lost when partition goes down l right ? Thanks On Feb 8, 2013 5:12 AM, "Jun Rao" wrote: > We have fixed this issue in 0.8. Withreplication factor 1, if the producer > doesn't care about partitio

Re: kafka replication blog

2013-02-07 Thread Jun Rao
We have fixed this issue in 0.8. Withreplication factor 1, if the producer doesn't care about partitioning by key, messages will be sent to partitions that are currently available. Thanks, Jun On Thu, Feb 7, 2013 at 3:11 PM, Michal Haris wrote: > Same here, summary was need as we have a fairly

Re: kafka replication blog

2013-02-07 Thread Michal Haris
Same here, summary was need as we have a fairly large ecosystem of multiple 0.7.2 clusters and I am planning to test upgrade to 0.8. However, one thing creeping at the back of my mind regarding 0.8 is something i have spotted in one thread few weeks ago namely that the rebalance behaviour of produ

Re: kafka replication blog

2013-02-07 Thread Vaibhav Puranik
Same here. Thanks a lot Jun. Regards, Vaibhav On Thu, Feb 7, 2013 at 10:38 AM, Felix GV wrote: > Thanks Jun! > > I hadn't been following the discussions regarding 0.8 and replication for a > little while and this was a great post to refresh my memory and get up to > speed on the current replica

Re: kafka replication blog

2013-02-07 Thread Felix GV
Thanks Jun! I hadn't been following the discussions regarding 0.8 and replication for a little while and this was a great post to refresh my memory and get up to speed on the current replication architecture's design. -- Felix On Tue, Feb 5, 2013 at 2:21 PM, Jun Rao wrote: > I just posted the