Also check for really long/bad GC pauses as another possibility. Not sure
your JDK and JVM_OPTS and if you are setting like this
https://kafka.apache.org/documentation.html#java or not. You need to find
some "spike" somewhere right before that error happens to track down what
is causing the timeouts.

On Tue, Sep 30, 2014 at 6:33 PM, Joe Stein <joe.st...@stealth.ly> wrote:

> It sounds like you have a much deeper rooted problem.  Is zookeeper
> swapping?  Something has to be causing this.  After you fix this symptom
> you will probably start to see constant leader elections and the isr
> shrinking/growing and constant consumer rebalancing (or at least every
> minute) and a herd affect up/down stream occuring.  You need to figure out
> what is causing the long session timeout and resolve that, IMHO.  Zookeeper
> health is the first place to look.  Next would be the network.
>
> /*******************************************
>  Joe Stein
>  Founder, Principal Consultant
>  Big Data Open Source Security LLC
>  http://www.stealth.ly
>  Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop>
> ********************************************/
>
> On Tue, Sep 30, 2014 at 4:57 PM, Seshadri, Balaji <
> balaji.sesha...@dish.com> wrote:
>
>> The zookeeper session timeout is 60 secs ,but that did not help.
>>
>> We are having broker crash and unresponsive, we got the "conditional
>> update" failed error when broker crashed which confirmed that it is because
>> of KAFKA-1382.
>>
>> server.log.2014-09-23:2014-09-23 13:54:48 ERROR utils.ZkUtils$ -
>> Conditional update of path
>> /brokers/topics/dish-promo-application-access/partitions/128/state with
>> data { "controller_epoch":40, "isr":[ 6, 1 ], "leader":1,
>> "leader_epoch":99, "version":1 } and expected version 150 failed due to
>> org.apache.zookeeper.KeeperException$BadVersionException: KeeperErrorCode =
>> BadVersion for
>> /brokers/topics/dish-promo-application-access/partitions/128/state
>>
>> We are in very old version 0.8-beta so it's not just patch but switching
>> to stable release version which also has the patch.
>>
>> -----Original Message-----
>> From: Joe Stein [mailto:joe.st...@stealth.ly]
>> Sent: Tuesday, September 30, 2014 2:01 PM
>> To: users@kafka.apache.org
>> Cc: Neha Narkhede
>> Subject: Re: BadVersion state in Kafka Logs
>>
>> Have you tried increasing your broker's zookeeper session timeout as a
>> work around for now to alleviate the issue?  Is that an option for you?
>> Assuming that is the culprit you are timing zk sessions out and bumping into
>> KAFKA-1382 on the reconnect? Not knowing enough about what is going on
>> with the cluster it is hard to say if anything negative will come from it
>> but seems like it might be a an approach to try... if you can figure out
>> what is causing the session to timeout and fix *that* it would be a
>> solution also.... if it is happening every couple days (as another email
>> thread
>> states) something is going on that may not just be fixed by a single
>> patch.
>>
>> /*******************************************
>>  Joe Stein
>>  Founder, Principal Consultant
>>  Big Data Open Source Security LLC
>>  http://www.stealth.ly
>>  Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop>
>> ********************************************/
>>
>> On Tue, Sep 30, 2014 at 11:49 AM, Seshadri, Balaji <
>> balaji.sesha...@dish.com
>> > wrote:
>>
>> > Hi Joe,
>> >
>> > I did not try on 0.8.1 branch ,I can try and see if it goes through
>> > when I get some breather.
>> >
>> > Thanks for initiating on 0.8.1.2.
>> >
>> > Thanks,
>> >
>> > Balaji
>> >
>> > -----Original Message-----
>> > From: Joe Stein [mailto:joe.st...@stealth.ly]
>> > Sent: Tuesday, September 30, 2014 9:34 AM
>> > To: users@kafka.apache.org
>> > Cc: Neha Narkhede
>> > Subject: Re: BadVersion state in Kafka Logs
>> >
>> > Does the patch in KAFKA-1382 apply on the 0.8.1 branch?  If not if you
>> > could make a patch that does would be great.
>> >
>> > I will kick off a discussion for KAFKA-1382 and the scala 2.11 for
>> > 0.8.1.2 release (and see what others may think we should do like the
>> > gradle changes I think we should do too for src release issues (and
>> > the jars in the repo)).  I will send that on dev/user in a little bit
>> > (please comment +1 community support please on that thread for the
>> release).
>> >
>> > /*******************************************
>> >  Joe Stein
>> >  Founder, Principal Consultant
>> >  Big Data Open Source Security LLC
>> >  http://www.stealth.ly
>> >  Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop>
>> > ********************************************/
>> >
>> > On Tue, Sep 30, 2014 at 11:10 AM, Seshadri, Balaji <
>> > balaji.sesha...@dish.com
>> > > wrote:
>> >
>> > > I would love to help you guys to make Kafka best in Pub/Sub, will
>> > > continue doing that whenever I can.
>> > >
>> > > Do we have 0.8.1.2 release tag  or should we apply patch on top of
>> > > 0.8.1.1 tag because we need this KAFKA-1382 JIRA ?.
>> > >
>> > > Balaji
>> > >
>> > > From: Neha Narkhede [mailto:neha.narkh...@gmail.com]
>> > > Sent: Monday, September 29, 2014 5:21 PM
>> > > To: Seshadri, Balaji
>> > > Cc: users@kafka.apache.org
>> > > Subject: Re: BadVersion state in Kafka Logs
>> > >
>> > > It is difficult to predict an exact date. Though all the discussions
>> > > of the progress and ETA are on the mailing list. You can follow the
>> > > discussions to know the details and/or offer to help out on the
>> > > outstanding issues.
>> > >
>> > > On Mon, Sep 29, 2014 at 3:48 PM, Seshadri, Balaji <
>> > > balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>> wrote:
>> > > Neha,
>> > >
>> > > Do you know the date in Oct when 0.8.2 is going to be out ?.
>> > >
>> > > Thanks,
>> > >
>> > > Balaji
>> > >
>> > > From: Neha Narkhede [mailto:neha.narkh...@gmail.com<mailto:
>> > > neha.narkh...@gmail.com>]
>> > > Sent: Thursday, September 25, 2014 1:08 PM
>> > > To: Seshadri, Balaji
>> > > Cc: users@kafka.apache.org<mailto:users@kafka.apache.org>
>> > >
>> > > Subject: Re: BadVersion state in Kafka Logs
>> > >
>> > > We are close to the release. I'd probably expect 0.8.2 sometime in
>> > October.
>> > >
>> > > On Thu, Sep 25, 2014 at 10:37 AM, Seshadri, Balaji <
>> > > balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>> wrote:
>> > > Hi Neha,
>> > >
>> > > Do you know when are you guys releasing 0.8.2 ?.
>> > >
>> > > Thanks,
>> > >
>> > > Balaji
>> > >
>> > > -----Original Message-----
>> > > From: Seshadri, Balaji [mailto:balaji.sesha...@dish.com<mailto:
>> > > balaji.sesha...@dish.com>]
>> > > Sent: Thursday, September 25, 2014 9:41 AM
>> > > To: users@kafka.apache.org<mailto:users@kafka.apache.org>
>> > > Subject: RE: BadVersion state in Kafka Logs
>> > >
>> > > Thanks for the replay.
>> > >
>> > > Please let me know if we can use trunk as 0.8.2 is not yet released.
>> > >
>> > > Balaji
>> > > ________________________________________
>> > > From: Neha Narkhede [neha.narkh...@gmail.com<mailto:
>> > > neha.narkh...@gmail.com>]
>> > > Sent: Wednesday, September 24, 2014 6:32 PM
>> > > To: users@kafka.apache.org<mailto:users@kafka.apache.org>
>> > > Subject: Re: BadVersion state in Kafka Logs
>> > >
>> > > From the logs you've attached, my guess is it's most likely due to
>> > > KAFKA-1382.
>> > >
>> > > Thanks,
>> > > Neha
>> > >
>> > > On Wed, Sep 24, 2014 at 10:48 AM, Seshadri, Balaji <
>> > > balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>
>> > > > wrote:
>> > >
>> > > > Hi,
>> > > >
>> > > >
>> > > >
>> > > > We got the below error in our logs and our consumers stopped
>> > > > consuming any data ?.It worked only after restart.
>> > > >
>> > > >
>> > > >
>> > > > We would like to confirm that it's because we are running with
>> > > > 0.8-beta version and not 0.8 release version to convince "THE MGMT"
>> > guys.
>> > > >
>> > > >
>> > > >
>> > > > Please let me know if it's this KAFKA-1382 causing the issue.
>> > > >
>> > > >
>> > > >
>> > > > Thanks,
>> > > >
>> > > >
>> > > >
>> > > > Balaji
>> > > >
>> > > >
>> > > >
>> > > > *From:* Gulia, Vikram
>> > > > *Sent:* Wednesday, September 24, 2014 8:43 AM
>> > > > *To:* Sharma, Navdeep; #IT-MAD DES; #IT-MAA
>> > > > *Cc:* Alam, Mohammad Shah
>> > > > *Subject:* RE: 9/23 prod issue - offline kafka partitions.
>> > > >
>> > > >
>> > > >
>> > > > Adding full MAA distro.
>> > > >
>> > > >
>> > > >
>> > > > DES Offshore looked in to the logs on kafka servers and seems like
>> > > > the issue we encountered yesterday may be described in these
>> > > > threads, please have a look -
>> > > >
>> > > >
>> > > >
>> > > > http://permalink.gmane.org/gmane.comp.apache.kafka.user/1904
>> > > >
>> > > >
>> > > >
>> > > > https://issues.apache.org/jira/browse/KAFKA-1382 (it describes the
>> > > > fix/patch which is available in
>> > > > 0.8.1.2/0.8.2<http://0.8.1.2/0.8.2>)
>> > > >
>> > > >
>> > > >
>> > > > Thank You,
>> > > >
>> > > > Vikram Gulia
>> > > >
>> > > >
>> > > >
>> > > > *From:* Sharma, Navdeep
>> > > > *Sent:* Wednesday, September 24, 2014 6:53 AM
>> > > > *To:* Gulia, Vikram; #IT-MAD DES
>> > > > *Cc:* #IT-MAA Offshore; Alam, Mohammad Shah
>> > > > *Subject:* RE: 9/23 prod issue - offline kafka partitions.
>> > > >
>> > > >
>> > > >
>> > > > Hi Vikram,
>> > > >
>> > > >
>> > > >
>> > > > We analyzed  below mentioned issue with MAA-Offshore (Abhishek)
>> > > > and found that the error occurred only on 23 Sept. This is  not
>> > > > historical as we checked last 4 days logs.
>> > > >
>> > > >
>> > > >
>> > > > It looks like that consumer got stopped on September 22 2014 for
>> > > > Linux patching activty.MAA started consumer September  23 2014 at
>> > > > 1:00
>> > AM.
>> > > >
>> > > >
>> > > >
>> > > > *Issue *in  server log   *"* *BadVersion for
>> > > > /brokers/topics/rain-burn-in/partitions/121/state"*  but it is not
>> > > > present in previous 4 days logs.
>> > > >
>> > > > More detail of this error can be found at-
>> > > >
>> > > > http://permalink.gmane.org/gmane.comp.apache.kafka.user/1904
>> > > >
>> > > >
>> > > >
>> > > > We are not sure about data loss in this scenario and working on
>> this.
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > > Let us know if any concerns.
>> > > >
>> > > >
>> > > >
>> > > > [image: cid:image001.gif@01CF7B0A.03F21580]
>> > > >
>> > > > Navdeep Sharma
>> > > > Developer - offshore,  Middleware Applications & Development o
>> India:
>> > > > 0120-4532000 - 2234
>> > > > c: +91-9911698102<tel:%2B91-9911698102>
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > >
>> > > > *From:* Gulia, Vikram
>> > > > *Sent:* Tuesday, September 23, 2014 6:17 PM
>> > > > *To:* #IT-MAD DES
>> > > > *Subject:* FW: 9/23 prod issue - offline kafka partitions.
>> > > >
>> > > >
>> > > >
>> > > > DES Offshore dev,
>> > > >
>> > > >
>> > > >
>> > > > Please work with MAA offshore to monitor the kafka broker as we
>> > > > had this incident where lot of partitions went offline around 1.45
>> > > > PM MST and MAA has to restart the kafka servers. We may have lost
>> > > > messages and we need to see if there is a way to figure out what
>> > > > was
>> > the impact.
>> > > >
>> > > >
>> > > >
>> > > > Also, check the logs for kafka servers and see if we can figure
>> > > > out why did partitions go offline or are un-available? Let us know
>> > > > if you find anything relevant.
>> > > >
>> > > >
>> > > >
>> > > > Thank You,
>> > > >
>> > > > Vikram Gulia
>> > > >
>> > > >
>> > > >
>> > > > *From:* Nielsen, Andy
>> > > > *Sent:* Tuesday, September 23, 2014 5:04 PM
>> > > > *To:* #IT-MAD DES; Gulia, Vikram
>> > > > *Cc:* #IT-MAA
>> > > > *Subject:* 9/23 prod issue - offline kafka partitions.
>> > > >
>> > > >
>> > > >
>> > > > desadmin@pc1mwdpl01 ~/bin $ ./kafka.sh topic
>> > > > --unavailable-partitions
>> > > >
>> > > > topic: account-access   partition: 21   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: account-access   partition: 51   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: account-access   partition: 81   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: account-access   partition: 111  leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: account-activated        partition: 13   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: account-activated        partition: 43   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: account-activated        partition: 73   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: account-activated        partition: 103  leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: account-adjustment-issued        partition: 27   leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: account-adjustment-issued        partition: 57   leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: account-adjustment-issued        partition: 87   leader: -1
>> > > > replicas: 4,6,1 isr: 1
>> > > >
>> > > > topic: account-adjustment-issued        partition: 117  leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: account-created  partition: 11   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: account-created  partition: 41   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: account-created  partition: 71   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: account-created  partition: 101  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: account-info-updated     partition: 7    leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: account-info-updated     partition: 37   leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: account-info-updated     partition: 67   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: account-info-updated     partition: 97   leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: account-info-updated     partition: 127  leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: application-access       partition: 21   leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: application-access       partition: 51   leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: application-access       partition: 81   leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: application-access       partition: 111  leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: bill-generated   partition: 3    leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: bill-generated   partition: 33   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: bill-generated   partition: 63   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: bill-generated   partition: 93   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: bill-generated   partition: 123  leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: collected-event  partition: 29   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: collected-event  partition: 59   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: collected-event  partition: 89   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: collected-event  partition: 119  leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: customer-cues    partition: 27   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: customer-cues    partition: 57   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: customer-cues    partition: 87   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: customer-cues    partition: 117  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: dish-promo-application-access    partition: 23   leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: dish-promo-application-access    partition: 53   leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: dish-promo-application-access    partition: 83   leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: dish-promo-application-access    partition: 113  leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: event-response   partition: 2    leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: event-response   partition: 32   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: event-response   partition: 62   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: event-response   partition: 92   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: event-response   partition: 122  leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: leads-service    partition: 24   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: leads-service    partition: 54   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: leads-service    partition: 84   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: leads-service    partition: 114  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: logprod_v3       partition: 3    leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: logprod_v3       partition: 33   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: logprod_v3       partition: 63   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: logprod_v3       partition: 93   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: logprod_v3       partition: 123  leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: online-account-registration-attempted    partition: 21
>>  leader:
>> > > > -1      replicas: 4,6,1 isr:
>> > > >
>> > > > topic: online-account-registration-attempted    partition: 51
>>  leader:
>> > > > -1      replicas: 4,6,1 isr: 1
>> > > >
>> > > > topic: online-account-registration-attempted    partition: 81
>>  leader:
>> > > > -1      replicas: 4,6,1 isr:
>> > > >
>> > > > topic: online-account-registration-attempted    partition: 111
>> leader:
>> > > > -1      replicas: 4,6,1 isr:
>> > > >
>> > > > topic: order-cancelled  partition: 29   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-cancelled  partition: 59   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-cancelled  partition: 89   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-cancelled  partition: 119  leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: order-completed  partition: 24   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-completed  partition: 54   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-completed  partition: 84   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: order-completed  partition: 114  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-created    partition: 25   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-created    partition: 55   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-created    partition: 85   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-created    partition: 115  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-modified   partition: 8    leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: order-modified   partition: 38   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-modified   partition: 68   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-modified   partition: 98   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-modified   partition: 128  leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: order-request    partition: 24   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-request    partition: 54   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-request    partition: 84   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: order-request    partition: 114  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-response   partition: 27   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: order-response   partition: 57   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-response   partition: 87   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: order-response   partition: 117  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: outbound-call-attempted  partition: 13   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: outbound-call-attempted  partition: 43   leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: outbound-call-attempted  partition: 73   leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: outbound-call-attempted  partition: 103  leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: outbound-communications  partition: 4    leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: outbound-communications  partition: 34   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: outbound-communications  partition: 64   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: outbound-communications  partition: 94   leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: outbound-communications  partition: 124  leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > > topic: postal-mail-undeliverable        partition: 15   leader: -1
>> > > > replicas: 4,6,1 isr: 1
>> > > >
>> > > > topic: postal-mail-undeliverable        partition: 45   leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: postal-mail-undeliverable        partition: 75   leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: postal-mail-undeliverable        partition: 105  leader: -1
>> > > > replicas: 4,6,1 isr:
>> > > >
>> > > > topic: rain-burn-in     partition: 4    leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: rain-burn-in     partition: 34   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: rain-burn-in     partition: 64   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: rain-burn-in     partition: 94   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: rain-burn-in     partition: 124  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: rain-enhanced    partition: 26   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: rain-enhanced    partition: 56   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: rain-enhanced    partition: 86   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: rain-enhanced    partition: 116  leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: rain-listener    partition: 23   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: rain-listener    partition: 53   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: rain-listener    partition: 83   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: rain-listener    partition: 113  leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: rain-load-test   partition: 8    leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: rain-load-test   partition: 38   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: rain-load-test   partition: 68   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: rain-load-test   partition: 98   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: rain-load-test   partition: 128  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: submit-agreement partition: 2    leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: submit-agreement partition: 32   leader: -1      replicas:
>> 4,6,1
>> > > > isr: 1
>> > > >
>> > > > topic: submit-agreement partition: 62   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: submit-agreement partition: 92   leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: submit-agreement partition: 122  leader: -1      replicas:
>> 4,6,1
>> > > > isr:
>> > > >
>> > > > topic: threshold-exceeded       partition: 14   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: threshold-exceeded       partition: 44   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: threshold-exceeded       partition: 74   leader: -1
>> > replicas:
>> > > > 4,6,1 isr:
>> > > >
>> > > > topic: threshold-exceeded       partition: 104  leader: -1
>> > replicas:
>> > > > 4,6,1 isr: 1
>> > > >
>> > > >
>> > > >
>> > > > *Andy Nielsen*
>> > > >
>> > > > *Middleware Application Admin*
>> > > >
>> > > > *303-723-2347<tel:303-723-2347> <303-723-2347<tel:303-723-2347>>*
>> > > >
>> > > > *cell:720-971-2856<tel:720-971-2856>
>> > > > <720-971-2856<tel:720-971-2856>>*
>> > > >
>> > > >
>> > > >
>> > >
>> > >
>> > >
>> >
>>
>
>

Reply via email to