I did some quick tests and the mbean values look reasonable. On the
producer side, produce msg/sec is actually for all brokers.

Thanks,

Jun

On Fri, Jan 16, 2015 at 12:09 PM, svante karlsson <s...@csi.se> wrote:

> Disregard the previous message, it was send accidently..
>
> Jun,
>
> I don't know if it was an issue with graphite or the mbean and have not
> seen it since - and we have tried of several cases of failover.
>
> That said, I have the feeling that it was a kafka issue and I'm a bit
> suspicious about the new mbeans.
>
> I attach a screenshot from the grafana dashboard and if you look at the
> first graph (top left) at ~10% it shows the startup after upgrade.
>
> This is a 3 node cluster with a topic of 2 partitions. When we start up a
> single producer it produces messages to both partitions without message
> loss. I know that all messages are acked.
>
>  If you look at the "produce msg/sec" graph it seems to hit 2 servers (it's
> per broker) but "messages in rate" & "byte in rate",& "byte out rate" (all
> from the new mbeans) look as if the data only hits one broker. (those are
> also per broker)
>
> At 70% I restarted two brokers after each other. After that point all three
> graphs looks fine.
>
> I'm not at work now and can't dig into the graphite data but I now see that
> the "fetch follower" also looks strange
>
> I can't file it as a bug report as I can't reproduce it but I have a
> distinct feeling that I can't trust the new mbeans or have to find another
> explanation.
>
> regard it as an observation if someone else reports issues.
>
>
> thanks,
>
> svante
>
> 2015-01-16 20:56 GMT+01:00 svante karlsson <s...@csi.se>:
>
> > Jun,
> >
> > I don't know if it was an issue with graphite or the mbean but I have not
> > seen it since - and we have tried of several cases of failover and this
> > problem has only been seen once.
> >
> > That said, I have the feeling that it was a kafka issue and I'm a bit
> > suspicious about the new mbeans.
> >
> > I attach a screenshot from the grafana dashboard and if you look at the
> > first graph (top left) at ~10% it shows the startup after upgrade.
> >
> > This is a 3 node cluster with a topic of 2 partitions. When we start up a
> > single producer it produces messages to both partitions without message
> > loss. I know that all messages are acked.
> >
> >  If you look at the "produce message msg/sec" graph it seems to hit 2
> > servers (it's per broker)
> >
> >
> > Bad picture but
> >
> > 2015-01-16 18:05 GMT+01:00 Jun Rao <j...@confluent.io>:
> >
> >> Svante,
> >>
> >> I tested this out locally and the mbeans for those metrics do show up on
> >> startup. Can you reproduce the issue reliably? Also, is what you saw an
> >> issue with the mbean itself or graphite?
> >>
> >> Thanks,
> >>
> >> Jun
> >>
> >> On Fri, Jan 16, 2015 at 4:38 AM, svante karlsson <s...@csi.se> wrote:
> >>
> >> > I upgrade two small test cluster and I had two small issues but I'm,
> not
> >> > clear yet as to if those were an issue due to us using ansible to
> >> configure
> >> > and deploy the cluster.
> >> >
> >> > The first issue could be us doing something bad when distributing the
> >> > update (I updated, not reinstalled) but it should be easy for you to
> >> > disregard since it seems so trivial.
> >> >
> >> > We replace the kafka-server-start.sh with something else but we had
> the
> >> > line
> >> >
> >> > EXTRA_ARGS="-name kafkaServer -loggc"
> >> >
> >> > then kafka-run-class.sh exits without starting the VM and complains on
> >> > unknown options. ( both -name and -loggc ) - once we removed the
> >> EXTRA_ARGS
> >> > everything starts.
> >> >
> >> > as I said - everyone should have this issue if it was a problem...
> >> >
> >> >
> >> > The second thing is regarding the jmx beans. I reconfigured our
> graphite
> >> > monitoring and noticed that the following metrics stopped working on
> one
> >> > broker
> >> > - server.BrokerTopicMetrics.MessagesInPerSec.OneMinuteRate,
> >> > - server.BrokerTopicMetrics.ByteInPerSec.OneMinuteRate,
> >> > - server.BrokerTopicMetrics.ByteOutPerSec.OneMinuteRate
> >> >
> >> > I had graphs running and the it looked like the traffic was dropping
> on
> >> > those metrics but our producers was working without problems and the
> >> > metrics
> >> > network.RequestMetrics.Produce.RequestsPerSec.OneMinuteRate confirmed
> >> that
> >> > on all brokers.
> >> >
> >> > A restart of the offending broker brought the metrics online again.
> >> >
> >> > /svante
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> > 2015-01-16 3:42 GMT+01:00 Gwen Shapira <gshap...@cloudera.com>:
> >> >
> >> > > Would make sense to enable it after we have authorization feature
> and
> >> > > admins can control who can delete what.
> >> > >
> >> > > On Thu, Jan 15, 2015 at 6:32 PM, Jun Rao <j...@confluent.io> wrote:
> >> > > > Yes, I agree it's probably better not to enable
> >> "delete.topic.enable"
> >> > by
> >> > > > default.
> >> > > >
> >> > > > Thanks,
> >> > > >
> >> > > > Jun
> >> > > >
> >> > > > On Thu, Jan 15, 2015 at 6:29 PM, Joe Stein <joe.st...@stealth.ly>
> >> > wrote:
> >> > > >
> >> > > >> I think that is a change of behavior that organizations may get
> >> burned
> >> > > on.
> >> > > >> Right now there is no delete data feature. If an operations teams
> >> > > upgrades
> >> > > >> to 0.8.2 and someone decides to delete a topic then there will be
> >> data
> >> > > >> loss. The organization may not have wanted that to happen. I
> would
> >> > > argue to
> >> > > >> not have a way to "by default" delete data. There is something
> >> > > actionable
> >> > > >> about consciously turning on a feature that allows anyone with
> >> access
> >> > to
> >> > > >> kafka-topics (or zookeeper for that matter) to delete Kafka data.
> >> If
> >> > > folks
> >> > > >> want that feature then flip the switch prior to upgrade or after
> >> and
> >> > > >> rolling restart and have at it. By not setting it as default they
> >> will
> >> > > know
> >> > > >> they have to turn it on and figure out what they need to-do from
> a
> >> > > security
> >> > > >> perspective (until Kafka gives them that) to protect their data
> >> > (through
> >> > > >> network or other type of measures).
> >> > > >>
> >> > > >> On Thu, Jan 15, 2015 at 8:24 PM, Manikumar Reddy <
> >> > ku...@nmsworks.co.in>
> >> > > >> wrote:
> >> > > >>
> >> > > >> > Also can we remove "delete.topic.enable" config property and
> >> enable
> >> > > topic
> >> > > >> > deletion by default?
> >> > > >> > On Jan 15, 2015 10:07 PM, "Jun Rao" <j...@confluent.io> wrote:
> >> > > >> >
> >> > > >> > > Thanks for reporting this. I will remove that option in RC2.
> >> > > >> > >
> >> > > >> > > Jun
> >> > > >> > >
> >> > > >> > > On Thu, Jan 15, 2015 at 5:21 AM, Jaikiran Pai <
> >> > > >> jai.forums2...@gmail.com>
> >> > > >> > > wrote:
> >> > > >> > >
> >> > > >> > > > I just downloaded the Kafka binary and am trying this on my
> >> 32
> >> > bit
> >> > > >> JVM
> >> > > >> > > > (Java 7)? Trying to start Zookeeper or Kafka server keeps
> >> > failing
> >> > > >> with
> >> > > >> > > > "Unrecognized VM option 'UseCompressedOops'":
> >> > > >> > > >
> >> > > >> > > > ./zookeeper-server-start.sh ../config/zookeeper.properties
> >> > > >> > > > Unrecognized VM option 'UseCompressedOops'
> >> > > >> > > > Error: Could not create the Java Virtual Machine.
> >> > > >> > > > Error: A fatal exception has occurred. Program will exit.
> >> > > >> > > >
> >> > > >> > > > Same with the Kafka server startup scripts. My Java version
> >> is:
> >> > > >> > > >
> >> > > >> > > > java version "1.7.0_71"
> >> > > >> > > > Java(TM) SE Runtime Environment (build 1.7.0_71-b14)
> >> > > >> > > > Java HotSpot(TM) Server VM (build 24.71-b01, mixed mode)
> >> > > >> > > >
> >> > > >> > > > Should there be a check in the script, before adding this
> >> > option?
> >> > > >> > > >
> >> > > >> > > > -Jaikiran
> >> > > >> > > >
> >> > > >> > > > On Wednesday 14 January 2015 10:08 PM, Jun Rao wrote:
> >> > > >> > > >
> >> > > >> > > >> + users mailing list. It would be great if people can test
> >> this
> >> > > out
> >> > > >> > and
> >> > > >> > > >> report any blocker issues.
> >> > > >> > > >>
> >> > > >> > > >> Thanks,
> >> > > >> > > >>
> >> > > >> > > >> Jun
> >> > > >> > > >>
> >> > > >> > > >> On Tue, Jan 13, 2015 at 7:16 PM, Jun Rao <
> j...@confluent.io>
> >> > > wrote:
> >> > > >> > > >>
> >> > > >> > > >>  This is the first candidate for release of Apache Kafka
> >> > 0.8.2.0.
> >> > > >> > There
> >> > > >> > > >>> has been some changes since the 0.8.2 beta release,
> >> especially
> >> > > in
> >> > > >> the
> >> > > >> > > new
> >> > > >> > > >>> java producer api and jmx mbean names. It would be great
> if
> >> > > people
> >> > > >> > can
> >> > > >> > > >>> test
> >> > > >> > > >>> this out thoroughly. We are giving people 10 days for
> >> testing
> >> > > and
> >> > > >> > > voting.
> >> > > >> > > >>>
> >> > > >> > > >>> Release Notes for the 0.8.2.0 release
> >> > > >> > > >>> *https://people.apache.org/~junrao/kafka-0.8.2.0-
> >> > > >> > > >>> candidate1/RELEASE_NOTES.html
> >> > > >> > > >>> <https://people.apache.org/~junrao/kafka-0.8.2.0-
> >> > > >> > > >>> candidate1/RELEASE_NOTES.html>*
> >> > > >> > > >>>
> >> > > >> > > >>> *** Please download, test and vote by Friday, Jan 23h,
> 7pm
> >> PT
> >> > > >> > > >>>
> >> > > >> > > >>> Kafka's KEYS file containing PGP keys we use to sign the
> >> > > release:
> >> > > >> > > >>> *
> >> > > https://people.apache.org/~junrao/kafka-0.8.2.0-candidate1/KEYS
> >> > > >> > > >>> <
> >> > > https://people.apache.org/~junrao/kafka-0.8.2.0-candidate1/KEYS>*
> >> > > >> > in
> >> > > >> > > >>> addition to the md5, sha1
> >> > > >> > > >>> and sha2 (SHA256) checksum.
> >> > > >> > > >>>
> >> > > >> > > >>> * Release artifacts to be voted upon (source and binary):
> >> > > >> > > >>> *
> >> https://people.apache.org/~junrao/kafka-0.8.2.0-candidate1/
> >> > > >> > > >>> <
> >> https://people.apache.org/~junrao/kafka-0.8.2.0-candidate1/
> >> > >*
> >> > > >> > > >>>
> >> > > >> > > >>> * Maven artifacts to be voted upon prior to release:
> >> > > >> > > >>> *https://people.apache.org/~junrao/kafka-0.8.2.0-
> >> > > >> > > >>> candidate1/maven_staging/
> >> > > >> > > >>> <https://people.apache.org/~junrao/kafka-0.8.2.0-
> >> > > >> > > >>> candidate1/maven_staging/>*
> >> > > >> > > >>>
> >> > > >> > > >>> * scala-doc
> >> > > >> > > >>> *https://people.apache.org/~junrao/kafka-0.8.2.0-
> >> > > >> > > >>> candidate1/scaladoc/#package
> >> > > >> > > >>> <https://people.apache.org/~junrao/kafka-0.8.2.0-
> >> > > >> > > >>> candidate1/scaladoc/#package>*
> >> > > >> > > >>>
> >> > > >> > > >>> * java-doc
> >> > > >> > > >>> *
> >> > > >>
> >> https://people.apache.org/~junrao/kafka-0.8.2.0-candidate1/javadoc/
> >> > > >> > > >>> <
> >> > > >>
> >> https://people.apache.org/~junrao/kafka-0.8.2.0-candidate1/javadoc/
> >> > > >> > >*
> >> > > >> > > >>>
> >> > > >> > > >>> * The tag to be voted upon (off the 0.8.2 branch) is the
> >> > 0.8.2.0
> >> > > >> tag
> >> > > >> > > >>> *
> >> https://git-wip-us.apache.org/repos/asf?p=kafka.git;a=tag;h=
> >> > > >> > > >>> b0c7d579f8aeb5750573008040a42b7377a651d5
> >> > > >> > > >>> <
> >> https://git-wip-us.apache.org/repos/asf?p=kafka.git;a=tag;h=
> >> > > >> > > >>> b0c7d579f8aeb5750573008040a42b7377a651d5>*
> >> > > >> > > >>>
> >> > > >> > > >>> /*******************************************
> >> > > >> > > >>>
> >> > > >> > > >>> Thanks,
> >> > > >> > > >>>
> >> > > >> > > >>> Jun
> >> > > >> > > >>>
> >> > > >> > > >>>
> >> > > >> > > >
> >> > > >> > >
> >> > > >> >
> >> > > >>
> >> > >
> >> >
> >>
> >
> >
>

Reply via email to