Hi Liam Clarke, Sorry but this is bit unclear for me. Can you please elaborate your answer? I am beginner to Kafka. " Producers emit metrics via JMX ": - How to enable this? I have kafka-Manager. Can I make use of kafka-manager? How? “kafka.producer:type=producer-metrics,client-id=(.+),topic=(.+)record-send-rate” please help to explain this.
Regards, Sunil. -----Original Message----- From: Liam Clarke <liam.cla...@adscale.co.nz> Sent: Thursday, February 20, 2020 11:16 AM To: users@kafka.apache.org Subject: [External] Re: Urgent helep please! How to measure producer and consumer throughput for single partition? This mail originated from an external party outside DBS - mailto:users-return-39554-sunilchaudhari=dbs....@kafka.apache.org. Do not click on links or open attachments unless you recognize the sender and know the content is safe. Hi Sunil, Producers emit metrics via JMX that will help you, assuming that your producers are using a round robin partition assignment strategy, you could divide this metric by your number of partitions, kafka.producer:type=producer-metrics,client-id=(.+),topic=(.+)record-send-rate Kind regards, Liam Clarke On Thu, 20 Feb. 2020, 5:57 pm Sunil CHAUDHARI, <mailto:sunilchaudh...@dbs.com.invalid> wrote: > Hi > I was referring to the article by Mr. June Rao about partitions in > kafka cluster. > https://www.confluent.io/blog/how-choose-number-topics-partitions-kafk > a-cluster/ > > "A rough formula for picking the number of partitions is based on > throughput. You measure the throughout that you can achieve on a > single partition for production (call it p) and consumption (call it > c). Let's say your target throughput is t. Then you need to have at > least max(t/p, t/c) partitions." > > I have the data pipeline as below. > > Filebeat-->Kafka-->Logstash-->Elasticsearch > There are many filebeat agents sending data to kafka. I want to > understand , how can I measure the events per seconds getting written > to Kafka? This will help me to know 'p' in above formula. > I can measure the consumer throughput by monitoring logsatsh pipelines > on Kibana. So it will give me 'c' in above formula. > > I know target throughput in my cluster, that is 't'. 30k events/s. > > Please let me know if I am going wrong? > > Regards, > Sunil. > CONFIDENTIAL NOTE: > The information contained in this email is intended only for the use > of the individual or entity named above and may contain information > that is privileged, confidential and exempt from disclosure under applicable > law. > If the reader of this message is not the intended recipient, you are > hereby notified that any dissemination, distribution or copying of > this communication is strictly prohibited. If you have received this > message in error, please immediately notify the sender and delete the mail. > Thank you. > CONFIDENTIAL NOTE: The information contained in this email is intended only for the use of the individual or entity named above and may contain information that is privileged, confidential and exempt from disclosure under applicable law. If the reader of this message is not the intended recipient, you are hereby notified that any dissemination, distribution or copying of this communication is strictly prohibited. If you have received this message in error, please immediately notify the sender and delete the mail. Thank you.