Not that I know of - but others may advise otherwise.
The change from KIP-551 is fairly self-contained and can be backported
well though.

Thanks,
Alexandre

Le dim. 16 mai 2021 à 14:51, Pushkar Deole <pdeole2...@gmail.com> a écrit :
>
> thanks Alexandre... currently we are using kafka 2.5.0, so is there any
> metric that can be used from 2.5.0?
>
> On Sun, May 16, 2021 at 6:02 PM Alexandre Dupriez <
> alexandre.dupr...@gmail.com> wrote:
>
> > Hi Pushkar,
> >
> > If you are using Linux and Kafka 2.6.0+, the closest metric to what
> > you are looking for is TotalDiskReadBytes [1], which measures data
> > transfer at the block layer.
> > Assuming your consumers are doing tail reads and there is no other
> > activity which requires loading pages from the disk on your system
> > (including log compaction from Kafka), you can determine if you are
> > effectively hitting the disk or not.
> >
> > [1]
> > https://cwiki.apache.org/confluence/display/KAFKA/KIP-551%3A+Expose+disk+read+and+write+metrics
> >
> > Thanks,
> > Alexandre
> >
> > Le sam. 15 mai 2021 à 05:49, Pushkar Deole <pdeole2...@gmail.com> a écrit
> > :
> > >
> > > Hi All,
> > >
> > > is there any metric that I can use to check whether the memory allocated
> > > for kafka is sufficient for the given load on the brokers and whether
> > kafka
> > > is optimally making use of page cache for consumer fetch reads which are
> > > not going to disk for each read slowing down the overall consumer
> > > processing ad thus increasing consumer lag?
> > >
> > > which metric can tell that i should assign more memory to brokers?
> >

Reply via email to