Hi Pushkar, If you are using Linux and Kafka 2.6.0+, the closest metric to what you are looking for is TotalDiskReadBytes [1], which measures data transfer at the block layer. Assuming your consumers are doing tail reads and there is no other activity which requires loading pages from the disk on your system (including log compaction from Kafka), you can determine if you are effectively hitting the disk or not.
[1] https://cwiki.apache.org/confluence/display/KAFKA/KIP-551%3A+Expose+disk+read+and+write+metrics Thanks, Alexandre Le sam. 15 mai 2021 à 05:49, Pushkar Deole <pdeole2...@gmail.com> a écrit : > > Hi All, > > is there any metric that I can use to check whether the memory allocated > for kafka is sufficient for the given load on the brokers and whether kafka > is optimally making use of page cache for consumer fetch reads which are > not going to disk for each read slowing down the overall consumer > processing ad thus increasing consumer lag? > > which metric can tell that i should assign more memory to brokers?