[ https://issues.apache.org/jira/browse/KAFKA-1024?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13751740#comment-13751740 ]
Jason Toffaletti commented on KAFKA-1024: ----------------------------------------- I may have spoken too soon, it might just be taking more time to reach critical levels. The GC activity reported by jvisualvm is slowly creeping up from rarely peaking at 0.1% to steadily hovering around 1.3% over the past hour and heap size has climbed from 50MB to over 80MB. {noformat} Object Histogram: num #instances #bytes Class description -------------------------------------------------------------------------- 1: 488074 26414648 char[] 2: 15191 20018704 byte[] 3: 486709 11681016 java.lang.String 4: 63839 8693080 * MethodKlass 5: 63839 8252816 * ConstMethodKlass 6: 93927 7522528 java.util.HashMap$Entry[] 7: 5468 5868952 * ConstantPoolKlass 8: 106269 5714200 java.lang.Object[] 9: 91418 5119408 java.util.HashMap 10: 5468 4572616 * InstanceKlassKlass 11: 4538 3346656 * ConstantPoolCacheKlass 12: 79237 3169480 java.util.TreeMap$Entry 13: 45404 2912840 java.util.Hashtable$Entry[] 14: 90667 2901344 java.util.Vector 15: 45352 2539712 java.util.Hashtable 16: 45330 1813200 java.security.ProtectionDomain 17: 14911 1650304 int[] ... Total : 2327288 140772592 Heap traversal took 46.218 seconds. {noformat} I'll update this in a few hours to get an idea where the difference is. My messages aren't complex, I'm using DefaultEncoder with KeyedMessage<byte[], byte[]> and null keys, so I thought I had a pretty good idea what size the messages are. I'm running this test on dev machines being used for many other purposes so it is very likely zookeeper will be slow and I'll end up hitting the max queue depth during a metadata refresh. However, I've not seen dropped messages on any producers yet. > possible memory leak in 0.8 beta1 producer with G1GC > ---------------------------------------------------- > > Key: KAFKA-1024 > URL: https://issues.apache.org/jira/browse/KAFKA-1024 > Project: Kafka > Issue Type: Bug > Affects Versions: 0.8 > Environment: java version "1.7.0_17" > Java(TM) SE Runtime Environment (build 1.7.0_17-b02) > Java HotSpot(TM) 64-Bit Server VM (build 23.7-b01, mixed mode) > Reporter: Jason Toffaletti > > I have this in my pom.xml > <dependency> > <groupId>org.apache.kafka</groupId> > <artifactId>kafka_2.9.2</artifactId> > <version>0.8.0-beta1</version> > </dependency> > <dependency> > <groupId>org.scala-lang</groupId> > <artifactId>scala-library</artifactId> > <version>2.9.2</version> > </dependency> > <dependency> > <groupId>org.xerial.snappy</groupId> > <artifactId>snappy-java</artifactId> > <version>1.0.4.1</version> > </dependency> > I'm using snappy compression codec and producing about 7k msg/sec on average. > I'm producing batches up to 10 messages per batch with null keys to a topic > with 16 partitions, no replication. Xms and Xmx are 64MB and I'm using > XX:+UseG1GC. After about 12 hours of operation heap usage hits right up > against the 64MB limit and the producer drops to about 4k msg/sec because of > the GC pressure. When I restart the process the heap usage goes back to > normal (around 30MB) and the producer does 7k msg/sec again. > What else can I provide to help debug this? -- This message is automatically generated by JIRA. If you think it was sent incorrectly, please contact your JIRA administrators For more information on JIRA, see: http://www.atlassian.com/software/jira