Thanks for the help Guozhang! however i realized that the exception and actual problem is totally different. The problem was the client was not set with SSL truststore while server is SSLenabled. I also found this open bug on kafka https://issues.apache.org/jira/browse/KAFKA-4493 After setting the SSL properties on stream, I am able to get it up and running.
@kafka developers, I think the problem is very misleading and should be fixed as soon as possible, or a proper exception should be thrown. On Thu, May 28, 2020 at 9:46 AM Guozhang Wang <wangg...@gmail.com> wrote: > Hello Pushkar, > > I think the memory pressure may not come from the topic data consumption, > but from rocksDB used for materializing the global table. Note rocksDB > allocates large chunk of memory beforehand in mem-table / page cache / > reader cache with default configs. You can get some detailed information > from this KIP: > > https://cwiki.apache.org/confluence/display/KAFKA/KIP-607%3A+Add+Metrics+to+Kafka+Streams+to+Record+the+Memory+Used+by+RocksDB > > > Guozhang > > > On Wed, May 27, 2020 at 8:44 PM Pushkar Deole <pdeole2...@gmail.com> > wrote: > > > Hello All, > > > > I am using Stream DSL API just to create a GlobalKTable backed by a > topic. > > The topology is simple, just create a global table from a topic and > that's > > it (pasted below code snippet), when I run this service on K8S cluster > > (container in a pod), the service gets OutOfMemoryError during > > kafkaStreams.start() method call (exception trace pasted below). Note > that > > the topic is newly created so there is no data in the topic. POD memory > was > > set initially to 500MiB which I doubled to 1000MiB but no luck. > > kafka-streams and kafka-clients jar at 2.3.1 version. Broker might be a > > version ahead I think 2.4 but that should not be an issue. Any help would > > be appreciated since I am blocked at this point. > > > > Properties props = new Properties(); > > props.put(StreamsConfig.APPLICATION_ID_CONFIG, DEFAULT_APPLICATION_ID); > > props.put(StreamsConfig.BOOTSTRAP_SERVERS_CONFIG, theKafkaServers); > > StreamsBuilder streamsBuilder = new StreamsBuilder(); > > GlobalKTable<String, Map<String, String>> groupCacheTable = > > streamsBuilder.globalTable(GROUP_CACHE_TOPIC, > > Consumed.with(Serdes.String(), GroupCacheSerdes.groupCache()), > > Materialized.as(GROUP_CACHE_STORE_NAME)); > > Topology groupCacheTopology = streamsBuilder.build(); > > kafkaStreams = new KafkaStreams(groupCacheTopology, props); > > kafkaStreams.start(); > > > > Runtime.getRuntime().addShutdownHook(new Thread(() -> { > > LOG.info("Stopping the stream"); > > kafkaStreams.close(); > > })); > > > > > > > {"@timestamp":"2020-05-28T03:11:39.719+00:00","@version":"1","message":"stream-client > > [DsiApplication-3db7d0ab-c84b-4e39-9038-61fa235786b9] State transition > from > > CREATED to > > > > > REBALANCING","logger_name":"org.apache.kafka.streams.KafkaStreams","thread_name":"main","level":"INFO","level_value":20000} > > > > > {"@timestamp":"2020-05-28T03:11:43.532+00:00","@version":"1","message":"Uncaught > > exception in thread 'kafka-admin-client-thread | > > > > > DsiApplication-3db7d0ab-c84b-4e39-9038-61fa235786b9-admin':","logger_name":"org.apache.kafka.common.utils.KafkaThread","thread_name":"kafka-admin-client-thread > > | > > > > > DsiApplication-3db7d0ab-c84b-4e39-9038-61fa235786b9-admin","level":"ERROR","level_value":40000,"stack_trace":"java.lang.OutOfMemoryError: > > Java heap space\n\tat java.base/java.nio.HeapByteBuffer.<init>(Unknown > > Source)\n\tat java.base/java.nio.ByteBuffer.allocate(Unknown > Source)\n\tat > > > > > org.apache.kafka.common.memory.MemoryPool$1.tryAllocate(MemoryPool.java:30)\n\tat > > > > > org.apache.kafka.common.network.NetworkReceive.readFrom(NetworkReceive.java:112)\n\tat > > > > > org.apache.kafka.common.network.KafkaChannel.receive(KafkaChannel.java:424)\n\tat > > > > > org.apache.kafka.common.network.KafkaChannel.read(KafkaChannel.java:385)\n\tat > > > > > org.apache.kafka.common.network.Selector.attemptRead(Selector.java:651)\n\tat > > > > > org.apache.kafka.common.network.Selector.pollSelectionKeys(Selector.java:572)\n\tat > > org.apache.kafka.common.network.Selector.poll(Selector.java:483)\n\tat > > org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:539)\n\tat > > > > > org.apache.kafka.clients.admin.KafkaAdminClient$AdminClientRunnable.run(KafkaAdminClient.java:1152)\n\tat > > java.base/java.lang.Thread.run(Unknown Source)\n"} > > > > > {"@timestamp":"2020-05-28T03:11:44.641+00:00","@version":"1","message":"Uncaught > > exception in thread 'kafka-producer-network-thread | > > > > > DsiApplication-3db7d0ab-c84b-4e39-9038-61fa235786b9-StreamThread-1-producer':","logger_name":"org.apache.kafka.common.utils.KafkaThread","thread_name":"kafka-producer-network-thread > > | > > > > > DsiApplication-3db7d0ab-c84b-4e39-9038-61fa235786b9-StreamThread-1-producer","level":"ERROR","level_value":40000,"stack_trace":"java.lang.OutOfMemoryError: > > Java heap space\n\tat java.base/java.nio.HeapByteBuffer.<init>(Unknown > > Source)\n\tat java.base/java.nio.ByteBuffer.allocate(Unknown > Source)\n\tat > > > > > org.apache.kafka.common.memory.MemoryPool$1.tryAllocate(MemoryPool.java:30)\n\tat > > > > > org.apache.kafka.common.network.NetworkReceive.readFrom(NetworkReceive.java:112)\n\tat > > > > > org.apache.kafka.common.network.KafkaChannel.receive(KafkaChannel.java:424)\n\tat > > > > > org.apache.kafka.common.network.KafkaChannel.read(KafkaChannel.java:385)\n\tat > > > > > org.apache.kafka.common.network.Selector.attemptRead(Selector.java:651)\n\tat > > > > > org.apache.kafka.common.network.Selector.pollSelectionKeys(Selector.java:572)\n\tat > > org.apache.kafka.common.network.Selector.poll(Selector.java:483)\n\tat > > org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:539)\n\tat > > > > > org.apache.kafka.clients.producer.internals.Sender.runOnce(Sender.java:335)\n\tat > > > > > org.apache.kafka.clients.producer.internals.Sender.run(Sender.java:244)\n\tat > > java.base/java.lang.Thread.run(Unknown Source)\n"} > > > > > {"@timestamp":"2020-05-28T03:11:45.017+00:00","@version":"1","message":"Opening > > store group-cache-store in regular > > > > > mode","logger_name":"org.apache.kafka.streams.state.internals.RocksDBTimestampedStore","thread_name":"DsiApplication-3db7d0ab-c84b-4e39-9038-61fa235786b9-GlobalStreamThread","level":"INFO","level_value":20000} > > > > > {"@timestamp":"2020-05-28T03:11:45.020+00:00","@version":"1","message":"global-stream-thread > > [DsiApplication-3db7d0ab-c84b-4e39-9038-61fa235786b9-GlobalStreamThread] > > Restoring state for global store > > > > > group-cache-store","logger_name":"org.apache.kafka.streams.processor.internals.GlobalStateManagerImpl","thread_name":"DsiApplication-3db7d0ab-c84b-4e39-9038-61fa235786b9-GlobalStreamThread","level":"INFO","level_value":20000} > > > > > -- > -- Guozhang >