Gérald Quintana created KAFKA-4491:
--------------------------------------

             Summary: OOME when Java client can not connect to brokers
                 Key: KAFKA-4491
                 URL: https://issues.apache.org/jira/browse/KAFKA-4491
             Project: Kafka
          Issue Type: Bug
          Components: clients
    Affects Versions: 0.10.0.1
         Environment: Any (Linux|Windows)
            Reporter: Gérald Quintana


Scenario: The broker cluster switched to SSL protocol but not the clients, this 
should have raised connection failures, but instead the client dies with OOME.

Sample code to easily reproce the problem:

{code}
public class Main {
    private static final Logger LOGGER = LoggerFactory.getLogger(Main.class);
    public static void main(String[] args) throws InterruptedException {
        int threads = 10;
        ExecutorService executorService = Executors.newFixedThreadPool(threads);
        for (int i = 0; i < threads; i++) {
            executorService.execute(new PrintConsumer("testgroup"+i, 
"testtopic"+i));
        }
        Thread.sleep(300000L);
        executorService.shutdown();
    }

    private static class PrintConsumer implements Runnable {
        private final String groupId;
        private final String topic;
        private final AtomicBoolean running = new AtomicBoolean(true);
        public PrintConsumer(String groupId, String topic) {
            this.groupId = groupId;
            this.topic = topic;
        }

        @Override
        public void run() {
            Properties props = new Properties();
            props.put(ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG, 
"kafka1:9093,kafka2:9093");
            props.put(ConsumerConfig.GROUP_ID_CONFIG, groupId);
            props.put(ConsumerConfig.CLIENT_ID_CONFIG,"testclient");
            props.put(ConsumerConfig.KEY_DESERIALIZER_CLASS_CONFIG, 
"org.apache.kafka.common.serialization.StringDeserializer");
            props.put(ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG, 
"org.apache.kafka.common.serialization.StringDeserializer");
            //props.put("ssl.truststore.location","/opt/truststore.jks");
            //props.put("ssl.truststore.password", "localhost");
            //props.put("security.protocol", "SSL");
            while (running.get()) {
                LOGGER.info("Connecting "+topic);
                try (KafkaConsumer<String, String> consumer = new 
KafkaConsumer<>(props)) {
                    consumer.subscribe(Collections.singleton(topic));
                    while (running.get()) {
                        ConsumerRecords<String, String> records = 
consumer.poll(100);
                        for (ConsumerRecord<String, String> record : records)
                            LOGGER.info("offset = %d, key = %s, value = %s%n", 
record.offset(), record.key(), record.value());
                    }
                } catch (Exception e) {
                    LOGGER.warn("Exception in "+groupId, e);
                }
            }
        }

        public void stop() {
            running.set(false);
        }
    }
}
{code}
Thrown exception:
{code}
java.lang.OutOfMemoryError: Java heap space
        at java.nio.HeapByteBuffer.<init>(HeapByteBuffer.java:57)
        at java.nio.ByteBuffer.allocate(ByteBuffer.java:335)
        at 
org.apache.kafka.common.network.NetworkReceive.readFromReadableChannel(NetworkReceive.java:93)
        at 
org.apache.kafka.common.network.NetworkReceive.readFrom(NetworkReceive.java:71)
        at 
org.apache.kafka.common.network.KafkaChannel.receive(KafkaChannel.java:154)
        at 
org.apache.kafka.common.network.KafkaChannel.read(KafkaChannel.java:135)
        at 
org.apache.kafka.common.network.Selector.pollSelectionKeys(Selector.java:323)
        at org.apache.kafka.common.network.Selector.poll(Selector.java:283)
        at org.apache.kafka.clients.NetworkClient.poll(NetworkClient.java:260)
        at 
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.clientPoll(ConsumerNetworkClient.java:360)
        at 
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:224)
        at 
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:192)
        at 
org.apache.kafka.clients.consumer.internals.ConsumerNetworkClient.poll(ConsumerNetworkClient.java:163)
        at 
org.apache.kafka.clients.consumer.internals.AbstractCoordinator.ensureCoordinatorReady(AbstractCoordinator.java:179)
        at 
org.apache.kafka.clients.consumer.KafkaConsumer.pollOnce(KafkaConsumer.java:973)
        at 
org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:937)
        at Main$PrintConsumer.run(Main.java:51)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
{code}
Increasing heap size doesn't help, the Kafka client swallows everything.
As soon as you fix configuration (uncomment SSL properties in configuration), 
everything is alright.
The allocation path seems to be:
{code} 
this     - value: byte[] #1
 <- hb (Java frame)     - class: java.nio.HeapByteBuffer, value: byte[] #1
  <- buffer (Java frame)     - class: 
org.apache.kafka.common.network.NetworkReceive, value: java.nio.HeapByteBuffer 
#5
   <- receive (Java frame)     - class: 
org.apache.kafka.common.network.KafkaChannel, value: 
org.apache.kafka.common.network.NetworkReceive #5
    <- attachment (Java frame)     - class: sun.nio.ch.SelectionKeyImpl, value: 
org.apache.kafka.common.network.KafkaChannel #5
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to