[
https://issues.apache.org/jira/browse/KAFKA-1041?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Neha Narkhede reassigned KAFKA-1041:
------------------------------------
Assignee: Neha Narkhede (was: Jun Rao)
> Number of file handles increases indefinitely in producer if broker host is
> unresolvable
> ----------------------------------------------------------------------------------------
>
> Key: KAFKA-1041
> URL: https://issues.apache.org/jira/browse/KAFKA-1041
> Project: Kafka
> Issue Type: Bug
> Components: producer
> Affects Versions: 0.8
> Environment: *unix*
> Reporter: Rajasekar Elango
> Assignee: Neha Narkhede
> Labels: features
>
> We found a issue that if broker host is un resolvable, the number of file
> handle keep increasing for every message we produce and eventually it uses up
> all available files handles in operating system. If broker itself is not
> running and broker host name is resolvable, open file handles count stays
> flat.
> lsof output shows number of these open file handles continue to grow for
> every message we produce.
> java 19631 relango 81u sock 0,6 0t0
> 196966526 can't identify protocol
> I can easily reproduce this with console producer, If I run console producer
> with right hostname and if broker is not running, the console producer will
> exit after three tries. But If I run console producer with unresolvable
> broker, it throws below exception and continues to wait for user input, every
> time I enter new message, it opens socket and file handle count keeps
> increasing..
> Here is Exception in producer
> ERROR fetching topic metadata for topics [Set(test-1378245487417)] from
> broker [ArrayBuffer(id:0,host:localhost1,port:6667)] failed
> (kafka.utils.Utils$)
> kafka.common.KafkaException: fetching topic metadata for topics
> [Set(test-1378245487417)] from broker
> [ArrayBuffer(id:0,host:localhost1,port:6667)] failed
> at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:51)
> at
> kafka.producer.BrokerPartitionInfo.updateInfo(BrokerPartitionInfo.scala:82)
> at
> kafka.producer.async.DefaultEventHandler$$anonfun$handle$2.apply$mcV$sp(DefaultEventHandler.scala:79)
> at kafka.utils.Utils$.swallow(Utils.scala:186)
> at kafka.utils.Logging$class.swallowError(Logging.scala:105)
> at kafka.utils.Utils$.swallowError(Utils.scala:45)
> at
> kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:79)
> at
> kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104)
> at
> kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87)
> at
> kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:67)
> at scala.collection.immutable.Stream.foreach(Stream.scala:526)
> at
> kafka.producer.async.ProducerSendThread.processEvents(ProducerSendThread.scala:66)
> at
> kafka.producer.async.ProducerSendThread.run(ProducerSendThread.scala:44)
> Caused by: java.nio.channels.UnresolvedAddressException
> at sun.nio.ch.Net.checkAddress(Net.java:30)
> at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:487)
> at kafka.network.BlockingChannel.connect(BlockingChannel.scala:59)
> at kafka.producer.SyncProducer.connect(SyncProducer.scala:151)
> at
> kafka.producer.SyncProducer.getOrMakeConnection(SyncProducer.scala:166)
> at
> kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:73)
> at kafka.producer.SyncProducer.send(SyncProducer.scala:117)
> at kafka.client.ClientUtils$.fetchTopicMetadata(ClientUtils.scala:37)
> ... 12 more
--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira