I confirmed. - Controller.log After kafka broker stated. Kafka broker continues outputting below log. Error is continued. ======= [2014-04-09 23:55:07,709] ERROR [Controller-0-to-broker-0-send-thread], Controller 0's connection to broker id:0,host:localhost,port:19092 was unsuccessful (kafka.controller.RequestSendThread) java.net.ConnectException: Connection refused at sun.nio.ch.Net.connect0(Native Method) at sun.nio.ch.Net.connect(Net.java:465) at sun.nio.ch.Net.connect(Net.java:457) at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:670) at kafka.network.BlockingChannel.connect(BlockingChannel.scala:57) at kafka.controller.RequestSendThread.connectToBroker(ControllerChannelManager.scala:173) at kafka.controller.RequestSendThread.liftedTree1$1(ControllerChannelManager.scala:140) at kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:131) at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:51) [2014-04-09 23:55:08,009] ERROR [Controller-0-to-broker-0-send-thread], Controller 0 epoch 7 failed to send UpdateMetadata request with correlation id 3 to broker id:0,host:localhost,port:19092. Reconnecting to broker. (kafka.controller.RequestSendThread) java.nio.channels.ClosedChannelException at kafka.network.BlockingChannel.send(BlockingChannel.scala:89) at kafka.controller.RequestSendThread.liftedTree1$1(ControllerChannelManager.scala:132) at kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:131) at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:51) =======
- state change log When kafka broker started. Below log outputed each partition. ======= [2014-04-09 23:53:43,341] TRACE Controller 0 epoch 7 started leader election for partition [TestTopic,0] (state.change.logger) kafka.common.NoReplicaOnlineException: No replica for partition [TestTopic,0] is alive. Live brokers are: [Set()], Assigned replicas are: [List(0)] at kafka.controller.OfflinePartitionLeaderSelector.selectLeader(PartitionLeaderSelector.scala:61) at kafka.controller.PartitionStateMachine.electLeaderForPartition(PartitionStateMachine.scala:335) at kafka.controller.PartitionStateMachine.kafka$controller$PartitionStateMachine$$handleStateChange(PartitionStateMachine.scala:184) at kafka.controller.PartitionStateMachine$$anonfun$triggerOnlinePartitionStateChange$3.apply(PartitionStateMachine.scala:98) at kafka.controller.PartitionStateMachine$$anonfun$triggerOnlinePartitionStateChange$3.apply(PartitionStateMachine.scala:95) at scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:743) at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:95) at scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:95) at scala.collection.Iterator$class.foreach(Iterator.scala:772) at scala.collection.mutable.HashTable$$anon$1.foreach(HashTable.scala:157) at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:190) at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:45) at scala.collection.mutable.HashMap.foreach(HashMap.scala:95) at scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:742) at kafka.controller.PartitionStateMachine.triggerOnlinePartitionStateChange(PartitionStateMachine.scala:95) at kafka.controller.PartitionStateMachine.startup(PartitionStateMachine.scala:68) at kafka.controller.KafkaController.onControllerFailover(KafkaController.scala:311) at kafka.controller.KafkaController$$anonfun$1.apply$mcV$sp(KafkaController.scala:161) at kafka.server.ZookeeperLeaderElector.elect(ZookeeperLeaderElector.scala:63) at kafka.server.ZookeeperLeaderElector$$anonfun$startup$1.apply$mcZ$sp(ZookeeperLeaderElector.scala:49) at kafka.server.ZookeeperLeaderElector$$anonfun$startup$1.apply(ZookeeperLeaderElector.scala:47) at kafka.server.ZookeeperLeaderElector$$anonfun$startup$1.apply(ZookeeperLeaderElector.scala:47) at kafka.utils.Utils$.inLock(Utils.scala:538) at kafka.server.ZookeeperLeaderElector.startup(ZookeeperLeaderElector.scala:47) at kafka.controller.KafkaController$$anonfun$startup$1.apply$mcV$sp(KafkaController.scala:630) at kafka.controller.KafkaController$$anonfun$startup$1.apply(KafkaController.scala:626) at kafka.controller.KafkaController$$anonfun$startup$1.apply(KafkaController.scala:626) at kafka.utils.Utils$.inLock(Utils.scala:538) at kafka.controller.KafkaController.startup(KafkaController.scala:626) at kafka.server.KafkaServer.startup(KafkaServer.scala:96) at kafka.server.KafkaServerStartable.startup(KafkaServerStartable.scala:34) at kafka.Kafka$.main(Kafka.scala:46) at kafka.Kafka.main(Kafka.scala) ======= regards. 2014-04-09 13:03 GMT+09:00 Jun Rao <jun...@gmail.com>: > Any error in the broker, controller and state change log? > > Thanks, > > Jun > > > On Tue, Apr 8, 2014 at 3:08 PM, Sotaro Kimura <rfbrin...@gmail.com> wrote: > > > Hi, all. > > > > Sorry for my late reply. > > > > I tried advertised.host.name/advertised.port. > > But producer failed fetching metadata from kafka broker. > > > > Configuration is below. > > - Kafka cluster in DataCenter(each machine has Grobal IP Address). > > Firewall allows only ssh port(22). > > ex) 192.168.100.100 > > - Kafka producer exists out of DataCenter. > > ex) 172.16.0.100 > > > > Broker additional setting is below. > > advertised.host.name=localhost > > advertised.port=19092 > > > > Kafka producer operation is below. > > > > - SSH port fowarding > > ======================== > > # ssh root@192.168.100.100 -L 19092:192.168.100.100:9092 -L 12181: > > 192.168.100.100:2181 > > // Kafka broker and zookeeper ssh port fowarding. > > ======================== > > > > - Producer data put(failed) > > ======================== > > # ./kafka-topics.sh --create --zookeeper localhost:12181 > > --replication-factor 1 --partitions 3 --topic TestTopic > > Created topic "TestTopic". > > // Succeed create topic over ssh port fowarding. > > # ./kafka-topics.sh --list --zookeeper localhost:12181 > > TestTopic > > // Succeed get topic list over ssh port fowarding. > > # ./kafka-console-producer.sh --broker-list localhost:19092 --topic > > TestTopic > > SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". > > SLF4J: Defaulting to no-operation (NOP) logger implementation > > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for > further > > details. > > Test > > [2014-04-06 20:42:01,352] WARN Error while fetching metadata > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic > > [TestTopic]: class kafka.common.LeaderNotAvailableException > > (kafka.producer.BrokerPartitionInfo) > > [2014-04-06 20:42:01,364] WARN Error while fetching metadata > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic > > [TestTopic]: class kafka.common.LeaderNotAvailableException > > (kafka.producer.BrokerPartitionInfo) > > [2014-04-06 20:42:01,365] ERROR Failed to collate messages by topic, > > partition due to: Failed to fetch topic metadata for topic: TestTopic > > (kafka.producer.async.DefaultEventHandler) > > [2014-04-06 20:42:01,476] WARN Error while fetching metadata > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic > > [TestTopic]: class kafka.common.LeaderNotAvailableException > > (kafka.producer.BrokerPartitionInfo) > > [2014-04-06 20:42:01,481] WARN Error while fetching metadata > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic > > [TestTopic]: class kafka.common.LeaderNotAvailableException > > (kafka.producer.BrokerPartitionInfo) > > [2014-04-06 20:42:01,481] ERROR Failed to collate messages by topic, > > partition due to: Failed to fetch topic metadata for topic: TestTopic > > (kafka.producer.async.DefaultEventHandler) > > [2014-04-06 20:42:01,589] WARN Error while fetching metadata > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic > > [TestTopic]: class kafka.common.LeaderNotAvailableException > > (kafka.producer.BrokerPartitionInfo) > > [2014-04-06 20:42:01,595] WARN Error while fetching metadata > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic > > [TestTopic]: class kafka.common.LeaderNotAvailableException > > (kafka.producer.BrokerPartitionInfo) > > [2014-04-06 20:42:01,595] ERROR Failed to collate messages by topic, > > partition due to: Failed to fetch topic metadata for topic: TestTopic > > (kafka.producer.async.DefaultEventHandler) > > [2014-04-06 20:42:01,709] WARN Error while fetching metadata > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic > > [TestTopic]: class kafka.common.LeaderNotAvailableException > > (kafka.producer.BrokerPartit > > ionInfo) > > [2014-04-06 20:42:01,719] WARN Error while fetching metadata > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic > > [TestTopic]: class kafka.common.LeaderNotAvailableException > > (kafka.producer.BrokerPartitionInfo) > > [2014-04-06 20:42:01,719] ERROR Failed to collate messages by topic, > > partition due to: Failed to fetch topic metadata for topic: TestTopic > > (kafka.producer.async.DefaultEventHandler) > > [2014-04-06 20:42:01,826] WARN Error while fetching metadata > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic > > [TestTopic]: class kafka.common.LeaderNotAvailableException > > (kafka.producer.BrokerPartitionInfo) > > [2014-04-06 20:42:01,829] ERROR Failed to send requests for topics > > TestTopic with correlation ids in [0,8] > > (kafka.producer.async.DefaultEventHandler) > > [2014-04-06 20:42:01,830] ERROR Error in handling batch of 2 events > > (kafka.producer.async.ProducerSendThread) > > kafka.common.FailedToSendMessageException: Failed to send messages after > 3 > > tries. > > at > > > > > kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90) > > at > > > > > kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104) > > at > > > > > kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87) > > at > > > > > kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:67) > > at scala.collection.immutable.Stream.foreach(Stream.scala:526) > > at > > > > > kafka.producer.async.ProducerSendThread.processEvents(ProducerSendThread.scala:66) > > at > kafka.producer.async.ProducerSendThread.run(ProducerSendThread.scala:44) > > ======================== > > > > > > If advertised setting is not defined, > > producer connection is refused.(below error) > > So I think advertised setting is working. > > > > ======================== > > # ./kafka-console-producer.sh --broker-list localhost:19092 --topic > > TestTopic > > SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". > > SLF4J: Defaulting to no-operation (NOP) logger implementation > > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for > further > > details. > > Test > > [2014-04-08 21:11:44,231] ERROR Producer connection to localhost:9092 > > unsuccessful (kafka.producer.SyncProducer) > > java.net.ConnectException: Connection refused > > at sun.nio.ch.Net.connect0(Native Method) > > at sun.nio.ch.Net.connect(Net.java:465) > > at sun.nio.ch.Net.connect(Net.java:457) > > ======================== > > > > Is there anyone you find the cause of error. > > > > regards. > > > > > > > > > > 2014-03-20 8:00 GMT+09:00 Sotaro Kimura <rfbrin...@gmail.com>: > > > > > Thanks for quick reply. > > > > > > > > > advertised.host.name/advertised.port setting is this. > > > I confirmed only producer setting, so I couldn't find these properties. > > > > > > I will try it. Thank you. > > > > > > > > > 2014-03-20 7:50 GMT+09:00 Joe Stein <joe.st...@stealth.ly>: > > > > > > With 0.8.1 update these two properties in the server.properties config > of > > >> the broker to be 172.16.0.100 and 19092 respectively > > >> > > >> # Hostname the broker will advertise to producers and consumers. If > not > > >> set, it uses the > > >> # value for "host.name" if configured. Otherwise, it will use the > > value > > >> returned from > > >> # java.net.InetAddress.getCanonicalHostName(). > > >> #advertised.host.name=<hostname routable by clients> > > >> > > >> # The port to publish to ZooKeeper for clients to use. If this is not > > set, > > >> # it will publish the same port that the broker binds to. > > >> #advertised.port=<port accessible by clients> > > >> > > >> /******************************************* > > >> Joe Stein > > >> Founder, Principal Consultant > > >> Big Data Open Source Security LLC > > >> http://www.stealth.ly > > >> Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop> > > >> ********************************************/ > > >> > > >> > > >> On Wed, Mar 19, 2014 at 6:39 PM, Sotaro Kimura <rfbrin...@gmail.com> > > >> wrote: > > >> > > >> > Hi, all. > > >> > > > >> > I'm trying Log produce to Remote Kafka Cluster over SSH Tunneling. > > >> > > > >> > Configuration is below. > > >> > - Kafka cluster in DataCenter(each machine has Grobal IP Address). > > >> > Firewall allows only ssh port(22). > > >> > ex) 192.168.100.100 > > >> > - Kafka producer exists out of DataCenter. > > >> > ex) 172.16.0.100 > > >> > > > >> > I tried produce over SSH Tunneling. > > >> > Producer successed getting meta data from brokers. > > >> > But meta data is Address which producer can't access directry. > > >> > > > >> > Detail is below. > > >> > 1. Ssh connect producer machine to DataCenter. > > >> > 172.16.0.100 > 192.168.100.100:22 > > >> > 2. Create SSH Tunnel over ssh connection. > > >> > 172.16.0.100's localhost:19092 to 192.168.100.100:9092(Kafka > receive > > >> > port). > > >> > 3. Producer connect to localhost:19092. > > >> > Producer get metadata from brokers(192.168.100.100:9092). > > >> > 4. Producer failed to connect brokers(192.168.100.100:9092). > > >> > > > >> > In such situations, how I can produce to Remote Kafka Cluster? > > >> > > > >> > -- > > >> > ################################# > > >> > 木村 宗太郎 Sotaro Kimura > > >> > <rfbrin...@gmail.com> > > >> > ################################# > > >> > > > >> > > > > > > > > > > > > -- > > > ################################# > > > 木村 宗太郎 Sotaro Kimura > > > <rfbrin...@gmail.com> > > > ################################# > > > > > > > > > > > -- > > ################################# > > 木村 宗太郎 Sotaro Kimura > > <rfbrin...@gmail.com> > > ################################# > > > -- ################################# 木村 宗太郎 Sotaro Kimura <rfbrin...@gmail.com> #################################