I think setting these is not a good idea b/c only apply to the specific
client where you've setup the tunnel.  Other clients cannot use these
settings

advertised.host.name=localhost
advertised.port=19092

You probably need to figure out another way such as
1) Setting up a local mapping on your producer that matches the remote
host:port
    - Edit /etc/hosts to map 172.16.0.100 to localhost
    - Forward localhost:9092 to localhost:19092
2) Change the producer client to translate the broker metadata it gets back
3) Create a proxy on the remote side that does all the talking to Kafka.
 Change your producer to only talk to the proxy



On Wed, Apr 9, 2014 at 8:04 AM, Sotaro Kimura <rfbrin...@gmail.com> wrote:

> I confirmed.
>
> - Controller.log
> After kafka broker stated.
> Kafka broker continues outputting below log.
> Error is continued.
> =======
> [2014-04-09 23:55:07,709] ERROR [Controller-0-to-broker-0-send-thread],
> Controller 0's connection to broker id:0,host:localhost,port:19092 was
> unsuccessful (kafka.controller.RequestSendThread)
> java.net.ConnectException: Connection refused
>         at sun.nio.ch.Net.connect0(Native Method)
>         at sun.nio.ch.Net.connect(Net.java:465)
>         at sun.nio.ch.Net.connect(Net.java:457)
>         at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:670)
>         at kafka.network.BlockingChannel.connect(BlockingChannel.scala:57)
>         at
>
> kafka.controller.RequestSendThread.connectToBroker(ControllerChannelManager.scala:173)
>         at
>
> kafka.controller.RequestSendThread.liftedTree1$1(ControllerChannelManager.scala:140)
>         at
>
> kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:131)
>         at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:51)
> [2014-04-09 23:55:08,009] ERROR [Controller-0-to-broker-0-send-thread],
> Controller 0 epoch 7 failed to send UpdateMetadata request with correlation
> id 3 to broker id:0,host:localhost,port:19092. Reconnecting to broker.
> (kafka.controller.RequestSendThread)
> java.nio.channels.ClosedChannelException
>         at kafka.network.BlockingChannel.send(BlockingChannel.scala:89)
>         at
>
> kafka.controller.RequestSendThread.liftedTree1$1(ControllerChannelManager.scala:132)
>         at
>
> kafka.controller.RequestSendThread.doWork(ControllerChannelManager.scala:131)
>         at kafka.utils.ShutdownableThread.run(ShutdownableThread.scala:51)
> =======
>
> - state change log
> When kafka broker started.
> Below log outputed each partition.
> =======
> [2014-04-09 23:53:43,341] TRACE Controller 0 epoch 7 started leader
> election for partition [TestTopic,0] (state.change.logger)
> kafka.common.NoReplicaOnlineException: No replica for partition
> [TestTopic,0] is alive. Live brokers are: [Set()], Assigned replicas are:
> [List(0)]
>         at
>
> kafka.controller.OfflinePartitionLeaderSelector.selectLeader(PartitionLeaderSelector.scala:61)
>         at
>
> kafka.controller.PartitionStateMachine.electLeaderForPartition(PartitionStateMachine.scala:335)
>         at
>
> kafka.controller.PartitionStateMachine.kafka$controller$PartitionStateMachine$$handleStateChange(PartitionStateMachine.scala:184)
>         at
>
> kafka.controller.PartitionStateMachine$$anonfun$triggerOnlinePartitionStateChange$3.apply(PartitionStateMachine.scala:98)
>         at
>
> kafka.controller.PartitionStateMachine$$anonfun$triggerOnlinePartitionStateChange$3.apply(PartitionStateMachine.scala:95)
>         at
>
> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:743)
>         at
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:95)
>         at
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:95)
>         at scala.collection.Iterator$class.foreach(Iterator.scala:772)
>         at
> scala.collection.mutable.HashTable$$anon$1.foreach(HashTable.scala:157)
>         at
> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:190)
>         at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:45)
>         at scala.collection.mutable.HashMap.foreach(HashMap.scala:95)
>         at
>
> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:742)
>         at
>
> kafka.controller.PartitionStateMachine.triggerOnlinePartitionStateChange(PartitionStateMachine.scala:95)
>         at
>
> kafka.controller.PartitionStateMachine.startup(PartitionStateMachine.scala:68)
>         at
>
> kafka.controller.KafkaController.onControllerFailover(KafkaController.scala:311)
>         at
>
> kafka.controller.KafkaController$$anonfun$1.apply$mcV$sp(KafkaController.scala:161)
>         at
> kafka.server.ZookeeperLeaderElector.elect(ZookeeperLeaderElector.scala:63)
>         at
>
> kafka.server.ZookeeperLeaderElector$$anonfun$startup$1.apply$mcZ$sp(ZookeeperLeaderElector.scala:49)
>         at
>
> kafka.server.ZookeeperLeaderElector$$anonfun$startup$1.apply(ZookeeperLeaderElector.scala:47)
>         at
>
> kafka.server.ZookeeperLeaderElector$$anonfun$startup$1.apply(ZookeeperLeaderElector.scala:47)
>         at kafka.utils.Utils$.inLock(Utils.scala:538)
>         at
>
> kafka.server.ZookeeperLeaderElector.startup(ZookeeperLeaderElector.scala:47)
>         at
>
> kafka.controller.KafkaController$$anonfun$startup$1.apply$mcV$sp(KafkaController.scala:630)
>         at
>
> kafka.controller.KafkaController$$anonfun$startup$1.apply(KafkaController.scala:626)
>         at
>
> kafka.controller.KafkaController$$anonfun$startup$1.apply(KafkaController.scala:626)
>         at kafka.utils.Utils$.inLock(Utils.scala:538)
>         at
> kafka.controller.KafkaController.startup(KafkaController.scala:626)
>         at kafka.server.KafkaServer.startup(KafkaServer.scala:96)
>         at
> kafka.server.KafkaServerStartable.startup(KafkaServerStartable.scala:34)
>         at kafka.Kafka$.main(Kafka.scala:46)
>         at kafka.Kafka.main(Kafka.scala)
> =======
>
> regards.
>
>
> 2014-04-09 13:03 GMT+09:00 Jun Rao <jun...@gmail.com>:
>
> > Any error in the broker, controller and state change log?
> >
> > Thanks,
> >
> > Jun
> >
> >
> > On Tue, Apr 8, 2014 at 3:08 PM, Sotaro Kimura <rfbrin...@gmail.com>
> wrote:
> >
> > > Hi, all.
> > >
> > > Sorry for my late reply.
> > >
> > > I tried advertised.host.name/advertised.port.
> > > But producer failed fetching metadata from kafka broker.
> > >
> > > Configuration is below.
> > > - Kafka cluster in DataCenter(each machine has Grobal IP Address).
> > > Firewall allows only ssh port(22).
> > > ex) 192.168.100.100
> > > - Kafka producer exists out of DataCenter.
> > > ex) 172.16.0.100
> > >
> > > Broker additional setting is below.
> > > advertised.host.name=localhost
> > > advertised.port=19092
> > >
> > > Kafka producer operation is below.
> > >
> > > - SSH port fowarding
> > > ========================
> > > # ssh root@192.168.100.100 -L 19092:192.168.100.100:9092 -L 12181:
> > > 192.168.100.100:2181
> > > // Kafka broker and zookeeper ssh port fowarding.
> > > ========================
> > >
> > > - Producer data put(failed)
> > > ========================
> > > # ./kafka-topics.sh --create --zookeeper localhost:12181
> > > --replication-factor 1 --partitions 3 --topic TestTopic
> > > Created topic "TestTopic".
> > > // Succeed create topic over ssh port fowarding.
> > > # ./kafka-topics.sh --list --zookeeper localhost:12181
> > > TestTopic
> > > // Succeed get topic list over ssh port fowarding.
> > > # ./kafka-console-producer.sh --broker-list localhost:19092 --topic
> > > TestTopic
> > > SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
> > > SLF4J: Defaulting to no-operation (NOP) logger implementation
> > > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for
> > further
> > > details.
> > > Test
> > > [2014-04-06 20:42:01,352] WARN Error while fetching metadata
> > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic
> > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic
> > > [TestTopic]: class kafka.common.LeaderNotAvailableException
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2014-04-06 20:42:01,364] WARN Error while fetching metadata
> > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic
> > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic
> > > [TestTopic]: class kafka.common.LeaderNotAvailableException
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2014-04-06 20:42:01,365] ERROR Failed to collate messages by topic,
> > > partition due to: Failed to fetch topic metadata for topic: TestTopic
> > > (kafka.producer.async.DefaultEventHandler)
> > > [2014-04-06 20:42:01,476] WARN Error while fetching metadata
> > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic
> > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic
> > > [TestTopic]: class kafka.common.LeaderNotAvailableException
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2014-04-06 20:42:01,481] WARN Error while fetching metadata
> > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic
> > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic
> > > [TestTopic]: class kafka.common.LeaderNotAvailableException
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2014-04-06 20:42:01,481] ERROR Failed to collate messages by topic,
> > > partition due to: Failed to fetch topic metadata for topic: TestTopic
> > > (kafka.producer.async.DefaultEventHandler)
> > > [2014-04-06 20:42:01,589] WARN Error while fetching metadata
> > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic
> > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic
> > > [TestTopic]: class kafka.common.LeaderNotAvailableException
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2014-04-06 20:42:01,595] WARN Error while fetching metadata
> > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic
> > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic
> > > [TestTopic]: class kafka.common.LeaderNotAvailableException
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2014-04-06 20:42:01,595] ERROR Failed to collate messages by topic,
> > > partition due to: Failed to fetch topic metadata for topic: TestTopic
> > > (kafka.producer.async.DefaultEventHandler)
> > > [2014-04-06 20:42:01,709] WARN Error while fetching metadata
> > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic
> > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic
> > > [TestTopic]: class kafka.common.LeaderNotAvailableException
> > > (kafka.producer.BrokerPartit
> > > ionInfo)
> > > [2014-04-06 20:42:01,719] WARN Error while fetching metadata
> > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic
> > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic
> > > [TestTopic]: class kafka.common.LeaderNotAvailableException
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2014-04-06 20:42:01,719] ERROR Failed to collate messages by topic,
> > > partition due to: Failed to fetch topic metadata for topic: TestTopic
> > > (kafka.producer.async.DefaultEventHandler)
> > > [2014-04-06 20:42:01,826] WARN Error while fetching metadata
> > > [{TopicMetadata for topic TestTopic ->No partition metadata for topic
> > > TestTopic due to kafka.common.LeaderNotAvailableException}] for topic
> > > [TestTopic]: class kafka.common.LeaderNotAvailableException
> > > (kafka.producer.BrokerPartitionInfo)
> > > [2014-04-06 20:42:01,829] ERROR Failed to send requests for topics
> > > TestTopic with correlation ids in [0,8]
> > > (kafka.producer.async.DefaultEventHandler)
> > > [2014-04-06 20:42:01,830] ERROR Error in handling batch of 2 events
> > > (kafka.producer.async.ProducerSendThread)
> > > kafka.common.FailedToSendMessageException: Failed to send messages
> after
> > 3
> > > tries.
> > > at
> > >
> > >
> >
> kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:90)
> > > at
> > >
> > >
> >
> kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104)
> > > at
> > >
> > >
> >
> kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87)
> > > at
> > >
> > >
> >
> kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:67)
> > > at scala.collection.immutable.Stream.foreach(Stream.scala:526)
> > > at
> > >
> > >
> >
> kafka.producer.async.ProducerSendThread.processEvents(ProducerSendThread.scala:66)
> > > at
> > kafka.producer.async.ProducerSendThread.run(ProducerSendThread.scala:44)
> > > ========================
> > >
> > >
> > > If advertised setting is not defined,
> > > producer connection is refused.(below error)
> > > So I think advertised setting is working.
> > >
> > > ========================
> > > # ./kafka-console-producer.sh --broker-list localhost:19092 --topic
> > > TestTopic
> > > SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder".
> > > SLF4J: Defaulting to no-operation (NOP) logger implementation
> > > SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder for
> > further
> > > details.
> > > Test
> > > [2014-04-08 21:11:44,231] ERROR Producer connection to localhost:9092
> > > unsuccessful (kafka.producer.SyncProducer)
> > > java.net.ConnectException: Connection refused
> > > at sun.nio.ch.Net.connect0(Native Method)
> > > at sun.nio.ch.Net.connect(Net.java:465)
> > > at sun.nio.ch.Net.connect(Net.java:457)
> > > ========================
> > >
> > > Is there anyone you find the cause of error.
> > >
> > > regards.
> > >
> > >
> > >
> > >
> > > 2014-03-20 8:00 GMT+09:00 Sotaro Kimura <rfbrin...@gmail.com>:
> > >
> > > > Thanks for quick reply.
> > > >
> > > >
> > > > advertised.host.name/advertised.port setting is this.
> > > > I confirmed only producer setting, so I couldn't find these
> properties.
> > > >
> > > > I will try it. Thank you.
> > > >
> > > >
> > > > 2014-03-20 7:50 GMT+09:00 Joe Stein <joe.st...@stealth.ly>:
> > > >
> > > > With 0.8.1 update these two properties in the server.properties
> config
> > of
> > > >> the broker to be 172.16.0.100 and 19092 respectively
> > > >>
> > > >> # Hostname the broker will advertise to producers and consumers. If
> > not
> > > >> set, it uses the
> > > >> # value for "host.name" if configured.  Otherwise, it will use the
> > > value
> > > >> returned from
> > > >> # java.net.InetAddress.getCanonicalHostName().
> > > >> #advertised.host.name=<hostname routable by clients>
> > > >>
> > > >> # The port to publish to ZooKeeper for clients to use. If this is
> not
> > > set,
> > > >> # it will publish the same port that the broker binds to.
> > > >> #advertised.port=<port accessible by clients>
> > > >>
> > > >> /*******************************************
> > > >>  Joe Stein
> > > >>  Founder, Principal Consultant
> > > >>  Big Data Open Source Security LLC
> > > >>  http://www.stealth.ly
> > > >>  Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop>
> > > >> ********************************************/
> > > >>
> > > >>
> > > >> On Wed, Mar 19, 2014 at 6:39 PM, Sotaro Kimura <rfbrin...@gmail.com
> >
> > > >> wrote:
> > > >>
> > > >> > Hi, all.
> > > >> >
> > > >> > I'm trying Log produce to Remote Kafka Cluster over SSH Tunneling.
> > > >> >
> > > >> > Configuration is below.
> > > >> > - Kafka cluster in DataCenter(each machine has Grobal IP Address).
> > > >> > Firewall allows only ssh port(22).
> > > >> > ex) 192.168.100.100
> > > >> > - Kafka producer exists out of DataCenter.
> > > >> > ex) 172.16.0.100
> > > >> >
> > > >> > I tried produce over SSH Tunneling.
> > > >> > Producer successed getting meta data from brokers.
> > > >> > But meta data is Address which producer can't access directry.
> > > >> >
> > > >> > Detail is below.
> > > >> > 1. Ssh connect producer machine to DataCenter.
> > > >> > 172.16.0.100 > 192.168.100.100:22
> > > >> > 2. Create SSH Tunnel over ssh connection.
> > > >> > 172.16.0.100's localhost:19092 to 192.168.100.100:9092(Kafka
> > receive
> > > >> > port).
> > > >> > 3. Producer connect to localhost:19092.
> > > >> > Producer get metadata from brokers(192.168.100.100:9092).
> > > >> > 4. Producer failed to connect brokers(192.168.100.100:9092).
> > > >> >
> > > >> > In such situations, how I can produce to Remote Kafka Cluster?
> > > >> >
> > > >> > --
> > > >> > #################################
> > > >> > 木村 宗太郎 Sotaro Kimura
> > > >> > <rfbrin...@gmail.com>
> > > >> > #################################
> > > >> >
> > > >>
> > > >
> > > >
> > > >
> > > > --
> > > > #################################
> > > > 木村 宗太郎 Sotaro Kimura
> > > > <rfbrin...@gmail.com>
> > > > #################################
> > > >
> > >
> > >
> > >
> > > --
> > > #################################
> > > 木村 宗太郎 Sotaro Kimura
> > > <rfbrin...@gmail.com>
> > > #################################
> > >
> >
>
>
>
> --
> #################################
> 木村 宗太郎 Sotaro Kimura
> <rfbrin...@gmail.com>
> #################################
>

Reply via email to