Thanks for the input Ismael.

I will try and let you know.

Also need your valuable inputs for the below issue:)

i am not able to run kafka-topics.sh(0.9.0.0 version)

[root@localhost bin]# ./kafka-topics.sh --list --zookeeper localhost:2181
[2015-12-28 12:41:32,589] WARN SASL configuration failed:
javax.security.auth.login.LoginException: No key to store Will continue
connection to Zookeeper server without SASL authentication, if Zookeeper
server allows it. (org.apache.zookeeper.ClientCnxn)
^Z

I am sure the key is present in its keytab file ( I have cross verified
using kinit command as well).

Am i missing anything while calling the kafka-topics.sh??



On Mon, Dec 28, 2015 at 3:53 PM, Ismael Juma <isma...@gmail.com> wrote:

> Hi Prabhu,
>
> kafka-console-consumer.sh uses the old consumer by default, but only the
> new consumer supports security. Use --new-consumer to change this.
>
> Hope this helps.
>
> Ismael
> On 28 Dec 2015 05:48, "prabhu v" <prabhuvrajp...@gmail.com> wrote:
>
> > Hi Experts,
> >
> > I am getting the below error when running the consumer
> > "kafka-console-consumer.sh" .
> >
> > I am using the new version 0.9.0.1.
> > Topic name: test
> >
> >
> > [2015-12-28 06:13:34,409] WARN
> >
> >
> [console-consumer-61657_localhost-1451283204993-5512891d-leader-finder-thread],
> > Failed to find leader for Set([test,0])
> > (kafka.consumer.ConsumerFetcherManager$LeaderFinderThread)
> > kafka.common.BrokerEndPointNotAvailableException: End point PLAINTEXT not
> > found for broker 0
> >         at kafka.cluster.Broker.getBrokerEndPoint(Broker.scala:136)
> >
> >
> > Please find the current configuration below.
> >
> > Configuration:
> >
> >
> > [root@localhost config]# grep -v "^#" consumer.properties
> > zookeeper.connect=localhost:2181
> > zookeeper.connection.timeout.ms=60000
> > group.id=test-consumer-group
> > security.protocol=SASL_PLAINTEXT
> > sasl.kerberos.service.name="kafka"
> >
> >
> > [root@localhost config]# grep -v "^#" producer.properties
> > metadata.broker.list=localhost:9094,localhost:9095
> > producer.type=sync
> > compression.codec=none
> > serializer.class=kafka.serializer.DefaultEncoder
> > security.protocol=SASL_PLAINTEXT
> > sasl.kerberos.service.name="kafka"
> >
> > [root@localhost config]# grep -v "^#" server1.properties
> >
> > broker.id=0
> > listeners=SASL_PLAINTEXT://localhost:9094
> > delete.topic.enable=true
> > num.network.threads=3
> > num.io.threads=8
> > socket.send.buffer.bytes=102400
> > socket.receive.buffer.bytes=102400
> > socket.request.max.bytes=104857600
> > log.dirs=/data/kafka_2.11-0.9.0.0/kafka-logs
> > num.partitions=1
> > num.recovery.threads.per.data.dir=1
> > log.retention.hours=168
> > log.segment.bytes=1073741824
> > log.retention.check.interval.ms=300000
> > log.cleaner.enable=false
> > zookeeper.connect=localhost:2181
> > zookeeper.connection.timeout.ms=60000
> > inter.broker.protocol.version=0.9.0.0
> > security.inter.broker.protocol=SASL_PLAINTEXT
> > allow.everyone.if.no.acl.found=true
> >
> >
> > [root@localhost config]# grep -v "^#" server4.properties
> > broker.id=1
> > listeners=SASL_PLAINTEXT://localhost:9095
> > delete.topic.enable=true
> > num.network.threads=3
> > num.io.threads=8
> > socket.send.buffer.bytes=102400
> > socket.receive.buffer.bytes=102400
> > socket.request.max.bytes=104857600
> > log.dirs=/data/kafka_2.11-0.9.0.0/kafka-logs-1
> > num.partitions=1
> > num.recovery.threads.per.data.dir=1
> > log.retention.hours=168
> > log.segment.bytes=1073741824
> > log.retention.check.interval.ms=300000
> > log.cleaner.enable=false
> > zookeeper.connect=localhost:2181
> > zookeeper.connection.timeout.ms=60000
> > inter.broker.protocol.version=0.9.0.0
> > security.inter.broker.protocol=SASL_PLAINTEXT
> > zookeeper.sasl.client=zkclient
> >
> > [root@localhost config]# grep -v "^#" zookeeper.properties
> > dataDir=/data/kafka_2.11-0.9.0.0/zookeeper
> > clientPort=2181
> > maxClientCnxns=0
> > requireClientAuthScheme=sasl
> >
> authProvider.1=org.apache.zookeeper.server.auth.SASLAuthenticationProvider
> > jaasLoginRenew=3600000
> >
> >
> > Need your valuable inputs on this issue.
> > --
> > Regards,
> >
> > Prabhu.V
> >
>



-- 
Regards,

Prabhu.V

Reply via email to