Thanks for the input Ismael. I will try and let you know.
Also need your valuable inputs for the below issue:) i am not able to run kafka-topics.sh(0.9.0.0 version) [root@localhost bin]# ./kafka-topics.sh --list --zookeeper localhost:2181 [2015-12-28 12:41:32,589] WARN SASL configuration failed: javax.security.auth.login.LoginException: No key to store Will continue connection to Zookeeper server without SASL authentication, if Zookeeper server allows it. (org.apache.zookeeper.ClientCnxn) ^Z I am sure the key is present in its keytab file ( I have cross verified using kinit command as well). Am i missing anything while calling the kafka-topics.sh?? On Mon, Dec 28, 2015 at 3:53 PM, Ismael Juma <isma...@gmail.com> wrote: > Hi Prabhu, > > kafka-console-consumer.sh uses the old consumer by default, but only the > new consumer supports security. Use --new-consumer to change this. > > Hope this helps. > > Ismael > On 28 Dec 2015 05:48, "prabhu v" <prabhuvrajp...@gmail.com> wrote: > > > Hi Experts, > > > > I am getting the below error when running the consumer > > "kafka-console-consumer.sh" . > > > > I am using the new version 0.9.0.1. > > Topic name: test > > > > > > [2015-12-28 06:13:34,409] WARN > > > > > [console-consumer-61657_localhost-1451283204993-5512891d-leader-finder-thread], > > Failed to find leader for Set([test,0]) > > (kafka.consumer.ConsumerFetcherManager$LeaderFinderThread) > > kafka.common.BrokerEndPointNotAvailableException: End point PLAINTEXT not > > found for broker 0 > > at kafka.cluster.Broker.getBrokerEndPoint(Broker.scala:136) > > > > > > Please find the current configuration below. > > > > Configuration: > > > > > > [root@localhost config]# grep -v "^#" consumer.properties > > zookeeper.connect=localhost:2181 > > zookeeper.connection.timeout.ms=60000 > > group.id=test-consumer-group > > security.protocol=SASL_PLAINTEXT > > sasl.kerberos.service.name="kafka" > > > > > > [root@localhost config]# grep -v "^#" producer.properties > > metadata.broker.list=localhost:9094,localhost:9095 > > producer.type=sync > > compression.codec=none > > serializer.class=kafka.serializer.DefaultEncoder > > security.protocol=SASL_PLAINTEXT > > sasl.kerberos.service.name="kafka" > > > > [root@localhost config]# grep -v "^#" server1.properties > > > > broker.id=0 > > listeners=SASL_PLAINTEXT://localhost:9094 > > delete.topic.enable=true > > num.network.threads=3 > > num.io.threads=8 > > socket.send.buffer.bytes=102400 > > socket.receive.buffer.bytes=102400 > > socket.request.max.bytes=104857600 > > log.dirs=/data/kafka_2.11-0.9.0.0/kafka-logs > > num.partitions=1 > > num.recovery.threads.per.data.dir=1 > > log.retention.hours=168 > > log.segment.bytes=1073741824 > > log.retention.check.interval.ms=300000 > > log.cleaner.enable=false > > zookeeper.connect=localhost:2181 > > zookeeper.connection.timeout.ms=60000 > > inter.broker.protocol.version=0.9.0.0 > > security.inter.broker.protocol=SASL_PLAINTEXT > > allow.everyone.if.no.acl.found=true > > > > > > [root@localhost config]# grep -v "^#" server4.properties > > broker.id=1 > > listeners=SASL_PLAINTEXT://localhost:9095 > > delete.topic.enable=true > > num.network.threads=3 > > num.io.threads=8 > > socket.send.buffer.bytes=102400 > > socket.receive.buffer.bytes=102400 > > socket.request.max.bytes=104857600 > > log.dirs=/data/kafka_2.11-0.9.0.0/kafka-logs-1 > > num.partitions=1 > > num.recovery.threads.per.data.dir=1 > > log.retention.hours=168 > > log.segment.bytes=1073741824 > > log.retention.check.interval.ms=300000 > > log.cleaner.enable=false > > zookeeper.connect=localhost:2181 > > zookeeper.connection.timeout.ms=60000 > > inter.broker.protocol.version=0.9.0.0 > > security.inter.broker.protocol=SASL_PLAINTEXT > > zookeeper.sasl.client=zkclient > > > > [root@localhost config]# grep -v "^#" zookeeper.properties > > dataDir=/data/kafka_2.11-0.9.0.0/zookeeper > > clientPort=2181 > > maxClientCnxns=0 > > requireClientAuthScheme=sasl > > > authProvider.1=org.apache.zookeeper.server.auth.SASLAuthenticationProvider > > jaasLoginRenew=3600000 > > > > > > Need your valuable inputs on this issue. > > -- > > Regards, > > > > Prabhu.V > > > -- Regards, Prabhu.V