Hi Prabhu, kafka-console-consumer.sh uses the old consumer by default, but only the new consumer supports security. Use --new-consumer to change this.
Hope this helps. Ismael On 28 Dec 2015 05:48, "prabhu v" <prabhuvrajp...@gmail.com> wrote: > Hi Experts, > > I am getting the below error when running the consumer > "kafka-console-consumer.sh" . > > I am using the new version 0.9.0.1. > Topic name: test > > > [2015-12-28 06:13:34,409] WARN > > [console-consumer-61657_localhost-1451283204993-5512891d-leader-finder-thread], > Failed to find leader for Set([test,0]) > (kafka.consumer.ConsumerFetcherManager$LeaderFinderThread) > kafka.common.BrokerEndPointNotAvailableException: End point PLAINTEXT not > found for broker 0 > at kafka.cluster.Broker.getBrokerEndPoint(Broker.scala:136) > > > Please find the current configuration below. > > Configuration: > > > [root@localhost config]# grep -v "^#" consumer.properties > zookeeper.connect=localhost:2181 > zookeeper.connection.timeout.ms=60000 > group.id=test-consumer-group > security.protocol=SASL_PLAINTEXT > sasl.kerberos.service.name="kafka" > > > [root@localhost config]# grep -v "^#" producer.properties > metadata.broker.list=localhost:9094,localhost:9095 > producer.type=sync > compression.codec=none > serializer.class=kafka.serializer.DefaultEncoder > security.protocol=SASL_PLAINTEXT > sasl.kerberos.service.name="kafka" > > [root@localhost config]# grep -v "^#" server1.properties > > broker.id=0 > listeners=SASL_PLAINTEXT://localhost:9094 > delete.topic.enable=true > num.network.threads=3 > num.io.threads=8 > socket.send.buffer.bytes=102400 > socket.receive.buffer.bytes=102400 > socket.request.max.bytes=104857600 > log.dirs=/data/kafka_2.11-0.9.0.0/kafka-logs > num.partitions=1 > num.recovery.threads.per.data.dir=1 > log.retention.hours=168 > log.segment.bytes=1073741824 > log.retention.check.interval.ms=300000 > log.cleaner.enable=false > zookeeper.connect=localhost:2181 > zookeeper.connection.timeout.ms=60000 > inter.broker.protocol.version=0.9.0.0 > security.inter.broker.protocol=SASL_PLAINTEXT > allow.everyone.if.no.acl.found=true > > > [root@localhost config]# grep -v "^#" server4.properties > broker.id=1 > listeners=SASL_PLAINTEXT://localhost:9095 > delete.topic.enable=true > num.network.threads=3 > num.io.threads=8 > socket.send.buffer.bytes=102400 > socket.receive.buffer.bytes=102400 > socket.request.max.bytes=104857600 > log.dirs=/data/kafka_2.11-0.9.0.0/kafka-logs-1 > num.partitions=1 > num.recovery.threads.per.data.dir=1 > log.retention.hours=168 > log.segment.bytes=1073741824 > log.retention.check.interval.ms=300000 > log.cleaner.enable=false > zookeeper.connect=localhost:2181 > zookeeper.connection.timeout.ms=60000 > inter.broker.protocol.version=0.9.0.0 > security.inter.broker.protocol=SASL_PLAINTEXT > zookeeper.sasl.client=zkclient > > [root@localhost config]# grep -v "^#" zookeeper.properties > dataDir=/data/kafka_2.11-0.9.0.0/zookeeper > clientPort=2181 > maxClientCnxns=0 > requireClientAuthScheme=sasl > authProvider.1=org.apache.zookeeper.server.auth.SASLAuthenticationProvider > jaasLoginRenew=3600000 > > > Need your valuable inputs on this issue. > -- > Regards, > > Prabhu.V >