Robert, Sorry I gave the information about wrong topic. Here is the right one.
balajirajagopalan@megatron-server02:/usr/share/kafka_2.11-0.8.2.1/bin$ ./kafka-topics.sh --describe --topic location --zookeeper (someserver) Topic:location PartitionCount:150 ReplicationFactor:1 Configs: Topic: location Partition: 0 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 1 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 2 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 3 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 4 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 5 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 6 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 7 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 8 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 9 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 10 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 11 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 12 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 13 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 14 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 15 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 16 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 17 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 18 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 19 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 20 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 21 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 22 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 23 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 24 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 25 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 26 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 27 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 28 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 29 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 30 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 31 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 32 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 33 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 34 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 35 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 36 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 37 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 38 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 39 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 40 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 41 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 42 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 43 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 44 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 45 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 46 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 47 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 48 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 49 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 50 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 51 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 52 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 53 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 54 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 55 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 56 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 57 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 58 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 59 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 60 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 61 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 62 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 63 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 64 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 65 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 66 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 67 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 68 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 69 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 70 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 71 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 72 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 73 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 74 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 75 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 76 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 77 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 78 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 79 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 80 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 81 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 82 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 83 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 84 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 85 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 86 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 87 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 88 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 89 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 90 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 91 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 92 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 93 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 94 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 95 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 96 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 97 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 98 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 99 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 100 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 101 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 102 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 103 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 104 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 105 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 106 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 107 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 108 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 109 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 110 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 111 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 112 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 113 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 114 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 115 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 116 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 117 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 118 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 119 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 120 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 121 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 122 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 123 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 124 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 125 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 126 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 127 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 128 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 129 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 130 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 131 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 132 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 133 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 134 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 135 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 136 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 137 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 138 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 139 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 140 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 141 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 142 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 143 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 144 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 145 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 146 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 147 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 148 Leader: 0 Replicas: 0 Isr: 0 Topic: location Partition: 149 Leader: 0 Replicas: 0 Isr: 0 On Wed, Apr 20, 2016 at 10:35 PM, Balaji Rajagopalan < balaji.rajagopa...@olacabs.com> wrote: > /usr/share/kafka_2.11-0.8.2.1/bin$ ./kafka-topics.sh --describe --topic > capi --zookeeper (someserver) > > Topic:capi PartitionCount:1 ReplicationFactor:1 Configs: > > Topic: capi Partition: 0 Leader: 0 Replicas: 0 Isr: 0 > > > There are no events to consume from this topic, this I confirm by running > the console consumer. > > ./kafka-console-consumer.sh --topic topicname --zookeeper (some server) > > > The flink connector is other consumer. This is happening in our > pre-production machines consistently, I will also try to reproduce this > locally. > > java.lang.RuntimeException: Unable to find a leader for partitions: > [FetchPartition {topic=location, partition=58, offset=-915623761776}, > FetchPartition {topic=location, partition=60, offset=-915623761776}, > FetchPartition {topic=location, partition=54, offset=-915623761776}, > FetchPartition {topic=location, partition=56, offset=-915623761776}, > FetchPartition {topic=location, partition=66, offset=-915623761776}, > FetchPartition {topic=location, partition=68, offset=-915623761776}, > FetchPartition {topic=location, partition=62, offset=-915623761776}, > FetchPartition {topic=location, partition=64, offset=-915623761776}, > FetchPartition {topic=location, partition=74, offset=-915623761776}, > FetchPartition {topic=location, partition=76, offset=-915623761776}, > FetchPartition {topic=location, partition=70, offset=-915623761776}, > FetchPartition {topic=location, partition=72, offset=-915623761776}, > FetchPartition {topic=location, partition=82, offset=-915623761776}, > FetchPartition {topic=location, partition=84, offset=-915623761776}, > FetchPartition {topic=location, partition=78, offset=-915623761776}, > FetchPartition {topic=location, partition=80, offset=-915623761776}, > FetchPartition {topic=location, partition=26, offset=-915623761776}, > FetchPartition {topic=location, partition=28, offset=-915623761776}, > FetchPartition {topic=location, partition=22, offset=-915623761776}, > FetchPartition {topic=location, partition=24, offset=-915623761776}, > FetchPartition {topic=location, partition=34, offset=-915623761776}, > FetchPartition {topic=location, partition=36, offset=-915623761776}, > FetchPartition {topic=location, partition=30, offset=-915623761776}, > FetchPartition {topic=location, partition=32, offset=-915623761776}, > FetchPartition {topic=location, partition=42, offset=-915623761776}, > FetchPartition {topic=location, partition=44, offset=-915623761776}, > FetchPartition {topic=location, partition=38, offset=-915623761776}, > FetchPartition {topic=location, partition=40, offset=-915623761776}, > FetchPartition {topic=location, partition=50, offset=-915623761776}, > FetchPartition {topic=location, partition=52, offset=-915623761776}, > FetchPartition {topic=location, partition=46, offset=-915623761776}, > FetchPartition {topic=location, partition=48, offset=-915623761776}, > FetchPartition {topic=location, partition=122, offset=-915623761776}, > FetchPartition {topic=location, partition=124, offset=-915623761776}, > FetchPartition {topic=location, partition=118, offset=-915623761776}, > FetchPartition {topic=location, partition=120, offset=-915623761776}, > FetchPartition {topic=location, partition=2, offset=-915623761776}, > FetchPartition {topic=location, partition=130, offset=-915623761776}, > FetchPartition {topic=location, partition=4, offset=-915623761776}, > FetchPartition {topic=location, partition=132, offset=-915623761776}, > FetchPartition {topic=location, partition=126, offset=-915623761776}, > FetchPartition {topic=location, partition=0, offset=-915623761776}, > FetchPartition {topic=location, partition=128, offset=-915623761776}, > FetchPartition {topic=location, partition=10, offset=-915623761776}, > FetchPartition {topic=location, partition=138, offset=-915623761776}, > FetchPartition {topic=location, partition=12, offset=-915623761776}, > FetchPartition {topic=location, partition=140, offset=-915623761776}, > FetchPartition {topic=location, partition=6, offset=-915623761776}, > FetchPartition {topic=location, partition=134, offset=-915623761776}, > FetchPartition {topic=location, partition=8, offset=-915623761776}, > FetchPartition {topic=location, partition=136, offset=-915623761776}, > FetchPartition {topic=location, partition=18, offset=-915623761776}, > FetchPartition {topic=location, partition=146, offset=-915623761776}, > FetchPartition {topic=location, partition=20, offset=-915623761776}, > FetchPartition {topic=location, partition=148, offset=-915623761776}, > FetchPartition {topic=location, partition=14, offset=-915623761776}, > FetchPartition {topic=location, partition=142, offset=-915623761776}, > FetchPartition {topic=location, partition=16, offset=-915623761776}, > FetchPartition {topic=location, partition=144, offset=-915623761776}, > FetchPartition {topic=location, partition=90, offset=-915623761776}, > FetchPartition {topic=location, partition=92, offset=-915623761776}, > FetchPartition {topic=location, partition=86, offset=-915623761776}, > FetchPartition {topic=location, partition=88, offset=-915623761776}, > FetchPartition {topic=location, partition=98, offset=-915623761776}, > FetchPartition {topic=location, partition=100, offset=-915623761776}, > FetchPartition {topic=location, partition=94, offset=-915623761776}, > FetchPartition {topic=location, partition=96, offset=-915623761776}, > FetchPartition {topic=location, partition=106, offset=-915623761776}, > FetchPartition {topic=location, partition=108, offset=-915623761776}, > FetchPartition {topic=location, partition=102, offset=-915623761776}, > FetchPartition {topic=location, partition=104, offset=-915623761776}, > FetchPartition {topic=location, partition=114, offset=-915623761776}, > FetchPartition {topic=location, partition=116, offset=-915623761776}, > FetchPartition {topic=location, partition=110, offset=-915623761776}, > FetchPartition {topic=location, partition=112, offset=-915623761776}] > > at > org.apache.flink.streaming.connectors.kafka.internals.LegacyFetcher.findLeaderForPartitions(LegacyFetcher.java:323) > > at > org.apache.flink.streaming.connectors.kafka.internals.LegacyFetcher.run(LegacyFetcher.java:162) > > at > org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.run(FlinkKafkaConsumer08.java:316) > > at > org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:78) > > at > org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:56) > > at > org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:224) > > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559) > > at java.lang.Thread.run(Thread.java:745) > > On Wed, Apr 20, 2016 at 1:12 PM, Robert Metzger <rmetz...@apache.org> > wrote: > >> Hi, >> I just tried it with Kafka 0.8.2.0 and 0.8.2.1 and for both versions >> everything worked fine. >> How many partitions does your topic have? >> >> Can you send me the full logs of the Kafka consumer? >> >> On Tue, Apr 19, 2016 at 6:05 PM, Balaji Rajagopalan < >> balaji.rajagopa...@olacabs.com> wrote: >> >>> Flink version : 1.0.0 >>> Kafka version : 0.8.2.1 >>> >>> Try to use a topic which has no message posted to it, at the time flink >>> starts. >>> >>> On Tue, Apr 19, 2016 at 5:41 PM, Robert Metzger <rmetz...@apache.org> >>> wrote: >>> >>>> Can you provide me with the exact Flink and Kafka versions you are >>>> using and the steps to reproduce the issue? >>>> >>>> On Tue, Apr 19, 2016 at 2:06 PM, Balaji Rajagopalan < >>>> balaji.rajagopa...@olacabs.com> wrote: >>>> >>>>> It does not seem to fully work if there is no data in the kafka >>>>> stream, the flink application emits this error and bails, could this be >>>>> missed use case in the fix. >>>>> >>>>> On Tue, Apr 19, 2016 at 3:58 PM, Robert Metzger <rmetz...@apache.org> >>>>> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> I'm sorry, the documentation in the JIRA issue is a bit incorrect. >>>>>> The issue has been fixed in all versions including and after 1.0.0. >>>>>> Earlier >>>>>> releases (0.10, 0.9) will fail when the leader changes. >>>>>> However, you don't necessarily need to upgrade to Flink 1.0.0 to >>>>>> resolve the issue: With checkpointing enabled, your job will fail on a >>>>>> leader change, then Flink will restart the Kafka consumers and they'll >>>>>> find >>>>>> the new leaders. >>>>>> Starting from Flink 1.0.0 the Kafka consumer will handle leader >>>>>> changes without failing. >>>>>> >>>>>> Regards, >>>>>> Robert >>>>>> >>>>>> On Tue, Apr 19, 2016 at 12:17 PM, Balaji Rajagopalan < >>>>>> balaji.rajagopa...@olacabs.com> wrote: >>>>>> >>>>>>> I am facing this exception repeatedly while trying to consume from >>>>>>> kafka topic. It seems it was reported in 1.0.0 and fixed in 1.0.0, how >>>>>>> can >>>>>>> I be sure that is fixed in the version of flink that I am using, does >>>>>>> it >>>>>>> require me to install patch updates ? >>>>>>> >>>>>>> Caused by: java.lang.RuntimeException: Unable to find a leader for >>>>>>> partitions: [FetchPartition {topic=capi, partition=0, >>>>>>> offset=-915623761776}] >>>>>>> at >>>>>>> org.apache.flink.streaming.connectors.kafka.internals.LegacyFetcher.findLeaderForPartitions(LegacyFetcher.java:323) >>>>>>> at >>>>>>> org.apache.flink.streaming.connectors.kafka.internals.LegacyFetcher.run(LegacyFetcher.java:162) >>>>>>> at >>>>>>> org.apache.flink.streaming.connectors.kafka.FlinkKafkaConsumer08.run(FlinkKafkaConsumer08.java:316) >>>>>>> at >>>>>>> org.apache.flink.streaming.api.operators.StreamSource.run(StreamSource.java:78) >>>>>>> at >>>>>>> org.apache.flink.streaming.runtime.tasks.SourceStreamTask.run(SourceStreamTask.java:56) >>>>>>> at >>>>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:224) >>>>>>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:559) >>>>>>> at java.lang.Thread.run(Thread.java:745) >>>>>>> >>>>>>> https://issues.apache.org/jira/browse/FLINK-3368 >>>>>>> >>>>>> >>>>>> >>>>> >>>> >>> >> >