Even after changing the fetch wait max ms the same thing is repeting
just that some partitions have the owners now, i mean
bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker --group group1
--zkconnect zkhost:zkport --topic testtopic
Group Topic Pid Offset
logSize Lag Owner
group1 testtopic 0 253 253 0
group1_xxxx-1397216047177-6f419d28-0
group1 testtopic 1 268 268 0
group1_xxxx-1397216047177-6f419d28-1
group1 testtopic 2 258 258 0
group1_xxxx-1397216047177-6f419d28-2
group1 testtopic 3 265 265 0 none
group1 testtopic 4 262 262 0 none
group1 testtopic 5 296 296 0 none
group1 testtopic 6 249 249 0
group1_xxxx1-1397219018735-a2cf60df-0
group1 testtopic 7 272 272 0
group1_xxxx1-1397219018735-a2cf60df-1
group1 testtopic 8 242 242 0
group1_xxxx1-1397219018735-a2cf60df-2
group1 testtopic 9 263 263 0
group1_xxxx1-1397279906839-12366659-0
group1 testtopic 10 294 294 0
group1_xxxx1-1397279906839-12366659-1
group1 testtopic 11 254 254 0
group1_xxxx1-1397279906839-12366659-2
and i dont see any exception after
"end rebalancing consumer group1_ip-XXXX-82095a28 try #9"
i just see the below log
12 Apr 2014 00:18:58,987 INFO [Thread-321]
[kafka.consumer.ZookeeperConsumerConnector] [group1_ip-xxxx-82095a28],
Rebalancing attempt failed. Clearing the cache before the next
rebalancing operation is triggered
12 Apr 2014 00:18:58,987 INFO [Thread-321]
[kafka.consumer.ConsumerFetcherManager]
[ConsumerFetcherManager-1397279912725] Stopping leader finder thread
12 Apr 2014 00:18:58,987 INFO [Thread-321]
[kafka.consumer.ConsumerFetcherManager]
[ConsumerFetcherManager-1397279912725] Stopping all fetchers
12 Apr 2014 00:18:58,987 INFO [Thread-321]
[kafka.consumer.ConsumerFetcherManager]
[ConsumerFetcherManager-1397279912725] All connections stopped
12 Apr 2014 00:18:58,987 INFO [Thread-321]
[kafka.consumer.ZookeeperConsumerConnector] [group1_xxxx-82095a28],
Cleared all relevant queues for this fetcher
12 Apr 2014 00:18:58,988 DEBUG [Thread-321]
[kafka.consumer.ConsumerIterator] Clearing the current data chunk for
this consumer iterator
12 Apr 2014 00:18:58,988 DEBUG [Thread-321]
[kafka.consumer.ConsumerIterator] Clearing the current data chunk for
this consumer iterator
12 Apr 2014 00:18:58,988 DEBUG [Thread-321]
[kafka.consumer.ConsumerIterator] Clearing the current data chunk for
this consumer iterator
12 Apr 2014 00:18:58,988 INFO [Thread-321]
[kafka.consumer.ZookeeperConsumerConnector] [group1_xxxx-82095a28],
Cleared the data chunks in all the consumer message iterators
12 Apr 2014 00:18:58,988 INFO [Thread-321]
[kafka.consumer.ZookeeperConsumerConnector] [group1_xxxx-82095a28],
Committing all offsets after clearing the fetcher queues
after this kafka consumer will not log a thing.Not even the exception.(I
have put my consumer log level to debug)
Thanks
Arjun Narasimha Kota
On Saturday 12 April 2014 08:41 AM, Jun Rao wrote:
Console consumer also uses the high level consumer. Could you try setting
fetch.wait.max.ms to 100ms?
Thanks,
Jun
On Fri, Apr 11, 2014 at 9:56 AM, Arjun Kota <ar...@socialtwist.com> wrote:
Console consumer works fine. Its the high level java consumer which is
giving this problem.
Thanks
Arjun narasimha kota
On Apr 11, 2014 8:42 PM, "Jun Rao" <jun...@gmail.com> wrote:
We may have a bug that doesn't observe etch.min.bytes accurately. So a
lower fetch.wait.max.ms will improve consumer latency.
Could you run a console consumer and see if you have the same issue? That
will tell us if this is a server side issue or an issue just in your
consumer.
Thanks,
Jun
On Thu, Apr 10, 2014 at 10:28 PM, Arjun <ar...@socialtwist.com> wrote:
i changed the time to 60 seconds even now i see the same result. The
Consumer is not consuming the messages.
Thanks
Arjun Narasimha Kota
On Friday 11 April 2014 10:36 AM, Arjun wrote:
yup i will change the value and recheck. Thanks for the help.
thanks
Arjun Narasimha Kota
On Friday 11 April 2014 10:28 AM, Guozhang Wang wrote:
What I tried to say is that it may be caused by your
"fetch.wait.max.ms"="180000"
too large. Try a small value and see if that helps.
On Thu, Apr 10, 2014 at 9:44 PM, Arjun <ar...@socialtwist.com>
wrote:
Hi,
I could not see any out of memory exceptions in the broker logs. One
thing
i can see is i may have configured consumer poorly. If its not too
much to
ask can u let me know the changes i have to do for over coming this
problem.
Thanks
Arjun Narasimha Kota
On Friday 11 April 2014 10:04 AM, Guozhang Wang wrote:
Hi Ajrun,
It seems to be the cause:
https://issues.apache.org/jira/browse/KAFKA-1016
Guozhang
On Thu, Apr 10, 2014 at 9:21 PM, Arjun <ar...@socialtwist.com>
wrote:
I hope this one would give u a better idea.
bin/kafka-run-class.sh kafka.tools.ConsumerOffsetChecker --group
group1
--zkconnect zkhost:port --topic testtopic
Group Topic Pid Offset logSize
Lag Owner
group1 testtopic 0 253 253 0
group1_ip-xx-1397188061429-b5ff1205-0
group1 testtopic 1 267 267 0
group1_ip-xx-1397188061429-b5ff1205-0
group1 testtopic 2 254 254 0
group1_ip-xx-1397188061429-b5ff1205-0
group1 testtopic 3 265 265 0
group1_ip-xx-1397188061429-b5ff1205-0
group1 testtopic 4 261 261 0
group1_ip-xx-1397188061429-b5ff1205-1
group1 testtopic 5 294 294 0
group1_ip-xx-1397188061429-b5ff1205-1
group1 testtopic 6 248 248 0
group1_ip-xx-1397188061429-b5ff1205-1
group1 testtopic 7 271 271 0
group1_ip-xx-1397188061429-b5ff1205-1
group1 testtopic 8 240 240 0
group1_ip-xx-1397188061429-b5ff1205-2
group1 testtopic 9 261 261 0
group1_ip-xx-1397188061429-b5ff1205-2
group1 testtopic 10 290 290 0
group1_ip-xx-1397188061429-b5ff1205-2
group1 testtopic 11 250 251 1
group1_ip-xx-1397188061429-b5ff1205-2
If you see the output, in the last line the lag is 1 for that
partition.
I
just send one message. This topic is not new as you see there are
lot
of
messages which have accumlated from yesterday. This one message
will
not
be
consumed by consumer what so ever. But if i send some 10 messages
then
all
the messages are consumed.
Please let me know if i have to change any consumer properties.
My consumer properties are :
"fetch.wait.max.ms"="180000"
"fetch.min.bytes" = "1"
"auto.offset.reset" = "smallest"
"auto.commit.enable"= "false"
"fetch.message.max.bytes" = "1048576"
Thanks
Arjun Narasimha Kota
On Friday 11 April 2014 06:23 AM, Arjun Kota wrote:
The consumer uses do specific topics.
On Apr 11, 2014 6:23 AM, "Arjun Kota" <ar...@socialtwist.com
<mailto:
ar...@socialtwist.com>> wrote:
Yes the message shows up on the server.
On Apr 11, 2014 12:07 AM, "Guozhang Wang" <
wangg...@gmail.com
<mailto:wangg...@gmail.com>> wrote:
Hi Arjun,
If you only send one message, does that message show up
on
the
server? Does
you consumer use wildcard topics or specific topics?
Guozhang
On Thu, Apr 10, 2014 at 9:20 AM, Arjun <
ar...@socialtwist.com
<mailto:ar...@socialtwist.com>> wrote:
> But we have auto offset reset to smallest not
largest,
even
then this
> issue arises? If so is there any work around?
>
> Thanks
> Arjun NArasimha Kota
>
>
> On Thursday 10 April 2014 09:39 PM, Guozhang Wang
wrote:
>
>> It could be https://issues.apache.org/
jira/browse/KAFKA-1006.
>>
>> Guozhang
>>
>>
>> On Thu, Apr 10, 2014 at 8:50 AM, Arjun
<ar...@socialtwist.com <mailto:ar...@socialtwist.com>>
wrote:
>>
>> its auto created
>>> but even after topic creation this is the scenario
>>>
>>> Arjun
>>>
>>> On Thursday 10 April 2014 08:41 PM, Guozhang Wang
wrote:
>>>
>>> Hi Arjun,
>>>>
>>>> Did you manually create the topic or use
auto.topic.creation?
>>>>
>>>> Guozhang
>>>>
>>>>
>>>> On Thu, Apr 10, 2014 at 7:39 AM, Arjun
<ar...@socialtwist.com <mailto:ar...@socialtwist.com>>
wrote:
>>>>
>>>> Hi,
>>>>
>>>>> We have 3 node kafka 0.8 setup with zookeepers
ensemble.
We use high
>>>>> level
>>>>> consumer with auto commit offset false. I am
facing
some
peculiar
>>>>> problem
>>>>> with kafka. When i send some 10-20 messages or so
the
consumer starts
>>>>> to
>>>>> consume the messages. But if i send only one
message to
kafka, then
>>>>> even
>>>>> though consumer is active it is not trying to
fetch
the
message. There
>>>>> is
>>>>> nothing in logs, just the messages are being
fetched
by
the kafka
>>>>> consumer.
>>>>> The messages are there in the Kafka server. Can
some
one
let me know
>>>>> where
>>>>> i am doing wrong.
>>>>>
>>>>>
>>>>> Thanks
>>>>> Arjun Narasimha Kota
>>>>>
>>>>>
>>>>>
>>>>
>>
>
--
-- Guozhang