Pablo, if you see this again, can you take a thread-dump of your
consumer and verify that the fetchers to all the brokers are still
alive as well as the corresponding iterator threads? It could be that
your consumer ran into some decoder error or some other exception
(although in general that should show up in the log).

Thanks,

Joel

On Thu, Jul 24, 2014 at 03:47:58PM -0400, Joe Stein wrote:
> For the consumer you should see logs like
> 
> "Connecting to zookeeper instance at " + config.zkConnect
> "begin registering consumer " + consumerIdString + " in ZK
> consumerThreadId + " successfully owned partition " + partition + " for
> topic " + topic
> "starting auto committer every " + config.autoCommitIntervalMs + " ms"
> 
> all coming from kafka.consumer.ZookeeperConsumerConnector
> 
> /*******************************************
>  Joe Stein
>  Founder, Principal Consultant
>  Big Data Open Source Security LLC
>  http://www.stealth.ly
>  Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop>
> ********************************************/
> 
> 
> On Thu, Jul 24, 2014 at 3:30 PM, Pablo Picko <p...@pitchsider.com> wrote:
> 
> > Hey guys..
> >
> > I have my my log level set to info, saying that I am not seeing much logs
> > at all for kafka on startup i see detail about the serializer.class my
> > producer uses but very little consumer related logs is there anything I
> > should always see if my log config is correct for the info level
> >
> > In relation to my settings for the number of streams here is my code.
> >
> > *Map<String, Integer> topicCountMap = new HashMap<>();*
> >
> > *//i've 20 threads for the topic for the 20 partitions with no replicas*
> >
> > *topicCountMap.put(topicConsumer.getTopic(),
> > topicConsumer.getNumThreads());*
> >
> >
> > *consumerConnector.createMessageStreams(topicCountMap);*
> >
> > Thanks
> >
> > Pablo
> >
> >
> > On 24 Jul 2014 18:34, "Joe Stein" <joe.st...@stealth.ly> wrote:
> >
> > > What is the value for what you are setting for your number of streams
> > when
> > > calling createMessageStreamsByFilter or if using createMessageStreams for
> > > the TopicCount ( topic -> numberOfStreams )?
> > >
> > > How are you threading the iterator on each stream?
> > >
> > > /*******************************************
> > >  Joe Stein
> > >  Founder, Principal Consultant
> > >  Big Data Open Source Security LLC
> > >  http://www.stealth.ly
> > >  Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop>
> > > ********************************************/
> > >
> > >
> > > On Thu, Jul 24, 2014 at 1:05 PM, Pablo Picko <p...@pitchsider.com>
> > wrote:
> > >
> > > > Guozhang
> > > >
> > > > I didn't no. I did spot other people with similar symptoms to my
> > problem
> > > > mentioning your suggestion too but I don't see anything in the log to
> > > > suggest it rebalanced. It could very well be the reason but I can't see
> > > > anything suggesting it is yet.
> > > >
> > > > Thanks
> > > > Pablo
> > > > On 24 Jul 2014 17:57, "Guozhang Wang" <wangg...@gmail.com> wrote:
> > > >
> > > > > Pablo,
> > > > >
> > > > > Do you see any rebalance related logs in consumers?
> > > > >
> > > > > Guozhang
> > > > >
> > > > >
> > > > > On Thu, Jul 24, 2014 at 9:02 AM, Pablo Picko <p...@pitchsider.com>
> > > > wrote:
> > > > >
> > > > > > Hey Guozhang
> > > > > >
> > > > > > Thanks for the reply, No nothing at all in the logs to suggest
> > > anything
> > > > > > went wrong.
> > > > > >
> > > > > > Its really puzzling as to what's happened. When I restarted the
> > > > consumer
> > > > > > everything worked again.
> > > > > >
> > > > > > Prior to the restart I even stopped the producer for a bit. However
> > > any
> > > > > > messages that got assigned to the Broker C never got processed.
> > When
> > > I
> > > > > ran
> > > > > > the console consumer script before I restarted it was able to all
> > > print
> > > > > all
> > > > > > messages, including messages on broker C. It seems to be that for
> > my
> > > > > > consumers consumergroup one of the brokers messages just became
> > > > > > inaccessible.
> > > > > >
> > > > > > Thanks
> > > > > > Pablo
> > > > > >
> > > > > >
> > > > > > On Thu, Jul 24, 2014 at 4:12 PM, Guozhang Wang <wangg...@gmail.com
> > >
> > > > > wrote:
> > > > > >
> > > > > > > Hi Pablo,
> > > > > > >
> > > > > > > During the period did you see any exception/errors on Broker C's
> > > logs
> > > > > and
> > > > > > > the consumer logs also?
> > > > > > >
> > > > > > > Guozhang
> > > > > > >
> > > > > > >
> > > > > > > On Thu, Jul 24, 2014 at 6:23 AM, Pablo Picko <
> > p...@pitchsider.com>
> > > > > > wrote:
> > > > > > >
> > > > > > > > Hello all
> > > > > > > >
> > > > > > > > Some background.
> > > > > > > >
> > > > > > > > I have a 3 kafka brokers A,B and C, there is a kafka topic
> > called
> > > > > topic
> > > > > > > > with 20 partitions (no replicas).
> > > > > > > >
> > > > > > > > Everything has been working fine for about a week when suddenly
> > > all
> > > > > the
> > > > > > > > data sent to partitions belonging to broker C are not seen by
> > the
> > > > > > > Consumer
> > > > > > > > the consumer is using the high level consumer and does not look
> > > > much
> > > > > > > > different to the sample provided in the documentation.
> > > > > > > >
> > > > > > > > When I inspected the topic i can see that all the partitions
> > are
> > > > > > lagging
> > > > > > > > behind. A restart (og the consumer) seems to sort it out but I
> > am
> > > > > > stumped
> > > > > > > > as to whats doing on any help appreciated.
> > > > > > > >
> > > > > > > >
> > > > > > > > Thanks
> > > > > > > > Pablo
> > > > > > > >
> > > > > > >
> > > > > > >
> > > > > > >
> > > > > > > --
> > > > > > > -- Guozhang
> > > > > > >
> > > > > >
> > > > >
> > > > >
> > > > >
> > > > > --
> > > > > -- Guozhang
> > > > >
> > > >
> > >
> >

Reply via email to