For the consumer you should see logs like "Connecting to zookeeper instance at " + config.zkConnect "begin registering consumer " + consumerIdString + " in ZK consumerThreadId + " successfully owned partition " + partition + " for topic " + topic "starting auto committer every " + config.autoCommitIntervalMs + " ms"
all coming from kafka.consumer.ZookeeperConsumerConnector /******************************************* Joe Stein Founder, Principal Consultant Big Data Open Source Security LLC http://www.stealth.ly Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop> ********************************************/ On Thu, Jul 24, 2014 at 3:30 PM, Pablo Picko <p...@pitchsider.com> wrote: > Hey guys.. > > I have my my log level set to info, saying that I am not seeing much logs > at all for kafka on startup i see detail about the serializer.class my > producer uses but very little consumer related logs is there anything I > should always see if my log config is correct for the info level > > In relation to my settings for the number of streams here is my code. > > *Map<String, Integer> topicCountMap = new HashMap<>();* > > *//i've 20 threads for the topic for the 20 partitions with no replicas* > > *topicCountMap.put(topicConsumer.getTopic(), > topicConsumer.getNumThreads());* > > > *consumerConnector.createMessageStreams(topicCountMap);* > > Thanks > > Pablo > > > On 24 Jul 2014 18:34, "Joe Stein" <joe.st...@stealth.ly> wrote: > > > What is the value for what you are setting for your number of streams > when > > calling createMessageStreamsByFilter or if using createMessageStreams for > > the TopicCount ( topic -> numberOfStreams )? > > > > How are you threading the iterator on each stream? > > > > /******************************************* > > Joe Stein > > Founder, Principal Consultant > > Big Data Open Source Security LLC > > http://www.stealth.ly > > Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop> > > ********************************************/ > > > > > > On Thu, Jul 24, 2014 at 1:05 PM, Pablo Picko <p...@pitchsider.com> > wrote: > > > > > Guozhang > > > > > > I didn't no. I did spot other people with similar symptoms to my > problem > > > mentioning your suggestion too but I don't see anything in the log to > > > suggest it rebalanced. It could very well be the reason but I can't see > > > anything suggesting it is yet. > > > > > > Thanks > > > Pablo > > > On 24 Jul 2014 17:57, "Guozhang Wang" <wangg...@gmail.com> wrote: > > > > > > > Pablo, > > > > > > > > Do you see any rebalance related logs in consumers? > > > > > > > > Guozhang > > > > > > > > > > > > On Thu, Jul 24, 2014 at 9:02 AM, Pablo Picko <p...@pitchsider.com> > > > wrote: > > > > > > > > > Hey Guozhang > > > > > > > > > > Thanks for the reply, No nothing at all in the logs to suggest > > anything > > > > > went wrong. > > > > > > > > > > Its really puzzling as to what's happened. When I restarted the > > > consumer > > > > > everything worked again. > > > > > > > > > > Prior to the restart I even stopped the producer for a bit. However > > any > > > > > messages that got assigned to the Broker C never got processed. > When > > I > > > > ran > > > > > the console consumer script before I restarted it was able to all > > print > > > > all > > > > > messages, including messages on broker C. It seems to be that for > my > > > > > consumers consumergroup one of the brokers messages just became > > > > > inaccessible. > > > > > > > > > > Thanks > > > > > Pablo > > > > > > > > > > > > > > > On Thu, Jul 24, 2014 at 4:12 PM, Guozhang Wang <wangg...@gmail.com > > > > > > wrote: > > > > > > > > > > > Hi Pablo, > > > > > > > > > > > > During the period did you see any exception/errors on Broker C's > > logs > > > > and > > > > > > the consumer logs also? > > > > > > > > > > > > Guozhang > > > > > > > > > > > > > > > > > > On Thu, Jul 24, 2014 at 6:23 AM, Pablo Picko < > p...@pitchsider.com> > > > > > wrote: > > > > > > > > > > > > > Hello all > > > > > > > > > > > > > > Some background. > > > > > > > > > > > > > > I have a 3 kafka brokers A,B and C, there is a kafka topic > called > > > > topic > > > > > > > with 20 partitions (no replicas). > > > > > > > > > > > > > > Everything has been working fine for about a week when suddenly > > all > > > > the > > > > > > > data sent to partitions belonging to broker C are not seen by > the > > > > > > Consumer > > > > > > > the consumer is using the high level consumer and does not look > > > much > > > > > > > different to the sample provided in the documentation. > > > > > > > > > > > > > > When I inspected the topic i can see that all the partitions > are > > > > > lagging > > > > > > > behind. A restart (og the consumer) seems to sort it out but I > am > > > > > stumped > > > > > > > as to whats doing on any help appreciated. > > > > > > > > > > > > > > > > > > > > > Thanks > > > > > > > Pablo > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > > > -- Guozhang > > > > > > > > > > > > > > > > > > > > > > > > > > > -- > > > > -- Guozhang > > > > > > > > > >