Thanks! Ismael
On Mon, Jan 25, 2016 at 4:03 PM, Han JU <ju.han.fe...@gmail.com> wrote: > Issue created: https://issues.apache.org/jira/browse/KAFKA-3146 > > 2016-01-25 16:07 GMT+01:00 Han JU <ju.han.fe...@gmail.com>: > > > Hi Bruno, > > > > Can you tell me a little bit more about that? A seek() in the > > `onPartitionAssigned`? > > > > Thanks. > > > > 2016-01-25 10:51 GMT+01:00 Han JU <ju.han.fe...@gmail.com>: > > > >> Ok I'll create a JIRA issue on this. > >> > >> Thanks! > >> > >> 2016-01-23 21:47 GMT+01:00 Bruno Rassaerts <bruno.rassae...@novazone.be > >: > >> > >>> +1 here > >>> > >>> As a workaround we seek to the current offset which resets the current > >>> clients internal states and everything continues. > >>> > >>> Regards, > >>> Bruno Rassaerts | Freelance Java Developer > >>> > >>> Novazone, Edingsesteenweg 302, B-1755 Gooik, Belgium > >>> T: +32(0)54/26.02.03 - M:+32(0)477/39.01.15 > >>> bruno.rassae...@novazone.be -www.novazone.be > >>> > >>> > On 23 Jan 2016, at 17:52, Ismael Juma <ism...@juma.me.uk> wrote: > >>> > > >>> > Hi, > >>> > > >>> > Can you please file an issue in JIRA so that we make sure this is > >>> > investigated? > >>> > > >>> > Ismael > >>> > > >>> >> On Fri, Jan 22, 2016 at 3:13 PM, Han JU <ju.han.fe...@gmail.com> > >>> wrote: > >>> >> > >>> >> Hi, > >>> >> > >>> >> I'm prototyping with the new consumer API of kafka 0.9 and I'm > >>> particularly > >>> >> interested in the `ConsumerRebalanceListener`. > >>> >> > >>> >> My test setup is like the following: > >>> >> - 5M messages pre-loaded in one node kafka 0.9 > >>> >> - 12 partitions, auto offset commit set to false > >>> >> - in `onPartitionsRevoked`, commit offset and flush the local state > >>> >> > >>> >> The test run is like the following: > >>> >> - launch one process with 2 consumers and let it consume for a > while > >>> >> - launch another process with 2 consumers, this triggers a > >>> rebalancing, > >>> >> and let these 2 processes run until messages are all consumed > >>> >> > >>> >> The code is here: > https://gist.github.com/darkjh/fe1e5a5387bf13b4d4dd > >>> >> > >>> >> So at first, the 2 consumers of the first process each got 6 > >>> partitions. > >>> >> And after the rebalancing, each consumer got 3 partitions. It's > >>> confirmed > >>> >> by logging inside the `onPartitionAssigned` callback. > >>> >> > >>> >> But after the rebalancing, one of the 2 consumers of the first > >>> process stop > >>> >> receiving messages, even if it has partitions assigned to: > >>> >> > >>> >> balance-1 pulled 7237 msgs ... > >>> >> balance-0 pulled 7263 msgs ... > >>> >> 2016-01-22 15:50:37,533 [INFO] [pool-1-thread-2] > >>> >> o.a.k.c.c.i.AbstractCoordinator - Attempt to heart beat failed since > >>> the > >>> >> group is rebalancing, try to re-join group. > >>> >> balance-1 flush @ 536637 > >>> >> balance-1 committed offset for List(balance-11, balance-10, > balance-9, > >>> >> balance-8, balance-7, balance-6) > >>> >> 2016-01-22 15:50:37,575 [INFO] [pool-1-thread-1] > >>> >> o.a.k.c.c.i.AbstractCoordinator - Attempt to heart beat failed since > >>> the > >>> >> group is rebalancing, try to re-join group. > >>> >> balance-0 flush @ 543845 > >>> >> balance-0 committed offset for List(balance-5, balance-4, balance-3, > >>> >> balance-2, balance-1, balance-0) > >>> >> balance-0 got assigned List(balance-5, balance-4, balance-3) > >>> >> balance-1 got assigned List(balance-11, balance-10, balance-9) > >>> >> balance-1 pulled 3625 msgs ... > >>> >> balance-0 pulled 3621 msgs ... > >>> >> balance-0 pulled 3631 msgs ... > >>> >> balance-0 pulled 3631 msgs ... > >>> >> balance-1 pulled 0 msgs ... > >>> >> balance-0 pulled 3643 msgs ... > >>> >> balance-0 pulled 3643 msgs ... > >>> >> balance-1 pulled 0 msgs ... > >>> >> balance-0 pulled 3622 msgs ... > >>> >> balance-0 pulled 3632 msgs ... > >>> >> balance-1 pulled 0 msgs ... > >>> >> balance-0 pulled 3637 msgs ... > >>> >> balance-0 pulled 3641 msgs ... > >>> >> balance-0 pulled 3640 msgs ... > >>> >> balance-1 pulled 0 msgs ... > >>> >> balance-0 pulled 3632 msgs ... > >>> >> balance-0 pulled 3630 msgs ... > >>> >> balance-1 pulled 0 msgs ... > >>> >> ...... > >>> >> > >>> >> `balance-0` and `balance-1` are the names of the consumer thread. So > >>> after > >>> >> the rebalancing, thread `balance-1` continues to poll but no message > >>> >> arrive, given that it has got 3 partitions assigned to after the > >>> >> rebalancing. > >>> >> > >>> >> Finally other 3 consumers pulls all their partitions' message, the > >>> >> situation is like > >>> >> > >>> >> GROUP, TOPIC, PARTITION, CURRENT OFFSET, LOG END OFFSET, LAG, OWNER > >>> >> balance-test, balance, 9, 417467, 417467, 0, consumer-2_/127.0.0.1 > >>> >> balance-test, balance, 10, 417467, 417467, 0, consumer-2_/127.0.0.1 > >>> >> balance-test, balance, 11, 417467, 417467, 0, consumer-2_/127.0.0.1 > >>> >> balance-test, balance, 6, 180269, 417467, 237198, consumer-2_/ > >>> 127.0.0.1 > >>> >> balance-test, balance, 7, 180036, 417468, 237432, consumer-2_/ > >>> 127.0.0.1 > >>> >> balance-test, balance, 8, 180197, 417467, 237270, consumer-2_/ > >>> 127.0.0.1 > >>> >> balance-test, balance, 3, 417467, 417467, 0, consumer-1_/127.0.0.1 > >>> >> balance-test, balance, 4, 417468, 417468, 0, consumer-1_/127.0.0.1 > >>> >> balance-test, balance, 5, 417468, 417468, 0, consumer-1_/127.0.0.1 > >>> >> balance-test, balance, 0, 417467, 417467, 0, consumer-1_/127.0.0.1 > >>> >> balance-test, balance, 1, 417467, 417467, 0, consumer-1_/127.0.0.1 > >>> >> balance-test, balance, 2, 417467, 417467, 0, consumer-1_/127.0.0.1 > >>> >> > >>> >> So you can see, partition [6, 7, 8] still has messages, but the > >>> consumer > >>> >> can't pull them after the rebalancing. > >>> >> > >>> >> I've tried 0.9.0.0 release, trunk and 0.9.0 branch, for both > >>> server/broker > >>> >> and client. > >>> >> > >>> >> I hope the code is clear enough to illustrate/reproduce the problem. > >>> It's > >>> >> quite a surprise for me because this is the main feature of the new > >>> >> consumer API, but it does not seem to work properly. > >>> >> Feel free to talk to me for any details. > >>> >> -- > >>> >> *JU Han* > >>> >> > >>> >> Software Engineer @ Teads.tv > >>> >> > >>> >> +33 0619608888 > >>> >> > >>> > >> > >> > >> > >> -- > >> *JU Han* > >> > >> Software Engineer @ Teads.tv > >> > >> +33 0619608888 > >> > > > > > > > > -- > > *JU Han* > > > > Software Engineer @ Teads.tv > > > > +33 0619608888 > > > > > > -- > *JU Han* > > Software Engineer @ Teads.tv > > +33 0619608888 >