Do you have the request log turned on? If so, what's total time taken for the corresponding fetch request?
Thanks, Jun On Sat, Jan 11, 2014 at 4:38 AM, Gerrit Jansen van Vuuren < gerrit...@gmail.com> wrote: > I'm also seeing the following. > > I consume the data in the queue. > Then after 10 seconds send another fetch request (with the incremented > offset), and never receives a response from the broker, my code eventually > times out (after 30seconds). > > The broker writes Expiring fetch request Name: FetchRequest; Version: 0; > CorrelationId: 1389443537; ClientId: 1; ReplicaId: -1; MaxWait: 1000 ms; > MinBytes: 1 bytes; RequestInfo: [ping,0] -> > PartitionFetchInfo(187,1048576). > > This corresponds with the timed out fetch request. > > > > > > > On Sat, Jan 11, 2014 at 12:19 PM, Gerrit Jansen van Vuuren < > gerrit...@gmail.com> wrote: > > > Hi, > > > > > > No the offsets are not the same. I've printed out the values to see this, > > and its not the case. > > > > > > > > On Fri, Jan 10, 2014 at 5:02 PM, Jun Rao <jun...@gmail.com> wrote: > > > >> Are the offset used in the 2 fetch requests the same? If so, you will > get > >> the same messages twice. You consumer is responsible for advancing the > >> offsets after consumption. > >> > >> Thanks, > >> > >> Jun > >> > >> > >> On Thu, Jan 9, 2014 at 1:00 PM, Gerrit Jansen van Vuuren < > >> gerrit...@gmail.com> wrote: > >> > >> > Hi, > >> > > >> > I'm writing a custom consumer for kafka 0.8. > >> > Everything works except for the following: > >> > > >> > a. connect, send fetch, read all results > >> > b. send fetch > >> > c. send fetch > >> > d. send fetch > >> > e. via the console publisher, publish 2 messages > >> > f. send fetch :corr-id 1 > >> > g. read 2 messages published :offsets [10 11] :corr-id 1 > >> > h. send fetch :corr-id 2 > >> > i. read 2 messages published :offsets [10 11] :corr-id 2 > >> > j. send fetch ... > >> > > >> > The problem is I get the messages sent twice as a response to two > >> separate > >> > fetch requests. The correlation id is distinct so it cannot be that I > >> read > >> > the response twice. The offsets of the 2 messages are are the same so > >> they > >> > are duplicates, and its not the producer sending the messages twice. > >> > > >> > Note: the same connection is kept open the whole time, and I send > >> > block,receive then send again, after the first 2 messages are read, > the > >> > offsets are incremented and the next fetch will ask kafka to give it > >> > messages from the new offsets. > >> > > >> > any ideas of why kafka would be sending the messages again on the > second > >> > fetch request? > >> > > >> > Regards, > >> > Gerrit > >> > > >> > > > > >