[ https://issues.apache.org/jira/browse/KAFKA-2500?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14735602#comment-14735602 ]
Jason Gustafson commented on KAFKA-2500: ---------------------------------------- [~willf] Thanks for the explanation. I'm a little unclear on your requirements, though. It seems like there could also be new records pushed in the time that it takes for the fetch response to be returned, right? It only reduces the window. Or am I missing something? Anyway, my inclination is to reconsider this after the initial release since it doesn't seem essential for most use cases. I'd be interested whether [~guozhang] has any thoughts. > Make logEndOffset available in the 0.8.3 Consumer > ------------------------------------------------- > > Key: KAFKA-2500 > URL: https://issues.apache.org/jira/browse/KAFKA-2500 > Project: Kafka > Issue Type: Sub-task > Components: consumer > Affects Versions: 0.8.3 > Reporter: Will Funnell > Assignee: Jason Gustafson > Priority: Critical > Fix For: 0.8.3 > > > Originally created in the old consumer here: > https://issues.apache.org/jira/browse/KAFKA-1977 > The requirement is to create a snapshot from the Kafka topic but NOT do > continual reads after that point. For example you might be creating a backup > of the data to a file. > This ticket covers the addition of the functionality to the new consumer. > In order to achieve that, a recommended solution by Joel Koshy and Jay Kreps > was to expose the high watermark, as maxEndOffset, from the FetchResponse > object through to each MessageAndMetadata object in order to be aware when > the consumer has reached the end of each partition. > The submitted patch achieves this by adding the maxEndOffset to the > PartitionTopicInfo, which is updated when a new message arrives in the > ConsumerFetcherThread and then exposed in MessageAndMetadata. > See here for discussion: > http://search-hadoop.com/m/4TaT4TpJy71 -- This message was sent by Atlassian JIRA (v6.3.4#6332)