Srikanth, KafkaStreams uses the new consumers, thus you need to use
> bin/kafka-consumer-groups.sh --new-consumer --bootstrap-server localhost:9092 > --list instead of "--zookeeper XXX" AbstractTask.commit() is called each "commit.interval.ms". A task is requested to persist all buffered data on this call; ie, flush state data to persistent storage and flush pending writes to Kafka. After all tasks executed .commit(), KafkaStreams app will commit the current source topic offsets to Kafka. This is a regular KafkaConsumer commit. If you use the latest trunc version you can get consumer/producer metrics via KafkaStreams.metrics(). For getting the Context object, you need to write you own custom processor and add it via stream.process(...); -Matthias On 06/02/2016 06:54 PM, Srikanth wrote: > Matthias, > > """bin/kafka-consumer-groups.sh --zookeeper localhost:2181/kafka10 > --list""" output didn't show the group I used in streams app. > Also, AbstractTask.java had a commit() API. That made me wonder if offset > management was overridden too. > > I'm trying out KafkaStreams for one new streaming app we are working on. > We'll most likely stick to DSL for that. > Does the DSL expose any stat or debug info? Or any way to access the > underlying Context? > > Srikanth > > On Thu, Jun 2, 2016 at 9:30 AM, Matthias J. Sax <matth...@confluent.io> > wrote: > >> Hi Srikanth, >> >> I am not exactly sure if I understand your question correctly. >> >> One way to track the progress is to get the current record offset (you >> can obtain it in the low lever Processor API via the provided Context >> object). >> >> Otherwise, on commit, all writes to intermediate topics are flushed to >> Kafka and the source offsets get committed to Kafka, too. >> >> A KafkaStream application internally uses the standard high level Java >> KafkaConsumer (all instances of a single application belong to the same >> consumer group) and standard Java KafkaProducer. >> >> So you can use standard Kafka tools to access this information. >> >> Does this answer your question? >> >> -Matthias >> >> On 05/31/2016 09:10 PM, Srikanth wrote: >>> Hi, >>> >>> How can I track the progress of a kafka streaming job? >>> The only reference I see is "commit.interval.ms" which controls how >> often >>> offset is committed. >>> By default where is it committed and is there a tool to read it back? May >>> be something similar to bin/kafka-consumer-groups.sh. >>> >>> I'd like to look at details for source & intermediate topics too. >>> >>> Srikanth >>> >> >> >
signature.asc
Description: OpenPGP digital signature