I don't think anything exists like this in Kafka (or contrib), but it would be a useful addition! Personally, I have written this exact thing at previous jobs.

As for the Hadoop consumer, since there is a FileSystem implementation for S3 in Hadoop, it should be possible. The Hadoop consumer works by writing out data files containing the Kafka messages along side offset files which contain the last offset read for each partition. If it is re-consuming from zero each time you run it, it means it's not finding the offset files from the previous run.

Having used it a bit, the Hadoop consumer is certainly an area that could use improvement.

HTH,
David

On 12/27/12 4:41 AM, Pratyush Chandra wrote:
Hi,

I am looking for a S3 based consumer, which can write all the received
events to S3 bucket (say every minute). Something similar to Flume HDFSSink
http://flume.apache.org/FlumeUserGuide.html#hdfs-sink
I have tried evaluating hadoop-consumer in contrib folder. But it seems to
be more for offline processing, which will fetch everything from offset 0
at once and replace it in S3 bucket.
Any help would be appreciated ?


Reply via email to