Because we need to do exploratory data analysis and machine learning. We need to backup the messages somewhere so that the data scientists can query/load them.
So we need something like a router that just opens up a new consumer group which just keeps on storing them to S3. On Tue, Dec 6, 2016 at 5:05 PM, Sharninder Khera <sharnin...@gmail.com> wrote: > Why not just have a parallel consumer read all messages from whichever > topics you're interested in and store them wherever you want to? You don't > need to "backup" Kafka messages. > > _____________________________ > From: Aseem Bansal <asmbans...@gmail.com> > Sent: Tuesday, December 6, 2016 4:55 PM > Subject: Storing Kafka Message JSON to deep storage like S3 > To: <users@kafka.apache.org> > > > Hi > > Has anyone done a storage of Kafka JSON messages to deep storage like S3. > We are looking to back up all of our raw Kafka JSON messages for > Exploration. S3, HDFS, MongoDB come to mind initially. > > I know that it can be stored in kafka itself but storing them in Kafka > itself does not seem like a good option as we won't be able to query it and > the configurations of machines containing kafka will have to be increased > as we go. Something like S3 we won't have to manage. > > > > >