Flume is one of the options to do this, yes.  There are a (growing) number of 
tools you could use to get logs in to Kafka, depending on your environment, 
needs, level of skill, deployment size, applications, log volume:  
(in no particular order)

        Kafkacat : https://github.com/edenhill/kafkacat
        Logstash: https://www.elastic.co/products/logstash
        Klogger (from my team): https://github.com/blackberry/Klogger
        Rsyslog: 
http://www.rsyslog.com/doc/master/configuration/modules/omkafka.html
        Log4j: 
http://stackoverflow.com/questions/22034895/how-to-use-kafka-0-8-log4j-appender

Depending on your app, you may need to read the logs in from the filesystem to 
send to kafka.  Better though, would be not bothering with the filesystem 
read/write, and using something that will listen to syslog/use log4j, and write 
the log data directly to Kafka.  If you have control over the application code, 
you could also look at tacking in a producer library (kafkas provided ones, or 
others (https://github.com/blackberry/Krackle))   and writing directly to kafka 
from the app - no need for log4j, syslog, filesystem logs.

If you're starting off, testing is nice and simple using logstash or klogger - 
you can take existing logs and just grab them from where they are, and put them 
in to kafka without changing any of your existing workflow/code/configuration.

There are lots of solutions, you need to experiment to find the one that fits 
your environment best.

Cheers,

T.      

-----Original Message-----
From: Snehalata Nagaje [mailto:snehalata.nag...@harbingergroup.com] 
Sent: Friday, July 31, 2015 3:32
To: users@kafka.apache.org
Subject: How to use kafka for log aggregation



Hi All, 

I want to use kafka for log aggregation, how can we use it? 

Do we need to manually read log files and post this logs into kafka using 
producer API? 

Can Apache Flume help in it? 

Thanks, 
Snehalata 

Reply via email to