Re: Where to put live model and business logic in Hadoop/Flink BigData system

2016-05-06 Thread Fabian Hueske
Hi Palle, this sounds indeed like a good use case for Flink. Depending on the complexity of the aggregated historical views, you can implement a Flink DataStream program which builds the views on the fly, i.e., you do not need to periodically trigger MR/Flink/Spark batch jobs to compute the views

Re: Where to put live model and business logic in Hadoop/Flink BigData system

2016-05-06 Thread Deepak Sharma
I see the flow to be as below: LogStash->Log Stream->Flink ->Kafka->Live Model | Mongo/HBASE The Live Model will again be Flink streaming data sets from Kakfa. There you analyze the incoming stream for the certain value and

RE: Where to put live model and business logic in Hadoop/Flink BigData system

2016-05-06 Thread Lohith Samaga M
Grüßen / Sincères salutations M. Lohith Samaga -Original Message- From: pa...@sport.dk [mailto:pa...@sport.dk Sent: Friday, May 06, 2016 16.23 To: user@flink.apache.org Subject: Where to put live model and business logic in Hadoop/Flink BigData system Hi there. We are putting together

Where to put live model and business logic in Hadoop/Flink BigData system

2016-05-06 Thread palle
Hi there. We are putting together some BigData components for handling a large amount of incoming data from different log files and perform some analysis on the data. All data being fed into the system will go into HDFS. We plan on using Logstash, Kafka and Flink for bringing data from the log