Hi list,

I'm new to Flink, and I find this project very interesting. I have
experience with Apache Spark, and for I've seen so far I find that Flink
provides an API at a similar abstraction level but based on single record
processing instead of batch processing. I've read in Quora that Flink
extends stream processing to batch processing, while Spark extends batch
processing to streaming. Therefore I find Flink specially attractive for
low latency stream processing. Anyway, I would appreciate if someone could
give some indication about where I could find a list of hardware
requirements for the slave nodes in a Flink cluster. Something along the
lines of https://spark.apache.org/docs/latest/hardware-provisioning.html.
Spark is known for having quite high minimal memory requirements (8GB RAM
and 8 cores minimum), and I was wondering if it is also the case for Flink.
Lower memory requirements would be very interesting for building small
Flink clusters for educational purposes, or for small projects.

Apart from that, I wonder if there is some blog post by the comunity about
transitioning from Spark to Flink. I think it could be interesting, as
there are some similarities in the APIs, but also deep differences in the
underlying approaches. I was thinking in something like Breeze's cheatsheet
comparing its matrix operatations with those available in Matlab and Numpy
https://github.com/scalanlp/breeze/wiki/Linear-Algebra-Cheat-Sheet, or like
http://rosettacode.org/wiki/Factorial. Just an idea anyway. Also, any
pointer to some online course, book or training for Flink besides the
official programming guides would be much appreciated

Thanks in advance for help

Greetings,

Juan

Reply via email to