Hi, Elias,
Thanks a lot to put up the patch for the simple job running in Kubernetes!
As Kartik mentioned, that is well aligned w/ our goal to make Samza job
launching easier. I am glad that we actually share a lot of common ideas
from independent minds. Let me try to give my opinions on this:
1.
How about writing out a lineage record to a different topic from every
samza job? Throw a little meta data along with the data such that every
job that touches a piece of data also writes out read/wrote records to a
separate tracking topic.
A read/wrote record would look something like:
{orig_id,
---
This is an automatically generated e-mail. To reply, visit:
https://reviews.apache.org/r/40857/#review108656
---
Ship it!
- Jake Maes
On Dec. 2, 2015, 7:53 a.m., Aleksandar Birc
Hi Anton,
Samza doesn’t have the same concept of an ack as Storm does built in. This
could be seen as a good or bad thing.
On one hand, the ack is very expensive in storm, on the other hand you can very
easily do what you are describing.
Samza topologies aren't DAGs, you can have jobs that feed