Hi Dominik,
the Web UI shows you the status of a checkpoint [0], so it might be
possible to retrieve the information via REST calls. Usually, you should
perform a savepoint for planned restarts. If a savepoint is successful
you can be sure to restart from it.
Otherwise the platform from data Artisans might be interesting for you
[1], it aims to improve the deployment for streaming application
lifecycles (disclaimer: I work for them).
Regards,
Timo
[0]
https://ci.apache.org/projects/flink/flink-docs-release-1.4/monitoring/checkpoint_monitoring.html
[1] https://data-artisans.com/da-platform-2
Am 11/22/17 um 10:41 AM schrieb domi...@dbruhn.de:
Hey,
we are running Flink 1.3.2 with streaming jobs and we are running into
issues when we are restarting a complete job (which can happen due to
various reasons: upgrading of the job, restarting of the cluster,
failures). The problem is that there is no automated way to find out
from which checkpoint-metadata (so externalized checkpoint) we should
resume. There can always be the situation that we are left with
multiple of those files: Now you want to use the most recent one which
is successfully written.
Is there any tooling available already which picks the latest good
checkpoint? Or at least a tool/commandline which we can use to
validate that a checkpoint is valid so we can pick the latest one?
How are others handling this? Manually?
Would be happy to get some input there,
Dominik