Hello,

I am trying to use Flink HA mode inside kubernetes
<https://ci.apache.org/projects/flink/flink-docs-master/docs/deployment/ha/kubernetes_ha/>
 in standalone
<https://ci.apache.org/projects/flink/flink-docs-master/docs/deployment/resource-providers/standalone/overview/#application-mode>
mode.
The Job ID is always constant, "00000000000000000000000000000000". In
situation where we restart the job (Not from a check-point or savepoint),
we see errors like
"""

Caused by: org.apache.hadoop.fs.FileAlreadyExistsException:
'<PATH>/flink-checkpoints/00000000000000000000000000000000/chk-1/_metadata'
already exists

"""
where checkpoints have not been created since the restart of Job .

My question:
* Is the recommended way to set a new unique "checkpoint path" every time
we update Job and restart necessary k8 resources (say not restarted from
checkpoint-savepoint)? Or GC checkpoints during deletion and reload from
savepoint if required? Looking for a standard recommendation.
* Is there a way I can override the JobID to be unique and indicate it is a
complete restart in HA mode?


Thanks,
Harsh

Reply via email to