Hey!

Nice to hear that it works.

A bit of info is now visible in the web dashboard now, as of that PR:
https://github.com/apache/flink/pull/1453

Is that what you had in mind?

Greetings,
Stephan


On Sat, Jan 2, 2016 at 4:53 PM, Gyula Fóra <gyf...@apache.org> wrote:

> Ok, I could figure out the problem, it was my fault :). The issue was that
> I was running a short testing job and the sources finished before
> triggering the checkpoint. So the folder was created for the job in S3 but
> since we didn't write anything to it is shown as a file in S3.
>
> Maybe it would be good to give some info to the user in case the source is
> finished when the checkpoint is triggered.
>
> On the bright side, it seems to work well, also with the savepoints :)
>
> Cheers
> Gyula
>
> Gyula Fóra <gyf...@apache.org> ezt írta (időpont: 2016. jan. 2., Szo,
> 11:57):
>
> > Hey,
> >
> > I am trying to checkpoint my streaming job to S3 but it seems that the
> > checkpoints never complete but also I don't get any error in the logs.
> >
> > The state backend connects properly to S3 apparently as it creates the
> > following file in the given S3 directory :
> >
> > 95560b1acf5307bc3096020071c83230_$folder$    (this is a file, not a
> folder)
> >
> > The job id is 95560b1acf5307bc3096020071c83230, but that filename is odd
> > and might cause the problem.
> > It seems that the backend doesnt properly create a folder for the job
> > checkpoints for the job id.
> >
> > Does anyone have any idea what might cause this problem?
> >
> > Thanks,
> > Gyula
> >
> >
> >
> >
> >
>

Reply via email to