Hi team,

Does anyone have any ideas?

On Thu, May 12, 2022 at 9:20 PM tao xiao <xiaotao...@gmail.com> wrote:

> Forgot to mention the Flink version is 1.13.2 and we use kubernetes native
> mode
>
> On Thu, May 12, 2022 at 9:18 PM tao xiao <xiaotao...@gmail.com> wrote:
>
>> Hi team,
>>
>> I met a weird issue when a job tries to recover from JM failure.  The
>> success checkpoint before JM crashed is 41205
>>
>> ```
>>
>> {"log":"2022-05-10 14:55:40,663 INFO  
>> org.apache.flink.runtime.checkpoint.CheckpointCoordinator    [] - Completed 
>> checkpoint 41205 for job 00000000000000000000000000000000 (9453840 bytes in 
>> 1922 ms).\n","stream":"stdout","time":"2022-05-10T14:55:40.663286893Z"}
>>
>> ```
>>
>> However JM tries to recover the job with an old checkpoint 41051 which
>> doesn't exist that leads to unrecoverable state
>>
>> ```
>>
>> "2022-05-10 14:59:38,949 INFO  
>> org.apache.flink.runtime.checkpoint.DefaultCompletedCheckpointStore [] - 
>> Trying to retrieve checkpoint 41051.\n"
>>
>> ```
>>
>> Full log attached
>>
>> --
>> Regards,
>> Tao
>>
>
>
> --
> Regards,
> Tao
>


-- 
Regards,
Tao

Reply via email to