INK-11843
>
> On Tue, Mar 31, 2020 at 6:49 AM Dinesh J wrote:
>
>> Hi Yang,
>> I am attaching one full jobmanager log for a job which I reran today.
>> This a job that tries to read from savepoint.
>> Same error message "leader election onging" is displayed.
rship, it may because of some problem of
> zookeeper.
>
> Maybe you need to share the complete jobmanager logs so that we could know
> what
> is happening in the jobmanager.
>
>
> Best,
> Yang
>
>
> Dinesh J 于2020年3月31日周二 上午3:46写道:
>
>> HI Yang,
>> Tha
ility.zookeeper.client.connection-timeout"
> and "high-availability.zookeeper.client.session-timeout". Please keep in
> mind that too small value
> will also cause unexpected failover because of network problem.
>
>
> Best,
> Yang
>
> Dinesh J 于2020年3月25日周三 下午4:
he failover not
> work, e.g. no leader is elected or a job is not restarted after the current
> leader failure?
>
> Best,
> Andrey
>
> On Sun, Mar 22, 2020 at 11:14 AM Dinesh J wrote:
>
>> Attaching the job manager log for reference.
>>
>> 2020-03-22 11:39:02,
, Mar 22, 2020 at 1:25 PM Dinesh J wrote:
> Hi all,
> We have single job yarn flink cluster setup with High Availability.
> Sometimes job manager failure successfully restarts next attempt from
> current checkpoint.
> But occasionally we are getting below error.
>
> {"err
Hi all,
We have single job yarn flink cluster setup with High Availability.
Sometimes job manager failure successfully restarts next attempt from
current checkpoint.
But occasionally we are getting below error.
{"errors":["Service temporarily unavailable due to an ongoing leader
election. Please r