Hi Yufei,

I cannot remember exactly the changes in this area between Flink 1.10.0 and
Flink 1.12.0. It sounds a bit as if we were not releasing memory segments
fast enough or had a memory leak. One thing to try out is to increase the
restart delay to see whether it is the first problem. Alternatively, you
can also try to bisect the commits in between these versions. If you have a
test failing reliably, this shouldn't take too long. Maybe Piotr knows
about a fix which could have solved this problem.

Cheers,
Till

On Fri, Dec 25, 2020 at 3:05 AM Yangze Guo <karma...@gmail.com> wrote:

> Hi, Yufei.
>
> Can you reproduce this issue in 1.10.0? The deterministic slot sharing
> introduced in 1.12.0 is one possible reason. Before 1.12.0, the
> distribution of tasks in slots is not determined. Even if the network
> buffers are enough from the perspective of the cluster. Bad
> distribution of tasks can lead to the "insufficient network buffer" as
> well.
>
> Best,
> Yangze Guo
>
> On Fri, Dec 25, 2020 at 12:54 AM Yufei Liu <liuyufei9...@gmail.com> wrote:
> >
> > Hey,
> > I’ve found that job will throw “java.io.IOException: Insufficient number
> of network buffers: required 51, but only 1 available” after job retstart,
> and I’ve observed TM use much more network buffers than before.
> > My internal branch is under 1.10.0 can easily  reproduce, but I use
> 1.12.0 doesn’t have this issue. I Think maybe was already fixed after some
> PR, I'm curious about what can lead to this problem?
> >
> > Best.
> > YuFei.
>

Reply via email to