Thanks for the confirmation, I’ve created Jira ticket to track this issue [1]
https://issues.apache.org/jira/browse/FLINK-14952 <https://issues.apache.org/jira/browse/FLINK-14952> Piotrek > On 26 Nov 2019, at 11:10, yingjie <yjclove...@gmail.com> wrote: > > The new BlockingSubpartition implementation in 1.9 uses mmap for data reading > by default which means it steals memory from OS. The mmaped region memory is > managed by JVM, so there should be no OutOfMemory problem reported by JVM > and the OS memory is also not exhausted, so there should be no kernal OOM. > I think Piotr's suspicion is right, yarn tracked the memory used and killed > the TM (the mmap region is also part of the process memory). > > Giving a strict resource restriction to the container (larger than the yarn > limit) which can avoid memory steal or using file instead of mmap as pointed > out by Piotr can solve the problem. > > I think Flink may need to restrict the amount of memory can be stolen. > > > > -- > Sent from: > http://apache-flink-user-mailing-list-archive.2336050.n4.nabble.com/