Each executor on the screenshot has 25GB memory remaining . What was the reason to store 170-500 MB to disk if executor has 25GB memory available?
On Thu, May 12, 2016 at 5:12 PM, Takeshi Yamamuro <linguin....@gmail.com> wrote: > Hi, > > Not sure this is a correct answer though, seems `UnifiedMemoryManager` > spills > some blocks of RDDs into disk when execution memory runs short. > > // maropu > > On Fri, May 13, 2016 at 6:16 AM, Alexander Pivovarov <apivova...@gmail.com > > wrote: > >> Hello Everyone >> >> I use Spark 1.6.0 on YARN (EMR-4.3.0) >> >> I use MEMORY_AND_DISK_SER StorageLevel for my RDD. And I use Kryo >> Serializer >> >> I noticed that Spark uses Disk to store some RDD blocks even if Executors >> have lots memory available. See the screenshot >> http://postimg.org/image/gxpsw1fk1/ >> >> Any ideas why it might happen? >> >> Thank you >> Alex >> > > > > -- > --- > Takeshi Yamamuro >