I see a bug in executer memory allocation in the standalone cluster, but I 
can't find which part of the spark code causes this problem. That why's I 
decided to raise this issue here.
Assume you have 3 workers with 10 CPU cores and 10 Gigabyte memories. Assume 
also you have 2 spark jobs that run on this cluster of workers, and these jobs 
configs set as below:
-----------------
job-1:
executer-memory: 5g
executer-CPU: 4
max-cores: 8
------------------
job-2:
executer-memory: 6g
executer-CPU: 4
max-cores: 8
------------------
In this situation, We expect that if we submit both of these jobs, the first 
job that submits get  2 executers which each of them has 4 CPU core and 5g 
memory, and the second job gets only one executer on thirds worker who has 4 
CPU core and 6g memory because worker 1 and worker 2 doesn't have enough memory 
to accept the second job. But surprisingly, we see that one of the first or 
second workers creates an executor for job-2, and the worker's consuming memory 
goes beyond what's allocated to that and gets 11g memory from the operating 
system.
Is this behavior normal? I think this can cause some undefined behavior problem 
in the cluster.
---------------------------------------------------------------------
To unsubscribe e-mail: user-unsubscr...@spark.apache.org

Reply via email to