You would presently do:

mpirun —map-by core:pe=4

to get what you are seeking. If we don’t already set that qualifier when we see 
“cpus_per_task”, then we probably should do so as there isn’t any reason to 
make you set it twice (well, other than trying to track which envar slurm is 
using now).


> On Oct 5, 2015, at 12:38 PM, marcin.krotkiewski 
> <marcin.krotkiew...@gmail.com> wrote:
> 
> Yet another question about cpu binding under SLURM environment..
> 
> Short version: will OpenMPI support SLURM_CPUS_PER_TASK for the purpose of 
> cpu binding?
> 
> 
> Full version: When you allocate a job like, e.g., this
> 
> salloc --ntasks=2 --cpus-per-task=4
> 
> SLURM will allocate 8 cores in total, 4 for each 'assumed' MPI tasks. This is 
> useful for hybrid jobs, where each MPI process spawns some internal worker 
> threads (e.g., OpenMP). The intention is that there are 2 MPI procs started, 
> each of them 'bound' to 4 cores. SLURM will also set an environment variable
> 
> SLURM_CPUS_PER_TASK=4
> 
> which should (probably?) be taken into account by the method that launches 
> the MPI processes to figure out the cpuset. In case of OpenMPI + mpirun I 
> think something should happen in orte/mca/ras/slurm/ras_slurm_module.c, where 
> the variable _is_ actually parsed. Unfortunately, it is never really used...
> 
> As a result, cpuset of all tasks started on a given compute node includes all 
> CPU cores of all MPI tasks on that node, just as provided by SLURM (in the 
> above example - 8). In general, there is no simple way for the user code in 
> the MPI procs to 'split' the cores between themselves. I imagine the original 
> intention to support this in OpenMPI was something like
> 
> mpirun --bind-to subtask_cpuset
> 
> with an artificial bind target that would cause OpenMPI to divide the 
> allocated cores between the mpi tasks. Is this right? If so, it seems that at 
> this point this is not implemented. Is there plans to do this? If no, does 
> anyone know another way to achieve that?
> 
> Thanks a lot!
> 
> Marcin
> 
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2015/10/27803.php

Reply via email to