Ah, yes - thanks! It’s been so long since I played with that option I honestly forgot :-)
Hope you are doing well ! Ralph > On Oct 5, 2015, at 4:04 PM, tmish...@jcity.maeda.co.jp wrote: > > Hi Ralph, it's been a long time. > > The option "map-by core" does not work when pe=N > 1 is specified. > So, you should use "map-by slot:pe=N" as far as I remember. > > Regards, > Tetsuya Mishima > > 2015/10/06 5:40:33、"users"さんは「Re: [OMPI users] Hybrid OpenMPI+OpenMP > tasks using SLURM」で書きました >> Hmmm…okay, try -map-by socket:pe=4 >> >> We’ll still hit the asymmetric topology issue, but otherwise this should > work >> >> >>> On Oct 5, 2015, at 1:25 PM, marcin.krotkiewski > <marcin.krotkiew...@gmail.com> wrote: >>> >>> Ralph, >>> >>> Thank you for a fast response! Sounds very good, unfortunately I get an > error: >>> >>> $ mpirun --map-by core:pe=4 ./affinity >>> > -------------------------------------------------------------------------- >>> A request for multiple cpus-per-proc was given, but a directive >>> was also give to map to an object level that cannot support that >>> directive. >>> >>> Please specify a mapping level that has more than one cpu, or >>> else let us define a default mapping that will allow multiple >>> cpus-per-proc. >>> > -------------------------------------------------------------------------- >>> >>> I have allocated my slurm job as >>> >>> salloc --ntasks=2 --cpus-per-task=4 >>> >>> I have checked in 1.10.0 and 1.10.1rc1. >>> >>> >>> >>> >>> On 10/05/2015 09:58 PM, Ralph Castain wrote: >>>> You would presently do: >>>> >>>> mpirun —map-by core:pe=4 >>>> >>>> to get what you are seeking. If we don’t already set that qualifier > when we see “cpus_per_task”, then we probably should do so as there isn’t > any reason to make you set it twice (well, other than >> trying to track which envar slurm is using now). >>>> >>>> >>>>> On Oct 5, 2015, at 12:38 PM, marcin.krotkiewski > <marcin.krotkiew...@gmail.com> wrote: >>>>> >>>>> Yet another question about cpu binding under SLURM environment.. >>>>> >>>>> Short version: will OpenMPI support SLURM_CPUS_PER_TASK for the > purpose of cpu binding? >>>>> >>>>> >>>>> Full version: When you allocate a job like, e.g., this >>>>> >>>>> salloc --ntasks=2 --cpus-per-task=4 >>>>> >>>>> SLURM will allocate 8 cores in total, 4 for each 'assumed' MPI tasks. > This is useful for hybrid jobs, where each MPI process spawns some internal > worker threads (e.g., OpenMP). The intention is >> that there are 2 MPI procs started, each of them 'bound' to 4 cores. > SLURM will also set an environment variable >>>>> >>>>> SLURM_CPUS_PER_TASK=4 >>>>> >>>>> which should (probably?) be taken into account by the method that > launches the MPI processes to figure out the cpuset. In case of OpenMPI + > mpirun I think something should happen in >> orte/mca/ras/slurm/ras_slurm_module.c, where the variable _is_ actually > parsed. Unfortunately, it is never really used... >>>>> >>>>> As a result, cpuset of all tasks started on a given compute node > includes all CPU cores of all MPI tasks on that node, just as provided by > SLURM (in the above example - 8). In general, there is >> no simple way for the user code in the MPI procs to 'split' the cores > between themselves. I imagine the original intention to support this in > OpenMPI was something like >>>>> >>>>> mpirun --bind-to subtask_cpuset >>>>> >>>>> with an artificial bind target that would cause OpenMPI to divide the > allocated cores between the mpi tasks. Is this right? If so, it seems that > at this point this is not implemented. Is there >> plans to do this? If no, does anyone know another way to achieve that? >>>>> >>>>> Thanks a lot! >>>>> >>>>> Marcin >>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> Link to this post: > http://www.open-mpi.org/community/lists/users/2015/10/27803.php >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> Link to this post: > http://www.open-mpi.org/community/lists/users/2015/10/27804.php >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>> Link to this post: > http://www.open-mpi.org/community/lists/users/2015/10/27805.php >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/usersLink to > this post: http://www.open-mpi.org/community/lists/users/2015/10/27806.php > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/10/27809.php