I'm doing quite well, thank you. I'm involved in a big project and so very busy now.
But I still try to keep watching these mailing lists. Regards, Tetsuya Mishima 2015/10/06 8:17:33、"users"さんは「Re: [OMPI users] Hybrid OpenMPI+OpenMP tasks using SLURM」で書きました > Ah, yes - thanks! It’s been so long since I played with that option I honestly forgot :-) > > Hope you are doing well ! > Ralph > > > On Oct 5, 2015, at 4:04 PM, tmish...@jcity.maeda.co.jp wrote: > > > > Hi Ralph, it's been a long time. > > > > The option "map-by core" does not work when pe=N > 1 is specified. > > So, you should use "map-by slot:pe=N" as far as I remember. > > > > Regards, > > Tetsuya Mishima > > > > 2015/10/06 5:40:33、"users"さんは「Re: [OMPI users] Hybrid OpenMPI +OpenMP > > tasks using SLURM」で書きました > >> Hmmm…okay, try -map-by socket:pe=4 > >> > >> We’ll still hit the asymmetric topology issue, but otherwise this should > > work > >> > >> > >>> On Oct 5, 2015, at 1:25 PM, marcin.krotkiewski > > <marcin.krotkiew...@gmail.com> wrote: > >>> > >>> Ralph, > >>> > >>> Thank you for a fast response! Sounds very good, unfortunately I get an > > error: > >>> > >>> $ mpirun --map-by core:pe=4 ./affinity > >>> > > -------------------------------------------------------------------------- > >>> A request for multiple cpus-per-proc was given, but a directive > >>> was also give to map to an object level that cannot support that > >>> directive. > >>> > >>> Please specify a mapping level that has more than one cpu, or > >>> else let us define a default mapping that will allow multiple > >>> cpus-per-proc. > >>> > > -------------------------------------------------------------------------- > >>> > >>> I have allocated my slurm job as > >>> > >>> salloc --ntasks=2 --cpus-per-task=4 > >>> > >>> I have checked in 1.10.0 and 1.10.1rc1. > >>> > >>> > >>> > >>> > >>> On 10/05/2015 09:58 PM, Ralph Castain wrote: > >>>> You would presently do: > >>>> > >>>> mpirun —map-by core:pe=4 > >>>> > >>>> to get what you are seeking. If we don’t already set that qualifier > > when we see “cpus_per_task”, then we probably should do so as there isn’t > > any reason to make you set it twice (well, other than > >> trying to track which envar slurm is using now). > >>>> > >>>> > >>>>> On Oct 5, 2015, at 12:38 PM, marcin.krotkiewski > > <marcin.krotkiew...@gmail.com> wrote: > >>>>> > >>>>> Yet another question about cpu binding under SLURM environment.. > >>>>> > >>>>> Short version: will OpenMPI support SLURM_CPUS_PER_TASK for the > > purpose of cpu binding? > >>>>> > >>>>> > >>>>> Full version: When you allocate a job like, e.g., this > >>>>> > >>>>> salloc --ntasks=2 --cpus-per-task=4 > >>>>> > >>>>> SLURM will allocate 8 cores in total, 4 for each 'assumed' MPI tasks. > > This is useful for hybrid jobs, where each MPI process spawns some internal > > worker threads (e.g., OpenMP). The intention is > >> that there are 2 MPI procs started, each of them 'bound' to 4 cores. > > SLURM will also set an environment variable > >>>>> > >>>>> SLURM_CPUS_PER_TASK=4 > >>>>> > >>>>> which should (probably?) be taken into account by the method that > > launches the MPI processes to figure out the cpuset. In case of OpenMPI + > > mpirun I think something should happen in > >> orte/mca/ras/slurm/ras_slurm_module.c, where the variable _is_ actually > > parsed. Unfortunately, it is never really used... > >>>>> > >>>>> As a result, cpuset of all tasks started on a given compute node > > includes all CPU cores of all MPI tasks on that node, just as provided by > > SLURM (in the above example - 8). In general, there is > >> no simple way for the user code in the MPI procs to 'split' the cores > > between themselves. I imagine the original intention to support this in > > OpenMPI was something like > >>>>> > >>>>> mpirun --bind-to subtask_cpuset > >>>>> > >>>>> with an artificial bind target that would cause OpenMPI to divide the > > allocated cores between the mpi tasks. Is this right? If so, it seems that > > at this point this is not implemented. Is there > >> plans to do this? If no, does anyone know another way to achieve that? > >>>>> > >>>>> Thanks a lot! > >>>>> > >>>>> Marcin > >>>>> > >>>>> > >>>>> > >>>>> _______________________________________________ > >>>>> users mailing list > >>>>> us...@open-mpi.org > >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > >>>>> Link to this post: > > http://www.open-mpi.org/community/lists/users/2015/10/27803.php > >>>> _______________________________________________ > >>>> users mailing list > >>>> us...@open-mpi.org > >>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > >>>> Link to this post: > > http://www.open-mpi.org/community/lists/users/2015/10/27804.php > >>> > >>> _______________________________________________ > >>> users mailing list > >>> us...@open-mpi.org > >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > >>> Link to this post: > > http://www.open-mpi.org/community/lists/users/2015/10/27805.php > >> > >> _______________________________________________ > >> users mailing list > >> us...@open-mpi.org > >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/usersLink to > > this post: http://www.open-mpi.org/community/lists/users/2015/10/27806.php > > _______________________________________________ > > users mailing list > > us...@open-mpi.org > > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > > Link to this post: http://www.open-mpi.org/community/lists/users/2015/10/27809.php > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/usersLink to this post: http://www.open-mpi.org/community/lists/users/2015/10/27810.php