> On Oct 20, 2015, at 5:47 AM, Daniel Letai <d...@letai.org.il> wrote: > > Thanks for the reply, > > On 10/13/2015 04:04 PM, Ralph Castain wrote: >>> On Oct 12, 2015, at 6:10 AM, Daniel Letai <d...@letai.org.il> wrote: >>> >>> Hi, >>> After upgrading to 1.8.8 I can no longer see the map. When looking at the >>> man page for mpirun, display-map no longer exists. Is there a way to show >>> the map in 1.8.8 ? >> I don’t know why/how it got dropped from the man page, but the display-map >> option certainly still exists - do “mpirun -h” to see the full list of >> options, and you’ll see it is there. I’ll ensure it gets restored to the man >> page in the 1.10 series as the 1.8 series is complete. > Thanks for clarifying, >> >>> Another issue - I'd like to map 2 process per node - 1 to each socket. >>> What is the current "correct" syntax? --map-by ppr:2:node doesn't guarantee >>> 1 per Socket. --map-by ppr:1:socket doesn't guarantee 2 per node. I assume >>> it's something obvious, but the documentation is somewhat lacking. >>> I'd like to know the general syntax - even if I have 4 socket nodes I'd >>> still like to map only 2 procs per node. >> That’s a tough one. I’m not sure there is a way to do that right now. >> Probably something we’d have to add. Out of curiosity, if you have 4 sockets >> and only 2 procs, would you want each proc bound to 2 of the 4 sockets? Or >> are you expecting them to be bound to only 1 socket (thus leaving 2 sockets >> idle), or simply leave them unbound? > I have 2 pci devices (gpu) per node. I need 1 proc per socket to be bound to > that socket and "talk" to it's respective gpu, so no matter how many sockets > I have - I must distribute the procs 2 per node, each in it's own socket > (actually, each in it's own numa) and be bound. > > So I expect them to be "bound to only 1 socket (thus leaving 2 sockets idle)”.
Are the gpu’s always near the same sockets for every node? If so, you might be able to use the cpu-set option to restrict us to those sockets, and then just "—map-by ppr:2:node —bind-to socket" -cpu-set|--cpu-set <arg0> Comma-separated list of ranges specifying logical cpus allocated to this job [default: none] > > I might run other jobs on the idle sockets (depending on mem utilization) but > that's not an immediate concern at this time. >> >>> Combining with numa/dist to hca/dist to gpu will be very helpful too. >> Definitely no way to do this one today. >> >>> Thanks, >>> >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>> Link to this post: >>> http://www.open-mpi.org/community/lists/users/2015/10/27860.php >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2015/10/27861.php > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/10/27898.php