<\begin hijacking this thread> I second Saliya's thanks to Tetsuya. I've been following this thread, to learn a bit more about how to use hardware locality with OpenMPI effectively. [I am still using "--bycore"+"--bind-to-core" in most cases, and "--cpus-per-proc" occasionally when in hybrid MPI+OpenMP mode.]
When it comes to hardware locality, the syntax and the functionality has changed fast and significantly in the recent past. Hence, it would be great if the OpenMPI web page could provide pointers for the type of external documentation that Tetsuya just sent. Perhaps also some additional guidelines and comments on what is available on each release/series of OpenMPI, and how to use these options. There is some material about hwloc, but I can't see much about lama ( which means "mud" in my first language :) ). We can hardly learn things like that from the mpiexec man page alone, although it has very good examples. Thank you, Gus Correa <\end hijacking of this thread> On 03/27/2014 11:38 AM, Saliya Ekanayake wrote:
Thank you, this is really helpful. Saliya On Thu, Mar 27, 2014 at 5:11 AM, <tmish...@jcity.maeda.co.jp <mailto:tmish...@jcity.maeda.co.jp>> wrote: Mapping and binding is related to so called process affinity. It's a bit difficult for me to explain ... So please see this URL below(especially the first half part of it - from 1 to 20 pages): http://www.slideshare.net/jsquyres/open-mpi-explorations-in-process-affinity-eurompi13-presentation Although these slides by Jeff are the explanation for LAMA, which is another mapping system installed in the openmpi-1.7 series, I guess you can easily understand what is mapping and binding in general terms. Tetsuya > Thank you Tetsuya - it worked. > > Btw. what's the difference between mapping and binding? I think I am bit confused here. > > Thank you, > Saliya > > > On Thu, Mar 27, 2014 at 4:19 AM, <tmish...@jcity.maeda.co.jp <mailto:tmish...@jcity.maeda.co.jp>>wrote: > > > Hi Saliya, > > What you want to do is map-by node. So please try below: > > -np 2 --map-by node:pe=4 --bind-to core > > You might not need to add --bind-to core, because it's default binding. > > Tetsuya > > > Hi, > > > > I see in v.1.7.5rc5 --cpus-per-proc is deprecated and is advised to > replace by --map-by <obj>:PE=N. > > I've tried this but I couldn't get the expected allocation of procs. > > > > For example I was running 2 procs on 2 nodes each with 2 sockets where a > socket has 4 cores. I wanted 1 proc per node and bound to all cores in one > of the sockets. I could get this by using > > > > --bind-to core: --map-by ppr:1:node --cpus-per-proc 4 -np 2 > > > > Then it'll show bindings as > > > > [i51:32274] MCW rank 0 bound to socket 0[core 0[hwt 0]], socket 0[core 1 > [hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 0]]: > [B/B/B/B][./././.] > > [i52:31765] MCW rank 1 bound to socket 0[core 0[hwt 0]], socket 0[core 1 > [hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 0]]: > [B/B/B/B][./././.] > > > > > > Is there a better way without using -cpus-per-proc as suggested to get > the same effect? > > > > Thank you, > > Saliya > > > > > > > > -- > > Saliya Ekanayake esal...@gmail.com <mailto:esal...@gmail.com> > > Cell 812-391-4914 <tel:812-391-4914> Home 812-961-6383 <tel:812-961-6383> > > http://saliya.org_______________________________________________ > > users mailing list > > users@open-mpi.orghttp://www.open-mpi.org/mailman/listinfo.cgi/users <http://www.open-mpi.org/mailman/listinfo.cgi/users> > > _______________________________________________ > users mailing list > us...@open-mpi.org <mailto:us...@open-mpi.org> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > -- > Saliya Ekanayake esal...@gmail.com <mailto:esal...@gmail.com> > Cell 812-391-4914 <tel:812-391-4914> Home 812-961-6383 <tel:812-961-6383> > http://saliya.org_______________________________________________ > users mailing list > users@open-mpi.orghttp://www.open-mpi.org/mailman/listinfo.cgi/users <http://www.open-mpi.org/mailman/listinfo.cgi/users> _______________________________________________ users mailing list us...@open-mpi.org <mailto:us...@open-mpi.org> http://www.open-mpi.org/mailman/listinfo.cgi/users -- Saliya Ekanayake esal...@gmail.com <mailto:esal...@gmail.com> Cell 812-391-4914 Home 812-961-6383 http://saliya.org _______________________________________________ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users