<\begin hijacking this thread>

I second Saliya's thanks to Tetsuya.
I've been following this thread, to learn a bit more about
how to use hardware locality with OpenMPI effectively.
[I am still using "--bycore"+"--bind-to-core" in most cases,
and "--cpus-per-proc" occasionally when in hybrid MPI+OpenMP mode.]

When it comes to hardware locality,
the syntax and the functionality has changed fast and significantly
in the recent past.
Hence, it would be great if the OpenMPI web page could provide pointers
for the type of external documentation that Tetsuya just sent.
Perhaps also some additional guidelines and comments
on what is available on each release/series of OpenMPI,
and how to use these options.

There is some material about hwloc,
but I can't see much about lama ( which means "mud" in my
first language :) ).
We can hardly learn things like that from the mpiexec man page
alone, although it has very good examples.

Thank you,
Gus Correa

<\end hijacking of this thread>

On 03/27/2014 11:38 AM, Saliya Ekanayake wrote:
Thank you, this is really helpful.

Saliya


On Thu, Mar 27, 2014 at 5:11 AM, <tmish...@jcity.maeda.co.jp
<mailto:tmish...@jcity.maeda.co.jp>> wrote:



    Mapping and binding is related to so called process affinity.
    It's a bit difficult for me to explain ...

    So please see this URL below(especially the first half part
    of it - from 1 to 20 pages):
    
http://www.slideshare.net/jsquyres/open-mpi-explorations-in-process-affinity-eurompi13-presentation

    Although these slides by Jeff are the explanation for LAMA,
    which is another mapping system installed in the openmpi-1.7
    series, I guess you can easily understand what is mapping and
    binding in general terms.

    Tetsuya

     > Thank you Tetsuya - it worked.
     >
     > Btw. what's the difference between mapping and binding? I think I
    am bit
    confused here.
     >
     > Thank you,
     > Saliya
     >
     >
     > On Thu, Mar 27, 2014 at 4:19 AM,  <tmish...@jcity.maeda.co.jp
    <mailto:tmish...@jcity.maeda.co.jp>>wrote:
     >
     >
     > Hi Saliya,
     >
     > What you want to do is map-by node. So please try below:
     >
     > -np 2 --map-by node:pe=4 --bind-to core
     >
     > You might not need to add --bind-to core, because it's default
    binding.
     >
     > Tetsuya
     >
     > > Hi,
     > >
     > > I see in v.1.7.5rc5 --cpus-per-proc is deprecated and is advised to
     > replace by --map-by <obj>:PE=N.
     > > I've tried this but I couldn't get the expected allocation of
    procs.
     > >
     > > For example I was running 2 procs on 2 nodes each with 2
    sockets where
    a
     > socket has 4 cores. I wanted 1 proc per node and bound to all
    cores in
    one
     > of the sockets. I could get this by using
     > >
     > > --bind-to core: --map-by ppr:1:node --cpus-per-proc 4 -np 2
     > >
     > > Then it'll show bindings as
     > >
     > > [i51:32274] MCW rank 0 bound to socket 0[core 0[hwt 0]], socket
    0[core
    1
     > [hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 0]]:
     > [B/B/B/B][./././.]
     > > [i52:31765] MCW rank 1 bound to socket 0[core 0[hwt 0]], socket
    0[core
    1
     > [hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 0]]:
     > [B/B/B/B][./././.]
     > >
     > >
     > > Is there a better way without using -cpus-per-proc as suggested
    to get
     > the same effect?
     > >
     > > Thank you,
     > > Saliya
     > >
     > >
     > >
     > > --
     > > Saliya Ekanayake esal...@gmail.com <mailto:esal...@gmail.com>
     > > Cell 812-391-4914 <tel:812-391-4914> Home 812-961-6383
    <tel:812-961-6383>
     > > http://saliya.org_______________________________________________
     > > users mailing list
     > >
    users@open-mpi.orghttp://www.open-mpi.org/mailman/listinfo.cgi/users
    <http://www.open-mpi.org/mailman/listinfo.cgi/users>
     >
     > _______________________________________________
     > users mailing list
     > us...@open-mpi.org <mailto:us...@open-mpi.org>
     > http://www.open-mpi.org/mailman/listinfo.cgi/users
     >
     >
     >
     > --
     > Saliya Ekanayake esal...@gmail.com <mailto:esal...@gmail.com>
     > Cell 812-391-4914 <tel:812-391-4914> Home 812-961-6383
    <tel:812-961-6383>
     > http://saliya.org_______________________________________________
     > users mailing list
     >
    users@open-mpi.orghttp://www.open-mpi.org/mailman/listinfo.cgi/users
    <http://www.open-mpi.org/mailman/listinfo.cgi/users>

    _______________________________________________
    users mailing list
    us...@open-mpi.org <mailto:us...@open-mpi.org>
    http://www.open-mpi.org/mailman/listinfo.cgi/users




--
Saliya Ekanayake esal...@gmail.com <mailto:esal...@gmail.com>
Cell 812-391-4914 Home 812-961-6383
http://saliya.org


_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


Reply via email to