Note that the lama mapper described in those slides may not work as it hasn't been maintained in a while. However, you can use the map-by and bind-to options to do the same things.
If you want to disable binding, you can do so by adding "--bind-to none" to the cmd line, or via the MCA param "hwloc_base_binding_policy=none" If you want to bind your process to multiple cores (say one per thread), then you can use "--map-by core:pe=N". Many hybrid users prefer to bind to a socket, with one process for each socket - that can be done with "--map-by socket --bind-to socket". This keeps all the threads in the same NUMA domain. If you aren't sure that each socket is its own NUMA domain, you can alternatively "--map-by numa --bind-to numa" - that'll keep you in your own NUMA domain regardless of whether that's at the socket level or elsewhere. I'm working on adding a full description of the mapping/binding system to our web site. HTH Ralph On May 23, 2014, at 6:22 AM, Brock Palen <bro...@umich.edu> wrote: > Albert, > > Actually doing affinity correctly for hybrid got easier in OpenMPI 1.7 and > newer, In the past you had to make a lot of assumptions, stride by node etc, > > Now you can define a layout: > > http://blogs.cisco.com/performance/eurompi13-cisco-slides-open-mpi-process-affinity-user-interface/ > > Brock Palen > www.umich.edu/~brockp > CAEN Advanced Computing > XSEDE Campus Champion > bro...@umich.edu > (734)936-1985 > > > > On May 23, 2014, at 9:19 AM, Albert Solernou <albert.soler...@oerc.ox.ac.uk> > wrote: > >> Hi, >> after compiling and installing OpenMPI 1.8.1, I find that OpenMPI is pinning >> processes onto cores. Although this may be >> desirable on some cases, it is a complete disaster when runnning hybrid >> OpenMP-MPI applications. Therefore, I want to disable this behaviour, but >> don't know how. >> >> I configured OpenMPI with: >> ./configure \ >> --prefix=$OPENMPI_HOME \ >> --with-psm \ >> --with-tm=/system/software/arcus/lib/PBS/11.3.0.121723 \ >> --enable-mpirun-prefix-by-default \ >> --enable-mpi-thread-multiple >> >> and: >> ompi_info | grep paffinity >> does not report anything. However, >> mpirun -np 2 --report-bindings ./wrk >> reports bindings: >> [login3:04574] MCW rank 1 bound to socket 0[core 1[hwt 0-1]]: >> [../BB/../../../../../..][../../../../../../../..] >> [login3:04574] MCW rank 0 bound to socket 0[core 0[hwt 0-1]]: >> [BB/../../../../../../..][../../../../../../../..] >> but they cannot be disabled as: >> mpirun -np 2 --bind-to-none ./wrk >> returns: >> mpirun: Error: unknown option "--bind-to-none" >> >> Any idea on what went wrong? >> >> Best, >> Albert >> >> -- >> --------------------------------- >> Dr. Albert Solernou >> Research Associate >> Oxford Supercomputing Centre, >> University of Oxford >> Tel: +44 (0)1865 610631 >> --------------------------------- >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users