Hi,
thanks a lot for your quick answers, and I see my error, it is "--bind-to none" instead of "--bind-to-none".

However, I need to be able to run "mpirun -np 2" without any binding argument and get a "--bind-to none" behaviour. I don't know if I can export an environment variable to do that, and I don't mind to re-compile with some flag I missed or to alter the code.

Any suggestion?

Albert

On 23/05/14 14:32, Ralph Castain wrote:
Note that the lama mapper described in those slides may not work as it hasn't 
been maintained in a while. However, you can use the map-by and bind-to options 
to do the same things.

If you want to disable binding, you can do so by adding "--bind-to none" to the cmd line, 
or via the MCA param "hwloc_base_binding_policy=none"

If you want to bind your process to multiple cores (say one per thread), then you can use "--map-by 
core:pe=N". Many hybrid users prefer to bind to a socket, with one process for each socket - that can be 
done with "--map-by socket --bind-to socket". This keeps all the threads in the same NUMA domain. 
If you aren't sure that each socket is its own NUMA domain, you can alternatively "--map-by numa 
--bind-to numa" - that'll keep you in your own NUMA domain regardless of whether that's at the socket 
level or elsewhere.

I'm working on adding a full description of the mapping/binding system to our 
web site.

HTH
Ralph

On May 23, 2014, at 6:22 AM, Brock Palen <bro...@umich.edu> wrote:

Albert,

Actually doing affinity correctly for hybrid got easier in OpenMPI 1.7 and 
newer,  In the past you had to make a lot of assumptions, stride by node etc,

Now you can define a layout:

http://blogs.cisco.com/performance/eurompi13-cisco-slides-open-mpi-process-affinity-user-interface/

Brock Palen
www.umich.edu/~brockp
CAEN Advanced Computing
XSEDE Campus Champion
bro...@umich.edu
(734)936-1985



On May 23, 2014, at 9:19 AM, Albert Solernou <albert.soler...@oerc.ox.ac.uk> 
wrote:

Hi,
after compiling and installing OpenMPI 1.8.1, I find that OpenMPI is pinning 
processes onto cores. Although this may be
desirable on some cases, it is a complete disaster when runnning hybrid 
OpenMP-MPI applications. Therefore, I want to disable this behaviour, but don't 
know how.

I configured OpenMPI with:
./configure \
        --prefix=$OPENMPI_HOME \
        --with-psm \
        --with-tm=/system/software/arcus/lib/PBS/11.3.0.121723 \
        --enable-mpirun-prefix-by-default \
        --enable-mpi-thread-multiple

and:
ompi_info | grep paffinity
does not report anything. However,
mpirun -np 2 --report-bindings ./wrk
reports bindings:
[login3:04574] MCW rank 1 bound to socket 0[core 1[hwt 0-1]]: 
[../BB/../../../../../..][../../../../../../../..]
[login3:04574] MCW rank 0 bound to socket 0[core 0[hwt 0-1]]: 
[BB/../../../../../../..][../../../../../../../..]
but they cannot be disabled as:
mpirun -np 2 --bind-to-none ./wrk
returns:
mpirun: Error: unknown option "--bind-to-none"

Any idea on what went wrong?

Best,
Albert

--
---------------------------------
Dr. Albert Solernou
Research Associate
Oxford Supercomputing Centre,
University of Oxford
Tel: +44 (0)1865 610631
---------------------------------
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


--
---------------------------------
  Dr. Albert Solernou
  Research Associate
  Oxford Supercomputing Centre,
  University of Oxford
  Tel: +44 (0)1865 610631
---------------------------------

Reply via email to