On May 23, 2014, at 6:58 AM, Albert Solernou <albert.soler...@oerc.ox.ac.uk> 
wrote:

> Hi,
> thanks a lot for your quick answers, and I see my error, it is "--bind-to 
> none" instead of "--bind-to-none".
> 
> However, I need to be able to run "mpirun -np 2" without any binding argument 
> and get a "--bind-to none" behaviour. I don't know if I can export an 
> environment variable to do that, and I don't mind to re-compile with some 
> flag I missed or to alter the code.
> 
> Any suggestion?

Obviously, you could set an variable in your environment, so I'm assuming there 
is some other limitation in play here? If so, you could always put the MCA 
param in the default mca param file - we'll pick it up from there.

> 
> Albert
> 
> On 23/05/14 14:32, Ralph Castain wrote:
>> Note that the lama mapper described in those slides may not work as it 
>> hasn't been maintained in a while. However, you can use the map-by and 
>> bind-to options to do the same things.
>> 
>> If you want to disable binding, you can do so by adding "--bind-to none" to 
>> the cmd line, or via the MCA param "hwloc_base_binding_policy=none"
>> 
>> If you want to bind your process to multiple cores (say one per thread), 
>> then you can use "--map-by core:pe=N". Many hybrid users prefer to bind to a 
>> socket, with one process for each socket - that can be done with "--map-by 
>> socket --bind-to socket". This keeps all the threads in the same NUMA 
>> domain. If you aren't sure that each socket is its own NUMA domain, you can 
>> alternatively "--map-by numa --bind-to numa" - that'll keep you in your own 
>> NUMA domain regardless of whether that's at the socket level or elsewhere.
>> 
>> I'm working on adding a full description of the mapping/binding system to 
>> our web site.
>> 
>> HTH
>> Ralph
>> 
>> On May 23, 2014, at 6:22 AM, Brock Palen <bro...@umich.edu> wrote:
>> 
>>> Albert,
>>> 
>>> Actually doing affinity correctly for hybrid got easier in OpenMPI 1.7 and 
>>> newer,  In the past you had to make a lot of assumptions, stride by node 
>>> etc,
>>> 
>>> Now you can define a layout:
>>> 
>>> http://blogs.cisco.com/performance/eurompi13-cisco-slides-open-mpi-process-affinity-user-interface/
>>> 
>>> Brock Palen
>>> www.umich.edu/~brockp
>>> CAEN Advanced Computing
>>> XSEDE Campus Champion
>>> bro...@umich.edu
>>> (734)936-1985
>>> 
>>> 
>>> 
>>> On May 23, 2014, at 9:19 AM, Albert Solernou 
>>> <albert.soler...@oerc.ox.ac.uk> wrote:
>>> 
>>>> Hi,
>>>> after compiling and installing OpenMPI 1.8.1, I find that OpenMPI is 
>>>> pinning processes onto cores. Although this may be
>>>> desirable on some cases, it is a complete disaster when runnning hybrid 
>>>> OpenMP-MPI applications. Therefore, I want to disable this behaviour, but 
>>>> don't know how.
>>>> 
>>>> I configured OpenMPI with:
>>>> ./configure \
>>>>        --prefix=$OPENMPI_HOME \
>>>>        --with-psm \
>>>>        --with-tm=/system/software/arcus/lib/PBS/11.3.0.121723 \
>>>>        --enable-mpirun-prefix-by-default \
>>>>        --enable-mpi-thread-multiple
>>>> 
>>>> and:
>>>> ompi_info | grep paffinity
>>>> does not report anything. However,
>>>> mpirun -np 2 --report-bindings ./wrk
>>>> reports bindings:
>>>> [login3:04574] MCW rank 1 bound to socket 0[core 1[hwt 0-1]]: 
>>>> [../BB/../../../../../..][../../../../../../../..]
>>>> [login3:04574] MCW rank 0 bound to socket 0[core 0[hwt 0-1]]: 
>>>> [BB/../../../../../../..][../../../../../../../..]
>>>> but they cannot be disabled as:
>>>> mpirun -np 2 --bind-to-none ./wrk
>>>> returns:
>>>> mpirun: Error: unknown option "--bind-to-none"
>>>> 
>>>> Any idea on what went wrong?
>>>> 
>>>> Best,
>>>> Albert
>>>> 
>>>> --
>>>> ---------------------------------
>>>> Dr. Albert Solernou
>>>> Research Associate
>>>> Oxford Supercomputing Centre,
>>>> University of Oxford
>>>> Tel: +44 (0)1865 610631
>>>> ---------------------------------
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
> 
> -- 
> ---------------------------------
>  Dr. Albert Solernou
>  Research Associate
>  Oxford Supercomputing Centre,
>  University of Oxford
>  Tel: +44 (0)1865 610631
> ---------------------------------
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to