Re: [OMPI users] Trying to map to socket #0 on each node

2015-12-07 Thread Carl Ponder

*On 12/06/2015 11:07 AM, Carl Ponder wrote:*


I'm trying to run a multi-node job but I want to map all of the
processes to cores on socket #0 only.
I'm having a hard time figuring out how to do this, the obvious
combinations

mpirun -n  8 -npernode 4 -report-bindings ...
mpirun -n  8 -npernode 4 --map-by core -report-bindings ...
mpirun -n  8 -npernode 4 -cpu-set S0 -report-bindings ...
mpirun -n  8 --map-by ppr:4:node,ppr:4:socket -report-bindings ...
mpirun -n  8 -npernode 4 -bind-to slot=0:0,2,4,6
-report-bindings ...
mpirun -n  8 -npernode 4 -bind-to slot=0:0,0:2,0:4,0:6
-report-bindings ...
mpirun -n  8 -npernode 4 --map-by core:PE=2 -bind-to core
-report-bindings ...

all are reported as having conflicting resource requirements.


*On 12/06/2015 11:28 AM, Ralph Castain wrote:*


You want "-bind-to socket -slot-list=0,2,4,6"
Or if you want each process bound to a single core on the socket,
then change “socket” to “core” in the above


So far I can't get this to work. Using the above form

   mpirun -n 8 *-bind-to socket --slot-list 0,2,4,6* -report-bindings ...

it says that it's a mis-specification:

   Conflicting directives for binding policy are causing the policy
   to be redefined:

  New policy:   socket
  Prior policy:  SOCKET

   Please check that only one policy is defined.

If I treat the socket-binding as redundant and just use this

   mpirun -n 8 -*-slot-list 0,2,4,6* -report-bindings ...

it looks like it's ignoring slots 0,2,4,6 available on the second node:

   A rank is missing its location specification:

  Rank:0
  Rank file:   (null)

   All processes must have their location specified in the rank file.
   Either
   add an entry to the file, or provide a default slot_list to use for
   any unspecified ranks.

(One question is whether it is interacting with Torque correctly).
Trying to force it to split the processes across nodes

   mpirun -n 8 *-npernode 4 --slot-list 0,2,4,6* -report-bindings 

gives

   Conflicting directives for mapping policy are causing the policy
   to be redefined:

  New policy:   RANK_FILE
  Prior policy:  UNKNOWN

   Please check that only one policy is defined.

Do you know what to do here? I'm using OpenMPI 1.10.1.
Thanks,

Carl


---
This email message is for the sole use of the intended recipient(s) and may 
contain
confidential information.  Any unauthorized review, use, disclosure or 
distribution
is prohibited.  If you are not the intended recipient, please contact the 
sender by
reply email and destroy all copies of the original message.
---


Re: [OMPI users] Trying to map to socket #0 on each node

2015-12-07 Thread Nick Papior
Couldn't it be that the slot list should be 0,1,2,3?
It depends on the setup.
You can get some more information about _what it does_ by using
--report-bindings (when/if it succeeds).


2015-12-07 16:18 GMT+01:00 Carl Ponder :

> *On 12/06/2015 11:07 AM, Carl Ponder wrote:*
>
> I'm trying to run a multi-node job but I want to map all of the processes
> to cores on socket #0 only.
> I'm having a hard time figuring out how to do this, the obvious
> combinations
>
> mpirun -n  8 -npernode 4 -report-bindings ...
> mpirun -n  8 -npernode 4 --map-by core -report-bindings ...
> mpirun -n  8 -npernode 4 -cpu-set S0 -report-bindings ...
> mpirun -n  8 --map-by ppr:4:node,ppr:4:socket -report-bindings ...
> mpirun -n  8 -npernode 4 -bind-to slot=0:0,2,4,6 -report-bindings ...
> mpirun -n  8 -npernode 4 -bind-to slot=0:0,0:2,0:4,0:6 -report-bindings ...
> mpirun -n  8 -npernode 4 --map-by core:PE=2 -bind-to core -report-bindings
> ...
>
> all are reported as having conflicting resource requirements.
>
> * On 12/06/2015 11:28 AM, Ralph Castain wrote:*
>
> You want "-bind-to socket -slot-list=0,2,4,6"
> Or if you want each process bound to a single core on the socket, then
> change “socket” to “core” in the above
>
> So far I can't get this to work. Using the above form
>
> mpirun -n 8 *-bind-to socket --slot-list 0,2,4,6* -report-bindings ...
>
> it says that it's a mis-specification:
>
> Conflicting directives for binding policy are causing the policy
> to be redefined:
>
>   New policy:   socket
>   Prior policy:  SOCKET
>
> Please check that only one policy is defined.
>
> If I treat the socket-binding as redundant and just use this
>
> mpirun -n 8 -*-slot-list 0,2,4,6* -report-bindings ...
>
> it looks like it's ignoring slots 0,2,4,6 available on the second node:
>
> A rank is missing its location specification:
>
>   Rank:0
>   Rank file:   (null)
>
> All processes must have their location specified in the rank file. Either
> add an entry to the file, or provide a default slot_list to use for
> any unspecified ranks.
>
> (One question is whether it is interacting with Torque correctly).
> Trying to force it to split the processes across nodes
>
> mpirun -n 8 *-npernode 4 --slot-list 0,2,4,6* -report-bindings 
>
> gives
>
> Conflicting directives for mapping policy are causing the policy
> to be redefined:
>
>   New policy:   RANK_FILE
>   Prior policy:  UNKNOWN
>
> Please check that only one policy is defined.
>
> Do you know what to do here? I'm using OpenMPI 1.10.1.
> Thanks,
>
> Carl
>
> --
> This email message is for the sole use of the intended recipient(s) and
> may contain confidential information.  Any unauthorized review, use,
> disclosure or distribution is prohibited.  If you are not the intended
> recipient, please contact the sender by reply email and destroy all copies
> of the original message.
> --
>
> ___
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2015/12/28139.php
>



-- 
Kind regards Nick


Re: [OMPI users] my_sense in ompi_osc_sm_module_t not always protected by OPAL_HAVE_POSIX_THREADS

2015-12-07 Thread Åke Sandgren
The #if OPAL_HAVE_POSIX_THREADS is still there around my_sense in 
osc_sm.h in 1.10.1


On 06/29/2015 05:42 PM, Åke Sandgren wrote:

Yeah, i thought so. Well code reductions are good when correct :-)

On 06/29/2015 05:39 PM, Nathan Hjelm wrote:


Open MPI has required posix threads for some time. The check for
OPAL_HAVE_POSIX_THREADS in ompi/mca/osc/sm/osc_sm.h is stale and should
be removed. I will clean that out in master, 1.8, and 1.10.



--
Ake Sandgren, HPC2N, Umea University, S-90187 Umea, Sweden
Internet: a...@hpc2n.umu.se   Phone: +46 90 7866134 Fax: +46 90-580 14
Mobile: +46 70 7716134 WWW: http://www.hpc2n.umu.se