Thankyou Ralph.   i guess the information I did not have in my head was
that   core = physical core (not hyperthreaded core)

On 18 July 2016 at 14:45, Ralph Castain <r...@open-mpi.org> wrote:

> It sounds like you just want to bind procs to cores since each core is
> composed of 2 HTs. So a simple “--map-by core --bind-to core" should do the
> trick.
>
> FWIW: the affinity settings are controlled by the bind-to <foo> option.
> You can use “mpirun -h”  to get the list of supported options and a little
> explanation:
>
> *--bind-to <foo>* Bind processes to the specified object, defaults to
> *core*. Supported options include slot, hwthread, core, l1cache, l2cache,
> l3cache, socket, numa, board, and none.
>
> https://www.open-mpi.org/doc/current/man1/mpirun.1.php#sect9
>
>
>
>
> On Jul 17, 2016, at 11:25 PM, John Hearns <hear...@googlemail.com> wrote:
>
> Please can someone point me towards the affinity settings for:
> OpenMPI 1.10   used with Slurm version 15
>
> I have some nodes with 2630-v4 processors.
> So 10 cores per socket / 20 hyperthreads
> Hyperthreading is enabled.
> I would like to set affinity for 20 processes per node,
> so that the processes are pinned to every second HT core - ie one process
> per physical thread.
>
> I'm sure this is quite easy...
>
> Thankyou
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2016/07/29674.php
>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2016/07/29676.php
>

Reply via email to