I’m afraid none of the current options is going to do that right now. I’ll put a note on my to-do list to look at this, but I can’t promise when I’ll get to it.
> On Apr 24, 2017, at 3:13 AM, Heinz-Ado Arnolds <arno...@mpa-garching.mpg.de> > wrote: > > Dear Ralph, > > thanks for this new hint. Unfortunately I don't see how that would fulfill > all my requirements: > > I like to have 8 OpenMPI jobs for 2 nodes -> 4 OpenMPI jobs per node -> 2 per > socket, each executing one OpenMP job with 5 threads > > mpirun -np 8 --map-by ppr:4:node:pe=5 ... > > How can I connect this with the constraint 1 threat per core: > > [pascal-3-06:14965] ... > [B./B./B./B./B./../../../../..][../../../../../../../../../..] > [pascal-3-06:14965] ... > [../../../../../B./B./B./B./B.][../../../../../../../../../..] > [pascal-3-06:14965] ... > [../../../../../../../../../..][B./B./B./B./B./../../../../..] > [pascal-3-06:14965] ... > [../../../../../../../../../..][../../../../../B./B./B./B./B./] > [pascal-3-07:21027] ... > [B./B./B./B./B./../../../../..][../../../../../../../../../..] > [pascal-3-07:21027] ... > [../../../../../B./B./B./B./B.][../../../../../../../../../..] > [pascal-3-07:21027] ... > [../../../../../../../../../..][B./B./B./B./B./../../../../..] > [pascal-3-07:21027] ... > [../../../../../../../../../..][../../../../../B./B./B./B./B./] > > Cheers, > > Ado > > On 22.04.2017 16:45, r...@open-mpi.org wrote: >> Sorry for delayed response. I’m glad that option solved the problem. We’ll >> have to look at that configure option - shouldn’t be too hard. >> >> As for the mapping you requested - no problem! Here’s the cmd line: >> >> mpirun --map-by ppr:1:core --bind-to hwthread >> >> Ralph >> >>> On Apr 19, 2017, at 2:51 AM, Heinz-Ado Arnolds >>> <arno...@mpa-garching.mpg.de> wrote: >>> >>> Dear Ralph, dear Gilles, >>> >>> thanks a lot for your help! The hints to use ":pe=<n>" and to install >>> libnuma have been the keys to solve my problems. >>> >>> Perhaps it would not be a bad idea to include --enable-libnuma in the >>> configure help, and make it a default, so that one has to specify >>> --disable-libnuma if he really likes to work without numactl. The option is >>> already checked in configure (framework in >>> opal/mca/hwloc/hwloc1112/hwloc/config/hwloc.m4). >>> >>> One qestion remains: I now get a binding like >>> [pascal-3-06:03036] MCW rank 0 bound to socket 0[core 0[hwt 0-1]], socket >>> 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 3[hwt 0-1]], >>> socket 0[core 4[hwt 0-1]]: >>> [BB/BB/BB/BB/BB/../../../../..][../../../../../../../../../..] >>> and OpenMP uses just "hwt 0" of each core, what is very welcome. But is >>> there a way to get a binding like >>> [pascal-3-06:03036] MCW rank 0 bound to socket 0[core 0[hwt 0]], socket >>> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 0]], socket >>> 0[core 4[hwt 0]]: >>> [B./B./B./B./B./../../../../..][../../../../../../../../../..] >>> from OpenMPI directly? >>> >>> Cheers and thanks again, >>> >>> Ado >>> >>> On 13.04.2017 17:34, r...@open-mpi.org wrote: >>>> Yeah, we need libnuma to set the memory binding. There is a param to turn >>>> off the warning if installing libnuma is problematic, but it helps your >>>> performance if the memory is kept local to the proc >>>> >>>>> On Apr 13, 2017, at 8:26 AM, Heinz-Ado Arnolds >>>>> <arno...@mpa-garching.mpg.de> wrote: >>>>> >>>>> Dear Ralph, >>>>> >>>>> thanks a lot for this valuable advice. Binding now works like expected! >>>>> >>>>> Since adding the ":pe=" option I'm getting warnings >>>>> >>>>> WARNING: a request was made to bind a process. While the system >>>>> supports binding the process itself, at least one node does NOT >>>>> support binding memory to the process location. >>>>> >>>>> Node: pascal-1-05 >>>>> ... >>>>> >>>>> even if I choose parameters so that binding is like exactly as before >>>>> without ":pe=". I don't have libnuma installed on the cluster. Might that >>>>> really be the cause of the warning? >>>>> >>>>> Thanks a lot, have a nice Easter days >>>>> >>>>> Ado >>>>> >>>>> On 13.04.2017 15:49, r...@open-mpi.org wrote: >>>>>> You can always specify a particular number of cpus to use for each >>>>>> process by adding it to the map-by directive: >>>>>> >>>>>> mpirun -np 8 --map-by ppr:2:socket:pe=5 --use-hwthread-cpus >>>>>> -report-bindings --mca plm_rsh_agent "qrsh" ./myid >>>>>> >>>>>> would map 2 processes to each socket, binding each process to 5 HTs on >>>>>> that socket (since you told us to treat HTs as independent cpus). If you >>>>>> want us to bind to you 5 cores, then you need to remove that >>>>>> --use-hwthread-cpus directive. >>>>>> >>>>>> As I said earlier in this thread, we are actively working with the >>>>>> OpenMP folks on a mechanism by which the two sides can coordinate these >>>>>> actions so it will be easier to get the desired behavior. For now, >>>>>> though, hopefully this will suffice. >>>>>> >>>>>>> On Apr 13, 2017, at 6:31 AM, Heinz-Ado Arnolds >>>>>>> <arno...@mpa-garching.mpg.de> wrote: >>>>>>> >>>>>>> On 13.04.2017 15:20, gil...@rist.or.jp wrote: >>>>>>> ... >>>>>>>> in your second case, there are 2 things >>>>>>>> - MPI binds to socket, that is why two MPI tasks are assigned the same >>>>>>>> hyperthreads >>>>>>>> - the GNU OpenMP runtime looks unable to figure out 2 processes use >>>>>>>> the >>>>>>>> same cores, and hence end up binding >>>>>>>> the OpenMP threads to the same cores. >>>>>>>> my best bet is you should bind a MPI tasks to 5 cores instead of one >>>>>>>> socket. >>>>>>>> i do not know the syntax off hand, and i am sure Ralph will help you >>>>>>>> with that >>>>>>> >>>>>>> Thanks, would be great if someone has that syntax. >>>>>>> >>>>>>> Cheers, >>>>>>> >>>>>>> Ado >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org >>>>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org >>>>>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >>>>>> >>>>> >>>> >>>> >>> >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://rfd.newmexicoconsortium.org/mailman/listinfo/users >> >> > _______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users