Hi

> > You'll want to look at orte/mca/rmaps/rank_file/rmaps_rank_file.c 
> > - the bit map is now computed in mpirun and then sent to the daemons
> 
> Actually, I'm getting lost in this code.  Anyhow, I don't think
> the problem is related to Solaris.  I think it's also on Linux. 
> E.g., I can reproduce the problem with 1.9a1r28035 on Linux using 
> GCC compilers.
> 
> Siegmar: can you confirm this is a problem also on Linux?  E.g.,
> with OMPI 1.9, on one of your Linux nodes (linpc0?) try
> 
>      % cat myrankfile
>      rank 0=linpc0 slot=0:1
>      % mpirun --report-bindings --rankfile myrankfile numactl --show
> 
> For me, the binding I get is not 0:1 but 0,1.

I get the following outputs for openmpi-1.6.4rc4 (without your patch)
and openmpi-1.9 (both compiled with Sun C 5.12).

linpc1 rankfiles 96 cat rf_1_linux
rank 0=linpc1 slot=0:1
linpc1 rankfiles 97 mpirun --report-bindings --rankfile rf_1_linux \
  numactl --show
[linpc1:15882] MCW rank 0 bound to socket 0[core 1]:
  [. B][. .] (slot list 0:1)
policy: preferred
preferred node: 0
physcpubind: 1 
cpubind: 0 
nodebind: 0 
membind: 0 1 
linpc1 rankfiles 98 ompi_info | grep "MPI:"
                Open MPI: 1.6.4rc4r28022



linpc1 rankfiles 96 mpirun --report-bindings --rankfile rf_1_linux \
  numactl --show
[linpc1:16061] MCW rank 0 bound to socket 0[core 0[hwt 0]],
  socket 0[core 1[hwt 0]]: [B/B][./.]
policy: default
preferred node: current
physcpubind: 0 1 
cpubind: 0 
nodebind: 0 
membind: 0 1 
linpc1 rankfiles 97 ompi_info | grep "MPI:"
                Open MPI: 1.9a1r28035


Today I will build the current versions of openmpi-1.6.4rc4 and
openmpi-1.9, so that I can test a rankfile with two machines
tomorrow.


Kind regards

Siegmar

Reply via email to