Hmmm...no, it worked just fine for me. It sounds like something else is going 
on.

Try configuring OMPI with --enable-debug, and then add -mca plm_base_verbose 10 
to get a better sense of what is going on.


On Jul 14, 2014, at 10:27 AM, Ralph Castain <r...@open-mpi.org> wrote:

> I confess I haven't tested no_tree_spawn in ages, so it is quite possible it 
> has suffered bit rot. I can try to take a look at it in a bit
> 
> 
> On Jul 14, 2014, at 10:13 AM, Ricardo Fernández-Perea 
> <rfernandezpe...@gmail.com> wrote:
> 
>> Thank you for the fast answer 
>> 
>> While that resolve my problem with cross ssh authentication   a command as
>> 
>> /opt/openmpi/bin/mpirun  --mca mtl mx --mca pml cm --mca 
>> plm_rsh_no_tree_spawn 1 -hostfile hostfile ompi_info
>> 
>> just hung with no output and although there is a ssh connexion no orte 
>> program is initiated in the destination nodes
>> 
>> and while 
>> 
>> /opt/openmpi/bin/mpirun  -host host18 ompi_info
>> 
>> works
>> 
>> /opt/openmpi/bin/mpirun  --mca plm_rsh_no_tree_spawn 1 -host host18 ompi_info
>> 
>> hangs, is there some condition in the use of this parameter.
>> 
>> Yours truly
>> 
>> Ricardo 
>> 
>> 
>> 
>> On Mon, Jul 14, 2014 at 6:35 PM, Ralph Castain <r...@open-mpi.org> wrote:
>> During the 1.7 series and for all follow-on series, OMPI changed to a mode 
>> where it launches a daemon on all allocated nodes at the startup of mpirun. 
>> This allows us to determine the hardware topology of the nodes and take that 
>> into account when mapping. You can override that behavior by either adding 
>> --novm to your cmd line (which will impact your mapping/binding options), or 
>> by specifying the hosts to use by editing your hostfile, or adding --host 
>> host1,host2 to your cmd line
>> 
>> The rsh launcher defaults to a tree-based pattern, thus requiring that we be 
>> able to ssh from one compute node to another. You can change that to a less 
>> scalable direct mode by adding
>> 
>> --mca plm_rsh_no_tree_spawn 1
>> 
>> to the cmd line
>> 
>> 
>> On Jul 14, 2014, at 9:21 AM, Ricardo Fernández-Perea 
>> <rfernandezpe...@gmail.com> wrote:
>> 
>> > I'm trying to update to openMPI 1.8.1 thru ssh  and Myrinet
>> >
>> > running a command as
>> >
>> > /opt/openmpi/bin/mpirun --verbose --mca mtl mx --mca pml cm  -hostfile 
>> > hostfile -np 16
>> >
>> > when the hostfile contain only two nodes as
>> >
>> > host1 slots=8 max-slots=8
>> > host2 slots=8 max-slots=8
>> >
>> > it runs perfectly but when the hostfile has a third node as
>> >
>> >
>> > host1 slots=8 max-slots=8
>> > host2 slots=8 max-slots=8
>> > host3 slots=8 max-slots=8
>> >
>> > it try to establish an ssh connection between  the running hosts1 and 
>> > host3 that should not run any process  that fails hanging the process 
>> > without signaling.
>> >
>> >
>> > my ompi_info is as follow
>> >
>> >                 Package: Open MPI XXX Distribution
>> >                 Open MPI: 1.8.1
>> >   Open MPI repo revision: r31483
>> >    Open MPI release date: Apr 22, 2014
>> >                 Open RTE: 1.8.1
>> >   Open RTE repo revision: r31483
>> >    Open RTE release date: Apr 22, 2014
>> >                     OPAL: 1.8.1
>> >       OPAL repo revision: r31483
>> >        OPAL release date: Apr 22, 2014
>> >                  MPI API: 3.0
>> >             Ident string: 1.8.1
>> >                   Prefix: /opt/openmpi
>> >  Configured architecture: x86_64-apple-darwin9.8.0
>> >           Configure host: XXXX
>> >            Configured by: XXXX
>> >            Configured on: Thu Jun 12 10:37:33 CEST 2014
>> >           Configure host: XXXX
>> >                 Built by: XXXX
>> >                 Built on: Thu Jun 12 11:13:16 CEST 2014
>> >               Built host: XXXX
>> >               C bindings: yes
>> >             C++ bindings: yes
>> >              Fort mpif.h: yes (single underscore)
>> >             Fort use mpi: yes (full: ignore TKR)
>> >        Fort use mpi size: deprecated-ompi-info-value
>> >         Fort use mpi_f08: yes
>> >  Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
>> >                           limitations in the ifort compiler, does not 
>> > support
>> >                           the following: array subsections, direct passthru
>> >                           (where possible) to underlying Open MPI's C
>> >                           functionality
>> >   Fort mpi_f08 subarrays: no
>> >            Java bindings: no
>> >   Wrapper compiler rpath: unnecessary
>> >               C compiler: icc
>> >      C compiler absolute: /opt/intel/Compiler/11.1/080/bin/intel64/icc
>> >   C compiler family name: INTEL
>> >       C compiler version: 1110.20091130
>> >             C++ compiler: icpc
>> >    C++ compiler absolute: /opt/intel/Compiler/11.1/080/bin/intel64/icpc
>> >            Fort compiler: ifort
>> >        Fort compiler abs: /opt/intel/Compiler/11.1/080/bin/intel64/ifort
>> >          Fort ignore TKR: yes (!DEC$ ATTRIBUTES NO_ARG_CHECK ::)
>> >    Fort 08 assumed shape: no
>> >       Fort optional args: yes
>> >       Fort BIND(C) (all): yes
>> >       Fort ISO_C_BINDING: yes
>> >  Fort SUBROUTINE BIND(C): yes
>> >        Fort TYPE,BIND(C): yes
>> >  Fort T,BIND(C,name="a"): yes
>> >             Fort PRIVATE: yes
>> >           Fort PROTECTED: yes
>> >            Fort ABSTRACT: yes
>> >        Fort ASYNCHRONOUS: yes
>> >           Fort PROCEDURE: yes
>> >  Fort f08 using wrappers: yes
>> >              C profiling: yes
>> >            C++ profiling: yes
>> >    Fort mpif.h profiling: yes
>> >   Fort use mpi profiling: yes
>> >    Fort use mpi_f08 prof: yes
>> >           C++ exceptions: no
>> >           Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: 
>> > yes,
>> >                           OMPI progress: no, ORTE progress: yes, Event lib:
>> >                           yes)
>> >            Sparse Groups: no
>> >   Internal debug support: no
>> >   MPI interface warnings: yes
>> >      MPI parameter check: runtime
>> > Memory profiling support: no
>> > Memory debugging support: no
>> >          libltdl support: yes
>> >    Heterogeneous support: no
>> >  mpirun default --prefix: no
>> >          MPI I/O support: yes
>> >        MPI_WTIME support: gettimeofday
>> >      Symbol vis. support: yes
>> >    Host topology support: yes
>> >           MPI extensions:
>> >    FT Checkpoint support: no (checkpoint thread: no)
>> >    C/R Enabled Debugging: no
>> >      VampirTrace support: yes
>> >   MPI_MAX_PROCESSOR_NAME: 256
>> >     MPI_MAX_ERROR_STRING: 256
>> >      MPI_MAX_OBJECT_NAME: 64
>> >         MPI_MAX_INFO_KEY: 36
>> >         MPI_MAX_INFO_VAL: 256
>> >        MPI_MAX_PORT_NAME: 1024
>> >   MPI_MAX_DATAREP_STRING: 128
>> >
>> >
>> > _______________________________________________
>> > users mailing list
>> > us...@open-mpi.org
>> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> > Link to this post: 
>> > http://www.open-mpi.org/community/lists/users/2014/07/24764.php
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2014/07/24765.php
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2014/07/24766.php
> 

Reply via email to