I confess I haven't tested no_tree_spawn in ages, so it is quite possible it 
has suffered bit rot. I can try to take a look at it in a bit


On Jul 14, 2014, at 10:13 AM, Ricardo Fernández-Perea 
<rfernandezpe...@gmail.com> wrote:

> Thank you for the fast answer 
> 
> While that resolve my problem with cross ssh authentication   a command as
> 
> /opt/openmpi/bin/mpirun  --mca mtl mx --mca pml cm --mca 
> plm_rsh_no_tree_spawn 1 -hostfile hostfile ompi_info
> 
> just hung with no output and although there is a ssh connexion no orte 
> program is initiated in the destination nodes
> 
> and while 
> 
> /opt/openmpi/bin/mpirun  -host host18 ompi_info
> 
> works
> 
> /opt/openmpi/bin/mpirun  --mca plm_rsh_no_tree_spawn 1 -host host18 ompi_info
> 
> hangs, is there some condition in the use of this parameter.
> 
> Yours truly
> 
> Ricardo 
> 
> 
> 
> On Mon, Jul 14, 2014 at 6:35 PM, Ralph Castain <r...@open-mpi.org> wrote:
> During the 1.7 series and for all follow-on series, OMPI changed to a mode 
> where it launches a daemon on all allocated nodes at the startup of mpirun. 
> This allows us to determine the hardware topology of the nodes and take that 
> into account when mapping. You can override that behavior by either adding 
> --novm to your cmd line (which will impact your mapping/binding options), or 
> by specifying the hosts to use by editing your hostfile, or adding --host 
> host1,host2 to your cmd line
> 
> The rsh launcher defaults to a tree-based pattern, thus requiring that we be 
> able to ssh from one compute node to another. You can change that to a less 
> scalable direct mode by adding
> 
> --mca plm_rsh_no_tree_spawn 1
> 
> to the cmd line
> 
> 
> On Jul 14, 2014, at 9:21 AM, Ricardo Fernández-Perea 
> <rfernandezpe...@gmail.com> wrote:
> 
> > I'm trying to update to openMPI 1.8.1 thru ssh  and Myrinet
> >
> > running a command as
> >
> > /opt/openmpi/bin/mpirun --verbose --mca mtl mx --mca pml cm  -hostfile 
> > hostfile -np 16
> >
> > when the hostfile contain only two nodes as
> >
> > host1 slots=8 max-slots=8
> > host2 slots=8 max-slots=8
> >
> > it runs perfectly but when the hostfile has a third node as
> >
> >
> > host1 slots=8 max-slots=8
> > host2 slots=8 max-slots=8
> > host3 slots=8 max-slots=8
> >
> > it try to establish an ssh connection between  the running hosts1 and host3 
> > that should not run any process  that fails hanging the process without 
> > signaling.
> >
> >
> > my ompi_info is as follow
> >
> >                 Package: Open MPI XXX Distribution
> >                 Open MPI: 1.8.1
> >   Open MPI repo revision: r31483
> >    Open MPI release date: Apr 22, 2014
> >                 Open RTE: 1.8.1
> >   Open RTE repo revision: r31483
> >    Open RTE release date: Apr 22, 2014
> >                     OPAL: 1.8.1
> >       OPAL repo revision: r31483
> >        OPAL release date: Apr 22, 2014
> >                  MPI API: 3.0
> >             Ident string: 1.8.1
> >                   Prefix: /opt/openmpi
> >  Configured architecture: x86_64-apple-darwin9.8.0
> >           Configure host: XXXX
> >            Configured by: XXXX
> >            Configured on: Thu Jun 12 10:37:33 CEST 2014
> >           Configure host: XXXX
> >                 Built by: XXXX
> >                 Built on: Thu Jun 12 11:13:16 CEST 2014
> >               Built host: XXXX
> >               C bindings: yes
> >             C++ bindings: yes
> >              Fort mpif.h: yes (single underscore)
> >             Fort use mpi: yes (full: ignore TKR)
> >        Fort use mpi size: deprecated-ompi-info-value
> >         Fort use mpi_f08: yes
> >  Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
> >                           limitations in the ifort compiler, does not 
> > support
> >                           the following: array subsections, direct passthru
> >                           (where possible) to underlying Open MPI's C
> >                           functionality
> >   Fort mpi_f08 subarrays: no
> >            Java bindings: no
> >   Wrapper compiler rpath: unnecessary
> >               C compiler: icc
> >      C compiler absolute: /opt/intel/Compiler/11.1/080/bin/intel64/icc
> >   C compiler family name: INTEL
> >       C compiler version: 1110.20091130
> >             C++ compiler: icpc
> >    C++ compiler absolute: /opt/intel/Compiler/11.1/080/bin/intel64/icpc
> >            Fort compiler: ifort
> >        Fort compiler abs: /opt/intel/Compiler/11.1/080/bin/intel64/ifort
> >          Fort ignore TKR: yes (!DEC$ ATTRIBUTES NO_ARG_CHECK ::)
> >    Fort 08 assumed shape: no
> >       Fort optional args: yes
> >       Fort BIND(C) (all): yes
> >       Fort ISO_C_BINDING: yes
> >  Fort SUBROUTINE BIND(C): yes
> >        Fort TYPE,BIND(C): yes
> >  Fort T,BIND(C,name="a"): yes
> >             Fort PRIVATE: yes
> >           Fort PROTECTED: yes
> >            Fort ABSTRACT: yes
> >        Fort ASYNCHRONOUS: yes
> >           Fort PROCEDURE: yes
> >  Fort f08 using wrappers: yes
> >              C profiling: yes
> >            C++ profiling: yes
> >    Fort mpif.h profiling: yes
> >   Fort use mpi profiling: yes
> >    Fort use mpi_f08 prof: yes
> >           C++ exceptions: no
> >           Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: yes,
> >                           OMPI progress: no, ORTE progress: yes, Event lib:
> >                           yes)
> >            Sparse Groups: no
> >   Internal debug support: no
> >   MPI interface warnings: yes
> >      MPI parameter check: runtime
> > Memory profiling support: no
> > Memory debugging support: no
> >          libltdl support: yes
> >    Heterogeneous support: no
> >  mpirun default --prefix: no
> >          MPI I/O support: yes
> >        MPI_WTIME support: gettimeofday
> >      Symbol vis. support: yes
> >    Host topology support: yes
> >           MPI extensions:
> >    FT Checkpoint support: no (checkpoint thread: no)
> >    C/R Enabled Debugging: no
> >      VampirTrace support: yes
> >   MPI_MAX_PROCESSOR_NAME: 256
> >     MPI_MAX_ERROR_STRING: 256
> >      MPI_MAX_OBJECT_NAME: 64
> >         MPI_MAX_INFO_KEY: 36
> >         MPI_MAX_INFO_VAL: 256
> >        MPI_MAX_PORT_NAME: 1024
> >   MPI_MAX_DATAREP_STRING: 128
> >
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> > Link to this post: 
> > http://www.open-mpi.org/community/lists/users/2014/07/24764.php
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2014/07/24765.php
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2014/07/24766.php

Reply via email to