Thank you for the fast answer

While that resolve my problem with cross ssh authentication   a command as

/opt/openmpi/bin/mpirun  --mca mtl mx --mca pml cm --mca
plm_rsh_no_tree_spawn 1 -hostfile hostfile ompi_info

just hung with no output and although there is a ssh connexion no orte
program is initiated in the destination nodes

and while

/opt/openmpi/bin/mpirun  -host host18 ompi_info

works

/opt/openmpi/bin/mpirun  --mca plm_rsh_no_tree_spawn 1 -host host18
ompi_info

hangs, is there some condition in the use of this parameter.

Yours truly

Ricardo



On Mon, Jul 14, 2014 at 6:35 PM, Ralph Castain <r...@open-mpi.org> wrote:

> During the 1.7 series and for all follow-on series, OMPI changed to a mode
> where it launches a daemon on all allocated nodes at the startup of mpirun.
> This allows us to determine the hardware topology of the nodes and take
> that into account when mapping. You can override that behavior by either
> adding --novm to your cmd line (which will impact your mapping/binding
> options), or by specifying the hosts to use by editing your hostfile, or
> adding --host host1,host2 to your cmd line
>
> The rsh launcher defaults to a tree-based pattern, thus requiring that we
> be able to ssh from one compute node to another. You can change that to a
> less scalable direct mode by adding
>
> --mca plm_rsh_no_tree_spawn 1
>
> to the cmd line
>
>
> On Jul 14, 2014, at 9:21 AM, Ricardo Fernández-Perea <
> rfernandezpe...@gmail.com> wrote:
>
> > I'm trying to update to openMPI 1.8.1 thru ssh  and Myrinet
> >
> > running a command as
> >
> > /opt/openmpi/bin/mpirun --verbose --mca mtl mx --mca pml cm  -hostfile
> hostfile -np 16
> >
> > when the hostfile contain only two nodes as
> >
> > host1 slots=8 max-slots=8
> > host2 slots=8 max-slots=8
> >
> > it runs perfectly but when the hostfile has a third node as
> >
> >
> > host1 slots=8 max-slots=8
> > host2 slots=8 max-slots=8
> > host3 slots=8 max-slots=8
> >
> > it try to establish an ssh connection between  the running hosts1 and
> host3 that should not run any process  that fails hanging the process
> without signaling.
> >
> >
> > my ompi_info is as follow
> >
> >                 Package: Open MPI XXX Distribution
> >                 Open MPI: 1.8.1
> >   Open MPI repo revision: r31483
> >    Open MPI release date: Apr 22, 2014
> >                 Open RTE: 1.8.1
> >   Open RTE repo revision: r31483
> >    Open RTE release date: Apr 22, 2014
> >                     OPAL: 1.8.1
> >       OPAL repo revision: r31483
> >        OPAL release date: Apr 22, 2014
> >                  MPI API: 3.0
> >             Ident string: 1.8.1
> >                   Prefix: /opt/openmpi
> >  Configured architecture: x86_64-apple-darwin9.8.0
> >           Configure host: XXXX
> >            Configured by: XXXX
> >            Configured on: Thu Jun 12 10:37:33 CEST 2014
> >           Configure host: XXXX
> >                 Built by: XXXX
> >                 Built on: Thu Jun 12 11:13:16 CEST 2014
> >               Built host: XXXX
> >               C bindings: yes
> >             C++ bindings: yes
> >              Fort mpif.h: yes (single underscore)
> >             Fort use mpi: yes (full: ignore TKR)
> >        Fort use mpi size: deprecated-ompi-info-value
> >         Fort use mpi_f08: yes
> >  Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
> >                           limitations in the ifort compiler, does not
> support
> >                           the following: array subsections, direct
> passthru
> >                           (where possible) to underlying Open MPI's C
> >                           functionality
> >   Fort mpi_f08 subarrays: no
> >            Java bindings: no
> >   Wrapper compiler rpath: unnecessary
> >               C compiler: icc
> >      C compiler absolute: /opt/intel/Compiler/11.1/080/bin/intel64/icc
> >   C compiler family name: INTEL
> >       C compiler version: 1110.20091130
> >             C++ compiler: icpc
> >    C++ compiler absolute: /opt/intel/Compiler/11.1/080/bin/intel64/icpc
> >            Fort compiler: ifort
> >        Fort compiler abs: /opt/intel/Compiler/11.1/080/bin/intel64/ifort
> >          Fort ignore TKR: yes (!DEC$ ATTRIBUTES NO_ARG_CHECK ::)
> >    Fort 08 assumed shape: no
> >       Fort optional args: yes
> >       Fort BIND(C) (all): yes
> >       Fort ISO_C_BINDING: yes
> >  Fort SUBROUTINE BIND(C): yes
> >        Fort TYPE,BIND(C): yes
> >  Fort T,BIND(C,name="a"): yes
> >             Fort PRIVATE: yes
> >           Fort PROTECTED: yes
> >            Fort ABSTRACT: yes
> >        Fort ASYNCHRONOUS: yes
> >           Fort PROCEDURE: yes
> >  Fort f08 using wrappers: yes
> >              C profiling: yes
> >            C++ profiling: yes
> >    Fort mpif.h profiling: yes
> >   Fort use mpi profiling: yes
> >    Fort use mpi_f08 prof: yes
> >           C++ exceptions: no
> >           Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support:
> yes,
> >                           OMPI progress: no, ORTE progress: yes, Event
> lib:
> >                           yes)
> >            Sparse Groups: no
> >   Internal debug support: no
> >   MPI interface warnings: yes
> >      MPI parameter check: runtime
> > Memory profiling support: no
> > Memory debugging support: no
> >          libltdl support: yes
> >    Heterogeneous support: no
> >  mpirun default --prefix: no
> >          MPI I/O support: yes
> >        MPI_WTIME support: gettimeofday
> >      Symbol vis. support: yes
> >    Host topology support: yes
> >           MPI extensions:
> >    FT Checkpoint support: no (checkpoint thread: no)
> >    C/R Enabled Debugging: no
> >      VampirTrace support: yes
> >   MPI_MAX_PROCESSOR_NAME: 256
> >     MPI_MAX_ERROR_STRING: 256
> >      MPI_MAX_OBJECT_NAME: 64
> >         MPI_MAX_INFO_KEY: 36
> >         MPI_MAX_INFO_VAL: 256
> >        MPI_MAX_PORT_NAME: 1024
> >   MPI_MAX_DATAREP_STRING: 128
> >
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> > Link to this post:
> http://www.open-mpi.org/community/lists/users/2014/07/24764.php
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2014/07/24765.php
>

Reply via email to