Thank you for the fast answer While that resolve my problem with cross ssh authentication a command as
/opt/openmpi/bin/mpirun --mca mtl mx --mca pml cm --mca plm_rsh_no_tree_spawn 1 -hostfile hostfile ompi_info just hung with no output and although there is a ssh connexion no orte program is initiated in the destination nodes and while /opt/openmpi/bin/mpirun -host host18 ompi_info works /opt/openmpi/bin/mpirun --mca plm_rsh_no_tree_spawn 1 -host host18 ompi_info hangs, is there some condition in the use of this parameter. Yours truly Ricardo On Mon, Jul 14, 2014 at 6:35 PM, Ralph Castain <r...@open-mpi.org> wrote: > During the 1.7 series and for all follow-on series, OMPI changed to a mode > where it launches a daemon on all allocated nodes at the startup of mpirun. > This allows us to determine the hardware topology of the nodes and take > that into account when mapping. You can override that behavior by either > adding --novm to your cmd line (which will impact your mapping/binding > options), or by specifying the hosts to use by editing your hostfile, or > adding --host host1,host2 to your cmd line > > The rsh launcher defaults to a tree-based pattern, thus requiring that we > be able to ssh from one compute node to another. You can change that to a > less scalable direct mode by adding > > --mca plm_rsh_no_tree_spawn 1 > > to the cmd line > > > On Jul 14, 2014, at 9:21 AM, Ricardo Fernández-Perea < > rfernandezpe...@gmail.com> wrote: > > > I'm trying to update to openMPI 1.8.1 thru ssh and Myrinet > > > > running a command as > > > > /opt/openmpi/bin/mpirun --verbose --mca mtl mx --mca pml cm -hostfile > hostfile -np 16 > > > > when the hostfile contain only two nodes as > > > > host1 slots=8 max-slots=8 > > host2 slots=8 max-slots=8 > > > > it runs perfectly but when the hostfile has a third node as > > > > > > host1 slots=8 max-slots=8 > > host2 slots=8 max-slots=8 > > host3 slots=8 max-slots=8 > > > > it try to establish an ssh connection between the running hosts1 and > host3 that should not run any process that fails hanging the process > without signaling. > > > > > > my ompi_info is as follow > > > > Package: Open MPI XXX Distribution > > Open MPI: 1.8.1 > > Open MPI repo revision: r31483 > > Open MPI release date: Apr 22, 2014 > > Open RTE: 1.8.1 > > Open RTE repo revision: r31483 > > Open RTE release date: Apr 22, 2014 > > OPAL: 1.8.1 > > OPAL repo revision: r31483 > > OPAL release date: Apr 22, 2014 > > MPI API: 3.0 > > Ident string: 1.8.1 > > Prefix: /opt/openmpi > > Configured architecture: x86_64-apple-darwin9.8.0 > > Configure host: XXXX > > Configured by: XXXX > > Configured on: Thu Jun 12 10:37:33 CEST 2014 > > Configure host: XXXX > > Built by: XXXX > > Built on: Thu Jun 12 11:13:16 CEST 2014 > > Built host: XXXX > > C bindings: yes > > C++ bindings: yes > > Fort mpif.h: yes (single underscore) > > Fort use mpi: yes (full: ignore TKR) > > Fort use mpi size: deprecated-ompi-info-value > > Fort use mpi_f08: yes > > Fort mpi_f08 compliance: The mpi_f08 module is available, but due to > > limitations in the ifort compiler, does not > support > > the following: array subsections, direct > passthru > > (where possible) to underlying Open MPI's C > > functionality > > Fort mpi_f08 subarrays: no > > Java bindings: no > > Wrapper compiler rpath: unnecessary > > C compiler: icc > > C compiler absolute: /opt/intel/Compiler/11.1/080/bin/intel64/icc > > C compiler family name: INTEL > > C compiler version: 1110.20091130 > > C++ compiler: icpc > > C++ compiler absolute: /opt/intel/Compiler/11.1/080/bin/intel64/icpc > > Fort compiler: ifort > > Fort compiler abs: /opt/intel/Compiler/11.1/080/bin/intel64/ifort > > Fort ignore TKR: yes (!DEC$ ATTRIBUTES NO_ARG_CHECK ::) > > Fort 08 assumed shape: no > > Fort optional args: yes > > Fort BIND(C) (all): yes > > Fort ISO_C_BINDING: yes > > Fort SUBROUTINE BIND(C): yes > > Fort TYPE,BIND(C): yes > > Fort T,BIND(C,name="a"): yes > > Fort PRIVATE: yes > > Fort PROTECTED: yes > > Fort ABSTRACT: yes > > Fort ASYNCHRONOUS: yes > > Fort PROCEDURE: yes > > Fort f08 using wrappers: yes > > C profiling: yes > > C++ profiling: yes > > Fort mpif.h profiling: yes > > Fort use mpi profiling: yes > > Fort use mpi_f08 prof: yes > > C++ exceptions: no > > Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: > yes, > > OMPI progress: no, ORTE progress: yes, Event > lib: > > yes) > > Sparse Groups: no > > Internal debug support: no > > MPI interface warnings: yes > > MPI parameter check: runtime > > Memory profiling support: no > > Memory debugging support: no > > libltdl support: yes > > Heterogeneous support: no > > mpirun default --prefix: no > > MPI I/O support: yes > > MPI_WTIME support: gettimeofday > > Symbol vis. support: yes > > Host topology support: yes > > MPI extensions: > > FT Checkpoint support: no (checkpoint thread: no) > > C/R Enabled Debugging: no > > VampirTrace support: yes > > MPI_MAX_PROCESSOR_NAME: 256 > > MPI_MAX_ERROR_STRING: 256 > > MPI_MAX_OBJECT_NAME: 64 > > MPI_MAX_INFO_KEY: 36 > > MPI_MAX_INFO_VAL: 256 > > MPI_MAX_PORT_NAME: 1024 > > MPI_MAX_DATAREP_STRING: 128 > > > > > > _______________________________________________ > > users mailing list > > us...@open-mpi.org > > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > > Link to this post: > http://www.open-mpi.org/community/lists/users/2014/07/24764.php > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2014/07/24765.php >