are you saying both IP are the ones of the VM on which mpirun is running ?
orted is only launched on all the machines *except* the one running mpirun.

can you double/triple check the IPs are ok and unique ?
for example, mpirun --host <internal IP> /sbin/ifconfig -a
can you also make sure Open MPI is installed on all your VMs in the same
directory ?
also make sure Open MPI has all the dependencies on all the VMs
ssh xxx ldd `which orted`
should show no missing dependency

generally speaking, I recommend you configure Open MPI with
--enable-mpirun-prefix-by-default

you can also try to replace
mpirun
with
`which mpirun`
or
mpirun --prefix <path to Open MPI>

Cheers,

Gilles

On Thursday, June 2, 2016, Ping Wang <ping.w...@asc-s.de> wrote:

> Hi,
>
> I've installed Open MPI v1.10.2. Every VM on the cloud has two IPs
> (internal IP, public IP).
> When I run: mpirun --host <internal IP> hostname, the output is the
> hostname of the VM.
> But when I run: mpirun --host <public IP> hostname, the output is
>
> bash: orted: command not found
> --------------------------------------------------------------------------
> ORTE was unable to reliably start one or more daemons.
> This usually is caused by:
>
> * not finding the required libraries and/or binaries on
>   one or more nodes. Please check your PATH and LD_LIBRARY_PATH
>   settings, or configure OMPI with --enable-orterun-prefix-by-default
>
> * lack of authority to execute on one or more specified nodes.
>   Please verify your allocation and authorities.
>
> * the inability to write startup files into /tmp
> (--tmpdir/orte_tmpdir_base).
>   Please check with your sys admin to determine the correct location to
> use.
>
> *  compilation of the orted with dynamic libraries when static are required
>   (e.g., on Cray). Please check your configure cmd line and consider using
>   one of the contrib/platform definitions for your system type.
>
> * an inability to create a connection back to mpirun due to a
>   lack of common network interfaces and/or no route found between
>   them. Please check network connectivity (including firewalls
>   and network routing requirements).
>
> Both IPs are the IP of the VM where MPI is running. Did I do something
> wrong in the configuration?
>
> Thanks for any help.
>
> Ping
>
> -----Ursprüngliche Nachricht-----
> Von: users [mailto:users-boun...@open-mpi.org <javascript:;>] Im Auftrag
> von Jeff Squyres (jsquyres)
> Gesendet: Mittwoch, 1. Juni 2016 15:02
> An: Open MPI User's List
> Betreff: Re: [OMPI users] Firewall settings for MPI communication
>
> In addition, you might want to consider upgrading to Open MPI v1.10.x
> (v1.6.x is fairly ancient).
>
> > On Jun 1, 2016, at 7:46 AM, Gilles Gouaillardet <
> gilles.gouaillar...@gmail.com <javascript:;>> wrote:
> >
> > which network are your VMs using for communications ?
> > if this is tcp, then you also have to specify a restricted set of
> > allowed ports for the tcp btl
> >
> > that would be something like
> > mpirun --mca btl_tcp_dynamic_ports 49990-50010 ...
> >
> > please double check the Open MPI 1.6.5 parameter and syntax with
> > ompi_info --all (or check the archives, I think I posted the correct
> > command line a few weeks ago)
> >
> > Cheers,
> >
> > Gilles
> >
> > On Wednesday, June 1, 2016, Ping Wang <ping.w...@asc-s.de <javascript:;>>
> wrote:
> > I'm using Open MPI 1.6.5 to run OpenFOAM in parallel on several VMs on
> > a cloud. mpirun hangs without any error messages. I think this is a
> > firewall issue. Because when I open all the TCP ports(1-65535) in the
> > security group of VMs, mpirun works well. However I was suggested to
> > open as less ports as possible. So I have to limit MPI to run on a
> > range of ports. I opened the port range 49990-50010 for MPI
> > communication. And use command
> >
> >
> >
> > mpirun --mca oob_tcp_dynamic_ports 49990-50010 -np 4 --hostfile machines
> simpleFoam –parallel.
> >
> >
> >
> > But it still hangs. How can I specify a port range that OpenMPI will
> use? I appreciate any help you can provide.
> >
> >
> >
> > Best,
> >
> > Ping Wang
> >
> >
> >
> > <image001.png>
> >
> > ------------------------------------------------------
> >
> > Ping Wang
> >
> > Automotive Simulation Center Stuttgart e.V.
> >
> > Nobelstraße 15
> >
> > D-70569 Stuttgart
> >
> > Telefon: +49 711 699659-14
> >
> > Fax: +49 711 699659-29
> >
> > E-Mail: ping.w...@asc-s.de <javascript:;>
> >
> > Web: http://www.asc-s.de
> >
> > Social Media: <image002.gif>/asc.stuttgart
> >
> > ------------------------------------------------------
> >
> >
> >
> >
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org <javascript:;>
> > Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
> > Link to this post:
> > http://www.open-mpi.org/community/lists/users/2016/06/29340.php
>
>
> --
> Jeff Squyres
> jsquy...@cisco.com <javascript:;>
> For corporate legal information go to:
> http://www.cisco.com/web/about/doing_business/legal/cri/
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org <javascript:;>
> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2016/06/29342.php
>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org <javascript:;>
> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2016/06/29349.php

Reply via email to