Hi,

I've installed Open MPI v1.10.2. Every VM on the cloud has two IPs (internal 
IP, public IP).
When I run: mpirun --host <internal IP> hostname, the output is the hostname of 
the VM.
But when I run: mpirun --host <public IP> hostname, the output is

bash: orted: command not found
--------------------------------------------------------------------------
ORTE was unable to reliably start one or more daemons.
This usually is caused by:

* not finding the required libraries and/or binaries on
  one or more nodes. Please check your PATH and LD_LIBRARY_PATH
  settings, or configure OMPI with --enable-orterun-prefix-by-default

* lack of authority to execute on one or more specified nodes.
  Please verify your allocation and authorities.

* the inability to write startup files into /tmp (--tmpdir/orte_tmpdir_base).
  Please check with your sys admin to determine the correct location to use.

*  compilation of the orted with dynamic libraries when static are required
  (e.g., on Cray). Please check your configure cmd line and consider using
  one of the contrib/platform definitions for your system type.

* an inability to create a connection back to mpirun due to a
  lack of common network interfaces and/or no route found between
  them. Please check network connectivity (including firewalls
  and network routing requirements).

Both IPs are the IP of the VM where MPI is running. Did I do something wrong in 
the configuration?

Thanks for any help.

Ping

-----Ursprüngliche Nachricht-----
Von: users [mailto:users-boun...@open-mpi.org] Im Auftrag von Jeff Squyres 
(jsquyres)
Gesendet: Mittwoch, 1. Juni 2016 15:02
An: Open MPI User's List
Betreff: Re: [OMPI users] Firewall settings for MPI communication

In addition, you might want to consider upgrading to Open MPI v1.10.x (v1.6.x 
is fairly ancient).

> On Jun 1, 2016, at 7:46 AM, Gilles Gouaillardet 
> <gilles.gouaillar...@gmail.com> wrote:
>
> which network are your VMs using for communications ?
> if this is tcp, then you also have to specify a restricted set of
> allowed ports for the tcp btl
>
> that would be something like
> mpirun --mca btl_tcp_dynamic_ports 49990-50010 ...
>
> please double check the Open MPI 1.6.5 parameter and syntax with
> ompi_info --all (or check the archives, I think I posted the correct
> command line a few weeks ago)
>
> Cheers,
>
> Gilles
>
> On Wednesday, June 1, 2016, Ping Wang <ping.w...@asc-s.de> wrote:
> I'm using Open MPI 1.6.5 to run OpenFOAM in parallel on several VMs on
> a cloud. mpirun hangs without any error messages. I think this is a
> firewall issue. Because when I open all the TCP ports(1-65535) in the
> security group of VMs, mpirun works well. However I was suggested to
> open as less ports as possible. So I have to limit MPI to run on a
> range of ports. I opened the port range 49990-50010 for MPI
> communication. And use command
>
>
>
> mpirun --mca oob_tcp_dynamic_ports 49990-50010 -np 4 --hostfile machines 
> simpleFoam –parallel.
>
>
>
> But it still hangs. How can I specify a port range that OpenMPI will use? I 
> appreciate any help you can provide.
>
>
>
> Best,
>
> Ping Wang
>
>
>
> <image001.png>
>
> ------------------------------------------------------
>
> Ping Wang
>
> Automotive Simulation Center Stuttgart e.V.
>
> Nobelstraße 15
>
> D-70569 Stuttgart
>
> Telefon: +49 711 699659-14
>
> Fax: +49 711 699659-29
>
> E-Mail: ping.w...@asc-s.de
>
> Web: http://www.asc-s.de
>
> Social Media: <image002.gif>/asc.stuttgart
>
> ------------------------------------------------------
>
>
>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2016/06/29340.php


--
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/

_______________________________________________
users mailing list
us...@open-mpi.org
Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
http://www.open-mpi.org/community/lists/users/2016/06/29342.php



Reply via email to