Did you configure —enable-debug? We aren’t seeing any of the debug output, so I 
suspect not.


> On Mar 28, 2015, at 12:56 PM, LOTFIFAR F. <foad.lotfi...@durham.ac.uk> wrote:
> 
> I have done it and it is the results:
> 
> ubuntu@fehg-node-0:~$ mpirun -host fehg-node-7 -mca oob_base_verbose 100 -mca 
> state_base_verbose 10 hostname
> [fehg-node-0:30034] mca: base: components_open: Looking for oob components
> [fehg-node-0:30034] mca: base: components_open: opening oob components
> [fehg-node-0:30034] mca: base: components_open: found loaded component tcp
> [fehg-node-0:30034] mca: base: components_open: component tcp register 
> function successful
> [fehg-node-0:30034] mca: base: components_open: component tcp open function 
> successful
> [fehg-node-7:31138] mca: base: components_open: Looking for oob components
> [fehg-node-7:31138] mca: base: components_open: opening oob components
> [fehg-node-7:31138] mca: base: components_open: found loaded component tcp
> [fehg-node-7:31138] mca: base: components_open: component tcp register 
> function successful
> [fehg-node-7:31138] mca: base: components_open: component tcp open function 
> successful
> 
> freeze ...
> 
> Regards
> 
> From: users [users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>] 
> on behalf of LOTFIFAR F. [foad.lotfi...@durham.ac.uk 
> <mailto:foad.lotfi...@durham.ac.uk>]
> Sent: 28 March 2015 18:49
> To: Open MPI Users
> Subject: Re: [OMPI users] Connection problem on Linux cluster
> 
> fehg_node_1 and fehg-node-7 are the same. it is just a typo. 
> 
> Correction: VM names are fehg-node-0 and fehg-node-7.
> 
> 
> Regards,
> 
> From: users [users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>] 
> on behalf of Ralph Castain [r...@open-mpi.org <mailto:r...@open-mpi.org>]
> Sent: 28 March 2015 18:23
> To: Open MPI Users
> Subject: Re: [OMPI users] Connection problem on Linux cluster
> 
> Just to be clear: do you have two physical nodes? Or just one physical node 
> and you are running two VMs on it?
> 
>> On Mar 28, 2015, at 10:51 AM, LOTFIFAR F. <foad.lotfi...@durham.ac.uk 
>> <mailto:foad.lotfi...@durham.ac.uk>> wrote:
>> 
>> I have a floating IP for accessing nodes from outside of the cluster and 
>> internal ip addresses. I tried to run the jobs with both of them (both ip 
>> addresses) but it makes no difference. 
>> I have just installed openmpi 1.6.5 to see how does this version works. In 
>> this case I get nothing and I have to press Crtl+c. not output or error is 
>> shown.
>> 
>> 
>> From: users [users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>] 
>> on behalf of Ralph Castain [r...@open-mpi.org <mailto:r...@open-mpi.org>]
>> Sent: 28 March 2015 17:03
>> To: Open MPI Users
>> Subject: Re: [OMPI users] Connection problem on Linux cluster
>> 
>> You mentioned running this in a VM - is that IP address correct for getting 
>> across the VMs?
>> 
>> 
>>> On Mar 28, 2015, at 8:38 AM, LOTFIFAR F. <foad.lotfi...@durham.ac.uk 
>>> <mailto:foad.lotfi...@durham.ac.uk>> wrote:
>>> 
>>> Hi , 
>>> 
>>> I am wondering how can I solve this problem. 
>>> System Spec:
>>> 1- Linux cluster with two nodes (master and slave) with Ubuntu 12.04 LTS 
>>> 32bit.
>>> 2- openmpi 1.8.4
>>> 
>>> I do a simple test running on fehg_node_0:
>>> > mpirun -host fehg_node_0,fehg_node_1 hello_world -mca oob_base_verbose 20
>>> 
>>> and I get the following error:
>>> 
>>> A process or daemon was unable to complete a TCP connection
>>> to another process:
>>>   Local host:    fehg-node-0
>>>   Remote host:   10.104.5.40
>>> This is usually caused by a firewall on the remote host. Please
>>> check that any firewall (e.g., iptables) has been disabled and
>>> try again.
>>> ------------------------------------------------------------
>>> --------------------------------------------------------------------------
>>> ORTE was unable to reliably start one or more daemons.
>>> This usually is caused by:
>>> 
>>> * not finding the required libraries and/or binaries on
>>>   one or more nodes. Please check your PATH and LD_LIBRARY_PATH
>>>   settings, or configure OMPI with --enable-orterun-prefix-by-default
>>> 
>>> * lack of authority to execute on one or more specified nodes.
>>>   Please verify your allocation and authorities.
>>> 
>>> * the inability to write startup files into /tmp 
>>> (--tmpdir/orte_tmpdir_base).
>>>   Please check with your sys admin to determine the correct location to use.
>>> 
>>> *  compilation of the orted with dynamic libraries when static are required
>>>   (e.g., on Cray). Please check your configure cmd line and consider using
>>>   one of the contrib/platform definitions for your system type.
>>> 
>>> * an inability to create a connection back to mpirun due to a
>>>   lack of common network interfaces and/or no route found between
>>>   them. Please check network connectivity (including firewalls
>>>   and network routing requirements).
>>> 
>>> Verbose:
>>> 1- I have full access to the VMs on the cluster and setup everything myself
>>> 2- Firewall and iptables are all disabled on the nodes
>>> 3- nodes can ssh to each other with  no problem
>>> 4- non-interactive bash calls works fine i.e. when I run ssh othernode env 
>>> | grep PATH from both nodes, both PATH and LD_LIBRARY_PATH are set correctly
>>> 5- I have checked the posts, a similar problem reported for Solaris but I 
>>> could not find a clue about mine. 
>>> 6- run with --enable-orterun-prefix-by-default does not make any changes.
>>> 7-  I see orte is running on the other node when I check processes, but 
>>> nothing happens after that and the error happens.
>>> 
>>> Regards,
>>> Karos
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2015/03/26555.php 
>>> <http://www.open-mpi.org/community/lists/users/2015/03/26555.php>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2015/03/26557.php 
>> <http://www.open-mpi.org/community/lists/users/2015/03/26557.php>
> _______________________________________________
> users mailing list
> us...@open-mpi.org <mailto:us...@open-mpi.org>
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2015/03/26562.php 
> <http://www.open-mpi.org/community/lists/users/2015/03/26562.php>

Reply via email to