In OMPI 1.9a1r32604 I get much better results:
$ time mpirun --mca oob_tcp_if_include ib0 -np 1 ./hello_c
Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
semenov@compiler-2 Distribution, ident: 1.9a1r32604, repo rev: r32604, Aug 26, 
2014 (nightly snapshot tarball), 146)
real 0m4.166s
user 0m0.034s
sys 0m0.079s


Thu, 28 Aug 2014 13:10:02 +0400 от Timur Ismagilov <tismagi...@mail.ru>:
>I enclosure 2 files with output of two foloowing commands (OMPI 1.9a1r32570)
>$time mpirun --leave-session-attached -mca oob_base_verbose 100 -np 1 
>./hello_c >& out1.txt 
>(Hello, world, I am ....)
>real 1m3.952s
>user 0m0.035s
>sys 0m0.107s
>$time mpirun --leave-session-attached -mca oob_base_verbose 100 --mca 
>oob_tcp_if_include ib0 -np 1 ./hello_c >& out2.txt 
>(no Hello, word, I am ....)
>real 0m9.337s
>user 0m0.059s
>sys 0m0.098s
>Wed, 27 Aug 2014 06:31:02 -0700 от Ralph Castain <r...@open-mpi.org>:
>>How bizarre. Please add "--leave-session-attached -mca oob_base_verbose 100" 
>>to your cmd line
>>
>>On Aug 27, 2014, at 4:31 AM, Timur Ismagilov < tismagi...@mail.ru > wrote:
>>>When i try to specify oob with --mca oob_tcp_if_include <one of interface 
>>>from ifconfig>, i alwase get error:
>>>$ mpirun  --mca oob_tcp_if_include ib0 -np 1 ./hello_c
>>>--------------------------------------------------------------------------
>>>An ORTE daemon has unexpectedly failed after launch and before
>>>communicating back to mpirun. This could be caused by a number
>>>of factors, including an inability to create a connection back
>>>to mpirun due to a lack of common network interfaces and/or no
>>>route found between them. Please check network connectivity
>>>(including firewalls and network routing requirements).
>>>-------------------------------------------------------------------------
>>>
>>>Earlier, in ompi 1.8.1, I can not run mpi jobs without " --mca 
>>>oob_tcp_if_include ib0 "... but now(ompi 1.9.a1) with this flag i get above 
>>>error.
>>>
>>>Here is an output of ifconfig
>>>$ ifconfig
>>>eth1 Link encap:Ethernet HWaddr 00:15:17:EE:89:E1  
>>>inet addr:10.0.251.53 Bcast:10.0.251.255 Mask:255.255.255.0
>>>UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
>>>RX packets:215087433 errors:0 dropped:0 overruns:0 frame:0
>>>TX packets:2648 errors:0 dropped:0 overruns:0 carrier:0
>>>collisions:0 txqueuelen:1000  
>>>RX bytes:26925754883 (25.0 GiB) TX bytes:137971 (134.7 KiB)
>>>Memory:b2c00000-b2c20000
>>>eth2 Link encap:Ethernet HWaddr 00:02:C9:04:73:F8  
>>>inet addr:10.0.0.4 Bcast:10.0.0.255 Mask:255.255.255.0
>>>UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
>>>RX packets:4892833125 errors:0 dropped:0 overruns:0 frame:0
>>>TX packets:8708606918 errors:0 dropped:0 overruns:0 carrier:0
>>>collisions:0 txqueuelen:1000  
>>>RX bytes:1823986502132 (1.6 TiB) TX bytes:11957754120037 (10.8 TiB)
>>>eth2.911 Link encap:Ethernet HWaddr 00:02:C9:04:73:F8  
>>>inet addr:93.180.7.38 Bcast:93.180.7.63 Mask:255.255.255.224
>>>UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
>>>RX packets:3746454225 errors:0 dropped:0 overruns:0 frame:0
>>>TX packets:1131917608 errors:0 dropped:3 overruns:0 carrier:0
>>>collisions:0 txqueuelen:0  
>>>RX bytes:285174723322 (265.5 GiB) TX bytes:11523163526058 (10.4 TiB)
>>>eth3 Link encap:Ethernet HWaddr 00:02:C9:04:73:F9  
>>>inet addr:10.2.251.14 Bcast:10.2.251.255 Mask:255.255.255.0
>>>UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
>>>RX packets:591156692 errors:0 dropped:56 overruns:56 frame:56
>>>TX packets:679729229 errors:0 dropped:0 overruns:0 carrier:0
>>>collisions:0 txqueuelen:1000  
>>>RX bytes:324195989293 (301.9 GiB) TX bytes:770299202886 (717.3 GiB)
>>>Ifconfig uses the ioctl access method to get the full address information, 
>>>which limits hardware addresses to 8 bytes.
>>>Because Infiniband address has 20 bytes, only the first 8 bytes are 
>>>displayed correctly.
>>>Ifconfig is obsolete! For replacement check ip.
>>>ib0 Link encap:InfiniBand HWaddr 
>>>80:00:00:48:FE:80:00:00:00:00:00:00:00:00:00:00:00:00:00:00  
>>>inet addr:10.128.0.4 Bcast:10.128.255.255 Mask:255.255.0.0
>>>UP BROADCAST RUNNING MULTICAST MTU:2044 Metric:1
>>>RX packets:10843859 errors:0 dropped:0 overruns:0 frame:0
>>>TX packets:8089839 errors:0 dropped:15 overruns:0 carrier:0
>>>collisions:0 txqueuelen:1024  
>>>RX bytes:939249464 (895.7 MiB) TX bytes:886054008 (845.0 MiB)
>>>lo Link encap:Local Loopback  
>>>inet addr:127.0.0.1 Mask:255.0.0.0
>>>UP LOOPBACK RUNNING MTU:16436 Metric:1
>>>RX packets:31235107 errors:0 dropped:0 overruns:0 frame:0
>>>TX packets:31235107 errors:0 dropped:0 overruns:0 carrier:0
>>>collisions:0 txqueuelen:0  
>>>RX bytes:132750916041 (123.6 GiB) TX bytes:132750916041 (123.6 GiB)
>>>
>>>
>>>
>>>Tue, 26 Aug 2014 09:48:35 -0700 от Ralph Castain < r...@open-mpi.org >:
>>>>I think something may be messed up with your installation. I went ahead and 
>>>>tested this on a Slurm 2.5.4 cluster, and got the following:
>>>>
>>>>$ time mpirun -np 1 --host bend001 ./hello
>>>>Hello, World, I am 0 of 1 [0 local peers]: get_cpubind: 0 bitmap 0,12
>>>>
>>>>real 0m0.086s
>>>>user 0m0.039s
>>>>sys 0m0.046s
>>>>
>>>>$ time mpirun -np 1 --host bend002 ./hello
>>>>Hello, World, I am 0 of 1 [0 local peers]: get_cpubind: 0 bitmap 0,12
>>>>
>>>>real 0m0.528s
>>>>user 0m0.021s
>>>>sys 0m0.023s
>>>>
>>>>Which is what I would have expected. With --host set to the local host, no 
>>>>daemons are being launched and so the time is quite short (just spent 
>>>>mapping and fork/exec). With --host set to a single remote host, you have 
>>>>the time it takes Slurm to launch our daemon on the remote host, so you get 
>>>>about half of a second.
>>>>
>>>>IIRC, you were having some problems with the OOB setup. If you specify the 
>>>>TCP interface to use, does your time come down?
>>>>
>>>>
>>>>On Aug 26, 2014, at 8:32 AM, Timur Ismagilov < tismagi...@mail.ru > wrote:
>>>>>I'm using slurm 2.5.6
>>>>>
>>>>>$salloc -N8 --exclusive -J ompi -p test
>>>>>$ srun hostname
>>>>>node1-128-21
>>>>>node1-128-24
>>>>>node1-128-22
>>>>>node1-128-26
>>>>>node1-128-27
>>>>>node1-128-20
>>>>>node1-128-25
>>>>>node1-128-23
>>>>>$ time mpirun -np 1 --host node1-128-21 ./hello_c
>>>>>Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
>>>>>semenov@compiler-2 Distribution, ident: 1.9a1r32570, repo rev: r32570, Aug 
>>>>>21, 2014 (nightly snapshot tarball), 146)
>>>>>real 1m3.932s
>>>>>user 0m0.035s
>>>>>sys 0m0.072s
>>>>>
>>>>>
>>>>>Tue, 26 Aug 2014 07:03:58 -0700 от Ralph Castain < r...@open-mpi.org >:
>>>>>>hmmm....what is your allocation like? do you have a large hostfile, for 
>>>>>>example?
>>>>>>
>>>>>>if you add a --host argument that contains just the local host, what is 
>>>>>>the time for that scenario?
>>>>>>
>>>>>>On Aug 26, 2014, at 6:27 AM, Timur Ismagilov < tismagi...@mail.ru > wrote:
>>>>>>>Hello!
>>>>>>>Here is my time results:
>>>>>>>$time mpirun -n 1 ./hello_c
>>>>>>>Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
>>>>>>>semenov@compiler-2 Distribution, ident: 1.9a1r32570, repo rev: r32570, 
>>>>>>>Aug 21, 2014 (nightly snapshot tarball), 146)
>>>>>>>real 1m3.985s
>>>>>>>user 0m0.031s
>>>>>>>sys 0m0.083s
>>>>>>>
>>>>>>>
>>>>>>>Fri, 22 Aug 2014 07:43:03 -0700 от Ralph Castain < r...@open-mpi.org >:
>>>>>>>>I'm also puzzled by your timing statement - I can't replicate it:
>>>>>>>>
>>>>>>>>07:41:43    $ time mpirun -n 1 ./hello_c
>>>>>>>>Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
>>>>>>>>rhc@bend001 Distribution, ident: 1.9a1r32577, repo rev: r32577, 
>>>>>>>>Unreleased developer copy, 125)
>>>>>>>>
>>>>>>>>real 0m0.547s
>>>>>>>>user 0m0.043s
>>>>>>>>sys 0m0.046s
>>>>>>>>
>>>>>>>>The entire thing ran in 0.5 seconds
>>>>>>>>
>>>>>>>>
>>>>>>>>On Aug 22, 2014, at 6:33 AM, Mike Dubman < mi...@dev.mellanox.co.il > 
>>>>>>>>wrote:
>>>>>>>>>Hi,
>>>>>>>>>The default delimiter is ";" . You can change delimiter with 
>>>>>>>>>mca_base_env_list_delimiter.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>On Fri, Aug 22, 2014 at 2:59 PM, Timur Ismagilov   < 
>>>>>>>>>tismagi...@mail.ru >   wrote:
>>>>>>>>>>Hello!
>>>>>>>>>>If i use latest night snapshot:
>>>>>>>>>>$ ompi_info -V
>>>>>>>>>>Open MPI v1.9a1r32570
>>>>>>>>>>*  In programm hello_c initialization takes ~1 min
>>>>>>>>>>In ompi 1.8.2rc4 and ealier it takes ~1 sec(or less)
>>>>>>>>>>*  if i use 
>>>>>>>>>>$mpirun  --mca mca_base_env_list 
>>>>>>>>>>'MXM_SHM_KCOPY_MODE=off,OMP_NUM_THREADS=8' --map-by slot:pe=8 -np 1 
>>>>>>>>>>./hello_c
>>>>>>>>>>i got error 
>>>>>>>>>>config_parser.c:657  MXM  ERROR Invalid value for SHM_KCOPY_MODE: 
>>>>>>>>>>'off,OMP_NUM_THREADS=8'. Expected: [off|knem|cma|autodetect]
>>>>>>>>>>but with -x all works fine (but with warn)
>>>>>>>>>>$mpirun  -x MXM_SHM_KCOPY_MODE=off -x OMP_NUM_THREADS=8 -np 1 
>>>>>>>>>>./hello_c
>>>>>>>>>>WARNING: The mechanism by which environment variables are explicitly
>>>>>>>>>>..............
>>>>>>>>>>..............
>>>>>>>>>>..............
>>>>>>>>>>Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
>>>>>>>>>>semenov@compiler-2 Distribution, ident: 1.9a1r32570, repo rev: 
>>>>>>>>>>r32570, Aug 21, 2014 (nightly snapshot tarball), 146)
>>>>>>>>>>Thu, 21 Aug 2014 06:26:13 -0700 от Ralph Castain < r...@open-mpi.org 
>>>>>>>>>>>:
>>>>>>>>>>>Not sure I understand. The problem has been fixed in both the trunk 
>>>>>>>>>>>and the 1.8 branch now, so you should be able to work with either of 
>>>>>>>>>>>those nightly builds.
>>>>>>>>>>>
>>>>>>>>>>>On Aug 21, 2014, at 12:02 AM, Timur Ismagilov < tismagi...@mail.ru > 
>>>>>>>>>>>wrote:
>>>>>>>>>>>>Have i I any opportunity to run mpi jobs?
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>Wed, 20 Aug 2014 10:48:38 -0700 от Ralph Castain < 
>>>>>>>>>>>>r...@open-mpi.org >:
>>>>>>>>>>>>>yes, i know - it is cmr'd
>>>>>>>>>>>>>
>>>>>>>>>>>>>On Aug 20, 2014, at 10:26 AM, Mike Dubman < 
>>>>>>>>>>>>>mi...@dev.mellanox.co.il > wrote:
>>>>>>>>>>>>>>btw, we get same error in v1.8 branch as well.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>On Wed, Aug 20, 2014 at 8:06 PM, Ralph Castain   < 
>>>>>>>>>>>>>>r...@open-mpi.org >   wrote:
>>>>>>>>>>>>>>>It was not yet fixed - but should be now.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>On Aug 20, 2014, at 6:39 AM, Timur Ismagilov < 
>>>>>>>>>>>>>>>tismagi...@mail.ru > wrote:
>>>>>>>>>>>>>>>>Hello!
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>As i can see, the bug is fixed, but in Open MPI v1.9a1r32516  i 
>>>>>>>>>>>>>>>>still have the problem
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>a)
>>>>>>>>>>>>>>>>$ mpirun  -np 1 ./hello_c
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>An ORTE daemon has unexpectedly failed after launch and before
>>>>>>>>>>>>>>>>communicating back to mpirun. This could be caused by a number
>>>>>>>>>>>>>>>>of factors, including an inability to create a connection back
>>>>>>>>>>>>>>>>to mpirun due to a lack of common network interfaces and/or no
>>>>>>>>>>>>>>>>route found between them. Please check network connectivity
>>>>>>>>>>>>>>>>(including firewalls and network routing requirements).
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>b)
>>>>>>>>>>>>>>>>$ mpirun --mca oob_tcp_if_include ib0 -np 1 ./hello_c
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>An ORTE daemon has unexpectedly failed after launch and before
>>>>>>>>>>>>>>>>communicating back to mpirun. This could be caused by a number
>>>>>>>>>>>>>>>>of factors, including an inability to create a connection back
>>>>>>>>>>>>>>>>to mpirun due to a lack of common network interfaces and/or no
>>>>>>>>>>>>>>>>route found between them. Please check network connectivity
>>>>>>>>>>>>>>>>(including firewalls and network routing requirements).
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>c)
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>$ mpirun --mca oob_tcp_if_include ib0 -debug-daemons --mca 
>>>>>>>>>>>>>>>>plm_base_verbose 5 -mca oob_base_verbose 10 -mca 
>>>>>>>>>>>>>>>>rml_base_verbose 10 -np 1 ./hello_c
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Querying component 
>>>>>>>>>>>>>>>>[isolated]
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Query of component 
>>>>>>>>>>>>>>>>[isolated] set priority to 0
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Querying component 
>>>>>>>>>>>>>>>>[rsh]
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Query of component 
>>>>>>>>>>>>>>>>[rsh] set priority to 10
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Querying component 
>>>>>>>>>>>>>>>>[slurm]
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Query of component 
>>>>>>>>>>>>>>>>[slurm] set priority to 75
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:base:select:( plm) Selected component 
>>>>>>>>>>>>>>>>[slurm]
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_register: registering 
>>>>>>>>>>>>>>>>oob components
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_register: found loaded 
>>>>>>>>>>>>>>>>component tcp
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_register: component 
>>>>>>>>>>>>>>>>tcp register function successful
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_open: opening oob 
>>>>>>>>>>>>>>>>components
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_open: found loaded 
>>>>>>>>>>>>>>>>component tcp
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_open: component tcp 
>>>>>>>>>>>>>>>>open function successful
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:oob:select: checking available component 
>>>>>>>>>>>>>>>>tcp
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:oob:select: Querying component [tcp]
>>>>>>>>>>>>>>>>[compiler-2:14673] oob:tcp: component_available called
>>>>>>>>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
>>>>>>>>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 2 KERNEL INDEX 3 FAMILY: V4
>>>>>>>>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 3 KERNEL INDEX 4 FAMILY: V4
>>>>>>>>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 4 KERNEL INDEX 5 FAMILY: V4
>>>>>>>>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 5 KERNEL INDEX 6 FAMILY: V4
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] oob:tcp:init adding 10.128.0.4 
>>>>>>>>>>>>>>>>to our list of V4 connections
>>>>>>>>>>>>>>>>[compiler-2:14673] WORKING INTERFACE 6 KERNEL INDEX 7 FAMILY: V4
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] TCP STARTUP
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] attempting to bind to IPv4 
>>>>>>>>>>>>>>>>port 0
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] assigned IPv4 port 59460
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:oob:select: Adding component to end
>>>>>>>>>>>>>>>>[compiler-2:14673] mca:oob:select: Found 1 active transports
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_register: registering 
>>>>>>>>>>>>>>>>rml components
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_register: found loaded 
>>>>>>>>>>>>>>>>component oob
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_register: component 
>>>>>>>>>>>>>>>>oob has no register or open function
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_open: opening rml 
>>>>>>>>>>>>>>>>components
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_open: found loaded 
>>>>>>>>>>>>>>>>component oob
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: components_open: component oob 
>>>>>>>>>>>>>>>>open function successful
>>>>>>>>>>>>>>>>[compiler-2:14673] orte_rml_base_select: initializing rml 
>>>>>>>>>>>>>>>>component oob
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>30 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>15 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>32 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>33 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>5 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>10 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>12 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>9 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>34 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>2 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>21 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>22 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>45 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>46 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>1 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting recv
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] posting persistent recv on tag 
>>>>>>>>>>>>>>>>27 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>Daemon was launched on node1-128-01 - beginning to initialize
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>WARNING: An invalid value was given for oob_tcp_if_include. This
>>>>>>>>>>>>>>>>value will be ignored.
>>>>>>>>>>>>>>>>Local host: node1-128-01
>>>>>>>>>>>>>>>>Value: "ib0"
>>>>>>>>>>>>>>>>Message: Invalid specification (missing "/")
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>None of the TCP networks specified to be included for 
>>>>>>>>>>>>>>>>out-of-band communications
>>>>>>>>>>>>>>>>could be found:
>>>>>>>>>>>>>>>>Value given:
>>>>>>>>>>>>>>>>Please revise the specification and try again.
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>No network interfaces were found for out-of-band 
>>>>>>>>>>>>>>>>communications. We require
>>>>>>>>>>>>>>>>at least one available network for out-of-band messaging.
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>It looks like orte_init failed for some reason; your parallel 
>>>>>>>>>>>>>>>>process is
>>>>>>>>>>>>>>>>likely to abort. There are many reasons that a parallel process 
>>>>>>>>>>>>>>>>can
>>>>>>>>>>>>>>>>fail during orte_init; some of which are due to configuration or
>>>>>>>>>>>>>>>>environment problems. This failure appears to be an internal 
>>>>>>>>>>>>>>>>failure;
>>>>>>>>>>>>>>>>here's some additional information (which may only be relevant 
>>>>>>>>>>>>>>>>to an
>>>>>>>>>>>>>>>>Open MPI developer):
>>>>>>>>>>>>>>>>orte_oob_base_select failed
>>>>>>>>>>>>>>>>--> Returned value (null) (-43) instead of ORTE_SUCCESS
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>srun: error: node1-128-01: task 0: Exited with exit code 213
>>>>>>>>>>>>>>>>srun: Terminating job step 661215.0
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>An ORTE daemon has unexpectedly failed after launch and before
>>>>>>>>>>>>>>>>communicating back to mpirun. This could be caused by a number
>>>>>>>>>>>>>>>>of factors, including an inability to create a connection back
>>>>>>>>>>>>>>>>to mpirun due to a lack of common network interfaces and/or no
>>>>>>>>>>>>>>>>route found between them. Please check network connectivity
>>>>>>>>>>>>>>>>(including firewalls and network routing requirements).
>>>>>>>>>>>>>>>>--------------------------------------------------------------------------
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] orted_cmd: received halt_vm cmd
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: close: component oob closed
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: close: unloading component oob
>>>>>>>>>>>>>>>>[compiler-2:14673] [[49095,0],0] TCP SHUTDOWN
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: close: component tcp closed
>>>>>>>>>>>>>>>>[compiler-2:14673] mca: base: close: unloading component tcp
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>Tue, 12 Aug 2014 18:33:24 +0000 от "Jeff Squyres (jsquyres)" < 
>>>>>>>>>>>>>>>>jsquy...@cisco.com >:
>>>>>>>>>>>>>>>>>I filed the following ticket:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>     https://svn.open-mpi.org/trac/ompi/ticket/4857
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>On Aug 12, 2014, at 12:39 PM, Jeff Squyres (jsquyres) < 
>>>>>>>>>>>>>>>>>jsquy...@cisco.com > wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> (please keep the users list CC'ed)
>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>> We talked about this on the weekly engineering call today. 
>>>>>>>>>>>>>>>>>> Ralph has an idea what is happening -- I need to do a little 
>>>>>>>>>>>>>>>>>> investigation today and file a bug. I'll make sure you're 
>>>>>>>>>>>>>>>>>> CC'ed on the bug ticket.
>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>> On Aug 12, 2014, at 12:27 PM, Timur Ismagilov < 
>>>>>>>>>>>>>>>>>> tismagi...@mail.ru > wrote:
>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>> I don't have this error in OMPI 1.9a1r32252 and OMPI 1.8.1 
>>>>>>>>>>>>>>>>>>> (with --mca oob_tcp_if_include ib0), but in all latest 
>>>>>>>>>>>>>>>>>>> night snapshots i got this error.
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>> Tue, 12 Aug 2014 13:08:12 +0000 от "Jeff Squyres 
>>>>>>>>>>>>>>>>>>> (jsquyres)" < jsquy...@cisco.com >:
>>>>>>>>>>>>>>>>>>> Are you running any kind of firewall on the node where 
>>>>>>>>>>>>>>>>>>> mpirun is invoked? Open MPI needs to be able to use 
>>>>>>>>>>>>>>>>>>> arbitrary TCP ports between the servers on which it runs.
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>> This second mail seems to imply a bug in OMPI's 
>>>>>>>>>>>>>>>>>>> oob_tcp_if_include param handling, however -- it's supposed 
>>>>>>>>>>>>>>>>>>> to be able to handle an interface name (not just a network 
>>>>>>>>>>>>>>>>>>> specification).
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>> Ralph -- can you have a look?
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>> On Aug 12, 2014, at 8:41 AM, Timur Ismagilov < 
>>>>>>>>>>>>>>>>>>> tismagi...@mail.ru > wrote:
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> When i add --mca oob_tcp_if_include ib0 (infiniband 
>>>>>>>>>>>>>>>>>>>> interface) to mpirun (as it was here:   
>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/07/24857.php
>>>>>>>>>>>>>>>>>>>>    ) i got this output:
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Querying 
>>>>>>>>>>>>>>>>>>>> component [isolated]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Query of 
>>>>>>>>>>>>>>>>>>>> component [isolated] set priority to 0
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Querying 
>>>>>>>>>>>>>>>>>>>> component [rsh]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Query of 
>>>>>>>>>>>>>>>>>>>> component [rsh] set priority to 10
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Querying 
>>>>>>>>>>>>>>>>>>>> component [slurm]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Query of 
>>>>>>>>>>>>>>>>>>>> component [slurm] set priority to 75
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:base:select:( plm) Selected 
>>>>>>>>>>>>>>>>>>>> component [slurm]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: 
>>>>>>>>>>>>>>>>>>>> registering oob components
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: found 
>>>>>>>>>>>>>>>>>>>> loaded component tcp
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: 
>>>>>>>>>>>>>>>>>>>> component tcp register function successful
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: opening oob 
>>>>>>>>>>>>>>>>>>>> components
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: found 
>>>>>>>>>>>>>>>>>>>> loaded component tcp
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: component 
>>>>>>>>>>>>>>>>>>>> tcp open function successful
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:oob:select: checking available 
>>>>>>>>>>>>>>>>>>>> component tcp
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:oob:select: Querying component [tcp]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] oob:tcp: component_available called
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 1 KERNEL INDEX 1 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 2 KERNEL INDEX 3 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 3 KERNEL INDEX 4 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 4 KERNEL INDEX 5 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 5 KERNEL INDEX 6 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] oob:tcp:init adding 
>>>>>>>>>>>>>>>>>>>> 10.128.0.4 to our list of V4 connections
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] WORKING INTERFACE 6 KERNEL INDEX 7 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] TCP STARTUP
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] attempting to bind to 
>>>>>>>>>>>>>>>>>>>> IPv4 port 0
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] assigned IPv4 port 53883
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:oob:select: Adding component to end
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca:oob:select: Found 1 active 
>>>>>>>>>>>>>>>>>>>> transports
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: 
>>>>>>>>>>>>>>>>>>>> registering rml components
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: found 
>>>>>>>>>>>>>>>>>>>> loaded component oob
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_register: 
>>>>>>>>>>>>>>>>>>>> component oob has no register or open function
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: opening rml 
>>>>>>>>>>>>>>>>>>>> components
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: found 
>>>>>>>>>>>>>>>>>>>> loaded component oob
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: components_open: component 
>>>>>>>>>>>>>>>>>>>> oob open function successful
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] orte_rml_base_select: initializing rml 
>>>>>>>>>>>>>>>>>>>> component oob
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 30 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 15 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 32 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 33 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 5 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 10 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 12 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 9 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 34 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 2 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 21 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 22 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 45 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 46 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 1 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 27 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-128-01 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-128-02 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> WARNING: An invalid value was given for 
>>>>>>>>>>>>>>>>>>>> oob_tcp_if_include. This
>>>>>>>>>>>>>>>>>>>> value will be ignored.
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> Local host: node1-128-01
>>>>>>>>>>>>>>>>>>>> Value: "ib0"
>>>>>>>>>>>>>>>>>>>> Message: Invalid specification (missing "/")
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> WARNING: An invalid value was given for 
>>>>>>>>>>>>>>>>>>>> oob_tcp_if_include. This
>>>>>>>>>>>>>>>>>>>> value will be ignored.
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> Local host: node1-128-02
>>>>>>>>>>>>>>>>>>>> Value: "ib0"
>>>>>>>>>>>>>>>>>>>> Message: Invalid specification (missing "/")
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> None of the TCP networks specified to be included for 
>>>>>>>>>>>>>>>>>>>> out-of-band communications
>>>>>>>>>>>>>>>>>>>> could be found:
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> Value given:
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> Please revise the specification and try again.
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> None of the TCP networks specified to be included for 
>>>>>>>>>>>>>>>>>>>> out-of-band communications
>>>>>>>>>>>>>>>>>>>> could be found:
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> Value given:
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> Please revise the specification and try again.
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> No network interfaces were found for out-of-band 
>>>>>>>>>>>>>>>>>>>> communications. We require
>>>>>>>>>>>>>>>>>>>> at least one available network for out-of-band messaging.
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> No network interfaces were found for out-of-band 
>>>>>>>>>>>>>>>>>>>> communications. We require
>>>>>>>>>>>>>>>>>>>> at least one available network for out-of-band messaging.
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> It looks like orte_init failed for some reason; your 
>>>>>>>>>>>>>>>>>>>> parallel process is
>>>>>>>>>>>>>>>>>>>> likely to abort. There are many reasons that a parallel 
>>>>>>>>>>>>>>>>>>>> process can
>>>>>>>>>>>>>>>>>>>> fail during orte_init; some of which are due to 
>>>>>>>>>>>>>>>>>>>> configuration or
>>>>>>>>>>>>>>>>>>>> environment problems. This failure appears to be an 
>>>>>>>>>>>>>>>>>>>> internal failure;
>>>>>>>>>>>>>>>>>>>> here's some additional information (which may only be 
>>>>>>>>>>>>>>>>>>>> relevant to an
>>>>>>>>>>>>>>>>>>>> Open MPI developer):
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> orte_oob_base_select failed
>>>>>>>>>>>>>>>>>>>> --> Returned value (null) (-43) instead of ORTE_SUCCESS
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> It looks like orte_init failed for some reason; your 
>>>>>>>>>>>>>>>>>>>> parallel process is
>>>>>>>>>>>>>>>>>>>> likely to abort. There are many reasons that a parallel 
>>>>>>>>>>>>>>>>>>>> process can
>>>>>>>>>>>>>>>>>>>> fail during orte_init; some of which are due to 
>>>>>>>>>>>>>>>>>>>> configuration or
>>>>>>>>>>>>>>>>>>>> environment problems. This failure appears to be an 
>>>>>>>>>>>>>>>>>>>> internal failure;
>>>>>>>>>>>>>>>>>>>> here's some additional information (which may only be 
>>>>>>>>>>>>>>>>>>>> relevant to an
>>>>>>>>>>>>>>>>>>>> Open MPI developer):
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> orte_oob_base_select failed
>>>>>>>>>>>>>>>>>>>> --> Returned value (null) (-43) instead of ORTE_SUCCESS
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> srun: error: node1-128-02: task 1: Exited with exit code 
>>>>>>>>>>>>>>>>>>>> 213
>>>>>>>>>>>>>>>>>>>> srun: Terminating job step 657300.0
>>>>>>>>>>>>>>>>>>>> srun: error: node1-128-01: task 0: Exited with exit code 
>>>>>>>>>>>>>>>>>>>> 213
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> An ORTE daemon has unexpectedly failed after launch and 
>>>>>>>>>>>>>>>>>>>> before
>>>>>>>>>>>>>>>>>>>> communicating back to mpirun. This could be caused by a 
>>>>>>>>>>>>>>>>>>>> number
>>>>>>>>>>>>>>>>>>>> of factors, including an inability to create a connection 
>>>>>>>>>>>>>>>>>>>> back
>>>>>>>>>>>>>>>>>>>> to mpirun due to a lack of common network interfaces 
>>>>>>>>>>>>>>>>>>>> and/or no
>>>>>>>>>>>>>>>>>>>> route found between them. Please check network connectivity
>>>>>>>>>>>>>>>>>>>> (including firewalls and network routing requirements).
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] orted_cmd: received 
>>>>>>>>>>>>>>>>>>>> halt_vm cmd
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: close: component oob closed
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: close: unloading component 
>>>>>>>>>>>>>>>>>>>> oob
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] [[42190,0],0] TCP SHUTDOWN
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: close: component tcp closed
>>>>>>>>>>>>>>>>>>>> [compiler-2:08792] mca: base: close: unloading component 
>>>>>>>>>>>>>>>>>>>> tcp
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> Tue, 12 Aug 2014 16:14:58 +0400 от Timur Ismagilov < 
>>>>>>>>>>>>>>>>>>>> tismagi...@mail.ru >:
>>>>>>>>>>>>>>>>>>>> Hello!
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> I have Open MPI v1.8.2rc4r32485
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> When i run hello_c, I got this error message
>>>>>>>>>>>>>>>>>>>> $mpirun -np 2 hello_c
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> An ORTE daemon has unexpectedly failed after launch and 
>>>>>>>>>>>>>>>>>>>> before
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> communicating back to mpirun. This could be caused by a 
>>>>>>>>>>>>>>>>>>>> number
>>>>>>>>>>>>>>>>>>>> of factors, including an inability to create a connection 
>>>>>>>>>>>>>>>>>>>> back
>>>>>>>>>>>>>>>>>>>> to mpirun due to a lack of common network interfaces 
>>>>>>>>>>>>>>>>>>>> and/or no
>>>>>>>>>>>>>>>>>>>> route found between them. Please check network connectivity
>>>>>>>>>>>>>>>>>>>> (including firewalls and network routing requirements).
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> When i run with --debug-daemons --mca plm_base_verbose 5 
>>>>>>>>>>>>>>>>>>>> -mca oob_base_verbose 10 -mca rml_base_verbose 10 i got 
>>>>>>>>>>>>>>>>>>>> this output:
>>>>>>>>>>>>>>>>>>>> $mpirun --debug-daemons --mca plm_base_verbose 5 -mca 
>>>>>>>>>>>>>>>>>>>> oob_base_verbose 10 -mca rml_base_verbose 10 -np 2 hello_c
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Querying 
>>>>>>>>>>>>>>>>>>>> component [isolated]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Query of 
>>>>>>>>>>>>>>>>>>>> component [isolated] set priority to 0
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Querying 
>>>>>>>>>>>>>>>>>>>> component [rsh]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Query of 
>>>>>>>>>>>>>>>>>>>> component [rsh] set priority to 10
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Querying 
>>>>>>>>>>>>>>>>>>>> component [slurm]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Query of 
>>>>>>>>>>>>>>>>>>>> component [slurm] set priority to 75
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:base:select:( plm) Selected 
>>>>>>>>>>>>>>>>>>>> component [slurm]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: 
>>>>>>>>>>>>>>>>>>>> registering oob components
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: found 
>>>>>>>>>>>>>>>>>>>> loaded component tcp
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: 
>>>>>>>>>>>>>>>>>>>> component tcp register function successful
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: opening oob 
>>>>>>>>>>>>>>>>>>>> components
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: found 
>>>>>>>>>>>>>>>>>>>> loaded component tcp
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: component 
>>>>>>>>>>>>>>>>>>>> tcp open function successful
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:oob:select: checking available 
>>>>>>>>>>>>>>>>>>>> component tcp
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:oob:select: Querying component [tcp]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] oob:tcp: component_available called
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 1 KERNEL INDEX 1 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 2 KERNEL INDEX 3 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 
>>>>>>>>>>>>>>>>>>>> 10.0.251.53 to our list of V4 connections
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 3 KERNEL INDEX 4 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 
>>>>>>>>>>>>>>>>>>>> 10.0.0.4 to our list of V4 connections
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 4 KERNEL INDEX 5 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 
>>>>>>>>>>>>>>>>>>>> 10.2.251.14 to our list of V4 connections
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 5 KERNEL INDEX 6 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 
>>>>>>>>>>>>>>>>>>>> 10.128.0.4 to our list of V4 connections
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] WORKING INTERFACE 6 KERNEL INDEX 7 
>>>>>>>>>>>>>>>>>>>> FAMILY: V4
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 
>>>>>>>>>>>>>>>>>>>> 93.180.7.38 to our list of V4 connections
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] TCP STARTUP
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] attempting to bind to 
>>>>>>>>>>>>>>>>>>>> IPv4 port 0
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] assigned IPv4 port 38420
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:oob:select: Adding component to end
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca:oob:select: Found 1 active 
>>>>>>>>>>>>>>>>>>>> transports
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: 
>>>>>>>>>>>>>>>>>>>> registering rml components
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: found 
>>>>>>>>>>>>>>>>>>>> loaded component oob
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_register: 
>>>>>>>>>>>>>>>>>>>> component oob has no register or open function
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: opening rml 
>>>>>>>>>>>>>>>>>>>> components
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: found 
>>>>>>>>>>>>>>>>>>>> loaded component oob
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: components_open: component 
>>>>>>>>>>>>>>>>>>>> oob open function successful
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] orte_rml_base_select: initializing rml 
>>>>>>>>>>>>>>>>>>>> component oob
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 30 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 15 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 32 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 33 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 5 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 10 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 12 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 9 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 34 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 2 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 21 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 22 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 45 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 46 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 1 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] posting persistent recv 
>>>>>>>>>>>>>>>>>>>> on tag 27 for peer [[WILDCARD],WILDCARD]
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-130-08 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-130-03 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-130-05 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-130-02 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-130-01 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-130-04 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-130-07 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> Daemon was launched on node1-130-06 - beginning to 
>>>>>>>>>>>>>>>>>>>> initialize
>>>>>>>>>>>>>>>>>>>> Daemon [[42202,0],3] checking in as pid 7178 on host 
>>>>>>>>>>>>>>>>>>>> node1-130-03
>>>>>>>>>>>>>>>>>>>> [node1-130-03:07178] [[42202,0],3] orted: up and running - 
>>>>>>>>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>>>>>>>>> Daemon [[42202,0],2] checking in as pid 13581 on host 
>>>>>>>>>>>>>>>>>>>> node1-130-02
>>>>>>>>>>>>>>>>>>>> [node1-130-02:13581] [[42202,0],2] orted: up and running - 
>>>>>>>>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>>>>>>>>> Daemon [[42202,0],1] checking in as pid 17220 on host 
>>>>>>>>>>>>>>>>>>>> node1-130-01
>>>>>>>>>>>>>>>>>>>> [node1-130-01:17220] [[42202,0],1] orted: up and running - 
>>>>>>>>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>>>>>>>>> Daemon [[42202,0],5] checking in as pid 6663 on host 
>>>>>>>>>>>>>>>>>>>> node1-130-05
>>>>>>>>>>>>>>>>>>>> [node1-130-05:06663] [[42202,0],5] orted: up and running - 
>>>>>>>>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>>>>>>>>> Daemon [[42202,0],8] checking in as pid 6683 on host 
>>>>>>>>>>>>>>>>>>>> node1-130-08
>>>>>>>>>>>>>>>>>>>> [node1-130-08:06683] [[42202,0],8] orted: up and running - 
>>>>>>>>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>>>>>>>>> Daemon [[42202,0],7] checking in as pid 7877 on host 
>>>>>>>>>>>>>>>>>>>> node1-130-07
>>>>>>>>>>>>>>>>>>>> [node1-130-07:07877] [[42202,0],7] orted: up and running - 
>>>>>>>>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>>>>>>>>> Daemon [[42202,0],4] checking in as pid 7735 on host 
>>>>>>>>>>>>>>>>>>>> node1-130-04
>>>>>>>>>>>>>>>>>>>> [node1-130-04:07735] [[42202,0],4] orted: up and running - 
>>>>>>>>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>>>>>>>>> Daemon [[42202,0],6] checking in as pid 8451 on host 
>>>>>>>>>>>>>>>>>>>> node1-130-06
>>>>>>>>>>>>>>>>>>>> [node1-130-06:08451] [[42202,0],6] orted: up and running - 
>>>>>>>>>>>>>>>>>>>> waiting for commands!
>>>>>>>>>>>>>>>>>>>> srun: error: node1-130-03: task 2: Exited with exit code 1
>>>>>>>>>>>>>>>>>>>> srun: Terminating job step 657040.1
>>>>>>>>>>>>>>>>>>>> srun: error: node1-130-02: task 1: Exited with exit code 1
>>>>>>>>>>>>>>>>>>>> slurmd[node1-130-04]: *** STEP 657040.1 KILLED AT 
>>>>>>>>>>>>>>>>>>>> 2014-08-12T12:59:07 WITH SIGNAL 9 ***
>>>>>>>>>>>>>>>>>>>> slurmd[node1-130-07]: *** STEP 657040.1 KILLED AT 
>>>>>>>>>>>>>>>>>>>> 2014-08-12T12:59:07 WITH SIGNAL 9 ***
>>>>>>>>>>>>>>>>>>>> slurmd[node1-130-06]: *** STEP 657040.1 KILLED AT 
>>>>>>>>>>>>>>>>>>>> 2014-08-12T12:59:07 WITH SIGNAL 9 ***
>>>>>>>>>>>>>>>>>>>> srun: Job step aborted: Waiting up to 2 seconds for job 
>>>>>>>>>>>>>>>>>>>> step to finish.
>>>>>>>>>>>>>>>>>>>> srun: error: node1-130-01: task 0: Exited with exit code 1
>>>>>>>>>>>>>>>>>>>> srun: error: node1-130-05: task 4: Exited with exit code 1
>>>>>>>>>>>>>>>>>>>> srun: error: node1-130-08: task 7: Exited with exit code 1
>>>>>>>>>>>>>>>>>>>> srun: error: node1-130-07: task 6: Exited with exit code 1
>>>>>>>>>>>>>>>>>>>> srun: error: node1-130-04: task 3: Killed
>>>>>>>>>>>>>>>>>>>> srun: error: node1-130-06: task 5: Killed
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> An ORTE daemon has unexpectedly failed after launch and 
>>>>>>>>>>>>>>>>>>>> before
>>>>>>>>>>>>>>>>>>>> communicating back to mpirun. This could be caused by a 
>>>>>>>>>>>>>>>>>>>> number
>>>>>>>>>>>>>>>>>>>> of factors, including an inability to create a connection 
>>>>>>>>>>>>>>>>>>>> back
>>>>>>>>>>>>>>>>>>>> to mpirun due to a lack of common network interfaces 
>>>>>>>>>>>>>>>>>>>> and/or no
>>>>>>>>>>>>>>>>>>>> route found between them. Please check network connectivity
>>>>>>>>>>>>>>>>>>>> (including firewalls and network routing requirements).
>>>>>>>>>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] orted_cmd: received 
>>>>>>>>>>>>>>>>>>>> halt_vm cmd
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: close: component oob closed
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: close: unloading component 
>>>>>>>>>>>>>>>>>>>> oob
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] [[42202,0],0] TCP SHUTDOWN
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: close: component tcp closed
>>>>>>>>>>>>>>>>>>>> [compiler-2:08780] mca: base: close: unloading component 
>>>>>>>>>>>>>>>>>>>> tcp
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>>>>> users mailing list
>>>>>>>>>>>>>>>>>>>>   us...@open-mpi.org
>>>>>>>>>>>>>>>>>>>> Subscription:   
>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>>>>>>>>> Link to this post:   
>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/08/24987.php
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>>>>> users mailing list
>>>>>>>>>>>>>>>>>>>>   us...@open-mpi.org
>>>>>>>>>>>>>>>>>>>> Subscription:   
>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>>>>>>>>> Link to this post:   
>>>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/08/24988.php
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>> --  
>>>>>>>>>>>>>>>>>>> Jeff Squyres
>>>>>>>>>>>>>>>>>>>   jsquy...@cisco.com
>>>>>>>>>>>>>>>>>>> For corporate legal information go to:   
>>>>>>>>>>>>>>>>>>> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>> --  
>>>>>>>>>>>>>>>>>> Jeff Squyres
>>>>>>>>>>>>>>>>>>   jsquy...@cisco.com
>>>>>>>>>>>>>>>>>> For corporate legal information go to:   
>>>>>>>>>>>>>>>>>> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>>> users mailing list
>>>>>>>>>>>>>>>>>>   us...@open-mpi.org
>>>>>>>>>>>>>>>>>> Subscription:   
>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>>>>>>> Link to this post:   
>>>>>>>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/08/25001.php
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>--  
>>>>>>>>>>>>>>>>>Jeff Squyres
>>>>>>>>>>>>>>>>>jsquy...@cisco.com
>>>>>>>>>>>>>>>>>For corporate legal information go to:   
>>>>>>>>>>>>>>>>>http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>_______________________________________________
>>>>>>>>>>>>>>>>users mailing list
>>>>>>>>>>>>>>>>us...@open-mpi.org
>>>>>>>>>>>>>>>>Subscription:   
>>>>>>>>>>>>>>>>http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>>>>>Link to this post:   
>>>>>>>>>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25086.php
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>_______________________________________________
>>>>>>>>>>>>>>>users mailing list
>>>>>>>>>>>>>>>us...@open-mpi.org
>>>>>>>>>>>>>>>Subscription:   
>>>>>>>>>>>>>>>http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>>>>Link to this post:   
>>>>>>>>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25093.php
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>_______________________________________________
>>>>>>>>>>>>>>users mailing list
>>>>>>>>>>>>>>us...@open-mpi.org
>>>>>>>>>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>>>Link to this post:   
>>>>>>>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25094.php
>>>>>>>>>>>>>_______________________________________________
>>>>>>>>>>>>>users mailing list
>>>>>>>>>>>>>us...@open-mpi.org
>>>>>>>>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>>Link to this post:   
>>>>>>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25095.php
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>_______________________________________________
>>>>>>>>>>>>users mailing list
>>>>>>>>>>>>us...@open-mpi.org
>>>>>>>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>Link to this post:   
>>>>>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25105.php
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>_______________________________________________
>>>>>>>>>>users mailing list
>>>>>>>>>>us...@open-mpi.org
>>>>>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>Link to this post:   
>>>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25127.php
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>--  
>>>>>>>>>
>>>>>>>>>Kind Regards,
>>>>>>>>>
>>>>>>>>>M. _______________________________________________
>>>>>>>>>users mailing list
>>>>>>>>>us...@open-mpi.org
>>>>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>Link to this post:   
>>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25128.php
>>>>>>>>_______________________________________________
>>>>>>>>users mailing list
>>>>>>>>us...@open-mpi.org
>>>>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>Link to this post:   
>>>>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25129.php
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>----------------------------------------------------------------------
>>>>>
>>>>>
>>>>>_______________________________________________
>>>>>users mailing list
>>>>>us...@open-mpi.org
>>>>>Subscription:   http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>Link to this post:   
>>>>>http://www.open-mpi.org/community/lists/users/2014/08/25154.php
>>>
>>>
>>>
>
>
>
>_______________________________________________
>users mailing list
>us...@open-mpi.org
>Subscription:  http://www.open-mpi.org/mailman/listinfo.cgi/users
>Link to this post:  
>http://www.open-mpi.org/community/lists/users/2014/08/25177.php




Reply via email to