I think something may be messed up with your installation. I went ahead and 
tested this on a Slurm 2.5.4 cluster, and got the following:

$ time mpirun -np 1 --host bend001 ./hello
Hello, World, I am 0 of 1 [0 local peers]: get_cpubind: 0 bitmap 0,12

real    0m0.086s
user    0m0.039s
sys     0m0.046s

$ time mpirun -np 1 --host bend002 ./hello
Hello, World, I am 0 of 1 [0 local peers]: get_cpubind: 0 bitmap 0,12

real    0m0.528s
user    0m0.021s
sys     0m0.023s

Which is what I would have expected. With --host set to the local host, no 
daemons are being launched and so the time is quite short (just spent mapping 
and fork/exec). With --host set to a single remote host, you have the time it 
takes Slurm to launch our daemon on the remote host, so you get about half of a 
second.

IIRC, you were having some problems with the OOB setup. If you specify the TCP 
interface to use, does your time come down?


On Aug 26, 2014, at 8:32 AM, Timur Ismagilov <tismagi...@mail.ru> wrote:

> I'm using slurm 2.5.6
> 
> $salloc -N8 --exclusive -J ompi -p test
> 
> $ srun hostname
> node1-128-21
> node1-128-24
> node1-128-22
> node1-128-26
> node1-128-27
> node1-128-20
> node1-128-25
> node1-128-23
> 
> $ time mpirun -np 1 --host node1-128-21 ./hello_c
> Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
> semenov@compiler-2 Distribution, ident: 1.9a1r32570, repo rev: r32570, Aug 
> 21, 2014 (nightly snapshot tarball), 146)
> 
> real 1m3.932s
> user 0m0.035s
> sys 0m0.072s
> 
> 
> 
> 
> Tue, 26 Aug 2014 07:03:58 -0700 от Ralph Castain <r...@open-mpi.org>:
> hmmm....what is your allocation like? do you have a large hostfile, for 
> example?
> 
> if you add a --host argument that contains just the local host, what is the 
> time for that scenario?
> 
> On Aug 26, 2014, at 6:27 AM, Timur Ismagilov <tismagi...@mail.ru> wrote:
> 
>> Hello!
>> Here is my time results:
>> 
>> $time mpirun -n 1 ./hello_c
>> Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
>> semenov@compiler-2 Distribution, ident: 1.9a1r32570, repo rev: r32570, Aug 
>> 21, 2014 (nightly snapshot tarball), 146)
>> 
>> real 1m3.985s
>> user 0m0.031s
>> sys 0m0.083s
>> 
>> 
>> 
>> 
>> Fri, 22 Aug 2014 07:43:03 -0700 от Ralph Castain <r...@open-mpi.org>:
>> I'm also puzzled by your timing statement - I can't replicate it:
>> 
>> 07:41:43  $ time mpirun -n 1 ./hello_c
>> Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI rhc@bend001 
>> Distribution, ident: 1.9a1r32577, repo rev: r32577, Unreleased developer 
>> copy, 125)
>> 
>> real 0m0.547s
>> user 0m0.043s
>> sys  0m0.046s
>> 
>> The entire thing ran in 0.5 seconds
>> 
>> 
>> On Aug 22, 2014, at 6:33 AM, Mike Dubman <mi...@dev.mellanox.co.il> wrote:
>> 
>>> Hi,
>>> The default delimiter is ";" . You can change delimiter with 
>>> mca_base_env_list_delimiter.
>>> 
>>> 
>>> 
>>> On Fri, Aug 22, 2014 at 2:59 PM, Timur Ismagilov <tismagi...@mail.ru> wrote:
>>> Hello!
>>> If i use latest night snapshot:
>>> $ ompi_info -V
>>> Open MPI v1.9a1r32570
>>> 
>>> In programm hello_c initialization takes ~1 min
>>> In ompi 1.8.2rc4 and ealier it takes ~1 sec(or less)
>>> if i use 
>>> $mpirun  --mca mca_base_env_list 'MXM_SHM_KCOPY_MODE=off,OMP_NUM_THREADS=8' 
>>> --map-by slot:pe=8 -np 1 ./hello_c
>>> i got error 
>>> config_parser.c:657  MXM  ERROR Invalid value for SHM_KCOPY_MODE: 
>>> 'off,OMP_NUM_THREADS=8'. Expected: [off|knem|cma|autodetect]
>>> but with -x all works fine (but with warn)
>>> $mpirun  -x MXM_SHM_KCOPY_MODE=off -x OMP_NUM_THREADS=8 -np 1 ./hello_c
>>> WARNING: The mechanism by which environment variables are explicitly
>>> ..............
>>> ..............
>>> ..............
>>> Hello, world, I am 0 of 1, (Open MPI v1.9a1, package: Open MPI 
>>> semenov@compiler-2 Distribution, ident: 1.9a1r32570, repo rev: r32570, Aug 
>>> 21, 2014 (nightly snapshot tarball), 146)
>>> 
>>> 
>>> Thu, 21 Aug 2014 06:26:13 -0700 от Ralph Castain <r...@open-mpi.org>:
>>> Not sure I understand. The problem has been fixed in both the trunk and the 
>>> 1.8 branch now, so you should be able to work with either of those nightly 
>>> builds.
>>> 
>>> On Aug 21, 2014, at 12:02 AM, Timur Ismagilov <tismagi...@mail.ru> wrote:
>>> 
>>>> Have i I any opportunity to run mpi jobs?
>>>> 
>>>> 
>>>> Wed, 20 Aug 2014 10:48:38 -0700 от Ralph Castain <r...@open-mpi.org>:
>>>> yes, i know - it is cmr'd
>>>> 
>>>> On Aug 20, 2014, at 10:26 AM, Mike Dubman <mi...@dev.mellanox.co.il> wrote:
>>>> 
>>>>> btw, we get same error in v1.8 branch as well.
>>>>> 
>>>>> 
>>>>> On Wed, Aug 20, 2014 at 8:06 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>>> It was not yet fixed - but should be now.
>>>>> 
>>>>> On Aug 20, 2014, at 6:39 AM, Timur Ismagilov <tismagi...@mail.ru> wrote:
>>>>> 
>>>>>> Hello!
>>>>>> 
>>>>>> As i can see, the bug is fixed, but in Open MPI v1.9a1r32516  i still 
>>>>>> have the problem
>>>>>> 
>>>>>> a)
>>>>>> $ mpirun  -np 1 ./hello_c
>>>>>> 
>>>>>> --------------------------------------------------------------------------
>>>>>> An ORTE daemon has unexpectedly failed after launch and before
>>>>>> communicating back to mpirun. This could be caused by a number
>>>>>> of factors, including an inability to create a connection back
>>>>>> to mpirun due to a lack of common network interfaces and/or no
>>>>>> route found between them. Please check network connectivity
>>>>>> (including firewalls and network routing requirements).
>>>>>> --------------------------------------------------------------------------
>>>>>> 
>>>>>> b)
>>>>>> $ mpirun --mca oob_tcp_if_include ib0 -np 1 ./hello_c
>>>>>> --------------------------------------------------------------------------
>>>>>> An ORTE daemon has unexpectedly failed after launch and before
>>>>>> communicating back to mpirun. This could be caused by a number
>>>>>> of factors, including an inability to create a connection back
>>>>>> to mpirun due to a lack of common network interfaces and/or no
>>>>>> route found between them. Please check network connectivity
>>>>>> (including firewalls and network routing requirements).
>>>>>> --------------------------------------------------------------------------
>>>>>> 
>>>>>> c)
>>>>>> 
>>>>>> $ mpirun --mca oob_tcp_if_include ib0 -debug-daemons --mca 
>>>>>> plm_base_verbose 5 -mca oob_base_verbose 10 -mca rml_base_verbose 10 -np 
>>>>>> 1 ./hello_c
>>>>>> 
>>>>>> [compiler-2:14673] mca:base:select:( plm) Querying component [isolated]
>>>>>> [compiler-2:14673] mca:base:select:( plm) Query of component [isolated] 
>>>>>> set priority to 0
>>>>>> [compiler-2:14673] mca:base:select:( plm) Querying component [rsh]
>>>>>> [compiler-2:14673] mca:base:select:( plm) Query of component [rsh] set 
>>>>>> priority to 10
>>>>>> [compiler-2:14673] mca:base:select:( plm) Querying component [slurm]
>>>>>> [compiler-2:14673] mca:base:select:( plm) Query of component [slurm] set 
>>>>>> priority to 75
>>>>>> [compiler-2:14673] mca:base:select:( plm) Selected component [slurm]
>>>>>> [compiler-2:14673] mca: base: components_register: registering oob 
>>>>>> components
>>>>>> [compiler-2:14673] mca: base: components_register: found loaded 
>>>>>> component tcp
>>>>>> [compiler-2:14673] mca: base: components_register: component tcp 
>>>>>> register function successful
>>>>>> [compiler-2:14673] mca: base: components_open: opening oob components
>>>>>> [compiler-2:14673] mca: base: components_open: found loaded component tcp
>>>>>> [compiler-2:14673] mca: base: components_open: component tcp open 
>>>>>> function successful
>>>>>> [compiler-2:14673] mca:oob:select: checking available component tcp
>>>>>> [compiler-2:14673] mca:oob:select: Querying component [tcp]
>>>>>> [compiler-2:14673] oob:tcp: component_available called
>>>>>> [compiler-2:14673] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
>>>>>> [compiler-2:14673] WORKING INTERFACE 2 KERNEL INDEX 3 FAMILY: V4
>>>>>> [compiler-2:14673] WORKING INTERFACE 3 KERNEL INDEX 4 FAMILY: V4
>>>>>> [compiler-2:14673] WORKING INTERFACE 4 KERNEL INDEX 5 FAMILY: V4
>>>>>> [compiler-2:14673] WORKING INTERFACE 5 KERNEL INDEX 6 FAMILY: V4
>>>>>> [compiler-2:14673] [[49095,0],0] oob:tcp:init adding 10.128.0.4 to our 
>>>>>> list of V4 connections
>>>>>> [compiler-2:14673] WORKING INTERFACE 6 KERNEL INDEX 7 FAMILY: V4
>>>>>> [compiler-2:14673] [[49095,0],0] TCP STARTUP
>>>>>> [compiler-2:14673] [[49095,0],0] attempting to bind to IPv4 port 0
>>>>>> [compiler-2:14673] [[49095,0],0] assigned IPv4 port 59460
>>>>>> [compiler-2:14673] mca:oob:select: Adding component to end
>>>>>> [compiler-2:14673] mca:oob:select: Found 1 active transports
>>>>>> [compiler-2:14673] mca: base: components_register: registering rml 
>>>>>> components
>>>>>> [compiler-2:14673] mca: base: components_register: found loaded 
>>>>>> component oob
>>>>>> [compiler-2:14673] mca: base: components_register: component oob has no 
>>>>>> register or open function
>>>>>> [compiler-2:14673] mca: base: components_open: opening rml components
>>>>>> [compiler-2:14673] mca: base: components_open: found loaded component oob
>>>>>> [compiler-2:14673] mca: base: components_open: component oob open 
>>>>>> function successful
>>>>>> [compiler-2:14673] orte_rml_base_select: initializing rml component oob
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 30 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 15 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 32 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 33 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 5 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 10 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 12 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 9 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 34 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 2 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 21 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 22 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 45 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 46 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 1 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> [compiler-2:14673] [[49095,0],0] posting recv
>>>>>> [compiler-2:14673] [[49095,0],0] posting persistent recv on tag 27 for 
>>>>>> peer [[WILDCARD],WILDCARD]
>>>>>> Daemon was launched on node1-128-01 - beginning to initialize
>>>>>> --------------------------------------------------------------------------
>>>>>> WARNING: An invalid value was given for oob_tcp_if_include. This
>>>>>> value will be ignored.
>>>>>> 
>>>>>> Local host: node1-128-01
>>>>>> Value: "ib0"
>>>>>> Message: Invalid specification (missing "/")
>>>>>> --------------------------------------------------------------------------
>>>>>> --------------------------------------------------------------------------
>>>>>> None of the TCP networks specified to be included for out-of-band 
>>>>>> communications
>>>>>> could be found:
>>>>>> 
>>>>>> Value given:
>>>>>> 
>>>>>> Please revise the specification and try again.
>>>>>> --------------------------------------------------------------------------
>>>>>> --------------------------------------------------------------------------
>>>>>> No network interfaces were found for out-of-band communications. We 
>>>>>> require
>>>>>> at least one available network for out-of-band messaging.
>>>>>> --------------------------------------------------------------------------
>>>>>> --------------------------------------------------------------------------
>>>>>> It looks like orte_init failed for some reason; your parallel process is
>>>>>> likely to abort. There are many reasons that a parallel process can
>>>>>> fail during orte_init; some of which are due to configuration or
>>>>>> environment problems. This failure appears to be an internal failure;
>>>>>> here's some additional information (which may only be relevant to an
>>>>>> Open MPI developer):
>>>>>> 
>>>>>> orte_oob_base_select failed
>>>>>> --> Returned value (null) (-43) instead of ORTE_SUCCESS
>>>>>> --------------------------------------------------------------------------
>>>>>> srun: error: node1-128-01: task 0: Exited with exit code 213
>>>>>> srun: Terminating job step 661215.0
>>>>>> --------------------------------------------------------------------------
>>>>>> An ORTE daemon has unexpectedly failed after launch and before
>>>>>> communicating back to mpirun. This could be caused by a number
>>>>>> of factors, including an inability to create a connection back
>>>>>> to mpirun due to a lack of common network interfaces and/or no
>>>>>> route found between them. Please check network connectivity
>>>>>> (including firewalls and network routing requirements).
>>>>>> --------------------------------------------------------------------------
>>>>>> [compiler-2:14673] [[49095,0],0] orted_cmd: received halt_vm cmd
>>>>>> [compiler-2:14673] mca: base: close: component oob closed
>>>>>> [compiler-2:14673] mca: base: close: unloading component oob
>>>>>> [compiler-2:14673] [[49095,0],0] TCP SHUTDOWN
>>>>>> [compiler-2:14673] mca: base: close: component tcp closed
>>>>>> [compiler-2:14673] mca: base: close: unloading component tcp
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> Tue, 12 Aug 2014 18:33:24 +0000 от "Jeff Squyres (jsquyres)" 
>>>>>> <jsquy...@cisco.com>:
>>>>>> I filed the following ticket:
>>>>>> 
>>>>>>     https://svn.open-mpi.org/trac/ompi/ticket/4857
>>>>>> 
>>>>>> 
>>>>>> On Aug 12, 2014, at 12:39 PM, Jeff Squyres (jsquyres) 
>>>>>> <jsquy...@cisco.com> wrote:
>>>>>> 
>>>>>> > (please keep the users list CC'ed)
>>>>>> > 
>>>>>> > We talked about this on the weekly engineering call today. Ralph has 
>>>>>> > an idea what is happening -- I need to do a little investigation today 
>>>>>> > and file a bug. I'll make sure you're CC'ed on the bug ticket.
>>>>>> > 
>>>>>> > 
>>>>>> > 
>>>>>> > On Aug 12, 2014, at 12:27 PM, Timur Ismagilov <tismagi...@mail.ru> 
>>>>>> > wrote:
>>>>>> > 
>>>>>> >> I don't have this error in OMPI 1.9a1r32252 and OMPI 1.8.1 (with 
>>>>>> >> --mca oob_tcp_if_include ib0), but in all latest night snapshots i 
>>>>>> >> got this error.
>>>>>> >> 
>>>>>> >> 
>>>>>> >> Tue, 12 Aug 2014 13:08:12 +0000 от "Jeff Squyres (jsquyres)" 
>>>>>> >> <jsquy...@cisco.com>:
>>>>>> >> Are you running any kind of firewall on the node where mpirun is 
>>>>>> >> invoked? Open MPI needs to be able to use arbitrary TCP ports between 
>>>>>> >> the servers on which it runs.
>>>>>> >> 
>>>>>> >> This second mail seems to imply a bug in OMPI's oob_tcp_if_include 
>>>>>> >> param handling, however -- it's supposed to be able to handle an 
>>>>>> >> interface name (not just a network specification).
>>>>>> >> 
>>>>>> >> Ralph -- can you have a look?
>>>>>> >> 
>>>>>> >> 
>>>>>> >> On Aug 12, 2014, at 8:41 AM, Timur Ismagilov <tismagi...@mail.ru> 
>>>>>> >> wrote:
>>>>>> >> 
>>>>>> >>> When i add --mca oob_tcp_if_include ib0 (infiniband interface) to 
>>>>>> >>> mpirun (as it was here: 
>>>>>> >>> http://www.open-mpi.org/community/lists/users/2014/07/24857.php ) i 
>>>>>> >>> got this output:
>>>>>> >>> 
>>>>>> >>> [compiler-2:08792] mca:base:select:( plm) Querying component 
>>>>>> >>> [isolated]
>>>>>> >>> [compiler-2:08792] mca:base:select:( plm) Query of component 
>>>>>> >>> [isolated] set priority to 0
>>>>>> >>> [compiler-2:08792] mca:base:select:( plm) Querying component [rsh]
>>>>>> >>> [compiler-2:08792] mca:base:select:( plm) Query of component [rsh] 
>>>>>> >>> set priority to 10
>>>>>> >>> [compiler-2:08792] mca:base:select:( plm) Querying component [slurm]
>>>>>> >>> [compiler-2:08792] mca:base:select:( plm) Query of component [slurm] 
>>>>>> >>> set priority to 75
>>>>>> >>> [compiler-2:08792] mca:base:select:( plm) Selected component [slurm]
>>>>>> >>> [compiler-2:08792] mca: base: components_register: registering oob 
>>>>>> >>> components
>>>>>> >>> [compiler-2:08792] mca: base: components_register: found loaded 
>>>>>> >>> component tcp
>>>>>> >>> [compiler-2:08792] mca: base: components_register: component tcp 
>>>>>> >>> register function successful
>>>>>> >>> [compiler-2:08792] mca: base: components_open: opening oob components
>>>>>> >>> [compiler-2:08792] mca: base: components_open: found loaded 
>>>>>> >>> component tcp
>>>>>> >>> [compiler-2:08792] mca: base: components_open: component tcp open 
>>>>>> >>> function successful
>>>>>> >>> [compiler-2:08792] mca:oob:select: checking available component tcp
>>>>>> >>> [compiler-2:08792] mca:oob:select: Querying component [tcp]
>>>>>> >>> [compiler-2:08792] oob:tcp: component_available called
>>>>>> >>> [compiler-2:08792] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
>>>>>> >>> [compiler-2:08792] WORKING INTERFACE 2 KERNEL INDEX 3 FAMILY: V4
>>>>>> >>> [compiler-2:08792] WORKING INTERFACE 3 KERNEL INDEX 4 FAMILY: V4
>>>>>> >>> [compiler-2:08792] WORKING INTERFACE 4 KERNEL INDEX 5 FAMILY: V4
>>>>>> >>> [compiler-2:08792] WORKING INTERFACE 5 KERNEL INDEX 6 FAMILY: V4
>>>>>> >>> [compiler-2:08792] [[42190,0],0] oob:tcp:init adding 10.128.0.4 to 
>>>>>> >>> our list of V4 connections
>>>>>> >>> [compiler-2:08792] WORKING INTERFACE 6 KERNEL INDEX 7 FAMILY: V4
>>>>>> >>> [compiler-2:08792] [[42190,0],0] TCP STARTUP
>>>>>> >>> [compiler-2:08792] [[42190,0],0] attempting to bind to IPv4 port 0
>>>>>> >>> [compiler-2:08792] [[42190,0],0] assigned IPv4 port 53883
>>>>>> >>> [compiler-2:08792] mca:oob:select: Adding component to end
>>>>>> >>> [compiler-2:08792] mca:oob:select: Found 1 active transports
>>>>>> >>> [compiler-2:08792] mca: base: components_register: registering rml 
>>>>>> >>> components
>>>>>> >>> [compiler-2:08792] mca: base: components_register: found loaded 
>>>>>> >>> component oob
>>>>>> >>> [compiler-2:08792] mca: base: components_register: component oob has 
>>>>>> >>> no register or open function
>>>>>> >>> [compiler-2:08792] mca: base: components_open: opening rml components
>>>>>> >>> [compiler-2:08792] mca: base: components_open: found loaded 
>>>>>> >>> component oob
>>>>>> >>> [compiler-2:08792] mca: base: components_open: component oob open 
>>>>>> >>> function successful
>>>>>> >>> [compiler-2:08792] orte_rml_base_select: initializing rml component 
>>>>>> >>> oob
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 30 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 15 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 32 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 33 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 5 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 10 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 12 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 9 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 34 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 2 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 21 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 22 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 45 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 46 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 1 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting recv
>>>>>> >>> [compiler-2:08792] [[42190,0],0] posting persistent recv on tag 27 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> Daemon was launched on node1-128-01 - beginning to initialize
>>>>>> >>> Daemon was launched on node1-128-02 - beginning to initialize
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> WARNING: An invalid value was given for oob_tcp_if_include. This
>>>>>> >>> value will be ignored.
>>>>>> >>> 
>>>>>> >>> Local host: node1-128-01
>>>>>> >>> Value: "ib0"
>>>>>> >>> Message: Invalid specification (missing "/")
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> WARNING: An invalid value was given for oob_tcp_if_include. This
>>>>>> >>> value will be ignored.
>>>>>> >>> 
>>>>>> >>> Local host: node1-128-02
>>>>>> >>> Value: "ib0"
>>>>>> >>> Message: Invalid specification (missing "/")
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> None of the TCP networks specified to be included for out-of-band 
>>>>>> >>> communications
>>>>>> >>> could be found:
>>>>>> >>> 
>>>>>> >>> Value given:
>>>>>> >>> 
>>>>>> >>> Please revise the specification and try again.
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> None of the TCP networks specified to be included for out-of-band 
>>>>>> >>> communications
>>>>>> >>> could be found:
>>>>>> >>> 
>>>>>> >>> Value given:
>>>>>> >>> 
>>>>>> >>> Please revise the specification and try again.
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> No network interfaces were found for out-of-band communications. We 
>>>>>> >>> require
>>>>>> >>> at least one available network for out-of-band messaging.
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> No network interfaces were found for out-of-band communications. We 
>>>>>> >>> require
>>>>>> >>> at least one available network for out-of-band messaging.
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> It looks like orte_init failed for some reason; your parallel 
>>>>>> >>> process is
>>>>>> >>> likely to abort. There are many reasons that a parallel process can
>>>>>> >>> fail during orte_init; some of which are due to configuration or
>>>>>> >>> environment problems. This failure appears to be an internal failure;
>>>>>> >>> here's some additional information (which may only be relevant to an
>>>>>> >>> Open MPI developer):
>>>>>> >>> 
>>>>>> >>> orte_oob_base_select failed
>>>>>> >>> --> Returned value (null) (-43) instead of ORTE_SUCCESS
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> It looks like orte_init failed for some reason; your parallel 
>>>>>> >>> process is
>>>>>> >>> likely to abort. There are many reasons that a parallel process can
>>>>>> >>> fail during orte_init; some of which are due to configuration or
>>>>>> >>> environment problems. This failure appears to be an internal failure;
>>>>>> >>> here's some additional information (which may only be relevant to an
>>>>>> >>> Open MPI developer):
>>>>>> >>> 
>>>>>> >>> orte_oob_base_select failed
>>>>>> >>> --> Returned value (null) (-43) instead of ORTE_SUCCESS
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> srun: error: node1-128-02: task 1: Exited with exit code 213
>>>>>> >>> srun: Terminating job step 657300.0
>>>>>> >>> srun: error: node1-128-01: task 0: Exited with exit code 213
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> An ORTE daemon has unexpectedly failed after launch and before
>>>>>> >>> communicating back to mpirun. This could be caused by a number
>>>>>> >>> of factors, including an inability to create a connection back
>>>>>> >>> to mpirun due to a lack of common network interfaces and/or no
>>>>>> >>> route found between them. Please check network connectivity
>>>>>> >>> (including firewalls and network routing requirements).
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> [compiler-2:08792] [[42190,0],0] orted_cmd: received halt_vm cmd
>>>>>> >>> [compiler-2:08792] mca: base: close: component oob closed
>>>>>> >>> [compiler-2:08792] mca: base: close: unloading component oob
>>>>>> >>> [compiler-2:08792] [[42190,0],0] TCP SHUTDOWN
>>>>>> >>> [compiler-2:08792] mca: base: close: component tcp closed
>>>>>> >>> [compiler-2:08792] mca: base: close: unloading component tcp
>>>>>> >>> 
>>>>>> >>> 
>>>>>> >>> 
>>>>>> >>> Tue, 12 Aug 2014 16:14:58 +0400 от Timur Ismagilov 
>>>>>> >>> <tismagi...@mail.ru>:
>>>>>> >>> Hello!
>>>>>> >>> 
>>>>>> >>> I have Open MPI v1.8.2rc4r32485
>>>>>> >>> 
>>>>>> >>> When i run hello_c, I got this error message
>>>>>> >>> $mpirun -np 2 hello_c
>>>>>> >>> 
>>>>>> >>> An ORTE daemon has unexpectedly failed after launch and before
>>>>>> >>> 
>>>>>> >>> communicating back to mpirun. This could be caused by a number
>>>>>> >>> of factors, including an inability to create a connection back
>>>>>> >>> to mpirun due to a lack of common network interfaces and/or no
>>>>>> >>> route found between them. Please check network connectivity
>>>>>> >>> (including firewalls and network routing requirements).
>>>>>> >>> 
>>>>>> >>> When i run with --debug-daemons --mca plm_base_verbose 5 -mca 
>>>>>> >>> oob_base_verbose 10 -mca rml_base_verbose 10 i got this output:
>>>>>> >>> $mpirun --debug-daemons --mca plm_base_verbose 5 -mca 
>>>>>> >>> oob_base_verbose 10 -mca rml_base_verbose 10 -np 2 hello_c
>>>>>> >>> 
>>>>>> >>> [compiler-2:08780] mca:base:select:( plm) Querying component 
>>>>>> >>> [isolated]
>>>>>> >>> [compiler-2:08780] mca:base:select:( plm) Query of component 
>>>>>> >>> [isolated] set priority to 0
>>>>>> >>> [compiler-2:08780] mca:base:select:( plm) Querying component [rsh]
>>>>>> >>> [compiler-2:08780] mca:base:select:( plm) Query of component [rsh] 
>>>>>> >>> set priority to 10
>>>>>> >>> [compiler-2:08780] mca:base:select:( plm) Querying component [slurm]
>>>>>> >>> [compiler-2:08780] mca:base:select:( plm) Query of component [slurm] 
>>>>>> >>> set priority to 75
>>>>>> >>> [compiler-2:08780] mca:base:select:( plm) Selected component [slurm]
>>>>>> >>> [compiler-2:08780] mca: base: components_register: registering oob 
>>>>>> >>> components
>>>>>> >>> [compiler-2:08780] mca: base: components_register: found loaded 
>>>>>> >>> component tcp
>>>>>> >>> [compiler-2:08780] mca: base: components_register: component tcp 
>>>>>> >>> register function successful
>>>>>> >>> [compiler-2:08780] mca: base: components_open: opening oob components
>>>>>> >>> [compiler-2:08780] mca: base: components_open: found loaded 
>>>>>> >>> component tcp
>>>>>> >>> [compiler-2:08780] mca: base: components_open: component tcp open 
>>>>>> >>> function successful
>>>>>> >>> [compiler-2:08780] mca:oob:select: checking available component tcp
>>>>>> >>> [compiler-2:08780] mca:oob:select: Querying component [tcp]
>>>>>> >>> [compiler-2:08780] oob:tcp: component_available called
>>>>>> >>> [compiler-2:08780] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
>>>>>> >>> [compiler-2:08780] WORKING INTERFACE 2 KERNEL INDEX 3 FAMILY: V4
>>>>>> >>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 10.0.251.53 to 
>>>>>> >>> our list of V4 connections
>>>>>> >>> [compiler-2:08780] WORKING INTERFACE 3 KERNEL INDEX 4 FAMILY: V4
>>>>>> >>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 10.0.0.4 to our 
>>>>>> >>> list of V4 connections
>>>>>> >>> [compiler-2:08780] WORKING INTERFACE 4 KERNEL INDEX 5 FAMILY: V4
>>>>>> >>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 10.2.251.14 to 
>>>>>> >>> our list of V4 connections
>>>>>> >>> [compiler-2:08780] WORKING INTERFACE 5 KERNEL INDEX 6 FAMILY: V4
>>>>>> >>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 10.128.0.4 to 
>>>>>> >>> our list of V4 connections
>>>>>> >>> [compiler-2:08780] WORKING INTERFACE 6 KERNEL INDEX 7 FAMILY: V4
>>>>>> >>> [compiler-2:08780] [[42202,0],0] oob:tcp:init adding 93.180.7.38 to 
>>>>>> >>> our list of V4 connections
>>>>>> >>> [compiler-2:08780] [[42202,0],0] TCP STARTUP
>>>>>> >>> [compiler-2:08780] [[42202,0],0] attempting to bind to IPv4 port 0
>>>>>> >>> [compiler-2:08780] [[42202,0],0] assigned IPv4 port 38420
>>>>>> >>> [compiler-2:08780] mca:oob:select: Adding component to end
>>>>>> >>> [compiler-2:08780] mca:oob:select: Found 1 active transports
>>>>>> >>> [compiler-2:08780] mca: base: components_register: registering rml 
>>>>>> >>> components
>>>>>> >>> [compiler-2:08780] mca: base: components_register: found loaded 
>>>>>> >>> component oob
>>>>>> >>> [compiler-2:08780] mca: base: components_register: component oob has 
>>>>>> >>> no register or open function
>>>>>> >>> [compiler-2:08780] mca: base: components_open: opening rml components
>>>>>> >>> [compiler-2:08780] mca: base: components_open: found loaded 
>>>>>> >>> component oob
>>>>>> >>> [compiler-2:08780] mca: base: components_open: component oob open 
>>>>>> >>> function successful
>>>>>> >>> [compiler-2:08780] orte_rml_base_select: initializing rml component 
>>>>>> >>> oob
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 30 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 15 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 32 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 33 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 5 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 10 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 12 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 9 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 34 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 2 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 21 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 22 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 45 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 46 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 1 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting recv
>>>>>> >>> [compiler-2:08780] [[42202,0],0] posting persistent recv on tag 27 
>>>>>> >>> for peer [[WILDCARD],WILDCARD]
>>>>>> >>> Daemon was launched on node1-130-08 - beginning to initialize
>>>>>> >>> Daemon was launched on node1-130-03 - beginning to initialize
>>>>>> >>> Daemon was launched on node1-130-05 - beginning to initialize
>>>>>> >>> Daemon was launched on node1-130-02 - beginning to initialize
>>>>>> >>> Daemon was launched on node1-130-01 - beginning to initialize
>>>>>> >>> Daemon was launched on node1-130-04 - beginning to initialize
>>>>>> >>> Daemon was launched on node1-130-07 - beginning to initialize
>>>>>> >>> Daemon was launched on node1-130-06 - beginning to initialize
>>>>>> >>> Daemon [[42202,0],3] checking in as pid 7178 on host node1-130-03
>>>>>> >>> [node1-130-03:07178] [[42202,0],3] orted: up and running - waiting 
>>>>>> >>> for commands!
>>>>>> >>> Daemon [[42202,0],2] checking in as pid 13581 on host node1-130-02
>>>>>> >>> [node1-130-02:13581] [[42202,0],2] orted: up and running - waiting 
>>>>>> >>> for commands!
>>>>>> >>> Daemon [[42202,0],1] checking in as pid 17220 on host node1-130-01
>>>>>> >>> [node1-130-01:17220] [[42202,0],1] orted: up and running - waiting 
>>>>>> >>> for commands!
>>>>>> >>> Daemon [[42202,0],5] checking in as pid 6663 on host node1-130-05
>>>>>> >>> [node1-130-05:06663] [[42202,0],5] orted: up and running - waiting 
>>>>>> >>> for commands!
>>>>>> >>> Daemon [[42202,0],8] checking in as pid 6683 on host node1-130-08
>>>>>> >>> [node1-130-08:06683] [[42202,0],8] orted: up and running - waiting 
>>>>>> >>> for commands!
>>>>>> >>> Daemon [[42202,0],7] checking in as pid 7877 on host node1-130-07
>>>>>> >>> [node1-130-07:07877] [[42202,0],7] orted: up and running - waiting 
>>>>>> >>> for commands!
>>>>>> >>> Daemon [[42202,0],4] checking in as pid 7735 on host node1-130-04
>>>>>> >>> [node1-130-04:07735] [[42202,0],4] orted: up and running - waiting 
>>>>>> >>> for commands!
>>>>>> >>> Daemon [[42202,0],6] checking in as pid 8451 on host node1-130-06
>>>>>> >>> [node1-130-06:08451] [[42202,0],6] orted: up and running - waiting 
>>>>>> >>> for commands!
>>>>>> >>> srun: error: node1-130-03: task 2: Exited with exit code 1
>>>>>> >>> srun: Terminating job step 657040.1
>>>>>> >>> srun: error: node1-130-02: task 1: Exited with exit code 1
>>>>>> >>> slurmd[node1-130-04]: *** STEP 657040.1 KILLED AT 
>>>>>> >>> 2014-08-12T12:59:07 WITH SIGNAL 9 ***
>>>>>> >>> slurmd[node1-130-07]: *** STEP 657040.1 KILLED AT 
>>>>>> >>> 2014-08-12T12:59:07 WITH SIGNAL 9 ***
>>>>>> >>> slurmd[node1-130-06]: *** STEP 657040.1 KILLED AT 
>>>>>> >>> 2014-08-12T12:59:07 WITH SIGNAL 9 ***
>>>>>> >>> srun: Job step aborted: Waiting up to 2 seconds for job step to 
>>>>>> >>> finish.
>>>>>> >>> srun: error: node1-130-01: task 0: Exited with exit code 1
>>>>>> >>> srun: error: node1-130-05: task 4: Exited with exit code 1
>>>>>> >>> srun: error: node1-130-08: task 7: Exited with exit code 1
>>>>>> >>> srun: error: node1-130-07: task 6: Exited with exit code 1
>>>>>> >>> srun: error: node1-130-04: task 3: Killed
>>>>>> >>> srun: error: node1-130-06: task 5: Killed
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> An ORTE daemon has unexpectedly failed after launch and before
>>>>>> >>> communicating back to mpirun. This could be caused by a number
>>>>>> >>> of factors, including an inability to create a connection back
>>>>>> >>> to mpirun due to a lack of common network interfaces and/or no
>>>>>> >>> route found between them. Please check network connectivity
>>>>>> >>> (including firewalls and network routing requirements).
>>>>>> >>> --------------------------------------------------------------------------
>>>>>> >>> [compiler-2:08780] [[42202,0],0] orted_cmd: received halt_vm cmd
>>>>>> >>> [compiler-2:08780] mca: base: close: component oob closed
>>>>>> >>> [compiler-2:08780] mca: base: close: unloading component oob
>>>>>> >>> [compiler-2:08780] [[42202,0],0] TCP SHUTDOWN
>>>>>> >>> [compiler-2:08780] mca: base: close: component tcp closed
>>>>>> >>> [compiler-2:08780] mca: base: close: unloading component tcp
>>>>>> >>> 
>>>>>> >>> _______________________________________________
>>>>>> >>> users mailing list
>>>>>> >>> us...@open-mpi.org
>>>>>> >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> >>> Link to this post: 
>>>>>> >>> http://www.open-mpi.org/community/lists/users/2014/08/24987.php
>>>>>> >>> 
>>>>>> >>> 
>>>>>> >>> 
>>>>>> >>> _______________________________________________
>>>>>> >>> users mailing list
>>>>>> >>> us...@open-mpi.org
>>>>>> >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> >>> Link to this post: 
>>>>>> >>> http://www.open-mpi.org/community/lists/users/2014/08/24988.php
>>>>>> >> 
>>>>>> >> 
>>>>>> >> -- 
>>>>>> >> Jeff Squyres
>>>>>> >> jsquy...@cisco.com
>>>>>> >> For corporate legal information go to: 
>>>>>> >> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>> >> 
>>>>>> >> 
>>>>>> >> 
>>>>>> >> 
>>>>>> >> 
>>>>>> > 
>>>>>> > 
>>>>>> > -- 
>>>>>> > Jeff Squyres
>>>>>> > jsquy...@cisco.com
>>>>>> > For corporate legal information go to: 
>>>>>> > http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>> > 
>>>>>> > _______________________________________________
>>>>>> > users mailing list
>>>>>> > us...@open-mpi.org
>>>>>> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> > Link to this post: 
>>>>>> > http://www.open-mpi.org/community/lists/users/2014/08/25001.php
>>>>>> 
>>>>>> 
>>>>>> -- 
>>>>>> Jeff Squyres
>>>>>> jsquy...@cisco.com
>>>>>> For corporate legal information go to: 
>>>>>> http://www.cisco.com/web/about/doing_business/legal/cri/
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> Link to this post: 
>>>>>> http://www.open-mpi.org/community/lists/users/2014/08/25086.php
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post: 
>>>>> http://www.open-mpi.org/community/lists/users/2014/08/25093.php
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post: 
>>>>> http://www.open-mpi.org/community/lists/users/2014/08/25094.php
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> Link to this post: 
>>>> http://www.open-mpi.org/community/lists/users/2014/08/25095.php
>>>> 
>>>> 
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> Link to this post: 
>>>> http://www.open-mpi.org/community/lists/users/2014/08/25105.php
>>> 
>>> 
>>> 
>>> 
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2014/08/25127.php
>>> 
>>> 
>>> 
>>> -- 
>>> 
>>> Kind Regards,
>>> 
>>> M.
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2014/08/25128.php
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2014/08/25129.php
>> 
>> 
>> 
> 
> 
> 
> 
> 
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2014/08/25154.php

Reply via email to