Yes, that would indeed break things. The 1.5 series isn't correctly checking 
connections across multiple interfaces until it finds one that works - it just 
uses the first one it sees. :-(

The solution is to specify -mca oob_tcp_if_include ib0. This will direct the 
run-time wireup across the IP over IB interface.

You will also need the -mca btl_tcp_if_include ib0 as well so the MPI comm goes 
exclusively over that network. Specifying both include and exclude should 
generate an error as those are mutually exclusive options - I think this was 
also missed in early 1.5 releases and was recently patched.

HTH
Ralph


On Nov 23, 2011, at 12:14 PM, TERRY DONTJE wrote:

> On 11/23/2011 2:02 PM, Paul Kapinos wrote:
>> 
>> Hello Ralph, hello all, 
>> 
>> Two news, as usual a good and a bad one. 
>> 
>> The good: we believe to find out *why* it hangs 
>> 
>> The bad: it seem for me, this is a bug or at least undocumented feature of 
>> Open MPI /1.5.x. 
>> 
>> In detail: 
>> As said, we see mystery hang-ups if starting on some nodes using some 
>> permutation of hostnames. Usually removing "some bad" nodes helps, sometimes 
>> a permutation of node names in the hostfile is enough(!). The behaviour is 
>> reproducible. 
>> 
>> The machines have at least 2 networks: 
>> 
>> *eth0* is used for installation, monitoring, ... - this ethernet is very 
>> slim 
>> 
>> *ib0* - is the "IP over IB" interface and is used for everything: the file 
>> systems, ssh and so on. The hostnames are bound to the ib0 network; our idea 
>> was not to use eth0 for MPI at all. 
>> 
>> all machines are available from any over ib0 (are in one network). 
>> 
>> But on eth0 there are at least two different networks; especially the 
>> computer linuxbsc025 is in different network than the others and is not 
>> reachable from other nodes over eth0! (but reachable over ib0. The name used 
>> in the hostfile is resolved to the IP of ib0 ). 
>> 
>> So I believe that Open MPI /1.5.x tries to communicate over eth0 and cannot 
>> do it, and hangs. The /1.4.3 does not hang, so this issue is 1.5.x-specific 
>> (seen in 1.5.3 and 1.5.4). A bug? 
>> 
>> I also tried to disable the eth0 completely: 
>> 
>> $ mpiexec -mca btl_tcp_if_exclude eth0,lo  -mca btl_tcp_if_include ib0 ... 
>> 
> I believe if you give "-mca btl_tcp_if_include ib0" you do not need to 
> specify the exclude parameter.
>> ...but this does not help. All right, the above command should disable the 
>> usage of eth0 for MPI communication itself, but it hangs just before the MPI 
>> is started, isn't it? (because one process lacks, the MPI_INIT cannot be 
>> passed) 
>> 
> By "just before the MPI is started" do you mean while orte is launching the 
> processes.
> I wonder if you need to specify "-mca oob_tcp_if_include ib0" also but I 
> think that may depend on which oob you are using.
>> Now a question: is there a way to forbid the mpiexec to use some interfaces 
>> at all? 
>> 
>> Best wishes, 
>> 
>> Paul Kapinos 
>> 
>> P.S. Of course we know about the good idea to bring all nodes into the same 
>> net on eth0, but at this point it is impossible due of technical 
>> reason[s]... 
>> 
>> P.S.2 I'm not sure that the issue is really rooted in the above mentioned 
>> misconfiguration of eth0, but I have no better idea at this point... 
>> 
>> 
>>>> The map seem to be correctly build, also the output if the daemons seem to 
>>>> be the same (see helloworld.txt) 
>>> 
>>> Unfortunately, it appears that OMPI was not built with --enable-debug as 
>>> there is no debug info in the output. Without a debug installation of OMPI, 
>>> the ability to determine the problem is pretty limited. 
>> 
>> well, this will be the next option we will activate. We also have another 
>> issue here, on (not) using uDAPL.. 
>> 
>> 
>>> 
>>> 
>>>>> You should also try putting that long list of nodes in a hostfile - see 
>>>>> if that makes a difference. 
>>>>> It will process the nodes thru a different code path, so if there is some 
>>>>> problem in --host, 
>>>>> this will tell us. 
>>>> No, with the host file instead of host list on command line the behaviour 
>>>> is the same. 
>>>> 
>>>> But, I just found out that the 1.4.3 does *not* hang on this 
>>>> constellation. The next thing I will try will be the installation of 1.5.4 
>>>> :o) 
>>>> 
>>>> Best, 
>>>> 
>>>> Paul 
>>>> 
>>>> P.S. started: 
>>>> 
>>>> $ /opt/MPI/openmpi-1.5.3/linux/intel/bin/mpiexec --hostfile hostfile-mini 
>>>> -mca odls_base_verbose 5 --leave-session-attached --display-map  
>>>> helloworld 2>&1 | tee helloworld.txt 
>>>> 
>>>> 
>>>> 
>>>>> On Nov 21, 2011, at 9:33 AM, Paul Kapinos wrote: 
>>>>>> Hello Open MPI volks, 
>>>>>> 
>>>>>> We use OpenMPI 1.5.3 on our pretty new 1800+ nodes InfiniBand cluster, 
>>>>>> and we have some strange hangups if starting OpenMPI processes. 
>>>>>> 
>>>>>> The nodes are named linuxbsc001,linuxbsc002,... (with some lacuna due of 
>>>>>>  offline nodes). Each node is accessible from each other over SSH 
>>>>>> (without password), also MPI programs between any two nodes are checked 
>>>>>> to run. 
>>>>>> 
>>>>>> 
>>>>>> So long, I tried to start some bigger number of processes, one process 
>>>>>> per node: 
>>>>>> $ mpiexec -np NN  --host linuxbsc001,linuxbsc002,... MPI_FastTest.exe 
>>>>>> 
>>>>>> Now the problem: there are some constellations of names in the host list 
>>>>>> on which mpiexec reproducible hangs forever; and more surprising: other 
>>>>>> *permutation* of the *same* node names may run without any errors! 
>>>>>> 
>>>>>> Example: the command in laueft.txt runs OK, the command in haengt.txt 
>>>>>> hangs. Note: the only difference is that the node linuxbsc025 is put on 
>>>>>> the end of the host list. Amazed, too? 
>>>>>> 
>>>>>> Looking on the particular nodes during the above mpiexec hangs, we found 
>>>>>> the orted daemons started on *each* node and the binary on all but one 
>>>>>> node (orted.txt, MPI_FastTest.txt). 
>>>>>> Again amazing that the node with no user process started (leading to 
>>>>>> hangup in MPI_Init of all processes and thus to hangup, I believe) was 
>>>>>> always the same, linuxbsc005, which is NOT the permuted item 
>>>>>> linuxbsc025... 
>>>>>> 
>>>>>> This behaviour is reproducible. The hang-on only occure if the started 
>>>>>> application is a MPI application ("hostname" does not hang). 
>>>>>> 
>>>>>> 
>>>>>> Any Idea what is gonna on? 
>>>>>> 
>>>>>> 
>>>>>> Best, 
>>>>>> 
>>>>>> Paul Kapinos 
>>>>>> 
>>>>>> 
>>>>>> P.S: no alias names used, all names are real ones 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> -- 
>>>>>> Dipl.-Inform. Paul Kapinos   -   High Performance Computing, 
>>>>>> RWTH Aachen University, Center for Computing and Communication 
>>>>>> Seffenter Weg 23,  D 52074  Aachen (Germany) 
>>>>>> Tel: +49 241/80-24915 
>>>>>> linuxbsc001: STDOUT: 24323 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc002: STDOUT:  2142 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc003: STDOUT: 69266 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc004: STDOUT: 58899 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc006: STDOUT: 68255 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc007: STDOUT: 62026 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc008: STDOUT: 54221 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc009: STDOUT: 55482 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc010: STDOUT: 59380 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc011: STDOUT: 58312 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc014: STDOUT: 56013 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc016: STDOUT: 58563 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc017: STDOUT: 54693 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc018: STDOUT: 54187 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc020: STDOUT: 55811 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc021: STDOUT: 54982 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc022: STDOUT: 50032 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc023: STDOUT: 54044 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc024: STDOUT: 51247 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc025: STDOUT: 18575 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc027: STDOUT: 48969 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc028: STDOUT: 52397 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc029: STDOUT: 52780 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc030: STDOUT: 47537 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc031: STDOUT: 54609 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> linuxbsc032: STDOUT: 52833 ?        SLl    0:00 MPI_FastTest.exe 
>>>>>> $ timex /opt/MPI/openmpi-1.5.3/linux/intel/bin/mpiexec -np 27  --host 
>>>>>> linuxbsc001,linuxbsc002,linuxbsc003,linuxbsc004,linuxbsc005,linuxbsc006,linuxbsc007,linuxbsc008,linuxbsc009,linuxbsc010,linuxbsc011,linuxbsc014,linuxbsc016,linuxbsc017,linuxbsc018,linuxbsc020,linuxbsc021,linuxbsc022,linuxbsc023,linuxbsc024,linuxbsc025,linuxbsc027,linuxbsc028,linuxbsc029,linuxbsc030,linuxbsc031,linuxbsc032
>>>>>>  MPI_FastTest.exe 
>>>>>> $ timex /opt/MPI/openmpi-1.5.3/linux/intel/bin/mpiexec -np 27  --host 
>>>>>> linuxbsc001,linuxbsc002,linuxbsc003,linuxbsc004,linuxbsc005,linuxbsc006,linuxbsc007,linuxbsc008,linuxbsc009,linuxbsc010,linuxbsc011,linuxbsc014,linuxbsc016,linuxbsc017,linuxbsc018,linuxbsc020,linuxbsc021,linuxbsc022,linuxbsc023,linuxbsc024,linuxbsc027,linuxbsc028,linuxbsc029,linuxbsc030,linuxbsc031,linuxbsc032,linuxbsc025
>>>>>>  MPI_FastTest.exe 
>>>>>> linuxbsc001: STDOUT: 24322 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 1 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc002: STDOUT:  2141 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 2 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc003: STDOUT: 69265 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 3 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc004: STDOUT: 58898 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 4 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc005: STDOUT: 65642 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 5 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc006: STDOUT: 68254 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 6 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc007: STDOUT: 62025 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 7 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc008: STDOUT: 54220 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 8 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc009: STDOUT: 55481 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 9 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc010: STDOUT: 59379 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 10 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc011: STDOUT: 58311 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 11 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc014: STDOUT: 56012 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 12 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc016: STDOUT: 58562 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 13 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc017: STDOUT: 54692 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 14 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc018: STDOUT: 54186 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 15 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc020: STDOUT: 55810 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 16 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc021: STDOUT: 54981 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 17 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc022: STDOUT: 50031 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 18 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc023: STDOUT: 54043 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 19 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc024: STDOUT: 51246 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 20 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc025: STDOUT: 18574 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 21 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc027: STDOUT: 48968 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 22 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc028: STDOUT: 52396 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 23 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc029: STDOUT: 52779 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 24 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc030: STDOUT: 47536 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 25 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc031: STDOUT: 54608 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 26 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> linuxbsc032: STDOUT: 52832 ?        Ss     0:00 
>>>>>> /opt/MPI/openmpi-1.5.3/linux/intel/bin/orted --daemonize -mca ess env 
>>>>>> -mca orte_ess_jobid 751435776 -mca orte_ess_vpid 27 -mca 
>>>>>> orte_ess_num_procs 28 --hnp-uri 751435776.0;tcp://134.61.194.2:33210 
>>>>>> -mca plm rsh 
>>>>>> _______________________________________________ 
>>>>>> users mailing list 
>>>>>> us...@open-mpi.org 
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>>>> _______________________________________________ 
>>>>> users mailing list 
>>>>> us...@open-mpi.org 
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>>> 
>>>> -- 
>>>> Dipl.-Inform. Paul Kapinos   -   High Performance Computing, 
>>>> RWTH Aachen University, Center for Computing and Communication 
>>>> Seffenter Weg 23,  D 52074  Aachen (Germany) 
>>>> Tel: +49 241/80-24915 
>>>> linuxbsc005 slots=1 
>>>> linuxbsc006 slots=1 
>>>> linuxbsc007 slots=1 
>>>> linuxbsc008 slots=1 
>>>> linuxbsc009 slots=1 
>>>> linuxbsc010 slots=1 
>>>> linuxbsc011 slots=1 
>>>> linuxbsc014 slots=1 
>>>> linuxbsc016 slots=1 
>>>> linuxbsc017 slots=1 
>>>> linuxbsc018 slots=1 
>>>> linuxbsc020 slots=1 
>>>> linuxbsc021 slots=1 
>>>> linuxbsc022 slots=1 
>>>> linuxbsc023 slots=1 
>>>> linuxbsc024 slots=1 
>>>> linuxbsc025 slots=1[linuxc2.rz.RWTH-Aachen.DE:22229] mca:base:select:( 
>>>> odls) Querying component [default] 
>>>> [linuxc2.rz.RWTH-Aachen.DE:22229] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxc2.rz.RWTH-Aachen.DE:22229] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> 
>>>> ========================   JOB MAP   ======================== 
>>>> 
>>>> Data for node: linuxbsc005    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 0 
>>>> 
>>>> Data for node: linuxbsc006    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 1 
>>>> 
>>>> Data for node: linuxbsc007    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 2 
>>>> 
>>>> Data for node: linuxbsc008    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 3 
>>>> 
>>>> Data for node: linuxbsc009    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 4 
>>>> 
>>>> Data for node: linuxbsc010    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 5 
>>>> 
>>>> Data for node: linuxbsc011    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 6 
>>>> 
>>>> Data for node: linuxbsc014    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 7 
>>>> 
>>>> Data for node: linuxbsc016    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 8 
>>>> 
>>>> Data for node: linuxbsc017    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 9 
>>>> 
>>>> Data for node: linuxbsc018    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 10 
>>>> 
>>>> Data for node: linuxbsc020    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 11 
>>>> 
>>>> Data for node: linuxbsc021    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 12 
>>>> 
>>>> Data for node: linuxbsc022    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 13 
>>>> 
>>>> Data for node: linuxbsc023    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 14 
>>>> 
>>>> Data for node: linuxbsc024    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 15 
>>>> 
>>>> Data for node: linuxbsc025    Num procs: 1 
>>>>     Process OMPI jobid: [87,1] Process rank: 16 
>>>> 
>>>> ============================================================= 
>>>> [linuxbsc007.rz.RWTH-Aachen.DE:07574] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc007.rz.RWTH-Aachen.DE:07574] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc007.rz.RWTH-Aachen.DE:07574] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc016.rz.RWTH-Aachen.DE:03146] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc016.rz.RWTH-Aachen.DE:03146] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc016.rz.RWTH-Aachen.DE:03146] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc005.rz.RWTH-Aachen.DE:22051] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc005.rz.RWTH-Aachen.DE:22051] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc005.rz.RWTH-Aachen.DE:22051] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc011.rz.RWTH-Aachen.DE:07131] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc011.rz.RWTH-Aachen.DE:07131] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc011.rz.RWTH-Aachen.DE:07131] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc025.rz.RWTH-Aachen.DE:43153] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc025.rz.RWTH-Aachen.DE:43153] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc025.rz.RWTH-Aachen.DE:43153] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc017.rz.RWTH-Aachen.DE:05044] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc017.rz.RWTH-Aachen.DE:05044] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc017.rz.RWTH-Aachen.DE:05044] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc018.rz.RWTH-Aachen.DE:01840] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc018.rz.RWTH-Aachen.DE:01840] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc018.rz.RWTH-Aachen.DE:01840] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc024.rz.RWTH-Aachen.DE:79549] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc024.rz.RWTH-Aachen.DE:79549] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc024.rz.RWTH-Aachen.DE:79549] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc022.rz.RWTH-Aachen.DE:73501] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc022.rz.RWTH-Aachen.DE:73501] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc022.rz.RWTH-Aachen.DE:73501] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc023.rz.RWTH-Aachen.DE:03364] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc023.rz.RWTH-Aachen.DE:03364] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc023.rz.RWTH-Aachen.DE:03364] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc006.rz.RWTH-Aachen.DE:16811] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc006.rz.RWTH-Aachen.DE:16811] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc006.rz.RWTH-Aachen.DE:16811] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc014.rz.RWTH-Aachen.DE:10206] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc014.rz.RWTH-Aachen.DE:10206] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc014.rz.RWTH-Aachen.DE:10206] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc008.rz.RWTH-Aachen.DE:00858] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc008.rz.RWTH-Aachen.DE:00858] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc008.rz.RWTH-Aachen.DE:00858] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc010.rz.RWTH-Aachen.DE:09727] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc010.rz.RWTH-Aachen.DE:09727] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc010.rz.RWTH-Aachen.DE:09727] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc020.rz.RWTH-Aachen.DE:06680] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc020.rz.RWTH-Aachen.DE:06680] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc020.rz.RWTH-Aachen.DE:06680] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc009.rz.RWTH-Aachen.DE:05145] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc009.rz.RWTH-Aachen.DE:05145] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc009.rz.RWTH-Aachen.DE:05145] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> [linuxbsc021.rz.RWTH-Aachen.DE:01405] mca:base:select:( odls) Querying 
>>>> component [default] 
>>>> [linuxbsc021.rz.RWTH-Aachen.DE:01405] mca:base:select:( odls) Query of 
>>>> component [default] set priority to 1 
>>>> [linuxbsc021.rz.RWTH-Aachen.DE:01405] mca:base:select:( odls) Selected 
>>>> component [default] 
>>>> _______________________________________________ 
>>>> users mailing list 
>>>> us...@open-mpi.org 
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> 
>>> 
>>> _______________________________________________ 
>>> users mailing list 
>>> us...@open-mpi.org 
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> 
>> 
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> -- 
> <Mail Attachment.gif>
> Terry D. Dontje | Principal Software Engineer
> Developer Tools Engineering | +1.781.442.2631
> Oracle - Performance Technologies
> 95 Network Drive, Burlington, MA 01803
> Email terry.don...@oracle.com
> 
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to