Why do it the hard way? I'll look at the FAQ because that definitely isn't a recommended thing to do - better to use -host to specify the subset, or just specify the desired mapping using all the various mappers we provide.
On Nov 13, 2013, at 6:39 PM, tmish...@jcity.maeda.co.jp wrote: > > > Sorry for cross-post. > > Nodefile is very simple which consists of 8 lines: > > node08 > node08 > node08 > node08 > node08 > node08 > node08 > node08 > > Therefore, NPROCS=8 > > My aim is to modify the allocation as you pointed out. According to Openmpi > FAQ, > proper subset of the hosts allocated to the Torque / PBS Pro job should be > allowed. > > tmishima > >> Please - can you answer my question on script2? What is the value of > NPROCS? >> >> Why would you want to do it this way? Are you planning to modify the > allocation?? That generally is a bad idea as it can confuse the system >> >> >> On Nov 13, 2013, at 5:55 PM, tmish...@jcity.maeda.co.jp wrote: >> >>> >>> >>> Since what I really want is to run script2 correctly, please let us >>> concentrate script2. >>> >>> I'm not an expert of the inside of openmpi. What I can do is just >>> obsabation >>> from the outside. I doubt these lines are strange, especially the last > one. >>> >>> [node08.cluster:26952] mca:rmaps:rr: mapping job [56581,1] >>> [node08.cluster:26952] [[56581,0],0] Starting with 1 nodes in list >>> [node08.cluster:26952] [[56581,0],0] Filtering thru apps >>> [node08.cluster:26952] [[56581,0],0] Retained 1 nodes in list >>> [node08.cluster:26952] [[56581,0],0] Removing node node08 slots 0 inuse > 0 >>> >>> These lines come from this part of orte_rmaps_base_get_target_nodes >>> in rmaps_base_support_fns.c: >>> >>> } else if (node->slots <= node->slots_inuse && >>> (ORTE_MAPPING_NO_OVERSUBSCRIBE & >>> ORTE_GET_MAPPING_DIRECTIVE(policy))) { >>> /* remove the node as fully used */ >>> OPAL_OUTPUT_VERBOSE((5, >>> orte_rmaps_base_framework.framework_output, >>> "%s Removing node %s slots %d inuse > %d", >>> ORTE_NAME_PRINT(ORTE_PROC_MY_NAME), >>> node->name, node->slots, node-> >>> slots_inuse)); >>> opal_list_remove_item(allocated_nodes, item); >>> OBJ_RELEASE(item); /* "un-retain" it */ >>> >>> I wonder why node->slots and node->slots_inuse is 0, which I can read >>> from the above line "Removing node node08 slots 0 inuse 0". >>> >>> Or I'm not sure but >>> "else if (node->slots <= node->slots_inuse &&" should be >>> "else if (node->slots < node->slots_inuse &&" ? >>> >>> tmishima >>> >>>> On Nov 13, 2013, at 4:43 PM, tmish...@jcity.maeda.co.jp wrote: >>>> >>>>> >>>>> >>>>> Yes, the node08 has 8 slots but the process I run is also 8. >>>>> >>>>> #PBS -l nodes=node08:ppn=8 >>>>> >>>>> Therefore, I think it should allow this allocation. Is that right? >>>> >>>> Correct >>>> >>>>> >>>>> My question is why scritp1 works and script2 does not. They are >>>>> almost same. >>>>> >>>>> #PBS -l nodes=node08:ppn=8 >>>>> export OMP_NUM_THREADS=1 >>>>> cd $PBS_O_WORKDIR >>>>> cp $PBS_NODEFILE pbs_hosts >>>>> NPROCS=`wc -l < pbs_hosts` >>>>> >>>>> #SCRITP1 >>>>> mpirun -report-bindings -bind-to core Myprog >>>>> >>>>> #SCRIPT2 >>>>> mpirun -machinefile pbs_hosts -np ${NPROCS} -report-bindings -bind-to >>> core >>>> >>>> This version is not only reading the PBS allocation, but also invoking >>> the hostfile filter on top of it. Different code path. I'll take a look > - >>> it should still match up assuming NPROCS=8. Any >>>> possibility that it is a different number? I don't recall, but isn't >>> there some extra lines in the nodefile - e.g., comments? >>>> >>>> >>>>> Myprog >>>>> >>>>> tmishima >>>>> >>>>>> I guess here's my confusion. If you are using only one node, and > that >>>>> node has 8 allocated slots, then we will not allow you to run more > than >>> 8 >>>>> processes on that node unless you specifically provide >>>>>> the --oversubscribe flag. This is because you are operating in a >>> managed >>>>> environment (in this case, under Torque), and so we treat the >>> allocation as >>>>> "mandatory" by default. >>>>>> >>>>>> I suspect that is the issue here, in which case the system is > behaving >>> as >>>>> it should. >>>>>> >>>>>> Is the above accurate? >>>>>> >>>>>> >>>>>> On Nov 13, 2013, at 4:11 PM, Ralph Castain <r...@open-mpi.org> wrote: >>>>>> >>>>>>> It has nothing to do with LAMA as you aren't using that mapper. >>>>>>> >>>>>>> How many nodes are in this allocation? >>>>>>> >>>>>>> On Nov 13, 2013, at 4:06 PM, tmish...@jcity.maeda.co.jp wrote: >>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Hi Ralph, this is an additional information. >>>>>>>> >>>>>>>> Here is the main part of output by adding "-mca rmaps_base_verbose >>>>> 50". >>>>>>>> >>>>>>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm >>>>>>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm creating > map >>>>>>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm only HNP in >>>>>>>> allocation >>>>>>>> [node08.cluster:26952] mca:rmaps: mapping job [56581,1] >>>>>>>> [node08.cluster:26952] mca:rmaps: creating new map for job > [56581,1] >>>>>>>> [node08.cluster:26952] mca:rmaps:ppr: job [56581,1] not using ppr >>>>> mapper >>>>>>>> [node08.cluster:26952] [[56581,0],0] rmaps:seq mapping job > [56581,1] >>>>>>>> [node08.cluster:26952] mca:rmaps:seq: job [56581,1] not using seq >>>>> mapper >>>>>>>> [node08.cluster:26952] mca:rmaps:resilient: cannot perform initial >>> map >>>>> of >>>>>>>> job [56581,1] - no fault groups >>>>>>>> [node08.cluster:26952] mca:rmaps:mindist: job [56581,1] not using >>>>> mindist >>>>>>>> mapper >>>>>>>> [node08.cluster:26952] mca:rmaps:rr: mapping job [56581,1] >>>>>>>> [node08.cluster:26952] [[56581,0],0] Starting with 1 nodes in list >>>>>>>> [node08.cluster:26952] [[56581,0],0] Filtering thru apps >>>>>>>> [node08.cluster:26952] [[56581,0],0] Retained 1 nodes in list >>>>>>>> [node08.cluster:26952] [[56581,0],0] Removing node node08 slots 0 >>>>> inuse 0 >>>>>>>> >>>>>>>> From this result, I guess it's related to oversubscribe. >>>>>>>> So I added "-oversubscribe" and rerun, then it worked well as show >>>>> below: >>>>>>>> >>>>>>>> [node08.cluster:27019] [[56774,0],0] Starting with 1 nodes in list >>>>>>>> [node08.cluster:27019] [[56774,0],0] Filtering thru apps >>>>>>>> [node08.cluster:27019] [[56774,0],0] Retained 1 nodes in list >>>>>>>> [node08.cluster:27019] AVAILABLE NODES FOR MAPPING: >>>>>>>> [node08.cluster:27019] node: node08 daemon: 0 >>>>>>>> [node08.cluster:27019] [[56774,0],0] Starting bookmark at node >>> node08 >>>>>>>> [node08.cluster:27019] [[56774,0],0] Starting at node node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:rr: mapping by slot for job >>> [56774,1] >>>>>>>> slots 1 num_procs 8 >>>>>>>> [node08.cluster:27019] mca:rmaps:rr:slot working node node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:rr:slot node node08 is full - >>>>> skipping >>>>>>>> [node08.cluster:27019] mca:rmaps:rr:slot job [56774,1] is >>>>> oversubscribed - >>>>>>>> performing second pass >>>>>>>> [node08.cluster:27019] mca:rmaps:rr:slot working node node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:rr:slot adding up to 8 procs to >>> node >>>>>>>> node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:base: computing vpids by slot for >>> job >>>>>>>> [56774,1] >>>>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 0 to node >>> node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 1 to node >>> node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 2 to node >>> node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 3 to node >>> node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 4 to node >>> node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 5 to node >>> node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 6 to node >>> node08 >>>>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 7 to node >>> node08 >>>>>>>> >>>>>>>> I think something is wrong with treatment of oversubscription, > which >>>>> might >>>>>>>> be >>>>>>>> related to "#3893: LAMA mapper has problems" >>>>>>>> >>>>>>>> tmishima >>>>>>>> >>>>>>>>> Hmmm...looks like we aren't getting your allocation. Can you > rerun >>>>> and >>>>>>>> add -mca ras_base_verbose 50? >>>>>>>>> >>>>>>>>> On Nov 12, 2013, at 11:30 PM, tmish...@jcity.maeda.co.jp wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Hi Ralph, >>>>>>>>>> >>>>>>>>>> Here is the output of "-mca plm_base_verbose 5". >>>>>>>>>> >>>>>>>>>> [node08.cluster:23573] mca:base:select:( plm) Querying > component >>>>> [rsh] >>>>>>>>>> [node08.cluster:23573] [[INVALID],INVALID] plm:rsh_lookup on >>>>>>>>>> agent /usr/bin/rsh path NULL >>>>>>>>>> [node08.cluster:23573] mca:base:select:( plm) Query of > component >>>>> [rsh] >>>>>>>> set >>>>>>>>>> priority to 10 >>>>>>>>>> [node08.cluster:23573] mca:base:select:( plm) Querying > component >>>>>>>> [slurm] >>>>>>>>>> [node08.cluster:23573] mca:base:select:( plm) Skipping > component >>>>>>>> [slurm]. >>>>>>>>>> Query failed to return a module >>>>>>>>>> [node08.cluster:23573] mca:base:select:( plm) Querying > component >>>>> [tm] >>>>>>>>>> [node08.cluster:23573] mca:base:select:( plm) Query of > component >>>>> [tm] >>>>>>>> set >>>>>>>>>> priority to 75 >>>>>>>>>> [node08.cluster:23573] mca:base:select:( plm) Selected > component >>>>> [tm] >>>>>>>>>> [node08.cluster:23573] plm:base:set_hnp_name: initial bias 23573 >>>>>>>> nodename >>>>>>>>>> hash 85176670 >>>>>>>>>> [node08.cluster:23573] plm:base:set_hnp_name: final jobfam 59480 >>>>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:receive start comm >>>>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_job >>>>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm >>>>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm creating >>> map >>>>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm only HNP > in >>>>>>>>>> allocation >>>>>>>>>> >>>>>>>> >>>>> >>> > -------------------------------------------------------------------------- >>>>>>>>>> All nodes which are allocated for this job are already filled. >>>>>>>>>> >>>>>>>> >>>>> >>> > -------------------------------------------------------------------------- >>>>>>>>>> >>>>>>>>>> Here, openmpi's configuration is as follows: >>>>>>>>>> >>>>>>>>>> ./configure \ >>>>>>>>>> --prefix=/home/mishima/opt/mpi/openmpi-1.7.4a1-pgi13.10 \ >>>>>>>>>> --with-tm \ >>>>>>>>>> --with-verbs \ >>>>>>>>>> --disable-ipv6 \ >>>>>>>>>> --disable-vt \ >>>>>>>>>> --enable-debug \ >>>>>>>>>> CC=pgcc CFLAGS="-tp k8-64e" \ >>>>>>>>>> CXX=pgCC CXXFLAGS="-tp k8-64e" \ >>>>>>>>>> F77=pgfortran FFLAGS="-tp k8-64e" \ >>>>>>>>>> FC=pgfortran FCFLAGS="-tp k8-64e" >>>>>>>>>> >>>>>>>>>>> Hi Ralph, >>>>>>>>>>> >>>>>>>>>>> Okey, I can help you. Please give me some time to report the >>>>> output. >>>>>>>>>>> >>>>>>>>>>> Tetsuya Mishima >>>>>>>>>>> >>>>>>>>>>>> I can try, but I have no way of testing Torque any more - so > all >>> I >>>>>>>> can >>>>>>>>>> do >>>>>>>>>>> is a code review. If you can build --enable-debug and add -mca >>>>>>>>>>> plm_base_verbose 5 to your cmd line, I'd appreciate seeing the >>>>>>>>>>>> output. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Nov 12, 2013, at 9:58 PM, tmish...@jcity.maeda.co.jp wrote: >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Hi Ralph, >>>>>>>>>>>>> >>>>>>>>>>>>> Thank you for your quick response. >>>>>>>>>>>>> >>>>>>>>>>>>> I'd like to report one more regressive issue about Torque >>> support >>>>> of >>>>>>>>>>>>> openmpi-1.7.4a1r29646, which might be related to "#3893: LAMA >>>>> mapper >>>>>>>>>>>>> has problems" I reported a few days ago. >>>>>>>>>>>>> >>>>>>>>>>>>> The script below does not work with openmpi-1.7.4a1r29646, >>>>>>>>>>>>> although it worked with openmpi-1.7.3 as I told you before. >>>>>>>>>>>>> >>>>>>>>>>>>> #!/bin/sh >>>>>>>>>>>>> #PBS -l nodes=node08:ppn=8 >>>>>>>>>>>>> export OMP_NUM_THREADS=1 >>>>>>>>>>>>> cd $PBS_O_WORKDIR >>>>>>>>>>>>> cp $PBS_NODEFILE pbs_hosts >>>>>>>>>>>>> NPROCS=`wc -l < pbs_hosts` >>>>>>>>>>>>> mpirun -machinefile pbs_hosts -np ${NPROCS} -report-bindings >>>>>>>> -bind-to >>>>>>>>>>> core >>>>>>>>>>>>> Myprog >>>>>>>>>>>>> >>>>>>>>>>>>> If I drop "-machinefile pbs_hosts -np ${NPROCS} ", then it >>> works >>>>>>>>>> fine. >>>>>>>>>>>>> Since this happens without lama request, I guess it's not the >>>>>>>> problem >>>>>>>>>>>>> in lama itself. Anyway, please look into this issue as well. >>>>>>>>>>>>> >>>>>>>>>>>>> Regards, >>>>>>>>>>>>> Tetsuya Mishima >>>>>>>>>>>>> >>>>>>>>>>>>>> Done - thanks! >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Nov 12, 2013, at 7:35 PM, tmish...@jcity.maeda.co.jp > wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Dear openmpi developers, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I got a segmentation fault in traial use of >>>>> openmpi-1.7.4a1r29646 >>>>>>>>>>> built >>>>>>>>>>>>> by >>>>>>>>>>>>>>> PGI13.10 as shown below: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> [mishima@manage testbed-openmpi-1.7.3]$ mpirun -np 4 >>>>>>>> -cpus-per-proc >>>>>>>>>> 2 >>>>>>>>>>>>>>> -report-bindings mPre >>>>>>>>>>>>>>> [manage.cluster:23082] MCW rank 2 bound to socket 0[core 4 >>> [hwt >>>>>>>> 0]], >>>>>>>>>>>>> socket >>>>>>>>>>>>>>> 0[core 5[hwt 0]]: [././././B/B][./././././.] >>>>>>>>>>>>>>> [manage.cluster:23082] MCW rank 3 bound to socket 1[core 6 >>> [hwt >>>>>>>> 0]], >>>>>>>>>>>>> socket >>>>>>>>>>>>>>> 1[core 7[hwt 0]]: [./././././.][B/B/./././.] >>>>>>>>>>>>>>> [manage.cluster:23082] MCW rank 0 bound to socket 0[core 0 >>> [hwt >>>>>>>> 0]], >>>>>>>>>>>>> socket >>>>>>>>>>>>>>> 0[core 1[hwt 0]]: [B/B/./././.][./././././.] >>>>>>>>>>>>>>> [manage.cluster:23082] MCW rank 1 bound to socket 0[core 2 >>> [hwt >>>>>>>> 0]], >>>>>>>>>>>>> socket >>>>>>>>>>>>>>> 0[core 3[hwt 0]]: [././B/B/./.][./././././.] >>>>>>>>>>>>>>> [manage:23082] *** Process received signal *** >>>>>>>>>>>>>>> [manage:23082] Signal: Segmentation fault (11) >>>>>>>>>>>>>>> [manage:23082] Signal code: Address not mapped (1) >>>>>>>>>>>>>>> [manage:23082] Failing at address: 0x34 >>>>>>>>>>>>>>> [manage:23082] *** End of error message *** >>>>>>>>>>>>>>> Segmentation fault (core dumped) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> [mishima@manage testbed-openmpi-1.7.3]$ gdb mpirun > core.23082 >>>>>>>>>>>>>>> GNU gdb (GDB) CentOS (7.0.1-42.el5.centos.1) >>>>>>>>>>>>>>> Copyright (C) 2009 Free Software Foundation, Inc. >>>>>>>>>>>>>>> ... >>>>>>>>>>>>>>> Core was generated by `mpirun -np 4 -cpus-per-proc 2 >>>>>>>>>> -report-bindings >>>>>>>>>>>>>>> mPre'. >>>>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault. >>>>>>>>>>>>>>> #0 0x00002b5f861c9c4f in recv_connect >>> (mod=0x5f861ca20b00007f, >>>>>>>>>>>>> sd=32767, >>>>>>>>>>>>>>> hdr=0x1ca20b00007fff25) at ./oob_tcp.c:631 >>>>>>>>>>>>>>> 631 peer = OBJ_NEW(mca_oob_tcp_peer_t); >>>>>>>>>>>>>>> (gdb) where >>>>>>>>>>>>>>> #0 0x00002b5f861c9c4f in recv_connect >>> (mod=0x5f861ca20b00007f, >>>>>>>>>>>>> sd=32767, >>>>>>>>>>>>>>> hdr=0x1ca20b00007fff25) at ./oob_tcp.c:631 >>>>>>>>>>>>>>> #1 0x00002b5f861ca20b in recv_handler (sd=1778385023, >>>>>>>> flags=32767, >>>>>>>>>>>>>>> cbdata=0x8eb06a00007fff25) at ./oob_tcp.c:760 >>>>>>>>>>>>>>> #2 0x00002b5f848eb06a in event_process_active_single_queue >>>>>>>>>>>>>>> (base=0x5f848eb27000007f, activeq=0x848eb27000007fff) >>>>>>>>>>>>>>> at ./event.c:1366 >>>>>>>>>>>>>>> #3 0x00002b5f848eb270 in event_process_active >>>>>>>>>>>>> (base=0x5f848eb84900007f) >>>>>>>>>>>>>>> at ./event.c:1435 >>>>>>>>>>>>>>> #4 0x00002b5f848eb849 in opal_libevent2021_event_base_loop >>>>>>>>>>>>>>> (base=0x4077a000007f, flags=32767) at ./event.c:1645 >>>>>>>>>>>>>>> #5 0x00000000004077a0 in orterun (argc=7, >>> argv=0x7fff25bbd4a8) >>>>>>>>>>>>>>> at ./orterun.c:1030 >>>>>>>>>>>>>>> #6 0x00000000004067fb in main (argc=7, > argv=0x7fff25bbd4a8) >>>>>>>>>>>>> at ./main.c:13 >>>>>>>>>>>>>>> (gdb) quit >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> The line 627 in orte/mca/oob/tcp/oob_tcp.c is apparently >>>>>>>>>> unnecessary, >>>>>>>>>>>>> which >>>>>>>>>>>>>>> causes the segfault. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 624 /* lookup the corresponding process > */>>>>>>>>>>>>> 625 peer = mca_oob_tcp_peer_lookup(mod, &hdr->origin); >>>>>>>>>>>>>>> 626 if (NULL == peer) { >>>>>>>>>>>>>>> 627 ui64 = (uint64_t*)(&peer->name); >>>>>>>>>>>>>>> 628 opal_output_verbose(OOB_TCP_DEBUG_CONNECT, >>>>>>>>>>>>>>> orte_oob_base_framework.framework_output, >>>>>>>>>>>>>>> 629 "%s >>> mca_oob_tcp_recv_connect: >>>>>>>>>>>>>>> connection from new peer", >>>>>>>>>>>>>>> 630 ORTE_NAME_PRINT >>>>>>>>>>> (ORTE_PROC_MY_NAME)); >>>>>>>>>>>>>>> 631 peer = OBJ_NEW(mca_oob_tcp_peer_t); >>>>>>>>>>>>>>> 632 peer->mod = mod; >>>>>>>>>>>>>>> 633 peer->name = hdr->origin; >>>>>>>>>>>>>>> 634 peer->state = MCA_OOB_TCP_ACCEPTING; >>>>>>>>>>>>>>> 635 ui64 = (uint64_t*)(&peer->name); >>>>>>>>>>>>>>> 636 if (OPAL_SUCCESS != >>>>> opal_hash_table_set_value_uint64 >>>>>>>>>>>>> (&mod-> >>>>>>>>>>>>>>> peers, (*ui64), peer)) { >>>>>>>>>>>>>>> 637 OBJ_RELEASE(peer); >>>>>>>>>>>>>>> 638 return; >>>>>>>>>>>>>>> 639 } >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Please fix this mistake in the next release. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>> Tetsuya Mishima >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>>> >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing list >>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> users mailing list >>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> users mailing list >>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> us...@open-mpi.org >>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> us...@open-mpi.org >>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> us...@open-mpi.org >>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list>> us...@open-mpi.org >>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users