Sorry for cross-post.

Nodefile is very simple which consists of 8 lines:

node08
node08
node08
node08
node08
node08
node08
node08

Therefore, NPROCS=8

My aim is to modify the allocation as you pointed out. According to Openmpi
FAQ,
proper subset of the hosts allocated to the Torque / PBS Pro job should be
allowed.

tmishima

> Please - can you answer my question on script2? What is the value of
NPROCS?
>
> Why would you want to do it this way? Are you planning to modify the
allocation?? That generally is a bad idea as it can confuse the system
>
>
> On Nov 13, 2013, at 5:55 PM, tmish...@jcity.maeda.co.jp wrote:
>
> >
> >
> > Since what I really want is to run script2 correctly, please let us
> > concentrate script2.
> >
> > I'm not an expert of the inside of openmpi. What I can do is just
> > obsabation
> > from the outside. I doubt these lines are strange, especially the last
one.
> >
> > [node08.cluster:26952] mca:rmaps:rr: mapping job [56581,1]
> > [node08.cluster:26952] [[56581,0],0] Starting with 1 nodes in list
> > [node08.cluster:26952] [[56581,0],0] Filtering thru apps
> > [node08.cluster:26952] [[56581,0],0] Retained 1 nodes in list
> > [node08.cluster:26952] [[56581,0],0] Removing node node08 slots 0 inuse
0
> >
> > These lines come from this part of orte_rmaps_base_get_target_nodes
> > in rmaps_base_support_fns.c:
> >
> >        } else if (node->slots <= node->slots_inuse &&
> >                   (ORTE_MAPPING_NO_OVERSUBSCRIBE &
> > ORTE_GET_MAPPING_DIRECTIVE(policy))) {
> >            /* remove the node as fully used */
> >            OPAL_OUTPUT_VERBOSE((5,
> > orte_rmaps_base_framework.framework_output,
> >                                 "%s Removing node %s slots %d inuse
%d",
> >                                 ORTE_NAME_PRINT(ORTE_PROC_MY_NAME),
> >                                 node->name, node->slots, node->
> > slots_inuse));
> >            opal_list_remove_item(allocated_nodes, item);
> >            OBJ_RELEASE(item);  /* "un-retain" it */
> >
> > I wonder why node->slots and node->slots_inuse is 0, which I can read
> > from the above line "Removing node node08 slots 0 inuse 0".
> >
> > Or I'm not sure but
> >   "else if (node->slots <= node->slots_inuse &&" should be
> >   "else if (node->slots < node->slots_inuse &&" ?
> >
> > tmishima
> >
> >> On Nov 13, 2013, at 4:43 PM, tmish...@jcity.maeda.co.jp wrote:
> >>
> >>>
> >>>
> >>> Yes, the node08 has 8 slots but the process I run is also 8.
> >>>
> >>> #PBS -l nodes=node08:ppn=8
> >>>
> >>> Therefore, I think it should allow this allocation. Is that right?
> >>
> >> Correct
> >>
> >>>
> >>> My question is why scritp1 works and script2 does not. They are
> >>> almost same.
> >>>
> >>> #PBS -l nodes=node08:ppn=8
> >>> export OMP_NUM_THREADS=1
> >>> cd $PBS_O_WORKDIR
> >>> cp $PBS_NODEFILE pbs_hosts
> >>> NPROCS=`wc -l < pbs_hosts`
> >>>
> >>> #SCRITP1
> >>> mpirun -report-bindings -bind-to core Myprog
> >>>
> >>> #SCRIPT2
> >>> mpirun -machinefile pbs_hosts -np ${NPROCS} -report-bindings -bind-to
> > core
> >>
> >> This version is not only reading the PBS allocation, but also invoking
> > the hostfile filter on top of it. Different code path. I'll take a look
-
> > it should still match up assuming NPROCS=8. Any
> >> possibility that it is a different number? I don't recall, but isn't
> > there some extra lines in the nodefile - e.g., comments?
> >>
> >>
> >>> Myprog
> >>>
> >>> tmishima
> >>>
> >>>> I guess here's my confusion. If you are using only one node, and
that
> >>> node has 8 allocated slots, then we will not allow you to run more
than
> > 8
> >>> processes on that node unless you specifically provide
> >>>> the --oversubscribe flag. This is because you are operating in a
> > managed
> >>> environment (in this case, under Torque), and so we treat the
> > allocation as
> >>> "mandatory" by default.
> >>>>
> >>>> I suspect that is the issue here, in which case the system is
behaving
> > as
> >>> it should.
> >>>>
> >>>> Is the above accurate?
> >>>>
> >>>>
> >>>> On Nov 13, 2013, at 4:11 PM, Ralph Castain <r...@open-mpi.org> wrote:
> >>>>
> >>>>> It has nothing to do with LAMA as you aren't using that mapper.
> >>>>>
> >>>>> How many nodes are in this allocation?
> >>>>>
> >>>>> On Nov 13, 2013, at 4:06 PM, tmish...@jcity.maeda.co.jp wrote:
> >>>>>
> >>>>>>
> >>>>>>
> >>>>>> Hi Ralph, this is an additional information.
> >>>>>>
> >>>>>> Here is the main part of output by adding "-mca rmaps_base_verbose
> >>> 50".
> >>>>>>
> >>>>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm
> >>>>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm creating
map
> >>>>>> [node08.cluster:26952] [[56581,0],0] plm:base:setup_vm only HNP in
> >>>>>> allocation
> >>>>>> [node08.cluster:26952] mca:rmaps: mapping job [56581,1]
> >>>>>> [node08.cluster:26952] mca:rmaps: creating new map for job
[56581,1]
> >>>>>> [node08.cluster:26952] mca:rmaps:ppr: job [56581,1] not using ppr
> >>> mapper
> >>>>>> [node08.cluster:26952] [[56581,0],0] rmaps:seq mapping job
[56581,1]
> >>>>>> [node08.cluster:26952] mca:rmaps:seq: job [56581,1] not using seq
> >>> mapper
> >>>>>> [node08.cluster:26952] mca:rmaps:resilient: cannot perform initial
> > map
> >>> of
> >>>>>> job [56581,1] - no fault groups
> >>>>>> [node08.cluster:26952] mca:rmaps:mindist: job [56581,1] not using
> >>> mindist
> >>>>>> mapper
> >>>>>> [node08.cluster:26952] mca:rmaps:rr: mapping job [56581,1]
> >>>>>> [node08.cluster:26952] [[56581,0],0] Starting with 1 nodes in list
> >>>>>> [node08.cluster:26952] [[56581,0],0] Filtering thru apps
> >>>>>> [node08.cluster:26952] [[56581,0],0] Retained 1 nodes in list
> >>>>>> [node08.cluster:26952] [[56581,0],0] Removing node node08 slots 0
> >>> inuse 0
> >>>>>>
> >>>>>> From this result, I guess it's related to oversubscribe.
> >>>>>> So I added "-oversubscribe" and rerun, then it worked well as show
> >>> below:
> >>>>>>
> >>>>>> [node08.cluster:27019] [[56774,0],0] Starting with 1 nodes in list
> >>>>>> [node08.cluster:27019] [[56774,0],0] Filtering thru apps
> >>>>>> [node08.cluster:27019] [[56774,0],0] Retained 1 nodes in list
> >>>>>> [node08.cluster:27019] AVAILABLE NODES FOR MAPPING:
> >>>>>> [node08.cluster:27019]     node: node08 daemon: 0
> >>>>>> [node08.cluster:27019] [[56774,0],0] Starting bookmark at node
> > node08
> >>>>>> [node08.cluster:27019] [[56774,0],0] Starting at node node08
> >>>>>> [node08.cluster:27019] mca:rmaps:rr: mapping by slot for job
> > [56774,1]
> >>>>>> slots 1 num_procs 8
> >>>>>> [node08.cluster:27019] mca:rmaps:rr:slot working node node08
> >>>>>> [node08.cluster:27019] mca:rmaps:rr:slot node node08 is full -
> >>> skipping
> >>>>>> [node08.cluster:27019] mca:rmaps:rr:slot job [56774,1] is
> >>> oversubscribed -
> >>>>>> performing second pass
> >>>>>> [node08.cluster:27019] mca:rmaps:rr:slot working node node08
> >>>>>> [node08.cluster:27019] mca:rmaps:rr:slot adding up to 8 procs to
> > node
> >>>>>> node08
> >>>>>> [node08.cluster:27019] mca:rmaps:base: computing vpids by slot for
> > job
> >>>>>> [56774,1]
> >>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 0 to node
> > node08
> >>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 1 to node
> > node08
> >>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 2 to node
> > node08
> >>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 3 to node
> > node08
> >>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 4 to node
> > node08
> >>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 5 to node
> > node08
> >>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 6 to node
> > node08
> >>>>>> [node08.cluster:27019] mca:rmaps:base: assigning rank 7 to node
> > node08
> >>>>>>
> >>>>>> I think something is wrong with treatment of oversubscription,
which
> >>> might
> >>>>>> be
> >>>>>> related to "#3893: LAMA mapper has problems"
> >>>>>>
> >>>>>> tmishima
> >>>>>>
> >>>>>>> Hmmm...looks like we aren't getting your allocation. Can you
rerun
> >>> and
> >>>>>> add -mca ras_base_verbose 50?
> >>>>>>>
> >>>>>>> On Nov 12, 2013, at 11:30 PM, tmish...@jcity.maeda.co.jp wrote:
> >>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> Hi Ralph,
> >>>>>>>>
> >>>>>>>> Here is the output of "-mca plm_base_verbose 5".
> >>>>>>>>
> >>>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Querying
component
> >>> [rsh]
> >>>>>>>> [node08.cluster:23573] [[INVALID],INVALID] plm:rsh_lookup on
> >>>>>>>> agent /usr/bin/rsh path NULL
> >>>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Query of
component
> >>> [rsh]
> >>>>>> set
> >>>>>>>> priority to 10
> >>>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Querying
component
> >>>>>> [slurm]
> >>>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Skipping
component
> >>>>>> [slurm].
> >>>>>>>> Query failed to return a module
> >>>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Querying
component
> >>> [tm]
> >>>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Query of
component
> >>> [tm]
> >>>>>> set
> >>>>>>>> priority to 75
> >>>>>>>> [node08.cluster:23573] mca:base:select:(  plm) Selected
component
> >>> [tm]
> >>>>>>>> [node08.cluster:23573] plm:base:set_hnp_name: initial bias 23573
> >>>>>> nodename
> >>>>>>>> hash 85176670
> >>>>>>>> [node08.cluster:23573] plm:base:set_hnp_name: final jobfam 59480
> >>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:receive start comm
> >>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_job
> >>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm
> >>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm creating
> > map
> >>>>>>>> [node08.cluster:23573] [[59480,0],0] plm:base:setup_vm only HNP
in
> >>>>>>>> allocation
> >>>>>>>>
> >>>>>>
> >>>
> >
--------------------------------------------------------------------------
> >>>>>>>> All nodes which are allocated for this job are already filled.
> >>>>>>>>
> >>>>>>
> >>>
> >
--------------------------------------------------------------------------
> >>>>>>>>
> >>>>>>>> Here, openmpi's configuration is as follows:
> >>>>>>>>
> >>>>>>>> ./configure \
> >>>>>>>> --prefix=/home/mishima/opt/mpi/openmpi-1.7.4a1-pgi13.10 \
> >>>>>>>> --with-tm \
> >>>>>>>> --with-verbs \
> >>>>>>>> --disable-ipv6 \
> >>>>>>>> --disable-vt \
> >>>>>>>> --enable-debug \
> >>>>>>>> CC=pgcc CFLAGS="-tp k8-64e" \
> >>>>>>>> CXX=pgCC CXXFLAGS="-tp k8-64e" \
> >>>>>>>> F77=pgfortran FFLAGS="-tp k8-64e" \
> >>>>>>>> FC=pgfortran FCFLAGS="-tp k8-64e"
> >>>>>>>>
> >>>>>>>>> Hi Ralph,
> >>>>>>>>>
> >>>>>>>>> Okey, I can help you. Please give me some time to report the
> >>> output.
> >>>>>>>>>
> >>>>>>>>> Tetsuya Mishima
> >>>>>>>>>
> >>>>>>>>>> I can try, but I have no way of testing Torque any more - so
all
> > I
> >>>>>> can
> >>>>>>>> do
> >>>>>>>>> is a code review. If you can build --enable-debug and add -mca
> >>>>>>>>> plm_base_verbose 5 to your cmd line, I'd appreciate seeing the
> >>>>>>>>>> output.
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> On Nov 12, 2013, at 9:58 PM, tmish...@jcity.maeda.co.jp wrote:
> >>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> Hi Ralph,
> >>>>>>>>>>>
> >>>>>>>>>>> Thank you for your quick response.
> >>>>>>>>>>>
> >>>>>>>>>>> I'd like to report one more regressive issue about Torque
> > support
> >>> of
> >>>>>>>>>>> openmpi-1.7.4a1r29646, which might be related to "#3893: LAMA
> >>> mapper
> >>>>>>>>>>> has problems" I reported a few days ago.
> >>>>>>>>>>>
> >>>>>>>>>>> The script below does not work with openmpi-1.7.4a1r29646,
> >>>>>>>>>>> although it worked with openmpi-1.7.3 as I told you before.
> >>>>>>>>>>>
> >>>>>>>>>>> #!/bin/sh
> >>>>>>>>>>> #PBS -l nodes=node08:ppn=8
> >>>>>>>>>>> export OMP_NUM_THREADS=1
> >>>>>>>>>>> cd $PBS_O_WORKDIR
> >>>>>>>>>>> cp $PBS_NODEFILE pbs_hosts
> >>>>>>>>>>> NPROCS=`wc -l < pbs_hosts`
> >>>>>>>>>>> mpirun -machinefile pbs_hosts -np ${NPROCS} -report-bindings
> >>>>>> -bind-to
> >>>>>>>>> core
> >>>>>>>>>>> Myprog
> >>>>>>>>>>>
> >>>>>>>>>>> If I drop "-machinefile pbs_hosts -np ${NPROCS} ", then it
> > works
> >>>>>>>> fine.
> >>>>>>>>>>> Since this happens without lama request, I guess it's not the
> >>>>>> problem
> >>>>>>>>>>> in lama itself. Anyway, please look into this issue as well.
> >>>>>>>>>>>
> >>>>>>>>>>> Regards,
> >>>>>>>>>>> Tetsuya Mishima
> >>>>>>>>>>>
> >>>>>>>>>>>> Done - thanks!
> >>>>>>>>>>>>
> >>>>>>>>>>>> On Nov 12, 2013, at 7:35 PM, tmish...@jcity.maeda.co.jp
wrote:
> >>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Dear openmpi developers,
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> I got a segmentation fault in traial use of
> >>> openmpi-1.7.4a1r29646
> >>>>>>>>> built
> >>>>>>>>>>> by
> >>>>>>>>>>>>> PGI13.10 as shown below:
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> [mishima@manage testbed-openmpi-1.7.3]$ mpirun -np 4
> >>>>>> -cpus-per-proc
> >>>>>>>> 2
> >>>>>>>>>>>>> -report-bindings mPre
> >>>>>>>>>>>>> [manage.cluster:23082] MCW rank 2 bound to socket 0[core 4
> > [hwt
> >>>>>> 0]],
> >>>>>>>>>>> socket
> >>>>>>>>>>>>> 0[core 5[hwt 0]]: [././././B/B][./././././.]
> >>>>>>>>>>>>> [manage.cluster:23082] MCW rank 3 bound to socket 1[core 6
> > [hwt
> >>>>>> 0]],
> >>>>>>>>>>> socket
> >>>>>>>>>>>>> 1[core 7[hwt 0]]: [./././././.][B/B/./././.]
> >>>>>>>>>>>>> [manage.cluster:23082] MCW rank 0 bound to socket 0[core 0
> > [hwt
> >>>>>> 0]],
> >>>>>>>>>>> socket
> >>>>>>>>>>>>> 0[core 1[hwt 0]]: [B/B/./././.][./././././.]
> >>>>>>>>>>>>> [manage.cluster:23082] MCW rank 1 bound to socket 0[core 2
> > [hwt
> >>>>>> 0]],
> >>>>>>>>>>> socket
> >>>>>>>>>>>>> 0[core 3[hwt 0]]: [././B/B/./.][./././././.]
> >>>>>>>>>>>>> [manage:23082] *** Process received signal ***
> >>>>>>>>>>>>> [manage:23082] Signal: Segmentation fault (11)
> >>>>>>>>>>>>> [manage:23082] Signal code: Address not mapped (1)
> >>>>>>>>>>>>> [manage:23082] Failing at address: 0x34
> >>>>>>>>>>>>> [manage:23082] *** End of error message ***
> >>>>>>>>>>>>> Segmentation fault (core dumped)
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> [mishima@manage testbed-openmpi-1.7.3]$ gdb mpirun
core.23082
> >>>>>>>>>>>>> GNU gdb (GDB) CentOS (7.0.1-42.el5.centos.1)
> >>>>>>>>>>>>> Copyright (C) 2009 Free Software Foundation, Inc.
> >>>>>>>>>>>>> ...
> >>>>>>>>>>>>> Core was generated by `mpirun -np 4 -cpus-per-proc 2
> >>>>>>>> -report-bindings
> >>>>>>>>>>>>> mPre'.
> >>>>>>>>>>>>> Program terminated with signal 11, Segmentation fault.
> >>>>>>>>>>>>> #0  0x00002b5f861c9c4f in recv_connect
> > (mod=0x5f861ca20b00007f,
> >>>>>>>>>>> sd=32767,
> >>>>>>>>>>>>> hdr=0x1ca20b00007fff25) at ./oob_tcp.c:631
> >>>>>>>>>>>>> 631             peer = OBJ_NEW(mca_oob_tcp_peer_t);
> >>>>>>>>>>>>> (gdb) where
> >>>>>>>>>>>>> #0  0x00002b5f861c9c4f in recv_connect
> > (mod=0x5f861ca20b00007f,
> >>>>>>>>>>> sd=32767,
> >>>>>>>>>>>>> hdr=0x1ca20b00007fff25) at ./oob_tcp.c:631
> >>>>>>>>>>>>> #1  0x00002b5f861ca20b in recv_handler (sd=1778385023,
> >>>>>> flags=32767,
> >>>>>>>>>>>>> cbdata=0x8eb06a00007fff25) at ./oob_tcp.c:760
> >>>>>>>>>>>>> #2  0x00002b5f848eb06a in event_process_active_single_queue
> >>>>>>>>>>>>> (base=0x5f848eb27000007f, activeq=0x848eb27000007fff)
> >>>>>>>>>>>>> at ./event.c:1366
> >>>>>>>>>>>>> #3  0x00002b5f848eb270 in event_process_active
> >>>>>>>>>>> (base=0x5f848eb84900007f)
> >>>>>>>>>>>>> at ./event.c:1435
> >>>>>>>>>>>>> #4  0x00002b5f848eb849 in opal_libevent2021_event_base_loop
> >>>>>>>>>>>>> (base=0x4077a000007f, flags=32767) at ./event.c:1645
> >>>>>>>>>>>>> #5  0x00000000004077a0 in orterun (argc=7,
> > argv=0x7fff25bbd4a8)
> >>>>>>>>>>>>> at ./orterun.c:1030
> >>>>>>>>>>>>> #6  0x00000000004067fb in main (argc=7,
argv=0x7fff25bbd4a8)
> >>>>>>>>>>> at ./main.c:13
> >>>>>>>>>>>>> (gdb) quit
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> The line 627 in orte/mca/oob/tcp/oob_tcp.c is apparently
> >>>>>>>> unnecessary,
> >>>>>>>>>>> which
> >>>>>>>>>>>>> causes the segfault.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> 624      /* lookup the corresponding process
*/>>>>>>>>>>>>> 625      peer = mca_oob_tcp_peer_lookup(mod, &hdr->origin);
> >>>>>>>>>>>>> 626      if (NULL == peer) {
> >>>>>>>>>>>>> 627          ui64 = (uint64_t*)(&peer->name);
> >>>>>>>>>>>>> 628          opal_output_verbose(OOB_TCP_DEBUG_CONNECT,
> >>>>>>>>>>>>> orte_oob_base_framework.framework_output,
> >>>>>>>>>>>>> 629                              "%s
> > mca_oob_tcp_recv_connect:
> >>>>>>>>>>>>> connection from new peer",
> >>>>>>>>>>>>> 630                              ORTE_NAME_PRINT
> >>>>>>>>> (ORTE_PROC_MY_NAME));
> >>>>>>>>>>>>> 631          peer = OBJ_NEW(mca_oob_tcp_peer_t);
> >>>>>>>>>>>>> 632          peer->mod = mod;
> >>>>>>>>>>>>> 633          peer->name = hdr->origin;
> >>>>>>>>>>>>> 634          peer->state = MCA_OOB_TCP_ACCEPTING;
> >>>>>>>>>>>>> 635          ui64 = (uint64_t*)(&peer->name);
> >>>>>>>>>>>>> 636          if (OPAL_SUCCESS !=
> >>> opal_hash_table_set_value_uint64
> >>>>>>>>>>> (&mod->
> >>>>>>>>>>>>> peers, (*ui64), peer)) {
> >>>>>>>>>>>>> 637              OBJ_RELEASE(peer);
> >>>>>>>>>>>>> 638              return;
> >>>>>>>>>>>>> 639          }
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Please fix this mistake in the next release.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Regards,
> >>>>>>>>>>>>> Tetsuya Mishima
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> _______________________________________________
> >>>>>>>>>>>>> users mailing list
> >>>>>>>>>>>>> us...@open-mpi.org
> >>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>>>>>
> >>>>>>>>>>>> _______________________________________________
> >>>>>>>>>>>> users mailing list
> >>>>>>>>>>>> us...@open-mpi.org
> >>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>>>>
> >>>>>>>>>>> _______________________________________________
> >>>>>>>>>>> users mailing list
> >>>>>>>>>>> us...@open-mpi.org
> >>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>>>
> >>>>>>>>>> _______________________________________________
> >>>>>>>>>> users mailing list
> >>>>>>>>>> us...@open-mpi.org
> >>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>>
> >>>>>>>>> _______________________________________________
> >>>>>>>>> users mailing list
> >>>>>>>>> us...@open-mpi.org
> >>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>>
> >>>>>>>> _______________________________________________
> >>>>>>>> users mailing list
> >>>>>>>> us...@open-mpi.org
> >>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>>
> >>>>>>> _______________________________________________
> >>>>>>> users mailing list
> >>>>>>> us...@open-mpi.org
> >>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>>
> >>>>>> _______________________________________________
> >>>>>> users mailing list
> >>>>>> us...@open-mpi.org
> >>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>>>
> >>>>
> >>>> _______________________________________________
> >>>> users mailing list>> us...@open-mpi.org
> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>
> >>> _______________________________________________
> >>> users mailing list
> >>> us...@open-mpi.org
> >>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>
> >> _______________________________________________
> >> users mailing list
> >> us...@open-mpi.org
> >> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to