Okay, I'll dig into it - must be a bug in my code.

Sorry for the problem! Thanks for patience in tracking it down...
Ralph

On Wed, Jul 15, 2009 at 7:28 AM, Lenny Verkhovsky <
lenny.verkhov...@gmail.com> wrote:

> Thanks, Ralph,
> I guess your guess was correct, here is the display map.
>
>
> $cat rankfile
> rank 0=+n1 slot=0
> rank 1=+n0 slot=0
> $cat appfile
> -np 1 -host witch1 ./hello_world
> -np 1 -host witch2 ./hello_world
> $mpirun -np 2 -rf rankfile --display-allocation  -app appfile
>
> ======================   ALLOCATED NODES   ======================
>
>  Data for node: Name: dellix7   Num slots: 0    Max slots: 0
>  Data for node: Name: witch1    Num slots: 1    Max slots: 0
>  Data for node: Name: witch2    Num slots: 1    Max slots: 0
>
> =================================================================
>
> --------------------------------------------------------------------------
> Rankfile claimed host +n1 by index that is bigger than number of allocated
> hosts.
>
>
> On Wed, Jul 15, 2009 at 4:10 PM, Ralph Castain <r...@open-mpi.org> wrote:
>
>> What is supposed to happen is this:
>>
>> 1. each line of the appfile causes us to create a new app_context. We
>> store the provided -host info in that object.
>>
>> 2. when we create the "allocation", we cycle through -all- the
>> app_contexts and add -all- of their -host info into the list of allocated
>> nodes
>>
>> 3. when we get_target_nodes, we start with the entire list of allocated
>> nodes, and then use -host for that app_context to filter down to the hosts
>> allowed for that specific app_context
>>
>> So you should have to only provide -np 1 and 1 host on each line. My guess
>> is that the rankfile mapper isn't correctly behaving for multiple
>> app_contexts.
>>
>> Add --display-allocation to your mpirun cmd line for the "not working" cse
>> and let's see what mpirun thinks the total allocation is - I'll bet that
>> both nodes show up, which would tell us that my "guess" is correct. Then
>> I'll know what needs to be fixed.
>>
>> Thanks
>> Ralph
>>
>>
>>
>> On Wed, Jul 15, 2009 at 6:08 AM, Lenny Verkhovsky <
>> lenny.verkhov...@gmail.com> wrote:
>>
>>>  Same result.
>>> I still suspect that rankfile claims for node in small hostlist provided
>>> by line in the app file, and not from the hostlist provided by mpirun on HNP
>>> node.
>>> According to my suspections your proposal should not work(and it does
>>> not), since in appfile line I provide np=1, and 1 host, while rankfile tries
>>> to allocate all ranks (np=2).
>>>
>>> $orte/mca/rmaps/rank_file/rmaps_rank_file.c at line 338
>>>
>>>   if(ORTE_SUCCESS != (rc = orte_rmaps_base_get_target_nodes(&node_list,
>>> &num_slots, app,
>>>
>>> map->policy))) {
>>>
>>> node_list will be partial, according to app, and not full provided by
>>> mpirun cmd. If I didnt provide hostlist in the appfile line, mpirun uses
>>> local host and not hosts from the hostfile.
>>>
>>>
>>> Tell me if I am wrong by expecting the following behaivor
>>>
>>> I provide to mpirun NP, full_hostlist, full_rankfile, appfile
>>> I provide in appfile only partial NP and partial hostlist.
>>> and it works.
>>>
>>> Currently, in order to get it working I need to provide full hostlist in
>>> the appfile. Which is quit a problematic.
>>>
>>>
>>> $mpirun -np 2 -rf rankfile -app appfile
>>>
>>> --------------------------------------------------------------------------
>>> Rankfile claimed host +n1 by index that is bigger than number of
>>> allocated hosts.
>>>
>>> --------------------------------------------------------------------------
>>> [dellix7:17277] [[23928,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>> ../../../../../orte/mca/rmaps/rank_file/rmaps_rank_file.c at line 422
>>> [dellix7:17277] [[23928,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>> ../../../../orte/mca/rmaps/base/rmaps_base_map_job.c at line 85
>>> [dellix7:17277] [[23928,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>> ../../../../orte/mca/plm/base/plm_base_launch_support.c at line 103
>>> [dellix7:17277] [[23928,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>> ../../../../../orte/mca/plm/rsh/plm_rsh_module.c at line 1001
>>>
>>>
>>> Thanks
>>> Lenny.
>>>
>>>
>>> On Wed, Jul 15, 2009 at 2:02 PM, Ralph Castain <r...@open-mpi.org> wrote:
>>>
>>>> Try your "not working" example without the -H on the mpirun cmd line -
>>>> i.e.,, just use "mpirun -np 2 -rf rankfile -app appfile". Does that work?
>>>> Sorry to have to keep asking you to try things - I don't have a setup
>>>> here where I can test this as everything is RM managed.
>>>>
>>>>
>>>>  On Jul 15, 2009, at 12:09 AM, Lenny Verkhovsky wrote:
>>>>
>>>>
>>>> Thanks Ralph, after playing with prefixes it worked,
>>>>
>>>> I still have a problem running app file with rankfile, by providing full
>>>> hostlist in mpirun command and not in app file.
>>>> Is is planned behaviour, or it can be fixed ?
>>>>
>>>> See Working example:
>>>>
>>>> $cat rankfile
>>>> rank 0=+n1 slot=0
>>>> rank 1=+n0 slot=0
>>>> $cat appfile
>>>> -np 1 -H witch1,witch2  ./hello_world
>>>> -np 1 -H witch1,witch2 ./hello_world
>>>>
>>>> $mpirun -rf rankfile -app appfile
>>>> Hello world! I'm 1 of 2 on witch1
>>>> Hello world! I'm 0 of 2 on witch2
>>>>
>>>> See NOT working example:
>>>>
>>>> $cat appfile
>>>> -np 1 -H witch1 ./hello_world
>>>> -np 1 -H witch2 ./hello_world
>>>> $mpirun  -np 2 -H witch1,witch2 -rf rankfile -app appfile
>>>>
>>>> --------------------------------------------------------------------------
>>>> Rankfile claimed host +n1 by index that is bigger than number of
>>>> allocated hosts.
>>>>
>>>> --------------------------------------------------------------------------
>>>> [dellix7:16405] [[24080,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>>> ../../../../../orte/mca/rmaps/rank_file/rmaps_rank_file.c at line 422
>>>> [dellix7:16405] [[24080,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>>> ../../../../orte/mca/rmaps/base/rmaps_base_map_job.c at line 85
>>>> [dellix7:16405] [[24080,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>>> ../../../../orte/mca/plm/base/plm_base_launch_support.c at line 103
>>>> [dellix7:16405] [[24080,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>>> ../../../../../orte/mca/plm/rsh/plm_rsh_module.c at line 1001
>>>>
>>>>
>>>>
>>>> On Wed, Jul 15, 2009 at 6:58 AM, Ralph Castain <r...@open-mpi.org>wrote:
>>>>
>>>>> Took a deeper look into this, and I think that your first guess was
>>>>> correct.
>>>>> When we changed hostfile and -host to be per-app-context options, it
>>>>> became necessary for you to put that info in the appfile itself. So try
>>>>> adding it there. What you would need in your appfile is the following:
>>>>>
>>>>> -np 1 -H witch1 hostname
>>>>> -np 1 -H witch2 hostname
>>>>>
>>>>> That should get you what you want.
>>>>> Ralph
>>>>>
>>>>>  On Jul 14, 2009, at 10:29 AM, Lenny Verkhovsky wrote:
>>>>>
>>>>>  No,  it's not working as I expect , unless I expect somthing wrong .
>>>>> ( sorry for the long PATH, I needed to provide it )
>>>>>
>>>>> $LD_LIBRARY_PATH=/hpc/home/USERS/lennyb/work/svn/ompi/trunk/build_x86-64/install/lib/
>>>>> /hpc/home/USERS/lennyb/work/svn/ompi/trunk/build_x86-64/install/bin/mpirun
>>>>> -np 2 -H witch1,witch2 hostname
>>>>> witch1
>>>>> witch2
>>>>>
>>>>> $LD_LIBRARY_PATH=/hpc/home/USERS/lennyb/work/svn/ompi/trunk/build_x86-64/install/lib/
>>>>> /hpc/home/USERS/lennyb/work/svn/ompi/trunk/build_x86-64/install/bin/mpirun
>>>>> -np 2 -H witch1,witch2 -app appfile
>>>>> dellix7
>>>>> dellix7
>>>>>  $cat appfile
>>>>> -np 1 hostname
>>>>> -np 1 hostname
>>>>>
>>>>>
>>>>> On Tue, Jul 14, 2009 at 7:08 PM, Ralph Castain <r...@open-mpi.org>wrote:
>>>>>
>>>>>> Run it without the appfile, just putting the apps on the cmd line -
>>>>>> does it work right then?
>>>>>>
>>>>>>  On Jul 14, 2009, at 10:04 AM, Lenny Verkhovsky wrote:
>>>>>>
>>>>>>  additional info
>>>>>> I am running mpirun on hostA, and providing hostlist with hostB and
>>>>>> hostC.
>>>>>> I expect that each application would run on hostB and hostC, but I get
>>>>>> all of them running on hostA.
>>>>>>  dellix7$cat appfile
>>>>>> -np 1 hostname
>>>>>> -np 1 hostname
>>>>>>  dellix7$mpirun -np 2 -H witch1,witch2 -app appfile
>>>>>> dellix7
>>>>>> dellix7
>>>>>>  Thanks
>>>>>> Lenny.
>>>>>>
>>>>>> On Tue, Jul 14, 2009 at 4:59 PM, Ralph Castain <r...@open-mpi.org>wrote:
>>>>>>
>>>>>>> Strange - let me have a look at it later today. Probably something
>>>>>>> simple that another pair of eyes might spot.
>>>>>>>  On Jul 14, 2009, at 7:43 AM, Lenny Verkhovsky wrote:
>>>>>>>
>>>>>>>  Seems like connected problem:
>>>>>>> I can't use rankfile with app, even after all those fixes ( working
>>>>>>> with trunk 1.4a1r21657).
>>>>>>> This is my case :
>>>>>>>
>>>>>>> $cat rankfile
>>>>>>> rank 0=+n1 slot=0
>>>>>>> rank 1=+n0 slot=0
>>>>>>> $cat appfile
>>>>>>> -np 1 hostname
>>>>>>> -np 1 hostname
>>>>>>> $mpirun -np 2 -H witch1,witch2 -rf rankfile -app appfile
>>>>>>>
>>>>>>> --------------------------------------------------------------------------
>>>>>>> Rankfile claimed host +n1 by index that is bigger than number of
>>>>>>> allocated hosts.
>>>>>>>
>>>>>>> --------------------------------------------------------------------------
>>>>>>> [dellix7:13414] [[10851,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>>>>>> ../../../../../orte/mca/rmaps/rank_file/rmaps_rank_file.c at line 422
>>>>>>> [dellix7:13414] [[10851,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>>>>>> ../../../../orte/mca/rmaps/base/rmaps_base_map_job.c at line 85
>>>>>>> [dellix7:13414] [[10851,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>>>>>> ../../../../orte/mca/plm/base/plm_base_launch_support.c at line 103
>>>>>>> [dellix7:13414] [[10851,0],0] ORTE_ERROR_LOG: Bad parameter in file
>>>>>>> ../../../../../orte/mca/plm/rsh/plm_rsh_module.c at line 1001
>>>>>>>
>>>>>>>
>>>>>>> The problem is, that rankfile mapper tries to find an appropriate
>>>>>>> host in the partial ( and not full ) hostlist.
>>>>>>>
>>>>>>> Any suggestions how to fix it?
>>>>>>>
>>>>>>> Thanks
>>>>>>> Lenny.
>>>>>>>
>>>>>>> On Wed, May 13, 2009 at 1:55 AM, Ralph Castain <r...@open-mpi.org>wrote:
>>>>>>>
>>>>>>>> Okay, I fixed this today too....r21219
>>>>>>>>
>>>>>>>>
>>>>>>>> On May 11, 2009, at 11:27 PM, Anton Starikov wrote:
>>>>>>>>
>>>>>>>> Now there is another problem :)
>>>>>>>>>
>>>>>>>>> You can try oversubscribe node. At least by 1 task.
>>>>>>>>> If you hostfile and rank file limit you at N procs, you can ask
>>>>>>>>> mpirun for N+1 and it wil be not rejected.
>>>>>>>>> Although in reality there will be N tasks.
>>>>>>>>> So, if your hostfile limit is 4, then "mpirun -np 4" and "mpirun
>>>>>>>>> -np 5" both works, but in both cases there are only 4 tasks. It isn't
>>>>>>>>> crucial, because there is nor real oversubscription, but there is 
>>>>>>>>> still some
>>>>>>>>> bug which can affect something in future.
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Anton Starikov.
>>>>>>>>>
>>>>>>>>> On May 12, 2009, at 1:45 AM, Ralph Castain wrote:
>>>>>>>>>
>>>>>>>>> This is fixed as of r21208.
>>>>>>>>>>
>>>>>>>>>> Thanks for reporting it!
>>>>>>>>>> Ralph
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On May 11, 2009, at 12:51 PM, Anton Starikov wrote:
>>>>>>>>>>
>>>>>>>>>> Although removing this check solves problem of having more slots
>>>>>>>>>>> in rankfile than necessary, there is another problem.
>>>>>>>>>>>
>>>>>>>>>>> If I set rmaps_base_no_oversubscribe=1 then if, for example:
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> hostfile:
>>>>>>>>>>>
>>>>>>>>>>> node01
>>>>>>>>>>> node01
>>>>>>>>>>> node02
>>>>>>>>>>> node02
>>>>>>>>>>>
>>>>>>>>>>> rankfile:
>>>>>>>>>>>
>>>>>>>>>>> rank 0=node01 slot=1
>>>>>>>>>>> rank 1=node01 slot=0
>>>>>>>>>>> rank 2=node02 slot=1
>>>>>>>>>>> rank 3=node02 slot=0
>>>>>>>>>>>
>>>>>>>>>>> mpirun -np 4 ./something
>>>>>>>>>>>
>>>>>>>>>>> complains with:
>>>>>>>>>>>
>>>>>>>>>>> "There are not enough slots available in the system to satisfy
>>>>>>>>>>> the 4 slots
>>>>>>>>>>> that were requested by the application"
>>>>>>>>>>>
>>>>>>>>>>> but "mpirun -np 3 ./something" will work though. It works, when
>>>>>>>>>>> you ask for 1 CPU less. And the same behavior in any case (shared 
>>>>>>>>>>> nodes,
>>>>>>>>>>> non-shared nodes, multi-node)
>>>>>>>>>>>
>>>>>>>>>>> If you switch off rmaps_base_no_oversubscribe, then it works and
>>>>>>>>>>> all affinities set as it requested in rankfile, there is no
>>>>>>>>>>> oversubscription.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Anton.
>>>>>>>>>>>
>>>>>>>>>>> On May 5, 2009, at 3:08 PM, Ralph Castain wrote:
>>>>>>>>>>>
>>>>>>>>>>> Ah - thx for catching that, I'll remove that check. It no longer
>>>>>>>>>>>> is required.
>>>>>>>>>>>>
>>>>>>>>>>>> Thx!
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, May 5, 2009 at 7:04 AM, Lenny Verkhovsky <
>>>>>>>>>>>> lenny.verkhov...@gmail.com> wrote:
>>>>>>>>>>>> According to the code it does cares.
>>>>>>>>>>>>
>>>>>>>>>>>> $vi orte/mca/rmaps/rank_file/rmaps_rank_file.c +572
>>>>>>>>>>>>
>>>>>>>>>>>> ival = orte_rmaps_rank_file_value.ival;
>>>>>>>>>>>> if ( ival > (np-1) ) {
>>>>>>>>>>>> orte_show_help("help-rmaps_rank_file.txt", "bad-rankfile", true,
>>>>>>>>>>>> ival, rankfile);
>>>>>>>>>>>> rc = ORTE_ERR_BAD_PARAM;
>>>>>>>>>>>> goto unlock;
>>>>>>>>>>>> }
>>>>>>>>>>>>
>>>>>>>>>>>> If I remember correctly, I used an array to map ranks, and since
>>>>>>>>>>>> the length of array is NP, maximum index must be less than np, so 
>>>>>>>>>>>> if you
>>>>>>>>>>>> have the number of rank > NP, you have no place to put it inside 
>>>>>>>>>>>> array.
>>>>>>>>>>>>
>>>>>>>>>>>> "Likewise, if you have more procs than the rankfile specifies,
>>>>>>>>>>>> we map the additional procs either byslot (default) or bynode (if 
>>>>>>>>>>>> you
>>>>>>>>>>>> specify that option). So the rankfile doesn't need to contain an 
>>>>>>>>>>>> entry for
>>>>>>>>>>>> every proc."  - Correct point.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Lenny.
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On 5/5/09, Ralph Castain <r...@open-mpi.org> wrote: Sorry Lenny,
>>>>>>>>>>>> but that isn't correct. The rankfile mapper doesn't care if the 
>>>>>>>>>>>> rankfile
>>>>>>>>>>>> contains additional info - it only maps up to the number of 
>>>>>>>>>>>> processes, and
>>>>>>>>>>>> ignores anything beyond that number. So there is no need to remove 
>>>>>>>>>>>> the
>>>>>>>>>>>> additional info.
>>>>>>>>>>>>
>>>>>>>>>>>> Likewise, if you have more procs than the rankfile specifies, we
>>>>>>>>>>>> map the additional procs either byslot (default) or bynode (if you 
>>>>>>>>>>>> specify
>>>>>>>>>>>> that option). So the rankfile doesn't need to contain an entry for 
>>>>>>>>>>>> every
>>>>>>>>>>>> proc.
>>>>>>>>>>>>
>>>>>>>>>>>> Just don't want to confuse folks.
>>>>>>>>>>>> Ralph
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Tue, May 5, 2009 at 5:59 AM, Lenny Verkhovsky <
>>>>>>>>>>>> lenny.verkhov...@gmail.com> wrote:
>>>>>>>>>>>> Hi,
>>>>>>>>>>>> maximum rank number must be less then np.
>>>>>>>>>>>> if np=1 then there is only rank 0 in the system, so rank 1 is
>>>>>>>>>>>> invalid.
>>>>>>>>>>>> please remove "rank 1=node2 slot=*" from the rankfile
>>>>>>>>>>>> Best regards,
>>>>>>>>>>>> Lenny.
>>>>>>>>>>>>
>>>>>>>>>>>> On Mon, May 4, 2009 at 11:14 AM, Geoffroy Pignot <
>>>>>>>>>>>> geopig...@gmail.com> wrote:
>>>>>>>>>>>> Hi ,
>>>>>>>>>>>>
>>>>>>>>>>>> I got the openmpi-1.4a1r21095.tar.gz tarball, but unfortunately
>>>>>>>>>>>> my command doesn't work
>>>>>>>>>>>>
>>>>>>>>>>>> cat rankf:
>>>>>>>>>>>> rank 0=node1 slot=*
>>>>>>>>>>>> rank 1=node2 slot=*
>>>>>>>>>>>>
>>>>>>>>>>>> cat hostf:
>>>>>>>>>>>> node1 slots=2
>>>>>>>>>>>> node2 slots=2
>>>>>>>>>>>>
>>>>>>>>>>>> mpirun  --rankfile rankf --hostfile hostf  --host node1 -n 1
>>>>>>>>>>>> hostname : --host node2 -n 1 hostname
>>>>>>>>>>>>
>>>>>>>>>>>> Error, invalid rank (1) in the rankfile (rankf)
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> [r011n006:28986] [[45541,0],0] ORTE_ERROR_LOG: Bad parameter in
>>>>>>>>>>>> file rmaps_rank_file.c at line 403
>>>>>>>>>>>> [r011n006:28986] [[45541,0],0] ORTE_ERROR_LOG: Bad parameter in
>>>>>>>>>>>> file base/rmaps_base_map_job.c at line 86
>>>>>>>>>>>> [r011n006:28986] [[45541,0],0] ORTE_ERROR_LOG: Bad parameter in
>>>>>>>>>>>> file base/plm_base_launch_support.c at line 86
>>>>>>>>>>>> [r011n006:28986] [[45541,0],0] ORTE_ERROR_LOG: Bad parameter in
>>>>>>>>>>>> file plm_rsh_module.c at line 1016
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Ralph, could you tell me if my command syntax is correct or not
>>>>>>>>>>>> ? if not, give me the expected one ?
>>>>>>>>>>>>
>>>>>>>>>>>> Regards
>>>>>>>>>>>>
>>>>>>>>>>>> Geoffroy
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> 2009/4/30 Geoffroy Pignot <geopig...@gmail.com>
>>>>>>>>>>>>
>>>>>>>>>>>> Immediately Sir !!! :)
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks again Ralph
>>>>>>>>>>>>
>>>>>>>>>>>> Geoffroy
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> ------------------------------
>>>>>>>>>>>>
>>>>>>>>>>>> Message: 2
>>>>>>>>>>>> Date: Thu, 30 Apr 2009 06:45:39 -0600
>>>>>>>>>>>> From: Ralph Castain <r...@open-mpi.org>
>>>>>>>>>>>> Subject: Re: [OMPI users] 1.3.1 -rf rankfile behaviour ??
>>>>>>>>>>>> To: Open MPI Users <us...@open-mpi.org>
>>>>>>>>>>>> Message-ID:
>>>>>>>>>>>>    <71d2d8cc0904300545v61a42fe1k50086d2704d0f...@mail.gmail.com
>>>>>>>>>>>> >
>>>>>>>>>>>> Content-Type: text/plain; charset="iso-8859-1"
>>>>>>>>>>>>
>>>>>>>>>>>> I believe this is fixed now in our development trunk - you can
>>>>>>>>>>>> download any
>>>>>>>>>>>> tarball starting from last night and give it a try, if you like.
>>>>>>>>>>>> Any
>>>>>>>>>>>> feedback would be appreciated.
>>>>>>>>>>>>
>>>>>>>>>>>> Ralph
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On Apr 14, 2009, at 7:57 AM, Ralph Castain wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Ah now, I didn't say it -worked-, did I? :-)
>>>>>>>>>>>>
>>>>>>>>>>>> Clearly a bug exists in the program. I'll try to take a look at
>>>>>>>>>>>> it (if Lenny
>>>>>>>>>>>> doesn't get to it first), but it won't be until later in the
>>>>>>>>>>>> week.
>>>>>>>>>>>>
>>>>>>>>>>>> On Apr 14, 2009, at 7:18 AM, Geoffroy Pignot wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> I agree with you Ralph , and that 's what I expect from openmpi
>>>>>>>>>>>> but my
>>>>>>>>>>>> second example shows that it's not working
>>>>>>>>>>>>
>>>>>>>>>>>> cat hostfile.0
>>>>>>>>>>>> r011n002 slots=4
>>>>>>>>>>>> r011n003 slots=4
>>>>>>>>>>>>
>>>>>>>>>>>> cat rankfile.0
>>>>>>>>>>>> rank 0=r011n002 slot=0
>>>>>>>>>>>> rank 1=r011n003 slot=1
>>>>>>>>>>>>
>>>>>>>>>>>> mpirun --hostfile hostfile.0 -rf rankfile.0 -n 1 hostname : -n 1
>>>>>>>>>>>> hostname
>>>>>>>>>>>> ### CRASHED
>>>>>>>>>>>>
>>>>>>>>>>>> > > Error, invalid rank (1) in the rankfile (rankfile.0)
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > > [r011n002:25129] [[63976,0],0] ORTE_ERROR_LOG: Bad parameter
>>>>>>>>>>>> in file
>>>>>>>>>>>> > > rmaps_rank_file.c at line 404
>>>>>>>>>>>> > > [r011n002:25129] [[63976,0],0] ORTE_ERROR_LOG: Bad parameter
>>>>>>>>>>>> in file
>>>>>>>>>>>> > > base/rmaps_base_map_job.c at line 87
>>>>>>>>>>>> > > [r011n002:25129] [[63976,0],0] ORTE_ERROR_LOG: Bad parameter
>>>>>>>>>>>> in file
>>>>>>>>>>>> > > base/plm_base_launch_support.c at line 77
>>>>>>>>>>>> > > [r011n002:25129] [[63976,0],0] ORTE_ERROR_LOG: Bad parameter
>>>>>>>>>>>> in file
>>>>>>>>>>>> > > plm_rsh_module.c at line 985
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > > A daemon (pid unknown) died unexpectedly on signal 1  while
>>>>>>>>>>>> > attempting to
>>>>>>>>>>>> > > launch so we are aborting.
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > There may be more information reported by the environment
>>>>>>>>>>>> (see
>>>>>>>>>>>> > above).
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > This may be because the daemon was unable to find all the
>>>>>>>>>>>> needed
>>>>>>>>>>>> > shared
>>>>>>>>>>>> > > libraries on the remote node. You may set your
>>>>>>>>>>>> LD_LIBRARY_PATH to
>>>>>>>>>>>> > have the
>>>>>>>>>>>> > > location of the shared libraries on the remote nodes and
>>>>>>>>>>>> this will
>>>>>>>>>>>> > > automatically be forwarded to the remote nodes.
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > > orterun noticed that the job aborted, but has no info as to
>>>>>>>>>>>> the
>>>>>>>>>>>> > process
>>>>>>>>>>>> > > that caused that situation.
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > > orterun: clean termination accomplished
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Message: 4
>>>>>>>>>>>> Date: Tue, 14 Apr 2009 06:55:58 -0600
>>>>>>>>>>>> From: Ralph Castain <r...@lanl.gov>
>>>>>>>>>>>> Subject: Re: [OMPI users] 1.3.1 -rf rankfile behaviour ??
>>>>>>>>>>>> To: Open MPI Users <us...@open-mpi.org>
>>>>>>>>>>>> Message-ID: <f6290ada-a196-43f0-a853-cbcb802d8...@lanl.gov>
>>>>>>>>>>>> Content-Type: text/plain; charset="us-ascii"; Format="flowed";
>>>>>>>>>>>>   DelSp="yes"
>>>>>>>>>>>>
>>>>>>>>>>>> The rankfile cuts across the entire job - it isn't applied on an
>>>>>>>>>>>> app_context basis. So the ranks in your rankfile must correspond
>>>>>>>>>>>> to
>>>>>>>>>>>> the eventual rank of each process in the cmd line.
>>>>>>>>>>>>
>>>>>>>>>>>> Unfortunately, that means you have to count ranks. In your case,
>>>>>>>>>>>> you
>>>>>>>>>>>> only have four, so that makes life easier. Your rankfile would
>>>>>>>>>>>> look
>>>>>>>>>>>> something like this:
>>>>>>>>>>>>
>>>>>>>>>>>> rank 0=r001n001 slot=0
>>>>>>>>>>>> rank 1=r001n002 slot=1
>>>>>>>>>>>> rank 2=r001n001 slot=1
>>>>>>>>>>>> rank 3=r001n002 slot=2
>>>>>>>>>>>>
>>>>>>>>>>>> HTH
>>>>>>>>>>>> Ralph
>>>>>>>>>>>>
>>>>>>>>>>>> On Apr 14, 2009, at 12:19 AM, Geoffroy Pignot wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> > Hi,
>>>>>>>>>>>> >
>>>>>>>>>>>> > I agree that my examples are not very clear. What I want to do
>>>>>>>>>>>> is to
>>>>>>>>>>>> > launch a multiexes application (masters-slaves) and benefit
>>>>>>>>>>>> from the
>>>>>>>>>>>> > processor affinity.
>>>>>>>>>>>> > Could you show me how to convert this command , using -rf
>>>>>>>>>>>> option
>>>>>>>>>>>> > (whatever the affinity is)
>>>>>>>>>>>> >
>>>>>>>>>>>> > mpirun -n 1 -host r001n001 master.x options1  : -n 1 -host
>>>>>>>>>>>> r001n002
>>>>>>>>>>>> > master.x options2 : -n 1 -host r001n001 slave.x options3 : -n
>>>>>>>>>>>> 1 -
>>>>>>>>>>>> > host r001n002 slave.x options4
>>>>>>>>>>>> >
>>>>>>>>>>>> > Thanks for your help
>>>>>>>>>>>> >
>>>>>>>>>>>> > Geoffroy
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>> > Message: 2
>>>>>>>>>>>> > Date: Sun, 12 Apr 2009 18:26:35 +0300
>>>>>>>>>>>> > From: Lenny Verkhovsky <lenny.verkhov...@gmail.com>
>>>>>>>>>>>> > Subject: Re: [OMPI users] 1.3.1 -rf rankfile behaviour ??
>>>>>>>>>>>> > To: Open MPI Users <us...@open-mpi.org>
>>>>>>>>>>>> > Message-ID:
>>>>>>>>>>>> >        <
>>>>>>>>>>>> 453d39990904120826t2e1d1d33l7bb1fe3de65b5...@mail.gmail.com>
>>>>>>>>>>>> > Content-Type: text/plain; charset="iso-8859-1"
>>>>>>>>>>>> >
>>>>>>>>>>>> > Hi,
>>>>>>>>>>>> >
>>>>>>>>>>>> > The first "crash" is OK, since your rankfile has ranks 0 and 1
>>>>>>>>>>>> > defined,
>>>>>>>>>>>> > while n=1, which means only rank 0 is present and can be
>>>>>>>>>>>> allocated.
>>>>>>>>>>>> >
>>>>>>>>>>>> > NP must be >= the largest rank in rankfile.
>>>>>>>>>>>> >
>>>>>>>>>>>> > What exactly are you trying to do ?
>>>>>>>>>>>> >
>>>>>>>>>>>> > I tried to recreate your seqv but all I got was
>>>>>>>>>>>> >
>>>>>>>>>>>> > ~/work/svn/ompi/trunk/build_x86-64/install/bin/mpirun
>>>>>>>>>>>> --hostfile
>>>>>>>>>>>> > hostfile.0
>>>>>>>>>>>> > -rf rankfile.0 -n 1 hostname : -rf rankfile.1 -n 1 hostname
>>>>>>>>>>>> > [witch19:30798] mca: base: component_find: paffinity
>>>>>>>>>>>> > "mca_paffinity_linux"
>>>>>>>>>>>> > uses an MCA interface that is not recognized (component MCA
>>>>>>>>>>>> v1.0.0 !=
>>>>>>>>>>>> > supported MCA v2.0.0) -- ignored
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > It looks like opal_init failed for some reason; your parallel
>>>>>>>>>>>> > process is
>>>>>>>>>>>> > likely to abort. There are many reasons that a parallel
>>>>>>>>>>>> process can
>>>>>>>>>>>> > fail during opal_init; some of which are due to configuration
>>>>>>>>>>>> or
>>>>>>>>>>>> > environment problems. This failure appears to be an internal
>>>>>>>>>>>> failure;
>>>>>>>>>>>> > here's some additional information (which may only be relevant
>>>>>>>>>>>> to an
>>>>>>>>>>>> > Open MPI developer):
>>>>>>>>>>>> >
>>>>>>>>>>>> >  opal_carto_base_select failed
>>>>>>>>>>>> >  --> Returned value -13 instead of OPAL_SUCCESS
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > [witch19:30798] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found
>>>>>>>>>>>> in file
>>>>>>>>>>>> > ../../orte/runtime/orte_init.c at line 78
>>>>>>>>>>>> > [witch19:30798] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found
>>>>>>>>>>>> in file
>>>>>>>>>>>> > ../../orte/orted/orted_main.c at line 344
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > A daemon (pid 11629) died unexpectedly with status 243 while
>>>>>>>>>>>> > attempting
>>>>>>>>>>>> > to launch so we are aborting.
>>>>>>>>>>>> >
>>>>>>>>>>>> > There may be more information reported by the environment (see
>>>>>>>>>>>> above).
>>>>>>>>>>>> >
>>>>>>>>>>>> > This may be because the daemon was unable to find all the
>>>>>>>>>>>> needed
>>>>>>>>>>>> > shared
>>>>>>>>>>>> > libraries on the remote node. You may set your LD_LIBRARY_PATH
>>>>>>>>>>>> to
>>>>>>>>>>>> > have the
>>>>>>>>>>>> > location of the shared libraries on the remote nodes and this
>>>>>>>>>>>> will
>>>>>>>>>>>> > automatically be forwarded to the remote nodes.
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > mpirun noticed that the job aborted, but has no info as to the
>>>>>>>>>>>> process
>>>>>>>>>>>> > that caused that situation.
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > mpirun: clean termination accomplished
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>> > Lenny.
>>>>>>>>>>>> >
>>>>>>>>>>>> >
>>>>>>>>>>>> > On 4/10/09, Geoffroy Pignot <geopig...@gmail.com> wrote:
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > Hi ,
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > I am currently testing the process affinity capabilities of
>>>>>>>>>>>> > openmpi and I
>>>>>>>>>>>> > > would like to know if the rankfile behaviour I will describe
>>>>>>>>>>>> below
>>>>>>>>>>>> > is normal
>>>>>>>>>>>> > > or not ?
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > cat hostfile.0
>>>>>>>>>>>> > > r011n002 slots=4
>>>>>>>>>>>> > > r011n003 slots=4
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > cat rankfile.0
>>>>>>>>>>>> > > rank 0=r011n002 slot=0
>>>>>>>>>>>> > > rank 1=r011n003 slot=1
>>>>>>>>>>>> > >
>>>>>>>>>>>> > >
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>>
>>>>>>>>>>>> ##################################################################################
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > mpirun --hostfile hostfile.0 -rf rankfile.0 -n 2  hostname
>>>>>>>>>>>> ### OK
>>>>>>>>>>>> > > r011n002
>>>>>>>>>>>> > > r011n003
>>>>>>>>>>>> > >
>>>>>>>>>>>> > >
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>>
>>>>>>>>>>>> ##################################################################################
>>>>>>>>>>>> > > but
>>>>>>>>>>>> > > mpirun --hostfile hostfile.0 -rf rankfile.0 -n 1 hostname :
>>>>>>>>>>>> -n 1
>>>>>>>>>>>> > hostname
>>>>>>>>>>>> > > ### CRASHED
>>>>>>>>>>>> > > *
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > > Error, invalid rank (1) in the rankfile (rankfile.0)
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > > [r011n002:25129] [[63976,0],0] ORTE_ERROR_LOG: Bad parameter
>>>>>>>>>>>> in file
>>>>>>>>>>>> > > rmaps_rank_file.c at line 404
>>>>>>>>>>>> > > [r011n002:25129] [[63976,0],0] ORTE_ERROR_LOG: Bad parameter
>>>>>>>>>>>> in file
>>>>>>>>>>>> > > base/rmaps_base_map_job.c at line 87
>>>>>>>>>>>> > > [r011n002:25129] [[63976,0],0] ORTE_ERROR_LOG: Bad parameter
>>>>>>>>>>>> in file
>>>>>>>>>>>> > > base/plm_base_launch_support.c at line 77
>>>>>>>>>>>> > > [r011n002:25129] [[63976,0],0] ORTE_ERROR_LOG: Bad parameter
>>>>>>>>>>>> in file
>>>>>>>>>>>> > > plm_rsh_module.c at line 985
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > > A daemon (pid unknown) died unexpectedly on signal 1  while
>>>>>>>>>>>> > attempting to
>>>>>>>>>>>> > > launch so we are aborting.
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > There may be more information reported by the environment
>>>>>>>>>>>> (see
>>>>>>>>>>>> > above).
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > This may be because the daemon was unable to find all the
>>>>>>>>>>>> needed
>>>>>>>>>>>> > shared
>>>>>>>>>>>> > > libraries on the remote node. You may set your
>>>>>>>>>>>> LD_LIBRARY_PATH to
>>>>>>>>>>>> > have the
>>>>>>>>>>>> > > location of the shared libraries on the remote nodes and
>>>>>>>>>>>> this will
>>>>>>>>>>>> > > automatically be forwarded to the remote nodes.
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > > orterun noticed that the job aborted, but has no info as to
>>>>>>>>>>>> the
>>>>>>>>>>>> > process
>>>>>>>>>>>> > > that caused that situation.
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>>> > > orterun: clean termination accomplished
>>>>>>>>>>>> > > *
>>>>>>>>>>>> > > It seems that the rankfile option is not propagted to the
>>>>>>>>>>>> second
>>>>>>>>>>>> > command
>>>>>>>>>>>> > > line ; there is no global understanding of the ranking
>>>>>>>>>>>> inside a
>>>>>>>>>>>> > mpirun
>>>>>>>>>>>> > > command.
>>>>>>>>>>>> > >
>>>>>>>>>>>> > >
>>>>>>>>>>>> > >
>>>>>>>>>>>> >
>>>>>>>>>>>>
>>>>>>>>>>>> ##################################################################################
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > Assuming that , I tried to provide a rankfile to each
>>>>>>>>>>>> command line:
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > cat rankfile.0
>>>>>>>>>>>> > > rank 0=r011n002 slot=0
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > cat rankfile.1
>>>>>>>>>>>> > > rank 0=r011n003 slot=1
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > mpirun --hostfile hostfile.0 -rf rankfile.0 -n 1 hostname :
>>>>>>>>>>>> -rf
>>>>>>>>>>>> > rankfile.1
>>>>>>>>>>>> > > -n 1 hostname ### CRASHED
>>>>>>>>>>>> > > *[r011n002:28778] *** Process received signal ***
>>>>>>>>>>>> > > [r011n002:28778] Signal: Segmentation fault (11)
>>>>>>>>>>>> > > [r011n002:28778] Signal code: Address not mapped (1)
>>>>>>>>>>>> > > [r011n002:28778] Failing at address: 0x34
>>>>>>>>>>>> > > [r011n002:28778] [ 0] [0xffffe600]
>>>>>>>>>>>> > > [r011n002:28778] [ 1]
>>>>>>>>>>>> > > /tmp/HALMPI/openmpi-1.3.1/lib/libopen-rte.so.
>>>>>>>>>>>> > 0(orte_odls_base_default_get_add_procs_data+0x55d)
>>>>>>>>>>>> > > [0x5557decd]
>>>>>>>>>>>> > > [r011n002:28778] [ 2]
>>>>>>>>>>>> > > /tmp/HALMPI/openmpi-1.3.1/lib/libopen-rte.so.
>>>>>>>>>>>> > 0(orte_plm_base_launch_apps+0x117)
>>>>>>>>>>>> > > [0x555842a7]
>>>>>>>>>>>> > > [r011n002:28778] [ 3] /tmp/HALMPI/openmpi-1.3.1/lib/openmpi/
>>>>>>>>>>>> > mca_plm_rsh.so
>>>>>>>>>>>> > > [0x556098c0]
>>>>>>>>>>>> > > [r011n002:28778] [ 4] /tmp/HALMPI/openmpi-1.3.1/bin/orterun
>>>>>>>>>>>> > [0x804aa27]
>>>>>>>>>>>> > > [r011n002:28778] [ 5] /tmp/HALMPI/openmpi-1.3.1/bin/orterun
>>>>>>>>>>>> > [0x804a022]
>>>>>>>>>>>> > > [r011n002:28778] [ 6] /lib/libc.so.6(__libc_start_main+0xdc)
>>>>>>>>>>>> > [0x9f1dec]
>>>>>>>>>>>> > > [r011n002:28778] [ 7] /tmp/HALMPI/openmpi-1.3.1/bin/orterun
>>>>>>>>>>>> > [0x8049f71]
>>>>>>>>>>>> > > [r011n002:28778] *** End of error message ***
>>>>>>>>>>>> > > Segmentation fault (core dumped)*
>>>>>>>>>>>> > >
>>>>>>>>>>>> > >
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > I hope that I've found a bug because it would be very
>>>>>>>>>>>> important
>>>>>>>>>>>> > for me to
>>>>>>>>>>>> > > have this kind of capabiliy .
>>>>>>>>>>>> > > Launch a multiexe mpirun command line and be able to bind my
>>>>>>>>>>>> exes
>>>>>>>>>>>> > and
>>>>>>>>>>>> > > sockets together.
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > Thanks in advance for your help
>>>>>>>>>>>> > >
>>>>>>>>>>>> > > Geoffroy
>>>>>>>>>>>> > _______________________________________________
>>>>>>>>>>>> > users mailing list
>>>>>>>>>>>> > us...@open-mpi.org
>>>>>>>>>>>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>> -------------- next part --------------
>>>>>>>>>>>> HTML attachment scrubbed and removed
>>>>>>>>>>>>
>>>>>>>>>>>> ------------------------------
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>> End of users Digest, Vol 1202, Issue 2
>>>>>>>>>>>> **************************************
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> -------------- next part --------------
>>>>>>>>>>>> HTML attachment scrubbed and removed
>>>>>>>>>>>>
>>>>>>>>>>>> ------------------------------
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>> End of users Digest, Vol 1218, Issue 2
>>>>>>>>>>>> **************************************
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> users mailing list
>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>>>>>>>>>> users mailing list
>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> users mailing list
>>>>>>>>> us...@open-mpi.org
>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>
>>>>
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>
>>>
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>>
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to