Have you tried putting the "<" as a separate parameter? In other words,
since you are specifying the argv, you have to specify each of them
separately. So it looks more like:

"mpirun", "-n", "1", "myapp", "<", "stdinfile"

Does that work?
Ralph


On Wed, Dec 17, 2014 at 8:07 AM, Alex A. Schmidt <a...@ufsm.br> wrote:
>
> Ralph,
>
> I am afraid I will have to insist on i/o redirection matter
> for the spawnee process.
>
> I have a "child" mpi code that do just 2 things: read the 3 parameters
> passed to it and print them, and then read data from stdin and show it.
> So, if "stdin_file" is a text file with two lines, say:
>
> 10
> 20
>
> executing "mpirun -n 1 child A B < stdin_file" wiil ouput two lines:
>
> [A]  [B]  []
> 10  20
>
> On the other hand , calling "child" from MPI_Comm_spawn("child",args,...)
> where
>
> args(1) = "A"
> args(2) = "B"
> args(3) ="< stdin_file"
> args(4) = " "
>
> will make "child" outputs only 1 line
>
> [A] [B] [< stdin_file]
>
> and then fails because there is not stdin data to read from.
>
> Please, note that surprisingly the whole string "< stdin_file" is
> interpreted
> as a third parameter to "child" and not a stdin...
>
> Alex
>
>
>
>
>
>
>
> 2014-12-15 17:26 GMT-02:00 Alex A. Schmidt <a...@ufsm.br>:
>
>> Ralph,
>>
>> I guess you mean "call mpi_comm_spawn( 'siesta', '< infile' , 2 ,...)"
>>
>> to execute 'mpirun -n 2 siesta < infile' on the spawnee side. That was
>> my first choice. Well, siesta behaves as if no stdin file was present...
>>
>> Alex
>>
>>
>> 2014-12-15 17:07 GMT-02:00 Ralph Castain <r...@open-mpi.org>:
>>>
>>> You should be able to just include that in your argv that you pass to
>>> the Comm_spawn API.
>>>
>>>
>>> On Mon, Dec 15, 2014 at 9:27 AM, Alex A. Schmidt <a...@ufsm.br> wrote:
>>>
>>>> George,
>>>>
>>>> Thanks for the tip. In fact, calling mpi_comm_spawn right away with MPI
>>>> _COMM_SELF
>>>> has worked for me just as well -- no subgroups needed at all.
>>>>
>>>> I am testing this openmpi app named "siesta" in parallel. The source
>>>> code is available,
>>>> so making it "spawn ready" by adding the pair mpi_comm_get_parent + 
>>>> mpi_comm_disconnect
>>>>
>>>> into the main code can be done.  If it works, maybe the siesta's
>>>> developers can be convinced
>>>> to add this feature in a future release.
>>>>
>>>> However, siesta is launched only by specifying input/output files with
>>>> i/o redirection like
>>>>
>>>> mpirun -n <*some number*>  siesta < infile > outfile
>>>>
>>>> So far, I could not find anything about how to set an stdin file for an
>>>> spawnee process.
>>>> Specifiyng it in a app context file doesn't seem to work. Can it be
>>>> done? Maybe through
>>>> an MCA parameter?
>>>>
>>>> Alex
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> 2014-12-15 2:43 GMT-02:00 George Bosilca <bosi...@icl.utk.edu>:
>>>>>
>>>>> Alex,
>>>>>
>>>>> The code looks good, and is 100% MPI standard accurate.
>>>>>
>>>>> I would change the way you create the subcoms in the parent. You do a
>>>>> lot of useless operations, as you can achieve exactly the same outcome 
>>>>> (one
>>>>> communicator per node), either by duplicating MPI_COMM_SELF or doing an
>>>>> MPI_Comm_split with the color equal to your rank.
>>>>>
>>>>>   George.
>>>>>
>>>>>
>>>>> On Sun, Dec 14, 2014 at 2:20 AM, Alex A. Schmidt <a...@ufsm.br> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> Sorry, guys. I don't think the newbie here can follow any discussion
>>>>>> beyond basic mpi...
>>>>>>
>>>>>> Anyway, if I add the pair
>>>>>>
>>>>>> call MPI_COMM_GET_PARENT(mpi_comm_parent,ierror)
>>>>>> call MPI_COMM_DISCONNECT(mpi_comm_parent,ierror)
>>>>>>
>>>>>> on the spawnee side I get the proper response in the spawning
>>>>>> processes.
>>>>>>
>>>>>> Please, take a look at the attached toy codes parent.F and child.F
>>>>>> I've been playing with. 'mpirun -n 2 parent' seems to work as
>>>>>> expected.
>>>>>>
>>>>>> Alex
>>>>>>
>>>>>> 2014-12-13 23:46 GMT-02:00 Gilles Gouaillardet <
>>>>>> gilles.gouaillar...@gmail.com>:
>>>>>>>
>>>>>>> Alex,
>>>>>>>
>>>>>>> Are you calling MPI_Comm_disconnect in the 3 "master" tasks and with
>>>>>>> the same remote communicator ?
>>>>>>>
>>>>>>> I also read the man page again, and MPI_Comm_disconnect does not
>>>>>>> ensure the remote processes have finished or called 
>>>>>>> MPI_Comm_disconnect, so
>>>>>>> that might not be the thing you need.
>>>>>>> George, can you please comment on that ?
>>>>>>>
>>>>>>> Cheers,
>>>>>>>
>>>>>>> Gilles
>>>>>>>
>>>>>>> George Bosilca <bosi...@icl.utk.edu> wrote:
>>>>>>> MPI_Comm_disconnect should be a local operation, there is no reason
>>>>>>> for it to deadlock. I looked at the code and everything is local with 
>>>>>>> the
>>>>>>> exception of a call to PMIX.FENCE. Can you attach to your deadlocked
>>>>>>> processes and confirm that they are stopped in the pmix.fence?
>>>>>>>
>>>>>>>   George.
>>>>>>>
>>>>>>>
>>>>>>> On Sat, Dec 13, 2014 at 8:47 AM, Alex A. Schmidt <a...@ufsm.br>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi
>>>>>>>>
>>>>>>>> Sorry, I was calling mpi_comm_disconnect on the group comm handler,
>>>>>>>> not
>>>>>>>> on the intercomm handler returned from the spawn call as it should
>>>>>>>> be.
>>>>>>>>
>>>>>>>> Well, calling the disconnect on the intercomm handler does halt the
>>>>>>>> spwaner
>>>>>>>> side but the wait is never completed since, as George points out,
>>>>>>>> there is no
>>>>>>>> disconnect call being made on the spawnee side.... and that brings
>>>>>>>> me back
>>>>>>>> to the beginning of the problem since, being a third party app,
>>>>>>>> that call would
>>>>>>>> never be there. I guess an mpi wrapper to deal with that could be
>>>>>>>> made for
>>>>>>>> the app, but I fell the wrapper itself, at the end, would face the
>>>>>>>> same problem
>>>>>>>> we face right now.
>>>>>>>>
>>>>>>>> My application is a genetic algorithm code that search optimal
>>>>>>>> configuration
>>>>>>>> (minimum or maximum energy) of cluster of atoms. The work flow
>>>>>>>> bottleneck
>>>>>>>> is the calculation of the cluster energy. For the cases which an
>>>>>>>> analytical
>>>>>>>> potential is available the calculation can be made internally and
>>>>>>>> the workload
>>>>>>>> is distributed among slaves nodes from a master node. This is also
>>>>>>>> done
>>>>>>>> when an analytical potential is not available and the energy
>>>>>>>> calculation must
>>>>>>>> be done externally by a quantum chemistry code like dftb+, siesta
>>>>>>>> and Gaussian.
>>>>>>>> So far, we have been running these codes in serial mode. No need to
>>>>>>>> say that
>>>>>>>> we could do a lot better if they could be executed in parallel.
>>>>>>>>
>>>>>>>> I am not familiar with DMRAA but it seems to be the right choice to
>>>>>>>> deal with
>>>>>>>> job schedulers as it covers the ones I am interested in (pbs/torque
>>>>>>>> and loadlever).
>>>>>>>>
>>>>>>>> Alex
>>>>>>>>
>>>>>>>> 2014-12-13 7:49 GMT-02:00 Gilles Gouaillardet <
>>>>>>>> gilles.gouaillar...@gmail.com>:
>>>>>>>>>
>>>>>>>>> George is right about the semantic
>>>>>>>>>
>>>>>>>>> However i am surprised it returns immediatly...
>>>>>>>>> That should either work or hang imho
>>>>>>>>>
>>>>>>>>> The second point is no more mpi related, and is batch manager
>>>>>>>>> specific.
>>>>>>>>>
>>>>>>>>> You will likely find a submit parameter to make the command block
>>>>>>>>> until the job completes. Or you can write your own wrapper.
>>>>>>>>> Or you can retrieve the jobid and qstat periodically to get the
>>>>>>>>> job state.
>>>>>>>>> If an api is available, this is also an option.
>>>>>>>>>
>>>>>>>>> Cheers,
>>>>>>>>>
>>>>>>>>> Gilles
>>>>>>>>>
>>>>>>>>> George Bosilca <bosi...@icl.utk.edu> wrote:
>>>>>>>>> You have to call MPI_Comm_disconnect on both sides of the
>>>>>>>>> intercommunicator. On the spawner processes you should call it on the
>>>>>>>>> intercom, while on the spawnees you should call it on the
>>>>>>>>> MPI_Comm_get_parent.
>>>>>>>>>
>>>>>>>>>   George.
>>>>>>>>>
>>>>>>>>> On Dec 12, 2014, at 20:43 , Alex A. Schmidt <a...@ufsm.br> wrote:
>>>>>>>>>
>>>>>>>>> Gilles,
>>>>>>>>>
>>>>>>>>> MPI_comm_disconnect seem to work but not quite.
>>>>>>>>> The call to it returns almost immediatly while
>>>>>>>>> the spawn processes keep piling up in the background
>>>>>>>>> until they are all done...
>>>>>>>>>
>>>>>>>>> I think system('env -i qsub...') to launch the third party apps
>>>>>>>>> would take the execution of every call back to the scheduler
>>>>>>>>> queue. How would I track each one for their completion?
>>>>>>>>>
>>>>>>>>> Alex
>>>>>>>>>
>>>>>>>>> 2014-12-12 22:35 GMT-02:00 Gilles Gouaillardet <
>>>>>>>>> gilles.gouaillar...@gmail.com>:
>>>>>>>>>>
>>>>>>>>>> Alex,
>>>>>>>>>>
>>>>>>>>>> You need MPI_Comm_disconnect at least.
>>>>>>>>>> I am not sure if this is 100% correct nor working.
>>>>>>>>>>
>>>>>>>>>> If you are using third party apps, why dont you do something like
>>>>>>>>>> system("env -i qsub ...")
>>>>>>>>>> with the right options to make qsub blocking or you manually wait
>>>>>>>>>> for the end of the job ?
>>>>>>>>>>
>>>>>>>>>> That looks like a much cleaner and simpler approach to me.
>>>>>>>>>>
>>>>>>>>>> Cheers,
>>>>>>>>>>
>>>>>>>>>> Gilles
>>>>>>>>>>
>>>>>>>>>> "Alex A. Schmidt" <a...@ufsm.br> wrote:
>>>>>>>>>> Hello Gilles,
>>>>>>>>>>
>>>>>>>>>> Ok, I believe I have a simple toy app running as I think it
>>>>>>>>>> should:
>>>>>>>>>> 'n' parent processes running under mpi_comm_world, each one
>>>>>>>>>> spawning its own 'm' child processes (each child group work
>>>>>>>>>> together nicely, returning the expected result for an mpi_
>>>>>>>>>> allreduce call).
>>>>>>>>>>
>>>>>>>>>> Now, as I mentioned before, the apps I want to run in the spawned
>>>>>>>>>> processes are third party mpi apps and I don't think it will be
>>>>>>>>>> possible
>>>>>>>>>> to exchange messages with them from my app. So, I do I tell
>>>>>>>>>> when the spawned processes have finnished running? All I have to
>>>>>>>>>> work
>>>>>>>>>> with is the intercommunicator returned from the mpi_comm_spawn
>>>>>>>>>> call...
>>>>>>>>>>
>>>>>>>>>> Alex
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 2014-12-12 2:42 GMT-02:00 Alex A. Schmidt <a...@ufsm.br>:
>>>>>>>>>>>
>>>>>>>>>>> Gilles,
>>>>>>>>>>>
>>>>>>>>>>> Well, yes, I guess....
>>>>>>>>>>>
>>>>>>>>>>> I'll do tests with the real third party apps and let you know.
>>>>>>>>>>> These are huge quantum chemistry codes (dftb+, siesta and
>>>>>>>>>>> Gaussian)
>>>>>>>>>>> which greatly benefits from a parallel environment. My code is
>>>>>>>>>>> just
>>>>>>>>>>> a front end to use those, but since we have a lot of data to
>>>>>>>>>>> process
>>>>>>>>>>> it also benefits from a parallel environment.
>>>>>>>>>>>
>>>>>>>>>>> Alex
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> 2014-12-12 2:30 GMT-02:00 Gilles Gouaillardet <
>>>>>>>>>>> gilles.gouaillar...@iferc.org>:
>>>>>>>>>>>>
>>>>>>>>>>>>  Alex,
>>>>>>>>>>>>
>>>>>>>>>>>> just to make sure ...
>>>>>>>>>>>> this is the behavior you expected, right ?
>>>>>>>>>>>>
>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>
>>>>>>>>>>>> Gilles
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On 2014/12/12 13:27, Alex A. Schmidt wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Gilles,
>>>>>>>>>>>>
>>>>>>>>>>>> Ok, very nice!
>>>>>>>>>>>>
>>>>>>>>>>>> When I excute
>>>>>>>>>>>>
>>>>>>>>>>>> do rank=1,3
>>>>>>>>>>>>     call  MPI_Comm_spawn('hello_world','
>>>>>>>>>>>> ',5,MPI_INFO_NULL,rank,MPI_COMM_WORLD,my_intercomm,MPI_ERRCODES_IGNORE,status)
>>>>>>>>>>>> enddo
>>>>>>>>>>>>
>>>>>>>>>>>> I do get 15 instances of the 'hello_world' app running: 5 for each 
>>>>>>>>>>>> parent
>>>>>>>>>>>> rank 1, 2 and 3.
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks a lot, Gilles.
>>>>>>>>>>>>
>>>>>>>>>>>> Best regargs,
>>>>>>>>>>>>
>>>>>>>>>>>> Alex
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> 2014-12-12 1:32 GMT-02:00 Gilles Gouaillardet 
>>>>>>>>>>>> <gilles.gouaillar...@iferc.org
>>>>>>>>>>>>
>>>>>>>>>>>>  :
>>>>>>>>>>>>
>>>>>>>>>>>>  Alex,
>>>>>>>>>>>>
>>>>>>>>>>>> just ask MPI_Comm_spawn to start (up to) 5 tasks via the maxprocs
>>>>>>>>>>>> parameter :
>>>>>>>>>>>>
>>>>>>>>>>>>        int MPI_Comm_spawn(char *command, char *argv[], int 
>>>>>>>>>>>> maxprocs,
>>>>>>>>>>>> MPI_Info info,
>>>>>>>>>>>>                          int root, MPI_Comm comm, MPI_Comm 
>>>>>>>>>>>> *intercomm,
>>>>>>>>>>>>                          int array_of_errcodes[])
>>>>>>>>>>>>
>>>>>>>>>>>> INPUT PARAMETERS
>>>>>>>>>>>>        maxprocs
>>>>>>>>>>>>               - maximum number of processes to start (integer, 
>>>>>>>>>>>> significant
>>>>>>>>>>>> only at root)
>>>>>>>>>>>>
>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>
>>>>>>>>>>>> Gilles
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On 2014/12/12 12:23, Alex A. Schmidt wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Hello Gilles,
>>>>>>>>>>>>
>>>>>>>>>>>> Thanks for your reply. The "env -i PATH=..." stuff seems to work!!!
>>>>>>>>>>>>
>>>>>>>>>>>> call system("sh -c 'env -i PATH=/usr/lib64/openmpi/bin:/bin mpirun 
>>>>>>>>>>>> -n 2
>>>>>>>>>>>> hello_world' ")
>>>>>>>>>>>>
>>>>>>>>>>>> did produce the expected result with a simple openmi "hello_world" 
>>>>>>>>>>>> code I
>>>>>>>>>>>> wrote.
>>>>>>>>>>>>
>>>>>>>>>>>> I might be harder though with the real third party app I have in 
>>>>>>>>>>>> mind. And
>>>>>>>>>>>> I realize
>>>>>>>>>>>> getting passed over a job scheduler with this approach might not 
>>>>>>>>>>>> work at
>>>>>>>>>>>> all...
>>>>>>>>>>>>
>>>>>>>>>>>> I have looked at the MPI_Comm_spawn call but I failed to 
>>>>>>>>>>>> understand how it
>>>>>>>>>>>> could help here. For instance, can I use it to launch an mpi app 
>>>>>>>>>>>> with the
>>>>>>>>>>>> option "-n 5" ?
>>>>>>>>>>>>
>>>>>>>>>>>> Alex
>>>>>>>>>>>>
>>>>>>>>>>>> 2014-12-12 0:36 GMT-02:00 Gilles Gouaillardet 
>>>>>>>>>>>> <gilles.gouaillar...@iferc.org
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>  :
>>>>>>>>>>>>
>>>>>>>>>>>>  Alex,
>>>>>>>>>>>>
>>>>>>>>>>>> can you try something like
>>>>>>>>>>>> call system(sh -c 'env -i /.../mpirun -np 2 /.../app_name')
>>>>>>>>>>>>
>>>>>>>>>>>> -i start with an empty environment
>>>>>>>>>>>> that being said, you might need to set a few environment variables
>>>>>>>>>>>> manually :
>>>>>>>>>>>> env -i PATH=/bin ...
>>>>>>>>>>>>
>>>>>>>>>>>> and that being also said, this "trick" could be just a bad idea :
>>>>>>>>>>>> you might be using a scheduler, and if you empty the environment, 
>>>>>>>>>>>> the
>>>>>>>>>>>> scheduler
>>>>>>>>>>>> will not be aware of the "inside" run.
>>>>>>>>>>>>
>>>>>>>>>>>> on top of that, invoking system might fail depending on the 
>>>>>>>>>>>> interconnect
>>>>>>>>>>>> you use.
>>>>>>>>>>>>
>>>>>>>>>>>> Bottom line, i believe Ralph's reply is still valid, even if five 
>>>>>>>>>>>> years
>>>>>>>>>>>> have passed :
>>>>>>>>>>>> changing your workflow, or using MPI_Comm_spawn is a much better 
>>>>>>>>>>>> approach.
>>>>>>>>>>>>
>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>
>>>>>>>>>>>> Gilles
>>>>>>>>>>>>
>>>>>>>>>>>> On 2014/12/12 11:22, Alex A. Schmidt wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Dear OpenMPI users,
>>>>>>>>>>>>
>>>>>>>>>>>> Regarding to this previous 
>>>>>>>>>>>> post<http://www.open-mpi.org/community/lists/users/2009/06/9560.php>
>>>>>>>>>>>>  <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> 
>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> 
>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> 
>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> 
>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> 
>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> 
>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> 
>>>>>>>>>>>> from 2009,
>>>>>>>>>>>> I wonder if the reply
>>>>>>>>>>>> from Ralph Castain is still valid. My need is similar but quite 
>>>>>>>>>>>> simpler:
>>>>>>>>>>>> to make a system call from an openmpi fortran application to run a
>>>>>>>>>>>> third party openmpi application. I don't need to exchange mpi 
>>>>>>>>>>>> messages
>>>>>>>>>>>> with the application. I just need to read the resulting output file
>>>>>>>>>>>> generated
>>>>>>>>>>>> by it. I have tried to do the following system call from my 
>>>>>>>>>>>> fortran openmpi
>>>>>>>>>>>> code:
>>>>>>>>>>>>
>>>>>>>>>>>> call system("sh -c 'mpirun -n 2 app_name")
>>>>>>>>>>>>
>>>>>>>>>>>> but I get
>>>>>>>>>>>>
>>>>>>>>>>>> **********************************************************
>>>>>>>>>>>>
>>>>>>>>>>>> Open MPI does not support recursive calls of mpirun
>>>>>>>>>>>>
>>>>>>>>>>>> **********************************************************
>>>>>>>>>>>>
>>>>>>>>>>>> Is there a way to make this work?
>>>>>>>>>>>>
>>>>>>>>>>>> Best regards,
>>>>>>>>>>>>
>>>>>>>>>>>> Alex
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing listus...@open-mpi.org
>>>>>>>>>>>>
>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>> Link to this post: 
>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25966.php
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing listus...@open-mpi.org
>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>> Link to this 
>>>>>>>>>>>> post:http://www.open-mpi.org/community/lists/users/2014/12/25967.php
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing listus...@open-mpi.org
>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>>
>>>>>>>>>>>> Link to this post: 
>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25968.php
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing listus...@open-mpi.org
>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>> Link to this 
>>>>>>>>>>>> post:http://www.open-mpi.org/community/lists/users/2014/12/25969.php
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing listus...@open-mpi.org
>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>> Link to this post: 
>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25970.php
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>> users mailing list
>>>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>>>> Subscription:
>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>>> Link to this post:
>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25971.php
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>>>>>>>>>> users mailing list
>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>> Link to this post:
>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25974.php
>>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> users mailing list
>>>>>>>>> us...@open-mpi.org
>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>> Link to this post:
>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25975.php
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> users mailing list
>>>>>>>>> us...@open-mpi.org
>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>> Link to this post:
>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25978.php
>>>>>>>>>
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>> Link to this post:
>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25979.php
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> Link to this post:
>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25981.php
>>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> Link to this post:
>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25982.php
>>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> Link to this post:
>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25991.php
>>>>>
>>>>
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> Link to this post:
>>>> http://www.open-mpi.org/community/lists/users/2014/12/26002.php
>>>>
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> Link to this post:
>>> http://www.open-mpi.org/community/lists/users/2014/12/26003.php
>>>
>>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2014/12/26010.php
>

Reply via email to