Ralph, Sorry, "<" as an element of argv to mpi_comm_spawn is interpreted just the same, as another parameter by the spawnee process.
But I am confused: wouldn't it be redundant to put "mpirun" "-n" "1" "myapp" as elements of argv, considering role of the other parameters of mpi_comm_spawn like the 1st and 3rd ? Imho, it seems to me that stdin and stdout should be keys to "info" and the respective filenames the values of those keys, if that is possible to implement... Alex 2014-12-17 15:16 GMT-02:00 Ralph Castain <r...@open-mpi.org>: > > Have you tried putting the "<" as a separate parameter? In other words, > since you are specifying the argv, you have to specify each of them > separately. So it looks more like: > > "mpirun", "-n", "1", "myapp", "<", "stdinfile" > > Does that work? > Ralph > > > On Wed, Dec 17, 2014 at 8:07 AM, Alex A. Schmidt <a...@ufsm.br> wrote: > >> Ralph, >> >> I am afraid I will have to insist on i/o redirection matter >> for the spawnee process. >> >> I have a "child" mpi code that do just 2 things: read the 3 parameters >> passed to it and print them, and then read data from stdin and show it. >> So, if "stdin_file" is a text file with two lines, say: >> >> 10 >> 20 >> >> executing "mpirun -n 1 child A B < stdin_file" wiil ouput two lines: >> >> [A] [B] [] >> 10 20 >> >> On the other hand , calling "child" from MPI_Comm_spawn("child",args,...) >> where >> >> args(1) = "A" >> args(2) = "B" >> args(3) ="< stdin_file" >> args(4) = " " >> >> will make "child" outputs only 1 line >> >> [A] [B] [< stdin_file] >> >> and then fails because there is not stdin data to read from. >> >> Please, note that surprisingly the whole string "< stdin_file" is >> interpreted >> as a third parameter to "child" and not a stdin... >> >> Alex >> >> >> >> >> >> >> >> 2014-12-15 17:26 GMT-02:00 Alex A. Schmidt <a...@ufsm.br>: >> >>> Ralph, >>> >>> I guess you mean "call mpi_comm_spawn( 'siesta', '< infile' , 2 ,...)" >>> >>> to execute 'mpirun -n 2 siesta < infile' on the spawnee side. That was >>> my first choice. Well, siesta behaves as if no stdin file was present... >>> >>> Alex >>> >>> >>> 2014-12-15 17:07 GMT-02:00 Ralph Castain <r...@open-mpi.org>: >>>> >>>> You should be able to just include that in your argv that you pass to >>>> the Comm_spawn API. >>>> >>>> >>>> On Mon, Dec 15, 2014 at 9:27 AM, Alex A. Schmidt <a...@ufsm.br> wrote: >>>> >>>>> George, >>>>> >>>>> Thanks for the tip. In fact, calling mpi_comm_spawn right away with >>>>> MPI_COMM_SELF >>>>> has worked for me just as well -- no subgroups needed at all. >>>>> >>>>> I am testing this openmpi app named "siesta" in parallel. The source >>>>> code is available, >>>>> so making it "spawn ready" by adding the pair mpi_comm_get_parent + >>>>> mpi_comm_disconnect >>>>> into the main code can be done. If it works, maybe the siesta's >>>>> developers can be convinced >>>>> to add this feature in a future release. >>>>> >>>>> However, siesta is launched only by specifying input/output files >>>>> with i/o redirection like >>>>> >>>>> mpirun -n <*some number*> siesta < infile > outfile >>>>> >>>>> So far, I could not find anything about how to set an stdin file for >>>>> an spawnee process. >>>>> Specifiyng it in a app context file doesn't seem to work. Can it be >>>>> done? Maybe through >>>>> an MCA parameter? >>>>> >>>>> Alex >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> 2014-12-15 2:43 GMT-02:00 George Bosilca <bosi...@icl.utk.edu>: >>>>>> >>>>>> Alex, >>>>>> >>>>>> The code looks good, and is 100% MPI standard accurate. >>>>>> >>>>>> I would change the way you create the subcoms in the parent. You do a >>>>>> lot of useless operations, as you can achieve exactly the same outcome >>>>>> (one >>>>>> communicator per node), either by duplicating MPI_COMM_SELF or doing an >>>>>> MPI_Comm_split with the color equal to your rank. >>>>>> >>>>>> George. >>>>>> >>>>>> >>>>>> On Sun, Dec 14, 2014 at 2:20 AM, Alex A. Schmidt <a...@ufsm.br> wrote: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> Sorry, guys. I don't think the newbie here can follow any discussion >>>>>>> beyond basic mpi... >>>>>>> >>>>>>> Anyway, if I add the pair >>>>>>> >>>>>>> call MPI_COMM_GET_PARENT(mpi_comm_parent,ierror) >>>>>>> call MPI_COMM_DISCONNECT(mpi_comm_parent,ierror) >>>>>>> >>>>>>> on the spawnee side I get the proper response in the spawning >>>>>>> processes. >>>>>>> >>>>>>> Please, take a look at the attached toy codes parent.F and child.F >>>>>>> I've been playing with. 'mpirun -n 2 parent' seems to work as >>>>>>> expected. >>>>>>> >>>>>>> Alex >>>>>>> >>>>>>> 2014-12-13 23:46 GMT-02:00 Gilles Gouaillardet < >>>>>>> gilles.gouaillar...@gmail.com>: >>>>>>>> >>>>>>>> Alex, >>>>>>>> >>>>>>>> Are you calling MPI_Comm_disconnect in the 3 "master" tasks and >>>>>>>> with the same remote communicator ? >>>>>>>> >>>>>>>> I also read the man page again, and MPI_Comm_disconnect does not >>>>>>>> ensure the remote processes have finished or called >>>>>>>> MPI_Comm_disconnect, so >>>>>>>> that might not be the thing you need. >>>>>>>> George, can you please comment on that ? >>>>>>>> >>>>>>>> Cheers, >>>>>>>> >>>>>>>> Gilles >>>>>>>> >>>>>>>> George Bosilca <bosi...@icl.utk.edu> wrote: >>>>>>>> MPI_Comm_disconnect should be a local operation, there is no reason >>>>>>>> for it to deadlock. I looked at the code and everything is local with >>>>>>>> the >>>>>>>> exception of a call to PMIX.FENCE. Can you attach to your deadlocked >>>>>>>> processes and confirm that they are stopped in the pmix.fence? >>>>>>>> >>>>>>>> George. >>>>>>>> >>>>>>>> >>>>>>>> On Sat, Dec 13, 2014 at 8:47 AM, Alex A. Schmidt <a...@ufsm.br> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi >>>>>>>>> >>>>>>>>> Sorry, I was calling mpi_comm_disconnect on the group comm >>>>>>>>> handler, not >>>>>>>>> on the intercomm handler returned from the spawn call as it should >>>>>>>>> be. >>>>>>>>> >>>>>>>>> Well, calling the disconnect on the intercomm handler does halt >>>>>>>>> the spwaner >>>>>>>>> side but the wait is never completed since, as George points out, >>>>>>>>> there is no >>>>>>>>> disconnect call being made on the spawnee side.... and that brings >>>>>>>>> me back >>>>>>>>> to the beginning of the problem since, being a third party app, >>>>>>>>> that call would >>>>>>>>> never be there. I guess an mpi wrapper to deal with that could be >>>>>>>>> made for >>>>>>>>> the app, but I fell the wrapper itself, at the end, would face the >>>>>>>>> same problem >>>>>>>>> we face right now. >>>>>>>>> >>>>>>>>> My application is a genetic algorithm code that search optimal >>>>>>>>> configuration >>>>>>>>> (minimum or maximum energy) of cluster of atoms. The work flow >>>>>>>>> bottleneck >>>>>>>>> is the calculation of the cluster energy. For the cases which an >>>>>>>>> analytical >>>>>>>>> potential is available the calculation can be made internally and >>>>>>>>> the workload >>>>>>>>> is distributed among slaves nodes from a master node. This is also >>>>>>>>> done >>>>>>>>> when an analytical potential is not available and the energy >>>>>>>>> calculation must >>>>>>>>> be done externally by a quantum chemistry code like dftb+, siesta >>>>>>>>> and Gaussian. >>>>>>>>> So far, we have been running these codes in serial mode. No need >>>>>>>>> to say that >>>>>>>>> we could do a lot better if they could be executed in parallel. >>>>>>>>> >>>>>>>>> I am not familiar with DMRAA but it seems to be the right choice >>>>>>>>> to deal with >>>>>>>>> job schedulers as it covers the ones I am interested in >>>>>>>>> (pbs/torque and loadlever). >>>>>>>>> >>>>>>>>> Alex >>>>>>>>> >>>>>>>>> 2014-12-13 7:49 GMT-02:00 Gilles Gouaillardet < >>>>>>>>> gilles.gouaillar...@gmail.com>: >>>>>>>>>> >>>>>>>>>> George is right about the semantic >>>>>>>>>> >>>>>>>>>> However i am surprised it returns immediatly... >>>>>>>>>> That should either work or hang imho >>>>>>>>>> >>>>>>>>>> The second point is no more mpi related, and is batch manager >>>>>>>>>> specific. >>>>>>>>>> >>>>>>>>>> You will likely find a submit parameter to make the command block >>>>>>>>>> until the job completes. Or you can write your own wrapper. >>>>>>>>>> Or you can retrieve the jobid and qstat periodically to get the >>>>>>>>>> job state. >>>>>>>>>> If an api is available, this is also an option. >>>>>>>>>> >>>>>>>>>> Cheers, >>>>>>>>>> >>>>>>>>>> Gilles >>>>>>>>>> >>>>>>>>>> George Bosilca <bosi...@icl.utk.edu> wrote: >>>>>>>>>> You have to call MPI_Comm_disconnect on both sides of the >>>>>>>>>> intercommunicator. On the spawner processes you should call it on the >>>>>>>>>> intercom, while on the spawnees you should call it on the >>>>>>>>>> MPI_Comm_get_parent. >>>>>>>>>> >>>>>>>>>> George. >>>>>>>>>> >>>>>>>>>> On Dec 12, 2014, at 20:43 , Alex A. Schmidt <a...@ufsm.br> wrote: >>>>>>>>>> >>>>>>>>>> Gilles, >>>>>>>>>> >>>>>>>>>> MPI_comm_disconnect seem to work but not quite. >>>>>>>>>> The call to it returns almost immediatly while >>>>>>>>>> the spawn processes keep piling up in the background >>>>>>>>>> until they are all done... >>>>>>>>>> >>>>>>>>>> I think system('env -i qsub...') to launch the third party apps >>>>>>>>>> would take the execution of every call back to the scheduler >>>>>>>>>> queue. How would I track each one for their completion? >>>>>>>>>> >>>>>>>>>> Alex >>>>>>>>>> >>>>>>>>>> 2014-12-12 22:35 GMT-02:00 Gilles Gouaillardet < >>>>>>>>>> gilles.gouaillar...@gmail.com>: >>>>>>>>>>> >>>>>>>>>>> Alex, >>>>>>>>>>> >>>>>>>>>>> You need MPI_Comm_disconnect at least. >>>>>>>>>>> I am not sure if this is 100% correct nor working. >>>>>>>>>>> >>>>>>>>>>> If you are using third party apps, why dont you do something like >>>>>>>>>>> system("env -i qsub ...") >>>>>>>>>>> with the right options to make qsub blocking or you manually >>>>>>>>>>> wait for the end of the job ? >>>>>>>>>>> >>>>>>>>>>> That looks like a much cleaner and simpler approach to me. >>>>>>>>>>> >>>>>>>>>>> Cheers, >>>>>>>>>>> >>>>>>>>>>> Gilles >>>>>>>>>>> >>>>>>>>>>> "Alex A. Schmidt" <a...@ufsm.br> wrote: >>>>>>>>>>> Hello Gilles, >>>>>>>>>>> >>>>>>>>>>> Ok, I believe I have a simple toy app running as I think it >>>>>>>>>>> should: >>>>>>>>>>> 'n' parent processes running under mpi_comm_world, each one >>>>>>>>>>> spawning its own 'm' child processes (each child group work >>>>>>>>>>> together nicely, returning the expected result for an mpi_ >>>>>>>>>>> allreduce call). >>>>>>>>>>> >>>>>>>>>>> Now, as I mentioned before, the apps I want to run in the >>>>>>>>>>> spawned >>>>>>>>>>> processes are third party mpi apps and I don't think it will be >>>>>>>>>>> possible >>>>>>>>>>> to exchange messages with them from my app. So, I do I tell >>>>>>>>>>> when the spawned processes have finnished running? All I have to >>>>>>>>>>> work >>>>>>>>>>> with is the intercommunicator returned from the mpi_comm_spawn >>>>>>>>>>> call... >>>>>>>>>>> >>>>>>>>>>> Alex >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> 2014-12-12 2:42 GMT-02:00 Alex A. Schmidt <a...@ufsm.br>: >>>>>>>>>>>> >>>>>>>>>>>> Gilles, >>>>>>>>>>>> >>>>>>>>>>>> Well, yes, I guess.... >>>>>>>>>>>> >>>>>>>>>>>> I'll do tests with the real third party apps and let you know. >>>>>>>>>>>> These are huge quantum chemistry codes (dftb+, siesta and >>>>>>>>>>>> Gaussian) >>>>>>>>>>>> which greatly benefits from a parallel environment. My code is >>>>>>>>>>>> just >>>>>>>>>>>> a front end to use those, but since we have a lot of data to >>>>>>>>>>>> process >>>>>>>>>>>> it also benefits from a parallel environment. >>>>>>>>>>>> >>>>>>>>>>>> Alex >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> 2014-12-12 2:30 GMT-02:00 Gilles Gouaillardet < >>>>>>>>>>>> gilles.gouaillar...@iferc.org>: >>>>>>>>>>>>> >>>>>>>>>>>>> Alex, >>>>>>>>>>>>> >>>>>>>>>>>>> just to make sure ... >>>>>>>>>>>>> this is the behavior you expected, right ? >>>>>>>>>>>>> >>>>>>>>>>>>> Cheers, >>>>>>>>>>>>> >>>>>>>>>>>>> Gilles >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On 2014/12/12 13:27, Alex A. Schmidt wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> Gilles, >>>>>>>>>>>>> >>>>>>>>>>>>> Ok, very nice! >>>>>>>>>>>>> >>>>>>>>>>>>> When I excute >>>>>>>>>>>>> >>>>>>>>>>>>> do rank=1,3 >>>>>>>>>>>>> call MPI_Comm_spawn('hello_world',' >>>>>>>>>>>>> ',5,MPI_INFO_NULL,rank,MPI_COMM_WORLD,my_intercomm,MPI_ERRCODES_IGNORE,status) >>>>>>>>>>>>> enddo >>>>>>>>>>>>> >>>>>>>>>>>>> I do get 15 instances of the 'hello_world' app running: 5 for >>>>>>>>>>>>> each parent >>>>>>>>>>>>> rank 1, 2 and 3. >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks a lot, Gilles. >>>>>>>>>>>>> >>>>>>>>>>>>> Best regargs, >>>>>>>>>>>>> >>>>>>>>>>>>> Alex >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> 2014-12-12 1:32 GMT-02:00 Gilles Gouaillardet >>>>>>>>>>>>> <gilles.gouaillar...@iferc.org >>>>>>>>>>>>> >>>>>>>>>>>>> : >>>>>>>>>>>>> >>>>>>>>>>>>> Alex, >>>>>>>>>>>>> >>>>>>>>>>>>> just ask MPI_Comm_spawn to start (up to) 5 tasks via the maxprocs >>>>>>>>>>>>> parameter : >>>>>>>>>>>>> >>>>>>>>>>>>> int MPI_Comm_spawn(char *command, char *argv[], int >>>>>>>>>>>>> maxprocs, >>>>>>>>>>>>> MPI_Info info, >>>>>>>>>>>>> int root, MPI_Comm comm, MPI_Comm >>>>>>>>>>>>> *intercomm, >>>>>>>>>>>>> int array_of_errcodes[]) >>>>>>>>>>>>> >>>>>>>>>>>>> INPUT PARAMETERS >>>>>>>>>>>>> maxprocs >>>>>>>>>>>>> - maximum number of processes to start (integer, >>>>>>>>>>>>> significant >>>>>>>>>>>>> only at root) >>>>>>>>>>>>> >>>>>>>>>>>>> Cheers, >>>>>>>>>>>>> >>>>>>>>>>>>> Gilles >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On 2014/12/12 12:23, Alex A. Schmidt wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> Hello Gilles, >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks for your reply. The "env -i PATH=..." stuff seems to >>>>>>>>>>>>> work!!! >>>>>>>>>>>>> >>>>>>>>>>>>> call system("sh -c 'env -i PATH=/usr/lib64/openmpi/bin:/bin >>>>>>>>>>>>> mpirun -n 2 >>>>>>>>>>>>> hello_world' ") >>>>>>>>>>>>> >>>>>>>>>>>>> did produce the expected result with a simple openmi >>>>>>>>>>>>> "hello_world" code I >>>>>>>>>>>>> wrote. >>>>>>>>>>>>> >>>>>>>>>>>>> I might be harder though with the real third party app I have in >>>>>>>>>>>>> mind. And >>>>>>>>>>>>> I realize >>>>>>>>>>>>> getting passed over a job scheduler with this approach might not >>>>>>>>>>>>> work at >>>>>>>>>>>>> all... >>>>>>>>>>>>> >>>>>>>>>>>>> I have looked at the MPI_Comm_spawn call but I failed to >>>>>>>>>>>>> understand how it >>>>>>>>>>>>> could help here. For instance, can I use it to launch an mpi app >>>>>>>>>>>>> with the >>>>>>>>>>>>> option "-n 5" ? >>>>>>>>>>>>> >>>>>>>>>>>>> Alex >>>>>>>>>>>>> >>>>>>>>>>>>> 2014-12-12 0:36 GMT-02:00 Gilles Gouaillardet >>>>>>>>>>>>> <gilles.gouaillar...@iferc.org >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> : >>>>>>>>>>>>> >>>>>>>>>>>>> Alex, >>>>>>>>>>>>> >>>>>>>>>>>>> can you try something like >>>>>>>>>>>>> call system(sh -c 'env -i /.../mpirun -np 2 /.../app_name') >>>>>>>>>>>>> >>>>>>>>>>>>> -i start with an empty environment >>>>>>>>>>>>> that being said, you might need to set a few environment variables >>>>>>>>>>>>> manually : >>>>>>>>>>>>> env -i PATH=/bin ... >>>>>>>>>>>>> >>>>>>>>>>>>> and that being also said, this "trick" could be just a bad idea : >>>>>>>>>>>>> you might be using a scheduler, and if you empty the environment, >>>>>>>>>>>>> the >>>>>>>>>>>>> scheduler >>>>>>>>>>>>> will not be aware of the "inside" run. >>>>>>>>>>>>> >>>>>>>>>>>>> on top of that, invoking system might fail depending on the >>>>>>>>>>>>> interconnect >>>>>>>>>>>>> you use. >>>>>>>>>>>>> >>>>>>>>>>>>> Bottom line, i believe Ralph's reply is still valid, even if five >>>>>>>>>>>>> years >>>>>>>>>>>>> have passed : >>>>>>>>>>>>> changing your workflow, or using MPI_Comm_spawn is a much better >>>>>>>>>>>>> approach. >>>>>>>>>>>>> >>>>>>>>>>>>> Cheers, >>>>>>>>>>>>> >>>>>>>>>>>>> Gilles >>>>>>>>>>>>> >>>>>>>>>>>>> On 2014/12/12 11:22, Alex A. Schmidt wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> Dear OpenMPI users, >>>>>>>>>>>>> >>>>>>>>>>>>> Regarding to this previous >>>>>>>>>>>>> post<http://www.open-mpi.org/community/lists/users/2009/06/9560.php> >>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> >>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> >>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> >>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> >>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> >>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> >>>>>>>>>>>>> <http://www.open-mpi.org/community/lists/users/2009/06/9560.php> >>>>>>>>>>>>> from 2009, >>>>>>>>>>>>> I wonder if the reply >>>>>>>>>>>>> from Ralph Castain is still valid. My need is similar but quite >>>>>>>>>>>>> simpler: >>>>>>>>>>>>> to make a system call from an openmpi fortran application to run a >>>>>>>>>>>>> third party openmpi application. I don't need to exchange mpi >>>>>>>>>>>>> messages >>>>>>>>>>>>> with the application. I just need to read the resulting output >>>>>>>>>>>>> file >>>>>>>>>>>>> generated >>>>>>>>>>>>> by it. I have tried to do the following system call from my >>>>>>>>>>>>> fortran openmpi >>>>>>>>>>>>> code: >>>>>>>>>>>>> >>>>>>>>>>>>> call system("sh -c 'mpirun -n 2 app_name") >>>>>>>>>>>>> >>>>>>>>>>>>> but I get >>>>>>>>>>>>> >>>>>>>>>>>>> ********************************************************** >>>>>>>>>>>>> >>>>>>>>>>>>> Open MPI does not support recursive calls of mpirun >>>>>>>>>>>>> >>>>>>>>>>>>> ********************************************************** >>>>>>>>>>>>> >>>>>>>>>>>>> Is there a way to make this work? >>>>>>>>>>>>> >>>>>>>>>>>>> Best regards, >>>>>>>>>>>>> >>>>>>>>>>>>> Alex >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing listus...@open-mpi.org >>>>>>>>>>>>> >>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25966.php >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing listus...@open-mpi.org >>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>> Link to this >>>>>>>>>>>>> post:http://www.open-mpi.org/community/lists/users/2014/12/25967.php >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing listus...@open-mpi.org >>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>> >>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25968.php >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing listus...@open-mpi.org >>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>> Link to this >>>>>>>>>>>>> post:http://www.open-mpi.org/community/lists/users/2014/12/25969.php >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing listus...@open-mpi.org >>>>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25970.php >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing list >>>>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>>>> Subscription: >>>>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>>>> Link to this post: >>>>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25971.php >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> users mailing list >>>>>>>>>>> us...@open-mpi.org >>>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>>> Link to this post: >>>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25974.php >>>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> us...@open-mpi.org >>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>> Link to this post: >>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25975.php >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> us...@open-mpi.org >>>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>>> Link to this post: >>>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25978.php >>>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> us...@open-mpi.org >>>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>>> Link to this post: >>>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25979.php >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> us...@open-mpi.org >>>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>>> Link to this post: >>>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25981.php >>>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> us...@open-mpi.org >>>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>>> Link to this post: >>>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25982.php >>>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> us...@open-mpi.org >>>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>>> Link to this post: >>>>>> http://www.open-mpi.org/community/lists/users/2014/12/25991.php >>>>>> >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org >>>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> Link to this post: >>>>> http://www.open-mpi.org/community/lists/users/2014/12/26002.php >>>>> >>>> >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> Link to this post: >>>> http://www.open-mpi.org/community/lists/users/2014/12/26003.php >>>> >>> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2014/12/26010.php >> > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2014/12/26011.php >