http://www.open-mpi.org/nightly/trunk/
I'm not sure this patch will solve your problem, but it is worth a try. On Mar 1, 2010, at 3:51 AM, Federico Golfrè Andreasi wrote: > Ok, thank you ! > > where can I found instructions for download the developer's copy of OpenMPI, > if it is possibile? > > I'd like to test it just to be sure that the problem is solved, with that > patch. > > Can you let me know where that patch is available? > > Thank you very much, > > Federico > > > > > 2010/2/27 Ralph Castain <r...@open-mpi.org> > Okay, thanks. It's the same problem as the other person encountered. > Basically, it looks to OMPI as if you are launching > 128 independent app > contexts, and our arrays were limited to 128. > > He has provided a patch that I'll review (couple of things I'd rather change) > and then apply to our developer's trunk. I would expect it to migrate over to > the 1.4 release series at some point (can't guarantee which one). > > > On Feb 27, 2010, at 6:47 AM, Federico Golfrè Andreasi wrote: > >> Hi, >> >> the program is executed as one application on 129 cpus defined by the >> hostfile. >> Than rank 0, inside the code, execute another program with 129 cpus, with a >> one-to-one relation, rank0 of the spawined process runs on the same host of >> rank0 of the spawning one and so on... >> Excuting the spawning program does not give any problem, >> but in the moment of spawning (with more than 128 cpus) it holds. >> >> Thank you! >> >> Federico >> >> >> >> >> 2010/2/27 Ralph Castain <r...@open-mpi.org> >> Since another user was doing something that caused a similar problem, >> perhaps we are missing a key piece of info here. Are you launching one >> app_context across 128 nodes? Or are you launching 128 app_contexts, each on >> a separate node? >> >> >> On Feb 26, 2010, at 10:23 AM, Federico Golfrè Andreasi wrote: >> >>> I'm doing some tests and it seems that is not able to do a spawn multiple >>> with more than 128 nodes. >>> >>> It just hold, with no error message. >>> >>> What do you think? What can I try to understand the problem. >>> >>> Thanks, >>> >>> Federico >>> >>> >>> >>> >>> 2010/2/26 Ralph Castain <r...@open-mpi.org> >>> No known limitations of which we are aware...the variables are all set to >>> int32_t, so INT32_MAX would be the only limit I can imagine. In which case, >>> you'll run out of memory long before you hit it. >>> >>> >>> 2010/2/26 Federico Golfrè Andreasi <federico.gol...@gmail.com> >>> HI ! >>> >>> have you ever did some analysis to understand if there is a limitation in >>> the number of nodes usable with OpenMPI-v1.4 ? >>> Using also the functions MPI_Comm_spawn o MPI_Comm_spawn_multiple. >>> >>> Thanks, >>> Federico >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users