Hi Ralf, sorry for the delay in the answer but I encountered some difficulties to access to internet since yesterday.
I have tried all your suggestions but I continue to experience problems. Actually, I have a problem with bjs on the one hand that I may submit to a bproc forum and I still spawn problem on the other hand. Let's focus first on the spawn problem. Even with a "bjssub -i bash" or "bjssub -n 1 -i bash" command, I continue to have the log: mpirun -np 1 main_exe machine10 main_exe: Begining of main_exe main_exe: Call MPI_Init main_exe: MPI_Info_set soft result=0 main_exe: MPI_Info_set node result=0 main_exe: Call MPI_Comm_spawn_multiple() -------------------------------------------------------------------------- Some of the requested hosts are not included in the current allocation for the application: ./spawned_exe The requested hosts were: machine10 Verify that you have mapped the allocated resources properly using the --host specification. -------------------------------------------------------------------------- [setics10:07250] [0,0,0] ORTE_ERROR_LOG: Out of resource in file base/rmaps_base_node.c at line 210 [setics10:07250] [0,0,0] ORTE_ERROR_LOG: Out of resource in file rmaps_rr.c at line 331 This problem is observed whatever the slave node is on the same machine than the master or not. On the bjs side of the problem. I have run bjssub under gdb and I could observed that I did not go into the code part that setenv NODES variable, so I stayed with the default value NODES=0. The question is, is the spawn problem a result of the bjs problem ? or are they two independant problems ? The good thing would be to find some other people with a debian platform, bproc, bjs and openmpi active. So that, we could check if I have made something wrong during the installation phase or if there is really a incompatibility problem in open mpi. Thank you so much for all you support, I wish it is not succesful yet. Regards. Herve List-Post: users@lists.open-mpi.org Date: Fri, 03 Nov 2006 14:10:20 -0700 From: Ralph H Castain <r...@lanl.gov> Subject: Re: [OMPI users] MPI_Comm_spawn multiple bproc support To: "Open MPI Users <us...@open-mpi.org>" <us...@open-mpi.org> Message-ID: <c170fe4c.59b3%...@lanl.gov> Content-Type: text/plain; charset="ISO-8859-1" Okay, I picked up some further info that may help you. >> The "bjsub -i /bin/env" only sets up the NODES for the session of >> /bin/env. Probably what he wants is "bjssub -i /bin/bash" and start >> bpsh/mpirun from the new shell. I would recommend doing as they suggest. Also, they noted that you failed to specify the number of nodes you wanted on the bjssub command line. As a result, the system gave you only one node (hence the NODES=0 instead of NODES=0, 1). If you do a "man bjssub", or a "bjssub --help", you should (hopefully) find out how to specify the desired number of nodes. Hope that helps. Ralph On 11/2/06 6:46 AM, "Ralph Castain" <r...@lanl.gov> wrote: > I truly appreciate your patience. Let me talk to some of our Bproc folks and > see if they can tell me what is going on. I agree - I would have expected > the NODES to be 0,1. The fact that you are getting just 0 explains the > behavior you are seeing with Open MPI. > > I also know (though I don't the command syntax) that you can get a long-term > allocation from bjs (i.e., one that continues until you logout). Let me dig > a little and see how that is done. > > Again, I appreciate your patience. > Ralph > > > On 11/2/06 6:32 AM, "hpe...@infonie.fr" <hpe...@infonie.fr> wrote: > >> I again Ralf, >> >>> I gather you have access to bjs? Could you use bjs to get a node allocation, >>> and then send me a printout of the environment? >> >> I have slightly changed my cluster configuration for something like: >> master is running on a machine call: machine10 >> node 0 is running on a machine call: machine10 (same as master then) >> node 1 is running on a machine call: machine14 >> >> node 0 and 1 are up >> >> My bjs configration allocates node 0 and 1 to the default pool >> <---------------> >> pool default >> policy simple >> nodes 0-1 >> <-----------------> >> >> Be default, when I run "env" in a terminal, NODES variable is not present. >> If I run env under a job submission command like "bjsub -i env", then I can >> see the following new environments variable. >> NODES=0 >> JOBID=27 (for instance) >> BPROC_RANK=0000000 >> BPROC_PROGNAME=/usr/bin/env >> >> When the command is over, NODES is unset again. >> >> What is strange is that I would have expected that NODES=0,1. I do not know >> if >> you bjs users have the same behaviour. >> >> Hopefully, it is the kind of information you were expecting. >> >> Regards. >> >> Herve >> >> >> >> >> --------------------- ALICE SECURITE ENFANTS --------------------- >> Prot?gez vos enfants des dangers d'Internet en installant S?curit? Enfants, >> le >> contr?le parental d'Alice. >> http://www.aliceadsl.fr/securitepc/default_copa.asp >> >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users ------------------------------ Message: 2 List-Post: users@lists.open-mpi.org Date: Sat, 4 Nov 2006 14:04:54 +0100 (CET) From: <pgar...@eside.deusto.es> Subject: [OMPI users] Technical inquiry To: us...@open-mpi.org Message-ID: <4444924729pgar...@eside.deusto.es> Content-Type: text/plain; charset="iso-8859-1" Hi, everydoby. Good afternoon. I've just configured and installed the openmpi-1.1.2 on a kubuntu GNU/linux, and I'm trying now to compile the hello.c example without results. > root@kubuntu:/home/livestrong/mpi/test# uname -a > Linux kubuntu 2.6.15-23-386 #1 PREEMPT Tue May 23 13:49:40 UTC 2006 > i686 GNU/Linux Hello.c ------- #include "/usr/lib/mpich-mpd/include/mpi.h" #include <stdio.h> int main (int argc, char** argv) { MPI_Init(&argc, &argv); printf("Hello word.\n"); MPI_Finalize(); return(0); } The error that I'm finding is this: root@kubuntu:/home/livestrong/mpi/prueba# mpirun -np 2 hello 0 - MPI_INIT : MPIRUN chose the wrong device ch_p4; program needs device ch_p4mpd /usr/lib/mpich/bin/mpirun.ch_p4: line 243: 16625 Segmentation fault "/home/livestrong/mpi/prueba/hello" -p4pg "/home/livestrong/mpi/prueba/PI16545" -p4wd "/home/livestrong/mpi/prueba" Does anybody know what it can be the problem? Regards and thank you very much in advance. Pablo. PD: I send the ompi_info output and the config.log to you. Besides -------------- next part -------------- A non-text attachment was scrubbed... Name: question.tar.gz Type: application/octet-stream Size: 59009 bytes Desc: Url : http://www.open-mpi.org/MailArchives/users/attachments/20061104/dd281cc5/attachment.obj ------------------------------ _______________________________________________ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users End of users Digest, Vol 425, Issue 1 ************************************* --------------------- ALICE SECURITE ENFANTS --------------------- Protégez vos enfants des dangers d'Internet en installant Sécurité Enfants, le contrôle parental d'Alice. http://www.aliceadsl.fr/securitepc/default_copa.asp