It confirms what we were saying - your application is not compiled against OMPI
You'll have to dig a little to figure out why that is happening - could be a path issue. On Apr 29, 2010, at 2:56 AM, Pankatz, Klaus wrote: > Hi Ralph, > > thanks for your advise. Finally I configured openmpi with ./configure > --prefix=... --enable-debug. > According to your suggestion I ran the hello_world with > mpirun -np 4 -mca ess_base_verbose 5 and the output is as follows: > I don't know what happend there... > > ****** > [marvin:00373] mca:base:select:( ess) Querying component [env] > [marvin:00373] mca:base:select:( ess) Skipping component [env]. Query failed > to return a module > [marvin:00373] mca:base:select:( ess) Querying component [hnp] > [marvin:00373] mca:base:select:( ess) Query of component [hnp] set priority > to 100 > [marvin:00373] mca:base:select:( ess) Querying component [singleton] > [marvin:00373] mca:base:select:( ess) Skipping component [singleton]. Query > failed to return a module > [marvin:00373] mca:base:select:( ess) Querying component [slurm] > [marvin:00373] mca:base:select:( ess) Skipping component [slurm]. Query > failed to return a module > [marvin:00373] mca:base:select:( ess) Querying component [tool] > [marvin:00373] mca:base:select:( ess) Skipping component [tool]. Query > failed to return a module > [marvin:00373] mca:base:select:( ess) Selected component [hnp] > Hello World! I'm number 0 of 1 running on host marvin > Hello World! I'm number 0 of 1 running on host marvin > Hello World! I'm number 0 of 1 running on host marvin > Hello World! I'm number 0 of 1 running on host marvin > **** > > ________________________________________ > Von: users-boun...@open-mpi.org [users-boun...@open-mpi.org] im Auftrag von > Ralph Castain [r...@open-mpi.org] > Gesendet: Freitag, 23. April 2010 17:04 > An: Open MPI Users > Betreff: Re: [OMPI users] mpirun -np 4 hello_world; on a eight processor > shared memory machine produces wrong output > > Is this build configured --enable-debug? If not, can you reconfigure it? > > If you can, you could run it with -mca ess_base_verbose 5 to see if it is > picking up the correct modules. > > It really looks like your application was built with an older version, or > compiled against something like mpich. > > On Apr 23, 2010, at 8:54 AM, Pankatz, Klaus wrote: > > Allright, I've ran a mpirun -np 4 env. And I see OMPI_COMM_WORLD_RANK 0 to 3. > So far so good. > OMPI_COMM_WORLD_SIZE=4 everytime, I think thats correct. > OMPI_MCA_mpi_yield_when_idle=0 everytime zero > OMPI_MCA_orte_app_num=0 everytime zero > > Am 23.04.2010 um 14:54 schrieb Terry Dontje: > > Ok can you do an "mpirun -np 4 env" you should seeOMPI_COMM_WORLD_RANK range > 0 thru 3. I am curious if you even see OMPI_* env-vars and if you do is this > one 0 for all procs? > > --td > > Pankatz, Klaus wrote: > > Yeah, I sure that I use the right mpirun. > > which mpirun leads to /usr/users/pankatz/OPENmpi/bin/mpirun which is the > right one. > ________________________________________ > Von: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org> > [users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>] im Auftrag > von Terry Dontje [terry.don...@oracle.com<mailto:terry.don...@oracle.com>] > Gesendet: Freitag, 23. April 2010 14:29 > An: Open MPI Users > Betreff: Re: [OMPI users] mpirun -np 4 hello_world; on a eight processor > shared memory machine produces wrong output > > This looks like you are using an mpirun or mpiexec from mvapich to run an > executable compiled with OMPI. Can you make sure that you are using the > right mpirun? > > --td > > Pankatz, Klaus wrote: > > Yes, I did that. > > It ist basically the same problem with a Fortran version of this little > program. With that I used the mpif90 command of openMPI. > ________________________________________ > Von: > users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto:users-boun...@open-mpi.org><mailto:users-boun...@open-mpi.org> > > [users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto:users-boun...@open-mpi.org><mailto:users-boun...@open-mpi.org>] > im Auftrag von Reuti > [re...@staff.uni-marburg.de<mailto:re...@staff.uni-marburg.de><mailto:re...@staff.uni-marburg.de><mailto:re...@staff.uni-marburg.de>] > Gesendet: Freitag, 23. April 2010 14:15 > An: Open MPI Users > Betreff: Re: [OMPI users] mpirun -np 4 hello_world; on a eight processor > shared memory machine produces wrong output > > Hi, > > Am 23.04.2010 um 14:06 schrieb Pankatz, Klaus: > > > > Hi all, > > there's a problem with openMPI on my machine. When I simply try to run this > little hello_world-program on multiple processors, the output isn't as > expected: > ***** > C code: > #include <mpi.h> > #include <stdio.h> > #include <unistd.h> > int main(int argc, char **argv) > { > int size,rank; > char hostname[50]; > MPI_Init(&argc,&argv); > MPI_Comm_rank(MPI_COMM_WORLD, &rank); //Who am I? > MPI_Comm_size(MPI_COMM_WORLD, &size); //How many processes? > gethostname (hostname, 50); > printf ("Hello World! I'm number %2d of %2d running on host %s\n", > rank, size, hostname); > MPI_Finalize(); > return 0; > } > **** > > Command: mpirun -np 4 a.out > > > > the mpirun (better, use: mpiexec) is the one from the Open MPI, and you also > used its version mpicc to compile the program? > > -- Reuti > > > > > Output: > Hello World! I'm number 0 of 1 running on host marvin > Hello World! I'm number 0 of 1 running on host marvin > Hello World! I'm number 0 of 1 running on host marvin > Hello World! I'm number 0 of 1 running on host marvin > > It should be more or less: > Hello World! I'm number 1 of 4 running on host marvin > Hello World! I'm number 2 of 4 running on host marvin > .... > > OpenMPI-version 1.4.1 compiled with Lahey Fortran 95 (lf95). > OpenMPI was compiled "out of the box" only changing to the Lahey compiler > with a setenv $FC lf95 > > The System: Linux marvin 2.6.27.6-1 #1 SMP Sat Nov 15 20:19:04 CET 2008 > x86_64 GNU/Linux > > Compiler: Lahey/Fujitsu Linux64 Fortran Compiler Release L8.10a > > Thanks very much! > Klaus > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org><mailto:us...@open-mpi.org> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org><mailto:us...@open-mpi.org> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org><mailto:us...@open-mpi.org> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > -- > [cid:part1.05010106.04050301@oracle.com] > Terry D. Dontje | Principal Software Engineer > Developer Tools Engineering | +1.650.633.7054 > Oracle - Performance Technologies > 95 Network Drive, Burlington, MA 01803 > Email > terry.don...@oracle.com<mailto:terry.don...@oracle.com><mailto:terry.don...@oracle.com><mailto:terry.don...@oracle.com> > > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > -- > <ATT00001..gif> > Terry D. Dontje | Principal Software Engineer > Developer Tools Engineering | +1.650.633.7054 > Oracle - Performance Technologies > 95 Network Drive, Burlington, MA 01803 > Email terry.don...@oracle.com<mailto:terry.don...@oracle.com> > > <ATT00002..txt> > > _______________________________________________ > users mailing list > us...@open-mpi.org<mailto:us...@open-mpi.org> > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users