I didn't use OpenMPI.
On Mon, Apr 14, 2014 at 2:37 PM, Jeff Squyres (jsquyres) <jsquy...@cisco.com > wrote: > This can also happen when you compile your application with one MPI > implementation (e.g., Open MPI), but then mistakenly use the "mpirun" (or > "mpiexec") from a different MPI implementation (e.g., MPICH). > > > On Apr 14, 2014, at 2:32 PM, Djordje Romanic <djord...@gmail.com> wrote: > > > I compiled it with: x86_64 Linux, gfortran compiler with gcc (dmpar). > dmpar - distributed memory option. > > > > Attached is the self-generated configuration file. The architecture > specification settings start at line 107. I didn't use Open MPI (shared > memory option). > > > > > > On Mon, Apr 14, 2014 at 1:23 PM, Dave Goodell (dgoodell) < > dgood...@cisco.com> wrote: > > On Apr 14, 2014, at 12:15 PM, Djordje Romanic <djord...@gmail.com> > wrote: > > > > > When I start wrf with mpirun -np 4 ./wrf.exe, I get this: > > > ------------------------------------------------- > > > starting wrf task 0 of 1 > > > starting wrf task 0 of 1 > > > starting wrf task 0 of 1 > > > starting wrf task 0 of 1 > > > ------------------------------------------------- > > > This indicates that it is not using 4 processors, but 1. > > > > > > Any idea what might be the problem? > > > > It could be that you compiled WRF with a different MPI implementation > than you are using to run it (e.g., MPICH vs. Open MPI). > > > > -Dave > > > > _______________________________________________ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > <configure.wrf>_______________________________________________ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > -- > Jeff Squyres > jsquy...@cisco.com > For corporate legal information go to: > http://www.cisco.com/web/about/doing_business/legal/cri/ > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >