Hi Jeff I ran the NAS Parallel Bechmark and it gives for me -bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$ mpirun -np 1 cg.A.1 -------------------------------------------------------------------------- [0,1,0]: uDAPL on host SERVSOLARIS was unable to find any NICs. Another transport will be used instead, although this may result in lower performance. -------------------------------------------------------------------------- NAS Parallel Benchmarks 3.2 -- CG Benchmark
Size: 14000 Iterations: 15 Number of active processes: 1 Number of nonzeroes per row: 11 Eigenvalue shift: .200E+02 Benchmark completed VERIFICATION SUCCESSFUL Zeta is 0.171302350540E+02 Error is 0.512264003323E-13 CG Benchmark Completed. Class = A Size = 14000 Iterations = 15 Time in seconds = 3.02 Total processes = 1 Compiled procs = 1 Mop/s total = 495.93 Mop/s/process = 495.93 Operation type = floating point Verification = SUCCESSFUL Version = 3.2 Compile date = 11 Jun 2007 -bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$ mpirun -np 2 cg.A.2 -------------------------------------------------------------------------- [0,1,0]: uDAPL on host SERVSOLARIS was unable to find any NICs. Another transport will be used instead, although this may result in lower performance. -------------------------------------------------------------------------- -------------------------------------------------------------------------- [0,1,1]: uDAPL on host SERVSOLARIS was unable to find any NICs. Another transport will be used instead, although this may result in lower performance. -------------------------------------------------------------------------- NAS Parallel Benchmarks 3.2 -- CG Benchmark Size: 14000 Iterations: 15 Number of active processes: 2 Number of nonzeroes per row: 11 Eigenvalue shift: .200E+02 Benchmark completed VERIFICATION SUCCESSFUL Zeta is 0.171302350540E+02 Error is 0.522633719989E-13 CG Benchmark Completed. Class = A Size = 14000 Iterations = 15 Time in seconds = 2.47 Total processes = 2 Compiled procs = 2 Mop/s total = 606.32 Mop/s/process = 303.16 Operation type = floating point Verification = SUCCESSFUL Version = 3.2 Compile date = 11 Jun 2007 You can remark that the scalling is not so good like yours. Maibe I am having comunications problems between processors. You can also remark that I am faster on one process concared to your processor. Victor --- Jeff Pummill <jpum...@uark.edu> wrote: > Perfect! Thanks Jeff! > > The NAS Parallel Benchmark on a dual core AMD > machine now returns this... > [jpummil@localhost bin]$ mpirun -np 1 cg.A.1 > NAS Parallel Benchmarks 3.2 -- CG Benchmark > CG Benchmark Completed. > Class = A > Size = 14000 > Iterations = 15 > Time in seconds = 4.75 > Total processes = 1 > Compiled procs = 1 > Mop/s total = 315.32 > > ...and... > > [jpummil@localhost bin]$ mpirun -np 2 cg.A.2 > NAS Parallel Benchmarks 3.2 -- CG Benchmark > CG Benchmark Completed. > Class = A > Size = 14000 > Iterations = 15 > Time in seconds = 2.48 > Total processes = 2 > Compiled procs = 2 > Mop/s total = 604.46 > > Not quite linear, but one must account for all of > the OS traffic that > one core or the other must deal with. > > > Jeff F. Pummill > Senior Linux Cluster Administrator > University of Arkansas > Fayetteville, Arkansas 72701 > (479) 575 - 4590 > http://hpc.uark.edu > > "A supercomputer is a device for turning > compute-bound > problems into I/O-bound problems." -Seymour Cray > > > Jeff Squyres wrote: > > Just remove the -L and -l arguments -- OMPI's > "mpif90" (and other > > wrapper compilers) will do all that magic for you. > > > > Many -L/-l arguments in MPI application Makefiles > are throwbacks to > > older versions of MPICH wrapper compilers that > didn't always work > > properly. Those days are long gone; most (all?) > MPI wrapper > > compilers do not need you to specify -L/-l these > days. > > > > > > > > On Jun 10, 2007, at 3:08 PM, Jeff Pummill wrote: > > > > > >> Maybe the "dumb question" of the week, but here > goes... > >> > >> I am trying to compile a piece of code (NPB) > under OpenMPI and I am > >> having a problem with specifying the right > library. Possibly > >> something I need to define in a LD_LIBRARY_PATH > statement? > >> > >> Using Gnu mpich, the line looked like this... > >> > >> FMPI_LIB = -L/opt/mpich/gnu/lib/ -lmpich > >> > >> I tried to replace this with... > >> > >> FMPI_LIB = -L/usr/lib/openmpi/ -llibmpi > >> > >> to which the make responded... > >> mpif90 -O -o ../bin/cg.A.2 cg.o > ../common/print_results.o ../common/ > >> randdp.o ../common/timers.o -L/usr/lib/openmpi/ > -llibmpi > >> /usr/bin/ld: cannot find -llibmpi > >> collect2: ld returned 1 exit status > >> > >> Wrong library file? Setup or path issue? > >> > >> -- > >> Jeff F. Pummill > >> Senior Linux Cluster Administrator > >> University of Arkansas > >> > >> _______________________________________________ > >> users mailing list > >> us...@open-mpi.org > >> > http://www.open-mpi.org/mailman/listinfo.cgi/users > >> > > > > > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users ____________________________________________________________________________________ Get your own web address. Have a HUGE year through Yahoo! Small Business. http://smallbusiness.yahoo.com/domains/?p=BESTDEAL