Hi Jeff
I ran the NAS Parallel Bechmark and it gives for me
-bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$
mpirun -np 1 cg.A.1
---------------------------------------------------------------------
-----
[0,1,0]: uDAPL on host SERVSOLARIS was unable to find
any NICs.
Another transport will be used instead, although this
may result in
lower performance.
---------------------------------------------------------------------
-----
NAS Parallel Benchmarks 3.2 -- CG Benchmark
Size: 14000
Iterations: 15
Number of active processes: 1
Number of nonzeroes per row: 11
Eigenvalue shift: .200E+02
Benchmark completed
VERIFICATION SUCCESSFUL
Zeta is 0.171302350540E+02
Error is 0.512264003323E-13
CG Benchmark Completed.
Class = A
Size = 14000
Iterations = 15
Time in seconds = 3.02
Total processes = 1
Compiled procs = 1
Mop/s total = 495.93
Mop/s/process = 495.93
Operation type = floating point
Verification = SUCCESSFUL
Version = 3.2
Compile date = 11 Jun 2007
-bash%/export/home/vmarian/fortran/benchmarks/NPB3.2/NPB3.2-MPI/bin$
mpirun -np 2 cg.A.2
---------------------------------------------------------------------
-----
[0,1,0]: uDAPL on host SERVSOLARIS was unable to find
any NICs.
Another transport will be used instead, although this
may result in
lower performance.
---------------------------------------------------------------------
-----
---------------------------------------------------------------------
-----
[0,1,1]: uDAPL on host SERVSOLARIS was unable to find
any NICs.
Another transport will be used instead, although this
may result in
lower performance.
---------------------------------------------------------------------
-----
NAS Parallel Benchmarks 3.2 -- CG Benchmark
Size: 14000
Iterations: 15
Number of active processes: 2
Number of nonzeroes per row: 11
Eigenvalue shift: .200E+02
Benchmark completed
VERIFICATION SUCCESSFUL
Zeta is 0.171302350540E+02
Error is 0.522633719989E-13
CG Benchmark Completed.
Class = A
Size = 14000
Iterations = 15
Time in seconds = 2.47
Total processes = 2
Compiled procs = 2
Mop/s total = 606.32
Mop/s/process = 303.16
Operation type = floating point
Verification = SUCCESSFUL
Version = 3.2
Compile date = 11 Jun 2007
You can remark that the scalling is not so good
like yours. Maibe I am having comunications problems
between processors.
You can also remark that I am faster on one process
concared to your processor.
Victor
--- Jeff Pummill <jpum...@uark.edu> wrote:
Perfect! Thanks Jeff!
The NAS Parallel Benchmark on a dual core AMD
machine now returns this...
[jpummil@localhost bin]$ mpirun -np 1 cg.A.1
NAS Parallel Benchmarks 3.2 -- CG Benchmark
CG Benchmark Completed.
Class = A
Size = 14000
Iterations = 15
Time in seconds = 4.75
Total processes = 1
Compiled procs = 1
Mop/s total = 315.32
...and...
[jpummil@localhost bin]$ mpirun -np 2 cg.A.2
NAS Parallel Benchmarks 3.2 -- CG Benchmark
CG Benchmark Completed.
Class = A
Size = 14000
Iterations = 15
Time in seconds = 2.48
Total processes = 2
Compiled procs = 2
Mop/s total = 604.46
Not quite linear, but one must account for all of
the OS traffic that
one core or the other must deal with.
Jeff F. Pummill
Senior Linux Cluster Administrator
University of Arkansas
Fayetteville, Arkansas 72701
(479) 575 - 4590
http://hpc.uark.edu
"A supercomputer is a device for turning
compute-bound
problems into I/O-bound problems." -Seymour Cray
Jeff Squyres wrote:
Just remove the -L and -l arguments -- OMPI's
"mpif90" (and other
wrapper compilers) will do all that magic for you.
Many -L/-l arguments in MPI application Makefiles
are throwbacks to
older versions of MPICH wrapper compilers that
didn't always work
properly. Those days are long gone; most (all?)
MPI wrapper
compilers do not need you to specify -L/-l these
days.
On Jun 10, 2007, at 3:08 PM, Jeff Pummill wrote:
Maybe the "dumb question" of the week, but here
goes...
I am trying to compile a piece of code (NPB)
under OpenMPI and I am
having a problem with specifying the right
library. Possibly
something I need to define in a LD_LIBRARY_PATH
statement?
Using Gnu mpich, the line looked like this...
FMPI_LIB = -L/opt/mpich/gnu/lib/ -lmpich
I tried to replace this with...
FMPI_LIB = -L/usr/lib/openmpi/ -llibmpi
to which the make responded...
mpif90 -O -o ../bin/cg.A.2 cg.o
../common/print_results.o ../common/
randdp.o ../common/timers.o -L/usr/lib/openmpi/
-llibmpi
/usr/bin/ld: cannot find -llibmpi
collect2: ld returned 1 exit status
Wrong library file? Setup or path issue?
--
Jeff F. Pummill
Senior Linux Cluster Administrator
University of Arkansas
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
_____________________________________________________________________
_______________
Get your own web address.
Have a HUGE year through Yahoo! Small Business.
http://smallbusiness.yahoo.com/domains/?p=BESTDEAL
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users