Re: [OMPI users] compiling mpptest using OpenMPI

2007-02-16 Thread Eric Thibodeau
Hello devs,

Thought I would let you know there seems to be a problem with 1.2b3r13112 when 
running the "bisection" test on a Tyan VX50 machine (the 8 DualCore model with 
32Gigs of RAM).

OpenMPI was compiled with (as seen from config.log):
configure:116866: running /bin/sh './configure'  CFLAGS="-O3 -DNDEBUG 
-finline-functions -fno-strict-aliasing -pthread" CPPFLAGS=" " FFLAGS="" 
LDFLAGS=" " --enable-shared --disable-static  
--prefix=/export/livia/home/parallel/eric/openmpi_x86_64 --with-mpi=open_mpi 
--cache-file=/dev/null --srcdir=.

MPPTEST (1.3c) was compiled with:
./configure --with-mpi=$HOME/openmpi_`uname -m`

...which, for some reason, works fine on that system that doesn't have any 
other MPI implementation (ie: doesn't have LAM-MPI as per this thread).

Then I ran a few tests but this one ran for over it's allowed time (1800 
seconds and was going over 50minutes...) and was up to 16Gigs of RAM:

orterun -np 9 ./perftest-1.3c/mpptest -max_run_time 1800 -bisect -size 0 4096 1 
-gnuplot -fname HyperTransport/Global_bisect_0_4096_1.gpl

I had to CTRL-\ the process as CTRL-C wasn't sufficient. 2 mpptest processes 
and 1 orterun process were using 100% CPU ou of of the 16 cores. 

If any of this can be indicative of an OpenMPI bug and if I can help in 
tracking it down, don't hesitate to ask for details.

And, finally, Anthony, thanks for the MPICC and --with-mpich pointers, I will 
try those to simplify the build process!

Eric

Le jeudi 15 février 2007 19:51, Anthony Chan a écrit :
> 
> As long as mpicc is working, try configuring mpptest as
> 
> mpptest/configure MPICC=/bin/mpicc
> 
> or
> 
> mpptest/configure  --with-mpich=
> 
> A.Chan
> 
> On Thu, 15 Feb 2007, Eric Thibodeau wrote:
> 
> > Hi Jeff,
> >
> > Thanks for your response, I eventually figured it out, here is the
> > only way I got mpptest to compile:
> >
> > export LD_LIBRARY_PATH="$HOME/openmpi_`uname -m`/lib"
> > CC="$HOME/openmpi_`uname -m`/bin/mpicc" ./configure 
> > --with-mpi="$HOME/openmpi_`uname -m`"
> >
> > And, yes I know I should use the mpicc wrapper and all (I do RTFM :P ) but
> > mpptest is less than cooperative and hasn't been updated lately AFAIK.
> >
> > I'll keep you posted on some results as I get some results out (testing
> > TCP/IP as well as the HyperTransport on a Tyan Beast). Up to now, LAM-MPI
> > seems less efficient at async communications and shows no improovments
> > with persistant communications under TCP/IP. OpenMPI, on the other hand,
> > seems more efficient using persistant communications when in a
> > HyperTransport (shmem) environment... I know I am crossing many test
> > boudaries but I will post some PNGs of my results (as well as how I got to
> > them ;)
> >
> > Eric
> >
> > On Thu, 15 Feb 2007, Jeff Squyres wrote:
> >
> > > I think you want to add $HOME/openmpi_`uname -m`/lib to your
> > > LD_LIBRARY_PATH.  This should allow executables created by mpicc (or
> > > any derivation thereof, such as extracting flags via showme) to find
> > > the Right shared libraries.
> > >
> > > Let us know if that works for you.
> > >
> > > FWIW, we do recommend using the wrapper compilers over extracting the
> > > flags via --showme whenever possible (it's just simpler and should do
> > > what you need).
> > >
> > >
> > > On Feb 15, 2007, at 3:38 PM, Eric Thibodeau wrote:
> > >
> > > > Hello all,
> > > >
> > > >
> > > > I have been attempting to compile mpptest on my nodes in vain. Here
> > > > is my current setup:
> > > >
> > > >
> > > > Openmpi is in "$HOME/openmpi_`uname -m`" which translates to "/
> > > > export/home/eric/openmpi_i686/". I tried the following approaches
> > > > (you can see some of these were out of desperation):
> > > >
> > > >
> > > > CFLAGS=`mpicc --showme:compile` LDFLAGS=`mpicc --showme:link` ./
> > > > configure
> > > >
> > > >
> > > > Configure fails on:
> > > >
> > > > checking whether the C compiler works... configure: error: cannot
> > > > run C compiled programs.
> > > >
> > > >
> > > > The log shows that:
> > > >
> > > > ./a.out: error while loading shared libraries: liborte.so.0: cannot
> > > > open shared object file: No such file or directory
> > > >
> > > >
> > > >
> > > > CC="/export/home/eric/openmpi_i686/bin/mpicc" ./configure --with-
> > > > mpi=$HOME/openmpi_`uname -m`
> > > >
> > > > Same problems as above...
> > > >
> > > >
> > > > LDFLAGS="$HOME/openmpi_`uname -m`/lib" ./configure --with-mpi=$HOME/
> > > > openmpi_`uname -m`
> > > >
> > > >
> > > > Configure fails on:
> > > >
> > > > checking for C compiler default output file name... configure:
> > > > error: C compiler cannot create executables
> > > >
> > > >
> > > > And...finally (not that all of this was done in the presented order):
> > > >
> > > > ./configure --with-mpi=$HOME/openmpi_`uname -m`
> > > >
> > > >
> > > > Which ends with:
> > > >
> > > >
> > > > checking for library containing MPI_Init... no
> > > >
> > > > configure: error: Could not find MPI library
> > > >
> > > >
> > > > Anyone can help m

Re: [OMPI users] compiling mpptest using OpenMPI

2007-02-16 Thread Jeff Squyres
Could you try one of the later nightly 1.2 tarballs?  We just fixed a  
shared memory race condition, for example:


http://www.open-mpi.org/nightly/v1.2/


On Feb 16, 2007, at 12:12 AM, Eric Thibodeau wrote:


Hello devs,

Thought I would let you know there seems to be a problem with  
1.2b3r13112 when running the "bisection" test on a Tyan VX50  
machine (the 8 DualCore model with 32Gigs of RAM).


OpenMPI was compiled with (as seen from config.log):
configure:116866: running /bin/sh './configure'  CFLAGS="-O3 - 
DNDEBUG -finline-functions -fno-strict-aliasing -pthread"  
CPPFLAGS=" " FFLAGS="" LDFLAGS=" " --enable-shared --disable- 
static  --prefix=/export/livia/home/parallel/eric/openmpi_x86_64 -- 
with-mpi=open_mpi --cache-file=/dev/null --srcdir=.


MPPTEST (1.3c) was compiled with:
./configure --with-mpi=$HOME/openmpi_`uname -m`

...which, for some reason, works fine on that system that doesn't  
have any other MPI implementation (ie: doesn't have LAM-MPI as per  
this thread).


Then I ran a few tests but this one ran for over it's allowed time  
(1800 seconds and was going over 50minutes...) and was up to 16Gigs  
of RAM:


orterun -np 9 ./perftest-1.3c/mpptest -max_run_time 1800 -bisect - 
size 0 4096 1 -gnuplot -fname HyperTransport/ 
Global_bisect_0_4096_1.gpl


I had to CTRL-\ the process as CTRL-C wasn't sufficient. 2 mpptest  
processes and 1 orterun process were using 100% CPU ou of of the 16  
cores.


If any of this can be indicative of an OpenMPI bug and if I can  
help in tracking it down, don't hesitate to ask for details.


And, finally, Anthony, thanks for the MPICC and --with-mpich  
pointers, I will try those to simplify the build process!


Eric

Le jeudi 15 février 2007 19:51, Anthony Chan a écrit :


As long as mpicc is working, try configuring mpptest as

mpptest/configure MPICC=/bin/mpicc

or

mpptest/configure  --with-mpich=

A.Chan

On Thu, 15 Feb 2007, Eric Thibodeau wrote:


Hi Jeff,

Thanks for your response, I eventually figured it out, here is the
only way I got mpptest to compile:

export LD_LIBRARY_PATH="$HOME/openmpi_`uname -m`/lib"
CC="$HOME/openmpi_`uname -m`/bin/mpicc" ./configure --with- 
mpi="$HOME/openmpi_`uname -m`"


And, yes I know I should use the mpicc wrapper and all (I do  
RTFM :P ) but
mpptest is less than cooperative and hasn't been updated lately  
AFAIK.


I'll keep you posted on some results as I get some results out  
(testing
TCP/IP as well as the HyperTransport on a Tyan Beast). Up to now,  
LAM-MPI
seems less efficient at async communications and shows no  
improovments
with persistant communications under TCP/IP. OpenMPI, on the  
other hand,

seems more efficient using persistant communications when in a
HyperTransport (shmem) environment... I know I am crossing many test
boudaries but I will post some PNGs of my results (as well as how  
I got to

them ;)

Eric

On Thu, 15 Feb 2007, Jeff Squyres wrote:


I think you want to add $HOME/openmpi_`uname -m`/lib to your
LD_LIBRARY_PATH.  This should allow executables created by mpicc  
(or
any derivation thereof, such as extracting flags via showme) to  
find

the Right shared libraries.

Let us know if that works for you.

FWIW, we do recommend using the wrapper compilers over  
extracting the
flags via --showme whenever possible (it's just simpler and  
should do

what you need).


On Feb 15, 2007, at 3:38 PM, Eric Thibodeau wrote:


Hello all,


I have been attempting to compile mpptest on my nodes in vain.  
Here

is my current setup:


Openmpi is in "$HOME/openmpi_`uname -m`" which translates to "/
export/home/eric/openmpi_i686/". I tried the following approaches
(you can see some of these were out of desperation):


CFLAGS=`mpicc --showme:compile` LDFLAGS=`mpicc --showme:link` ./
configure


Configure fails on:

checking whether the C compiler works... configure: error: cannot
run C compiled programs.


The log shows that:

./a.out: error while loading shared libraries: liborte.so.0:  
cannot

open shared object file: No such file or directory



CC="/export/home/eric/openmpi_i686/bin/mpicc" ./configure --with-
mpi=$HOME/openmpi_`uname -m`

Same problems as above...


LDFLAGS="$HOME/openmpi_`uname -m`/lib" ./configure --with-mpi= 
$HOME/

openmpi_`uname -m`


Configure fails on:

checking for C compiler default output file name... configure:
error: C compiler cannot create executables


And...finally (not that all of this was done in the presented  
order):


./configure --with-mpi=$HOME/openmpi_`uname -m`


Which ends with:


checking for library containing MPI_Init... no

configure: error: Could not find MPI library


Anyone can help me with this one...?


Note that LAM-MPI is also installed on these systems...


Eric Thibodeau


___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users






___
users mailing list
us...@open-mpi.org
http://www.open-mp