Hello devs, Thought I would let you know there seems to be a problem with 1.2b3r13112 when running the "bisection" test on a Tyan VX50 machine (the 8 DualCore model with 32Gigs of RAM).
OpenMPI was compiled with (as seen from config.log): configure:116866: running /bin/sh './configure' CFLAGS="-O3 -DNDEBUG -finline-functions -fno-strict-aliasing -pthread" CPPFLAGS=" " FFLAGS="" LDFLAGS=" " --enable-shared --disable-static --prefix=/export/livia/home/parallel/eric/openmpi_x86_64 --with-mpi=open_mpi --cache-file=/dev/null --srcdir=. MPPTEST (1.3c) was compiled with: ./configure --with-mpi=$HOME/openmpi_`uname -m` ...which, for some reason, works fine on that system that doesn't have any other MPI implementation (ie: doesn't have LAM-MPI as per this thread). Then I ran a few tests but this one ran for over it's allowed time (1800 seconds and was going over 50minutes...) and was up to 16Gigs of RAM: orterun -np 9 ./perftest-1.3c/mpptest -max_run_time 1800 -bisect -size 0 4096 1 -gnuplot -fname HyperTransport/Global_bisect_0_4096_1.gpl I had to CTRL-\ the process as CTRL-C wasn't sufficient. 2 mpptest processes and 1 orterun process were using 100% CPU ou of of the 16 cores. If any of this can be indicative of an OpenMPI bug and if I can help in tracking it down, don't hesitate to ask for details. And, finally, Anthony, thanks for the MPICC and --with-mpich pointers, I will try those to simplify the build process! Eric Le jeudi 15 février 2007 19:51, Anthony Chan a écrit : > > As long as mpicc is working, try configuring mpptest as > > mpptest/configure MPICC=<OpenMPI-install-dir>/bin/mpicc > > or > > mpptest/configure --with-mpich=<OpenMPI-install-dir> > > A.Chan > > On Thu, 15 Feb 2007, Eric Thibodeau wrote: > > > Hi Jeff, > > > > Thanks for your response, I eventually figured it out, here is the > > only way I got mpptest to compile: > > > > export LD_LIBRARY_PATH="$HOME/openmpi_`uname -m`/lib" > > CC="$HOME/openmpi_`uname -m`/bin/mpicc" ./configure > > --with-mpi="$HOME/openmpi_`uname -m`" > > > > And, yes I know I should use the mpicc wrapper and all (I do RTFM :P ) but > > mpptest is less than cooperative and hasn't been updated lately AFAIK. > > > > I'll keep you posted on some results as I get some results out (testing > > TCP/IP as well as the HyperTransport on a Tyan Beast). Up to now, LAM-MPI > > seems less efficient at async communications and shows no improovments > > with persistant communications under TCP/IP. OpenMPI, on the other hand, > > seems more efficient using persistant communications when in a > > HyperTransport (shmem) environment... I know I am crossing many test > > boudaries but I will post some PNGs of my results (as well as how I got to > > them ;) > > > > Eric > > > > On Thu, 15 Feb 2007, Jeff Squyres wrote: > > > > > I think you want to add $HOME/openmpi_`uname -m`/lib to your > > > LD_LIBRARY_PATH. This should allow executables created by mpicc (or > > > any derivation thereof, such as extracting flags via showme) to find > > > the Right shared libraries. > > > > > > Let us know if that works for you. > > > > > > FWIW, we do recommend using the wrapper compilers over extracting the > > > flags via --showme whenever possible (it's just simpler and should do > > > what you need). > > > > > > > > > On Feb 15, 2007, at 3:38 PM, Eric Thibodeau wrote: > > > > > > > Hello all, > > > > > > > > > > > > I have been attempting to compile mpptest on my nodes in vain. Here > > > > is my current setup: > > > > > > > > > > > > Openmpi is in "$HOME/openmpi_`uname -m`" which translates to "/ > > > > export/home/eric/openmpi_i686/". I tried the following approaches > > > > (you can see some of these were out of desperation): > > > > > > > > > > > > CFLAGS=`mpicc --showme:compile` LDFLAGS=`mpicc --showme:link` ./ > > > > configure > > > > > > > > > > > > Configure fails on: > > > > > > > > checking whether the C compiler works... configure: error: cannot > > > > run C compiled programs. > > > > > > > > > > > > The log shows that: > > > > > > > > ./a.out: error while loading shared libraries: liborte.so.0: cannot > > > > open shared object file: No such file or directory > > > > > > > > > > > > > > > > CC="/export/home/eric/openmpi_i686/bin/mpicc" ./configure --with- > > > > mpi=$HOME/openmpi_`uname -m` > > > > > > > > Same problems as above... > > > > > > > > > > > > LDFLAGS="$HOME/openmpi_`uname -m`/lib" ./configure --with-mpi=$HOME/ > > > > openmpi_`uname -m` > > > > > > > > > > > > Configure fails on: > > > > > > > > checking for C compiler default output file name... configure: > > > > error: C compiler cannot create executables > > > > > > > > > > > > And...finally (not that all of this was done in the presented order): > > > > > > > > ./configure --with-mpi=$HOME/openmpi_`uname -m` > > > > > > > > > > > > Which ends with: > > > > > > > > > > > > checking for library containing MPI_Init... no > > > > > > > > configure: error: Could not find MPI library > > > > > > > > > > > > Anyone can help me with this one...? > > > > > > > > > > > > Note that LAM-MPI is also installed on these systems... > > > > > > > > > > > > Eric Thibodeau > > > > > > > > > > > > _______________________________________________ > > > > users mailing list > > > > us...@open-mpi.org > > > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > > > > > > > > > > _______________________________________________ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > -- Eric Thibodeau Neural Bucket Solutions Inc. T. (514) 736-1436 C. (514) 710-0517