For Linux, this isn't too big of a problem, but you might want to take a look at the output of "mpicc -showme" to get an idea of what compiler flags / libraries would be added if you used the wrapper compilers. I think for Linux the only one that might at all matter is -pthread.

But I didn't mention it before because I'm pretty sure that's not the issue...

Brian


On Apr 10, 2006, at 10:24 AM, David Gunter wrote:

The problem with doing it that way is that is disallows our in-hose code teams from using their compilers of choice. Prior to open-mpi we have been using LA-MPI. LA-MPI has always been compiled in such a way that it wouldn't matter what other compilers were used to build mpi applications provided the necessary include and lib files were linked in the compilations.

-david

On Apr 10, 2006, at 8:00 AM, Ralph Castain wrote:

I'm not an expert on the configure system, but one thing jumps out at me immediately - you used "gcc" to compile your program. You really need to use "mpicc" to do so.

I think that might be the source of your errors.

Ralph


David Gunter wrote:
After much fiddling around, I managed to create a version of open- mpi that would actually build. Unfortunately, I can't run the simplest of applications with it. Here's the setup I used: export CC=gcc export CXX=g++ export FC=gfortran export F77=gfortran export CFLAGS="-m32" export CXXFLAGS="-m32" export FFLAGS="-m32" export FCFLAGS="-m32" export LDFLAGS="-L/usr/lib" ./configure -- prefix=/net/scratch1/dog/flash64/openmpi/ openmpi-1.0.2-32b -- build=i686-pc-linux-gnu --with-bproc --with-g m --enable-io-romio --with-romio --with-io-romio-flags='--build=i686- pc-linux-gnu' Configure completes, as does 'make' and then 'make install'. Next I tried to compile a simple MPI_Send test application, which fails to run: (flashc 104%) gcc -m32 -I/net/scratch1/dog/flash64/ openmpi/ openmpi-1.0.2-32b/include -o send4 send4.c -L/net/ scratch1/dog/ flash64/openmpi/openmpi-1.0.2-32b/lib -lmpi /net/ scratch1/dog/flash64/openmpi/openmpi-1.0.2-32b/lib/libopal.so.0: warning: epoll_wait is not implemented and will always fail /net/ scratch1/dog/flash64/openmpi/openmpi-1.0.2-32b/lib/libopal.so.0: warning: epoll_ctl is not implemented and will always fail /net/ scratch1/dog/flash64/openmpi/openmpi-1.0.2-32b/lib/libopal.so.0: warning: epoll_create is not implemented and will always fail (flashc 105%) which mpiexec /net/scratch1/dog/flash64/openmpi/ openmpi-1.0.2-32b/bin/mpiexec (flashc 106%) mpiexec -n 4 ./send4 [flashc.lanl.gov:32373] mca: base: component_find: unable to open: / lib/libc.so.6: version `GLIBC_2.3.4' not found (required by /net/ scratch1/dog/flash64/openmpi/openmpi-1.0.2-32b/lib/ openmpi/ mca_paffinity_linux.so) (ignored) [flashc.lanl.gov: 32373] mca: base: component_find: unable to open: libbproc.so.4: cannot open shared object file: No such file or directory (ignored) [flashc.lanl.gov:32373] mca: base: component_find: unable to open: libbproc.so.4: cannot open shared object file: No such file or directory (ignored) [flashc.lanl.gov:32373] mca: base: component_find: unable to open: libbproc.so.4: cannot open shared object file: No such file or directory (ignored) [flashc.lanl.gov:32373] mca: base: component_find: unable to open: libbproc.so.4: cannot open shared object file: No such file or directory (ignored) [flashc.lanl.gov:32373] mca: base: component_find: unable to open: libbproc.so.4: cannot open shared object file: No such file or directory (ignored) mpiexec: relocation error: /net/scratch1/dog/flash64/openmpi/ openmpi-1.0.2-32b/lib/openmpi/mca_soh_bproc.so: undefined symbol: bproc_nodelist I'm still open to suggestions. -david On Apr 10, 2006, at 7:11 AM, David R. (Chip) Kent IV wrote:
When running the tests, is the LD_LIBRARY_PATH getting set to lib64 instead of lib or something like that? Chip On Sat, Apr 08, 2006 at 02:45:01AM -0600, David Gunter wrote:
I am trying to build a 32-bit compatible OpenMPI for our 64-bit Bproc Opteron systems. I saw the thread from last August- September 2005 regarding this but didn't see where it ever succeeded or if any of the problems had been fixed. Most importantly, romio is required to work as well. Is this possible and how is it done? Here's what I have tried so far: setenv CFLAGS -m32 setenv CXXFLAGS -m32 setenv FFLAGS -m32 setenv F90FLAGS -m32 I have used the '--build=i686-pc-linux- gnu' option to the configure setup as well as --with-io-romio- flags="--build=i686-pc-linux-gnu" configure halts with errors when trying to run the Fortran 77 tests. If I remove those env settings and just use the --build option, configure will proceed to the end but the make will eventually halt with errors due to a mix of lib64 libs being accessed at some point. Any ideas? -david -- David Gunter CCN-8: HPC Environments: Parallel Tools Team Los Alamos National Laboratory ------------------------------------------------------------ listmanager [ptools_team] Options: To: listmana...@listserv.lanl.gov Body: ptools_team ------------------------------------------------------------
-- ----------------------------------------------------- David R. "Chip" Kent IV Parallel Tools Team High Performance Computing Environments Group (CCN-8) Los Alamos National Laboratory (505)665-5021 drk...@lanl.gov ----------------------------------------------------- This message is "Technical data or Software Publicly Available" or "Correspondence".
_______________________________________________ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/ listinfo.cgi/users
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to