Hmmm….you might try removing the -enable-mpi-threads from the configure to be safe.
FWIW: I have a Centos6 system myself, and I have no problems running OMPI on it (1.4 or 1.5). I can try building it the same way you do and see what happens. On Mar 13, 2012, at 5:22 PM, Joshua Baker-LePain wrote: > On Tue, 13 Mar 2012 at 10:57pm, Gutierrez, Samuel K wrote > >> Fooey. What compiler are you using to build Open MPI and how are you >> configuring your build? > > I'm using gcc as packaged by RH/CentOS 6.2: > > [jlb@opt200 1.4.5-2]$ gcc --version > gcc (GCC) 4.4.6 20110731 (Red Hat 4.4.6-3) > > I actually tried 2 custom builds of Open MPI 1.4.5. For the first I tried to > stick close to the options in RH's compat-openmpi SRPM: > > ./configure --prefix=$HOME/ompi-1.4.5 --enable-mpi-threads > --enable-openib-ibcm --with-sge --with-libltdl=external --with-valgrind > --enable-memchecker --with-psm=no --with-esmtp LDFLAGS='-Wl,-z,noexecstack' > > That resulted in the backtrace I sent previously: > #0 0x00002b0099ec4c4c in mca_btl_sm_component_progress () > from /netapp/sali/jlb/ompi-1.4.5/lib/openmpi/mca_btl_sm.so > #1 0x00002b00967737ca in opal_progress () > from /netapp/sali/jlb/ompi-1.4.5/lib/libopen-pal.so.0 > #2 0x00002b00975ef8d5 in barrier () > from /netapp/sali/jlb/ompi-1.4.5/lib/openmpi/mca_grpcomm_bad.so > #3 0x00002b009628da24 in ompi_mpi_init () > from /netapp/sali/jlb/ompi-1.4.5/lib/libmpi.so.0 > #4 0x00002b00962b24f0 in PMPI_Init () > from /netapp/sali/jlb/ompi-1.4.5/lib/libmpi.so.0 > #5 0x0000000000400826 in main (argc=1, argv=0x7fff9fe113f8) > at mpihello-long.c:11 > > For kicks, I tried a 2nd compile of 1.4.5 with a bare minimum of options: > > ./configure --prefix=$HOME/ompi-1.4.5 --with-sge > > That resulted in a slightly different backtrace that seems to be missing a > bit: > #0 0x00002b7bbc8681d0 in ?? () > #1 <signal handler called> > #2 0x00002b7bbd2b8f6c in mca_btl_sm_component_progress () > from /netapp/sali/jlb/ompi-1.4.5/lib/openmpi/mca_btl_sm.so > #3 0x00002b7bb9b2feda in opal_progress () > from /netapp/sali/jlb/ompi-1.4.5/lib/libopen-pal.so.0 > #4 0x00002b7bba9a98d5 in barrier () > from /netapp/sali/jlb/ompi-1.4.5/lib/openmpi/mca_grpcomm_bad.so > #5 0x00002b7bb965d426 in ompi_mpi_init () > from /netapp/sali/jlb/ompi-1.4.5/lib/libmpi.so.0 > #6 0x00002b7bb967cba0 in PMPI_Init () > from /netapp/sali/jlb/ompi-1.4.5/lib/libmpi.so.0 > #7 0x0000000000400826 in main (argc=1, argv=0x7fff93634788) > at mpihello-long.c:11 > >> Can you also run with a debug build of Open MPI so we can see the line >> numbers? > > I'll do that first thing tomorrow. > >>>> Another question. How reproducible is this on your system? >>> >>> In my testing today, it's been 100% reproducible. >> >> That's surprising. > > Heh. You're telling me. > > Thanks for taking an interest in this. > > -- > Joshua Baker-LePain > QB3 Shared Cluster Sysadmin > UCSF > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users