Hi Jeff,

Thanks, I will give that a try and repost to the list if problems
remain.  I am kind of surprised that compiling with -fpic is not done by
default on a Linux system, since OpenMPI builds as a set of shared
library .so files.  Normally you want position-independent code in such
libraries so that, among other reasons, static objects are handled
properly.

Regards, Julian C.


On Wed, 2007-06-13 at 11:59 -0400, Jeff Squyres wrote:
> Bummer -- I thought I had replied to that one (perhaps I'm thinking  
> that multiple people have posted this and I've replied to some but  
> not all of them).
> 
> Brock is correct that using "-fpic" to compile your MPI C++ app  
> should solve the problem.  This information *used* to be posted on  
> the PGI web site in their support section, but I can't seem to find  
> it any more.
> 
> As far as I understand the issue, it's a PGI compiler issue, not an  
> OMPI issue.
> 
> 
> 
> On Jun 13, 2007, at 12:38 AM, Julian Cummings wrote:
> 
> > Hello,
> >
> > This is a follow up to a message originally posted by Andrew J  
> > Caird on
> > 2006-08-16.  No one ever replied to Andrew's message, and I am  
> > experiencing
> > exactly the same problem with a more recent version of OpenMPI  
> > (1.2.1) and
> > the PGI compiler (7.0).  Essentially, the problem is that if you  
> > link an MPI
> > application against the mpi_cxx library, at run time you will get a  
> > failure
> > with each process giving the following message:
> >
> > C++ runtime abort: internal error: static object marked for  
> > destruction more
> > than once
> >
> > If your MPI application does not utilize the MPI C++ bindings, you  
> > can link
> > without this library and the runtime errors will go away.
> >
> > Since this problem was reported long ago and no one ever replied to  
> > the
> > report, I would assume that this is a bug either in the mpi_cxx  
> > library or
> > in the way it is built under the PGI compiler.  I could not figure  
> > out how
> > to submit a bug report to the open-mpi bug tracking system, so I  
> > hope that
> > this message to the users list will suffice.  I am attaching my  
> > ompi_info
> > --all output to this message.  I am running on a Myrinet-based Linux
> > cluster, but the particulars are not relevant for this problem.   
> > You can
> > replicate the problem with any trivial MPI application code, such  
> > as the
> > standard "hello" program using the standard C interface.  I am  
> > attaching my
> > hello.c source code.  Compile with "mpicxx -o hello hello.c" and  
> > run with
> > "mpirun -np 1 ./hello".  The runtime error disappears if you  
> > compile with
> > "mpicc -o hello hello.c" to avoid linking against the mpi_cxx library.
> >
> > Please let me know if there is any fix available for this problem.
> >
> > Regards, Julian C.
> > <ompi_info.txt.gz>
> > <hello.c>
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
-- 
Dr. Julian C. Cummings                       E-mail: cummi...@cacr.caltech.edu
California Institute of Technology           Phone:  626-395-2543
1200 E. California Blvd., Mail Code 158-79   Fax:    626-584-5917
Pasadena, CA 91125                           Office: 125 Powell-Booth

Reply via email to