Sorry for the delay in replying -- we've been quite busy trying to get OMPI v1.2 out the door!

Are you sure that you build BLACS properly with Open MPI? Check this FAQ item:

   http://www.open-mpi.org/faq/?category=mpi-apps#blacs

In particular, note that there are items in Bmake.inc that you need to set properly or BLACS won't work properly with Open MPI.


On Feb 20, 2007, at 4:25 AM, Kobotov, Alexander V wrote:

Hello all,



I built BLACS on Itanium using Intel compilers under linux (2.6.9-34.EL). But it fails default BLACS Fortran tests (xFbtest), C tests (xCbtest) are ok. I’ve tried different configurations combining OpenMPI-1.1.2 or OpenMPI-1.1.4, ICC 9.1.038 or ICC 8.1.38, IFORT 9.1.33 or IFORT 8.1.34, but all results were the same. OpenMPI is built using 9.1 compilers. Also I’ve tried the same using em64t compiler on Intel XEON – all tests were passed. MPICH2 on IPF also works fine.



Is that an OpenMPI bug? Maybe some workaround exists?



Bmake.inc is attached.

Below is output I’ve got (Don’t pay attention to blacs warnings, they are normal for MPI):

===[ begin of: xFbtest output ]=====================================

-bash-3.00$  mpirun -np 4 xFbtest_MPI-LINUX-0

BLACS WARNING 'No need to set message ID range due to MPI communicator.'

from {-1,-1}, pnum=1, Contxt=-1, on line 18 of file 'blacs_set_.c'.



BLACS WARNING 'No need to set message ID range due to MPI communicator.'

from {-1,-1}, pnum=3, Contxt=-1, on line 18 of file 'blacs_set_.c'.



BLACS WARNING 'No need to set message ID range due to MPI communicator.'

from {-1,-1}, pnum=0, Contxt=-1, on line 18 of file 'blacs_set_.c'.



BLACS WARNING 'No need to set message ID range due to MPI communicator.'

from {-1,-1}, pnum=2, Contxt=-1, on line 18 of file 'blacs_set_.c'.



[comp-pvfs-0-7.local:30119] *** An error occurred in MPI_Comm_group

[comp-pvfs-0-7.local:30118] *** An error occurred in MPI_Comm_group

[comp-pvfs-0-7.local:30118] *** on communicator MPI_COMM_WORLD

[comp-pvfs-0-7.local:30118] *** MPI_ERR_COMM: invalid communicator

[comp-pvfs-0-7.local:30119] *** on communicator MPI_COMM_WORLD

[comp-pvfs-0-7.local:30119] *** MPI_ERR_COMM: invalid communicator

[comp-pvfs-0-7.local:30119] *** MPI_ERRORS_ARE_FATAL (goodbye)

[comp-pvfs-0-7.local:30116] *** An error occurred in MPI_Comm_group

[comp-pvfs-0-7.local:30116] *** on communicator MPI_COMM_WORLD

[comp-pvfs-0-7.local:30118] *** MPI_ERRORS_ARE_FATAL (goodbye)

[comp-pvfs-0-7.local:30116] *** MPI_ERR_COMM: invalid communicator

[comp-pvfs-0-7.local:30116] *** MPI_ERRORS_ARE_FATAL (goodbye)

[comp-pvfs-0-7.local:30117] *** An error occurred in MPI_Comm_group

[comp-pvfs-0-7.local:30117] *** on communicator MPI_COMM_WORLD

[comp-pvfs-0-7.local:30117] *** MPI_ERR_COMM: invalid communicator

[comp-pvfs-0-7.local:30117] *** MPI_ERRORS_ARE_FATAL (goodbye)

forrtl: error (78): process killed (SIGTERM)

forrtl: error (78): process killed (SIGTERM)

forrtl: error (78): process killed (SIGTERM)

forrtl: error (78): process killed (SIGTERM)

===[ end of: xFbtest output ]=====================================





W.B.R.,

Kobotov Alexander





<Bmake.inc>
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


--
Jeff Squyres
Server Virtualization Business Unit
Cisco Systems


Reply via email to