Maybe I can chip in, We use OpenMPI 1.10.1 with Intel /2016.1.0.423501 without problems.
I could not get 1.10.0 to work, one reason is: http://www.open-mpi.org/community/lists/users/2015/09/27655.php On a side-note, please note that if you require scalapack you may need to follow this approach: https://software.intel.com/en-us/forums/intel-math-kernel-library/topic/590302 2015-11-19 11:24 GMT+01:00 <michael.rach...@dlr.de>: > Sorry, Gilles, > > > > I cannot update to more recent versions, because what I used is the > newest combination of OpenMPI and Intel-Ftn available on that cluster. > > > > When looking at the list of improvements on the OpenMPI website for > OpenMPI 1.10.1 compared to 1.10.0, I do not remember having seen this item > to be corrected. > > > > Greeting > > Michael Rachner > > > > > > *Von:* users [mailto:users-boun...@open-mpi.org] *Im Auftrag von *Gilles > Gouaillardet > *Gesendet:* Donnerstag, 19. November 2015 10:21 > *An:* Open MPI Users > *Betreff:* Re: [OMPI users] Bug in Fortran-module MPI of OpenMPI 1.10.0 > with Intel-Ftn-compiler > > > > Michael, > > I remember i saw similar reports. > > Could you give a try to the latest v1.10.1 ? > And if that still does not work, can you upgrade icc suite and give it an > other try ? > > I cannot remember whether this is an ifort bug or the way ompi uses > fortran... > > Btw, any reason why you do not > Use mpi_f08 ? > > HTH > > Gilles > > michael.rach...@dlr.de wrote: > > Dear developers of OpenMPI, > > > > I am trying to run our parallelized Ftn-95 code on a Linux cluster with > OpenMPI-1-10.0 and Intel-16.0.0 Fortran compiler. > > In the code I use the module MPI (“use MPI”-stmts). > > > > However I am not able to compile the code, because of compiler error > messages like this: > > > > /src_SPRAY/mpi_wrapper.f90(2065): error #6285: There is no matching > specific subroutin for this generic subroutine call. [MPI_REDUCE] > > > > > > The problem seems for me to be this one: > > > > The interfaces in the module MPI for the MPI-routines do not accept a send > or receive buffer array, which is > > actually a variable, an array element or a constant (like MPI_IN_PLACE). > > > > Example 1: > > This does not work (gives the compiler error message: error > #6285: There is no matching specific subroutin for this generic subroutine > call ) > > ivar=123 ! ß ivar is an integer variable, not an array > > *call* MPI_BCAST( ivar, 1, MPI_INTEGER, 0, MPI_COMM_WORLD), > ierr_mpi ) ! ß- this should work, but is not accepted by the compiler > > > > only this cumbersome workaround works: > > ivar=123 > > allocate( iarr(1) ) > > iarr(1) = ivar > > * call* MPI_BCAST( iarr, 1, MPI_INTEGER, 0, MPI_COMM_WORLD, > ierr_mpi ) ! ß- this workaround works > > ivar = iarr(1) > > deallocate( iarr(1) ) > > > > Example 2: > > Any call of an MPI-routine with MPI_IN_PLACE does not work, like that > coding: > > > > *if*(lmaster) *then* > > *call* MPI_REDUCE( MPI_IN_PLACE, rbuffarr, nelem, MPI_REAL8, > MPI_MAX & ! ß- this should work, but is not accepted by the compiler > > ,0_INT4, MPI_COMM_WORLD, ierr_mpi > ) > > *else* ! slaves > > *call* MPI_REDUCE( rbuffarr, rdummyarr, nelem, MPI_REAL8, MPI_MAX > & > > ,0_INT4, MPI_COMM_WORLD, ierr_mpi ) > > *endif* > > > > This results in this compiler error message: > > > > /src_SPRAY/mpi_wrapper.f90(2122): error #6285: There is no matching > specific subroutine for this generic subroutine call. [MPI_REDUCE] > > call MPI_REDUCE( MPI_IN_PLACE, rbuffarr, nelem, MPI_REAL8, > MPI_MAX & > > -------------^ > > > > > > In our code I observed the bug with MPI_BCAST, MPI_REDUCE, MPI_ALLREDUCE, > > but probably there may be other MPI-routines with the same kind of bug. > > > > This bug occurred for : OpenMPI-1.10.0 > with Intel-16.0.0 > > In contrast, this bug did NOT occur for: OpenMPI-1.8.8 with > Intel-16.0.0 > > > OpenMPI-1.8.8 with Intel-15.0.3 > > > OpenMPI-1.10.0 with > gfortran-5.2.0 > > > > Greetings > > Michael Rachner > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/11/28052.php > -- Kind regards Nick