FWIW: the usual reason for differences between IMPI and OMPI lies in OMPI’s default settings, which are focused on ensuring it can run anywhere as opposed to being optimized for a specific platform. You might look at the MCA params relating to your environment and networks to optimize them for your cluster
> On Nov 23, 2015, at 8:07 AM, michael.rach...@dlr.de wrote: > > Dear Gilles, > > In the meantime the administrators have installed (Thanks!) OpenMPI-1.10.1 > with Intel-16.0.0 on the cluster. > I have tested it with our code: It works. > The time spent for MPI-data transmission was the same as with > OpenMPI-1.8.3&Intel-14.0.4, but was ~20% higher than with > IMPI-5.1.1&Intel-16.0.0 > for the same case running on 3 nodes and 8 procs per node. > > Greetings > Michael Rachner > > > Von: users [mailto:users-boun...@open-mpi.org > <mailto:users-boun...@open-mpi.org>] Im Auftrag von Gilles Gouaillardet > Gesendet: Freitag, 20. November 2015 00:53 > An: Open MPI Users > Betreff: Re: [OMPI users] Bug in Fortran-module MPI of OpenMPI 1.10.0 with > Intel-Ftn-compiler > > Michael, > > in the mean time, you can use 'mpi_f08' instead of 'use mpi' > this is really a f90 binding issue, and f08 is safe > > Cheers, > > Gilles > > On 11/19/2015 10:21 PM, michael.rach...@dlr.de > <mailto:michael.rach...@dlr.de> wrote: > Thank You, Nick and Gilles, > > I hope the administrators of the cluster will be so kind and will update > OpenMPI for me (and others) soon. > > Greetings > Michael > > Von: users [mailto:users-boun...@open-mpi.org > <mailto:users-boun...@open-mpi.org>] Im Auftrag von Gilles Gouaillardet > Gesendet: Donnerstag, 19. November 2015 12:59 > An: Open MPI Users > Betreff: Re: [OMPI users] Bug in Fortran-module MPI of OpenMPI 1.10.0 with > Intel-Ftn-compiler > > Thanks Nick for the pointer ! > > Michael, > > good news is you do not have to upgrade ifort, > but you have to update to 1.10.1 > (intel 16 changed the way gcc pragmas are handled, and ompi has been made > aware in 1.10.1) > 1.10.1 fixes many bugs from 1.10.0, so I strongly encourage anyone to use > 1.10.1 > > Cheers, > > Gilles > > On Thursday, November 19, 2015, Nick Papior <nickpap...@gmail.com > <mailto:nickpap...@gmail.com>> wrote: > Maybe I can chip in, > > We use OpenMPI 1.10.1 with Intel /2016.1.0.423501 without problems. > > I could not get 1.10.0 to work, one reason is: > http://www.open-mpi.org/community/lists/users/2015/09/27655.php > <http://www.open-mpi.org/community/lists/users/2015/09/27655.php> > > On a side-note, please note that if you require scalapack you may need to > follow this approach: > https://software.intel.com/en-us/forums/intel-math-kernel-library/topic/590302 > > <https://software.intel.com/en-us/forums/intel-math-kernel-library/topic/590302> > > 2015-11-19 11:24 GMT+01:00 <michael.rach...@dlr.de > <mailto:michael.rach...@dlr.de>>: > Sorry, Gilles, > > I cannot update to more recent versions, because what I used is the newest > combination of OpenMPI and Intel-Ftn available on that cluster. > > When looking at the list of improvements on the OpenMPI website for OpenMPI > 1.10.1 compared to 1.10.0, I do not remember having seen this item to be > corrected. > > Greeting > Michael Rachner > > > Von: users [mailto: <mailto:users-boun...@open-mpi.org> > <>users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>] Im Auftrag > von Gilles Gouaillardet > Gesendet: Donnerstag, 19. November 2015 10:21 > An: Open MPI Users > Betreff: Re: [OMPI users] Bug in Fortran-module MPI of OpenMPI 1.10.0 with > Intel-Ftn-compiler > > Michael, > > I remember i saw similar reports. > > Could you give a try to the latest v1.10.1 ? > And if that still does not work, can you upgrade icc suite and give it an > other try ? > > I cannot remember whether this is an ifort bug or the way ompi uses fortran... > > Btw, any reason why you do not > Use mpi_f08 ? > > HTH > > Gilles > > <>michael.rach...@dlr.de <mailto:michael.rach...@dlr.de> wrote: > Dear developers of OpenMPI, > > I am trying to run our parallelized Ftn-95 code on a Linux cluster with > OpenMPI-1-10.0 and Intel-16.0.0 Fortran compiler. > In the code I use the module MPI (“use MPI”-stmts). > > However I am not able to compile the code, because of compiler error messages > like this: > > /src_SPRAY/mpi_wrapper.f90(2065): error #6285: There is no matching specific > subroutin for this generic subroutine call. [MPI_REDUCE] > > > The problem seems for me to be this one: > > The interfaces in the module MPI for the MPI-routines do not accept a send or > receive buffer array, which is > actually a variable, an array element or a constant (like MPI_IN_PLACE). > > Example 1: > This does not work (gives the compiler error message: error #6285: > There is no matching specific subroutin for this generic subroutine call ) > ivar=123 ! ß ivar is an integer variable, not an array > call MPI_BCAST( ivar, 1, MPI_INTEGER, 0, MPI_COMM_WORLD), ierr_mpi > ) ! ß- this should work, but is not accepted by the compiler > > only this cumbersome workaround works: > ivar=123 > allocate( iarr(1) ) > iarr(1) = ivar > call MPI_BCAST( iarr, 1, MPI_INTEGER, 0, MPI_COMM_WORLD, ierr_mpi ) > ! ß- this workaround works > ivar = iarr(1) > deallocate( iarr(1) ) > > Example 2: > Any call of an MPI-routine with MPI_IN_PLACE does not work, like that > coding: > > if(lmaster) then > call MPI_REDUCE( MPI_IN_PLACE, rbuffarr, nelem, MPI_REAL8, MPI_MAX & > ! ß- this should work, but is not accepted by the compiler > ,0_INT4, MPI_COMM_WORLD, ierr_mpi ) > else ! slaves > call MPI_REDUCE( rbuffarr, rdummyarr, nelem, MPI_REAL8, MPI_MAX & > ,0_INT4, MPI_COMM_WORLD, ierr_mpi ) > endif > > This results in this compiler error message: > > /src_SPRAY/mpi_wrapper.f90(2122): error #6285: There is no matching > specific subroutine for this generic subroutine call. [MPI_REDUCE] > call MPI_REDUCE( MPI_IN_PLACE, rbuffarr, nelem, MPI_REAL8, > MPI_MAX & > -------------^ > > > In our code I observed the bug with MPI_BCAST, MPI_REDUCE, MPI_ALLREDUCE, > but probably there may be other MPI-routines with the same kind of bug. > > This bug occurred for : OpenMPI-1.10.0 > with Intel-16.0.0 > In contrast, this bug did NOT occur for: OpenMPI-1.8.8 with > Intel-16.0.0 > > OpenMPI-1.8.8 with Intel-15.0.3 > > OpenMPI-1.10.0 with gfortran-5.2.0 > > Greetings > Michael Rachner > > _______________________________________________ > users mailing list > <>us...@open-mpi.org <mailto:us...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > <http://www.open-mpi.org/mailman/listinfo.cgi/users> > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/11/28052.php > <http://www.open-mpi.org/community/lists/users/2015/11/28052.php> > > > > -- > Kind regards Nick > > > > _______________________________________________ > users mailing list > us...@open-mpi.org <mailto:us...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > <http://www.open-mpi.org/mailman/listinfo.cgi/users> > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/11/28056.php > <http://www.open-mpi.org/community/lists/users/2015/11/28056.php> > > _______________________________________________ > users mailing list > us...@open-mpi.org <mailto:us...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > <http://www.open-mpi.org/mailman/listinfo.cgi/users> > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/11/28098.php > <http://www.open-mpi.org/community/lists/users/2015/11/28098.php>