Hi all,
I'm going crazy about a possible bug in my code. I'm using a derived mpi
datatype in a sendrecv function.
The problem is that the memory footprint of my code is growing as time
increases.
The problem is not showing with a regular datatype, as MPI_DOUBLE.
I don't have this problem for openmpi 1.8.4, but it's present for 1.10.1
and 2.0.1
The key parts of the code are (i'm using a 1D array with a macro in
order to be 3D) :
Definition of the datatype:
MPI_Type_vector( Ny, 1, Nx, MPI_DOUBLE, &mpi.MPI_COL );
MPI_Type_commit( &mpi.MPI_COL ) ;
And the sendrecv part:
MPI_Sendrecv( &(thebigone[_(1,0,k)]) , 1 , mpi.MPI_COL ,
mpi.left , 3, \
&(thebigone[_(Nx-1,0,k)]) , 1 , mpi.MPI_COL ,
mpi.right, 3, \
mpi.com, &mpi.stat );
Is it coming from my code ?
I isolated the communications in a small code (500 lines). I can give it
in order to reproduce the problem.
Thanks,
Yann
---
L'absence de virus dans ce courrier électronique a été vérifiée par le logiciel
antivirus Avast.
https://www.avast.com/antivirus
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users