Hi George, Thanks for your response. I confirmed that your commit fixed the issue.
Regards, Akihiro On 2016/05/22 5:05, George Bosilca wrote:
16d9f71d01cc should provide a fix for this issue. George. On Sat, May 21, 2016 at 12:08 PM, Akihiro Tabuchi <tabu...@hpcs.cs.tsukuba.ac.jp <mailto:tabu...@hpcs.cs.tsukuba.ac.jp>> wrote: Hi Gilles, Thanks for your quick response and patch. After applying the patch to 1.10.2, the test code and our program which uses nested hvector type ran without error. I hope the patch will be applied to future releases. Regards, Akihiro On 2016/05/21 23:15, Gilles Gouaillardet wrote: Here are attached two patches (one for master, one for v1.10) please consider these as experimental ones : - they cannot hurt - they might not always work - they will likely allocate a bit more memory than necessary - if something goes wrong, it will hopefully be caught soon enough in a new assert clause Cheers, Gilles On Sat, May 21, 2016 at 9:19 PM, Gilles Gouaillardet <gilles.gouaillar...@gmail.com <mailto:gilles.gouaillar...@gmail.com>> wrote: Tabuchi-san, thanks for the report. this is indeed a bug i was able to reproduce on my linux laptop (for some unknown reasons, there is no crash on OS X ) ompi_datatype_pack_description_length malloc 88 bytes for the datatype description, but 96 bytes are required. this causes a memory corruption with undefined side effects (crash in MPI_Type_free, or in MPI_Win_unlock) iirc, we made some changes to ensure data is always aligned (Sparc processors require this), and we could have missed some stuff, and hence malloc less bytes than required. Cheers, Gilles On Sat, May 21, 2016 at 5:50 PM, Akihiro Tabuchi <tabu...@hpcs.cs.tsukuba.ac.jp <mailto:tabu...@hpcs.cs.tsukuba.ac.jp>> wrote: Hi, At OpenMPI 1.10.2, MPI_Type_free crashes with a many nested derived type after using MPI_Put/Get with the datatype as target_datatype. The test code is attached. In the code, MPI_Type_free crashes if N_NEST >= 4. This problem occurs at OpenMPI 1.8.5 or later. There is no problem at OpenMPI 1.8.4, MPICH 3.2, and MVAPICH 2.1. Does anyone know about the problem? Regards, Akihiro -- Akihiro Tabuchi HPCS Lab, Univ. of Tsukuba tabu...@hpcs.cs.tsukuba.ac.jp <mailto:tabu...@hpcs.cs.tsukuba.ac.jp> _______________________________________________ users mailing list us...@open-mpi.org <mailto:us...@open-mpi.org> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/05/29260.php _______________________________________________ users mailing list us...@open-mpi.org <mailto:us...@open-mpi.org> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/05/29262.php -- Akihiro Tabuchi HPCS Lab, Univ. of Tsukuba tabu...@hpcs.cs.tsukuba.ac.jp <mailto:tabu...@hpcs.cs.tsukuba.ac.jp> _______________________________________________ users mailing list us...@open-mpi.org <mailto:us...@open-mpi.org> Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/05/29263.php _______________________________________________ users mailing list us...@open-mpi.org Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users Link to this post: http://www.open-mpi.org/community/lists/users/2016/05/29266.php
-- Akihiro Tabuchi HPCS Lab, Univ. of Tsukuba tabu...@hpcs.cs.tsukuba.ac.jp