Joseph, at first glance, there is a memory corruption (!) the first printf should be 0 -> 100, instead of 0 -> 3200
this is very odd because nelems is const, and the compiler might not even allocate this variable. I also noted some counter intuitive stuff in your test program (which still looks valid to me) neighbor = (rank +1) / size; should it be neighbor = (rank + 1) % size; instead ? the first loop is for (elem=0; elem < nelems-1; elem++) ... it could be for (elem=0; elem < nelems; elem++) ... the second loop uses disp_set, and I guess you meant to use disp_set2 I will try to reproduce this crash. which compiler (vendor and version) are you using ? which compiler options do you pass to mpicc ? Cheers, Gilles On Thursday, August 25, 2016, Joseph Schuchart <schuch...@hlrs.de> wrote: > All, > > It seems there is a regression in the handling of dynamic windows between > Open MPI 1.10.3 and 2.0.0. I am attaching a test case that works fine with > Open MPI 1.8.3 and fail with version 2.0.0 with the following output: > > === > [0] MPI_Get 0 -> 3200 on first memory region > [cl3fr1:7342] *** An error occurred in MPI_Get > [cl3fr1:7342] *** reported by process [908197889,0] > [cl3fr1:7342] *** on win rdma window 3 > [cl3fr1:7342] *** MPI_ERR_RMA_RANGE: invalid RMA address range > [cl3fr1:7342] *** MPI_ERRORS_ARE_FATAL (processes in this win will now > abort, > [cl3fr1:7342] *** and potentially your MPI job) > === > > Expected output is: > === > [0] MPI_Get 0 -> 100 on first memory region: > [0] Done. > [0] MPI_Get 0 -> 100 on second memory region: > [0] Done. > === > > The code allocates a dynamic window and attaches two memory regions to it > before accessing both memory regions using MPI_Get. With Open MPI 2.0.0, > only access to the both memory regions fails. Access to the first memory > region only succeeds if the second memory region is not attached. With Open > MPI 1.10.3, all MPI operations succeed. > > Please let me know if you need any additional information or think that my > code example is not standard compliant. > > Best regards > Joseph > > > -- > Dipl.-Inf. Joseph Schuchart > High Performance Computing Center Stuttgart (HLRS) > Nobelstr. 19 > D-70569 Stuttgart > > Tel.: +49(0)711-68565890 > Fax: +49(0)711-6856832 > E-Mail: schuch...@hlrs.de > >
_______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users