I can reproduce this with Intel 17 and MPICH on Mac OSX so it's not an
Open-MPI issue.  I added VOLATILE to the shared memory objects to prevent
Fortran compiler optimizations as well as a bunch of MPI_Win_sync calls
(after replacing fence with lock_all/unlock_all), but neither changed the
outcome.

While I claim a modest understanding of MPI-3 RMA and Fortran 2008,
unfortunately, I have never figured out how to use MPI-3 shared memory from
Fortran, which is especially unfortunate since it seems to be a fantastic
source of frustration to both real users such as yourself and MPI+Fortran
standard experts (Rolf).

Sorry for the unsatisfying response, but my suspicion is that this is a
program correctness issue.  I can't point to any error, but I've ruled out
the obvious alternatives.

Jeff

On Tue, Oct 25, 2016 at 11:29 AM, Tom Rosmond <rosm...@reachone.com> wrote:

> All:
>
> I am trying to understand the use of the shared memory features of MPI-3
> that allow direct sharing of the memory space of on-node processes.
> Attached are 2 small test programs, one written in C (testmpi3.c), the
> other F95 (testmpi3.f90) .  They are solving the identical 'halo' exchange
> problem.  'testmpi3.c' is a simplified version of an example program from a
> presentation by Mark Lubin of Intel.  I wrote 'testmpi3.f90' to mimic the C
> version.
>
>  Also attached are 2 text files of the compile, execution, and output of
> the respective programs:
>
> CC_testmpi3.txt
> F95_testmpi3.txt
>
> Note: All 4 files are contained in the attached 'testmpi3.tar.gz'.
>
> Comparing the outputs of each version, it is clear that the shared memory
> copies in 'testmpi3.c' are working correctly, but not in 'testmpi3.f90'.
> As far as I can tell, the 2 programs are equivalent up to line 134 of
> 'testmpi3.c' and lines 97-101 of 'testmpi3.f90'. I thought the calls to
> 'c_f_pointer' would produce Fortran pointers that would access the correct
> shared memory addresses as the C-pointers do in 'testmpi3.c', but clearly
> that isn't happening. Can anyone explain why not, and what is needed to
> make this happen. Any suggestions are welcome.
>
> My environment:
>  Scientific Linux 6.8
>  INTEL FORTRAN and ICC version 15.0.2.164
>  OPEN-MPI 2.0.1
>
>
> T. Rosmond
>
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>



-- 
Jeff Hammond
jeff.scie...@gmail.com
http://jeffhammond.github.io/
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to