Dear Gilles, Thank you! Indeed, this appears to address the issue in my test.
I have yet to run a full regression test on our actual code to ensure that there are no other side effects. I don’t expect any, though. ** Interestingly, removing '-lmpi_usempif08 -lmpi_usempi_ignore_tkr’ actually fixes a second regression that I had found only last night and not yet reported. Specifically, I have a very well tested mpi_alltoallv call (part of our standard regression tests). Versions of this exact call live in multiple codes, including libraries. With the unpatched OpenMPI version on the same macbook and in a specific case, this suddenly resulted in: [egr-vwb3-mbp2:3573] *** An error occurred in MPI_Alltoallv [egr-vwb3-mbp2:3573] *** reported by process [1710358529,0] [egr-vwb3-mbp2:3573] *** on communicator MPI COMMUNICATOR 7 SPLIT FROM 0 [egr-vwb3-mbp2:3573] *** MPI_ERR_TRUNCATE: message truncated [egr-vwb3-mbp2:3573] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort, [egr-vwb3-mbp2:3573] *** and potentially your MPI job) [egr-vwb3-mbp2.local:03572] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal [egr-vwb3-mbp2.local:03572] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages This regression is also fixed. I do not have a simple test case for this behavior yet since I was still investigating it. My plan was to create a test case, but extracting such a case from the larger code we have is more difficult. I can still try to create a test case for this as well but I’ll be watching with great interest if the earlier MPI_IN_PLACE regression can be nailed down and fixed in a more general way. Perhaps the two are related. Thanks again, Best wishes Volker > On Jul 27, 2017, at 10:46 AM, Gilles Gouaillardet <gil...@rist.or.jp> wrote: > > Volker, > > > since you are only using > > include 'mpif.h' > > > a workaround is you edit your /.../share/openmpi/mpifort-wrapper-data.txt > and simply remove '-lmpi_usempif08 -lmpi_usempi_ignore_tkr' > > > Cheers, > > Gilles > > On 7/27/2017 3:28 PM, Volker Blum wrote: >> Thanks! >> >> If you wish, please also keep me posted. >> >> Best wishes >> Volker >> >>> On Jul 27, 2017, at 7:50 AM, Gilles Gouaillardet >>> <gilles.gouaillar...@gmail.com> wrote: >>> >>> Thanks Jeff for your offer, i will contact you off-list later >>> >>> >>> i tried a gcc+gfortran and gcc+ifort on both linux and OS X >>> so far, only gcc+ifort on OS X is failing >>> i will try icc+ifort on OS X from now >>> >>> short story, MPI_IN_PLACE is not recognized as such by the ompi >>> fortran wrapper, and i do not know why. >>> >>> the attached program can be used to evidence the issue. >>> >>> >>> Cheers, >>> >>> Gilles >>> >>> On Thu, Jul 27, 2017 at 2:15 PM, Volker Blum <volker.b...@duke.edu> wrote: >>>> Thanks! That’s great. Sounds like the exact combination I have here. >>>> >>>> Thanks also to George. Sorry that the test did not trigger on a more >>>> standard platform - that would have simplified things. >>>> >>>> Best wishes >>>> Volker >>>> >>>>> On Jul 27, 2017, at 3:56 AM, Gilles Gouaillardet <gil...@rist.or.jp> >>>>> wrote: >>>>> >>>>> Folks, >>>>> >>>>> >>>>> I am able to reproduce the issue on OS X (Sierra) with stock gcc (aka >>>>> clang) and ifort 17.0.4 >>>>> >>>>> >>>>> i will investigate this from now >>>>> >>>>> >>>>> Cheers, >>>>> >>>>> Gilles >>>>> >>>>> On 7/27/2017 9:28 AM, George Bosilca wrote: >>>>>> Volker, >>>>>> >>>>>> Unfortunately, I can't replicate with icc. I tried on a x86_64 box with >>>>>> Intel compiler chain 17.0.4 20170411 to no avail. I also tested the >>>>>> 3.0.0-rc1 tarball and the current master, and you test completes without >>>>>> errors on all cases. >>>>>> >>>>>> Once you figure out an environment where you can consistently replicate >>>>>> the issue, I would suggest to attach to the processes and: >>>>>> - make sure the MPI_IN_PLACE as seen through the Fortran layer matches >>>>>> what the C layer expects >>>>>> - what is the collective algorithm used by Open MPI >>>>>> >>>>>> I have a "Fortran 101" level question. When you pass an array a(:) as >>>>>> argument, what exactly gets passed via the Fortran interface to the >>>>>> corresponding C function ? >>>>>> >>>>>> George. >>>>>> >>>>>> On Wed, Jul 26, 2017 at 1:55 PM, Volker Blum <volker.b...@duke.edu >>>>>> <mailto:volker.b...@duke.edu>> wrote: >>>>>> >>>>>> Thanks! Yes, trying with Intel 2017 would be very nice. >>>>>> >>>>>>> On Jul 26, 2017, at 6:12 PM, George Bosilca <bosi...@icl.utk.edu >>>>>> <mailto:bosi...@icl.utk.edu>> wrote: >>>>>>> No, I don't have (or used where they were available) the Intel >>>>>> compiler. I used clang and gfortran. I can try on a Linux box with >>>>>> the Intel 2017 compilers. >>>>>>> George. >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Wed, Jul 26, 2017 at 11:59 AM, Volker Blum >>>>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>> Did you use Intel Fortran 2017 as well? >>>>>>> >>>>>>> (I’m asking because I did see the same issue with a combination >>>>>> of an earlier Intel Fortran 2017 version and OpenMPI on an >>>>>> Intel/Infiniband Linux HPC machine … but not Intel Fortran 2016 on >>>>>> the same machine. Perhaps I can revive my access to that >>>>>> combination somehow.) >>>>>>> Best wishes >>>>>>> Volker >>>>>>> >>>>>>>> On Jul 26, 2017, at 5:55 PM, George Bosilca >>>>>> <bosi...@icl.utk.edu <mailto:bosi...@icl.utk.edu>> wrote: >>>>>>>> I thought that maybe the underlying allreduce algorithm fails >>>>>> to support MPI_IN_PLACE correctly, but I can't replicate on any >>>>>> machine (including OSX) with any number of processes. >>>>>>>> George. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Wed, Jul 26, 2017 at 10:59 AM, Volker Blum >>>>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>>> Thanks! >>>>>>>> >>>>>>>> I tried ‘use mpi’, which compiles fine. >>>>>>>> >>>>>>>> Same result as with ‘include mpif.h', in that the output is >>>>>>>> >>>>>>>> * MPI_IN_PLACE does not appear to work as intended. >>>>>>>> * Checking whether MPI_ALLREDUCE works at all. >>>>>>>> * Without MPI_IN_PLACE, MPI_ALLREDUCE appears to work. >>>>>>>> >>>>>>>> Hm. Any other thoughts? >>>>>>>> >>>>>>>> Thanks again! >>>>>>>> Best wishes >>>>>>>> Volker >>>>>>>> >>>>>>>>> On Jul 26, 2017, at 4:06 PM, Gilles Gouaillardet >>>>>> <gilles.gouaillar...@gmail.com >>>>>> <mailto:gilles.gouaillar...@gmail.com>> wrote: >>>>>>>>> Volker, >>>>>>>>> >>>>>>>>> With mpi_f08, you have to declare >>>>>>>>> >>>>>>>>> Type(MPI_Comm) :: mpi_comm_global >>>>>>>>> >>>>>>>>> (I am afk and not 100% sure of the syntax) >>>>>>>>> >>>>>>>>> A simpler option is to >>>>>>>>> >>>>>>>>> use mpi >>>>>>>>> >>>>>>>>> Cheers, >>>>>>>>> >>>>>>>>> Gilles >>>>>>>>> >>>>>>>>> Volker Blum <volker.b...@duke.edu >>>>>> <mailto:volker.b...@duke.edu>> wrote: >>>>>>>>>> Hi Gilles, >>>>>>>>>> >>>>>>>>>> Thank you very much for the response! >>>>>>>>>> >>>>>>>>>> Unfortunately, I don’t have access to a different system >>>>>> with the issue right now. As I said, it’s not new; it just keeps >>>>>> creeping up unexpectedly again on different platforms. What >>>>>> puzzles me is that I’ve encountered the same problem with low but >>>>>> reasonable frequency over a period of now over five years. >>>>>>>>>> We can’t require F’08 in our application, unfortunately, >>>>>> since this standard is too new. Since we maintain a large >>>>>> application that has to run on a broad range of platforms, Fortran >>>>>> 2008 would not work for many of our users. In a few years, this >>>>>> will be different, but not yet. >>>>>>>>>> On gfortran: In our own tests, unfortunately, Intel Fortran >>>>>> consistently produced much faster executable code in the past. The >>>>>> latter observation may also change someday, but for us, the >>>>>> performance difference was an important constraint. >>>>>>>>>> I did suspect mpif.h, too. Not sure how to best test this >>>>>> hypothesis, however. >>>>>>>>>> Just replacing >>>>>>>>>> >>>>>>>>>>> include 'mpif.h' >>>>>>>>>>> with >>>>>>>>>>> use mpi_f08 >>>>>>>>>> did not work, for me. >>>>>>>>>> >>>>>>>>>> This produces a number of compilation errors: >>>>>>>>>> >>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> mpif90 >>>>>> check_mpi_in_place_08.f90 -o check_mpi_in_place_08.x >>>>>>>>>> check_mpi_in_place_08.f90(55): error #6303: The assignment >>>>>> operation or the binary expression operation is invalid for the >>>>>> data types of the two operands. [MPI_COMM_WORLD] >>>>>>>>>> mpi_comm_global = MPI_COMM_WORLD >>>>>>>>>> ----------------------^ >>>>>>>>>> check_mpi_in_place_08.f90(57): error #6285: There is no >>>>>> matching specific subroutine for this generic subroutine call. >>>>>> [MPI_COMM_SIZE] >>>>>>>>>> call MPI_COMM_SIZE(mpi_comm_global, n_tasks, mpierr) >>>>>>>>>> ---------^ >>>>>>>>>> check_mpi_in_place_08.f90(58): error #6285: There is no >>>>>> matching specific subroutine for this generic subroutine call. >>>>>> [MPI_COMM_RANK] >>>>>>>>>> call MPI_COMM_RANK(mpi_comm_global, myid, mpierr) >>>>>>>>>> ---------^ >>>>>>>>>> check_mpi_in_place_08.f90(75): error #6285: There is no >>>>>> matching specific subroutine for this generic subroutine call. >>>>>> [MPI_ALLREDUCE] >>>>>>>>>> call MPI_ALLREDUCE(MPI_IN_PLACE, & >>>>>>>>>> ---------^ >>>>>>>>>> check_mpi_in_place_08.f90(94): error #6285: There is no >>>>>> matching specific subroutine for this generic subroutine call. >>>>>> [MPI_ALLREDUCE] >>>>>>>>>> call MPI_ALLREDUCE(check_success, aux_check_success, 1, >>>>>> MPI_LOGICAL, & >>>>>>>>>> ---------^ >>>>>>>>>> check_mpi_in_place_08.f90(119): error #6285: There is no >>>>>> matching specific subroutine for this generic subroutine call. >>>>>> [MPI_ALLREDUCE] >>>>>>>>>> call MPI_ALLREDUCE(test_data(:), & >>>>>>>>>> ------------^ >>>>>>>>>> check_mpi_in_place_08.f90(140): error #6285: There is no >>>>>> matching specific subroutine for this generic subroutine call. >>>>>> [MPI_ALLREDUCE] >>>>>>>>>> call MPI_ALLREDUCE(check_conventional_mpi, >>>>>> aux_check_success, 1, MPI_LOGICAL, & >>>>>>>>>> ------------^ >>>>>>>>>> compilation aborted for check_mpi_in_place_08.f90 (code 1) >>>>>>>>>> >>>>>>>>>> This is an interesting result, however … what might I be >>>>>> missing? Another use statement? >>>>>>>>>> Best wishes >>>>>>>>>> Volker >>>>>>>>>> >>>>>>>>>>> On Jul 26, 2017, at 2:53 PM, Gilles Gouaillardet >>>>>> <gilles.gouaillar...@gmail.com >>>>>> <mailto:gilles.gouaillar...@gmail.com>> wrote: >>>>>>>>>>> Volker, >>>>>>>>>>> >>>>>>>>>>> thanks, i will have a look at it >>>>>>>>>>> >>>>>>>>>>> meanwhile, if you can reproduce this issue on a more >>>>>> mainstream >>>>>>>>>>> platform (e.g. linux + gfortran) please let me know. >>>>>>>>>>> >>>>>>>>>>> since you are using ifort, Open MPI was built with Fortran >>>>>> 2008 >>>>>>>>>>> bindings, so you can replace >>>>>>>>>>> include 'mpif.h' >>>>>>>>>>> with >>>>>>>>>>> use mpi_f08 >>>>>>>>>>> and who knows, that might solve your issue >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Cheers, >>>>>>>>>>> >>>>>>>>>>> Gilles >>>>>>>>>>> >>>>>>>>>>> On Wed, Jul 26, 2017 at 5:22 PM, Volker Blum >>>>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>>>>>>> Dear Gilles, >>>>>>>>>>>> >>>>>>>>>>>> Thank you very much for the fast answer. >>>>>>>>>>>> >>>>>>>>>>>> Darn. I feared it might not occur on all platforms, since >>>>>> my former Macbook >>>>>>>>>>>> (with an older OpenMPI version) no longer exhibited the >>>>>> problem, a different >>>>>>>>>>>> Linux/Intel Machine did last December, etc. >>>>>>>>>>>> >>>>>>>>>>>> On this specific machine, the configure line is >>>>>>>>>>>> >>>>>>>>>>>> ./configure CC=gcc FC=ifort F77=ifort >>>>>>>>>>>> >>>>>>>>>>>> ifort version 17.0.4 >>>>>>>>>>>> >>>>>>>>>>>> blum:/Users/blum/software/openmpi-3.0.0rc1> gcc -v >>>>>>>>>>>> Configured with: >>>>>> --prefix=/Applications/Xcode.app/Contents/Developer/usr >>>>>>>>>>>> --with-gxx-include-dir=/usr/include/c++/4.2.1 >>>>>>>>>>>> Apple LLVM version 8.1.0 (clang-802.0.42) >>>>>>>>>>>> Target: x86_64-apple-darwin16.6.0 >>>>>>>>>>>> Thread model: posix >>>>>>>>>>>> InstalledDir: >>>>>>>>>>>> >>>>>> >>>>>> /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin >>>>>>>>>>>> The full test program is appended. >>>>>>>>>>>> >>>>>>>>>>>> Compilation: >>>>>>>>>>>> >>>>>>>>>>>> mpif90 check_mpi_in_place.f90 >>>>>>>>>>>> >>>>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> which mpif90 >>>>>>>>>>>> /usr/local/openmpi-3.0.0rc1/bin/mpif90 >>>>>>>>>>>> >>>>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> which mpirun >>>>>>>>>>>> /usr/local/openmpi-3.0.0rc1/bin/mpirun >>>>>>>>>>>> >>>>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> mpirun -np >>>>>> 2 a.out >>>>>>>>>>>> * MPI_IN_PLACE does not appear to work as intended. >>>>>>>>>>>> * Checking whether MPI_ALLREDUCE works at all. >>>>>>>>>>>> * Without MPI_IN_PLACE, MPI_ALLREDUCE appears to work. >>>>>>>>>>>> >>>>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> mpirun -np >>>>>> 1 a.out >>>>>>>>>>>> * MPI_IN_PLACE does not appear to work as intended. >>>>>>>>>>>> * Checking whether MPI_ALLREDUCE works at all. >>>>>>>>>>>> * Without MPI_IN_PLACE, MPI_ALLREDUCE appears to work. >>>>>>>>>>>> >>>>>>>>>>>> Hopefully, no trivial mistakes in the testcase. I just >>>>>> spent a few days >>>>>>>>>>>> tracing this issue through a fairly large code, which is >>>>>> where the issue >>>>>>>>>>>> originally arose (and leads to wrong numbers). >>>>>>>>>>>> >>>>>>>>>>>> Best wishes >>>>>>>>>>>> Volker >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>>> On Jul 26, 2017, at 9:46 AM, Gilles Gouaillardet >>>>>>>>>>>>> <gilles.gouaillar...@gmail.com >>>>>> <mailto:gilles.gouaillar...@gmail.com>> wrote: >>>>>>>>>>>>> Volker, >>>>>>>>>>>>> >>>>>>>>>>>>> i was unable to reproduce this issue on linux >>>>>>>>>>>>> >>>>>>>>>>>>> can you please post your full configure command line, >>>>>> your gnu >>>>>>>>>>>>> compiler version and the full test program ? >>>>>>>>>>>>> >>>>>>>>>>>>> also, how many mpi tasks are you running ? >>>>>>>>>>>>> >>>>>>>>>>>>> Cheers, >>>>>>>>>>>>> >>>>>>>>>>>>> Gilles >>>>>>>>>>>>> >>>>>>>>>>>>> On Wed, Jul 26, 2017 at 4:25 PM, Volker Blum >>>>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>> >>>>>>>>>>>>>> I tried openmpi-3.0.0rc1.tar.gz using Intel Fortran >>>>>> 2017 and gcc on a >>>>>>>>>>>>>> current MacOS system. For this version, it seems to me >>>>>> that MPI_IN_PLACE >>>>>>>>>>>>>> returns incorrect results (while other MPI >>>>>> implementations, including some >>>>>>>>>>>>>> past OpenMPI versions, work fine). >>>>>>>>>>>>>> >>>>>>>>>>>>>> This can be seen with a simple Fortran example code, >>>>>> shown below. In the >>>>>>>>>>>>>> test, the values of all entries of an array “test_data” >>>>>> should be 1.0d0 if >>>>>>>>>>>>>> the behavior were as intended. However, the version of >>>>>> OpenMPI I have >>>>>>>>>>>>>> returns 0.d0 instead. >>>>>>>>>>>>>> >>>>>>>>>>>>>> I’ve seen this behavior on some other compute platforms >>>>>> too, in the past, >>>>>>>>>>>>>> so it wasn’t new to me. Still, I thought that this >>>>>> time, I’d ask. Any >>>>>>>>>>>>>> thoughts? >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thank you, >>>>>>>>>>>>>> Best wishes >>>>>>>>>>>>>> Volker >>>>>>>>>>>>>> >>>>>>>>>>>>>> ! size of test data array >>>>>>>>>>>>>> integer :: n_data >>>>>>>>>>>>>> >>>>>>>>>>>>>> ! array that contains test data for MPI_IN_PLACE >>>>>>>>>>>>>> real*8, allocatable :: test_data(:) >>>>>>>>>>>>>> >>>>>>>>>>>>>> integer :: mpierr >>>>>>>>>>>>>> >>>>>>>>>>>>>> n_data = 10 >>>>>>>>>>>>>> >>>>>>>>>>>>>> allocate(test_data(n_data),stat=mpierr) >>>>>>>>>>>>>> >>>>>>>>>>>>>> ! seed test data array for allreduce call below >>>>>>>>>>>>>> if (myid.eq.0) then >>>>>>>>>>>>>> test_data(:) = 1.d0 >>>>>>>>>>>>>> else >>>>>>>>>>>>>> test_data(:) = 0.d0 >>>>>>>>>>>>>> end if >>>>>>>>>>>>>> >>>>>>>>>>>>>> ! Sum the test_data array over all MPI tasks >>>>>>>>>>>>>> call MPI_ALLREDUCE(MPI_IN_PLACE, & >>>>>>>>>>>>>> test_data(:), & >>>>>>>>>>>>>> n_data, & >>>>>>>>>>>>>> MPI_DOUBLE_PRECISION, & >>>>>>>>>>>>>> MPI_SUM, & >>>>>>>>>>>>>> mpi_comm_global, & >>>>>>>>>>>>>> mpierr ) >>>>>>>>>>>>>> >>>>>>>>>>>>>> ! The value of all entries of test_data should now be >>>>>> 1.d0 on all MPI >>>>>>>>>>>>>> tasks. >>>>>>>>>>>>>> ! If that is not the case, then the MPI_IN_PLACE flag >>>>>> may be broken. >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Volker Blum >>>>>>>>>>>>>> Associate Professor >>>>>>>>>>>>>> Ab Initio Materials Simulations >>>>>>>>>>>>>> Duke University, MEMS Department >>>>>>>>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC >>>>>> 27708, USA >>>>>>>>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>>>>>>>> https://aims.pratt.duke.edu >>>>>>>>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>>>>>>>> Twitter: Aimsduke >>>>>>>>>>>>>> >>>>>>>>>>>>>> Office: 1111 Hudson Hall >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e=> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing list >>>>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e=> >>>>>>>>>>>> Volker Blum >>>>>>>>>>>> Associate Professor >>>>>>>>>>>> Ab Initio Materials Simulations >>>>>>>>>>>> Duke University, MEMS Department >>>>>>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC >>>>>> 27708, USA >>>>>>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>>>>>> https://aims.pratt.duke.edu >>>>>>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>>>>>> Twitter: Aimsduke >>>>>>>>>>>> >>>>>>>>>>>> Office: 1111 Hudson Hall >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> users mailing list >>>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>>> >>>>>> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e=> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> users mailing list >>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>> >>>>>> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e=> >>>>>>>>>> Volker Blum >>>>>>>>>> Associate Professor >>>>>>>>>> Ab Initio Materials Simulations >>>>>>>>>> Duke University, MEMS Department >>>>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC >>>>>> 27708, USA >>>>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>>>> https://aims.pratt.duke.edu >>>>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>>>> Twitter: Aimsduke >>>>>>>>>> >>>>>>>>>> Office: 1111 Hudson Hall >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>> >>>>>> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e=> >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>> >>>>>> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e=> >>>>>>>> Volker Blum >>>>>>>> Associate Professor >>>>>>>> Ab Initio Materials Simulations >>>>>>>> Duke University, MEMS Department >>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>>>>>> >>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>> https://aims.pratt.duke.edu >>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>> Twitter: Aimsduke >>>>>>>> >>>>>>>> Office: 1111 Hudson Hall >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>>>>> >>>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e=> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>> >>>>>> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=4W_1_DI3QU9PRT2fhTXyUeXiu7HvTSNzX48E-9ifoTc&s=i3f7Olcbyor4Pu0hv6YlgO10cJ_XvOR13zZn7PPIGto&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=4W_1_DI3QU9PRT2fhTXyUeXiu7HvTSNzX48E-9ifoTc&s=i3f7Olcbyor4Pu0hv6YlgO10cJ_XvOR13zZn7PPIGto&e=> >>>>>>> Volker Blum >>>>>>> Associate Professor >>>>>>> Ab Initio Materials Simulations >>>>>>> Duke University, MEMS Department >>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>>>>> >>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>> https://aims.pratt.duke.edu >>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>> Twitter: Aimsduke >>>>>>> >>>>>>> Office: 1111 Hudson Hall >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e=> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>> >>>>>> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h6zBsOI45o8fovfy43A2FyCt-fL_yVpNbVSf1OA8CrQ&s=YNADyKbvRnxPmnDVmVlYmsYgZEr8m-etPBXLHPRkflw&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h6zBsOI45o8fovfy43A2FyCt-fL_yVpNbVSf1OA8CrQ&s=YNADyKbvRnxPmnDVmVlYmsYgZEr8m-etPBXLHPRkflw&e=> >>>>>> >>>>>> Volker Blum >>>>>> Associate Professor >>>>>> Ab Initio Materials Simulations >>>>>> Duke University, MEMS Department >>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>>>> >>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>> https://aims.pratt.duke.edu >>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>> Twitter: Aimsduke >>>>>> >>>>>> Office: 1111 Hudson Hall >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e=> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.open-mpi.org >>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>> Volker Blum >>>> Associate Professor >>>> Ab Initio Materials Simulations >>>> Duke University, MEMS Department >>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>> >>>> volker.b...@duke.edu >>>> https://aims.pratt.duke.edu >>>> +1 (919) 660 5279 >>>> Twitter: Aimsduke >>>> >>>> Office: 1111 Hudson Hall >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIFaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h3HNno_BCKRMuxrZ26zNcyS_JTV7oDefvQtaN7NF3Kk&s=ENCO2vndR26kHOuem57zq1jvk1Tl1ftPBQKcQ2xIpwQ&e= >>> <test_constants.f90>_______________________________________________ >>> users mailing list >>> users@lists.open-mpi.org >>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h3HNno_BCKRMuxrZ26zNcyS_JTV7oDefvQtaN7NF3Kk&s=ENCO2vndR26kHOuem57zq1jvk1Tl1ftPBQKcQ2xIpwQ&e= >> Volker Blum >> Associate Professor >> Ab Initio Materials Simulations >> Duke University, MEMS Department >> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >> >> volker.b...@duke.edu >> https://aims.pratt.duke.edu >> +1 (919) 660 5279 >> Twitter: Aimsduke >> >> Office: 1111 Hudson Hall >> >> >> >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=_NZvZhsoGaDbJEC8nwG5oDMiMtSi6cOdtsEboB-uXKQ&s=gVgnzXkfNMEN3UwBtvjPBhrTJPNJs2ty-7pEpHKvImI&e= >> > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=_NZvZhsoGaDbJEC8nwG5oDMiMtSi6cOdtsEboB-uXKQ&s=gVgnzXkfNMEN3UwBtvjPBhrTJPNJs2ty-7pEpHKvImI&e= Volker Blum Associate Professor Ab Initio Materials Simulations Duke University, MEMS Department 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA volker.b...@duke.edu https://aims.pratt.duke.edu +1 (919) 660 5279 Twitter: Aimsduke Office: 1111 Hudson Hall _______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users