Thanks! If you wish, please also keep me posted.
Best wishes Volker > On Jul 27, 2017, at 7:50 AM, Gilles Gouaillardet > <gilles.gouaillar...@gmail.com> wrote: > > Thanks Jeff for your offer, i will contact you off-list later > > > i tried a gcc+gfortran and gcc+ifort on both linux and OS X > so far, only gcc+ifort on OS X is failing > i will try icc+ifort on OS X from now > > short story, MPI_IN_PLACE is not recognized as such by the ompi > fortran wrapper, and i do not know why. > > the attached program can be used to evidence the issue. > > > Cheers, > > Gilles > > On Thu, Jul 27, 2017 at 2:15 PM, Volker Blum <volker.b...@duke.edu> wrote: >> Thanks! That’s great. Sounds like the exact combination I have here. >> >> Thanks also to George. Sorry that the test did not trigger on a more >> standard platform - that would have simplified things. >> >> Best wishes >> Volker >> >>> On Jul 27, 2017, at 3:56 AM, Gilles Gouaillardet <gil...@rist.or.jp> wrote: >>> >>> Folks, >>> >>> >>> I am able to reproduce the issue on OS X (Sierra) with stock gcc (aka >>> clang) and ifort 17.0.4 >>> >>> >>> i will investigate this from now >>> >>> >>> Cheers, >>> >>> Gilles >>> >>> On 7/27/2017 9:28 AM, George Bosilca wrote: >>>> Volker, >>>> >>>> Unfortunately, I can't replicate with icc. I tried on a x86_64 box with >>>> Intel compiler chain 17.0.4 20170411 to no avail. I also tested the >>>> 3.0.0-rc1 tarball and the current master, and you test completes without >>>> errors on all cases. >>>> >>>> Once you figure out an environment where you can consistently replicate >>>> the issue, I would suggest to attach to the processes and: >>>> - make sure the MPI_IN_PLACE as seen through the Fortran layer matches >>>> what the C layer expects >>>> - what is the collective algorithm used by Open MPI >>>> >>>> I have a "Fortran 101" level question. When you pass an array a(:) as >>>> argument, what exactly gets passed via the Fortran interface to the >>>> corresponding C function ? >>>> >>>> George. >>>> >>>> On Wed, Jul 26, 2017 at 1:55 PM, Volker Blum <volker.b...@duke.edu >>>> <mailto:volker.b...@duke.edu>> wrote: >>>> >>>> Thanks! Yes, trying with Intel 2017 would be very nice. >>>> >>>>> On Jul 26, 2017, at 6:12 PM, George Bosilca <bosi...@icl.utk.edu >>>> <mailto:bosi...@icl.utk.edu>> wrote: >>>>> >>>>> No, I don't have (or used where they were available) the Intel >>>> compiler. I used clang and gfortran. I can try on a Linux box with >>>> the Intel 2017 compilers. >>>>> >>>>> George. >>>>> >>>>> >>>>> >>>>> On Wed, Jul 26, 2017 at 11:59 AM, Volker Blum >>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>> Did you use Intel Fortran 2017 as well? >>>>> >>>>> (I’m asking because I did see the same issue with a combination >>>> of an earlier Intel Fortran 2017 version and OpenMPI on an >>>> Intel/Infiniband Linux HPC machine … but not Intel Fortran 2016 on >>>> the same machine. Perhaps I can revive my access to that >>>> combination somehow.) >>>>> >>>>> Best wishes >>>>> Volker >>>>> >>>>>> On Jul 26, 2017, at 5:55 PM, George Bosilca >>>> <bosi...@icl.utk.edu <mailto:bosi...@icl.utk.edu>> wrote: >>>>>> >>>>>> I thought that maybe the underlying allreduce algorithm fails >>>> to support MPI_IN_PLACE correctly, but I can't replicate on any >>>> machine (including OSX) with any number of processes. >>>>>> >>>>>> George. >>>>>> >>>>>> >>>>>> >>>>>> On Wed, Jul 26, 2017 at 10:59 AM, Volker Blum >>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>> Thanks! >>>>>> >>>>>> I tried ‘use mpi’, which compiles fine. >>>>>> >>>>>> Same result as with ‘include mpif.h', in that the output is >>>>>> >>>>>> * MPI_IN_PLACE does not appear to work as intended. >>>>>> * Checking whether MPI_ALLREDUCE works at all. >>>>>> * Without MPI_IN_PLACE, MPI_ALLREDUCE appears to work. >>>>>> >>>>>> Hm. Any other thoughts? >>>>>> >>>>>> Thanks again! >>>>>> Best wishes >>>>>> Volker >>>>>> >>>>>>> On Jul 26, 2017, at 4:06 PM, Gilles Gouaillardet >>>> <gilles.gouaillar...@gmail.com >>>> <mailto:gilles.gouaillar...@gmail.com>> wrote: >>>>>>> >>>>>>> Volker, >>>>>>> >>>>>>> With mpi_f08, you have to declare >>>>>>> >>>>>>> Type(MPI_Comm) :: mpi_comm_global >>>>>>> >>>>>>> (I am afk and not 100% sure of the syntax) >>>>>>> >>>>>>> A simpler option is to >>>>>>> >>>>>>> use mpi >>>>>>> >>>>>>> Cheers, >>>>>>> >>>>>>> Gilles >>>>>>> >>>>>>> Volker Blum <volker.b...@duke.edu >>>> <mailto:volker.b...@duke.edu>> wrote: >>>>>>>> Hi Gilles, >>>>>>>> >>>>>>>> Thank you very much for the response! >>>>>>>> >>>>>>>> Unfortunately, I don’t have access to a different system >>>> with the issue right now. As I said, it’s not new; it just keeps >>>> creeping up unexpectedly again on different platforms. What >>>> puzzles me is that I’ve encountered the same problem with low but >>>> reasonable frequency over a period of now over five years. >>>>>>>> >>>>>>>> We can’t require F’08 in our application, unfortunately, >>>> since this standard is too new. Since we maintain a large >>>> application that has to run on a broad range of platforms, Fortran >>>> 2008 would not work for many of our users. In a few years, this >>>> will be different, but not yet. >>>>>>>> >>>>>>>> On gfortran: In our own tests, unfortunately, Intel Fortran >>>> consistently produced much faster executable code in the past. The >>>> latter observation may also change someday, but for us, the >>>> performance difference was an important constraint. >>>>>>>> >>>>>>>> I did suspect mpif.h, too. Not sure how to best test this >>>> hypothesis, however. >>>>>>>> >>>>>>>> Just replacing >>>>>>>> >>>>>>>>> include 'mpif.h' >>>>>>>>> with >>>>>>>>> use mpi_f08 >>>>>>>> >>>>>>>> did not work, for me. >>>>>>>> >>>>>>>> This produces a number of compilation errors: >>>>>>>> >>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> mpif90 >>>> check_mpi_in_place_08.f90 -o check_mpi_in_place_08.x >>>>>>>> check_mpi_in_place_08.f90(55): error #6303: The assignment >>>> operation or the binary expression operation is invalid for the >>>> data types of the two operands. [MPI_COMM_WORLD] >>>>>>>> mpi_comm_global = MPI_COMM_WORLD >>>>>>>> ----------------------^ >>>>>>>> check_mpi_in_place_08.f90(57): error #6285: There is no >>>> matching specific subroutine for this generic subroutine call. >>>> [MPI_COMM_SIZE] >>>>>>>> call MPI_COMM_SIZE(mpi_comm_global, n_tasks, mpierr) >>>>>>>> ---------^ >>>>>>>> check_mpi_in_place_08.f90(58): error #6285: There is no >>>> matching specific subroutine for this generic subroutine call. >>>> [MPI_COMM_RANK] >>>>>>>> call MPI_COMM_RANK(mpi_comm_global, myid, mpierr) >>>>>>>> ---------^ >>>>>>>> check_mpi_in_place_08.f90(75): error #6285: There is no >>>> matching specific subroutine for this generic subroutine call. >>>> [MPI_ALLREDUCE] >>>>>>>> call MPI_ALLREDUCE(MPI_IN_PLACE, & >>>>>>>> ---------^ >>>>>>>> check_mpi_in_place_08.f90(94): error #6285: There is no >>>> matching specific subroutine for this generic subroutine call. >>>> [MPI_ALLREDUCE] >>>>>>>> call MPI_ALLREDUCE(check_success, aux_check_success, 1, >>>> MPI_LOGICAL, & >>>>>>>> ---------^ >>>>>>>> check_mpi_in_place_08.f90(119): error #6285: There is no >>>> matching specific subroutine for this generic subroutine call. >>>> [MPI_ALLREDUCE] >>>>>>>> call MPI_ALLREDUCE(test_data(:), & >>>>>>>> ------------^ >>>>>>>> check_mpi_in_place_08.f90(140): error #6285: There is no >>>> matching specific subroutine for this generic subroutine call. >>>> [MPI_ALLREDUCE] >>>>>>>> call MPI_ALLREDUCE(check_conventional_mpi, >>>> aux_check_success, 1, MPI_LOGICAL, & >>>>>>>> ------------^ >>>>>>>> compilation aborted for check_mpi_in_place_08.f90 (code 1) >>>>>>>> >>>>>>>> This is an interesting result, however … what might I be >>>> missing? Another use statement? >>>>>>>> >>>>>>>> Best wishes >>>>>>>> Volker >>>>>>>> >>>>>>>>> On Jul 26, 2017, at 2:53 PM, Gilles Gouaillardet >>>> <gilles.gouaillar...@gmail.com >>>> <mailto:gilles.gouaillar...@gmail.com>> wrote: >>>>>>>>> >>>>>>>>> Volker, >>>>>>>>> >>>>>>>>> thanks, i will have a look at it >>>>>>>>> >>>>>>>>> meanwhile, if you can reproduce this issue on a more >>>> mainstream >>>>>>>>> platform (e.g. linux + gfortran) please let me know. >>>>>>>>> >>>>>>>>> since you are using ifort, Open MPI was built with Fortran >>>> 2008 >>>>>>>>> bindings, so you can replace >>>>>>>>> include 'mpif.h' >>>>>>>>> with >>>>>>>>> use mpi_f08 >>>>>>>>> and who knows, that might solve your issue >>>>>>>>> >>>>>>>>> >>>>>>>>> Cheers, >>>>>>>>> >>>>>>>>> Gilles >>>>>>>>> >>>>>>>>> On Wed, Jul 26, 2017 at 5:22 PM, Volker Blum >>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>>>>> Dear Gilles, >>>>>>>>>> >>>>>>>>>> Thank you very much for the fast answer. >>>>>>>>>> >>>>>>>>>> Darn. I feared it might not occur on all platforms, since >>>> my former Macbook >>>>>>>>>> (with an older OpenMPI version) no longer exhibited the >>>> problem, a different >>>>>>>>>> Linux/Intel Machine did last December, etc. >>>>>>>>>> >>>>>>>>>> On this specific machine, the configure line is >>>>>>>>>> >>>>>>>>>> ./configure CC=gcc FC=ifort F77=ifort >>>>>>>>>> >>>>>>>>>> ifort version 17.0.4 >>>>>>>>>> >>>>>>>>>> blum:/Users/blum/software/openmpi-3.0.0rc1> gcc -v >>>>>>>>>> Configured with: >>>> --prefix=/Applications/Xcode.app/Contents/Developer/usr >>>>>>>>>> --with-gxx-include-dir=/usr/include/c++/4.2.1 >>>>>>>>>> Apple LLVM version 8.1.0 (clang-802.0.42) >>>>>>>>>> Target: x86_64-apple-darwin16.6.0 >>>>>>>>>> Thread model: posix >>>>>>>>>> InstalledDir: >>>>>>>>>> >>>> >>>> /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin >>>>>>>>>> >>>>>>>>>> The full test program is appended. >>>>>>>>>> >>>>>>>>>> Compilation: >>>>>>>>>> >>>>>>>>>> mpif90 check_mpi_in_place.f90 >>>>>>>>>> >>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> which mpif90 >>>>>>>>>> /usr/local/openmpi-3.0.0rc1/bin/mpif90 >>>>>>>>>> >>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> which mpirun >>>>>>>>>> /usr/local/openmpi-3.0.0rc1/bin/mpirun >>>>>>>>>> >>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> mpirun -np >>>> 2 a.out >>>>>>>>>> * MPI_IN_PLACE does not appear to work as intended. >>>>>>>>>> * Checking whether MPI_ALLREDUCE works at all. >>>>>>>>>> * Without MPI_IN_PLACE, MPI_ALLREDUCE appears to work. >>>>>>>>>> >>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> mpirun -np >>>> 1 a.out >>>>>>>>>> * MPI_IN_PLACE does not appear to work as intended. >>>>>>>>>> * Checking whether MPI_ALLREDUCE works at all. >>>>>>>>>> * Without MPI_IN_PLACE, MPI_ALLREDUCE appears to work. >>>>>>>>>> >>>>>>>>>> Hopefully, no trivial mistakes in the testcase. I just >>>> spent a few days >>>>>>>>>> tracing this issue through a fairly large code, which is >>>> where the issue >>>>>>>>>> originally arose (and leads to wrong numbers). >>>>>>>>>> >>>>>>>>>> Best wishes >>>>>>>>>> Volker >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>>> On Jul 26, 2017, at 9:46 AM, Gilles Gouaillardet >>>>>>>>>>> <gilles.gouaillar...@gmail.com >>>> <mailto:gilles.gouaillar...@gmail.com>> wrote: >>>>>>>>>>> >>>>>>>>>>> Volker, >>>>>>>>>>> >>>>>>>>>>> i was unable to reproduce this issue on linux >>>>>>>>>>> >>>>>>>>>>> can you please post your full configure command line, >>>> your gnu >>>>>>>>>>> compiler version and the full test program ? >>>>>>>>>>> >>>>>>>>>>> also, how many mpi tasks are you running ? >>>>>>>>>>> >>>>>>>>>>> Cheers, >>>>>>>>>>> >>>>>>>>>>> Gilles >>>>>>>>>>> >>>>>>>>>>> On Wed, Jul 26, 2017 at 4:25 PM, Volker Blum >>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>>>>>>> Hi, >>>>>>>>>>>> >>>>>>>>>>>> I tried openmpi-3.0.0rc1.tar.gz using Intel Fortran >>>> 2017 and gcc on a >>>>>>>>>>>> current MacOS system. For this version, it seems to me >>>> that MPI_IN_PLACE >>>>>>>>>>>> returns incorrect results (while other MPI >>>> implementations, including some >>>>>>>>>>>> past OpenMPI versions, work fine). >>>>>>>>>>>> >>>>>>>>>>>> This can be seen with a simple Fortran example code, >>>> shown below. In the >>>>>>>>>>>> test, the values of all entries of an array “test_data” >>>> should be 1.0d0 if >>>>>>>>>>>> the behavior were as intended. However, the version of >>>> OpenMPI I have >>>>>>>>>>>> returns 0.d0 instead. >>>>>>>>>>>> >>>>>>>>>>>> I’ve seen this behavior on some other compute platforms >>>> too, in the past, >>>>>>>>>>>> so it wasn’t new to me. Still, I thought that this >>>> time, I’d ask. Any >>>>>>>>>>>> thoughts? >>>>>>>>>>>> >>>>>>>>>>>> Thank you, >>>>>>>>>>>> Best wishes >>>>>>>>>>>> Volker >>>>>>>>>>>> >>>>>>>>>>>> ! size of test data array >>>>>>>>>>>> integer :: n_data >>>>>>>>>>>> >>>>>>>>>>>> ! array that contains test data for MPI_IN_PLACE >>>>>>>>>>>> real*8, allocatable :: test_data(:) >>>>>>>>>>>> >>>>>>>>>>>> integer :: mpierr >>>>>>>>>>>> >>>>>>>>>>>> n_data = 10 >>>>>>>>>>>> >>>>>>>>>>>> allocate(test_data(n_data),stat=mpierr) >>>>>>>>>>>> >>>>>>>>>>>> ! seed test data array for allreduce call below >>>>>>>>>>>> if (myid.eq.0) then >>>>>>>>>>>> test_data(:) = 1.d0 >>>>>>>>>>>> else >>>>>>>>>>>> test_data(:) = 0.d0 >>>>>>>>>>>> end if >>>>>>>>>>>> >>>>>>>>>>>> ! Sum the test_data array over all MPI tasks >>>>>>>>>>>> call MPI_ALLREDUCE(MPI_IN_PLACE, & >>>>>>>>>>>> test_data(:), & >>>>>>>>>>>> n_data, & >>>>>>>>>>>> MPI_DOUBLE_PRECISION, & >>>>>>>>>>>> MPI_SUM, & >>>>>>>>>>>> mpi_comm_global, & >>>>>>>>>>>> mpierr ) >>>>>>>>>>>> >>>>>>>>>>>> ! The value of all entries of test_data should now be >>>> 1.d0 on all MPI >>>>>>>>>>>> tasks. >>>>>>>>>>>> ! If that is not the case, then the MPI_IN_PLACE flag >>>> may be broken. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Volker Blum >>>>>>>>>>>> Associate Professor >>>>>>>>>>>> Ab Initio Materials Simulations >>>>>>>>>>>> Duke University, MEMS Department >>>>>>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC >>>> 27708, USA >>>>>>>>>>>> >>>>>>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>>>>>> https://aims.pratt.duke.edu >>>>>>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>>>>>> Twitter: Aimsduke >>>>>>>>>>>> >>>>>>>>>>>> Office: 1111 Hudson Hall >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> users mailing list >>>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>>> >>>>>>>>>>>> >>>> >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e= >>>> >>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e=> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> users mailing list >>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>> >>>>>>>>>>> >>>> >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e= >>>> >>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e=> >>>>>>>>>> >>>>>>>>>> Volker Blum >>>>>>>>>> Associate Professor >>>>>>>>>> Ab Initio Materials Simulations >>>>>>>>>> Duke University, MEMS Department >>>>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC >>>> 27708, USA >>>>>>>>>> >>>>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>>>> https://aims.pratt.duke.edu >>>>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>>>> Twitter: Aimsduke >>>>>>>>>> >>>>>>>>>> Office: 1111 Hudson Hall >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>> >>>> >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e= >>>> >>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e=> >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>> >>>> >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e= >>>> >>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e=> >>>>>>>> >>>>>>>> Volker Blum >>>>>>>> Associate Professor >>>>>>>> Ab Initio Materials Simulations >>>>>>>> Duke University, MEMS Department >>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC >>>> 27708, USA >>>>>>>> >>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>> https://aims.pratt.duke.edu >>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>> Twitter: Aimsduke >>>>>>>> >>>>>>>> Office: 1111 Hudson Hall >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>> >>>> >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e= >>>> >>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e=> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>> >>>> >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e= >>>> >>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e=> >>>>>> >>>>>> Volker Blum >>>>>> Associate Professor >>>>>> Ab Initio Materials Simulations >>>>>> Duke University, MEMS Department >>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>>>> >>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>> https://aims.pratt.duke.edu >>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>> Twitter: Aimsduke >>>>>> >>>>>> Office: 1111 Hudson Hall >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>>> >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e=> >>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>> >>>> >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=4W_1_DI3QU9PRT2fhTXyUeXiu7HvTSNzX48E-9ifoTc&s=i3f7Olcbyor4Pu0hv6YlgO10cJ_XvOR13zZn7PPIGto&e= >>>> >>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=4W_1_DI3QU9PRT2fhTXyUeXiu7HvTSNzX48E-9ifoTc&s=i3f7Olcbyor4Pu0hv6YlgO10cJ_XvOR13zZn7PPIGto&e=> >>>>> >>>>> Volker Blum >>>>> Associate Professor >>>>> Ab Initio Materials Simulations >>>>> Duke University, MEMS Department >>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>>> >>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>> https://aims.pratt.duke.edu >>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>> Twitter: Aimsduke >>>>> >>>>> Office: 1111 Hudson Hall >>>>> >>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>> >>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e=> >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>> >>>> >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h6zBsOI45o8fovfy43A2FyCt-fL_yVpNbVSf1OA8CrQ&s=YNADyKbvRnxPmnDVmVlYmsYgZEr8m-etPBXLHPRkflw&e= >>>> >>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h6zBsOI45o8fovfy43A2FyCt-fL_yVpNbVSf1OA8CrQ&s=YNADyKbvRnxPmnDVmVlYmsYgZEr8m-etPBXLHPRkflw&e=> >>>> >>>> Volker Blum >>>> Associate Professor >>>> Ab Initio Materials Simulations >>>> Duke University, MEMS Department >>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>> >>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>> https://aims.pratt.duke.edu >>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>> Twitter: Aimsduke >>>> >>>> Office: 1111 Hudson Hall >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>> >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>> >>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e=> >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>> >>> _______________________________________________ >>> users mailing list >>> users@lists.open-mpi.org >>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >> >> Volker Blum >> Associate Professor >> Ab Initio Materials Simulations >> Duke University, MEMS Department >> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >> >> volker.b...@duke.edu >> https://aims.pratt.duke.edu >> +1 (919) 660 5279 >> Twitter: Aimsduke >> >> Office: 1111 Hudson Hall >> >> >> >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIFaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h3HNno_BCKRMuxrZ26zNcyS_JTV7oDefvQtaN7NF3Kk&s=ENCO2vndR26kHOuem57zq1jvk1Tl1ftPBQKcQ2xIpwQ&e= >> > <test_constants.f90>_______________________________________________ > users mailing list > users@lists.open-mpi.org > https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h3HNno_BCKRMuxrZ26zNcyS_JTV7oDefvQtaN7NF3Kk&s=ENCO2vndR26kHOuem57zq1jvk1Tl1ftPBQKcQ2xIpwQ&e= Volker Blum Associate Professor Ab Initio Materials Simulations Duke University, MEMS Department 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA volker.b...@duke.edu https://aims.pratt.duke.edu +1 (919) 660 5279 Twitter: Aimsduke Office: 1111 Hudson Hall _______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users