Update: > I have yet to run a full regression test on our actual code to ensure that > there are no other side effects. I don’t expect any, though.
Indeed. Fixes all regressions that I had observed. Best wishes Volker > On Jul 27, 2017, at 11:47 AM, Volker Blum <volker.b...@duke.edu> wrote: > > Dear Gilles, > > Thank you! Indeed, this appears to address the issue in my test. > > I have yet to run a full regression test on our actual code to ensure that > there are no other side effects. I don’t expect any, though. > > ** > > Interestingly, removing '-lmpi_usempif08 -lmpi_usempi_ignore_tkr’ actually > fixes a second regression that I had found only last night and not yet > reported. > > Specifically, I have a very well tested mpi_alltoallv call (part of our > standard regression tests). Versions of this exact call live in multiple > codes, including libraries. With the unpatched OpenMPI version on the same > macbook and in a specific case, this suddenly resulted in: > > [egr-vwb3-mbp2:3573] *** An error occurred in MPI_Alltoallv > [egr-vwb3-mbp2:3573] *** reported by process [1710358529,0] > [egr-vwb3-mbp2:3573] *** on communicator MPI COMMUNICATOR 7 SPLIT FROM 0 > [egr-vwb3-mbp2:3573] *** MPI_ERR_TRUNCATE: message truncated > [egr-vwb3-mbp2:3573] *** MPI_ERRORS_ARE_FATAL (processes in this communicator > will now abort, > [egr-vwb3-mbp2:3573] *** and potentially your MPI job) > [egr-vwb3-mbp2.local:03572] 1 more process has sent help message > help-mpi-errors.txt / mpi_errors_are_fatal > [egr-vwb3-mbp2.local:03572] Set MCA parameter "orte_base_help_aggregate" to 0 > to see all help / error messages > > This regression is also fixed. > > I do not have a simple test case for this behavior yet since I was still > investigating it. My plan was to create a test case, but extracting such a > case from the larger code we have is more difficult. > > I can still try to create a test case for this as well but I’ll be watching > with great interest if the earlier MPI_IN_PLACE regression can be nailed down > and fixed in a more general way. Perhaps the two are related. > > Thanks again, > Best wishes > Volker > >> On Jul 27, 2017, at 10:46 AM, Gilles Gouaillardet <gil...@rist.or.jp> wrote: >> >> Volker, >> >> >> since you are only using >> >> include 'mpif.h' >> >> >> a workaround is you edit your /.../share/openmpi/mpifort-wrapper-data.txt >> and simply remove '-lmpi_usempif08 -lmpi_usempi_ignore_tkr' >> >> >> Cheers, >> >> Gilles >> >> On 7/27/2017 3:28 PM, Volker Blum wrote: >>> Thanks! >>> >>> If you wish, please also keep me posted. >>> >>> Best wishes >>> Volker >>> >>>> On Jul 27, 2017, at 7:50 AM, Gilles Gouaillardet >>>> <gilles.gouaillar...@gmail.com> wrote: >>>> >>>> Thanks Jeff for your offer, i will contact you off-list later >>>> >>>> >>>> i tried a gcc+gfortran and gcc+ifort on both linux and OS X >>>> so far, only gcc+ifort on OS X is failing >>>> i will try icc+ifort on OS X from now >>>> >>>> short story, MPI_IN_PLACE is not recognized as such by the ompi >>>> fortran wrapper, and i do not know why. >>>> >>>> the attached program can be used to evidence the issue. >>>> >>>> >>>> Cheers, >>>> >>>> Gilles >>>> >>>> On Thu, Jul 27, 2017 at 2:15 PM, Volker Blum <volker.b...@duke.edu> wrote: >>>>> Thanks! That’s great. Sounds like the exact combination I have here. >>>>> >>>>> Thanks also to George. Sorry that the test did not trigger on a more >>>>> standard platform - that would have simplified things. >>>>> >>>>> Best wishes >>>>> Volker >>>>> >>>>>> On Jul 27, 2017, at 3:56 AM, Gilles Gouaillardet <gil...@rist.or.jp> >>>>>> wrote: >>>>>> >>>>>> Folks, >>>>>> >>>>>> >>>>>> I am able to reproduce the issue on OS X (Sierra) with stock gcc (aka >>>>>> clang) and ifort 17.0.4 >>>>>> >>>>>> >>>>>> i will investigate this from now >>>>>> >>>>>> >>>>>> Cheers, >>>>>> >>>>>> Gilles >>>>>> >>>>>> On 7/27/2017 9:28 AM, George Bosilca wrote: >>>>>>> Volker, >>>>>>> >>>>>>> Unfortunately, I can't replicate with icc. I tried on a x86_64 box with >>>>>>> Intel compiler chain 17.0.4 20170411 to no avail. I also tested the >>>>>>> 3.0.0-rc1 tarball and the current master, and you test completes >>>>>>> without errors on all cases. >>>>>>> >>>>>>> Once you figure out an environment where you can consistently replicate >>>>>>> the issue, I would suggest to attach to the processes and: >>>>>>> - make sure the MPI_IN_PLACE as seen through the Fortran layer matches >>>>>>> what the C layer expects >>>>>>> - what is the collective algorithm used by Open MPI >>>>>>> >>>>>>> I have a "Fortran 101" level question. When you pass an array a(:) as >>>>>>> argument, what exactly gets passed via the Fortran interface to the >>>>>>> corresponding C function ? >>>>>>> >>>>>>> George. >>>>>>> >>>>>>> On Wed, Jul 26, 2017 at 1:55 PM, Volker Blum <volker.b...@duke.edu >>>>>>> <mailto:volker.b...@duke.edu>> wrote: >>>>>>> >>>>>>> Thanks! Yes, trying with Intel 2017 would be very nice. >>>>>>> >>>>>>>> On Jul 26, 2017, at 6:12 PM, George Bosilca <bosi...@icl.utk.edu >>>>>>> <mailto:bosi...@icl.utk.edu>> wrote: >>>>>>>> No, I don't have (or used where they were available) the Intel >>>>>>> compiler. I used clang and gfortran. I can try on a Linux box with >>>>>>> the Intel 2017 compilers. >>>>>>>> George. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Wed, Jul 26, 2017 at 11:59 AM, Volker Blum >>>>>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>>> Did you use Intel Fortran 2017 as well? >>>>>>>> >>>>>>>> (I’m asking because I did see the same issue with a combination >>>>>>> of an earlier Intel Fortran 2017 version and OpenMPI on an >>>>>>> Intel/Infiniband Linux HPC machine … but not Intel Fortran 2016 on >>>>>>> the same machine. Perhaps I can revive my access to that >>>>>>> combination somehow.) >>>>>>>> Best wishes >>>>>>>> Volker >>>>>>>> >>>>>>>>> On Jul 26, 2017, at 5:55 PM, George Bosilca >>>>>>> <bosi...@icl.utk.edu <mailto:bosi...@icl.utk.edu>> wrote: >>>>>>>>> I thought that maybe the underlying allreduce algorithm fails >>>>>>> to support MPI_IN_PLACE correctly, but I can't replicate on any >>>>>>> machine (including OSX) with any number of processes. >>>>>>>>> George. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, Jul 26, 2017 at 10:59 AM, Volker Blum >>>>>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>>>> Thanks! >>>>>>>>> >>>>>>>>> I tried ‘use mpi’, which compiles fine. >>>>>>>>> >>>>>>>>> Same result as with ‘include mpif.h', in that the output is >>>>>>>>> >>>>>>>>> * MPI_IN_PLACE does not appear to work as intended. >>>>>>>>> * Checking whether MPI_ALLREDUCE works at all. >>>>>>>>> * Without MPI_IN_PLACE, MPI_ALLREDUCE appears to work. >>>>>>>>> >>>>>>>>> Hm. Any other thoughts? >>>>>>>>> >>>>>>>>> Thanks again! >>>>>>>>> Best wishes >>>>>>>>> Volker >>>>>>>>> >>>>>>>>>> On Jul 26, 2017, at 4:06 PM, Gilles Gouaillardet >>>>>>> <gilles.gouaillar...@gmail.com >>>>>>> <mailto:gilles.gouaillar...@gmail.com>> wrote: >>>>>>>>>> Volker, >>>>>>>>>> >>>>>>>>>> With mpi_f08, you have to declare >>>>>>>>>> >>>>>>>>>> Type(MPI_Comm) :: mpi_comm_global >>>>>>>>>> >>>>>>>>>> (I am afk and not 100% sure of the syntax) >>>>>>>>>> >>>>>>>>>> A simpler option is to >>>>>>>>>> >>>>>>>>>> use mpi >>>>>>>>>> >>>>>>>>>> Cheers, >>>>>>>>>> >>>>>>>>>> Gilles >>>>>>>>>> >>>>>>>>>> Volker Blum <volker.b...@duke.edu >>>>>>> <mailto:volker.b...@duke.edu>> wrote: >>>>>>>>>>> Hi Gilles, >>>>>>>>>>> >>>>>>>>>>> Thank you very much for the response! >>>>>>>>>>> >>>>>>>>>>> Unfortunately, I don’t have access to a different system >>>>>>> with the issue right now. As I said, it’s not new; it just keeps >>>>>>> creeping up unexpectedly again on different platforms. What >>>>>>> puzzles me is that I’ve encountered the same problem with low but >>>>>>> reasonable frequency over a period of now over five years. >>>>>>>>>>> We can’t require F’08 in our application, unfortunately, >>>>>>> since this standard is too new. Since we maintain a large >>>>>>> application that has to run on a broad range of platforms, Fortran >>>>>>> 2008 would not work for many of our users. In a few years, this >>>>>>> will be different, but not yet. >>>>>>>>>>> On gfortran: In our own tests, unfortunately, Intel Fortran >>>>>>> consistently produced much faster executable code in the past. The >>>>>>> latter observation may also change someday, but for us, the >>>>>>> performance difference was an important constraint. >>>>>>>>>>> I did suspect mpif.h, too. Not sure how to best test this >>>>>>> hypothesis, however. >>>>>>>>>>> Just replacing >>>>>>>>>>> >>>>>>>>>>>> include 'mpif.h' >>>>>>>>>>>> with >>>>>>>>>>>> use mpi_f08 >>>>>>>>>>> did not work, for me. >>>>>>>>>>> >>>>>>>>>>> This produces a number of compilation errors: >>>>>>>>>>> >>>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> mpif90 >>>>>>> check_mpi_in_place_08.f90 -o check_mpi_in_place_08.x >>>>>>>>>>> check_mpi_in_place_08.f90(55): error #6303: The assignment >>>>>>> operation or the binary expression operation is invalid for the >>>>>>> data types of the two operands. [MPI_COMM_WORLD] >>>>>>>>>>> mpi_comm_global = MPI_COMM_WORLD >>>>>>>>>>> ----------------------^ >>>>>>>>>>> check_mpi_in_place_08.f90(57): error #6285: There is no >>>>>>> matching specific subroutine for this generic subroutine call. >>>>>>> [MPI_COMM_SIZE] >>>>>>>>>>> call MPI_COMM_SIZE(mpi_comm_global, n_tasks, mpierr) >>>>>>>>>>> ---------^ >>>>>>>>>>> check_mpi_in_place_08.f90(58): error #6285: There is no >>>>>>> matching specific subroutine for this generic subroutine call. >>>>>>> [MPI_COMM_RANK] >>>>>>>>>>> call MPI_COMM_RANK(mpi_comm_global, myid, mpierr) >>>>>>>>>>> ---------^ >>>>>>>>>>> check_mpi_in_place_08.f90(75): error #6285: There is no >>>>>>> matching specific subroutine for this generic subroutine call. >>>>>>> [MPI_ALLREDUCE] >>>>>>>>>>> call MPI_ALLREDUCE(MPI_IN_PLACE, & >>>>>>>>>>> ---------^ >>>>>>>>>>> check_mpi_in_place_08.f90(94): error #6285: There is no >>>>>>> matching specific subroutine for this generic subroutine call. >>>>>>> [MPI_ALLREDUCE] >>>>>>>>>>> call MPI_ALLREDUCE(check_success, aux_check_success, 1, >>>>>>> MPI_LOGICAL, & >>>>>>>>>>> ---------^ >>>>>>>>>>> check_mpi_in_place_08.f90(119): error #6285: There is no >>>>>>> matching specific subroutine for this generic subroutine call. >>>>>>> [MPI_ALLREDUCE] >>>>>>>>>>> call MPI_ALLREDUCE(test_data(:), & >>>>>>>>>>> ------------^ >>>>>>>>>>> check_mpi_in_place_08.f90(140): error #6285: There is no >>>>>>> matching specific subroutine for this generic subroutine call. >>>>>>> [MPI_ALLREDUCE] >>>>>>>>>>> call MPI_ALLREDUCE(check_conventional_mpi, >>>>>>> aux_check_success, 1, MPI_LOGICAL, & >>>>>>>>>>> ------------^ >>>>>>>>>>> compilation aborted for check_mpi_in_place_08.f90 (code 1) >>>>>>>>>>> >>>>>>>>>>> This is an interesting result, however … what might I be >>>>>>> missing? Another use statement? >>>>>>>>>>> Best wishes >>>>>>>>>>> Volker >>>>>>>>>>> >>>>>>>>>>>> On Jul 26, 2017, at 2:53 PM, Gilles Gouaillardet >>>>>>> <gilles.gouaillar...@gmail.com >>>>>>> <mailto:gilles.gouaillar...@gmail.com>> wrote: >>>>>>>>>>>> Volker, >>>>>>>>>>>> >>>>>>>>>>>> thanks, i will have a look at it >>>>>>>>>>>> >>>>>>>>>>>> meanwhile, if you can reproduce this issue on a more >>>>>>> mainstream >>>>>>>>>>>> platform (e.g. linux + gfortran) please let me know. >>>>>>>>>>>> >>>>>>>>>>>> since you are using ifort, Open MPI was built with Fortran >>>>>>> 2008 >>>>>>>>>>>> bindings, so you can replace >>>>>>>>>>>> include 'mpif.h' >>>>>>>>>>>> with >>>>>>>>>>>> use mpi_f08 >>>>>>>>>>>> and who knows, that might solve your issue >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Cheers, >>>>>>>>>>>> >>>>>>>>>>>> Gilles >>>>>>>>>>>> >>>>>>>>>>>> On Wed, Jul 26, 2017 at 5:22 PM, Volker Blum >>>>>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>>>>>>>> Dear Gilles, >>>>>>>>>>>>> >>>>>>>>>>>>> Thank you very much for the fast answer. >>>>>>>>>>>>> >>>>>>>>>>>>> Darn. I feared it might not occur on all platforms, since >>>>>>> my former Macbook >>>>>>>>>>>>> (with an older OpenMPI version) no longer exhibited the >>>>>>> problem, a different >>>>>>>>>>>>> Linux/Intel Machine did last December, etc. >>>>>>>>>>>>> >>>>>>>>>>>>> On this specific machine, the configure line is >>>>>>>>>>>>> >>>>>>>>>>>>> ./configure CC=gcc FC=ifort F77=ifort >>>>>>>>>>>>> >>>>>>>>>>>>> ifort version 17.0.4 >>>>>>>>>>>>> >>>>>>>>>>>>> blum:/Users/blum/software/openmpi-3.0.0rc1> gcc -v >>>>>>>>>>>>> Configured with: >>>>>>> --prefix=/Applications/Xcode.app/Contents/Developer/usr >>>>>>>>>>>>> --with-gxx-include-dir=/usr/include/c++/4.2.1 >>>>>>>>>>>>> Apple LLVM version 8.1.0 (clang-802.0.42) >>>>>>>>>>>>> Target: x86_64-apple-darwin16.6.0 >>>>>>>>>>>>> Thread model: posix >>>>>>>>>>>>> InstalledDir: >>>>>>>>>>>>> >>>>>>> >>>>>>> /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin >>>>>>>>>>>>> The full test program is appended. >>>>>>>>>>>>> >>>>>>>>>>>>> Compilation: >>>>>>>>>>>>> >>>>>>>>>>>>> mpif90 check_mpi_in_place.f90 >>>>>>>>>>>>> >>>>>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> which mpif90 >>>>>>>>>>>>> /usr/local/openmpi-3.0.0rc1/bin/mpif90 >>>>>>>>>>>>> >>>>>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> which mpirun >>>>>>>>>>>>> /usr/local/openmpi-3.0.0rc1/bin/mpirun >>>>>>>>>>>>> >>>>>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> mpirun -np >>>>>>> 2 a.out >>>>>>>>>>>>> * MPI_IN_PLACE does not appear to work as intended. >>>>>>>>>>>>> * Checking whether MPI_ALLREDUCE works at all. >>>>>>>>>>>>> * Without MPI_IN_PLACE, MPI_ALLREDUCE appears to work. >>>>>>>>>>>>> >>>>>>>>>>>>> blum:/Users/blum/codes/fhi-aims/openmpi_test> mpirun -np >>>>>>> 1 a.out >>>>>>>>>>>>> * MPI_IN_PLACE does not appear to work as intended. >>>>>>>>>>>>> * Checking whether MPI_ALLREDUCE works at all. >>>>>>>>>>>>> * Without MPI_IN_PLACE, MPI_ALLREDUCE appears to work. >>>>>>>>>>>>> >>>>>>>>>>>>> Hopefully, no trivial mistakes in the testcase. I just >>>>>>> spent a few days >>>>>>>>>>>>> tracing this issue through a fairly large code, which is >>>>>>> where the issue >>>>>>>>>>>>> originally arose (and leads to wrong numbers). >>>>>>>>>>>>> >>>>>>>>>>>>> Best wishes >>>>>>>>>>>>> Volker >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>> On Jul 26, 2017, at 9:46 AM, Gilles Gouaillardet >>>>>>>>>>>>>> <gilles.gouaillar...@gmail.com >>>>>>> <mailto:gilles.gouaillar...@gmail.com>> wrote: >>>>>>>>>>>>>> Volker, >>>>>>>>>>>>>> >>>>>>>>>>>>>> i was unable to reproduce this issue on linux >>>>>>>>>>>>>> >>>>>>>>>>>>>> can you please post your full configure command line, >>>>>>> your gnu >>>>>>>>>>>>>> compiler version and the full test program ? >>>>>>>>>>>>>> >>>>>>>>>>>>>> also, how many mpi tasks are you running ? >>>>>>>>>>>>>> >>>>>>>>>>>>>> Cheers, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Gilles >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Wed, Jul 26, 2017 at 4:25 PM, Volker Blum >>>>>>> <volker.b...@duke.edu <mailto:volker.b...@duke.edu>> wrote: >>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I tried openmpi-3.0.0rc1.tar.gz using Intel Fortran >>>>>>> 2017 and gcc on a >>>>>>>>>>>>>>> current MacOS system. For this version, it seems to me >>>>>>> that MPI_IN_PLACE >>>>>>>>>>>>>>> returns incorrect results (while other MPI >>>>>>> implementations, including some >>>>>>>>>>>>>>> past OpenMPI versions, work fine). >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> This can be seen with a simple Fortran example code, >>>>>>> shown below. In the >>>>>>>>>>>>>>> test, the values of all entries of an array “test_data” >>>>>>> should be 1.0d0 if >>>>>>>>>>>>>>> the behavior were as intended. However, the version of >>>>>>> OpenMPI I have >>>>>>>>>>>>>>> returns 0.d0 instead. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I’ve seen this behavior on some other compute platforms >>>>>>> too, in the past, >>>>>>>>>>>>>>> so it wasn’t new to me. Still, I thought that this >>>>>>> time, I’d ask. Any >>>>>>>>>>>>>>> thoughts? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thank you, >>>>>>>>>>>>>>> Best wishes >>>>>>>>>>>>>>> Volker >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ! size of test data array >>>>>>>>>>>>>>> integer :: n_data >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ! array that contains test data for MPI_IN_PLACE >>>>>>>>>>>>>>> real*8, allocatable :: test_data(:) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> integer :: mpierr >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> n_data = 10 >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> allocate(test_data(n_data),stat=mpierr) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ! seed test data array for allreduce call below >>>>>>>>>>>>>>> if (myid.eq.0) then >>>>>>>>>>>>>>> test_data(:) = 1.d0 >>>>>>>>>>>>>>> else >>>>>>>>>>>>>>> test_data(:) = 0.d0 >>>>>>>>>>>>>>> end if >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ! Sum the test_data array over all MPI tasks >>>>>>>>>>>>>>> call MPI_ALLREDUCE(MPI_IN_PLACE, & >>>>>>>>>>>>>>> test_data(:), & >>>>>>>>>>>>>>> n_data, & >>>>>>>>>>>>>>> MPI_DOUBLE_PRECISION, & >>>>>>>>>>>>>>> MPI_SUM, & >>>>>>>>>>>>>>> mpi_comm_global, & >>>>>>>>>>>>>>> mpierr ) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ! The value of all entries of test_data should now be >>>>>>> 1.d0 on all MPI >>>>>>>>>>>>>>> tasks. >>>>>>>>>>>>>>> ! If that is not the case, then the MPI_IN_PLACE flag >>>>>>> may be broken. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Volker Blum >>>>>>>>>>>>>>> Associate Professor >>>>>>>>>>>>>>> Ab Initio Materials Simulations >>>>>>>>>>>>>>> Duke University, MEMS Department >>>>>>>>>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC >>>>>>> 27708, USA >>>>>>>>>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>>>>>>>>> https://aims.pratt.duke.edu >>>>>>>>>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>>>>>>>>> Twitter: Aimsduke >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Office: 1111 Hudson Hall >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e=> >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> users mailing list >>>>>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=QLtQXqnGSkgQnmgI4RxZXa9R6FhMmgj2FLN452Q0Wis&s=BeracGkSHhIyI_bjKJqPHCqMuP-Se2pRmbiNfugkdK8&e=> >>>>>>>>>>>>> Volker Blum >>>>>>>>>>>>> Associate Professor >>>>>>>>>>>>> Ab Initio Materials Simulations >>>>>>>>>>>>> Duke University, MEMS Department >>>>>>>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC >>>>>>> 27708, USA >>>>>>>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>>>>>>> https://aims.pratt.duke.edu >>>>>>>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>>>>>>> Twitter: Aimsduke >>>>>>>>>>>>> >>>>>>>>>>>>> Office: 1111 Hudson Hall >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> users mailing list >>>>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>>>> >>>>>>> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e=> >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> users mailing list >>>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>>> >>>>>>> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=e9pjil1vV3SDa40dQJww0p-d0LhgyQzX_kPNhmz-oUE&s=Y4hrMiRzNuObkpm0vPojCqr6Cx6uS_wLxNyAfUaBz70&e=> >>>>>>>>>>> Volker Blum >>>>>>>>>>> Associate Professor >>>>>>>>>>> Ab Initio Materials Simulations >>>>>>>>>>> Duke University, MEMS Department >>>>>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC >>>>>>> 27708, USA >>>>>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>>>>> https://aims.pratt.duke.edu >>>>>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>>>>> Twitter: Aimsduke >>>>>>>>>>> >>>>>>>>>>> Office: 1111 Hudson Hall >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> users mailing list >>>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>>> >>>>>>> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e=> >>>>>>>>>> _______________________________________________ >>>>>>>>>> users mailing list >>>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>>> >>>>>>> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=HDubxzHgm3hz7-NQfgobz7rkGf0LWBTlGGqdgSoPCC4&s=2D1Arirt92pKR6i2-4KQKZ8YhSnZ2TPkouQQePHvNf0&e=> >>>>>>>>> Volker Blum >>>>>>>>> Associate Professor >>>>>>>>> Ab Initio Materials Simulations >>>>>>>>> Duke University, MEMS Department >>>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>>>>>>> >>>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>>> https://aims.pratt.duke.edu >>>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>>> Twitter: Aimsduke >>>>>>>>> >>>>>>>>> Office: 1111 Hudson Hall >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>>>>>> >>>>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e=> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> users mailing list >>>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>>> >>>>>>> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=4W_1_DI3QU9PRT2fhTXyUeXiu7HvTSNzX48E-9ifoTc&s=i3f7Olcbyor4Pu0hv6YlgO10cJ_XvOR13zZn7PPIGto&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=4W_1_DI3QU9PRT2fhTXyUeXiu7HvTSNzX48E-9ifoTc&s=i3f7Olcbyor4Pu0hv6YlgO10cJ_XvOR13zZn7PPIGto&e=> >>>>>>>> Volker Blum >>>>>>>> Associate Professor >>>>>>>> Ab Initio Materials Simulations >>>>>>>> Duke University, MEMS Department >>>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>>>>>> >>>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>>> https://aims.pratt.duke.edu >>>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>>> Twitter: Aimsduke >>>>>>>> >>>>>>>> Office: 1111 Hudson Hall >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>>>>> >>>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e=> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> users mailing list >>>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>>> >>>>>>> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h6zBsOI45o8fovfy43A2FyCt-fL_yVpNbVSf1OA8CrQ&s=YNADyKbvRnxPmnDVmVlYmsYgZEr8m-etPBXLHPRkflw&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h6zBsOI45o8fovfy43A2FyCt-fL_yVpNbVSf1OA8CrQ&s=YNADyKbvRnxPmnDVmVlYmsYgZEr8m-etPBXLHPRkflw&e=> >>>>>>> >>>>>>> Volker Blum >>>>>>> Associate Professor >>>>>>> Ab Initio Materials Simulations >>>>>>> Duke University, MEMS Department >>>>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>>>>> >>>>>>> volker.b...@duke.edu <mailto:volker.b...@duke.edu> >>>>>>> https://aims.pratt.duke.edu >>>>>>> +1 (919) 660 5279 <tel:%2B1%20%28919%29%20660%205279> >>>>>>> Twitter: Aimsduke >>>>>>> >>>>>>> Office: 1111 Hudson Hall >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >>>>>>> >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>>>> >>>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e=> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> _______________________________________________ >>>>>>> users mailing list >>>>>>> users@lists.open-mpi.org >>>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> users@lists.open-mpi.org >>>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=0X7DUjFfbjVVpI_r3F9MFV4Cs6DpbwqBCobJBfWB-DI&s=i7T7XLEvUGLBfkNjzFjixk1JmT_QB0nGMM2j_uhDWyQ&e= >>>>> Volker Blum >>>>> Associate Professor >>>>> Ab Initio Materials Simulations >>>>> Duke University, MEMS Department >>>>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>>>> >>>>> volker.b...@duke.edu >>>>> https://aims.pratt.duke.edu >>>>> +1 (919) 660 5279 >>>>> Twitter: Aimsduke >>>>> >>>>> Office: 1111 Hudson Hall >>>>> >>>>> >>>>> >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> users@lists.open-mpi.org >>>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIFaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h3HNno_BCKRMuxrZ26zNcyS_JTV7oDefvQtaN7NF3Kk&s=ENCO2vndR26kHOuem57zq1jvk1Tl1ftPBQKcQ2xIpwQ&e= >>>> <test_constants.f90>_______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org >>>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwICAg&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=h3HNno_BCKRMuxrZ26zNcyS_JTV7oDefvQtaN7NF3Kk&s=ENCO2vndR26kHOuem57zq1jvk1Tl1ftPBQKcQ2xIpwQ&e= >>> Volker Blum >>> Associate Professor >>> Ab Initio Materials Simulations >>> Duke University, MEMS Department >>> 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA >>> >>> volker.b...@duke.edu >>> https://aims.pratt.duke.edu >>> +1 (919) 660 5279 >>> Twitter: Aimsduke >>> >>> Office: 1111 Hudson Hall >>> >>> >>> >>> >>> _______________________________________________ >>> users mailing list >>> users@lists.open-mpi.org >>> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=_NZvZhsoGaDbJEC8nwG5oDMiMtSi6cOdtsEboB-uXKQ&s=gVgnzXkfNMEN3UwBtvjPBhrTJPNJs2ty-7pEpHKvImI&e= >>> >> >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org >> https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=_NZvZhsoGaDbJEC8nwG5oDMiMtSi6cOdtsEboB-uXKQ&s=gVgnzXkfNMEN3UwBtvjPBhrTJPNJs2ty-7pEpHKvImI&e= > > Volker Blum > Associate Professor > Ab Initio Materials Simulations > Duke University, MEMS Department > 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA > > volker.b...@duke.edu > https://aims.pratt.duke.edu > +1 (919) 660 5279 > Twitter: Aimsduke > > Office: 1111 Hudson Hall > > > > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://urldefense.proofpoint.com/v2/url?u=https-3A__rfd.newmexicoconsortium.org_mailman_listinfo_users&d=DwIGaQ&c=imBPVzF25OnBgGmVOlcsiEgHoG1i6YHLR0Sj_gZ4adc&r=I9QLPu689VeINkpRod6EQfprr_v-FLoLsAuSXIHhDsk&m=W7rZfrifaK6ckq8le_u6DtgDRCh4aktkT3ZXaCDnTNk&s=QohFSlUihCo0BvXuDZPf_a3bHR1j3UGdY-55Qr13SMc&e= Volker Blum Associate Professor Ab Initio Materials Simulations Duke University, MEMS Department 144 Hudson Hall, Box 90300, Duke University, Durham, NC 27708, USA volker.b...@duke.edu https://aims.pratt.duke.edu +1 (919) 660 5279 Twitter: Aimsduke Office: 1111 Hudson Hall _______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users