On Jan 18, 2019, at 12:43 PM, Matt Thompson <fort...@gmail.com> wrote: > > With some help, I managed to build an Open MPI 4.0.0 with:
We can discuss each of these params to let you know what they are. > ./configure --disable-wrapper-rpath --disable-wrapper-runpath Did you have a reason for disabling these? They're generally good things. What they do is add linker flags to the wrapper compilers (i.e., mpicc and friends) that basically put a default path to find libraries at run time (that can/will in most cases override LD_LIBRARY_PATH -- but you can override these linked-in-default-paths if you want/need to). > --with-psm2 Ensure that Open MPI can include support for the PSM2 library, and abort configure if it cannot. > --with-slurm Ensure that Open MPI can include support for SLURM, and abort configure if it cannot. > --enable-mpi1-compatibility Add support for MPI_Address and other MPI-1 functions that have since been deleted from the MPI 3.x specification. > --with-ucx Ensure that Open MPI can include support for UCX, and abort configure if it cannot. > --with-pmix=/usr/nlocal/pmix/2.1 Tells Open MPI to use the PMIx that is installed at /usr/nlocal/pmix/2.1 (instead of using the PMIx that is bundled internally to Open MPI's source code tree/expanded tarball). Unless you have a reason to use the external PMIx, the internal/bundled PMIx is usually sufficient. > --with-libevent=/usr Same as previous; change "pmix" to "libevent" (i.e., use the external libevent instead of the bundled libevent). > CC=icc CXX=icpc FC=ifort Specify the exact compilers to use. > The MPI 1 is because I need to build HDF5 eventually and I added psm2 because > it's an Omnipath cluster. The libevent was probably a red herring as > libevent-devel wasn't installed on the system. It was eventually, and I just > didn't remove the flag. And I saw no errors in the build! Might as well remove the --with-libevent if you don't need it. > However, I seem to have built an Open MPI that doesn't work: > > (1099)(master) $ mpirun --version > mpirun (Open MPI) 4.0.0 > > Report bugs to http://www.open-mpi.org/community/help/ > (1100)(master) $ mpirun -np 4 ./helloWorld.mpi3.SLES12.OMPI400.exe > > It just sits there...forever. Can the gurus here help me figure out what I > managed to break? Perhaps I added too much to my configure line? Not enough? There could be a few things going on here. Are you running inside a SLURM job? E.g., in a "salloc" job, or in an "sbatch" script? -- Jeff Squyres jsquy...@cisco.com _______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users