Hi All, I installed gromacs 4.6.2 using the following cmake options:
cmake -DCMAKE_INSTALL_PREFIX=/home/kaushik/gromacs_executable/gromacs-new-mpi -DGMX_MPI=on After successful installation, when I run a test with mpirun mdrun, the program breaks with the following error: -------------------------------------------------------------------------- It looks like orte_init failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during orte_init; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): orte_ess_set_name failed --> Returned value Data unpack would read past end of buffer (-26) instead of ORTE_SUCCESS -------------------------------------------------------------------------- [n132:11207] [[10503,1],0] ORTE_ERROR_LOG: Data unpack would read past end of buffer in file ../../../orte/util/nidmap.c at line 371 [n132:11207] [[10503,1],0] ORTE_ERROR_LOG: Data unpack would read past end of buffer in file ../../../../../orte/mca/ess/base/ess_base_nidmap.c at line 62 [n132:11207] [[10503,1],0] ORTE_ERROR_LOG: Data unpack would read past end of buffer in file ../../../../../../orte/mca/ess/env/ess_env_module.c at line 173 [n132:11207] [[10503,1],0] ORTE_ERROR_LOG: Data unpack would read past end of buffer in file ../../../orte/runtime/orte_init.c at line 132 *** The MPI_Init() function was called before MPI_INIT was invoked. *** This is disallowed by the MPI standard. *** Your MPI job will now abort. MPI related forums suggest that error points to gromacs compiling against an older-mpi-library than the current distribution used during running mpirun mdrun. when I run an ldd on mdrun, I see : linux-vdso.so.1 => (0x00007fff571ff000) libgmxpreprocess.so.8 => /state/partition1/home/kaushik/gromacs_executable/gromacs-new-mpi/bin/./../lib/libgmxpreprocess.so.8 (0x00007fab42706000) libmd.so.8 => /state/partition1/home/kaushik/gromacs_executable/gromacs-new-mpi/bin/./../lib/libmd.so.8 (0x00007fab421f4000) libgmx.so.8 => /state/partition1/home/kaushik/gromacs_executable/gromacs-new-mpi/bin/./../lib/libgmx.so.8 (0x00007fab41a0c000) libfftw3f.so.3 => /opt/fftw/3.3.2/single_prec/lib/libfftw3f.so.3 (0x00007fab4170e000) libmpi_cxx.so.0 => /usr/lib/libmpi_cxx.so.0 (0x00007fab414e7000) libmpi.so.0 => /usr/lib/libmpi.so.0 (0x00007fab41237000) libdl.so.2 => /lib/libdl.so.2 (0x00007fab41033000) libm.so.6 => /lib/libm.so.6 (0x00007fab40db0000) libnuma.so.1 => /usr/lib/libnuma.so.1 (0x00007fab40ba8000) librt.so.1 => /lib/librt.so.1 (0x00007fab409a0000) libnsl.so.1 => /lib/libnsl.so.1 (0x00007fab40787000) libutil.so.1 => /lib/libutil.so.1 (0x00007fab40584000) libpthread.so.0 => /lib/libpthread.so.0 (0x00007fab40368000) libgomp.so.1 => /usr/local/lib64/libgomp.so.1 (0x00007fab40159000) libc.so.6 => /lib/libc.so.6 (0x00007fab3fdf7000) libstdc++.so.6 => /usr/local/lib64/libstdc++.so.6 (0x00007fab3faef000) libgcc_s.so.1 => /usr/local/lib64/libgcc_s.so.1 (0x00007fab3f8da000) libopen-rte.so.0 => /usr/lib/libopen-rte.so.0 (0x00007fab3f68e000) /lib64/ld-linux-x86-64.so.2 (0x00007fab429f8000) libopen-pal.so.0 => /usr/lib/libopen-pal.so.0 (0x00007fab3f438000) Our current open-mpi libraries are in /usr/local/lib and not /usr/lib; I checked the CMakeCache.txt to check for the paths of the mpi related environment variables and stuff: //Extra MPI libraries to link against MPI_EXTRA_LIBRARY:STRING=/usr/local/lib/libmpi.so;/usr/lib/libdl.so;/usr/lib/libm.so;/usr/lib/libnuma.so;/usr/lib/librt.so;/usr/lib/libnsl.so;/usr/lib/libutil.so;/usr/lib/libm.so;/usr/lib/libdl.so //MPI include path MPI_INCLUDE_PATH:STRING=/usr/local/include //MPI library to link against MPI_LIBRARY:FILEPATH=/usr/local/lib/libmpi_cxx.so //MPI linking flags MPI_LINK_FLAGS:STRING=-Wl,--export-dynamic When all these correctly point to the /usr/local/lib; why is GROMACS compiling with /usr/lib/libmpi.so which is old? Ares there any other environment variables/locations that I am failing to set correctly during compilation? Thanks, Kaushik -- View this message in context: http://gromacs.5086.x6.nabble.com/gromacs-4-6-2-MPI-distribution-location-problems-tp5009002.html Sent from the GROMACS Users Forum mailing list archive at Nabble.com. -- gmx-users mailing list gmx-users@gromacs.org http://lists.gromacs.org/mailman/listinfo/gmx-users * Please search the archive at http://www.gromacs.org/Support/Mailing_Lists/Search before posting! * Please don't post (un)subscribe requests to the list. Use the www interface or send it to gmx-users-requ...@gromacs.org. * Can't post? Read http://www.gromacs.org/Support/Mailing_Lists