Are you able to run on a single node? Is there any chance you can upgrade your Open MPI? 1.10 is ancient and isn't really supported any more. 4.0.2 is the current version.
On Dec 5, 2019, at 7:15 PM, Guido granda muñoz <guidogra...@gmail.com<mailto:guidogra...@gmail.com>> wrote: Hello Jeff, Thank you for replying. I ran it using PBS like this: #!/bin/bash #PBS -l nodes=2:ppn=32 #PBS -N cond_0_h3 #PBS -o cond_0_h3.o #PBS -e cond_0_h3.e PATH=$PATH:/usr/mpi/intel/openmpi-1.10.3/bin LD_LIBRARY_PATH=/share/apps/composerxe-2011.2.137/lib/intel64:$LD_LIBRARY_PATH cd $PBS_O_WORKDIR mpirun -np 64 ./flash4 Besides that, my .bashrc has the following lines: # Source global definitions if [ -f /etc/bashrc ]; then . /etc/bashrc fi #irya.guido.intel export PATH=$PATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/bin export INCLUDE=$INCLUDE:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/include export CPATH=$CPATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/include export LIBRARY_PATH=$LIBRARY_PATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib ###intel open-mpi system### export PATH=$PATH:/usr/mpi/intel/openmpi-1.10.3/bin export INCLUDE=$INCLUDE:/usr/mpi/intel/openmpi-1.10.3/include export CPATH=$CPATH:/usr/mpi/intel/openmpi-1.10.3/include export LIBRARY_PATH=$LIBRARY_PATH:/usr/mpi/intel/openmpi-1.10.3/lib export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/mpi/intel/openmpi-1.10.3/lib ##anaconda 3 ## export PATH="$PATH:/home/guido/anaconda3/bin" # commented out by conda initialize # Intel export PATH=$PATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/ export INCLUDE=$INCLUDE:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/ export CPATH=$CPATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/ export LIBRARY_PATH=$LIBRARY_PATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/ export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/ How Can I check that what you suggested was the reason for this error? Cheers, El jue., 5 dic. 2019 a las 18:02, Jeff Squyres (jsquyres) (<jsquy...@cisco.com<mailto:jsquy...@cisco.com>>) escribió: How did you try to execute your application? An error message like this can mean that you accidentally mixed versions of Open MPI within your run (e.g., used Open MPI va.b.c on node A but used Open MPI vx.y.z on node B). > On Dec 5, 2019, at 5:28 PM, Guido granda muñoz via users > <users@lists.open-mpi.org<mailto:users@lists.open-mpi.org>> wrote: > > Hello open-mpi users, > I'm getting some problem while using openmpi-1.10.3. The executable was > compiled using : (ldd output) > > linux-vdso.so.1 => (0x00007fffd9e8b000) > libhdf5.so.10 => > /home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib/libhdf5.so.10 > (0x00002ac4313c4000) > libhdf5_fortran.so.10 => > /home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib/libhdf5_fortran.so.10 > (0x00002ac4319ca000) > libz.so.1 => /lib64/libz.so.1 (0x00002ac431c2d000) > libmpi_usempif08.so.11 => > /usr/mpi/intel/openmpi-1.10.3/lib/libmpi_usempif08.so.11 (0x00002ac431e44000) > libmpi_usempi_ignore_tkr.so.6 => > /usr/mpi/intel/openmpi-1.10.3/lib/libmpi_usempi_ignore_tkr.so.6 > (0x00002ac432077000) > libmpi_mpifh.so.12 => /usr/mpi/intel/openmpi-1.10.3/lib/libmpi_mpifh.so.12 > (0x00002ac432280000) > libmpi.so.12 => /usr/mpi/intel/openmpi-1.10.3/lib/libmpi.so.12 > (0x00002ac4324df000) > libm.so.6 => /lib64/libm.so.6 (0x00002ac4327e7000) > libpthread.so.0 => /lib64/libpthread.so.0 (0x00002ac432a6b000) > libc.so.6 => /lib64/libc.so.6 (0x00002ac432c88000) > libgcc_s.so.1 => /lib64/libgcc_s.so.1 (0x00002ac43301d000) > libdl.so.2 => /lib64/libdl.so.2 (0x00002ac433233000) > libimf.so => > /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libimf.so > (0x00002ac433437000) > libsvml.so => > /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libsvml.so > (0x00002ac43381b000) > libintlc.so.5 => > /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libintlc.so.5 > (0x00002ac433ec3000) > libifport.so.5 => > /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libifport.so.5 > (0x00002ac434013000) > libifcore.so.5 => > /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libifcore.so.5 > (0x00002ac43414c000) > libopen-rte.so.12 => /usr/mpi/intel/openmpi-1.10.3/lib/libopen-rte.so.12 > (0x00002ac4343ad000) > libopen-pal.so.13 => /usr/mpi/intel/openmpi-1.10.3/lib/libopen-pal.so.13 > (0x00002ac43464e000) > librt.so.1 => /lib64/librt.so.1 (0x00002ac43496a000) > libutil.so.1 => /lib64/libutil.so.1 (0x00002ac434b72000) > libifcoremt.so.5 => > /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libifcoremt.so.5 > (0x00002ac434d76000) > /lib64/ld-linux-x86-64.so.2 (0x00002ac4311a2000) > > When I run it, I get the following message: > > [compute-0-34.local:17553] [[5279,0],0] mca_oob_tcp_recv_handler: invalid > message type: 15 > -------------------------------------------------------------------------- > mpirun noticed that the job aborted, but has no info as to the process > that caused that situation. > > The executable was also compiled using hdf5-1.8.20 > I really don't know wht this error means, could please help me? > Cheers, > -- > Guido -- Jeff Squyres jsquy...@cisco.com<mailto:jsquy...@cisco.com> -- Guido -- Jeff Squyres jsquy...@cisco.com<mailto:jsquy...@cisco.com>