Are you able to run on a single node?

Is there any chance you can upgrade your Open MPI?  1.10 is ancient and isn't 
really supported any more.  4.0.2 is the current version.


On Dec 5, 2019, at 7:15 PM, Guido granda muñoz 
<guidogra...@gmail.com<mailto:guidogra...@gmail.com>> wrote:

Hello Jeff,

Thank you for replying. I ran it using PBS like this:

#!/bin/bash
#PBS -l nodes=2:ppn=32
#PBS -N cond_0_h3
#PBS -o cond_0_h3.o
#PBS -e cond_0_h3.e


PATH=$PATH:/usr/mpi/intel/openmpi-1.10.3/bin
LD_LIBRARY_PATH=/share/apps/composerxe-2011.2.137/lib/intel64:$LD_LIBRARY_PATH
cd $PBS_O_WORKDIR
mpirun -np 64 ./flash4

Besides that, my .bashrc has the following lines:
# Source global definitions
if [ -f /etc/bashrc ]; then
        . /etc/bashrc
fi
#irya.guido.intel
export PATH=$PATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/bin
export 
INCLUDE=$INCLUDE:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/include
export 
CPATH=$CPATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/include
export 
LIBRARY_PATH=$LIBRARY_PATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib
export 
LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib
###intel open-mpi system###
export PATH=$PATH:/usr/mpi/intel/openmpi-1.10.3/bin
export INCLUDE=$INCLUDE:/usr/mpi/intel/openmpi-1.10.3/include
export CPATH=$CPATH:/usr/mpi/intel/openmpi-1.10.3/include
export LIBRARY_PATH=$LIBRARY_PATH:/usr/mpi/intel/openmpi-1.10.3/lib
export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/mpi/intel/openmpi-1.10.3/lib
##anaconda 3 ##
export PATH="$PATH:/home/guido/anaconda3/bin"  # commented out by conda 
initialize
# Intel
export 
PATH=$PATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/
export 
INCLUDE=$INCLUDE:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/
export 
CPATH=$CPATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/
export 
LIBRARY_PATH=$LIBRARY_PATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/
export 
LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/

How Can I check that what you suggested was the reason for this error?
Cheers,

El jue., 5 dic. 2019 a las 18:02, Jeff Squyres (jsquyres) 
(<jsquy...@cisco.com<mailto:jsquy...@cisco.com>>) escribió:
How did you try to execute your application?

An error message like this can mean that you accidentally mixed versions of 
Open MPI within your run (e.g., used Open MPI va.b.c on node A but used Open 
MPI vx.y.z on node B).


> On Dec 5, 2019, at 5:28 PM, Guido granda muñoz via users 
> <users@lists.open-mpi.org<mailto:users@lists.open-mpi.org>> wrote:
>
> Hello open-mpi users,
> I'm getting some problem while using openmpi-1.10.3. The executable was 
> compiled using : (ldd output)
>
> linux-vdso.so.1 =>  (0x00007fffd9e8b000)
> libhdf5.so.10 => 
> /home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib/libhdf5.so.10 
> (0x00002ac4313c4000)
> libhdf5_fortran.so.10 => 
> /home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib/libhdf5_fortran.so.10
>  (0x00002ac4319ca000)
> libz.so.1 => /lib64/libz.so.1 (0x00002ac431c2d000)
> libmpi_usempif08.so.11 => 
> /usr/mpi/intel/openmpi-1.10.3/lib/libmpi_usempif08.so.11 (0x00002ac431e44000)
> libmpi_usempi_ignore_tkr.so.6 => 
> /usr/mpi/intel/openmpi-1.10.3/lib/libmpi_usempi_ignore_tkr.so.6 
> (0x00002ac432077000)
> libmpi_mpifh.so.12 => /usr/mpi/intel/openmpi-1.10.3/lib/libmpi_mpifh.so.12 
> (0x00002ac432280000)
> libmpi.so.12 => /usr/mpi/intel/openmpi-1.10.3/lib/libmpi.so.12 
> (0x00002ac4324df000)
> libm.so.6 => /lib64/libm.so.6 (0x00002ac4327e7000)
> libpthread.so.0 => /lib64/libpthread.so.0 (0x00002ac432a6b000)
> libc.so.6 => /lib64/libc.so.6 (0x00002ac432c88000)
> libgcc_s.so.1 => /lib64/libgcc_s.so.1 (0x00002ac43301d000)
> libdl.so.2 => /lib64/libdl.so.2 (0x00002ac433233000)
> libimf.so => 
> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libimf.so
>  (0x00002ac433437000)
> libsvml.so => 
> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libsvml.so
>  (0x00002ac43381b000)
> libintlc.so.5 => 
> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libintlc.so.5
>  (0x00002ac433ec3000)
> libifport.so.5 => 
> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libifport.so.5
>  (0x00002ac434013000)
> libifcore.so.5 => 
> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libifcore.so.5
>  (0x00002ac43414c000)
> libopen-rte.so.12 => /usr/mpi/intel/openmpi-1.10.3/lib/libopen-rte.so.12 
> (0x00002ac4343ad000)
> libopen-pal.so.13 => /usr/mpi/intel/openmpi-1.10.3/lib/libopen-pal.so.13 
> (0x00002ac43464e000)
> librt.so.1 => /lib64/librt.so.1 (0x00002ac43496a000)
> libutil.so.1 => /lib64/libutil.so.1 (0x00002ac434b72000)
> libifcoremt.so.5 => 
> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libifcoremt.so.5
>  (0x00002ac434d76000)
> /lib64/ld-linux-x86-64.so.2 (0x00002ac4311a2000)
>
> When I run it, I get the following message:
>
> [compute-0-34.local:17553] [[5279,0],0] mca_oob_tcp_recv_handler: invalid 
> message type: 15
> --------------------------------------------------------------------------
> mpirun noticed that the job aborted, but has no info as to the process
> that caused that situation.
>
> The executable was also compiled using hdf5-1.8.20
> I really don't know wht this error means, could please help me?
> Cheers,
> --
> Guido


--
Jeff Squyres
jsquy...@cisco.com<mailto:jsquy...@cisco.com>



--
Guido


--
Jeff Squyres
jsquy...@cisco.com<mailto:jsquy...@cisco.com>

Reply via email to