Hello,
I get the same error when running in a single node. I will try to use the
last version. Is there way to check if different versions of open mpi were
used in different nodes?
Cheers,

El jue., 5 dic. 2019 a las 19:10, Jeff Squyres (jsquyres) (<
jsquy...@cisco.com>) escribió:

> Are you able to run on a single node?
>
> Is there any chance you can upgrade your Open MPI?  1.10 is ancient and
> isn't really supported any more.  4.0.2 is the current version.
>
>
> On Dec 5, 2019, at 7:15 PM, Guido granda muñoz <guidogra...@gmail.com>
> wrote:
>
> Hello Jeff,
>
> Thank you for replying. I ran it using PBS like this:
>
> #!/bin/bash
> #PBS -l nodes=2:ppn=32
> #PBS -N cond_0_h3
> #PBS -o cond_0_h3.o
> #PBS -e cond_0_h3.e
>
>
> PATH=$PATH:/usr/mpi/intel/openmpi-1.10.3/bin
>
> LD_LIBRARY_PATH=/share/apps/composerxe-2011.2.137/lib/intel64:$LD_LIBRARY_PATH
> cd $PBS_O_WORKDIR
> mpirun -np 64 ./flash4
>
> Besides that, my .bashrc has the following lines:
> # Source global definitions
> if [ -f /etc/bashrc ]; then
>         . /etc/bashrc
> fi
> #irya.guido.intel
> export PATH=$PATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/bin
> export
> INCLUDE=$INCLUDE:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/include
> export
> CPATH=$CPATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/include
> export
> LIBRARY_PATH=$LIBRARY_PATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib
> export
> LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib
> ###intel open-mpi system###
> export PATH=$PATH:/usr/mpi/intel/openmpi-1.10.3/bin
> export INCLUDE=$INCLUDE:/usr/mpi/intel/openmpi-1.10.3/include
> export CPATH=$CPATH:/usr/mpi/intel/openmpi-1.10.3/include
> export LIBRARY_PATH=$LIBRARY_PATH:/usr/mpi/intel/openmpi-1.10.3/lib
> export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/usr/mpi/intel/openmpi-1.10.3/lib
> ##anaconda 3 ##
> export PATH="$PATH:/home/guido/anaconda3/bin"  # commented out by conda
> initialize
> # Intel
> export
> PATH=$PATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/
> export
> INCLUDE=$INCLUDE:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/
> export
> CPATH=$CPATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/
> export
> LIBRARY_PATH=$LIBRARY_PATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/
> export
> LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/share/apps/composerxe-2011.2.137/composerxe-2011.2.137/bin/intel64/
>
> How Can I check that what you suggested was the reason for this error?
> Cheers,
>
> El jue., 5 dic. 2019 a las 18:02, Jeff Squyres (jsquyres) (<
> jsquy...@cisco.com>) escribió:
>
>> How did you try to execute your application?
>>
>> An error message like this can mean that you accidentally mixed versions
>> of Open MPI within your run (e.g., used Open MPI va.b.c on node A but used
>> Open MPI vx.y.z on node B).
>>
>>
>> > On Dec 5, 2019, at 5:28 PM, Guido granda muñoz via users <
>> users@lists.open-mpi.org> wrote:
>> >
>> > Hello open-mpi users,
>> > I'm getting some problem while using openmpi-1.10.3. The executable was
>> compiled using : (ldd output)
>> >
>> > linux-vdso.so.1 =>  (0x00007fffd9e8b000)
>> > libhdf5.so.10 =>
>> /home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib/libhdf5.so.10
>> (0x00002ac4313c4000)
>> > libhdf5_fortran.so.10 =>
>> /home/guido/libraries/compiled_with_intel/hdf5-1.8.20/lib/libhdf5_fortran.so.10
>> (0x00002ac4319ca000)
>> > libz.so.1 => /lib64/libz.so.1 (0x00002ac431c2d000)
>> > libmpi_usempif08.so.11 =>
>> /usr/mpi/intel/openmpi-1.10.3/lib/libmpi_usempif08.so.11
>> (0x00002ac431e44000)
>> > libmpi_usempi_ignore_tkr.so.6 =>
>> /usr/mpi/intel/openmpi-1.10.3/lib/libmpi_usempi_ignore_tkr.so.6
>> (0x00002ac432077000)
>> > libmpi_mpifh.so.12 =>
>> /usr/mpi/intel/openmpi-1.10.3/lib/libmpi_mpifh.so.12 (0x00002ac432280000)
>> > libmpi.so.12 => /usr/mpi/intel/openmpi-1.10.3/lib/libmpi.so.12
>> (0x00002ac4324df000)
>> > libm.so.6 => /lib64/libm.so.6 (0x00002ac4327e7000)
>> > libpthread.so.0 => /lib64/libpthread.so.0 (0x00002ac432a6b000)
>> > libc.so.6 => /lib64/libc.so.6 (0x00002ac432c88000)
>> > libgcc_s.so.1 => /lib64/libgcc_s.so.1 (0x00002ac43301d000)
>> > libdl.so.2 => /lib64/libdl.so.2 (0x00002ac433233000)
>> > libimf.so =>
>> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libimf.so
>> (0x00002ac433437000)
>> > libsvml.so =>
>> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libsvml.so
>> (0x00002ac43381b000)
>> > libintlc.so.5 =>
>> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libintlc.so.5
>> (0x00002ac433ec3000)
>> > libifport.so.5 =>
>> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libifport.so.5
>> (0x00002ac434013000)
>> > libifcore.so.5 =>
>> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libifcore.so.5
>> (0x00002ac43414c000)
>> > libopen-rte.so.12 =>
>> /usr/mpi/intel/openmpi-1.10.3/lib/libopen-rte.so.12 (0x00002ac4343ad000)
>> > libopen-pal.so.13 =>
>> /usr/mpi/intel/openmpi-1.10.3/lib/libopen-pal.so.13 (0x00002ac43464e000)
>> > librt.so.1 => /lib64/librt.so.1 (0x00002ac43496a000)
>> > libutil.so.1 => /lib64/libutil.so.1 (0x00002ac434b72000)
>> > libifcoremt.so.5 =>
>> /share/apps/composerxe-2011.2.137/composerxe-2011.2.137/compiler/lib/intel64/libifcoremt.so.5
>> (0x00002ac434d76000)
>> > /lib64/ld-linux-x86-64.so.2 (0x00002ac4311a2000)
>> >
>> > When I run it, I get the following message:
>> >
>> > [compute-0-34.local:17553] [[5279,0],0] mca_oob_tcp_recv_handler:
>> invalid message type: 15
>> >
>> --------------------------------------------------------------------------
>> > mpirun noticed that the job aborted, but has no info as to the process
>> > that caused that situation.
>> >
>> > The executable was also compiled using hdf5-1.8.20
>> > I really don't know wht this error means, could please help me?
>> > Cheers,
>> > --
>> > Guido
>>
>>
>> --
>> Jeff Squyres
>> jsquy...@cisco.com
>>
>>
>
> --
> Guido
>
>
>
> --
> Jeff Squyres
> jsquy...@cisco.com
>
>

-- 
Guido

Reply via email to