, which is clearly an unknown switch
(plus uppercase o three is a known one)
At the end of the configure, make sure Fortran bindings are generated.
If the link error persists, you can
ldd /.../libmpi_mpifh.so | grep igatherv
and confirm the symbol does indeed exists
Cheers,
Gilles
On Tu
.com/users@lists.open-mpi.org/msg10375.html__;!!Nmw4Hv0!mHuOLtyAuSwka2PpXEsyHXX5N2xzBT7digyvflX1ZBibtEeKFnwlHKCk3JyuA2ePQ7Gy0N8$>
I hope this helps,
Gus Correa
On Thu, Jan 14, 2021 at 5:45 PM Passant A. Hafez via users
mailto:users@lists.open-mpi.org>> wrote:
Hello,
I'm having an error when trying to build OMPI 4.0.3 (also tried 4.1) with
Hello,
I'm having an error when trying to build OMPI 4.0.3 (also tried 4.1) with PGI
20.1
./configure CPP=cpp CC=pgcc CXX=pgc++ F77=pgf77 FC=pgf90 --prefix=$PREFIX
--with-ucx=$UCX_HOME --with-slurm --with-pmi=/opt/slurm/cluster/ibex/install
--with-cuda=$CUDATOOLKIT_HOME
in the make install
n MPI User's List
Cc: Passant A. Hafez
Subject: Re: [OMPI users] Debug OMPI errors
I'm not sure exactly what you are asking -- can you be more specific?
Are you asking if Open MPI can emit more detail when an error occurs and the
job aborts?
> On Jul 28, 2019, at 4:12 AM, Passant
Hello all,
I was wondering if I can enable some reasonable level of debugging for OMPI
errors, especially in the cases that just report that a process is killed (for
example MPI_ABORT was invoked) and that's it.
All the best,
--
Passant
___
users
rs] undefined reference error related to ucx
Passant,
UCX 1.6.0 is not yet officially released, and it seems Open MPI
(4.0.1) does not support it yet, and some porting is needed.
Cheers,
Gilles
On Tue, Jun 25, 2019 at 5:13 PM Passant A. Hafez via users
wrote:
>
> Hello,
>
>
> I&
Hello,
I'm trying to build ompi 4.0.1 with external ucx 1.6.0 but I'm getting
../../../opal/.libs/libopen-pal.so: undefined reference to
`uct_ep_create_connected'
collect2: error: ld returned 1 exit status
configure line for ompi
./configure --prefix=/opt/ompi401_ucx16 --with-slurm --with-hwlo
, and not able to
guarantee that all other processes were killed!
srun: error: cn603-20-l: tasks 0-1: Exited with exit code 1
I'm suspecting Slurm, but anyways, how can I troubleshoot this?
The program is a simple MPI Hello World code.
All the best,
--
Passant A. Hafez | HPC Applica
ull output of your job ?
Cheers,
Gilles
On 3/12/2019 7:59 AM, Passant A. Hafez wrote:
>
> Hello,
>
>
> So we now have Slurm 18.08.6-2 compiled with PMIx 3.1.2
>
> then I installed openmpi 4.0.0 with:
>
> --with-slurm --with-pmix=internal --with-libevent=internal
] Local abort before MPI_INIT completed completed
successfully, but am not able to aggregate error messages, and not able to
guarantee that all other processes were killed!
for each process, please advise! what's going wrong here?
All the best,
--
Passant A. Hafez | HPC Applica
Thank you so much for help, Gilles!
All the best,
--
Passant A. Hafez | HPC Applications Specialist
KAUST Supercomputing Core Laboratory (KSL)
King Abdullah University of Science and Technology
Building 1, Al-Khawarizmi, Room 0123
Mobile : +966 (0) 55-247-9568
Mobile : +20 (0) 106-146-9644
=external
to be on the safe side I also invite you to pass --with-hwloc=external
to the configure command line
Cheers,
Gilles
On Sun, Feb 24, 2019 at 1:54 AM Passant A. Hafez
wrote:
>
> Hello Gilles,
>
> Here are some details:
>
> Slurm 18.08.4
>
> PMIx 2.2.1 (
)
PMIx support: Internal
Tested also different installations for 3.1.2 and got errors similar to
400ext_2x2.out
(NOT-SUPPORTED in file event/pmix_event_registration.c at line 101)
All the best,
--
Passant A. Hafez | HPC Applications Specialist
KAUST Supercomputing Core Laboratory (KSL)
King
ersions of each PMIx and Open MPI that
should be working well with Slurm 18.08, it'd be great.
Also, what is the difference between using internal vs external PMIx
installations?
All the best,
--
Passant A. Hafez | HPC Applications Specialist
KAUST Supercomputing Core Laboratory (KSL
14 matches
Mail list logo