Re: [OMPI users] Error with building OMPI with PGI

2021-01-19 Thread Passant A. Hafez via users
, which is clearly an unknown switch (plus uppercase o three is a known one) At the end of the configure, make sure Fortran bindings are generated. If the link error persists, you can ldd /.../libmpi_mpifh.so | grep igatherv and confirm the symbol does indeed exists Cheers, Gilles On Tu

Re: [OMPI users] Error with building OMPI with PGI

2021-01-19 Thread Passant A. Hafez via users
.com/users@lists.open-mpi.org/msg10375.html__;!!Nmw4Hv0!mHuOLtyAuSwka2PpXEsyHXX5N2xzBT7digyvflX1ZBibtEeKFnwlHKCk3JyuA2ePQ7Gy0N8$> I hope this helps, Gus Correa On Thu, Jan 14, 2021 at 5:45 PM Passant A. Hafez via users mailto:users@lists.open-mpi.org>> wrote: Hello, I'm having an error when trying to build OMPI 4.0.3 (also tried 4.1) with

[OMPI users] Error with building OMPI with PGI

2021-01-14 Thread Passant A. Hafez via users
Hello, I'm having an error when trying to build OMPI 4.0.3 (also tried 4.1) with PGI 20.1 ./configure CPP=cpp CC=pgcc CXX=pgc++ F77=pgf77 FC=pgf90 --prefix=$PREFIX --with-ucx=$UCX_HOME --with-slurm --with-pmi=/opt/slurm/cluster/ibex/install --with-cuda=$CUDATOOLKIT_HOME in the make install

Re: [OMPI users] Debug OMPI errors

2019-08-04 Thread Passant A. Hafez via users
n MPI User's List Cc: Passant A. Hafez Subject: Re: [OMPI users] Debug OMPI errors I'm not sure exactly what you are asking -- can you be more specific? Are you asking if Open MPI can emit more detail when an error occurs and the job aborts? > On Jul 28, 2019, at 4:12 AM, Passant

[OMPI users] Debug OMPI errors

2019-07-28 Thread Passant A. Hafez via users
Hello all, I was wondering if I can enable some reasonable level of debugging for OMPI errors, especially in the cases that just report that a process is killed (for example MPI_ABORT was invoked) and that's it. All the best, -- Passant ___ users

Re: [OMPI users] undefined reference error related to ucx

2019-06-25 Thread Passant A. Hafez via users
rs] undefined reference error related to ucx Passant, UCX 1.6.0 is not yet officially released, and it seems Open MPI (4.0.1) does not support it yet, and some porting is needed. Cheers, Gilles On Tue, Jun 25, 2019 at 5:13 PM Passant A. Hafez via users wrote: > > Hello, > > > I&

[OMPI users] undefined reference error related to ucx

2019-06-25 Thread Passant A. Hafez via users
Hello, I'm trying to build ompi 4.0.1 with external ucx 1.6.0 but I'm getting ../../../opal/.libs/libopen-pal.so: undefined reference to `uct_ep_create_connected' collect2: error: ld returned 1 exit status configure line for ompi ./configure --prefix=/opt/ompi401_ucx16 --with-slurm --with-hwlo

Re: [OMPI users] Building PMIx and Slurm support

2019-03-12 Thread Passant A. Hafez
, and not able to guarantee that all other processes were killed! srun: error: cn603-20-l: tasks 0-1: Exited with exit code 1 I'm suspecting Slurm, but anyways, how can I troubleshoot this? The program is a simple MPI Hello World code. All the best, -- Passant A. Hafez | HPC Applica

Re: [OMPI users] Building PMIx and Slurm support

2019-03-11 Thread Passant A. Hafez
ull output of your job ? Cheers, Gilles On 3/12/2019 7:59 AM, Passant A. Hafez wrote: > > Hello, > > > So we now have Slurm 18.08.6-2 compiled with PMIx 3.1.2 > > then I installed openmpi 4.0.0 with: > > --with-slurm --with-pmix=internal --with-libevent=internal

Re: [OMPI users] Building PMIx and Slurm support

2019-03-11 Thread Passant A. Hafez
] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed! for each process, please advise! what's going wrong here? All the best, -- Passant A. Hafez | HPC Applica

Re: [OMPI users] Building PMIx and Slurm support

2019-02-24 Thread Passant A. Hafez
Thank you so much for help, Gilles! All the best, -- Passant A. Hafez | HPC Applications Specialist KAUST Supercomputing Core Laboratory (KSL) King Abdullah University of Science and Technology Building 1, Al-Khawarizmi, Room 0123 Mobile : +966 (0) 55-247-9568 Mobile : +20 (0) 106-146-9644

Re: [OMPI users] Building PMIx and Slurm support

2019-02-23 Thread Passant A. Hafez
=external to be on the safe side I also invite you to pass --with-hwloc=external to the configure command line Cheers, Gilles On Sun, Feb 24, 2019 at 1:54 AM Passant A. Hafez wrote: > > Hello Gilles, > > Here are some details: > > Slurm 18.08.4 > > PMIx 2.2.1 (

Re: [OMPI users] Building PMIx and Slurm support

2019-02-23 Thread Passant A. Hafez
) PMIx support: Internal Tested also different installations for 3.1.2 and got errors similar to 400ext_2x2.out (NOT-SUPPORTED in file event/pmix_event_registration.c at line 101) All the best, -- Passant A. Hafez | HPC Applications Specialist KAUST Supercomputing Core Laboratory (KSL) King

[OMPI users] Building PMIx and Slurm support

2019-02-23 Thread Passant A. Hafez
ersions of each PMIx and Open MPI that should be working well with Slurm 18.08, it'd be great. Also, what is the difference between using internal vs external PMIx installations? All the best, -- Passant A. Hafez | HPC Applications Specialist KAUST Supercomputing Core Laboratory (KSL