[OMPI users] 3.1.1 Bindings Change

2018-07-03 Thread Matt Thompson
Dear Open MPI Gurus,

In the latest 3.1.1 announcement, I saw:

- Fix dummy variable names for the mpi and mpi_f08 Fortran bindings to
  match the MPI standard.  This may break applications which use
  name-based parameters in Fortran which used our internal names
  rather than those documented in the MPI standard.

Is there an example of this change somewhere (in the Git issues or another
place)? I don't think we have anything in our software that would be hit by
this (since we test/run our code with Intel MPI, MPT as well as Open MPI),
but I want to be sure we don't have some hidden #ifdef OPENMPI somewhere.

Matt

-- 
Matt Thompson
   “The fact is, this is about us identifying what we do best and
   finding more ways of doing less of it better” -- Director of Better Anna
Rampton
___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Re: [OMPI users] [EXTERNAL] Re: OpenMPI 3.1.0 Lock Up on POWER9 w/ CUDA9.2

2018-07-03 Thread Nathan Hjelm via users

Found this issue. PR #5374 fixes it. Will make its way into the v3.0.x and 
v3.1.x release series.

-Nathan

On Jul 02, 2018, at 02:36 PM, Nathan Hjelm  wrote:


The result should be the same with v3.1.1. I will investigate on our Coral test 
systems.

-Nathan

On Jul 02, 2018, at 02:23 PM, "Hammond, Simon David via users" 
 wrote:

Howard,

 

This fixed the issue with OpenMPI 3.1.0. Do you want me to try the same with 
3.1.1 as well?

 

S.

 

-- 

Si Hammond

Scalable Computer Architectures

Sandia National Laboratories, NM, USA

 

 

From: users  on behalf of Howard Pritchard 

Reply-To: Open MPI Users 
Date: Monday, July 2, 2018 at 1:34 PM
To: Open MPI Users 
Subject: Re: [OMPI users] [EXTERNAL] Re: OpenMPI 3.1.0 Lock Up on POWER9 w/ 
CUDA9.2

 

HI Si,

 

Could you add --disable-builtin-atomics

 

to the configure options and see if the hang goes away?

 

Howard

 

 

2018-07-02 8:48 GMT-06:00 Jeff Squyres (jsquyres) via users 
:

Simon --

You don't currently have another Open MPI installation in your PATH / 
LD_LIBRARY_PATH, do you?

I have seen dependency library loads cause "make check" to get confused, and 
instead of loading the libraries from the build tree, actually load some -- but not all 
-- of the required OMPI/ORTE/OPAL/etc. libraries from an installation tree.  Hilarity 
ensues (to include symptoms such as running forever).

Can you double check that you have no Open MPI libraries in your LD_LIBRARY_PATH before 
running "make check" on the build tree?





On Jun 30, 2018, at 3:18 PM, Hammond, Simon David via users 
 wrote:

Nathan,

Same issue with OpenMPI 3.1.1 on POWER9 with GCC 7.2.0 and CUDA9.2.

S.

--
Si Hammond
Scalable Computer Architectures
Sandia National Laboratories, NM, USA
[Sent from remote connection, excuse typos]


On 6/16/18, 10:10 PM, "Nathan Hjelm"  wrote:

    Try the latest nightly tarball for v3.1.x. Should be fixed. 


On Jun 16, 2018, at 5:48 PM, Hammond, Simon David via users 
 wrote:

The output from the test in question is:

Single thread test. Time: 0 s 10182 us 10 nsec/poppush
Atomics thread finished. Time: 0 s 169028 us 169 nsec/poppush


S.

--
Si Hammond
Scalable Computer Architectures
Sandia National Laboratories, NM, USA
[Sent from remote connection, excuse typos]


On 6/16/18, 5:45 PM, "Hammond, Simon David"  wrote:

   Hi OpenMPI Team,

   We have recently updated an install of OpenMPI on POWER9 system (configuration details 
below). We migrated from OpenMPI 2.1 to OpenMPI 3.1. We seem to have a symptom where code 
than ran before is now locking up and making no progress, getting stuck in wait-all 
operations. While I think it's prudent for us to root cause this a little more, I have 
gone back and rebuilt MPI and re-run the "make check" tests. The opal_fifo test 
appears to hang forever. I am not sure if this is the cause of our issue but wanted to 
report that we are seeing this on our system.

   OpenMPI 3.1.0 Configuration:

   ./configure 
--prefix=/home/projects/ppc64le-pwr9-nvidia/openmpi/3.1.0-nomxm/gcc/7.2.0/cuda/9.2.88
 --with-cuda=$CUDA_ROOT --enable-mpi-java --enable-java 
--with-lsf=/opt/lsf/10.1 
--with-lsf-libdir=/opt/lsf/10.1/linux3.10-glibc2.17-ppc64le/lib --with-verbs

   GCC versions are 7.2.0, built by our team. CUDA is 9.2.88 from NVIDIA for 
POWER9 (standard download from their website). We enable IBM's JDK 8.0.0.
   RedHat: Red Hat Enterprise Linux Server release 7.5 (Maipo)

   Output:

   make[3]: Entering directory `/home/sdhammo/openmpi/openmpi-3.1.0/test/class'
   make[4]: Entering directory `/home/sdhammo/openmpi/openmpi-3.1.0/test/class'
   PASS: ompi_rb_tree
   PASS: opal_bitmap
   PASS: opal_hash_table
   PASS: opal_proc_table
   PASS: opal_tree
   PASS: opal_list
   PASS: opal_value_array
   PASS: opal_pointer_array
   PASS: opal_lifo
   

   Output from Top:

   20   0   73280   4224   2560 S 800.0  0.0  17:22.94 lt-opal_fifo

   -- 
   Si Hammond

   Scalable Computer Architectures
   Sandia National Laboratories, NM, USA
   [Sent from remote connection, excuse typos]




___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users



___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users



--
Jeff Squyres
jsquy...@cisco.com


___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

 

___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users
___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users