Re: [OMPI users] Several Bcast calls in a loop causing the code to hang

2015-02-20 Thread Sachin Krishnan
Hello Josh,

The command i use to compile the code is:

mpicc bcast_loop.c


To run the code I use:

mpirun -np 2 ./a.out

Output is unpredictable. It gets stuck at different places.

Im attaching lstopo and ompi_info outputs. Do you need any other info?


lstopo-no-graphics output:

Machine (3433MB)

  Socket L#0 + L3 L#0 (8192KB)

L2 L#0 (256KB) + L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0

  PU L#0 (P#0)

  PU L#1 (P#4)

L2 L#1 (256KB) + L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1

  PU L#2 (P#1)

  PU L#3 (P#5)

L2 L#2 (256KB) + L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2

  PU L#4 (P#2)

  PU L#5 (P#6)

L2 L#3 (256KB) + L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3

  PU L#6 (P#3)

  PU L#7 (P#7)

  HostBridge L#0

PCI 8086:0162

  GPU L#0 "card0"

  GPU L#1 "renderD128"

  GPU L#2 "controlD64"

PCI 8086:1502

  Net L#3 "eth0"

PCI 8086:1e02

  Block L#4 "sda"

  Block L#5 "sr0"


ompi_info output:


 Package: Open MPI builduser@anatol Distribution

Open MPI: 1.8.4

  Open MPI repo revision: v1.8.3-330-g0344f04

   Open MPI release date: Dec 19, 2014

Open RTE: 1.8.4

  Open RTE repo revision: v1.8.3-330-g0344f04

   Open RTE release date: Dec 19, 2014

OPAL: 1.8.4

  OPAL repo revision: v1.8.3-330-g0344f04

   OPAL release date: Dec 19, 2014

 MPI API: 3.0

Ident string: 1.8.4

  Prefix: /usr

 Configured architecture: i686-pc-linux-gnu

  Configure host: anatol

   Configured by: builduser

   Configured on: Sat Dec 20 17:00:34 PST 2014

  Configure host: anatol

Built by: builduser

Built on: Sat Dec 20 17:12:16 PST 2014

  Built host: anatol

  C bindings: yes

C++ bindings: yes

 Fort mpif.h: yes (all)

Fort use mpi: yes (full: ignore TKR)

   Fort use mpi size: deprecated-ompi-info-value

Fort use mpi_f08: yes

 Fort mpi_f08 compliance: The mpi_f08 module is available, but due to

  limitations in the /usr/bin/gfortran compiler,
does

  not support the following: array subsections,

  direct passthru (where possible) to underlying
Open

  MPI's C functionality

  Fort mpi_f08 subarrays: no

   Java bindings: no

  Wrapper compiler rpath: runpath

  C compiler: gcc

 C compiler absolute: /usr/bin/gcc

  C compiler family name: GNU

  C compiler version: 4.9.2

C++ compiler: g++

   C++ compiler absolute: /usr/bin/g++

   Fort compiler: /usr/bin/gfortran

   Fort compiler abs:

 Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::)

   Fort 08 assumed shape: yes

  Fort optional args: yes

  Fort INTERFACE: yes

Fort ISO_FORTRAN_ENV: yes

   Fort STORAGE_SIZE: yes

  Fort BIND(C) (all): yes

  Fort ISO_C_BINDING: yes

 Fort SUBROUTINE BIND(C): yes

   Fort TYPE,BIND(C): yes

 Fort T,BIND(C,name="a"): yes

Fort PRIVATE: yes

  Fort PROTECTED: yes

   Fort ABSTRACT: yes

   Fort ASYNCHRONOUS: yes

  Fort PROCEDURE: yes

   Fort C_FUNLOC: yes

 Fort f08 using wrappers: yes

 Fort MPI_SIZEOF: yes

 C profiling: yes

   C++ profiling: yes

   Fort mpif.h profiling: yes

  Fort use mpi profiling: yes

   Fort use mpi_f08 prof: yes

  C++ exceptions: no

  Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: yes,

  OMPI progress: no, ORTE progress: yes, Event lib:

  yes)

   Sparse Groups: no

  Internal debug support: no

  MPI interface warnings: yes

 MPI parameter check: runtime

Memory profiling support: no

Memory debugging support: no

 libltdl support: yes

   Heterogeneous support: no

 mpirun default --prefix: no

 MPI I/O support: yes

   MPI_WTIME support: gettimeofday

 Symbol vis. support: yes

   Host topology support: yes

  MPI extensions:

   FT Checkpoint support: no (checkpoint thread: no)

   C/R Enabled Debugging: no

 VampirTrace support: yes

  MPI_MAX_PROCESSOR_NAME: 256

MPI_MAX_ERROR_STRING: 256

 MPI_MAX_OBJECT_NAME: 64

MPI_MAX_INFO_KEY: 36

MPI_MAX_INFO_VAL: 256

   MPI_MAX_PORT_NAME: 1024

  MPI_MAX_DATAREP_STRING: 128

   MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.8.4)

MCA compress: bzip (MCA v2.0, API v2.0, Component v1.8.4)

MCA compress: gzip (MCA v2.0, API v2.0, Component v1.8.4)

 MCA crs: none (MCA v2.0, API v2.0, Component v1.8.4)

  MCA db: hash (MCA v2.0, API v1.0, Component v1.8.4)

  MCA db: print (MCA v2.0, API v1.0, Component v1

[OMPI users] Fix small typo in man

2015-02-20 Thread Harald Servat

Hello,

  may I suggest you to apply this small patch to fix a typo in the 
MPI_Intercomm_merge entry manual?


Best regards


WARNING / LEGAL TEXT: This message is intended only for the use of the
individual or entity to which it is addressed and may contain
information which is privileged, confidential, proprietary, or exempt
from disclosure under applicable law. If you are not the intended
recipient or the person responsible for delivering the message to the
intended recipient, you are strictly prohibited from disclosing,
distributing, copying, or in any way using this message. If you have
received this communication in error, please notify the sender and
destroy and delete any copies you may have received.

http://www.bsc.es/disclaimer--- MPI_Intercomm_merge.3in	2014-11-01 01:29:50.0 +0100
+++ MPI_Intercomm_merge.3in.new	2015-02-20 15:08:55.072114384 +0100
@@ -13,7 +13,7 @@
 .nf
 #include 
 int MPI_Intercomm_merge(MPI_Comm \fIintercomm\fP, int\fI high\fP, 
- MPI_Comm\fI *newintracomm\fP
+ MPI_Comm\fI *newintracomm\fP)

 .fi
 .SH Fortran Syntax


Re: [OMPI users] Fix small typo in man

2015-02-20 Thread Jeff Squyres (jsquyres)
Harald --

Many thanks for noticing and sending a patch.  I have committed the fix, and 
slated it for the next Open MPI release:

   https://github.com/open-mpi/ompi-release/pull/188


> On Feb 20, 2015, at 9:36 AM, Harald Servat  wrote:
> 
> Hello,
> 
>  may I suggest you to apply this small patch to fix a typo in the 
> MPI_Intercomm_merge entry manual?
> 
> Best regards
> 
> 
> WARNING / LEGAL TEXT: This message is intended only for the use of the
> individual or entity to which it is addressed and may contain
> information which is privileged, confidential, proprietary, or exempt
> from disclosure under applicable law. If you are not the intended
> recipient or the person responsible for delivering the message to the
> intended recipient, you are strictly prohibited from disclosing,
> distributing, copying, or in any way using this message. If you have
> received this communication in error, please notify the sender and
> destroy and delete any copies you may have received.
> 
> http://www.bsc.es/disclaimer___
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2015/02/26364.php


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/



Re: [OMPI users] Several Bcast calls in a loop causing the code to hang

2015-02-20 Thread Joshua Ladd
Sachin,

Are you running this on a single host or across multiple hosts (i.e. are
you communicating between processes via networking.) If it's on a single
host, then it might be an issue with shared memory.

Josh

On Fri, Feb 20, 2015 at 1:51 AM, Sachin Krishnan  wrote:

> Hello Josh,
>
> The command i use to compile the code is:
>
> mpicc bcast_loop.c
>
>
> To run the code I use:
>
> mpirun -np 2 ./a.out
>
> Output is unpredictable. It gets stuck at different places.
>
> Im attaching lstopo and ompi_info outputs. Do you need any other info?
>
>
> lstopo-no-graphics output:
>
> Machine (3433MB)
>
>   Socket L#0 + L3 L#0 (8192KB)
>
> L2 L#0 (256KB) + L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0
>
>   PU L#0 (P#0)
>
>   PU L#1 (P#4)
>
> L2 L#1 (256KB) + L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1
>
>   PU L#2 (P#1)
>
>   PU L#3 (P#5)
>
> L2 L#2 (256KB) + L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2
>
>   PU L#4 (P#2)
>
>   PU L#5 (P#6)
>
> L2 L#3 (256KB) + L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3
>
>   PU L#6 (P#3)
>
>   PU L#7 (P#7)
>
>   HostBridge L#0
>
> PCI 8086:0162
>
>   GPU L#0 "card0"
>
>   GPU L#1 "renderD128"
>
>   GPU L#2 "controlD64"
>
> PCI 8086:1502
>
>   Net L#3 "eth0"
>
> PCI 8086:1e02
>
>   Block L#4 "sda"
>
>   Block L#5 "sr0"
>
>
> ompi_info output:
>
>
>  Package: Open MPI builduser@anatol Distribution
>
> Open MPI: 1.8.4
>
>   Open MPI repo revision: v1.8.3-330-g0344f04
>
>Open MPI release date: Dec 19, 2014
>
> Open RTE: 1.8.4
>
>   Open RTE repo revision: v1.8.3-330-g0344f04
>
>Open RTE release date: Dec 19, 2014
>
> OPAL: 1.8.4
>
>   OPAL repo revision: v1.8.3-330-g0344f04
>
>OPAL release date: Dec 19, 2014
>
>  MPI API: 3.0
>
> Ident string: 1.8.4
>
>   Prefix: /usr
>
>  Configured architecture: i686-pc-linux-gnu
>
>   Configure host: anatol
>
>Configured by: builduser
>
>Configured on: Sat Dec 20 17:00:34 PST 2014
>
>   Configure host: anatol
>
> Built by: builduser
>
> Built on: Sat Dec 20 17:12:16 PST 2014
>
>   Built host: anatol
>
>   C bindings: yes
>
> C++ bindings: yes
>
>  Fort mpif.h: yes (all)
>
> Fort use mpi: yes (full: ignore TKR)
>
>Fort use mpi size: deprecated-ompi-info-value
>
> Fort use mpi_f08: yes
>
>  Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
>
>   limitations in the /usr/bin/gfortran compiler,
> does
>
>   not support the following: array subsections,
>
>   direct passthru (where possible) to underlying
> Open
>
>   MPI's C functionality
>
>   Fort mpi_f08 subarrays: no
>
>Java bindings: no
>
>   Wrapper compiler rpath: runpath
>
>   C compiler: gcc
>
>  C compiler absolute: /usr/bin/gcc
>
>   C compiler family name: GNU
>
>   C compiler version: 4.9.2
>
> C++ compiler: g++
>
>C++ compiler absolute: /usr/bin/g++
>
>Fort compiler: /usr/bin/gfortran
>
>Fort compiler abs:
>
>  Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::)
>
>Fort 08 assumed shape: yes
>
>   Fort optional args: yes
>
>   Fort INTERFACE: yes
>
> Fort ISO_FORTRAN_ENV: yes
>
>Fort STORAGE_SIZE: yes
>
>   Fort BIND(C) (all): yes
>
>   Fort ISO_C_BINDING: yes
>
>  Fort SUBROUTINE BIND(C): yes
>
>Fort TYPE,BIND(C): yes
>
>  Fort T,BIND(C,name="a"): yes
>
> Fort PRIVATE: yes
>
>   Fort PROTECTED: yes
>
>Fort ABSTRACT: yes
>
>Fort ASYNCHRONOUS: yes
>
>   Fort PROCEDURE: yes
>
>Fort C_FUNLOC: yes
>
>  Fort f08 using wrappers: yes
>
>  Fort MPI_SIZEOF: yes
>
>  C profiling: yes
>
>C++ profiling: yes
>
>Fort mpif.h profiling: yes
>
>   Fort use mpi profiling: yes
>
>Fort use mpi_f08 prof: yes
>
>   C++ exceptions: no
>
>   Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support:
> yes,
>
>   OMPI progress: no, ORTE progress: yes, Event lib:
>
>   yes)
>
>Sparse Groups: no
>
>   Internal debug support: no
>
>   MPI interface warnings: yes
>
>  MPI parameter check: runtime
>
> Memory profiling support: no
>
> Memory debugging support: no
>
>  libltdl support: yes
>
>Heterogeneous support: no
>
>  mpirun default --prefix: no
>
>  MPI I/O support: yes
>
>MPI_WTIME support: gettimeofday
>
>  Symbol vis. support: yes
>
>Host topology support: yes
>
>   MPI extensions:
>
>FT Checkpoint support: no (checkpoint thread: no)
>
>C/R Enabled Debugging: no
>
>  VampirTrace support: y

Re: [OMPI users] Several Bcast calls in a loop causing the code to hang

2015-02-20 Thread Sachin Krishnan
Josh,

Thanks for the help.
I'm running on a single host. How do I confirm that it is an issue with the
shared memory?

Sachin

On Fri, Feb 20, 2015 at 11:58 PM, Joshua Ladd  wrote:

> Sachin,
>
> Are you running this on a single host or across multiple hosts (i.e. are
> you communicating between processes via networking.) If it's on a single
> host, then it might be an issue with shared memory.
>
> Josh
>
> On Fri, Feb 20, 2015 at 1:51 AM, Sachin Krishnan 
> wrote:
>
>> Hello Josh,
>>
>> The command i use to compile the code is:
>>
>> mpicc bcast_loop.c
>>
>>
>> To run the code I use:
>>
>> mpirun -np 2 ./a.out
>>
>> Output is unpredictable. It gets stuck at different places.
>>
>> Im attaching lstopo and ompi_info outputs. Do you need any other info?
>>
>>
>> lstopo-no-graphics output:
>>
>> Machine (3433MB)
>>
>>   Socket L#0 + L3 L#0 (8192KB)
>>
>> L2 L#0 (256KB) + L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0
>>
>>   PU L#0 (P#0)
>>
>>   PU L#1 (P#4)
>>
>> L2 L#1 (256KB) + L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1
>>
>>   PU L#2 (P#1)
>>
>>   PU L#3 (P#5)
>>
>> L2 L#2 (256KB) + L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2
>>
>>   PU L#4 (P#2)
>>
>>   PU L#5 (P#6)
>>
>> L2 L#3 (256KB) + L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3
>>
>>   PU L#6 (P#3)
>>
>>   PU L#7 (P#7)
>>
>>   HostBridge L#0
>>
>> PCI 8086:0162
>>
>>   GPU L#0 "card0"
>>
>>   GPU L#1 "renderD128"
>>
>>   GPU L#2 "controlD64"
>>
>> PCI 8086:1502
>>
>>   Net L#3 "eth0"
>>
>> PCI 8086:1e02
>>
>>   Block L#4 "sda"
>>
>>   Block L#5 "sr0"
>>
>>
>> ompi_info output:
>>
>>
>>  Package: Open MPI builduser@anatol Distribution
>>
>> Open MPI: 1.8.4
>>
>>   Open MPI repo revision: v1.8.3-330-g0344f04
>>
>>Open MPI release date: Dec 19, 2014
>>
>> Open RTE: 1.8.4
>>
>>   Open RTE repo revision: v1.8.3-330-g0344f04
>>
>>Open RTE release date: Dec 19, 2014
>>
>> OPAL: 1.8.4
>>
>>   OPAL repo revision: v1.8.3-330-g0344f04
>>
>>OPAL release date: Dec 19, 2014
>>
>>  MPI API: 3.0
>>
>> Ident string: 1.8.4
>>
>>   Prefix: /usr
>>
>>  Configured architecture: i686-pc-linux-gnu
>>
>>   Configure host: anatol
>>
>>Configured by: builduser
>>
>>Configured on: Sat Dec 20 17:00:34 PST 2014
>>
>>   Configure host: anatol
>>
>> Built by: builduser
>>
>> Built on: Sat Dec 20 17:12:16 PST 2014
>>
>>   Built host: anatol
>>
>>   C bindings: yes
>>
>> C++ bindings: yes
>>
>>  Fort mpif.h: yes (all)
>>
>> Fort use mpi: yes (full: ignore TKR)
>>
>>Fort use mpi size: deprecated-ompi-info-value
>>
>> Fort use mpi_f08: yes
>>
>>  Fort mpi_f08 compliance: The mpi_f08 module is available, but due to
>>
>>   limitations in the /usr/bin/gfortran compiler,
>> does
>>
>>   not support the following: array subsections,
>>
>>   direct passthru (where possible) to underlying
>> Open
>>
>>   MPI's C functionality
>>
>>   Fort mpi_f08 subarrays: no
>>
>>Java bindings: no
>>
>>   Wrapper compiler rpath: runpath
>>
>>   C compiler: gcc
>>
>>  C compiler absolute: /usr/bin/gcc
>>
>>   C compiler family name: GNU
>>
>>   C compiler version: 4.9.2
>>
>> C++ compiler: g++
>>
>>C++ compiler absolute: /usr/bin/g++
>>
>>Fort compiler: /usr/bin/gfortran
>>
>>Fort compiler abs:
>>
>>  Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::)
>>
>>Fort 08 assumed shape: yes
>>
>>   Fort optional args: yes
>>
>>   Fort INTERFACE: yes
>>
>> Fort ISO_FORTRAN_ENV: yes
>>
>>Fort STORAGE_SIZE: yes
>>
>>   Fort BIND(C) (all): yes
>>
>>   Fort ISO_C_BINDING: yes
>>
>>  Fort SUBROUTINE BIND(C): yes
>>
>>Fort TYPE,BIND(C): yes
>>
>>  Fort T,BIND(C,name="a"): yes
>>
>> Fort PRIVATE: yes
>>
>>   Fort PROTECTED: yes
>>
>>Fort ABSTRACT: yes
>>
>>Fort ASYNCHRONOUS: yes
>>
>>   Fort PROCEDURE: yes
>>
>>Fort C_FUNLOC: yes
>>
>>  Fort f08 using wrappers: yes
>>
>>  Fort MPI_SIZEOF: yes
>>
>>  C profiling: yes
>>
>>C++ profiling: yes
>>
>>Fort mpif.h profiling: yes
>>
>>   Fort use mpi profiling: yes
>>
>>Fort use mpi_f08 prof: yes
>>
>>   C++ exceptions: no
>>
>>   Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support:
>> yes,
>>
>>   OMPI progress: no, ORTE progress: yes, Event
>> lib:
>>
>>   yes)
>>
>>Sparse Groups: no
>>
>>   Internal debug support: no
>>
>>   MPI interface warnings: yes
>>
>>  MPI parameter check: runtime
>>
>> Memory profiling support: no
>