Re: [OMPI users] Several Bcast calls in a loop causing the code to hang
Hello Josh, The command i use to compile the code is: mpicc bcast_loop.c To run the code I use: mpirun -np 2 ./a.out Output is unpredictable. It gets stuck at different places. Im attaching lstopo and ompi_info outputs. Do you need any other info? lstopo-no-graphics output: Machine (3433MB) Socket L#0 + L3 L#0 (8192KB) L2 L#0 (256KB) + L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0 PU L#0 (P#0) PU L#1 (P#4) L2 L#1 (256KB) + L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1 PU L#2 (P#1) PU L#3 (P#5) L2 L#2 (256KB) + L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2 PU L#4 (P#2) PU L#5 (P#6) L2 L#3 (256KB) + L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3 PU L#6 (P#3) PU L#7 (P#7) HostBridge L#0 PCI 8086:0162 GPU L#0 "card0" GPU L#1 "renderD128" GPU L#2 "controlD64" PCI 8086:1502 Net L#3 "eth0" PCI 8086:1e02 Block L#4 "sda" Block L#5 "sr0" ompi_info output: Package: Open MPI builduser@anatol Distribution Open MPI: 1.8.4 Open MPI repo revision: v1.8.3-330-g0344f04 Open MPI release date: Dec 19, 2014 Open RTE: 1.8.4 Open RTE repo revision: v1.8.3-330-g0344f04 Open RTE release date: Dec 19, 2014 OPAL: 1.8.4 OPAL repo revision: v1.8.3-330-g0344f04 OPAL release date: Dec 19, 2014 MPI API: 3.0 Ident string: 1.8.4 Prefix: /usr Configured architecture: i686-pc-linux-gnu Configure host: anatol Configured by: builduser Configured on: Sat Dec 20 17:00:34 PST 2014 Configure host: anatol Built by: builduser Built on: Sat Dec 20 17:12:16 PST 2014 Built host: anatol C bindings: yes C++ bindings: yes Fort mpif.h: yes (all) Fort use mpi: yes (full: ignore TKR) Fort use mpi size: deprecated-ompi-info-value Fort use mpi_f08: yes Fort mpi_f08 compliance: The mpi_f08 module is available, but due to limitations in the /usr/bin/gfortran compiler, does not support the following: array subsections, direct passthru (where possible) to underlying Open MPI's C functionality Fort mpi_f08 subarrays: no Java bindings: no Wrapper compiler rpath: runpath C compiler: gcc C compiler absolute: /usr/bin/gcc C compiler family name: GNU C compiler version: 4.9.2 C++ compiler: g++ C++ compiler absolute: /usr/bin/g++ Fort compiler: /usr/bin/gfortran Fort compiler abs: Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::) Fort 08 assumed shape: yes Fort optional args: yes Fort INTERFACE: yes Fort ISO_FORTRAN_ENV: yes Fort STORAGE_SIZE: yes Fort BIND(C) (all): yes Fort ISO_C_BINDING: yes Fort SUBROUTINE BIND(C): yes Fort TYPE,BIND(C): yes Fort T,BIND(C,name="a"): yes Fort PRIVATE: yes Fort PROTECTED: yes Fort ABSTRACT: yes Fort ASYNCHRONOUS: yes Fort PROCEDURE: yes Fort C_FUNLOC: yes Fort f08 using wrappers: yes Fort MPI_SIZEOF: yes C profiling: yes C++ profiling: yes Fort mpif.h profiling: yes Fort use mpi profiling: yes Fort use mpi_f08 prof: yes C++ exceptions: no Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: yes, OMPI progress: no, ORTE progress: yes, Event lib: yes) Sparse Groups: no Internal debug support: no MPI interface warnings: yes MPI parameter check: runtime Memory profiling support: no Memory debugging support: no libltdl support: yes Heterogeneous support: no mpirun default --prefix: no MPI I/O support: yes MPI_WTIME support: gettimeofday Symbol vis. support: yes Host topology support: yes MPI extensions: FT Checkpoint support: no (checkpoint thread: no) C/R Enabled Debugging: no VampirTrace support: yes MPI_MAX_PROCESSOR_NAME: 256 MPI_MAX_ERROR_STRING: 256 MPI_MAX_OBJECT_NAME: 64 MPI_MAX_INFO_KEY: 36 MPI_MAX_INFO_VAL: 256 MPI_MAX_PORT_NAME: 1024 MPI_MAX_DATAREP_STRING: 128 MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.8.4) MCA compress: bzip (MCA v2.0, API v2.0, Component v1.8.4) MCA compress: gzip (MCA v2.0, API v2.0, Component v1.8.4) MCA crs: none (MCA v2.0, API v2.0, Component v1.8.4) MCA db: hash (MCA v2.0, API v1.0, Component v1.8.4) MCA db: print (MCA v2.0, API v1.0, Component v1
[OMPI users] Fix small typo in man
Hello, may I suggest you to apply this small patch to fix a typo in the MPI_Intercomm_merge entry manual? Best regards WARNING / LEGAL TEXT: This message is intended only for the use of the individual or entity to which it is addressed and may contain information which is privileged, confidential, proprietary, or exempt from disclosure under applicable law. If you are not the intended recipient or the person responsible for delivering the message to the intended recipient, you are strictly prohibited from disclosing, distributing, copying, or in any way using this message. If you have received this communication in error, please notify the sender and destroy and delete any copies you may have received. http://www.bsc.es/disclaimer--- MPI_Intercomm_merge.3in 2014-11-01 01:29:50.0 +0100 +++ MPI_Intercomm_merge.3in.new 2015-02-20 15:08:55.072114384 +0100 @@ -13,7 +13,7 @@ .nf #include int MPI_Intercomm_merge(MPI_Comm \fIintercomm\fP, int\fI high\fP, - MPI_Comm\fI *newintracomm\fP + MPI_Comm\fI *newintracomm\fP) .fi .SH Fortran Syntax
Re: [OMPI users] Fix small typo in man
Harald -- Many thanks for noticing and sending a patch. I have committed the fix, and slated it for the next Open MPI release: https://github.com/open-mpi/ompi-release/pull/188 > On Feb 20, 2015, at 9:36 AM, Harald Servat wrote: > > Hello, > > may I suggest you to apply this small patch to fix a typo in the > MPI_Intercomm_merge entry manual? > > Best regards > > > WARNING / LEGAL TEXT: This message is intended only for the use of the > individual or entity to which it is addressed and may contain > information which is privileged, confidential, proprietary, or exempt > from disclosure under applicable law. If you are not the intended > recipient or the person responsible for delivering the message to the > intended recipient, you are strictly prohibited from disclosing, > distributing, copying, or in any way using this message. If you have > received this communication in error, please notify the sender and > destroy and delete any copies you may have received. > > http://www.bsc.es/disclaimer___ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/02/26364.php -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
Re: [OMPI users] Several Bcast calls in a loop causing the code to hang
Sachin, Are you running this on a single host or across multiple hosts (i.e. are you communicating between processes via networking.) If it's on a single host, then it might be an issue with shared memory. Josh On Fri, Feb 20, 2015 at 1:51 AM, Sachin Krishnan wrote: > Hello Josh, > > The command i use to compile the code is: > > mpicc bcast_loop.c > > > To run the code I use: > > mpirun -np 2 ./a.out > > Output is unpredictable. It gets stuck at different places. > > Im attaching lstopo and ompi_info outputs. Do you need any other info? > > > lstopo-no-graphics output: > > Machine (3433MB) > > Socket L#0 + L3 L#0 (8192KB) > > L2 L#0 (256KB) + L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0 > > PU L#0 (P#0) > > PU L#1 (P#4) > > L2 L#1 (256KB) + L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1 > > PU L#2 (P#1) > > PU L#3 (P#5) > > L2 L#2 (256KB) + L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2 > > PU L#4 (P#2) > > PU L#5 (P#6) > > L2 L#3 (256KB) + L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3 > > PU L#6 (P#3) > > PU L#7 (P#7) > > HostBridge L#0 > > PCI 8086:0162 > > GPU L#0 "card0" > > GPU L#1 "renderD128" > > GPU L#2 "controlD64" > > PCI 8086:1502 > > Net L#3 "eth0" > > PCI 8086:1e02 > > Block L#4 "sda" > > Block L#5 "sr0" > > > ompi_info output: > > > Package: Open MPI builduser@anatol Distribution > > Open MPI: 1.8.4 > > Open MPI repo revision: v1.8.3-330-g0344f04 > >Open MPI release date: Dec 19, 2014 > > Open RTE: 1.8.4 > > Open RTE repo revision: v1.8.3-330-g0344f04 > >Open RTE release date: Dec 19, 2014 > > OPAL: 1.8.4 > > OPAL repo revision: v1.8.3-330-g0344f04 > >OPAL release date: Dec 19, 2014 > > MPI API: 3.0 > > Ident string: 1.8.4 > > Prefix: /usr > > Configured architecture: i686-pc-linux-gnu > > Configure host: anatol > >Configured by: builduser > >Configured on: Sat Dec 20 17:00:34 PST 2014 > > Configure host: anatol > > Built by: builduser > > Built on: Sat Dec 20 17:12:16 PST 2014 > > Built host: anatol > > C bindings: yes > > C++ bindings: yes > > Fort mpif.h: yes (all) > > Fort use mpi: yes (full: ignore TKR) > >Fort use mpi size: deprecated-ompi-info-value > > Fort use mpi_f08: yes > > Fort mpi_f08 compliance: The mpi_f08 module is available, but due to > > limitations in the /usr/bin/gfortran compiler, > does > > not support the following: array subsections, > > direct passthru (where possible) to underlying > Open > > MPI's C functionality > > Fort mpi_f08 subarrays: no > >Java bindings: no > > Wrapper compiler rpath: runpath > > C compiler: gcc > > C compiler absolute: /usr/bin/gcc > > C compiler family name: GNU > > C compiler version: 4.9.2 > > C++ compiler: g++ > >C++ compiler absolute: /usr/bin/g++ > >Fort compiler: /usr/bin/gfortran > >Fort compiler abs: > > Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::) > >Fort 08 assumed shape: yes > > Fort optional args: yes > > Fort INTERFACE: yes > > Fort ISO_FORTRAN_ENV: yes > >Fort STORAGE_SIZE: yes > > Fort BIND(C) (all): yes > > Fort ISO_C_BINDING: yes > > Fort SUBROUTINE BIND(C): yes > >Fort TYPE,BIND(C): yes > > Fort T,BIND(C,name="a"): yes > > Fort PRIVATE: yes > > Fort PROTECTED: yes > >Fort ABSTRACT: yes > >Fort ASYNCHRONOUS: yes > > Fort PROCEDURE: yes > >Fort C_FUNLOC: yes > > Fort f08 using wrappers: yes > > Fort MPI_SIZEOF: yes > > C profiling: yes > >C++ profiling: yes > >Fort mpif.h profiling: yes > > Fort use mpi profiling: yes > >Fort use mpi_f08 prof: yes > > C++ exceptions: no > > Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: > yes, > > OMPI progress: no, ORTE progress: yes, Event lib: > > yes) > >Sparse Groups: no > > Internal debug support: no > > MPI interface warnings: yes > > MPI parameter check: runtime > > Memory profiling support: no > > Memory debugging support: no > > libltdl support: yes > >Heterogeneous support: no > > mpirun default --prefix: no > > MPI I/O support: yes > >MPI_WTIME support: gettimeofday > > Symbol vis. support: yes > >Host topology support: yes > > MPI extensions: > >FT Checkpoint support: no (checkpoint thread: no) > >C/R Enabled Debugging: no > > VampirTrace support: y
Re: [OMPI users] Several Bcast calls in a loop causing the code to hang
Josh, Thanks for the help. I'm running on a single host. How do I confirm that it is an issue with the shared memory? Sachin On Fri, Feb 20, 2015 at 11:58 PM, Joshua Ladd wrote: > Sachin, > > Are you running this on a single host or across multiple hosts (i.e. are > you communicating between processes via networking.) If it's on a single > host, then it might be an issue with shared memory. > > Josh > > On Fri, Feb 20, 2015 at 1:51 AM, Sachin Krishnan > wrote: > >> Hello Josh, >> >> The command i use to compile the code is: >> >> mpicc bcast_loop.c >> >> >> To run the code I use: >> >> mpirun -np 2 ./a.out >> >> Output is unpredictable. It gets stuck at different places. >> >> Im attaching lstopo and ompi_info outputs. Do you need any other info? >> >> >> lstopo-no-graphics output: >> >> Machine (3433MB) >> >> Socket L#0 + L3 L#0 (8192KB) >> >> L2 L#0 (256KB) + L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0 >> >> PU L#0 (P#0) >> >> PU L#1 (P#4) >> >> L2 L#1 (256KB) + L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1 >> >> PU L#2 (P#1) >> >> PU L#3 (P#5) >> >> L2 L#2 (256KB) + L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2 >> >> PU L#4 (P#2) >> >> PU L#5 (P#6) >> >> L2 L#3 (256KB) + L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3 >> >> PU L#6 (P#3) >> >> PU L#7 (P#7) >> >> HostBridge L#0 >> >> PCI 8086:0162 >> >> GPU L#0 "card0" >> >> GPU L#1 "renderD128" >> >> GPU L#2 "controlD64" >> >> PCI 8086:1502 >> >> Net L#3 "eth0" >> >> PCI 8086:1e02 >> >> Block L#4 "sda" >> >> Block L#5 "sr0" >> >> >> ompi_info output: >> >> >> Package: Open MPI builduser@anatol Distribution >> >> Open MPI: 1.8.4 >> >> Open MPI repo revision: v1.8.3-330-g0344f04 >> >>Open MPI release date: Dec 19, 2014 >> >> Open RTE: 1.8.4 >> >> Open RTE repo revision: v1.8.3-330-g0344f04 >> >>Open RTE release date: Dec 19, 2014 >> >> OPAL: 1.8.4 >> >> OPAL repo revision: v1.8.3-330-g0344f04 >> >>OPAL release date: Dec 19, 2014 >> >> MPI API: 3.0 >> >> Ident string: 1.8.4 >> >> Prefix: /usr >> >> Configured architecture: i686-pc-linux-gnu >> >> Configure host: anatol >> >>Configured by: builduser >> >>Configured on: Sat Dec 20 17:00:34 PST 2014 >> >> Configure host: anatol >> >> Built by: builduser >> >> Built on: Sat Dec 20 17:12:16 PST 2014 >> >> Built host: anatol >> >> C bindings: yes >> >> C++ bindings: yes >> >> Fort mpif.h: yes (all) >> >> Fort use mpi: yes (full: ignore TKR) >> >>Fort use mpi size: deprecated-ompi-info-value >> >> Fort use mpi_f08: yes >> >> Fort mpi_f08 compliance: The mpi_f08 module is available, but due to >> >> limitations in the /usr/bin/gfortran compiler, >> does >> >> not support the following: array subsections, >> >> direct passthru (where possible) to underlying >> Open >> >> MPI's C functionality >> >> Fort mpi_f08 subarrays: no >> >>Java bindings: no >> >> Wrapper compiler rpath: runpath >> >> C compiler: gcc >> >> C compiler absolute: /usr/bin/gcc >> >> C compiler family name: GNU >> >> C compiler version: 4.9.2 >> >> C++ compiler: g++ >> >>C++ compiler absolute: /usr/bin/g++ >> >>Fort compiler: /usr/bin/gfortran >> >>Fort compiler abs: >> >> Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::) >> >>Fort 08 assumed shape: yes >> >> Fort optional args: yes >> >> Fort INTERFACE: yes >> >> Fort ISO_FORTRAN_ENV: yes >> >>Fort STORAGE_SIZE: yes >> >> Fort BIND(C) (all): yes >> >> Fort ISO_C_BINDING: yes >> >> Fort SUBROUTINE BIND(C): yes >> >>Fort TYPE,BIND(C): yes >> >> Fort T,BIND(C,name="a"): yes >> >> Fort PRIVATE: yes >> >> Fort PROTECTED: yes >> >>Fort ABSTRACT: yes >> >>Fort ASYNCHRONOUS: yes >> >> Fort PROCEDURE: yes >> >>Fort C_FUNLOC: yes >> >> Fort f08 using wrappers: yes >> >> Fort MPI_SIZEOF: yes >> >> C profiling: yes >> >>C++ profiling: yes >> >>Fort mpif.h profiling: yes >> >> Fort use mpi profiling: yes >> >>Fort use mpi_f08 prof: yes >> >> C++ exceptions: no >> >> Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: >> yes, >> >> OMPI progress: no, ORTE progress: yes, Event >> lib: >> >> yes) >> >>Sparse Groups: no >> >> Internal debug support: no >> >> MPI interface warnings: yes >> >> MPI parameter check: runtime >> >> Memory profiling support: no >