Sachin, Are you running this on a single host or across multiple hosts (i.e. are you communicating between processes via networking.) If it's on a single host, then it might be an issue with shared memory.
Josh On Fri, Feb 20, 2015 at 1:51 AM, Sachin Krishnan <sachk...@gmail.com> wrote: > Hello Josh, > > The command i use to compile the code is: > > mpicc bcast_loop.c > > > To run the code I use: > > mpirun -np 2 ./a.out > > Output is unpredictable. It gets stuck at different places. > > Im attaching lstopo and ompi_info outputs. Do you need any other info? > > > lstopo-no-graphics output: > > Machine (3433MB) > > Socket L#0 + L3 L#0 (8192KB) > > L2 L#0 (256KB) + L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0 > > PU L#0 (P#0) > > PU L#1 (P#4) > > L2 L#1 (256KB) + L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1 > > PU L#2 (P#1) > > PU L#3 (P#5) > > L2 L#2 (256KB) + L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2 > > PU L#4 (P#2) > > PU L#5 (P#6) > > L2 L#3 (256KB) + L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3 > > PU L#6 (P#3) > > PU L#7 (P#7) > > HostBridge L#0 > > PCI 8086:0162 > > GPU L#0 "card0" > > GPU L#1 "renderD128" > > GPU L#2 "controlD64" > > PCI 8086:1502 > > Net L#3 "eth0" > > PCI 8086:1e02 > > Block L#4 "sda" > > Block L#5 "sr0" > > > ompi_info output: > > > Package: Open MPI builduser@anatol Distribution > > Open MPI: 1.8.4 > > Open MPI repo revision: v1.8.3-330-g0344f04 > > Open MPI release date: Dec 19, 2014 > > Open RTE: 1.8.4 > > Open RTE repo revision: v1.8.3-330-g0344f04 > > Open RTE release date: Dec 19, 2014 > > OPAL: 1.8.4 > > OPAL repo revision: v1.8.3-330-g0344f04 > > OPAL release date: Dec 19, 2014 > > MPI API: 3.0 > > Ident string: 1.8.4 > > Prefix: /usr > > Configured architecture: i686-pc-linux-gnu > > Configure host: anatol > > Configured by: builduser > > Configured on: Sat Dec 20 17:00:34 PST 2014 > > Configure host: anatol > > Built by: builduser > > Built on: Sat Dec 20 17:12:16 PST 2014 > > Built host: anatol > > C bindings: yes > > C++ bindings: yes > > Fort mpif.h: yes (all) > > Fort use mpi: yes (full: ignore TKR) > > Fort use mpi size: deprecated-ompi-info-value > > Fort use mpi_f08: yes > > Fort mpi_f08 compliance: The mpi_f08 module is available, but due to > > limitations in the /usr/bin/gfortran compiler, > does > > not support the following: array subsections, > > direct passthru (where possible) to underlying > Open > > MPI's C functionality > > Fort mpi_f08 subarrays: no > > Java bindings: no > > Wrapper compiler rpath: runpath > > C compiler: gcc > > C compiler absolute: /usr/bin/gcc > > C compiler family name: GNU > > C compiler version: 4.9.2 > > C++ compiler: g++ > > C++ compiler absolute: /usr/bin/g++ > > Fort compiler: /usr/bin/gfortran > > Fort compiler abs: > > Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::) > > Fort 08 assumed shape: yes > > Fort optional args: yes > > Fort INTERFACE: yes > > Fort ISO_FORTRAN_ENV: yes > > Fort STORAGE_SIZE: yes > > Fort BIND(C) (all): yes > > Fort ISO_C_BINDING: yes > > Fort SUBROUTINE BIND(C): yes > > Fort TYPE,BIND(C): yes > > Fort T,BIND(C,name="a"): yes > > Fort PRIVATE: yes > > Fort PROTECTED: yes > > Fort ABSTRACT: yes > > Fort ASYNCHRONOUS: yes > > Fort PROCEDURE: yes > > Fort C_FUNLOC: yes > > Fort f08 using wrappers: yes > > Fort MPI_SIZEOF: yes > > C profiling: yes > > C++ profiling: yes > > Fort mpif.h profiling: yes > > Fort use mpi profiling: yes > > Fort use mpi_f08 prof: yes > > C++ exceptions: no > > Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: > yes, > > OMPI progress: no, ORTE progress: yes, Event lib: > > yes) > > Sparse Groups: no > > Internal debug support: no > > MPI interface warnings: yes > > MPI parameter check: runtime > > Memory profiling support: no > > Memory debugging support: no > > libltdl support: yes > > Heterogeneous support: no > > mpirun default --prefix: no > > MPI I/O support: yes > > MPI_WTIME support: gettimeofday > > Symbol vis. support: yes > > Host topology support: yes > > MPI extensions: > > FT Checkpoint support: no (checkpoint thread: no) > > C/R Enabled Debugging: no > > VampirTrace support: yes > > MPI_MAX_PROCESSOR_NAME: 256 > > MPI_MAX_ERROR_STRING: 256 > > MPI_MAX_OBJECT_NAME: 64 > > MPI_MAX_INFO_KEY: 36 > > MPI_MAX_INFO_VAL: 256 > > MPI_MAX_PORT_NAME: 1024 > > MPI_MAX_DATAREP_STRING: 128 > > MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.8.4) > > MCA compress: bzip (MCA v2.0, API v2.0, Component v1.8.4) > > MCA compress: gzip (MCA v2.0, API v2.0, Component v1.8.4) > > MCA crs: none (MCA v2.0, API v2.0, Component v1.8.4) > > MCA db: hash (MCA v2.0, API v1.0, Component v1.8.4) > > MCA db: print (MCA v2.0, API v1.0, Component v1.8.4) > > MCA event: libevent2021 (MCA v2.0, API v2.0, Component > v1.8.4) > > MCA hwloc: external (MCA v2.0, API v2.0, Component v1.8.4) > > MCA if: posix_ipv4 (MCA v2.0, API v2.0, Component v1.8.4) > > MCA if: linux_ipv6 (MCA v2.0, API v2.0, Component v1.8.4) > > MCA installdirs: env (MCA v2.0, API v2.0, Component v1.8.4) > > MCA installdirs: config (MCA v2.0, API v2.0, Component v1.8.4) > > MCA memchecker: valgrind (MCA v2.0, API v2.0, Component v1.8.4) > > MCA memory: linux (MCA v2.0, API v2.0, Component v1.8.4) > > MCA pstat: linux (MCA v2.0, API v2.0, Component v1.8.4) > > MCA sec: basic (MCA v2.0, API v1.0, Component v1.8.4) > > MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.8.4) > > MCA shmem: posix (MCA v2.0, API v2.0, Component v1.8.4) > > MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.8.4) > > MCA timer: linux (MCA v2.0, API v2.0, Component v1.8.4) > > MCA dfs: app (MCA v2.0, API v1.0, Component v1.8.4) > > MCA dfs: orted (MCA v2.0, API v1.0, Component v1.8.4) > > MCA dfs: test (MCA v2.0, API v1.0, Component v1.8.4) > > MCA errmgr: default_app (MCA v2.0, API v3.0, Component > v1.8.4) > > MCA errmgr: default_hnp (MCA v2.0, API v3.0, Component > v1.8.4) > > MCA errmgr: default_orted (MCA v2.0, API v3.0, Component > > v1.8.4) > > MCA errmgr: default_tool (MCA v2.0, API v3.0, Component > v1.8.4) > > MCA ess: env (MCA v2.0, API v3.0, Component v1.8.4) > > MCA ess: hnp (MCA v2.0, API v3.0, Component v1.8.4) > > MCA ess: singleton (MCA v2.0, API v3.0, Component v1.8.4) > > MCA ess: tool (MCA v2.0, API v3.0, Component v1.8.4) > > MCA filem: raw (MCA v2.0, API v2.0, Component v1.8.4) > > MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.8.4) > > MCA iof: hnp (MCA v2.0, API v2.0, Component v1.8.4) > > MCA iof: mr_hnp (MCA v2.0, API v2.0, Component v1.8.4) > > MCA iof: mr_orted (MCA v2.0, API v2.0, Component v1.8.4) > > MCA iof: orted (MCA v2.0, API v2.0, Component v1.8.4) > > MCA iof: tool (MCA v2.0, API v2.0, Component v1.8.4) > > MCA odls: default (MCA v2.0, API v2.0, Component v1.8.4) > > MCA oob: tcp (MCA v2.0, API v2.0, Component v1.8.4) > > MCA plm: isolated (MCA v2.0, API v2.0, Component v1.8.4) > > MCA plm: rsh (MCA v2.0, API v2.0, Component v1.8.4) > > MCA ras: loadleveler (MCA v2.0, API v2.0, Component > v1.8.4) > > MCA ras: simulator (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rmaps: lama (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rmaps: mindist (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rmaps: ppr (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rmaps: round_robin (MCA v2.0, API v2.0, Component > v1.8.4) > > MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rmaps: staged (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rml: oob (MCA v2.0, API v2.0, Component v1.8.4) > > MCA routed: binomial (MCA v2.0, API v2.0, Component v1.8.4) > > MCA routed: debruijn (MCA v2.0, API v2.0, Component v1.8.4) > > MCA routed: direct (MCA v2.0, API v2.0, Component v1.8.4) > > MCA routed: radix (MCA v2.0, API v2.0, Component v1.8.4) > > MCA state: app (MCA v2.0, API v1.0, Component v1.8.4) > > MCA state: hnp (MCA v2.0, API v1.0, Component v1.8.4) > > MCA state: novm (MCA v2.0, API v1.0, Component v1.8.4) > > MCA state: orted (MCA v2.0, API v1.0, Component v1.8.4) > > MCA state: staged_hnp (MCA v2.0, API v1.0, Component v1.8.4) > > MCA state: staged_orted (MCA v2.0, API v1.0, Component > v1.8.4) > > MCA state: tool (MCA v2.0, API v1.0, Component v1.8.4) > > MCA allocator: basic (MCA v2.0, API v2.0, Component v1.8.4) > > MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.8.4) > > MCA bcol: basesmuma (MCA v2.0, API v2.0, Component v1.8.4) > > MCA bcol: ptpcoll (MCA v2.0, API v2.0, Component v1.8.4) > > MCA bml: r2 (MCA v2.0, API v2.0, Component v1.8.4) > > MCA btl: self (MCA v2.0, API v2.0, Component v1.8.4) > > MCA btl: sm (MCA v2.0, API v2.0, Component v1.8.4) > > MCA btl: tcp (MCA v2.0, API v2.0, Component v1.8.4) > > MCA btl: vader (MCA v2.0, API v2.0, Component v1.8.4) > > MCA coll: basic (MCA v2.0, API v2.0, Component v1.8.4) > > MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.8.4) > > MCA coll: inter (MCA v2.0, API v2.0, Component v1.8.4) > > MCA coll: libnbc (MCA v2.0, API v2.0, Component v1.8.4) > > MCA coll: ml (MCA v2.0, API v2.0, Component v1.8.4) > > MCA coll: self (MCA v2.0, API v2.0, Component v1.8.4) > > MCA coll: sm (MCA v2.0, API v2.0, Component v1.8.4) > > MCA coll: tuned (MCA v2.0, API v2.0, Component v1.8.4) > > MCA dpm: orte (MCA v2.0, API v2.0, Component v1.8.4) > > MCA fbtl: posix (MCA v2.0, API v2.0, Component v1.8.4) > > MCA fcoll: dynamic (MCA v2.0, API v2.0, Component v1.8.4) > > MCA fcoll: individual (MCA v2.0, API v2.0, Component v1.8.4) > > MCA fcoll: static (MCA v2.0, API v2.0, Component v1.8.4) > > MCA fcoll: two_phase (MCA v2.0, API v2.0, Component v1.8.4) > > MCA fcoll: ylib (MCA v2.0, API v2.0, Component v1.8.4) > > MCA fs: ufs (MCA v2.0, API v2.0, Component v1.8.4) > > MCA io: ompio (MCA v2.0, API v2.0, Component v1.8.4) > > MCA io: romio (MCA v2.0, API v2.0, Component v1.8.4) > > MCA mpool: grdma (MCA v2.0, API v2.0, Component v1.8.4) > > MCA mpool: sm (MCA v2.0, API v2.0, Component v1.8.4) > > MCA osc: rdma (MCA v2.0, API v3.0, Component v1.8.4) > > MCA osc: sm (MCA v2.0, API v3.0, Component v1.8.4) > > MCA pml: v (MCA v2.0, API v2.0, Component v1.8.4) > > MCA pml: bfo (MCA v2.0, API v2.0, Component v1.8.4) > > MCA pml: cm (MCA v2.0, API v2.0, Component v1.8.4) > > MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.8.4) > > MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rcache: vma (MCA v2.0, API v2.0, Component v1.8.4) > > MCA rte: orte (MCA v2.0, API v2.0, Component v1.8.4) > > MCA sbgp: basesmsocket (MCA v2.0, API v2.0, Component > v1.8.4) > > MCA sbgp: basesmuma (MCA v2.0, API v2.0, Component v1.8.4) > > MCA sbgp: p2p (MCA v2.0, API v2.0, Component v1.8.4) > > MCA sharedfp: individual (MCA v2.0, API v2.0, Component v1.8.4) > > MCA sharedfp: lockedfile (MCA v2.0, API v2.0, Component v1.8.4) > > MCA sharedfp: sm (MCA v2.0, API v2.0, Component v1.8.4) > > MCA topo: basic (MCA v2.0, API v2.1, Component v1.8.4) > > MCA vprotocol: pessimist (MCA v2.0, API v2.0, Component v1.8.4) > > Sachin > > >Sachin, > > >Can you, please, provide a command line? Additional information about > your > >system could be helpful also. > > >Josh > > >>On Wed, Feb 18, 2015 at 3:43 AM, Sachin Krishnan <sachkris_at_[hidden]> > wrote: > > >> Hello, > >> > >> I am new to MPI and also this list. > >> I wrote an MPI code with several MPI_Bcast calls in a loop. My code was > >> getting stuck at random points, ie it was not systematic. After a few > hours > >> of debugging and googling, I found that the issue may be with the > several > >> MPI_Bcast calls in a loop. > >> > >> I stumbled on this test code which can reproduce the issue: > >> https://github.com/fintler/ompi/blob/master/orte/test/mpi/bcast_loop.c > >> > >> Im using OpenMPI v1.8.4 installed from official Arch Linux repo. > >> > >> Is it a known issue with OpenMPI? > >> Is it some problem with the way openmpi is configured in my system? > >> > >> Thanks in advance. > >> > >> Sachin > >> > >> > >> > >> _______________________________________________ > >> users mailing list > >> users_at_[hidden] > >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > >> Link to this post: > >> http://www.open-mpi.org/community/lists/users/2015/02/26338.php > >> > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/02/26363.php >