Hello Josh, The command i use to compile the code is:
mpicc bcast_loop.c To run the code I use: mpirun -np 2 ./a.out Output is unpredictable. It gets stuck at different places. Im attaching lstopo and ompi_info outputs. Do you need any other info? lstopo-no-graphics output: Machine (3433MB) Socket L#0 + L3 L#0 (8192KB) L2 L#0 (256KB) + L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0 PU L#0 (P#0) PU L#1 (P#4) L2 L#1 (256KB) + L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1 PU L#2 (P#1) PU L#3 (P#5) L2 L#2 (256KB) + L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2 PU L#4 (P#2) PU L#5 (P#6) L2 L#3 (256KB) + L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3 PU L#6 (P#3) PU L#7 (P#7) HostBridge L#0 PCI 8086:0162 GPU L#0 "card0" GPU L#1 "renderD128" GPU L#2 "controlD64" PCI 8086:1502 Net L#3 "eth0" PCI 8086:1e02 Block L#4 "sda" Block L#5 "sr0" ompi_info output: Package: Open MPI builduser@anatol Distribution Open MPI: 1.8.4 Open MPI repo revision: v1.8.3-330-g0344f04 Open MPI release date: Dec 19, 2014 Open RTE: 1.8.4 Open RTE repo revision: v1.8.3-330-g0344f04 Open RTE release date: Dec 19, 2014 OPAL: 1.8.4 OPAL repo revision: v1.8.3-330-g0344f04 OPAL release date: Dec 19, 2014 MPI API: 3.0 Ident string: 1.8.4 Prefix: /usr Configured architecture: i686-pc-linux-gnu Configure host: anatol Configured by: builduser Configured on: Sat Dec 20 17:00:34 PST 2014 Configure host: anatol Built by: builduser Built on: Sat Dec 20 17:12:16 PST 2014 Built host: anatol C bindings: yes C++ bindings: yes Fort mpif.h: yes (all) Fort use mpi: yes (full: ignore TKR) Fort use mpi size: deprecated-ompi-info-value Fort use mpi_f08: yes Fort mpi_f08 compliance: The mpi_f08 module is available, but due to limitations in the /usr/bin/gfortran compiler, does not support the following: array subsections, direct passthru (where possible) to underlying Open MPI's C functionality Fort mpi_f08 subarrays: no Java bindings: no Wrapper compiler rpath: runpath C compiler: gcc C compiler absolute: /usr/bin/gcc C compiler family name: GNU C compiler version: 4.9.2 C++ compiler: g++ C++ compiler absolute: /usr/bin/g++ Fort compiler: /usr/bin/gfortran Fort compiler abs: Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::) Fort 08 assumed shape: yes Fort optional args: yes Fort INTERFACE: yes Fort ISO_FORTRAN_ENV: yes Fort STORAGE_SIZE: yes Fort BIND(C) (all): yes Fort ISO_C_BINDING: yes Fort SUBROUTINE BIND(C): yes Fort TYPE,BIND(C): yes Fort T,BIND(C,name="a"): yes Fort PRIVATE: yes Fort PROTECTED: yes Fort ABSTRACT: yes Fort ASYNCHRONOUS: yes Fort PROCEDURE: yes Fort C_FUNLOC: yes Fort f08 using wrappers: yes Fort MPI_SIZEOF: yes C profiling: yes C++ profiling: yes Fort mpif.h profiling: yes Fort use mpi profiling: yes Fort use mpi_f08 prof: yes C++ exceptions: no Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: yes, OMPI progress: no, ORTE progress: yes, Event lib: yes) Sparse Groups: no Internal debug support: no MPI interface warnings: yes MPI parameter check: runtime Memory profiling support: no Memory debugging support: no libltdl support: yes Heterogeneous support: no mpirun default --prefix: no MPI I/O support: yes MPI_WTIME support: gettimeofday Symbol vis. support: yes Host topology support: yes MPI extensions: FT Checkpoint support: no (checkpoint thread: no) C/R Enabled Debugging: no VampirTrace support: yes MPI_MAX_PROCESSOR_NAME: 256 MPI_MAX_ERROR_STRING: 256 MPI_MAX_OBJECT_NAME: 64 MPI_MAX_INFO_KEY: 36 MPI_MAX_INFO_VAL: 256 MPI_MAX_PORT_NAME: 1024 MPI_MAX_DATAREP_STRING: 128 MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.8.4) MCA compress: bzip (MCA v2.0, API v2.0, Component v1.8.4) MCA compress: gzip (MCA v2.0, API v2.0, Component v1.8.4) MCA crs: none (MCA v2.0, API v2.0, Component v1.8.4) MCA db: hash (MCA v2.0, API v1.0, Component v1.8.4) MCA db: print (MCA v2.0, API v1.0, Component v1.8.4) MCA event: libevent2021 (MCA v2.0, API v2.0, Component v1.8.4) MCA hwloc: external (MCA v2.0, API v2.0, Component v1.8.4) MCA if: posix_ipv4 (MCA v2.0, API v2.0, Component v1.8.4) MCA if: linux_ipv6 (MCA v2.0, API v2.0, Component v1.8.4) MCA installdirs: env (MCA v2.0, API v2.0, Component v1.8.4) MCA installdirs: config (MCA v2.0, API v2.0, Component v1.8.4) MCA memchecker: valgrind (MCA v2.0, API v2.0, Component v1.8.4) MCA memory: linux (MCA v2.0, API v2.0, Component v1.8.4) MCA pstat: linux (MCA v2.0, API v2.0, Component v1.8.4) MCA sec: basic (MCA v2.0, API v1.0, Component v1.8.4) MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.8.4) MCA shmem: posix (MCA v2.0, API v2.0, Component v1.8.4) MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.8.4) MCA timer: linux (MCA v2.0, API v2.0, Component v1.8.4) MCA dfs: app (MCA v2.0, API v1.0, Component v1.8.4) MCA dfs: orted (MCA v2.0, API v1.0, Component v1.8.4) MCA dfs: test (MCA v2.0, API v1.0, Component v1.8.4) MCA errmgr: default_app (MCA v2.0, API v3.0, Component v1.8.4) MCA errmgr: default_hnp (MCA v2.0, API v3.0, Component v1.8.4) MCA errmgr: default_orted (MCA v2.0, API v3.0, Component v1.8.4) MCA errmgr: default_tool (MCA v2.0, API v3.0, Component v1.8.4) MCA ess: env (MCA v2.0, API v3.0, Component v1.8.4) MCA ess: hnp (MCA v2.0, API v3.0, Component v1.8.4) MCA ess: singleton (MCA v2.0, API v3.0, Component v1.8.4) MCA ess: tool (MCA v2.0, API v3.0, Component v1.8.4) MCA filem: raw (MCA v2.0, API v2.0, Component v1.8.4) MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.8.4) MCA iof: hnp (MCA v2.0, API v2.0, Component v1.8.4) MCA iof: mr_hnp (MCA v2.0, API v2.0, Component v1.8.4) MCA iof: mr_orted (MCA v2.0, API v2.0, Component v1.8.4) MCA iof: orted (MCA v2.0, API v2.0, Component v1.8.4) MCA iof: tool (MCA v2.0, API v2.0, Component v1.8.4) MCA odls: default (MCA v2.0, API v2.0, Component v1.8.4) MCA oob: tcp (MCA v2.0, API v2.0, Component v1.8.4) MCA plm: isolated (MCA v2.0, API v2.0, Component v1.8.4) MCA plm: rsh (MCA v2.0, API v2.0, Component v1.8.4) MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.8.4) MCA ras: simulator (MCA v2.0, API v2.0, Component v1.8.4) MCA rmaps: lama (MCA v2.0, API v2.0, Component v1.8.4) MCA rmaps: mindist (MCA v2.0, API v2.0, Component v1.8.4) MCA rmaps: ppr (MCA v2.0, API v2.0, Component v1.8.4) MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.8.4) MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.8.4) MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.8.4) MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.8.4) MCA rmaps: staged (MCA v2.0, API v2.0, Component v1.8.4) MCA rml: oob (MCA v2.0, API v2.0, Component v1.8.4) MCA routed: binomial (MCA v2.0, API v2.0, Component v1.8.4) MCA routed: debruijn (MCA v2.0, API v2.0, Component v1.8.4) MCA routed: direct (MCA v2.0, API v2.0, Component v1.8.4) MCA routed: radix (MCA v2.0, API v2.0, Component v1.8.4) MCA state: app (MCA v2.0, API v1.0, Component v1.8.4) MCA state: hnp (MCA v2.0, API v1.0, Component v1.8.4) MCA state: novm (MCA v2.0, API v1.0, Component v1.8.4) MCA state: orted (MCA v2.0, API v1.0, Component v1.8.4) MCA state: staged_hnp (MCA v2.0, API v1.0, Component v1.8.4) MCA state: staged_orted (MCA v2.0, API v1.0, Component v1.8.4) MCA state: tool (MCA v2.0, API v1.0, Component v1.8.4) MCA allocator: basic (MCA v2.0, API v2.0, Component v1.8.4) MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.8.4) MCA bcol: basesmuma (MCA v2.0, API v2.0, Component v1.8.4) MCA bcol: ptpcoll (MCA v2.0, API v2.0, Component v1.8.4) MCA bml: r2 (MCA v2.0, API v2.0, Component v1.8.4) MCA btl: self (MCA v2.0, API v2.0, Component v1.8.4) MCA btl: sm (MCA v2.0, API v2.0, Component v1.8.4) MCA btl: tcp (MCA v2.0, API v2.0, Component v1.8.4) MCA btl: vader (MCA v2.0, API v2.0, Component v1.8.4) MCA coll: basic (MCA v2.0, API v2.0, Component v1.8.4) MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.8.4) MCA coll: inter (MCA v2.0, API v2.0, Component v1.8.4) MCA coll: libnbc (MCA v2.0, API v2.0, Component v1.8.4) MCA coll: ml (MCA v2.0, API v2.0, Component v1.8.4) MCA coll: self (MCA v2.0, API v2.0, Component v1.8.4) MCA coll: sm (MCA v2.0, API v2.0, Component v1.8.4) MCA coll: tuned (MCA v2.0, API v2.0, Component v1.8.4) MCA dpm: orte (MCA v2.0, API v2.0, Component v1.8.4) MCA fbtl: posix (MCA v2.0, API v2.0, Component v1.8.4) MCA fcoll: dynamic (MCA v2.0, API v2.0, Component v1.8.4) MCA fcoll: individual (MCA v2.0, API v2.0, Component v1.8.4) MCA fcoll: static (MCA v2.0, API v2.0, Component v1.8.4) MCA fcoll: two_phase (MCA v2.0, API v2.0, Component v1.8.4) MCA fcoll: ylib (MCA v2.0, API v2.0, Component v1.8.4) MCA fs: ufs (MCA v2.0, API v2.0, Component v1.8.4) MCA io: ompio (MCA v2.0, API v2.0, Component v1.8.4) MCA io: romio (MCA v2.0, API v2.0, Component v1.8.4) MCA mpool: grdma (MCA v2.0, API v2.0, Component v1.8.4) MCA mpool: sm (MCA v2.0, API v2.0, Component v1.8.4) MCA osc: rdma (MCA v2.0, API v3.0, Component v1.8.4) MCA osc: sm (MCA v2.0, API v3.0, Component v1.8.4) MCA pml: v (MCA v2.0, API v2.0, Component v1.8.4) MCA pml: bfo (MCA v2.0, API v2.0, Component v1.8.4) MCA pml: cm (MCA v2.0, API v2.0, Component v1.8.4) MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.8.4) MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.8.4) MCA rcache: vma (MCA v2.0, API v2.0, Component v1.8.4) MCA rte: orte (MCA v2.0, API v2.0, Component v1.8.4) MCA sbgp: basesmsocket (MCA v2.0, API v2.0, Component v1.8.4) MCA sbgp: basesmuma (MCA v2.0, API v2.0, Component v1.8.4) MCA sbgp: p2p (MCA v2.0, API v2.0, Component v1.8.4) MCA sharedfp: individual (MCA v2.0, API v2.0, Component v1.8.4) MCA sharedfp: lockedfile (MCA v2.0, API v2.0, Component v1.8.4) MCA sharedfp: sm (MCA v2.0, API v2.0, Component v1.8.4) MCA topo: basic (MCA v2.0, API v2.1, Component v1.8.4) MCA vprotocol: pessimist (MCA v2.0, API v2.0, Component v1.8.4) Sachin >Sachin, >Can you, please, provide a command line? Additional information about your >system could be helpful also. >Josh >>On Wed, Feb 18, 2015 at 3:43 AM, Sachin Krishnan <sachkris_at_[hidden]> wrote: >> Hello, >> >> I am new to MPI and also this list. >> I wrote an MPI code with several MPI_Bcast calls in a loop. My code was >> getting stuck at random points, ie it was not systematic. After a few hours >> of debugging and googling, I found that the issue may be with the several >> MPI_Bcast calls in a loop. >> >> I stumbled on this test code which can reproduce the issue: >> https://github.com/fintler/ompi/blob/master/orte/test/mpi/bcast_loop.c >> >> Im using OpenMPI v1.8.4 installed from official Arch Linux repo. >> >> Is it a known issue with OpenMPI? >> Is it some problem with the way openmpi is configured in my system? >> >> Thanks in advance. >> >> Sachin >> >> >> >> _______________________________________________ >> users mailing list >> users_at_[hidden] >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2015/02/26338.php >>