Hello,
on win32 openmpi 1.4.3, I have a slave process that reaches this pseudo-code
and then blocks and the CPU usage for that process stays at 25% all the time (I
have a quadcore processor). When I set the affinity to 1 of the cores, that
core is 100% busy because of my slave process.
main()
{
....
.....
MPI_ISEND
std::cout<< "about to get broadcast"<<std::endl;
MPI_Bcast of an integer
std::cout<< " broadcast received"<<std::endl;
...
}
The first printout is seen but not the next which makes me thing the process is
inside the MPI_Bcast call. Should the CPU be 100% busy while this call is
waiting for the broadcast message to arrive?
Any ideas? below the output of ompi-info:
------------------------------------------------------------------------------------------------------------------------------------------------------------------
Package: Open MPI
Distribution
Open MPI: 1.4.3
Open MPI SVN revision: r23834
Open MPI release date: Oct 05, 2010
Open RTE: 1.4.3
Open RTE SVN revision: r23834
Open RTE release date: Oct 05, 2010
OPAL: 1.4.3
OPAL SVN revision: r23834
OPAL release date: Oct 05, 2010
Ident string: 1.4.3
Prefix: C:/Program Files/openmpi
Configured architecture: x86 Windows-5.1
Configure host: LC12-003-D-055A
Configured by: hicham.mouline
Configured on: 18:07 19/11/2010
Configure host:
Built by: hicham.mouline
Built on: 18:07 19/11/2010
Built host:
C bindings: yes
C++ bindings: yes
Fortran77 bindings: no
Fortran90 bindings: no
Fortran90 bindings size: na
C compiler: C:/Program Files/Microsoft Visual Studio
9.0/VC/bin/cl.exe
C compiler absolute: C:/Program Files/Microsoft Visual Studio
9.0/VC/bin/cl.exe
C++ compiler: C:/Program Files/Microsoft Visual Studio
9.0/VC/bin/cl.exe
C++ compiler absolute: C:/Program Files/Microsoft Visual Studio
9.0/VC/bin/cl.exe
Fortran77 compiler: CMAKE_Fortran_COMPILER-NOTFOUND
Fortran77 compiler abs: none
Fortran90 compiler:
Fortran90 compiler abs: none
C profiling: yes
C++ profiling: yes
Fortran77 profiling: no
Fortran90 profiling: no
C++ exceptions: no
Thread support: no
Sparse Groups: no
Internal debug support: no
MPI parameter check: runtime
Memory profiling support: no
Memory debugging support: no
libltdl support: no
Heterogeneous support: no
mpirun default --prefix: yes
MPI I/O support: yes
MPI_WTIME support: gettimeofday
Symbol visibility support: yes
FT Checkpoint support: yes (checkpoint thread: no)
MCA backtrace: none (MCA v2.0, API v2.0, Component v1.4.3)
MCA paffinity: windows (MCA v2.0, API v2.0, Component v1.4.3)
MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.4.3)
MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.4.3)
MCA timer: windows (MCA v2.0, API v2.0, Component v1.4.3)
MCA installdirs: windows (MCA v2.0, API v2.0, Component v1.4.3)
MCA installdirs: env (MCA v2.0, API v2.0, Component v1.4.3)
MCA installdirs: config (MCA v2.0, API v2.0, Component v1.4.3)
MCA crs: none (MCA v2.0, API v2.0, Component v1.4.3)
MCA dpm: orte (MCA v2.0, API v2.0, Component v1.4.3)
MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.4.3)
MCA allocator: basic (MCA v2.0, API v2.0, Component v1.4.3)
MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.4.3)
MCA coll: basic (MCA v2.0, API v2.0, Component v1.4.3)
MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.4.3)
MCA coll: self (MCA v2.0, API v2.0, Component v1.4.3)
MCA coll: sm (MCA v2.0, API v2.0, Component v1.4.3)
MCA coll: sync (MCA v2.0, API v2.0, Component v1.4.3)
MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.4.3)
MCA mpool: sm (MCA v2.0, API v2.0, Component v1.4.3)
MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.4.3)
MCA bml: r2 (MCA v2.0, API v2.0, Component v1.4.3)
MCA btl: self (MCA v2.0, API v2.0, Component v1.4.3)
MCA btl: sm (MCA v2.0, API v2.0, Component v1.4.3)
MCA btl: tcp (MCA v2.0, API v2.0, Component v1.4.3)
MCA topo: unity (MCA v2.0, API v2.0, Component v1.4.3)
MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.4.3)
MCA osc: rdma (MCA v2.0, API v2.0, Component v1.4.3)
MCA iof: hnp (MCA v2.0, API v2.0, Component v1.4.3)
MCA iof: orted (MCA v2.0, API v2.0, Component v1.4.3)
MCA iof: tool (MCA v2.0, API v2.0, Component v1.4.3)
MCA oob: tcp (MCA v2.0, API v2.0, Component v1.4.3)
MCA odls: process (MCA v2.0, API v2.0, Component v1.4.3)
MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.4.3)
MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.4.3)
MCA rml: ftrm (MCA v2.0, API v2.0, Component v1.4.3)
MCA rml: oob (MCA v2.0, API v2.0, Component v1.4.3)
MCA routed: binomial (MCA v2.0, API v2.0, Component v1.4.3)
MCA routed: linear (MCA v2.0, API v2.0, Component v1.4.3)
MCA plm: process (MCA v2.0, API v2.0, Component v1.4.3)
MCA errmgr: default (MCA v2.0, API v2.0, Component v1.4.3)
MCA ess: env (MCA v2.0, API v2.0, Component v1.4.3)
MCA ess: hnp (MCA v2.0, API v2.0, Component v1.4.3)
MCA ess: singleton (MCA v2.0, API v2.0, Component v1.4.3)
MCA ess: tool (MCA v2.0, API v2.0, Component v1.4.3)
MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.4.3)
regards,