On Wed, Sep 19, 2007 at 01:58:35PM -0600, Edmund Sumbar wrote:
> I'm trying to run skampi-5.0.1-r0191 under PBS
> over IB with the command line
> 
>    mpirun -np 2 ./skampi -i coll.ski -o coll_ib.sko
Can you add 
choose_barrier_synchronization()
to coll.ski and try again? It looks like this one:
https://svn.open-mpi.org/trac/ompi/ticket/1015

> 
> The pt2pt and mmisc tests run to completion.
> The coll and onesided tests, on the other hand,
> start to produce output but then seem to hang.
> Actually, the cpus appear to be busy doing
> something (I don't know what), but output stops.
> The tests should only last the order of minutes
> but I end up deleting the job after about 15 min.
> 
> All test run to completion with --mca btl tcp,self
> 
> Any suggestions as to how to diagnose this problem?
> Are there any known issues with OpenMPI/IB and the
> SKaMPI benchmark?
> 
> (BTW, skampi works with mvapich2)
> 
> System details follow...
> 
> -- 
> Ed[mund [Sumbar]]
> AICT Research Support, Univ of Alberta
> 
> 
> $ uname -a
> Linux opteron-cluster.nic.ualberta.ca 2.6.21-smp #1 SMP Tue Aug 7 12:45:20 
> MDT 2007 x86_64 x86_64 x86_64 GNU/Linux
> 
> $ ./configure --prefix=/usr/local/openmpi-1.2.3 --with-tm=/opt/torque 
> --with-openib=/usr/lib --with-libnuma=/usr/lib64
> 
> $ ompi_info
>                  Open MPI: 1.2.3
>     Open MPI SVN revision: r15136
>                  Open RTE: 1.2.3
>     Open RTE SVN revision: r15136
>                      OPAL: 1.2.3
>         OPAL SVN revision: r15136
>                    Prefix: /usr/local/openmpi-1.2.3
>   Configured architecture: x86_64-unknown-linux-gnu
>             Configured by: esumbar
>             Configured on: Mon Sep 17 10:00:35 MDT 2007
>            Configure host: opteron-cluster.nic.ualberta.ca
>                  Built by: esumbar
>                  Built on: Mon Sep 17 10:05:09 MDT 2007
>                Built host: opteron-cluster.nic.ualberta.ca
>                C bindings: yes
>              C++ bindings: yes
>        Fortran77 bindings: yes (all)
>        Fortran90 bindings: yes
>   Fortran90 bindings size: small
>                C compiler: gcc
>       C compiler absolute: /usr/bin/gcc
>              C++ compiler: g++
>     C++ compiler absolute: /usr/bin/g++
>        Fortran77 compiler: gfortran
>    Fortran77 compiler abs: /usr/bin/gfortran
>        Fortran90 compiler: gfortran
>    Fortran90 compiler abs: /usr/bin/gfortran
>               C profiling: yes
>             C++ profiling: yes
>       Fortran77 profiling: yes
>       Fortran90 profiling: yes
>            C++ exceptions: no
>            Thread support: posix (mpi: no, progress: no)
>    Internal debug support: no
>       MPI parameter check: runtime
> Memory profiling support: no
> Memory debugging support: no
>           libltdl support: yes
>     Heterogeneous support: yes
>   mpirun default --prefix: no
>             MCA backtrace: execinfo (MCA v1.0, API v1.0, Component v1.2.3)
>                MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component v1.2.3)
>             MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2.3)
>             MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.2.3)
>             MCA maffinity: libnuma (MCA v1.0, API v1.0, Component v1.2.3)
>                 MCA timer: linux (MCA v1.0, API v1.0, Component v1.2.3)
>           MCA installdirs: env (MCA v1.0, API v1.0, Component v1.2.3)
>           MCA installdirs: config (MCA v1.0, API v1.0, Component v1.2.3)
>             MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
>             MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
>                  MCA coll: basic (MCA v1.0, API v1.0, Component v1.2.3)
>                  MCA coll: self (MCA v1.0, API v1.0, Component v1.2.3)
>                  MCA coll: sm (MCA v1.0, API v1.0, Component v1.2.3)
>                  MCA coll: tuned (MCA v1.0, API v1.0, Component v1.2.3)
>                    MCA io: romio (MCA v1.0, API v1.0, Component v1.2.3)
>                 MCA mpool: rdma (MCA v1.0, API v1.0, Component v1.2.3)
>                 MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA pml: cm (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2.3)
>                MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA btl: openib (MCA v1.0, API v1.0.1, Component v1.2.3)
>                   MCA btl: self (MCA v1.0, API v1.0.1, Component v1.2.3)
>                   MCA btl: sm (MCA v1.0, API v1.0.1, Component v1.2.3)
>                   MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0)
>                  MCA topo: unity (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.2.3)
>                MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2.3)
>                MCA errmgr: orted (MCA v1.0, API v1.3, Component v1.2.3)
>                MCA errmgr: proxy (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA gpr: null (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA gpr: replica (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA iof: proxy (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA iof: svc (MCA v1.0, API v1.0, Component v1.2.3)
>                    MCA ns: proxy (MCA v1.0, API v2.0, Component v1.2.3)
>                    MCA ns: replica (MCA v1.0, API v2.0, Component v1.2.3)
>                   MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
>                   MCA ras: dash_host (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA ras: gridengine (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA ras: localhost (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA ras: slurm (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA ras: tm (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA rds: hostfile (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA rds: proxy (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA rds: resfile (MCA v1.0, API v1.3, Component v1.2.3)
>                 MCA rmaps: round_robin (MCA v1.0, API v1.3, Component v1.2.3)
>                  MCA rmgr: proxy (MCA v1.0, API v2.0, Component v1.2.3)
>                  MCA rmgr: urm (MCA v1.0, API v2.0, Component v1.2.3)
>                   MCA rml: oob (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA pls: gridengine (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA pls: proxy (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA pls: rsh (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA pls: slurm (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA pls: tm (MCA v1.0, API v1.3, Component v1.2.3)
>                   MCA sds: env (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA sds: pipe (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA sds: seed (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA sds: singleton (MCA v1.0, API v1.0, Component v1.2.3)
>                   MCA sds: slurm (MCA v1.0, API v1.0, Component v1.2.3)
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

--
                        Gleb.

Reply via email to