Re: [O-MPI users] Open MPI @SC2005

2005-11-16 Thread Konstantin Karganov
> The Open MPI Team will be at SC this year.
> I am co-hosting a BOF entitled "Why MPI Makes You Scream!  And how can
> we simplify parallel debugging?"
By the way (though it may be offtopic here)...

Does anybody know some statistics on MPI usage. e.i. :
- percentage of C, C++, Fortran MPI programs
- percentage of SPMD and MPMD programs
- MPI function an function groups usage (how many people use
topologies, asynchronous IO, communicator attributes etc.)
- average number of nodes for running and for debugging

I suppose this can be interesting for all MPI community as a valuable
feedback from those who use MPI for real applications. There was a
survey on SC'93 on the usage of debugging tools (
http://citeseer.ist.psu.edu/pancake94what.html ) and It might be
possible (if its not too late now) to do something alike.

--
Best regards,
Konstantin.



Re: [O-MPI users] Open MPI @SC2005

2005-11-16 Thread Edgar Gabriel

Konstantin Karganov wrote:


The Open MPI Team will be at SC this year.
I am co-hosting a BOF entitled "Why MPI Makes You Scream!  And how can
we simplify parallel debugging?"
   


By the way (though it may be offtopic here)...

Does anybody know some statistics on MPI usage. e.i. :
- percentage of C, C++, Fortran MPI programs
- percentage of SPMD and MPMD programs
- MPI function an function groups usage (how many people use
topologies, asynchronous IO, communicator attributes etc.)
- average number of nodes for running and for debugging

I suppose this can be interesting for all MPI community as a valuable
feedback from those who use MPI for real applications. There was a
survey on SC'93 on the usage of debugging tools (
http://citeseer.ist.psu.edu/pancake94what.html ) and It might be
possible (if its not too late now) to do something alike.
 

There has been at least one project analysing these kind of things, look 
e.g. at

http://www.hlrs.de/people/rabenseifner/publ/publications.html

for  papers about automatic MPI counter profiling. These were studies 
about the usage of MPI at two different HPC centers.


Best regards
Edgar



--
Best regards,
Konstantin.

___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
 




--
Edgar Gabriel
Assistant Professor
Department of Computer Science  email:gabr...@cs.uh.edu
University of Houston   http://www.cs.uh.edu/~gabriel
Philip G. Hoffman Hall, Room 524Tel: +1 (713) 743-3857
Houston, TX-77204, USA  Fax: +1 (713) 743-3335




[O-MPI users] Anyone installed openmpi in Redhat 4?

2005-11-16 Thread Clement Chu

Hi,

   Did anyone installed openmpi in Redhat Core 4?  I got a major 
problem to run mpi program with openmpi in RH 4.  I would like to share 
your experience.


Regards,
Clement

--
Clement Kam Man Chu
Research Assistant
School of Computer Science & Software Engineering
Monash University, Caulfield Campus
Ph: 61 3 9903 1964



Re: [O-MPI users] Error on mpirun in Redhat Fedora Core 4

2005-11-16 Thread Jeff Squyres

Clement --

Sorry for the delay in replying.  We're running around crazy here at 
SC, which pretty much keeps us away from e-mail except early in the 
morning and late at night.


We fixed a bunch of things in the sm btl as of r8136 (someone reported 
similar issues as you, and we took the exchange off-list to fix).  The 
problems could definitely affect correctness and cause segv's similar 
to what you were seeing (see 
http://www.open-mpi.org/community/lists/users/2005/11/0326.php for a 
little more info).


I notice that you're running 8113 here -- could you try the latest 
nightly snapshot or rc and see if the same problems occur?


Thanks for your patience!


On Nov 14, 2005, at 4:51 AM, Clement Chu wrote:


Hi Jeff,

   I tried the rc6 and trunk nightly 8150.  I got the same problem.  I 
copied the message from terminal as below.


[clement@localhost testmpi]$ ompi_info
   Open MPI: 1.1a1r8113
  Open MPI SVN revision: r8113
   Open RTE: 1.1a1r8113
  Open RTE SVN revision: r8113
   OPAL: 1.1a1r8113
  OPAL SVN revision: r8113
 Prefix: /home/clement/openmpi/
Configured architecture: i686-pc-linux-gnu
  Configured by: clement
  Configured on: Mon Nov 14 10:12:12 EST 2005
 Configure host: localhost
   Built by: clement
   Built on: Mon Nov 14 10:28:21 EST 2005
 Built host: localhost
 C bindings: yes
   C++ bindings: yes
 Fortran77 bindings: yes (all)
 Fortran90 bindings: yes
 C compiler: gcc
C compiler absolute: /usr/bin/gcc
   C++ compiler: g++
  C++ compiler absolute: /usr/bin/g++
 Fortran77 compiler: gfortran
 Fortran77 compiler abs: /usr/bin/gfortran
 Fortran90 compiler: gfortran
 Fortran90 compiler abs: /usr/bin/gfortran
C profiling: yes
  C++ profiling: yes
Fortran77 profiling: yes
Fortran90 profiling: yes
 C++ exceptions: no
 Thread support: posix (mpi: no, progress: no)
 Internal debug support: no
MPI parameter check: runtime
Memory profiling support: no
Memory debugging support: no
libltdl support: 1
 MCA memory: malloc_hooks (MCA v1.0, API v1.0, Component 
v1.1)

  MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.1)
  MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.1)
  MCA timer: linux (MCA v1.0, API v1.0, Component v1.1)
  MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
  MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
   MCA coll: basic (MCA v1.0, API v1.0, Component v1.1)
   MCA coll: hierarch (MCA v1.0, API v1.0, Component v1.1)
   MCA coll: self (MCA v1.0, API v1.0, Component v1.1)
   MCA coll: sm (MCA v1.0, API v1.0, Component v1.1)
 MCA io: romio (MCA v1.0, API v1.0, Component v1.1)
  MCA mpool: sm (MCA v1.0, API v1.0, Component v1.1)
MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.1)
MCA pml: teg (MCA v1.0, API v1.0, Component v1.1)
MCA pml: uniq (MCA v1.0, API v1.0, Component v1.1)
MCA ptl: self (MCA v1.0, API v1.0, Component v1.1)
MCA ptl: sm (MCA v1.0, API v1.0, Component v1.1)
MCA ptl: tcp (MCA v1.0, API v1.0, Component v1.1)
MCA btl: self (MCA v1.0, API v1.0, Component v1.1)
MCA btl: sm (MCA v1.0, API v1.0, Component v1.1)
MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0)
   MCA topo: unity (MCA v1.0, API v1.0, Component v1.1)
MCA gpr: null (MCA v1.0, API v1.0, Component v1.1)
MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.1)
MCA gpr: replica (MCA v1.0, API v1.0, Component v1.1)
MCA iof: proxy (MCA v1.0, API v1.0, Component v1.1)
MCA iof: svc (MCA v1.0, API v1.0, Component v1.1)
 MCA ns: proxy (MCA v1.0, API v1.0, Component v1.1)
 MCA ns: replica (MCA v1.0, API v1.0, Component v1.1)
MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
MCA ras: dash_host (MCA v1.0, API v1.0, Component v1.1)
MCA ras: hostfile (MCA v1.0, API v1.0, Component v1.1)
MCA ras: localhost (MCA v1.0, API v1.0, Component v1.1)
MCA ras: slurm (MCA v1.0, API v1.0, Component v1.1)
MCA rds: hostfile (MCA v1.0, API v1.0, Component v1.1)
MCA rds: resfile (MCA v1.0, API v1.0, Component v1.1)
  MCA rmaps: round_robin (MCA v1.0, API v1.0, Component 
v1.1)

   MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.1)
   MCA rmgr: urm (MCA v1.0, API v1.0, Component v1.1)
MCA rml: oob (MCA v1.0, API v1.0, Component v1.1)
MCA pls: fork (MCA v1.0, API v1.0, Component v1.1)
MCA

Re: [O-MPI users] Error on mpirun in Redhat Fedora Core 4

2005-11-16 Thread Mostyn Lewis

I get this on rh9 ONLY if I leave out a -hostfile  option
on mpirun, otherwise it works fine.
This is an old Red Hat.

Regards,
Mostyn

On Wed, 16 Nov 2005, Jeff Squyres wrote:


Clement --

Sorry for the delay in replying.  We're running around crazy here at
SC, which pretty much keeps us away from e-mail except early in the
morning and late at night.

We fixed a bunch of things in the sm btl as of r8136 (someone reported
similar issues as you, and we took the exchange off-list to fix).  The
problems could definitely affect correctness and cause segv's similar
to what you were seeing (see
http://www.open-mpi.org/community/lists/users/2005/11/0326.php for a
little more info).

I notice that you're running 8113 here -- could you try the latest
nightly snapshot or rc and see if the same problems occur?

Thanks for your patience!


On Nov 14, 2005, at 4:51 AM, Clement Chu wrote:


Hi Jeff,

   I tried the rc6 and trunk nightly 8150.  I got the same problem.  I
copied the message from terminal as below.

[clement@localhost testmpi]$ ompi_info
   Open MPI: 1.1a1r8113
  Open MPI SVN revision: r8113
   Open RTE: 1.1a1r8113
  Open RTE SVN revision: r8113
   OPAL: 1.1a1r8113
  OPAL SVN revision: r8113
 Prefix: /home/clement/openmpi/
Configured architecture: i686-pc-linux-gnu
  Configured by: clement
  Configured on: Mon Nov 14 10:12:12 EST 2005
 Configure host: localhost
   Built by: clement
   Built on: Mon Nov 14 10:28:21 EST 2005
 Built host: localhost
 C bindings: yes
   C++ bindings: yes
 Fortran77 bindings: yes (all)
 Fortran90 bindings: yes
 C compiler: gcc
C compiler absolute: /usr/bin/gcc
   C++ compiler: g++
  C++ compiler absolute: /usr/bin/g++
 Fortran77 compiler: gfortran
 Fortran77 compiler abs: /usr/bin/gfortran
 Fortran90 compiler: gfortran
 Fortran90 compiler abs: /usr/bin/gfortran
C profiling: yes
  C++ profiling: yes
Fortran77 profiling: yes
Fortran90 profiling: yes
 C++ exceptions: no
 Thread support: posix (mpi: no, progress: no)
 Internal debug support: no
MPI parameter check: runtime
Memory profiling support: no
Memory debugging support: no
libltdl support: 1
 MCA memory: malloc_hooks (MCA v1.0, API v1.0, Component
v1.1)
  MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.1)
  MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.1)
  MCA timer: linux (MCA v1.0, API v1.0, Component v1.1)
  MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0)
  MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0)
   MCA coll: basic (MCA v1.0, API v1.0, Component v1.1)
   MCA coll: hierarch (MCA v1.0, API v1.0, Component v1.1)
   MCA coll: self (MCA v1.0, API v1.0, Component v1.1)
   MCA coll: sm (MCA v1.0, API v1.0, Component v1.1)
 MCA io: romio (MCA v1.0, API v1.0, Component v1.1)
  MCA mpool: sm (MCA v1.0, API v1.0, Component v1.1)
MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.1)
MCA pml: teg (MCA v1.0, API v1.0, Component v1.1)
MCA pml: uniq (MCA v1.0, API v1.0, Component v1.1)
MCA ptl: self (MCA v1.0, API v1.0, Component v1.1)
MCA ptl: sm (MCA v1.0, API v1.0, Component v1.1)
MCA ptl: tcp (MCA v1.0, API v1.0, Component v1.1)
MCA btl: self (MCA v1.0, API v1.0, Component v1.1)
MCA btl: sm (MCA v1.0, API v1.0, Component v1.1)
MCA btl: tcp (MCA v1.0, API v1.0, Component v1.0)
   MCA topo: unity (MCA v1.0, API v1.0, Component v1.1)
MCA gpr: null (MCA v1.0, API v1.0, Component v1.1)
MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.1)
MCA gpr: replica (MCA v1.0, API v1.0, Component v1.1)
MCA iof: proxy (MCA v1.0, API v1.0, Component v1.1)
MCA iof: svc (MCA v1.0, API v1.0, Component v1.1)
 MCA ns: proxy (MCA v1.0, API v1.0, Component v1.1)
 MCA ns: replica (MCA v1.0, API v1.0, Component v1.1)
MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0)
MCA ras: dash_host (MCA v1.0, API v1.0, Component v1.1)
MCA ras: hostfile (MCA v1.0, API v1.0, Component v1.1)
MCA ras: localhost (MCA v1.0, API v1.0, Component v1.1)
MCA ras: slurm (MCA v1.0, API v1.0, Component v1.1)
MCA rds: hostfile (MCA v1.0, API v1.0, Component v1.1)
MCA rds: resfile (MCA v1.0, API v1.0, Component v1.1)
  MCA rmaps: round_robin (MCA v1.0, API v1.0, Component
v1.1)
   MCA rmgr: proxy (MCA v1.0, API v1.0, Component v1.1)
   MCA rmgr: urm (MCA v1.0, API v1.0, Compon

[O-MPI users] Configuring port

2005-11-16 Thread Enrique Curchitser

Hi,

I put together a small cluster (4 computers) which has one head node  
that sees the world
and 3 that are on a private network.  If I want to use the head node  
(which has 2 NICs)
as part of the ring, how do I tell it to go over the NIC that is part  
of the private network?
In mpich, there is a modification to the hostfile, but I did not see  
any such thing with OpenMPI


Thanks,

Enrique

_
Enrique Curchitser
ec...@columbia.edu





Re: [O-MPI users] Configuring port

2005-11-16 Thread Troy Telford
On Wed, 16 Nov 2005 14:16:20 -0700, Enrique Curchitser  
 wrote:

Hi,

I put together a small cluster (4 computers) which has one head node
that sees the world
and 3 that are on a private network.  If I want to use the head node
(which has 2 NICs)
as part of the ring, how do I tell it to go over the NIC that is part
of the private network?
In mpich, there is a modification to the hostfile, but I did not see
any such thing with OpenMPI


Here's an entry from an earlier post to this list (in the archives) that I  
believe applies:


http://www.open-mpi.org/community/lists/users/2005/11/0337.php


Re: [O-MPI users] Configuring port

2005-11-16 Thread Enrique Curchitser

Thank you.  It did the trick!  Now on to optimizing the communications
over gig-ethernet.


_
Enrique Curchitser
ec...@columbia.edu



On Nov 16, 2005, at 4:29 PM, Troy Telford wrote:


On Wed, 16 Nov 2005 14:16:20 -0700, Enrique Curchitser
 wrote:


Hi,

I put together a small cluster (4 computers) which has one head node
that sees the world
and 3 that are on a private network.  If I want to use the head node
(which has 2 NICs)
as part of the ring, how do I tell it to go over the NIC that is part
of the private network?
In mpich, there is a modification to the hostfile, but I did not see
any such thing with OpenMPI



Here's an entry from an earlier post to this list (in the archives)  
that I

believe applies:

http://www.open-mpi.org/community/lists/users/2005/11/0337.php
___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users