[OMPI users] odd network behavior

2008-01-15 Thread Mark Kosmowski
Dear Open-MPI Community: I have a 3 node cluster, each a dual opteron workstation running OpenSUSE 10.1 64-bit. The node names are LT, SGT and PFC. When I start an mpirun job from either SGT or PFC, things work as they are supposed to. However, if I start the same job from LT, the jobs hangs at

Re: [OMPI users] mixed myrinet/non-myrinet nodes

2008-01-15 Thread Aurélien Bouteiller
Using the --mca btl ^mx totally prevents use of mx interface. So everybody uses tcp (even mx capable nodes). If you want a mixed configuration you have to enforce use of the ob1 pml, but let the mx btl be used where it is suitable (it will be disabled at runtime if it can't run). You're pro

Re: [OMPI users] mixed myrinet/non-myrinet nodes

2008-01-15 Thread M D Jones
Hmm, that is the way that I expected it to work as well - we see the warnings also, but closely followed by the errors (I've been trying both 1.2.5 and a recent 1.3 snapshot with the same behavior). You don't have the mx driver loaded on the nodes that do not have a myrinet card, do you? Our mx

Re: [OMPI users] mixed myrinet/non-myrinet nodes

2008-01-15 Thread 8mj6tc902
We also have a mixed myrinet/ip cluster, and maybe I'm missing some nuance of your configuration, but openmpi seems to work fine for me "as is" with no --mca options across mixed nodes (there's a bunch of warnings at the beginning where the non-mx nodes realize they don't have myrinet cards and the

Re: [OMPI users] mixed myrinet/non-myrinet nodes

2008-01-15 Thread Patrick Geoffray
Hi Matt, M Jones wrote: I thought that we would be able to use a single open-mpi build to support both networks (and users would be able to request mx nodes if they need them using the batch queuing system, which they are already accustomed to). Am I missing something (or just doing I don't k

Re: [OMPI users] mixed myrinet/non-myrinet nodes

2008-01-15 Thread M D Jones
Hmm, that combination seems to hang on me - but '--mca pml ob1 --mca btl ^mx' does indeed do the trick. Many thanks! Matt On Tue, 15 Jan 2008, George Bosilca wrote: This case actually works. We run into it few days ago, when we discovered that one of the compute nodes in a cluster didn't get

Re: [OMPI users] mixed myrinet/non-myrinet nodes

2008-01-15 Thread George Bosilca
This case actually works. We run into it few days ago, when we discovered that one of the compute nodes in a cluster didn't get his Myrinet card installed properly ... The performance were horrible but the application run to completion. You will have to use the following flags: --mca pml ob

Re: [OMPI users] mixed myrinet/non-myrinet nodes

2008-01-15 Thread Gleb Natapov
On Tue, Jan 15, 2008 at 09:49:40AM -0500, M Jones wrote: > Hi, > >We have a mixed environment in which roughly 2/3 of the nodes > in our cluster have myrinet (mx 1.2.1), while the full cluster has > gigE. Running open-mpi exclusively on myrinet nodes or exclusively > on non-myrinet nodes is f

[OMPI users] mixed myrinet/non-myrinet nodes

2008-01-15 Thread M Jones
Hi, We have a mixed environment in which roughly 2/3 of the nodes in our cluster have myrinet (mx 1.2.1), while the full cluster has gigE. Running open-mpi exclusively on myrinet nodes or exclusively on non-myrinet nodes is fine, but mixing the two nodes types results in a runtime error (PML a

Re: [OMPI users] getenv issue

2008-01-15 Thread Prakash Velayutham
Hi Ralph, Sorry that I did not come back to clean up this request. Actually, it was a stupid user (my) error. I had not included stdlib.h in my source. Sorry again and thanks for the effort. Prakash On Jan 14, 2008, at 11:12 PM, Jeff Squyres wrote: Sorry, this mail slipped by me. The mo

[OMPI users] MPI_Finalize segmentation fault with MPI_Intercomm_merge

2008-01-15 Thread Rajesh Sudarsan
Hi, I am trying to write a simple code which does the following - A master process running on 'n' processors spawn 4 processes using the MPI_Comm_spawn_multiple command. After spawning, the intercommunicator between the master and the spawned processes are merged using MPI_Intercomm_merge to create