Dear Open-MPI Community:
I have a 3 node cluster, each a dual opteron workstation running
OpenSUSE 10.1 64-bit. The node names are LT, SGT and PFC. When I
start an mpirun job from either SGT or PFC, things work as they are
supposed to. However, if I start the same job from LT, the jobs hangs
at
Using the --mca btl ^mx totally prevents use of mx interface. So
everybody uses tcp (even mx capable nodes). If you want a mixed
configuration you have to enforce use of the ob1 pml, but let the mx
btl be used where it is suitable (it will be disabled at runtime if it
can't run). You're pro
Hmm, that is the way that I expected it to work as well -
we see the warnings also, but closely followed by the
errors (I've been trying both 1.2.5 and a recent 1.3
snapshot with the same behavior). You don't have the
mx driver loaded on the nodes that do not have a myrinet
card, do you? Our mx
We also have a mixed myrinet/ip cluster, and maybe I'm missing some
nuance of your configuration, but openmpi seems to work fine for me "as
is" with no --mca options across mixed nodes (there's a bunch of
warnings at the beginning where the non-mx nodes realize they don't have
myrinet cards and the
Hi Matt,
M Jones wrote:
I thought that we would be able to use a single open-mpi build to
support both networks (and users would be able to request mx nodes if
they need them using the batch queuing system, which they are
already accustomed to). Am I missing something (or just doing
I don't k
Hmm, that combination seems to hang on me - but
'--mca pml ob1 --mca btl ^mx' does indeed do the trick.
Many thanks!
Matt
On Tue, 15 Jan 2008, George Bosilca wrote:
This case actually works. We run into it few days ago, when we discovered
that one of the compute nodes in a cluster didn't get
This case actually works. We run into it few days ago, when we
discovered that one of the compute nodes in a cluster didn't get his
Myrinet card installed properly ... The performance were horrible but
the application run to completion.
You will have to use the following flags: --mca pml ob
On Tue, Jan 15, 2008 at 09:49:40AM -0500, M Jones wrote:
> Hi,
>
>We have a mixed environment in which roughly 2/3 of the nodes
> in our cluster have myrinet (mx 1.2.1), while the full cluster has
> gigE. Running open-mpi exclusively on myrinet nodes or exclusively
> on non-myrinet nodes is f
Hi,
We have a mixed environment in which roughly 2/3 of the nodes
in our cluster have myrinet (mx 1.2.1), while the full cluster has
gigE. Running open-mpi exclusively on myrinet nodes or exclusively
on non-myrinet nodes is fine, but mixing the two nodes types
results in a runtime error (PML a
Hi Ralph,
Sorry that I did not come back to clean up this request. Actually, it
was a stupid user (my) error. I had not included stdlib.h in my
source. Sorry again and thanks for the effort.
Prakash
On Jan 14, 2008, at 11:12 PM, Jeff Squyres wrote:
Sorry, this mail slipped by me.
The mo
Hi,
I am trying to write a simple code which does the following -
A master process running on 'n' processors spawn 4 processes using the
MPI_Comm_spawn_multiple command. After spawning, the intercommunicator
between the master and the spawned processes are merged using
MPI_Intercomm_merge to create
11 matches
Mail list logo