On 3 April 2009 at 03:33, Jerome BENOIT wrote:
| The above submission works the same on my clusters.
| But in fact, my issue involve interconnection between the nodes of the 
clusters:
| in the above examples involve no connection between nodes.
| 
| My cluster is a cluster of quadcore computers:
| if in the sbatch script
| 
| #SBATCH --nodes=7
| #SBATCH --ntasks=15
| 
| is replaced by
| 
| #SBATCH --nodes=1
| #SBATCH --ntasks=4
| 
| everything is fine as no interconnection is involved.
| 
| Can you test the inconnection part of the story ?

Again, think about in terms of layers. You have a problem with slurm on top
of Open MPI.  

So before blaming Open MPI, I would try something like this:

~$ orterun -np 2 -H abc,xyz /tmp/jerome_hw
Hello world! I am 1 of 2 and my name is `abc'
Hello world! I am 0 of 2 and my name is `xyz'
~$

ie whether the simple MPI example can be launched successfully on two nodes or 
not.

Dirk

-- 
Three out of two people have difficulties with fractions.

Reply via email to