Hi George,
I think the confusion was my fault because --mca pml teg did not
produce errors and gave almost the same performance as Mpich2 v 1.02p1.
The reason why I cannot do what you suggest below is because the
.openmpi/mca-params.conf file if I am not mistaken would reside in my
home NFS share directory. I have installed a new 5.01 beta version of
Oscar and /home/allan is a shared directory of my head node where the
openmpi installation resides.[/home/allan/openmpi with paths in the
.bash_profile and .bashrc files] I would have to do an individual 16
installations of open mpi on each node for /opt/openmpi and the
mca-params file residing in there. Tell me if I am wrong. I might have
to do this as this is a heterogenous cluster with different brands of
ethernet cards and CPU's.
But it's a good test bed and I have no problems installing Oscar 4.2 on it.
See my later post Hpl and TCP today where I tried 0b1 without mca pml
teg and so on and get a good performance with 15 nodes and open mpi rc6.
Thank you very much,
Regards,
Allan
Message: 1
List-Post: users@lists.open-mpi.org
Date: Mon, 14 Nov 2005 16:10:36 -0500 (Eastern Standard Time)
From: George Bosilca <bosi...@cs.utk.edu>
Subject: Re: [O-MPI users] HPL and TCP
To: Open MPI Users <us...@open-mpi.org>
Message-ID: <Pine.WNT.4.63.0511141603140.3384@bosilca>
Content-Type: TEXT/PLAIN; charset=US-ASCII; format=flowed
Allan,
If there are 2 Ethernet cards it's better if you can point to the one you
want to use. For that you can modify the .openmpi/mca-params.conf file in
your home directory. All of the options can go in this file so you will
not have to specify them on the mpirun command every time.
I give you here a small example that contain the host file (from where
open mpi will pick the nodes) as well as the BTL configuration.
btl_base_include=tcp,sm,self
btl_tcp_if_include=eth0
rds_hostfile_path = /home/bosilca/.openmpi/machinefile
On the first line I specify that Open MPI is allowed to use the TCP,
shared memory and self devices. Self should always be specified otherwise
any communication to the same process will fail (it's out loopback
device).
The second line specify that the TCP BTL is allowed to use only the eth0
interface. This line has to reflect your own configuration.
Finally the 3th one give the full path to the hostfile file.
Thanks,
george.
On Mon, 14 Nov 2005, Allan Menezes wrote:
Dear Jeff, Sorry I could not test the cluster earlier but I am having
problems with one compute node.(I will have to replace it!). So I will have
to repeat this test with 15 nodes. Yes I had 4 NIC cards on the head node and
it was only eth3 that was the gigabit NIC which was communicating to other
eth1 gigabit Nics on the compute nodes through a gigabit switch. So though I
did not specify the ethernet interface in the switch --mca pml teg I was
getting good performance but in --mca btl tcp not specifying the interface
seems to create problems. I wiped out the Linux FC3 installation and tried
again with Oscar 4.2 but am having problems with --mca btl tcp switch. mpirun
--mca btl tcp --prefix /home/allan/openmpi --hostfile aa -np 16 ./xhpl The
hostfile aa contains the 16 hosts a1.lightning.net to a16.lightning.net. So
to recap the cluster is only connected to itself through the giga bit 16 port
switch through gigabit ethernet cards to form a LAN with an IP for each.
There is an extra ethernet card built into the compute motherboards that is
10/100Mbps that is not connected to anything yet. Please can you tell me the
right mpirun command line for btl tcp for my setup? Is the hostfile right?
for the mpirun command above? Should it include a1.lightning.net which is the
head node from where I am invoking mpirun? Or should it not have the head
node? Thank you, Allan Message: 2 Date: Sun, 13 Nov 2005 15:51:30 -0500 From:
Jeff Squyres <jsquy...@open-mpi.org> Subject: Re: [O-MPI users] HPL anf TCP
To: Open MPI Users <us...@open-mpi.org> Message-ID:
<f143e44670c59a2f345708e6e0fad...@open-mpi.org> Content-Type: text/plain;
charset=US-ASCII; format=flowed On Nov 3, 2005, at 8:35 PM, Allan Menezes
wrote:
1. No, I have 4 NICs on the head node and two on each of the 15 other
compute nodes. I use the realtek 8169 gigabit ethernet cards on the
compute nodes as eth1 or eth0(one only) connected to a gigabit ethernet
switch with bisection bandwidth of 32Gbps and a sk98lin driver 3Com built
in gigabit ethernet NIC card on the head node(eth3). The other ethernet
cards 10/100M on the head node handle a network laser printer(eth0) and
eth2 (10/100M) internet access. Eth1 is a spare 10/100M which I can
remove. The compute nodes each have two ethernet cards one 10/100Mbps
ethernet not connected to anything(built in to M/B) and a PCI realtek 8169
gigabit ethernet connected to the TCP network LAN(Gigabit). When I tried
it without the switches -mca pml teg the maximum performace I would get
with it was 9GFlops for P=4 Q=4 N=approx 12- 16 thousand and NB
ridiculously low at 10 block size. If I tried bigger block sizes it would
run for along time for large N ~ 16,000 unless I killed xhpl. I use atlas
BLAS 3.7.11 libs compiled for each node and linked to HPL when creating
xhpl. I also use open mpi mpicc in the hpl make file for compile and link
both. Maybe I should according to the new faq use the TCP switch to use
eth3 on the head node?
So if I'm reading that right, there's only one network that connects the head
node and the compute nodes, right?
That's right!
Allan