Re: [OMPI users] Open MPI 1.2.3 spec file

2007-07-10 Thread Alex Tumanov
On 7/9/07, Jeff Squyres wrote: On Jul 6, 2007, at 12:05 PM, Alex Tumanov wrote: > Eureka! I managed to get it working despite the incorrect _initial_ > ./configure invocation. For those interested, here are my compilation > options: > # cat ompi_build.sh > #!/bin/sh > > r

Re: [OMPI users] Open MPI 1.2.3 spec file

2007-07-06 Thread Alex Tumanov
uot;_defaultdocdir %{_prefix}/share" \ -D "mflags all" openmpi-1.2.3-1.src.rpm Thanks for looking :) Alex. On 7/5/07, Alex Tumanov wrote: Actually, tried compiling the RPM again, and at the very top, noticed that the ./configure is called with --sysconfdir set to /opt/o

Re: [OMPI users] Open MPI 1.2.3 spec file

2007-07-05 Thread Alex Tumanov
Actually, tried compiling the RPM again, and at the very top, noticed that the ./configure is called with --sysconfdir set to /opt/openmpi instead of the new name provided. All other parameters are correct! Any ideas? ./configure --build=x86_64-redhat-linux-gnu --host=x86_64-redhat-linux-gnu --ta

[OMPI users] Open MPI 1.2.3 spec file

2007-07-05 Thread Alex Tumanov
Greetings, The spec file provided in the latest stable src RPM makes it possible to change the name of the resulting RPM. I tried to make use of that, but ran into some issues. Specifically, the resulting RPM does not have the etc directory (and sample config files in it). rpmbuild complained abo

Re: [OMPI users] mixing MX and TCP

2007-06-09 Thread Alex Tumanov
forgive a trivial question, but what's a multi-rail? On 6/8/07, George Bosilca wrote: A fix for this problem is now available on the trunk. Please use any revision after 14963 and your problem will vanish [I hope!]. There are now some additional parameters which allow you to select which Myrine

[OMPI users] where did liborte/libopal go?

2007-05-15 Thread Alex Tumanov
Greetings, Very quick question: I noticed that liborte and libopal shared libraries are no longer present in the 1.2 version of Open MPI. Can I establish backward compatibility symlinks for applications that were compiled with Open MPI 1.1.4 and were linked with liborte and libopal? If so , which

Re: [OMPI users] NetPipe benchmark & spanning multiple interconnects

2007-02-14 Thread Alex Tumanov
nies, which probably do not have any vested interest in making their interconnects synergistically coexist with the ones of their competitors or with interconnects the companies are dropping support for. Many thanks, Alex. On Feb 12, 2007, at 6:48 PM, Alex Tumanov wrote: > Anyone else

[OMPI users] problems with HPLinpack over myrinet MX-10G

2007-02-14 Thread Alex Tumanov
Hello, I recently tried running HPLinpack, compiled with OMPI, over myrinet MX interconnect. Running a simple hello world program works, but XHPL fails with an error occurring when it tries to MPI_Send: # mpirun -np 4 -H l0-0,c0-2 --prefix $MPIHOME --mca btl mx,self /opt/hpl/openmpi-hpl/bin/xhpl

Re: [OMPI users] NetPipe benchmark & spanning multiple interconnects

2007-02-12 Thread Alex Tumanov
. On 2/8/07, Alex Tumanov wrote: Thanks for your insight George. > Strange, the latency is supposed to be there too. Anyway, the latency > is only used to determine which one is faster, in order to use it for > small messages. I searched the code base for mca parameter registering

Re: [OMPI users] first time user - can run mpi job SMP but not over cluster

2007-02-08 Thread Alex Tumanov
mpirun --prefix /opt/openmpi -mca oob_tcp_include eth0 -mca btl_tcp_if_include eth0 --hostfile ~/work/openmpi_hostfile -np 4 hostname Could a section be added to the FAQ mentioning that the firewall service should be shutdown over the mpi interface and that the two -mca switches should be used?

Re: [OMPI users] first time user - can run mpi job SMP but not over cluster

2007-02-08 Thread Alex Tumanov
I have added the following line to my .bashrc: export OMPIFLAGS="-mca oob_tcp_include eth0 -mca btl_tcp_if_include eth0 --hostfile ~/work/openmpi_hostfile" and have verified that mpirun $OMPIFLAGS -np 4 hostname works. Is there a better way of accomplishing this, or is this a matter of there be

Re: [OMPI users] Open MPI/OpenIB Error/Problem

2007-02-08 Thread Alex Tumanov
Hello Troy, This issue is fairly common and has to do with the maximum amount of memory allowed to be allocated. See this FAQ for more detail: http://www.open-mpi.org/faq/?category=openfabrics#ib-locked-pages I had the same issue and the FAQ resolves it. Good luck, Alex. On 2/8/07, Troy Telfor

Re: [OMPI users] NetPipe benchmark & spanning multiple interconnects

2007-02-08 Thread Alex Tumanov
gt; btl_tcp_if_include = eth0,eth1 >> btl_tcp_max_rdma_size = 524288 >> >> btl_tcp_latency_eth0 = 47 >> btl_tcp_bandwidth_eth0 = 587 >> >> btl_tcp_latency_eth1 = 51 >> btl_tcp_bandwidth_eth1 = 233 >> >> Something similar has to be done for op

Re: [OMPI users] NetPipe benchmark & spanning multiple interconnects

2007-02-08 Thread Alex Tumanov
p_latency_eth0 = 47 btl_tcp_bandwidth_eth0 = 587 btl_tcp_latency_eth1 = 51 btl_tcp_bandwidth_eth1 = 233 Something similar has to be done for openib and gm, in order to allow us to strip the messages correctly. Thanks, george. On Feb 8, 2007, at 12:02 PM, Alex Tumanov wrote: > Hel

[OMPI users] NetPipe benchmark & spanning multiple interconnects

2007-02-08 Thread Alex Tumanov
Hello Jeff. Thanks for pointing out NetPipe to me. I've played around with it a little in hope to see clear evidence/effect of message striping in OpenMPI. Unfortunately, what I saw is that the result of running NPmpi over several interconnects is identical to running it over a single fastest one

Re: [OMPI users] first time user - can run mpi job SMP but not over cluster

2007-02-07 Thread Alex Tumanov
Hello, mpirun -np 2 myprogram inputfile >outputfile There can be a whole host of issues with the way you run your executable and/or the way you have the environment setup. First of all, when you ssh into the node, does the environment automatically get updated with correct Open MPI paths? I.e.

Re: [OMPI users] [OMPI Users] OpenMPI 1.1.4 over ethernet fails

2007-02-06 Thread Alex Tumanov
Thanks for your reply, Jeff. > It never occurred to me that the headnode would try to communicate > with the slave using infiniband interfaces... Orthogonally, what are The problem here is that since your IB IP addresses are "public" (meaning that they're not in the IETF defined ranges for priv

Re: [OMPI users] running OpenMPI jobs over Myrinet gm interconnect

2007-02-06 Thread Alex Tumanov
Thank you for your reply, Reese! What version of GM are you running? # rpm -qa |egrep "^gm-[0-9]+|^gm-devel" gm-2.0.24-1 gm-devel-2.0.24-1 Is this too old? And are you sure that gm_board_info shows all the nodes that are listed in your machine file? Yes, that was the issue - bad cable connec

[OMPI users] running OpenMPI jobs over Myrinet gm interconnect

2007-02-05 Thread Alex Tumanov
Hello list, I'm having difficulty with running a simple hello world OpenMPI program over Myrinet gm interconnect - please see the log at the end of this email. The error is tripped by a call to the function gm_global_id_to_node_id( gm_btl->port, gm_endpoint->endpoint_addr.global_id,

Re: [OMPI users] [OMPI Users] OpenMPI 1.1.4 over ethernet fails

2007-02-02 Thread Alex Tumanov
That really did fix it, George: # mpirun --prefix $MPIHOME -hostfile ~/testdir/hosts --mca btl tcp,self --mca btl_tcp_if_exclude ib0,ib1 ~/testdir/hello Hello from Alex' MPI test program Process 0 on dr11.lsf.platform.com out of 2 Hello from Alex' MPI test program Process 1 on compute-0-0.local o

Re: [OMPI users] [OMPI Users] OpenMPI 1.1.4 over ethernet fails

2007-02-01 Thread Alex Tumanov
nodes? - Galen On Feb 1, 2007, at 2:50 PM, Alex Tumanov wrote: > Hi, > > I have kept doing my own investigation and recompiled OpenMPI to have > only the barebones functionality with no support for any interconnects > other than ethernet: > # rpmbuild --rebuild --define="

Re: [OMPI users] [OMPI Users] OpenMPI 1.1.4 over ethernet fails

2007-02-01 Thread Alex Tumanov
replicate the error, I ran # mpirun -hostfile ~/testdir/hosts --mca btl tcp,self ~/testdir/hello In this case, you can even omit the runtime mca param specifications: # mpirun -hostfile ~/testdir/hosts ~/testdir/hello Thanks for reading this. I hope I've provided enough information. Sincerel

[OMPI users] [OMPI Users] OpenMPI 1.1.4 over ethernet fails

2007-02-01 Thread Alex Tumanov
Hello, I have tried a very basic test on a 2 node "cluster" consisting of 2 dell boxes. One of them is dual CPU Intel(R) Xeon(TM) CPU 2.80GHz with 1GB of RAM and the slave node is quad-CPU Intel(R) Xeon(TM) CPU 3.40GHz with 2GB of RAM. Both have Infiniband cards and Gig-E. The slave node is conne