Hello Brad, 

I removed the openib specifier from the btl lists in the appfile and the
application ran fine over ethernet.  And yes, to confirm, if I attempt
to mix systems with IB and systems without IB, the application
hangs.    

Thanks, 

Ryan



Hello Ryan, 


I have been running a similar heterogeneous setup in my lab; i.e., a mix
of 
ppc64 and x86_64 systems connected by ethernet and InfiniBand. In trying
to 
replicate your problem, what I see is that it is not an issue of
processor 
heterogeneity, but rather an issue with heterogeneous transports. Can
you 
remove the openib specifier from the btl lists in the appfile and try 
again? I.e., force all inter-system communications over ethernet? For
me, 
that works. But, if I mix systems with IB with systems without IB, I,
too, 
see a hang...even if the processor architectures are the same. If you
could 
confirm that your case is the same, then we can make sure we're only
chasing 
one problem and not two. 


Thanks, 
--Brad 


Brad Benton 
IBM 


On Thu, May 1, 2008 at 11:02 AM, Ryan Buckley ; 21426
<rbuckley_at_[hidden]> 
wrote: 


> Hello,
>
> I am trying to run a simple Hello World MPI application in a
> heterogeneous environment. The machines include 1 x86 machine with a
> standard 1Gb ethernet connection and 2 ppc machines with standard 1Gb
> ethernet as well as a 10Gb ethernet (Infiniband) switch between the
two.
> The Hello World program is the same hello_c.c that is included in the
> examples directory of the Open MPI installation.
>
> The goal is that I would like to run heterogeneous applications
between
> the three aforementioned machines in the following manner:
>
> The x86 machine will use tcp to communicate to the 2 ppc machines,
> while the ppc machines will communicate with one another via the
10GbE.
>
> x86 <--tcp--> ppc_1
> x86 <--tcp--> ppc_2
> ppc1 <--openib--> ppc_2
>
> I am currently using a machfile set up as follows,
>
> # cat machfile
> <ppc_host_1>
> <ppc_host_2>
> <x86_host>
>
> In addition I am using an appfile set up as follows,
>
> # cat appfile
> -np 1 --hostfile machfile --host <ppc_host_1> --mca btl
> sm,self,tcp,openib /path/to/ppc/openmpi-1.2.5/examples/hello
> -np 1 --hostfile machfile --host <ppc_host_2> --mca btl
> sm,self,tcp,openib /path/to/ppc/openmpi-1.2.5/examples/hello
> -np 1 --hostfile machfile --host <x86_host> --mca btl
> sm,self,tcp /path/to/x86/openmpi-1.2.5/examples/hello
>
> I am running on the command line via
>
> # mpirun --app appfile
>
> I've also attached the output from 'ompi_info --all' from all
machines.
>
> Any suggestions would be much appreciated.
>
> Thanks,
>
> Ryan
>
>
> _______________________________________________
> users mailing list
> users_at_[hidden]
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to