On Jun 28, 2011, at 1:46 PM, Bill Johnstone wrote: > I have a heterogeneous network of InfiniBand-equipped hosts which are all > connected to the same backbone switch, an older SDR 10 Gb/s unit. > > One set of nodes uses the Mellanox "ib_mthca" driver, while the other uses > the "mlx4" driver. > > This is on Linux 2.6.32, with Open MPI 1.5.3 . > > When I run Open MPI across these node types, I get an error message of the > form: > > Open MPI detected two different OpenFabrics transport types in the same > Infiniband network. > Such mixed network trasport configuration is not supported by Open MPI. > > Local host: compute-chassis-1-node-01 > Local adapter: mthca0 (vendor 0x5ad, part ID 25208) > Local transport type: MCA_BTL_OPENIB_TRANSPORT_UNKNOWN
Wow, that's cool ("UNKNOWN"). Are you using an old version of OFED or something? Mellanox -- how can this happen? > Remote host: compute-chassis-3-node-01 > Remote Adapter: (vendor 0x2c9, part ID 26428) > Remote transport type: MCA_BTL_OPENIB_TRANSPORT_IB > > Two questions: > > 1. Why is this occurring if both adapters have all the OpenIB software set > up? Is it because Open MPI is trying to use functionality such as ConnectX > with the newer hardware, which is incompatible with older hardware, or is it > something more mundane? It's basically a mismatch of IB capabilities -- Open MPI is trying to use more advanced features in some nodes and not in others. > 2. How can I use IB amongst these heterogeneous nodes? Mellanox will need to answer this question... It might be able to be done, but I don't know how offhand. The first issue is to figure out why you're getting TRANSPORT_UNKNOWN on the one node. -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/