Thanks for the feedback! Sorry the lack of a clear error message caused you pain. :-(

FWIW, we do have this feature in the upcoming v1.3 release. For example:

[6:46] svbu-mpi:~/mpi % mpirun -np 2 --mca btl openibbbb,self hello
--------------------------------------------------------------------------
A requested component was not found, or was unable to be opened.  This
means that this component is either not installed or is unable to be
used on your system (e.g., sometimes this means that shared libraries
that the component requires are unable to be found/loaded).  Note that
Open MPI stopped checking at the first component that it did not find.

Host:      svbu-mpi.cisco.com
Framework: btl
Component: openibbbb
--------------------------------------------------------------------------

Hopefully, this will help such situations in the future.



On Aug 4, 2008, at 9:23 AM, Bernhard Hermann wrote:

Hi,

We would like to request a warning/error message to be printed whenever
a component (like "openib") is specified, but not actually loaded.

Our case: Running a job via "tcp" over IB went OK, trying to use
"openib" failed. It turned out that while compiling the "openib" module wasn't made (missing path), so OpenMPI croaked when trying to access it.

This was the error that was of little help to us in tracing the problem:
--------------------------------------------------------------------------
Process 0.1.7 is unable to reach 0.1.0 for MPI communication.
If you specified the use of a BTL component, you may have
forgotten a component (such as "self") in the list of
usable components.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
It looks like MPI_INIT failed for some reason; your parallel process is
likely to abort.  There are many reasons that a parallel process can
fail during MPI_INIT; some of which are due to configuration or environment
problems.  This failure appears to be an internal failure; here's some
additional information (which may only be relevant to an Open MPI
developer):

 PML add procs failed
 --> Returned "Unreachable" (-12) instead of "Success" (0)
--------------------------------------------------------------------------
*** An error occurred in MPI_Init
*** before MPI was initialized
*** MPI_ERRORS_ARE_FATAL (goodbye)




Unfortunately we can't provide a patch and would be glad if someone
could implement and include this.

Thanks for your attention,
best regards,
Bernhard HERMANN

init.at informationstechnologie GmbH.

Mariahilfer Str. 61/1/10
1060 Wien/Vienna
Austria - EU

Tel: +43 1 522 53 77 - 35

http://www.init.at
herm...@init.at

_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users


--
Jeff Squyres
Cisco Systems

Reply via email to