Hi, I set up a Linux Cluster with differnt Distributions ( 1x Debian Lenny, 4x OpenSuse11.2 ) and openmpi-1.4.1 , all my test applications ran perfekt.
Now I decided to create a Debian-Live System (Lenny) with openmpi-1.4.1, to include some more Pc's in our Student-Pool, and always get the folowing errors: #: mpirun --hostfile my_hostfile -np 4 hello_c Hello, world, I am 2 of 4 Dell-19 (256) Hello, world, I am 0 of 4 Dell-19 (256) Hello, world, I am 3 of 4 Dlive (256) Hello, world, I am 1 of 4 Dlive (256) [Dell-19:9199] *** An error occurred in MPI_Barrier [Dell-19:9199] *** on communicator MPI_COMM_WORLD [Dell-19:9199] *** MPI_ERR_IN_STATUS: error code in status [Dell-19:9199] *** MPI_ERRORS_ARE_FATAL (your MPI job will now abort) -------------------------------------------------------------------------- mpirun has exited due to process rank 2 with PID 9199 on node Dell-19 exiting without calling "finalize". This may have caused other processes in the application to be terminated by signals sent by mpirun (as reported here). -------------------------------------------------------------------------- [Dell-19:09196] 1 more process has sent help message help-mpi-errors.txt / mpi_errors_are_fatal [Dell-19:09196] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages I've got no more idea how to fix this. Thanks in advance horst. -- Horst Piening IT-Administrator Fb15 Universitaet Hamburg Bundesstr. 55 20146 Hamburg Email: pien...@geowiss.uni-hamburg.de Tel.: (040) 42838-7015