Arthur,
Your email does not contain enough information to pinpoint the problem.
However, there are several hints that tent to indicate a problem in your
application.
1. in the collective communication that succeed, the MPI_Intercomm_merge,
the processes are doing [at least] one MPI_Allreduce foll
Artur,
do you check all the error codes returned by MPI_Comm_spawn_multiple ?
(so you can confirm the requested number of tasks was spawned)
since the error occurs only on the first MPI_Send, you might want to
retrieve rank and size and print them right before MPI_Send, just to make
sure the comm
Artur,
in OpenMPI, MPI_Comm is an opaque pointer, so strictly speaking, high value
might not be an issue.
can you have your failed processes generate a core and post the stack trace
?
btw, do you MPI_Send on the intra communicator created by
MPI_Intercomm_merge ?
what is the minimal config neede
Hi,
I have a problem with my application that is based on dynamic process
management. The scenario related to process creation is as follows:
1. All processes call MPI_Comm_spawn_multiple to spawn additional
single process per each node.
2. Parent processes call MPI_Intercomm_merge.
3. C