Le 05/06/2011 00:15, Fengguang Song a écrit :
> Hi,
>
> I'm confronting a problem when using OpenMPI 1.5.1 on a GPU cluster. My
> program uses MPI to exchange data
> between nodes, and uses cudaMemcpyAsync to exchange data between Host and GPU
> devices within a node.
> When the MPI message size
Hi Brice,
Thank you! I saw your previous discussion and actually have tried "--mca
btl_openib_flags 304".
It didn't solve the problem unfortunately. In our case, the MPI buffer is
different from the cudaMemcpy
buffer and we do manually copy between them. I'm still trying to figure out how
to co