Re: [OMPI users] System hang-up on MPI_Reduce

2009-11-11 Thread Ralph Castain
You are welcome to stick barriers in - doesn't hurt anything other than performance. On Nov 11, 2009, at 3:00 AM, Glembek Ondřej wrote: Thanx for your reply... My coll_sync_priority is set to 50. See the dump of ompi_info -- param coll sync below... Does sticking barriers hurt anything or

Re: [OMPI users] System hang-up on MPI_Reduce

2009-11-11 Thread Glembek Ondřej
Thanx for your reply... My coll_sync_priority is set to 50. See the dump of ompi_info --param coll sync below... Does sticking barriers hurt anything or is it just a cosmetic thing??? I'm fine with this solution... Thanx Ondrej $ompi_info --param coll sync MCA coll: para

Re: [OMPI users] System hang-up on MPI_Reduce

2009-11-10 Thread Ralph Castain
Yeah, that is "normal". It has to do with unexpected messages. When you have procs running at significantly different speeds, the various operations get far enough out of sync that the memory consumed by recvd messages not yet processed grows too large. Instead of sticking barriers into you

[OMPI users] System hang-up on MPI_Reduce

2009-11-10 Thread Glembek Ondřej
Hi, I am using MPI_Reduce operation on 122880x400 matrix of doubles. The parallel job runs on 32 machines, each having different processor in terms of speed, but the architecture and OS is the same on all machines (x86_64). The task is a typical map-and-reduce, i.e. each of the processes