Re: [OMPI users] Segmentation fault when Send/Recv, onheterogeneouscluster (32/64 bit machines)

2010-03-08 Thread Terry Dontje
We (Oracle) have not done that much extensive limits testing going between 32 to 64bit applications. Most of the testing we've done is more around endianess (SPARC vs x86_64). Though the below is kind of interesting. Sounds like the eager limit isn't being normalized on the 64 bit machines.

Re: [OMPI users] Why might MPI_Recv trip PSM_MQ_RECVREQS_MAX ?

2010-03-08 Thread Rainer Keller
Hello Jonathan, Your are using Infinipath's PSM library and the corresponding MTL/psm and therefore the corresponding upper-layer PML/cm. In fact, this _is_ calling into the psm's irecv() function, which explains the error triggered in the psm library. Not knowing the degree of parallelism of Y

Re: [OMPI users] Segmentation fault when Send/Recv, onheterogeneouscluster (32/64 bit machines)

2010-03-08 Thread TRINH Minh Hieu
Hello, I changed the test code (hetero.c, in attach) so that the master (where data is centralized) can be rank 1 or 2. I tested with a master,rank 2 or rank 1 : same probleme, when the master is a 64bit machine, as soon as it receive data from a 32bit machines it got segfault. no probleme with a

Re: [OMPI users] Why might MPI_Recv trip PSM_MQ_RECVREQS_MAX ?

2010-03-08 Thread Richard Treumann
The program Jonathan offers as an example is valid use of MPI standard send. With this message size it is fair to assume the implementation is doing standard send with an eager send. The MPI standard is explicit about how eager send, as a undercover option for standard send, must work. When the

Re: [OMPI users] low efficiency when we use --am ft-enable-cr tocheckpoint

2010-03-08 Thread 马少杰
Thank you so quick response, I will try it 2010-03-09 马少杰 发件人: Joshua Hursey 发送时间: 2010-03-05 22:10:52 收件人: Open MPI Users 抄送: 主题: Re: [OMPI users] low efficiency when we use --am ft-enable-cr tocheckpoint On Mar 5, 2010, at 3:15 AM, 马少杰 wrote: > Dear Sir: > - What version of Op