Marcelo -- Many thanks for this report! I have duplicated your problem (doh!), and have opened bug #232 about it.
https://svn.open-mpi.org/trac/ompi/ticket/232 On 7/29/06 5:35 PM, "Marcelo Stival" <marsti...@gmail.com> wrote: > Here (attached) is a programm that reproduce the error reported... > > Now I updated trunk , but got the same... (output attached again) > > I also tested with 1.1 stable release... > With 1.1 the program blocks without any error output... > > The program try to send 512K messages of 8bytes. > If reduced to 1k messages of 8bytes, it works fine... > > > Thanks > Marcelo > > > On 7/28/06, Jeff Squyres <jsquy...@cisco.com> wrote: >> >> Marcelo -- >> >> Can you send your code that is failing? I'm unable to reproduce with some >> toy programs here. >> >> I also notice that you're running a somewhat old version of and OMPI SVN >> checkout of the trunk. Can you update to the most recent version? The >> trunk is not guaranteed to be stable, and we did have some stability >> problems recently -- you might want to upgrade to the most recent version >> (today seems to be ok) and/or try one of the nightly or prerelease >> tarballs >> in the 1.1 branch. >> >> >> On 7/26/06 6:18 PM, "Marcelo Stival" <marsti...@gmail.com> wrote: >> >>> Hi, >>> >>> I got a problem with ompi when sending large number of messages from >>> process A to process B. >>> Process A only send... and B only receive (the buffers are reused) >>> >>> int n = 4 * 1024;//number of iterations (messages to be sent) >> consecutively >>> int len = 8; //len of each message >>> >>> Process A (rank 0): >>> for (i=0; i < n; i++){ >>> MPI_Send( sbuffer, len, MPI_BYTE,to,i,MPI_COMM_WORLD); >>> } >>> Process B (rank 1): >>> for (i=0; i < n; i++){ >>> MPI_Recv(rbuffer,len,MPI_BYTE,recv_from , i,MPI_COMM_WORLD, >> &status); >>> } >>> (It's a benchmark program... will run with increasing messages sizes.. ) >>> (I tried with the same tag on all iterations - and got the same) >>> >>> It works fine for n (number of messages) equals to 3k (for example), but >> do >>> not work with n equals to 4k (for messages of 8 bytes 4k iterations >> seems to >>> be the treshould). >>> >>> The error messages (complete output attached): >>> malloc debug: Request for 8396964 bytes failed (class/ompi_free_list.c, >> 142) >>> mpptest: btl_tcp_endpoint.c:624: mca_btl_tcp_endpoint_recv_handler: >>> Assertion `0 >>> == btl_endpoint->endpoint_cache_length' failed. >>> Signal:6 info.si_errno:0(Success) si_code:-6() >>> >>> >>> Considerations: >>> It works for synchronous send (MPI_Ssend). >>> It works with MPICH2 ( 1.0.3). >>> It is a benchmark program, I want to flood the network to measure the >>> bandwidth ... (for different message sizes) >>> >>> >>> Thanks >>> >>> Marcelo >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> -- >> Jeff Squyres >> Server Virtualization Business Unit >> Cisco Systems >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres Server Virtualization Business Unit Cisco Systems