Re: [OMPI users] Continuous poll/select using btl sm (svn 1.4a1r18899)

2008-08-17 Thread George Bosilca
Mostyn, There was a problem with the SM BTL. Please upgrade to at least 19315 and [hopefully] your application will run to completion. Thanks, george. On Jul 24, 2008, at 3:39 AM, Mostyn Lewis wrote: Hello, Using a very recent svn version (1.4a1r18899) I'm getting a non- terminati

Re: [OMPI users] memory leak in alltoallw

2008-08-17 Thread George Bosilca
Dave, Thanks for your report. As you discovered we had a memory leak in the MPI_Alltoallw. A very small one, but it was there. Basically, we didn't release two internal arrays of data-types, used to convert from the Fortran data-types (as supplied by the user) to their C version (as requi

Re: [OMPI users] Q: OpenMPI's use of /tmp and hanging apps via FS problems?

2008-08-17 Thread George Bosilca
Brian, Some of these files are used for startup, while others are used during application execution (such as the back-end for shared memory files). Over the years we had a lot of discussions about this topic, and so far we have two ways to help people deal with such situations. However, f

Re: [OMPI users] MPI_ERR_TRUNCATE with MPI_Revc without Infinipath

2008-08-17 Thread George Bosilca
Tom, I did the same modification as you on the osu_latency and the resulting application run to completion. I don't get any TRUNCATE error messages. I'm using the latest version of Open MPI (1.4a1r19313). There was a bug that might be related to your problem but our commit log shows it wa

Re: [OMPI users] MPI_Type_struct for structs with dynamic arrays

2008-08-17 Thread George Bosilca
Jitendra, There is a problem with the addresses you provide to MPI_Type_struct. For all arrays instead of giving the pointer to the array, you provide a pointer to the pointer in the individual struct. Try the following MPI_Get_address(&parentPop[0].indiv[0].intvar[0], &disp[0]); instead of

Re: [OMPI users] bug in MPI_File_get_position_shared ?

2008-08-17 Thread Yvan Fournier
Thanks. I had also posted the bug on the MPICH2 list, and received an aswer from the ROMIO maintainers: the issue seems to be related to NFS file locking bugs. I had been testing on an NFS system, and when I re-tested under a local (ext3) file system, I did not reproduce the bug. I had been exper

Re: [OMPI users] TCP Bandwidth

2008-08-17 Thread Steve Wise
Andy Georgi wrote: Hello again ;), after getting acceptable latency on our Chelsio S320E-CXA adapters we now want to check if we can also tune the bandwidth. On TCP level (measured via iperf) we get 1.15 GB/s, on MPI level (measured via MPI-Ping-Pong) just 930 MB/s. We already set btl_tcp_sndb

Re: [OMPI users] TCP Latency

2008-08-17 Thread Steve Wise
With OpenMPI 1.3 / iWARP you should get around 8us latency using mpi pingpong tests. Andy Georgi wrote: Thanks again for all the answers. It seems that were was a bug in the driver in combination with Suse Linux Enterprise Server 10. It was fixed with version 1.0.146. Now we have 12us with NP

[OMPI users] TCP Bandwidth

2008-08-17 Thread Andy Georgi
Hello again ;), after getting acceptable latency on our Chelsio S320E-CXA adapters we now want to check if we can also tune the bandwidth. On TCP level (measured via iperf) we get 1.15 GB/s, on MPI level (measured via MPI-Ping-Pong) just 930 MB/s. We already set btl_tcp_sndbuf and btl_tcp_rcvb