Mostyn,
There was a problem with the SM BTL. Please upgrade to at least 19315
and [hopefully] your application will run to completion.
Thanks,
george.
On Jul 24, 2008, at 3:39 AM, Mostyn Lewis wrote:
Hello,
Using a very recent svn version (1.4a1r18899) I'm getting a non-
terminati
Dave,
Thanks for your report. As you discovered we had a memory leak in the
MPI_Alltoallw. A very small one, but it was there. Basically, we
didn't release two internal arrays of data-types, used to convert from
the Fortran data-types (as supplied by the user) to their C version
(as requi
Brian,
Some of these files are used for startup, while others are used during
application execution (such as the back-end for shared memory files).
Over the years we had a lot of discussions about this topic, and so
far we have two ways to help people deal with such situations.
However, f
Tom,
I did the same modification as you on the osu_latency and the
resulting application run to completion. I don't get any TRUNCATE
error messages. I'm using the latest version of Open MPI (1.4a1r19313).
There was a bug that might be related to your problem but our commit
log shows it wa
Jitendra,
There is a problem with the addresses you provide to MPI_Type_struct.
For all arrays instead of giving the pointer to the array, you provide
a pointer to the pointer in the individual struct.
Try the following
MPI_Get_address(&parentPop[0].indiv[0].intvar[0], &disp[0]);
instead of
Thanks.
I had also posted the bug on the MPICH2 list, and received an
aswer from the ROMIO maintainers: the issue seems to be related to
NFS file locking bugs. I had been testing on an NFS system, and
when I re-tested under a local (ext3) file system, I did not reproduce
the bug.
I had been exper
Andy Georgi wrote:
Hello again ;),
after getting acceptable latency on our Chelsio S320E-CXA adapters we
now want to check if we can
also tune the bandwidth. On TCP level (measured via iperf) we get 1.15
GB/s, on MPI level (measured
via MPI-Ping-Pong) just 930 MB/s. We already set btl_tcp_sndb
With OpenMPI 1.3 / iWARP you should get around 8us latency using mpi
pingpong tests.
Andy Georgi wrote:
Thanks again for all the answers. It seems that were was a bug in the
driver in combination with
Suse Linux Enterprise Server 10. It was fixed with version 1.0.146.
Now we have 12us with NP
Hello again ;),
after getting acceptable latency on our Chelsio S320E-CXA adapters we now want
to check if we can
also tune the bandwidth. On TCP level (measured via iperf) we get 1.15 GB/s, on
MPI level (measured
via MPI-Ping-Pong) just 930 MB/s. We already set btl_tcp_sndbuf and
btl_tcp_rcvb